[dpdk-dev] [PATCH 40/56] net/sfc: maintain management event queue

Andrew Rybchenko arybchenko at solarflare.com
Mon Nov 21 16:00:54 CET 2016


The event queue is required for device level events (e.g. link status
change) and flush events.
Provide thread-safe function to poll the event queue since it may be
really done from different contexts.

Reviewed-by: Andy Moreton <amoreton at solarflare.com>
Signed-off-by: Andrew Rybchenko <arybchenko at solarflare.com>
---
 drivers/net/sfc/efx/sfc.h    |  1 +
 drivers/net/sfc/efx/sfc_ev.c | 50 ++++++++++++++++++++++++++++++++++++++++----
 drivers/net/sfc/efx/sfc_ev.h |  2 ++
 3 files changed, 49 insertions(+), 4 deletions(-)

diff --git a/drivers/net/sfc/efx/sfc.h b/drivers/net/sfc/efx/sfc.h
index eb8c071..39f15b1 100644
--- a/drivers/net/sfc/efx/sfc.h
+++ b/drivers/net/sfc/efx/sfc.h
@@ -144,6 +144,7 @@ struct sfc_adapter {
 	struct sfc_evq_info		*evq_info;
 
 	unsigned int			mgmt_evq_index;
+	rte_spinlock_t			mgmt_evq_lock;
 };
 
 /*
diff --git a/drivers/net/sfc/efx/sfc_ev.c b/drivers/net/sfc/efx/sfc_ev.c
index aa04b34..1734b1e 100644
--- a/drivers/net/sfc/efx/sfc_ev.c
+++ b/drivers/net/sfc/efx/sfc_ev.c
@@ -213,6 +213,19 @@ sfc_ev_qpoll(struct sfc_evq *evq)
 	/* Poll-mode driver does not re-prime the event queue for interrupts */
 }
 
+void
+sfc_ev_mgmt_qpoll(struct sfc_adapter *sa)
+{
+	if (rte_spinlock_trylock(&sa->mgmt_evq_lock)) {
+		struct sfc_evq *mgmt_evq = sa->evq_info[sa->mgmt_evq_index].evq;
+
+		if (mgmt_evq->init_state == SFC_EVQ_STARTED)
+			sfc_ev_qpoll(mgmt_evq);
+
+		rte_spinlock_unlock(&sa->mgmt_evq_lock);
+	}
+}
+
 int
 sfc_ev_qprime(struct sfc_evq *evq)
 {
@@ -324,13 +337,26 @@ sfc_ev_start(struct sfc_adapter *sa)
 	if (rc != 0)
 		goto fail_ev_init;
 
+	/* Start management EVQ used for global events */
+	rte_spinlock_lock(&sa->mgmt_evq_lock);
+
+	rc = sfc_ev_qstart(sa, sa->mgmt_evq_index);
+	if (rc != 0)
+		goto fail_mgmt_evq_start;
+
+	rte_spinlock_unlock(&sa->mgmt_evq_lock);
+
 	/*
-	 * Rx/Tx event queues are started/stopped when corresponding queue
-	 * is started/stopped.
+	 * Rx/Tx event queues are started/stopped when corresponding
+	 * Rx/Tx queue is started/stopped.
 	 */
 
 	return 0;
 
+fail_mgmt_evq_start:
+	rte_spinlock_unlock(&sa->mgmt_evq_lock);
+	efx_ev_fini(sa->nic);
+
 fail_ev_init:
 	sfc_log_init(sa, "failed %d", rc);
 	return rc;
@@ -345,8 +371,17 @@ sfc_ev_stop(struct sfc_adapter *sa)
 
 	/* Make sure that all event queues are stopped */
 	sw_index = sa->evq_count;
-	while (--sw_index >= 0)
-		sfc_ev_qstop(sa, sw_index);
+	while (--sw_index >= 0) {
+		if (sw_index == sa->mgmt_evq_index) {
+			/* Locks are required for the management EVQ */
+			rte_spinlock_lock(&sa->mgmt_evq_lock);
+			sfc_ev_qstop(sa, sa->mgmt_evq_index);
+			rte_spinlock_unlock(&sa->mgmt_evq_lock);
+		} else {
+			sfc_ev_qstop(sa, sw_index);
+		}
+	}
+
 
 	efx_ev_fini(sa->nic);
 }
@@ -444,6 +479,7 @@ sfc_ev_init(struct sfc_adapter *sa)
 
 	sa->evq_count = sfc_ev_qcount(sa);
 	sa->mgmt_evq_index = 0;
+	rte_spinlock_init(&sa->mgmt_evq_lock);
 
 	/* Allocate EVQ info array */
 	rc = ENOMEM;
@@ -459,6 +495,11 @@ sfc_ev_init(struct sfc_adapter *sa)
 			goto fail_ev_qinit_info;
 	}
 
+	rc = sfc_ev_qinit(sa, sa->mgmt_evq_index, SFC_MGMT_EVQ_ENTRIES,
+			  sa->socket_id);
+	if (rc != 0)
+		goto fail_mgmt_evq_init;
+
 	/*
 	 * Rx/Tx event queues are created/destroyed when corresponding
 	 * Rx/Tx queue is created/destroyed.
@@ -466,6 +507,7 @@ sfc_ev_init(struct sfc_adapter *sa)
 
 	return 0;
 
+fail_mgmt_evq_init:
 fail_ev_qinit_info:
 	while (sw_index-- > 0)
 		sfc_ev_qfini_info(sa, sw_index);
diff --git a/drivers/net/sfc/efx/sfc_ev.h b/drivers/net/sfc/efx/sfc_ev.h
index 140a436..f7bcf01 100644
--- a/drivers/net/sfc/efx/sfc_ev.h
+++ b/drivers/net/sfc/efx/sfc_ev.h
@@ -132,6 +132,8 @@ void sfc_ev_qstop(struct sfc_adapter *sa, unsigned int sw_index);
 int sfc_ev_qprime(struct sfc_evq *evq);
 void sfc_ev_qpoll(struct sfc_evq *evq);
 
+void sfc_ev_mgmt_qpoll(struct sfc_adapter *sa);
+
 #ifdef __cplusplus
 }
 #endif
-- 
2.5.5



More information about the dev mailing list