[dpdk-dev] [PATCH] event/sw: remove stale IQ references when reconfigured

Pavan Nikhilesh pbhagavatula at caviumnetworks.com
Wed Jan 10 11:26:16 CET 2018


On Tue, Jan 09, 2018 at 10:19:35AM -0600, Gage Eads wrote:
> This commit fixes a bug in which, when the sw PMD is reconfigured, it would
> leave stale IQ chunk pointers in each queue's IQ structure. Now, the PMD
> initializes all IQs at eventdev start time and releases all IQ chunk
> pointers at eventdev stop time (which has the consequence that any events
> in a queue when the eventdev is stopped will be lost). This approach should
> be resilient to any reconfiguration done between the stop and start, such
> as adding or removing queues.
>
> This commit also fixes two potential issues in iq_chunk.h. iq_init()
> now initializes the IQ's count field to 0, and iq_dequeue_burst() sets
> iq->head to the appropriate next pointer.
>
> Fixes: 5b5e476e59a4 ("event/sw: use dynamically-sized IQs")
> Signed-off-by: Gage Eads <gage.eads at intel.com>
> ---
>  drivers/event/sw/iq_chunk.h | 14 +++++++++-
>  drivers/event/sw/sw_evdev.c | 62 ++++++++++++++++++++++++++++++++-------------
>  2 files changed, 57 insertions(+), 19 deletions(-)
>
> diff --git a/drivers/event/sw/iq_chunk.h b/drivers/event/sw/iq_chunk.h
> index 29f5a35..2fa724c 100644
> --- a/drivers/event/sw/iq_chunk.h
> +++ b/drivers/event/sw/iq_chunk.h
> @@ -73,12 +73,24 @@ iq_free_chunk(struct sw_evdev *sw, struct sw_queue_chunk *chunk)
>  }
>
>  static __rte_always_inline void
> +iq_free_chunk_list(struct sw_evdev *sw, struct sw_queue_chunk *head)
> +{
> +	while (head) {
> +		struct sw_queue_chunk *next;
> +		next = head->next;
> +		iq_free_chunk(sw, head);
> +		head = next;
> +	}
> +}
> +
> +static __rte_always_inline void
>  iq_init(struct sw_evdev *sw, struct sw_iq *iq)
>  {
>  	iq->head = iq_alloc_chunk(sw);
>  	iq->tail = iq->head;
>  	iq->head_idx = 0;
>  	iq->tail_idx = 0;
> +	iq->count = 0;
>  }
>
>  static __rte_always_inline void
> @@ -154,7 +166,7 @@ iq_dequeue_burst(struct sw_evdev *sw,
>
>  done:
>  	if (unlikely(index == SW_EVS_PER_Q_CHUNK)) {
> -		struct sw_queue_chunk *next = iq->head->next;
> +		struct sw_queue_chunk *next = current->next;
>  		iq_free_chunk(sw, current);
>  		iq->head = next;
>  		iq->head_idx = 0;
> diff --git a/drivers/event/sw/sw_evdev.c b/drivers/event/sw/sw_evdev.c
> index 1ef6340..7430a5d 100644
> --- a/drivers/event/sw/sw_evdev.c
> +++ b/drivers/event/sw/sw_evdev.c
> @@ -245,9 +245,6 @@ qid_init(struct sw_evdev *sw, unsigned int idx, int type,
>  	char buf[IQ_ROB_NAMESIZE];
>  	struct sw_qid *qid = &sw->qids[idx];
>
> -	for (i = 0; i < SW_IQS_MAX; i++)
> -		iq_init(sw, &qid->iq[i]);
> -
>  	/* Initialize the FID structures to no pinning (-1), and zero packets */
>  	const struct sw_fid_t fid = {.cq = -1, .pcount = 0};
>  	for (i = 0; i < RTE_DIM(qid->fids); i++)
> @@ -325,11 +322,6 @@ qid_init(struct sw_evdev *sw, unsigned int idx, int type,
>  	return 0;
>
>  cleanup:
> -	for (i = 0; i < SW_IQS_MAX; i++) {
> -		if (qid->iq[i].head)
> -			iq_free_chunk(sw, qid->iq[i].head);
> -	}
> -
>  	if (qid->reorder_buffer) {
>  		rte_free(qid->reorder_buffer);
>  		qid->reorder_buffer = NULL;
> @@ -348,13 +340,6 @@ sw_queue_release(struct rte_eventdev *dev, uint8_t id)
>  {
>  	struct sw_evdev *sw = sw_pmd_priv(dev);
>  	struct sw_qid *qid = &sw->qids[id];
> -	uint32_t i;
> -
> -	for (i = 0; i < SW_IQS_MAX; i++) {
> -		if (!qid->iq[i].head)
> -			continue;
> -		iq_free_chunk(sw, qid->iq[i].head);
> -	}
>
>  	if (qid->type == RTE_SCHED_TYPE_ORDERED) {
>  		rte_free(qid->reorder_buffer);
> @@ -388,6 +373,41 @@ sw_queue_setup(struct rte_eventdev *dev, uint8_t queue_id,
>  }
>
>  static void
> +sw_init_qid_iqs(struct sw_evdev *sw)
> +{
> +	int i, j;
> +
> +	/* Initialize the IQ memory of all configured qids */
> +	for (i = 0; i < RTE_EVENT_MAX_QUEUES_PER_DEV; i++) {
> +		struct sw_qid *qid = &sw->qids[i];
> +
> +		if (!qid->initialized)
> +			continue;
> +
> +		for (j = 0; j < SW_IQS_MAX; j++)
> +			iq_init(sw, &qid->iq[j]);
> +	}
> +}
> +
> +static void
> +sw_clean_qid_iqs(struct sw_evdev *sw)
> +{
> +	int i, j;
> +
> +	/* Release the IQ memory of all configured qids */
> +	for (i = 0; i < RTE_EVENT_MAX_QUEUES_PER_DEV; i++) {
> +		struct sw_qid *qid = &sw->qids[i];
> +
> +		for (j = 0; j < SW_IQS_MAX; j++) {
> +			if (!qid->iq[j].head)
> +				continue;
> +			iq_free_chunk_list(sw, qid->iq[j].head);
> +			qid->iq[j].head = NULL;
> +		}
> +	}
> +}
> +
> +static void
>  sw_queue_def_conf(struct rte_eventdev *dev, uint8_t queue_id,
>  				 struct rte_event_queue_conf *conf)
>  {
> @@ -434,7 +454,10 @@ sw_dev_configure(const struct rte_eventdev *dev)
>  	num_chunks = ((SW_INFLIGHT_EVENTS_TOTAL/SW_EVS_PER_Q_CHUNK)+1) +
>  			sw->qid_count*SW_IQS_MAX*2;
>
> -	/* If this is a reconfiguration, free the previous IQ allocation */
> +	/* If this is a reconfiguration, free the previous IQ allocation. All
> +	 * IQ chunk references were cleaned out of the QIDs in sw_stop(), and
> +	 * will be reinitialized in sw_start().
> +	 */
>  	if (sw->chunks)
>  		rte_free(sw->chunks);
>
> @@ -667,8 +690,8 @@ sw_start(struct rte_eventdev *dev)
>
>  	/* check all queues are configured and mapped to ports*/
>  	for (i = 0; i < sw->qid_count; i++)
> -		if (sw->qids[i].iq[0].head == NULL ||
> -				sw->qids[i].cq_num_mapped_cqs == 0) {
> +		if (!sw->qids[i].initialized ||
> +		    sw->qids[i].cq_num_mapped_cqs == 0) {
>  			SW_LOG_ERR("Queue %d not configured\n", i);
>  			return -ENOLINK;
>  		}
> @@ -689,6 +712,8 @@ sw_start(struct rte_eventdev *dev)
>  		}
>  	}
>
> +	sw_init_qid_iqs(sw);
> +
>  	if (sw_xstats_init(sw) < 0)
>  		return -EINVAL;
>
> @@ -702,6 +727,7 @@ static void
>  sw_stop(struct rte_eventdev *dev)
>  {
>  	struct sw_evdev *sw = sw_pmd_priv(dev);
> +	sw_clean_qid_iqs(sw);
>  	sw_xstats_uninit(sw);
>  	sw->started = 0;
>  	rte_smp_wmb();
> --
> 2.7.4
>

Sw eventdev works well alongside Rx adapter now.

Cheers,
Pavan.

Acked-by: Pavan Nikhilesh <pbhagavatula at caviumnetworks.com>


More information about the dev mailing list