[dpdk-dev] [PATCH] igb: fix crash with offload on 82575 chipset

Ananyev, Konstantin konstantin.ananyev at intel.com
Fri Mar 25 15:06:51 CET 2016



> -----Original Message-----
> From: dev [mailto:dev-bounces at dpdk.org] On Behalf Of Olivier Matz
> Sent: Friday, March 25, 2016 10:32 AM
> To: dev at dpdk.org
> Cc: Lu, Wenzhuo
> Subject: [dpdk-dev] [PATCH] igb: fix crash with offload on 82575 chipset
> 
> On the 82575 chipset, there is a pool of global TX contexts instead of 2
> per queues on 82576. See Table A-1 "Changes in Programming Interface
> Relative to 82575" of Intel® 82576EB GbE Controller datasheet (*).
> 
> In the driver, the contexts are attributed to a TX queue: 0-1 for txq0,
> 2-3 for txq1, and so on.
> 
> In igbe_set_xmit_ctx(), the variable ctx_curr contains the index of the
> per-queue context (0 or 1), and ctx_idx contains the index to be given
> to the hardware (0 to 7). The size of txq->ctx_cache[] is 2, and must
> be indexed with ctx_curr to avoid an out-of-bound access.
> 
> Also, the index returned by what_advctx_update() is the per-queue
> index (0 or 1), so we need to add txq->ctx_start before sending it
> to the hardware.
> 
> (*) The datasheets says 16 global contexts, however the IDX fields in TX
>     descriptors are 3 bits, which gives a total of 8 contexts. The
>     driver assumes there are 8 contexts on 82575: 2 per queues, 4 txqs.
> 
> Fixes: 4c8db5f09a ("igb: enable TSO support")
> Fixes: af75078fec ("first public release")
> Signed-off-by: Olivier Matz <olivier.matz at 6wind.com>
> ---
>  drivers/net/e1000/igb_rxtx.c | 6 +++---
>  1 file changed, 3 insertions(+), 3 deletions(-)
> 
> diff --git a/drivers/net/e1000/igb_rxtx.c b/drivers/net/e1000/igb_rxtx.c
> index e527895..529dba4 100644
> --- a/drivers/net/e1000/igb_rxtx.c
> +++ b/drivers/net/e1000/igb_rxtx.c
> @@ -325,9 +325,9 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq,
>  	}
> 
>  	txq->ctx_cache[ctx_curr].flags = ol_flags;
> -	txq->ctx_cache[ctx_idx].tx_offload.data =
> +	txq->ctx_cache[ctx_curr].tx_offload.data =
>  		tx_offload_mask.data & tx_offload.data;
> -	txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask;
> +	txq->ctx_cache[ctx_curr].tx_offload_mask = tx_offload_mask;
> 
>  	ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl);
>  	vlan_macip_lens = (uint32_t)tx_offload.data;
> @@ -450,7 +450,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
>  			ctx = what_advctx_update(txq, tx_ol_req, tx_offload);
>  			/* Only allocate context descriptor if required*/
>  			new_ctx = (ctx == IGB_CTX_NUM);
> -			ctx = txq->ctx_curr;
> +			ctx = txq->ctx_curr + txq->ctx_start;
>  			tx_last = (uint16_t) (tx_last + new_ctx);
>  		}
>  		if (tx_last >= txq->nb_tx_desc)
> --

Acked-by: Konstantin Ananyev <konstantin.ananyev at intel.com>

> 2.1.4



More information about the dev mailing list