[dpdk-stable] [dpdk-dev] [PATCH v3 2/3] net/af_xdp: use correct fill queue addresses
Ye Xiaolong
xiaolong.ye at intel.com
Thu Feb 13 04:09:17 CET 2020
On 02/10, Ciara Loftus wrote:
>The fill queue addresses should start at the beginning of the mempool
>object instead of the beginning of the mbuf. This is because the umem
>frame headroom includes the mp hdrobj size. Starting at this point
>ensures AF_XDP doesn't write past the available room in the frame, in
>the case of larger packets which are close to the size of the mbuf.
>
>Fixes: d8a210774e1d ("net/af_xdp: support unaligned umem chunks")
>Cc: stable at dpdk.org
>
>Signed-off-by: Ciara Loftus <ciara.loftus at intel.com>
>---
> drivers/net/af_xdp/rte_eth_af_xdp.c | 25 +++++++++++++++++--------
> 1 file changed, 17 insertions(+), 8 deletions(-)
>
>diff --git a/drivers/net/af_xdp/rte_eth_af_xdp.c b/drivers/net/af_xdp/rte_eth_af_xdp.c
>index 8b189119c..1e98cd44f 100644
>--- a/drivers/net/af_xdp/rte_eth_af_xdp.c
>+++ b/drivers/net/af_xdp/rte_eth_af_xdp.c
>@@ -172,7 +172,8 @@ reserve_fill_queue_zc(struct xsk_umem_info *umem, uint16_t reserve_size,
> uint64_t addr;
>
> fq_addr = xsk_ring_prod__fill_addr(fq, idx++);
>- addr = (uint64_t)bufs[i] - (uint64_t)umem->buffer;
>+ addr = (uint64_t)bufs[i] - (uint64_t)umem->buffer -
>+ umem->mb_pool->header_size;
> *fq_addr = addr;
> }
>
>@@ -271,8 +272,11 @@ af_xdp_rx_zc(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
> addr = xsk_umem__extract_addr(addr);
>
> bufs[i] = (struct rte_mbuf *)
>- xsk_umem__get_data(umem->buffer, addr);
>- bufs[i]->data_off = offset - sizeof(struct rte_mbuf);
>+ xsk_umem__get_data(umem->buffer, addr +
>+ umem->mb_pool->header_size);
>+ bufs[i]->data_off = offset - sizeof(struct rte_mbuf) -
>+ rte_pktmbuf_priv_size(umem->mb_pool) -
>+ umem->mb_pool->header_size;
>
> rte_pktmbuf_pkt_len(bufs[i]) = len;
> rte_pktmbuf_data_len(bufs[i]) = len;
>@@ -385,7 +389,8 @@ pull_umem_cq(struct xsk_umem_info *umem, int size)
> #if defined(XDP_UMEM_UNALIGNED_CHUNK_FLAG)
> addr = xsk_umem__extract_addr(addr);
> rte_pktmbuf_free((struct rte_mbuf *)
>- xsk_umem__get_data(umem->buffer, addr));
>+ xsk_umem__get_data(umem->buffer,
>+ addr + umem->mb_pool->header_size));
> #else
> rte_ring_enqueue(umem->buf_ring, (void *)addr);
> #endif
>@@ -443,9 +448,11 @@ af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
> }
> desc = xsk_ring_prod__tx_desc(&txq->tx, idx_tx);
> desc->len = mbuf->pkt_len;
>- addr = (uint64_t)mbuf - (uint64_t)umem->buffer;
>+ addr = (uint64_t)mbuf - (uint64_t)umem->buffer -
>+ umem->mb_pool->header_size;
> offset = rte_pktmbuf_mtod(mbuf, uint64_t) -
>- (uint64_t)mbuf;
>+ (uint64_t)mbuf +
>+ umem->mb_pool->header_size;
> offset = offset << XSK_UNALIGNED_BUF_OFFSET_SHIFT;
> desc->addr = addr | offset;
> count++;
>@@ -466,9 +473,11 @@ af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
> desc = xsk_ring_prod__tx_desc(&txq->tx, idx_tx);
> desc->len = mbuf->pkt_len;
>
>- addr = (uint64_t)local_mbuf - (uint64_t)umem->buffer;
>+ addr = (uint64_t)local_mbuf - (uint64_t)umem->buffer -
>+ umem->mb_pool->header_size;
> offset = rte_pktmbuf_mtod(local_mbuf, uint64_t) -
>- (uint64_t)local_mbuf;
>+ (uint64_t)local_mbuf +
>+ umem->mb_pool->header_size;
> pkt = xsk_umem__get_data(umem->buffer, addr + offset);
> offset = offset << XSK_UNALIGNED_BUF_OFFSET_SHIFT;
> desc->addr = addr | offset;
>--
>2.17.1
>
Reviewed-by: Xiaolong Ye <xiaolong.ye at intel.com>
More information about the stable
mailing list