[dpdk-dev] [PATCH] The VMXNET3 PMD can't receive packet suddenly after a lot of traffic coming in
Yong Wang
yongwang at vmware.com
Tue Dec 22 22:36:15 CET 2015
On 7/23/15, 5:53 PM, "dev on behalf of Marco Lee" <dev-bounces at dpdk.org on behalf of mac_leehk at yahoo.com.hk> wrote:
>The RX of VMXNET3 PMD will have deadlock when a lot of traffic coming in.
>The root cause is due to mbuf allocation fail in vmxnet3_post_rx_bufs()
>and there is no error handling when it is called from vmxnet3_recv_pkts().
>The RXD will not have "free" mbuf for it but the counter still increment.
Can you describe what counter this refers to?
>Finally, no packet can be received.
>
>This fix is allocate the mbuf first, if the allocation is failed,
>then reuse the old mbuf. If the allocation is success,
>the vmxnet3_post_rx_bufs() will call vmxnet3_renew_desc()
>and RXD will be renew inside.
I didn’t see this part of logic implemented.
>
>Signed-off-by: Marco Lee <mac_leehk at yahoo.com.hk/marco.lee at ruckuswireless.com>
>---
> drivers/net/vmxnet3/vmxnet3_rxtx.c | 37 +++++++++++++++++++++++++++++++++++-
> 1 file changed, 36 insertions(+), 1 deletion(-)
>
>diff --git a/drivers/net/vmxnet3/vmxnet3_rxtx.c b/drivers/net/vmxnet3/vmxnet3_rxtx.c
>index 39ad6ef..cbed438 100644
>--- a/drivers/net/vmxnet3/vmxnet3_rxtx.c
>+++ b/drivers/net/vmxnet3/vmxnet3_rxtx.c
>@@ -421,6 +421,35 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
> return nb_tx;
> }
>
>+static inline void
>+vmxnet3_renew_desc(vmxnet3_rx_queue_t *rxq, uint8_t ring_id,
>+ struct rte_mbuf *mbuf)
>+{
>+ uint32_t val = 0;
>+ struct vmxnet3_cmd_ring *ring = &rxq->cmd_ring[ring_id];
>+
Remove this blank line.
>+ struct Vmxnet3_RxDesc *rxd;
>+ vmxnet3_buf_info_t *buf_info = &ring->buf_info[ring->next2fill];
>+
>+ rxd = (struct Vmxnet3_RxDesc *)(ring->base + ring->next2fill);
>+
nit: this can be merged with the above definition.
>+ if (ring->rid == 0)
>+ val = VMXNET3_RXD_BTYPE_HEAD;
>+ else
>+ val = VMXNET3_RXD_BTYPE_BODY;
>+
Remove the trailing space here.
>+
>+ buf_info->m = mbuf;
>+ buf_info->len = (uint16_t)(mbuf->buf_len - RTE_PKTMBUF_HEADROOM);
>+ buf_info->bufPA = RTE_MBUF_DATA_DMA_ADDR_DEFAULT(mbuf);
>+
>+ rxd->addr = buf_info->bufPA;
>+ rxd->btype = val;
>+ rxd->len = buf_info->len;
>+ rxd->gen = ring->gen;
>+
>+ vmxnet3_cmd_ring_adv_next2fill(ring);
>+}
> /*
> * Allocates mbufs and clusters. Post rx descriptors with buffer details
> * so that device can receive packets in those buffers.
>@@ -578,6 +607,8 @@ vmxnet3_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t nb_pkts)
> if (nb_rx >= nb_pkts)
> break;
>
>+ struct rte_mbuf *rep;
What does rep mean? Can you rename it to something easier to understand
(say newm, or m2)?
Also, please move the definition to the top of this block.
>+ rep = rte_rxmbuf_alloc(rxq->mb_pool);
> idx = rcd->rxdIdx;
> ring_idx = (uint8_t)((rcd->rqID == rxq->qid1) ? 0 : 1);
> rxd = (Vmxnet3_RxDesc *)rxq->cmd_ring[ring_idx].base + idx;
>@@ -651,13 +682,17 @@ vmxnet3_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t nb_pkts)
>
> vmxnet3_rx_offload(rcd, rxm);
>
>+ if (unlikely(rep == NULL)) {
>+ rep = rxm;
>+ goto rcd_done;
>+ }
Should this be moved earlier? Also need to update the rx_buf_alloc_failure
counter.
> rx_pkts[nb_rx++] = rxm;
> rcd_done:
> rxq->cmd_ring[ring_idx].next2comp = idx;
> VMXNET3_INC_RING_IDX_ONLY(rxq->cmd_ring[ring_idx].next2comp, rxq->cmd_ring[ring_idx].size);
>
> /* It's time to allocate some new buf and renew descriptors */
>- vmxnet3_post_rx_bufs(rxq, ring_idx);
>+ vmxnet3_renew_desc(rxq, ring_idx, rep);
> if (unlikely(rxq->shared->ctrl.updateRxProd)) {
> VMXNET3_WRITE_BAR0_REG(hw, rxprod_reg[ring_idx] + (rxq->queue_id * VMXNET3_REG_ALIGN),
> rxq->cmd_ring[ring_idx].next2fill);
>--
>1.7.9.5
>
More information about the dev
mailing list