[dpdk-dev] [PATCH v2 1/2] net/ixgbe: calculate the correct number of received packets in bulk alloc function
Ananyev, Konstantin
konstantin.ananyev at intel.com
Wed Feb 8 20:53:13 CET 2017
> -----Original Message-----
> From: dev [mailto:dev-bounces at dpdk.org] On Behalf Of Ananyev, Konstantin
> Sent: Wednesday, February 8, 2017 6:54 PM
> To: Yigit, Ferruh <ferruh.yigit at intel.com>; Jianbo Liu <jianbo.liu at linaro.org>; dev at dpdk.org; Zhang, Helin <helin.zhang at intel.com>;
> jerin.jacob at caviumnetworks.com
> Subject: Re: [dpdk-dev] [PATCH v2 1/2] net/ixgbe: calculate the correct number of received packets in bulk alloc function
>
> Hi Ferruh,
>
> >
> > On 2/4/2017 1:26 PM, Ananyev, Konstantin wrote:
> > >>
> > >> To get better performance, Rx bulk alloc recv function will scan 8 descs
> > >> in one time, but the statuses are not consistent on ARM platform because
> > >> the memory allocated for Rx descriptors is cacheable hugepages.
> > >> This patch is to calculate the number of received packets by scan DD bit
> > >> sequentially, and stops when meeting the first packet with DD bit unset.
> > >>
> > >> Signed-off-by: Jianbo Liu <jianbo.liu at linaro.org>
> > >> ---
> > >> drivers/net/ixgbe/ixgbe_rxtx.c | 16 +++++++++-------
> > >> 1 file changed, 9 insertions(+), 7 deletions(-)
> > >>
> > >> diff --git a/drivers/net/ixgbe/ixgbe_rxtx.c b/drivers/net/ixgbe/ixgbe_rxtx.c
> > >> index 36f1c02..613890e 100644
> > >> --- a/drivers/net/ixgbe/ixgbe_rxtx.c
> > >> +++ b/drivers/net/ixgbe/ixgbe_rxtx.c
> > >> @@ -1460,17 +1460,19 @@ static inline int __attribute__((always_inline))
> > >> for (i = 0; i < RTE_PMD_IXGBE_RX_MAX_BURST;
> > >> i += LOOK_AHEAD, rxdp += LOOK_AHEAD, rxep += LOOK_AHEAD) {
> > >> /* Read desc statuses backwards to avoid race condition */
> > >> - for (j = LOOK_AHEAD-1; j >= 0; --j)
> > >> + for (j = 0; j < LOOK_AHEAD; j++)
> > >> s[j] = rte_le_to_cpu_32(rxdp[j].wb.upper.status_error);
> > >>
> > >> - for (j = LOOK_AHEAD - 1; j >= 0; --j)
> > >> - pkt_info[j] = rte_le_to_cpu_32(rxdp[j].wb.lower.
> > >> - lo_dword.data);
> > >> + rte_smp_rmb();
> > >>
> > >> /* Compute how many status bits were set */
> > >> - nb_dd = 0;
> > >> - for (j = 0; j < LOOK_AHEAD; ++j)
> > >> - nb_dd += s[j] & IXGBE_RXDADV_STAT_DD;
> > >> + for (nb_dd = 0; nb_dd < LOOK_AHEAD &&
> > >> + (s[nb_dd] & IXGBE_RXDADV_STAT_DD); nb_dd++)
> > >> + ;
> > >> +
> > >> + for (j = 0; j < nb_dd; j++)
> > >> + pkt_info[j] = rte_le_to_cpu_32(rxdp[j].wb.lower.
> > >> + lo_dword.data);
> > >>
> > >> nb_rx += nb_dd;
> > >>
> > >> --
> > >
> > > Acked-by: Konstantin Ananyev <konstantin.ananyev at intel.com>
> >
> > Hi Konstantin,
> >
> > Is the ack valid for v3 and both patches?
>
> No, I didn't look into the second one in details.
> It is ARM specific, and I left it for people who are more familiar with ARM then me :)
> Konstantin
Actually, I had a quick look after your mail.
+ /* A.1 load 1 pkts desc */
+ descs[0] = vld1q_u64((uint64_t *)(rxdp));
+ rte_smp_rmb();
/* B.2 copy 2 mbuf point into rx_pkts */
vst1q_u64((uint64_t *)&rx_pkts[pos], mbp1);
@@ -271,10 +270,11 @@
/* B.1 load 1 mbuf point */
mbp2 = vld1q_u64((uint64_t *)&sw_ring[pos + 2]);
- descs[2] = vld1q_u64((uint64_t *)(rxdp + 2));
- /* B.1 load 2 mbuf point */
descs[1] = vld1q_u64((uint64_t *)(rxdp + 1));
- descs[0] = vld1q_u64((uint64_t *)(rxdp));
+
+ /* A.1 load 2 pkts descs */
+ descs[2] = vld1q_u64((uint64_t *)(rxdp + 2));
+ descs[3] = vld1q_u64((uint64_t *)(rxdp + 3));
Assuming that on all ARM-NEON platforms 16B reads are atomic,
I think there is no need for smp_rmb() after the desc[0] read.
What looks more appropriate to me:
descs[0] = vld1q_u64((uint64_t *)(rxdp));
descs[1] = vld1q_u64((uint64_t *)(rxdp + 1));
descs[2] = vld1q_u64((uint64_t *)(rxdp + 2));
descs[3] = vld1q_u64((uint64_t *)(rxdp + 3));
rte_smp_rmb();
...
But, as I said would be good if some ARM guys have a look here.
Konstantin
>
> >
> > Thanks,
> > ferruh
> >
> > >
> > >> 1.8.3.1
> > >
More information about the dev
mailing list