[dpdk-dev] [PATCH 1/6] cxgbe: Optimize forwarding performance for 40G
Rahul Lakkireddy
rahul.lakkireddy at chelsio.com
Mon Oct 5 14:42:08 CEST 2015
Hi Konstantin,
On Monday, October 10/05/15, 2015 at 04:46:40 -0700, Ananyev, Konstantin wrote:
>
>
> > -----Original Message-----
> > From: dev [mailto:dev-bounces at dpdk.org] On Behalf Of Rahul Lakkireddy
> > Sent: Monday, October 05, 2015 11:06 AM
> > To: Aaron Conole
> > Cc: dev at dpdk.org; Felix Marti; Kumar A S; Nirranjan Kirubaharan
> > Subject: Re: [dpdk-dev] [PATCH 1/6] cxgbe: Optimize forwarding performance for 40G
> >
> > Hi Aaron,
> >
> > On Friday, October 10/02/15, 2015 at 14:48:28 -0700, Aaron Conole wrote:
> > > Hi Rahul,
> > >
> > > Rahul Lakkireddy <rahul.lakkireddy at chelsio.com> writes:
> > >
> > > > Update sge initialization with respect to free-list manager configuration
> > > > and ingress arbiter. Also update refill logic to refill mbufs only after
> > > > a certain threshold for rx. Optimize tx packet prefetch and free.
> > > <<snip>>
> > > > for (i = 0; i < sd->coalesce.idx; i++) {
> > > > - rte_pktmbuf_free(sd->coalesce.mbuf[i]);
> > > > + struct rte_mbuf *tmp = sd->coalesce.mbuf[i];
> > > > +
> > > > + do {
> > > > + struct rte_mbuf *next = tmp->next;
> > > > +
> > > > + rte_pktmbuf_free_seg(tmp);
> > > > + tmp = next;
> > > > + } while (tmp);
> > > > sd->coalesce.mbuf[i] = NULL;
> > > Pardon my ignorance here, but rte_pktmbuf_free does this work. I can't
> > > actually see much difference between your rewrite of this block, and
> > > the implementation of rte_pktmbuf_free() (apart from moving your branch
> > > to the end of the function). Did your microbenchmarking really show this
> > > as an improvement?
> > >
> > > Thanks for your time,
> > > Aaron
> >
> > rte_pktmbuf_free calls rte_mbuf_sanity_check which does a lot of
> > checks.
>
> Only when RTE_LIBRTE_MBUF_DEBUG is enabled in your config.
> By default it is switched off.
Right. I clearly missed this.
I am running with default config only btw.
>
> > This additional check seems redundant for single segment
> > packets since rte_pktmbuf_free_seg also performs rte_mbuf_sanity_check.
> >
> > Several PMDs already prefer to use rte_pktmbuf_free_seg directly over
> > rte_pktmbuf_free as it is faster.
>
> Other PMDs use rte_pktmbuf_free_seg() as each TD has an associated
> with it segment. So as HW is done with the TD, SW frees associated segment.
> In your case I don't see any point in re-implementing rte_pktmbuf_free() manually,
> and I don't think it would be any faster.
>
> Konstantin
As I mentioned below, I am clearly seeing a difference of 1 Mpps. And 1
Mpps is not a small difference IMHO.
When running l3fwd with 8 queues, I also collected a perf report.
When using rte_pktmbuf_free, I see that it eats up around 6% cpu as
below in perf top report:-
--------------------
32.00% l3fwd [.] cxgbe_poll
22.25% l3fwd [.] t4_eth_xmit
20.30% l3fwd [.] main_loop
6.77% l3fwd [.] rte_pktmbuf_free
4.86% l3fwd [.] refill_fl_usembufs
2.00% l3fwd [.] write_sgl
.....
--------------------
While, when using rte_pktmbuf_free_seg directly, I don't see above
problem. perf top report now comes as:-
-------------------
33.36% l3fwd [.] cxgbe_poll
32.69% l3fwd [.] t4_eth_xmit
19.05% l3fwd [.] main_loop
5.21% l3fwd [.] refill_fl_usembufs
2.40% l3fwd [.] write_sgl
....
-------------------
I obviously missed the debug flag for rte_mbuf_sanity_check.
However, there is a clear difference of 1 Mpps. I don't know if its the
change between while construct used in rte_pktmbuf_free and the
do..while construct that I used - is making the difference.
>
> >
> > The forwarding perf. improvement with only this particular block is
> > around 1 Mpps for 64B packets when using l3fwd with 8 queues.
> >
> > Thanks,
> > Rahul
More information about the dev
mailing list