[dpdk-dev] mlx5 reta size is dynamic

Nélio Laranjeiro nelio.laranjeiro at 6wind.com
Thu Mar 22 09:54:41 CET 2018


On Thu, Mar 22, 2018 at 06:52:53AM +0000, Hanoch Haim (hhaim) wrote:
> Hi Yongseok, 
> 
> 
> RSS has a DPDK API,application can ask for the reta table size and
> configure it. In your case you are assuming specific use case and
> change the size dynamically which solve 90% of the use-cases but break
> the 10% use-case. 
> Instead, you could provide the application a consistent API and with
> that 100% of the applications can work with no issue. This is what
> happen with Intel (ixgbe/i40e)
> Another minor issue the rss_key_size return as zero but internally it
> is 40 bytes

Hi Hanoch,

Legacy DPDK API has always considered there is only a single indirection
table aka. RETA whereas this is not true [1][2] on this device.

On MLX5 there is an indirection table per Hash Rx queue according to the
list of queues making part of it.
The Hash Rx queue is configured to make the hash with configured
information:
 - Algorithm,
 - key
 - hash field (Verbs hash field)
 - Indirection table
An Hash Rx queue cannot handle multiple RSS configuration, we have an
Hash Rx queue per protocol and thus a full configuration per protocol.

In such situation, changing the RETA means stopping the traffic,
destroying every single flow, hash Rx queue, indirection table to remake
everything with the new configuration.
Until then, we always recommended to any application to restart the port
on this device after a RETA update to apply this new configuration.

Since the flow API is the new way to configure flows, application should
move to this new one instead of using old API for such behavior.
We should also remove such devop from the PMD to avoid any confusion.

Regards,

> Thanks,
> Hanoh
> 
> -----Original Message-----
> From: Yongseok Koh [mailto:yskoh at mellanox.com] 
> Sent: Wednesday, March 21, 2018 11:48 PM
> To: Hanoch Haim (hhaim)
> Cc: dev at dpdk.org
> Subject: Re: [dpdk-dev] mlx5 reta size is dynamic
> 
> On Wed, Mar 21, 2018 at 06:56:33PM +0000, Hanoch Haim (hhaim) wrote:
> > Hi mlx5 driver expert,
> > 
> > DPDK: 17.11
> > Any reason mlx5 driver change the rate table size dynamically based on 
> > the rx- queues# ?
> 
> The device only supports 2^n-sized indirection table. For example, if the number of Rx queues is 6, device can't have 1-1 mapping but the size of ind tbl could be 8, 16, 32 and so on. If we configure it as 8 for example, 2 out of 6 queues will have 1/4 of traffic while the rest 4 queues receives 1/8. We thought it was too much disparity and preferred setting the max size in order to mitigate the imbalance.
> 
> > There is a hidden assumption that the user wants to distribute the 
> > packets evenly which is not always correct.
> 
> But it is mostly correct because RSS is used for uniform distribution. The decision wasn't made based on our speculation but by many request from multiple customers.
> 
> > /* If the requested number of RX queues is not a power of two, use the
> >           * maximum indirection table size for better balancing.
> >           * The result is always rounded to the next power of two. */
> >           reta_idx_n = (1 << log2above((rxqs_n & (rxqs_n - 1)) ?
> >                                            priv->ind_table_max_size :
> >                                            rxqs_n));
> 
> Thanks,
> Yongseok

[1] https://dpdk.org/ml/archives/dev/2015-October/024668.html
[2] https://dpdk.org/ml/archives/dev/2015-October/024669.html

-- 
Nélio Laranjeiro
6WIND


More information about the dev mailing list