[dpdk-stable] [dpdk-dev] net/mlx5: mellanox cx5/cx6-dx increment "rx_phy_discard_packets" with DPDK rte_flow actions COUNT & DROP
Arthas
kangzy1982 at qq.com
Thu May 6 09:59:42 CEST 2021
Hareware: CX5/CX6 DX + Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz DPDK version: 19.11.8/20.11/21.05-rc1&2
testpmd with case:
testpmd> flow create 0 ingress pattern eth / ipv4 / udp dst is 53 / end actions count / drop / end
testpmd> flow create 0 ingress pattern eth / ipv4 / udp src is 53 / end actions count / drop / end
testpmd> flow create 0 ingress pattern eth / ipv4 / tcp / end actions count /drop end
testpmd> flow list 0
ID Group Prio Attr Rule
0 0 0 i-- ETH IPV4 UDP => COUNT DROP
1 0 0 i-- ETH IPV4 UDP => COUNT DROP
2 0 0 i-- ETH IPV4 UDP => COUNT DROP
or
testpmd> flow create 0 ingress pattern eth / ipv4 / udp dst is 53 / end actions count / rss / end
testpmd> flow create 0 ingress pattern eth / ipv4 / udp src is 53 / end actions count / rss / end
testpmd> flow create 0 ingress pattern eth / ipv4 / udp / end actions count / rss / end
testpmd> flow list 0
ID Group Prio Attr Rule
0 0 0 i-- ETH IPV4 UDP => COUNT RSS
1 0 0 i-- ETH IPV4 UDP => COUNT RSS
2 0 0 i-- ETH IPV4 UDP => COUNT RSS
testpmd>
as soon as NIC create more than 1 flow , CX5/CX6-dx NIC will increment 'rx_phy_discard_packets'.
only 1 flow no problem!
Is this a CX5/CX6-DX hardware issue?
or Is it a DPDK mlx5 pmd bugs?
Best Regards!
KANG
More information about the stable
mailing list