[dts] [PATCH V1] Add test suite about pvp multi path vhost single core performance
Wang, Yinan
yinan.wang at intel.com
Tue Mar 5 06:37:19 CET 2019
-----Original Message-----
From: dts [mailto:dts-bounces at dpdk.org] On Behalf Of lihong
Sent: 2019年3月4日 4:37
To: dts at dpdk.org
Cc: Ma, LihongX <lihongx.ma at intel.com>
Subject: [dts] [PATCH V1] Add test suite about pvp multi path vhost single core performance
Signed-off-by: lihong <lihongx.ma at intel.com>
Acked-by: Wang, Yinan <yinan.wang at intel.com>
---
...vp_multi_paths_vhost_single_core_performance.py | 285 +++++++++++++++++++++
1 file changed, 285 insertions(+)
create mode 100644 tests/TestSuite_pvp_multi_paths_vhost_single_core_performance.py
diff --git a/tests/TestSuite_pvp_multi_paths_vhost_single_core_performance.py b/tests/TestSuite_pvp_multi_paths_vhost_single_core_performance.py
new file mode 100644
index 0000000..c085e6e
--- /dev/null
+++ b/tests/TestSuite_pvp_multi_paths_vhost_single_core_performance.py
@@ -0,0 +1,285 @@
+# BSD LICENSE
+#
+# Copyright(c) 2010-2019 Intel Corporation. All rights reserved.
+# All rights reserved.
+#
+# Redistribution and use in source and binary forms, with or without #
+modification, are permitted provided that the following conditions #
+are met:
+#
+# * Redistributions of source code must retain the above copyright
+# notice, this list of conditions and the following disclaimer.
+# * Redistributions in binary form must reproduce the above copyright
+# notice, this list of conditions and the following disclaimer in
+# the documentation and/or other materials provided with the
+# distribution.
+# * Neither the name of Intel Corporation nor the names of its
+# contributors may be used to endorse or promote products derived
+# from this software without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS #
+"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT #
+LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR #
+A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT #
+OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, #
+SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT #
+LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, #
+DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY #
+THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT #
+(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE #
+OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+"""
+DPDK Test suite.
+Test PVP vhost single core performance using virtio_user on 7 tx/rx path.
+"""
+
+import utils
+from test_case import TestCase
+from settings import HEADER_SIZE
+
+
+class TestPVPMultiPathVhostPerformance(TestCase):
+
+ def set_up_all(self):
+ """
+ Run at the start of each test suite.
+ """
+ self.frame_sizes = [64, 128, 256, 512, 1024, 1518]
+ self.test_cycles = {'Mpps': {}, 'pct': {}}
+ self.core_config = "1S/5C/1T"
+ self.number_of_ports = 1
+ self.headers_size = HEADER_SIZE['eth'] + HEADER_SIZE['ip'] + \
+ HEADER_SIZE['udp']
+ self.dut_ports = self.dut.get_ports()
+ self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
+ ports = []
+ for port in xrange(self.number_of_ports):
+ ports.append(self.dut_ports[port])
+ self.core_list = self.dut.get_core_list(
+ self.core_config, socket=self.ports_socket)
+ self.core_list_user = self.core_list[0:3]
+ self.core_list_host = self.core_list[3:5]
+ self.port_mask = utils.create_mask(ports)
+ self.core_mask_user = utils.create_mask(self.core_list_user)
+ self.core_mask_host = utils.create_mask(self.core_list_host)
+ if self.dut.cores[len(self.dut.cores) - 1]['socket'] == '0':
+ self.socket_mem = '1024'
+ else:
+ self.socket_mem = '1024,1024'
+
+ def set_up(self):
+ """
+ Run before each test case.
+ """
+ # Prepare the result table
+ self.table_header = ['Frame']
+ self.table_header.append("Mode")
+ self.table_header.append("Mpps")
+ self.table_header.append("% linerate")
+ self.result_table_create(self.table_header)
+
+ def send_and_verify(self, case_info, frame_size):
+ """
+ Send packet with packet generator and verify
+ """
+ payload_size = frame_size - self.headers_size
+ tgen_input = []
+ for port in xrange(self.number_of_ports):
+ rx_port = self.tester.get_local_port(
+ self.dut_ports[port % self.number_of_ports])
+ tx_port = self.tester.get_local_port(
+ self.dut_ports[(port) % self.number_of_ports])
+ destination_mac = self.dut.get_mac_address(
+ self.dut_ports[(port) % self.number_of_ports])
+ self.tester.scapy_append(
+ 'wrpcap("l2fwd_%d.pcap", [Ether(dst="%s")/IP()/UDP()/("X"*%d)])' %
+ (port, destination_mac, payload_size))
+
+ tgen_input.append((tx_port, rx_port, "l2fwd_%d.pcap" %
+ port))
+
+ self.tester.scapy_execute()
+ _, pps = self.tester.traffic_generator_throughput(tgen_input)
+ Mpps = pps / 1000000.0
+ self.verify(Mpps > 0, "%s can not receive packets of frame size
+ %d" % (self.running_case, frame_size))
+
+ throughput = Mpps * 100 / \
+ float(self.wirespeed(self.nic, frame_size,
+ self.number_of_ports))
+
+ results_row = [frame_size]
+ results_row.append(case_info)
+ results_row.append(Mpps)
+ results_row.append(throughput)
+ self.result_table_add(results_row)
+
+ def start_vhost_testpmd(self):
+ """
+ start testpmd on vhost
+ """
+ # Clean the execution ENV
+ self.dut.send_expect("rm -rf ./vhost.out", "#")
+ self.dut.send_expect("rm -rf ./vhost-net*", "#")
+ self.dut.send_expect("killall -s INT testpmd", "#")
+ self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+ command_line_client = "./x86_64-native-linuxapp-gcc/app/testpmd -n %d -c %s --socket-mem " + \
+ " %s --legacy-mem --no-pci --file-prefix=vhost --vdev " + \
+ "'net_vhost0,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024"
+ command_line_client = command_line_client % (
+ self.dut.get_memory_channels(), self.core_mask_host, self.socket_mem)
+ self.vhost.send_expect(command_line_client, "testpmd> ", 120)
+ self.vhost.send_expect("set fwd mac", "testpmd> ", 120)
+ self.vhost.send_expect("start", "testpmd> ", 120)
+
+ def start_virtio_testpmd(self, args):
+ """
+ start testpmd on virtio
+ """
+ command_line_user = "./x86_64-native-linuxapp-gcc/app/testpmd -n %d -c %s " + \
+ " --socket-mem %s --legacy-mem --file-prefix=virtio " + \
+ "--vdev=virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,%s " + \
+ "-- -i %s --nb-cores=2 --txd=1024 --rxd=1024"
+ command_line_user = command_line_user % (
+ self.dut.get_memory_channels(), self.core_mask_user, self.socket_mem, args["version"], args["path"])
+ self.vhost_user.send_expect(command_line_user, "testpmd> ", 120)
+ self.vhost_user.send_expect("set fwd io", "testpmd> ", 120)
+ self.vhost_user.send_expect("start", "testpmd> ", 120)
+
+ def close_all_testpmd(self):
+ """
+ close all testpmd of vhost and virtio
+ """
+ self.vhost.send_expect("quit", "#", 60)
+ self.vhost_user.send_expect("quit", "#", 60)
+
+ def close_all_session(self):
+ """
+ close all session of vhost and vhost-user
+ """
+ self.dut.close_session(self.vhost_user)
+ self.dut.close_session(self.vhost)
+
+ def test_perf_vhost_single_core_virtio11_mergeable(self):
+ """
+ performance for Vhost PVP virtio 1.1 Mergeable Path.
+ """
+ virtio_pmd_arg = {"version": "in_order=0,packed_vq=1,mrg_rxbuf=1",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip"}
+ self.vhost_user = self.dut.new_session(suite="user")
+ self.vhost = self.dut.new_session(suite="vhost")
+ for frame_size in self.frame_sizes:
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("virtio_1.1_mergeable on", frame_size)
+ self.close_all_testpmd()
+
+ self.close_all_session()
+ self.result_table_print()
+
+ def test_perf_vhost_single_core_virtio11_no_mergeable(self):
+ """
+ performance for Vhost PVP virtio1.1 Normal Path.
+ """
+ virtio_pmd_arg = {"version": "in_order=0,packed_vq=1,mrg_rxbuf=0",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip"}
+ self.vhost_user = self.dut.new_session(suite="user")
+ self.vhost = self.dut.new_session(suite="vhost")
+ for frame_size in self.frame_sizes:
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("virtio_1.1_mergeable off", frame_size)
+ self.close_all_testpmd()
+ self.close_all_session()
+ self.result_table_print()
+
+ def test_perf_vhost_single_core_inorder_mergeable(self):
+ """
+ performance for Vhost PVP In_order mergeable Path.
+ """
+ virtio_pmd_arg = {"version": "in_order=1,mrg_rxbuf=1",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip"}
+ self.vhost_user = self.dut.new_session(suite="user")
+ self.vhost = self.dut.new_session(suite="vhost")
+ for frame_size in self.frame_sizes:
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("inoder mergeable on", frame_size)
+ self.close_all_testpmd()
+ self.close_all_session()
+ self.result_table_print()
+
+ def test_perf_vhost_single_core_inorder_no_mergeable(self):
+ """
+ performance for Vhost PVP In_order no_mergeable Path.
+ """
+ virtio_pmd_arg = {"version": "in_order=1,mrg_rxbuf=0",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip"}
+ self.vhost_user = self.dut.new_session(suite="user")
+ self.vhost = self.dut.new_session(suite="vhost")
+ for frame_size in self.frame_sizes:
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("inoder mergeable off", frame_size)
+ self.close_all_testpmd()
+ self.close_all_session()
+ self.result_table_print()
+
+ def test_perf_vhost_single_core_mergeable(self):
+ """
+ performance for Vhost PVP Mergeable Path.
+ """
+ virtio_pmd_arg = {"version": "in_order=0,mrg_rxbuf=1",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip"}
+ self.vhost_user = self.dut.new_session(suite="user")
+ self.vhost = self.dut.new_session(suite="vhost")
+ for frame_size in self.frame_sizes:
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("mergeable on", frame_size)
+ self.close_all_testpmd()
+ self.close_all_session()
+ self.result_table_print()
+
+ def test_perf_vhost_single_core_normal(self):
+ """
+ performance for Vhost PVP Normal Path.
+ """
+ virtio_pmd_arg = {"version": "in_order=0,mrg_rxbuf=0",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip"}
+ self.vhost_user = self.dut.new_session(suite="user")
+ self.vhost = self.dut.new_session(suite="vhost")
+ for frame_size in self.frame_sizes:
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("normal", frame_size)
+ self.close_all_testpmd()
+ self.close_all_session()
+ self.result_table_print()
+
+ def test_perf_vhost_single_core_vector_rx(self):
+ """
+ performance for Vhost PVP Vector_RX Path
+ """
+ virtio_pmd_arg = {"version": "in_order=0,mrg_rxbuf=0",
+ "path": "--tx-offloads=0x0"}
+ self.vhost_user = self.dut.new_session(suite="user")
+ self.vhost = self.dut.new_session(suite="vhost")
+ for frame_size in self.frame_sizes:
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("vector rx", frame_size)
+ self.close_all_testpmd()
+ self.close_all_session()
+ self.result_table_print()
+
+ def tear_down(self):
+ """
+ Run after each test case.
+ """
+ self.dut.send_expect("killall -s INT testpmd", "#")
+ self.close_all_session()
+
+ def tear_down_all(self):
+ """
+ Run after each test suite.
+ """
+ pass
--
2.7.4
More information about the dts
mailing list