[dts] [PATCH V2] add testsuite of loopback multi queues
lihong
lihongx.ma at intel.com
Fri Apr 12 00:29:55 CEST 2019
Signed-off-by: lihong <lihongx.ma at intel.com>
---
tests/TestSuite_loopback_multi_queues.py | 308 +++++++++++++++++++++++++++++++
1 file changed, 308 insertions(+)
create mode 100644 tests/TestSuite_loopback_multi_queues.py
diff --git a/tests/TestSuite_loopback_multi_queues.py b/tests/TestSuite_loopback_multi_queues.py
new file mode 100644
index 0000000..5ab17fd
--- /dev/null
+++ b/tests/TestSuite_loopback_multi_queues.py
@@ -0,0 +1,308 @@
+# BSD LICENSE
+#
+# Copyright(c) 2010-2019 Intel Corporation. All rights reserved.
+# All rights reserved.
+#
+# Redistribution and use in source and binary forms, with or without
+# modification, are permitted provided that the following conditions
+# are met:
+#
+# * Redistributions of source code must retain the above copyright
+# notice, this list of conditions and the following disclaimer.
+# * Redistributions in binary form must reproduce the above copyright
+# notice, this list of conditions and the following disclaimer in
+# the documentation and/or other materials provided with the
+# distribution.
+# * Neither the name of Intel Corporation nor the names of its
+# contributors may be used to endorse or promote products derived
+# from this software without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+"""
+DPDK Test suite.
+Test vhost/virtio-user loopback multi-queues on 8 tx/rx path.
+Includes Mergeable, Normal, Vector_RX, Inorder mergeable,
+Inorder no-mergeable, Virtio 1.1 mergeable, Virtio 1.1 no-mergeable Path,
+Virtio 1.1 inorder no-mergeable Path.
+"""
+
+import utils
+import time
+import re
+from test_case import TestCase
+
+
+class TestLoopbackMultiQueues(TestCase):
+
+ def set_up_all(self):
+ """
+ Run at the start of each test suite.
+ """
+ self.frame_sizes = [64, 128, 256, 512, 1024, 1518]
+ self.core_config = "1S/18C/1T"
+ self.nb_cores = 2
+ self.queue_number = 2
+ self.cores_num = len([n for n in self.dut.cores if int(n['socket']) == 0])
+ self.verify(self.cores_num >= 18,
+ "There has not enought cores to test this case")
+ self.core_list = self.dut.get_core_list(self.core_config)
+ self.logger.info("you can config packet_size in file %s.cfg," % self.suite_name + \
+ "in region 'suite' like packet_sizes=[64, 128, 256]")
+ # get the frame_sizes from cfg file
+ if 'packet_sizes' in self.get_suite_cfg():
+ self.frame_sizes = self.get_suite_cfg()['packet_sizes']
+
+ def set_up(self):
+ """
+ Run before each test case.
+ """
+ # Prepare the result table
+ self.dut.send_expect("rm -rf ./vhost-net*", "#")
+ self.dut.send_expect("killall -s INT testpmd", "#")
+ self.table_header = ["Frame", "Mode", "Throughput(Mpps)"]
+ self.result_table_create(self.table_header)
+
+ self.vhost = self.dut.new_session(suite="vhost")
+ self.virtio_user = self.dut.new_session(suite="virtio-user")
+
+ def get_core_mask(self):
+ """
+ get the coremask about vhost and virito depend on the queue number
+ """
+ self.core_list_user = self.core_list[0:self.nb_cores + 1]
+ self.core_list_host = self.core_list[self.nb_cores + 1:2 * self.nb_cores + 2]
+ self.core_mask_user = utils.create_mask(self.core_list_user)
+ self.core_mask_host = utils.create_mask(self.core_list_host)
+
+ def start_vhost_testpmd(self):
+ """
+ start testpmd on vhost
+ """
+ command_line_client = self.dut.target + "/app/testpmd -n %d -c %s --socket-mem 1024,1024" + \
+ " --legacy-mem --no-pci --file-prefix=vhost --vdev " + \
+ "'net_vhost0,iface=vhost-net,queues=%d' -- -i --nb-cores=%d " + \
+ "--rxq=%d --txq=%d --txd=1024 --rxd=1024"
+ command_line_client = command_line_client % (
+ self.dut.get_memory_channels(), self.core_mask_host, self.queue_number,
+ self.nb_cores, self.queue_number, self.queue_number)
+ self.vhost.send_expect(command_line_client, "testpmd> ", 120)
+ self.vhost.send_expect("set fwd mac", "testpmd> ", 120)
+
+ def start_virtio_testpmd(self, args):
+ """
+ start testpmd on virtio
+ """
+ command_line_user = self.dut.target + "/app/testpmd -n %d -c %s " + \
+ " --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio " + \
+ "--vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=%d,%s " + \
+ "-- -i %s --rss-ip --nb-cores=%d --rxq=%d --txq=%d --txd=1024 --rxd=1024"
+ command_line_user = command_line_user % (
+ self.dut.get_memory_channels(), self.core_mask_user, self.queue_number,
+ args["version"], args["path"], self.nb_cores, self.queue_number, self.queue_number)
+ self.virtio_user.send_expect(command_line_user, "testpmd> ", 120)
+ self.virtio_user.send_expect("set fwd mac", "testpmd> ", 120)
+ self.virtio_user.send_expect("start", "testpmd> ", 120)
+
+ def calculate_avg_throughput(self):
+ """
+ calculate the average throughput
+ """
+ results = 0.0
+ results_row = []
+ for i in range(10):
+ out = self.vhost.send_expect("show port stats all", "testpmd>", 60)
+ time.sleep(5)
+ lines = re.search("Rx-pps:\s*(\d*)", out)
+ result = lines.group(1)
+ results += float(result)
+ Mpps = results / (1000000 * 10)
+ self.verify(Mpps > 0, "port can not receive packets")
+ return Mpps
+
+ def update_result_table(self, frame_size, case_info, Mpps):
+ results_row = [frame_size]
+ results_row.append(case_info)
+ results_row.append(Mpps)
+ self.result_table_add(results_row)
+
+ def check_packets_of_each_queue(self, frame_size):
+ """
+ check each queue has receive packets
+ """
+ out = self.vhost.send_expect("stop", "testpmd> ", 60)
+ for queue_index in range(0, self.queue_number):
+ queue = "Queue= %d" % queue_index
+ index = out.find(queue)
+ rx = re.search("RX-packets:\s*(\d*)", out[index:])
+ tx = re.search("TX-packets:\s*(\d*)", out[index:])
+ rx_packets = int(rx.group(1))
+ tx_packets = int(tx.group(1))
+ self.verify(rx_packets > 0 and tx_packets > 0,
+ "The queue %d rx-packets or tx-packets is 0 about " %
+ queue_index + \
+ "frame_size:%d, rx-packets:%d, tx-packets:%d" %
+ (frame_size, rx_packets, tx_packets))
+
+ self.vhost.send_expect("clear port stats all", "testpmd> ", 60)
+
+ def send_and_verify(self, case_info):
+ """
+ start to send packets and calculate avg throughput
+ """
+ for frame_size in self.frame_sizes:
+ self.vhost.send_expect("set txpkts %d" % frame_size, "testpmd> ", 30)
+ self.vhost.send_expect("start tx_first 32", "testpmd> ", 30)
+ Mpps = self.calculate_avg_throughput()
+ self.update_result_table(frame_size, case_info, Mpps)
+ self.check_packets_of_each_queue(frame_size)
+ self.result_table_print()
+
+ def close_all_testpmd(self):
+ """
+ close all testpmd of vhost and virtio
+ """
+ self.vhost.send_expect("quit", "#", 60)
+ self.virtio_user.send_expect("quit", "#", 60)
+
+ def close_all_session(self):
+ """
+ close all session of vhost and vhost-user
+ """
+ self.dut.close_session(self.virtio_user)
+ self.dut.close_session(self.vhost)
+
+ def test_loopback_multi_queue_virtio11_mergeable(self):
+ """
+ performance for Vhost PVP virtio 1.1 Mergeable Path.
+ """
+ virtio_pmd_arg = {"version": "in_order=0,packed_vq=1,mrg_rxbuf=1",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+ self.nb_cores = 2
+ self.queue_number = 2
+ self.get_core_mask()
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("virtio_1.1 mergeable on")
+ self.close_all_testpmd()
+
+ def test_loopback_multi_queue_virtio11_normal(self):
+ """
+ performance for Vhost PVP virtio1.1 Normal Path.
+ """
+ virtio_pmd_arg = {"version": "in_order=0,packed_vq=1,mrg_rxbuf=0",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+ self.nb_cores = 2
+ self.queue_number = 2
+ self.get_core_mask()
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("virtio_1.1 normal")
+ self.close_all_testpmd()
+
+ def test_loopback_multi_queue_virtio11_inorder(self):
+ """
+ performance for Vhost PVP virtio1.1 inorder Path.
+ """
+ virtio_pmd_arg = {"version": "in_order=1,packed_vq=1,mrg_rxbuf=0",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+ self.nb_cores = 2
+ self.queue_number = 2
+ self.get_core_mask()
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("virtio_1.1 inorder")
+ self.close_all_testpmd()
+
+ def test_loopback_multi_queue_inorder_mergeable(self):
+ """
+ performance for Vhost PVP In_order mergeable Path.
+ """
+ virtio_pmd_arg = {"version": "packed_vq=0,in_order=1,mrg_rxbuf=1",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+ self.nb_cores = 2
+ self.queue_number = 2
+ self.get_core_mask()
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("inoder mergeable on")
+ self.close_all_testpmd()
+
+ def test_loopback_multi_queue_inorder_no_mergeable(self):
+ """
+ performance for Vhost PVP In_order no_mergeable Path.
+ """
+ virtio_pmd_arg = {"version": "packed_vq=0,in_order=1,mrg_rxbuf=0",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+ self.nb_cores = 2
+ self.queue_number = 2
+ self.get_core_mask()
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("inoder mergeable off")
+ self.close_all_testpmd()
+
+ def test_loopback_mulit_queue_mergeable(self):
+ """
+ performance for Vhost PVP Mergeable Path.
+ """
+ virtio_pmd_arg = {"version": "packed_vq=0,in_order=0,mrg_rxbuf=1",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+ self.nb_cores = 8
+ self.queue_number = 8
+ self.get_core_mask()
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("virito mergeable")
+ self.close_all_testpmd()
+
+ def test_loopback_multi_queue_normal(self):
+ """
+ performance for Vhost PVP Normal Path.
+ """
+ virtio_pmd_arg = {"version": "packed_vq=0,in_order=0,mrg_rxbuf=0",
+ "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+ self.nb_cores = 8
+ self.queue_number = 8
+ self.get_core_mask()
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("virtio normal")
+ self.close_all_testpmd()
+
+ def test_loopback_multi_queue_vector_rx(self):
+ """
+ performance for Vhost PVP Vector_RX Path
+ """
+ virtio_pmd_arg = {"version": "packed_vq=0,in_order=0,mrg_rxbuf=0",
+ "path": "--tx-offloads=0x0"}
+ self.nb_cores = 8
+ self.queue_number = 8
+ self.get_core_mask()
+ self.start_vhost_testpmd()
+ self.start_virtio_testpmd(virtio_pmd_arg)
+ self.send_and_verify("virito vector rx")
+ self.close_all_testpmd()
+
+ def tear_down(self):
+ """
+ Run after each test case.
+ """
+ self.dut.send_expect("killall -s INT testpmd", "#")
+ self.close_all_session()
+
+ def tear_down_all(self):
+ """
+ Run after each test suite.
+ """
+ pass
--
2.7.4
More information about the dts
mailing list