[dts] [PATCH V1] add testsuite of loopback multi queues

lihong lihongx.ma at intel.com
Wed Apr 3 20:08:09 CEST 2019


Signed-off-by: lihong <lihongx.ma at intel.com>
---
 tests/TestSuite_loopback_multi_queues.py | 303 +++++++++++++++++++++++++++++++
 1 file changed, 303 insertions(+)
 create mode 100644 tests/TestSuite_loopback_multi_queues.py

diff --git a/tests/TestSuite_loopback_multi_queues.py b/tests/TestSuite_loopback_multi_queues.py
new file mode 100644
index 0000000..8ec464a
--- /dev/null
+++ b/tests/TestSuite_loopback_multi_queues.py
@@ -0,0 +1,303 @@
+# BSD LICENSE
+#
+# Copyright(c) 2010-2019 Intel Corporation. All rights reserved.
+# All rights reserved.
+#
+# Redistribution and use in source and binary forms, with or without
+# modification, are permitted provided that the following conditions
+# are met:
+#
+#   * Redistributions of source code must retain the above copyright
+#     notice, this list of conditions and the following disclaimer.
+#   * Redistributions in binary form must reproduce the above copyright
+#     notice, this list of conditions and the following disclaimer in
+#     the documentation and/or other materials provided with the
+#     distribution.
+#   * Neither the name of Intel Corporation nor the names of its
+#     contributors may be used to endorse or promote products derived
+#     from this software without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+"""
+DPDK Test suite.
+Test vhost/virtio-user loopback multi-queues on 8 tx/rx path.
+Includes Mergeable, Normal, Vector_RX, Inorder mergeable,
+Inorder no-mergeable, Virtio 1.1 mergeable, Virtio 1.1 no-mergeable Path,
+Virtio 1.1 inorder no-mergeable Path.
+"""
+
+import utils
+import time
+import re
+from test_case import TestCase
+
+
+class TestLoopbackMultiQueues(TestCase):
+
+    def set_up_all(self):
+        """
+        Run at the start of each test suite.
+        """
+        self.frame_sizes = [64, 128, 256, 512, 1024, 1518]
+        self.core_config = "1S/18C/1T"
+        self.nb_cores = 2
+        self.queue_number = 2
+        self.dut_ports = self.dut.get_ports()
+        self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
+        self.core_list = self.dut.get_core_list(
+            self.core_config, socket=self.ports_socket)
+
+    def set_up(self):
+        """
+        Run before each test case.
+        """
+        # Prepare the result table
+        self.dut.send_expect("rm -rf ./vhost-net*", "#")
+        self.dut.send_expect("killall -s INT testpmd", "#")
+        self.table_header = ["Frame", "Mode", "Throughput(Mpps)"]
+        self.result_table_create(self.table_header)
+
+        self.vhost = self.dut.new_session(suite="vhost")
+        self.virtio_user = self.dut.new_session(suite="virtio-user")
+
+    def get_core_mask(self):
+        """
+        get the coremask about vhost and virito depend on the queue number
+        """
+        self.core_list_user = self.core_list[0:self.nb_cores + 1]
+        self.core_list_host = self.core_list[self.nb_cores + 1:2 * self.nb_cores + 2]
+        self.core_mask_user = utils.create_mask(self.core_list_user)
+        self.core_mask_host = utils.create_mask(self.core_list_host)
+
+    def start_vhost_testpmd(self):
+        """
+        start testpmd on vhost
+        """
+        command_line_client = self.dut.target + "/app/testpmd -n %d -c %s --socket-mem 1024,1024" + \
+                              " --legacy-mem --no-pci --file-prefix=vhost --vdev " + \
+                              "'net_vhost0,iface=vhost-net,queues=%d' -- -i --nb-cores=%d " + \
+                              "--rxq=%d --txq=%d --txd=1024 --rxd=1024"
+        command_line_client = command_line_client % (
+            self.dut.get_memory_channels(), self.core_mask_host, self.queue_number,
+            self.nb_cores, self.queue_number, self.queue_number)
+        self.vhost.send_expect(command_line_client, "testpmd> ", 120)
+        self.vhost.send_expect("set fwd mac", "testpmd> ", 120)
+
+    def start_virtio_testpmd(self, args):
+        """
+        start testpmd on virtio
+        """
+        command_line_user = self.dut.target + "/app/testpmd -n %d -c %s " + \
+                            " --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio " + \
+                            "--vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=%d,%s " + \
+                            "-- -i %s --rss-ip --nb-cores=%d --rxq=%d --txq=%d --txd=1024 --rxd=1024"
+        command_line_user = command_line_user % (
+            self.dut.get_memory_channels(), self.core_mask_user, self.queue_number,
+            args["version"], args["path"], self.nb_cores, self.queue_number, self.queue_number)
+        self.virtio_user.send_expect(command_line_user, "testpmd> ", 120)
+        self.virtio_user.send_expect("set fwd mac", "testpmd> ", 120)
+        self.virtio_user.send_expect("start", "testpmd> ", 120)
+
+    def calculate_avg_throughput(self):
+        """
+        calculate the average throughput
+        """
+        results = 0.0
+        results_row = []
+        for i in range(10):
+            out = self.vhost.send_expect("show port stats all", "testpmd>", 60)
+            time.sleep(5)
+            lines = re.search("Rx-pps:\s*(\d*)", out)
+            result = lines.group(1)
+            results += float(result)
+        Mpps = results / (1000000 * 10)
+        self.verify(Mpps > 0, "port can not receive packets")
+        return Mpps
+
+    def update_result_table(self, frame_size, case_info, Mpps):
+        results_row = [frame_size]
+        results_row.append(case_info)
+        results_row.append(Mpps)
+        self.result_table_add(results_row)
+
+    def check_packets_of_each_queue(self, frame_size):
+        """
+        check each queue has receive packets
+        """
+        out = self.vhost.send_expect("stop", "testpmd> ", 60)
+        for queue_index in range(0, self.queue_number):
+            queue = "Queue= %d" % queue_index
+            index = out.find(queue)
+            rx = re.search("RX-packets:\s*(\d*)", out[index:])
+            tx = re.search("TX-packets:\s*(\d*)", out[index:])
+            rx_packets = int(rx.group(1))
+            tx_packets = int(tx.group(1))
+            self.verify(rx_packets > 0 and tx_packets > 0,
+                   "The queue %d rx-packets or tx-packets is 0 about " %
+                   queue_index + \
+                   "frame_size:%d, rx-packets:%d, tx-packets:%d" %
+                   (frame_size, rx_packets, tx_packets))
+
+        self.vhost.send_expect("clear port stats all", "testpmd> ", 60)
+
+    def send_and_verify(self, case_info):
+        """
+        start to send packets and calculate avg throughput
+        """
+        for frame_size in self.frame_sizes:
+            self.vhost.send_expect("set txpkts %d" % frame_size, "testpmd> ", 30)
+            self.vhost.send_expect("start tx_first 32", "testpmd> ", 30)
+            Mpps = self.calculate_avg_throughput()
+            self.update_result_table(frame_size, case_info, Mpps)
+            self.check_packets_of_each_queue(frame_size)
+        self.result_table_print()
+
+    def close_all_testpmd(self):
+        """
+        close all testpmd of vhost and virtio
+        """
+        self.vhost.send_expect("quit", "#", 60)
+        self.virtio_user.send_expect("quit", "#", 60)
+
+    def close_all_session(self):
+        """
+        close all session of vhost and vhost-user
+        """
+        self.dut.close_session(self.virtio_user)
+        self.dut.close_session(self.vhost)
+
+    def test_loopback_multi_queue_virtio11_mergeable(self):
+        """
+        performance for Vhost PVP virtio 1.1 Mergeable Path.
+        """
+        virtio_pmd_arg = {"version": "in_order=0,packed_vq=1,mrg_rxbuf=1",
+                            "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+        self.nb_cores = 2
+        self.queue_number = 2
+        self.get_core_mask()
+        self.start_vhost_testpmd()
+        self.start_virtio_testpmd(virtio_pmd_arg)
+        self.send_and_verify("virtio_1.1 mergeable on")
+        self.close_all_testpmd()
+
+    def test_loopback_multi_queue_virtio11_normal(self):
+        """
+        performance for Vhost PVP virtio1.1 Normal Path.
+        """
+        virtio_pmd_arg = {"version": "in_order=0,packed_vq=1,mrg_rxbuf=0",
+                            "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+        self.nb_cores = 2
+        self.queue_number = 2
+        self.get_core_mask()
+        self.start_vhost_testpmd()
+        self.start_virtio_testpmd(virtio_pmd_arg)
+        self.send_and_verify("virtio_1.1 normal")
+        self.close_all_testpmd()
+
+    def test_loopback_multi_queue_virtio11_inorder(self):
+        """
+        performance for Vhost PVP virtio1.1 inorder Path.
+        """
+        virtio_pmd_arg = {"version": "in_order=1,packed_vq=1,mrg_rxbuf=0",
+                            "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+        self.nb_cores = 2
+        self.queue_number = 2
+        self.get_core_mask()
+        self.start_vhost_testpmd()
+        self.start_virtio_testpmd(virtio_pmd_arg)
+        self.send_and_verify("virtio_1.1 inorder")
+        self.close_all_testpmd()
+
+    def test_loopback_multi_queue_inorder_mergeable(self):
+        """
+        performance for Vhost PVP In_order mergeable Path.
+        """
+        virtio_pmd_arg = {"version": "packed_vq=0,in_order=1,mrg_rxbuf=1",
+                            "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+        self.nb_cores = 2
+        self.queue_number = 2
+        self.get_core_mask()
+        self.start_vhost_testpmd()
+        self.start_virtio_testpmd(virtio_pmd_arg)
+        self.send_and_verify("inoder mergeable on")
+        self.close_all_testpmd()
+
+    def test_loopback_multi_queue_inorder_no_mergeable(self):
+        """
+        performance for Vhost PVP In_order no_mergeable Path.
+        """
+        virtio_pmd_arg = {"version": "packed_vq=0,in_order=1,mrg_rxbuf=0",
+                        "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+        self.nb_cores = 2
+        self.queue_number = 2
+        self.get_core_mask()
+        self.start_vhost_testpmd()
+        self.start_virtio_testpmd(virtio_pmd_arg)
+        self.send_and_verify("inoder mergeable off")
+        self.close_all_testpmd()
+
+    def test_loopback_mulit_queue_mergeable(self):
+        """
+        performance for Vhost PVP Mergeable Path.
+        """
+        virtio_pmd_arg = {"version": "packed_vq=0,in_order=0,mrg_rxbuf=1",
+                            "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+        self.nb_cores = 8
+        self.queue_number = 8
+        self.get_core_mask()
+        self.start_vhost_testpmd()
+        self.start_virtio_testpmd(virtio_pmd_arg)
+        self.send_and_verify("virito mergeable")
+        self.close_all_testpmd()
+
+    def test_loopback_multi_queue_normal(self):
+        """
+        performance for Vhost PVP Normal Path.
+        """
+        virtio_pmd_arg = {"version": "packed_vq=0,in_order=0,mrg_rxbuf=0",
+                            "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"}
+        self.nb_cores = 8
+        self.queue_number = 8
+        self.get_core_mask()
+        self.start_vhost_testpmd()
+        self.start_virtio_testpmd(virtio_pmd_arg)
+        self.send_and_verify("virtio normal")
+        self.close_all_testpmd()
+
+    def test_loopback_multi_queue_vector_rx(self):
+        """
+        performance for Vhost PVP Vector_RX Path
+        """
+        virtio_pmd_arg = {"version": "packed_vq=0,in_order=0,mrg_rxbuf=0",
+                            "path": "--tx-offloads=0x0"}
+        self.nb_cores = 8
+        self.queue_number = 8
+        self.get_core_mask()
+        self.start_vhost_testpmd()
+        self.start_virtio_testpmd(virtio_pmd_arg)
+        self.send_and_verify("virito vector rx")
+        self.close_all_testpmd()
+
+    def tear_down(self):
+        """
+        Run after each test case.
+        """
+        self.dut.send_expect("killall -s INT testpmd", "#")
+        self.close_all_session()
+
+    def tear_down_all(self):
+        """
+        Run after each test suite.
+        """
+        pass
-- 
2.7.4



More information about the dts mailing list