[dts] [PATCH V1] Add testsuite of pvp qemu multi path port restart
Tu, Lijuan
lijuan.tu at intel.com
Mon Apr 15 20:24:20 CEST 2019
Applied, thanks
> -----Original Message-----
> From: dts [mailto:dts-bounces at dpdk.org] On Behalf Of lihong
> Sent: Wednesday, April 3, 2019 11:20 AM
> To: dts at dpdk.org
> Cc: Ma, LihongX <lihongx.ma at intel.com>
> Subject: [dts] [PATCH V1] Add testsuite of pvp qemu multi path port restart
>
> Signed-off-by: lihong <lihongx.ma at intel.com>
> ---
> .../TestSuite_pvp_qemu_multi_paths_port_restart.py | 326
> +++++++++++++++++++++
> 1 file changed, 326 insertions(+)
> create mode 100644
> tests/TestSuite_pvp_qemu_multi_paths_port_restart.py
>
> diff --git a/tests/TestSuite_pvp_qemu_multi_paths_port_restart.py
> b/tests/TestSuite_pvp_qemu_multi_paths_port_restart.py
> new file mode 100644
> index 0000000..38380ed
> --- /dev/null
> +++ b/tests/TestSuite_pvp_qemu_multi_paths_port_restart.py
> @@ -0,0 +1,326 @@
> +#
> +# BSD LICENSE
> +#
> +# Copyright(c) 2010-2019 Intel Corporation. All rights reserved.
> +# All rights reserved.
> +#
> +# Redistribution and use in source and binary forms, with or without #
> +modification, are permitted provided that the following conditions #
> +are met:
> +#
> +# * Redistributions of source code must retain the above copyright
> +# notice, this list of conditions and the following disclaimer.
> +# * Redistributions in binary form must reproduce the above copyright
> +# notice, this list of conditions and the following disclaimer in
> +# the documentation and/or other materials provided with the
> +# distribution.
> +# * Neither the name of Intel Corporation nor the names of its
> +# contributors may be used to endorse or promote products derived
> +# from this software without specific prior written permission.
> +#
> +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND
> CONTRIBUTORS #
> +"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT #
> +LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
> FOR #
> +A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
> COPYRIGHT #
> +OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
> INCIDENTAL, #
> +SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
> #
> +LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF
> USE, #
> +DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
> ON ANY #
> +THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT #
> +(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
> USE #
> +OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
> +
> +
> +"""
> +DPDK Test suite.
> +Benchmark pvp qemu test with 3 RX/TX PATHs, includes Mergeable, Normal,
> +Vector_RX.
> +Cover virtio 1.0 and virtio 0.95.Also cover port restart test with each
> +path """
> +import utils
> +import time
> +import re
> +from settings import HEADER_SIZE
> +from virt_common import VM
> +from test_case import TestCase
> +
> +
> +class TestPVPQemuMultiPathPortRestart(TestCase):
> +
> + def set_up_all(self):
> + """
> + Run at the start of each test suite.
> + """
> + self.frame_sizes = [64, 128, 256, 512, 1024, 1280, 1518]
> + self.core_config = "1S/3C/1T"
> + self.dut_ports = self.dut.get_ports()
> + self.verify(len(self.dut_ports) >= 1, "Insufficient ports for testing")
> + # get core mask
> + self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
> + self.core_list = self.dut.get_core_list(
> + self.core_config, socket=self.ports_socket)
> + self.core_mask = utils.create_mask(self.core_list)
> + self.dst_mac = self.dut.get_mac_address(self.dut_ports[0])
> + self.vm_dut = None
> + self.virtio1_mac = "52:54:00:00:00:01"
> +
> + def set_up(self):
> + """
> + Run before each test case.
> + """
> + # Clean the execution ENV
> + self.dut.send_expect("rm -rf ./vhost.out", "#")
> + self.dut.send_expect("killall -s INT testpmd", "#")
> + self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
> + # Prepare the result table
> + self.table_header = ["FrameSize(B)", "Mode",
> + "Throughput(Mpps)", "% linerate", "Cycle"]
> + self.result_table_create(self.table_header)
> +
> + self.vhost = self.dut.new_session(suite="vhost-user")
> +
> + def start_vhost_testpmd(self):
> + """
> + start testpmd on vhost
> + """
> + self.dut.send_expect("killall -s INT testpmd", "#")
> + self.dut.send_expect("rm -rf ./vhost-net*", "#")
> + command_client = self.dut.target + "/app/testpmd " + \
> + " -n %d -c %s --socket-mem 1024,1024 " + \
> + " --legacy-mem --file-prefix=vhost " + \
> + " --vdev 'net_vhost0,iface=vhost-net,queues=1' " + \
> + " -- -i --nb-cores=1 --txd=1024 --rxd=1024"
> + command_line_client = command_client % (
> + self.dut.get_memory_channels(), self.core_mask)
> + self.vhost.send_expect(command_line_client, "testpmd> ", 120)
> + self.vhost.send_expect("set fwd mac", "testpmd> ", 120)
> + self.vhost.send_expect("start", "testpmd> ", 120)
> +
> + def start_vm_testpmd(self, path):
> + """
> + start testpmd in vm depend on different path
> + """
> + if path == "mergeable":
> + command = self.dut.target + "/app/testpmd " + \
> + "-c 0x3 -n 3 -- -i " + \
> + "--nb-cores=1 --txd=1024 --rxd=1024"
> + elif path == "normal":
> + command = self.dut.target + "/app/testpmd " + \
> + "-c 0x3 -n 3 -- -i " + \
> + "--tx-offloads=0x0 --enable-hw-vlan-strip " + \
> + "--nb-cores=1 --txd=1024 --rxd=1024"
> + elif path == "vector_rx":
> + command = self.dut.target + "/app/testpmd " + \
> + "-c 0x3 -n 3 -- -i " + \
> + "--nb-cores=1 --txd=1024 --rxd=1024"
> + self.vm_dut.send_expect(command, "testpmd> ", 30)
> + self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> +
> + def start_one_vm(self, modem=0, mergeable=0):
> + """
> + start qemu
> + """
> + self.vm = VM(self.dut, 'vm0', 'vhost_sample')
> + vm_params = {}
> + vm_params['driver'] = 'vhost-user'
> + vm_params['opt_path'] = './vhost-net'
> + vm_params['opt_mac'] = self.virtio1_mac
> + if modem == 1 and mergeable == 0:
> + vm_params['opt_settings'] = "disable-
> modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024"
> + elif modem == 1 and mergeable == 1:
> + vm_params['opt_settings'] = "disable-
> modern=false,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024"
> + elif modem == 0 and mergeable == 0:
> + vm_params['opt_settings'] = "disable-
> modern=true,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024"
> + elif modem == 0 and mergeable == 1:
> + vm_params['opt_settings'] = "disable-
> modern=true,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024"
> + self.vm.set_vm_device(**vm_params)
> +
> + try:
> + self.vm_dut = self.vm.start()
> + if self.vm_dut is None:
> + raise Exception("Set up VM ENV failed")
> + except Exception as e:
> + self.logger.error("ERROR: Failure for %s" % str(e))
> +
> + def check_port_throughput_after_port_stop(self):
> + """
> + check the throughput after port stop
> + """
> + loop = 1
> + while(loop <= 5):
> + out = self.vhost.send_expect("show port stats 0", "testpmd>", 60)
> + lines = re.search("Rx-pps:\s*(\d*)", out)
> + result = lines.group(1)
> + if result == "0":
> + break
> + time.sleep(3)
> + loop = loop + 1
> + self.verify(result == "0", "port stop failed, it alse can
> + recevie data after stop.")
> +
> + def check_port_link_status_after_port_restart(self):
> + """
> + check the link status after port restart
> + """
> + loop = 1
> + while(loop <= 5):
> + out = self.vhost.send_expect("show port info all", "testpmd> ", 120)
> + port_status = re.findall("Link\s*status:\s*([a-z]*)", out)
> + if("down" not in port_status):
> + break
> + time.sleep(3)
> + loop = loop + 1
> +
> + self.verify("down" not in port_status, "port can not up after
> + restart")
> +
> + def port_restart(self):
> + self.vhost.send_expect("stop", "testpmd> ", 120)
> + self.vhost.send_expect("port stop 0", "testpmd> ", 120)
> + self.check_port_throughput_after_port_stop()
> + self.vhost.send_expect("clear port stats all", "testpmd> ", 120)
> + self.vhost.send_expect("port start all", "testpmd> ", 120)
> + self.check_port_link_status_after_port_restart()
> + self.vhost.send_expect("start", "testpmd> ", 120)
> +
> + def update_table_info(self, case_info, frame_size, Mpps, throughtput,
> Cycle):
> + results_row = [frame_size]
> + results_row.append(case_info)
> + results_row.append(Mpps)
> + results_row.append(throughtput)
> + results_row.append(Cycle)
> + self.result_table_add(results_row)
> +
> + def calculate_avg_throughput(self, frame_size):
> + """
> + start to send packet and get the throughput
> + """
> + payload = frame_size - HEADER_SIZE['eth'] - HEADER_SIZE['ip']
> + flow =
> '[Ether(dst="%s")/IP(src="192.168.4.1",dst="192.168.3.1")/("X"*%d)]' % (
> + self.dst_mac, payload)
> + self.tester.scapy_append('wrpcap("pvp_multipath.pcap", %s)' % flow)
> + self.tester.scapy_execute()
> +
> + tgenInput = []
> + port = self.tester.get_local_port(self.dut_ports[0])
> + tgenInput.append((port, port, "pvp_multipath.pcap"))
> + _, pps = self.tester.traffic_generator_throughput(tgenInput, delay=30)
> + Mpps = pps / 1000000.0
> + self.verify(Mpps > 0, "can not receive packets of frame size %d" %
> (frame_size))
> + throughput = Mpps * 100 / \
> + float(self.wirespeed(self.nic, frame_size, 1))
> + return Mpps, throughput
> +
> + def send_and_verify(self, case_info):
> + """
> + start to send packets and verify it
> + """
> + for frame_size in self.frame_sizes:
> + info = "Running test %s, and %d frame size." % (self.running_case,
> frame_size)
> + self.logger.info(info)
> +
> + Mpps, throughput = self.calculate_avg_throughput(frame_size)
> + self.update_table_info(case_info, frame_size, Mpps,
> + throughput, "Before Restart")
> +
> + self.port_restart()
> + Mpps, throughput = self.calculate_avg_throughput(frame_size)
> + self.update_table_info(case_info, frame_size, Mpps,
> + throughput, "After Restart")
> +
> + def close_all_testpmd(self):
> + """
> + close testpmd about vhost-user and vm_testpmd
> + """
> + self.vhost.send_expect("quit", "#", 60)
> + self.vm_dut.send_expect("quit", "#", 60)
> +
> + def close_session(self):
> + """
> + close session of vhost-user
> + """
> + self.dut.close_session(self.vhost)
> +
> + def test_perf_pvp_qemu_mergeable_mac(self):
> + """
> + performance for [frame_sizes] and restart port on virtio 0.95
> mergeable path
> + """
> + self.start_vhost_testpmd()
> + self.start_one_vm(modem=0, mergeable=1)
> + self.start_vm_testpmd(path="mergeable")
> + self.send_and_verify("virtio0.95 mergeable")
> + self.close_all_testpmd()
> + self.result_table_print()
> + self.vm.stop()
> +
> + def test_perf_pvp_qemu_normal_mac(self):
> + """
> + performance for [frame_sizes] and restart port ob virtio0.95 normal
> path
> + """
> + self.start_vhost_testpmd()
> + self.start_one_vm(modem=0, mergeable=0)
> + self.start_vm_testpmd(path="normal")
> + self.send_and_verify("virtio0.95 normal")
> + self.close_all_testpmd()
> + self.result_table_print()
> + self.vm.stop()
> +
> + def test_perf_pvp_qemu_vector_rx_mac(self):
> + """
> + performance for [frame_sizes] and restart port on virtio0.95 vector_rx
> + """
> + self.start_vhost_testpmd()
> + self.start_one_vm(modem=0, mergeable=0)
> + self.start_vm_testpmd(path="vector_rx")
> + self.send_and_verify("virtio0.95 vector_rx")
> + self.close_all_testpmd()
> + self.result_table_print()
> + self.vm.stop()
> +
> + def test_perf_pvp_qemu_modern_mergeable_mac(self):
> + """
> + performance for [frame_sizes] and restart port on virtio1.0 mergeable
> path
> + """
> + self.start_vhost_testpmd()
> + self.start_one_vm(modem=1, mergeable=1)
> + self.start_vm_testpmd(path="mergeable")
> + self.send_and_verify("virtio1.0 mergeable")
> + self.close_all_testpmd()
> + self.result_table_print()
> + self.vm.stop()
> +
> + def test_perf_pvp_qemu_modern_normal_path(self):
> + """
> + performance for [frame_sizes] and restart port on virito1.0 normal
> path
> + """
> + self.start_vhost_testpmd()
> + self.start_one_vm(modem=1, mergeable=0)
> + self.start_vm_testpmd(path="normal")
> + self.send_and_verify("virtio1.0 normal")
> + self.close_all_testpmd()
> + self.result_table_print()
> + self.vm.stop()
> +
> + def test_perf_pvp_qemu_modern_vector_rx_mac(self):
> + """
> + performance for frame_sizes and restart port on virtio1.0 vector rx
> + """
> + self.start_vhost_testpmd()
> + self.start_one_vm(modem=1, mergeable=0)
> + self.start_vm_testpmd(path="vector_rx")
> + self.send_and_verify("virtio1.0 vector_rx")
> + self.close_all_testpmd()
> + self.result_table_print()
> + self.vm.stop()
> +
> + def tear_down(self):
> + """
> + Run after each test case.
> + """
> + self.dut.send_expect("killall -s INT testpmd", "#")
> + self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
> + self.close_session()
> + time.sleep(2)
> +
> + def tear_down_all(self):
> + """
> + Run after each test suite.
> + """
> --
> 2.7.4
More information about the dts
mailing list