[dts][PATCH V2 2/3]tests: add i40e_multi_driver_perf test script
Tu, Lijuan
lijuan.tu at intel.com
Tue Apr 11 10:16:41 CEST 2023
> -----Original Message-----
> From: Yingya Han <yingyax.han at intel.com>
> Sent: Tuesday, April 4, 2023 11:15 AM
> To: dts at dpdk.org
> Cc: Han, YingyaX <yingyax.han at intel.com>
> Subject: [dts][PATCH V2 2/3]tests: add i40e_multi_driver_perf test script
>
> Signed-off-by: Yingya Han <yingyax.han at intel.com>
> ---
> tests/TestSuite_i40e_multi_driver_perf.py | 341 ++++++++++++++++++++++
> 1 file changed, 341 insertions(+)
> create mode 100644 tests/TestSuite_i40e_multi_driver_perf.py
>
> diff --git a/tests/TestSuite_i40e_multi_driver_perf.py
> b/tests/TestSuite_i40e_multi_driver_perf.py
> new file mode 100644
> index 00000000..a1be9053
> --- /dev/null
> +++ b/tests/TestSuite_i40e_multi_driver_perf.py
> @@ -0,0 +1,341 @@
> +# SPDX-License-Identifier: BSD-3-Clause # Copyright(c) 2023 Intel
> +Corporation #
> +
> +"""
> +DPDK Test suite.
> +"""
> +
> +import os
> +import re
> +import time
> +from copy import deepcopy
> +
> +from framework.exception import VerifyFailure from framework.packet
> +import Packet from framework.pktgen import TRANSMIT_CONT from
> +framework.pmd_output import PmdOutput from framework.settings import
> +HEADER_SIZE from framework.test_case import TestCase from
> +framework.utils import convert_int2ip, convert_ip2int
> +
> +
> +class TestI40eMultiDriverPerf(TestCase):
> + #
> + # Test cases.
> + #
> + def set_up_all(self):
> + """
> + Run at the start of each test suite.
> + """
> + self.verify(
> + self.nic in ["I40E_25G-25G_SFP28", "I40E_40G-QSFP_A"],
> + "NIC Unsupported: " + str(self.nic),
> + )
I think the suite also support other 710 series cards.
And checklist or supportlist should be added.
> + self.rx_desc_size = self.get_suite_cfg().get("compile_rx_desc", 32)
> + err_msg = (
> + "Rx desc only has 16B and 32B size, %d is not valid" % self.rx_desc_size
> + )
> + self.verify(self.rx_desc_size == 16 or self.rx_desc_size == 32, err_msg)
> + if self.rx_desc_size == 16:
> + extra_options = "-Dc_args=-DRTE_LIBRTE_I40E_16BYTE_RX_DESC"
> + self.dut.build_install_dpdk(self.target, extra_options=extra_options)
I don't see the requirement in test plan, "set RX descriptor to 16", Is it needed ?
> + self.dut_ports = self.dut.get_ports(self.nic)
> + self.verify(len(self.dut_ports) >= 1, "At least 1 port is required to test")
> + # get socket and cores
> + self.socket = self.dut.get_numa_id(self.dut_ports[0])
> + cores = self.dut.get_core_list("1S/8C/1T", socket=self.socket)
> + self.verify(cores, "Requested 8 cores failed")
> + self.pmdout = PmdOutput(self.dut)
> + self.core_offset = 3
> + self.test_content =
> +self.get_test_content_from_cfg(self.get_suite_cfg())
> +
> + def set_up(self):
> + """
> + Run before each test case.
> + """
> + self.test_result = {"header": [], "data": []}
> +
> + def flows(self):
> + """
> + Return a list of packets that implements the flows described.
> + """
> + return [
> + "198.18.0.0/24",
> + "198.18.1.0/24",
> + "198.18.2.0/24",
> + "198.18.3.0/24",
> + "198.18.4.0/24",
> + "198.18.5.0/24",
> + "198.18.6.0/24",
> + "198.18.7.0/24",
> + ]
> +
> + def parse_test_config(self, config):
> + """
> + [n]C/[mT]-[i]Q
> + n: how many physical core use for polling.
> + m: how many cpu thread use for polling, if Hyper-threading disabled
> + in BIOS, m equals n, if enabled, m is 2 times as n.
> + i: how many queues use per port, so total queues = i x nb_port
> + """
> + pat = "(.*)/(.*)-(.*)"
> + result = re.findall(pat, config)
> + if not result:
> + msg = f"{config} is wrong format, please check"
> + raise VerifyFailure(msg)
> + cores, threads, queue = result[0]
> + _thread_num = int(int(threads[:-1]) // int(cores[:-1]))
> +
> + _thread = str(_thread_num) + "T"
> + _cores = str(self.core_offset + int(cores[:-1])) + "C"
> + cores_config = "/".join(["1S", _cores, _thread])
> + queues_per_port = int(queue[:-1])
> + return cores_config, _thread_num, queues_per_port
> +
> + def get_test_configs(self, test_parameters):
> + configs = []
> + frame_sizes_grp = []
> + nb_desc = self.get_suite_cfg().get("rxtx_queue_size")
> + for test_item, frame_sizes in sorted(test_parameters.items()):
> + _frame_sizes = [int(frame_size) for frame_size in frame_sizes]
> + frame_sizes_grp.extend([int(item) for item in _frame_sizes])
> + cores, thread_num, queues = self.parse_test_config(test_item)
> + corelist = self.dut.get_core_list(cores, self.socket)
> + core_list = corelist[(self.core_offset - 1) * thread_num :]
> + if "2T" in cores:
> + core_list = core_list[1:2] + core_list[0::2] + core_list[1::2][1:]
> + _core_list = core_list[thread_num - 1 :]
> + configs.append(
> + [
> + test_item,
> + _core_list,
> + [
> + " --txd={0} --rxd={0}".format(nb_desc)
> + + " --rxq={0} --txq={0}".format(queues)
> + + " --nb-cores={}".format(len(core_list) - thread_num)
> + ],
> + ]
> + )
> + return configs, sorted(set(frame_sizes_grp))
> +
> + def get_test_content_from_cfg(self, test_content):
> + test_content["flows"] = self.flows()
> + configs, frame_sizes =
> self.get_test_configs(test_content["test_parameters"])
> + test_content["configs"] = configs
> + test_content["frame_sizes"] = frame_sizes
> + return test_content
> +
> + def get_mac_layer(self, port_id):
> + smac = "02:00:00:00:00:0%d" % port_id
> + dmac = "52:00:00:00:00:0%d" % port_id
> + layer = {
> + "ether": {
> + "dst": dmac,
> + "src": smac,
> + },
> + }
> + return layer
> +
> + def get_ipv4_config(self, config):
> + netaddr, mask = config.split("/")
> + ip_range = int("1" * (32 - int(mask)), 2)
> + start_ip = convert_int2ip(convert_ip2int(netaddr) + 1)
> + end_ip = convert_int2ip(convert_ip2int(start_ip) + ip_range - 1)
> + layers = {
> + "ipv4": {
> + "src": start_ip,
> + },
> + }
> + fields_config = {
> + "ip": {
> + "src": {
> + "start": start_ip,
> + "end": end_ip,
> + "step": 1,
> + "action": "random",
> + },
> + },
> + }
> + return layers, fields_config
> +
> + def preset_flows_configs(self):
> + flows = self.test_content.get("flows")
> + flows_configs = []
> + for index, config in enumerate(flows):
> + if index >= len(self.dut_ports):
> + break
> + port_id = self.dut_ports[index]
> + _layer = self.get_mac_layer(port_id)
> + _layer2, fields_config = self.get_ipv4_config(config)
> + _layer.update(_layer2)
> + flows_configs.append([_layer, fields_config])
> + return flows_configs
> +
> + def preset_streams(self):
> + frame_sizes = self.test_content.get("frame_sizes")
> + test_streams = {}
> + flows_configs = self.preset_flows_configs()
> + for frame_size in frame_sizes:
> + for flow_config in flows_configs:
> + _layers, fields_config = flow_config
> + pkt = self.config_stream(_layers, frame_size)
> + test_streams.setdefault(frame_size, []).append([pkt, fields_config])
> + return test_streams
> +
> + def config_stream(self, layers, frame_size):
> + """
> + Prepare traffic flow
> + """
> + headers_size = sum([HEADER_SIZE[x] for x in ["eth", "ip"]])
> + payload_size = frame_size - headers_size
> + # set streams for traffic
> + pkt_config = {
> + "type": "IP_RAW",
> + "pkt_layers": {"raw": {"payload": ["58"] * payload_size}},
> + }
> + pkt_config["pkt_layers"].update(layers)
> + pkt_type = pkt_config.get("type")
> + pkt_layers = pkt_config.get("pkt_layers")
> + pkt = Packet(pkt_type=pkt_type)
> + for layer in list(pkt_layers.keys()):
> + pkt.config_layer(layer, pkt_layers[layer])
> +
> + return pkt.pktgen.pkt
> +
> + def add_stream_to_pktgen(self, streams, option):
> + def port(index):
> + p = self.tester.get_local_port(self.dut_ports[index])
> + return p
> +
> + topos = (
> + [
> + [port(index), port(index - 1)]
> + if index % 2
> + else [port(index), port(index + 1)]
> + for index, _ in enumerate(self.dut_ports)
> + ]
> + if len(self.dut_ports) > 1
> + else [[port(0), port(0)]]
> + )
> + stream_ids = []
> + step = int(len(streams) / len(self.dut_ports))
> + for cnt, stream in enumerate(streams):
> + pkt, fields_config = stream
> + index = cnt // step
> + txport, rxport = topos[index]
> + _option = deepcopy(option)
> + _option["pcap"] = pkt
> + if fields_config:
> + _option["fields_config"] = fields_config
> + stream_id = self.tester.pktgen.add_stream(txport, rxport, pkt)
> + self.tester.pktgen.config_stream(stream_id, _option)
> + stream_ids.append(stream_id)
> + return stream_ids
> +
> + def start_testpmd(self, core_list, pci_para, eal):
> + self.pmdout.start_testpmd(core_list, eal, pci_para, socket=self.socket)
> + self.pmdout.wait_link_status_up("all", timeout=60)
> + self.dut.send_expect("start", "testpmd> ", 15)
> +
> + def throughput(self, frame_size):
> + streams = self.stream.get(frame_size)
> + # set traffic option
> + duration = self.test_content.get("test_duration")
> + traffic_stop_wait_time = self.test_content.get("traffic_stop_wait_time", 0)
> + # clear streams before add new streams
> + self.tester.pktgen.clear_streams()
> + # ser stream into pktgen
> + stream_option = {
> + "stream_config": {
> + "txmode": {},
> + "transmit_mode": TRANSMIT_CONT,
> + "rate": 100,
> + }
> + }
> + traffic_option = {
> + "method": "throughput",
> + "duration": duration,
> + }
> + stream_ids = self.add_stream_to_pktgen(streams, stream_option)
> + # run packet generator
> + result = self.tester.pktgen.measure(stream_ids, traffic_option)
> + time.sleep(traffic_stop_wait_time)
> + # statistics result
> + _, pps = result
> + self.verify(pps > 0, "No traffic detected")
> + self.logger.info(
> + "Throughput of "
> + + "framesize: {}, is: {} Mpps".format(frame_size, pps / 1000000)
> + )
> + return result
> +
> + def display_result(self, datas):
> + # display result table
> + header_row = ["Fwd Core", "Frame Size", "Throughput", "Rate"]
> + self.test_result["header"] = header_row
> + self.result_table_create(header_row)
> + self.test_result["data"] = []
> + for data in datas:
> + config, frame_size, result = data
> + _, pps = result
> + pps /= 1000000.0
> + linerate = self.wirespeed(self.nic, frame_size, len(self.dut_ports))
> + percentage = pps * 100 / linerate
> + data_row = [
> + config,
> + frame_size,
> + "{:.3f} Mpps".format(pps),
> + "{:.3f}%".format(percentage),
> + ]
> + self.result_table_add(data_row)
> + self.test_result["data"].append(data_row)
> + self.result_table_print()
> +
> + def perf_test(self):
> + """
> + Benchmarking test
> + """
> + self.stream = self.preset_streams()
> + # ports allow list
> + pci_para = ""
> + for port_id in self.dut_ports:
> + pci_para += (
> + "-a "
> + + self.dut.ports_info[port_id]["port"].pci
> + + ",support-multi-driver=1 "
> + )
> + results = []
> + for config, core_list, eal in self.test_content["configs"]:
> + self.logger.info(
> + ("Executing Test Using cores: {0} of config {1}, ").format(
> + core_list, config
> + )
> + )
> + self.start_testpmd(core_list, pci_para, eal[0])
> + for frame_size in self.test_content["frame_sizes"]:
> + self.logger.info("Test running at framesize: {}".format(frame_size))
> + result = self.throughput(frame_size)
> + if result:
> + results.append([config, frame_size, result])
> + self.dut.send_expect("stop", "testpmd> ", 15)
> + self.dut.send_expect("quit", "# ", 15)
> + self.display_result(results)
> +
> + def test_perf_enable_multi_driver(self):
> + self.perf_test()
> +
> + def tear_down(self):
> + """
> + Run after each test case.
> + """
> + pass
> +
> + def tear_down_all(self):
> + """
> + Run after each test suite.
> + """
> + # resume setting
> + if self.rx_desc_size == 16:
> + self.dut.build_install_dpdk(self.target)
> + self.dut.kill_all()
> --
> 2.34.1
More information about the dts
mailing list