[dts] [PATCH]tests: discard TestSuite_ip_pipeline.py
Peng Yuan
yuan.peng at intel.com
Mon Jan 28 07:20:46 CET 2019
Discard the old automation script TestSuite_ip_pipeline.py,
and will submit a new TestSuite_ip_pipeline.py
Signed-off-by: Peng Yuan <yuan.peng at intel.com>
diff --git a/tests/TestSuite_ip_pipeline.py b/tests/TestSuite_ip_pipeline.py
deleted file mode 100644
index 76a477f..0000000
--- a/tests/TestSuite_ip_pipeline.py
+++ /dev/null
@@ -1,581 +0,0 @@
-# BSD LICENSE
-#
-# Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
-# All rights reserved.
-#
-# Redistribution and use in source and binary forms, with or without
-# modification, are permitted provided that the following conditions
-# are met:
-#
-# * Redistributions of source code must retain the above copyright
-# notice, this list of conditions and the following disclaimer.
-# * Redistributions in binary form must reproduce the above copyright
-# notice, this list of conditions and the following disclaimer in
-# the documentation and/or other materials provided with the
-# distribution.
-# * Neither the name of Intel Corporation nor the names of its
-# contributors may be used to endorse or promote products derived
-# from this software without specific prior written permission.
-#
-# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-"""
-DPDK Test suite.
-Test userland 10Gb PMD
-"""
-
-from scapy.layers.inet import Ether, IP, TCP
-from scapy.utils import struct, socket, PcapWriter
-from settings import HEADER_SIZE
-from test_case import TestCase
-from time import sleep
-import utils
-
-
-class TestIPPipeline(TestCase):
- payload_watermark = 'TestPF'
-
- frame_sizes = [64, 65, 128, 1024]
- """Sizes of the frames to be sent"""
-
- number_of_frames = [1, 3, 63, 64, 65, 127, 128]
- """Number of frames in the pcap file to be created"""
-
- incremental_ip_address = [True, False]
- """True if the IP address is incremented in the frames"""
-
- inter = [0, 0.7]
- """Interval between frames sent in seconds"""
-
- dummy_pcap = 'dummy.pcap'
-
- def increment_ip_addr(self, ip_address, increment):
-
- ip2int = lambda ipstr: struct.unpack('!I', socket.inet_aton(ipstr))[0]
- x = ip2int(ip_address)
- int2ip = lambda n: socket.inet_ntoa(struct.pack('!I', n))
- return int2ip(x + increment)
-
- def create_tcp_ipv4_frame(
- self, ip_id, src_ip_addr, dst_ip_addr, frame_size,
- src_mac_addr='00:00:0A:00:0B:00',
- dst_mac_addr='00:00:0A:00:0A:00'):
-
- payload_size = frame_size - HEADER_SIZE['eth'] - HEADER_SIZE['ip'] -\
- HEADER_SIZE['tcp'] - \
- len(TestIPPipeline.payload_watermark)
-
- if payload_size < 0:
- payload_size = 0
-
- frame = Ether() / IP() / TCP(flags="") / (TestIPPipeline.payload_watermark +
- "X" * payload_size)
- frame[Ether].src = src_mac_addr
- frame[Ether].dst = dst_mac_addr
-
- frame[IP].src = src_ip_addr
- frame[IP].dst = dst_ip_addr
- frame[IP].id = ip_id
-
- # TCP ports always 0
- frame[TCP].sport = 0
- frame[TCP].dport = 0
-
- return frame
-
- def create_pcap_file_from_frames(self, file_name, frames):
-
- writer = PcapWriter(file_name, append=False)
-
- for frame in frames:
- writer.write(frame)
-
- writer.close()
-
- def create_pcap_file(self, file_name, frame_size, number_of_frames,
- incremental_ip_address,
- src_ip="0.0.0.0",
- dst_ip="0.0.0.0"):
-
- current_frame = 0
- writer = PcapWriter(file_name, append=False)
-
- while current_frame < number_of_frames:
- ip_id = 0 # current_frame % 0x10000
-
- frame = self.create_tcp_ipv4_frame(ip_id, src_ip, dst_ip,
- frame_size)
- writer.write(frame)
-
- if incremental_ip_address:
- dst_ip = self.increment_ip_addr(dst_ip, 1)
-
- current_frame += 1
-
- writer.close()
-
- def create_passthrough_cfgfile(self):
- """
- Create configuration file for passthrough pipeline.
- Two ports are connected as follows: RXQ0.0 -> TXQ1.0, RXQ1.0 -> TXQ0.0.
- """
-
- self.dut.send_expect('echo [PIPELINE0] > /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo type = MASTER >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo core = 0 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo [PIPELINE1] >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo type = PASS-THROUGH >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo core = 1 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo pktq_in = RXQ0.0 RXQ1.0 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo pktq_out = TXQ1.0 TXQ0.0 >> /tmp/ip_pipeline.cfg', '#')
-
- def create_routing_cfgfile(self):
- """
- Create configuration file for ip routing pipeline.
- It is mainly to set ip header offset and arp key offset in the packet buffer.
- """
-
- self.dut.send_expect('echo [PIPELINE0] > /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo type = MASTER >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo core = 0 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo [PIPELINE1] >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo type = ROUTING >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo core = 1 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo pktq_in = RXQ0.0 RXQ1.0 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo pktq_out = TXQ0.0 TXQ1.0 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo encap = ethernet >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo ip_hdr_offset = 270 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo arp_key_offset = 128 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo n_arp_entries = 1000 >> /tmp/ip_pipeline.cfg', '#')
-
- def create_flow_cfgfile(self):
- """
- Create configuration file for flow classification pipeline.
- It is mainly to set key size, offset and mask to get the ipv4 5-tuple.
- """
-
- self.dut.send_expect('echo [PIPELINE0] > /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo type = MASTER >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo core = 0 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo [PIPELINE1] >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo type = FLOW_CLASSIFICATION >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo core = 1 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo pktq_in = RXQ0.0 RXQ1.0 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo pktq_out = TXQ0.0 TXQ1.0 SINK0 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo n_flows = 65536 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo key_size = 16 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo key_offset = 278 >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo key_mask = 00FF0000FFFFFFFFFFFFFFFFFFFFFFFF >> /tmp/ip_pipeline.cfg', '#')
- self.dut.send_expect('echo flowid_offset = 128 >> /tmp/ip_pipeline.cfg', '#')
-
- def start_ip_pipeline(self, ports):
- command_line = "./examples/ip_pipeline/build/ip_pipeline -p %s -f /tmp/ip_pipeline.cfg" % ports
-
- out = self.dut.send_expect(command_line, 'pipeline>', 60)
- sleep(5) # 'Initialization completed' is not the last output, some
- # seconds are still needed for init.
-
- self.verify("Aborted" not in out, "Error starting ip_pipeline")
- self.verify("PANIC" not in out, "Error starting ip_pipeline")
- self.verify("ERROR" not in out, "Error starting ip_pipeline")
-
- def quit_ip_pipeline(self):
- self.dut.send_expect("quit", "# ", 5)
-
- def tcpdump_start_sniffing(self, ifaces=[]):
- """
- Starts tcpdump in the background to sniff the tester interface where
- the packets are transmitted to and from the self.dut.
- All the captured packets are going to be stored in a file for a
- post-analysis.
- """
-
- for iface in ifaces:
- command = ('rm -f tcpdump_{0}.pcap').format(iface)
- self.tester.send_expect(command, '#')
- command = (
- 'tcpdump -w tcpdump_{0}.pcap -i {0} 2>tcpdump_{0}.out &').format(iface)
- self.tester.send_expect(command, '#')
-
- def tcpdump_stop_sniff(self):
- """
- Stops the tcpdump process running in the background.
- """
-
- self.tester.send_expect('killall tcpdump', '#')
- # For the [pid]+ Done tcpdump... message after killing the process
- sleep(1)
- self.tester.send_expect('echo "Cleaning buffer"', '#')
- sleep(1)
-
- def tcpdump_command(self, command, machine):
- """
- Sends a tcpdump related command and returns an integer from the output
- """
-
- if machine == 'dut':
- result = self.dut.send_expect(command, '#', alt_session=True)
- else:
- result = self.tester.send_expect(command, '#', alt_session=True)
-
- return int(result.strip())
-
- def number_of_packets(self, file_name, machine='tester'):
- """
- By reading the file generated by tcpdump it counts how many packets were
- forwarded by the sample app and received in the self.tester. The sample app
- will add a known MAC address for the test to look for.
- """
-
- command = ('tcpdump -A -nn -e -v -r %s 2>/dev/null | grep -c "%s"' %
- (file_name, TestIPPipeline.payload_watermark))
- return int(self.tcpdump_command(command, machine))
-
- def send_and_sniff_pcap_file(self, pcap_file, frames_number, from_port,
- to_port, inter=0):
- """
- Sent frames_number frames from the pcap_file with inter seconds of
- interval.
- Returns the number of received frames.
- """
-
- tx_port = self.tester.get_local_port(self.dut_ports[from_port])
- rx_port = self.tester.get_local_port(self.dut_ports[to_port])
- port0 = self.tester.get_local_port(self.dut_ports[0])
-
- tx_interface = self.tester.get_interface(tx_port)
- rx_interface = self.tester.get_interface(rx_port)
-
- self.tcpdump_start_sniffing([tx_interface, rx_interface])
-
- timeout = frames_number * inter + 2
- inter = ", inter=%d" % inter
-
- # Prepare the frames to be sent
- self.tester.scapy_foreground()
- self.tester.scapy_append('p = rdpcap("%s")' % (pcap_file))
- self.tester.scapy_append(
- 'sendp(p[:%s], iface="%s" %s)' % (frames_number,
- tx_interface,
- inter))
-
- # Execute scapy to sniff sniffing and send the frames
- self.tester.scapy_execute(timeout)
-
- self.tcpdump_stop_sniff()
-
- rx_stats = self.number_of_packets('tcpdump_%s.pcap' % rx_interface)
- tx_stats = self.number_of_packets('tcpdump_%s.pcap' % tx_interface)
-
- # Do not count the sent frames in the tx_interface
- tx_stats = tx_stats - frames_number
-
- if port0 == tx_port:
- return {'rx0': tx_stats, 'rx1': rx_stats}
- else:
- return {'rx0': rx_stats, 'rx1': tx_stats}
-
- def check_results(self, stats, expected):
- """
- This function check that the received packet numbers of port0 and port1 match the expected.
- expected = [Rx0, Rx1]
- """
-
- for port in ['rx0', 'rx1']:
- self.verify(stats[port] == expected[port],
- 'Frames expected (%s) and received (%s) mismatch on %s port' % (
- expected[port], stats[port], port))
-
- def pipeline_command(self, command):
- out = self.dut.send_expect(command, 'pipeline>')
- self.verify("arguments" not in out, "Incorrect arguments: '%s'" % command)
- self.verify("Invalid" not in out, "Invalid argument: '%s'" % command)
- self.verify("Syntax error" not in out, "Syntax error: '%s'" % command)
- return out
-
- def pipeline_add_flow(self, port, src_ip, dst_ip, src_port, dst_port, flowid,
- protocol=6):
- command = 'p 1 flow add ipv4 %s %s %d %d %d port %d id %d' % (src_ip, dst_ip, src_port,
- dst_port, protocol, port, flowid)
- out = self.pipeline_command(command)
- self.verify("failed" not in out, "Add flow error")
-
- def pipeline_del_flow(self, src_ip, dst_ip, src_port, dst_port,
- protocol=6):
- command = 'p 1 flow del ipv4 %s %s %d %d %d' % (src_ip, dst_ip, src_port,
- dst_port, protocol)
- out = self.pipeline_command(command)
- self.verify("failed" not in out, "Del flow error")
-
- def pipeline_add_route(self, port, src_ip, netmask, gw_ip):
- command = 'p 1 route add %s %d port %d ether %s' % (src_ip, netmask, port, gw_ip)
- out = self.pipeline_command(command)
- self.verify("failed" not in out, "Add route error")
-
- def pipeline_del_route(self, src_ip, netmask):
- command = 'p 1 route del %s %d' % (src_ip, netmask)
- out = self.pipeline_command(command)
- self.verify("failed" not in out, "Del route error")
-
- def set_up_all(self):
- """
- Run at the start of each test suite.
-
- PMD prerequisites.
- """
-
- self.dut_ports = self.dut.get_ports()
- self.verify(len(self.dut_ports) >= 2,
- "Insufficient ports for speed testing")
-
- out = self.dut.build_dpdk_apps("./examples/ip_pipeline")
- self.verify("Error" not in out, "Compilation error")
-
- self.ports_mask = utils.create_mask(
- [self.dut_ports[0], self.dut_ports[1]])
- self.coremask = "0x3e" # IP Pipeline app requires FIVE cores
-
- self.dut.setup_memory(4096)
-
- def set_up(self):
- """
- Run before each test case.
- """
- pass
-
- def test_incremental_ip(self):
- """
- Testing that frames with incremental IP addresses pass through the
- pipeline regardless the frames_number and the speed.
- """
- pcap_file = 'ip_pipeline.pcap'
- frame_size = 64
-
- self.create_passthrough_cfgfile()
- self.start_ip_pipeline(ports=self.ports_mask)
- self.dut.send_expect(
- 'run examples/ip_pipeline/config/ip_pipeline.sh', 'pipeline>', 10)
-
- # Create a PCAP file containing the maximum frames_number of frames needed
- # with fixed size and incremental IP
- self.create_pcap_file(pcap_file, frame_size,
- max(TestIPPipeline.number_of_frames), True)
- self.tester.session.copy_file_to(pcap_file)
-
- for frames_number in TestIPPipeline.number_of_frames:
- for inter in TestIPPipeline.inter:
- print utils.BLUE(
- "\tNumber of frames %d, interval %.1f" % (frames_number,
- inter))
- stats = self.send_and_sniff_pcap_file(pcap_file, frames_number,
- 1, 0, inter)
-
- expected = {'rx0': frames_number, 'rx1': 0}
- self.check_results(stats, expected)
-
- stats = self.send_and_sniff_pcap_file(pcap_file, frames_number,
- 0, 1, inter)
-
- expected = {'rx0': 0, 'rx1': frames_number}
- self.check_results(stats, expected)
-
- def test_frame_sizes(self):
- """
- Testing that frames with different sizes pass through the pipeline.
- """
- pcap_file = 'ip_pipeline.pcap'
- frames_number = 100
- inter = 0.5
-
- self.create_passthrough_cfgfile()
- self.start_ip_pipeline(ports=self.ports_mask)
- self.dut.send_expect(
- 'run examples/ip_pipeline/config/ip_pipeline.sh', 'pipeline>', 10)
-
- for frame_size in TestIPPipeline.frame_sizes:
-
- # Create a PCAP file containing the fixed number of frames above
- # with variable size and incremental IP
- self.create_pcap_file(pcap_file, frame_size, 100, True)
- self.tester.session.copy_file_to(pcap_file)
-
- print utils.BLUE("\tFrame size %d, interval %.1f" % (frame_size,
- inter))
-
- stats = self.send_and_sniff_pcap_file(pcap_file, frames_number,
- 1, 0, inter)
-
- expected = {'rx0': frames_number, 'rx1': 0}
- self.check_results(stats, expected)
-
- stats = self.send_and_sniff_pcap_file(pcap_file, frames_number,
- 0, 1, inter)
-
- expected = {'rx0': 0, 'rx1': frames_number}
- self.check_results(stats, expected)
-
- def test_flow_management(self):
- """
- Add several flows and check only frames with matching IPs passes
- """
- pcap_file = 'ip_pipeline.pcap'
- frame_size = 64
-
- ip_addrs = [
- '0.0.0.0', '0.0.0.1', '0.0.0.127', '0.0.0.128', '0.0.0.255',
- '0.0.1.0', '0.0.127.0', '0.0.128.0', '0.0.129.0', '0.0.255.0',
- '0.127.0.0', '0.127.1.0', '0.127.127.0', '0.127.255.0',
- '0.127.255.255']
-
- frames = []
-
- for addr in ip_addrs:
- frames.append(self.create_tcp_ipv4_frame(0, '0.0.0.0', addr,
- frame_size))
-
- self.create_flow_cfgfile()
- self.create_pcap_file_from_frames(pcap_file, frames)
- self.tester.session.copy_file_to(pcap_file)
-
- # Start ip_pipeline app and setup defaults
- self.start_ip_pipeline(ports=self.ports_mask)
-
- # default to SINK0
- self.pipeline_command('p 1 flow add default 2')
-
- # Check that no traffic pass though
- stats = self.send_and_sniff_pcap_file(pcap_file, len(frames),
- 1, 0, 0.2)
- expected = {'rx0': 0, 'rx1': 0}
- self.check_results(stats, expected)
-
- # Add the flows
- flows_added = 0
- for addrs in ip_addrs:
- self.pipeline_add_flow(0, '0.0.0.0', addrs, 0, 0, flows_added)
- flows_added += 1
-
- # Check that traffic matching flows pass though
- stats = self.send_and_sniff_pcap_file(pcap_file, len(frames),
- 1, 0, 0.2)
- expected = {'rx0': flows_added, 'rx1': 0}
- self.check_results(stats, expected)
-
- # Remove flows
- for addrs in ip_addrs:
- self.pipeline_del_flow('0.0.0.0', addrs, 0, 0)
- flows_added -= 1
-
- # Check that traffic matching flows pass though
- stats = self.send_and_sniff_pcap_file(pcap_file, len(frames),
- 1, 0, 0.2)
- expected = {'rx0': flows_added, 'rx1': 0}
- self.check_results(stats, expected)
-
- out = self.dut.send_expect('flow print', 'pipeline>')
- self.verify("=> Port =" not in out, "Flow found after deletion")
-
- # Check that again no traffic pass though
- stats = self.send_and_sniff_pcap_file(pcap_file, len(frames),
- 1, 0, 0.2)
- expected = {'rx0': 0, 'rx1': 0}
- self.check_results(stats, expected)
-
- self.quit_ip_pipeline()
-
- def test_route_management(self):
- """
- Add several flows and check only frames with matching IPs passes
- """
- pcap_file = 'ip_pipeline.pcap'
- frame_size = 64
-
- default_setup = ['p 1 arp add 0 0.0.0.1 0a:0b:0c:0d:0e:0f',
- 'p 1 arp add 1 0.128.0.1 1a:1b:1c:1d:1e:1f']
-
- ip_addrs = [
- '0.0.0.0', '0.0.0.1', '0.0.0.127', '0.0.0.128', '0.0.0.255',
- '0.0.1.0', '0.0.127.0', '0.0.128.0', '0.0.129.0', '0.0.255.0',
- '0.127.0.0', '0.127.1.0', '0.127.127.0', '0.127.255.0',
- '0.127.255.255']
-
- frames = []
-
- for addr in ip_addrs:
- frames.append(self.create_tcp_ipv4_frame(0, '0.0.0.0', addr,
- frame_size))
-
- self.create_pcap_file_from_frames(pcap_file, frames)
- self.tester.session.copy_file_to(pcap_file)
-
- self.create_routing_cfgfile()
- # Start ip_pipeline app and setup defaults
- self.start_ip_pipeline(ports=self.ports_mask)
- for command in default_setup:
- self.pipeline_command(command)
-
- # Check that no traffic pass though
- stats = self.send_and_sniff_pcap_file(pcap_file, len(frames),
- 1, 0, 0.2)
- expected = {'rx0': 0, 'rx1': 0}
- self.check_results(stats, expected)
-
- # Add the routes
- routes_added = 0
- for addr in ip_addrs:
- self.pipeline_add_route(0, addr, 32, '0.0.0.1')
- routes_added += 1
-
- # Check that traffic matching routes pass though
- stats = self.send_and_sniff_pcap_file(pcap_file, len(frames),
- 1, 0, 0.2)
-
- expected = {'rx0': routes_added, 'rx1': 0}
- self.check_results(stats, expected)
-
- # Remove routes
- for addr in ip_addrs:
- self.pipeline_del_route(addr, 32)
- routes_added -= 1
-
- # Check that traffic matching flows pass though
- stats = self.send_and_sniff_pcap_file(pcap_file, len(frames),
- 1, 0, 0.2)
- expected = {'rx0': routes_added, 'rx1': 0}
- self.check_results(stats, expected)
-
- out = self.dut.send_expect('route print', 'pipeline>')
- self.verify("Destination = " not in out, "Route found after deletion")
-
- # Check that again no traffic pass though
- stats = self.send_and_sniff_pcap_file(pcap_file, len(frames),
- 1, 0, 0.2)
- expected = {'rx0': 0, 'rx1': 0}
- self.check_results(stats, expected)
-
- self.quit_ip_pipeline()
-
- def tear_down(self):
- """
- Run after each test case.
- """
- self.quit_ip_pipeline()
-
- def tear_down_all(self):
- """
- Run after each test suite.
- """
- self.dut.send_expect("rm -f /tmp/ip_pipeline.cfg", "#")
- self.dut.send_expect("rm -f /tmp/ip_pipeline.cfg.out", "#")
- out = self.dut.build_dpdk_apps("./examples/ip_pipeline")
- self.verify("Error" not in out, "Compilation error")
--
2.5.0
More information about the dts
mailing list