[V3,3/3] tests/vhost_user_interrupt_cbdma: add new testsuite

Message ID 20221122065251.2894138-1-weix.ling@intel.com (mailing list archive)
State Accepted
Headers
Series add vhost_user_interrupt_cbdma |

Checks

Context Check Description
ci/Intel-dts-format-test success Testing OK
ci/Intel-dts-pylama-test success Testing OK
ci/Intel-dts-doc-test success Testing OK
ci/Intel-dts-suite-test warning SKIPPED

Commit Message

Ling, WeiX Nov. 22, 2022, 6:52 a.m. UTC
  Add new vhost_user_interrupt_cbdma testsuite to test the virtio 
enqueue and dequeue use l3fwd-power with split ring and packed
ring path and CBDMA.

Signed-off-by: Wei Ling <weix.ling@intel.com>
---
 tests/TestSuite_vhost_user_interrupt_cbdma.py | 275 ++++++++++++++++++
 1 file changed, 275 insertions(+)
 create mode 100644 tests/TestSuite_vhost_user_interrupt_cbdma.py
  

Comments

Tu, Lijuan Nov. 30, 2022, 6:48 a.m. UTC | #1
On Tue, 22 Nov 2022 14:52:51 +0800, Wei Ling <weix.ling@intel.com> wrote:
> Add new vhost_user_interrupt_cbdma testsuite to test the virtio 
> enqueue and dequeue use l3fwd-power with split ring and packed
> ring path and CBDMA.
> 
> Signed-off-by: Wei Ling <weix.ling@intel.com>


Series applied, thanks
  

Patch

diff --git a/tests/TestSuite_vhost_user_interrupt_cbdma.py b/tests/TestSuite_vhost_user_interrupt_cbdma.py
new file mode 100644
index 00000000..ce95e94b
--- /dev/null
+++ b/tests/TestSuite_vhost_user_interrupt_cbdma.py
@@ -0,0 +1,275 @@ 
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(c) 2022 Intel Corporation
+#
+
+"""
+DPDK Test suite.
+Vhost enqueue interrupt need test with l3fwd-power sample
+"""
+
+import re
+import time
+
+import framework.utils as utils
+from framework.test_case import TestCase
+
+
+class TestVhostUserInterruptCbdma(TestCase):
+    def set_up_all(self):
+        """
+        Run at the start of each test suite.
+
+        """
+        self.queues = 1
+        self.cores_num = len([n for n in self.dut.cores if int(n["socket"]) == 0])
+        self.vmac = "00:11:22:33:44:10"
+        self.pci_info = self.dut.ports_info[0]["pci"]
+        self.prepare_l3fwd_power()
+        self.app_l3fwd_power_path = self.dut.apps_name["l3fwd-power"]
+        self.app_testpmd_path = self.dut.apps_name["test-pmd"]
+        self.testpmd_name = self.app_testpmd_path.split("/")[-1]
+        self.l3fwdpower_name = self.app_l3fwd_power_path.split("/")[-1]
+
+        self.dut_ports = self.dut.get_ports()
+        self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
+        # get cbdma device
+        self.cbdma_dev_infos = []
+        self.dmas_info = None
+        self.device_str = None
+
+    def set_up(self):
+        """
+        Run before each test case.
+        """
+        # Clean the execution ENV
+        self.verify_info = []
+        self.dut.send_expect("killall -s INT %s" % self.testpmd_name, "#")
+        self.dut.send_expect("killall %s" % self.l3fwdpower_name, "#")
+        self.dut.send_expect("rm -rf ./vhost-net*", "#")
+        self.vhost = self.dut.new_session(suite="vhost-l3fwd")
+        self.virtio_user = self.dut.new_session(suite="virtio-user")
+
+    def prepare_l3fwd_power(self):
+        out = self.dut.build_dpdk_apps("examples/l3fwd-power")
+        self.verify("Error" not in out, "compilation l3fwd-power error")
+
+    def get_core_list(self):
+        """
+        get core list depend on the core number
+        """
+        need_num = 2 * self.queues + 1
+        self.core_config = "1S/%dC/1T" % need_num
+        self.verify(
+            self.cores_num >= need_num, "There has not enought cores to test this case"
+        )
+        core_list = self.dut.get_core_list(self.core_config)
+        self.core_list_virtio = core_list[0 : self.queues + 1]
+        self.core_list_l3fwd = core_list[self.queues + 1 : need_num]
+
+    def lanuch_virtio_user(self, packed=False):
+        """
+        launch virtio-user with server mode
+        """
+        vdev = (
+            "net_virtio_user0,mac=%s,path=./vhost-net,server=1,queues=%d"
+            % (self.vmac, self.queues)
+            if not packed
+            else "net_virtio_user0,mac=%s,path=./vhost-net,server=1,queues=%d,packed_vq=1"
+            % (self.vmac, self.queues)
+        )
+        eal_params = self.dut.create_eal_parameters(
+            cores=self.core_list_virtio, prefix="virtio", no_pci=True, vdevs=[vdev]
+        )
+
+        if self.check_2M_env:
+            eal_params += " --single-file-segments"
+        para = " -- -i --rxq=%d --txq=%d --rss-ip" % (self.queues, self.queues)
+        command_line_client = self.app_testpmd_path + " " + eal_params + para
+        self.virtio_user.send_expect(
+            command_line_client, "waiting for client connection...", 120
+        )
+
+    def get_cbdma_ports_info_and_bind_to_dpdk(self, cbdma_num):
+        """
+        get all cbdma ports
+        """
+        out = self.dut.send_expect(
+            "./usertools/dpdk-devbind.py --status-dev dma", "# ", 30
+        )
+        device_info = out.split("\n")
+        for device in device_info:
+            pci_info = re.search("\s*(0000:\S*:\d*.\d*)", device)
+            if pci_info is not None:
+                dev_info = pci_info.group(1)
+                # the numa id of ioat dev, only add the device which
+                # on same socket with nic dev
+                bus = int(dev_info[5:7], base=16)
+                if bus >= 128:
+                    cur_socket = 1
+                else:
+                    cur_socket = 0
+                if self.ports_socket == cur_socket:
+                    self.cbdma_dev_infos.append(pci_info.group(1))
+        self.verify(
+            len(self.cbdma_dev_infos) >= cbdma_num,
+            "There no enough cbdma device to run this suite",
+        )
+        used_cbdma = self.cbdma_dev_infos[0:cbdma_num]
+        tx_dmas_info = ""
+        for dmas in used_cbdma:
+            number = used_cbdma.index(dmas)
+            dmas = "txq{}@{};".format(number, dmas)
+            tx_dmas_info += dmas
+        rx_dmas_info = ""
+        for dmas in used_cbdma:
+            number = used_cbdma.index(dmas)
+            dmas = "rxq{}@{};".format(number, dmas)
+            rx_dmas_info += dmas
+        dmas_info = tx_dmas_info + rx_dmas_info
+        self.dmas_info = dmas_info[:-1]
+        self.device_str = " ".join(used_cbdma)
+        self.dut.send_expect(
+            "./usertools/dpdk-devbind.py --force --bind=%s %s"
+            % (self.drivername, self.device_str),
+            "# ",
+            60,
+        )
+
+    def bind_cbdma_device_to_kernel(self):
+        if self.device_str is not None:
+            self.dut.send_expect("modprobe ioatdma", "# ")
+            self.dut.send_expect(
+                "./usertools/dpdk-devbind.py -u %s" % self.device_str, "# ", 30
+            )
+            self.dut.send_expect(
+                "./usertools/dpdk-devbind.py --force --bind=ioatdma  %s"
+                % self.device_str,
+                "# ",
+                60,
+            )
+
+    @property
+    def check_2M_env(self):
+        out = self.dut.send_expect(
+            "cat /proc/meminfo |grep Hugepagesize|awk '{print($2)}'", "# "
+        )
+        return True if out == "2048" else False
+
+    def lanuch_l3fwd_power(self):
+        """
+        launch l3fwd-power with a virtual vhost device
+        """
+        self.logger.info("Launch l3fwd_sample sample:")
+        # config the interrupt cores
+        config_info = ""
+        for i in range(self.queues):
+            if config_info != "":
+                config_info += ","
+            config_info += "(0,%d,%s)" % (i, self.core_list_l3fwd[i])
+            info = {"core": self.core_list_l3fwd[i], "port": 0, "queue": i}
+            self.verify_info.append(info)
+
+        example_cmd = self.app_l3fwd_power_path + " "
+        example_cmd += " --log-level=9 "
+        self.get_cbdma_ports_info_and_bind_to_dpdk(4)
+        vdev = "'net_vhost0,iface=vhost-net,queues=%d,client=1,dmas=[%s]'" % (
+            self.queues,
+            self.dmas_info,
+        )
+        eal_params = self.dut.create_eal_parameters(
+            cores=self.core_list_l3fwd,
+            ports=self.cbdma_dev_infos[0:4],
+            vdevs=[vdev],
+        )
+        para = " -- -p 0x1 --parse-ptype 1 --config '%s' --interrupt-only" % config_info
+        command_line_client = example_cmd + eal_params + para
+        self.vhost.get_session_before(timeout=2)
+        self.vhost.send_expect(command_line_client, "POWER", 40)
+        time.sleep(10)
+        out = self.vhost.get_session_before()
+        if "Error" in out and "Error opening" not in out:
+            self.logger.error("Launch l3fwd-power sample error")
+        else:
+            self.logger.info("Launch l3fwd-power sample finished")
+
+    def check_vhost_core_status(self, status):
+        """
+        check the cpu status
+        """
+        out = self.vhost.get_session_before()
+        for i in range(len(self.verify_info)):
+            if status == "waked up":
+                info = "lcore %s is waked up from rx interrupt on port %d queue %d"
+                info = info % (
+                    self.verify_info[i]["core"],
+                    self.verify_info[i]["port"],
+                    self.verify_info[i]["queue"],
+                )
+            elif status == "sleeps":
+                info = (
+                    "lcore %s sleeps until interrupt triggers"
+                    % self.verify_info[i]["core"]
+                )
+            self.verify(info in out, "The CPU status not right for %s" % info)
+            self.logger.info(info)
+
+    def send_and_verify(self):
+        """
+        start to send packets and check the cpu status
+        stop and restart to send packets and check the cpu status
+        """
+        self.virtio_user.send_expect("start", "testpmd> ", 20)
+        self.check_vhost_core_status("waked up")
+
+        self.virtio_user.send_expect("stop", "testpmd> ", 20)
+        self.check_vhost_core_status("sleeps")
+
+        self.virtio_user.send_expect("start", "testpmd> ", 20)
+        self.check_vhost_core_status("waked up")
+
+    def close_testpmd_and_session(self):
+        self.virtio_user.send_expect("quit", "#", 20)
+        self.dut.close_session(self.vhost)
+        self.dut.close_session(self.virtio_user)
+
+    def test_wake_up_split_ring_vhost_user_core_with_l3fwd_power_sample_when_multi_queues_enabled_and_cbdma_enabled(
+        self,
+    ):
+        """
+        Test Case1: Wake up split ring vhost-user cores with l3fwd-power sample when multi queues and cbdma are enabled
+        """
+        self.queues = 4
+        self.get_core_list()
+        self.lanuch_virtio_user(packed=False)
+        self.lanuch_l3fwd_power()
+        self.virtio_user.send_expect("set fwd txonly", "testpmd> ", 20)
+        self.send_and_verify()
+
+    def test_wake_up_packed_ring_vhost_user_core_with_l3fwd_power_sample_when_multi_queues_enabled_and_cbdma_enabled(
+        self,
+    ):
+        """
+        Test Case2: Wake up packed ring vhost-user cores with l3fwd-power sample when multi queues and cbdma are enabled
+        """
+        self.queues = 4
+        self.get_core_list()
+        self.lanuch_virtio_user(packed=True)
+        self.lanuch_l3fwd_power()
+        self.virtio_user.send_expect("set fwd txonly", "testpmd> ", 20)
+        self.send_and_verify()
+
+    def tear_down(self):
+        """
+        Run after each test case.
+        """
+        self.close_testpmd_and_session()
+        self.dut.send_expect("killall %s" % self.l3fwdpower_name, "#")
+        self.dut.send_expect("killall -s INT %s" % self.testpmd_name, "#")
+        self.dut.kill_all()
+        self.bind_cbdma_device_to_kernel()
+
+    def tear_down_all(self):
+        """
+        Run after each test suite.
+        """
+        pass