Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/121348/?format=api
https://patches.dpdk.org/api/patches/121348/?format=api", "web_url": "https://patches.dpdk.org/project/dts/patch/20221223073545.756712-1-weix.ling@intel.com/", "project": { "id": 3, "url": "https://patches.dpdk.org/api/projects/3/?format=api", "name": "DTS", "link_name": "dts", "list_id": "dts.dpdk.org", "list_email": "dts@dpdk.org", "web_url": "", "scm_url": "git://dpdk.org/tools/dts", "webscm_url": "http://git.dpdk.org/tools/dts/", "list_archive_url": "https://inbox.dpdk.org/dts", "list_archive_url_format": "https://inbox.dpdk.org/dts/{}", "commit_url_format": "" }, "msgid": "<20221223073545.756712-1-weix.ling@intel.com>", "list_archive_url": "https://inbox.dpdk.org/dts/20221223073545.756712-1-weix.ling@intel.com", "date": "2022-12-23T07:35:45", "name": "[V1] optimization pvp_qemu_multi_paths_port_restart testplan and testsuite", "commit_ref": null, "pull_url": null, "state": "superseded", "archived": false, "hash": "4813932d05df090ccff787b27b371a8cd7d10a4e", "submitter": { "id": 1828, "url": "https://patches.dpdk.org/api/people/1828/?format=api", "name": "Ling, WeiX", "email": "weix.ling@intel.com" }, "delegate": null, "mbox": "https://patches.dpdk.org/project/dts/patch/20221223073545.756712-1-weix.ling@intel.com/mbox/", "series": [ { "id": 26258, "url": "https://patches.dpdk.org/api/series/26258/?format=api", "web_url": "https://patches.dpdk.org/project/dts/list/?series=26258", "date": "2022-12-23T07:35:45", "name": "[V1] optimization pvp_qemu_multi_paths_port_restart testplan and testsuite", "version": 1, "mbox": "https://patches.dpdk.org/series/26258/mbox/" } ], "comments": "https://patches.dpdk.org/api/patches/121348/comments/", "check": "fail", "checks": "https://patches.dpdk.org/api/patches/121348/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "<dts-bounces@dpdk.org>", "X-Original-To": "patchwork@inbox.dpdk.org", "Delivered-To": "patchwork@inbox.dpdk.org", "Received": [ "from mails.dpdk.org (mails.dpdk.org [217.70.189.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id 3E443A0093;\n\tFri, 23 Dec 2022 08:44:35 +0100 (CET)", "from mails.dpdk.org (localhost [127.0.0.1])\n\tby mails.dpdk.org (Postfix) with ESMTP id 145EE40698;\n\tFri, 23 Dec 2022 08:44:35 +0100 (CET)", "from mga12.intel.com (mga12.intel.com [192.55.52.136])\n by mails.dpdk.org (Postfix) with ESMTP id 757A240141\n for <dts@dpdk.org>; Fri, 23 Dec 2022 08:44:32 +0100 (CET)", "from fmsmga008.fm.intel.com ([10.253.24.58])\n by fmsmga106.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 22 Dec 2022 23:44:31 -0800", "from unknown (HELO localhost.localdomain) ([10.239.252.222])\n by fmsmga008-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 22 Dec 2022 23:44:30 -0800" ], "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/simple;\n d=intel.com; i=@intel.com; q=dns/txt; s=Intel;\n t=1671781472; x=1703317472;\n h=from:to:cc:subject:date:message-id:mime-version:\n content-transfer-encoding;\n bh=V10zQFntW27lq+J4tGw2glY0tJgOpYk1TtunWN6XBME=;\n b=Bp0vjQqdxl7FaQbZOERaqlAvh+8HARZwLtWGoVgc2GmvEbkjs+/32nW7\n IM3F4ushYs/8TUBRZY5rgBssoLqpMojbq2PkGpyhgKTOEydt0OAWRk/pS\n tjjv1vLDDRrrWAVUeu/3VIkHPUG0kAx51c7qL+GL8uSj20m2hi7Z4gzDi\n ZNBl2uPCBzAXz0k4xLQWRsD4fl8GfIqde8m/+lm3IHTHzona2/Lyr3Pn5\n aMcq5g93tYl9gxaJ6qsQSv/0ldhRNdfcHk/FTh80I5trfz6kYb4ChLHc/\n XQoXQ73WzF+t0cf42HZ0/z5PfvCyYldQhOAMP9z/4ICemIG+XHbwrcROS g==;", "X-IronPort-AV": [ "E=McAfee;i=\"6500,9779,10569\"; a=\"299965197\"", "E=Sophos;i=\"5.96,267,1665471600\"; d=\"scan'208\";a=\"299965197\"", "E=McAfee;i=\"6500,9779,10569\"; a=\"715430324\"", "E=Sophos;i=\"5.96,267,1665471600\"; d=\"scan'208\";a=\"715430324\"" ], "From": "Wei Ling <weix.ling@intel.com>", "To": "dts@dpdk.org", "Cc": "Wei Ling <weix.ling@intel.com>", "Subject": "[dts][PATCH V1] optimization pvp_qemu_multi_paths_port_restart\n testplan and testsuite", "Date": "Fri, 23 Dec 2022 15:35:45 +0800", "Message-Id": "<20221223073545.756712-1-weix.ling@intel.com>", "X-Mailer": "git-send-email 2.25.1", "MIME-Version": "1.0", "Content-Transfer-Encoding": "8bit", "X-BeenThere": "dts@dpdk.org", "X-Mailman-Version": "2.1.29", "Precedence": "list", "List-Id": "test suite reviews and discussions <dts.dpdk.org>", "List-Unsubscribe": "<https://mails.dpdk.org/options/dts>,\n <mailto:dts-request@dpdk.org?subject=unsubscribe>", "List-Archive": "<http://mails.dpdk.org/archives/dts/>", "List-Post": "<mailto:dts@dpdk.org>", "List-Help": "<mailto:dts-request@dpdk.org?subject=help>", "List-Subscribe": "<https://mails.dpdk.org/listinfo/dts>,\n <mailto:dts-request@dpdk.org?subject=subscribe>", "Errors-To": "dts-bounces@dpdk.org" }, "content": "1.Add `disable-modern=false` parameter in vitio0.95 testcases.\n2.Add `-a 0000:af:00.0` in start vhost-user testpmd.\n3.Add `-a 0000:04:00.0,vectorized=1` in virtio0.95 and virtio1.0\nvector_rx path case.\n\nSigned-off-by: Wei Ling <weix.ling@intel.com>\n---\n ...emu_multi_paths_port_restart_test_plan.rst | 108 +++++++++---------\n ...Suite_pvp_qemu_multi_paths_port_restart.py | 4 +-\n 2 files changed, 56 insertions(+), 56 deletions(-)", "diff": "diff --git a/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst b/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst\nindex 017ea5f0..a621738d 100644\n--- a/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst\n+++ b/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst\n@@ -19,27 +19,27 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG\n Test Case 1: pvp test with virtio 0.95 mergeable path\n =====================================================\n \n-1. Bind one port to vfio-pci, then launch testpmd by below command::\n+1. Bind 1 NIC port to vfio-pci, then launch testpmd by below command::\n \n rm -rf vhost-net*\n- ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 \\\n- --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \\\n- -i --nb-cores=1 --txd=1024 --rxd=1024\n+ ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 -a 0000:af:00.0 \\\n+ --vdev 'eth_vhost0,iface=vhost-net,queues=1' \\\n+ -- -i --nb-cores=1 --txd=1024 --rxd=1024\n testpmd>set fwd mac\n testpmd>start\n \n 2. Launch VM with mrg_rxbuf feature on::\n \n- qemu-system-x86_64 -name vm2 -enable-kvm -cpu host -smp 2 -m 4096 \\\n+ qemu-system-x86_64 -name vm0 -enable-kvm -cpu host -smp 2 -m 4096 \\\n -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on \\\n -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \\\n- -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \\\n- -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n- -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \\\n- -net user,hostfwd=tcp:127.0.0.1:6002-:22 \\\n+ -chardev socket,path=/tmp/vm0_qga0.sock,server,nowait,id=vm0_qga0 -device virtio-serial \\\n+ -device virtserialport,chardev=vm0_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n+ -monitor unix:/tmp/vm0_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \\\n+ -net user,hostfwd=tcp:127.0.0.1:6000-:22 \\\n -chardev socket,id=char0,path=./vhost-net \\\n -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \\\n- -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \\\n+ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=true,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \\\n -vnc :10\n \n 3. On VM, bind virtio net to vfio-pci and run testpmd::\n@@ -66,26 +66,26 @@ Test Case 1: pvp test with virtio 0.95 mergeable path\n Test Case 2: pvp test with virtio 0.95 normal path\n ==================================================\n \n-1. Bind one port to vfio-pci, then launch testpmd by below command::\n+1. Bind 1 NIC port to vfio-pci, then launch testpmd by below command::\n \n rm -rf vhost-net*\n- ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 \\\n- --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \\\n- -i --nb-cores=1 --txd=1024 --rxd=1024\n+ ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 -a 0000:af:00.0 \\\n+ --vdev 'eth_vhost0,iface=vhost-net,queues=1' \\\n+ -- -i --nb-cores=1 --txd=1024 --rxd=1024\n testpmd>set fwd mac\n testpmd>start\n \n 2. Launch VM with mrg_rxbuf feature off::\n \n- qemu-system-x86_64 -name vm2 -enable-kvm -cpu host -smp 2 -m 4096 \\\n+ qemu-system-x86_64 -name vm0 -enable-kvm -cpu host -smp 2 -m 4096 \\\n -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on \\\n -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \\\n- -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \\\n- -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n- -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \\\n+ -chardev socket,path=/tmp/vm0_qga0.sock,server,nowait,id=vm0_qga0 -device virtio-serial \\\n+ -device virtserialport,chardev=vm0_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n+ -monitor unix:/tmp/vm0_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6000-:22 \\\n -chardev socket,id=char0,path=./vhost-net \\\n -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \\\n- -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \\\n+ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=true,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \\\n -vnc :10\n \n 3. On VM, bind virtio net to vfio-pci and run testpmd with tx-offloads::\n@@ -112,31 +112,31 @@ Test Case 2: pvp test with virtio 0.95 normal path\n Test Case 3: pvp test with virtio 0.95 vrctor_rx path\n =====================================================\n \n-1. Bind one port to vfio-pci, then launch testpmd by below command::\n+1. Bind 1 NIC port to vfio-pci, then launch testpmd by below command::\n \n rm -rf vhost-net*\n- ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 \\\n- --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \\\n- -i --nb-cores=1 --txd=1024 --rxd=1024\n+ ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 -a 0000:af:00.0 \\\n+ --vdev 'eth_vhost0,iface=vhost-net,queues=1' \\\n+ -- -i --nb-cores=1 --txd=1024 --rxd=1024\n testpmd>set fwd mac\n testpmd>start\n \n 2. Launch VM with mrg_rxbuf feature off::\n \n- qemu-system-x86_64 -name vm2 -enable-kvm -cpu host -smp 2 -m 4096 \\\n+ qemu-system-x86_64 -name vm0 -enable-kvm -cpu host -smp 2 -m 4096 \\\n -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on \\\n -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \\\n- -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \\\n- -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n- -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \\\n+ -chardev socket,path=/tmp/vm0_qga0.sock,server,nowait,id=vm0_qga0 -device virtio-serial \\\n+ -device virtserialport,chardev=vm0_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n+ -monitor unix:/tmp/vm0_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6000-:22 \\\n -chardev socket,id=char0,path=./vhost-net \\\n -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \\\n- -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \\\n+ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=true,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \\\n -vnc :10\n \n 3. On VM, bind virtio net to vfio-pci and run testpmd without ant tx-offloads::\n \n- ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 3 -- -i \\\n+ ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 3 -a 0000:04:00.0,vectorized=1 -- -i \\\n --nb-cores=1 --txd=1024 --rxd=1024\n testpmd>set fwd mac\n testpmd>start\n@@ -158,23 +158,23 @@ Test Case 3: pvp test with virtio 0.95 vrctor_rx path\n Test Case 4: pvp test with virtio 1.0 mergeable path\n ====================================================\n \n-1. Bind one port to vfio-pci, then launch testpmd by below command::\n+1. Bind 1 NIC port to vfio-pci, then launch testpmd by below command::\n \n rm -rf vhost-net*\n- ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 \\\n- --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \\\n- -i --nb-cores=1 --txd=1024 --rxd=1024\n+ ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 -a 0000:af:00.0 \\\n+ --vdev 'eth_vhost0,iface=vhost-net,queues=1' \\\n+ -- -i --nb-cores=1 --txd=1024 --rxd=1024\n testpmd>set fwd mac\n testpmd>start\n \n 2. Launch VM with 1 virtio, note: we need add \"disable-modern=false\" to enable virtio 1.0::\n \n- qemu-system-x86_64 -name vm2 -enable-kvm -cpu host -smp 2 -m 4096 \\\n+ qemu-system-x86_64 -name vm0 -enable-kvm -cpu host -smp 2 -m 4096 \\\n -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on \\\n -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \\\n- -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \\\n- -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n- -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \\\n+ -chardev socket,path=/tmp/vm0_qga0.sock,server,nowait,id=vm0_qga0 -device virtio-serial \\\n+ -device virtserialport,chardev=vm0_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n+ -monitor unix:/tmp/vm0_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6000-:22 \\\n -chardev socket,id=char0,path=./vhost-net \\\n -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \\\n -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \\\n@@ -204,23 +204,23 @@ Test Case 4: pvp test with virtio 1.0 mergeable path\n Test Case 5: pvp test with virtio 1.0 normal path\n =================================================\n \n-1. Bind one port to vfio-pci, then launch testpmd by below command::\n+1. Bind 1 NIC port to vfio-pci, then launch testpmd by below command::\n \n rm -rf vhost-net*\n- ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 \\\n- --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \\\n- -i --nb-cores=1 --txd=1024 --rxd=1024\n+ ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 -a 0000:af:00.0 \\\n+ --vdev 'eth_vhost0,iface=vhost-net,queues=1' \\\n+ -- -i --nb-cores=1 --txd=1024 --rxd=1024\n testpmd>set fwd mac\n testpmd>start\n \n 2. Launch VM with 1 virtio, note: we need add \"disable-modern=false\" to enable virtio 1.0::\n \n- qemu-system-x86_64 -name vm2 -enable-kvm -cpu host -smp 2 -m 4096 \\\n+ qemu-system-x86_64 -name vm0 -enable-kvm -cpu host -smp 2 -m 4096 \\\n -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on \\\n -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \\\n- -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \\\n- -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n- -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \\\n+ -chardev socket,path=/tmp/vm0_qga0.sock,server,nowait,id=vm0_qga0 -device virtio-serial \\\n+ -device virtserialport,chardev=vm0_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n+ -monitor unix:/tmp/vm0_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6000-:22 \\\n -chardev socket,id=char0,path=./vhost-net \\\n -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \\\n -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \\\n@@ -250,23 +250,23 @@ Test Case 5: pvp test with virtio 1.0 normal path\n Test Case 6: pvp test with virtio 1.0 vrctor_rx path\n ====================================================\n \n-1. Bind one port to vfio-pci, then launch testpmd by below command::\n+1. Bind 1 NIC port to vfio-pci, then launch testpmd by below command::\n \n rm -rf vhost-net*\n- ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 \\\n- --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \\\n- -i --nb-cores=1 --txd=1024 --rxd=1024\n+ ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0xe -n 4 -a 0000:af:00.0 \\\n+ --vdev 'eth_vhost0,iface=vhost-net,queues=1' \\\n+ -- -i --nb-cores=1 --txd=1024 --rxd=1024\n testpmd>set fwd mac\n testpmd>start\n \n 2. Launch VM with 1 virtio, note: we need add \"disable-modern=false\" to enable virtio 1.0::\n \n- qemu-system-x86_64 -name vm2 -enable-kvm -cpu host -smp 2 -m 4096 \\\n+ qemu-system-x86_64 -name vm0 -enable-kvm -cpu host -smp 2 -m 4096 \\\n -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on \\\n -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \\\n- -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \\\n- -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n- -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \\\n+ -chardev socket,path=/tmp/vm0_qga0.sock,server,nowait,id=vm0_qga0 -device virtio-serial \\\n+ -device virtserialport,chardev=vm0_qga0,name=org.qemu.guest_agent.2 -daemonize \\\n+ -monitor unix:/tmp/vm0_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6000-:22 \\\n -chardev socket,id=char0,path=./vhost-net \\\n -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \\\n -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \\\n@@ -274,7 +274,7 @@ Test Case 6: pvp test with virtio 1.0 vrctor_rx path\n \n 3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads::\n \n- ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 3 -- -i \\\n+ ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 3 -a 0000:04:00.0,vectorized=1 -- -i \\\n --nb-cores=1 --txd=1024 --rxd=1024\n testpmd>set fwd mac\n testpmd>start\ndiff --git a/tests/TestSuite_pvp_qemu_multi_paths_port_restart.py b/tests/TestSuite_pvp_qemu_multi_paths_port_restart.py\nindex 2b753eb1..9ae83dfe 100644\n--- a/tests/TestSuite_pvp_qemu_multi_paths_port_restart.py\n+++ b/tests/TestSuite_pvp_qemu_multi_paths_port_restart.py\n@@ -101,8 +101,8 @@ class TestPVPQemuMultiPathPortRestart(TestCase):\n )\n elif path == \"vector_rx\":\n command = (\n- self.path + \"-c 0x3 -n 3 -- -i \" + \"--nb-cores=1 --txd=1024 --rxd=1024\"\n- )\n+ self.path + \"-c 0x3 -n 3 -a %s,vectorized=1 -- -i \" + \"--nb-cores=1 --txd=1024 --rxd=1024\"\n+ ) % self.vm_dut.get_port_pci(0)\n self.vm_dut.send_expect(command, \"testpmd> \", 30)\n self.vm_dut.send_expect(\"set fwd mac\", \"testpmd> \", 30)\n self.vm_dut.send_expect(\"start\", \"testpmd> \", 30)\n", "prefixes": [ "V1" ] }{ "id": 121348, "url": "