Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/95343/?format=api
https://patches.dpdk.org/api/patches/95343/?format=api", "web_url": "https://patches.dpdk.org/project/dts/patch/20210706181653.4660-1-leweix.yang@intel.com/", "project": { "id": 3, "url": "https://patches.dpdk.org/api/projects/3/?format=api", "name": "DTS", "link_name": "dts", "list_id": "dts.dpdk.org", "list_email": "dts@dpdk.org", "web_url": "", "scm_url": "git://dpdk.org/tools/dts", "webscm_url": "http://git.dpdk.org/tools/dts/", "list_archive_url": "https://inbox.dpdk.org/dts", "list_archive_url_format": "https://inbox.dpdk.org/dts/{}", "commit_url_format": "" }, "msgid": "<20210706181653.4660-1-leweix.yang@intel.com>", "list_archive_url": "https://inbox.dpdk.org/dts/20210706181653.4660-1-leweix.yang@intel.com", "date": "2021-07-06T18:16:53", "name": "[V1] tests/TestSuite_vm2vm_virtio_net_perf:change case", "commit_ref": null, "pull_url": null, "state": "accepted", "archived": false, "hash": "6e6590cd8cc6df0caeabc9d0cd134e09e44e6143", "submitter": { "id": 2225, "url": "https://patches.dpdk.org/api/people/2225/?format=api", "name": "Lewei Yang", "email": "leweix.yang@intel.com" }, "delegate": null, "mbox": "https://patches.dpdk.org/project/dts/patch/20210706181653.4660-1-leweix.yang@intel.com/mbox/", "series": [ { "id": 17657, "url": "https://patches.dpdk.org/api/series/17657/?format=api", "web_url": "https://patches.dpdk.org/project/dts/list/?series=17657", "date": "2021-07-06T18:16:53", "name": "[V1] tests/TestSuite_vm2vm_virtio_net_perf:change case", "version": 1, "mbox": "https://patches.dpdk.org/series/17657/mbox/" } ], "comments": "https://patches.dpdk.org/api/patches/95343/comments/", "check": "pending", "checks": "https://patches.dpdk.org/api/patches/95343/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "<dts-bounces@dpdk.org>", "X-Original-To": "patchwork@inbox.dpdk.org", "Delivered-To": "patchwork@inbox.dpdk.org", "Received": [ "from mails.dpdk.org (mails.dpdk.org [217.70.189.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id 394DCA0C47;\n\tTue, 6 Jul 2021 11:42:45 +0200 (CEST)", "from [217.70.189.124] (localhost [127.0.0.1])\n\tby mails.dpdk.org (Postfix) with ESMTP id 31DAA4120E;\n\tTue, 6 Jul 2021 11:42:44 +0200 (CEST)", "from mga02.intel.com (mga02.intel.com [134.134.136.20])\n by mails.dpdk.org (Postfix) with ESMTP id BF31740688\n for <dts@dpdk.org>; Tue, 6 Jul 2021 11:42:42 +0200 (CEST)", "from fmsmga008.fm.intel.com ([10.253.24.58])\n by orsmga101.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 06 Jul 2021 02:42:40 -0700", "from unknown (HELO localhost.localdomain) ([10.240.183.102])\n by fmsmga008-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 06 Jul 2021 02:42:38 -0700" ], "X-IronPort-AV": [ "E=McAfee;i=\"6200,9189,10036\"; a=\"196252470\"", "E=Sophos;i=\"5.83,328,1616482800\"; d=\"scan'208\";a=\"196252470\"", "E=Sophos;i=\"5.83,328,1616482800\"; d=\"scan'208\";a=\"460634584\"" ], "From": "Yang Lewei <leweix.yang@intel.com>", "To": "dts@dpdk.org", "Cc": "YangLewei <leweix.yang@intel.com>", "Date": "Tue, 6 Jul 2021 18:16:53 +0000", "Message-Id": "<20210706181653.4660-1-leweix.yang@intel.com>", "X-Mailer": "git-send-email 2.17.1", "Subject": "[dts] [PATCH V1] tests/TestSuite_vm2vm_virtio_net_perf:change case", "X-BeenThere": "dts@dpdk.org", "X-Mailman-Version": "2.1.29", "Precedence": "list", "List-Id": "test suite reviews and discussions <dts.dpdk.org>", "List-Unsubscribe": "<https://mails.dpdk.org/options/dts>,\n <mailto:dts-request@dpdk.org?subject=unsubscribe>", "List-Archive": "<http://mails.dpdk.org/archives/dts/>", "List-Post": "<mailto:dts@dpdk.org>", "List-Help": "<mailto:dts-request@dpdk.org?subject=help>", "List-Subscribe": "<https://mails.dpdk.org/listinfo/dts>,\n <mailto:dts-request@dpdk.org?subject=subscribe>", "Errors-To": "dts-bounces@dpdk.org", "Sender": "\"dts\" <dts-bounces@dpdk.org>" }, "content": "From: YangLewei <leweix.yang@intel.com>\n\nupdate for cases with imix pkts for better coverage\n\nSigned-off-by: YangLewei <leweix.yang@intel.com>\n---\n tests/TestSuite_vm2vm_virtio_net_perf.py | 83 ++++++++----------------\n 1 file changed, 26 insertions(+), 57 deletions(-)", "diff": "diff --git a/tests/TestSuite_vm2vm_virtio_net_perf.py b/tests/TestSuite_vm2vm_virtio_net_perf.py\nindex d1e94093..0ec542af 100644\n--- a/tests/TestSuite_vm2vm_virtio_net_perf.py\n+++ b/tests/TestSuite_vm2vm_virtio_net_perf.py\n@@ -225,11 +225,11 @@ class TestVM2VMVirtioNetPerf(TestCase):\n \n # add -f g param, use Gbits/sec report teste result\n if iperf_mode == \"tso\":\n- iperf_server = \"iperf -f g -s -i 1\"\n- iperf_client = \"iperf -f g -c 1.1.1.2 -i 1 -t 60\"\n+ iperf_server = \"iperf -s -i 1\"\n+ iperf_client = \"iperf -c 1.1.1.2 -i 1 -t 60\"\n elif iperf_mode == 'ufo':\n- iperf_server = \"iperf -f g -s -u -i 1\"\n- iperf_client = \"iperf -f g -c 1.1.1.2 -i 1 -t 30 -P 4 -u -b 1G -l 9000\"\n+ iperf_server = \"iperf -s -u -i 1\"\n+ iperf_client = \"iperf -c 1.1.1.2 -i 1 -t 30 -P 4 -u -b 1G -l 9000\"\n self.vm_dut[0].send_expect(\"%s > iperf_server.log &\" % iperf_server, \"\", 10)\n self.vm_dut[1].send_expect(\"%s > iperf_client.log &\" % iperf_client, \"\", 60)\n time.sleep(90)\n@@ -252,6 +252,7 @@ class TestVM2VMVirtioNetPerf(TestCase):\n iperfdata = re.compile('\\S*\\s*[M|G]bits/sec').findall(fmsg)\n # the last data of iperf is the ave data from 0-30 sec\n self.verify(len(iperfdata) != 0, \"The iperf data between to vms is 0\")\n+ self.verify((iperfdata[-1]).split()[1] == \"Gbits/sec\", \"The iperf data is %s,Can't reach Gbits/sec\" % iperfdata[-1])\n self.logger.info(\"The iperf data between vms is %s\" % iperfdata[-1])\n \n # put the result to table\n@@ -357,7 +358,7 @@ class TestVM2VMVirtioNetPerf(TestCase):\n \"\"\"\n TestCase1: VM2VM split ring vhost-user/virtio-net test with tcp traffic\n \"\"\"\n- self.vm_args = \"disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on\"\n+ self.vm_args = \"disable-modern=false,mrg_rxbuf=off,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on\"\n self.prepare_test_env(cbdma=False, no_pci=True, client_mode=False, enable_queues=1, nb_cores=2,\n server_mode=False, opt_queue=1, combined=False, rxq_txq=None)\n self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')\n@@ -521,35 +522,19 @@ class TestVM2VMVirtioNetPerf(TestCase):\n \n self.logger.info(\"Launch vhost-testpmd with CBDMA and used 8 queue\")\n self.vm_args = \"disable-modern=false,mrg_rxbuf=on,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on,packed=on\"\n- self.prepare_test_env(cbdma=True, no_pci=False, client_mode=True, enable_queues=8, nb_cores=4,\n- server_mode=True, opt_queue=8, combined=True, rxq_txq=8)\n- self.check_scp_file_valid_between_vms()\n- iperf_data_cbdma_enable_8_queue = self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')\n- ipef_result.append(['Enable', 'mergeable path', 8, iperf_data_cbdma_enable_8_queue])\n-\n- self.logger.info(\"Re-launch without CBDMA and used 8 queue\")\n- self.vhost.send_expect(\"quit\", \"# \", 30)\n- self.start_vhost_testpmd(cbdma=False, no_pci=False, client_mode=True, enable_queues=8, nb_cores=4, rxq_txq=8)\n- self.check_scp_file_valid_between_vms()\n- iperf_data_cbdma_disable_8_queue = self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')\n- ipef_result.append(['Disable', 'mergeable path', 8, iperf_data_cbdma_disable_8_queue])\n-\n- self.logger.info(\"Re-launch without CBDMA and used 1 queue\")\n- self.vhost.send_expect(\"quit\", \"# \", 30)\n- self.start_vhost_testpmd(cbdma=False, no_pci=False, client_mode=True, enable_queues=8, nb_cores=4, rxq_txq=1)\n- self.config_vm_env(combined=True, rxq_txq=1)\n- self.check_scp_file_valid_between_vms()\n- iperf_data_cbdma_disable_1_queue = self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')\n- ipef_result.append(['Disable', 'mergeable path', 1, iperf_data_cbdma_disable_1_queue])\n-\n- self.table_header = ['CBDMA Enable/Disable', 'Mode', 'rxq/txq', 'Gbits/sec']\n- self.result_table_create(self.table_header)\n+ self.prepare_test_env(cbdma=True, no_pci=False, client_mode=False, enable_queues=8, nb_cores=4,\n+ server_mode=False, opt_queue=8, combined=True, rxq_txq=8)\n+ for i in range(0,5):\n+ self.check_scp_file_valid_between_vms()\n+ iperf_data_cbdma_enable_8_queue = self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')\n+ ipef_result.append(['Enable_%d' % i, 'mergeable path', 8, iperf_data_cbdma_enable_8_queue])\n+ if i > 0:\n+ self.verify(abs(ipef_result[0][3]-ipef_result[i][3])/ipef_result[0][3] < 0.2 ,\"Performance fluctuates too much\")\n+ self.table_header = ['CBDMA Enable/Disable', 'Mode', 'rxq/txq', 'Gbits/sec']\n+ self.result_table_create(self.table_header)\n for table_row in ipef_result:\n self.result_table_add(table_row)\n self.result_table_print()\n- self.verify(iperf_data_cbdma_enable_8_queue > iperf_data_cbdma_disable_8_queue, \\\n- \"CMDMA enable: %s is lower than CBDMA disable: %s\" % (\n- iperf_data_cbdma_enable_8_queue, iperf_data_cbdma_disable_8_queue))\n \n def test_vm2vm_packed_ring_with_no_mergeable_path_check_large_packet_and_cbdma_enable_8queue(self):\n \"\"\"\n@@ -561,35 +546,19 @@ class TestVM2VMVirtioNetPerf(TestCase):\n \n self.logger.info(\"Launch vhost-testpmd with CBDMA and used 8 queue\")\n self.vm_args = \"disable-modern=false,mrg_rxbuf=off,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on,packed=on\"\n- self.prepare_test_env(cbdma=True, no_pci=False, client_mode=True, enable_queues=8, nb_cores=4,\n- server_mode=True, opt_queue=8, combined=True, rxq_txq=8)\n- self.check_scp_file_valid_between_vms()\n- iperf_data_cbdma_enable_8_queue = self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')\n- ipef_result.append(['Enable', 'mergeable path', 8, iperf_data_cbdma_enable_8_queue])\n-\n- self.logger.info(\"Re-launch without CBDMA and used 8 queue\")\n- self.vhost.send_expect(\"quit\", \"# \", 30)\n- self.start_vhost_testpmd(cbdma=False, no_pci=False, client_mode=True, enable_queues=8, nb_cores=4, rxq_txq=8)\n- self.check_scp_file_valid_between_vms()\n- iperf_data_cbdma_disable_8_queue = self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')\n- ipef_result.append(['Disable', 'mergeable path', 8, iperf_data_cbdma_disable_8_queue])\n-\n- self.logger.info(\"Re-launch without CBDMA and used 1 queue\")\n- self.vhost.send_expect(\"quit\", \"# \", 30)\n- self.start_vhost_testpmd(cbdma=False, no_pci=False, client_mode=True, enable_queues=8, nb_cores=4, rxq_txq=1)\n- self.config_vm_env(combined=True, rxq_txq=1)\n- self.check_scp_file_valid_between_vms()\n- iperf_data_cbdma_disable_1_queue = self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')\n- ipef_result.append(['Disable', 'mergeable path', 1, iperf_data_cbdma_disable_1_queue])\n-\n- self.table_header = ['CBDMA Enable/Disable', 'Mode', 'rxq/txq', 'Gbits/sec']\n- self.result_table_create(self.table_header)\n+ self.prepare_test_env(cbdma=True, no_pci=False, client_mode=False, enable_queues=8, nb_cores=4,\n+ server_mode=False, opt_queue=8, combined=True, rxq_txq=8)\n+ for i in range(0,5):\n+ self.check_scp_file_valid_between_vms()\n+ iperf_data_cbdma_enable_8_queue = self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')\n+ ipef_result.append(['Enable', 'mergeable path', 8, iperf_data_cbdma_enable_8_queue])\n+ if i > 0:\n+ self.verify(abs(ipef_result[0][3]-ipef_result[i][3])/ipef_result[0][3] < 0.2 ,\"Performance fluctuates too much\")\n+ self.table_header = ['CBDMA Enable/Disable', 'Mode', 'rxq/txq', 'Gbits/sec']\n+ self.result_table_create(self.table_header)\n for table_row in ipef_result:\n self.result_table_add(table_row)\n self.result_table_print()\n- self.verify(iperf_data_cbdma_enable_8_queue > iperf_data_cbdma_disable_8_queue, \\\n- \"CMDMA enable: %s is lower than CBDMA disable: %s\" % (\n- iperf_data_cbdma_enable_8_queue, iperf_data_cbdma_disable_8_queue))\n \n def tear_down(self):\n \"\"\"\n", "prefixes": [ "V1" ] }{ "id": 95343, "url": "