Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/83370/?format=api
https://patches.dpdk.org/api/patches/83370/?format=api", "web_url": "https://patches.dpdk.org/project/dts/patch/20201102092120.9146-2-haiyangx.zhao@intel.com/", "project": { "id": 3, "url": "https://patches.dpdk.org/api/projects/3/?format=api", "name": "DTS", "link_name": "dts", "list_id": "dts.dpdk.org", "list_email": "dts@dpdk.org", "web_url": "", "scm_url": "git://dpdk.org/tools/dts", "webscm_url": "http://git.dpdk.org/tools/dts/", "list_archive_url": "https://inbox.dpdk.org/dts", "list_archive_url_format": "https://inbox.dpdk.org/dts/{}", "commit_url_format": "" }, "msgid": "<20201102092120.9146-2-haiyangx.zhao@intel.com>", "list_archive_url": "https://inbox.dpdk.org/dts/20201102092120.9146-2-haiyangx.zhao@intel.com", "date": "2020-11-02T09:21:13", "name": "[V3,1/8] tests/rte_flow_common: add a common module to process rss test", "commit_ref": null, "pull_url": null, "state": "changes-requested", "archived": false, "hash": "fccb15d3eeff63bf88f306b08ab06ec5f59a887b", "submitter": { "id": 1978, "url": "https://patches.dpdk.org/api/people/1978/?format=api", "name": "Zhao, HaiyangX", "email": "haiyangx.zhao@intel.com" }, "delegate": null, "mbox": "https://patches.dpdk.org/project/dts/patch/20201102092120.9146-2-haiyangx.zhao@intel.com/mbox/", "series": [ { "id": 13558, "url": "https://patches.dpdk.org/api/series/13558/?format=api", "web_url": "https://patches.dpdk.org/project/dts/list/?series=13558", "date": "2020-11-02T09:21:12", "name": "tests: update or add rss related suites", "version": 3, "mbox": "https://patches.dpdk.org/series/13558/mbox/" } ], "comments": "https://patches.dpdk.org/api/patches/83370/comments/", "check": "pending", "checks": "https://patches.dpdk.org/api/patches/83370/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "<dts-bounces@dpdk.org>", "X-Original-To": "patchwork@inbox.dpdk.org", "Delivered-To": "patchwork@inbox.dpdk.org", "Received": [ "from dpdk.org (dpdk.org [92.243.14.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id BBAD9A04E7;\n\tMon, 2 Nov 2020 10:22:42 +0100 (CET)", "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id CB463BBB4;\n\tMon, 2 Nov 2020 10:22:40 +0100 (CET)", "from mga07.intel.com (mga07.intel.com [134.134.136.100])\n by dpdk.org (Postfix) with ESMTP id ACF8A72ED\n for <dts@dpdk.org>; Mon, 2 Nov 2020 10:22:38 +0100 (CET)", "from fmsmga008.fm.intel.com ([10.253.24.58])\n by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 02 Nov 2020 01:22:37 -0800", "from unknown (HELO localhost.localdomain) ([10.240.183.55])\n by fmsmga008-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 02 Nov 2020 01:22:35 -0800" ], "IronPort-SDR": [ "\n l4WUTG3xKFlOGRdDU4/jn6uMla19GrkhQXyVWFbmouNATdwXypFSV6afMDuxuuFRFCPvubBM0S\n v2yz8uW3xQTw==", "\n yfBwgbfZg2VEubh6GITLlXfKleyU3G+VZRd+hsBmJENUey4n1i3elXBKLX35gKQr+G14cVlPMw\n 6ITyRQR2VYzg==" ], "X-IronPort-AV": [ "E=McAfee;i=\"6000,8403,9792\"; a=\"233027959\"", "E=Sophos;i=\"5.77,444,1596524400\"; d=\"scan'208\";a=\"233027959\"", "E=Sophos;i=\"5.77,444,1596524400\"; d=\"scan'208\";a=\"305386210\"" ], "X-Amp-Result": "SKIPPED(no attachment in message)", "X-Amp-File-Uploaded": "False", "From": "Haiyang Zhao <haiyangx.zhao@intel.com>", "To": "dts@dpdk.org,\n\tqi.fu@intel.com", "Cc": "Haiyang Zhao <haiyangx.zhao@intel.com>", "Date": "Mon, 2 Nov 2020 17:21:13 +0800", "Message-Id": "<20201102092120.9146-2-haiyangx.zhao@intel.com>", "X-Mailer": "git-send-email 2.17.1", "In-Reply-To": "<20201102092120.9146-1-haiyangx.zhao@intel.com>", "References": "<20201102092120.9146-1-haiyangx.zhao@intel.com>", "Subject": "[dts] [PATCH V3 1/8] tests/rte_flow_common: add a common module to\n\tprocess rss test", "X-BeenThere": "dts@dpdk.org", "X-Mailman-Version": "2.1.15", "Precedence": "list", "List-Id": "test suite reviews and discussions <dts.dpdk.org>", "List-Unsubscribe": "<https://mails.dpdk.org/options/dts>,\n <mailto:dts-request@dpdk.org?subject=unsubscribe>", "List-Archive": "<http://mails.dpdk.org/archives/dts/>", "List-Post": "<mailto:dts@dpdk.org>", "List-Help": "<mailto:dts-request@dpdk.org?subject=help>", "List-Subscribe": "<https://mails.dpdk.org/listinfo/dts>,\n <mailto:dts-request@dpdk.org?subject=subscribe>", "Errors-To": "dts-bounces@dpdk.org", "Sender": "\"dts\" <dts-bounces@dpdk.org>" }, "content": "*.add a class named RssProcessing to process rss tests.\n\nSigned-off-by: Haiyang Zhao <haiyangx.zhao@intel.com>\n---\n tests/rte_flow_common.py | 391 +++++++++++++++++++++++++++++++++++++++\n 1 file changed, 391 insertions(+)", "diff": "diff --git a/tests/rte_flow_common.py b/tests/rte_flow_common.py\nindex 8a4cded..43c0eab 100644\n--- a/tests/rte_flow_common.py\n+++ b/tests/rte_flow_common.py\n@@ -33,6 +33,7 @@ import json\n import time\n import re\n from utils import GREEN, RED\n+from packet import Packet\n \n CVL_TXQ_RXQ_NUMBER = 16\n \n@@ -672,3 +673,393 @@ def check_pf_rss_queue(out, count):\n return True\n else:\n return False\n+\n+\n+class RssProcessing(object):\n+ def __init__(self, test_case, pmd_output, tester_ifaces, rxq):\n+ self.test_case = test_case\n+ self.pmd_output = pmd_output\n+ self.tester_ifaces = tester_ifaces\n+ self.rxq = rxq\n+ self.logger = test_case.logger\n+ self.pkt = Packet()\n+ self.verify = self.test_case.verify\n+ self.pass_flag = 'passed'\n+ self.fail_flag = 'failed'\n+ self.current_saved_hash = ''\n+ self.hash_records = {}\n+ self.handle_output_methods = {\n+ 'save_hash': self.save_hash,\n+ 'save_or_no_hash': self.save_or_no_hash,\n+ 'check_hash_different': self.check_hash_different,\n+ 'check_no_hash_or_different': self.check_no_hash_or_different,\n+ 'check_hash_same': self.check_hash_same,\n+ 'check_no_hash': self.check_no_hash,\n+ }\n+ self.error_msgs = []\n+\n+ def save_hash(self, out, key='', port_id=0):\n+ hashes, rss_distribute = self.get_hash_verify_rss_distribute(out, port_id)\n+ if len(key) != 0:\n+ self.hash_records[key] = hashes\n+ self.current_saved_hash = hashes\n+ if not rss_distribute:\n+ error_msg = 'the packet do not distribute by rss'\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+\n+ def save_or_no_hash(self, out, key='', port_id=0):\n+ hashes, queues = self.get_hash_and_queues(out, port_id)\n+ if len(hashes) == 0:\n+ self.logger.info('There no hash value passed as expected')\n+ if set(queues) != {'0x0'}:\n+ error_msg = 'received queues should all be 0, but are {}'.format(queues)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ return\n+ if len(key) != 0:\n+ self.hash_records[key] = hashes\n+ self.current_saved_hash = hashes\n+ if not self.verify_rss_distribute(hashes, queues):\n+ error_msg = 'the packet do not distribute by rss'\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+\n+ def check_hash_different(self, out, key='', port_id=0):\n+ hashes, rss_distribute = self.get_hash_verify_rss_distribute(out, port_id)\n+ if len(key) == 0:\n+ if hashes == self.current_saved_hash:\n+ error_msg = 'hash value {} should be different ' \\\n+ 'with current saved hash {}'.format(hashes, self.current_saved_hash)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ else:\n+ if hashes == self.hash_records[key]:\n+ error_msg = 'hash value {} should be different ' \\\n+ 'with {} {}'.format(hashes, key, self.hash_records[key])\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ if not rss_distribute:\n+ error_msg = 'the packet do not distribute by rss'\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+\n+ def check_no_hash(self, out, port_id=0):\n+ hashes, queues = self.get_hash_and_queues(out, port_id)\n+ if len(hashes) != 0:\n+ error_msg = 'hash value {} should be empty'.format(hashes)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ elif set(queues) != {'0x0'}:\n+ error_msg = 'received queues should all be 0, but are {}'.format(queues)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+\n+ def check_no_hash_or_different(self, out, key='', port_id=0):\n+ hashes, queues = self.get_hash_and_queues(out, port_id)\n+ if len(hashes) == 0:\n+ self.logger.info('There no hash value passed as expected')\n+ if set(queues) != {'0x0'}:\n+ error_msg = 'received queues should all be 0, but are {}'.format(queues)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ return\n+ if len(key) == 0:\n+ if hashes == self.current_saved_hash:\n+ error_msg = 'hash value {} should be different ' \\\n+ 'with current saved hash {}'.format(hashes, self.current_saved_hash)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ else:\n+ if hashes == self.hash_records[key]:\n+ error_msg = 'hash value {} should be different ' \\\n+ 'with {} {}'.format(hashes, key, self.hash_records[key])\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+\n+ def check_hash_same(self, out, key='', port_id=0):\n+ hashes, rss_distribute = self.get_hash_verify_rss_distribute(out, port_id)\n+ if len(key) == 0:\n+ if hashes != self.current_saved_hash:\n+ error_msg = 'hash value {} should be same ' \\\n+ 'with current saved hash {}'.format(hashes, self.current_saved_hash)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ else:\n+ if hashes != self.hash_records[key]:\n+ error_msg = 'hash value {} should be same ' \\\n+ 'with {} {}'.format(hashes, key, self.hash_records[key])\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ if not rss_distribute:\n+ error_msg = 'the packet do not distribute by rss'\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+\n+ def check_hash_same_or_no_hash(self, out, key='', port_id=0):\n+ hashes, rss_distribute = self.get_hash_verify_rss_distribute(out, port_id)\n+ if len(hashes) != 0:\n+ error_msg = 'hash value {} should be empty'.format(hashes)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ return\n+ elif set(rss_distribute) != {'0x0'}:\n+ error_msg = 'received queues should all be 0, but are {}'.format(rss_distribute)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ return\n+ if len(key) == 0:\n+ if hashes != self.current_saved_hash:\n+ error_msg = 'hash value {} should be same ' \\\n+ 'with current saved hash {}'.format(hashes, self.current_saved_hash)\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ else:\n+ if hashes != self.hash_records[key]:\n+ error_msg = 'hash value {} should be same ' \\\n+ 'with {} {}'.format(hashes, key, self.hash_records[key])\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+ if not rss_distribute:\n+ error_msg = 'the packet do not distribute by rss'\n+ self.logger.error(error_msg)\n+ self.error_msgs.append(error_msg)\n+\n+ def verify_rss_distribute(self, hashes, queues):\n+ if len(hashes) != len(queues):\n+ self.logger.warning('hash length {} != queue length {}'.format(hashes, queues))\n+ return False\n+ for i in range(len(hashes)):\n+ if int(hashes[i], 16) % self.rxq != int(queues[i], 16):\n+ self.logger.warning('hash values {} mod total queues {} != queue {}'\n+ .format(hashes[i], self.rxq, queues[i]))\n+ return False\n+ return True\n+\n+ def get_hash_verify_rss_distribute(self, out, port_id=0):\n+ hashes, queues = self.get_hash_and_queues(out, port_id)\n+ if len(hashes) == 0:\n+ return [], False\n+ return hashes, self.verify_rss_distribute(hashes, queues)\n+\n+ def get_hash_and_queues(self, out, port_id=0):\n+ hash_pattern = re.compile('port\\s%s/queue\\s\\d+:\\sreceived\\s\\d+\\spackets.+?\\n.*RSS\\shash=(\\w+)\\s-\\sRSS\\squeue=(\\w+)' % port_id)\n+ hash_infos = hash_pattern.findall(out)\n+ self.logger.info('hash_infos: {}'.format(hash_infos))\n+ if len(hash_infos) == 0:\n+ queue_pattern = re.compile('Receive\\squeue=(\\w+)')\n+ queues = queue_pattern.findall(out)\n+ return [], queues\n+ # hashes = [int(hash_info[0], 16) for hash_info in hash_infos]\n+ hashes = [hash_info[0].strip() for hash_info in hash_infos]\n+ queues = [hash_info[1].strip() for hash_info in hash_infos]\n+ return hashes, queues\n+\n+ def send_pkt_get_output(self, pkts, port_id=0, count=1, interval=0):\n+ self.pkt.update_pkt(pkts)\n+ tx_port = self.tester_ifaces[0] if port_id == 0 else self.tester_ifaces[1]\n+ self.logger.info('----------send packet-------------')\n+ self.logger.info('{}'.format(pkts))\n+ self.pkt.send_pkt(crb=self.test_case.tester, tx_port=tx_port, count=count, interval=interval)\n+ out = self.pmd_output.get_output(timeout=1)\n+ pkt_pattern = 'port\\s%d/queue\\s\\d+:\\sreceived\\s(\\d+)\\spackets.+?\\n.*length=\\d{2,}\\s' % port_id\n+ reveived_data = re.findall(pkt_pattern, out)\n+ reveived_pkts = sum(map(int, [i[0] for i in reveived_data]))\n+ if isinstance(pkts, list):\n+ self.verify(reveived_pkts == len(pkts) * count,\n+ 'expect received %d pkts, but get %d instead' % (len(pkts) * count, reveived_pkts))\n+ else:\n+ self.verify(reveived_pkts == 1 * count,\n+ 'expect received %d pkts, but get %d instead' % (1 * count, reveived_pkts))\n+ return out\n+\n+ def send_pkt_get_hash_queues(self, pkts, port_id=0, count=1, interval=0):\n+ output = self.send_pkt_get_output(pkts, port_id, count, interval)\n+ hashes, queues = self.get_hash_and_queues(output, port_id)\n+ return hashes, queues\n+\n+ def create_rule(self, rule: (list, str), check_stats=True, msg=None):\n+ p = re.compile(r\"Flow rule #(\\d+) created\")\n+ rule_list = list()\n+ if isinstance(rule, list):\n+ for i in rule:\n+ out = self.pmd_output.execute_cmd(i, timeout=1)\n+ if msg:\n+ self.verify(msg in out, \"failed: expect %s in %s\" % (msg, out))\n+ m = p.search(out)\n+ if m:\n+ rule_list.append(m.group(1))\n+ else:\n+ rule_list.append(False)\n+ elif isinstance(rule, str):\n+ out = self.pmd_output.execute_cmd(rule, timeout=1)\n+ if msg:\n+ self.verify(msg in out, \"failed: expect %s in %s\" % (msg, out))\n+ m = p.search(out)\n+ if m:\n+ rule_list.append(m.group(1))\n+ else:\n+ rule_list.append(False)\n+ else:\n+ raise Exception(\"unsupported rule type, only accept list or str\")\n+ if check_stats:\n+ self.verify(all(rule_list), \"some rules create failed, result %s\" % rule_list)\n+ elif not check_stats:\n+ self.verify(not any(rule_list), \"all rules should create failed, result %s\" % rule_list)\n+ return rule_list\n+\n+ def validate_rule(self, rule, check_stats=True, check_msg=None):\n+ flag = 'Flow rule validated'\n+ if isinstance(rule, str):\n+ if 'create' in rule:\n+ rule = rule.replace('create', 'validate')\n+ out = self.pmd_output.execute_cmd(rule, timeout=1)\n+ if check_stats:\n+ self.verify(flag in out.strip(), \"rule %s validated failed, result %s\" % (rule, out))\n+ else:\n+ if check_msg:\n+ self.verify(flag not in out.strip() and check_msg in out.strip(),\n+ \"rule %s validate should failed with msg: %s, but result %s\" % (rule, check_msg, out))\n+ else:\n+ self.verify(flag not in out.strip(), \"rule %s validate should failed, result %s\" % (rule, out))\n+ elif isinstance(rule, list):\n+ for r in rule:\n+ if 'create' in r:\n+ r = r.replace('create', 'validate')\n+ out = self.pmd_output.execute_cmd(r, timeout=1)\n+ if check_stats:\n+ self.verify(flag in out.strip(), \"rule %s validated failed, result %s\" % (r, out))\n+ else:\n+ if not check_msg:\n+ self.verify(flag not in out.strip(), \"rule %s validate should failed, result %s\" % (r, out))\n+ else:\n+ self.verify(flag not in out.strip() and check_msg in out.strip(),\n+ \"rule %s should validate failed with msg: %s, but result %s\" % (\n+ r, check_msg, out))\n+\n+ def check_rule(self, port_id=0, stats=True, rule_list=None):\n+ out = self.pmd_output.execute_cmd(\"flow list %s\" % port_id)\n+ p = re.compile(r\"ID\\s+Group\\s+Prio\\s+Attr\\s+Rule\")\n+ matched = p.search(out)\n+ if stats:\n+ self.verify(matched, \"flow rule on port %s is not existed\" % port_id)\n+ if rule_list:\n+ p2 = re.compile(\"^(\\d+)\\s\")\n+ li = out.splitlines()\n+ res = list(filter(bool, list(map(p2.match, li))))\n+ result = [i.group(1) for i in res]\n+ self.verify(set(rule_list).issubset(set(result)),\n+ \"check rule list failed. expect %s, result %s\" % (rule_list, result))\n+ else:\n+ if matched:\n+ if rule_list:\n+ res_li = [i.split()[0].strip() for i in out.splitlines() if re.match('\\d', i)]\n+ self.verify(not set(rule_list).issubset(res_li), 'rule specified should not in result.')\n+ else:\n+ raise Exception('expect no rule listed')\n+ else:\n+ self.verify(not matched, \"flow rule on port %s is existed\" % port_id)\n+\n+ def destroy_rule(self, port_id=0, rule_id=None):\n+ if rule_id is None:\n+ rule_id = 0\n+ if isinstance(rule_id, list):\n+ for i in rule_id:\n+ out = self.test_case.dut.send_command(\"flow destroy %s rule %s\" % (port_id, i), timeout=1)\n+ p = re.compile(r\"Flow rule #(\\d+) destroyed\")\n+ m = p.search(out)\n+ self.verify(m, \"flow rule %s delete failed\" % rule_id)\n+ else:\n+ out = self.test_case.dut.send_command(\"flow destroy %s rule %s\" % (port_id, rule_id), timeout=1)\n+ p = re.compile(r\"Flow rule #(\\d+) destroyed\")\n+ m = p.search(out)\n+ self.verify(m, \"flow rule %s delete failed\" % rule_id)\n+\n+ def handle_actions(self, output, actions, port_id=0):\n+ if isinstance(actions, dict) or isinstance(actions, str):\n+ actions = [actions]\n+ for action in actions: # [{}]\n+ self.logger.info('action: {}\\n'.format(action))\n+ if isinstance(action, str):\n+ if action in self.handle_output_methods:\n+ self.handle_output_methods[action](output, port_id=port_id)\n+ else:\n+ for method in action: # {'save': ''}\n+ if method in self.handle_output_methods:\n+ if method == 'check_no_hash':\n+ self.check_no_hash(output, port_id=port_id)\n+ else:\n+ self.handle_output_methods[method](output, action[method], port_id=port_id)\n+\n+ def handle_tests(self, tests, port_id=0):\n+ out = ''\n+ for test in tests:\n+ if 'send_packet' in test:\n+ out = self.send_pkt_get_output(test['send_packet'], port_id)\n+ if 'action' in test:\n+ self.handle_actions(out, test['action'])\n+\n+ def handle_rss_case(self, case_info):\n+ # clear hash_records before each sub case\n+ self.hash_records = {}\n+ self.error_msgs = []\n+ self.current_saved_hash = ''\n+ sub_case_name = case_info.get('sub_casename')\n+ self.logger.info('===================Test sub case: {}================'.format(sub_case_name))\n+ port_id = case_info.get('port_id') if case_info.get('port_id') else 0\n+ rules = case_info.get('rule') if case_info.get('rule') else []\n+ rule_ids = []\n+ if 'pre-test' in case_info:\n+ self.logger.info('------------handle pre-test--------------')\n+ self.handle_tests(case_info['pre-test'], port_id)\n+\n+ # handle tests\n+ tests = case_info['test']\n+ self.logger.info('------------handle test--------------')\n+ # validate rule\n+ if rules:\n+ self.validate_rule(rule=rules, check_stats=True)\n+ rule_ids = self.create_rule(rule=case_info['rule'], check_stats=True)\n+ self.check_rule(port_id=port_id, rule_list=rule_ids)\n+ self.handle_tests(tests, port_id)\n+\n+ # handle post-test\n+ if 'post-test' in case_info:\n+ self.logger.info('------------handle post-test--------------')\n+ self.destroy_rule(port_id=port_id, rule_id=rule_ids)\n+ self.check_rule(port_id=port_id, stats=False)\n+ self.handle_tests(case_info['post-test'], port_id)\n+ if self.error_msgs:\n+ self.verify(False, str(self.error_msgs[:500]))\n+\n+ def handle_rss_distribute_cases(self, cases_info):\n+ sub_cases_result = dict()\n+ if not isinstance(cases_info, list):\n+ cases_info = [cases_info]\n+\n+ for case_info in cases_info:\n+ try:\n+ # self.handle_rss_distribute_case(case_info=case_info)\n+ self.handle_rss_case(case_info=case_info)\n+ except Exception as e:\n+ self.logger.warning('sub_case %s failed: %s' % (case_info['sub_casename'], e))\n+ sub_cases_result[case_info['sub_casename']] = self.fail_flag\n+ else:\n+ self.logger.info('sub_case %s passed' % case_info['sub_casename'])\n+ sub_cases_result[case_info['sub_casename']] = self.pass_flag\n+ finally:\n+ self.pmd_output.execute_cmd('flow flush 0')\n+ pass_rate = round(list(sub_cases_result.values()).count(self.pass_flag) / len(sub_cases_result), 4) * 100\n+ self.logger.info(sub_cases_result)\n+ # self.logger.info('%s pass rate is: %s' % (self.test_case.running_case, pass_rate))\n+ self.logger.info('pass rate is: %s' % pass_rate)\n+ self.verify(pass_rate == 100.00, 'some subcases failed')\n+\n+ @staticmethod\n+ def get_ipv6_template_by_ipv4(template):\n+ if isinstance(template, dict):\n+ template = [template]\n+ ipv6_template = [eval(str(element).replace('eth / ipv4', 'eth / ipv6')\n+ .replace('IP()', 'IPv6()').replace('mac_ipv4', 'mac_ipv6'))\n+ for element in template]\n+ return ipv6_template\n", "prefixes": [ "V3", "1/8" ] }{ "id": 83370, "url": "