get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/83370/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 83370,
    "url": "https://patches.dpdk.org/api/patches/83370/?format=api",
    "web_url": "https://patches.dpdk.org/project/dts/patch/20201102092120.9146-2-haiyangx.zhao@intel.com/",
    "project": {
        "id": 3,
        "url": "https://patches.dpdk.org/api/projects/3/?format=api",
        "name": "DTS",
        "link_name": "dts",
        "list_id": "dts.dpdk.org",
        "list_email": "dts@dpdk.org",
        "web_url": "",
        "scm_url": "git://dpdk.org/tools/dts",
        "webscm_url": "http://git.dpdk.org/tools/dts/",
        "list_archive_url": "https://inbox.dpdk.org/dts",
        "list_archive_url_format": "https://inbox.dpdk.org/dts/{}",
        "commit_url_format": ""
    },
    "msgid": "<20201102092120.9146-2-haiyangx.zhao@intel.com>",
    "list_archive_url": "https://inbox.dpdk.org/dts/20201102092120.9146-2-haiyangx.zhao@intel.com",
    "date": "2020-11-02T09:21:13",
    "name": "[V3,1/8] tests/rte_flow_common: add a common module to process rss test",
    "commit_ref": null,
    "pull_url": null,
    "state": "changes-requested",
    "archived": false,
    "hash": "fccb15d3eeff63bf88f306b08ab06ec5f59a887b",
    "submitter": {
        "id": 1978,
        "url": "https://patches.dpdk.org/api/people/1978/?format=api",
        "name": "Zhao, HaiyangX",
        "email": "haiyangx.zhao@intel.com"
    },
    "delegate": null,
    "mbox": "https://patches.dpdk.org/project/dts/patch/20201102092120.9146-2-haiyangx.zhao@intel.com/mbox/",
    "series": [
        {
            "id": 13558,
            "url": "https://patches.dpdk.org/api/series/13558/?format=api",
            "web_url": "https://patches.dpdk.org/project/dts/list/?series=13558",
            "date": "2020-11-02T09:21:12",
            "name": "tests: update or add rss related suites",
            "version": 3,
            "mbox": "https://patches.dpdk.org/series/13558/mbox/"
        }
    ],
    "comments": "https://patches.dpdk.org/api/patches/83370/comments/",
    "check": "pending",
    "checks": "https://patches.dpdk.org/api/patches/83370/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<dts-bounces@dpdk.org>",
        "X-Original-To": "patchwork@inbox.dpdk.org",
        "Delivered-To": "patchwork@inbox.dpdk.org",
        "Received": [
            "from dpdk.org (dpdk.org [92.243.14.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id BBAD9A04E7;\n\tMon,  2 Nov 2020 10:22:42 +0100 (CET)",
            "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id CB463BBB4;\n\tMon,  2 Nov 2020 10:22:40 +0100 (CET)",
            "from mga07.intel.com (mga07.intel.com [134.134.136.100])\n by dpdk.org (Postfix) with ESMTP id ACF8A72ED\n for <dts@dpdk.org>; Mon,  2 Nov 2020 10:22:38 +0100 (CET)",
            "from fmsmga008.fm.intel.com ([10.253.24.58])\n by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 02 Nov 2020 01:22:37 -0800",
            "from unknown (HELO localhost.localdomain) ([10.240.183.55])\n by fmsmga008-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 02 Nov 2020 01:22:35 -0800"
        ],
        "IronPort-SDR": [
            "\n l4WUTG3xKFlOGRdDU4/jn6uMla19GrkhQXyVWFbmouNATdwXypFSV6afMDuxuuFRFCPvubBM0S\n v2yz8uW3xQTw==",
            "\n yfBwgbfZg2VEubh6GITLlXfKleyU3G+VZRd+hsBmJENUey4n1i3elXBKLX35gKQr+G14cVlPMw\n 6ITyRQR2VYzg=="
        ],
        "X-IronPort-AV": [
            "E=McAfee;i=\"6000,8403,9792\"; a=\"233027959\"",
            "E=Sophos;i=\"5.77,444,1596524400\"; d=\"scan'208\";a=\"233027959\"",
            "E=Sophos;i=\"5.77,444,1596524400\"; d=\"scan'208\";a=\"305386210\""
        ],
        "X-Amp-Result": "SKIPPED(no attachment in message)",
        "X-Amp-File-Uploaded": "False",
        "From": "Haiyang Zhao <haiyangx.zhao@intel.com>",
        "To": "dts@dpdk.org,\n\tqi.fu@intel.com",
        "Cc": "Haiyang Zhao <haiyangx.zhao@intel.com>",
        "Date": "Mon,  2 Nov 2020 17:21:13 +0800",
        "Message-Id": "<20201102092120.9146-2-haiyangx.zhao@intel.com>",
        "X-Mailer": "git-send-email 2.17.1",
        "In-Reply-To": "<20201102092120.9146-1-haiyangx.zhao@intel.com>",
        "References": "<20201102092120.9146-1-haiyangx.zhao@intel.com>",
        "Subject": "[dts] [PATCH V3 1/8] tests/rte_flow_common: add a common module to\n\tprocess rss test",
        "X-BeenThere": "dts@dpdk.org",
        "X-Mailman-Version": "2.1.15",
        "Precedence": "list",
        "List-Id": "test suite reviews and discussions <dts.dpdk.org>",
        "List-Unsubscribe": "<https://mails.dpdk.org/options/dts>,\n <mailto:dts-request@dpdk.org?subject=unsubscribe>",
        "List-Archive": "<http://mails.dpdk.org/archives/dts/>",
        "List-Post": "<mailto:dts@dpdk.org>",
        "List-Help": "<mailto:dts-request@dpdk.org?subject=help>",
        "List-Subscribe": "<https://mails.dpdk.org/listinfo/dts>,\n <mailto:dts-request@dpdk.org?subject=subscribe>",
        "Errors-To": "dts-bounces@dpdk.org",
        "Sender": "\"dts\" <dts-bounces@dpdk.org>"
    },
    "content": "*.add a class named RssProcessing to process rss tests.\n\nSigned-off-by: Haiyang Zhao <haiyangx.zhao@intel.com>\n---\n tests/rte_flow_common.py | 391 +++++++++++++++++++++++++++++++++++++++\n 1 file changed, 391 insertions(+)",
    "diff": "diff --git a/tests/rte_flow_common.py b/tests/rte_flow_common.py\nindex 8a4cded..43c0eab 100644\n--- a/tests/rte_flow_common.py\n+++ b/tests/rte_flow_common.py\n@@ -33,6 +33,7 @@ import json\n import time\n import re\n from utils import GREEN, RED\n+from packet import Packet\n \n CVL_TXQ_RXQ_NUMBER = 16\n \n@@ -672,3 +673,393 @@ def check_pf_rss_queue(out, count):\n         return True\n     else:\n         return False\n+\n+\n+class RssProcessing(object):\n+    def __init__(self, test_case, pmd_output, tester_ifaces, rxq):\n+        self.test_case = test_case\n+        self.pmd_output = pmd_output\n+        self.tester_ifaces = tester_ifaces\n+        self.rxq = rxq\n+        self.logger = test_case.logger\n+        self.pkt = Packet()\n+        self.verify = self.test_case.verify\n+        self.pass_flag = 'passed'\n+        self.fail_flag = 'failed'\n+        self.current_saved_hash = ''\n+        self.hash_records = {}\n+        self.handle_output_methods = {\n+            'save_hash': self.save_hash,\n+            'save_or_no_hash': self.save_or_no_hash,\n+            'check_hash_different': self.check_hash_different,\n+            'check_no_hash_or_different': self.check_no_hash_or_different,\n+            'check_hash_same': self.check_hash_same,\n+            'check_no_hash': self.check_no_hash,\n+        }\n+        self.error_msgs = []\n+\n+    def save_hash(self, out, key='', port_id=0):\n+        hashes, rss_distribute = self.get_hash_verify_rss_distribute(out, port_id)\n+        if len(key) != 0:\n+            self.hash_records[key] = hashes\n+        self.current_saved_hash = hashes\n+        if not rss_distribute:\n+            error_msg = 'the packet do not distribute by rss'\n+            self.logger.error(error_msg)\n+            self.error_msgs.append(error_msg)\n+\n+    def save_or_no_hash(self, out, key='', port_id=0):\n+        hashes, queues = self.get_hash_and_queues(out, port_id)\n+        if len(hashes) == 0:\n+            self.logger.info('There no hash value passed as expected')\n+            if set(queues) != {'0x0'}:\n+                error_msg = 'received queues should all be 0, but are {}'.format(queues)\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+            return\n+        if len(key) != 0:\n+            self.hash_records[key] = hashes\n+        self.current_saved_hash = hashes\n+        if not self.verify_rss_distribute(hashes, queues):\n+            error_msg = 'the packet do not distribute by rss'\n+            self.logger.error(error_msg)\n+            self.error_msgs.append(error_msg)\n+\n+    def check_hash_different(self, out, key='', port_id=0):\n+        hashes, rss_distribute = self.get_hash_verify_rss_distribute(out, port_id)\n+        if len(key) == 0:\n+            if hashes == self.current_saved_hash:\n+                error_msg = 'hash value {} should be different ' \\\n+                            'with current saved hash {}'.format(hashes, self.current_saved_hash)\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+        else:\n+            if hashes == self.hash_records[key]:\n+                error_msg = 'hash value {} should be different ' \\\n+                            'with {} {}'.format(hashes, key, self.hash_records[key])\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+        if not rss_distribute:\n+            error_msg = 'the packet do not distribute by rss'\n+            self.logger.error(error_msg)\n+            self.error_msgs.append(error_msg)\n+\n+    def check_no_hash(self, out, port_id=0):\n+        hashes, queues = self.get_hash_and_queues(out, port_id)\n+        if len(hashes) != 0:\n+            error_msg = 'hash value {} should be empty'.format(hashes)\n+            self.logger.error(error_msg)\n+            self.error_msgs.append(error_msg)\n+        elif set(queues) != {'0x0'}:\n+            error_msg = 'received queues should all be 0, but are {}'.format(queues)\n+            self.logger.error(error_msg)\n+            self.error_msgs.append(error_msg)\n+\n+    def check_no_hash_or_different(self, out, key='', port_id=0):\n+        hashes, queues = self.get_hash_and_queues(out, port_id)\n+        if len(hashes) == 0:\n+            self.logger.info('There no hash value passed as expected')\n+            if set(queues) != {'0x0'}:\n+                error_msg = 'received queues should all be 0, but are {}'.format(queues)\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+            return\n+        if len(key) == 0:\n+            if hashes == self.current_saved_hash:\n+                error_msg = 'hash value {} should be different ' \\\n+                            'with current saved hash {}'.format(hashes, self.current_saved_hash)\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+        else:\n+            if hashes == self.hash_records[key]:\n+                error_msg = 'hash value {} should be different ' \\\n+                            'with {} {}'.format(hashes, key, self.hash_records[key])\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+\n+    def check_hash_same(self, out, key='', port_id=0):\n+        hashes, rss_distribute = self.get_hash_verify_rss_distribute(out, port_id)\n+        if len(key) == 0:\n+            if hashes != self.current_saved_hash:\n+                error_msg = 'hash value {} should be same ' \\\n+                            'with current saved hash {}'.format(hashes, self.current_saved_hash)\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+        else:\n+            if hashes != self.hash_records[key]:\n+                error_msg = 'hash value {} should be same ' \\\n+                            'with {} {}'.format(hashes, key, self.hash_records[key])\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+        if not rss_distribute:\n+            error_msg = 'the packet do not distribute by rss'\n+            self.logger.error(error_msg)\n+            self.error_msgs.append(error_msg)\n+\n+    def check_hash_same_or_no_hash(self, out, key='', port_id=0):\n+        hashes, rss_distribute = self.get_hash_verify_rss_distribute(out, port_id)\n+        if len(hashes) != 0:\n+            error_msg = 'hash value {} should be empty'.format(hashes)\n+            self.logger.error(error_msg)\n+            self.error_msgs.append(error_msg)\n+            return\n+        elif set(rss_distribute) != {'0x0'}:\n+            error_msg = 'received queues should all be 0, but are {}'.format(rss_distribute)\n+            self.logger.error(error_msg)\n+            self.error_msgs.append(error_msg)\n+            return\n+        if len(key) == 0:\n+            if hashes != self.current_saved_hash:\n+                error_msg = 'hash value {} should be same ' \\\n+                            'with current saved hash {}'.format(hashes, self.current_saved_hash)\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+        else:\n+            if hashes != self.hash_records[key]:\n+                error_msg = 'hash value {} should be same ' \\\n+                            'with {} {}'.format(hashes, key, self.hash_records[key])\n+                self.logger.error(error_msg)\n+                self.error_msgs.append(error_msg)\n+        if not rss_distribute:\n+            error_msg = 'the packet do not distribute by rss'\n+            self.logger.error(error_msg)\n+            self.error_msgs.append(error_msg)\n+\n+    def verify_rss_distribute(self, hashes, queues):\n+        if len(hashes) != len(queues):\n+            self.logger.warning('hash length {} != queue length {}'.format(hashes, queues))\n+            return False\n+        for i in range(len(hashes)):\n+            if int(hashes[i], 16) % self.rxq != int(queues[i], 16):\n+                self.logger.warning('hash values {} mod total queues {} != queue {}'\n+                                    .format(hashes[i], self.rxq, queues[i]))\n+                return False\n+        return True\n+\n+    def get_hash_verify_rss_distribute(self, out, port_id=0):\n+        hashes, queues = self.get_hash_and_queues(out, port_id)\n+        if len(hashes) == 0:\n+            return [], False\n+        return hashes, self.verify_rss_distribute(hashes, queues)\n+\n+    def get_hash_and_queues(self, out, port_id=0):\n+        hash_pattern = re.compile('port\\s%s/queue\\s\\d+:\\sreceived\\s\\d+\\spackets.+?\\n.*RSS\\shash=(\\w+)\\s-\\sRSS\\squeue=(\\w+)' % port_id)\n+        hash_infos = hash_pattern.findall(out)\n+        self.logger.info('hash_infos: {}'.format(hash_infos))\n+        if len(hash_infos) == 0:\n+            queue_pattern = re.compile('Receive\\squeue=(\\w+)')\n+            queues = queue_pattern.findall(out)\n+            return [], queues\n+        # hashes = [int(hash_info[0], 16) for hash_info in hash_infos]\n+        hashes = [hash_info[0].strip() for hash_info in hash_infos]\n+        queues = [hash_info[1].strip() for hash_info in hash_infos]\n+        return hashes, queues\n+\n+    def send_pkt_get_output(self, pkts, port_id=0, count=1, interval=0):\n+        self.pkt.update_pkt(pkts)\n+        tx_port = self.tester_ifaces[0] if port_id == 0 else self.tester_ifaces[1]\n+        self.logger.info('----------send packet-------------')\n+        self.logger.info('{}'.format(pkts))\n+        self.pkt.send_pkt(crb=self.test_case.tester, tx_port=tx_port, count=count, interval=interval)\n+        out = self.pmd_output.get_output(timeout=1)\n+        pkt_pattern = 'port\\s%d/queue\\s\\d+:\\sreceived\\s(\\d+)\\spackets.+?\\n.*length=\\d{2,}\\s' % port_id\n+        reveived_data = re.findall(pkt_pattern, out)\n+        reveived_pkts = sum(map(int, [i[0] for i in reveived_data]))\n+        if isinstance(pkts, list):\n+            self.verify(reveived_pkts == len(pkts) * count,\n+                        'expect received %d pkts, but get %d instead' % (len(pkts) * count, reveived_pkts))\n+        else:\n+            self.verify(reveived_pkts == 1 * count,\n+                        'expect received %d pkts, but get %d instead' % (1 * count, reveived_pkts))\n+        return out\n+\n+    def send_pkt_get_hash_queues(self, pkts, port_id=0, count=1, interval=0):\n+        output = self.send_pkt_get_output(pkts, port_id, count, interval)\n+        hashes, queues = self.get_hash_and_queues(output, port_id)\n+        return hashes, queues\n+\n+    def create_rule(self, rule: (list, str), check_stats=True, msg=None):\n+        p = re.compile(r\"Flow rule #(\\d+) created\")\n+        rule_list = list()\n+        if isinstance(rule, list):\n+            for i in rule:\n+                out = self.pmd_output.execute_cmd(i, timeout=1)\n+                if msg:\n+                    self.verify(msg in out, \"failed: expect %s in %s\" % (msg, out))\n+                m = p.search(out)\n+                if m:\n+                    rule_list.append(m.group(1))\n+                else:\n+                    rule_list.append(False)\n+        elif isinstance(rule, str):\n+            out = self.pmd_output.execute_cmd(rule, timeout=1)\n+            if msg:\n+                self.verify(msg in out, \"failed: expect %s in %s\" % (msg, out))\n+            m = p.search(out)\n+            if m:\n+                rule_list.append(m.group(1))\n+            else:\n+                rule_list.append(False)\n+        else:\n+            raise Exception(\"unsupported rule type, only accept list or str\")\n+        if check_stats:\n+            self.verify(all(rule_list), \"some rules create failed, result %s\" % rule_list)\n+        elif not check_stats:\n+            self.verify(not any(rule_list), \"all rules should create failed, result %s\" % rule_list)\n+        return rule_list\n+\n+    def validate_rule(self, rule, check_stats=True, check_msg=None):\n+        flag = 'Flow rule validated'\n+        if isinstance(rule, str):\n+            if 'create' in rule:\n+                rule = rule.replace('create', 'validate')\n+            out = self.pmd_output.execute_cmd(rule, timeout=1)\n+            if check_stats:\n+                self.verify(flag in out.strip(), \"rule %s validated failed, result %s\" % (rule, out))\n+            else:\n+                if check_msg:\n+                    self.verify(flag not in out.strip() and check_msg in out.strip(),\n+                                \"rule %s validate should failed with msg: %s, but result %s\" % (rule, check_msg, out))\n+                else:\n+                    self.verify(flag not in out.strip(), \"rule %s validate should failed, result %s\" % (rule, out))\n+        elif isinstance(rule, list):\n+            for r in rule:\n+                if 'create' in r:\n+                    r = r.replace('create', 'validate')\n+                out = self.pmd_output.execute_cmd(r, timeout=1)\n+                if check_stats:\n+                    self.verify(flag in out.strip(), \"rule %s validated failed, result %s\" % (r, out))\n+                else:\n+                    if not check_msg:\n+                        self.verify(flag not in out.strip(), \"rule %s validate should failed, result %s\" % (r, out))\n+                    else:\n+                        self.verify(flag not in out.strip() and check_msg in out.strip(),\n+                                    \"rule %s should validate failed with msg: %s, but result %s\" % (\n+                                        r, check_msg, out))\n+\n+    def check_rule(self, port_id=0, stats=True, rule_list=None):\n+        out = self.pmd_output.execute_cmd(\"flow list %s\" % port_id)\n+        p = re.compile(r\"ID\\s+Group\\s+Prio\\s+Attr\\s+Rule\")\n+        matched = p.search(out)\n+        if stats:\n+            self.verify(matched, \"flow rule on port %s is not existed\" % port_id)\n+            if rule_list:\n+                p2 = re.compile(\"^(\\d+)\\s\")\n+                li = out.splitlines()\n+                res = list(filter(bool, list(map(p2.match, li))))\n+                result = [i.group(1) for i in res]\n+                self.verify(set(rule_list).issubset(set(result)),\n+                            \"check rule list failed. expect %s, result %s\" % (rule_list, result))\n+        else:\n+            if matched:\n+                if rule_list:\n+                    res_li = [i.split()[0].strip() for i in out.splitlines() if re.match('\\d', i)]\n+                    self.verify(not set(rule_list).issubset(res_li), 'rule specified should not in result.')\n+                else:\n+                    raise Exception('expect no rule listed')\n+            else:\n+                self.verify(not matched, \"flow rule on port %s is existed\" % port_id)\n+\n+    def destroy_rule(self, port_id=0, rule_id=None):\n+        if rule_id is None:\n+            rule_id = 0\n+        if isinstance(rule_id, list):\n+            for i in rule_id:\n+                out = self.test_case.dut.send_command(\"flow destroy %s rule %s\" % (port_id, i), timeout=1)\n+                p = re.compile(r\"Flow rule #(\\d+) destroyed\")\n+                m = p.search(out)\n+                self.verify(m, \"flow rule %s delete failed\" % rule_id)\n+        else:\n+            out = self.test_case.dut.send_command(\"flow destroy %s rule %s\" % (port_id, rule_id), timeout=1)\n+            p = re.compile(r\"Flow rule #(\\d+) destroyed\")\n+            m = p.search(out)\n+            self.verify(m, \"flow rule %s delete failed\" % rule_id)\n+\n+    def handle_actions(self, output, actions, port_id=0):\n+        if isinstance(actions, dict) or isinstance(actions, str):\n+            actions = [actions]\n+        for action in actions:  # [{}]\n+            self.logger.info('action: {}\\n'.format(action))\n+            if isinstance(action, str):\n+                if action in self.handle_output_methods:\n+                    self.handle_output_methods[action](output, port_id=port_id)\n+            else:\n+                for method in action:  # {'save': ''}\n+                    if method in self.handle_output_methods:\n+                        if method == 'check_no_hash':\n+                            self.check_no_hash(output, port_id=port_id)\n+                        else:\n+                            self.handle_output_methods[method](output, action[method], port_id=port_id)\n+\n+    def handle_tests(self, tests, port_id=0):\n+        out = ''\n+        for test in tests:\n+            if 'send_packet' in test:\n+                out = self.send_pkt_get_output(test['send_packet'], port_id)\n+            if 'action' in test:\n+                self.handle_actions(out, test['action'])\n+\n+    def handle_rss_case(self, case_info):\n+        # clear hash_records before each sub case\n+        self.hash_records = {}\n+        self.error_msgs = []\n+        self.current_saved_hash = ''\n+        sub_case_name = case_info.get('sub_casename')\n+        self.logger.info('===================Test sub case: {}================'.format(sub_case_name))\n+        port_id = case_info.get('port_id') if case_info.get('port_id') else 0\n+        rules = case_info.get('rule') if case_info.get('rule') else []\n+        rule_ids = []\n+        if 'pre-test' in case_info:\n+            self.logger.info('------------handle pre-test--------------')\n+            self.handle_tests(case_info['pre-test'], port_id)\n+\n+        # handle tests\n+        tests = case_info['test']\n+        self.logger.info('------------handle test--------------')\n+        # validate rule\n+        if rules:\n+            self.validate_rule(rule=rules, check_stats=True)\n+            rule_ids = self.create_rule(rule=case_info['rule'], check_stats=True)\n+            self.check_rule(port_id=port_id, rule_list=rule_ids)\n+        self.handle_tests(tests, port_id)\n+\n+        # handle post-test\n+        if 'post-test' in case_info:\n+            self.logger.info('------------handle post-test--------------')\n+            self.destroy_rule(port_id=port_id, rule_id=rule_ids)\n+            self.check_rule(port_id=port_id, stats=False)\n+            self.handle_tests(case_info['post-test'], port_id)\n+        if self.error_msgs:\n+            self.verify(False, str(self.error_msgs[:500]))\n+\n+    def handle_rss_distribute_cases(self, cases_info):\n+        sub_cases_result = dict()\n+        if not isinstance(cases_info, list):\n+            cases_info = [cases_info]\n+\n+        for case_info in cases_info:\n+            try:\n+                # self.handle_rss_distribute_case(case_info=case_info)\n+                self.handle_rss_case(case_info=case_info)\n+            except Exception as e:\n+                self.logger.warning('sub_case %s failed: %s' % (case_info['sub_casename'], e))\n+                sub_cases_result[case_info['sub_casename']] = self.fail_flag\n+            else:\n+                self.logger.info('sub_case %s passed' % case_info['sub_casename'])\n+                sub_cases_result[case_info['sub_casename']] = self.pass_flag\n+            finally:\n+                self.pmd_output.execute_cmd('flow flush 0')\n+        pass_rate = round(list(sub_cases_result.values()).count(self.pass_flag) / len(sub_cases_result), 4) * 100\n+        self.logger.info(sub_cases_result)\n+        # self.logger.info('%s pass rate is: %s' % (self.test_case.running_case, pass_rate))\n+        self.logger.info('pass rate is: %s' % pass_rate)\n+        self.verify(pass_rate == 100.00, 'some subcases failed')\n+\n+    @staticmethod\n+    def get_ipv6_template_by_ipv4(template):\n+        if isinstance(template, dict):\n+            template = [template]\n+        ipv6_template = [eval(str(element).replace('eth / ipv4', 'eth / ipv6')\n+                              .replace('IP()', 'IPv6()').replace('mac_ipv4', 'mac_ipv6'))\n+                         for element in template]\n+        return ipv6_template\n",
    "prefixes": [
        "V3",
        "1/8"
    ]
}