get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/94048/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 94048,
    "url": "http://patches.dpdk.org/api/patches/94048/?format=api",
    "web_url": "http://patches.dpdk.org/project/dts/patch/20210609164721.30190-1-zhiminx.huang@intel.com/",
    "project": {
        "id": 3,
        "url": "http://patches.dpdk.org/api/projects/3/?format=api",
        "name": "DTS",
        "link_name": "dts",
        "list_id": "dts.dpdk.org",
        "list_email": "dts@dpdk.org",
        "web_url": "",
        "scm_url": "git://dpdk.org/tools/dts",
        "webscm_url": "http://git.dpdk.org/tools/dts/",
        "list_archive_url": "https://inbox.dpdk.org/dts",
        "list_archive_url_format": "https://inbox.dpdk.org/dts/{}",
        "commit_url_format": ""
    },
    "msgid": "<20210609164721.30190-1-zhiminx.huang@intel.com>",
    "list_archive_url": "https://inbox.dpdk.org/dts/20210609164721.30190-1-zhiminx.huang@intel.com",
    "date": "2021-06-09T16:47:21",
    "name": "[V1] tests: add test suite for ipfragment",
    "commit_ref": null,
    "pull_url": null,
    "state": "superseded",
    "archived": false,
    "hash": "2f71b052ff758c9fcfcde89878f4aaaa154873fe",
    "submitter": {
        "id": 1685,
        "url": "http://patches.dpdk.org/api/people/1685/?format=api",
        "name": "Huang, ZhiminX",
        "email": "zhiminx.huang@intel.com"
    },
    "delegate": null,
    "mbox": "http://patches.dpdk.org/project/dts/patch/20210609164721.30190-1-zhiminx.huang@intel.com/mbox/",
    "series": [
        {
            "id": 17274,
            "url": "http://patches.dpdk.org/api/series/17274/?format=api",
            "web_url": "http://patches.dpdk.org/project/dts/list/?series=17274",
            "date": "2021-06-09T16:47:21",
            "name": "[V1] tests: add test suite for ipfragment",
            "version": 1,
            "mbox": "http://patches.dpdk.org/series/17274/mbox/"
        }
    ],
    "comments": "http://patches.dpdk.org/api/patches/94048/comments/",
    "check": "pending",
    "checks": "http://patches.dpdk.org/api/patches/94048/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<dts-bounces@dpdk.org>",
        "X-Original-To": "patchwork@inbox.dpdk.org",
        "Delivered-To": "patchwork@inbox.dpdk.org",
        "Received": [
            "from mails.dpdk.org (mails.dpdk.org [217.70.189.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id 1D5C8A0C46;\n\tWed,  9 Jun 2021 10:17:13 +0200 (CEST)",
            "from [217.70.189.124] (localhost [127.0.0.1])\n\tby mails.dpdk.org (Postfix) with ESMTP id E13D140FDF;\n\tWed,  9 Jun 2021 10:17:12 +0200 (CEST)",
            "from mga02.intel.com (mga02.intel.com [134.134.136.20])\n by mails.dpdk.org (Postfix) with ESMTP id BD75E40E3C\n for <dts@dpdk.org>; Wed,  9 Jun 2021 10:17:10 +0200 (CEST)",
            "from orsmga004.jf.intel.com ([10.7.209.38])\n by orsmga101.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 09 Jun 2021 01:16:54 -0700",
            "from unknown (HELO cvl_100g_103.icx.intel.com) ([10.240.183.103])\n by orsmga004-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 09 Jun 2021 01:16:52 -0700"
        ],
        "IronPort-SDR": [
            "\n hetVrFAj17HjhL9TUuXuwXtCXZrxqC7/Boed1DAXhVaIDidgLGF7erWrPDTSRcXF0pIou0R9fD\n wgy/FHl+bwlQ==",
            "\n NjV92U4Y5stvxzd60ZmQBqLoozr0AyUd2AkoYIvNshj/lqiHr7/O/rMK23zhvPpzgZECrCzo3k\n Nb95C2vaRezA=="
        ],
        "X-IronPort-AV": [
            "E=McAfee;i=\"6200,9189,10009\"; a=\"192137703\"",
            "E=Sophos;i=\"5.83,260,1616482800\"; d=\"scan'208\";a=\"192137703\"",
            "E=Sophos;i=\"5.83,260,1616482800\"; d=\"scan'208\";a=\"551905460\""
        ],
        "From": "Zhimin Huang <zhiminx.huang@intel.com>",
        "To": "dts@dpdk.org",
        "Cc": "qi.fu@intel.com,\n\tZhimin Huang <zhiminx.huang@intel.com>",
        "Date": "Thu, 10 Jun 2021 00:47:21 +0800",
        "Message-Id": "<20210609164721.30190-1-zhiminx.huang@intel.com>",
        "X-Mailer": "git-send-email 2.17.1",
        "MIME-Version": "1.0",
        "Content-Type": "text/plain; charset=UTF-8",
        "Content-Transfer-Encoding": "8bit",
        "Subject": "[dts] [PATCH V1] tests: add test suite for ipfragment",
        "X-BeenThere": "dts@dpdk.org",
        "X-Mailman-Version": "2.1.29",
        "Precedence": "list",
        "List-Id": "test suite reviews and discussions <dts.dpdk.org>",
        "List-Unsubscribe": "<https://mails.dpdk.org/options/dts>,\n <mailto:dts-request@dpdk.org?subject=unsubscribe>",
        "List-Archive": "<http://mails.dpdk.org/archives/dts/>",
        "List-Post": "<mailto:dts@dpdk.org>",
        "List-Help": "<mailto:dts-request@dpdk.org?subject=help>",
        "List-Subscribe": "<https://mails.dpdk.org/listinfo/dts>,\n <mailto:dts-request@dpdk.org?subject=subscribe>",
        "Errors-To": "dts-bounces@dpdk.org",
        "Sender": "\"dts\" <dts-bounces@dpdk.org>"
    },
    "content": "add new feature ipfragment test suite\nthis feature enable ipfragment pkt rss and fdir by IPID\n\nSigned-off-by: Zhimin Huang <zhiminx.huang@intel.com>\n---\n tests/TestSuite_ip_fragment.py | 755 +++++++++++++++++++++++++++++++++\n 1 file changed, 755 insertions(+)\n create mode 100644 tests/TestSuite_ip_fragment.py",
    "diff": "diff --git a/tests/TestSuite_ip_fragment.py b/tests/TestSuite_ip_fragment.py\nnew file mode 100644\nindex 00000000..1c104b46\n--- /dev/null\n+++ b/tests/TestSuite_ip_fragment.py\n@@ -0,0 +1,755 @@\n+# BSD LICENSE\n+#\n+# Copyright(c) 2021 Intel Corporation. All rights reserved.\n+# All rights reserved.\n+#\n+# Redistribution and use in source and binary forms, with or without\n+# modification, are permitted provided that the following conditions\n+# are met:\n+#\n+#   * Redistributions of source code must retain the above copyright\n+#     notice, this list of conditions and the following disclaimer.\n+#   * Redistributions in binary form must reproduce the above copyright\n+#     notice, this list of conditions and the following disclaimer in\n+#     the documentation and/or other materials provided with the\n+#     distribution.\n+#   * Neither the name of Intel Corporation nor the names of its\n+#     contributors may be used to endorse or promote products derived\n+#     from this software without specific prior written permission.\n+#\n+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS\n+# \"AS IS\" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT\n+# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR\n+# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT\n+# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,\n+# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT\n+# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,\n+# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY\n+# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT\n+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE\n+# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\n+\n+\n+from packet import Packet\n+from pmd_output import PmdOutput\n+from test_case import TestCase\n+import re\n+from utils import GREEN, RED\n+import time\n+from scapy.all import *\n+import rte_flow_common as rfc\n+\n+LAUNCH_QUEUE = 16\n+\n+tv_mac_ipv4_frag_fdir_queue_index = {\n+    \"name\": \"tv_mac_ipv4_frag_fdir_queue_index\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions queue index 1 / mark / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"queue\": 1, \"mark_id\": 0}\n+}\n+\n+tv_mac_ipv4_frag_fdir_rss_queues = {\n+    \"name\": \"tv_mac_ipv4_frag_fdir_rss_queues\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions rss queues 2 3 end / mark / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"queue\": [2, 3], \"mark_id\": 0}\n+}\n+\n+tv_mac_ipv4_frag_fdir_passthru = {\n+    \"name\": \"tv_mac_ipv4_frag_fdir_passthru\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions passthru / mark / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"mark_id\": 0}\n+}\n+\n+tv_mac_ipv4_frag_fdir_drop = {\n+    \"name\": \"tv_mac_ipv4_frag_fdir_drop\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions drop / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"drop\": True}\n+}\n+\n+tv_mac_ipv4_frag_fdir_mark_rss = {\n+    \"name\": \"tv_mac_ipv4_frag_fdir_mark_rss\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions mark / rss / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"mark_id\": 0, \"rss\": True}\n+}\n+\n+tv_mac_ipv4_frag_fdir_mark = {\n+    \"name\": \"tv_mac_ipv4_frag_fdir_mark\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions mark id 1 / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"mark_id\": 1}\n+}\n+\n+tvs_mac_ipv4_fragment_fdir = [\n+    tv_mac_ipv4_frag_fdir_queue_index,\n+    tv_mac_ipv4_frag_fdir_rss_queues,\n+    tv_mac_ipv4_frag_fdir_passthru,\n+    tv_mac_ipv4_frag_fdir_drop,\n+    tv_mac_ipv4_frag_fdir_mark_rss,\n+    tv_mac_ipv4_frag_fdir_mark,\n+]\n+\n+tvs_mac_ipv4_fragment_fdir_l2dst = [eval(str(element).replace('mac_ipv4_frag', 'mac_ipv4_frag_l2dst')\n+                                                     .replace('eth /', 'eth dst is 00:00:00:00:00:01 /')\n+                                                     .replace(\"Ether()\", \"Ether(dst='00:00:00:00:00:01')\")\n+                                         )\n+                                    for element in tvs_mac_ipv4_fragment_fdir]\n+\n+tvs_mac_ipv4_frag_fdir_with_l2 = tvs_mac_ipv4_fragment_fdir_l2dst\n+\n+tvs_mac_ipv4_fragment_fdir_l3src = [eval(str(element).replace('mac_ipv4_frag', 'mac_ipv4_frag_l3src')\n+                                                     .replace('ipv4 packet_id', 'ipv4 src is 192.168.1.1 packet_id')\n+                                                     .replace(\"IP(id=47750)\", \"IP(id=47750, src='192.168.1.1')\"))\n+                                    for element in tvs_mac_ipv4_fragment_fdir]\n+\n+tvs_mac_ipv4_fragment_fdir_l3dst = [eval(str(element).replace('mac_ipv4_frag', 'mac_ipv4_frag_l3dst')\n+                                                     .replace('ipv4 packet_id', 'ipv4 dst is 192.168.1.2 packet_id')\n+                                                     .replace(\"IP(id=47750)\", \"IP(id=47750, dst='192.168.1.2')\"))\n+                                    for element in tvs_mac_ipv4_fragment_fdir]\n+\n+tvs_mac_ipv4_frag_fdir_with_l3 = tvs_mac_ipv4_fragment_fdir_l3src + tvs_mac_ipv4_fragment_fdir_l3dst\n+\n+tv_mac_ipv6_frag_fdir_queue_index = {\n+    \"name\": \"tv_mac_ipv6_frag_fdir_queue_index\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffffffff packet_id mask 0xffffffff frag_data spec 0x0001 frag_data last 0xffff frag_data mask 0xffff / end actions queue index 1 / mark / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"queue\": 1, \"mark_id\": 0}\n+}\n+\n+tv_mac_ipv6_frag_fdir_rss_queues = {\n+    \"name\": \"tv_mac_ipv6_frag_fdir_rss_queues\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffffffff packet_id mask 0xffffffff frag_data spec 0x0001 frag_data last 0xffff frag_data mask 0xffff / end actions rss queues 2 3 end / mark / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"queue\": [2, 3], \"mark_id\": 0}\n+}\n+\n+tv_mac_ipv6_frag_fdir_passthru = {\n+    \"name\": \"tv_mac_ipv6_frag_fdir_passthru\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffffffff packet_id mask 0xffffffff frag_data spec 0x0001 frag_data last 0xffff frag_data mask 0xffff / end actions passthru / mark / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"mark_id\": 0}\n+}\n+\n+tv_mac_ipv6_frag_fdir_drop = {\n+    \"name\": \"tv_mac_ipv6_frag_fdir_drop\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffffffff packet_id mask 0xffffffff frag_data spec 0x0001 frag_data last 0xffff frag_data mask 0xffff / end actions drop / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"drop\": True}\n+}\n+\n+tv_mac_ipv6_frag_fdir_mark_rss = {\n+    \"name\": \"tv_mac_ipv6_frag_fdir_mark_rss\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffffffff packet_id mask 0xffffffff frag_data spec 0x0001 frag_data last 0xffff frag_data mask 0xffff / end actions mark / rss / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"mark_id\": 0, \"rss\": True}\n+}\n+\n+tv_mac_ipv6_frag_fdir_mark = {\n+    \"name\": \"tv_mac_ipv6_frag_fdir_mark\",\n+    \"rule\": \"flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffffffff packet_id mask 0xffffffff frag_data spec 0x0001 frag_data last 0xffff frag_data mask 0xffff / end actions mark id 1 / end\",\n+    \"scapy_str\": {\"matched\": [\"Ether()/IPv6()/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"],\n+                  \"unmatched\": [\"Ether()/IP(id=47750)/Raw('X'*666)\"]\n+                  },\n+    \"check_param\": {\"port_id\": 0, \"rxq\": LAUNCH_QUEUE, \"mark_id\": 1}\n+}\n+\n+tvs_mac_ipv6_fragment_fdir = [\n+    tv_mac_ipv6_frag_fdir_queue_index,\n+    tv_mac_ipv6_frag_fdir_rss_queues,\n+    tv_mac_ipv6_frag_fdir_passthru,\n+    tv_mac_ipv6_frag_fdir_drop,\n+    tv_mac_ipv6_frag_fdir_mark_rss,\n+    tv_mac_ipv6_frag_fdir_mark,\n+]\n+\n+tvs_mac_ipv6_fragment_fdir_l2dst = [eval(str(element).replace('mac_ipv6_frag', 'mac_ipv6_frag_l2dst')\n+                                                     .replace('eth /', 'eth dst is 00:00:00:00:00:01 /')\n+                                                     .replace(\"Ether()\", \"Ether(dst='00:00:00:00:00:01')\")\n+                                         )\n+                                    for element in tvs_mac_ipv6_fragment_fdir]\n+\n+tvs_mac_ipv6_frag_fdir_with_l2 = tvs_mac_ipv6_fragment_fdir_l2dst\n+\n+tvs_mac_ipv6_fragment_fdir_l3src = [eval(str(element).replace('mac_ipv6_frag', 'mac_ipv6_frag_l3src')\n+                                                     .replace('/ ipv6 /', '/ ipv6 src is 2001::1 /')\n+                                                     .replace(\"IPv6()\", \"IPv6(src='2001::1')\"))\n+                                    for element in tvs_mac_ipv6_fragment_fdir]\n+\n+tvs_mac_ipv6_fragment_fdir_l3dst = [eval(str(element).replace('mac_ipv6_frag', 'mac_ipv6_frag_l3dst')\n+                                                     .replace('/ ipv6 /', '/ ipv6 dst is 2001::2 /')\n+                                                     .replace(\"IPv6()\", \"IPv6(dst='2001::2')\"))\n+                                    for element in tvs_mac_ipv6_fragment_fdir]\n+\n+tvs_mac_ipv6_frag_fdir_with_l3 = tvs_mac_ipv6_fragment_fdir_l3src + tvs_mac_ipv6_fragment_fdir_l3dst\n+\n+tv_rss_basic_packets = {\n+    'ipv4_rss_fragment':\n+        \"Ether(src='00:11:22:33:44:55', dst='00:11:22:33:55:66')/IP(src='192.168.6.11', dst='10.11.12.13', id=47750)/Raw('X'*666)\",\n+    'ipv6_rss_fragment':\n+        \"Ether(src='00:11:22:33:44:55', dst='00:11:22:33:55:66')/IPv6(src='CDCD:910A:2222:5498:8475:1111:3900:1537', dst='CDCD:910A:2222:5498:8475:1111:3900:2020')/IPv6ExtHdrFragment(id=47750)/Raw('X'*666)\"\n+}\n+\n+tv_mac_ipv4_fragment_rss = {\n+    'sub_casename': 'tv_mac_ipv4_fragment_rss',\n+    'rule': 'flow create 0 ingress pattern eth / ipv4 / end actions rss types eth ipv4-frag end key_len 0 queues end / end',\n+    'port_id': 0,\n+    'test': [\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv4_rss_fragment'],\n+            'action': {'save_hash': 'ipv4'},\n+        },\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv4_rss_fragment'].replace('00:11:22:33:44:55', '00:11:22:33:44:66'),\n+            'action': {'check_hash_different': 'ipv4'},\n+        },\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv4_rss_fragment'].replace('192.168.6.11', '192.168.6.12'),\n+            'action': {'check_hash_different': 'ipv4'},\n+        },\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv4_rss_fragment'].replace('10.11.12.13', '10.11.12.14'),\n+            'action': {'check_hash_different': 'ipv4'},\n+        },\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv4_rss_fragment'].replace('id=47750', 'id=47751'),\n+            'action': {'check_hash_different': 'ipv4'},\n+        },\n+        {\n+            'send_packet': \"Ether()/IPv6()/IPv6ExtHdrFragment(id=47751)/Raw('X'*666)\",\n+            'action': {'check_no_hash': 'ipv4'},\n+        },\n+    ],\n+    'post-test': [\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv4_rss_fragment'],\n+            'action': {'check_no_hash': 'ipv4'},\n+        },\n+    ]\n+}\n+\n+tv_mac_ipv6_fragment_rss = {\n+    'sub_casename': 'tv_mac_ipv6_fragment_rss',\n+    'rule': 'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext / end actions rss types eth ipv6-frag end key_len 0 queues end / end',\n+    'port_id': 0,\n+    'test': [\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv6_rss_fragment'],\n+            'action': {'save_hash': 'ipv6'},\n+        },\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv6_rss_fragment'].replace('00:11:22:33:44:55', '00:11:22:33:44:66'),\n+            'action': {'check_hash_different': 'ipv6'},\n+        },\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv6_rss_fragment'].replace('3900:1537', '3900:1538'),\n+            'action': {'check_hash_different': 'ipv6'},\n+        },\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv6_rss_fragment'].replace('3900:2020', '3900:2021'),\n+            'action': {'check_hash_different': 'ipv6'},\n+        },\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv6_rss_fragment'].replace('id=47750', 'id=47751'),\n+            'action': {'check_hash_different': 'ipv6'},\n+        },\n+        {\n+            'send_packet': \"Ether()/IP(id=47750)/Raw('X'*666)\",\n+            'action': {'check_no_hash': 'ipv6'},\n+        },\n+    ],\n+    'post-test': [\n+        {\n+            'send_packet': tv_rss_basic_packets['ipv6_rss_fragment'],\n+            'action': {'check_no_hash': 'ipv6'},\n+        },\n+    ]\n+}\n+\n+\n+class TestIpFragment(TestCase):\n+    def set_up_all(self):\n+        self.ports = self.dut.get_ports(self.nic)\n+\n+        # init pkt\n+        self.pkt = Packet()\n+        # set default app parameter\n+        self.pmd_out = PmdOutput(self.dut)\n+        self.tester_mac = self.tester.get_mac(0)\n+        self.tester_port0 = self.tester.get_local_port(self.ports[0])\n+        self.tester_iface0 = self.tester.get_interface(self.tester_port0)\n+\n+        self.tester.send_expect('ifconfig {} up'.format(self.tester_iface0), '# ')\n+        self.param = '--rxq={} --txq={} --disable-rss --txd=384 --rxd=384'.format(LAUNCH_QUEUE, LAUNCH_QUEUE)\n+        self.param_fdir = '--rxq={} --txq={}'.format(LAUNCH_QUEUE, LAUNCH_QUEUE)\n+        self.vf_flag = False\n+        self.cores = self.dut.get_core_list(\"1S/4C/1T\")\n+\n+        self.ports_pci = [self.dut.ports_info[self.ports[0]]['pci']]\n+\n+        self.rssprocess = rfc.RssProcessing(self, self.pmd_out, [self.tester_iface0], LAUNCH_QUEUE)\n+\n+    def set_up(self):\n+        self.dut.bind_interfaces_linux('vfio-pci')\n+\n+    def setup_1pf_vfs_env(self):\n+        \"\"\"\n+        create vf and set vf mac\n+        \"\"\"\n+        self.vf_flag = True\n+        self.dut.bind_interfaces_linux('ice')\n+        self.pf_interface = self.dut.ports_info[0]['intf']\n+        self.dut.send_expect(\"ifconfig {} up\".format(self.pf_interface), \"# \")\n+        self.dut.generate_sriov_vfs_by_port(self.ports[0], 1, driver=self.kdriver)\n+        self.dut.send_expect('ip link set {} vf 0 mac 00:11:22:33:55:66'.format(self.pf_interface), '# ')\n+        self.vf_port = self.dut.ports_info[0]['vfs_port']\n+        self.verify(len(self.vf_port) != 0, \"VF create failed\")\n+        self.vf_driver = self.get_suite_cfg()['vf_driver']\n+        if self.vf_driver is None:\n+            self.vf_assign_method = 'vfio-pci'\n+        self.vf_port[0].bind_driver(self.vf_driver)\n+\n+        self.vf_ports_pci = [self.vf_port[0].pci]\n+\n+    def launch_testpmd(self, param_fdir=False):\n+        \"\"\"\n+        start testpmd with fdir or rss param, and pf or vf\n+\n+        :param param_fdir: True: Fdir param/False: rss param\n+        \"\"\"\n+        if param_fdir == True:\n+            if self.vf_flag == True:\n+                self.pmd_out.start_testpmd(cores=self.cores, ports=self.vf_ports_pci, param=self.param_fdir)\n+            else:\n+                self.pmd_out.start_testpmd(cores=self.cores, ports=self.ports_pci, param=self.param_fdir)\n+        else:\n+            if self.vf_flag == True:\n+                self.pmd_out.start_testpmd(cores=self.cores, ports=self.vf_ports_pci, param=self.param)\n+            else:\n+                self.pmd_out.start_testpmd(cores=self.cores, ports=self.ports_pci, param=self.param)\n+        self.dut.send_expect(\"set fwd rxonly\", \"testpmd> \")\n+        self.dut.send_expect(\"set verbose 1\", \"testpmd> \")\n+        self.dut.send_expect(\"start\", \"testpmd> \")\n+\n+    def _rte_flow_validate(self, vectors):\n+        \"\"\"\n+        FDIR test: validate/create rule, check match pkts and unmatched pkts, destroy rule...\n+\n+        :param vectors: test vectors\n+        \"\"\"\n+        test_results = {}\n+        for tv in vectors:\n+            try:\n+                self.logger.info(\"====================sub_case: {}=========================\".format(tv[\"name\"]))\n+                port_id = tv[\"check_param\"][\"port_id\"] if tv[\"check_param\"].get(\"port_id\") is not None else 0\n+                drop = tv[\"check_param\"].get(\"drop\")\n+                # create rule\n+                self.dut.send_expect(\"flow flush %d\" % port_id, \"testpmd> \", 120)\n+                rule_li = self.rssprocess.create_rule(tv[\"rule\"])\n+                # send and check match packets\n+                out1 = self.send_pkts_getouput(pkts=tv[\"scapy_str\"][\"matched\"], port_id=port_id, drop=drop)\n+                matched_queue = rfc.check_mark(out1, pkt_num=len(tv[\"scapy_str\"][\"matched\"])*2, check_param=tv[\n+                    \"check_param\"])\n+\n+                # send and check unmatched packets\n+                out2 = self.send_pkts_getouput(pkts=tv[\"scapy_str\"][\"unmatched\"], port_id=port_id, drop=drop)\n+                rfc.check_mark(out2, pkt_num=len(tv[\"scapy_str\"][\"unmatched\"])*2, check_param=tv[\"check_param\"], stats=False)\n+\n+                # list and destroy rule\n+                self.rssprocess.check_rule(port_id=tv[\"check_param\"][\"port_id\"], rule_list=rule_li)\n+                self.rssprocess.destroy_rule(rule_id=rule_li, port_id=port_id)\n+                # send matched packet\n+                out3 = self.send_pkts_getouput(pkts=tv[\"scapy_str\"][\"matched\"], port_id=port_id, drop=drop)\n+                matched_queue2 = rfc.check_mark(out3, pkt_num=len(tv[\"scapy_str\"][\"matched\"])*2,\n+                                                check_param=tv[\"check_param\"], stats=False)\n+                if tv[\"check_param\"].get(\"rss\"):\n+                    self.verify(matched_queue == matched_queue2 and None not in matched_queue,\n+                                \"send twice matched packet, received in deferent queues\")\n+                # check not rule exists\n+                self.rssprocess.check_rule(port_id=port_id, stats=False)\n+                test_results[tv[\"name\"]] = True\n+                self.logger.info((GREEN(\"case passed: %s\" % tv[\"name\"])))\n+            except Exception as e:\n+                self.logger.warning((RED(e)))\n+                self.dut.send_command(\"flow flush 0\", timeout=1)\n+                test_results[tv[\"name\"]] = False\n+                self.logger.info((GREEN(\"case failed: %s\" % tv[\"name\"])))\n+                continue\n+        self.output_result_list(test_results)\n+\n+    def _rss_distribute_validate(self, vectors):\n+        \"\"\"\n+        RSS test: validate/create rule, check different hash values, destroy rule...\n+\n+        :param vectors: test vectors\n+        \"\"\"\n+        test_results = {}\n+        if not isinstance(vectors, list):\n+            vectors = [vectors]\n+        for vector in vectors:\n+            try:\n+                self.rssprocess.error_msgs = []\n+                self.logger.info(\"====================sub_case: {}=======================\".format(vector[\"sub_casename\"]))\n+                port_id = vector.get(\"port_id\") if vector.get(\"port_id\") is not None else 0\n+                rules = vector.get('rule') if vector.get('rule') else []\n+                rule_ids = []\n+                # handle tests\n+                tests = vector['test']\n+                self.logger.info('------------handle test--------------')\n+                # validate rule\n+                if rules:\n+                    self.rssprocess.validate_rule(rule=rules, check_stats=True)\n+                    rule_ids = self.rssprocess.create_rule(rule=vector['rule'], check_stats=True)\n+                    self.rssprocess.check_rule(port_id=port_id, rule_list=rule_ids)\n+                for test in tests:\n+                    out1 = ''\n+                    if 'send_packet' in test:\n+                        out1 = self.send_pkts_getouput(test['send_packet'], port_id)\n+                    if 'action' in test:\n+                        self.rssprocess.handle_actions(out1, test['action'])\n+                # handle post-test\n+                if 'post-test' in vector:\n+                    self.logger.info('------------handle post-test--------------')\n+                    self.rssprocess.destroy_rule(port_id=port_id, rule_id=rule_ids)\n+                    self.rssprocess.check_rule(port_id=port_id, stats=False)\n+                    post_tests = vector['post-test']\n+                    for post_test in post_tests:\n+                        out2 = ''\n+                        if 'send_packet' in post_test:\n+                            out2 = self.send_pkts_getouput(post_test['send_packet'], port_id)\n+                        if 'action' in post_test:\n+                            self.rssprocess.handle_actions(out2, post_test['action'])\n+                if self.rssprocess.error_msgs:\n+                    self.verify(False, str(self.rssprocess.error_msgs[:500]))\n+            except Exception as e:\n+                self.logger.warning('sub_case %s failed: %s' % (vector['sub_casename'], e))\n+                test_results[vector['sub_casename']] = False\n+            else:\n+                pass\n+                self.logger.info('sub_case %s passed' % vector['sub_casename'])\n+                test_results[vector['sub_casename']] = True\n+            finally:\n+                self.pmd_out.execute_cmd('flow flush 0')\n+        self.output_result_list(test_results)\n+\n+    def output_result_list(self, result_dic):\n+        \"\"\"\n+        verified test result pass or failed\n+\n+        :param result_dic: FDIR or RSS result dic\n+        \"\"\"\n+        failed_cases = []\n+        for k, v in list(result_dic.items()):\n+            if not v:\n+                failed_cases.append(k)\n+        self.verify(all(result_dic.values()), \"{} failed\".format(failed_cases))\n+\n+    def send_pkts_getouput(self, pkts, port_id=0, drop=False):\n+        \"\"\"\n+        send fragment pkts and get output\n+\n+        :param pkts: pkts list\n+        :param port_id: default port_id = 0\n+        :param drop: if validate drop, need re-start testpmd to show output\n+        :return: return pmd output\n+        \"\"\"\n+        tx_port = self.tester_iface0 if port_id == 0 else self.tester_iface1\n+        time.sleep(1)\n+        if isinstance(pkts, str):\n+            pkts = [pkts]\n+        if drop:\n+            self.pmd_out.execute_cmd(\"clear port stats all\")\n+            time.sleep(1)\n+            self.send_pkt(pkts, tx_port=tx_port)\n+            out = self.pmd_out.execute_cmd(\"stop\")\n+            self.pmd_out.execute_cmd(\"start\")\n+        else:\n+            self.send_pkt(pkts, tx_port=tx_port)\n+            out = self.pmd_out.get_output()\n+        return out\n+\n+    def destroy_testpmd_and_vf(self):\n+        \"\"\"\n+        quit testpmd\n+        if vf testpmd, destroy the vfs and set vf_flag = false\n+        \"\"\"\n+        self.dut.send_expect(\"quit\", \"# \")\n+        if self.vf_flag == True:\n+            for port_id in self.ports:\n+                self.dut.destroy_sriov_vfs_by_port(port_id)\n+            self.vf_flag = False\n+\n+    def send_pkt(self, pkt, tx_port):\n+        \"\"\"\n+        send fragment pkts\n+\n+        :param pkt: pkts list\n+        :param tx_port: send pkts tester port\n+        \"\"\"\n+        self.logger.info(\"                  Send pkt:{}\".format(pkt))\n+        for i in range(len(pkt)):\n+            self.tester.scapy_session.send_expect(\n+                'p=eval(\"{}\")'.format(pkt[i]), '>>> ')\n+\n+            if 'IPv6ExtHdrFragment' in pkt[i]:\n+                self.tester.scapy_session.send_expect(\n+                    'pkts=fragment6(p, 500)', '>>> ')\n+            else:\n+                self.tester.scapy_session.send_expect(\n+                    'pkts=fragment(p, fragsize=500)', '>>> ')\n+            self.tester.scapy_session.send_expect(\n+                'sendp(pkts, iface=\"{}\")'.format(tx_port), '>>> ')\n+\n+    def tear_down(self):\n+        self.destroy_testpmd_and_vf()\n+        self.dut.kill_all()\n+\n+    def tear_down_all(self):\n+        self.dut.kill_all()\n+\n+    def test_mac_ipv4_frag_pf_fdir(self):\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv4_fragment_fdir)\n+\n+    def test_mac_ipv6_frag_pf_fdir(self):\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv6_fragment_fdir)\n+\n+    def test_mac_ipv4_frag_pf_fdir_with_l2(self):\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv4_frag_fdir_with_l2)\n+\n+    def test_mac_ipv4_frag_pf_fdir_with_l3(self):\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv4_frag_fdir_with_l3)\n+\n+    def test_mac_ipv6_frag_pf_fdir_with_l2(self):\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv6_frag_fdir_with_l2)\n+\n+    def test_mac_ipv6_frag_pf_fdir_with_l3(self):\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv6_frag_fdir_with_l3)\n+\n+    def test_mac_ipv4_frag_vf_fdir(self):\n+        self.setup_1pf_vfs_env()\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv4_fragment_fdir)\n+\n+    def test_mac_ipv6_frag_vf_fdir(self):\n+        self.setup_1pf_vfs_env()\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv6_fragment_fdir)\n+\n+    def test_mac_ipv4_frag_vf_fdir_with_l3(self):\n+        self.setup_1pf_vfs_env()\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv4_frag_fdir_with_l3)\n+\n+    def test_mac_ipv6_frag_vf_fdir_with_l3(self):\n+        self.setup_1pf_vfs_env()\n+        self.launch_testpmd(param_fdir=True)\n+        self._rte_flow_validate(tvs_mac_ipv6_frag_fdir_with_l3)\n+\n+    def test_mac_ipv4_frag_pf_rss(self):\n+        self.launch_testpmd(param_fdir=False)\n+        self._rss_distribute_validate(tv_mac_ipv4_fragment_rss)\n+\n+    def test_mac_ipv4_frag_vf_rss(self):\n+        self.setup_1pf_vfs_env()\n+        self.launch_testpmd(param_fdir=False)\n+        self._rss_distribute_validate(tv_mac_ipv4_fragment_rss)\n+\n+    def test_mac_ipv6_frag_pf_rss(self):\n+        self.launch_testpmd(param_fdir=False)\n+        self._rss_distribute_validate(tv_mac_ipv6_fragment_rss)\n+\n+    def test_mac_ipv6_frag_vf_rss(self):\n+        self.setup_1pf_vfs_env()\n+        self.launch_testpmd(param_fdir=False)\n+        self._rss_distribute_validate(tv_mac_ipv6_fragment_rss)\n+\n+    def exclusive_validation(self):\n+        result = True\n+        result_list = []\n+        rule_list_pf = [\n+            'flow create 0 ingress pattern eth dst is 00:11:22:33:44:55 / ipv4 src is 192.168.0.20 / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions queue index 2 / end']\n+        rule_list_vf = [\n+            'flow create 0 ingress pattern eth / ipv4 src is 192.168.0.20 / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions queue index 2 / end']\n+        pkt_pf = \"Ether(dst='00:11:22:33:44:55')/IP(src='192.168.0.20', id=47750)/Raw('X'*666)\"\n+        pkt_vf = \"Ether()/IP(src='192.168.0.20', id=47750)/Raw('X'*666)\"\n+\n+        self.logger.info('Subcase 1: exclusive validation fdir rule')\n+        self.launch_testpmd(param_fdir=True)\n+        try:\n+            if self.vf_flag:\n+                self.rssprocess.create_rule(rule_list_vf)\n+            else:\n+                self.rssprocess.create_rule(rule_list_pf)\n+        except Exception as e:\n+            self.logger.warning('Subcase 1 failed: %s' % e)\n+            result = False\n+        pkt = pkt_vf if self.vf_flag else pkt_pf\n+        out = self.send_pkts_getouput(pkt)\n+        hashes, queues = self.rssprocess.get_hash_and_queues(out)\n+        for queue in queues:\n+            if '0x2' != queue:\n+                result = False\n+                self.logger.error('Error: queue index {} != 2'.format(queue))\n+                continue\n+        result_list.append(result)\n+        self.dut.send_expect(\"quit\", \"# \")\n+        self.logger.info(\"*********subcase test result %s\" % result_list)\n+\n+        self.logger.info('Subcase 2: exclusive validation fdir rule')\n+        result = True\n+        self.launch_testpmd(param_fdir=True)\n+        try:\n+            if self.vf_flag:\n+                self.rssprocess.create_rule(rule_list_vf[::-1])\n+            else:\n+                self.rssprocess.create_rule(rule_list_pf[::-1])\n+        except Exception as e:\n+            self.logger.warning('Subcase 2 failed: %s' % e)\n+            result = False\n+        pkt = pkt_vf if self.vf_flag else pkt_pf\n+        out = self.send_pkts_getouput(pkt)\n+        hashes, queues = self.rssprocess.get_hash_and_queues(out)\n+        for queue in queues:\n+            if '0x2' != queue:\n+                result = False\n+                self.logger.error('Error: queue index {} != 2'.format(queue))\n+                continue\n+        result_list.append(result)\n+        self.dut.send_expect(\"quit\", \"# \")\n+        self.logger.info(\"*********subcase test result %s\" % result_list)\n+\n+        self.logger.info('Subcase 3: exclusive validation rss rule')\n+        result = True\n+        self.launch_testpmd()\n+        rule_list = [\n+            'flow create 0 ingress pattern eth / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end',\n+            'flow create 0 ingress pattern eth / ipv4 / end actions rss types ipv4-frag end key_len 0 queues end / end']\n+        pkt = [\"Ether()/IP(id=47750)/Raw('X'*666)\",\n+               \"Ether()/IP(id=47751)/Raw('X'*666)\"]\n+        try:\n+            self.rssprocess.create_rule(rule_list)\n+        except Exception as e:\n+            self.logger.warning('Subcase 3 failed: %s' % e)\n+            result = False\n+        out1 = self.send_pkts_getouput(pkt[0])\n+        hashes1, queues1 = self.rssprocess.get_hash_and_queues(out1)\n+        out2 = self.send_pkts_getouput(pkt[1])\n+        hashes2, queues2 = self.rssprocess.get_hash_and_queues(out2)\n+        if hashes1[0] != hashes1[1] and hashes2[0] != hashes2[1]:\n+            result = False\n+            self.logger.error(\"hash value is incorrect\")\n+        if hashes1[0] == hashes2[0]:\n+            result = False\n+            self.logger.error(\"hash value is incorrect\")\n+        result_list.append(result)\n+        self.dut.send_expect(\"quit\", \"# \")\n+        self.logger.info(\"*********subcase test result %s\" % result_list)\n+\n+        self.logger.info('Subcase 4: exclusive validation rss rule')\n+        result = True\n+        self.launch_testpmd()\n+        rule_list = [\n+            'flow create 0 ingress pattern eth / ipv4 / end actions rss types ipv4-frag end key_len 0 queues end / end',\n+            'flow create 0 ingress pattern eth / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end'\n+        ]\n+        pkt = [\"Ether()/IP(id=47750)/Raw('X'*666)\",\n+               \"Ether()/IP(id=47751)/Raw('X'*666)\"]\n+        try:\n+            self.rssprocess.create_rule(rule_list)\n+        except Exception as e:\n+            self.logger.warning('Subcase 3 failed: %s' % e)\n+        out1 = self.send_pkts_getouput(pkt[0])\n+        hashes1, queues1 = self.rssprocess.get_hash_and_queues(out1)\n+        out2 = self.send_pkts_getouput(pkt[1])\n+        hashes2, queues2 = self.rssprocess.get_hash_and_queues(out2)\n+        if hashes1[0] != hashes1[1] and hashes2[0] != hashes2[1]:\n+            result = False\n+            self.logger.error(\"hash value is incorrect\")\n+        if hashes1[0] != hashes2[0]:\n+            result = False\n+            self.logger.error(\"hash value is incorrect\")\n+        result_list.append(result)\n+        self.dut.send_expect(\"quit\", \"# \")\n+        self.logger.info(\"*********subcase test result %s\" % result_list)\n+        self.verify(all(result_list) is True, 'sub-case failed {}'.format(result_list))\n+\n+    def test_pf_exclusive_validation(self):\n+        self.exclusive_validation()\n+\n+    def test_vf_exclusive_validation(self):\n+        self.setup_1pf_vfs_env()\n+        self.exclusive_validation()\n+\n+    def test_negative_case(self):\n+        negative_rules = [\n+            'flow create 0 ingress pattern eth / ipv6 packet_id is 47750 fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 47750 fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions queue index 300 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 0x10000 fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 47750 fragment_offset spec 0x2 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 47750 fragment_offset spec 0x2000 fragment_offset last 0x1 fragment_offset mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 47750 fragment_offset spec 0x2000 fragment_offset last 0x1fff fragment_offset mask 0xf / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 47750 fragment_offset last 0x1fff fragment_offset mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 47750 fragment_offset spec 0x2000 fragment_offset mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 47750 fragment_offset spec 0x2000 fragment_offset last 0x1fff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 0x10000 / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id is 47750 / end actions queue index 300 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id spec 0xfff packet_id last 0x0 packet_id mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id last 0xffff packet_id mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 packet_id spec 0 packet_id last 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 / ipv6_frag_ext packet_id is 47750 frag_data spec 0x0001 frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id is 47750 frag_data spec 0xfff8 frag_data last 0x0001 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / frag_data spec 0x0001 frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id is 47750 frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id is 47750 frag_data spec 0x0001 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id is 47750 frag_data spec 0x0001 frag_data last 0xfff8 / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id is 47750 frag_data spec 0x0001 frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 300 / end',\n+            'flow create 0 ingress pattern eth / ipv4 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff frag_data spec 0x0001 frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0xffff packet_id last 0x0 packet_id mask 0xffff frag_data spec 0x0001 frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff frag_data spec 0xfff8 frag_data last 0x0001 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / packet_id last 0xffff packet_id mask 0xffff frag_data spec 0x0001 frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id mask 0xffff frag_data spec 0x0001 frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffff frag_data spec 0x0001 frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff frag_data last 0xfff8 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff frag_data spec 0x0001 frag_data last 0xfff8 / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id spec 0 packet_id last 0xffff packet_id mask 0xffff frag_data spec 0x0001 frag_data mask 0xffff / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv4 / ipv6_frag_ext packet_id is 47750 / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext packet_id is 0x10000 / end actions queue index 1 / end',\n+            'flow create 0 ingress pattern eth / ipv6 / end actions rss types ipv4-frag end key_len 0 queues end / end',\n+            'flow create 0 ingress pattern eth / ipv4 / ipv6_frag_ext / end actions rss types ipv6-frag end key_len 0 queues end / end',\n+            'flow create 0 ingress pattern eth / ipv6 / ipv6_frag_ext / end actions rss types ipv4-frag end key_len 0 queues end / end',\n+        ]\n+        self.launch_testpmd()\n+        self.rssprocess.create_rule(negative_rules, check_stats=False)\n",
    "prefixes": [
        "V1"
    ]
}