get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/83959/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 83959,
    "url": "http://patches.dpdk.org/api/patches/83959/?format=api",
    "web_url": "http://patches.dpdk.org/project/dpdk/patch/20201111064936.768604-17-jiawenwu@trustnetic.com/",
    "project": {
        "id": 1,
        "url": "http://patches.dpdk.org/api/projects/1/?format=api",
        "name": "DPDK",
        "link_name": "dpdk",
        "list_id": "dev.dpdk.org",
        "list_email": "dev@dpdk.org",
        "web_url": "http://core.dpdk.org",
        "scm_url": "git://dpdk.org/dpdk",
        "webscm_url": "http://git.dpdk.org/dpdk",
        "list_archive_url": "https://inbox.dpdk.org/dev",
        "list_archive_url_format": "https://inbox.dpdk.org/dev/{}",
        "commit_url_format": ""
    },
    "msgid": "<20201111064936.768604-17-jiawenwu@trustnetic.com>",
    "list_archive_url": "https://inbox.dpdk.org/dev/20201111064936.768604-17-jiawenwu@trustnetic.com",
    "date": "2020-11-11T06:49:15",
    "name": "[v2,16/37] net/txgbe: add FDIR parse normal rule",
    "commit_ref": null,
    "pull_url": null,
    "state": "changes-requested",
    "archived": true,
    "hash": "4d759e10e19eb28d963f0c3daaf2b24f9ed04024",
    "submitter": {
        "id": 1932,
        "url": "http://patches.dpdk.org/api/people/1932/?format=api",
        "name": "Jiawen Wu",
        "email": "jiawenwu@trustnetic.com"
    },
    "delegate": {
        "id": 319,
        "url": "http://patches.dpdk.org/api/users/319/?format=api",
        "username": "fyigit",
        "first_name": "Ferruh",
        "last_name": "Yigit",
        "email": "ferruh.yigit@amd.com"
    },
    "mbox": "http://patches.dpdk.org/project/dpdk/patch/20201111064936.768604-17-jiawenwu@trustnetic.com/mbox/",
    "series": [
        {
            "id": 13798,
            "url": "http://patches.dpdk.org/api/series/13798/?format=api",
            "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=13798",
            "date": "2020-11-11T06:49:00",
            "name": "net: add txgbe PMD part 2",
            "version": 2,
            "mbox": "http://patches.dpdk.org/series/13798/mbox/"
        }
    ],
    "comments": "http://patches.dpdk.org/api/patches/83959/comments/",
    "check": "warning",
    "checks": "http://patches.dpdk.org/api/patches/83959/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<dev-bounces@dpdk.org>",
        "X-Original-To": "patchwork@inbox.dpdk.org",
        "Delivered-To": "patchwork@inbox.dpdk.org",
        "Received": [
            "from dpdk.org (dpdk.org [92.243.14.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id A4A3BA09D2;\n\tWed, 11 Nov 2020 07:53:25 +0100 (CET)",
            "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id 3E8B1BE43;\n\tWed, 11 Nov 2020 07:48:01 +0100 (CET)",
            "from smtpbgau1.qq.com (smtpbgau1.qq.com [54.206.16.166])\n by dpdk.org (Postfix) with ESMTP id DD62A6CC0\n for <dev@dpdk.org>; Wed, 11 Nov 2020 07:47:47 +0100 (CET)",
            "from localhost.localdomain.com (unknown [183.129.236.74])\n by esmtp10.qq.com (ESMTP) with\n id ; Wed, 11 Nov 2020 14:47:41 +0800 (CST)"
        ],
        "X-QQ-mid": "bizesmtp27t1605077261t0x24jjd",
        "X-QQ-SSF": "01400000000000C0C000B00A0000000",
        "X-QQ-FEAT": "YKCDl5A3/aqLQvI37sMIuaKTjXv5ZO6evzKYrNRQPd5rsD/KhxSfQ8jUr35zz\n mogv0EHGi19SaLgx7zh2MOwUZW50V5J3YuTDaJwb8Xf08eB2nmE2/DYnEjMIPB6zJGxedzO\n zeYXeMoMbKG7HUFfLuVk/HnXtn+6+9pL3saT6YdVBtt7gUWfciQQkcIxBzvAigKATyQ+0EK\n o0Ff4WsUwAOp9xJv84PulgTGPdGqLmlGPrD7k/6hUdJlOVZSFzNWmr+W4bXEDpilsl7Rgys\n NEug+EvyzvGE7e9YUmbSREHzbDJ4d9QXIn4pY67yRGWwh2HrC0ISJ+yRu6sYrl4ib3Mwf8v\n 7zeU/JLhBUSDALrGwogc3qRrv8Teg==",
        "X-QQ-GoodBg": "2",
        "From": "Jiawen Wu <jiawenwu@trustnetic.com>",
        "To": "dev@dpdk.org",
        "Cc": "Jiawen Wu <jiawenwu@trustnetic.com>",
        "Date": "Wed, 11 Nov 2020 14:49:15 +0800",
        "Message-Id": "<20201111064936.768604-17-jiawenwu@trustnetic.com>",
        "X-Mailer": "git-send-email 2.18.4",
        "In-Reply-To": "<20201111064936.768604-1-jiawenwu@trustnetic.com>",
        "References": "<20201111064936.768604-1-jiawenwu@trustnetic.com>",
        "X-QQ-SENDSIZE": "520",
        "Feedback-ID": "bizesmtp:trustnetic.com:qybgforeign:qybgforeign6",
        "X-QQ-Bgrelay": "1",
        "Subject": "[dpdk-dev] [PATCH v2 16/37] net/txgbe: add FDIR parse normal rule",
        "X-BeenThere": "dev@dpdk.org",
        "X-Mailman-Version": "2.1.15",
        "Precedence": "list",
        "List-Id": "DPDK patches and discussions <dev.dpdk.org>",
        "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n <mailto:dev-request@dpdk.org?subject=unsubscribe>",
        "List-Archive": "<http://mails.dpdk.org/archives/dev/>",
        "List-Post": "<mailto:dev@dpdk.org>",
        "List-Help": "<mailto:dev-request@dpdk.org?subject=help>",
        "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n <mailto:dev-request@dpdk.org?subject=subscribe>",
        "Errors-To": "dev-bounces@dpdk.org",
        "Sender": "\"dev\" <dev-bounces@dpdk.org>"
    },
    "content": "Add support to parse flow for fdir filter.\n\nSigned-off-by: Jiawen Wu <jiawenwu@trustnetic.com>\n---\n drivers/net/txgbe/base/txgbe_type.h |  10 +\n drivers/net/txgbe/txgbe_flow.c      | 856 ++++++++++++++++++++++++++++\n 2 files changed, 866 insertions(+)",
    "diff": "diff --git a/drivers/net/txgbe/base/txgbe_type.h b/drivers/net/txgbe/base/txgbe_type.h\nindex 160d5253a..a73f66d39 100644\n--- a/drivers/net/txgbe/base/txgbe_type.h\n+++ b/drivers/net/txgbe/base/txgbe_type.h\n@@ -83,6 +83,16 @@ enum {\n #define TXGBE_ATR_L4TYPE_SCTP\t\t\t0x3\n #define TXGBE_ATR_TUNNEL_MASK\t\t\t0x10\n #define TXGBE_ATR_TUNNEL_ANY\t\t\t0x10\n+enum txgbe_atr_flow_type {\n+\tTXGBE_ATR_FLOW_TYPE_IPV4\t\t= 0x0,\n+\tTXGBE_ATR_FLOW_TYPE_UDPV4\t\t= 0x1,\n+\tTXGBE_ATR_FLOW_TYPE_TCPV4\t\t= 0x2,\n+\tTXGBE_ATR_FLOW_TYPE_SCTPV4\t\t= 0x3,\n+\tTXGBE_ATR_FLOW_TYPE_IPV6\t\t= 0x4,\n+\tTXGBE_ATR_FLOW_TYPE_UDPV6\t\t= 0x5,\n+\tTXGBE_ATR_FLOW_TYPE_TCPV6\t\t= 0x6,\n+\tTXGBE_ATR_FLOW_TYPE_SCTPV6\t\t= 0x7,\n+};\n \n /* Flow Director ATR input struct. */\n struct txgbe_atr_input {\ndiff --git a/drivers/net/txgbe/txgbe_flow.c b/drivers/net/txgbe/txgbe_flow.c\nindex 8589e3328..ba1be9f12 100644\n--- a/drivers/net/txgbe/txgbe_flow.c\n+++ b/drivers/net/txgbe/txgbe_flow.c\n@@ -27,6 +27,7 @@\n \n #define TXGBE_MIN_N_TUPLE_PRIO 1\n #define TXGBE_MAX_N_TUPLE_PRIO 7\n+#define TXGBE_MAX_FLX_SOURCE_OFF 62\n \n /**\n  * Endless loop will never happen with below assumption\n@@ -1242,3 +1243,858 @@ txgbe_parse_l2_tn_filter(struct rte_eth_dev *dev,\n \treturn ret;\n }\n \n+/* Parse to get the attr and action info of flow director rule. */\n+static int\n+txgbe_parse_fdir_act_attr(const struct rte_flow_attr *attr,\n+\t\t\t  const struct rte_flow_action actions[],\n+\t\t\t  struct txgbe_fdir_rule *rule,\n+\t\t\t  struct rte_flow_error *error)\n+{\n+\tconst struct rte_flow_action *act;\n+\tconst struct rte_flow_action_queue *act_q;\n+\tconst struct rte_flow_action_mark *mark;\n+\n+\t/* parse attr */\n+\t/* must be input direction */\n+\tif (!attr->ingress) {\n+\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_ATTR_INGRESS,\n+\t\t\tattr, \"Only support ingress.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\t/* not supported */\n+\tif (attr->egress) {\n+\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_ATTR_EGRESS,\n+\t\t\tattr, \"Not support egress.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\t/* not supported */\n+\tif (attr->transfer) {\n+\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_ATTR_TRANSFER,\n+\t\t\tattr, \"No support for transfer.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\t/* not supported */\n+\tif (attr->priority) {\n+\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_ATTR_PRIORITY,\n+\t\t\tattr, \"Not support priority.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\t/* check if the first not void action is QUEUE or DROP. */\n+\tact = next_no_void_action(actions, NULL);\n+\tif (act->type != RTE_FLOW_ACTION_TYPE_QUEUE &&\n+\t    act->type != RTE_FLOW_ACTION_TYPE_DROP) {\n+\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_ACTION,\n+\t\t\tact, \"Not supported action.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\tif (act->type == RTE_FLOW_ACTION_TYPE_QUEUE) {\n+\t\tact_q = (const struct rte_flow_action_queue *)act->conf;\n+\t\trule->queue = act_q->index;\n+\t} else { /* drop */\n+\t\t/* signature mode does not support drop action. */\n+\t\tif (rule->mode == RTE_FDIR_MODE_SIGNATURE) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ACTION,\n+\t\t\t\tact, \"Not supported action.\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\trule->fdirflags = TXGBE_FDIRPICMD_DROP;\n+\t}\n+\n+\t/* check if the next not void item is MARK */\n+\tact = next_no_void_action(actions, act);\n+\tif (act->type != RTE_FLOW_ACTION_TYPE_MARK &&\n+\t\tact->type != RTE_FLOW_ACTION_TYPE_END) {\n+\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_ACTION,\n+\t\t\tact, \"Not supported action.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\trule->soft_id = 0;\n+\n+\tif (act->type == RTE_FLOW_ACTION_TYPE_MARK) {\n+\t\tmark = (const struct rte_flow_action_mark *)act->conf;\n+\t\trule->soft_id = mark->id;\n+\t\tact = next_no_void_action(actions, act);\n+\t}\n+\n+\t/* check if the next not void item is END */\n+\tif (act->type != RTE_FLOW_ACTION_TYPE_END) {\n+\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_ACTION,\n+\t\t\tact, \"Not supported action.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\treturn 0;\n+}\n+\n+/* search next no void pattern and skip fuzzy */\n+static inline\n+const struct rte_flow_item *next_no_fuzzy_pattern(\n+\t\tconst struct rte_flow_item pattern[],\n+\t\tconst struct rte_flow_item *cur)\n+{\n+\tconst struct rte_flow_item *next =\n+\t\tnext_no_void_pattern(pattern, cur);\n+\twhile (1) {\n+\t\tif (next->type != RTE_FLOW_ITEM_TYPE_FUZZY)\n+\t\t\treturn next;\n+\t\tnext = next_no_void_pattern(pattern, next);\n+\t}\n+}\n+\n+static inline uint8_t signature_match(const struct rte_flow_item pattern[])\n+{\n+\tconst struct rte_flow_item_fuzzy *spec, *last, *mask;\n+\tconst struct rte_flow_item *item;\n+\tuint32_t sh, lh, mh;\n+\tint i = 0;\n+\n+\twhile (1) {\n+\t\titem = pattern + i;\n+\t\tif (item->type == RTE_FLOW_ITEM_TYPE_END)\n+\t\t\tbreak;\n+\n+\t\tif (item->type == RTE_FLOW_ITEM_TYPE_FUZZY) {\n+\t\t\tspec = item->spec;\n+\t\t\tlast = item->last;\n+\t\t\tmask = item->mask;\n+\n+\t\t\tif (!spec || !mask)\n+\t\t\t\treturn 0;\n+\n+\t\t\tsh = spec->thresh;\n+\n+\t\t\tif (!last)\n+\t\t\t\tlh = sh;\n+\t\t\telse\n+\t\t\t\tlh = last->thresh;\n+\n+\t\t\tmh = mask->thresh;\n+\t\t\tsh = sh & mh;\n+\t\t\tlh = lh & mh;\n+\n+\t\t\tif (!sh || sh > lh)\n+\t\t\t\treturn 0;\n+\n+\t\t\treturn 1;\n+\t\t}\n+\n+\t\ti++;\n+\t}\n+\n+\treturn 0;\n+}\n+\n+/**\n+ * Parse the rule to see if it is a IP or MAC VLAN flow director rule.\n+ * And get the flow director filter info BTW.\n+ * UDP/TCP/SCTP PATTERN:\n+ * The first not void item can be ETH or IPV4 or IPV6\n+ * The second not void item must be IPV4 or IPV6 if the first one is ETH.\n+ * The next not void item could be UDP or TCP or SCTP (optional)\n+ * The next not void item could be RAW (for flexbyte, optional)\n+ * The next not void item must be END.\n+ * A Fuzzy Match pattern can appear at any place before END.\n+ * Fuzzy Match is optional for IPV4 but is required for IPV6\n+ * MAC VLAN PATTERN:\n+ * The first not void item must be ETH.\n+ * The second not void item must be MAC VLAN.\n+ * The next not void item must be END.\n+ * ACTION:\n+ * The first not void action should be QUEUE or DROP.\n+ * The second not void optional action should be MARK,\n+ * mark_id is a uint32_t number.\n+ * The next not void action should be END.\n+ * UDP/TCP/SCTP pattern example:\n+ * ITEM\t\tSpec\t\t\tMask\n+ * ETH\t\tNULL\t\t\tNULL\n+ * IPV4\t\tsrc_addr 192.168.1.20\t0xFFFFFFFF\n+ *\t\tdst_addr 192.167.3.50\t0xFFFFFFFF\n+ * UDP/TCP/SCTP\tsrc_port\t80\t0xFFFF\n+ *\t\tdst_port\t80\t0xFFFF\n+ * FLEX\trelative\t0\t0x1\n+ *\t\tsearch\t\t0\t0x1\n+ *\t\treserved\t0\t0\n+ *\t\toffset\t\t12\t0xFFFFFFFF\n+ *\t\tlimit\t\t0\t0xFFFF\n+ *\t\tlength\t\t2\t0xFFFF\n+ *\t\tpattern[0]\t0x86\t0xFF\n+ *\t\tpattern[1]\t0xDD\t0xFF\n+ * END\n+ * MAC VLAN pattern example:\n+ * ITEM\t\tSpec\t\t\tMask\n+ * ETH\t\tdst_addr\n+\t\t{0xAC, 0x7B, 0xA1,\t{0xFF, 0xFF, 0xFF,\n+\t\t0x2C, 0x6D, 0x36}\t0xFF, 0xFF, 0xFF}\n+ * MAC VLAN\ttci\t0x2016\t\t0xEFFF\n+ * END\n+ * Other members in mask and spec should set to 0x00.\n+ * Item->last should be NULL.\n+ */\n+static int\n+txgbe_parse_fdir_filter_normal(struct rte_eth_dev *dev __rte_unused,\n+\t\t\t       const struct rte_flow_attr *attr,\n+\t\t\t       const struct rte_flow_item pattern[],\n+\t\t\t       const struct rte_flow_action actions[],\n+\t\t\t       struct txgbe_fdir_rule *rule,\n+\t\t\t       struct rte_flow_error *error)\n+{\n+\tconst struct rte_flow_item *item;\n+\tconst struct rte_flow_item_eth *eth_mask;\n+\tconst struct rte_flow_item_ipv4 *ipv4_spec;\n+\tconst struct rte_flow_item_ipv4 *ipv4_mask;\n+\tconst struct rte_flow_item_ipv6 *ipv6_spec;\n+\tconst struct rte_flow_item_ipv6 *ipv6_mask;\n+\tconst struct rte_flow_item_tcp *tcp_spec;\n+\tconst struct rte_flow_item_tcp *tcp_mask;\n+\tconst struct rte_flow_item_udp *udp_spec;\n+\tconst struct rte_flow_item_udp *udp_mask;\n+\tconst struct rte_flow_item_sctp *sctp_spec;\n+\tconst struct rte_flow_item_sctp *sctp_mask;\n+\tconst struct rte_flow_item_raw *raw_mask;\n+\tconst struct rte_flow_item_raw *raw_spec;\n+\tu32 ptype = 0;\n+\tuint8_t j;\n+\n+\tif (!pattern) {\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_ITEM_NUM,\n+\t\t\tNULL, \"NULL pattern.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\tif (!actions) {\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\t   RTE_FLOW_ERROR_TYPE_ACTION_NUM,\n+\t\t\t\t   NULL, \"NULL action.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\tif (!attr) {\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\t   RTE_FLOW_ERROR_TYPE_ATTR,\n+\t\t\t\t   NULL, \"NULL attribute.\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\t/**\n+\t * Some fields may not be provided. Set spec to 0 and mask to default\n+\t * value. So, we need not do anything for the not provided fields later.\n+\t */\n+\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\tmemset(&rule->mask, 0xFF, sizeof(struct txgbe_hw_fdir_mask));\n+\trule->mask.vlan_tci_mask = 0;\n+\trule->mask.flex_bytes_mask = 0;\n+\n+\t/**\n+\t * The first not void item should be\n+\t * MAC or IPv4 or TCP or UDP or SCTP.\n+\t */\n+\titem = next_no_fuzzy_pattern(pattern, NULL);\n+\tif (item->type != RTE_FLOW_ITEM_TYPE_ETH &&\n+\t    item->type != RTE_FLOW_ITEM_TYPE_IPV4 &&\n+\t    item->type != RTE_FLOW_ITEM_TYPE_IPV6 &&\n+\t    item->type != RTE_FLOW_ITEM_TYPE_TCP &&\n+\t    item->type != RTE_FLOW_ITEM_TYPE_UDP &&\n+\t    item->type != RTE_FLOW_ITEM_TYPE_SCTP) {\n+\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\titem, \"Not supported by fdir filter\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\tif (signature_match(pattern))\n+\t\trule->mode = RTE_FDIR_MODE_SIGNATURE;\n+\telse\n+\t\trule->mode = RTE_FDIR_MODE_PERFECT;\n+\n+\t/*Not supported last point for range*/\n+\tif (item->last) {\n+\t\trte_flow_error_set(error, EINVAL,\n+\t\t\tRTE_FLOW_ERROR_TYPE_UNSPECIFIED,\n+\t\t\titem, \"Not supported last point for range\");\n+\t\treturn -rte_errno;\n+\t}\n+\n+\t/* Get the MAC info. */\n+\tif (item->type == RTE_FLOW_ITEM_TYPE_ETH) {\n+\t\t/**\n+\t\t * Only support vlan and dst MAC address,\n+\t\t * others should be masked.\n+\t\t */\n+\t\tif (item->spec && !item->mask) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\n+\t\tif (item->mask) {\n+\t\t\trule->b_mask = TRUE;\n+\t\t\teth_mask = item->mask;\n+\n+\t\t\t/* Ether type should be masked. */\n+\t\t\tif (eth_mask->type ||\n+\t\t\t    rule->mode == RTE_FDIR_MODE_SIGNATURE) {\n+\t\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\t\treturn -rte_errno;\n+\t\t\t}\n+\n+\t\t\t/* If ethernet has meaning, it means MAC VLAN mode. */\n+\t\t\trule->mode = RTE_FDIR_MODE_PERFECT_MAC_VLAN;\n+\n+\t\t\t/**\n+\t\t\t * src MAC address must be masked,\n+\t\t\t * and don't support dst MAC address mask.\n+\t\t\t */\n+\t\t\tfor (j = 0; j < RTE_ETHER_ADDR_LEN; j++) {\n+\t\t\t\tif (eth_mask->src.addr_bytes[j] ||\n+\t\t\t\t\teth_mask->dst.addr_bytes[j] != 0xFF) {\n+\t\t\t\t\tmemset(rule, 0,\n+\t\t\t\t\tsizeof(struct txgbe_fdir_rule));\n+\t\t\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\t\t\treturn -rte_errno;\n+\t\t\t\t}\n+\t\t\t}\n+\n+\t\t\t/* When no VLAN, considered as full mask. */\n+\t\t\trule->mask.vlan_tci_mask = rte_cpu_to_be_16(0xEFFF);\n+\t\t}\n+\t\t/*** If both spec and mask are item,\n+\t\t * it means don't care about ETH.\n+\t\t * Do nothing.\n+\t\t */\n+\n+\t\t/**\n+\t\t * Check if the next not void item is vlan or ipv4.\n+\t\t * IPv6 is not supported.\n+\t\t */\n+\t\titem = next_no_fuzzy_pattern(pattern, item);\n+\t\tif (rule->mode == RTE_FDIR_MODE_PERFECT_MAC_VLAN) {\n+\t\t\tif (item->type != RTE_FLOW_ITEM_TYPE_VLAN) {\n+\t\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\t\treturn -rte_errno;\n+\t\t\t}\n+\t\t} else {\n+\t\t\tif (item->type != RTE_FLOW_ITEM_TYPE_IPV4 &&\n+\t\t\t\t\titem->type != RTE_FLOW_ITEM_TYPE_VLAN) {\n+\t\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\t\treturn -rte_errno;\n+\t\t\t}\n+\t\t}\n+\t}\n+\n+\t/* Get the IPV4 info. */\n+\tif (item->type == RTE_FLOW_ITEM_TYPE_IPV4) {\n+\t\t/**\n+\t\t * Set the flow type even if there's no content\n+\t\t * as we must have a flow type.\n+\t\t */\n+\t\trule->input.flow_type = TXGBE_ATR_FLOW_TYPE_IPV4;\n+\t\tptype = txgbe_ptype_table[TXGBE_PT_IPV4];\n+\t\t/*Not supported last point for range*/\n+\t\tif (item->last) {\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_UNSPECIFIED,\n+\t\t\t\titem, \"Not supported last point for range\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\t/**\n+\t\t * Only care about src & dst addresses,\n+\t\t * others should be masked.\n+\t\t */\n+\t\tif (!item->mask) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\trule->b_mask = TRUE;\n+\t\tipv4_mask = item->mask;\n+\t\tif (ipv4_mask->hdr.version_ihl ||\n+\t\t    ipv4_mask->hdr.type_of_service ||\n+\t\t    ipv4_mask->hdr.total_length ||\n+\t\t    ipv4_mask->hdr.packet_id ||\n+\t\t    ipv4_mask->hdr.fragment_offset ||\n+\t\t    ipv4_mask->hdr.time_to_live ||\n+\t\t    ipv4_mask->hdr.next_proto_id ||\n+\t\t    ipv4_mask->hdr.hdr_checksum) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\trule->mask.dst_ipv4_mask = ipv4_mask->hdr.dst_addr;\n+\t\trule->mask.src_ipv4_mask = ipv4_mask->hdr.src_addr;\n+\n+\t\tif (item->spec) {\n+\t\t\trule->b_spec = TRUE;\n+\t\t\tipv4_spec = item->spec;\n+\t\t\trule->input.dst_ip[0] =\n+\t\t\t\tipv4_spec->hdr.dst_addr;\n+\t\t\trule->input.src_ip[0] =\n+\t\t\t\tipv4_spec->hdr.src_addr;\n+\t\t}\n+\n+\t\t/**\n+\t\t * Check if the next not void item is\n+\t\t * TCP or UDP or SCTP or END.\n+\t\t */\n+\t\titem = next_no_fuzzy_pattern(pattern, item);\n+\t\tif (item->type != RTE_FLOW_ITEM_TYPE_TCP &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_UDP &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_SCTP &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_END &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_RAW) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t}\n+\n+\t/* Get the IPV6 info. */\n+\tif (item->type == RTE_FLOW_ITEM_TYPE_IPV6) {\n+\t\t/**\n+\t\t * Set the flow type even if there's no content\n+\t\t * as we must have a flow type.\n+\t\t */\n+\t\trule->input.flow_type = TXGBE_ATR_FLOW_TYPE_IPV6;\n+\t\tptype = txgbe_ptype_table[TXGBE_PT_IPV6];\n+\n+\t\t/**\n+\t\t * 1. must signature match\n+\t\t * 2. not support last\n+\t\t * 3. mask must not null\n+\t\t */\n+\t\tif (rule->mode != RTE_FDIR_MODE_SIGNATURE ||\n+\t\t    item->last ||\n+\t\t    !item->mask) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_UNSPECIFIED,\n+\t\t\t\titem, \"Not supported last point for range\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\n+\t\trule->b_mask = TRUE;\n+\t\tipv6_mask = item->mask;\n+\t\tif (ipv6_mask->hdr.vtc_flow ||\n+\t\t    ipv6_mask->hdr.payload_len ||\n+\t\t    ipv6_mask->hdr.proto ||\n+\t\t    ipv6_mask->hdr.hop_limits) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\n+\t\t/* check src addr mask */\n+\t\tfor (j = 0; j < 16; j++) {\n+\t\t\tif (ipv6_mask->hdr.src_addr[j] == UINT8_MAX) {\n+\t\t\t\trule->mask.src_ipv6_mask |= 1 << j;\n+\t\t\t} else if (ipv6_mask->hdr.src_addr[j] != 0) {\n+\t\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\t\treturn -rte_errno;\n+\t\t\t}\n+\t\t}\n+\n+\t\t/* check dst addr mask */\n+\t\tfor (j = 0; j < 16; j++) {\n+\t\t\tif (ipv6_mask->hdr.dst_addr[j] == UINT8_MAX) {\n+\t\t\t\trule->mask.dst_ipv6_mask |= 1 << j;\n+\t\t\t} else if (ipv6_mask->hdr.dst_addr[j] != 0) {\n+\t\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\t\treturn -rte_errno;\n+\t\t\t}\n+\t\t}\n+\n+\t\tif (item->spec) {\n+\t\t\trule->b_spec = TRUE;\n+\t\t\tipv6_spec = item->spec;\n+\t\t\trte_memcpy(rule->input.src_ip,\n+\t\t\t\t   ipv6_spec->hdr.src_addr, 16);\n+\t\t\trte_memcpy(rule->input.dst_ip,\n+\t\t\t\t   ipv6_spec->hdr.dst_addr, 16);\n+\t\t}\n+\n+\t\t/**\n+\t\t * Check if the next not void item is\n+\t\t * TCP or UDP or SCTP or END.\n+\t\t */\n+\t\titem = next_no_fuzzy_pattern(pattern, item);\n+\t\tif (item->type != RTE_FLOW_ITEM_TYPE_TCP &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_UDP &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_SCTP &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_END &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_RAW) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t}\n+\n+\t/* Get the TCP info. */\n+\tif (item->type == RTE_FLOW_ITEM_TYPE_TCP) {\n+\t\t/**\n+\t\t * Set the flow type even if there's no content\n+\t\t * as we must have a flow type.\n+\t\t */\n+\t\trule->input.flow_type |= TXGBE_ATR_L4TYPE_TCP;\n+\t\tptype = txgbe_ptype_table[TXGBE_PT_IPV4_TCP];\n+\t\t/*Not supported last point for range*/\n+\t\tif (item->last) {\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_UNSPECIFIED,\n+\t\t\t\titem, \"Not supported last point for range\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\t/**\n+\t\t * Only care about src & dst ports,\n+\t\t * others should be masked.\n+\t\t */\n+\t\tif (!item->mask) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\trule->b_mask = TRUE;\n+\t\ttcp_mask = item->mask;\n+\t\tif (tcp_mask->hdr.sent_seq ||\n+\t\t    tcp_mask->hdr.recv_ack ||\n+\t\t    tcp_mask->hdr.data_off ||\n+\t\t    tcp_mask->hdr.tcp_flags ||\n+\t\t    tcp_mask->hdr.rx_win ||\n+\t\t    tcp_mask->hdr.cksum ||\n+\t\t    tcp_mask->hdr.tcp_urp) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\trule->mask.src_port_mask = tcp_mask->hdr.src_port;\n+\t\trule->mask.dst_port_mask = tcp_mask->hdr.dst_port;\n+\n+\t\tif (item->spec) {\n+\t\t\trule->b_spec = TRUE;\n+\t\t\ttcp_spec = item->spec;\n+\t\t\trule->input.src_port =\n+\t\t\t\ttcp_spec->hdr.src_port;\n+\t\t\trule->input.dst_port =\n+\t\t\t\ttcp_spec->hdr.dst_port;\n+\t\t}\n+\n+\t\titem = next_no_fuzzy_pattern(pattern, item);\n+\t\tif (item->type != RTE_FLOW_ITEM_TYPE_RAW &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_END) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t}\n+\n+\t/* Get the UDP info */\n+\tif (item->type == RTE_FLOW_ITEM_TYPE_UDP) {\n+\t\t/**\n+\t\t * Set the flow type even if there's no content\n+\t\t * as we must have a flow type.\n+\t\t */\n+\t\trule->input.flow_type |= TXGBE_ATR_L4TYPE_UDP;\n+\t\tptype = txgbe_ptype_table[TXGBE_PT_IPV4_UDP];\n+\t\t/*Not supported last point for range*/\n+\t\tif (item->last) {\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_UNSPECIFIED,\n+\t\t\t\titem, \"Not supported last point for range\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\t/**\n+\t\t * Only care about src & dst ports,\n+\t\t * others should be masked.\n+\t\t */\n+\t\tif (!item->mask) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\trule->b_mask = TRUE;\n+\t\tudp_mask = item->mask;\n+\t\tif (udp_mask->hdr.dgram_len ||\n+\t\t    udp_mask->hdr.dgram_cksum) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\trule->mask.src_port_mask = udp_mask->hdr.src_port;\n+\t\trule->mask.dst_port_mask = udp_mask->hdr.dst_port;\n+\n+\t\tif (item->spec) {\n+\t\t\trule->b_spec = TRUE;\n+\t\t\tudp_spec = item->spec;\n+\t\t\trule->input.src_port =\n+\t\t\t\tudp_spec->hdr.src_port;\n+\t\t\trule->input.dst_port =\n+\t\t\t\tudp_spec->hdr.dst_port;\n+\t\t}\n+\n+\t\titem = next_no_fuzzy_pattern(pattern, item);\n+\t\tif (item->type != RTE_FLOW_ITEM_TYPE_RAW &&\n+\t\t    item->type != RTE_FLOW_ITEM_TYPE_END) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t}\n+\n+\t/* Get the SCTP info */\n+\tif (item->type == RTE_FLOW_ITEM_TYPE_SCTP) {\n+\t\t/**\n+\t\t * Set the flow type even if there's no content\n+\t\t * as we must have a flow type.\n+\t\t */\n+\t\trule->input.flow_type |= TXGBE_ATR_L4TYPE_SCTP;\n+\t\tptype = txgbe_ptype_table[TXGBE_PT_IPV4_SCTP];\n+\t\t/*Not supported last point for range*/\n+\t\tif (item->last) {\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_UNSPECIFIED,\n+\t\t\t\titem, \"Not supported last point for range\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\n+\t\t/**\n+\t\t * Only care about src & dst ports,\n+\t\t * others should be masked.\n+\t\t */\n+\t\tif (!item->mask) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\trule->b_mask = TRUE;\n+\t\tsctp_mask = item->mask;\n+\t\tif (sctp_mask->hdr.tag ||\n+\t\t\tsctp_mask->hdr.cksum) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\trule->mask.src_port_mask = sctp_mask->hdr.src_port;\n+\t\trule->mask.dst_port_mask = sctp_mask->hdr.dst_port;\n+\n+\t\tif (item->spec) {\n+\t\t\trule->b_spec = TRUE;\n+\t\t\tsctp_spec = item->spec;\n+\t\t\trule->input.src_port =\n+\t\t\t\tsctp_spec->hdr.src_port;\n+\t\t\trule->input.dst_port =\n+\t\t\t\tsctp_spec->hdr.dst_port;\n+\t\t}\n+\t\t/* others even sctp port is not supported */\n+\t\tsctp_mask = item->mask;\n+\t\tif (sctp_mask &&\n+\t\t\t(sctp_mask->hdr.src_port ||\n+\t\t\t sctp_mask->hdr.dst_port ||\n+\t\t\t sctp_mask->hdr.tag ||\n+\t\t\t sctp_mask->hdr.cksum)) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\n+\t\titem = next_no_fuzzy_pattern(pattern, item);\n+\t\tif (item->type != RTE_FLOW_ITEM_TYPE_RAW &&\n+\t\t\titem->type != RTE_FLOW_ITEM_TYPE_END) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t}\n+\n+\t/* Get the flex byte info */\n+\tif (item->type == RTE_FLOW_ITEM_TYPE_RAW) {\n+\t\t/* Not supported last point for range*/\n+\t\tif (item->last) {\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_UNSPECIFIED,\n+\t\t\t\titem, \"Not supported last point for range\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t\t/* mask should not be null */\n+\t\tif (!item->mask || !item->spec) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\n+\t\traw_mask = item->mask;\n+\n+\t\t/* check mask */\n+\t\tif (raw_mask->relative != 0x1 ||\n+\t\t    raw_mask->search != 0x1 ||\n+\t\t    raw_mask->reserved != 0x0 ||\n+\t\t    (uint32_t)raw_mask->offset != 0xffffffff ||\n+\t\t    raw_mask->limit != 0xffff ||\n+\t\t    raw_mask->length != 0xffff) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\n+\t\traw_spec = item->spec;\n+\n+\t\t/* check spec */\n+\t\tif (raw_spec->relative != 0 ||\n+\t\t    raw_spec->search != 0 ||\n+\t\t    raw_spec->reserved != 0 ||\n+\t\t    raw_spec->offset > TXGBE_MAX_FLX_SOURCE_OFF ||\n+\t\t    raw_spec->offset % 2 ||\n+\t\t    raw_spec->limit != 0 ||\n+\t\t    raw_spec->length != 2 ||\n+\t\t    /* pattern can't be 0xffff */\n+\t\t    (raw_spec->pattern[0] == 0xff &&\n+\t\t     raw_spec->pattern[1] == 0xff)) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\n+\t\t/* check pattern mask */\n+\t\tif (raw_mask->pattern[0] != 0xff ||\n+\t\t    raw_mask->pattern[1] != 0xff) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\n+\t\trule->mask.flex_bytes_mask = 0xffff;\n+\t\trule->input.flex_bytes =\n+\t\t\t(((uint16_t)raw_spec->pattern[1]) << 8) |\n+\t\t\traw_spec->pattern[0];\n+\t\trule->flex_bytes_offset = raw_spec->offset;\n+\t}\n+\n+\tif (item->type != RTE_FLOW_ITEM_TYPE_END) {\n+\t\t/* check if the next not void item is END */\n+\t\titem = next_no_fuzzy_pattern(pattern, item);\n+\t\tif (item->type != RTE_FLOW_ITEM_TYPE_END) {\n+\t\t\tmemset(rule, 0, sizeof(struct txgbe_fdir_rule));\n+\t\t\trte_flow_error_set(error, EINVAL,\n+\t\t\t\tRTE_FLOW_ERROR_TYPE_ITEM,\n+\t\t\t\titem, \"Not supported by fdir filter\");\n+\t\t\treturn -rte_errno;\n+\t\t}\n+\t}\n+\n+\trule->input.pkt_type = cpu_to_be16(txgbe_encode_ptype(ptype));\n+\n+\treturn txgbe_parse_fdir_act_attr(attr, actions, rule, error);\n+}\n+\n+static int\n+txgbe_parse_fdir_filter(struct rte_eth_dev *dev,\n+\t\t\tconst struct rte_flow_attr *attr,\n+\t\t\tconst struct rte_flow_item pattern[],\n+\t\t\tconst struct rte_flow_action actions[],\n+\t\t\tstruct txgbe_fdir_rule *rule,\n+\t\t\tstruct rte_flow_error *error)\n+{\n+\tint ret;\n+\tstruct txgbe_hw *hw = TXGBE_DEV_HW(dev);\n+\tenum rte_fdir_mode fdir_mode = dev->data->dev_conf.fdir_conf.mode;\n+\n+\tret = txgbe_parse_fdir_filter_normal(dev, attr, pattern,\n+\t\t\t\t\tactions, rule, error);\n+\tif (!ret)\n+\t\tgoto step_next;\n+\n+step_next:\n+\n+\tif (hw->mac.type == txgbe_mac_raptor &&\n+\t\trule->fdirflags == TXGBE_FDIRPICMD_DROP &&\n+\t\t(rule->input.src_port != 0 || rule->input.dst_port != 0))\n+\t\treturn -ENOTSUP;\n+\n+\tif (fdir_mode == RTE_FDIR_MODE_NONE ||\n+\t    fdir_mode != rule->mode)\n+\t\treturn -ENOTSUP;\n+\n+\tif (rule->queue >= dev->data->nb_rx_queues)\n+\t\treturn -ENOTSUP;\n+\n+\treturn ret;\n+}\n+\n",
    "prefixes": [
        "v2",
        "16/37"
    ]
}