From patchwork Tue Mar 17 15:38:22 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Venkat Duvvuru X-Patchwork-Id: 66808 X-Patchwork-Delegate: ajit.khaparde@broadcom.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id CB5D1A0566; Tue, 17 Mar 2020 16:45:01 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 55FF61C1BF; Tue, 17 Mar 2020 16:39:57 +0100 (CET) Received: from mail-wm1-f65.google.com (mail-wm1-f65.google.com [209.85.128.65]) by dpdk.org (Postfix) with ESMTP id C76441C193 for ; Tue, 17 Mar 2020 16:39:54 +0100 (CET) Received: by mail-wm1-f65.google.com with SMTP id a141so22537137wme.2 for ; Tue, 17 Mar 2020 08:39:54 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=broadcom.com; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=lnD9kzd6OE9F0ZDJ5N7+DeOUPW4MtS0Q5edQ5YBj/Mo=; b=VpFfb81/scMqjYTh+OPdQL1e3jmfB+jofURfogQuA2wX22MiV+d+L+ps5JszGhS0Qz jTJtcw3e+jxi9sjMbuy3LfN60080MoQlJBan8+G/jaPAEKVlnU0uySAX43lWDgpfhs5t /ZblNL5RVXNs6F9Q85ZH46TR3UBoBRHlUTkkU= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=lnD9kzd6OE9F0ZDJ5N7+DeOUPW4MtS0Q5edQ5YBj/Mo=; b=fA25mlXy5DAY0DavNtwAaaWtyQhZiZ5JQoDQoc7BEIJoBo3V7T2aGLgGgo/kmRHVFz 804bnIYWPYvXvuRTAVUAfOgsc38GcbLPoDIV+VHrqU8zD04pKfl6X+CyO0y4/VnUbTaq Vv7TCc4Zoeq+E0VQr8/7B6tsw4K3RWb3oRK+2aJ4XuHGVnEntoiU5Z9s/QXZdJ9cnG8d 8JiAZP/F0K0dpSaYlZiNJ8gqMknz9T6y0ETLsFNXSeURzK8d+eTUxAeJSJPnepxlXmqO yjWMVa7z0734xeewmbYdZ5eX103vrYkcikFELcoFiPgbjOmWgX4Mw2sNXTJkix8NaTFI +feg== X-Gm-Message-State: ANhLgQ1WB6B7UwrH+n77rIoKbPOLQS5VDBkmcLZvvXuaWuhqV+iroxXP dGzVUIGAUHPI263Sb78zoGf92FxXwoSVXE+MJ8LL0poJBbmJu3oKOJAzpLnuESIotUVoCGfNu4Y h7LPZz+deW3q1XMIfmknjyd9EAR3R9TbcpPFRbt3cTb4htnJ1KL7qa9kCVjhNIvH2v5vG X-Google-Smtp-Source: ADFU+vs9txCFU2d0eTO7pPNJ/H2w4zTMt7XPaqwu9s8fuAeqEfe95G9OGLbnpnJgQXwUh9ZxYcukAQ== X-Received: by 2002:a1c:9d41:: with SMTP id g62mr6141791wme.131.1584459593321; Tue, 17 Mar 2020 08:39:53 -0700 (PDT) Received: from S60.dhcp.broadcom.net ([192.19.234.250]) by smtp.gmail.com with ESMTPSA id q4sm5052142wro.56.2020.03.17.08.39.51 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Tue, 17 Mar 2020 08:39:52 -0700 (PDT) From: Venkat Duvvuru To: dev@dpdk.org Cc: Kishore Padmanabha , Venkat Duvvuru Date: Tue, 17 Mar 2020 21:08:22 +0530 Message-Id: <1584459511-5353-25-git-send-email-venkatkumar.duvvuru@broadcom.com> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1584459511-5353-1-git-send-email-venkatkumar.duvvuru@broadcom.com> References: <1584459511-5353-1-git-send-email-venkatkumar.duvvuru@broadcom.com> Subject: [dpdk-dev] [PATCH 24/33] net/bnxt: add support for rte flow item parsing X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" From: Kishore Padmanabha 1. Registers a callback handler for each rte_flow_item type, if it is supported 2. Iterates through each rte_flow_item till RTE_FLOW_ITEM_TYPE_END 3. Invokes the header call back handler 4. Each header call back handler will populate the respective fields in hdr_field & hdr_bitmap Signed-off-by: Kishore Padmanabha Signed-off-by: Venkat Duvvuru Reviewed-by: Lance Richardson Reviewed-by: Ajit Kumar Khaparde --- drivers/net/bnxt/Makefile | 1 + drivers/net/bnxt/tf_ulp/ulp_rte_parser.c | 767 ++++++++++++++++++++++++++ drivers/net/bnxt/tf_ulp/ulp_rte_parser.h | 120 ++++ drivers/net/bnxt/tf_ulp/ulp_template_db.c | 197 +++++++ drivers/net/bnxt/tf_ulp/ulp_template_db.h | 7 + drivers/net/bnxt/tf_ulp/ulp_template_struct.h | 26 + 6 files changed, 1118 insertions(+) create mode 100644 drivers/net/bnxt/tf_ulp/ulp_rte_parser.c create mode 100644 drivers/net/bnxt/tf_ulp/ulp_rte_parser.h diff --git a/drivers/net/bnxt/Makefile b/drivers/net/bnxt/Makefile index 9776987..29d45e7 100644 --- a/drivers/net/bnxt/Makefile +++ b/drivers/net/bnxt/Makefile @@ -64,6 +64,7 @@ SRCS-$(CONFIG_RTE_LIBRTE_BNXT_TRUFLOW) += tf_ulp/ulp_template_db.c SRCS-$(CONFIG_RTE_LIBRTE_BNXT_TRUFLOW) += tf_ulp/ulp_utils.c SRCS-$(CONFIG_RTE_LIBRTE_BNXT_TRUFLOW) += tf_ulp/ulp_mapper.c SRCS-$(CONFIG_RTE_LIBRTE_BNXT_TRUFLOW) += tf_ulp/ulp_matcher.c +SRCS-$(CONFIG_RTE_LIBRTE_BNXT_TRUFLOW) += tf_ulp/ulp_rte_parser.c # # Export include files # diff --git a/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c new file mode 100644 index 0000000..3ffdcbd --- /dev/null +++ b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c @@ -0,0 +1,767 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2014-2020 Broadcom + * All rights reserved. + */ + +#include "bnxt.h" +#include "ulp_template_db.h" +#include "ulp_template_struct.h" +#include "bnxt_tf_common.h" +#include "ulp_rte_parser.h" +#include "ulp_utils.h" +#include "tfp.h" + +/* Inline Func to read integer that is stored in big endian format */ +static inline void ulp_util_field_int_read(uint8_t *buffer, + uint32_t *val) +{ + uint32_t temp_val; + + memcpy(&temp_val, buffer, sizeof(uint32_t)); + *val = rte_be_to_cpu_32(temp_val); +} + +/* Inline Func to write integer that is stored in big endian format */ +static inline void ulp_util_field_int_write(uint8_t *buffer, + uint32_t val) +{ + uint32_t temp_val = rte_cpu_to_be_32(val); + + memcpy(buffer, &temp_val, sizeof(uint32_t)); +} + +/* + * Function to handle the parsing of RTE Flows and placing + * the RTE flow items into the ulp structures. + */ +int32_t +bnxt_ulp_rte_parser_hdr_parse(const struct rte_flow_item pattern[], + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field) +{ + const struct rte_flow_item *item = pattern; + uint32_t field_idx = BNXT_ULP_HDR_FIELD_LAST; + uint32_t vlan_idx = 0; + struct bnxt_ulp_rte_hdr_info *hdr_info; + + /* Parse all the items in the pattern */ + while (item && item->type != RTE_FLOW_ITEM_TYPE_END) { + /* get the header information from the flow_hdr_info table */ + hdr_info = &ulp_hdr_info[item->type]; + if (hdr_info->hdr_type == + BNXT_ULP_HDR_TYPE_NOT_SUPPORTED) { + BNXT_TF_DBG(ERR, + "Truflow parser does not support type %d\n", + item->type); + return BNXT_TF_RC_PARSE_ERR; + } else if (hdr_info->hdr_type == + BNXT_ULP_HDR_TYPE_SUPPORTED) { + /* call the registered callback handler */ + if (hdr_info->proto_hdr_func) { + if (hdr_info->proto_hdr_func(item, + hdr_bitmap, + hdr_field, + &field_idx, + &vlan_idx) != + BNXT_TF_RC_SUCCESS) { + return BNXT_TF_RC_ERROR; + } + } + } + item++; + } + return BNXT_TF_RC_SUCCESS; +} + +/* Function to handle the parsing of RTE Flow item PF Header. */ +static int32_t +ulp_rte_parser_svif_set(struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + enum rte_flow_item_type proto, + uint32_t svif, + uint32_t mask) +{ + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_SVIF)) { + BNXT_TF_DBG(ERR, + "SVIF already set," + " multiple sources not supported\n"); + return BNXT_TF_RC_ERROR; + } + + /* TBD: Check for any mapping errors for svif */ + /* Update the hdr_bitmap with BNXT_ULP_HDR_PROTO_SVIF. */ + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_SVIF); + + if (proto != RTE_FLOW_ITEM_TYPE_PF) { + memcpy(hdr_field[BNXT_ULP_HDR_FIELD_SVIF_INDEX].spec, + &svif, sizeof(svif)); + memcpy(hdr_field[BNXT_ULP_HDR_FIELD_SVIF_INDEX].mask, + &mask, sizeof(mask)); + hdr_field[BNXT_ULP_HDR_FIELD_SVIF_INDEX].size = sizeof(svif); + } + + return BNXT_TF_RC_SUCCESS; +} + +/* Function to handle the parsing of RTE Flow item PF Header. */ +int32_t +ulp_rte_pf_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx __rte_unused, + uint32_t *vlan_idx __rte_unused) +{ + return ulp_rte_parser_svif_set(hdr_bitmap, hdr_field, + item->type, 0, 0); +} + +/* Function to handle the parsing of RTE Flow item VF Header. */ +int32_t +ulp_rte_vf_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx __rte_unused, + uint32_t *vlan_idx __rte_unused) +{ + const struct rte_flow_item_vf *vf_spec, *vf_mask; + uint32_t svif = 0, mask = 0; + + vf_spec = item->spec; + vf_mask = item->mask; + + /* + * Copy the rte_flow_item for eth into hdr_field using ethernet + * header fields. + */ + if (vf_spec) + svif = vf_spec->id; + if (vf_mask) + mask = vf_mask->id; + + return ulp_rte_parser_svif_set(hdr_bitmap, hdr_field, + item->type, svif, mask); +} + +/* Function to handle the parsing of RTE Flow item port id Header. */ +int32_t +ulp_rte_port_id_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx __rte_unused, + uint32_t *vlan_idx __rte_unused) +{ + const struct rte_flow_item_port_id *port_spec, *port_mask; + uint32_t svif = 0, mask = 0; + + port_spec = item->spec; + port_mask = item->mask; + + /* + * Copy the rte_flow_item for Port into hdr_field using port id + * header fields. + */ + if (port_spec) + svif = port_spec->id; + if (port_mask) + mask = port_mask->id; + + return ulp_rte_parser_svif_set(hdr_bitmap, hdr_field, + item->type, svif, mask); +} + +/* Function to handle the parsing of RTE Flow item phy port Header. */ +int32_t +ulp_rte_phy_port_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx __rte_unused, + uint32_t *vlan_idx __rte_unused) +{ + const struct rte_flow_item_phy_port *port_spec, *port_mask; + uint32_t svif = 0, mask = 0; + + port_spec = item->spec; + port_mask = item->mask; + + /* Copy the rte_flow_item for phy port into hdr_field */ + if (port_spec) + svif = port_spec->index; + if (port_mask) + mask = port_mask->index; + + return ulp_rte_parser_svif_set(hdr_bitmap, hdr_field, + item->type, svif, mask); +} + +/* Function to handle the parsing of RTE Flow item Ethernet Header. */ +int32_t +ulp_rte_eth_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx) +{ + const struct rte_flow_item_eth *eth_spec, *eth_mask; + uint32_t idx = *field_idx; + uint32_t mdx = *field_idx; + uint64_t set_flag = 0; + + eth_spec = item->spec; + eth_mask = item->mask; + + /* + * Copy the rte_flow_item for eth into hdr_field using ethernet + * header fields + */ + if (eth_spec) { + hdr_field[idx].size = sizeof(eth_spec->dst.addr_bytes); + memcpy(hdr_field[idx++].spec, eth_spec->dst.addr_bytes, + sizeof(eth_spec->dst.addr_bytes)); + hdr_field[idx].size = sizeof(eth_spec->src.addr_bytes); + memcpy(hdr_field[idx++].spec, eth_spec->src.addr_bytes, + sizeof(eth_spec->src.addr_bytes)); + hdr_field[idx].size = sizeof(eth_spec->type); + memcpy(hdr_field[idx++].spec, ð_spec->type, + sizeof(eth_spec->type)); + } else { + idx += BNXT_ULP_PROTO_HDR_ETH_NUM; + } + + if (eth_mask) { + memcpy(hdr_field[mdx++].mask, eth_mask->dst.addr_bytes, + sizeof(eth_mask->dst.addr_bytes)); + memcpy(hdr_field[mdx++].mask, eth_mask->src.addr_bytes, + sizeof(eth_mask->src.addr_bytes)); + memcpy(hdr_field[mdx++].mask, ð_mask->type, + sizeof(eth_mask->type)); + } + /* Add number of vlan header elements */ + *field_idx = idx + BNXT_ULP_PROTO_HDR_VLAN_NUM; + *vlan_idx = idx; + + /* Update the hdr_bitmap with BNXT_ULP_HDR_PROTO_I_ETH */ + set_flag = ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_ETH); + if (set_flag) + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_ETH); + else + ULP_BITMAP_RESET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_ETH); + + /* update the hdr_bitmap with BNXT_ULP_HDR_PROTO_O_ETH */ + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_ETH); + + return BNXT_TF_RC_SUCCESS; +} + +/* Function to handle the parsing of RTE Flow item Vlan Header. */ +int32_t +ulp_rte_vlan_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx __rte_unused, + uint32_t *vlan_idx) +{ + const struct rte_flow_item_vlan *vlan_spec, *vlan_mask; + uint32_t idx = *vlan_idx; + uint32_t mdx = *vlan_idx; + uint16_t vlan_tag, priority; + uint32_t outer_vtag_num = 0, inner_vtag_num = 0; + uint8_t *outer_tag_buffer; + uint8_t *inner_tag_buffer; + + vlan_spec = item->spec; + vlan_mask = item->mask; + outer_tag_buffer = hdr_field[BNXT_ULP_HDR_FIELD_O_VTAG_NUM].spec; + inner_tag_buffer = hdr_field[BNXT_ULP_HDR_FIELD_I_VTAG_NUM].spec; + + /* + * Copy the rte_flow_item for vlan into hdr_field using Vlan + * header fields + */ + if (vlan_spec) { + vlan_tag = ntohs(vlan_spec->tci); + priority = htons(vlan_tag >> 13); + vlan_tag &= 0xfff; + vlan_tag = htons(vlan_tag); + + hdr_field[idx].size = sizeof(priority); + memcpy(hdr_field[idx++].spec, &priority, sizeof(priority)); + hdr_field[idx].size = sizeof(vlan_tag); + memcpy(hdr_field[idx++].spec, &vlan_tag, sizeof(vlan_tag)); + hdr_field[idx].size = sizeof(vlan_spec->inner_type); + memcpy(hdr_field[idx++].spec, &vlan_spec->inner_type, + sizeof(vlan_spec->inner_type)); + } else { + idx += BNXT_ULP_PROTO_HDR_S_VLAN_NUM; + } + + if (vlan_mask) { + vlan_tag = ntohs(vlan_mask->tci); + priority = htons(vlan_tag >> 13); + vlan_tag &= 0xfff; + vlan_tag = htons(vlan_tag); + + memcpy(hdr_field[mdx++].mask, &priority, sizeof(priority)); + memcpy(hdr_field[mdx++].mask, &vlan_tag, sizeof(vlan_tag)); + memcpy(hdr_field[mdx++].mask, &vlan_mask->inner_type, + sizeof(vlan_mask->inner_type)); + } + /* Set the vlan index to new incremented value */ + *vlan_idx = idx; + + /* Get the outer tag and inner tag counts */ + ulp_util_field_int_read(outer_tag_buffer, &outer_vtag_num); + ulp_util_field_int_read(inner_tag_buffer, &inner_vtag_num); + + /* Update the hdr_bitmap of the vlans */ + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_ETH) && + !ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_OO_VLAN)) { + /* Set the outer vlan bit and update the vlan tag num */ + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_OO_VLAN); + outer_vtag_num++; + ulp_util_field_int_write(outer_tag_buffer, outer_vtag_num); + hdr_field[BNXT_ULP_HDR_FIELD_O_VTAG_NUM].size = + sizeof(uint32_t); + } else if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_ETH) && + ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_OO_VLAN) && + !ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_OI_VLAN)) { + /* Set the outer vlan bit and update the vlan tag num */ + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_OI_VLAN); + outer_vtag_num++; + ulp_util_field_int_write(outer_tag_buffer, outer_vtag_num); + hdr_field[BNXT_ULP_HDR_FIELD_O_VTAG_NUM].size = + sizeof(uint32_t); + } else if (ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_O_ETH) && + ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_OO_VLAN) && + ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_OI_VLAN) && + ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_I_ETH) && + !ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_IO_VLAN)) { + /* Set the inner vlan bit and update the vlan tag num */ + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_IO_VLAN); + inner_vtag_num++; + ulp_util_field_int_write(inner_tag_buffer, inner_vtag_num); + hdr_field[BNXT_ULP_HDR_FIELD_I_VTAG_NUM].size = + sizeof(uint32_t); + } else if (ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_O_ETH) && + ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_OO_VLAN) && + ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_OI_VLAN) && + ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_I_ETH) && + ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_IO_VLAN) && + !ULP_BITMAP_ISSET(hdr_bitmap->bits, + BNXT_ULP_HDR_BIT_II_VLAN)) { + /* Set the inner vlan bit and update the vlan tag num */ + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_II_VLAN); + inner_vtag_num++; + ulp_util_field_int_write(inner_tag_buffer, inner_vtag_num); + hdr_field[BNXT_ULP_HDR_FIELD_I_VTAG_NUM].size = + sizeof(uint32_t); + } else { + BNXT_TF_DBG(ERR, "Error Parsing:Vlan hdr found withtout eth\n"); + return BNXT_TF_RC_ERROR; + } + return BNXT_TF_RC_SUCCESS; +} + +/* Function to handle the parsing of RTE Flow item IPV4 Header. */ +int32_t +ulp_rte_ipv4_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx __rte_unused) +{ + const struct rte_flow_item_ipv4 *ipv4_spec, *ipv4_mask; + uint32_t idx = *field_idx; + uint32_t mdx = *field_idx; + + ipv4_spec = item->spec; + ipv4_mask = item->mask; + + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_L3)) { + BNXT_TF_DBG(ERR, "Parse Error:Third L3 header not supported\n"); + return BNXT_TF_RC_ERROR; + } + + /* + * Copy the rte_flow_item for ipv4 into hdr_field using ipv4 + * header fields + */ + if (ipv4_spec) { + hdr_field[idx].size = sizeof(ipv4_spec->hdr.version_ihl); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.version_ihl, + sizeof(ipv4_spec->hdr.version_ihl)); + hdr_field[idx].size = sizeof(ipv4_spec->hdr.type_of_service); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.type_of_service, + sizeof(ipv4_spec->hdr.type_of_service)); + hdr_field[idx].size = sizeof(ipv4_spec->hdr.total_length); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.total_length, + sizeof(ipv4_spec->hdr.total_length)); + hdr_field[idx].size = sizeof(ipv4_spec->hdr.packet_id); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.packet_id, + sizeof(ipv4_spec->hdr.packet_id)); + hdr_field[idx].size = sizeof(ipv4_spec->hdr.fragment_offset); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.fragment_offset, + sizeof(ipv4_spec->hdr.fragment_offset)); + hdr_field[idx].size = sizeof(ipv4_spec->hdr.time_to_live); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.time_to_live, + sizeof(ipv4_spec->hdr.time_to_live)); + hdr_field[idx].size = sizeof(ipv4_spec->hdr.next_proto_id); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.next_proto_id, + sizeof(ipv4_spec->hdr.next_proto_id)); + hdr_field[idx].size = sizeof(ipv4_spec->hdr.hdr_checksum); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.hdr_checksum, + sizeof(ipv4_spec->hdr.hdr_checksum)); + hdr_field[idx].size = sizeof(ipv4_spec->hdr.src_addr); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.src_addr, + sizeof(ipv4_spec->hdr.src_addr)); + hdr_field[idx].size = sizeof(ipv4_spec->hdr.dst_addr); + memcpy(hdr_field[idx++].spec, &ipv4_spec->hdr.dst_addr, + sizeof(ipv4_spec->hdr.dst_addr)); + } else { + idx += BNXT_ULP_PROTO_HDR_IPV4_NUM; + } + + if (ipv4_mask) { + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.version_ihl, + sizeof(ipv4_mask->hdr.version_ihl)); + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.type_of_service, + sizeof(ipv4_mask->hdr.type_of_service)); + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.total_length, + sizeof(ipv4_mask->hdr.total_length)); + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.packet_id, + sizeof(ipv4_mask->hdr.packet_id)); + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.fragment_offset, + sizeof(ipv4_mask->hdr.fragment_offset)); + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.time_to_live, + sizeof(ipv4_mask->hdr.time_to_live)); + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.next_proto_id, + sizeof(ipv4_mask->hdr.next_proto_id)); + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.hdr_checksum, + sizeof(ipv4_mask->hdr.hdr_checksum)); + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.src_addr, + sizeof(ipv4_mask->hdr.src_addr)); + memcpy(hdr_field[mdx++].mask, &ipv4_mask->hdr.dst_addr, + sizeof(ipv4_mask->hdr.dst_addr)); + } + *field_idx = idx; /* Number of ipv4 header elements */ + + /* Set the ipv4 header bitmap and computed l3 header bitmaps */ + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_L3) || + ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV4) || + ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV6)) { + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_IPV4); + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_L3); + } else { + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV4); + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_L3); + } + return BNXT_TF_RC_SUCCESS; +} + +/* Function to handle the parsing of RTE Flow item IPV6 Header */ +int32_t +ulp_rte_ipv6_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx __rte_unused) +{ + const struct rte_flow_item_ipv6 *ipv6_spec, *ipv6_mask; + uint32_t idx = *field_idx; + uint32_t mdx = *field_idx; + + ipv6_spec = item->spec; + ipv6_mask = item->mask; + + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_L3)) { + BNXT_TF_DBG(ERR, "Parse Error: 3'rd L3 header not supported\n"); + return BNXT_TF_RC_ERROR; + } + + /* + * Copy the rte_flow_item for ipv4 into hdr_field using ipv4 + * header fields + */ + if (ipv6_spec) { + hdr_field[idx].size = sizeof(ipv6_spec->hdr.vtc_flow); + memcpy(hdr_field[idx++].spec, &ipv6_spec->hdr.vtc_flow, + sizeof(ipv6_spec->hdr.vtc_flow)); + hdr_field[idx].size = sizeof(ipv6_spec->hdr.payload_len); + memcpy(hdr_field[idx++].spec, &ipv6_spec->hdr.payload_len, + sizeof(ipv6_spec->hdr.payload_len)); + hdr_field[idx].size = sizeof(ipv6_spec->hdr.proto); + memcpy(hdr_field[idx++].spec, &ipv6_spec->hdr.proto, + sizeof(ipv6_spec->hdr.proto)); + hdr_field[idx].size = sizeof(ipv6_spec->hdr.hop_limits); + memcpy(hdr_field[idx++].spec, &ipv6_spec->hdr.hop_limits, + sizeof(ipv6_spec->hdr.hop_limits)); + hdr_field[idx].size = sizeof(ipv6_spec->hdr.src_addr); + memcpy(hdr_field[idx++].spec, &ipv6_spec->hdr.src_addr, + sizeof(ipv6_spec->hdr.src_addr)); + hdr_field[idx].size = sizeof(ipv6_spec->hdr.dst_addr); + memcpy(hdr_field[idx++].spec, &ipv6_spec->hdr.dst_addr, + sizeof(ipv6_spec->hdr.dst_addr)); + } else { + idx += BNXT_ULP_PROTO_HDR_IPV6_NUM; + } + + if (ipv6_mask) { + memcpy(hdr_field[mdx++].mask, &ipv6_mask->hdr.vtc_flow, + sizeof(ipv6_mask->hdr.vtc_flow)); + memcpy(hdr_field[mdx++].mask, &ipv6_mask->hdr.payload_len, + sizeof(ipv6_mask->hdr.payload_len)); + memcpy(hdr_field[mdx++].mask, &ipv6_mask->hdr.proto, + sizeof(ipv6_mask->hdr.proto)); + memcpy(hdr_field[mdx++].mask, &ipv6_mask->hdr.hop_limits, + sizeof(ipv6_mask->hdr.hop_limits)); + memcpy(hdr_field[mdx++].mask, &ipv6_mask->hdr.src_addr, + sizeof(ipv6_mask->hdr.src_addr)); + memcpy(hdr_field[mdx++].mask, &ipv6_mask->hdr.dst_addr, + sizeof(ipv6_mask->hdr.dst_addr)); + } + *field_idx = idx; /* add number of ipv6 header elements */ + + /* Set the ipv6 header bitmap and computed l3 header bitmaps */ + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_L3) || + ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV4) || + ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV6)) { + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_IPV6); + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_L3); + } else { + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV6); + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_L3); + } + return BNXT_TF_RC_SUCCESS; +} + +/* Function to handle the parsing of RTE Flow item UDP Header. */ +int32_t +ulp_rte_udp_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx __rte_unused) +{ + const struct rte_flow_item_udp *udp_spec, *udp_mask; + uint32_t idx = *field_idx; + uint32_t mdx = *field_idx; + + udp_spec = item->spec; + udp_mask = item->mask; + + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_L4)) { + BNXT_TF_DBG(ERR, "Parse Err:Third L4 header not supported\n"); + return BNXT_TF_RC_ERROR; + } + + /* + * Copy the rte_flow_item for ipv4 into hdr_field using ipv4 + * header fields + */ + if (udp_spec) { + hdr_field[idx].size = sizeof(udp_spec->hdr.src_port); + memcpy(hdr_field[idx++].spec, &udp_spec->hdr.src_port, + sizeof(udp_spec->hdr.src_port)); + hdr_field[idx].size = sizeof(udp_spec->hdr.dst_port); + memcpy(hdr_field[idx++].spec, &udp_spec->hdr.dst_port, + sizeof(udp_spec->hdr.dst_port)); + hdr_field[idx].size = sizeof(udp_spec->hdr.dgram_len); + memcpy(hdr_field[idx++].spec, &udp_spec->hdr.dgram_len, + sizeof(udp_spec->hdr.dgram_len)); + hdr_field[idx].size = sizeof(udp_spec->hdr.dgram_cksum); + memcpy(hdr_field[idx++].spec, &udp_spec->hdr.dgram_cksum, + sizeof(udp_spec->hdr.dgram_cksum)); + } else { + idx += BNXT_ULP_PROTO_HDR_UDP_NUM; + } + + if (udp_mask) { + memcpy(hdr_field[mdx++].mask, &udp_mask->hdr.src_port, + sizeof(udp_mask->hdr.src_port)); + memcpy(hdr_field[mdx++].mask, &udp_mask->hdr.dst_port, + sizeof(udp_mask->hdr.dst_port)); + memcpy(hdr_field[mdx++].mask, &udp_mask->hdr.dgram_len, + sizeof(udp_mask->hdr.dgram_len)); + memcpy(hdr_field[mdx++].mask, &udp_mask->hdr.dgram_cksum, + sizeof(udp_mask->hdr.dgram_cksum)); + } + *field_idx = idx; /* Add number of UDP header elements */ + + /* Set the udp header bitmap and computed l4 header bitmaps */ + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_L4) || + ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_UDP) || + ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_TCP)) { + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_UDP); + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_L4); + } else { + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_UDP); + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_L4); + } + return BNXT_TF_RC_SUCCESS; +} + +/* Function to handle the parsing of RTE Flow item TCP Header. */ +int32_t +ulp_rte_tcp_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx __rte_unused) +{ + const struct rte_flow_item_tcp *tcp_spec, *tcp_mask; + uint32_t idx = *field_idx; + uint32_t mdx = *field_idx; + + tcp_spec = item->spec; + tcp_mask = item->mask; + + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_L4)) { + BNXT_TF_DBG(ERR, "Parse Error:Third L4 header not supported\n"); + return BNXT_TF_RC_ERROR; + } + + /* + * Copy the rte_flow_item for ipv4 into hdr_field using ipv4 + * header fields + */ + if (tcp_spec) { + hdr_field[idx].size = sizeof(tcp_spec->hdr.src_port); + memcpy(hdr_field[idx++].spec, &tcp_spec->hdr.src_port, + sizeof(tcp_spec->hdr.src_port)); + hdr_field[idx].size = sizeof(tcp_spec->hdr.dst_port); + memcpy(hdr_field[idx++].spec, &tcp_spec->hdr.dst_port, + sizeof(tcp_spec->hdr.dst_port)); + hdr_field[idx].size = sizeof(tcp_spec->hdr.sent_seq); + memcpy(hdr_field[idx++].spec, &tcp_spec->hdr.sent_seq, + sizeof(tcp_spec->hdr.sent_seq)); + hdr_field[idx].size = sizeof(tcp_spec->hdr.recv_ack); + memcpy(hdr_field[idx++].spec, &tcp_spec->hdr.recv_ack, + sizeof(tcp_spec->hdr.recv_ack)); + hdr_field[idx].size = sizeof(tcp_spec->hdr.data_off); + memcpy(hdr_field[idx++].spec, &tcp_spec->hdr.data_off, + sizeof(tcp_spec->hdr.data_off)); + hdr_field[idx].size = sizeof(tcp_spec->hdr.tcp_flags); + memcpy(hdr_field[idx++].spec, &tcp_spec->hdr.tcp_flags, + sizeof(tcp_spec->hdr.tcp_flags)); + hdr_field[idx].size = sizeof(tcp_spec->hdr.rx_win); + memcpy(hdr_field[idx++].spec, &tcp_spec->hdr.rx_win, + sizeof(tcp_spec->hdr.rx_win)); + hdr_field[idx].size = sizeof(tcp_spec->hdr.cksum); + memcpy(hdr_field[idx++].spec, &tcp_spec->hdr.cksum, + sizeof(tcp_spec->hdr.cksum)); + hdr_field[idx].size = sizeof(tcp_spec->hdr.tcp_urp); + memcpy(hdr_field[idx++].spec, &tcp_spec->hdr.tcp_urp, + sizeof(tcp_spec->hdr.tcp_urp)); + } else { + idx += BNXT_ULP_PROTO_HDR_TCP_NUM; + } + + if (tcp_mask) { + memcpy(hdr_field[mdx++].mask, &tcp_mask->hdr.src_port, + sizeof(tcp_mask->hdr.src_port)); + memcpy(hdr_field[mdx++].mask, &tcp_mask->hdr.dst_port, + sizeof(tcp_mask->hdr.dst_port)); + memcpy(hdr_field[mdx++].mask, &tcp_mask->hdr.sent_seq, + sizeof(tcp_mask->hdr.sent_seq)); + memcpy(hdr_field[mdx++].mask, &tcp_mask->hdr.recv_ack, + sizeof(tcp_mask->hdr.recv_ack)); + memcpy(hdr_field[mdx++].mask, &tcp_mask->hdr.data_off, + sizeof(tcp_mask->hdr.data_off)); + memcpy(hdr_field[mdx++].mask, &tcp_mask->hdr.tcp_flags, + sizeof(tcp_mask->hdr.tcp_flags)); + memcpy(hdr_field[mdx++].mask, &tcp_mask->hdr.rx_win, + sizeof(tcp_mask->hdr.rx_win)); + memcpy(hdr_field[mdx++].mask, &tcp_mask->hdr.cksum, + sizeof(tcp_mask->hdr.cksum)); + memcpy(hdr_field[mdx++].mask, &tcp_mask->hdr.tcp_urp, + sizeof(tcp_mask->hdr.tcp_urp)); + } + *field_idx = idx; /* add number of TCP header elements */ + + /* Set the udp header bitmap and computed l4 header bitmaps */ + if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_L4) || + ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_UDP) || + ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_TCP)) { + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_TCP); + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_L4); + } else { + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_TCP); + ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_L4); + } + return BNXT_TF_RC_SUCCESS; +} + +/* Function to handle the parsing of RTE Flow item Vxlan Header. */ +int32_t +ulp_rte_vxlan_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdrbitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx __rte_unused) +{ + const struct rte_flow_item_vxlan *vxlan_spec, *vxlan_mask; + uint32_t idx = *field_idx; + uint32_t mdx = *field_idx; + + vxlan_spec = item->spec; + vxlan_mask = item->mask; + + /* + * Copy the rte_flow_item for vxlan into hdr_field using vxlan + * header fields + */ + if (vxlan_spec) { + hdr_field[idx].size = sizeof(vxlan_spec->flags); + memcpy(hdr_field[idx++].spec, &vxlan_spec->flags, + sizeof(vxlan_spec->flags)); + hdr_field[idx].size = sizeof(vxlan_spec->rsvd0); + memcpy(hdr_field[idx++].spec, &vxlan_spec->rsvd0, + sizeof(vxlan_spec->rsvd0)); + hdr_field[idx].size = sizeof(vxlan_spec->vni); + memcpy(hdr_field[idx++].spec, &vxlan_spec->vni, + sizeof(vxlan_spec->vni)); + hdr_field[idx].size = sizeof(vxlan_spec->rsvd1); + memcpy(hdr_field[idx++].spec, &vxlan_spec->rsvd1, + sizeof(vxlan_spec->rsvd1)); + } else { + idx += BNXT_ULP_PROTO_HDR_VXLAN_NUM; + } + + if (vxlan_mask) { + memcpy(hdr_field[mdx++].mask, &vxlan_mask->flags, + sizeof(vxlan_mask->flags)); + memcpy(hdr_field[mdx++].mask, &vxlan_mask->rsvd0, + sizeof(vxlan_mask->rsvd0)); + memcpy(hdr_field[mdx++].mask, &vxlan_mask->vni, + sizeof(vxlan_mask->vni)); + memcpy(hdr_field[mdx++].mask, &vxlan_mask->rsvd1, + sizeof(vxlan_mask->rsvd1)); + } + *field_idx = idx; /* Add number of vxlan header elements */ + + /* Update the hdr_bitmap with vxlan */ + ULP_BITMAP_SET(hdrbitmap->bits, BNXT_ULP_HDR_BIT_T_VXLAN); + return BNXT_TF_RC_SUCCESS; +} + +/* Function to handle the parsing of RTE Flow item void Header */ +int32_t +ulp_rte_void_hdr_handler(const struct rte_flow_item *item __rte_unused, + struct ulp_rte_hdr_bitmap *hdr_bit __rte_unused, + struct ulp_rte_hdr_field *hdr_field __rte_unused, + uint32_t *field_idx __rte_unused, + uint32_t *vlan_idx __rte_unused) +{ + return BNXT_TF_RC_SUCCESS; +} diff --git a/drivers/net/bnxt/tf_ulp/ulp_rte_parser.h b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.h new file mode 100644 index 0000000..3a7845d --- /dev/null +++ b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.h @@ -0,0 +1,120 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2014-2020 Broadcom + * All rights reserved. + */ + +#ifndef _ULP_RTE_PARSER_H_ +#define _ULP_RTE_PARSER_H_ + +#include +#include +#include +#include "ulp_template_db.h" +#include "ulp_template_struct.h" + +/* + * Function to handle the parsing of RTE Flows and placing + * the RTE flow items into the ulp structures. + */ +int32_t +bnxt_ulp_rte_parser_hdr_parse(const struct rte_flow_item pattern[], + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field); + +/* Function to handle the parsing of RTE Flow item PF Header. */ +int32_t +ulp_rte_pf_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item VF Header. */ +int32_t +ulp_rte_vf_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item port id Header. */ +int32_t +ulp_rte_port_id_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item port id Header. */ +int32_t +ulp_rte_phy_port_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the RTE item Ethernet Header. */ +int32_t +ulp_rte_eth_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item Vlan Header. */ +int32_t +ulp_rte_vlan_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item IPV4 Header. */ +int32_t +ulp_rte_ipv4_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item IPV6 Header. */ +int32_t +ulp_rte_ipv6_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item UDP Header. */ +int32_t +ulp_rte_udp_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item TCP Header. */ +int32_t +ulp_rte_tcp_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item Vxlan Header. */ +int32_t +ulp_rte_vxlan_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdrbitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +/* Function to handle the parsing of RTE Flow item void Header. */ +int32_t +ulp_rte_void_hdr_handler(const struct rte_flow_item *item, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); + +#endif /* _ULP_RTE_PARSER_H_ */ diff --git a/drivers/net/bnxt/tf_ulp/ulp_template_db.c b/drivers/net/bnxt/tf_ulp/ulp_template_db.c index 5981c74..1d15563 100644 --- a/drivers/net/bnxt/tf_ulp/ulp_template_db.c +++ b/drivers/net/bnxt/tf_ulp/ulp_template_db.c @@ -11,6 +11,8 @@ #include "ulp_template_db.h" #include "ulp_template_field_db.h" #include "ulp_template_struct.h" +#include "ulp_rte_parser.h" + uint32_t ulp_act_prop_map_table[] = { [BNXT_ULP_ACT_PROP_IDX_ENCAP_TUN_SZ] = BNXT_ULP_ACT_PROP_SZ_ENCAP_TUN_SZ, @@ -108,6 +110,201 @@ struct bnxt_ulp_device_params ulp_device_params[] = { } }; +struct bnxt_ulp_rte_hdr_info ulp_hdr_info[] = { + [RTE_FLOW_ITEM_TYPE_END] = { + .hdr_type = BNXT_ULP_HDR_TYPE_END, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_VOID] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_void_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_INVERT] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ANY] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_PF] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_pf_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_VF] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_vf_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_PHY_PORT] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_phy_port_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_PORT_ID] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_port_id_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_RAW] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ETH] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_eth_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_VLAN] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_vlan_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_IPV4] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_ipv4_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_IPV6] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_ipv6_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_ICMP] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_UDP] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_udp_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_TCP] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_tcp_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_SCTP] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_VXLAN] = { + .hdr_type = BNXT_ULP_HDR_TYPE_SUPPORTED, + .proto_hdr_func = ulp_rte_vxlan_hdr_handler + }, + [RTE_FLOW_ITEM_TYPE_E_TAG] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_NVGRE] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_MPLS] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_GRE] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_FUZZY] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_GTP] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_GTPC] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_GTPU] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ESP] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_GENEVE] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_VXLAN_GPE] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ARP_ETH_IPV4] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_IPV6_EXT] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ICMP6] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ICMP6_ND_NS] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ICMP6_ND_NA] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ICMP6_ND_OPT] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ICMP6_ND_OPT_SLA_ETH] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_ICMP6_ND_OPT_TLA_ETH] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_MARK] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_META] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_GRE_KEY] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_GTP_PSC] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_PPPOES] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_PPPOED] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_PPPOE_PROTO_ID] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_NSH] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_IGMP] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_AH] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + }, + [RTE_FLOW_ITEM_TYPE_HIGIG2] = { + .hdr_type = BNXT_ULP_HDR_TYPE_NOT_SUPPORTED, + .proto_hdr_func = NULL + } +}; + struct bnxt_ulp_mapper_tbl_list_info ulp_class_tmpl_list[] = { [((0 << BNXT_ULP_LOG2_MAX_NUM_DEV) | BNXT_ULP_DEVICE_ID_WH_PLUS)] = { .device_name = BNXT_ULP_DEVICE_ID_WH_PLUS, diff --git a/drivers/net/bnxt/tf_ulp/ulp_template_db.h b/drivers/net/bnxt/tf_ulp/ulp_template_db.h index f4850bf..906b542 100644 --- a/drivers/net/bnxt/tf_ulp/ulp_template_db.h +++ b/drivers/net/bnxt/tf_ulp/ulp_template_db.h @@ -115,6 +115,13 @@ enum bnxt_ulp_hdr_field { BNXT_ULP_HDR_FIELD_LAST = 4 }; +enum bnxt_ulp_hdr_type { + BNXT_ULP_HDR_TYPE_NOT_SUPPORTED = 0, + BNXT_ULP_HDR_TYPE_SUPPORTED = 1, + BNXT_ULP_HDR_TYPE_END = 2, + BNXT_ULP_HDR_TYPE_LAST = 3 +}; + enum bnxt_ulp_mask_opc { BNXT_ULP_MASK_OPC_SET_TO_CONSTANT = 0, BNXT_ULP_MASK_OPC_SET_TO_HDR_FIELD = 1, diff --git a/drivers/net/bnxt/tf_ulp/ulp_template_struct.h b/drivers/net/bnxt/tf_ulp/ulp_template_struct.h index 0e811ec..0699634 100644 --- a/drivers/net/bnxt/tf_ulp/ulp_template_struct.h +++ b/drivers/net/bnxt/tf_ulp/ulp_template_struct.h @@ -17,6 +17,18 @@ #include "rte_flow.h" #include "tf_core.h" +/* Number of fields for each protocol */ +#define BNXT_ULP_PROTO_HDR_SVIF_NUM 1 +#define BNXT_ULP_PROTO_HDR_ETH_NUM 3 +#define BNXT_ULP_PROTO_HDR_S_VLAN_NUM 3 +#define BNXT_ULP_PROTO_HDR_VLAN_NUM 6 +#define BNXT_ULP_PROTO_HDR_IPV4_NUM 10 +#define BNXT_ULP_PROTO_HDR_IPV6_NUM 6 +#define BNXT_ULP_PROTO_HDR_UDP_NUM 4 +#define BNXT_ULP_PROTO_HDR_TCP_NUM 9 +#define BNXT_ULP_PROTO_HDR_VXLAN_NUM 4 +#define BNXT_ULP_PROTO_HDR_MAX 128 + struct ulp_rte_hdr_bitmap { uint64_t bits; }; @@ -29,6 +41,20 @@ struct ulp_rte_hdr_field { uint32_t size; }; +/* Flow Parser Header Information Structure */ +struct bnxt_ulp_rte_hdr_info { + enum bnxt_ulp_hdr_type hdr_type; + /* Flow Parser Protocol Header Function Prototype */ + int (*proto_hdr_func)(const struct rte_flow_item *item_list, + struct ulp_rte_hdr_bitmap *hdr_bitmap, + struct ulp_rte_hdr_field *hdr_field, + uint32_t *field_idx, + uint32_t *vlan_idx); +}; + +/* Flow Parser Header Information Structure Array defined in template source*/ +extern struct bnxt_ulp_rte_hdr_info ulp_hdr_info[]; + struct bnxt_ulp_matcher_field_info { enum bnxt_ulp_fmf_mask mask_opcode; enum bnxt_ulp_fmf_spec spec_opcode;