From patchwork Thu Dec 17 09:22:51 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85280 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 56E45A09F0; Thu, 17 Dec 2020 10:28:27 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 2CAF2CA10; Thu, 17 Dec 2020 10:28:26 +0100 (CET) Received: from mga01.intel.com (mga01.intel.com [192.55.52.88]) by dpdk.org (Postfix) with ESMTP id 6864ACA10 for ; Thu, 17 Dec 2020 10:28:24 +0100 (CET) IronPort-SDR: 3XSswR7C2NslBIQptjU+08Ni2SVLFSoBo01fBrukc1Ei40mgmaOM9hQQWqu40d+MryFKx2pZ5G ZtXO3qFk/kxg== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="193610789" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="193610789" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga101.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:28:22 -0800 IronPort-SDR: /JG8AjLKj9EAqVsmqWUo4YVYp69TWsxQgmM9QfkTVkFPjhoDeF4qq2a4cswKYFEZxJvgGvNQXS nd40KgCzKtMg== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369745126" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:27:52 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:22:51 +0000 Message-Id: <20201217092312.27033-2-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 01/22] ethdev: fix MTU size exceeds max rx packet length X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" If max rx packet length is smaller then MTU + Ether overhead, that will drop all MTU size packets. Update the MTU size according to the max rx packet and Ether overhead. Fixes: 59d0ecdbf0e1 ("ethdev: MTU accessors") Signed-off-by: Steve Yang --- lib/librte_ethdev/rte_ethdev.c | 21 ++++++++++++++++++--- 1 file changed, 18 insertions(+), 3 deletions(-) diff --git a/lib/librte_ethdev/rte_ethdev.c b/lib/librte_ethdev/rte_ethdev.c index 17ddacc78d..ff6a1e675f 100644 --- a/lib/librte_ethdev/rte_ethdev.c +++ b/lib/librte_ethdev/rte_ethdev.c @@ -1292,6 +1292,7 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q, struct rte_eth_dev *dev; struct rte_eth_dev_info dev_info; struct rte_eth_conf orig_conf; + uint16_t overhead_len; int diag; int ret; @@ -1323,6 +1324,15 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q, if (ret != 0) goto rollback; + /* Get the real Ethernet overhead length */ + if (dev_info.max_mtu && + dev_info.max_mtu != UINT16_MAX && + dev_info.max_rx_pktlen && + dev_info.max_rx_pktlen > dev_info.max_mtu) + overhead_len = dev_info.max_rx_pktlen - dev_info.max_mtu; + else + overhead_len = RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN; + /* If number of queues specified by application for both Rx and Tx is * zero, use driver preferred values. This cannot be done individually * as it is valid for either Tx or Rx (but not both) to be zero. @@ -1410,13 +1420,18 @@ rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q, goto rollback; } } else { - if (dev_conf->rxmode.max_rx_pkt_len < RTE_ETHER_MIN_LEN || - dev_conf->rxmode.max_rx_pkt_len > RTE_ETHER_MAX_LEN) + uint16_t pktlen = dev_conf->rxmode.max_rx_pkt_len; + if (pktlen < RTE_ETHER_MIN_MTU + overhead_len || + pktlen > RTE_ETHER_MTU + overhead_len) /* Use default value */ dev->data->dev_conf.rxmode.max_rx_pkt_len = - RTE_ETHER_MAX_LEN; + RTE_ETHER_MTU + overhead_len; } + /* Scale the MTU size to adapt max_rx_pkt_len */ + dev->data->mtu = dev->data->dev_conf.rxmode.max_rx_pkt_len - + overhead_len; + /* * If LRO is enabled, check that the maximum aggregated packet * size is supported by the configured device. From patchwork Thu Dec 17 09:22:52 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85281 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 6ED48A09F0; Thu, 17 Dec 2020 10:28:52 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 4B4B0CA00; Thu, 17 Dec 2020 10:28:51 +0100 (CET) Received: from mga04.intel.com (mga04.intel.com [192.55.52.120]) by dpdk.org (Postfix) with ESMTP id 2FC35C9FC for ; Thu, 17 Dec 2020 10:28:48 +0100 (CET) IronPort-SDR: Ba4pGYzWo9RvaG6qOPJLRkYO2sAofPfC6vD80CCTz14Oz4N3VqoQIh1w4TocRgx8P6wETWxVGW xoI9sk8NfU+w== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="172650490" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="172650490" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:28:46 -0800 IronPort-SDR: UIbjj9aXlB19D2Q0+WFoNHnFfIU+C2Bv6XK8FQ9Qh2X97yH9WHOAniTOdnfAU17RccbbnhB2ex 56H4umGFPGXA== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369745979" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:28:22 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:22:52 +0000 Message-Id: <20201217092312.27033-3-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 02/22] app/testpmd: fix max rx packet length for VLAN packets X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" When the max rx packet length is smaller than the sum of mtu size and ether overhead size, it should be enlarged, otherwise the VLAN packets will be dropped. Removed the rx_offloads assignment for jumbo frame during command line parsing, and set the correct jumbo frame flag if MTU size is larger than the default value 'RTE_ETHER_MTU' within 'init_config()'. Fixes: 384161e00627 ("app/testpmd: adjust on the fly VLAN configuration") Fixes: 35b2d13fd6fd ("net: add rte prefix to ether defines") Fixes: ce17eddefc20 ("ethdev: introduce Rx queue offloads API") Fixes: 150c9ac2df13 ("app/testpmd: update Rx offload after setting MTU") Signed-off-by: Steve Yang --- app/test-pmd/cmdline.c | 6 ------ app/test-pmd/config.c | 2 +- app/test-pmd/parameters.c | 7 ++----- app/test-pmd/testpmd.c | 20 ++++++++++++++++++++ 4 files changed, 23 insertions(+), 12 deletions(-) diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c index 2ccbaa039e..65042fcff5 100644 --- a/app/test-pmd/cmdline.c +++ b/app/test-pmd/cmdline.c @@ -1886,7 +1886,6 @@ cmd_config_max_pkt_len_parsed(void *parsed_result, RTE_ETH_FOREACH_DEV(pid) { struct rte_port *port = &ports[pid]; - uint64_t rx_offloads = port->dev_conf.rxmode.offloads; if (!strcmp(res->name, "max-pkt-len")) { if (res->value < RTE_ETHER_MIN_LEN) { @@ -1898,11 +1897,6 @@ cmd_config_max_pkt_len_parsed(void *parsed_result, return; port->dev_conf.rxmode.max_rx_pkt_len = res->value; - if (res->value > RTE_ETHER_MAX_LEN) - rx_offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; - else - rx_offloads &= ~DEV_RX_OFFLOAD_JUMBO_FRAME; - port->dev_conf.rxmode.offloads = rx_offloads; } else { printf("Unknown parameter\n"); return; diff --git a/app/test-pmd/config.c b/app/test-pmd/config.c index 3f6c8642b1..1195f054f3 100644 --- a/app/test-pmd/config.c +++ b/app/test-pmd/config.c @@ -1434,7 +1434,7 @@ port_mtu_set(portid_t port_id, uint16_t mtu) * device supports jumbo frame. */ eth_overhead = dev_info.max_rx_pktlen - dev_info.max_mtu; - if (mtu > RTE_ETHER_MAX_LEN - eth_overhead) { + if (mtu > RTE_ETHER_MTU) { rte_port->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; rte_port->dev_conf.rxmode.max_rx_pkt_len = diff --git a/app/test-pmd/parameters.c b/app/test-pmd/parameters.c index 414a0068fb..df5eb10d84 100644 --- a/app/test-pmd/parameters.c +++ b/app/test-pmd/parameters.c @@ -834,12 +834,9 @@ launch_args_parse(int argc, char** argv) } if (!strcmp(lgopts[opt_idx].name, "max-pkt-len")) { n = atoi(optarg); - if (n >= RTE_ETHER_MIN_LEN) { + if (n >= RTE_ETHER_MIN_LEN) rx_mode.max_rx_pkt_len = (uint32_t) n; - if (n > RTE_ETHER_MAX_LEN) - rx_offloads |= - DEV_RX_OFFLOAD_JUMBO_FRAME; - } else + else rte_exit(EXIT_FAILURE, "Invalid max-pkt-len=%d - should be > %d\n", n, RTE_ETHER_MIN_LEN); diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c index 33a060dffd..fdb5c3c38b 100644 --- a/app/test-pmd/testpmd.c +++ b/app/test-pmd/testpmd.c @@ -1410,6 +1410,7 @@ init_config(void) struct rte_gro_param gro_param; uint32_t gso_types; uint16_t data_size; + uint16_t eth_overhead; bool warning = 0; int k; int ret; @@ -1446,6 +1447,25 @@ init_config(void) rte_exit(EXIT_FAILURE, "rte_eth_dev_info_get() failed\n"); + /* Update the max_rx_pkt_len to have MTU as RTE_ETHER_MTU */ + if (port->dev_info.max_mtu && + port->dev_info.max_mtu != UINT16_MAX && + port->dev_info.max_rx_pktlen && + port->dev_info.max_rx_pktlen > port->dev_info.max_mtu) + eth_overhead = port->dev_info.max_rx_pktlen - + port->dev_info.max_mtu; + else + eth_overhead = + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN; + + if (port->dev_conf.rxmode.max_rx_pkt_len <= + (uint32_t)(RTE_ETHER_MTU + eth_overhead)) + port->dev_conf.rxmode.max_rx_pkt_len = + RTE_ETHER_MTU + eth_overhead; + else + port->dev_conf.rxmode.offloads |= + DEV_RX_OFFLOAD_JUMBO_FRAME; + if (!(port->dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE)) port->dev_conf.txmode.offloads &= From patchwork Thu Dec 17 09:22:53 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85282 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 2CC59A09F0; Thu, 17 Dec 2020 10:29:22 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 114CACA1A; Thu, 17 Dec 2020 10:29:21 +0100 (CET) Received: from mga05.intel.com (mga05.intel.com [192.55.52.43]) by dpdk.org (Postfix) with ESMTP id 28FE7CA1A for ; Thu, 17 Dec 2020 10:29:18 +0100 (CET) IronPort-SDR: HVKnGRrgtuREZPjneU3i+UaftcqICqKPDeVzakca3YaJIBHuuWVRs5S2j2M2tLMqeGgkB3mj65 bwOUv/A2cwrg== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="259944034" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="259944034" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga105.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:29:17 -0800 IronPort-SDR: ursF79ucE+RixEUYk+SQkxCalLJ6lxgw7OsmBaslc6DrjSiOwgow/HGZ0ssHxNt6k24Nv2Fs2s SyTi97VBTv/A== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369746932" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:28:46 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:22:53 +0000 Message-Id: <20201217092312.27033-4-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 03/22] net/dpaa: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports VLAN tag. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: 25f854197abc ("net/dpaa: support jumbo frames") Signed-off-by: Steve Yang --- drivers/net/dpaa/dpaa_ethdev.c | 2 +- drivers/net/dpaa/dpaa_ethdev.h | 4 ++++ 2 files changed, 5 insertions(+), 1 deletion(-) diff --git a/drivers/net/dpaa/dpaa_ethdev.c b/drivers/net/dpaa/dpaa_ethdev.c index f00279e004..0c87c136d7 100644 --- a/drivers/net/dpaa/dpaa_ethdev.c +++ b/drivers/net/dpaa/dpaa_ethdev.c @@ -184,7 +184,7 @@ dpaa_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) return -EINVAL; } - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > DPAA_ETH_MAX_LEN) dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else diff --git a/drivers/net/dpaa/dpaa_ethdev.h b/drivers/net/dpaa/dpaa_ethdev.h index 659bceb467..a858b1372c 100644 --- a/drivers/net/dpaa/dpaa_ethdev.h +++ b/drivers/net/dpaa/dpaa_ethdev.h @@ -51,6 +51,10 @@ #define VLAN_TAG_SIZE 4 /** < Vlan Header Length */ #endif +#define DPAA_ETH_MAX_LEN (RTE_ETHER_MTU + \ + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + \ + VLAN_TAG_SIZE) + /* PCD frame queues */ #define DPAA_DEFAULT_NUM_PCD_QUEUES 1 #define DPAA_VSP_PROFILE_MAX_NUM 8 From patchwork Thu Dec 17 09:22:54 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85283 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id CCF20A09F0; Thu, 17 Dec 2020 10:29:59 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id AE9A2CA3A; Thu, 17 Dec 2020 10:29:58 +0100 (CET) Received: from mga04.intel.com (mga04.intel.com [192.55.52.120]) by dpdk.org (Postfix) with ESMTP id 16EBFCA38 for ; Thu, 17 Dec 2020 10:29:55 +0100 (CET) IronPort-SDR: BCG4xpiHJBaRVKKKsurloLfgpJG9r9HURl9RbhNqP0nETbuoxaxkYtntjet5R6IWsZ8yqafJT9 KI+4IjpQXhUg== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="172650635" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="172650635" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:29:42 -0800 IronPort-SDR: xa3TSwudyz9XAKH7rCmxUivVmeIrjL/NY8f2pTgRCZEp1bqFFPASd92S69cjv7zh4PeB6Bd+0x fg0//JIzvk0g== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369747758" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:29:17 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:22:54 +0000 Message-Id: <20201217092312.27033-5-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 04/22] net/dpaa2: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports VLAN tag. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: e16408499412 ("net/dpaa2: configure jumbo frames") Signed-off-by: Steve Yang --- drivers/net/dpaa2/dpaa2_ethdev.c | 2 +- drivers/net/dpaa2/dpaa2_ethdev.h | 4 ++++ 2 files changed, 5 insertions(+), 1 deletion(-) diff --git a/drivers/net/dpaa2/dpaa2_ethdev.c b/drivers/net/dpaa2/dpaa2_ethdev.c index ab6863300e..6f38da3cce 100644 --- a/drivers/net/dpaa2/dpaa2_ethdev.c +++ b/drivers/net/dpaa2/dpaa2_ethdev.c @@ -1420,7 +1420,7 @@ dpaa2_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) if (mtu < RTE_ETHER_MIN_MTU || frame_size > DPAA2_MAX_RX_PKT_LEN) return -EINVAL; - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > DPAA2_ETH_MAX_LEN) dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else diff --git a/drivers/net/dpaa2/dpaa2_ethdev.h b/drivers/net/dpaa2/dpaa2_ethdev.h index 8d82f74684..cacb11bd3e 100644 --- a/drivers/net/dpaa2/dpaa2_ethdev.h +++ b/drivers/net/dpaa2/dpaa2_ethdev.h @@ -26,6 +26,10 @@ #define DPAA2_RX_DEFAULT_NBDESC 512 +#define DPAA2_ETH_MAX_LEN (RTE_ETHER_MTU + \ + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + \ + VLAN_TAG_SIZE) + /*default tc to be used for ,congestion, distribution etc configuration. */ #define DPAA2_DEF_TC 0 From patchwork Thu Dec 17 09:22:55 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85284 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 7AEA4A09F0; Thu, 17 Dec 2020 10:30:15 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 43D09CA38; Thu, 17 Dec 2020 10:30:10 +0100 (CET) Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by dpdk.org (Postfix) with ESMTP id 0798BCA1A for ; Thu, 17 Dec 2020 10:30:07 +0100 (CET) IronPort-SDR: mFW6ZBhfkpgCVsuWoRZHkUMcxD6PLYQHcnoyeBZAMCnktyBFznnqce3lY5YkYXymSpCuDPbTDf mfcD52aF3zuw== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="239309085" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="239309085" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:30:06 -0800 IronPort-SDR: mhsaJ4PCz4xKxvifpRFafl56GGQ+TQtaavkGDceBSJB3dw3+OY8dI/Am350XXsVeMurVfCBrcI 9tnn1EuOYZ5w== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369748708" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:29:42 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:22:55 +0000 Message-Id: <20201217092312.27033-6-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 05/22] net/e1000: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports VLAN tag. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: ef990fb56e55 ("net/e1000: convert to new Rx offloads API") Signed-off-by: Steve Yang Acked-by: Jeff Guo --- drivers/net/e1000/e1000_ethdev.h | 2 +- drivers/net/e1000/em_ethdev.c | 5 ++--- drivers/net/e1000/igb_ethdev.c | 2 +- 3 files changed, 4 insertions(+), 5 deletions(-) diff --git a/drivers/net/e1000/e1000_ethdev.h b/drivers/net/e1000/e1000_ethdev.h index 4755a5f333..3b4d9c3ee6 100644 --- a/drivers/net/e1000/e1000_ethdev.h +++ b/drivers/net/e1000/e1000_ethdev.h @@ -97,7 +97,7 @@ */ #define E1000_ETH_OVERHEAD (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + \ VLAN_TAG_SIZE) - +#define E1000_ETH_MAX_LEN (RTE_ETHER_MTU + E1000_ETH_OVERHEAD) /* * Maximum number of Ring Descriptors. * diff --git a/drivers/net/e1000/em_ethdev.c b/drivers/net/e1000/em_ethdev.c index 8ee9422bf4..2036c6e917 100644 --- a/drivers/net/e1000/em_ethdev.c +++ b/drivers/net/e1000/em_ethdev.c @@ -1799,8 +1799,7 @@ eth_em_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) if (ret != 0) return ret; - frame_size = mtu + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + - VLAN_TAG_SIZE; + frame_size = mtu + E1000_ETH_OVERHEAD; /* check that mtu is within the allowed range */ if (mtu < RTE_ETHER_MIN_MTU || frame_size > dev_info.max_rx_pktlen) @@ -1816,7 +1815,7 @@ eth_em_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) rctl = E1000_READ_REG(hw, E1000_RCTL); /* switch to jumbo mode if needed */ - if (frame_size > RTE_ETHER_MAX_LEN) { + if (frame_size > E1000_ETH_MAX_LEN) { dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; rctl |= E1000_RCTL_LPE; diff --git a/drivers/net/e1000/igb_ethdev.c b/drivers/net/e1000/igb_ethdev.c index 647aa8d995..dfe87508c2 100644 --- a/drivers/net/e1000/igb_ethdev.c +++ b/drivers/net/e1000/igb_ethdev.c @@ -4369,7 +4369,7 @@ eth_igb_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) rctl = E1000_READ_REG(hw, E1000_RCTL); /* switch to jumbo mode if needed */ - if (frame_size > RTE_ETHER_MAX_LEN) { + if (frame_size > E1000_ETH_MAX_LEN) { dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; rctl |= E1000_RCTL_LPE; From patchwork Thu Dec 17 09:22:56 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85285 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 4FDD5A09F0; Thu, 17 Dec 2020 10:30:41 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 2B1F2CA3A; Thu, 17 Dec 2020 10:30:40 +0100 (CET) Received: from mga14.intel.com (mga14.intel.com [192.55.52.115]) by dpdk.org (Postfix) with ESMTP id 03BDACA1A for ; Thu, 17 Dec 2020 10:30:37 +0100 (CET) IronPort-SDR: ePHbY+IxCfiLdh3rUqUNvMkLQwIwQt8Zx20fApngRFcQgBrfwlTj3mnxKUYXFzYt8qlsb6wduP bl3g/GcMWJzw== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="174450114" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="174450114" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga103.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:30:34 -0800 IronPort-SDR: VPWnOrtmeZu0tgL4m/brcfwcI/qX+lIzkbUO01Cz3HwGL7wUbV3Kj09rkFkp+WH8tGhIP8GPUp QWLIDmIQ6tCg== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369749692" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:30:06 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:22:56 +0000 Message-Id: <20201217092312.27033-7-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 06/22] net/hns3: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports dual VLAN tags. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'HSN3_DEFAULT_FRAME_LEN'. Fixes: 1f5ca0b460cd ("net/hns3: support some device operations") Fixes: a5475d61fa34 ("net/hns3: support VF") Signed-off-by: Steve Yang --- drivers/net/hns3/hns3_ethdev.c | 2 +- drivers/net/hns3/hns3_ethdev_vf.c | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/drivers/net/hns3/hns3_ethdev.c b/drivers/net/hns3/hns3_ethdev.c index 7c34e382fb..10e0c0de46 100644 --- a/drivers/net/hns3/hns3_ethdev.c +++ b/drivers/net/hns3/hns3_ethdev.c @@ -2467,7 +2467,7 @@ hns3_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) } rte_spinlock_lock(&hw->lock); - is_jumbo_frame = frame_size > RTE_ETHER_MAX_LEN ? true : false; + is_jumbo_frame = frame_size > HNS3_DEFAULT_FRAME_LEN ? true : false; frame_size = RTE_MAX(frame_size, HNS3_DEFAULT_FRAME_LEN); /* diff --git a/drivers/net/hns3/hns3_ethdev_vf.c b/drivers/net/hns3/hns3_ethdev_vf.c index f09cabcd82..ef03fb1c4e 100644 --- a/drivers/net/hns3/hns3_ethdev_vf.c +++ b/drivers/net/hns3/hns3_ethdev_vf.c @@ -928,7 +928,7 @@ hns3vf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) rte_spinlock_unlock(&hw->lock); return ret; } - if (frame_size > RTE_ETHER_MAX_LEN) + if (mtu > RTE_ETHER_MTU) dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else From patchwork Thu Dec 17 09:22:57 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85286 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id C1321A09F0; Thu, 17 Dec 2020 10:31:06 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 8C3E3CA4A; Thu, 17 Dec 2020 10:31:05 +0100 (CET) Received: from mga01.intel.com (mga01.intel.com [192.55.52.88]) by dpdk.org (Postfix) with ESMTP id AF99ACA4A for ; Thu, 17 Dec 2020 10:31:03 +0100 (CET) IronPort-SDR: JkJpV9LR4Nt/p9W/2XRQIQtPLxJ3fM1FkjOaD1KRntFzFKmAeoohCCJiLXhYGoirtTyVPD4MdP 8P0bKWF9ciTg== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="193611139" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="193611139" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga101.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:31:01 -0800 IronPort-SDR: 5aJL5JW9gtTdEC+B4/7U/uNFhQu/GrQPL0V+4agAM0Rm+YCNHqwXxjLGfZ+3Rh5vrYk3cPNxhX 18FvPxDNrXlw== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369750664" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:30:34 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:22:57 +0000 Message-Id: <20201217092312.27033-8-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 07/22] net/i40e: fix the jumbo frame flag condition X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports dual VLAN tags. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: c1715402df8f ("i40evf: fix jumbo frame support") Fixes: 43e5488c0ac6 ("net/i40e: support MTU configuration") Fixes: a778a1fa2e4e ("i40e: set up and initialize flow director") Fixes: c3ac7c5b0b8a ("net/i40e: convert to new Rx offloads API") Signed-off-by: Steve Yang Acked-by: Jeff Guo --- drivers/net/i40e/i40e_ethdev.c | 2 +- drivers/net/i40e/i40e_ethdev.h | 1 + drivers/net/i40e/i40e_ethdev_vf.c | 10 +++++----- drivers/net/i40e/i40e_fdir.c | 2 +- drivers/net/i40e/i40e_rxtx.c | 8 ++++---- 5 files changed, 12 insertions(+), 11 deletions(-) diff --git a/drivers/net/i40e/i40e_ethdev.c b/drivers/net/i40e/i40e_ethdev.c index 2cb18ecc03..d4fcc6c853 100644 --- a/drivers/net/i40e/i40e_ethdev.c +++ b/drivers/net/i40e/i40e_ethdev.c @@ -11770,7 +11770,7 @@ i40e_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) return -EBUSY; } - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > I40E_ETH_MAX_LEN) dev_data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else diff --git a/drivers/net/i40e/i40e_ethdev.h b/drivers/net/i40e/i40e_ethdev.h index 696c5aaf7e..1f31a532a1 100644 --- a/drivers/net/i40e/i40e_ethdev.h +++ b/drivers/net/i40e/i40e_ethdev.h @@ -281,6 +281,7 @@ struct rte_flow { */ #define I40E_ETH_OVERHEAD \ (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + I40E_VLAN_TAG_SIZE * 2) +#define I40E_ETH_MAX_LEN (RTE_ETHER_MTU + I40E_ETH_OVERHEAD) #define I40E_RXTX_BYTES_H_16_BIT(bytes) ((bytes) & ~I40E_48_BIT_MASK) #define I40E_RXTX_BYTES_L_48_BIT(bytes) ((bytes) & I40E_48_BIT_MASK) diff --git a/drivers/net/i40e/i40e_ethdev_vf.c b/drivers/net/i40e/i40e_ethdev_vf.c index c26b036b85..85c65c09ee 100644 --- a/drivers/net/i40e/i40e_ethdev_vf.c +++ b/drivers/net/i40e/i40e_ethdev_vf.c @@ -1889,22 +1889,22 @@ i40evf_rxq_init(struct rte_eth_dev *dev, struct i40e_rx_queue *rxq) * Check if the jumbo frame and maximum packet length are set correctly */ if (dev_data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) { - if (rxq->max_pkt_len <= RTE_ETHER_MAX_LEN || + if (rxq->max_pkt_len <= I40E_ETH_MAX_LEN || rxq->max_pkt_len > I40E_FRAME_SIZE_MAX) { PMD_DRV_LOG(ERR, "maximum packet length must be " "larger than %u and smaller than %u, as jumbo " - "frame is enabled", (uint32_t)RTE_ETHER_MAX_LEN, + "frame is enabled", (uint32_t)I40E_ETH_MAX_LEN, (uint32_t)I40E_FRAME_SIZE_MAX); return I40E_ERR_CONFIG; } } else { if (rxq->max_pkt_len < RTE_ETHER_MIN_LEN || - rxq->max_pkt_len > RTE_ETHER_MAX_LEN) { + rxq->max_pkt_len > I40E_ETH_MAX_LEN) { PMD_DRV_LOG(ERR, "maximum packet length must be " "larger than %u and smaller than %u, as jumbo " "frame is disabled", (uint32_t)RTE_ETHER_MIN_LEN, - (uint32_t)RTE_ETHER_MAX_LEN); + (uint32_t)I40E_ETH_MAX_LEN); return I40E_ERR_CONFIG; } } @@ -2825,7 +2825,7 @@ i40evf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) return -EBUSY; } - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > I40E_ETH_MAX_LEN) dev_data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else diff --git a/drivers/net/i40e/i40e_fdir.c b/drivers/net/i40e/i40e_fdir.c index 50c0eee9f2..449b67b2a9 100644 --- a/drivers/net/i40e/i40e_fdir.c +++ b/drivers/net/i40e/i40e_fdir.c @@ -116,7 +116,7 @@ i40e_fdir_rx_queue_init(struct i40e_rx_queue *rxq) #endif rx_ctx.dtype = i40e_header_split_none; rx_ctx.hsplit_0 = I40E_HEADER_SPLIT_NONE; - rx_ctx.rxmax = RTE_ETHER_MAX_LEN; + rx_ctx.rxmax = I40E_ETH_MAX_LEN; rx_ctx.tphrdesc_ena = 1; rx_ctx.tphwdesc_ena = 1; rx_ctx.tphdata_ena = 1; diff --git a/drivers/net/i40e/i40e_rxtx.c b/drivers/net/i40e/i40e_rxtx.c index 5df9a9df56..b8859bbff2 100644 --- a/drivers/net/i40e/i40e_rxtx.c +++ b/drivers/net/i40e/i40e_rxtx.c @@ -2797,23 +2797,23 @@ i40e_rx_queue_config(struct i40e_rx_queue *rxq) RTE_MIN((uint32_t)(hw->func_caps.rx_buf_chain_len * rxq->rx_buf_len), data->dev_conf.rxmode.max_rx_pkt_len); if (data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) { - if (rxq->max_pkt_len <= RTE_ETHER_MAX_LEN || + if (rxq->max_pkt_len <= I40E_ETH_MAX_LEN || rxq->max_pkt_len > I40E_FRAME_SIZE_MAX) { PMD_DRV_LOG(ERR, "maximum packet length must " "be larger than %u and smaller than %u," "as jumbo frame is enabled", - (uint32_t)RTE_ETHER_MAX_LEN, + (uint32_t)I40E_ETH_MAX_LEN, (uint32_t)I40E_FRAME_SIZE_MAX); return I40E_ERR_CONFIG; } } else { if (rxq->max_pkt_len < RTE_ETHER_MIN_LEN || - rxq->max_pkt_len > RTE_ETHER_MAX_LEN) { + rxq->max_pkt_len > I40E_ETH_MAX_LEN) { PMD_DRV_LOG(ERR, "maximum packet length must be " "larger than %u and smaller than %u, " "as jumbo frame is disabled", (uint32_t)RTE_ETHER_MIN_LEN, - (uint32_t)RTE_ETHER_MAX_LEN); + (uint32_t)I40E_ETH_MAX_LEN); return I40E_ERR_CONFIG; } } From patchwork Thu Dec 17 09:22:58 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85287 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id EA76EA09F0; Thu, 17 Dec 2020 10:31:38 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id C1590CA4C; Thu, 17 Dec 2020 10:31:37 +0100 (CET) Received: from mga04.intel.com (mga04.intel.com [192.55.52.120]) by dpdk.org (Postfix) with ESMTP id 043F6CA4C for ; Thu, 17 Dec 2020 10:31:34 +0100 (CET) IronPort-SDR: 7swxz43poRNng6DF67nxT8Yd796NZcCAdJA7ZWx1CGRK4Hz8hFeiYrCeM/au/1jyoOHpVJX+r+ 4te//zj3DY5Q== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="172650982" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="172650982" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:31:33 -0800 IronPort-SDR: 2zKBFErrAC/WZytXjPUHCb1NsuakM4E+hfdJhMDyOK6lwNEe8w8VC/GqRkvsPpF0IUXQO3eACP h25Z8OomxHsA== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369751750" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:31:01 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:22:58 +0000 Message-Id: <20201217092312.27033-9-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 08/22] net/iavf: fix the jumbo frame flag condition X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports dual VLAN tags. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: 3fd7a3719c66 ("net/avf: enable ops for MTU setting") Fixes: 69dd4c3d0898 ("net/avf: enable queue and device") Signed-off-by: Steve Yang --- drivers/net/iavf/iavf.h | 1 + drivers/net/iavf/iavf_ethdev.c | 10 +++++----- 2 files changed, 6 insertions(+), 5 deletions(-) diff --git a/drivers/net/iavf/iavf.h b/drivers/net/iavf/iavf.h index 6d5912d8c1..3328bd9327 100644 --- a/drivers/net/iavf/iavf.h +++ b/drivers/net/iavf/iavf.h @@ -66,6 +66,7 @@ #define IAVF_VLAN_TAG_SIZE 4 #define IAVF_ETH_OVERHEAD \ (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + IAVF_VLAN_TAG_SIZE * 2) +#define IAVF_ETH_MAX_LEN (RTE_ETHER_MTU + IAVF_ETH_OVERHEAD) #define IAVF_32_BIT_WIDTH (CHAR_BIT * 4) #define IAVF_48_BIT_WIDTH (CHAR_BIT * 6) diff --git a/drivers/net/iavf/iavf_ethdev.c b/drivers/net/iavf/iavf_ethdev.c index 7e3c26a94e..658f28d353 100644 --- a/drivers/net/iavf/iavf_ethdev.c +++ b/drivers/net/iavf/iavf_ethdev.c @@ -418,23 +418,23 @@ iavf_init_rxq(struct rte_eth_dev *dev, struct iavf_rx_queue *rxq) * correctly. */ if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) { - if (max_pkt_len <= RTE_ETHER_MAX_LEN || + if (max_pkt_len <= IAVF_ETH_MAX_LEN || max_pkt_len > IAVF_FRAME_SIZE_MAX) { PMD_DRV_LOG(ERR, "maximum packet length must be " "larger than %u and smaller than %u, " "as jumbo frame is enabled", - (uint32_t)RTE_ETHER_MAX_LEN, + (uint32_t)IAVF_ETH_MAX_LEN, (uint32_t)IAVF_FRAME_SIZE_MAX); return -EINVAL; } } else { if (max_pkt_len < RTE_ETHER_MIN_LEN || - max_pkt_len > RTE_ETHER_MAX_LEN) { + max_pkt_len > IAVF_ETH_MAX_LEN) { PMD_DRV_LOG(ERR, "maximum packet length must be " "larger than %u and smaller than %u, " "as jumbo frame is disabled", (uint32_t)RTE_ETHER_MIN_LEN, - (uint32_t)RTE_ETHER_MAX_LEN); + (uint32_t)IAVF_ETH_MAX_LEN); return -EINVAL; } } @@ -1167,7 +1167,7 @@ iavf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) return -EBUSY; } - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > IAVF_ETH_MAX_LEN) dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else From patchwork Thu Dec 17 09:22:59 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85288 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 6BC70A09F0; Thu, 17 Dec 2020 10:32:07 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 43ABFCA4A; Thu, 17 Dec 2020 10:32:06 +0100 (CET) Received: from mga18.intel.com (mga18.intel.com [134.134.136.126]) by dpdk.org (Postfix) with ESMTP id 84208CA4A for ; Thu, 17 Dec 2020 10:32:04 +0100 (CET) IronPort-SDR: es7J4Sqac2jROLv7uq3JqCTW/sUKbGOgOIQJ2W0If//HGnimFe928QKBggTyW4p2vMRK3CUq2b xX7ytRZ6+N3A== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="162961443" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="162961443" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by orsmga106.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:32:02 -0800 IronPort-SDR: PgXuae33URubO/94NUh0dSUoGZR6QWiseoR4FZQqPO6iSnvfU+UCAQjXBKp9lmV+7sLoc/Q7Bt 5QGwlnFNz5Sw== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369752882" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:31:33 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:22:59 +0000 Message-Id: <20201217092312.27033-10-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 09/22] net/ice: fix the jumbo frame flag condition X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports dual VLAN tags. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: 84dc7a95a2d3 ("net/ice: enable flow director engine") Fixes: 1b009275e2c8 ("net/ice: add Rx queue init in DCF") Fixes: ae2bdd0219cb ("net/ice: support MTU setting") Fixes: 50370662b727 ("net/ice: support device and queue ops") Signed-off-by: Steve Yang --- drivers/net/ice/ice_dcf_ethdev.c | 8 ++++---- drivers/net/ice/ice_ethdev.c | 2 +- drivers/net/ice/ice_ethdev.h | 1 + drivers/net/ice/ice_rxtx.c | 10 +++++----- 4 files changed, 11 insertions(+), 10 deletions(-) diff --git a/drivers/net/ice/ice_dcf_ethdev.c b/drivers/net/ice/ice_dcf_ethdev.c index b0b2ecb0d6..e5c877805f 100644 --- a/drivers/net/ice/ice_dcf_ethdev.c +++ b/drivers/net/ice/ice_dcf_ethdev.c @@ -60,23 +60,23 @@ ice_dcf_init_rxq(struct rte_eth_dev *dev, struct ice_rx_queue *rxq) * correctly. */ if (dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) { - if (max_pkt_len <= RTE_ETHER_MAX_LEN || + if (max_pkt_len <= ICE_ETH_MAX_LEN || max_pkt_len > ICE_FRAME_SIZE_MAX) { PMD_DRV_LOG(ERR, "maximum packet length must be " "larger than %u and smaller than %u, " "as jumbo frame is enabled", - (uint32_t)RTE_ETHER_MAX_LEN, + (uint32_t)ICE_ETH_MAX_LEN, (uint32_t)ICE_FRAME_SIZE_MAX); return -EINVAL; } } else { if (max_pkt_len < RTE_ETHER_MIN_LEN || - max_pkt_len > RTE_ETHER_MAX_LEN) { + max_pkt_len > ICE_ETH_MAX_LEN) { PMD_DRV_LOG(ERR, "maximum packet length must be " "larger than %u and smaller than %u, " "as jumbo frame is disabled", (uint32_t)RTE_ETHER_MIN_LEN, - (uint32_t)RTE_ETHER_MAX_LEN); + (uint32_t)ICE_ETH_MAX_LEN); return -EINVAL; } } diff --git a/drivers/net/ice/ice_ethdev.c b/drivers/net/ice/ice_ethdev.c index 9a5d6a559f..d25aed54b3 100644 --- a/drivers/net/ice/ice_ethdev.c +++ b/drivers/net/ice/ice_ethdev.c @@ -3904,7 +3904,7 @@ ice_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) return -EBUSY; } - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > ICE_ETH_MAX_LEN) dev_data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else diff --git a/drivers/net/ice/ice_ethdev.h b/drivers/net/ice/ice_ethdev.h index 899f446cde..2b03c59671 100644 --- a/drivers/net/ice/ice_ethdev.h +++ b/drivers/net/ice/ice_ethdev.h @@ -135,6 +135,7 @@ */ #define ICE_ETH_OVERHEAD \ (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + ICE_VLAN_TAG_SIZE * 2) +#define ICE_ETH_MAX_LEN (RTE_ETHER_MTU + ICE_ETH_OVERHEAD) #define ICE_RXTX_BYTES_HIGH(bytes) ((bytes) & ~ICE_40_BIT_MASK) #define ICE_RXTX_BYTES_LOW(bytes) ((bytes) & ICE_40_BIT_MASK) diff --git a/drivers/net/ice/ice_rxtx.c b/drivers/net/ice/ice_rxtx.c index d052bd0f1b..c98328ce0b 100644 --- a/drivers/net/ice/ice_rxtx.c +++ b/drivers/net/ice/ice_rxtx.c @@ -246,23 +246,23 @@ ice_program_hw_rx_queue(struct ice_rx_queue *rxq) dev->data->dev_conf.rxmode.max_rx_pkt_len); if (rxmode->offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) { - if (rxq->max_pkt_len <= RTE_ETHER_MAX_LEN || + if (rxq->max_pkt_len <= ICE_ETH_MAX_LEN || rxq->max_pkt_len > ICE_FRAME_SIZE_MAX) { PMD_DRV_LOG(ERR, "maximum packet length must " "be larger than %u and smaller than %u," "as jumbo frame is enabled", - (uint32_t)RTE_ETHER_MAX_LEN, + (uint32_t)ICE_ETH_MAX_LEN, (uint32_t)ICE_FRAME_SIZE_MAX); return -EINVAL; } } else { if (rxq->max_pkt_len < RTE_ETHER_MIN_LEN || - rxq->max_pkt_len > RTE_ETHER_MAX_LEN) { + rxq->max_pkt_len > ICE_ETH_MAX_LEN) { PMD_DRV_LOG(ERR, "maximum packet length must be " "larger than %u and smaller than %u, " "as jumbo frame is disabled", (uint32_t)RTE_ETHER_MIN_LEN, - (uint32_t)RTE_ETHER_MAX_LEN); + (uint32_t)ICE_ETH_MAX_LEN); return -EINVAL; } } @@ -701,7 +701,7 @@ ice_fdir_program_hw_rx_queue(struct ice_rx_queue *rxq) rx_ctx.hbuf = rxq->rx_hdr_len >> ICE_RLAN_CTX_HBUF_S; rx_ctx.dtype = 0; /* No Header Split mode */ rx_ctx.dsize = 1; /* 32B descriptors */ - rx_ctx.rxmax = RTE_ETHER_MAX_LEN; + rx_ctx.rxmax = ICE_ETH_MAX_LEN; /* TPH: Transaction Layer Packet (TLP) processing hints */ rx_ctx.tphrdesc_ena = 1; rx_ctx.tphwdesc_ena = 1; From patchwork Thu Dec 17 09:23:00 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85289 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 6024DA09F0; Thu, 17 Dec 2020 10:32:34 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 48337CA59; Thu, 17 Dec 2020 10:32:33 +0100 (CET) Received: from mga14.intel.com (mga14.intel.com [192.55.52.115]) by dpdk.org (Postfix) with ESMTP id 4677ACA4C for ; Thu, 17 Dec 2020 10:32:32 +0100 (CET) IronPort-SDR: mZ0zrtlrzADQKnG1Mm/ZnKVGKB1g7+y7fvdYdC/KWOlg8paCr3YzRalWH7JoYPfKtG++zqKx1y 3r7CRD7xAqZw== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="174450383" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="174450383" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga103.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:32:30 -0800 IronPort-SDR: dlJ/jgr2XmMQGTKmLwZ++3R/SapL4ivH5MM4TfmWv86BRHtJ8ffq47kvgErxjC07tHemLMGDA2 bP5CHp3snhhQ== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369753849" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:32:02 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:00 +0000 Message-Id: <20201217092312.27033-11-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 10/22] net/ipn3ke: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports dual VLAN tags. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: 70d6b7f550f4 ("net/ipn3ke: add representor") Signed-off-by: Steve Yang Reviewed-by: Rosen Xu --- drivers/net/ipn3ke/ipn3ke_ethdev.h | 1 + drivers/net/ipn3ke/ipn3ke_representor.c | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/drivers/net/ipn3ke/ipn3ke_ethdev.h b/drivers/net/ipn3ke/ipn3ke_ethdev.h index 9b0cf309c8..a6815a9cca 100644 --- a/drivers/net/ipn3ke/ipn3ke_ethdev.h +++ b/drivers/net/ipn3ke/ipn3ke_ethdev.h @@ -640,6 +640,7 @@ ipn3ke_tm_ops_get(struct rte_eth_dev *ethdev, */ #define IPN3KE_ETH_OVERHEAD \ (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + IPN3KE_VLAN_TAG_SIZE * 2) +#define IPN3KE_ETH_MAX_LEN (RTE_ETHER_MTU + IPN3KE_ETH_OVERHEAD) #define IPN3KE_MAC_FRAME_SIZE_MAX 9728 #define IPN3KE_MAC_RX_FRAME_MAXLENGTH 0x00AE diff --git a/drivers/net/ipn3ke/ipn3ke_representor.c b/drivers/net/ipn3ke/ipn3ke_representor.c index 8a53602576..9e15cce34f 100644 --- a/drivers/net/ipn3ke/ipn3ke_representor.c +++ b/drivers/net/ipn3ke/ipn3ke_representor.c @@ -2801,7 +2801,7 @@ ipn3ke_rpst_mtu_set(struct rte_eth_dev *ethdev, uint16_t mtu) return -EBUSY; } - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > IPN3KE_ETH_MAX_LEN) dev_data->dev_conf.rxmode.offloads |= (uint64_t)(DEV_RX_OFFLOAD_JUMBO_FRAME); else From patchwork Thu Dec 17 09:23:01 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85290 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 2343BA09F0; Thu, 17 Dec 2020 10:33:06 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id F036ACA51; Thu, 17 Dec 2020 10:33:04 +0100 (CET) Received: from mga01.intel.com (mga01.intel.com [192.55.52.88]) by dpdk.org (Postfix) with ESMTP id 3BEFBCA4C for ; Thu, 17 Dec 2020 10:33:02 +0100 (CET) IronPort-SDR: 0rmFj+Fn21meqH+nG60kAUIyJU73slKjexaQ0GA9q4LLThKUdB8EQLMO0NJTka/QmiSZ26pS2s +mxRJGbNy9Ew== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="193611431" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="193611431" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga101.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:33:01 -0800 IronPort-SDR: ZptatX94hP3Pwckyg8nU6H8KGEYq6quV1GnYMxkl2CrG/JFk8Fnxk2v+88OnCpvOpLbgX0t/eN 97T5ECa7Jd9g== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369754923" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:32:30 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:01 +0000 Message-Id: <20201217092312.27033-12-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 11/22] net/octeontx: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports dual VLAN tags. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: 3151e6a687a3 ("net/octeontx: support MTU") Signed-off-by: Steve Yang Acked-by: Harman Kalra --- drivers/net/octeontx/octeontx_ethdev.c | 2 +- drivers/net/octeontx/octeontx_ethdev.h | 1 + 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/drivers/net/octeontx/octeontx_ethdev.c b/drivers/net/octeontx/octeontx_ethdev.c index 3ee7b043fd..81779885d5 100644 --- a/drivers/net/octeontx/octeontx_ethdev.c +++ b/drivers/net/octeontx/octeontx_ethdev.c @@ -552,7 +552,7 @@ octeontx_dev_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu) if (rc) return rc; - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > OCCTX_L2_MAX_LEN) nic->rx_offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else nic->rx_offloads &= ~DEV_RX_OFFLOAD_JUMBO_FRAME; diff --git a/drivers/net/octeontx/octeontx_ethdev.h b/drivers/net/octeontx/octeontx_ethdev.h index 7246fb6d1d..780a094ffa 100644 --- a/drivers/net/octeontx/octeontx_ethdev.h +++ b/drivers/net/octeontx/octeontx_ethdev.h @@ -44,6 +44,7 @@ /* ETH_HLEN+ETH_FCS+2*VLAN_HLEN */ #define OCCTX_L2_OVERHEAD (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + \ OCCTX_MAX_VTAG_ACT_SIZE) +#define OCCTX_L2_MAX_LEN (RTE_ETHER_MTU + OCCTX_L2_OVERHEAD) /* Since HW FRS includes NPC VTAG insertion space, user has reduced FRS */ #define OCCTX_MAX_FRS \ From patchwork Thu Dec 17 09:23:02 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85291 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id CF5B1A09F0; Thu, 17 Dec 2020 10:33:34 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id AB880CA63; Thu, 17 Dec 2020 10:33:33 +0100 (CET) Received: from mga03.intel.com (mga03.intel.com [134.134.136.65]) by dpdk.org (Postfix) with ESMTP id C5A84CA63 for ; Thu, 17 Dec 2020 10:33:30 +0100 (CET) IronPort-SDR: 9EiPmQo+viK2U8d4P0zzce0BncJZL+n2LcPEnAQUqQKVcx5z5ePOi/CXbyFWqGFGApdyObq2/L +tbmaX9S7H3A== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="175324736" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="175324736" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by orsmga103.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:33:30 -0800 IronPort-SDR: VsWRohVejLR9U2A5mgPpOc7WnzzyBUDISJxbUOd0U8GBBUkCYxxLYDFXkLcXD7Dd17QG1cRYJA B92EIrvQU43Q== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369755978" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:33:01 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:02 +0000 Message-Id: <20201217092312.27033-13-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 12/22] net/octeontx2: fix the jumbo frame flag condition for mtu X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports dual VLAN tags. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: 0e2efd02db58 ("net/octeontx2: add MTU set operation") Signed-off-by: Steve Yang Acked-by: Nithin Dabilpuram Acked-by: Sunil Kumar Kori --- drivers/net/octeontx2/otx2_ethdev.h | 2 ++ drivers/net/octeontx2/otx2_ethdev_ops.c | 2 +- 2 files changed, 3 insertions(+), 1 deletion(-) diff --git a/drivers/net/octeontx2/otx2_ethdev.h b/drivers/net/octeontx2/otx2_ethdev.h index 3b9871f4dc..99f0469d89 100644 --- a/drivers/net/octeontx2/otx2_ethdev.h +++ b/drivers/net/octeontx2/otx2_ethdev.h @@ -51,6 +51,8 @@ /* ETH_HLEN+ETH_FCS+2*VLAN_HLEN */ #define NIX_L2_OVERHEAD \ (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + 8) +#define NIX_L2_MAX_LEN \ + (RTE_ETHER_MTU + NIX_L2_OVERHEAD) /* HW config of frame size doesn't include FCS */ #define NIX_MAX_HW_FRS 9212 diff --git a/drivers/net/octeontx2/otx2_ethdev_ops.c b/drivers/net/octeontx2/otx2_ethdev_ops.c index b36d37b9f7..963cc285ed 100644 --- a/drivers/net/octeontx2/otx2_ethdev_ops.c +++ b/drivers/net/octeontx2/otx2_ethdev_ops.c @@ -58,7 +58,7 @@ otx2_nix_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu) if (rc) return rc; - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > NIX_L2_MAX_LEN) dev->rx_offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else dev->rx_offloads &= ~DEV_RX_OFFLOAD_JUMBO_FRAME; From patchwork Thu Dec 17 09:23:03 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85292 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id AAD8AA09F0; Thu, 17 Dec 2020 10:34:23 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 8EB68CA3A; Thu, 17 Dec 2020 10:34:22 +0100 (CET) Received: from mga01.intel.com (mga01.intel.com [192.55.52.88]) by dpdk.org (Postfix) with ESMTP id 6D46EC9BA for ; Thu, 17 Dec 2020 10:34:19 +0100 (CET) IronPort-SDR: h4ch0T9MQRc/J/usroE3ZcznI0MmRfWkjlIm7NwPkvAS2yCnZSgcioFreyp3qtychNCFXPzIb1 7rCVj7QG+Dnw== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="193611569" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="193611569" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga101.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:33:56 -0800 IronPort-SDR: nXuCtUYQI0GBMR6ur48CYQquG/CF527EGohpOzPQGVJsCegsaGc4PfhUDDe+nua8/+mjZe0UUH MOXjVBzj7ajg== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369756873" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:33:30 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:03 +0000 Message-Id: <20201217092312.27033-14-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 13/22] net/qede: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports dual VLAN tags. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: 200645ac7909 ("net/qede: set MTU") Signed-off-by: Steve Yang --- drivers/net/qede/qede_ethdev.c | 2 +- drivers/net/qede/qede_rxtx.h | 1 + 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/drivers/net/qede/qede_ethdev.c b/drivers/net/qede/qede_ethdev.c index 549013557c..6919378b8e 100644 --- a/drivers/net/qede/qede_ethdev.c +++ b/drivers/net/qede/qede_ethdev.c @@ -2367,7 +2367,7 @@ static int qede_set_mtu(struct rte_eth_dev *dev, uint16_t mtu) fp->rxq->rx_buf_size = rc; } } - if (max_rx_pkt_len > RTE_ETHER_MAX_LEN) + if (frame_size > QEDE_ETH_MAX_LEN) dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else dev->data->dev_conf.rxmode.offloads &= ~DEV_RX_OFFLOAD_JUMBO_FRAME; diff --git a/drivers/net/qede/qede_rxtx.h b/drivers/net/qede/qede_rxtx.h index d7ff870b20..fcb564a1bb 100644 --- a/drivers/net/qede/qede_rxtx.h +++ b/drivers/net/qede/qede_rxtx.h @@ -71,6 +71,7 @@ + (QEDE_LLC_SNAP_HDR_LEN) + 2) #define QEDE_MAX_ETHER_HDR_LEN (RTE_ETHER_HDR_LEN + QEDE_ETH_OVERHEAD) +#define QEDE_ETH_MAX_LEN (RTE_ETHER_MTU + QEDE_MAX_ETHER_HDR_LEN) #define QEDE_RSS_OFFLOAD_ALL (ETH_RSS_IPV4 |\ ETH_RSS_NONFRAG_IPV4_TCP |\ From patchwork Thu Dec 17 09:23:04 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85293 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id EE54AA09F0; Thu, 17 Dec 2020 10:34:46 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 7B735CA5C; Thu, 17 Dec 2020 10:34:25 +0100 (CET) Received: from mga11.intel.com (mga11.intel.com [192.55.52.93]) by dpdk.org (Postfix) with ESMTP id 83A4BCA4C for ; Thu, 17 Dec 2020 10:34:23 +0100 (CET) IronPort-SDR: GBySPzi5tVEvmArGCjbex0LbhTnTt/7KywzGEsbpHBsotjvJ8VmKxRhP9RBmRPLbf4JzHdWYAJ WnXaGVjnA7Rw== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="171717125" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="171717125" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga102.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:34:21 -0800 IronPort-SDR: VjyEcikXrk1L2fHXnRnUrQTQQAVo21U+dFG5YtezAFAFfVOTKqkSc3lo+hD3rJ8igigj/uEqfp MFtfA69eX/Fg== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369757791" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:33:56 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:04 +0000 Message-Id: <20201217092312.27033-15-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 14/22] net/sfc: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports VLAN tag. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU'. Fixes: ff6a1197c3b1 ("net/sfc: convert to new Rx offload API") Signed-off-by: Steve Yang --- drivers/net/sfc/sfc_ethdev.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/drivers/net/sfc/sfc_ethdev.c b/drivers/net/sfc/sfc_ethdev.c index 93fc7baa0d..f2f5336435 100644 --- a/drivers/net/sfc/sfc_ethdev.c +++ b/drivers/net/sfc/sfc_ethdev.c @@ -1017,7 +1017,7 @@ sfc_dev_set_mtu(struct rte_eth_dev *dev, uint16_t mtu) * The driver does not use it, but other PMDs update jumbo frame * flag and max_rx_pkt_len when MTU is set. */ - if (mtu > RTE_ETHER_MAX_LEN) { + if (mtu > RTE_ETHER_MTU) { struct rte_eth_rxmode *rxmode = &dev->data->dev_conf.rxmode; rxmode->offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; } From patchwork Thu Dec 17 09:23:05 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85294 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id E5524A09F0; Thu, 17 Dec 2020 10:35:07 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id EAD03CA44; Thu, 17 Dec 2020 10:34:48 +0100 (CET) Received: from mga09.intel.com (mga09.intel.com [134.134.136.24]) by dpdk.org (Postfix) with ESMTP id E487ACA2E for ; Thu, 17 Dec 2020 10:34:46 +0100 (CET) IronPort-SDR: jLhsFXDvfdSRPWoHQBmAAye5t+ucivR5YuERqCijX5dYuSUnLjC8buFoQRlQOsAMd0ARmNYSIv 2YosMGZfzXFg== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="175365173" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="175365173" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by orsmga102.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:34:45 -0800 IronPort-SDR: F2sfVT/vehUOTv+h2Ojs45uIlj7zedESW9n76MtjO366CJVZoMAMmp1Fmlm0kDH928+4L851Ob AKIF3xZsX7Sw== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369758730" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:34:21 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:05 +0000 Message-Id: <20201217092312.27033-16-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 15/22] net/thunderx: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition, but the Ether overhead is larger than 18 when it supports dual VLAN tags. That will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU' and overhead. Fixes: 65d9804edc05 ("net/thunderx: support MTU configuration") Signed-off-by: Steve Yang --- drivers/net/thunderx/base/nicvf_hw_defs.h | 1 + drivers/net/thunderx/nicvf_ethdev.c | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/drivers/net/thunderx/base/nicvf_hw_defs.h b/drivers/net/thunderx/base/nicvf_hw_defs.h index b12c8ec50a..adc8ec943d 100644 --- a/drivers/net/thunderx/base/nicvf_hw_defs.h +++ b/drivers/net/thunderx/base/nicvf_hw_defs.h @@ -176,6 +176,7 @@ #define NIC_HW_MAX_MTU (9190) #define NIC_HW_MAX_FRS (NIC_HW_MAX_MTU + NIC_HW_L2_OVERHEAD) #define NIC_HW_MAX_SEGS (12) +#define NIC_HW_L2_MAX_LEN (RTE_ETHER_MTU + NIC_HW_L2_OVERHEAD) /* Descriptor alignments */ #define NICVF_RBDR_BASE_ALIGN_BYTES (128) /* 7 bits */ diff --git a/drivers/net/thunderx/nicvf_ethdev.c b/drivers/net/thunderx/nicvf_ethdev.c index b6bb05e500..c2e7c334d4 100644 --- a/drivers/net/thunderx/nicvf_ethdev.c +++ b/drivers/net/thunderx/nicvf_ethdev.c @@ -176,7 +176,7 @@ nicvf_dev_set_mtu(struct rte_eth_dev *dev, uint16_t mtu) (frame_size + 2 * VLAN_TAG_SIZE > buffsz * NIC_HW_MAX_SEGS)) return -EINVAL; - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > NIC_HW_L2_MAX_LEN) rxmode->offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else rxmode->offloads &= ~DEV_RX_OFFLOAD_JUMBO_FRAME; From patchwork Thu Dec 17 09:23:06 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85295 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 41931A09F0; Thu, 17 Dec 2020 10:35:28 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 85D41CA3A; Thu, 17 Dec 2020 10:35:14 +0100 (CET) Received: from mga12.intel.com (mga12.intel.com [192.55.52.136]) by dpdk.org (Postfix) with ESMTP id 63897CA3A for ; Thu, 17 Dec 2020 10:35:12 +0100 (CET) IronPort-SDR: ad1vQaxz+51IOoXsyQfQ8kvdRJ85gW4ulmwan7EgN84H03BpiKGe4VOv+IA/22DioVcpIHS/OH A3ibIguL67Bw== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="154447679" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="154447679" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga106.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:35:10 -0800 IronPort-SDR: aOquo6DVvrnGZ3HnpaFj5jsm4UDFC0UafhX3QlDxbPMjEWwv1CKb1WzhCtBV+ODh93CWyV+Bf0 5L0fnn+fESyg== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369759692" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:34:46 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:06 +0000 Message-Id: <20201217092312.27033-17-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 16/22] net/ixgbe: fix the jumbo frame flag condition X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition. If the Ether overhead is larger than 18 when it supports VLAN tag, that will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will normalize the boundary condition with 'RTE_ETHER_MTU' and overhead even though current overhead is 18. Fixes: 59d0ecdbf0e1 ("ethdev: MTU accessors") Fixes: 95a27b3ba5f5 ("net/ixgbe: enable jumbo frame for VF") Signed-off-by: Steve Yang Acked-by: Jeff Guo --- drivers/net/ixgbe/ixgbe_ethdev.c | 2 +- drivers/net/ixgbe/ixgbe_ethdev.h | 3 +++ drivers/net/ixgbe/ixgbe_pf.c | 2 +- 3 files changed, 5 insertions(+), 2 deletions(-) diff --git a/drivers/net/ixgbe/ixgbe_ethdev.c b/drivers/net/ixgbe/ixgbe_ethdev.c index d7a1806ab8..fa0f5afd03 100644 --- a/drivers/net/ixgbe/ixgbe_ethdev.c +++ b/drivers/net/ixgbe/ixgbe_ethdev.c @@ -5173,7 +5173,7 @@ ixgbe_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) hlreg0 = IXGBE_READ_REG(hw, IXGBE_HLREG0); /* switch to jumbo mode if needed */ - if (frame_size > RTE_ETHER_MAX_LEN) { + if (frame_size > IXGBE_ETH_MAX_LEN) { dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; hlreg0 |= IXGBE_HLREG0_JUMBOEN; diff --git a/drivers/net/ixgbe/ixgbe_ethdev.h b/drivers/net/ixgbe/ixgbe_ethdev.h index 3d35ea791b..a0ce18ca24 100644 --- a/drivers/net/ixgbe/ixgbe_ethdev.h +++ b/drivers/net/ixgbe/ixgbe_ethdev.h @@ -104,6 +104,9 @@ /* The overhead from MTU to max frame size. */ #define IXGBE_ETH_OVERHEAD (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN) +/* The max frame size with default MTU */ +#define IXGBE_ETH_MAX_LEN (RTE_ETHER_MTU + IXGBE_ETH_OVERHEAD) + /* bit of VXLAN tunnel type | 7 bits of zeros | 8 bits of zeros*/ #define IXGBE_FDIR_VXLAN_TUNNEL_TYPE 0x8000 /* bit of NVGRE tunnel type | 7 bits of zeros | 8 bits of zeros*/ diff --git a/drivers/net/ixgbe/ixgbe_pf.c b/drivers/net/ixgbe/ixgbe_pf.c index 833863af5a..89698e8470 100644 --- a/drivers/net/ixgbe/ixgbe_pf.c +++ b/drivers/net/ixgbe/ixgbe_pf.c @@ -575,7 +575,7 @@ ixgbe_set_vf_lpe(struct rte_eth_dev *dev, __rte_unused uint32_t vf, uint32_t *ms IXGBE_MHADD_MFS_MASK) >> IXGBE_MHADD_MFS_SHIFT; if (max_frs < new_mtu) { hlreg0 = IXGBE_READ_REG(hw, IXGBE_HLREG0); - if (new_mtu > RTE_ETHER_MAX_LEN) { + if (new_mtu > IXGBE_ETH_MAX_LEN) { dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; hlreg0 |= IXGBE_HLREG0_JUMBOEN; From patchwork Thu Dec 17 09:23:07 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85296 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id C9AA1A09F0; Thu, 17 Dec 2020 10:35:51 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 41FC8CA57; Thu, 17 Dec 2020 10:35:49 +0100 (CET) Received: from mga09.intel.com (mga09.intel.com [134.134.136.24]) by dpdk.org (Postfix) with ESMTP id 2D47FCA5E for ; Thu, 17 Dec 2020 10:35:46 +0100 (CET) IronPort-SDR: 1vWtVvsIms1roHLQZq8THzjYqldcizA5sSQoGpYWDGF4gk2OwRNl6QEVRtySJfo/hQLqSFIz2a H1Kawbbl8Tig== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="175365356" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="175365356" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by orsmga102.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:35:44 -0800 IronPort-SDR: lSkvj+8vze4wP79MCut7rPkosH8ygFnUjih1NRokTti0/8+TbgDLD/idH6K7+jdAMXUjZiGMot b0w9HgW0JzRQ== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369760327" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:35:10 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:07 +0000 Message-Id: <20201217092312.27033-18-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 17/22] net/cxgbe: fix the jumbo frame flag condition X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition. If the Ether overhead is larger than 18 when it supports VLAN tag, that will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will normalize the boundary condition with 'RTE_ETHER_MTU' and overhead even though current overhead is 18. Fixes: 4b2eff452d2e ("cxgbe: enable jumbo frames") Fixes: 0ec33be4c857 ("cxgbe: allow to change mtu") Signed-off-by: Steve Yang --- drivers/net/cxgbe/cxgbe.h | 4 ++++ drivers/net/cxgbe/cxgbe_ethdev.c | 4 ++-- 2 files changed, 6 insertions(+), 2 deletions(-) diff --git a/drivers/net/cxgbe/cxgbe.h b/drivers/net/cxgbe/cxgbe.h index ef62af1c3f..7c89a028bf 100644 --- a/drivers/net/cxgbe/cxgbe.h +++ b/drivers/net/cxgbe/cxgbe.h @@ -19,6 +19,10 @@ #define CXGBE_MAX_RX_PKTLEN (9000 + RTE_ETHER_HDR_LEN + \ RTE_ETHER_CRC_LEN) /* max pkt */ +/* The max frame size with default MTU */ +#define CXGBE_ETH_MAX_LEN (RTE_ETHER_MTU + \ + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN) + /* Max poll time is 100 * 100msec = 10 sec */ #define CXGBE_LINK_STATUS_POLL_MS 100 /* 100ms */ #define CXGBE_LINK_STATUS_POLL_CNT 100 /* Max number of times to poll */ diff --git a/drivers/net/cxgbe/cxgbe_ethdev.c b/drivers/net/cxgbe/cxgbe_ethdev.c index 98d0362fa3..480d6f58a8 100644 --- a/drivers/net/cxgbe/cxgbe_ethdev.c +++ b/drivers/net/cxgbe/cxgbe_ethdev.c @@ -300,7 +300,7 @@ int cxgbe_dev_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu) return -EINVAL; /* set to jumbo mode if needed */ - if (new_mtu > RTE_ETHER_MAX_LEN) + if (new_mtu > CXGBE_ETH_MAX_LEN) eth_dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else @@ -669,7 +669,7 @@ int cxgbe_dev_rx_queue_setup(struct rte_eth_dev *eth_dev, rxq->fl.size = temp_nb_desc; /* Set to jumbo mode if necessary */ - if (pkt_len > RTE_ETHER_MAX_LEN) + if (pkt_len > CXGBE_ETH_MAX_LEN) eth_dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else From patchwork Thu Dec 17 09:23:08 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85297 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id D4222A09F0; Thu, 17 Dec 2020 10:36:15 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id CB719CA1A; Thu, 17 Dec 2020 10:36:09 +0100 (CET) Received: from mga11.intel.com (mga11.intel.com [192.55.52.93]) by dpdk.org (Postfix) with ESMTP id 42BB3C9BA for ; Thu, 17 Dec 2020 10:36:07 +0100 (CET) IronPort-SDR: +JOAZo/zWy3AkjJnmCzuJIaKzo/IfsCrZeXR43mHgd3e5kJ9PpEiLLjqPCsoQZPl9rMiyjT3Tf UQFAsvWTu0mg== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="171717520" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="171717520" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga102.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:36:05 -0800 IronPort-SDR: YmxQOdpfN6RUvM9es4bpEkIvT9zALy43VgWKs6PbXFsUyozGy30JLHCm3U29WB26weLsl14YeA Kl9JOtgBqzwA== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369760799" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:35:40 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:08 +0000 Message-Id: <20201217092312.27033-19-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 18/22] net/axgbe: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition. If the Ether overhead is larger than 18 when it supports VLAN tag, that will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will normalize the boundary condition with 'RTE_ETHER_MTU' and overhead even though current overhead is 18. Fixes: b58d8781fa1f ("net/axgbe: support setting MTU") Signed-off-by: Steve Yang --- drivers/net/axgbe/axgbe_ethdev.c | 2 +- drivers/net/axgbe/axgbe_ethdev.h | 6 ++++++ 2 files changed, 7 insertions(+), 1 deletion(-) diff --git a/drivers/net/axgbe/axgbe_ethdev.c b/drivers/net/axgbe/axgbe_ethdev.c index f524a3be96..599ae19940 100644 --- a/drivers/net/axgbe/axgbe_ethdev.c +++ b/drivers/net/axgbe/axgbe_ethdev.c @@ -1489,7 +1489,7 @@ static int axgb_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) dev->data->port_id); return -EBUSY; } - if (frame_size > RTE_ETHER_MAX_LEN) { + if (frame_size > AXGBE_ETH_MAX_LEN) { dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; val = 1; diff --git a/drivers/net/axgbe/axgbe_ethdev.h b/drivers/net/axgbe/axgbe_ethdev.h index 1481fd9ff3..a6226729fe 100644 --- a/drivers/net/axgbe/axgbe_ethdev.h +++ b/drivers/net/axgbe/axgbe_ethdev.h @@ -125,6 +125,12 @@ /* MDIO port types */ #define AXGMAC_MAX_C22_PORT 3 +/* The max frame size with default MTU */ +#define AXGBE_ETH_MAX_LEN ( \ + RTE_ETHER_MTU + \ + RTE_ETHER_HDR_LEN + \ + RTE_ETHER_CRC_LEN) + /* Helper macro for descriptor handling * Always use AXGBE_GET_DESC_DATA to access the descriptor data * since the index is free-running and needs to be and-ed From patchwork Thu Dec 17 09:23:09 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85298 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 90FB5A09F0; Thu, 17 Dec 2020 10:36:54 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 622D5CA2E; Thu, 17 Dec 2020 10:36:53 +0100 (CET) Received: from mga11.intel.com (mga11.intel.com [192.55.52.93]) by dpdk.org (Postfix) with ESMTP id 64EB6DE0 for ; Thu, 17 Dec 2020 10:36:51 +0100 (CET) IronPort-SDR: qUM/JWc0OL3+I6Yfni15iTavMS/u/uqy3thHzlc2bZuw8XLWiGxJbxro8rKJ8cWtxn4T1ve/ea 3sUZAMfyZbHA== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="171717606" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="171717606" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga102.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:36:29 -0800 IronPort-SDR: GrsCkdel8t0UX9jvUt7hJ398ewbPTXKhU4GXuGWpmY8jH8Mhm91zJG92BllUv8+DC2BOBO9d9l t1EO09O5Tbvw== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369761234" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:36:05 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:09 +0000 Message-Id: <20201217092312.27033-20-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 19/22] net/enetc: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition. If the Ether overhead is larger than 18 when it supports VLAN tag, that will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will normalize the boundary condition with 'RTE_ETHER_MTU' and overhead even though current overhead is 18. Fixes: 5d5589b0c858 ("net/enetc: support MTU update and jumbo frames") Signed-off-by: Steve Yang --- drivers/net/enetc/enetc.h | 4 ++++ drivers/net/enetc/enetc_ethdev.c | 2 +- 2 files changed, 5 insertions(+), 1 deletion(-) diff --git a/drivers/net/enetc/enetc.h b/drivers/net/enetc/enetc.h index 14ef3bc18b..7163633bce 100644 --- a/drivers/net/enetc/enetc.h +++ b/drivers/net/enetc/enetc.h @@ -29,6 +29,10 @@ /* maximum frame size supported */ #define ENETC_MAC_MAXFRM_SIZE 9600 +/* The max frame size with default MTU */ +#define ENETC_ETH_MAX_LEN (RTE_ETHER_MTU + \ + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN) + /* * upper_32_bits - return bits 32-63 of a number * @n: the number we're accessing diff --git a/drivers/net/enetc/enetc_ethdev.c b/drivers/net/enetc/enetc_ethdev.c index 6ff3022874..4d2c9c0474 100644 --- a/drivers/net/enetc/enetc_ethdev.c +++ b/drivers/net/enetc/enetc_ethdev.c @@ -677,7 +677,7 @@ enetc_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) return -EINVAL; } - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > ENETC_ETH_MAX_LEN) dev->data->dev_conf.rxmode.offloads &= DEV_RX_OFFLOAD_JUMBO_FRAME; else From patchwork Thu Dec 17 09:23:10 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85299 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id E749AA09F0; Thu, 17 Dec 2020 10:37:14 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 4A92ECA1C; Thu, 17 Dec 2020 10:36:59 +0100 (CET) Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by dpdk.org (Postfix) with ESMTP id F1A9ECA1A for ; Thu, 17 Dec 2020 10:36:56 +0100 (CET) IronPort-SDR: L3kE/6P2ConpZrZN/wooeQpIOvk2zE0i8F+2H9yQSyQGQBCe6elyV7BQX1kx4tiCs+9lNUKTjV FdF48kPR3yFQ== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="239310038" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="239310038" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:36:53 -0800 IronPort-SDR: /QydEZZp3mWe9PcJSWJQMTO5/ar+9hXoGOGAE3oiPBBvIUdwYbRV4lmyhXqVsUWM60P7GglbQv kbYuBkLaQojg== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369761633" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:36:29 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:10 +0000 Message-Id: <20201217092312.27033-21-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 20/22] net/hinic: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition. If the Ether overhead is larger than 18 when it supports VLAN tag, that will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will normalize the boundary condition with 'RTE_ETHER_MTU' and overhead even though current overhead is 18. Fixes: 254bd849b132 ("net/hinic: set jumbo frame offload flag") Signed-off-by: Steve Yang --- drivers/net/hinic/hinic_pmd_ethdev.c | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/drivers/net/hinic/hinic_pmd_ethdev.c b/drivers/net/hinic/hinic_pmd_ethdev.c index 62642354cf..5a2c171099 100644 --- a/drivers/net/hinic/hinic_pmd_ethdev.c +++ b/drivers/net/hinic/hinic_pmd_ethdev.c @@ -75,6 +75,9 @@ #define HINIC_PKTLEN_TO_MTU(pktlen) \ ((pktlen) - (ETH_HLEN + ETH_CRC_LEN)) +/* The max frame size with default MTU */ +#define HINIC_ETH_MAX_LEN (RTE_ETHER_MTU + ETH_HLEN + ETH_CRC_LEN) + /* lro numer limit for one packet */ #define HINIC_LRO_WQE_NUM_DEFAULT 8 @@ -1556,7 +1559,7 @@ static int hinic_dev_set_mtu(struct rte_eth_dev *dev, uint16_t mtu) /* update max frame size */ frame_size = HINIC_MTU_TO_PKTLEN(mtu); - if (frame_size > RTE_ETHER_MAX_LEN) + if (frame_size > HINIC_ETH_MAX_LEN) dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else From patchwork Thu Dec 17 09:23:11 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85300 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 93E31A09F0; Thu, 17 Dec 2020 10:37:33 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id DCE6BCA46; Thu, 17 Dec 2020 10:37:22 +0100 (CET) Received: from mga14.intel.com (mga14.intel.com [192.55.52.115]) by dpdk.org (Postfix) with ESMTP id 4C7AECA3A for ; Thu, 17 Dec 2020 10:37:21 +0100 (CET) IronPort-SDR: RNQAKYEOy8O/u+WoEIs9c0CszOK7gYcWDFPHS7Oy2z8065/APtBD+M+ptlOfWwLwL3I/MtrwU2 J8FD5BBnYH0g== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="174450983" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="174450983" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga103.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:37:19 -0800 IronPort-SDR: O/x5cVK7FAvoOw5qYXiXg27NIjFbDeEUH8x4bo8v8h3w2VZuRa7apuaWMi7WYxSyktIrLzvBR8 /lOcIFfAT7Ug== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369761909" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:36:53 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:11 +0000 Message-Id: <20201217092312.27033-22-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 21/22] net/nfp: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition. If the Ether overhead is larger than 18 when it supports VLAN tag, that will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will change the boundary condition with 'RTE_ETHER_MTU'. Fixes: d4a27a3b092a ("nfp: add basic features") Signed-off-by: Steve Yang --- drivers/net/nfp/nfp_net.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/drivers/net/nfp/nfp_net.c b/drivers/net/nfp/nfp_net.c index 1608bf5ea1..9ea24e5bda 100644 --- a/drivers/net/nfp/nfp_net.c +++ b/drivers/net/nfp/nfp_net.c @@ -1508,7 +1508,7 @@ nfp_net_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) } /* switch to jumbo mode if needed */ - if ((uint32_t)mtu > RTE_ETHER_MAX_LEN) + if ((uint32_t)mtu > RTE_ETHER_MTU) dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else dev->data->dev_conf.rxmode.offloads &= ~DEV_RX_OFFLOAD_JUMBO_FRAME; From patchwork Thu Dec 17 09:23:12 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Steve Yang X-Patchwork-Id: 85301 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 286E4A09F0; Thu, 17 Dec 2020 10:37:54 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id C99B7CA1A; Thu, 17 Dec 2020 10:37:48 +0100 (CET) Received: from mga04.intel.com (mga04.intel.com [192.55.52.120]) by dpdk.org (Postfix) with ESMTP id AF520DE0 for ; Thu, 17 Dec 2020 10:37:46 +0100 (CET) IronPort-SDR: 6OMlBaP7GTrTl2d14ashtX8jKczb6HPuYtTYgQlNmCIS9Jplv8RnP2CHYK4OHRJg6fz2kserE2 o0XBsxXC1hhQ== X-IronPort-AV: E=McAfee;i="6000,8403,9837"; a="172652058" X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="172652058" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:37:44 -0800 IronPort-SDR: pdJwtlRsJdL9vbqcs9INU/hxT03k1BIkIlZfcCcDd7llc7Uf81YbHeKhOP4rbESek1D3AoQ0l4 zUCU25cKv8OA== X-IronPort-AV: E=Sophos;i="5.78,426,1599548400"; d="scan'208";a="369761969" Received: from intel-npg-odc-srv01.cd.intel.com ([10.240.178.136]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Dec 2020 01:37:19 -0800 From: Steve Yang To: dev@dpdk.org Cc: wenzhuo.lu@intel.com, beilei.xing@intel.com, bernard.iremonger@intel.com, asomalap@amd.com, rahul.lakkireddy@chelsio.com, hemant.agrawal@nxp.com, sachin.saxena@oss.nxp.com, jia.guo@intel.com, haiyue.wang@intel.com, g.singh@nxp.com, xuanziyang2@huawei.com, cloud.wangxiaoyun@huawei.com, zhouguoyang@huawei.com, xavier.huwei@huawei.com, humin29@huawei.com, yisen.zhuang@huawei.com, oulijun@huawei.com, jingjing.wu@intel.com, qiming.yang@intel.com, qi.z.zhang@intel.com, rosen.xu@intel.com, sthotton@marvell.com, srinivasan@marvell.com, heinrich.kuhn@netronome.com, hkalra@marvell.com, jerinj@marvell.com, ndabilpuram@marvell.com, kirankumark@marvell.com, rmody@marvell.com, shshaikh@marvell.com, andrew.rybchenko@oktetlabs.ru, mczekaj@marvell.com, thomas@monjalon.net, ferruh.yigit@intel.com, ivan.boule@6wind.com, konstantin.ananyev@intel.com, samuel.gauthier@6wind.com, david.marchand@6wind.com, shahafs@mellanox.com, stephen@networkplumber.org, maxime.coquelin@redhat.com, olivier.matz@6wind.com, lihuisong@huawei.com, shreyansh.jain@nxp.com, wei.dai@intel.com, fengchunsong@huawei.com, chenhao164@huawei.com, tangchengchang@hisilicon.com, helin.zhang@intel.com, yanglong.wu@intel.com, xiaolong.ye@intel.com, ting.xu@intel.com, xiaoyun.li@intel.com, dan.wei@intel.com, andy.pei@intel.com, vattunuru@marvell.com, skori@marvell.com, sony.chacko@qlogic.com, bruce.richardson@intel.com, ivan.malov@oktetlabs.ru, rad@semihalf.com, slawomir.rosek@semihalf.com, kamil.rytarowski@caviumnetworks.com, wei.zhao1@intel.com, junyux.jiang@intel.com, kumaras@chelsio.com, girish.nandibasappa@amd.com, rolf.neugebauer@netronome.com, alejandro.lucero@netronome.com, Steve Yang Date: Thu, 17 Dec 2020 09:23:12 +0000 Message-Id: <20201217092312.27033-23-stevex.yang@intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20201217092312.27033-1-stevex.yang@intel.com> References: <20201209031628.29572-1-stevex.yang@intel.com> <20201217092312.27033-1-stevex.yang@intel.com> Subject: [dpdk-dev] [PATCH v2 22/22] net/liquidio: fix the jumbo frame flag condition for mtu set X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" The jumbo frame uses the 'RTE_ETHER_MAX_LEN' as boundary condition. If the Ether overhead is larger than 18 when it supports VLAN tag, that will cause the jumbo flag rx offload is wrong when MTU size is 'RTE_ETHER_MTU'. This fix will normalize the boundary condition with 'RTE_ETHER_MTU' and overhead even though current overhead is 18. Fixes: 9f1c00266d82 ("net/liquidio: add API to set MTU") Signed-off-by: Steve Yang --- drivers/net/liquidio/lio_ethdev.c | 2 +- drivers/net/liquidio/lio_ethdev.h | 3 +++ 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/drivers/net/liquidio/lio_ethdev.c b/drivers/net/liquidio/lio_ethdev.c index d4dd3768cd..eb0fdab45a 100644 --- a/drivers/net/liquidio/lio_ethdev.c +++ b/drivers/net/liquidio/lio_ethdev.c @@ -481,7 +481,7 @@ lio_dev_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu) return -1; } - if (frame_len > RTE_ETHER_MAX_LEN) + if (frame_len > LIO_ETH_MAX_LEN) eth_dev->data->dev_conf.rxmode.offloads |= DEV_RX_OFFLOAD_JUMBO_FRAME; else diff --git a/drivers/net/liquidio/lio_ethdev.h b/drivers/net/liquidio/lio_ethdev.h index 74cd2fb6c6..d33be1c44d 100644 --- a/drivers/net/liquidio/lio_ethdev.h +++ b/drivers/net/liquidio/lio_ethdev.h @@ -13,6 +13,9 @@ #define LIO_LSC_TIMEOUT 100000 /* 100000us (100ms) */ #define LIO_MAX_CMD_TIMEOUT 10000 /* 10000ms (10s) */ +/* The max frame size with default MTU */ +#define LIO_ETH_MAX_LEN (RTE_ETHER_MTU + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN) + #define LIO_DEV(_eth_dev) ((_eth_dev)->data->dev_private) /* LIO Response condition variable */