From patchwork Wed Aug 1 08:37:38 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Matan Azrad X-Patchwork-Id: 43499 X-Patchwork-Delegate: shahafs@mellanox.com Return-Path: X-Original-To: patchwork@dpdk.org Delivered-To: patchwork@dpdk.org Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 76E9A8DAF; Wed, 1 Aug 2018 10:37:43 +0200 (CEST) Received: from mellanox.co.il (mail-il-dmz.mellanox.com [193.47.165.129]) by dpdk.org (Postfix) with ESMTP id D7A098DA6 for ; Wed, 1 Aug 2018 10:37:41 +0200 (CEST) Received: from Internal Mail-Server by MTLPINE1 (envelope-from matan@mellanox.com) with ESMTPS (AES256-SHA encrypted); 1 Aug 2018 11:41:04 +0300 Received: from pegasus08.mtr.labs.mlnx (pegasus08.mtr.labs.mlnx [10.210.16.114]) by labmailer.mlnx (8.13.8/8.13.8) with ESMTP id w718beVq001251; Wed, 1 Aug 2018 11:37:41 +0300 From: Matan Azrad To: Yongseok Koh , Shahaf Shuler Cc: dev@dpdk.org Date: Wed, 1 Aug 2018 08:37:38 +0000 Message-Id: <1533112658-24390-1-git-send-email-matan@mellanox.com> X-Mailer: git-send-email 2.7.4 Subject: [dpdk-dev] [PATCH] net/mlx5: fix tunnel offloads X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" There are a dedicated QP attributes, tunnel offload flag and mask, which must be configured in order to allow part of the HW tunnel offloads. So, if a QP is pointed by a tunnel flow, the above QP attributes should be configured. The mask configuration is wrongly only performed if an internal RSS was configured by the user, while there is no reason to condition the tunnel offloads in RSS configurations. Consequently, some of the tunnel offloads was not performed by the HW when a tunnel flow was configured, for example, the packet tunnel types was not reported to the user. Replace the internal RSS condition with the tunnel flow condition. Fixes: df6afd377ace ("net/mlx5: remove useless arguments in hrxq API") Signed-off-by: Matan Azrad --- drivers/net/mlx5/mlx5_flow.c | 4 +++- drivers/net/mlx5/mlx5_rxq.c | 8 ++++---- drivers/net/mlx5/mlx5_rxtx.h | 3 ++- 3 files changed, 9 insertions(+), 6 deletions(-) diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c index adda4df..923fc28 100644 --- a/drivers/net/mlx5/mlx5_flow.c +++ b/drivers/net/mlx5/mlx5_flow.c @@ -2879,7 +2879,9 @@ mlx5_flow_apply(struct rte_eth_dev *dev, struct rte_flow *flow, MLX5_RSS_HASH_KEY_LEN, verbs->hash_fields, (*flow->queue), - flow->rss.queue_num); + flow->rss.queue_num, + !!(flow->layers & + MLX5_FLOW_LAYER_TUNNEL)); if (!hrxq) { rte_flow_error_set (error, rte_errno, diff --git a/drivers/net/mlx5/mlx5_rxq.c b/drivers/net/mlx5/mlx5_rxq.c index 16e1641..97b3e8e 100644 --- a/drivers/net/mlx5/mlx5_rxq.c +++ b/drivers/net/mlx5/mlx5_rxq.c @@ -1752,7 +1752,8 @@ struct mlx5_hrxq * mlx5_hrxq_new(struct rte_eth_dev *dev, const uint8_t *rss_key, uint32_t rss_key_len, uint64_t hash_fields, - const uint16_t *queues, uint32_t queues_n) + const uint16_t *queues, uint32_t queues_n, + int tunnel __rte_unused) { struct priv *priv = dev->data->dev_private; struct mlx5_hrxq *hrxq; @@ -1794,9 +1795,8 @@ mlx5_hrxq_new(struct rte_eth_dev *dev, .pd = priv->pd, }, &(struct mlx5dv_qp_init_attr){ - .comp_mask = (hash_fields & IBV_RX_HASH_INNER) ? - MLX5DV_QP_INIT_ATTR_MASK_QP_CREATE_FLAGS : - 0, + .comp_mask = tunnel ? + MLX5DV_QP_INIT_ATTR_MASK_QP_CREATE_FLAGS : 0, .create_flags = MLX5DV_QP_CREATE_TUNNEL_OFFLOADS, }); #else diff --git a/drivers/net/mlx5/mlx5_rxtx.h b/drivers/net/mlx5/mlx5_rxtx.h index 992e977..48ed2b2 100644 --- a/drivers/net/mlx5/mlx5_rxtx.h +++ b/drivers/net/mlx5/mlx5_rxtx.h @@ -283,7 +283,8 @@ void mlx5_ind_table_ibv_drop_release(struct rte_eth_dev *dev); struct mlx5_hrxq *mlx5_hrxq_new(struct rte_eth_dev *dev, const uint8_t *rss_key, uint32_t rss_key_len, uint64_t hash_fields, - const uint16_t *queues, uint32_t queues_n); + const uint16_t *queues, uint32_t queues_n, + int tunnel __rte_unused); struct mlx5_hrxq *mlx5_hrxq_get(struct rte_eth_dev *dev, const uint8_t *rss_key, uint32_t rss_key_len, uint64_t hash_fields,