From patchwork Thu Mar 7 10:12:09 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Gregory Etelson X-Patchwork-Id: 138091 X-Patchwork-Delegate: rasland@nvidia.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 2EE1943BA7; Thu, 7 Mar 2024 11:12:52 +0100 (CET) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 0EB3642E9D; Thu, 7 Mar 2024 11:12:52 +0100 (CET) Received: from NAM12-MW2-obe.outbound.protection.outlook.com (mail-mw2nam12on2064.outbound.protection.outlook.com [40.107.244.64]) by mails.dpdk.org (Postfix) with ESMTP id 31D7042E74; Thu, 7 Mar 2024 11:12:50 +0100 (CET) ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=oUPqbJGZefSzLS8uZBXwX8pCPPoORaMCJRd1hawFVkKXHQtzHtizghOLD8Aivv92GITPfLbII3MoiD/Zspf/uxozNlPmLLeBMgXqLKT1/bD1itfKGK/o4WaDjWkz9nC/t6kdOZdQWz9itr+Vf6e9EvGHn0U0hgxQRVGGxX5EA9JuoyOdB6tTk7BS/bIfYbvbTptu495kKS+XRjallMU1jmsNr+rCmYadOH7tJgYbJzEjYzKbHj8LGEuOGKNdGx+TsWcIe/YJuCGWqUizMXRIl5s/KhUAaEpxn0TxcdG2ObTRwonPArBTFrSjMhTAMkWWli6BnLccwdeEA50GTyjK7g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=FQ6COwgA7EDugiDitjOB2+j6se8K2sbrNMBNWEv1Kd0=; b=BZaMWI7IO7o+X73MpRLdGQAO8FTOHbzjGNcz5fu1X7F/kxB3v+dWmGjlkOPHs1n0+GLHWcDTWN9FdkgUhXL5D27/6OGISuBKZyovPTIQbDjf9ykn8TqY//OAkDgKZaHLbJVlud5m3AonwKJOX6q9FuMzJahhdGLlHsa9ZCVMsm4tepM1PIL6Lowzt4kTtsAA5/EunMn9FfM7/3E3YsyovnBgwWU5CuRKbv9b5+t2xavskArm8CjniH5E9uLGLKkv1HzTekZTW6sPSSXeYgO/4Vbq2BghWpeH7U7uAJQxIFh6I1ViYuvg8Rdzxif6+cKi8v9Vm5XnMs2O6QZkLzH2cA== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 216.228.117.160) smtp.rcpttodomain=dpdk.org smtp.mailfrom=nvidia.com; dmarc=pass (p=reject sp=reject pct=100) action=none header.from=nvidia.com; dkim=none (message not signed); arc=none (0) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=Nvidia.com; s=selector2; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=FQ6COwgA7EDugiDitjOB2+j6se8K2sbrNMBNWEv1Kd0=; b=kR/nGiWnOwBBH7b0SevY05yICDUyyqJ9jpId+l826CH+AFAWna3LuNJNRLV2lCaVi0L0YD1lj9ww8Y+L9QxS64W/nBwhJHxrLI8Y4ml3TTpjOlNyCIlDZoOMlU3wCpbQQw/q/UeIZ+mo6F5pSWbbkVtwj2dhR0gSGRVyMJoiKpCVQ+7rr+Du8g5EQfHm3Ne5FrDy2uvVGqW8JmYviFINQT6X2bWFyc28JbBxk7nrzS4+YuldZ7qSU+zSQdL4P8/MIBEbC/GG5hY1ZJt8PQyDzCbgF1OJbpfAE5mVTzXLG6GrxcIxp9EusfuhbomWz6h2jGgNK6UAdR76h58Hf8bMFg== Received: from SN6PR01CA0020.prod.exchangelabs.com (2603:10b6:805:b6::33) by MW6PR12MB8898.namprd12.prod.outlook.com (2603:10b6:303:246::8) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.7339.39; Thu, 7 Mar 2024 10:12:46 +0000 Received: from SA2PEPF00001509.namprd04.prod.outlook.com (2603:10b6:805:b6:cafe::4) by SN6PR01CA0020.outlook.office365.com (2603:10b6:805:b6::33) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.7362.26 via Frontend Transport; Thu, 7 Mar 2024 10:12:46 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 216.228.117.160) smtp.mailfrom=nvidia.com; dkim=none (message not signed) header.d=none;dmarc=pass action=none header.from=nvidia.com; Received-SPF: Pass (protection.outlook.com: domain of nvidia.com designates 216.228.117.160 as permitted sender) receiver=protection.outlook.com; client-ip=216.228.117.160; helo=mail.nvidia.com; pr=C Received: from mail.nvidia.com (216.228.117.160) by SA2PEPF00001509.mail.protection.outlook.com (10.167.242.41) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.7362.11 via Frontend Transport; Thu, 7 Mar 2024 10:12:45 +0000 Received: from rnnvmail201.nvidia.com (10.129.68.8) by mail.nvidia.com (10.129.200.66) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.41; Thu, 7 Mar 2024 02:12:34 -0800 Received: from nvidia.com (10.126.230.35) by rnnvmail201.nvidia.com (10.129.68.8) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.1258.12; Thu, 7 Mar 2024 02:12:31 -0800 From: Gregory Etelson To: CC: , , , , Dariusz Sosnowski , "Viacheslav Ovsiienko" , Ori Kam , Suanming Mou , Matan Azrad , Alexander Kozyrev Subject: [PATCH 1/3] net/mlx5: fix HWS meter actions availability Date: Thu, 7 Mar 2024 12:12:09 +0200 Message-ID: <20240307101211.1134473-2-getelson@nvidia.com> X-Mailer: git-send-email 2.39.2 In-Reply-To: <20240307101211.1134473-1-getelson@nvidia.com> References: <20240307101211.1134473-1-getelson@nvidia.com> MIME-Version: 1.0 X-Originating-IP: [10.126.230.35] X-ClientProxiedBy: rnnvmail203.nvidia.com (10.129.68.9) To rnnvmail201.nvidia.com (10.129.68.8) X-EOPAttributedMessage: 0 X-MS-PublicTrafficType: Email X-MS-TrafficTypeDiagnostic: SA2PEPF00001509:EE_|MW6PR12MB8898:EE_ X-MS-Office365-Filtering-Correlation-Id: e150df56-a143-4a26-4f2b-08dc3e8f2295 X-MS-Exchange-SenderADCheck: 1 X-MS-Exchange-AntiSpam-Relay: 0 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: nDNyd/JbyPFEZh21dnyuRj/nIgFVY/3q95dZDytmO/WuRZhVKcpD9o17U6N9saAe2JBEd66upXBvzA7xV6OP7QLaIkrW1lZNdp0cGMHZLkPhAuxN+vSoyEFHDswmk9iXL8Z07AxRFEqiOiKj6tW0I6Dpe/KxOqj3SLOQRuA0H4fShU1RuvOMx1+U3H91gJrma9nbC44rNfeeDlL9O3gV3uwT38P3Nkp/IXc5xhHyNa57qicYghDVr1GNDvJ/wvTW1SFkHMYHBkB9ujEYSifT9ZdpVIdeZ/p8q+4DMHm9AWV6E//hCPHv989QjOwxPy+BYR5n1JGpGLJ6rQW+QFmwxa/qPLzqutw8r21fK0PXevH+aY7okKIf7mQXhvfAqREym3OyHCusBzIz8M/U8e1qdVN4537U7yc+rxPtDrxhfw2ZjTQQACBIpJhok7uoQKplIg1G2jbQZrfwLJX462JiqvBOeZKD6MW4C689zuh4Iq1P/m+tjFtlu623LJSbVp9qtUzxkE2KRGygZJ/lk70EFfu4jPJXvgxF3iOcDnAG/Q23c7E07KFcsph0baVZ/ArjNDw92Fl8rCNZBOOVzmfEEuHM9uL5LUTbsYyL7R2dGia8DYwDsc8FbHmO0WDfnFRbEa2LhYwb56IFI36hdoQPVhRmitGFShyO+wGZmVMV2HENP4Qtdgr4TM+8oIqPoY+mKGlKooclY7WZ0seXhPpUD+BKiHzZFImwNUq0bk2s2Qa6mHq+4DdTc1cOaFJV322G X-Forefront-Antispam-Report: CIP:216.228.117.160; CTRY:US; LANG:en; SCL:1; SRV:; IPV:NLI; SFV:NSPM; H:mail.nvidia.com; PTR:dc6edge1.nvidia.com; CAT:NONE; SFS:(13230031)(376005)(36860700004)(82310400014); DIR:OUT; SFP:1101; X-OriginatorOrg: Nvidia.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 07 Mar 2024 10:12:45.9973 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: e150df56-a143-4a26-4f2b-08dc3e8f2295 X-MS-Exchange-CrossTenant-Id: 43083d15-7273-40c1-b7db-39efd9ccc17a X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=43083d15-7273-40c1-b7db-39efd9ccc17a; Ip=[216.228.117.160]; Helo=[mail.nvidia.com] X-MS-Exchange-CrossTenant-AuthSource: SA2PEPF00001509.namprd04.prod.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: MW6PR12MB8898 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Allow compilation of HWS meter code only on platforms that support HWS. Fixes: 24865366e495 ("net/mlx5: support flow meter action for HWS") Cc: stable@dpdk.org Signed-off-by: Gregory Etelson Acked-by: Dariusz Sosnowski --- drivers/net/mlx5/mlx5_flow_meter.c | 28 ++++++++++++++++++++++++---- 1 file changed, 24 insertions(+), 4 deletions(-) diff --git a/drivers/net/mlx5/mlx5_flow_meter.c b/drivers/net/mlx5/mlx5_flow_meter.c index c0578ce6e9..57de95b4b9 100644 --- a/drivers/net/mlx5/mlx5_flow_meter.c +++ b/drivers/net/mlx5/mlx5_flow_meter.c @@ -896,6 +896,7 @@ mlx5_flow_meter_profile_get(struct rte_eth_dev *dev, meter_profile_id); } +#if defined(HAVE_MLX5_HWS_SUPPORT) /** * Callback to add MTR profile with HWS. * @@ -981,6 +982,7 @@ mlx5_flow_meter_profile_hws_delete(struct rte_eth_dev *dev, memset(fmp, 0, sizeof(struct mlx5_flow_meter_profile)); return 0; } +#endif /** * Find policy by id. @@ -1123,6 +1125,7 @@ mlx5_flow_meter_policy_validate(struct rte_eth_dev *dev, return 0; } +#if defined(HAVE_MLX5_HWS_SUPPORT) /** * Callback to check MTR policy action validate for HWS * @@ -1159,6 +1162,7 @@ mlx5_flow_meter_policy_hws_validate(struct rte_eth_dev *dev, } return 0; } +#endif static int __mlx5_flow_meter_policy_delete(struct rte_eth_dev *dev, @@ -1485,6 +1489,7 @@ mlx5_flow_meter_policy_get(struct rte_eth_dev *dev, &policy_idx); } +#if defined(HAVE_MLX5_HWS_SUPPORT) /** * Callback to delete MTR policy for HWS. * @@ -1807,7 +1812,7 @@ mlx5_flow_meter_policy_hws_add(struct rte_eth_dev *dev, RTE_MTR_ERROR_TYPE_UNSPECIFIED, NULL, "Failed to create meter policy."); } - +#endif /** * Check meter validation. * @@ -2177,6 +2182,7 @@ mlx5_flow_meter_create(struct rte_eth_dev *dev, uint32_t meter_id, NULL, "Failed to create devx meter."); } +#if defined(HAVE_MLX5_HWS_SUPPORT) /** * Create meter rules. * @@ -2260,6 +2266,7 @@ mlx5_flow_meter_hws_create(struct rte_eth_dev *dev, uint32_t meter_id, __atomic_fetch_add(&policy->ref_cnt, 1, __ATOMIC_RELAXED); return 0; } +#endif static int mlx5_flow_meter_params_flush(struct rte_eth_dev *dev, @@ -2744,6 +2751,7 @@ static const struct rte_mtr_ops mlx5_flow_mtr_ops = { .stats_read = mlx5_flow_meter_stats_read, }; +#if defined(HAVE_MLX5_HWS_SUPPORT) static const struct rte_mtr_ops mlx5_flow_mtr_hws_ops = { .capabilities_get = mlx5_flow_mtr_cap_get, .meter_profile_add = mlx5_flow_meter_profile_hws_add, @@ -2762,6 +2770,7 @@ static const struct rte_mtr_ops mlx5_flow_mtr_hws_ops = { .stats_update = NULL, .stats_read = NULL, }; +#endif /** * Get meter operations. @@ -2777,12 +2786,16 @@ static const struct rte_mtr_ops mlx5_flow_mtr_hws_ops = { int mlx5_flow_meter_ops_get(struct rte_eth_dev *dev __rte_unused, void *arg) { +#if defined(HAVE_MLX5_HWS_SUPPORT) struct mlx5_priv *priv = dev->data->dev_private; if (priv->sh->config.dv_flow_en == 2) *(const struct rte_mtr_ops **)arg = &mlx5_flow_mtr_hws_ops; else *(const struct rte_mtr_ops **)arg = &mlx5_flow_mtr_ops; +#else + *(const struct rte_mtr_ops **)arg = &mlx5_flow_mtr_ops; +#endif return 0; } @@ -3161,7 +3174,6 @@ mlx5_flow_meter_flush(struct rte_eth_dev *dev, struct rte_mtr_error *error) struct mlx5_flow_meter_profile *fmp; struct mlx5_legacy_flow_meter *legacy_fm; struct mlx5_flow_meter_info *fm; - struct mlx5_flow_meter_policy *policy; struct mlx5_flow_meter_sub_policy *sub_policy; void *tmp; uint32_t i, mtr_idx, policy_idx; @@ -3229,15 +3241,20 @@ mlx5_flow_meter_flush(struct rte_eth_dev *dev, struct rte_mtr_error *error) mlx5_l3t_destroy(priv->policy_idx_tbl); priv->policy_idx_tbl = NULL; } +#if defined(HAVE_MLX5_HWS_SUPPORT) if (priv->mtr_policy_arr) { + struct mlx5_flow_meter_policy *policy; + for (i = 0; i < priv->mtr_config.nb_meter_policies; i++) { policy = mlx5_flow_meter_policy_find(dev, i, &policy_idx); - if (policy->initialized) + if (policy->initialized) { mlx5_flow_meter_policy_hws_delete(dev, i, error); + } } } +#endif if (priv->mtr_profile_tbl) { MLX5_L3T_FOREACH(priv->mtr_profile_tbl, i, entry) { fmp = entry; @@ -3251,14 +3268,17 @@ mlx5_flow_meter_flush(struct rte_eth_dev *dev, struct rte_mtr_error *error) mlx5_l3t_destroy(priv->mtr_profile_tbl); priv->mtr_profile_tbl = NULL; } +#if defined(HAVE_MLX5_HWS_SUPPORT) if (priv->mtr_profile_arr) { for (i = 0; i < priv->mtr_config.nb_meter_profiles; i++) { fmp = mlx5_flow_meter_profile_find(priv, i); - if (fmp->initialized) + if (fmp->initialized) { mlx5_flow_meter_profile_hws_delete(dev, i, error); + } } } +#endif /* Delete default policy table. */ mlx5_flow_destroy_def_policy(dev); if (priv->sh->refcnt == 1) From patchwork Thu Mar 7 10:12:10 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Gregory Etelson X-Patchwork-Id: 138092 X-Patchwork-Delegate: rasland@nvidia.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 88D0F43BA7; Thu, 7 Mar 2024 11:13:02 +0100 (CET) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 715EB42EB5; Thu, 7 Mar 2024 11:13:02 +0100 (CET) Received: from NAM12-DM6-obe.outbound.protection.outlook.com (mail-dm6nam12on2064.outbound.protection.outlook.com [40.107.243.64]) by mails.dpdk.org (Postfix) with ESMTP id 1972942E94; Thu, 7 Mar 2024 11:13:00 +0100 (CET) ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=YZa5VHOjNvrM9TwgUs1NMlbDPL62cjlOpJaeby9jUSvWJJzS/ykvw99IbsJYJvnUibDruIQGx7/i4ermLu21M+hasLdduYx6okeRi/ZwXY2d5YuQbWvrpCITLsEFmPXT7gnGvaVCdjAuTxtbMLF+n5yMfpzfloN5Qr8wcDeI45SeOZcuhHIjQ828WruxClNnBgw+mGz2D9tpRfbi3GxV2/okirRbh+FSfSpSc85ZoJEjDmtYgi8zCLOoZ5b5LaBPxMOhrag1ooO+lk+pVVhvpBwqz08zSedXYO9q+07NSClxpU9CatTvIu+OP0wa3nrmoGYk8uyFQ7UpVpbQhxJXHg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=BY96S1X7Np4e3CpN8LWzqOR7ESgR17U59xuo3sVYRQg=; b=YfhpFQyEn6M1YX1sU+ToNNDfmh3/UozYCGfDStJIxbbj9fwH03u/SlMISoaHkRTkr2XNdc+MkCqaR4KjVm/qz5sjE0zXP92IMtCjVvdKQytHhC/3meVYqzyV3WENkoJ1oreTreu1vdaQcB/OK11+6ckCSlb0jO2YTOEId5v2KzFaBQnrEi5RiSbQpFYM3PPzpOZ6Y+NqMGM7A+H+kC6L6Rv1dDDHZ/0HiL5/wkzob87h2757YTSvWVPQ+HdR9BtMD61sXH2k6zpQnhN5lzGRIalKPhLlFm5xcY92BnCn5P3uZTpVgQuJ7C3yNtlnH/IRlrWIfjq/pi4tdRLDMibL1w== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 216.228.117.160) smtp.rcpttodomain=dpdk.org smtp.mailfrom=nvidia.com; dmarc=pass (p=reject sp=reject pct=100) action=none header.from=nvidia.com; dkim=none (message not signed); arc=none (0) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=Nvidia.com; s=selector2; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=BY96S1X7Np4e3CpN8LWzqOR7ESgR17U59xuo3sVYRQg=; b=A3WcS/3Z/6LzZICikVeueT3lefUiflBELJ7LiUVkHThDS1n9J+qJVs8LRZlnXJOwH3Hf2m06mk1wUyy9dEoVvIXNUtaVnMEV1Dhk7DyoFizPDoM5zJMO+6aMkVS8pcXnfKHq9vcN8GlQJ1sWlY3R14LxiKI/1lcy8RXEs2J03BZBx3eSiOGxpjRkOnIX4hvTG29KNKG0jDKlEiapxRji1aqLCV/LEFjzMi8wNZKWSMuCau9CpHgkVGItR2NBVgzYSRP0iDzOBMo2wbgSBLixcXwCdKKzetR7FBbsUZMzqMmqU+5t2GnqytwWDIRSHIaQ4Xhtpkq9zu+9oywUF3ZgFA== Received: from SA9PR13CA0151.namprd13.prod.outlook.com (2603:10b6:806:28::6) by CH0PR12MB8551.namprd12.prod.outlook.com (2603:10b6:610:186::21) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.7362.26; Thu, 7 Mar 2024 10:12:56 +0000 Received: from SA2PEPF0000150A.namprd04.prod.outlook.com (2603:10b6:806:28::4) by SA9PR13CA0151.outlook.office365.com (2603:10b6:806:28::6) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.7386.6 via Frontend Transport; Thu, 7 Mar 2024 10:12:56 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 216.228.117.160) smtp.mailfrom=nvidia.com; dkim=none (message not signed) header.d=none;dmarc=pass action=none header.from=nvidia.com; Received-SPF: Pass (protection.outlook.com: domain of nvidia.com designates 216.228.117.160 as permitted sender) receiver=protection.outlook.com; client-ip=216.228.117.160; helo=mail.nvidia.com; pr=C Received: from mail.nvidia.com (216.228.117.160) by SA2PEPF0000150A.mail.protection.outlook.com (10.167.242.42) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.7362.11 via Frontend Transport; Thu, 7 Mar 2024 10:12:56 +0000 Received: from rnnvmail201.nvidia.com (10.129.68.8) by mail.nvidia.com (10.129.200.66) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.41; Thu, 7 Mar 2024 02:12:38 -0800 Received: from nvidia.com (10.126.230.35) by rnnvmail201.nvidia.com (10.129.68.8) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.1258.12; Thu, 7 Mar 2024 02:12:34 -0800 From: Gregory Etelson To: CC: , , , , Dariusz Sosnowski , "Viacheslav Ovsiienko" , Ori Kam , Suanming Mou , Matan Azrad , Alexander Kozyrev Subject: [PATCH 2/3] net/mlx5: fix sync meter processing in HWS setup Date: Thu, 7 Mar 2024 12:12:10 +0200 Message-ID: <20240307101211.1134473-3-getelson@nvidia.com> X-Mailer: git-send-email 2.39.2 In-Reply-To: <20240307101211.1134473-1-getelson@nvidia.com> References: <20240307101211.1134473-1-getelson@nvidia.com> MIME-Version: 1.0 X-Originating-IP: [10.126.230.35] X-ClientProxiedBy: rnnvmail203.nvidia.com (10.129.68.9) To rnnvmail201.nvidia.com (10.129.68.8) X-EOPAttributedMessage: 0 X-MS-PublicTrafficType: Email X-MS-TrafficTypeDiagnostic: SA2PEPF0000150A:EE_|CH0PR12MB8551:EE_ X-MS-Office365-Filtering-Correlation-Id: 7deb5e50-a216-476c-389a-08dc3e8f28d8 X-MS-Exchange-SenderADCheck: 1 X-MS-Exchange-AntiSpam-Relay: 0 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: zAa5ZaUSilWu8v3cd0XAWRqHbWVea20q0ZRNIUfOFn0DotM+jDgYoeChLIUgwjBJT6ZcCS7WDgR3j+3aWwHJzQZsA0O6/16mBoG39veCz1k2M0JhQkRgvmRMW/Oj0Hgnh58868fDdQPefvlZBDZyXuF4dYljr4atketjWnj2MKevEBq3rKcXK1qXMdBcmdb5qtCMDycB71i0d+8ELHRosP5Imv3EBbDuVCRMp4eNCGfgPVPuzDAKUULXNg+Ib/Z5XNfZikSIMhB8p4slN6pHGL67w4K9AC2ATMazzVoeKEMyeWEm3tQUL4jiTaVlxP3PrkWIDnob3bonROUqpBQZ3sINjw5H3dsn6C0O0kJoqWPs8tylOWvuQuHL2BHRZEX7nQJdRKEm0iynb4i2DXdjZJ0EgRVPTnyp4OjjqonOA4ESGJEHawWvGQHy5YSqV289VPYYbEV5woVqFQXV1cvEqndFr9cEVbs5nNdje+SC6/8gsg+0IcqXQSdjkdeB20JnuyYrYSW1hDvD7x+xpgK+tYuYxwWs3rh+C5GEzCp+ju86Daj09uMCpUTU1I/0djFFDRVK45U9MabJy1Q1qLw9YSC939C72LiwmRTOOFV/CeEcQDGwdvNjAw78misCeUDJCvkda5t1vKwqW817U6kBWJ1REAwlkVvSHEyV+8xZN0hRxkTzj7XGbCJdLGl55fT7BCoOSxj3P63cnxWvWr2hDIXDkx/iFKOqJz9lTLrO78pGln63bcZ98cJPdfXtiY5m X-Forefront-Antispam-Report: CIP:216.228.117.160; CTRY:US; LANG:en; SCL:1; SRV:; IPV:NLI; SFV:NSPM; H:mail.nvidia.com; PTR:dc6edge1.nvidia.com; CAT:NONE; SFS:(13230031)(376005)(36860700004)(82310400014); DIR:OUT; SFP:1101; X-OriginatorOrg: Nvidia.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 07 Mar 2024 10:12:56.4709 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: 7deb5e50-a216-476c-389a-08dc3e8f28d8 X-MS-Exchange-CrossTenant-Id: 43083d15-7273-40c1-b7db-39efd9ccc17a X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=43083d15-7273-40c1-b7db-39efd9ccc17a; Ip=[216.228.117.160]; Helo=[mail.nvidia.com] X-MS-Exchange-CrossTenant-AuthSource: SA2PEPF0000150A.namprd04.prod.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: CH0PR12MB8551 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Synchronous calls for meter ASO try to pull pending completions from CQ, submit WR and return to caller. That avoids delays between WR post and HW response. If the template API was activated, PMD will use control queue for sync operations. PMD has different formats for the `user_data` context in sync and async meter ASO calls. PMD port destruction procedure submits async operations to the port control queue and polls the queue CQs to clean HW responses. Port destruction can pull a meter ASO completion from control CQ. Such completion has sync format, but was processed by async handler. The patch implements sync meter ASO interface with async calls in the template API environment. Fixes: 48fbb0e93d06 ("net/mlx5: support flow meter mark indirect action with HWS") Cc: stable@dpdk.org Signed-off-by: Gregory Etelson Acked-by: Dariusz Sosnowski --- drivers/net/mlx5/mlx5.h | 35 +++++- drivers/net/mlx5/mlx5_flow_aso.c | 178 ++++++++++++++++++----------- drivers/net/mlx5/mlx5_flow_hw.c | 99 ++++++++-------- drivers/net/mlx5/mlx5_flow_meter.c | 27 +++-- 4 files changed, 216 insertions(+), 123 deletions(-) diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h index 2fb3bb65cc..6ff8f322e0 100644 --- a/drivers/net/mlx5/mlx5.h +++ b/drivers/net/mlx5/mlx5.h @@ -2033,6 +2033,30 @@ enum dr_dump_rec_type { DR_DUMP_REC_TYPE_PMD_COUNTER = 4430, }; +#if defined(HAVE_MLX5_HWS_SUPPORT) +static __rte_always_inline struct mlx5_hw_q_job * +flow_hw_job_get(struct mlx5_priv *priv, uint32_t queue) +{ + MLX5_ASSERT(priv->hw_q[queue].job_idx <= priv->hw_q[queue].size); + return priv->hw_q[queue].job_idx ? + priv->hw_q[queue].job[--priv->hw_q[queue].job_idx] : NULL; +} + +static __rte_always_inline void +flow_hw_job_put(struct mlx5_priv *priv, struct mlx5_hw_q_job *job, uint32_t queue) +{ + MLX5_ASSERT(priv->hw_q[queue].job_idx < priv->hw_q[queue].size); + priv->hw_q[queue].job[priv->hw_q[queue].job_idx++] = job; +} + +struct mlx5_hw_q_job * +mlx5_flow_action_job_init(struct mlx5_priv *priv, uint32_t queue, + const struct rte_flow_action_handle *handle, + void *user_data, void *query_data, + enum mlx5_hw_job_type type, + struct rte_flow_error *error); +#endif + /** * Indicates whether HW objects operations can be created by DevX. * @@ -2443,11 +2467,12 @@ int mlx5_aso_flow_hit_queue_poll_start(struct mlx5_dev_ctx_shared *sh); int mlx5_aso_flow_hit_queue_poll_stop(struct mlx5_dev_ctx_shared *sh); void mlx5_aso_queue_uninit(struct mlx5_dev_ctx_shared *sh, enum mlx5_access_aso_opc_mod aso_opc_mod); -int mlx5_aso_meter_update_by_wqe(struct mlx5_dev_ctx_shared *sh, uint32_t queue, - struct mlx5_aso_mtr *mtr, struct mlx5_mtr_bulk *bulk, - void *user_data, bool push); -int mlx5_aso_mtr_wait(struct mlx5_dev_ctx_shared *sh, uint32_t queue, - struct mlx5_aso_mtr *mtr); +int mlx5_aso_meter_update_by_wqe(struct mlx5_priv *priv, uint32_t queue, + struct mlx5_aso_mtr *mtr, + struct mlx5_mtr_bulk *bulk, + struct mlx5_hw_q_job *job, bool push); +int mlx5_aso_mtr_wait(struct mlx5_priv *priv, + struct mlx5_aso_mtr *mtr, bool is_tmpl_api); int mlx5_aso_ct_update_by_wqe(struct mlx5_dev_ctx_shared *sh, uint32_t queue, struct mlx5_aso_ct_action *ct, const struct rte_flow_action_conntrack *profile, diff --git a/drivers/net/mlx5/mlx5_flow_aso.c b/drivers/net/mlx5/mlx5_flow_aso.c index f311443472..ab9eb21e01 100644 --- a/drivers/net/mlx5/mlx5_flow_aso.c +++ b/drivers/net/mlx5/mlx5_flow_aso.c @@ -792,7 +792,7 @@ mlx5_aso_mtr_sq_enqueue_single(struct mlx5_dev_ctx_shared *sh, struct mlx5_aso_mtr *aso_mtr, struct mlx5_mtr_bulk *bulk, bool need_lock, - void *user_data, + struct mlx5_hw_q_job *job, bool push) { volatile struct mlx5_aso_wqe *wqe = NULL; @@ -819,7 +819,7 @@ mlx5_aso_mtr_sq_enqueue_single(struct mlx5_dev_ctx_shared *sh, rte_prefetch0(&sq->sq_obj.aso_wqes[(sq->head + 1) & mask]); /* Fill next WQE. */ fm = &aso_mtr->fm; - sq->elts[sq->head & mask].mtr = user_data ? user_data : aso_mtr; + sq->elts[sq->head & mask].user_data = job ? job : (void *)aso_mtr; if (aso_mtr->type == ASO_METER_INDIRECT) { if (likely(sh->config.dv_flow_en == 2)) pool = aso_mtr->pool; @@ -897,24 +897,6 @@ mlx5_aso_mtr_sq_enqueue_single(struct mlx5_dev_ctx_shared *sh, return 1; } -static void -mlx5_aso_mtrs_status_update(struct mlx5_aso_sq *sq, uint16_t aso_mtrs_nums) -{ - uint16_t size = 1 << sq->log_desc_n; - uint16_t mask = size - 1; - uint16_t i; - struct mlx5_aso_mtr *aso_mtr = NULL; - uint8_t exp_state = ASO_METER_WAIT; - - for (i = 0; i < aso_mtrs_nums; ++i) { - aso_mtr = sq->elts[(sq->tail + i) & mask].mtr; - MLX5_ASSERT(aso_mtr); - (void)__atomic_compare_exchange_n(&aso_mtr->state, - &exp_state, ASO_METER_READY, - false, __ATOMIC_RELAXED, __ATOMIC_RELAXED); - } -} - static void mlx5_aso_mtr_completion_handle(struct mlx5_aso_sq *sq, bool need_lock) { @@ -925,7 +907,7 @@ mlx5_aso_mtr_completion_handle(struct mlx5_aso_sq *sq, bool need_lock) uint32_t idx; uint32_t next_idx = cq->cq_ci & mask; uint16_t max; - uint16_t n = 0; + uint16_t i, n = 0; int ret; if (need_lock) @@ -957,7 +939,19 @@ mlx5_aso_mtr_completion_handle(struct mlx5_aso_sq *sq, bool need_lock) cq->cq_ci++; } while (1); if (likely(n)) { - mlx5_aso_mtrs_status_update(sq, n); + uint8_t exp_state = ASO_METER_WAIT; + struct mlx5_aso_mtr *aso_mtr; + __rte_unused bool verdict; + + for (i = 0; i < n; ++i) { + aso_mtr = sq->elts[(sq->tail + i) & mask].mtr; + MLX5_ASSERT(aso_mtr); + verdict = __atomic_compare_exchange_n(&aso_mtr->state, + &exp_state, ASO_METER_READY, + false, __ATOMIC_RELAXED, + __ATOMIC_RELAXED); + MLX5_ASSERT(verdict); + } sq->tail += n; rte_io_wmb(); cq->cq_obj.db_rec[0] = rte_cpu_to_be_32(cq->cq_ci); @@ -966,6 +960,82 @@ mlx5_aso_mtr_completion_handle(struct mlx5_aso_sq *sq, bool need_lock) rte_spinlock_unlock(&sq->sqsl); } +static __rte_always_inline struct mlx5_aso_sq * +mlx5_aso_mtr_select_sq(struct mlx5_dev_ctx_shared *sh, uint32_t queue, + struct mlx5_aso_mtr *mtr, bool *need_lock) +{ + struct mlx5_aso_sq *sq; + + if (likely(sh->config.dv_flow_en == 2) && + mtr->type == ASO_METER_INDIRECT) { + if (queue == MLX5_HW_INV_QUEUE) { + sq = &mtr->pool->sq[mtr->pool->nb_sq - 1]; + *need_lock = true; + } else { + sq = &mtr->pool->sq[queue]; + *need_lock = false; + } + } else { + sq = &sh->mtrmng->pools_mng.sq; + *need_lock = true; + } + return sq; +} + +#if defined(HAVE_MLX5_HWS_SUPPORT) +static void +mlx5_aso_poll_cq_mtr_hws(struct mlx5_priv *priv, struct mlx5_aso_sq *sq) +{ +#define MLX5_HWS_MTR_CMPL_NUM 4 + + int i, ret; + struct mlx5_aso_mtr *mtr; + uint8_t exp_state = ASO_METER_WAIT; + struct rte_flow_op_result res[MLX5_HWS_MTR_CMPL_NUM]; + __rte_unused bool verdict; + + rte_spinlock_lock(&sq->sqsl); +repeat: + ret = mlx5_aso_pull_completion(sq, res, MLX5_HWS_MTR_CMPL_NUM); + if (ret) { + for (i = 0; i < ret; i++) { + struct mlx5_hw_q_job *job = res[i].user_data; + + MLX5_ASSERT(job); + mtr = mlx5_ipool_get(priv->hws_mpool->idx_pool, + MLX5_INDIRECT_ACTION_IDX_GET(job->action)); + MLX5_ASSERT(mtr); + verdict = __atomic_compare_exchange_n(&mtr->state, + &exp_state, ASO_METER_READY, + false, __ATOMIC_RELAXED, + __ATOMIC_RELAXED); + MLX5_ASSERT(verdict); + flow_hw_job_put(priv, job, CTRL_QUEUE_ID(priv)); + } + if (ret == MLX5_HWS_MTR_CMPL_NUM) + goto repeat; + } + rte_spinlock_unlock(&sq->sqsl); + +#undef MLX5_HWS_MTR_CMPL_NUM +} +#else +static void +mlx5_aso_poll_cq_mtr_hws(__rte_unused struct mlx5_priv *priv, __rte_unused struct mlx5_aso_sq *sq) +{ + MLX5_ASSERT(false); +} +#endif + +static void +mlx5_aso_poll_cq_mtr_sws(__rte_unused struct mlx5_priv *priv, + struct mlx5_aso_sq *sq) +{ + mlx5_aso_mtr_completion_handle(sq, true); +} + +typedef void (*poll_cq_t)(struct mlx5_priv *, struct mlx5_aso_sq *); + /** * Update meter parameter by send WQE. * @@ -980,39 +1050,29 @@ mlx5_aso_mtr_completion_handle(struct mlx5_aso_sq *sq, bool need_lock) * 0 on success, a negative errno value otherwise and rte_errno is set. */ int -mlx5_aso_meter_update_by_wqe(struct mlx5_dev_ctx_shared *sh, uint32_t queue, - struct mlx5_aso_mtr *mtr, - struct mlx5_mtr_bulk *bulk, - void *user_data, - bool push) +mlx5_aso_meter_update_by_wqe(struct mlx5_priv *priv, uint32_t queue, + struct mlx5_aso_mtr *mtr, + struct mlx5_mtr_bulk *bulk, + struct mlx5_hw_q_job *job, bool push) { - struct mlx5_aso_sq *sq; - uint32_t poll_wqe_times = MLX5_MTR_POLL_WQE_CQE_TIMES; bool need_lock; + struct mlx5_dev_ctx_shared *sh = priv->sh; + struct mlx5_aso_sq *sq = + mlx5_aso_mtr_select_sq(sh, queue, mtr, &need_lock); + uint32_t poll_wqe_times = MLX5_MTR_POLL_WQE_CQE_TIMES; + poll_cq_t poll_mtr_cq = + job ? mlx5_aso_poll_cq_mtr_hws : mlx5_aso_poll_cq_mtr_sws; int ret; - if (likely(sh->config.dv_flow_en == 2) && - mtr->type == ASO_METER_INDIRECT) { - if (queue == MLX5_HW_INV_QUEUE) { - sq = &mtr->pool->sq[mtr->pool->nb_sq - 1]; - need_lock = true; - } else { - sq = &mtr->pool->sq[queue]; - need_lock = false; - } - } else { - sq = &sh->mtrmng->pools_mng.sq; - need_lock = true; - } if (queue != MLX5_HW_INV_QUEUE) { ret = mlx5_aso_mtr_sq_enqueue_single(sh, sq, mtr, bulk, - need_lock, user_data, push); + need_lock, job, push); return ret > 0 ? 0 : -1; } do { - mlx5_aso_mtr_completion_handle(sq, need_lock); + poll_mtr_cq(priv, sq); if (mlx5_aso_mtr_sq_enqueue_single(sh, sq, mtr, bulk, - need_lock, NULL, true)) + need_lock, job, true)) return 0; /* Waiting for wqe resource. */ rte_delay_us_sleep(MLX5_ASO_WQE_CQE_RESPONSE_DELAY); @@ -1036,32 +1096,22 @@ mlx5_aso_meter_update_by_wqe(struct mlx5_dev_ctx_shared *sh, uint32_t queue, * 0 on success, a negative errno value otherwise and rte_errno is set. */ int -mlx5_aso_mtr_wait(struct mlx5_dev_ctx_shared *sh, uint32_t queue, - struct mlx5_aso_mtr *mtr) +mlx5_aso_mtr_wait(struct mlx5_priv *priv, + struct mlx5_aso_mtr *mtr, bool is_tmpl_api) { + bool need_lock; struct mlx5_aso_sq *sq; + struct mlx5_dev_ctx_shared *sh = priv->sh; uint32_t poll_cqe_times = MLX5_MTR_POLL_WQE_CQE_TIMES; - uint8_t state; - bool need_lock; + uint8_t state = __atomic_load_n(&mtr->state, __ATOMIC_RELAXED); + poll_cq_t poll_mtr_cq = + is_tmpl_api ? mlx5_aso_poll_cq_mtr_hws : mlx5_aso_poll_cq_mtr_sws; - if (likely(sh->config.dv_flow_en == 2) && - mtr->type == ASO_METER_INDIRECT) { - if (queue == MLX5_HW_INV_QUEUE) { - sq = &mtr->pool->sq[mtr->pool->nb_sq - 1]; - need_lock = true; - } else { - sq = &mtr->pool->sq[queue]; - need_lock = false; - } - } else { - sq = &sh->mtrmng->pools_mng.sq; - need_lock = true; - } - state = __atomic_load_n(&mtr->state, __ATOMIC_RELAXED); if (state == ASO_METER_READY || state == ASO_METER_WAIT_ASYNC) return 0; + sq = mlx5_aso_mtr_select_sq(sh, MLX5_HW_INV_QUEUE, mtr, &need_lock); do { - mlx5_aso_mtr_completion_handle(sq, need_lock); + poll_mtr_cq(priv, sq); if (__atomic_load_n(&mtr->state, __ATOMIC_RELAXED) == ASO_METER_READY) return 0; diff --git a/drivers/net/mlx5/mlx5_flow_hw.c b/drivers/net/mlx5/mlx5_flow_hw.c index c1b09c9c03..8f004b5435 100644 --- a/drivers/net/mlx5/mlx5_flow_hw.c +++ b/drivers/net/mlx5/mlx5_flow_hw.c @@ -183,6 +183,12 @@ mlx5_flow_hw_aux_get_mtr_id(struct rte_flow_hw *flow, struct rte_flow_hw_aux *au return aux->orig.mtr_id; } +static __rte_always_inline struct mlx5_hw_q_job * +flow_hw_action_job_init(struct mlx5_priv *priv, uint32_t queue, + const struct rte_flow_action_handle *handle, + void *user_data, void *query_data, + enum mlx5_hw_job_type type, + struct rte_flow_error *error); static int mlx5_tbl_multi_pattern_process(struct rte_eth_dev *dev, struct rte_flow_template_table *tbl, @@ -384,21 +390,6 @@ flow_hw_q_dec_flow_ops(struct mlx5_priv *priv, uint32_t queue) q->ongoing_flow_ops--; } -static __rte_always_inline struct mlx5_hw_q_job * -flow_hw_job_get(struct mlx5_priv *priv, uint32_t queue) -{ - MLX5_ASSERT(priv->hw_q[queue].job_idx <= priv->hw_q[queue].size); - return priv->hw_q[queue].job_idx ? - priv->hw_q[queue].job[--priv->hw_q[queue].job_idx] : NULL; -} - -static __rte_always_inline void -flow_hw_job_put(struct mlx5_priv *priv, struct mlx5_hw_q_job *job, uint32_t queue) -{ - MLX5_ASSERT(priv->hw_q[queue].job_idx < priv->hw_q[queue].size); - priv->hw_q[queue].job[priv->hw_q[queue].job_idx++] = job; -} - static inline enum mlx5dr_matcher_insert_mode flow_hw_matcher_insert_mode_get(enum rte_flow_table_insertion_type insert_type) { @@ -1560,7 +1551,7 @@ flow_hw_meter_compile(struct rte_eth_dev *dev, acts->rule_acts[jump_pos].action = (!!group) ? acts->jump->hws_action : acts->jump->root_action; - if (mlx5_aso_mtr_wait(priv->sh, MLX5_HW_INV_QUEUE, aso_mtr)) + if (mlx5_aso_mtr_wait(priv, aso_mtr, true)) return -ENOMEM; return 0; } @@ -1637,7 +1628,7 @@ static rte_be32_t vlan_hdr_to_be32(const struct rte_flow_action *actions) static __rte_always_inline struct mlx5_aso_mtr * flow_hw_meter_mark_alloc(struct rte_eth_dev *dev, uint32_t queue, const struct rte_flow_action *action, - void *user_data, bool push, + struct mlx5_hw_q_job *job, bool push, struct rte_flow_error *error) { struct mlx5_priv *priv = dev->data->dev_private; @@ -1646,6 +1637,8 @@ flow_hw_meter_mark_alloc(struct rte_eth_dev *dev, uint32_t queue, struct mlx5_aso_mtr *aso_mtr; struct mlx5_flow_meter_info *fm; uint32_t mtr_id; + uintptr_t handle = (uintptr_t)MLX5_INDIRECT_ACTION_TYPE_METER_MARK << + MLX5_INDIRECT_ACTION_TYPE_OFFSET; if (priv->shared_host) { rte_flow_error_set(error, ENOTSUP, RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL, @@ -1669,15 +1662,16 @@ flow_hw_meter_mark_alloc(struct rte_eth_dev *dev, uint32_t queue, ASO_METER_WAIT : ASO_METER_WAIT_ASYNC; aso_mtr->offset = mtr_id - 1; aso_mtr->init_color = fm->color_aware ? RTE_COLORS : RTE_COLOR_GREEN; + job->action = (void *)(handle | mtr_id); /* Update ASO flow meter by wqe. */ - if (mlx5_aso_meter_update_by_wqe(priv->sh, queue, aso_mtr, - &priv->mtr_bulk, user_data, push)) { + if (mlx5_aso_meter_update_by_wqe(priv, queue, aso_mtr, + &priv->mtr_bulk, job, push)) { mlx5_ipool_free(pool->idx_pool, mtr_id); return NULL; } /* Wait for ASO object completion. */ if (queue == MLX5_HW_INV_QUEUE && - mlx5_aso_mtr_wait(priv->sh, MLX5_HW_INV_QUEUE, aso_mtr)) { + mlx5_aso_mtr_wait(priv, aso_mtr, true)) { mlx5_ipool_free(pool->idx_pool, mtr_id); return NULL; } @@ -1696,10 +1690,18 @@ flow_hw_meter_mark_compile(struct rte_eth_dev *dev, struct mlx5_priv *priv = dev->data->dev_private; struct mlx5_aso_mtr_pool *pool = priv->hws_mpool; struct mlx5_aso_mtr *aso_mtr; + struct mlx5_hw_q_job *job = + flow_hw_action_job_init(priv, queue, NULL, NULL, NULL, + MLX5_HW_Q_JOB_TYPE_CREATE, NULL); - aso_mtr = flow_hw_meter_mark_alloc(dev, queue, action, NULL, true, error); - if (!aso_mtr) + if (!job) + return -1; + aso_mtr = flow_hw_meter_mark_alloc(dev, queue, action, job, + true, error); + if (!aso_mtr) { + flow_hw_job_put(priv, job, queue); return -1; + } /* Compile METER_MARK action */ acts[aso_mtr_pos].action = pool->action; @@ -3275,7 +3277,7 @@ flow_hw_actions_construct(struct rte_eth_dev *dev, jump->root_action; flow->jump = jump; flow->flags |= MLX5_FLOW_HW_FLOW_FLAG_FATE_JUMP; - if (mlx5_aso_mtr_wait(priv->sh, MLX5_HW_INV_QUEUE, aso_mtr)) + if (mlx5_aso_mtr_wait(priv, aso_mtr, true)) return -1; break; case RTE_FLOW_ACTION_TYPE_AGE: @@ -4009,13 +4011,6 @@ flow_hw_pull_legacy_indirect_comp(struct rte_eth_dev *dev, struct mlx5_hw_q_job job->query.hw); aso_ct->state = ASO_CONNTRACK_READY; } - } else { - /* - * rte_flow_op_result::user data can point to - * struct mlx5_aso_mtr object as well - */ - if (queue != CTRL_QUEUE_ID(priv)) - MLX5_ASSERT(false); } } @@ -11007,7 +11002,8 @@ flow_hw_action_job_init(struct mlx5_priv *priv, uint32_t queue, { struct mlx5_hw_q_job *job; - MLX5_ASSERT(queue != MLX5_HW_INV_QUEUE); + if (queue == MLX5_HW_INV_QUEUE) + queue = CTRL_QUEUE_ID(priv); job = flow_hw_job_get(priv, queue); if (!job) { rte_flow_error_set(error, ENOMEM, @@ -11022,6 +11018,17 @@ flow_hw_action_job_init(struct mlx5_priv *priv, uint32_t queue, return job; } +struct mlx5_hw_q_job * +mlx5_flow_action_job_init(struct mlx5_priv *priv, uint32_t queue, + const struct rte_flow_action_handle *handle, + void *user_data, void *query_data, + enum mlx5_hw_job_type type, + struct rte_flow_error *error) +{ + return flow_hw_action_job_init(priv, queue, handle, user_data, query_data, + type, error); +} + static __rte_always_inline void flow_hw_action_finalize(struct rte_eth_dev *dev, uint32_t queue, struct mlx5_hw_q_job *job, @@ -11081,12 +11088,12 @@ flow_hw_action_handle_create(struct rte_eth_dev *dev, uint32_t queue, const struct rte_flow_action_age *age; struct mlx5_aso_mtr *aso_mtr; cnt_id_t cnt_id; - uint32_t mtr_id; uint32_t age_idx; bool push = flow_hw_action_push(attr); bool aso = false; + bool force_job = action->type == RTE_FLOW_ACTION_TYPE_METER_MARK; - if (attr) { + if (attr || force_job) { job = flow_hw_action_job_init(priv, queue, NULL, user_data, NULL, MLX5_HW_Q_JOB_TYPE_CREATE, error); @@ -11141,9 +11148,7 @@ flow_hw_action_handle_create(struct rte_eth_dev *dev, uint32_t queue, aso_mtr = flow_hw_meter_mark_alloc(dev, queue, action, job, push, error); if (!aso_mtr) break; - mtr_id = (MLX5_INDIRECT_ACTION_TYPE_METER_MARK << - MLX5_INDIRECT_ACTION_TYPE_OFFSET) | (aso_mtr->fm.meter_id); - handle = (struct rte_flow_action_handle *)(uintptr_t)mtr_id; + handle = (void *)(uintptr_t)job->action; break; case RTE_FLOW_ACTION_TYPE_RSS: handle = flow_dv_action_create(dev, conf, action, error); @@ -11158,7 +11163,7 @@ flow_hw_action_handle_create(struct rte_eth_dev *dev, uint32_t queue, NULL, "action type not supported"); break; } - if (job) { + if (job && !force_job) { job->action = handle; job->indirect_type = MLX5_HW_INDIRECT_TYPE_LEGACY; flow_hw_action_finalize(dev, queue, job, push, aso, @@ -11191,15 +11196,17 @@ mlx5_flow_update_meter_mark(struct rte_eth_dev *dev, uint32_t queue, fm->color_aware = meter_mark->color_mode; if (upd_meter_mark->state_valid) fm->is_enable = meter_mark->state; + aso_mtr->state = (queue == MLX5_HW_INV_QUEUE) ? + ASO_METER_WAIT : ASO_METER_WAIT_ASYNC; /* Update ASO flow meter by wqe. */ - if (mlx5_aso_meter_update_by_wqe(priv->sh, queue, + if (mlx5_aso_meter_update_by_wqe(priv, queue, aso_mtr, &priv->mtr_bulk, job, push)) return rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL, "Unable to update ASO meter WQE"); /* Wait for ASO object completion. */ if (queue == MLX5_HW_INV_QUEUE && - mlx5_aso_mtr_wait(priv->sh, MLX5_HW_INV_QUEUE, aso_mtr)) + mlx5_aso_mtr_wait(priv, aso_mtr, true)) return rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL, "Unable to wait for ASO meter CQE"); @@ -11245,8 +11252,9 @@ flow_hw_action_handle_update(struct rte_eth_dev *dev, uint32_t queue, int ret = 0; bool push = flow_hw_action_push(attr); bool aso = false; + bool force_job = type == MLX5_INDIRECT_ACTION_TYPE_METER_MARK; - if (attr) { + if (attr || force_job) { job = flow_hw_action_job_init(priv, queue, handle, user_data, NULL, MLX5_HW_Q_JOB_TYPE_UPDATE, error); @@ -11283,7 +11291,7 @@ flow_hw_action_handle_update(struct rte_eth_dev *dev, uint32_t queue, "action type not supported"); break; } - if (job) + if (job && !force_job) flow_hw_action_finalize(dev, queue, job, push, aso, ret == 0); return ret; } @@ -11326,8 +11334,9 @@ flow_hw_action_handle_destroy(struct rte_eth_dev *dev, uint32_t queue, bool push = flow_hw_action_push(attr); bool aso = false; int ret = 0; + bool force_job = type == MLX5_INDIRECT_ACTION_TYPE_METER_MARK; - if (attr) { + if (attr || force_job) { job = flow_hw_action_job_init(priv, queue, handle, user_data, NULL, MLX5_HW_Q_JOB_TYPE_DESTROY, error); @@ -11363,7 +11372,7 @@ flow_hw_action_handle_destroy(struct rte_eth_dev *dev, uint32_t queue, fm = &aso_mtr->fm; fm->is_enable = 0; /* Update ASO flow meter by wqe. */ - if (mlx5_aso_meter_update_by_wqe(priv->sh, queue, aso_mtr, + if (mlx5_aso_meter_update_by_wqe(priv, queue, aso_mtr, &priv->mtr_bulk, job, push)) { ret = -EINVAL; rte_flow_error_set(error, EINVAL, @@ -11373,7 +11382,7 @@ flow_hw_action_handle_destroy(struct rte_eth_dev *dev, uint32_t queue, } /* Wait for ASO object completion. */ if (queue == MLX5_HW_INV_QUEUE && - mlx5_aso_mtr_wait(priv->sh, MLX5_HW_INV_QUEUE, aso_mtr)) { + mlx5_aso_mtr_wait(priv, aso_mtr, true)) { ret = -EINVAL; rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_UNSPECIFIED, @@ -11397,7 +11406,7 @@ flow_hw_action_handle_destroy(struct rte_eth_dev *dev, uint32_t queue, "action type not supported"); break; } - if (job) + if (job && !force_job) flow_hw_action_finalize(dev, queue, job, push, aso, ret == 0); return ret; } diff --git a/drivers/net/mlx5/mlx5_flow_meter.c b/drivers/net/mlx5/mlx5_flow_meter.c index 57de95b4b9..4045c4c249 100644 --- a/drivers/net/mlx5/mlx5_flow_meter.c +++ b/drivers/net/mlx5/mlx5_flow_meter.c @@ -1897,12 +1897,12 @@ mlx5_flow_meter_action_modify(struct mlx5_priv *priv, if (sh->meter_aso_en) { fm->is_enable = !!is_enable; aso_mtr = container_of(fm, struct mlx5_aso_mtr, fm); - ret = mlx5_aso_meter_update_by_wqe(sh, MLX5_HW_INV_QUEUE, + ret = mlx5_aso_meter_update_by_wqe(priv, MLX5_HW_INV_QUEUE, aso_mtr, &priv->mtr_bulk, NULL, true); if (ret) return ret; - ret = mlx5_aso_mtr_wait(sh, MLX5_HW_INV_QUEUE, aso_mtr); + ret = mlx5_aso_mtr_wait(priv, aso_mtr, false); if (ret) return ret; } else { @@ -2148,7 +2148,7 @@ mlx5_flow_meter_create(struct rte_eth_dev *dev, uint32_t meter_id, /* If ASO meter supported, update ASO flow meter by wqe. */ if (priv->sh->meter_aso_en) { aso_mtr = container_of(fm, struct mlx5_aso_mtr, fm); - ret = mlx5_aso_meter_update_by_wqe(priv->sh, MLX5_HW_INV_QUEUE, + ret = mlx5_aso_meter_update_by_wqe(priv, MLX5_HW_INV_QUEUE, aso_mtr, &priv->mtr_bulk, NULL, true); if (ret) goto error; @@ -2210,6 +2210,7 @@ mlx5_flow_meter_hws_create(struct rte_eth_dev *dev, uint32_t meter_id, struct mlx5_flow_meter_info *fm; struct mlx5_flow_meter_policy *policy = NULL; struct mlx5_aso_mtr *aso_mtr; + struct mlx5_hw_q_job *job; int ret; if (!priv->mtr_profile_arr || @@ -2255,12 +2256,20 @@ mlx5_flow_meter_hws_create(struct rte_eth_dev *dev, uint32_t meter_id, fm->shared = !!shared; fm->initialized = 1; /* Update ASO flow meter by wqe. */ - ret = mlx5_aso_meter_update_by_wqe(priv->sh, MLX5_HW_INV_QUEUE, aso_mtr, - &priv->mtr_bulk, NULL, true); - if (ret) + job = mlx5_flow_action_job_init(priv, MLX5_HW_INV_QUEUE, NULL, NULL, + NULL, MLX5_HW_Q_JOB_TYPE_CREATE, NULL); + if (!job) + return -rte_mtr_error_set(error, ENOMEM, + RTE_MTR_ERROR_TYPE_MTR_ID, + NULL, "No job context."); + ret = mlx5_aso_meter_update_by_wqe(priv, MLX5_HW_INV_QUEUE, aso_mtr, + &priv->mtr_bulk, job, true); + if (ret) { + flow_hw_job_put(priv, job, MLX5_HW_INV_QUEUE); return -rte_mtr_error_set(error, ENOTSUP, - RTE_MTR_ERROR_TYPE_UNSPECIFIED, - NULL, "Failed to create devx meter."); + RTE_MTR_ERROR_TYPE_UNSPECIFIED, + NULL, "Failed to create devx meter."); + } fm->active_state = params->meter_enable; __atomic_fetch_add(&fm->profile->ref_cnt, 1, __ATOMIC_RELAXED); __atomic_fetch_add(&policy->ref_cnt, 1, __ATOMIC_RELAXED); @@ -2911,7 +2920,7 @@ mlx5_flow_meter_attach(struct mlx5_priv *priv, struct mlx5_aso_mtr *aso_mtr; aso_mtr = container_of(fm, struct mlx5_aso_mtr, fm); - if (mlx5_aso_mtr_wait(priv->sh, MLX5_HW_INV_QUEUE, aso_mtr)) { + if (mlx5_aso_mtr_wait(priv, aso_mtr, false)) { return rte_flow_error_set(error, ENOENT, RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL, From patchwork Thu Mar 7 10:12:11 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Gregory Etelson X-Patchwork-Id: 138093 X-Patchwork-Delegate: rasland@nvidia.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 1E82743BA7; Thu, 7 Mar 2024 11:13:11 +0100 (CET) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 60E6642EE0; Thu, 7 Mar 2024 11:13:05 +0100 (CET) Received: from NAM10-MW2-obe.outbound.protection.outlook.com (mail-mw2nam10on2072.outbound.protection.outlook.com [40.107.94.72]) by mails.dpdk.org (Postfix) with ESMTP id C05A442ED1; Thu, 7 Mar 2024 11:13:03 +0100 (CET) ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=EF4UzYeTa2x3Q0ofXS/naIHhWmGCcTUnidt4HOLOlVx6sx3NnkxnKnFoJi8woSx7Q0AN04VUmJAqE053SNfjRDRb9RrrVkFcX+p0JUBkWY2Nn9paMXwlkCPAMBba4AuhBh79M6y6EsIZfQWQSlCutbbkrMNWXdWLExebsfVH8xLQiph5fPJyriMt4EaZVJEAB3MYYEI6/B9WP1DVtX7opSc1MMZ0Mupf9Wt/tixubmWVRRbxImp3PrGHeSSvV5OEhRIiVFE3jIvfRxQ+i6bCHqDlRrzMFxfenT3OwoR5Uer37JsVqLdjsDEipg81PkV71k3xlX1zECjFvgrwX4t8hg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=PSWdCgDkWgCQJ3c8j5GWoj0qjmrPuQmyaXN3TbUxnfc=; b=Dut9VfGVhjkXTnD4/TgDv1F4x1q2l4V7w/LFdl58/dd1P7b+Oz5STrxacfYCUPKT029kBRuWvPcQQw/1NyA5cMT2rX1gi05111bk+cJ1x1fuqc+k2Z5A/xkzF1N9mg76ycplPwctp3IhTooE057NqhVRoWm+0XdmA37tlg3bos1ZrI+UM9EU+ZR5tHXUflNWo9oyc1AlqMzmNDFg3MS0Pfma/OgjpTbA4Aczv0fDTt/MIZWbOfnCkPEusQanrPT+U4tsMrcdGh2NypCH9RtUcWsVxlYaCST5MHBbnj5BcIUm0JxylYv1P/NdoR0ZlasZnmNf0sOQ/oqDJT5KGSxhGw== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 216.228.117.160) smtp.rcpttodomain=dpdk.org smtp.mailfrom=nvidia.com; dmarc=pass (p=reject sp=reject pct=100) action=none header.from=nvidia.com; dkim=none (message not signed); arc=none (0) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=Nvidia.com; s=selector2; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=PSWdCgDkWgCQJ3c8j5GWoj0qjmrPuQmyaXN3TbUxnfc=; b=SyKtcWxC8/Y5B2J59b08cXdMXN+UOgHcroJziZmccyd4KP2bc7Jeb38gJ4MHACSV1EKoWRtfJJpk5voanb1mO/lnLBgE1BWJeuF5T/9iiO22hNZqMYs6GvznZuWl0u2Tmdcrsxj0+4+O404Er4c+dJXDX3phjtq4xtcpsnKOuBUITz2h1rqj5Y5Y0qhcZRI3n1BsIoz+Lr1Ls+i5tIEPWSuFpbJ5TQcC0Ci/Bt1l65+Z/gwDrwT/tj1CL7BJ6EqTS45zeb0ueMxvwuGEy2toEePWw5QI7pGloYP9qsyRXj13Bwj9MF8fwx07XTF/P7/aWZ0Q5UrHuf7B049UuCTXvw== Received: from SA9PR13CA0177.namprd13.prod.outlook.com (2603:10b6:806:28::32) by CY8PR12MB8412.namprd12.prod.outlook.com (2603:10b6:930:6f::11) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.7362.26; Thu, 7 Mar 2024 10:13:01 +0000 Received: from SA2PEPF0000150A.namprd04.prod.outlook.com (2603:10b6:806:28:cafe::9c) by SA9PR13CA0177.outlook.office365.com (2603:10b6:806:28::32) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.7386.9 via Frontend Transport; Thu, 7 Mar 2024 10:13:01 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 216.228.117.160) smtp.mailfrom=nvidia.com; dkim=none (message not signed) header.d=none;dmarc=pass action=none header.from=nvidia.com; Received-SPF: Pass (protection.outlook.com: domain of nvidia.com designates 216.228.117.160 as permitted sender) receiver=protection.outlook.com; client-ip=216.228.117.160; helo=mail.nvidia.com; pr=C Received: from mail.nvidia.com (216.228.117.160) by SA2PEPF0000150A.mail.protection.outlook.com (10.167.242.42) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.7362.11 via Frontend Transport; Thu, 7 Mar 2024 10:13:01 +0000 Received: from rnnvmail201.nvidia.com (10.129.68.8) by mail.nvidia.com (10.129.200.66) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.41; Thu, 7 Mar 2024 02:12:42 -0800 Received: from nvidia.com (10.126.230.35) by rnnvmail201.nvidia.com (10.129.68.8) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.1258.12; Thu, 7 Mar 2024 02:12:38 -0800 From: Gregory Etelson To: CC: , , , , Dariusz Sosnowski , "Viacheslav Ovsiienko" , Ori Kam , Suanming Mou , Matan Azrad Subject: [PATCH 3/3] net/mlx5: fix indirect action async job initialization Date: Thu, 7 Mar 2024 12:12:11 +0200 Message-ID: <20240307101211.1134473-4-getelson@nvidia.com> X-Mailer: git-send-email 2.39.2 In-Reply-To: <20240307101211.1134473-1-getelson@nvidia.com> References: <20240307101211.1134473-1-getelson@nvidia.com> MIME-Version: 1.0 X-Originating-IP: [10.126.230.35] X-ClientProxiedBy: rnnvmail203.nvidia.com (10.129.68.9) To rnnvmail201.nvidia.com (10.129.68.8) X-EOPAttributedMessage: 0 X-MS-PublicTrafficType: Email X-MS-TrafficTypeDiagnostic: SA2PEPF0000150A:EE_|CY8PR12MB8412:EE_ X-MS-Office365-Filtering-Correlation-Id: 8051dcc4-7f5a-441e-024f-08dc3e8f2bcc X-MS-Exchange-SenderADCheck: 1 X-MS-Exchange-AntiSpam-Relay: 0 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: 5F/VBrvJ1nrM6jO6KJSPXsM2LVY+vsZRbRW61gU2fCja5yi0HGsz/4J6PDt7oNdG9jXfMc1/q/dKJprpvKSfTR4iZiCxWJLIz3KMtVG1hi1CBVrGTIDNweKv6LnAacmgyTBhvjN9pKlWZYpXuwysidadQqz+p4gq8n9dSFTfDISkL3N7UyxM3GFWZE8Ctw/vdPlbAzoTPceVNZyPJ1tDfisem4dlBf3gAcWaHqxd3e5fTunpjvQMfFU3LKklEgq/yQiLjme32TregPMN7QrZOnPrAKG0YRmQpQG7dPmx1BbiQkx2wXZTo+vzySg9BpzEf19OLMlfjo0hPe+fpt1mhbsBZNj3OdXSo13Pr7ZL3DKYhLvoZO9CEp3hrd3erke+g7x/cC9MdtUUMpPLevo39YWoryA+iCkE6GM/kcIbcZZKptiLMHPX0TXaNSEJdnFH9DZrh4W3ZqIp6dBlJbOkK40nRDJrDAt9Dt5XLGn6rjgldRUAAUFrE1Ni3U992Q+n1nNmInyhu/z9TLFCIqZB5U6CRxwdTJFDfEGDdpWGKc9L7hKS5nvlqpFbzF9rZQ3AmhbeF7eJf9gPX37xm5X+IWuNY+n/UGgz9TF2o9ad1QDX4hUDNChuA/Q75GRirP+01gH24a2jcARTr7GXdKPesUvM5RTS2paKWEJxbnEgIVHAVDk05EsvlxhssX29eAvGg3mI4Py5Kj7XDSXzFNSUQWtWGG9gigtsxplBY7W+R45Trv64rOZ70rOQNUHfHkUw X-Forefront-Antispam-Report: CIP:216.228.117.160; CTRY:US; LANG:en; SCL:1; SRV:; IPV:NLI; SFV:NSPM; H:mail.nvidia.com; PTR:dc6edge1.nvidia.com; CAT:NONE; SFS:(13230031)(82310400014)(36860700004)(376005); DIR:OUT; SFP:1101; X-OriginatorOrg: Nvidia.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 07 Mar 2024 10:13:01.4397 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: 8051dcc4-7f5a-441e-024f-08dc3e8f2bcc X-MS-Exchange-CrossTenant-Id: 43083d15-7273-40c1-b7db-39efd9ccc17a X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=43083d15-7273-40c1-b7db-39efd9ccc17a; Ip=[216.228.117.160]; Helo=[mail.nvidia.com] X-MS-Exchange-CrossTenant-AuthSource: SA2PEPF0000150A.namprd04.prod.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: CY8PR12MB8412 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org MLX5 PMD driver supports 2 types of indirect actions: legacy INDIRECT and INDIRECT_LIST. PMD has different handlers for each of indirection actions types. Therefore PMD marks async `job::indirect_type` with relevant value. PMD set the type only during indirect action creation. Legacy INDIRECT query could have get a job object used previously by INDIRECT_LIST action. In that case such job object was handled as INDIRECT_LIST because the `job::indirect_type` was not re-assigned. The patch sets `job::indirect_type` during the job initialization according to operation type. Fixes: 59155721936e ("net/mlx5: fix indirect flow completion processing") Cc: stable@dpdk.org Signed-off-by: Gregory Etelson Acked-by: Dariusz Sosnowski --- drivers/net/mlx5/mlx5_flow_hw.c | 24 +++++++++++++----------- 1 file changed, 13 insertions(+), 11 deletions(-) diff --git a/drivers/net/mlx5/mlx5_flow_hw.c b/drivers/net/mlx5/mlx5_flow_hw.c index 8f004b5435..b9ba05f695 100644 --- a/drivers/net/mlx5/mlx5_flow_hw.c +++ b/drivers/net/mlx5/mlx5_flow_hw.c @@ -188,6 +188,7 @@ flow_hw_action_job_init(struct mlx5_priv *priv, uint32_t queue, const struct rte_flow_action_handle *handle, void *user_data, void *query_data, enum mlx5_hw_job_type type, + enum mlx5_hw_indirect_type indirect_type, struct rte_flow_error *error); static int mlx5_tbl_multi_pattern_process(struct rte_eth_dev *dev, @@ -1692,7 +1693,8 @@ flow_hw_meter_mark_compile(struct rte_eth_dev *dev, struct mlx5_aso_mtr *aso_mtr; struct mlx5_hw_q_job *job = flow_hw_action_job_init(priv, queue, NULL, NULL, NULL, - MLX5_HW_Q_JOB_TYPE_CREATE, NULL); + MLX5_HW_Q_JOB_TYPE_CREATE, + MLX5_HW_INDIRECT_TYPE_LEGACY, NULL); if (!job) return -1; @@ -10998,6 +11000,7 @@ flow_hw_action_job_init(struct mlx5_priv *priv, uint32_t queue, const struct rte_flow_action_handle *handle, void *user_data, void *query_data, enum mlx5_hw_job_type type, + enum mlx5_hw_indirect_type indirect_type, struct rte_flow_error *error) { struct mlx5_hw_q_job *job; @@ -11015,6 +11018,7 @@ flow_hw_action_job_init(struct mlx5_priv *priv, uint32_t queue, job->action = handle; job->user_data = user_data; job->query.user = query_data; + job->indirect_type = indirect_type; return job; } @@ -11026,7 +11030,7 @@ mlx5_flow_action_job_init(struct mlx5_priv *priv, uint32_t queue, struct rte_flow_error *error) { return flow_hw_action_job_init(priv, queue, handle, user_data, query_data, - type, error); + type, MLX5_HW_INDIRECT_TYPE_LEGACY, error); } static __rte_always_inline void @@ -11096,7 +11100,7 @@ flow_hw_action_handle_create(struct rte_eth_dev *dev, uint32_t queue, if (attr || force_job) { job = flow_hw_action_job_init(priv, queue, NULL, user_data, NULL, MLX5_HW_Q_JOB_TYPE_CREATE, - error); + MLX5_HW_INDIRECT_TYPE_LEGACY, error); if (!job) return NULL; } @@ -11165,7 +11169,6 @@ flow_hw_action_handle_create(struct rte_eth_dev *dev, uint32_t queue, } if (job && !force_job) { job->action = handle; - job->indirect_type = MLX5_HW_INDIRECT_TYPE_LEGACY; flow_hw_action_finalize(dev, queue, job, push, aso, handle != NULL); } @@ -11257,7 +11260,7 @@ flow_hw_action_handle_update(struct rte_eth_dev *dev, uint32_t queue, if (attr || force_job) { job = flow_hw_action_job_init(priv, queue, handle, user_data, NULL, MLX5_HW_Q_JOB_TYPE_UPDATE, - error); + MLX5_HW_INDIRECT_TYPE_LEGACY, error); if (!job) return -rte_errno; } @@ -11339,7 +11342,7 @@ flow_hw_action_handle_destroy(struct rte_eth_dev *dev, uint32_t queue, if (attr || force_job) { job = flow_hw_action_job_init(priv, queue, handle, user_data, NULL, MLX5_HW_Q_JOB_TYPE_DESTROY, - error); + MLX5_HW_INDIRECT_TYPE_LEGACY, error); if (!job) return -rte_errno; } @@ -11663,7 +11666,7 @@ flow_hw_action_handle_query(struct rte_eth_dev *dev, uint32_t queue, if (attr) { job = flow_hw_action_job_init(priv, queue, handle, user_data, data, MLX5_HW_Q_JOB_TYPE_QUERY, - error); + MLX5_HW_INDIRECT_TYPE_LEGACY, error); if (!job) return -rte_errno; } @@ -11717,7 +11720,7 @@ flow_hw_async_action_handle_query_update job = flow_hw_action_job_init(priv, queue, handle, user_data, query, MLX5_HW_Q_JOB_TYPE_UPDATE_QUERY, - error); + MLX5_HW_INDIRECT_TYPE_LEGACY, error); if (!job) return -rte_errno; } @@ -12397,7 +12400,7 @@ flow_hw_async_action_list_handle_create(struct rte_eth_dev *dev, uint32_t queue, if (attr) { job = flow_hw_action_job_init(priv, queue, NULL, user_data, NULL, MLX5_HW_Q_JOB_TYPE_CREATE, - error); + MLX5_HW_INDIRECT_TYPE_LIST, error); if (!job) return NULL; } @@ -12417,7 +12420,6 @@ flow_hw_async_action_list_handle_create(struct rte_eth_dev *dev, uint32_t queue, } if (job) { job->action = handle; - job->indirect_type = MLX5_HW_INDIRECT_TYPE_LIST; flow_hw_action_finalize(dev, queue, job, push, false, handle != NULL); } @@ -12462,7 +12464,7 @@ flow_hw_async_action_list_handle_destroy if (attr) { job = flow_hw_action_job_init(priv, queue, NULL, user_data, NULL, MLX5_HW_Q_JOB_TYPE_DESTROY, - error); + MLX5_HW_INDIRECT_TYPE_LIST, error); if (!job) return rte_errno; }