From patchwork Tue Jul 6 13:14:48 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Shun Hao X-Patchwork-Id: 95383 X-Patchwork-Delegate: rasland@nvidia.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 299FBA0C47; Tue, 6 Jul 2021 15:15:45 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 09B574133D; Tue, 6 Jul 2021 15:15:45 +0200 (CEST) Received: from NAM11-DM6-obe.outbound.protection.outlook.com (mail-dm6nam11on2051.outbound.protection.outlook.com [40.107.223.51]) by mails.dpdk.org (Postfix) with ESMTP id 107F74120E for ; Tue, 6 Jul 2021 15:15:44 +0200 (CEST) ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=cngPqrU9hC+jR02gV06a6NFnU8A7oUDncks2CwCcnu8wk1sew8urGcBdIPZ5kVbudu7Pbm7oQgXo5rH5OrMcgpmtO2ff+4I0x1ftHlfKgTRDSy1k3UGMsuTqSaqd9v+MgMitALX1teSYAvsNFnyg39OkMRzD6WHZKEyU6FkiZniuGP0/t+Ort5KtFye9y61+OUIBm2BL3o5uyRVGeIsTA6Iq3eZsOUmN2e+6Z6O2J8sMc6bSoHGL7vYMzjO1KY6s+iKkTj7N9qSk2K5WUK9IfEai98At/LdQWsSKP2mxNJCphhHbTzn09hE41Jj61+zKRC8T//E8oIQMKpaFPoBlHw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=+UXyOBbNyYeBGOmsHZ2w2bBuk3MPrC0HDe1JLJV/D1k=; b=bHLIDgnY4EgFguTH3xmFuf/o6JQmF2k6r78PGM4mmhxmJ1hijBTJES32KHs9KrYLtoIVT6MSdyNDPdlOyERknwvsBrz96Zqgf6J7L/2AxXmoV9dlqQwf8sb3qR9spc9bU2uDGLC8TyDPgB1N4jiihGis7NRqJQ6s74BQBfoH32bvS6A7dfiurD0OUZKFC4/Ba2+e/lXW0iz2/rtrCDqQoBne+GMUXPqRBxzCbX743fH3cIqgHI/PYxkKBwBTZUmxHJZ29UjVvtijT5/sU+KfWR3Ex0BGuznfRZueUz4Y8/BUuC47sHAgib58az22hL4Yas/oNxwym9KTUryRMtMijA== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 216.228.112.34) smtp.rcpttodomain=monjalon.net smtp.mailfrom=nvidia.com; dmarc=pass (p=none sp=none pct=100) action=none header.from=nvidia.com; dkim=none (message not signed); arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=Nvidia.com; s=selector2; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=+UXyOBbNyYeBGOmsHZ2w2bBuk3MPrC0HDe1JLJV/D1k=; b=bOFzWDXp5YsV4cQhXt9CDVtAwQZRNb7204XweIE1FhM3/itS1YQJ2XdkqF9TnAbm5Cq8lFlA8kqCbxkE3qVXl40K6HHslt3CdXI0SIwgFFQpFbSdxu3E1Uz+qhWUbs+dnSC3AMsAyXQrVbCx2mjH1iQPvXkbLwHhdN4YguKHdllwXUNQUA1k8HQ7NMIlG/7lcFt95xP0eBhVHa+gP8PELzP8VNV2tWckORlabcOjwoq/xiYP/tU7Bw2y2WgI6aJq913EQL5cO350z1PqefKAeoOeORW+cs8epp8l3S6ql9EvUuc8f0PQSR7jS/POt6tysl8tUs5m9ihNHskZ19SIyQ== Received: from BN6PR14CA0021.namprd14.prod.outlook.com (2603:10b6:404:79::31) by BN6PR12MB1425.namprd12.prod.outlook.com (2603:10b6:404:1f::19) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.4287.31; Tue, 6 Jul 2021 13:15:40 +0000 Received: from BN8NAM11FT003.eop-nam11.prod.protection.outlook.com (2603:10b6:404:79:cafe::c4) by BN6PR14CA0021.outlook.office365.com (2603:10b6:404:79::31) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.4308.19 via Frontend Transport; Tue, 6 Jul 2021 13:15:40 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 216.228.112.34) smtp.mailfrom=nvidia.com; monjalon.net; dkim=none (message not signed) header.d=none;monjalon.net; dmarc=pass action=none header.from=nvidia.com; Received-SPF: Pass (protection.outlook.com: domain of nvidia.com designates 216.228.112.34 as permitted sender) receiver=protection.outlook.com; client-ip=216.228.112.34; helo=mail.nvidia.com; Received: from mail.nvidia.com (216.228.112.34) by BN8NAM11FT003.mail.protection.outlook.com (10.13.177.90) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384) id 15.20.4287.22 via Frontend Transport; Tue, 6 Jul 2021 13:15:40 +0000 Received: from nvidia.com (172.20.187.5) by HQMAIL107.nvidia.com (172.20.187.13) with Microsoft SMTP Server (TLS) id 15.0.1497.2; Tue, 6 Jul 2021 13:15:37 +0000 From: Shun Hao To: , , , "Shahaf Shuler" CC: , , Date: Tue, 6 Jul 2021 16:14:48 +0300 Message-ID: <20210706131450.30917-3-shunh@nvidia.com> X-Mailer: git-send-email 2.20.0 In-Reply-To: <20210706131450.30917-1-shunh@nvidia.com> References: <20210706131450.30917-1-shunh@nvidia.com> MIME-Version: 1.0 X-Originating-IP: [172.20.187.5] X-ClientProxiedBy: HQMAIL101.nvidia.com (172.20.187.10) To HQMAIL107.nvidia.com (172.20.187.13) X-EOPAttributedMessage: 0 X-MS-PublicTrafficType: Email X-MS-Office365-Filtering-Correlation-Id: e533dcd2-fcca-4746-5001-08d940802707 X-MS-TrafficTypeDiagnostic: BN6PR12MB1425: X-LD-Processed: 43083d15-7273-40c1-b7db-39efd9ccc17a,ExtAddr X-Microsoft-Antispam-PRVS: X-MS-Oob-TLC-OOBClassifiers: OLM:1227; X-MS-Exchange-SenderADCheck: 1 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: 5vob7CB0br+vtzwFXkysIKm2KYFMYKjlFmGboPkEa+rL4KWDg42z2I81ds/xBX8P9dBoY8FgvUbA+w1XxFVnTQEP76O1ekl8GDs5IRSv1qXos90j+cEaLHKZJy6DQ6nE6debSSQirFlN7FBzGooSLls/zxTccDEMJHl7uoiT44mtL0Ayk7GIcIvonGE2vHJhyph4qR1uPeZDUrjuPFI5sD6iI4ZV9vkrib/HvZfm1v9OZO0O7afRuRt0IToPfBocZMRLy+/3R3wbSh7rQ/fPXi4YUj94/igBVHayH6b5iNNaapygISsfM8xTm8Z3YrQl2JCn/oWEYQ2WMPkkkORINnOun1yMLIh8kTdcPsk6giJIce1f1LgiIKltEIN2eGq316Q2fWLsiVmXVIWV0yH6tQIWCfMOANK2pYOedJAXDN+m2DkExbBeVqV0IilrBguDz35UgE5/uDdkQXzU5hv93PqNpoTArBEdSdal5piPn5C86XWc35p/3FknKxswfFOpl2oTLmRXdT/jvYSYlmM/dPEZ8TMXicmZuJTaheZ8+/WXii3Vxp/MxyBphUwHJs+c0uSDKTqyqXo+w4jHhipG0KtnNRIHadATWSjRqiJ7BaoTeMvqJiBoYq6il/BQjSdbkRs+C3CE602i0tPxhtRrLROC2ipXfqqmSKMIq0D/OXXn3lON0lQABXJRBtI6k5MgdyrbV/wuKQJLf/bQEAjVWw== X-Forefront-Antispam-Report: CIP:216.228.112.34; CTRY:US; LANG:en; SCL:1; SRV:; IPV:NLI; SFV:NSPM; H:mail.nvidia.com; PTR:schybrid03.nvidia.com; CAT:NONE; SFS:(4636009)(39860400002)(376002)(136003)(346002)(396003)(46966006)(36840700001)(2906002)(107886003)(2616005)(30864003)(336012)(186003)(110136005)(8936002)(6286002)(54906003)(8676002)(36906005)(5660300002)(1076003)(55016002)(478600001)(4326008)(316002)(426003)(36860700001)(70206006)(6666004)(16526019)(70586007)(36756003)(47076005)(7636003)(83380400001)(356005)(82310400003)(26005)(82740400003)(6636002)(7696005)(86362001); DIR:OUT; SFP:1101; X-OriginatorOrg: Nvidia.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 06 Jul 2021 13:15:40.1799 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: e533dcd2-fcca-4746-5001-08d940802707 X-MS-Exchange-CrossTenant-Id: 43083d15-7273-40c1-b7db-39efd9ccc17a X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=43083d15-7273-40c1-b7db-39efd9ccc17a; Ip=[216.228.112.34]; Helo=[mail.nvidia.com] X-MS-Exchange-CrossTenant-AuthSource: BN8NAM11FT003.eop-nam11.prod.protection.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: BN6PR12MB1425 Subject: [dpdk-dev] [PATCH v1 2/4] net/mlx5: support meter hierarchy drop count X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" When using meter hierarchy with multiple meters, every meter may have drop counter, so a packet being set red color by one meter should be counted to that specific meter only. To support this, add tag action in the color rule so packet going to next new meter can have its meter id, so as to be counted to the correct drop counter in drop table. Signed-off-by: Shun Hao Acked-by: Matan Azrad --- drivers/net/mlx5/mlx5.h | 20 +- drivers/net/mlx5/mlx5_flow.c | 52 +++++- drivers/net/mlx5/mlx5_flow.h | 7 + drivers/net/mlx5/mlx5_flow_dv.c | 318 ++++++++++++++++++++++++++------ 4 files changed, 339 insertions(+), 58 deletions(-) diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h index 0c555f0b1f..e5c9ec0777 100644 --- a/drivers/net/mlx5/mlx5.h +++ b/drivers/net/mlx5/mlx5.h @@ -631,6 +631,20 @@ enum mlx5_meter_domain { MLX5_MTR_DOMAIN_EGRESS_BIT | \ MLX5_MTR_DOMAIN_TRANSFER_BIT) +/* The color tag rule structure. */ +struct mlx5_sub_policy_color_rule { + void *rule; + /* The color rule. */ + struct mlx5_flow_dv_matcher *matcher; + /* The color matcher. */ + TAILQ_ENTRY(mlx5_sub_policy_color_rule) next_port; + /**< Pointer to the next color rule structure. */ + int32_t src_port; + /* On which src port this rule applied. */ +}; + +TAILQ_HEAD(mlx5_sub_policy_color_rules, mlx5_sub_policy_color_rule); + /* * Meter sub-policy structure. * Each RSS TIR in meter policy need its own sub-policy resource. @@ -648,10 +662,8 @@ struct mlx5_flow_meter_sub_policy { /* Index to TIR resource. */ struct mlx5_flow_tbl_resource *jump_tbl[MLX5_MTR_RTE_COLORS]; /* Meter jump/drop table. */ - struct mlx5_flow_dv_matcher *color_matcher[RTE_COLORS]; - /* Matcher for Color. */ - void *color_rule[RTE_COLORS]; - /* Meter green/yellow/drop rule. */ + struct mlx5_sub_policy_color_rules color_rules[RTE_COLORS]; + /* List for the color rules. */ }; struct mlx5_meter_policy_acts { diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c index 6c4bfde098..3cd91a7e8c 100644 --- a/drivers/net/mlx5/mlx5_flow.c +++ b/drivers/net/mlx5/mlx5_flow.c @@ -3449,6 +3449,41 @@ flow_drv_meter_sub_policy_rss_prepare(struct rte_eth_dev *dev, return fops->meter_sub_policy_rss_prepare(dev, policy, rss_desc); } +/** + * Flow driver color tag rule API. This abstracts calling driver + * specific functions. Parent flow (rte_flow) should have driver + * type (drv_type). It will create the color tag rules in hierarchy meter. + * + * @param[in] dev + * Pointer to Ethernet device. + * @param[in, out] flow + * Pointer to flow structure. + * @param[in] fm + * Pointer to flow meter structure. + * @param[in] src_port + * The src port this extra rule should use. + * @param[in] item + * The src port id match item. + * @param[out] error + * Pointer to error structure. + */ +static int +flow_drv_mtr_hierarchy_rule_create(struct rte_eth_dev *dev, + struct rte_flow *flow, + struct mlx5_flow_meter_info *fm, + int32_t src_port, + const struct rte_flow_item *item, + struct rte_flow_error *error) +{ + const struct mlx5_flow_driver_ops *fops; + enum mlx5_flow_drv_type type = flow->drv_type; + + MLX5_ASSERT(type > MLX5_FLOW_TYPE_MIN && type < MLX5_FLOW_TYPE_MAX); + fops = flow_get_drv_ops(type); + return fops->meter_hierarchy_rule_create(dev, fm, + src_port, item, error); +} + /** * Get RSS action from the action list. * @@ -4773,6 +4808,15 @@ flow_meter_split_prep(struct rte_eth_dev *dev, pid_v, "Failed to get port info."); flow_src_port = port_priv->representor_id; + if (!fm->def_policy && wks->policy->is_hierarchy && + flow_src_port != priv->representor_id) { + if (flow_drv_mtr_hierarchy_rule_create(dev, + flow, fm, + flow_src_port, + items, + error)) + return -rte_errno; + } memcpy(sfx_items, items, sizeof(*sfx_items)); sfx_items++; break; @@ -5713,6 +5757,7 @@ flow_create_split_meter(struct rte_eth_dev *dev, bool has_mtr = false; bool has_modify = false; bool set_mtr_reg = true; + bool is_mtr_hierarchy = false; uint32_t meter_id = 0; uint32_t mtr_idx = 0; uint32_t mtr_flow_id = 0; @@ -5759,6 +5804,7 @@ flow_create_split_meter(struct rte_eth_dev *dev, EINVAL, RTE_FLOW_ERROR_TYPE_ACTION, NULL, "Failed to find terminal policy of hierarchy."); + is_mtr_hierarchy = true; } } /* @@ -5766,9 +5812,11 @@ flow_create_split_meter(struct rte_eth_dev *dev, * 1. There's no action in flow to change * packet (modify/encap/decap etc.), OR * 2. No drop count needed for this meter. - * no need to use regC to save meter id anymore. + * 3. It's not meter hierarchy. + * Then no need to use regC to save meter id anymore. */ - if (!fm->def_policy && (!has_modify || !fm->drop_cnt)) + if (!fm->def_policy && !is_mtr_hierarchy && + (!has_modify || !fm->drop_cnt)) set_mtr_reg = false; /* Prefix actions: meter, decap, encap, tag, jump, end. */ act_size = sizeof(struct rte_flow_action) * (actions_n + 6) + diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h index 09d6d609db..7d97c5880f 100644 --- a/drivers/net/mlx5/mlx5_flow.h +++ b/drivers/net/mlx5/mlx5_flow.h @@ -1161,6 +1161,12 @@ typedef struct mlx5_flow_meter_sub_policy * (struct rte_eth_dev *dev, struct mlx5_flow_meter_policy *mtr_policy, struct mlx5_flow_rss_desc *rss_desc[MLX5_MTR_RTE_COLORS]); +typedef int (*mlx5_flow_meter_hierarchy_rule_create_t) + (struct rte_eth_dev *dev, + struct mlx5_flow_meter_info *fm, + int32_t src_port, + const struct rte_flow_item *item, + struct rte_flow_error *error); typedef void (*mlx5_flow_destroy_sub_policy_with_rxq_t) (struct rte_eth_dev *dev, struct mlx5_flow_meter_policy *mtr_policy); @@ -1257,6 +1263,7 @@ struct mlx5_flow_driver_ops { mlx5_flow_create_def_policy_t create_def_policy; mlx5_flow_destroy_def_policy_t destroy_def_policy; mlx5_flow_meter_sub_policy_rss_prepare_t meter_sub_policy_rss_prepare; + mlx5_flow_meter_hierarchy_rule_create_t meter_hierarchy_rule_create; mlx5_flow_destroy_sub_policy_with_rxq_t destroy_sub_policy_with_rxq; mlx5_flow_counter_alloc_t counter_alloc; mlx5_flow_counter_free_t counter_free; diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c index d34f5214a8..119de09809 100644 --- a/drivers/net/mlx5/mlx5_flow_dv.c +++ b/drivers/net/mlx5/mlx5_flow_dv.c @@ -23,6 +23,7 @@ #include #include #include +#include #include #include @@ -13090,6 +13091,15 @@ flow_dv_translate(struct rte_eth_dev *dev, matcher.mask.size); matcher.priority = mlx5_get_matcher_priority(dev, attr, matcher.priority); + /** + * When creating meter drop flow in drop table, using original + * 5-tuple match, the matcher priority should be lower than + * mtr_id matcher. + */ + if (attr->group == MLX5_FLOW_TABLE_LEVEL_METER && + dev_flow->dv.table_id == MLX5_MTR_TABLE_ID_DROP && + matcher.priority <= MLX5_REG_BITS) + matcher.priority += MLX5_REG_BITS; /* reserved field no needs to be set to 0 here. */ tbl_key.is_fdb = attr->transfer; tbl_key.is_egress = attr->egress; @@ -14579,20 +14589,21 @@ __flow_dv_destroy_sub_policy_rules(struct rte_eth_dev *dev, struct mlx5_flow_meter_sub_policy *sub_policy) { struct mlx5_flow_tbl_data_entry *tbl; + struct mlx5_sub_policy_color_rule *color_rule; + void *tmp; int i; for (i = 0; i < RTE_COLORS; i++) { - if (sub_policy->color_rule[i]) { - claim_zero(mlx5_flow_os_destroy_flow - (sub_policy->color_rule[i])); - sub_policy->color_rule[i] = NULL; - } - if (sub_policy->color_matcher[i]) { - tbl = container_of(sub_policy->color_matcher[i]->tbl, - typeof(*tbl), tbl); + TAILQ_FOREACH_SAFE(color_rule, &sub_policy->color_rules[i], + next_port, tmp) { + claim_zero(mlx5_flow_os_destroy_flow(color_rule->rule)); + tbl = container_of(color_rule->matcher->tbl, + typeof(*tbl), tbl); mlx5_cache_unregister(&tbl->matchers, - &sub_policy->color_matcher[i]->entry); - sub_policy->color_matcher[i] = NULL; + &color_rule->matcher->entry); + TAILQ_REMOVE(&sub_policy->color_rules[i], + color_rule, next_port); + mlx5_free(color_rule); } } for (i = 0; i < MLX5_MTR_RTE_COLORS; i++) { @@ -14741,6 +14752,7 @@ __flow_dv_create_domain_policy_acts(struct rte_eth_dev *dev, sizeof(struct mlx5_modification_cmd) * (MLX5_MAX_MODIFY_NUM + 1)]; } mhdr_dummy; + struct mlx5_flow_dv_modify_hdr_resource *mhdr_res = &mhdr_dummy.res; egress = (domain == MLX5_MTR_DOMAIN_EGRESS) ? 1 : 0; transfer = (domain == MLX5_MTR_DOMAIN_TRANSFER) ? 1 : 0; @@ -14748,6 +14760,11 @@ __flow_dv_create_domain_policy_acts(struct rte_eth_dev *dev, memset(&dev_flow, 0, sizeof(struct mlx5_flow)); memset(&port_id_action, 0, sizeof(struct mlx5_flow_dv_port_id_action_resource)); + memset(mhdr_res, 0, sizeof(*mhdr_res)); + mhdr_res->ft_type = transfer ? MLX5DV_FLOW_TABLE_TYPE_FDB : + egress ? + MLX5DV_FLOW_TABLE_TYPE_NIC_TX : + MLX5DV_FLOW_TABLE_TYPE_NIC_RX; dev_flow.handle = &dh; dev_flow.dv.port_id_action = &port_id_action; dev_flow.external = true; @@ -14786,10 +14803,6 @@ __flow_dv_create_domain_policy_acts(struct rte_eth_dev *dev, break; } case RTE_FLOW_ACTION_TYPE_SET_TAG: - { - struct mlx5_flow_dv_modify_hdr_resource - *mhdr_res = &mhdr_dummy.res; - if (i >= MLX5_MTR_RTE_COLORS) return -rte_mtr_error_set(error, ENOTSUP, @@ -14797,12 +14810,6 @@ __flow_dv_create_domain_policy_acts(struct rte_eth_dev *dev, NULL, "cannot create policy " "set tag action for this color"); - memset(mhdr_res, 0, sizeof(*mhdr_res)); - mhdr_res->ft_type = transfer ? - MLX5DV_FLOW_TABLE_TYPE_FDB : - egress ? - MLX5DV_FLOW_TABLE_TYPE_NIC_TX : - MLX5DV_FLOW_TABLE_TYPE_NIC_RX; if (flow_dv_convert_action_set_tag (dev, mhdr_res, (const struct rte_flow_action_set_tag *) @@ -14818,20 +14825,8 @@ __flow_dv_create_domain_policy_acts(struct rte_eth_dev *dev, RTE_MTR_ERROR_TYPE_METER_POLICY, NULL, "cannot find policy " "set tag action"); - /* create modify action if needed. */ - dev_flow.dv.group = 1; - if (flow_dv_modify_hdr_resource_register - (dev, mhdr_res, &dev_flow, &flow_err)) - return -rte_mtr_error_set(error, - ENOTSUP, - RTE_MTR_ERROR_TYPE_METER_POLICY, - NULL, "cannot register policy " - "set tag action"); - act_cnt->modify_hdr = - dev_flow.handle->dvh.modify_hdr; action_flags |= MLX5_FLOW_ACTION_SET_TAG; break; - } case RTE_FLOW_ACTION_TYPE_DROP: { struct mlx5_flow_mtr_mng *mtrmng = @@ -15035,6 +15030,8 @@ __flow_dv_create_domain_policy_acts(struct rte_eth_dev *dev, const struct rte_flow_action_meter *mtr; struct mlx5_flow_meter_info *next_fm; struct mlx5_flow_meter_policy *next_policy; + struct rte_flow_action tag_action; + struct mlx5_rte_flow_action_set_tag set_tag; uint32_t next_mtr_idx = 0; mtr = act->conf; @@ -15052,6 +15049,30 @@ __flow_dv_create_domain_policy_acts(struct rte_eth_dev *dev, next_policy = mlx5_flow_meter_policy_find(dev, next_fm->policy_id, NULL); MLX5_ASSERT(next_policy); + if (next_fm->drop_cnt) { + set_tag.id = + (enum modify_reg) + mlx5_flow_get_reg_id(dev, + MLX5_MTR_ID, + 0, + (struct rte_flow_error *)error); + set_tag.offset = (priv->mtr_reg_share ? + MLX5_MTR_COLOR_BITS : 0); + set_tag.length = (priv->mtr_reg_share ? + MLX5_MTR_IDLE_BITS_IN_COLOR_REG : + MLX5_REG_BITS); + set_tag.data = next_mtr_idx; + tag_action.type = + (enum rte_flow_action_type) + MLX5_RTE_FLOW_ACTION_TYPE_TAG; + tag_action.conf = &set_tag; + if (flow_dv_convert_action_set_reg + (mhdr_res, &tag_action, + (struct rte_flow_error *)error)) + return -rte_errno; + action_flags |= + MLX5_FLOW_ACTION_SET_TAG; + } act_cnt->fate_action = MLX5_FLOW_FATE_MTR; act_cnt->next_mtr_id = next_fm->meter_id; act_cnt->next_sub_policy = NULL; @@ -15066,6 +15087,19 @@ __flow_dv_create_domain_policy_acts(struct rte_eth_dev *dev, RTE_MTR_ERROR_TYPE_METER_POLICY, NULL, "action type not supported"); } + if (action_flags & MLX5_FLOW_ACTION_SET_TAG) { + /* create modify action if needed. */ + dev_flow.dv.group = 1; + if (flow_dv_modify_hdr_resource_register + (dev, mhdr_res, &dev_flow, &flow_err)) + return -rte_mtr_error_set(error, + ENOTSUP, + RTE_MTR_ERROR_TYPE_METER_POLICY, + NULL, "cannot register policy " + "set tag action"); + act_cnt->modify_hdr = + dev_flow.handle->dvh.modify_hdr; + } } } return 0; @@ -15418,8 +15452,8 @@ __flow_dv_create_policy_flow(struct rte_eth_dev *dev, uint32_t color_reg_c_idx, enum rte_color color, void *matcher_object, int actions_n, void *actions, - bool match_src_port, void **rule, - const struct rte_flow_attr *attr) + bool match_src_port, const struct rte_flow_item *item, + void **rule, const struct rte_flow_attr *attr) { int ret; struct mlx5_flow_dv_match_params value = { @@ -15434,7 +15468,7 @@ __flow_dv_create_policy_flow(struct rte_eth_dev *dev, if (match_src_port && (priv->representor || priv->master)) { if (flow_dv_translate_item_port_id(dev, matcher.buf, - value.buf, NULL, attr)) { + value.buf, item, attr)) { DRV_LOG(ERR, "Failed to create meter policy flow with port."); return -1; @@ -15460,6 +15494,8 @@ __flow_dv_create_policy_matcher(struct rte_eth_dev *dev, struct mlx5_flow_meter_sub_policy *sub_policy, const struct rte_flow_attr *attr, bool match_src_port, + const struct rte_flow_item *item, + struct mlx5_flow_dv_matcher **policy_matcher, struct rte_flow_error *error) { struct mlx5_cache_entry *entry; @@ -15485,7 +15521,7 @@ __flow_dv_create_policy_matcher(struct rte_eth_dev *dev, if (match_src_port && (priv->representor || priv->master)) { if (flow_dv_translate_item_port_id(dev, matcher.mask.buf, - value.buf, NULL, attr)) { + value.buf, item, attr)) { DRV_LOG(ERR, "Failed to register meter drop matcher with port."); return -1; @@ -15503,7 +15539,7 @@ __flow_dv_create_policy_matcher(struct rte_eth_dev *dev, DRV_LOG(ERR, "Failed to register meter drop matcher."); return -1; } - sub_policy->color_matcher[priority] = + *policy_matcher = container_of(entry, struct mlx5_flow_dv_matcher, entry); return 0; } @@ -15531,6 +15567,7 @@ __flow_dv_create_domain_policy_rules(struct rte_eth_dev *dev, uint8_t egress, uint8_t transfer, bool match_src_port, struct mlx5_meter_policy_acts acts[RTE_COLORS]) { + struct mlx5_priv *priv = dev->data->dev_private; struct rte_flow_error flow_err; uint32_t color_reg_c_idx; struct rte_flow_attr attr = { @@ -15543,6 +15580,7 @@ __flow_dv_create_domain_policy_rules(struct rte_eth_dev *dev, }; int i; int ret = mlx5_flow_get_reg_id(dev, MLX5_MTR_COLOR, 0, &flow_err); + struct mlx5_sub_policy_color_rule *color_rule; if (ret < 0) return -1; @@ -15560,29 +15598,56 @@ __flow_dv_create_domain_policy_rules(struct rte_eth_dev *dev, /* Prepare matchers. */ color_reg_c_idx = ret; for (i = 0; i < RTE_COLORS; i++) { + TAILQ_INIT(&sub_policy->color_rules[i]); if (i == RTE_COLOR_YELLOW || !acts[i].actions_n) continue; + color_rule = mlx5_malloc(MLX5_MEM_ZERO, + sizeof(struct mlx5_sub_policy_color_rule), + 0, SOCKET_ID_ANY); + if (!color_rule) { + DRV_LOG(ERR, "No memory to create color rule."); + goto err_exit; + } + color_rule->src_port = priv->representor_id; attr.priority = i; - if (!sub_policy->color_matcher[i]) { - /* Create matchers for Color. */ - if (__flow_dv_create_policy_matcher(dev, - color_reg_c_idx, i, sub_policy, - &attr, match_src_port, &flow_err)) - return -1; + /* Create matchers for Color. */ + if (__flow_dv_create_policy_matcher(dev, + color_reg_c_idx, i, sub_policy, &attr, + (i != RTE_COLOR_RED ? match_src_port : false), + NULL, &color_rule->matcher, &flow_err)) { + DRV_LOG(ERR, "Failed to create color matcher."); + goto err_exit; } /* Create flow, matching color. */ - if (acts[i].actions_n) - if (__flow_dv_create_policy_flow(dev, + if (__flow_dv_create_policy_flow(dev, color_reg_c_idx, (enum rte_color)i, - sub_policy->color_matcher[i]->matcher_object, + color_rule->matcher->matcher_object, acts[i].actions_n, acts[i].dv_actions, - match_src_port, - &sub_policy->color_rule[i], - &attr)) - return -1; + (i != RTE_COLOR_RED ? match_src_port : false), + NULL, &color_rule->rule, + &attr)) { + DRV_LOG(ERR, "Failed to create color rule."); + goto err_exit; + } + TAILQ_INSERT_TAIL(&sub_policy->color_rules[i], + color_rule, next_port); } return 0; +err_exit: + if (color_rule) { + if (color_rule->rule) + mlx5_flow_os_destroy_flow(color_rule->rule); + if (color_rule->matcher) { + struct mlx5_flow_tbl_data_entry *tbl = + container_of(color_rule->matcher->tbl, + typeof(*tbl), tbl); + mlx5_cache_unregister(&tbl->matchers, + &color_rule->matcher->entry); + } + mlx5_free(color_rule); + } + return -1; } static int @@ -15734,8 +15799,6 @@ __flow_dv_create_policy_acts_rules(struct rte_eth_dev *dev, } } } - egress = (domain == MLX5_MTR_DOMAIN_EGRESS) ? 1 : 0; - transfer = (domain == MLX5_MTR_DOMAIN_TRANSFER) ? 1 : 0; if (__flow_dv_create_domain_policy_rules(dev, sub_policy, egress, transfer, match_src_port, acts)) { DRV_LOG(ERR, @@ -16291,6 +16354,156 @@ flow_dv_meter_sub_policy_rss_prepare(struct rte_eth_dev *dev, return NULL; } +/** + * Create the sub policy tag rule for all meters in hierarchy. + * + * @param[in] dev + * Pointer to Ethernet device. + * @param[in] fm + * Meter information table. + * @param[in] src_port + * The src port this extra rule should use. + * @param[in] item + * The src port match item. + * @param[out] error + * Perform verbose error reporting if not NULL. + * @return + * 0 on success, a negative errno value otherwise and rte_errno is set. + */ +static int +flow_dv_meter_hierarchy_rule_create(struct rte_eth_dev *dev, + struct mlx5_flow_meter_info *fm, + int32_t src_port, + const struct rte_flow_item *item, + struct rte_flow_error *error) +{ + struct mlx5_priv *priv = dev->data->dev_private; + struct mlx5_flow_meter_policy *mtr_policy; + struct mlx5_flow_meter_sub_policy *sub_policy; + struct mlx5_flow_meter_info *next_fm = NULL; + struct mlx5_flow_meter_policy *next_policy; + struct mlx5_flow_meter_sub_policy *next_sub_policy; + struct mlx5_flow_tbl_data_entry *tbl_data; + struct mlx5_sub_policy_color_rule *color_rule; + struct mlx5_meter_policy_acts acts; + uint32_t color_reg_c_idx; + bool mtr_first = (src_port != 0xffff) ? true : false; + struct rte_flow_attr attr = { + .group = MLX5_FLOW_TABLE_LEVEL_POLICY, + .priority = 0, + .ingress = 0, + .egress = 0, + .transfer = 1, + .reserved = 0, + }; + uint32_t domain = MLX5_MTR_DOMAIN_TRANSFER; + int i; + + mtr_policy = mlx5_flow_meter_policy_find(dev, fm->policy_id, NULL); + MLX5_ASSERT(mtr_policy); + if (!mtr_policy->is_hierarchy) + return 0; + next_fm = mlx5_flow_meter_find(priv, + mtr_policy->act_cnt[RTE_COLOR_GREEN].next_mtr_id, NULL); + if (!next_fm) { + return rte_flow_error_set(error, EINVAL, + RTE_FLOW_ERROR_TYPE_ACTION, NULL, + "Failed to find next meter in hierarchy."); + } + if (!next_fm->drop_cnt) + goto exit; + color_reg_c_idx = mlx5_flow_get_reg_id(dev, MLX5_MTR_COLOR, 0, error); + sub_policy = mtr_policy->sub_policys[domain][0]; + for (i = 0; i < RTE_COLORS; i++) { + bool rule_exist = false; + struct mlx5_meter_policy_action_container *act_cnt; + + if (i >= RTE_COLOR_YELLOW) + break; + TAILQ_FOREACH(color_rule, + &sub_policy->color_rules[i], next_port) + if (color_rule->src_port == src_port) { + rule_exist = true; + break; + } + if (rule_exist) + continue; + color_rule = mlx5_malloc(MLX5_MEM_ZERO, + sizeof(struct mlx5_sub_policy_color_rule), + 0, SOCKET_ID_ANY); + if (!color_rule) + return rte_flow_error_set(error, ENOMEM, + RTE_FLOW_ERROR_TYPE_ACTION, + NULL, "No memory to create tag color rule."); + color_rule->src_port = src_port; + attr.priority = i; + next_policy = mlx5_flow_meter_policy_find(dev, + next_fm->policy_id, NULL); + MLX5_ASSERT(next_policy); + next_sub_policy = next_policy->sub_policys[domain][0]; + tbl_data = container_of(next_sub_policy->tbl_rsc, + struct mlx5_flow_tbl_data_entry, tbl); + act_cnt = &mtr_policy->act_cnt[i]; + if (mtr_first) { + acts.dv_actions[0] = next_fm->meter_action; + acts.dv_actions[1] = act_cnt->modify_hdr->action; + } else { + acts.dv_actions[0] = act_cnt->modify_hdr->action; + acts.dv_actions[1] = next_fm->meter_action; + } + acts.dv_actions[2] = tbl_data->jump.action; + acts.actions_n = 3; + if (mlx5_flow_meter_attach(priv, next_fm, &attr, error)) { + next_fm = NULL; + goto err_exit; + } + if (__flow_dv_create_policy_matcher(dev, color_reg_c_idx, + i, sub_policy, &attr, true, item, + &color_rule->matcher, error)) { + rte_flow_error_set(error, errno, + RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL, + "Failed to create hierarchy meter matcher."); + goto err_exit; + } + if (__flow_dv_create_policy_flow(dev, color_reg_c_idx, + (enum rte_color)i, + color_rule->matcher->matcher_object, + acts.actions_n, acts.dv_actions, + true, item, + &color_rule->rule, &attr)) { + rte_flow_error_set(error, errno, + RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL, + "Failed to create hierarchy meter rule."); + goto err_exit; + } + TAILQ_INSERT_TAIL(&sub_policy->color_rules[i], + color_rule, next_port); + } +exit: + /** + * Recursive call to iterate all meters in hierarchy and + * create needed rules. + */ + return flow_dv_meter_hierarchy_rule_create(dev, next_fm, + src_port, item, error); +err_exit: + if (color_rule) { + if (color_rule->rule) + mlx5_flow_os_destroy_flow(color_rule->rule); + if (color_rule->matcher) { + struct mlx5_flow_tbl_data_entry *tbl = + container_of(color_rule->matcher->tbl, + typeof(*tbl), tbl); + mlx5_cache_unregister(&tbl->matchers, + &color_rule->matcher->entry); + } + mlx5_free(color_rule); + } + if (next_fm) + mlx5_flow_meter_detach(priv, next_fm); + return -rte_errno; +} + /** * Destroy the sub policy table with RX queue. * @@ -16966,6 +17179,7 @@ const struct mlx5_flow_driver_ops mlx5_flow_dv_drv_ops = { .create_def_policy = flow_dv_create_def_policy, .destroy_def_policy = flow_dv_destroy_def_policy, .meter_sub_policy_rss_prepare = flow_dv_meter_sub_policy_rss_prepare, + .meter_hierarchy_rule_create = flow_dv_meter_hierarchy_rule_create, .destroy_sub_policy_with_rxq = flow_dv_destroy_sub_policy_with_rxq, .counter_alloc = flow_dv_counter_allocate, .counter_free = flow_dv_counter_free,