diff mbox series

[1/1] net/bnxt: fix handling of null flow mask

Message ID 20210318195251.683450-1-lance.richardson@broadcom.com (mailing list archive)
State Accepted
Delegated to: Ajit Khaparde
Headers show
Series [1/1] net/bnxt: fix handling of null flow mask | expand

Checks

Context Check Description
ci/iol-mellanox-Performance success Performance Testing PASS
ci/iol-testing success Testing PASS
ci/intel-Testing success Testing PASS
ci/Intel-compilation success Compilation OK
ci/iol-intel-Performance success Performance Testing PASS
ci/iol-abi-testing success Testing PASS
ci/github-robot success github build: passed
ci/travis-robot success travis build: passed
ci/checkpatch success coding style OK

Commit Message

Lance Richardson March 18, 2021, 7:52 p.m. UTC
When the mask field of an rte_flow pattern item is NULL,
the default mask for that item type should be used.

Fixes: 5ef3b79fdfe6 ("net/bnxt: support flow filter ops")
Cc: stable@dpdk.org

Signed-off-by: Lance Richardson <lance.richardson@broadcom.com>
---
 drivers/net/bnxt/bnxt_flow.c | 47 +++++++++++++++++++++++++++---------
 1 file changed, 36 insertions(+), 11 deletions(-)

Comments

Ajit Khaparde March 19, 2021, 3:51 a.m. UTC | #1
On Thu, Mar 18, 2021 at 12:52 PM Lance Richardson
<lance.richardson@broadcom.com> wrote:
>
> When the mask field of an rte_flow pattern item is NULL,
> the default mask for that item type should be used.
>
> Fixes: 5ef3b79fdfe6 ("net/bnxt: support flow filter ops")
> Cc: stable@dpdk.org
>
> Signed-off-by: Lance Richardson <lance.richardson@broadcom.com>
Patch applied to dpdk-next-net-brcm.

> ---
>  drivers/net/bnxt/bnxt_flow.c | 47 +++++++++++++++++++++++++++---------
>  1 file changed, 36 insertions(+), 11 deletions(-)
>
> diff --git a/drivers/net/bnxt/bnxt_flow.c b/drivers/net/bnxt/bnxt_flow.c
> index a8f5d91fc4..e3906b4779 100644
> --- a/drivers/net/bnxt/bnxt_flow.c
> +++ b/drivers/net/bnxt/bnxt_flow.c
> @@ -188,11 +188,15 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp,
>                                 PMD_DRV_LOG(DEBUG, "Parse inner header\n");
>                         break;
>                 case RTE_FLOW_ITEM_TYPE_ETH:
> -                       if (!item->spec || !item->mask)
> +                       if (!item->spec)
>                                 break;
>
>                         eth_spec = item->spec;
> -                       eth_mask = item->mask;
> +
> +                       if (item->mask)
> +                               eth_mask = item->mask;
> +                       else
> +                               eth_mask = &rte_flow_item_eth_mask;
>
>                         /* Source MAC address mask cannot be partially set.
>                          * Should be All 0's or all 1's.
> @@ -281,7 +285,12 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp,
>                         break;
>                 case RTE_FLOW_ITEM_TYPE_VLAN:
>                         vlan_spec = item->spec;
> -                       vlan_mask = item->mask;
> +
> +                       if (item->mask)
> +                               vlan_mask = item->mask;
> +                       else
> +                               vlan_mask = &rte_flow_item_vlan_mask;
> +
>                         if (en & en_ethertype) {
>                                 rte_flow_error_set(error, EINVAL,
>                                                    RTE_FLOW_ERROR_TYPE_ITEM,
> @@ -324,11 +333,15 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp,
>                 case RTE_FLOW_ITEM_TYPE_IPV4:
>                         /* If mask is not involved, we could use EM filters. */
>                         ipv4_spec = item->spec;
> -                       ipv4_mask = item->mask;
>
> -                       if (!item->spec || !item->mask)
> +                       if (!item->spec)
>                                 break;
>
> +                       if (item->mask)
> +                               ipv4_mask = item->mask;
> +                       else
> +                               ipv4_mask = &rte_flow_item_ipv4_mask;
> +
>                         /* Only IP DST and SRC fields are maskable. */
>                         if (ipv4_mask->hdr.version_ihl ||
>                             ipv4_mask->hdr.type_of_service ||
> @@ -385,11 +398,15 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp,
>                         break;
>                 case RTE_FLOW_ITEM_TYPE_IPV6:
>                         ipv6_spec = item->spec;
> -                       ipv6_mask = item->mask;
>
> -                       if (!item->spec || !item->mask)
> +                       if (!item->spec)
>                                 break;
>
> +                       if (item->mask)
> +                               ipv6_mask = item->mask;
> +                       else
> +                               ipv6_mask = &rte_flow_item_ipv6_mask;
> +
>                         /* Only IP DST and SRC fields are maskable. */
>                         if (ipv6_mask->hdr.vtc_flow ||
>                             ipv6_mask->hdr.payload_len ||
> @@ -437,11 +454,15 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp,
>                         break;
>                 case RTE_FLOW_ITEM_TYPE_TCP:
>                         tcp_spec = item->spec;
> -                       tcp_mask = item->mask;
>
> -                       if (!item->spec || !item->mask)
> +                       if (!item->spec)
>                                 break;
>
> +                       if (item->mask)
> +                               tcp_mask = item->mask;
> +                       else
> +                               tcp_mask = &rte_flow_item_tcp_mask;
> +
>                         /* Check TCP mask. Only DST & SRC ports are maskable */
>                         if (tcp_mask->hdr.sent_seq ||
>                             tcp_mask->hdr.recv_ack ||
> @@ -482,11 +503,15 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp,
>                         break;
>                 case RTE_FLOW_ITEM_TYPE_UDP:
>                         udp_spec = item->spec;
> -                       udp_mask = item->mask;
>
> -                       if (!item->spec || !item->mask)
> +                       if (!item->spec)
>                                 break;
>
> +                       if (item->mask)
> +                               udp_mask = item->mask;
> +                       else
> +                               udp_mask = &rte_flow_item_udp_mask;
> +
>                         if (udp_mask->hdr.dgram_len ||
>                             udp_mask->hdr.dgram_cksum) {
>                                 rte_flow_error_set(error,
> --
> 2.25.1
>
diff mbox series

Patch

diff --git a/drivers/net/bnxt/bnxt_flow.c b/drivers/net/bnxt/bnxt_flow.c
index a8f5d91fc4..e3906b4779 100644
--- a/drivers/net/bnxt/bnxt_flow.c
+++ b/drivers/net/bnxt/bnxt_flow.c
@@ -188,11 +188,15 @@  bnxt_validate_and_parse_flow_type(struct bnxt *bp,
 				PMD_DRV_LOG(DEBUG, "Parse inner header\n");
 			break;
 		case RTE_FLOW_ITEM_TYPE_ETH:
-			if (!item->spec || !item->mask)
+			if (!item->spec)
 				break;
 
 			eth_spec = item->spec;
-			eth_mask = item->mask;
+
+			if (item->mask)
+				eth_mask = item->mask;
+			else
+				eth_mask = &rte_flow_item_eth_mask;
 
 			/* Source MAC address mask cannot be partially set.
 			 * Should be All 0's or all 1's.
@@ -281,7 +285,12 @@  bnxt_validate_and_parse_flow_type(struct bnxt *bp,
 			break;
 		case RTE_FLOW_ITEM_TYPE_VLAN:
 			vlan_spec = item->spec;
-			vlan_mask = item->mask;
+
+			if (item->mask)
+				vlan_mask = item->mask;
+			else
+				vlan_mask = &rte_flow_item_vlan_mask;
+
 			if (en & en_ethertype) {
 				rte_flow_error_set(error, EINVAL,
 						   RTE_FLOW_ERROR_TYPE_ITEM,
@@ -324,11 +333,15 @@  bnxt_validate_and_parse_flow_type(struct bnxt *bp,
 		case RTE_FLOW_ITEM_TYPE_IPV4:
 			/* If mask is not involved, we could use EM filters. */
 			ipv4_spec = item->spec;
-			ipv4_mask = item->mask;
 
-			if (!item->spec || !item->mask)
+			if (!item->spec)
 				break;
 
+			if (item->mask)
+				ipv4_mask = item->mask;
+			else
+				ipv4_mask = &rte_flow_item_ipv4_mask;
+
 			/* Only IP DST and SRC fields are maskable. */
 			if (ipv4_mask->hdr.version_ihl ||
 			    ipv4_mask->hdr.type_of_service ||
@@ -385,11 +398,15 @@  bnxt_validate_and_parse_flow_type(struct bnxt *bp,
 			break;
 		case RTE_FLOW_ITEM_TYPE_IPV6:
 			ipv6_spec = item->spec;
-			ipv6_mask = item->mask;
 
-			if (!item->spec || !item->mask)
+			if (!item->spec)
 				break;
 
+			if (item->mask)
+				ipv6_mask = item->mask;
+			else
+				ipv6_mask = &rte_flow_item_ipv6_mask;
+
 			/* Only IP DST and SRC fields are maskable. */
 			if (ipv6_mask->hdr.vtc_flow ||
 			    ipv6_mask->hdr.payload_len ||
@@ -437,11 +454,15 @@  bnxt_validate_and_parse_flow_type(struct bnxt *bp,
 			break;
 		case RTE_FLOW_ITEM_TYPE_TCP:
 			tcp_spec = item->spec;
-			tcp_mask = item->mask;
 
-			if (!item->spec || !item->mask)
+			if (!item->spec)
 				break;
 
+			if (item->mask)
+				tcp_mask = item->mask;
+			else
+				tcp_mask = &rte_flow_item_tcp_mask;
+
 			/* Check TCP mask. Only DST & SRC ports are maskable */
 			if (tcp_mask->hdr.sent_seq ||
 			    tcp_mask->hdr.recv_ack ||
@@ -482,11 +503,15 @@  bnxt_validate_and_parse_flow_type(struct bnxt *bp,
 			break;
 		case RTE_FLOW_ITEM_TYPE_UDP:
 			udp_spec = item->spec;
-			udp_mask = item->mask;
 
-			if (!item->spec || !item->mask)
+			if (!item->spec)
 				break;
 
+			if (item->mask)
+				udp_mask = item->mask;
+			else
+				udp_mask = &rte_flow_item_udp_mask;
+
 			if (udp_mask->hdr.dgram_len ||
 			    udp_mask->hdr.dgram_cksum) {
 				rte_flow_error_set(error,