net/octeontx2: perf improvement to rx vector func

Message ID 20191210120844.50017-1-ndabilpuram@marvell.com (mailing list archive)
State Accepted, archived
Delegated to: Jerin Jacob
Headers
Series net/octeontx2: perf improvement to rx vector func |

Checks

Context Check Description
ci/checkpatch success coding style OK
ci/Intel-compilation success Compilation OK
ci/travis-robot warning Travis build: failed
ci/iol-testing success Testing PASS
ci/iol-intel-Performance success Performance Testing PASS
ci/iol-mellanox-Performance success Performance Testing PASS

Commit Message

Nithin Dabilpuram Dec. 10, 2019, 12:08 p.m. UTC
  From: Jerin Jacob <jerinj@marvell.com>

Use scalar loads instead of vector loads for fields
that don't need any vector operations.

Signed-off-by: Andrew Pinski <apinski@marvell.com>
Signed-off-by: Jerin Jacob <jerinj@marvell.com>
Signed-off-by: Nithin Dabilpuram <ndabilpuram@marvell.com>
---
 drivers/net/octeontx2/otx2_rx.c | 48 ++++++++++++++++++++---------------------
 1 file changed, 24 insertions(+), 24 deletions(-)
  

Comments

Jerin Jacob Jan. 13, 2020, 7:40 a.m. UTC | #1
On Tue, Dec 10, 2019 at 5:39 PM Nithin Dabilpuram
<ndabilpuram@marvell.com> wrote:
>
> From: Jerin Jacob <jerinj@marvell.com>
>
> Use scalar loads instead of vector loads for fields
> that don't need any vector operations.
>
> Signed-off-by: Andrew Pinski <apinski@marvell.com>
> Signed-off-by: Jerin Jacob <jerinj@marvell.com>
> Signed-off-by: Nithin Dabilpuram <ndabilpuram@marvell.com>

Applied to dpdk-next-net-mrvl/master. Thanks



> ---
>  drivers/net/octeontx2/otx2_rx.c | 48 ++++++++++++++++++++---------------------
>  1 file changed, 24 insertions(+), 24 deletions(-)
>
> diff --git a/drivers/net/octeontx2/otx2_rx.c b/drivers/net/octeontx2/otx2_rx.c
> index 48565db..db4a221 100644
> --- a/drivers/net/octeontx2/otx2_rx.c
> +++ b/drivers/net/octeontx2/otx2_rx.c
> @@ -184,17 +184,21 @@ nix_recv_pkts_vector(void *rx_queue, struct rte_mbuf **rx_pkts,
>                 f3 = vqtbl1q_u8(cq3_w8, shuf_msk);
>
>                 /* Load CQE word0 and word 1 */
> -               uint64x2_t cq0_w0 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(0)));
> -               uint64x2_t cq1_w0 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(1)));
> -               uint64x2_t cq2_w0 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(2)));
> -               uint64x2_t cq3_w0 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(3)));
> +               uint64_t cq0_w0 = ((uint64_t *)(cq0 + CQE_SZ(0)))[0];
> +               uint64_t cq0_w1 = ((uint64_t *)(cq0 + CQE_SZ(0)))[1];
> +               uint64_t cq1_w0 = ((uint64_t *)(cq0 + CQE_SZ(1)))[0];
> +               uint64_t cq1_w1 = ((uint64_t *)(cq0 + CQE_SZ(1)))[1];
> +               uint64_t cq2_w0 = ((uint64_t *)(cq0 + CQE_SZ(2)))[0];
> +               uint64_t cq2_w1 = ((uint64_t *)(cq0 + CQE_SZ(2)))[1];
> +               uint64_t cq3_w0 = ((uint64_t *)(cq0 + CQE_SZ(3)))[0];
> +               uint64_t cq3_w1 = ((uint64_t *)(cq0 + CQE_SZ(3)))[1];
>
>                 if (flags & NIX_RX_OFFLOAD_RSS_F) {
>                         /* Fill rss in the rx_descriptor_fields1 */
> -                       f0 = vsetq_lane_u32(vgetq_lane_u32(cq0_w0, 0), f0, 3);
> -                       f1 = vsetq_lane_u32(vgetq_lane_u32(cq1_w0, 0), f1, 3);
> -                       f2 = vsetq_lane_u32(vgetq_lane_u32(cq2_w0, 0), f2, 3);
> -                       f3 = vsetq_lane_u32(vgetq_lane_u32(cq3_w0, 0), f3, 3);
> +                       f0 = vsetq_lane_u32(cq0_w0, f0, 3);
> +                       f1 = vsetq_lane_u32(cq1_w0, f1, 3);
> +                       f2 = vsetq_lane_u32(cq2_w0, f2, 3);
> +                       f3 = vsetq_lane_u32(cq3_w0, f3, 3);
>                         ol_flags0 = PKT_RX_RSS_HASH;
>                         ol_flags1 = PKT_RX_RSS_HASH;
>                         ol_flags2 = PKT_RX_RSS_HASH;
> @@ -206,25 +210,21 @@ nix_recv_pkts_vector(void *rx_queue, struct rte_mbuf **rx_pkts,
>
>                 if (flags & NIX_RX_OFFLOAD_PTYPE_F) {
>                         /* Fill packet_type in the rx_descriptor_fields1 */
> -                       f0 = vsetq_lane_u32(nix_ptype_get(lookup_mem,
> -                                           vgetq_lane_u64(cq0_w0, 1)), f0, 0);
> -                       f1 = vsetq_lane_u32(nix_ptype_get(lookup_mem,
> -                                           vgetq_lane_u64(cq1_w0, 1)), f1, 0);
> -                       f2 = vsetq_lane_u32(nix_ptype_get(lookup_mem,
> -                                           vgetq_lane_u64(cq2_w0, 1)), f2, 0);
> -                       f3 = vsetq_lane_u32(nix_ptype_get(lookup_mem,
> -                                           vgetq_lane_u64(cq3_w0, 1)), f3, 0);
> +                       f0 = vsetq_lane_u32(nix_ptype_get(lookup_mem, cq0_w1),
> +                                           f0, 0);
> +                       f1 = vsetq_lane_u32(nix_ptype_get(lookup_mem, cq1_w1),
> +                                           f1, 0);
> +                       f2 = vsetq_lane_u32(nix_ptype_get(lookup_mem, cq2_w1),
> +                                           f2, 0);
> +                       f3 = vsetq_lane_u32(nix_ptype_get(lookup_mem, cq3_w1),
> +                                           f3, 0);
>                 }
>
>                 if (flags & NIX_RX_OFFLOAD_CHECKSUM_F) {
> -                       ol_flags0 |= nix_rx_olflags_get(lookup_mem,
> -                                               vgetq_lane_u64(cq0_w0, 1));
> -                       ol_flags1 |= nix_rx_olflags_get(lookup_mem,
> -                                               vgetq_lane_u64(cq1_w0, 1));
> -                       ol_flags2 |= nix_rx_olflags_get(lookup_mem,
> -                                               vgetq_lane_u64(cq2_w0, 1));
> -                       ol_flags3 |= nix_rx_olflags_get(lookup_mem,
> -                                               vgetq_lane_u64(cq3_w0, 1));
> +                       ol_flags0 |= nix_rx_olflags_get(lookup_mem, cq0_w1);
> +                       ol_flags1 |= nix_rx_olflags_get(lookup_mem, cq1_w1);
> +                       ol_flags2 |= nix_rx_olflags_get(lookup_mem, cq2_w1);
> +                       ol_flags3 |= nix_rx_olflags_get(lookup_mem, cq3_w1);
>                 }
>
>                 if (flags & NIX_RX_OFFLOAD_VLAN_STRIP_F) {
> --
> 2.8.4
>
  

Patch

diff --git a/drivers/net/octeontx2/otx2_rx.c b/drivers/net/octeontx2/otx2_rx.c
index 48565db..db4a221 100644
--- a/drivers/net/octeontx2/otx2_rx.c
+++ b/drivers/net/octeontx2/otx2_rx.c
@@ -184,17 +184,21 @@  nix_recv_pkts_vector(void *rx_queue, struct rte_mbuf **rx_pkts,
 		f3 = vqtbl1q_u8(cq3_w8, shuf_msk);
 
 		/* Load CQE word0 and word 1 */
-		uint64x2_t cq0_w0 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(0)));
-		uint64x2_t cq1_w0 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(1)));
-		uint64x2_t cq2_w0 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(2)));
-		uint64x2_t cq3_w0 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(3)));
+		uint64_t cq0_w0 = ((uint64_t *)(cq0 + CQE_SZ(0)))[0];
+		uint64_t cq0_w1 = ((uint64_t *)(cq0 + CQE_SZ(0)))[1];
+		uint64_t cq1_w0 = ((uint64_t *)(cq0 + CQE_SZ(1)))[0];
+		uint64_t cq1_w1 = ((uint64_t *)(cq0 + CQE_SZ(1)))[1];
+		uint64_t cq2_w0 = ((uint64_t *)(cq0 + CQE_SZ(2)))[0];
+		uint64_t cq2_w1 = ((uint64_t *)(cq0 + CQE_SZ(2)))[1];
+		uint64_t cq3_w0 = ((uint64_t *)(cq0 + CQE_SZ(3)))[0];
+		uint64_t cq3_w1 = ((uint64_t *)(cq0 + CQE_SZ(3)))[1];
 
 		if (flags & NIX_RX_OFFLOAD_RSS_F) {
 			/* Fill rss in the rx_descriptor_fields1 */
-			f0 = vsetq_lane_u32(vgetq_lane_u32(cq0_w0, 0), f0, 3);
-			f1 = vsetq_lane_u32(vgetq_lane_u32(cq1_w0, 0), f1, 3);
-			f2 = vsetq_lane_u32(vgetq_lane_u32(cq2_w0, 0), f2, 3);
-			f3 = vsetq_lane_u32(vgetq_lane_u32(cq3_w0, 0), f3, 3);
+			f0 = vsetq_lane_u32(cq0_w0, f0, 3);
+			f1 = vsetq_lane_u32(cq1_w0, f1, 3);
+			f2 = vsetq_lane_u32(cq2_w0, f2, 3);
+			f3 = vsetq_lane_u32(cq3_w0, f3, 3);
 			ol_flags0 = PKT_RX_RSS_HASH;
 			ol_flags1 = PKT_RX_RSS_HASH;
 			ol_flags2 = PKT_RX_RSS_HASH;
@@ -206,25 +210,21 @@  nix_recv_pkts_vector(void *rx_queue, struct rte_mbuf **rx_pkts,
 
 		if (flags & NIX_RX_OFFLOAD_PTYPE_F) {
 			/* Fill packet_type in the rx_descriptor_fields1 */
-			f0 = vsetq_lane_u32(nix_ptype_get(lookup_mem,
-					    vgetq_lane_u64(cq0_w0, 1)), f0, 0);
-			f1 = vsetq_lane_u32(nix_ptype_get(lookup_mem,
-					    vgetq_lane_u64(cq1_w0, 1)), f1, 0);
-			f2 = vsetq_lane_u32(nix_ptype_get(lookup_mem,
-					    vgetq_lane_u64(cq2_w0, 1)), f2, 0);
-			f3 = vsetq_lane_u32(nix_ptype_get(lookup_mem,
-					    vgetq_lane_u64(cq3_w0, 1)), f3, 0);
+			f0 = vsetq_lane_u32(nix_ptype_get(lookup_mem, cq0_w1),
+					    f0, 0);
+			f1 = vsetq_lane_u32(nix_ptype_get(lookup_mem, cq1_w1),
+					    f1, 0);
+			f2 = vsetq_lane_u32(nix_ptype_get(lookup_mem, cq2_w1),
+					    f2, 0);
+			f3 = vsetq_lane_u32(nix_ptype_get(lookup_mem, cq3_w1),
+					    f3, 0);
 		}
 
 		if (flags & NIX_RX_OFFLOAD_CHECKSUM_F) {
-			ol_flags0 |= nix_rx_olflags_get(lookup_mem,
-						vgetq_lane_u64(cq0_w0, 1));
-			ol_flags1 |= nix_rx_olflags_get(lookup_mem,
-						vgetq_lane_u64(cq1_w0, 1));
-			ol_flags2 |= nix_rx_olflags_get(lookup_mem,
-						vgetq_lane_u64(cq2_w0, 1));
-			ol_flags3 |= nix_rx_olflags_get(lookup_mem,
-						vgetq_lane_u64(cq3_w0, 1));
+			ol_flags0 |= nix_rx_olflags_get(lookup_mem, cq0_w1);
+			ol_flags1 |= nix_rx_olflags_get(lookup_mem, cq1_w1);
+			ol_flags2 |= nix_rx_olflags_get(lookup_mem, cq2_w1);
+			ol_flags3 |= nix_rx_olflags_get(lookup_mem, cq3_w1);
 		}
 
 		if (flags & NIX_RX_OFFLOAD_VLAN_STRIP_F) {