net/mlx5: do not poll CQEs when no available elts

Message ID 20241206005811.948293-1-gahu@nvidia.com (mailing list archive)
State New
Delegated to: Raslan Darawsheh
Headers
Series net/mlx5: do not poll CQEs when no available elts |

Checks

Context Check Description
ci/checkpatch warning coding style issues
ci/loongarch-compilation success Compilation OK
ci/Intel-compilation success Compilation OK
ci/loongarch-unit-testing success Unit Testing PASS
ci/intel-Testing success Testing PASS
ci/iol-sample-apps-testing pending Testing pending
ci/intel-Functional success Functional PASS
ci/iol-intel-Performance success Performance Testing PASS
ci/iol-unit-amd64-testing success Testing PASS
ci/iol-mellanox-Performance success Performance Testing PASS
ci/iol-intel-Functional success Functional Testing PASS
ci/github-robot: build success github build: passed
ci/iol-unit-arm64-testing success Testing PASS
ci/iol-compile-arm64-testing success Testing PASS
ci/iol-compile-amd64-testing warning Testing issues

Commit Message

Gavin Hu Dec. 6, 2024, 12:58 a.m. UTC
In certain situations, the receive queue (rxq) fails to replenish its
internal ring with memory buffers (mbufs) from the pool. This can happen
when the pool has a limited number of mbufs allocated, and the user
application holds incoming packets for an extended period, resulting in a
delayed release of mbufs. Consequently, the pool becomes depleted,
preventing the rxq from replenishing from it.

There was a bug in the behavior of the vectorized rxq_cq_process_v routine,
which handled completion queue entries (CQEs) in batches of four. This
routine consistently accessed four mbufs from the internal queue ring,
regardless of whether they had been replenished. As a result, it could
access mbufs that no longer belonged to the poll mode driver (PMD).

The fix involves checking if there are four replenished mbufs available
before allowing rxq_cq_process_v to handle the batch. Once replenishment
succeeds during the polling process, the routine will resume its operation.

Fixes: 1ded26239aa0 ("net/mlx5: refactor vectorized Rx")
Cc: stable@dpdk.org

Reported-by: Changqi Dingluo <dingluochangqi.ck@bytedance.com>
Signed-off-by: Gavin Hu <gahu@nvidia.com>
---
 drivers/net/mlx5/mlx5_rxtx_vec.c | 3 +++
 1 file changed, 3 insertions(+)
  

Comments

Slava Ovsiienko Dec. 9, 2024, 8:10 a.m. UTC | #1
Acked-by: Viacheslav Ovsiienko <viacheslavo@nvidia.com>

> -----Original Message-----
> From: Gavin Hu <gahu@nvidia.com>
> Sent: Friday, December 6, 2024 2:58 AM
> To: dev@dpdk.org
> Cc: stable@dpdk.org; Dariusz Sosnowski <dsosnowski@nvidia.com>; Slava
> Ovsiienko <viacheslavo@nvidia.com>; Bing Zhao <bingz@nvidia.com>; Ori
> Kam <orika@nvidia.com>; Suanming Mou <suanmingm@nvidia.com>; Matan
> Azrad <matan@nvidia.com>; Alexander Kozyrev <akozyrev@nvidia.com>
> Subject: [PATCH] net/mlx5: do not poll CQEs when no available elts
> 
> In certain situations, the receive queue (rxq) fails to replenish its internal ring
> with memory buffers (mbufs) from the pool. This can happen when the pool
> has a limited number of mbufs allocated, and the user application holds
> incoming packets for an extended period, resulting in a delayed release of
> mbufs. Consequently, the pool becomes depleted, preventing the rxq from
> replenishing from it.
> 
> There was a bug in the behavior of the vectorized rxq_cq_process_v routine,
> which handled completion queue entries (CQEs) in batches of four. This
> routine consistently accessed four mbufs from the internal queue ring,
> regardless of whether they had been replenished. As a result, it could access
> mbufs that no longer belonged to the poll mode driver (PMD).
> 
> The fix involves checking if there are four replenished mbufs available before
> allowing rxq_cq_process_v to handle the batch. Once replenishment succeeds
> during the polling process, the routine will resume its operation.
> 
> Fixes: 1ded26239aa0 ("net/mlx5: refactor vectorized Rx")
> Cc: stable@dpdk.org
> 
> Reported-by: Changqi Dingluo <dingluochangqi.ck@bytedance.com>
> Signed-off-by: Gavin Hu <gahu@nvidia.com>
> ---
>  drivers/net/mlx5/mlx5_rxtx_vec.c | 3 +++
>  1 file changed, 3 insertions(+)
> 
> diff --git a/drivers/net/mlx5/mlx5_rxtx_vec.c
> b/drivers/net/mlx5/mlx5_rxtx_vec.c
> index 1872bf310c..1b701801c5 100644
> --- a/drivers/net/mlx5/mlx5_rxtx_vec.c
> +++ b/drivers/net/mlx5/mlx5_rxtx_vec.c
> @@ -325,6 +325,9 @@ rxq_burst_v(struct mlx5_rxq_data *rxq, struct
> rte_mbuf **pkts,
>  	/* Not to cross queue end. */
>  	pkts_n = RTE_MIN(pkts_n, q_n - elts_idx);
>  	pkts_n = RTE_MIN(pkts_n, q_n - cq_idx);
> +	/* Not to move past the allocated mbufs. */
> +	pkts_n = RTE_MIN(pkts_n, RTE_ALIGN_FLOOR(rxq->rq_ci - rxq-
> >rq_pi,
> +
> 	MLX5_VPMD_DESCS_PER_LOOP));
>  	if (!pkts_n) {
>  		*no_cq = !rcvd_pkt;
>  		return rcvd_pkt;
> --
> 2.18.2
  

Patch

diff --git a/drivers/net/mlx5/mlx5_rxtx_vec.c b/drivers/net/mlx5/mlx5_rxtx_vec.c
index 1872bf310c..1b701801c5 100644
--- a/drivers/net/mlx5/mlx5_rxtx_vec.c
+++ b/drivers/net/mlx5/mlx5_rxtx_vec.c
@@ -325,6 +325,9 @@  rxq_burst_v(struct mlx5_rxq_data *rxq, struct rte_mbuf **pkts,
 	/* Not to cross queue end. */
 	pkts_n = RTE_MIN(pkts_n, q_n - elts_idx);
 	pkts_n = RTE_MIN(pkts_n, q_n - cq_idx);
+	/* Not to move past the allocated mbufs. */
+	pkts_n = RTE_MIN(pkts_n, RTE_ALIGN_FLOOR(rxq->rq_ci - rxq->rq_pi,
+						MLX5_VPMD_DESCS_PER_LOOP));
 	if (!pkts_n) {
 		*no_cq = !rcvd_pkt;
 		return rcvd_pkt;