From patchwork Thu Feb 4 16:51:03 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ferruh Yigit X-Patchwork-Id: 87752 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id AAFFFA0524; Thu, 4 Feb 2021 17:51:14 +0100 (CET) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 88D752406D6; Thu, 4 Feb 2021 17:51:14 +0100 (CET) Received: from mga17.intel.com (mga17.intel.com [192.55.52.151]) by mails.dpdk.org (Postfix) with ESMTP id 23F7A2406C5; Thu, 4 Feb 2021 17:51:11 +0100 (CET) IronPort-SDR: wH0J9/Hl4fvAn9hLTR0iHSPtpdUybRWpDE4XlugnNh7cKU7wIIKiYQyLLwy2ggaNrWahtx/haz h9afLmnQCZnw== X-IronPort-AV: E=McAfee;i="6000,8403,9885"; a="161035695" X-IronPort-AV: E=Sophos;i="5.79,401,1602572400"; d="scan'208";a="161035695" Received: from fmsmga007.fm.intel.com ([10.253.24.52]) by fmsmga107.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 04 Feb 2021 08:51:09 -0800 IronPort-SDR: KkRIZF+7ky0eZr9zmxNp2ZhFzNyxO055ap4K9AYK3LBa1r6eORqbd769//zIgOwkCUZdkcezuW S4iRZTwGjdMQ== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.81,401,1610438400"; d="scan'208";a="356754821" Received: from silpixa00399752.ir.intel.com (HELO silpixa00399752.ger.corp.intel.com) ([10.237.222.27]) by fmsmga007.fm.intel.com with ESMTP; 04 Feb 2021 08:51:07 -0800 From: Ferruh Yigit To: Cian Ferriter Cc: Ferruh Yigit , dev@dpdk.org, stable@dpdk.org Date: Thu, 4 Feb 2021 16:51:03 +0000 Message-Id: <20210204165103.2355136-1-ferruh.yigit@intel.com> X-Mailer: git-send-email 2.29.2 In-Reply-To: <20210203154920.2449179-1-ferruh.yigit@intel.com> References: <20210203154920.2449179-1-ferruh.yigit@intel.com> MIME-Version: 1.0 Subject: [dpdk-dev] [PATCH v2] net/pcap: fix infinite Rx with large files X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Packet forwarding is not working when infinite Rx feature is used with large .pcap files that has high number of packets. The problem is number of allocated mbufs are less than the infinite Rx ring size, and all mbufs consumed to fill the ring, so there is no mbuf left for forwarding. Current logic can not detect that infinite Rx ring is not filled completely and no more mbufs left, and setup continues which leads silent fail on packet forwarding. There isn't much can be done when there is not enough mbuf for the given .pcap file, so additional checks added to detect the case and fail explicitly with an error log. Bugzilla ID: 595 Fixes: a3f5252e5cbd ("net/pcap: enable infinitely Rx a pcap file") Cc: stable@dpdk.org Reported-by: Cian Ferriter Signed-off-by: Ferruh Yigit Acked-by: Cian Ferriter --- v2: * Updated log message --- drivers/net/pcap/rte_eth_pcap.c | 40 ++++++++++++++++++++------------- 1 file changed, 25 insertions(+), 15 deletions(-) diff --git a/drivers/net/pcap/rte_eth_pcap.c b/drivers/net/pcap/rte_eth_pcap.c index c7751b7ba742..90f5d75ea87f 100644 --- a/drivers/net/pcap/rte_eth_pcap.c +++ b/drivers/net/pcap/rte_eth_pcap.c @@ -735,6 +735,17 @@ eth_stats_reset(struct rte_eth_dev *dev) return 0; } +static inline void +infinite_rx_ring_free(struct rte_ring *pkts) +{ + struct rte_mbuf *bufs; + + while (!rte_ring_dequeue(pkts, (void **)&bufs)) + rte_pktmbuf_free(bufs); + + rte_ring_free(pkts); +} + static int eth_dev_close(struct rte_eth_dev *dev) { @@ -753,7 +764,6 @@ eth_dev_close(struct rte_eth_dev *dev) if (internals->infinite_rx) { for (i = 0; i < dev->data->nb_rx_queues; i++) { struct pcap_rx_queue *pcap_q = &internals->rx_queue[i]; - struct rte_mbuf *pcap_buf; /* * 'pcap_q->pkts' can be NULL if 'eth_dev_close()' @@ -762,11 +772,7 @@ eth_dev_close(struct rte_eth_dev *dev) if (pcap_q->pkts == NULL) continue; - while (!rte_ring_dequeue(pcap_q->pkts, - (void **)&pcap_buf)) - rte_pktmbuf_free(pcap_buf); - - rte_ring_free(pcap_q->pkts); + infinite_rx_ring_free(pcap_q->pkts); } } @@ -835,21 +841,25 @@ eth_rx_queue_setup(struct rte_eth_dev *dev, while (eth_pcap_rx(pcap_q, bufs, 1)) { /* Check for multiseg mbufs. */ if (bufs[0]->nb_segs != 1) { - rte_pktmbuf_free(*bufs); - - while (!rte_ring_dequeue(pcap_q->pkts, - (void **)bufs)) - rte_pktmbuf_free(*bufs); - - rte_ring_free(pcap_q->pkts); - PMD_LOG(ERR, "Multiseg mbufs are not supported in infinite_rx " - "mode."); + infinite_rx_ring_free(pcap_q->pkts); + PMD_LOG(ERR, + "Multiseg mbufs are not supported in infinite_rx mode."); return -EINVAL; } rte_ring_enqueue_bulk(pcap_q->pkts, (void * const *)bufs, 1, NULL); } + + if (rte_ring_count(pcap_q->pkts) < pcap_pkt_count) { + infinite_rx_ring_free(pcap_q->pkts); + PMD_LOG(ERR, + "Not enough mbufs to accommodate packets in pcap file. " + "At least %" PRIu64 " mbufs per queue is required.", + pcap_pkt_count); + return -EINVAL; + } + /* * Reset the stats for this queue since eth_pcap_rx calls above * didn't result in the application receiving packets.