From patchwork Wed Mar 20 11:47:21 2019 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: David Marchand X-Patchwork-Id: 51388 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@dpdk.org Delivered-To: patchwork@dpdk.org Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 79BE25F1B; Wed, 20 Mar 2019 12:47:51 +0100 (CET) Received: from mx1.redhat.com (mx1.redhat.com [209.132.183.28]) by dpdk.org (Postfix) with ESMTP id E50FB5B2C; Wed, 20 Mar 2019 12:47:46 +0100 (CET) Received: from smtp.corp.redhat.com (int-mx07.intmail.prod.int.phx2.redhat.com [10.5.11.22]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 55BD93168919; Wed, 20 Mar 2019 11:47:46 +0000 (UTC) Received: from dmarchan.remote.csb (ovpn-204-129.brq.redhat.com [10.40.204.129]) by smtp.corp.redhat.com (Postfix) with ESMTP id 008E21001E71; Wed, 20 Mar 2019 11:47:44 +0000 (UTC) From: David Marchand To: dev@dpdk.org Cc: chas3@att.com, stable@dpdk.org Date: Wed, 20 Mar 2019 12:47:21 +0100 Message-Id: <1553082442-18850-5-git-send-email-david.marchand@redhat.com> In-Reply-To: <1553082442-18850-1-git-send-email-david.marchand@redhat.com> References: <1553082442-18850-1-git-send-email-david.marchand@redhat.com> X-Scanned-By: MIMEDefang 2.84 on 10.5.11.22 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.41]); Wed, 20 Mar 2019 11:47:46 +0000 (UTC) Subject: [dpdk-dev] [PATCH 4/5] net/bonding: fix incorrect packet count type for lacp X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Caught by code review, those variables are supposed to be on 16bits to avoid endless loops in the (unlikely?) case where the application asks for receiving more than 256 packets and the accumulated num_rx_total count reaches 256: uint16_t num_rx_total = 0; uint8_t j, k; j = num_rx_total; num_rx_total += rte_eth_rx_burst(); for (k = j; k < 2 && k < num_rx_total; k++) rte_prefetch0(rte_pktmbuf_mtod(bufs[k], void *)); while (j < num_rx_total) { j++; } Fixes: 46fb43683679 ("bond: add mode 4") Cc: stable@dpdk.org Signed-off-by: David Marchand --- drivers/net/bonding/rte_eth_bond_pmd.c | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/drivers/net/bonding/rte_eth_bond_pmd.c b/drivers/net/bonding/rte_eth_bond_pmd.c index bbe2568..ae66a70 100644 --- a/drivers/net/bonding/rte_eth_bond_pmd.c +++ b/drivers/net/bonding/rte_eth_bond_pmd.c @@ -405,8 +405,9 @@ uint8_t collecting; /* current slave collecting status */ const uint8_t promisc = internals->promiscuous_en; uint8_t subtype; - uint8_t j, k; uint16_t i; + uint16_t j; + uint16_t k; /* Copy slave list to protect against slave up/down changes during tx * bursting */