From patchwork Mon May 13 18:52:17 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Stephen Hemminger X-Patchwork-Id: 140046 X-Patchwork-Delegate: thomas@monjalon.net Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id B175D4401D; Mon, 13 May 2024 20:55:48 +0200 (CEST) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id A456A40A76; Mon, 13 May 2024 20:55:09 +0200 (CEST) Received: from mail-oo1-f44.google.com (mail-oo1-f44.google.com [209.85.161.44]) by mails.dpdk.org (Postfix) with ESMTP id 035A84067C for ; Mon, 13 May 2024 20:55:05 +0200 (CEST) Received: by mail-oo1-f44.google.com with SMTP id 006d021491bc7-5b2761611e8so2731614eaf.2 for ; Mon, 13 May 2024 11:55:04 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=networkplumber-org.20230601.gappssmtp.com; s=20230601; t=1715626504; x=1716231304; darn=dpdk.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=NO4DVKUdt+v+339Dm/0NF9uOVdDeR7qGw415YuijwCU=; b=zMzIT++lEq2OTXAJbjJvIKPC38AWwj28+u7ZbHulkxenHfmPyQZ05ENcxylkAxgXJW Xk+hLJV0SePpjLe/C7ZdwxMmKniAhZu3qIlrb9qsJ8zehoAoEFu6YKxfI5ec55qxZuYB uT8PB6ZG7i4C8YkID4prPLmj0zrd5qP4Vlj3qdAntPm5yaSVI2W8lDPbmvq4k/mMJrrO IeEKG+yyVgRFmEX+CqcqvgfLHrzzuxFCyAClxL8Vu5a+O0BweCoo7hq1RY+aG4bI6RId UXK/eypSM9yVLWba7HC1JMWiQgXLMrQRvC9HXg1p8vkKUCEOJWXXQGnbwCHVIv6DfDGu jYdA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1715626504; x=1716231304; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=NO4DVKUdt+v+339Dm/0NF9uOVdDeR7qGw415YuijwCU=; b=P0lBPwY2w2i0yxU6Xz5huc7oh4FXoW4PByHsamf9CnD1gWynsLTnqZSqtIbAjZ9Z7g 8xdiQBnMsGK1ggBYOYqiNHCZpr+zX33MAT7+shTQIiwWdBefWZYbfVfup07qPV0xNErz y/eJH9dyq4EQTnthZN6pFTeeL0/W4HOHARNF89PDwgzUrf/CndUxg/SG06HgKMfa0qZ/ duOutMsyNqATVgCkRaM+VIMJv5P9Vrh9WbQ4gTTYi3ozAjy7keHZ1t7sySwAcQTkd2Xm gVTkqSsfO9as2tswUBxSgri/YjjV5w8MNlVhybqWmsmYbFsjc6D9fo1kBmThA75lJTma aXaA== X-Gm-Message-State: AOJu0YzM1E6zZ4pTQxKjK5QXrgrdf6eryti2A7KLmqHQjSp7OvRNwEz4 xVykVJTf7wP1syfGeZOiSJ7J2WI/5XEjjaZA0U4TNuiuK/9fxmVIEL9xuekhk3P43H3CVx4C3K1 8ftohFg== X-Google-Smtp-Source: AGHT+IGLB9jRvCFk9WOaps3IU2RQMYqIkzzs/+NipiQ3rib2soW6VDjrh0RECnnGrwIWzuhGE2enmw== X-Received: by 2002:a05:6359:a3a5:b0:18d:e328:7e7f with SMTP id e5c5f4694b2df-193bd000215mr1146584255d.23.1715626504282; Mon, 13 May 2024 11:55:04 -0700 (PDT) Received: from hermes.local (204-195-96-226.wavecable.com. [204.195.96.226]) by smtp.gmail.com with ESMTPSA id 41be03b00d2f7-634103f7237sm8154680a12.71.2024.05.13.11.55.03 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 13 May 2024 11:55:03 -0700 (PDT) From: Stephen Hemminger To: dev@dpdk.org Cc: Stephen Hemminger , Bruce Richardson Subject: [RFC v2 7/7] net/ring: use generic SW stats Date: Mon, 13 May 2024 11:52:17 -0700 Message-ID: <20240513185448.120356-8-stephen@networkplumber.org> X-Mailer: git-send-email 2.43.0 In-Reply-To: <20240513185448.120356-1-stephen@networkplumber.org> References: <20240510050507.14381-1-stephen@networkplumber.org> <20240513185448.120356-1-stephen@networkplumber.org> MIME-Version: 1.0 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Use generic per-queue infrastructure. Signed-off-by: Stephen Hemminger --- drivers/net/ring/rte_eth_ring.c | 85 +++++++++++++++++---------------- 1 file changed, 44 insertions(+), 41 deletions(-) diff --git a/drivers/net/ring/rte_eth_ring.c b/drivers/net/ring/rte_eth_ring.c index 48953dd7a0..550b927392 100644 --- a/drivers/net/ring/rte_eth_ring.c +++ b/drivers/net/ring/rte_eth_ring.c @@ -7,6 +7,7 @@ #include "rte_eth_ring.h" #include #include +#include #include #include #include @@ -44,8 +45,8 @@ enum dev_action { struct ring_queue { struct rte_ring *rng; - uint64_t rx_pkts; - uint64_t tx_pkts; + + struct rte_eth_counters stats; }; struct pmd_internals { @@ -77,12 +78,13 @@ eth_ring_rx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs) { void **ptrs = (void *)&bufs[0]; struct ring_queue *r = q; - const uint16_t nb_rx = (uint16_t)rte_ring_dequeue_burst(r->rng, - ptrs, nb_bufs, NULL); - if (r->rng->flags & RING_F_SC_DEQ) - r->rx_pkts += nb_rx; - else - __atomic_fetch_add(&r->rx_pkts, nb_rx, __ATOMIC_RELAXED); + uint16_t i, nb_rx; + + nb_rx = (uint16_t)rte_ring_dequeue_burst(r->rng, ptrs, nb_bufs, NULL); + + for (i = 0; i < nb_rx; i++) + rte_eth_count_mbuf(&r->stats, bufs[i]); + return nb_rx; } @@ -90,13 +92,20 @@ static uint16_t eth_ring_tx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs) { void **ptrs = (void *)&bufs[0]; + uint32_t *sizes; struct ring_queue *r = q; - const uint16_t nb_tx = (uint16_t)rte_ring_enqueue_burst(r->rng, - ptrs, nb_bufs, NULL); - if (r->rng->flags & RING_F_SP_ENQ) - r->tx_pkts += nb_tx; - else - __atomic_fetch_add(&r->tx_pkts, nb_tx, __ATOMIC_RELAXED); + uint16_t i, nb_tx; + + sizes = alloca(sizeof(uint32_t) * nb_bufs); + + for (i = 0; i < nb_bufs; i++) + sizes[i] = rte_pktmbuf_pkt_len(bufs[i]); + + nb_tx = (uint16_t)rte_ring_enqueue_burst(r->rng, ptrs, nb_bufs, NULL); + + for (i = 0; i < nb_tx; i++) + rte_eth_count_packet(&r->stats, sizes[i]); + return nb_tx; } @@ -193,42 +202,33 @@ eth_dev_info(struct rte_eth_dev *dev, static int eth_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats) { - unsigned int i; - unsigned long rx_total = 0, tx_total = 0; - const struct pmd_internals *internal = dev->data->dev_private; - - for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && - i < dev->data->nb_rx_queues; i++) { - stats->q_ipackets[i] = internal->rx_ring_queues[i].rx_pkts; - rx_total += stats->q_ipackets[i]; - } - - for (i = 0; i < RTE_ETHDEV_QUEUE_STAT_CNTRS && - i < dev->data->nb_tx_queues; i++) { - stats->q_opackets[i] = internal->tx_ring_queues[i].tx_pkts; - tx_total += stats->q_opackets[i]; - } - - stats->ipackets = rx_total; - stats->opackets = tx_total; - - return 0; + return rte_eth_counters_stats_get(dev, offsetof(struct ring_queue, stats), + offsetof(struct ring_queue, stats), + stats); } static int eth_stats_reset(struct rte_eth_dev *dev) { - unsigned int i; - struct pmd_internals *internal = dev->data->dev_private; + return rte_eth_counters_reset(dev, offsetof(struct ring_queue, stats), + offsetof(struct ring_queue, stats)); +} - for (i = 0; i < dev->data->nb_rx_queues; i++) - internal->rx_ring_queues[i].rx_pkts = 0; - for (i = 0; i < dev->data->nb_tx_queues; i++) - internal->tx_ring_queues[i].tx_pkts = 0; +static int +eth_xstats_get_names(struct rte_eth_dev *dev, struct rte_eth_xstat_name *names, + __rte_unused unsigned int limit) +{ + return rte_eth_counters_xstats_get_names(dev, names); +} - return 0; +static int +eth_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats, unsigned int n) +{ + return rte_eth_counters_xstats_get(dev, offsetof(struct ring_queue, stats), + offsetof(struct ring_queue, stats), xstats, n); } + static void eth_mac_addr_remove(struct rte_eth_dev *dev __rte_unused, uint32_t index __rte_unused) @@ -339,6 +339,9 @@ static const struct eth_dev_ops ops = { .link_update = eth_link_update, .stats_get = eth_stats_get, .stats_reset = eth_stats_reset, + .xstats_get_names = eth_xstats_get_names, + .xstats_get = eth_xstats_get, + .xstats_reset = eth_stats_reset, .mac_addr_remove = eth_mac_addr_remove, .mac_addr_add = eth_mac_addr_add, .promiscuous_enable = eth_promiscuous_enable,