From patchwork Fri Aug 27 06:56:55 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Andrew Rybchenko X-Patchwork-Id: 97434 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id D263EA0C43; Fri, 27 Aug 2021 08:59:25 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 626884121F; Fri, 27 Aug 2021 08:58:34 +0200 (CEST) Received: from shelob.oktetlabs.ru (shelob.oktetlabs.ru [91.220.146.113]) by mails.dpdk.org (Postfix) with ESMTP id C8FC9411E9 for ; Fri, 27 Aug 2021 08:58:30 +0200 (CEST) Received: by shelob.oktetlabs.ru (Postfix, from userid 122) id 8E8A57F6DB; Fri, 27 Aug 2021 09:58:30 +0300 (MSK) X-Spam-Checker-Version: SpamAssassin 3.4.2 (2018-09-13) on shelob.oktetlabs.ru X-Spam-Level: X-Spam-Status: No, score=0.8 required=5.0 tests=ALL_TRUSTED, DKIM_ADSP_DISCARD, URIBL_BLOCKED autolearn=no autolearn_force=no version=3.4.2 Received: from aros.oktetlabs.ru (aros.oktetlabs.ru [192.168.38.17]) by shelob.oktetlabs.ru (Postfix) with ESMTP id 878BD7F6E1; Fri, 27 Aug 2021 09:57:34 +0300 (MSK) DKIM-Filter: OpenDKIM Filter v2.11.0 shelob.oktetlabs.ru 878BD7F6E1 Authentication-Results: shelob.oktetlabs.ru/878BD7F6E1; dkim=none; dkim-atps=neutral From: Andrew Rybchenko To: dev@dpdk.org Cc: Igor Romanov , Andy Moreton , Ivan Malov Date: Fri, 27 Aug 2021 09:56:55 +0300 Message-Id: <20210827065717.1838258-17-andrew.rybchenko@oktetlabs.ru> X-Mailer: git-send-email 2.30.2 In-Reply-To: <20210827065717.1838258-1-andrew.rybchenko@oktetlabs.ru> References: <20210827065717.1838258-1-andrew.rybchenko@oktetlabs.ru> MIME-Version: 1.0 Subject: [dpdk-dev] [PATCH 16/38] net/sfc: implement representor queue setup and release X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" From: Igor Romanov Implement queue creation and destruction both in port representors and representor proxy. Signed-off-by: Igor Romanov Signed-off-by: Andrew Rybchenko Reviewed-by: Andy Moreton Reviewed-by: Ivan Malov --- drivers/net/sfc/sfc_repr.c | 279 +++++++++++++++++++++++++++ drivers/net/sfc/sfc_repr_proxy.c | 132 +++++++++++++ drivers/net/sfc/sfc_repr_proxy.h | 22 +++ drivers/net/sfc/sfc_repr_proxy_api.h | 15 ++ 4 files changed, 448 insertions(+) diff --git a/drivers/net/sfc/sfc_repr.c b/drivers/net/sfc/sfc_repr.c index f684b1d7ef..b3876586cc 100644 --- a/drivers/net/sfc/sfc_repr.c +++ b/drivers/net/sfc/sfc_repr.c @@ -30,6 +30,25 @@ struct sfc_repr_shared { uint16_t switch_port_id; }; +struct sfc_repr_rxq { + /* Datapath members */ + struct rte_ring *ring; + + /* Non-datapath members */ + struct sfc_repr_shared *shared; + uint16_t queue_id; +}; + +struct sfc_repr_txq { + /* Datapath members */ + struct rte_ring *ring; + efx_mport_id_t egress_mport; + + /* Non-datapath members */ + struct sfc_repr_shared *shared; + uint16_t queue_id; +}; + /** Primary process representor private data */ struct sfc_repr { /** @@ -50,6 +69,14 @@ struct sfc_repr { SFC_GENERIC_LOG(ERR, __VA_ARGS__); \ } while (0) +#define sfcr_warn(sr, ...) \ + do { \ + const struct sfc_repr *_sr = (sr); \ + \ + (void)_sr; \ + SFC_GENERIC_LOG(WARNING, __VA_ARGS__); \ + } while (0) + #define sfcr_info(sr, ...) \ do { \ const struct sfc_repr *_sr = (sr); \ @@ -269,6 +296,243 @@ sfc_repr_dev_infos_get(struct rte_eth_dev *dev, return 0; } +static int +sfc_repr_ring_create(uint16_t pf_port_id, uint16_t repr_id, + const char *type_name, uint16_t qid, uint16_t nb_desc, + unsigned int socket_id, struct rte_ring **ring) +{ + char ring_name[RTE_RING_NAMESIZE]; + int ret; + + ret = snprintf(ring_name, sizeof(ring_name), "sfc_%u_repr_%u_%sq%u", + pf_port_id, repr_id, type_name, qid); + if (ret >= (int)sizeof(ring_name)) + return -ENAMETOOLONG; + + /* + * Single producer/consumer rings are used since the API for Tx/Rx + * packet burst for representors are guaranteed to be called from + * a single thread, and the user of the other end (representor proxy) + * is also single-threaded. + */ + *ring = rte_ring_create(ring_name, nb_desc, socket_id, + RING_F_SP_ENQ | RING_F_SC_DEQ); + if (*ring == NULL) + return -rte_errno; + + return 0; +} + +static int +sfc_repr_rx_qcheck_conf(struct sfc_repr *sr, + const struct rte_eth_rxconf *rx_conf) +{ + int ret = 0; + + sfcr_info(sr, "entry"); + + if (rx_conf->rx_thresh.pthresh != 0 || + rx_conf->rx_thresh.hthresh != 0 || + rx_conf->rx_thresh.wthresh != 0) { + sfcr_warn(sr, + "RxQ prefetch/host/writeback thresholds are not supported"); + } + + if (rx_conf->rx_free_thresh != 0) + sfcr_warn(sr, "RxQ free threshold is not supported"); + + if (rx_conf->rx_drop_en == 0) + sfcr_warn(sr, "RxQ drop disable is not supported"); + + if (rx_conf->rx_deferred_start) { + sfcr_err(sr, "Deferred start is not supported"); + ret = -EINVAL; + } + + sfcr_info(sr, "done: %s", rte_strerror(-ret)); + + return ret; +} + +static int +sfc_repr_rx_queue_setup(struct rte_eth_dev *dev, uint16_t rx_queue_id, + uint16_t nb_rx_desc, unsigned int socket_id, + __rte_unused const struct rte_eth_rxconf *rx_conf, + struct rte_mempool *mb_pool) +{ + struct sfc_repr_shared *srs = sfc_repr_shared_by_eth_dev(dev); + struct sfc_repr *sr = sfc_repr_by_eth_dev(dev); + struct sfc_repr_rxq *rxq; + int ret; + + sfcr_info(sr, "entry"); + + ret = sfc_repr_rx_qcheck_conf(sr, rx_conf); + if (ret != 0) + goto fail_check_conf; + + ret = -ENOMEM; + rxq = rte_zmalloc_socket("sfc-repr-rxq", sizeof(*rxq), + RTE_CACHE_LINE_SIZE, socket_id); + if (rxq == NULL) { + sfcr_err(sr, "%s() failed to alloc RxQ", __func__); + goto fail_rxq_alloc; + } + + rxq->shared = srs; + rxq->queue_id = rx_queue_id; + + ret = sfc_repr_ring_create(srs->pf_port_id, srs->repr_id, + "rx", rxq->queue_id, nb_rx_desc, + socket_id, &rxq->ring); + if (ret != 0) { + sfcr_err(sr, "%s() failed to create ring", __func__); + goto fail_ring_create; + } + + ret = sfc_repr_proxy_add_rxq(srs->pf_port_id, srs->repr_id, + rxq->queue_id, rxq->ring, mb_pool); + if (ret != 0) { + SFC_ASSERT(ret > 0); + ret = -ret; + sfcr_err(sr, "%s() failed to add proxy RxQ", __func__); + goto fail_proxy_add_rxq; + } + + dev->data->rx_queues[rx_queue_id] = rxq; + + sfcr_info(sr, "done"); + + return 0; + +fail_proxy_add_rxq: + rte_ring_free(rxq->ring); + +fail_ring_create: + rte_free(rxq); + +fail_rxq_alloc: +fail_check_conf: + sfcr_err(sr, "%s() failed: %s", __func__, rte_strerror(-ret)); + return ret; +} + +static void +sfc_repr_rx_queue_release(void *queue) +{ + struct sfc_repr_rxq *rxq = queue; + struct sfc_repr_shared *srs; + + if (rxq == NULL) + return; + + srs = rxq->shared; + sfc_repr_proxy_del_rxq(srs->pf_port_id, srs->repr_id, rxq->queue_id); + rte_ring_free(rxq->ring); + rte_free(rxq); +} + +static int +sfc_repr_tx_qcheck_conf(struct sfc_repr *sr, + const struct rte_eth_txconf *tx_conf) +{ + int ret = 0; + + sfcr_info(sr, "entry"); + + if (tx_conf->tx_rs_thresh != 0) + sfcr_warn(sr, "RS bit in transmit descriptor is not supported"); + + if (tx_conf->tx_free_thresh != 0) + sfcr_warn(sr, "TxQ free threshold is not supported"); + + if (tx_conf->tx_thresh.pthresh != 0 || + tx_conf->tx_thresh.hthresh != 0 || + tx_conf->tx_thresh.wthresh != 0) { + sfcr_warn(sr, + "prefetch/host/writeback thresholds are not supported"); + } + + if (tx_conf->tx_deferred_start) { + sfcr_err(sr, "Deferred start is not supported"); + ret = -EINVAL; + } + + sfcr_info(sr, "done: %s", rte_strerror(-ret)); + + return ret; +} + +static int +sfc_repr_tx_queue_setup(struct rte_eth_dev *dev, uint16_t tx_queue_id, + uint16_t nb_tx_desc, unsigned int socket_id, + const struct rte_eth_txconf *tx_conf) +{ + struct sfc_repr_shared *srs = sfc_repr_shared_by_eth_dev(dev); + struct sfc_repr *sr = sfc_repr_by_eth_dev(dev); + struct sfc_repr_txq *txq; + int ret; + + sfcr_info(sr, "entry"); + + ret = sfc_repr_tx_qcheck_conf(sr, tx_conf); + if (ret != 0) + goto fail_check_conf; + + ret = -ENOMEM; + txq = rte_zmalloc_socket("sfc-repr-txq", sizeof(*txq), + RTE_CACHE_LINE_SIZE, socket_id); + if (txq == NULL) + goto fail_txq_alloc; + + txq->shared = srs; + txq->queue_id = tx_queue_id; + + ret = sfc_repr_ring_create(srs->pf_port_id, srs->repr_id, + "tx", txq->queue_id, nb_tx_desc, + socket_id, &txq->ring); + if (ret != 0) + goto fail_ring_create; + + ret = sfc_repr_proxy_add_txq(srs->pf_port_id, srs->repr_id, + txq->queue_id, txq->ring, + &txq->egress_mport); + if (ret != 0) + goto fail_proxy_add_txq; + + dev->data->tx_queues[tx_queue_id] = txq; + + sfcr_info(sr, "done"); + + return 0; + +fail_proxy_add_txq: + rte_ring_free(txq->ring); + +fail_ring_create: + rte_free(txq); + +fail_txq_alloc: +fail_check_conf: + sfcr_err(sr, "%s() failed: %s", __func__, rte_strerror(-ret)); + return ret; +} + +static void +sfc_repr_tx_queue_release(void *queue) +{ + struct sfc_repr_txq *txq = queue; + struct sfc_repr_shared *srs; + + if (txq == NULL) + return; + + srs = txq->shared; + sfc_repr_proxy_del_txq(srs->pf_port_id, srs->repr_id, txq->queue_id); + rte_ring_free(txq->ring); + rte_free(txq); +} + static void sfc_repr_close(struct sfc_repr *sr) { @@ -287,6 +551,7 @@ sfc_repr_dev_close(struct rte_eth_dev *dev) { struct sfc_repr *sr = sfc_repr_by_eth_dev(dev); struct sfc_repr_shared *srs = sfc_repr_shared_by_eth_dev(dev); + unsigned int i; sfcr_info(sr, "entry"); @@ -303,6 +568,16 @@ sfc_repr_dev_close(struct rte_eth_dev *dev) break; } + for (i = 0; i < dev->data->nb_rx_queues; i++) { + sfc_repr_rx_queue_release(dev->data->rx_queues[i]); + dev->data->rx_queues[i] = NULL; + } + + for (i = 0; i < dev->data->nb_tx_queues; i++) { + sfc_repr_tx_queue_release(dev->data->tx_queues[i]); + dev->data->tx_queues[i] = NULL; + } + /* * Cleanup all resources. * Rollback primary process sfc_repr_eth_dev_init() below. @@ -326,6 +601,10 @@ static const struct eth_dev_ops sfc_repr_dev_ops = { .dev_configure = sfc_repr_dev_configure, .dev_close = sfc_repr_dev_close, .dev_infos_get = sfc_repr_dev_infos_get, + .rx_queue_setup = sfc_repr_rx_queue_setup, + .rx_queue_release = sfc_repr_rx_queue_release, + .tx_queue_setup = sfc_repr_tx_queue_setup, + .tx_queue_release = sfc_repr_tx_queue_release, }; diff --git a/drivers/net/sfc/sfc_repr_proxy.c b/drivers/net/sfc/sfc_repr_proxy.c index f64fa2efc7..6a89cca40a 100644 --- a/drivers/net/sfc/sfc_repr_proxy.c +++ b/drivers/net/sfc/sfc_repr_proxy.c @@ -528,3 +528,135 @@ sfc_repr_proxy_del_port(uint16_t pf_port_id, uint16_t repr_id) return rc; } + +int +sfc_repr_proxy_add_rxq(uint16_t pf_port_id, uint16_t repr_id, + uint16_t queue_id, struct rte_ring *rx_ring, + struct rte_mempool *mp) +{ + struct sfc_repr_proxy_port *port; + struct sfc_repr_proxy_rxq *rxq; + struct sfc_repr_proxy *rp; + struct sfc_adapter *sa; + + sa = sfc_get_adapter_by_pf_port_id(pf_port_id); + rp = sfc_repr_proxy_by_adapter(sa); + + sfc_log_init(sa, "entry"); + + port = sfc_repr_proxy_find_port(rp, repr_id); + if (port == NULL) { + sfc_err(sa, "%s() failed: no such port", __func__); + return ENOENT; + } + + rxq = &port->rxq[queue_id]; + if (rp->dp_rxq[queue_id].mp != NULL && rp->dp_rxq[queue_id].mp != mp) { + sfc_err(sa, "multiple mempools per queue are not supported"); + sfc_put_adapter(sa); + return ENOTSUP; + } + + rxq->ring = rx_ring; + rxq->mb_pool = mp; + rp->dp_rxq[queue_id].mp = mp; + rp->dp_rxq[queue_id].ref_count++; + + sfc_log_init(sa, "done"); + sfc_put_adapter(sa); + + return 0; +} + +void +sfc_repr_proxy_del_rxq(uint16_t pf_port_id, uint16_t repr_id, + uint16_t queue_id) +{ + struct sfc_repr_proxy_port *port; + struct sfc_repr_proxy_rxq *rxq; + struct sfc_repr_proxy *rp; + struct sfc_adapter *sa; + + sa = sfc_get_adapter_by_pf_port_id(pf_port_id); + rp = sfc_repr_proxy_by_adapter(sa); + + sfc_log_init(sa, "entry"); + + port = sfc_repr_proxy_find_port(rp, repr_id); + if (port == NULL) { + sfc_err(sa, "%s() failed: no such port", __func__); + return; + } + + rxq = &port->rxq[queue_id]; + + rxq->ring = NULL; + rxq->mb_pool = NULL; + rp->dp_rxq[queue_id].ref_count--; + if (rp->dp_rxq[queue_id].ref_count == 0) + rp->dp_rxq[queue_id].mp = NULL; + + sfc_log_init(sa, "done"); + sfc_put_adapter(sa); +} + +int +sfc_repr_proxy_add_txq(uint16_t pf_port_id, uint16_t repr_id, + uint16_t queue_id, struct rte_ring *tx_ring, + efx_mport_id_t *egress_mport) +{ + struct sfc_repr_proxy_port *port; + struct sfc_repr_proxy_txq *txq; + struct sfc_repr_proxy *rp; + struct sfc_adapter *sa; + + sa = sfc_get_adapter_by_pf_port_id(pf_port_id); + rp = sfc_repr_proxy_by_adapter(sa); + + sfc_log_init(sa, "entry"); + + port = sfc_repr_proxy_find_port(rp, repr_id); + if (port == NULL) { + sfc_err(sa, "%s() failed: no such port", __func__); + return ENOENT; + } + + txq = &port->txq[queue_id]; + + txq->ring = tx_ring; + + *egress_mport = port->egress_mport; + + sfc_log_init(sa, "done"); + sfc_put_adapter(sa); + + return 0; +} + +void +sfc_repr_proxy_del_txq(uint16_t pf_port_id, uint16_t repr_id, + uint16_t queue_id) +{ + struct sfc_repr_proxy_port *port; + struct sfc_repr_proxy_txq *txq; + struct sfc_repr_proxy *rp; + struct sfc_adapter *sa; + + sa = sfc_get_adapter_by_pf_port_id(pf_port_id); + rp = sfc_repr_proxy_by_adapter(sa); + + sfc_log_init(sa, "entry"); + + port = sfc_repr_proxy_find_port(rp, repr_id); + if (port == NULL) { + sfc_err(sa, "%s() failed: no such port", __func__); + return; + } + + txq = &port->txq[queue_id]; + + txq->ring = NULL; + + sfc_log_init(sa, "done"); + sfc_put_adapter(sa); +} diff --git a/drivers/net/sfc/sfc_repr_proxy.h b/drivers/net/sfc/sfc_repr_proxy.h index e4a6213c10..bd7ad7148a 100644 --- a/drivers/net/sfc/sfc_repr_proxy.h +++ b/drivers/net/sfc/sfc_repr_proxy.h @@ -12,8 +12,13 @@ #include +#include +#include + #include "efx.h" +#include "sfc_repr.h" + #ifdef __cplusplus extern "C" { #endif @@ -26,11 +31,27 @@ extern "C" { #define SFC_REPR_PROXY_NB_TXQ_MIN (1) #define SFC_REPR_PROXY_NB_TXQ_MAX (1) +struct sfc_repr_proxy_rxq { + struct rte_ring *ring; + struct rte_mempool *mb_pool; +}; + +struct sfc_repr_proxy_txq { + struct rte_ring *ring; +}; + struct sfc_repr_proxy_port { TAILQ_ENTRY(sfc_repr_proxy_port) entries; uint16_t repr_id; uint16_t rte_port_id; efx_mport_id_t egress_mport; + struct sfc_repr_proxy_rxq rxq[SFC_REPR_RXQ_MAX]; + struct sfc_repr_proxy_txq txq[SFC_REPR_TXQ_MAX]; +}; + +struct sfc_repr_proxy_dp_rxq { + struct rte_mempool *mp; + unsigned int ref_count; }; enum sfc_repr_proxy_mbox_op { @@ -54,6 +75,7 @@ struct sfc_repr_proxy { efx_mport_id_t mport_alias; struct sfc_repr_proxy_ports ports; bool started; + struct sfc_repr_proxy_dp_rxq dp_rxq[SFC_REPR_PROXY_NB_RXQ_MAX]; struct sfc_repr_proxy_mbox mbox; }; diff --git a/drivers/net/sfc/sfc_repr_proxy_api.h b/drivers/net/sfc/sfc_repr_proxy_api.h index af9009ca3c..d1c0760efa 100644 --- a/drivers/net/sfc/sfc_repr_proxy_api.h +++ b/drivers/net/sfc/sfc_repr_proxy_api.h @@ -12,6 +12,9 @@ #include +#include +#include + #include "efx.h" #ifdef __cplusplus @@ -23,6 +26,18 @@ int sfc_repr_proxy_add_port(uint16_t pf_port_id, uint16_t repr_id, const efx_mport_sel_t *mport_set); int sfc_repr_proxy_del_port(uint16_t pf_port_id, uint16_t repr_id); +int sfc_repr_proxy_add_rxq(uint16_t pf_port_id, uint16_t repr_id, + uint16_t queue_id, struct rte_ring *rx_ring, + struct rte_mempool *mp); +void sfc_repr_proxy_del_rxq(uint16_t pf_port_id, uint16_t repr_id, + uint16_t queue_id); + +int sfc_repr_proxy_add_txq(uint16_t pf_port_id, uint16_t repr_id, + uint16_t queue_id, struct rte_ring *tx_ring, + efx_mport_id_t *egress_mport); +void sfc_repr_proxy_del_txq(uint16_t pf_port_id, uint16_t repr_id, + uint16_t queue_id); + #ifdef __cplusplus } #endif