From patchwork Sat Sep 24 02:45:30 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Long Li X-Patchwork-Id: 116782 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id E9C52A054F; Sat, 24 Sep 2022 04:46:06 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 1B42A42BDA; Sat, 24 Sep 2022 04:46:02 +0200 (CEST) Received: from linux.microsoft.com (linux.microsoft.com [13.77.154.182]) by mails.dpdk.org (Postfix) with ESMTP id F0B7042BCE for ; Sat, 24 Sep 2022 04:45:53 +0200 (CEST) Received: by linux.microsoft.com (Postfix, from userid 1004) id 2D73020C32BB; Fri, 23 Sep 2022 19:45:53 -0700 (PDT) DKIM-Filter: OpenDKIM Filter v2.11.0 linux.microsoft.com 2D73020C32BB DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linuxonhyperv.com; s=default; t=1663987553; bh=0QdQCWax4pQN4W1EaIAes+/NKrj75LasxJsJxV4umUE=; h=From:To:Cc:Subject:Date:In-Reply-To:References:Reply-To:From; b=rnZekATDPtU4okrEOrDgQbpJL2R/xCLe8di6RcvQ/Sfvbs/bIv7/KDrql73xk0BH4 9Rtz6QdmuitcmuY15bNzApMusaBj3IV8k7NLy4NW1W2C3W8MjrkDpLd96yTbEnUCkg 1okISStuEok5O5zBdp7nehQmA8NE0hgTE/N5k0JQ= From: longli@linuxonhyperv.com To: Ferruh Yigit Cc: dev@dpdk.org, Ajay Sharma , Stephen Hemminger , Long Li Subject: [Patch v9 02/18] net/mana: device configuration and stop Date: Fri, 23 Sep 2022 19:45:30 -0700 Message-Id: <1663987546-15982-3-git-send-email-longli@linuxonhyperv.com> X-Mailer: git-send-email 1.8.3.1 In-Reply-To: <1663987546-15982-1-git-send-email-longli@linuxonhyperv.com> References: <1662674189-29524-1-git-send-email-longli@linuxonhyperv.com> <1663987546-15982-1-git-send-email-longli@linuxonhyperv.com> X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Reply-To: longli@microsoft.com Errors-To: dev-bounces@dpdk.org From: Long Li MANA defines its memory allocation functions to override IB layer default functions to allocate device queues. This patch adds the code for device configuration and stop. Signed-off-by: Long Li --- v2: Removed validation for offload settings in mana_dev_configure(). v8: Fix coding style to function definitions. drivers/net/mana/mana.c | 81 ++++++++++++++++++++++++++++++++++++++++- drivers/net/mana/mana.h | 4 ++ 2 files changed, 83 insertions(+), 2 deletions(-) diff --git a/drivers/net/mana/mana.c b/drivers/net/mana/mana.c index 1c317418b6..64069e2b9f 100644 --- a/drivers/net/mana/mana.c +++ b/drivers/net/mana/mana.c @@ -42,7 +42,85 @@ static rte_spinlock_t mana_shared_data_lock = RTE_SPINLOCK_INITIALIZER; int mana_logtype_driver; int mana_logtype_init; +/* + * Callback from rdma-core to allocate a buffer for a queue. + */ +void * +mana_alloc_verbs_buf(size_t size, void *data) +{ + void *ret; + size_t alignment = rte_mem_page_size(); + int socket = (int)(uintptr_t)data; + + DRV_LOG(DEBUG, "size=%zu socket=%d", size, socket); + + if (alignment == (size_t)-1) { + DRV_LOG(ERR, "Failed to get mem page size"); + rte_errno = ENOMEM; + return NULL; + } + + ret = rte_zmalloc_socket("mana_verb_buf", size, alignment, socket); + if (!ret && size) + rte_errno = ENOMEM; + return ret; +} + +void +mana_free_verbs_buf(void *ptr, void *data __rte_unused) +{ + rte_free(ptr); +} + +static int +mana_dev_configure(struct rte_eth_dev *dev) +{ + struct mana_priv *priv = dev->data->dev_private; + struct rte_eth_conf *dev_conf = &dev->data->dev_conf; + + if (dev_conf->rxmode.mq_mode & ETH_MQ_RX_RSS_FLAG) + dev_conf->rxmode.offloads |= DEV_RX_OFFLOAD_RSS_HASH; + + if (dev->data->nb_rx_queues != dev->data->nb_tx_queues) { + DRV_LOG(ERR, "Only support equal number of rx/tx queues"); + return -EINVAL; + } + + if (!rte_is_power_of_2(dev->data->nb_rx_queues)) { + DRV_LOG(ERR, "number of TX/RX queues must be power of 2"); + return -EINVAL; + } + + priv->num_queues = dev->data->nb_rx_queues; + + manadv_set_context_attr(priv->ib_ctx, MANADV_CTX_ATTR_BUF_ALLOCATORS, + (void *)((uintptr_t)&(struct manadv_ctx_allocators){ + .alloc = &mana_alloc_verbs_buf, + .free = &mana_free_verbs_buf, + .data = 0, + })); + + return 0; +} + +static int +mana_dev_close(struct rte_eth_dev *dev) +{ + struct mana_priv *priv = dev->data->dev_private; + int ret; + + ret = ibv_close_device(priv->ib_ctx); + if (ret) { + ret = errno; + return ret; + } + + return 0; +} + static const struct eth_dev_ops mana_dev_ops = { + .dev_configure = mana_dev_configure, + .dev_close = mana_dev_close, }; static const struct eth_dev_ops mana_dev_secondary_ops = { @@ -649,8 +727,7 @@ mana_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, static int mana_dev_uninit(struct rte_eth_dev *dev) { - RTE_SET_USED(dev); - return 0; + return mana_dev_close(dev); } /* diff --git a/drivers/net/mana/mana.h b/drivers/net/mana/mana.h index a2021ceb4a..0b4a828b0a 100644 --- a/drivers/net/mana/mana.h +++ b/drivers/net/mana/mana.h @@ -28,6 +28,7 @@ struct mana_process_priv { struct mana_priv { struct rte_eth_dev_data *dev_data; struct mana_process_priv *process_priv; + int num_queues; /* DPDK port */ uint16_t port_id; @@ -99,4 +100,7 @@ int mana_mp_req_verbs_cmd_fd(struct rte_eth_dev *dev); void mana_mp_req_on_rxtx(struct rte_eth_dev *dev, enum mana_mp_req_type type); +void *mana_alloc_verbs_buf(size_t size, void *data); +void mana_free_verbs_buf(void *ptr, void *data __rte_unused); + #endif