[RFC,26/29] net/qdma: add datapath burst API for VF

Message ID 20220706075219.517046-27-aman.kumar@vvdntech.in (mailing list archive)
State Changes Requested, archived
Delegated to: Thomas Monjalon
Headers
Series cover letter for net/qdma PMD |

Checks

Context Check Description
ci/checkpatch success coding style OK

Commit Message

Aman Kumar July 6, 2022, 7:52 a.m. UTC
  this patch implements routine for dev_start and
dev_stop PMD ops and adds support for RX/TX
datapath burst APIs for VF.

Signed-off-by: Aman Kumar <aman.kumar@vvdntech.in>
---
 drivers/net/qdma/qdma_vf_ethdev.c | 61 +++++++++++++++++++++++++++++++
 1 file changed, 61 insertions(+)
  

Patch

diff --git a/drivers/net/qdma/qdma_vf_ethdev.c b/drivers/net/qdma/qdma_vf_ethdev.c
index 5a54c00893..cbae4c9716 100644
--- a/drivers/net/qdma/qdma_vf_ethdev.c
+++ b/drivers/net/qdma/qdma_vf_ethdev.c
@@ -334,6 +334,39 @@  static int qdma_queue_context_invalidate(struct rte_eth_dev *dev, uint32_t qid,
 	return rv;
 }
 
+static int qdma_vf_dev_start(struct rte_eth_dev *dev)
+{
+	struct qdma_tx_queue *txq;
+	struct qdma_rx_queue *rxq;
+	uint32_t qid;
+	int err;
+
+	PMD_DRV_LOG(INFO, "qdma_dev_start: Starting\n");
+	/* prepare descriptor rings for operation */
+	for (qid = 0; qid < dev->data->nb_tx_queues; qid++) {
+		txq = (struct qdma_tx_queue *)dev->data->tx_queues[qid];
+
+		/* Deferred Queues should not start with dev_start */
+		if (!txq->tx_deferred_start) {
+			err = qdma_vf_dev_tx_queue_start(dev, qid);
+			if (err != 0)
+				return err;
+		}
+	}
+
+	for (qid = 0; qid < dev->data->nb_rx_queues; qid++) {
+		rxq = (struct qdma_rx_queue *)dev->data->rx_queues[qid];
+
+		/* Deferred Queues should not start with dev_start */
+		if (!rxq->rx_deferred_start) {
+			err = qdma_vf_dev_rx_queue_start(dev, qid);
+			if (err != 0)
+				return err;
+		}
+	}
+	return 0;
+}
+
 static int qdma_vf_dev_link_update(struct rte_eth_dev *dev,
 					__rte_unused int wait_to_complete)
 {
@@ -361,6 +394,24 @@  static int qdma_vf_dev_infos_get(__rte_unused struct rte_eth_dev *dev,
 	return 0;
 }
 
+static int qdma_vf_dev_stop(struct rte_eth_dev *dev)
+{
+	uint32_t qid;
+#ifdef RTE_LIBRTE_QDMA_DEBUG_DRIVER
+	struct qdma_pci_dev *qdma_dev = dev->data->dev_private;
+
+	PMD_DRV_LOG(INFO, "VF-%d(DEVFN) Stop H2C & C2H queues",
+			qdma_dev->func_id);
+#endif
+	/* reset driver's internal queue structures to default values */
+	for (qid = 0; qid < dev->data->nb_tx_queues; qid++)
+		qdma_vf_dev_tx_queue_stop(dev, qid);
+	for (qid = 0; qid < dev->data->nb_rx_queues; qid++)
+		qdma_vf_dev_rx_queue_stop(dev, qid);
+
+	return 0;
+}
+
 int qdma_vf_dev_close(struct rte_eth_dev *dev)
 {
 	struct qdma_pci_dev *qdma_dev = dev->data->dev_private;
@@ -371,6 +422,9 @@  int qdma_vf_dev_close(struct rte_eth_dev *dev)
 
 	PMD_DRV_LOG(INFO, "Closing all queues\n");
 
+	if (dev->data->dev_started)
+		qdma_vf_dev_stop(dev);
+
 	/* iterate over rx queues */
 	for (qid = 0; qid < dev->data->nb_rx_queues; ++qid) {
 		rxq = dev->data->rx_queues[qid];
@@ -729,6 +783,8 @@  int qdma_vf_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t qid)
 static struct eth_dev_ops qdma_vf_eth_dev_ops = {
 	.dev_configure        = qdma_vf_dev_configure,
 	.dev_infos_get        = qdma_vf_dev_infos_get,
+	.dev_start            = qdma_vf_dev_start,
+	.dev_stop             = qdma_vf_dev_stop,
 	.dev_close            = qdma_vf_dev_close,
 	.dev_reset            = qdma_vf_dev_reset,
 	.link_update          = qdma_vf_dev_link_update,
@@ -811,6 +867,8 @@  static int eth_qdma_vf_dev_init(struct rte_eth_dev *dev)
 	dma_priv->h2c_bypass_mode = 0;
 
 	dev->dev_ops = &qdma_vf_eth_dev_ops;
+	dev->rx_pkt_burst = &qdma_recv_pkts;
+	dev->tx_pkt_burst = &qdma_xmit_pkts;
 
 	dma_priv->config_bar_idx = DEFAULT_VF_CONFIG_BAR;
 	dma_priv->bypass_bar_idx = BAR_ID_INVALID;
@@ -913,6 +971,9 @@  static int eth_qdma_vf_dev_uninit(struct rte_eth_dev *dev)
 	if (rte_eal_process_type() != RTE_PROC_PRIMARY)
 		return -EPERM;
 
+	if (qdma_dev->dev_configured)
+		qdma_vf_dev_close(dev);
+
 	qdma_ethdev_offline(dev);
 
 	if (qdma_dev->reset_state != RESET_STATE_RECV_PF_RESET_REQ)