Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/114578/?format=api
http://patches.dpdk.org/api/patches/114578/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/patch/20220803113104.1184059-13-junfeng.guo@intel.com/", "project": { "id": 1, "url": "http://patches.dpdk.org/api/projects/1/?format=api", "name": "DPDK", "link_name": "dpdk", "list_id": "dev.dpdk.org", "list_email": "dev@dpdk.org", "web_url": "http://core.dpdk.org", "scm_url": "git://dpdk.org/dpdk", "webscm_url": "http://git.dpdk.org/dpdk", "list_archive_url": "https://inbox.dpdk.org/dev", "list_archive_url_format": "https://inbox.dpdk.org/dev/{}", "commit_url_format": "" }, "msgid": "<20220803113104.1184059-13-junfeng.guo@intel.com>", "list_archive_url": "https://inbox.dpdk.org/dev/20220803113104.1184059-13-junfeng.guo@intel.com", "date": "2022-08-03T11:31:03", "name": "[12/13] net/idpf: support write back based on ITR expire", "commit_ref": null, "pull_url": null, "state": "superseded", "archived": true, "hash": "ab95baae0da74701ef52da46312113491d395ba8", "submitter": { "id": 1785, "url": "http://patches.dpdk.org/api/people/1785/?format=api", "name": "Junfeng Guo", "email": "junfeng.guo@intel.com" }, "delegate": { "id": 3961, "url": "http://patches.dpdk.org/api/users/3961/?format=api", "username": "arybchenko", "first_name": "Andrew", "last_name": "Rybchenko", "email": "andrew.rybchenko@oktetlabs.ru" }, "mbox": "http://patches.dpdk.org/project/dpdk/patch/20220803113104.1184059-13-junfeng.guo@intel.com/mbox/", "series": [ { "id": 24188, "url": "http://patches.dpdk.org/api/series/24188/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=24188", "date": "2022-08-03T11:30:51", "name": "add support for idpf PMD in DPDK", "version": 1, "mbox": "http://patches.dpdk.org/series/24188/mbox/" } ], "comments": "http://patches.dpdk.org/api/patches/114578/comments/", "check": "warning", "checks": "http://patches.dpdk.org/api/patches/114578/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "<dev-bounces@dpdk.org>", "X-Original-To": "patchwork@inbox.dpdk.org", "Delivered-To": "patchwork@inbox.dpdk.org", "Received": [ "from mails.dpdk.org (mails.dpdk.org [217.70.189.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id 1A889A00C5;\n\tWed, 3 Aug 2022 13:32:48 +0200 (CEST)", "from [217.70.189.124] (localhost [127.0.0.1])\n\tby mails.dpdk.org (Postfix) with ESMTP id B432F42BDE;\n\tWed, 3 Aug 2022 13:31:43 +0200 (CEST)", "from mga05.intel.com (mga05.intel.com [192.55.52.43])\n by mails.dpdk.org (Postfix) with ESMTP id CBEAC42BD4\n for <dev@dpdk.org>; Wed, 3 Aug 2022 13:31:40 +0200 (CEST)", "from fmsmga003.fm.intel.com ([10.253.24.29])\n by fmsmga105.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 03 Aug 2022 04:31:40 -0700", "from dpdk-jf-ntb-v2.sh.intel.com ([10.67.118.246])\n by FMSMGA003.fm.intel.com with ESMTP; 03 Aug 2022 04:31:38 -0700" ], "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/simple;\n d=intel.com; i=@intel.com; q=dns/txt; s=Intel;\n t=1659526301; x=1691062301;\n h=from:to:cc:subject:date:message-id:in-reply-to:\n references:mime-version:content-transfer-encoding;\n bh=tqA3ezylEkEAMA2KJ7vJ9+WNYHeSQw+Im0o4BmdGdho=;\n b=AVU+O2f5NnUsI4tbZUQYSYBFoIN50h1MCTUAmPDYBgVd7AzkVcxaYHnv\n hH+51pwoKLRjT9oxuGRMGx4mT4YYxY/StKS3hpPkB7bXAf8kv0acdIPtM\n BOGdloy9Ent7KLeYWVb8swmoMFvn/yQUPl9fgjoZZBryVjz+veNCHTyEr\n 8bX/IQxwnIqT5b4NbnfjkPnhW2St0+emOga0W6otKol+/r1Wp/V2PmNM2\n 1tlETNWEYCH0dNTYSV3xMyY0aLmu1z/nm5lgWHNQoYvOLjNA5brI6dENK\n G24a9VYFUxcyzfWLBNg72uCJCAYtYQK1cxInxEBQuFGsqLh87JU4LAYuu w==;", "X-IronPort-AV": [ "E=McAfee;i=\"6400,9594,10427\"; a=\"375948544\"", "E=Sophos;i=\"5.93,214,1654585200\"; d=\"scan'208\";a=\"375948544\"", "E=Sophos;i=\"5.93,214,1654585200\"; d=\"scan'208\";a=\"692211133\"" ], "X-ExtLoop1": "1", "From": "Junfeng Guo <junfeng.guo@intel.com>", "To": "qi.z.zhang@intel.com,\n\tjingjing.wu@intel.com,\n\tbeilei.xing@intel.com", "Cc": "dev@dpdk.org,\n\tjunfeng.guo@intel.com", "Subject": "[PATCH 12/13] net/idpf: support write back based on ITR expire", "Date": "Wed, 3 Aug 2022 19:31:03 +0800", "Message-Id": "<20220803113104.1184059-13-junfeng.guo@intel.com>", "X-Mailer": "git-send-email 2.25.1", "In-Reply-To": "<20220803113104.1184059-1-junfeng.guo@intel.com>", "References": "<20220803113104.1184059-1-junfeng.guo@intel.com>", "MIME-Version": "1.0", "Content-Transfer-Encoding": "8bit", "X-BeenThere": "dev@dpdk.org", "X-Mailman-Version": "2.1.29", "Precedence": "list", "List-Id": "DPDK patches and discussions <dev.dpdk.org>", "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n <mailto:dev-request@dpdk.org?subject=unsubscribe>", "List-Archive": "<http://mails.dpdk.org/archives/dev/>", "List-Post": "<mailto:dev@dpdk.org>", "List-Help": "<mailto:dev-request@dpdk.org?subject=help>", "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n <mailto:dev-request@dpdk.org?subject=subscribe>", "Errors-To": "dev-bounces@dpdk.org" }, "content": "Force write-backs by setting WB_ON_ITR bit in DYN_CTL register,\nso that the packets can be received once at a time.\n\nSigned-off-by: Beilei Xing <beilei.xing@intel.com>\nSigned-off-by: Junfeng Guo <junfeng.guo@intel.com>\n---\n drivers/net/idpf/idpf_ethdev.c | 117 +++++++++++++++++++++++++++++++++\n drivers/net/idpf/idpf_ethdev.h | 8 +++\n drivers/net/idpf/idpf_vchnl.c | 108 ++++++++++++++++++++++++++++++\n 3 files changed, 233 insertions(+)", "diff": "diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c\nindex b934488d0b..1e9564a5a9 100644\n--- a/drivers/net/idpf/idpf_ethdev.c\n+++ b/drivers/net/idpf/idpf_ethdev.c\n@@ -512,6 +512,87 @@ idpf_dev_configure(struct rte_eth_dev *dev)\n \treturn ret;\n }\n \n+static int\n+idpf_config_rx_queues_irqs(struct rte_eth_dev *dev)\n+{\n+\tstruct idpf_vport *vport =\n+\t\t(struct idpf_vport *)dev->data->dev_private;\n+\tstruct virtchnl2_queue_vector *qv_map;\n+\tstruct iecm_hw *hw = &adapter->hw;\n+\tuint32_t dynctl_reg_start;\n+\tuint32_t itrn_reg_start;\n+\tuint32_t dynctl_val, itrn_val;\n+\tuint16_t i;\n+\n+\tqv_map = rte_zmalloc(\"qv_map\",\n+\t\t\tdev->data->nb_rx_queues *\n+\t\t\tsizeof(struct virtchnl2_queue_vector), 0);\n+\tif (!qv_map) {\n+\t\tPMD_DRV_LOG(ERR, \"Failed to allocate %d queue-vector map\",\n+\t\t\t dev->data->nb_rx_queues);\n+\t\tgoto qv_map_alloc_err;\n+\t}\n+\n+\t/* Rx interrupt disabled, Map interrupt only for writeback */\n+\n+\t/* The capability flags adapter->caps->other_caps here should be\n+\t * compared with bit VIRTCHNL2_CAP_WB_ON_ITR. The if condition should\n+\t * be updated when the FW can return correct flag bits.\n+\t */\n+\tif (adapter->caps->other_caps) {\n+\t\tdynctl_reg_start = vport->recv_vectors->vchunks.vchunks->dynctl_reg_start;\n+\t\titrn_reg_start = vport->recv_vectors->vchunks.vchunks->itrn_reg_start;\n+\t\tdynctl_val = IECM_READ_REG(hw, dynctl_reg_start);\n+\t\tPMD_DRV_LOG(DEBUG, \"Value of dynctl_reg_start is 0x%x\", dynctl_val);\n+\t\titrn_val = IECM_READ_REG(hw, itrn_reg_start);\n+\t\tPMD_DRV_LOG(DEBUG, \"Value of itrn_reg_start is 0x%x\", itrn_val);\n+\t\t/* Force write-backs by setting WB_ON_ITR bit in DYN_CTL\n+\t\t * register. WB_ON_ITR and INTENA are mutually exclusive\n+\t\t * bits. Setting WB_ON_ITR bits means TX and RX Descs\n+\t\t * are writen back based on ITR expiration irrespective\n+\t\t * of INTENA setting.\n+\t\t */\n+\t\t/* TBD: need to tune INTERVAL value for better performance. */\n+\t\tif (itrn_val)\n+\t\t\tIECM_WRITE_REG(hw,\n+\t\t\t\t dynctl_reg_start,\n+\t\t\t\t VIRTCHNL2_ITR_IDX_0 <<\n+\t\t\t\t PF_GLINT_DYN_CTL_ITR_INDX_S |\n+\t\t\t\t PF_GLINT_DYN_CTL_WB_ON_ITR_M |\n+\t\t\t\t itrn_val <<\n+\t\t\t\t PF_GLINT_DYN_CTL_INTERVAL_S);\n+\t\telse\n+\t\t\tIECM_WRITE_REG(hw,\n+\t\t\t\t dynctl_reg_start,\n+\t\t\t\t VIRTCHNL2_ITR_IDX_0 <<\n+\t\t\t\t PF_GLINT_DYN_CTL_ITR_INDX_S |\n+\t\t\t\t PF_GLINT_DYN_CTL_WB_ON_ITR_M |\n+\t\t\t\t IDPF_DFLT_INTERVAL <<\n+\t\t\t\t PF_GLINT_DYN_CTL_INTERVAL_S);\n+\t}\n+\tfor (i = 0; i < dev->data->nb_rx_queues; i++) {\n+\t\t/* map all queues to the same vector */\n+\t\tqv_map[i].queue_id = vport->chunks_info.rx_start_qid + i;\n+\t\tqv_map[i].vector_id =\n+\t\t\tvport->recv_vectors->vchunks.vchunks->start_vector_id;\n+\t}\n+\tvport->qv_map = qv_map;\n+\n+\tif (idpf_config_irq_map_unmap(vport, true)) {\n+\t\tPMD_DRV_LOG(ERR, \"config interrupt mapping failed\");\n+\t\tgoto config_irq_map_err;\n+\t}\n+\n+\treturn 0;\n+\n+config_irq_map_err:\n+\trte_free(vport->qv_map);\n+\tvport->qv_map = NULL;\n+\n+qv_map_alloc_err:\n+\treturn -1;\n+}\n+\n static int\n idpf_start_queues(struct rte_eth_dev *dev)\n {\n@@ -550,6 +631,9 @@ idpf_dev_start(struct rte_eth_dev *dev)\n {\n \tstruct idpf_vport *vport =\n \t\t(struct idpf_vport *)dev->data->dev_private;\n+\tuint16_t num_allocated_vectors =\n+\t\tadapter->caps->num_allocated_vectors;\n+\tuint16_t req_vecs_num;\n \n \tPMD_INIT_FUNC_TRACE();\n \n@@ -562,6 +646,23 @@ idpf_dev_start(struct rte_eth_dev *dev)\n \n \tvport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;\n \n+\treq_vecs_num = IDPF_DFLT_Q_VEC_NUM;\n+\tif (req_vecs_num + used_vecs_num > num_allocated_vectors) {\n+\t\tPMD_DRV_LOG(ERR, \"The accumulated request vectors' number should be less than %d\",\n+\t\t\t num_allocated_vectors);\n+\t\tgoto err_mtu;\n+\t}\n+\tif (idpf_alloc_vectors(vport, req_vecs_num)) {\n+\t\tPMD_DRV_LOG(ERR, \"Failed to allocate interrupt vectors\");\n+\t\tgoto err_mtu;\n+\t}\n+\tused_vecs_num += req_vecs_num;\n+\n+\tif (idpf_config_rx_queues_irqs(dev)) {\n+\t\tPMD_DRV_LOG(ERR, \"Failed to configure irqs\");\n+\t\tgoto err_mtu;\n+\t}\n+\n \tif (idpf_start_queues(dev)) {\n \t\tPMD_DRV_LOG(ERR, \"Failed to start queues\");\n \t\tgoto err_mtu;\n@@ -603,6 +704,12 @@ idpf_dev_stop(struct rte_eth_dev *dev)\n \n \tidpf_stop_queues(dev);\n \n+\tif (idpf_config_irq_map_unmap(vport, false))\n+\t\tPMD_DRV_LOG(ERR, \"config interrupt unmapping failed\");\n+\n+\tif (idpf_dealloc_vectors(vport))\n+\t\tPMD_DRV_LOG(ERR, \"deallocate interrupt vectors failed\");\n+\n \tvport->stopped = 1;\n \tdev->data->dev_started = 0;\n \n@@ -631,6 +738,16 @@ idpf_dev_close(struct rte_eth_dev *dev)\n \t\tvport->rss_key = NULL;\n \t}\n \n+\tif (vport->recv_vectors) {\n+\t\trte_free(vport->recv_vectors);\n+\t\tvport->recv_vectors = NULL;\n+\t}\n+\n+\tif (vport->qv_map) {\n+\t\trte_free(vport->qv_map);\n+\t\tvport->qv_map = NULL;\n+\t}\n+\n \treturn 0;\n }\n \ndiff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h\nindex c0cbf4c3c6..32520c03bb 100644\n--- a/drivers/net/idpf/idpf_ethdev.h\n+++ b/drivers/net/idpf/idpf_ethdev.h\n@@ -119,6 +119,11 @@ struct idpf_vport {\n \tuint8_t *rss_key;\n \tuint64_t rss_hf;\n \n+\t/* MSIX info*/\n+\tstruct virtchnl2_queue_vector *qv_map; /* queue vector mapping */\n+\tuint16_t max_vectors;\n+\tstruct virtchnl2_alloc_vectors *recv_vectors;\n+\n \t/* Chunk info */\n \tstruct idpf_chunks_info chunks_info;\n \n@@ -229,6 +234,9 @@ int idpf_ena_dis_queues(struct idpf_vport *vport, bool enable);\n int idpf_ena_dis_vport(struct idpf_vport *vport, bool enable);\n int idpf_query_stats(struct idpf_vport *vport,\n \t\t\tstruct virtchnl2_vport_stats **pstats);\n+int idpf_config_irq_map_unmap(struct idpf_vport *vport, bool map);\n+int idpf_alloc_vectors(struct idpf_vport *vport, uint16_t num_vectors);\n+int idpf_dealloc_vectors(struct idpf_vport *vport);\n \n #endif /* _IDPF_ETHDEV_H_ */\n \ndiff --git a/drivers/net/idpf/idpf_vchnl.c b/drivers/net/idpf/idpf_vchnl.c\nindex 563f8f649e..bfb3b08465 100644\n--- a/drivers/net/idpf/idpf_vchnl.c\n+++ b/drivers/net/idpf/idpf_vchnl.c\n@@ -220,6 +220,10 @@ idpf_execute_vc_cmd(struct idpf_adapter *adapter, struct idpf_cmd_info *args)\n \tcase VIRTCHNL2_OP_ENABLE_VPORT:\n \tcase VIRTCHNL2_OP_DISABLE_VPORT:\n \tcase VIRTCHNL2_OP_GET_STATS:\n+\tcase VIRTCHNL2_OP_MAP_QUEUE_VECTOR:\n+\tcase VIRTCHNL2_OP_UNMAP_QUEUE_VECTOR:\n+\tcase VIRTCHNL2_OP_ALLOC_VECTORS:\n+\tcase VIRTCHNL2_OP_DEALLOC_VECTORS:\n \t\t/* for init virtchnl ops, need to poll the response */\n \t\tdo {\n \t\t\tresult = idpf_read_msg_from_ipf(adapter,\n@@ -900,6 +904,110 @@ idpf_config_txq(struct idpf_vport *vport, uint16_t txq_id)\n \treturn err;\n }\n \n+int\n+idpf_config_irq_map_unmap(struct idpf_vport *vport, bool map)\n+{\n+\tstruct virtchnl2_queue_vector_maps *map_info;\n+\tstruct virtchnl2_queue_vector *vecmap;\n+\tuint16_t nb_rxq = vport->dev_data->nb_rx_queues;\n+\tstruct idpf_cmd_info args;\n+\tint len, i, err = 0;\n+\n+\tlen = sizeof(struct virtchnl2_queue_vector_maps) +\n+\t\t(nb_rxq - 1) * sizeof(struct virtchnl2_queue_vector);\n+\n+\tmap_info = rte_zmalloc(\"map_info\", len, 0);\n+\tif (!map_info)\n+\t\treturn -ENOMEM;\n+\n+\tmap_info->vport_id = vport->vport_id;\n+\tmap_info->num_qv_maps = nb_rxq;\n+\tfor (i = 0; i < nb_rxq; i++) {\n+\t\tvecmap = &map_info->qv_maps[i];\n+\t\tvecmap->queue_id = vport->qv_map[i].queue_id;\n+\t\tvecmap->vector_id = vport->qv_map[i].vector_id;\n+\t\tvecmap->itr_idx = VIRTCHNL2_ITR_IDX_0;\n+\t\tvecmap->queue_type = VIRTCHNL2_QUEUE_TYPE_RX;\n+\t}\n+\n+\targs.ops = map ? VIRTCHNL2_OP_MAP_QUEUE_VECTOR :\n+\t\tVIRTCHNL2_OP_UNMAP_QUEUE_VECTOR;\n+\targs.in_args = (u8 *)map_info;\n+\targs.in_args_size = len;\n+\targs.out_buffer = adapter->mbx_resp;\n+\targs.out_size = IDPF_DFLT_MBX_BUF_SIZE;\n+\terr = idpf_execute_vc_cmd(adapter, &args);\n+\tif (err)\n+\t\tPMD_DRV_LOG(ERR, \"Failed to execute command of VIRTCHNL2_OP_%s_QUEUE_VECTOR\",\n+\t\t\t map ? \"MAP\" : \"UNMAP\");\n+\n+\trte_free(map_info);\n+\treturn err;\n+}\n+\n+int\n+idpf_alloc_vectors(struct idpf_vport *vport, uint16_t num_vectors)\n+{\n+\tstruct virtchnl2_alloc_vectors *alloc_vec;\n+\tstruct idpf_cmd_info args;\n+\tint err, len;\n+\n+\tlen = sizeof(struct virtchnl2_alloc_vectors) +\n+\t\t(num_vectors - 1) * sizeof(struct virtchnl2_vector_chunk);\n+\talloc_vec = rte_zmalloc(\"alloc_vec\", len, 0);\n+\tif (!alloc_vec)\n+\t\treturn -ENOMEM;\n+\n+\talloc_vec->num_vectors = num_vectors;\n+\n+\targs.ops = VIRTCHNL2_OP_ALLOC_VECTORS;\n+\targs.in_args = (u8 *)alloc_vec;\n+\targs.in_args_size = sizeof(struct virtchnl2_alloc_vectors);\n+\targs.out_buffer = adapter->mbx_resp;\n+\targs.out_size = IDPF_DFLT_MBX_BUF_SIZE;\n+\terr = idpf_execute_vc_cmd(adapter, &args);\n+\tif (err)\n+\t\tPMD_DRV_LOG(ERR, \"Failed to execute command VIRTCHNL2_OP_ALLOC_VECTORS\");\n+\n+\tif (!vport->recv_vectors) {\n+\t\tvport->recv_vectors = rte_zmalloc(\"recv_vectors\", len, 0);\n+\t\tif (!vport->recv_vectors) {\n+\t\t\trte_free(alloc_vec);\n+\t\t\treturn -ENOMEM;\n+\t\t}\n+\t}\n+\n+\trte_memcpy(vport->recv_vectors, args.out_buffer, len);\n+\trte_free(alloc_vec);\n+\treturn err;\n+}\n+\n+int\n+idpf_dealloc_vectors(struct idpf_vport *vport)\n+{\n+\tstruct virtchnl2_alloc_vectors *alloc_vec;\n+\tstruct virtchnl2_vector_chunks *vcs;\n+\tstruct idpf_cmd_info args;\n+\tint err, len;\n+\n+\talloc_vec = vport->recv_vectors;\n+\tvcs = &alloc_vec->vchunks;\n+\n+\tlen = sizeof(struct virtchnl2_vector_chunks) +\n+\t\t(vcs->num_vchunks - 1) * sizeof(struct virtchnl2_vector_chunk);\n+\n+\targs.ops = VIRTCHNL2_OP_DEALLOC_VECTORS;\n+\targs.in_args = (u8 *)vcs;\n+\targs.in_args_size = len;\n+\targs.out_buffer = adapter->mbx_resp;\n+\targs.out_size = IDPF_DFLT_MBX_BUF_SIZE;\n+\terr = idpf_execute_vc_cmd(adapter, &args);\n+\tif (err)\n+\t\tPMD_DRV_LOG(ERR, \"Failed to execute command VIRTCHNL2_OP_DEALLOC_VECTORS\");\n+\n+\treturn err;\n+}\n+\n static int\n idpf_ena_dis_one_queue(struct idpf_vport *vport, uint16_t qid,\n \t\t uint32_t type, bool on)\n", "prefixes": [ "12/13" ] }{ "id": 114578, "url": "