Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/110646/?format=api
https://patches.dpdk.org/api/patches/110646/?format=api", "web_url": "https://patches.dpdk.org/project/dpdk/patch/20220505062335.25445-4-xuan.ding@intel.com/", "project": { "id": 1, "url": "https://patches.dpdk.org/api/projects/1/?format=api", "name": "DPDK", "link_name": "dpdk", "list_id": "dev.dpdk.org", "list_email": "dev@dpdk.org", "web_url": "http://core.dpdk.org", "scm_url": "git://dpdk.org/dpdk", "webscm_url": "http://git.dpdk.org/dpdk", "list_archive_url": "https://inbox.dpdk.org/dev", "list_archive_url_format": "https://inbox.dpdk.org/dev/{}", "commit_url_format": "" }, "msgid": "<20220505062335.25445-4-xuan.ding@intel.com>", "list_archive_url": "https://inbox.dpdk.org/dev/20220505062335.25445-4-xuan.ding@intel.com", "date": "2022-05-05T06:23:33", "name": "[v4,3/5] vhost: merge sync and async descriptor to mbuf filling", "commit_ref": null, "pull_url": null, "state": "superseded", "archived": true, "hash": "9d8d37ae2afa90a2834499cd4bf0cb23a87c9f66", "submitter": { "id": 1401, "url": "https://patches.dpdk.org/api/people/1401/?format=api", "name": "Ding, Xuan", "email": "xuan.ding@intel.com" }, "delegate": { "id": 2642, "url": "https://patches.dpdk.org/api/users/2642/?format=api", "username": "mcoquelin", "first_name": "Maxime", "last_name": "Coquelin", "email": "maxime.coquelin@redhat.com" }, "mbox": "https://patches.dpdk.org/project/dpdk/patch/20220505062335.25445-4-xuan.ding@intel.com/mbox/", "series": [ { "id": 22784, "url": "https://patches.dpdk.org/api/series/22784/?format=api", "web_url": "https://patches.dpdk.org/project/dpdk/list/?series=22784", "date": "2022-05-05T06:23:30", "name": "vhost: support async dequeue data path", "version": 4, "mbox": "https://patches.dpdk.org/series/22784/mbox/" } ], "comments": "https://patches.dpdk.org/api/patches/110646/comments/", "check": "success", "checks": "https://patches.dpdk.org/api/patches/110646/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "<dev-bounces@dpdk.org>", "X-Original-To": "patchwork@inbox.dpdk.org", "Delivered-To": "patchwork@inbox.dpdk.org", "Received": [ "from mails.dpdk.org (mails.dpdk.org [217.70.189.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id 8B24AA00BE;\n\tThu, 5 May 2022 08:31:07 +0200 (CEST)", "from [217.70.189.124] (localhost [127.0.0.1])\n\tby mails.dpdk.org (Postfix) with ESMTP id 0130542829;\n\tThu, 5 May 2022 08:30:53 +0200 (CEST)", "from mga02.intel.com (mga02.intel.com [134.134.136.20])\n by mails.dpdk.org (Postfix) with ESMTP id F047840042\n for <dev@dpdk.org>; Thu, 5 May 2022 08:30:47 +0200 (CEST)", "from orsmga007.jf.intel.com ([10.7.209.58])\n by orsmga101.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 04 May 2022 23:27:16 -0700", "from npg-dpdk-xuan-cbdma.sh.intel.com ([10.67.110.228])\n by orsmga007.jf.intel.com with ESMTP; 04 May 2022 23:27:14 -0700" ], "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/simple;\n d=intel.com; i=@intel.com; q=dns/txt; s=Intel;\n t=1651732248; x=1683268248;\n h=from:to:cc:subject:date:message-id:in-reply-to: references;\n bh=lkSDT58h1L5z/SScvUySl9U3gvqJJVjo1S7bX+D0odQ=;\n b=Hn7KAaBONs/IMhus6JO274q8f/WsntjAYt3In4xy1ZljSC/mPx0GlM85\n c8vqv+vGML4Qv+h6kWr8vgIhiV/HmmJu8R70GkBTDc2MkWJtoZ46t9wrw\n TnOis5ewsBm4aoXqJpBealaJfh1On1SB+nf3NYFX3SQQTAfL5LfU+dibo\n uvP4iIIS8KdrWrAWOwkwyxsauH6W2CRf2zad9BtS2bO+rHvP9zb+i4jOL\n tkT9gyMww1t1rSf51bzpit179kzPsIYZCAn+vHu1Uv6R6p6MxdMEpg28P\n Uoe/FvJR07CQlydTpjDGV5hg9NBcBlmObjoXKhAmx5uZBZ0zyREev8t1A w==;", "X-IronPort-AV": [ "E=McAfee;i=\"6400,9594,10337\"; a=\"255488244\"", "E=Sophos;i=\"5.91,200,1647327600\"; d=\"scan'208\";a=\"255488244\"", "E=Sophos;i=\"5.91,200,1647327600\"; d=\"scan'208\";a=\"563092991\"" ], "X-ExtLoop1": "1", "From": "xuan.ding@intel.com", "To": "maxime.coquelin@redhat.com,\n\tchenbo.xia@intel.com", "Cc": "dev@dpdk.org, jiayu.hu@intel.com, cheng1.jiang@intel.com,\n sunil.pai.g@intel.com, liangma@liangbit.com,\n Xuan Ding <xuan.ding@intel.com>", "Subject": "[PATCH v4 3/5] vhost: merge sync and async descriptor to mbuf filling", "Date": "Thu, 5 May 2022 06:23:33 +0000", "Message-Id": "<20220505062335.25445-4-xuan.ding@intel.com>", "X-Mailer": "git-send-email 2.17.1", "In-Reply-To": "<20220505062335.25445-1-xuan.ding@intel.com>", "References": "<20220407152546.38167-1-xuan.ding@intel.com>\n <20220505062335.25445-1-xuan.ding@intel.com>", "X-BeenThere": "dev@dpdk.org", "X-Mailman-Version": "2.1.29", "Precedence": "list", "List-Id": "DPDK patches and discussions <dev.dpdk.org>", "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n <mailto:dev-request@dpdk.org?subject=unsubscribe>", "List-Archive": "<http://mails.dpdk.org/archives/dev/>", "List-Post": "<mailto:dev@dpdk.org>", "List-Help": "<mailto:dev-request@dpdk.org?subject=help>", "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n <mailto:dev-request@dpdk.org?subject=subscribe>", "Errors-To": "dev-bounces@dpdk.org" }, "content": "From: Xuan Ding <xuan.ding@intel.com>\n\nThis patch refactors copy_desc_to_mbuf() used by the sync\npath to support both sync and async descriptor to mbuf filling.\n\nSigned-off-by: Xuan Ding <xuan.ding@intel.com>\nReviewed-by: Maxime Coquelin <maxime.coquelin@redhat.com>\n---\n lib/vhost/vhost.h | 1 +\n lib/vhost/virtio_net.c | 48 ++++++++++++++++++++++++++++++++----------\n 2 files changed, 38 insertions(+), 11 deletions(-)", "diff": "diff --git a/lib/vhost/vhost.h b/lib/vhost/vhost.h\nindex a9edc271aa..00744b234f 100644\n--- a/lib/vhost/vhost.h\n+++ b/lib/vhost/vhost.h\n@@ -180,6 +180,7 @@ struct async_inflight_info {\n \tstruct rte_mbuf *mbuf;\n \tuint16_t descs; /* num of descs inflight */\n \tuint16_t nr_buffers; /* num of buffers inflight for packed ring */\n+\tstruct virtio_net_hdr nethdr;\n };\n \n struct vhost_async {\ndiff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c\nindex a9e2dcd9ce..5904839d5c 100644\n--- a/lib/vhost/virtio_net.c\n+++ b/lib/vhost/virtio_net.c\n@@ -2487,10 +2487,10 @@ copy_vnet_hdr_from_desc(struct virtio_net_hdr *hdr,\n }\n \n static __rte_always_inline int\n-copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n+desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \t\t struct buf_vector *buf_vec, uint16_t nr_vec,\n \t\t struct rte_mbuf *m, struct rte_mempool *mbuf_pool,\n-\t\t bool legacy_ol_flags)\n+\t\t bool legacy_ol_flags, uint16_t slot_idx, bool is_async)\n {\n \tuint32_t buf_avail, buf_offset, buf_len;\n \tuint64_t buf_addr, buf_iova;\n@@ -2501,6 +2501,8 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \tstruct virtio_net_hdr *hdr = NULL;\n \t/* A counter to avoid desc dead loop chain */\n \tuint16_t vec_idx = 0;\n+\tstruct vhost_async *async = vq->async;\n+\tstruct async_inflight_info *pkts_info;\n \n \tbuf_addr = buf_vec[vec_idx].buf_addr;\n \tbuf_iova = buf_vec[vec_idx].buf_iova;\n@@ -2538,6 +2540,7 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \t\tif (unlikely(++vec_idx >= nr_vec))\n \t\t\tgoto error;\n \t\tbuf_addr = buf_vec[vec_idx].buf_addr;\n+\t\tbuf_iova = buf_vec[vec_idx].buf_iova;\n \t\tbuf_len = buf_vec[vec_idx].buf_len;\n \n \t\tbuf_offset = 0;\n@@ -2553,12 +2556,25 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \n \tmbuf_offset = 0;\n \tmbuf_avail = m->buf_len - RTE_PKTMBUF_HEADROOM;\n+\n+\tif (is_async) {\n+\t\tpkts_info = async->pkts_info;\n+\t\tif (async_iter_initialize(dev, async))\n+\t\t\treturn -1;\n+\t}\n+\n \twhile (1) {\n \t\tcpy_len = RTE_MIN(buf_avail, mbuf_avail);\n \n-\t\tsync_fill_seg(dev, vq, cur, mbuf_offset,\n-\t\t\t buf_addr + buf_offset,\n-\t\t\t buf_iova + buf_offset, cpy_len, false);\n+\t\tif (is_async) {\n+\t\t\tif (async_fill_seg(dev, vq, cur, mbuf_offset,\n+\t\t\t\t\t buf_iova + buf_offset, cpy_len, false) < 0)\n+\t\t\t\tgoto error;\n+\t\t} else {\n+\t\t\tsync_fill_seg(dev, vq, cur, mbuf_offset,\n+\t\t\t\t buf_addr + buf_offset,\n+\t\t\t\t buf_iova + buf_offset, cpy_len, false);\n+\t\t}\n \n \t\tmbuf_avail -= cpy_len;\n \t\tmbuf_offset += cpy_len;\n@@ -2607,11 +2623,20 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \tprev->data_len = mbuf_offset;\n \tm->pkt_len += mbuf_offset;\n \n-\tif (hdr)\n-\t\tvhost_dequeue_offload(dev, hdr, m, legacy_ol_flags);\n+\tif (is_async) {\n+\t\tasync_iter_finalize(async);\n+\t\tif (hdr)\n+\t\t\tpkts_info[slot_idx].nethdr = *hdr;\n+\t} else {\n+\t\tif (hdr)\n+\t\t\tvhost_dequeue_offload(dev, hdr, m, legacy_ol_flags);\n+\t}\n \n \treturn 0;\n error:\n+\tif (is_async)\n+\t\tasync_iter_cancel(async);\n+\n \treturn -1;\n }\n \n@@ -2743,8 +2768,8 @@ virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \t\t\tbreak;\n \t\t}\n \n-\t\terr = copy_desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts[i],\n-\t\t\t\tmbuf_pool, legacy_ol_flags);\n+\t\terr = desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts[i],\n+\t\t\t\t mbuf_pool, legacy_ol_flags, 0, false);\n \t\tif (unlikely(err)) {\n \t\t\tif (!allocerr_warned) {\n \t\t\t\tVHOST_LOG_DATA(ERR, \"(%s) failed to copy desc to mbuf.\\n\",\n@@ -2755,6 +2780,7 @@ virtio_dev_tx_split(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \t\t\ti++;\n \t\t\tbreak;\n \t\t}\n+\n \t}\n \n \tif (dropped)\n@@ -2936,8 +2962,8 @@ vhost_dequeue_single_packed(struct virtio_net *dev,\n \t\treturn -1;\n \t}\n \n-\terr = copy_desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts,\n-\t\t\t\tmbuf_pool, legacy_ol_flags);\n+\terr = desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts,\n+\t\t\t mbuf_pool, legacy_ol_flags, 0, false);\n \tif (unlikely(err)) {\n \t\tif (!allocerr_warned) {\n \t\t\tVHOST_LOG_DATA(ERR, \"(%s) failed to copy desc to mbuf.\\n\",\n", "prefixes": [ "v4", "3/5" ] }{ "id": 110646, "url": "