Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/59723/?format=api
http://patches.dpdk.org/api/patches/59723/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/patch/20190925171329.63734-13-yong.liu@intel.com/", "project": { "id": 1, "url": "http://patches.dpdk.org/api/projects/1/?format=api", "name": "DPDK", "link_name": "dpdk", "list_id": "dev.dpdk.org", "list_email": "dev@dpdk.org", "web_url": "http://core.dpdk.org", "scm_url": "git://dpdk.org/dpdk", "webscm_url": "http://git.dpdk.org/dpdk", "list_archive_url": "https://inbox.dpdk.org/dev", "list_archive_url_format": "https://inbox.dpdk.org/dev/{}", "commit_url_format": "" }, "msgid": "<20190925171329.63734-13-yong.liu@intel.com>", "list_archive_url": "https://inbox.dpdk.org/dev/20190925171329.63734-13-yong.liu@intel.com", "date": "2019-09-25T17:13:26", "name": "[v3,12/15] vhost: optimize dequeue function of packed ring", "commit_ref": null, "pull_url": null, "state": "superseded", "archived": true, "hash": "ed68613bcccd29281330826570cc7c9f612f0ab5", "submitter": { "id": 17, "url": "http://patches.dpdk.org/api/people/17/?format=api", "name": "Marvin Liu", "email": "yong.liu@intel.com" }, "delegate": { "id": 2642, "url": "http://patches.dpdk.org/api/users/2642/?format=api", "username": "mcoquelin", "first_name": "Maxime", "last_name": "Coquelin", "email": "maxime.coquelin@redhat.com" }, "mbox": "http://patches.dpdk.org/project/dpdk/patch/20190925171329.63734-13-yong.liu@intel.com/mbox/", "series": [ { "id": 6522, "url": "http://patches.dpdk.org/api/series/6522/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=6522", "date": "2019-09-25T17:13:14", "name": "vhost packed ring performance optimization", "version": 3, "mbox": "http://patches.dpdk.org/series/6522/mbox/" } ], "comments": "http://patches.dpdk.org/api/patches/59723/comments/", "check": "fail", "checks": "http://patches.dpdk.org/api/patches/59723/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "<dev-bounces@dpdk.org>", "X-Original-To": "patchwork@dpdk.org", "Delivered-To": "patchwork@dpdk.org", "Received": [ "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id BF1EC1BEE7;\n\tWed, 25 Sep 2019 11:33:10 +0200 (CEST)", "from mga04.intel.com (mga04.intel.com [192.55.52.120])\n\tby dpdk.org (Postfix) with ESMTP id DD7BB1BEC0\n\tfor <dev@dpdk.org>; Wed, 25 Sep 2019 11:32:59 +0200 (CEST)", "from fmsmga004.fm.intel.com ([10.253.24.48])\n\tby fmsmga104.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384;\n\t25 Sep 2019 02:32:48 -0700", "from npg-dpdk-virtual-marvin-dev.sh.intel.com ([10.67.119.142])\n\tby fmsmga004.fm.intel.com with ESMTP; 25 Sep 2019 02:32:46 -0700" ], "X-Amp-Result": "SKIPPED(no attachment in message)", "X-Amp-File-Uploaded": "False", "X-ExtLoop1": "1", "X-IronPort-AV": "E=Sophos;i=\"5.64,547,1559545200\"; d=\"scan'208\";a=\"213986289\"", "From": "Marvin Liu <yong.liu@intel.com>", "To": "maxime.coquelin@redhat.com, tiwei.bie@intel.com, zhihong.wang@intel.com, \n\tstephen@networkplumber.org, gavin.hu@arm.com", "Cc": "dev@dpdk.org,\n\tMarvin Liu <yong.liu@intel.com>", "Date": "Thu, 26 Sep 2019 01:13:26 +0800", "Message-Id": "<20190925171329.63734-13-yong.liu@intel.com>", "X-Mailer": "git-send-email 2.17.1", "In-Reply-To": "<20190925171329.63734-1-yong.liu@intel.com>", "References": "<20190919163643.24130-2-yong.liu@intel.com>\n\t<20190925171329.63734-1-yong.liu@intel.com>", "Subject": "[dpdk-dev] [PATCH v3 12/15] vhost: optimize dequeue function of\n\tpacked ring", "X-BeenThere": "dev@dpdk.org", "X-Mailman-Version": "2.1.15", "Precedence": "list", "List-Id": "DPDK patches and discussions <dev.dpdk.org>", "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n\t<mailto:dev-request@dpdk.org?subject=unsubscribe>", "List-Archive": "<http://mails.dpdk.org/archives/dev/>", "List-Post": "<mailto:dev@dpdk.org>", "List-Help": "<mailto:dev-request@dpdk.org?subject=help>", "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n\t<mailto:dev-request@dpdk.org?subject=subscribe>", "Errors-To": "dev-bounces@dpdk.org", "Sender": "\"dev\" <dev-bounces@dpdk.org>" }, "content": "Optimize vhost device Rx datapath by separate functions. No-chained\nand direct descriptors will be handled by batch and other will be\nhandled one by one as before.\n\nSigned-off-by: Marvin Liu <yong.liu@intel.com>", "diff": "diff --git a/lib/librte_vhost/virtio_net.c b/lib/librte_vhost/virtio_net.c\nindex 9ab95763a..20624efdc 100644\n--- a/lib/librte_vhost/virtio_net.c\n+++ b/lib/librte_vhost/virtio_net.c\n@@ -184,17 +184,6 @@ flush_dequeue_shadow_used_ring_packed(struct virtio_net *dev,\n \tvhost_log_cache_sync(dev, vq);\n }\n \n-static __rte_always_inline void\n-update_shadow_used_ring_packed(struct vhost_virtqueue *vq,\n-\t\t\t uint16_t desc_idx, uint32_t len, uint16_t count)\n-{\n-\tuint16_t i = vq->shadow_used_idx++;\n-\n-\tvq->shadow_used_packed[i].id = desc_idx;\n-\tvq->shadow_used_packed[i].len = len;\n-\tvq->shadow_used_packed[i].count = count;\n-}\n-\n static __rte_always_inline void\n flush_used_batch_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \tuint64_t *lens, uint16_t *ids, uint16_t flags)\n@@ -378,7 +367,7 @@ flush_enqueue_packed(struct virtio_net *dev,\n \t}\n }\n \n-static __rte_unused void\n+static __rte_always_inline void\n flush_dequeue_packed(struct virtio_net *dev, struct vhost_virtqueue *vq)\n {\n \tif (!vq->shadow_used_idx)\n@@ -1784,7 +1773,7 @@ vhost_dequeue_batch_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \treturn -1;\n }\n \n-static __rte_unused int\n+static __rte_always_inline int\n virtio_dev_tx_batch_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \tstruct rte_mempool *mbuf_pool, struct rte_mbuf **pkts)\n {\n@@ -1859,7 +1848,7 @@ vhost_dequeue_single_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \treturn 0;\n }\n \n-static __rte_unused int\n+static __rte_always_inline int\n virtio_dev_tx_single_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \tstruct rte_mempool *mbuf_pool, struct rte_mbuf **pkts)\n {\n@@ -1881,7 +1870,7 @@ virtio_dev_tx_single_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \treturn 0;\n }\n \n-static __rte_unused int\n+static __rte_always_inline int\n virtio_dev_tx_batch_packed_zmbuf(struct virtio_net *dev,\n \t\t\t\t\tstruct vhost_virtqueue *vq,\n \t\t\t\t\tstruct rte_mempool *mbuf_pool,\n@@ -1940,7 +1929,7 @@ virtio_dev_tx_batch_packed_zmbuf(struct virtio_net *dev,\n \treturn -1;\n }\n \n-static __rte_unused int\n+static __rte_always_inline int\n virtio_dev_tx_single_packed_zmbuf(struct virtio_net *dev,\n \tstruct vhost_virtqueue *vq, struct rte_mempool *mbuf_pool,\n \tstruct rte_mbuf **pkts)\n@@ -2017,118 +2006,74 @@ free_zmbuf(struct vhost_virtqueue *vq)\n }\n \n static __rte_noinline uint16_t\n-virtio_dev_tx_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,\n-\tstruct rte_mempool *mbuf_pool, struct rte_mbuf **pkts, uint16_t count)\n+virtio_dev_tx_packed_zmbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n+\tstruct rte_mempool *mbuf_pool, struct rte_mbuf **pkts, uint32_t count)\n {\n-\tuint16_t i;\n-\n-\tif (unlikely(dev->dequeue_zero_copy)) {\n-\t\tstruct zcopy_mbuf *zmbuf, *next;\n-\n-\t\tfor (zmbuf = TAILQ_FIRST(&vq->zmbuf_list);\n-\t\t zmbuf != NULL; zmbuf = next) {\n-\t\t\tnext = TAILQ_NEXT(zmbuf, next);\n+\tuint32_t pkt_idx = 0;\n+\tuint32_t remained = count;\n \n-\t\t\tif (mbuf_is_consumed(zmbuf->mbuf)) {\n-\t\t\t\tupdate_shadow_used_ring_packed(vq,\n-\t\t\t\t\t\tzmbuf->desc_idx,\n-\t\t\t\t\t\t0,\n-\t\t\t\t\t\tzmbuf->desc_count);\n+\tfree_zmbuf(vq);\n \n-\t\t\t\tTAILQ_REMOVE(&vq->zmbuf_list, zmbuf, next);\n-\t\t\t\trestore_mbuf(zmbuf->mbuf);\n-\t\t\t\trte_pktmbuf_free(zmbuf->mbuf);\n-\t\t\t\tput_zmbuf(zmbuf);\n-\t\t\t\tvq->nr_zmbuf -= 1;\n+\tdo {\n+\t\tif (remained >= PACKED_BATCH_SIZE) {\n+\t\t\tif (virtio_dev_tx_batch_packed_zmbuf(dev, vq,\n+\t\t\t\t\t\t\t mbuf_pool,\n+\t\t\t\t\t\t\t &pkts[pkt_idx])) {\n+\t\t\t\tpkt_idx += PACKED_BATCH_SIZE;\n+\t\t\t\tremained -= PACKED_BATCH_SIZE;\n+\t\t\t\tcontinue;\n \t\t\t}\n \t\t}\n+\t\tif (virtio_dev_tx_single_packed_zmbuf(dev, vq, mbuf_pool,\n+\t\t\t\t\t\t &pkts[pkt_idx]))\n+\t\t\tbreak;\n \n-\t\tif (likely(vq->shadow_used_idx)) {\n-\t\t\tflush_dequeue_shadow_used_ring_packed(dev, vq);\n-\t\t\tvhost_vring_call_packed(dev, vq);\n-\t\t}\n-\t}\n-\n-\tVHOST_LOG_DEBUG(VHOST_DATA, \"(%d) %s\\n\", dev->vid, __func__);\n+\t\tpkt_idx++;\n+\t\tremained--;\n+\t} while (remained);\n \n-\tcount = RTE_MIN(count, MAX_PKT_BURST);\n-\tVHOST_LOG_DEBUG(VHOST_DATA, \"(%d) about to dequeue %u buffers\\n\",\n-\t\t\tdev->vid, count);\n+\tif (pkt_idx)\n+\t\tvhost_vring_call_packed(dev, vq);\n \n-\tfor (i = 0; i < count; i++) {\n-\t\tstruct buf_vector buf_vec[BUF_VECTOR_MAX];\n-\t\tuint16_t buf_id;\n-\t\tuint32_t dummy_len;\n-\t\tuint16_t desc_count, nr_vec = 0;\n-\t\tint err;\n+\treturn pkt_idx;\n+}\n \n-\t\tif (unlikely(fill_vec_buf_packed(dev, vq,\n-\t\t\t\t\t\tvq->last_avail_idx, &desc_count,\n-\t\t\t\t\t\tbuf_vec, &nr_vec,\n-\t\t\t\t\t\t&buf_id, &dummy_len,\n-\t\t\t\t\t\tVHOST_ACCESS_RO) < 0))\n-\t\t\tbreak;\n+static __rte_noinline uint16_t\n+virtio_dev_tx_packed(struct virtio_net *dev, struct vhost_virtqueue *vq,\n+\tstruct rte_mempool *mbuf_pool, struct rte_mbuf **pkts, uint32_t count)\n+{\n+\tuint32_t pkt_idx = 0;\n+\tuint32_t remained = count;\n \n-\t\tif (likely(dev->dequeue_zero_copy == 0))\n-\t\t\tupdate_shadow_used_ring_packed(vq, buf_id, 0,\n-\t\t\t\t\tdesc_count);\n+\tdo {\n+\t\trte_prefetch0(&vq->desc_packed[vq->last_avail_idx &\n+\t\t (vq->size - 1)]);\n \n-\t\tpkts[i] = rte_pktmbuf_alloc(mbuf_pool);\n-\t\tif (unlikely(pkts[i] == NULL)) {\n-\t\t\tRTE_LOG(ERR, VHOST_DATA,\n-\t\t\t\t\"Failed to allocate memory for mbuf.\\n\");\n-\t\t\tbreak;\n+\t\tif (remained >= PACKED_BATCH_SIZE) {\n+\t\t\tif (!virtio_dev_tx_batch_packed(dev, vq, mbuf_pool,\n+\t\t\t\t\t\t\t&pkts[pkt_idx])) {\n+\t\t\t\tflush_dequeue_packed(dev, vq);\n+\t\t\t\tpkt_idx += PACKED_BATCH_SIZE;\n+\t\t\t\tremained -= PACKED_BATCH_SIZE;\n+\t\t\t\tcontinue;\n+\t\t\t}\n \t\t}\n \n-\t\terr = copy_desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts[i],\n-\t\t\t\tmbuf_pool);\n-\t\tif (unlikely(err)) {\n-\t\t\trte_pktmbuf_free(pkts[i]);\n+\t\tif (virtio_dev_tx_single_packed(dev, vq, mbuf_pool,\n+\t\t\t\t\t\t&pkts[pkt_idx]))\n \t\t\tbreak;\n-\t\t}\n-\n-\t\tif (unlikely(dev->dequeue_zero_copy)) {\n-\t\t\tstruct zcopy_mbuf *zmbuf;\n-\n-\t\t\tzmbuf = get_zmbuf(vq);\n-\t\t\tif (!zmbuf) {\n-\t\t\t\trte_pktmbuf_free(pkts[i]);\n-\t\t\t\tbreak;\n-\t\t\t}\n-\t\t\tzmbuf->mbuf = pkts[i];\n-\t\t\tzmbuf->desc_idx = buf_id;\n-\t\t\tzmbuf->desc_count = desc_count;\n \n-\t\t\t/*\n-\t\t\t * Pin lock the mbuf; we will check later to see\n-\t\t\t * whether the mbuf is freed (when we are the last\n-\t\t\t * user) or not. If that's the case, we then could\n-\t\t\t * update the used ring safely.\n-\t\t\t */\n-\t\t\trte_mbuf_refcnt_update(pkts[i], 1);\n-\n-\t\t\tvq->nr_zmbuf += 1;\n-\t\t\tTAILQ_INSERT_TAIL(&vq->zmbuf_list, zmbuf, next);\n-\t\t}\n-\n-\t\tvq->last_avail_idx += desc_count;\n-\t\tif (vq->last_avail_idx >= vq->size) {\n-\t\t\tvq->last_avail_idx -= vq->size;\n-\t\t\tvq->avail_wrap_counter ^= 1;\n-\t\t}\n-\t}\n+\t\tpkt_idx++;\n+\t\tremained--;\n+\t\tflush_dequeue_packed(dev, vq);\n+\t} while (remained);\n \n-\tif (likely(dev->dequeue_zero_copy == 0)) {\n-\t\tdo_data_copy_dequeue(vq);\n-\t\tif (unlikely(i < count))\n-\t\t\tvq->shadow_used_idx = i;\n-\t\tif (likely(vq->shadow_used_idx)) {\n-\t\t\tflush_dequeue_shadow_used_ring_packed(dev, vq);\n-\t\t\tvhost_vring_call_packed(dev, vq);\n-\t\t}\n+\tif (pkt_idx) {\n+\t\tif (vq->shadow_used_idx)\n+\t\t\tdo_data_copy_dequeue(vq);\n \t}\n \n-\treturn i;\n+\treturn pkt_idx;\n }\n \n uint16_t\n@@ -2204,9 +2149,14 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,\n \t\tcount -= 1;\n \t}\n \n-\tif (vq_is_packed(dev))\n-\t\tcount = virtio_dev_tx_packed(dev, vq, mbuf_pool, pkts, count);\n-\telse\n+\tif (vq_is_packed(dev)) {\n+\t\tif (unlikely(dev->dequeue_zero_copy))\n+\t\t\tcount = virtio_dev_tx_packed_zmbuf(dev, vq, mbuf_pool,\n+\t\t\t\t\t\t\t pkts, count);\n+\t\telse\n+\t\t\tcount = virtio_dev_tx_packed(dev, vq, mbuf_pool, pkts,\n+\t\t\t\t\t\t count);\n+\t} else\n \t\tcount = virtio_dev_tx_split(dev, vq, mbuf_pool, pkts, count);\n \n out:\n", "prefixes": [ "v3", "12/15" ] }{ "id": 59723, "url": "