Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/41668/?format=api
http://patches.dpdk.org/api/patches/41668/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/patch/20180627144959.17277-4-maxime.coquelin@redhat.com/", "project": { "id": 1, "url": "http://patches.dpdk.org/api/projects/1/?format=api", "name": "DPDK", "link_name": "dpdk", "list_id": "dev.dpdk.org", "list_email": "dev@dpdk.org", "web_url": "http://core.dpdk.org", "scm_url": "git://dpdk.org/dpdk", "webscm_url": "http://git.dpdk.org/dpdk", "list_archive_url": "https://inbox.dpdk.org/dev", "list_archive_url_format": "https://inbox.dpdk.org/dev/{}", "commit_url_format": "" }, "msgid": "<20180627144959.17277-4-maxime.coquelin@redhat.com>", "list_archive_url": "https://inbox.dpdk.org/dev/20180627144959.17277-4-maxime.coquelin@redhat.com", "date": "2018-06-27T14:49:55", "name": "[v3,3/7] vhost: use buffer vectors in dequeue path", "commit_ref": null, "pull_url": null, "state": "superseded", "archived": true, "hash": "42ad5304d8bc29a8ef3f90ad0d4495a482b46c91", "submitter": { "id": 512, "url": "http://patches.dpdk.org/api/people/512/?format=api", "name": "Maxime Coquelin", "email": "maxime.coquelin@redhat.com" }, "delegate": { "id": 2642, "url": "http://patches.dpdk.org/api/users/2642/?format=api", "username": "mcoquelin", "first_name": "Maxime", "last_name": "Coquelin", "email": "maxime.coquelin@redhat.com" }, "mbox": "http://patches.dpdk.org/project/dpdk/patch/20180627144959.17277-4-maxime.coquelin@redhat.com/mbox/", "series": [ { "id": 266, "url": "http://patches.dpdk.org/api/series/266/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=266", "date": "2018-06-27T14:49:52", "name": "vhost: generalize buffer vectors", "version": 3, "mbox": "http://patches.dpdk.org/series/266/mbox/" } ], "comments": "http://patches.dpdk.org/api/patches/41668/comments/", "check": "success", "checks": "http://patches.dpdk.org/api/patches/41668/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "<dev-bounces@dpdk.org>", "X-Original-To": "patchwork@dpdk.org", "Delivered-To": "patchwork@dpdk.org", "Received": [ "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id 9F3B91BF55;\n\tWed, 27 Jun 2018 16:50:22 +0200 (CEST)", "from mx1.redhat.com (mx3-rdu2.redhat.com [66.187.233.73])\n\tby dpdk.org (Postfix) with ESMTP id 233171BF43\n\tfor <dev@dpdk.org>; Wed, 27 Jun 2018 16:50:17 +0200 (CEST)", "from smtp.corp.redhat.com\n\t(int-mx06.intmail.prod.int.rdu2.redhat.com [10.11.54.6])\n\t(using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits))\n\t(No client certificate requested)\n\tby mx1.redhat.com (Postfix) with ESMTPS id B87CF818F047;\n\tWed, 27 Jun 2018 14:50:16 +0000 (UTC)", "from localhost.localdomain (ovpn-112-39.ams2.redhat.com\n\t[10.36.112.39])\n\tby smtp.corp.redhat.com (Postfix) with ESMTP id B401320389E0;\n\tWed, 27 Jun 2018 14:50:15 +0000 (UTC)" ], "From": "Maxime Coquelin <maxime.coquelin@redhat.com>", "To": "tiwei.bie@intel.com,\n\tzhihong.wang@intel.com,\n\tdev@dpdk.org", "Cc": "Maxime Coquelin <maxime.coquelin@redhat.com>", "Date": "Wed, 27 Jun 2018 16:49:55 +0200", "Message-Id": "<20180627144959.17277-4-maxime.coquelin@redhat.com>", "In-Reply-To": "<20180627144959.17277-1-maxime.coquelin@redhat.com>", "References": "<20180627144959.17277-1-maxime.coquelin@redhat.com>", "X-Scanned-By": "MIMEDefang 2.78 on 10.11.54.6", "X-Greylist": [ "Sender IP whitelisted, not delayed by milter-greylist-4.5.16\n\t(mx1.redhat.com [10.11.55.8]); Wed, 27 Jun 2018 14:50:16 +0000 (UTC)", "inspected by milter-greylist-4.5.16 (mx1.redhat.com\n\t[10.11.55.8]); \n\tWed, 27 Jun 2018 14:50:16 +0000 (UTC) for IP:'10.11.54.6'\n\tDOMAIN:'int-mx06.intmail.prod.int.rdu2.redhat.com'\n\tHELO:'smtp.corp.redhat.com' FROM:'maxime.coquelin@redhat.com' RCPT:''" ], "Subject": "[dpdk-dev] [PATCH v3 3/7] vhost: use buffer vectors in dequeue path", "X-BeenThere": "dev@dpdk.org", "X-Mailman-Version": "2.1.15", "Precedence": "list", "List-Id": "DPDK patches and discussions <dev.dpdk.org>", "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n\t<mailto:dev-request@dpdk.org?subject=unsubscribe>", "List-Archive": "<http://mails.dpdk.org/archives/dev/>", "List-Post": "<mailto:dev@dpdk.org>", "List-Help": "<mailto:dev-request@dpdk.org?subject=help>", "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n\t<mailto:dev-request@dpdk.org?subject=subscribe>", "Errors-To": "dev-bounces@dpdk.org", "Sender": "\"dev\" <dev-bounces@dpdk.org>" }, "content": "To ease packed ring layout integration, this patch makes\nthe dequeue path to re-use buffer vectors implemented for\nenqueue path.\n\nDoing this, copy_desc_to_mbuf() is now ring layout type\nagnostic.\n\nSigned-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>\n---\n lib/librte_vhost/virtio_net.c | 143 ++++++++++--------------------------------\n 1 file changed, 33 insertions(+), 110 deletions(-)", "diff": "diff --git a/lib/librte_vhost/virtio_net.c b/lib/librte_vhost/virtio_net.c\nindex ec4bcc400..4816e8003 100644\n--- a/lib/librte_vhost/virtio_net.c\n+++ b/lib/librte_vhost/virtio_net.c\n@@ -750,11 +750,9 @@ put_zmbuf(struct zcopy_mbuf *zmbuf)\n \n static __rte_always_inline int\n copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n-\t\t struct vring_desc *descs, uint16_t max_desc,\n-\t\t struct rte_mbuf *m, uint16_t desc_idx,\n-\t\t struct rte_mempool *mbuf_pool)\n+\t\t struct buf_vector *buf_vec, uint16_t nr_vec,\n+\t\t struct rte_mbuf *m, struct rte_mempool *mbuf_pool)\n {\n-\tstruct vring_desc *desc;\n \tuint64_t desc_addr, desc_gaddr;\n \tuint32_t desc_avail, desc_offset;\n \tuint32_t mbuf_avail, mbuf_offset;\n@@ -764,24 +762,18 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \tstruct virtio_net_hdr tmp_hdr;\n \tstruct virtio_net_hdr *hdr = NULL;\n \t/* A counter to avoid desc dead loop chain */\n-\tuint32_t nr_desc = 1;\n+\tuint16_t vec_idx = 0;\n \tstruct batch_copy_elem *batch_copy = vq->batch_copy_elems;\n \tint error = 0;\n \n-\tdesc = &descs[desc_idx];\n-\tif (unlikely((desc->len < dev->vhost_hlen)) ||\n-\t\t\t(desc->flags & VRING_DESC_F_INDIRECT)) {\n-\t\terror = -1;\n-\t\tgoto out;\n-\t}\n-\n-\tdesc_chunck_len = desc->len;\n-\tdesc_gaddr = desc->addr;\n+\tdesc_chunck_len = buf_vec[vec_idx].buf_len;\n+\tdesc_gaddr = buf_vec[vec_idx].buf_addr;\n \tdesc_addr = vhost_iova_to_vva(dev,\n \t\t\t\t\tvq, desc_gaddr,\n \t\t\t\t\t&desc_chunck_len,\n \t\t\t\t\tVHOST_ACCESS_RO);\n-\tif (unlikely(!desc_addr)) {\n+\tif (unlikely(buf_vec[vec_idx].buf_len < dev->vhost_hlen ||\n+\t\t\t\t!desc_addr)) {\n \t\terror = -1;\n \t\tgoto out;\n \t}\n@@ -828,16 +820,12 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \t * for Tx: the first for storing the header, and others\n \t * for storing the data.\n \t */\n-\tif (likely((desc->len == dev->vhost_hlen) &&\n-\t\t (desc->flags & VRING_DESC_F_NEXT) != 0)) {\n-\t\tdesc = &descs[desc->next];\n-\t\tif (unlikely(desc->flags & VRING_DESC_F_INDIRECT)) {\n-\t\t\terror = -1;\n+\tif (likely(buf_vec[vec_idx].buf_len == dev->vhost_hlen)) {\n+\t\tif (unlikely(++vec_idx >= nr_vec))\n \t\t\tgoto out;\n-\t\t}\n \n-\t\tdesc_chunck_len = desc->len;\n-\t\tdesc_gaddr = desc->addr;\n+\t\tdesc_chunck_len = buf_vec[vec_idx].buf_len;\n+\t\tdesc_gaddr = buf_vec[vec_idx].buf_addr;\n \t\tdesc_addr = vhost_iova_to_vva(dev,\n \t\t\t\t\t\t\tvq, desc_gaddr,\n \t\t\t\t\t\t\t&desc_chunck_len,\n@@ -848,10 +836,9 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \t\t}\n \n \t\tdesc_offset = 0;\n-\t\tdesc_avail = desc->len;\n-\t\tnr_desc += 1;\n+\t\tdesc_avail = buf_vec[vec_idx].buf_len;\n \t} else {\n-\t\tdesc_avail = desc->len - dev->vhost_hlen;\n+\t\tdesc_avail = buf_vec[vec_idx].buf_len - dev->vhost_hlen;\n \n \t\tif (unlikely(desc_chunck_len < dev->vhost_hlen)) {\n \t\t\tdesc_chunck_len = desc_avail;\n@@ -906,7 +893,8 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \t\t\tif (likely(cpy_len > MAX_BATCH_LEN ||\n \t\t\t\t vq->batch_copy_nb_elems >= vq->size ||\n \t\t\t\t (hdr && cur == m) ||\n-\t\t\t\t desc->len != desc_chunck_len)) {\n+\t\t\t\t buf_vec[vec_idx].buf_len !=\n+\t\t\t\t\t\tdesc_chunck_len)) {\n \t\t\t\trte_memcpy(rte_pktmbuf_mtod_offset(cur, void *,\n \t\t\t\t\t\t\t\t mbuf_offset),\n \t\t\t\t\t (void *)((uintptr_t)(desc_addr +\n@@ -933,22 +921,11 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \n \t\t/* This desc reaches to its end, get the next one */\n \t\tif (desc_avail == 0) {\n-\t\t\tif ((desc->flags & VRING_DESC_F_NEXT) == 0)\n+\t\t\tif (++vec_idx >= nr_vec)\n \t\t\t\tbreak;\n \n-\t\t\tif (unlikely(desc->next >= max_desc ||\n-\t\t\t\t ++nr_desc > max_desc)) {\n-\t\t\t\terror = -1;\n-\t\t\t\tgoto out;\n-\t\t\t}\n-\t\t\tdesc = &descs[desc->next];\n-\t\t\tif (unlikely(desc->flags & VRING_DESC_F_INDIRECT)) {\n-\t\t\t\terror = -1;\n-\t\t\t\tgoto out;\n-\t\t\t}\n-\n-\t\t\tdesc_chunck_len = desc->len;\n-\t\t\tdesc_gaddr = desc->addr;\n+\t\t\tdesc_chunck_len = buf_vec[vec_idx].buf_len;\n+\t\t\tdesc_gaddr = buf_vec[vec_idx].buf_addr;\n \t\t\tdesc_addr = vhost_iova_to_vva(dev,\n \t\t\t\t\t\t\tvq, desc_gaddr,\n \t\t\t\t\t\t\t&desc_chunck_len,\n@@ -961,7 +938,7 @@ copy_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,\n \t\t\trte_prefetch0((void *)(uintptr_t)desc_addr);\n \n \t\t\tdesc_offset = 0;\n-\t\t\tdesc_avail = desc->len;\n+\t\t\tdesc_avail = buf_vec[vec_idx].buf_len;\n \n \t\t\tPRINT_PACKET(dev, (uintptr_t)desc_addr,\n \t\t\t\t\t(uint32_t)desc_chunck_len, 0);\n@@ -1085,11 +1062,8 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,\n \tstruct virtio_net *dev;\n \tstruct rte_mbuf *rarp_mbuf = NULL;\n \tstruct vhost_virtqueue *vq;\n-\tuint32_t desc_indexes[MAX_PKT_BURST];\n-\tuint32_t used_idx;\n \tuint32_t i = 0;\n \tuint16_t free_entries;\n-\tuint16_t avail_idx;\n \n \tdev = get_device(vid);\n \tif (!dev)\n@@ -1135,7 +1109,6 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,\n \t\t\tnext = TAILQ_NEXT(zmbuf, next);\n \n \t\t\tif (mbuf_is_consumed(zmbuf->mbuf)) {\n-\t\t\t\tused_idx = vq->last_used_idx++ & (vq->size - 1);\n \t\t\t\tupdate_shadow_used_ring(vq, zmbuf->desc_idx, 0);\n \t\t\t\tnr_updated += 1;\n \n@@ -1182,89 +1155,43 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,\n \t}\n \n \tfree_entries = *((volatile uint16_t *)&vq->avail->idx) -\n-\t\t\tvq->last_avail_idx;\n+\t\tvq->last_avail_idx;\n \tif (free_entries == 0)\n \t\tgoto out;\n \n \tVHOST_LOG_DEBUG(VHOST_DATA, \"(%d) %s\\n\", dev->vid, __func__);\n \n-\t/* Prefetch available and used ring */\n-\tavail_idx = vq->last_avail_idx & (vq->size - 1);\n-\tused_idx = vq->last_used_idx & (vq->size - 1);\n-\trte_prefetch0(&vq->avail->ring[avail_idx]);\n-\trte_prefetch0(&vq->used->ring[used_idx]);\n-\n \tcount = RTE_MIN(count, MAX_PKT_BURST);\n \tcount = RTE_MIN(count, free_entries);\n \tVHOST_LOG_DEBUG(VHOST_DATA, \"(%d) about to dequeue %u buffers\\n\",\n \t\t\tdev->vid, count);\n \n-\t/* Retrieve all of the head indexes first to avoid caching issues. */\n \tfor (i = 0; i < count; i++) {\n-\t\tavail_idx = (vq->last_avail_idx + i) & (vq->size - 1);\n-\t\tused_idx = (vq->last_used_idx + i) & (vq->size - 1);\n-\t\tdesc_indexes[i] = vq->avail->ring[avail_idx];\n-\n-\t\tif (likely(dev->dequeue_zero_copy == 0))\n-\t\t\tupdate_shadow_used_ring(vq, desc_indexes[i], 0);\n-\t}\n-\n-\t/* Prefetch descriptor index. */\n-\trte_prefetch0(&vq->desc[desc_indexes[0]]);\n-\tfor (i = 0; i < count; i++) {\n-\t\tstruct vring_desc *desc, *idesc = NULL;\n-\t\tuint16_t sz, idx;\n-\t\tuint64_t dlen;\n+\t\tstruct buf_vector buf_vec[BUF_VECTOR_MAX];\n+\t\tuint16_t head_idx, dummy_len;\n+\t\tuint32_t nr_vec = 0;\n \t\tint err;\n \n-\t\tif (likely(i + 1 < count))\n-\t\t\trte_prefetch0(&vq->desc[desc_indexes[i + 1]]);\n-\n-\t\tif (vq->desc[desc_indexes[i]].flags & VRING_DESC_F_INDIRECT) {\n-\t\t\tdlen = vq->desc[desc_indexes[i]].len;\n-\t\t\tdesc = (struct vring_desc *)(uintptr_t)\n-\t\t\t\tvhost_iova_to_vva(dev, vq,\n-\t\t\t\t\t\tvq->desc[desc_indexes[i]].addr,\n-\t\t\t\t\t\t&dlen,\n-\t\t\t\t\t\tVHOST_ACCESS_RO);\n-\t\t\tif (unlikely(!desc))\n-\t\t\t\tbreak;\n-\n-\t\t\tif (unlikely(dlen < vq->desc[desc_indexes[i]].len)) {\n-\t\t\t\t/*\n-\t\t\t\t * The indirect desc table is not contiguous\n-\t\t\t\t * in process VA space, we have to copy it.\n-\t\t\t\t */\n-\t\t\t\tidesc = alloc_copy_ind_table(dev, vq,\n-\t\t\t\t\t\t&vq->desc[desc_indexes[i]]);\n-\t\t\t\tif (unlikely(!idesc))\n-\t\t\t\t\tbreak;\n-\n-\t\t\t\tdesc = idesc;\n-\t\t\t}\n+\t\tif (unlikely(fill_vec_buf(dev, vq,\n+\t\t\t\t\t\tvq->last_avail_idx + i,\n+\t\t\t\t\t\t&nr_vec, buf_vec,\n+\t\t\t\t\t\t&head_idx, &dummy_len) < 0))\n+\t\t\tbreak;\n \n-\t\t\trte_prefetch0(desc);\n-\t\t\tsz = vq->desc[desc_indexes[i]].len / sizeof(*desc);\n-\t\t\tidx = 0;\n-\t\t} else {\n-\t\t\tdesc = vq->desc;\n-\t\t\tsz = vq->size;\n-\t\t\tidx = desc_indexes[i];\n-\t\t}\n+\t\tif (likely(dev->dequeue_zero_copy == 0))\n+\t\t\tupdate_shadow_used_ring(vq, head_idx, 0);\n \n \t\tpkts[i] = rte_pktmbuf_alloc(mbuf_pool);\n \t\tif (unlikely(pkts[i] == NULL)) {\n \t\t\tRTE_LOG(ERR, VHOST_DATA,\n \t\t\t\t\"Failed to allocate memory for mbuf.\\n\");\n-\t\t\tfree_ind_table(idesc);\n \t\t\tbreak;\n \t\t}\n \n-\t\terr = copy_desc_to_mbuf(dev, vq, desc, sz, pkts[i], idx,\n-\t\t\t\t\tmbuf_pool);\n+\t\terr = copy_desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts[i],\n+\t\t\t\tmbuf_pool);\n \t\tif (unlikely(err)) {\n \t\t\trte_pktmbuf_free(pkts[i]);\n-\t\t\tfree_ind_table(idesc);\n \t\t\tbreak;\n \t\t}\n \n@@ -1274,11 +1201,10 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,\n \t\t\tzmbuf = get_zmbuf(vq);\n \t\t\tif (!zmbuf) {\n \t\t\t\trte_pktmbuf_free(pkts[i]);\n-\t\t\t\tfree_ind_table(idesc);\n \t\t\t\tbreak;\n \t\t\t}\n \t\t\tzmbuf->mbuf = pkts[i];\n-\t\t\tzmbuf->desc_idx = desc_indexes[i];\n+\t\t\tzmbuf->desc_idx = head_idx;\n \n \t\t\t/*\n \t\t\t * Pin lock the mbuf; we will check later to see\n@@ -1291,9 +1217,6 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,\n \t\t\tvq->nr_zmbuf += 1;\n \t\t\tTAILQ_INSERT_TAIL(&vq->zmbuf_list, zmbuf, next);\n \t\t}\n-\n-\t\tif (unlikely(!!idesc))\n-\t\t\tfree_ind_table(idesc);\n \t}\n \tvq->last_avail_idx += i;\n \n", "prefixes": [ "v3", "3/7" ] }{ "id": 41668, "url": "