get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/45073/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 45073,
    "url": "https://patches.dpdk.org/api/patches/45073/?format=api",
    "web_url": "https://patches.dpdk.org/project/dpdk/patch/20180921103308.16357-7-jfreimann@redhat.com/",
    "project": {
        "id": 1,
        "url": "https://patches.dpdk.org/api/projects/1/?format=api",
        "name": "DPDK",
        "link_name": "dpdk",
        "list_id": "dev.dpdk.org",
        "list_email": "dev@dpdk.org",
        "web_url": "http://core.dpdk.org",
        "scm_url": "git://dpdk.org/dpdk",
        "webscm_url": "http://git.dpdk.org/dpdk",
        "list_archive_url": "https://inbox.dpdk.org/dev",
        "list_archive_url_format": "https://inbox.dpdk.org/dev/{}",
        "commit_url_format": ""
    },
    "msgid": "<20180921103308.16357-7-jfreimann@redhat.com>",
    "list_archive_url": "https://inbox.dpdk.org/dev/20180921103308.16357-7-jfreimann@redhat.com",
    "date": "2018-09-21T10:33:03",
    "name": "[v6,06/11] net/virtio: implement transmit path for packed queues",
    "commit_ref": null,
    "pull_url": null,
    "state": "superseded",
    "archived": true,
    "hash": "b71fc74cbe72a3b6a680487ba669b8a4aaa02daa",
    "submitter": {
        "id": 745,
        "url": "https://patches.dpdk.org/api/people/745/?format=api",
        "name": "Jens Freimann",
        "email": "jfreimann@redhat.com"
    },
    "delegate": {
        "id": 2642,
        "url": "https://patches.dpdk.org/api/users/2642/?format=api",
        "username": "mcoquelin",
        "first_name": "Maxime",
        "last_name": "Coquelin",
        "email": "maxime.coquelin@redhat.com"
    },
    "mbox": "https://patches.dpdk.org/project/dpdk/patch/20180921103308.16357-7-jfreimann@redhat.com/mbox/",
    "series": [
        {
            "id": 1435,
            "url": "https://patches.dpdk.org/api/series/1435/?format=api",
            "web_url": "https://patches.dpdk.org/project/dpdk/list/?series=1435",
            "date": "2018-09-21T10:32:57",
            "name": "implement packed virtqueues",
            "version": 6,
            "mbox": "https://patches.dpdk.org/series/1435/mbox/"
        }
    ],
    "comments": "https://patches.dpdk.org/api/patches/45073/comments/",
    "check": "success",
    "checks": "https://patches.dpdk.org/api/patches/45073/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<dev-bounces@dpdk.org>",
        "X-Original-To": "patchwork@dpdk.org",
        "Delivered-To": "patchwork@dpdk.org",
        "Received": [
            "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id 0026A4CA5;\n\tFri, 21 Sep 2018 12:34:07 +0200 (CEST)",
            "from mx1.redhat.com (mx1.redhat.com [209.132.183.28])\n\tby dpdk.org (Postfix) with ESMTP id 7E4FA4CA0\n\tfor <dev@dpdk.org>; Fri, 21 Sep 2018 12:34:06 +0200 (CEST)",
            "from smtp.corp.redhat.com\n\t(int-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.12])\n\t(using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits))\n\t(No client certificate requested)\n\tby mx1.redhat.com (Postfix) with ESMTPS id E2CD7356C5;\n\tFri, 21 Sep 2018 10:34:05 +0000 (UTC)",
            "from localhost (dhcp-192-209.str.redhat.com [10.33.192.209])\n\tby smtp.corp.redhat.com (Postfix) with ESMTPS id 879787ED94;\n\tFri, 21 Sep 2018 10:34:01 +0000 (UTC)"
        ],
        "From": "Jens Freimann <jfreimann@redhat.com>",
        "To": "dev@dpdk.org",
        "Cc": "tiwei.bie@intel.com,\n\tmaxime.coquelin@redhat.com,\n\tGavin.Hu@arm.com",
        "Date": "Fri, 21 Sep 2018 12:33:03 +0200",
        "Message-Id": "<20180921103308.16357-7-jfreimann@redhat.com>",
        "In-Reply-To": "<20180921103308.16357-1-jfreimann@redhat.com>",
        "References": "<20180921103308.16357-1-jfreimann@redhat.com>",
        "X-Scanned-By": "MIMEDefang 2.79 on 10.5.11.12",
        "X-Greylist": "Sender IP whitelisted, not delayed by milter-greylist-4.5.16\n\t(mx1.redhat.com [10.5.110.30]); Fri, 21 Sep 2018 10:34:05 +0000 (UTC)",
        "Subject": "[dpdk-dev] [PATCH v6 06/11] net/virtio: implement transmit path for\n\tpacked queues",
        "X-BeenThere": "dev@dpdk.org",
        "X-Mailman-Version": "2.1.15",
        "Precedence": "list",
        "List-Id": "DPDK patches and discussions <dev.dpdk.org>",
        "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n\t<mailto:dev-request@dpdk.org?subject=unsubscribe>",
        "List-Archive": "<http://mails.dpdk.org/archives/dev/>",
        "List-Post": "<mailto:dev@dpdk.org>",
        "List-Help": "<mailto:dev-request@dpdk.org?subject=help>",
        "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n\t<mailto:dev-request@dpdk.org?subject=subscribe>",
        "Errors-To": "dev-bounces@dpdk.org",
        "Sender": "\"dev\" <dev-bounces@dpdk.org>"
    },
    "content": "This implements the transmit path for devices with\nsupport for packed virtqueues.\n\nAdd the feature bit and enable code to\nadd buffers to vring and mark descriptors as available.\n\nSigned-off-by: Jens Freiman <jfreimann@redhat.com>\n---\n drivers/net/virtio/virtio_ethdev.c |   8 +-\n drivers/net/virtio/virtio_ethdev.h |   2 +\n drivers/net/virtio/virtio_ring.h   |  15 +-\n drivers/net/virtio/virtio_rxtx.c   | 243 +++++++++++++++++++++++++++++\n drivers/net/virtio/virtqueue.h     |  17 +-\n 5 files changed, 280 insertions(+), 5 deletions(-)",
    "diff": "diff --git a/drivers/net/virtio/virtio_ethdev.c b/drivers/net/virtio/virtio_ethdev.c\nindex 29f3e1043..5c28af282 100644\n--- a/drivers/net/virtio/virtio_ethdev.c\n+++ b/drivers/net/virtio/virtio_ethdev.c\n@@ -384,6 +384,8 @@ virtio_init_queue(struct rte_eth_dev *dev, uint16_t vtpci_queue_idx)\n \tvq->hw = hw;\n \tvq->vq_queue_index = vtpci_queue_idx;\n \tvq->vq_nentries = vq_size;\n+\tif (vtpci_packed_queue(hw))\n+\t\tvq->vq_ring.avail_wrap_counter = 1;\n \n \t/*\n \t * Reserve a memzone for vring elements\n@@ -1338,7 +1340,11 @@ set_rxtx_funcs(struct rte_eth_dev *eth_dev)\n \t\teth_dev->rx_pkt_burst = &virtio_recv_pkts;\n \t}\n \n-\tif (hw->use_inorder_tx) {\n+\tif (vtpci_packed_queue(hw)) {\n+\t\tPMD_INIT_LOG(INFO, \"virtio: using virtio 1.1 Tx path on port %u\",\n+\t\t\teth_dev->data->port_id);\n+\t\teth_dev->tx_pkt_burst = virtio_xmit_pkts_packed;\n+\t} else if (hw->use_inorder_tx) {\n \t\tPMD_INIT_LOG(INFO, \"virtio: using inorder Tx path on port %u\",\n \t\t\teth_dev->data->port_id);\n \t\teth_dev->tx_pkt_burst = virtio_xmit_pkts_inorder;\ndiff --git a/drivers/net/virtio/virtio_ethdev.h b/drivers/net/virtio/virtio_ethdev.h\nindex b726ad108..04161b461 100644\n--- a/drivers/net/virtio/virtio_ethdev.h\n+++ b/drivers/net/virtio/virtio_ethdev.h\n@@ -79,6 +79,8 @@ uint16_t virtio_recv_mergeable_pkts_inorder(void *rx_queue,\n \n uint16_t virtio_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,\n \t\tuint16_t nb_pkts);\n+uint16_t virtio_xmit_pkts_packed(void *tx_queue, struct rte_mbuf **tx_pkts,\n+\t\tuint16_t nb_pkts);\n \n uint16_t virtio_xmit_pkts_inorder(void *tx_queue, struct rte_mbuf **tx_pkts,\n \t\tuint16_t nb_pkts);\ndiff --git a/drivers/net/virtio/virtio_ring.h b/drivers/net/virtio/virtio_ring.h\nindex b9e63d4d4..dbffd4dcd 100644\n--- a/drivers/net/virtio/virtio_ring.h\n+++ b/drivers/net/virtio/virtio_ring.h\n@@ -108,14 +108,25 @@ set_desc_avail(struct vring *vr, struct vring_desc_packed *desc)\n }\n \n static inline int\n-desc_is_used(struct vring_desc_packed *desc, struct vring *vr)\n+_desc_is_used(struct vring_desc_packed *desc)\n {\n \tuint16_t used, avail;\n \n \tused = !!(desc->flags & VRING_DESC_F_USED(1));\n \tavail = !!(desc->flags & VRING_DESC_F_AVAIL(1));\n \n-\treturn used == avail && used == vr->used_wrap_counter;\n+\treturn used == avail;\n+\n+}\n+\n+static inline int\n+desc_is_used(struct vring_desc_packed *desc, struct vring *vr)\n+{\n+\tuint16_t used;\n+\n+\tused = !!(desc->flags & VRING_DESC_F_USED(1));\n+\n+\treturn _desc_is_used(desc) && used == vr->used_wrap_counter;\n }\n \n /* The standard layout for the ring is a continuous chunk of memory which\ndiff --git a/drivers/net/virtio/virtio_rxtx.c b/drivers/net/virtio/virtio_rxtx.c\nindex eb891433e..ea6300563 100644\n--- a/drivers/net/virtio/virtio_rxtx.c\n+++ b/drivers/net/virtio/virtio_rxtx.c\n@@ -38,6 +38,7 @@\n #define  VIRTIO_DUMP_PACKET(m, len) do { } while (0)\n #endif\n \n+\n int\n virtio_dev_rx_queue_done(void *rxq, uint16_t offset)\n {\n@@ -165,6 +166,31 @@ virtqueue_dequeue_rx_inorder(struct virtqueue *vq,\n #endif\n \n /* Cleanup from completed transmits. */\n+static void\n+virtio_xmit_cleanup_packed(struct virtqueue *vq)\n+{\n+\tuint16_t idx;\n+\tuint16_t size = vq->vq_nentries;\n+\tstruct vring_desc_packed *desc = vq->vq_ring.desc_packed;\n+\tstruct vq_desc_extra *dxp;\n+\n+\tidx = vq->vq_used_cons_idx;\n+\twhile (_desc_is_used(&desc[idx]) &&\n+\t       vq->vq_free_cnt < size) {\n+\t\tdxp = &vq->vq_descx[idx];\n+\t\tvq->vq_free_cnt += dxp->ndescs;\n+\t\tidx += dxp->ndescs;\n+\t\tidx = idx >= size ? idx - size : idx;\n+\t\tif (idx == 0) {\n+\t\t\tvq->vq_ring.used_wrap_counter ^= 1;\n+\t\t}\n+\t\tif (dxp->cookie != NULL) {\n+\t\t\trte_pktmbuf_free(dxp->cookie);\n+\t\t\tdxp->cookie = NULL;\n+\t\t}\n+\t}\n+}\n+\n static void\n virtio_xmit_cleanup(struct virtqueue *vq, uint16_t num)\n {\n@@ -456,6 +482,129 @@ virtqueue_enqueue_xmit_inorder(struct virtnet_tx *txvq,\n \tvq->vq_desc_head_idx = idx & (vq->vq_nentries - 1);\n }\n \n+static inline void\n+virtqueue_enqueue_xmit_packed(struct virtnet_tx *txvq, struct rte_mbuf *cookie,\n+\t\t\tuint16_t needed, int use_indirect, int can_push,\n+\t\t\tint in_order)\n+{\n+\tstruct virtio_tx_region *txr = txvq->virtio_net_hdr_mz->addr;\n+\tstruct vq_desc_extra *dxp;\n+\tstruct virtqueue *vq = txvq->vq;\n+\tstruct vring_desc_packed *start_dp;\n+\tuint16_t seg_num = cookie->nb_segs;\n+\tuint16_t head_idx, idx, prev;\n+\tuint16_t head_id;\n+\tuint16_t head_size = vq->hw->vtnet_hdr_size;\n+\tstruct virtio_net_hdr *hdr;\n+\tint wrap_counter = vq->vq_ring.avail_wrap_counter;\n+\n+\thead_idx = vq->vq_desc_head_idx;\n+\tidx = head_idx;\n+\tdxp = &vq->vq_descx[idx];\n+\tdxp->cookie = (void *)cookie;\n+\tdxp->ndescs = needed;\n+\n+\tstart_dp = vq->vq_ring.desc_packed;\n+\thead_id = start_dp[head_idx].index;\n+\n+\tif (can_push) {\n+\t\t/* prepend cannot fail, checked by caller */\n+\t\thdr = (struct virtio_net_hdr *)\n+\t\t\trte_pktmbuf_prepend(cookie, head_size);\n+\t\t/* rte_pktmbuf_prepend() counts the hdr size to the pkt length,\n+\t\t * which is wrong. Below subtract restores correct pkt size.\n+\t\t */\n+\t\tcookie->pkt_len -= head_size;\n+\n+\t\t/* if offload disabled, it is not zeroed below, do it now */\n+\t\tif (!vq->hw->has_tx_offload) {\n+\t\t\tASSIGN_UNLESS_EQUAL(hdr->csum_start, 0);\n+\t\t\tASSIGN_UNLESS_EQUAL(hdr->csum_offset, 0);\n+\t\t\tASSIGN_UNLESS_EQUAL(hdr->flags, 0);\n+\t\t\tASSIGN_UNLESS_EQUAL(hdr->gso_type, 0);\n+\t\t\tASSIGN_UNLESS_EQUAL(hdr->gso_size, 0);\n+\t\t\tASSIGN_UNLESS_EQUAL(hdr->hdr_len, 0);\n+\t\t}\n+\t} else if (use_indirect) {\n+\t\t/* setup tx ring slot to point to indirect\n+\t\t * descriptor list stored in reserved region.\n+\t\t *\n+\t\t * the first slot in indirect ring is already preset\n+\t\t * to point to the header in reserved region\n+\t\t */\n+\t\tstart_dp[idx].addr  = txvq->virtio_net_hdr_mem +\n+\t\t\tRTE_PTR_DIFF(&txr[idx].tx_indir, txr);\n+\t\tstart_dp[idx].len   = (seg_num + 1) * sizeof(struct vring_desc);\n+\t\tstart_dp[idx].flags = VRING_DESC_F_INDIRECT;\n+\t\thdr = (struct virtio_net_hdr *)&txr[idx].tx_hdr;\n+\n+\t\t/* loop below will fill in rest of the indirect elements */\n+\t\tstart_dp = txr[idx].tx_indir_pq;\n+\t\tidx = 1;\n+\t} else {\n+\t\t/* setup first tx ring slot to point to header\n+\t\t * stored in reserved region.\n+\t\t */\n+\t\tstart_dp[idx].addr  = txvq->virtio_net_hdr_mem +\n+\t\t\tRTE_PTR_DIFF(&txr[idx].tx_hdr, txr);\n+\t\tstart_dp[idx].len   = vq->hw->vtnet_hdr_size;\n+\t\tstart_dp[idx].flags = VRING_DESC_F_NEXT |\n+\t\t\tVRING_DESC_F_AVAIL(vq->vq_ring.avail_wrap_counter) |\n+\t\t\tVRING_DESC_F_USED(!vq->vq_ring.used_wrap_counter); \n+\t\thdr = (struct virtio_net_hdr *)&txr[idx].tx_hdr;\n+\n+\t\tidx++;\n+\t}\n+\n+\tvirtqueue_xmit_offload(hdr, cookie, vq->hw->has_tx_offload);\n+\n+\tdo {\n+\t\tif (idx >= vq->vq_nentries) {\n+\t\t\tidx -= vq->vq_nentries;\n+\t\t\tvq->vq_ring.avail_wrap_counter ^= 1;\n+\t\t\tvq->vq_ring.used_wrap_counter ^= 1;\n+\t\t}\n+\t\tstart_dp[idx].addr  = VIRTIO_MBUF_DATA_DMA_ADDR(cookie, vq);\n+\t\tstart_dp[idx].len   = cookie->data_len;\n+\t\tstart_dp[idx].flags = VRING_DESC_F_NEXT |\n+\t\t\tVRING_DESC_F_AVAIL(vq->vq_ring.avail_wrap_counter) |\n+\t\t\tVRING_DESC_F_USED(!vq->vq_ring.used_wrap_counter); \n+\t\tidx++;\n+\t} while ((cookie = cookie->next) != NULL);\n+\n+\tif (use_indirect)\n+\t\tidx = vq->vq_ring.desc_packed[head_idx].index;\n+\n+\tif (idx >= vq->vq_nentries) {\n+\t\tidx -= vq->vq_nentries;\n+\t\tvq->vq_ring.avail_wrap_counter ^= 1;\n+\t\tvq->vq_ring.used_wrap_counter ^= 1;\n+\t}\n+\n+\tvq->vq_free_cnt = (uint16_t)(vq->vq_free_cnt - needed);\n+\n+\tif (needed > 1) {\n+\t\tprev = (idx - 1 > 0 ? idx - 1 : vq->vq_nentries) - 1;\n+\t\tstart_dp[prev].index = head_id;\n+\t\tstart_dp[prev].flags =\n+\t\t\t(VRING_DESC_F_AVAIL(wrap_counter) |\n+\t\t\t VRING_DESC_F_USED(!wrap_counter));\n+\t}\n+\tstart_dp[head_idx].flags =\n+\t\t(VRING_DESC_F_AVAIL(wrap_counter) |\n+\t\t VRING_DESC_F_USED(!wrap_counter));\n+\trte_smp_wmb();\n+\n+\tvq->vq_desc_head_idx = idx;\n+\tvq->vq_avail_idx = idx;\n+\n+\tif (!in_order) {\n+\t\tif (vq->vq_desc_head_idx == VQ_RING_DESC_CHAIN_END)\n+\t\t\tvq->vq_desc_tail_idx = idx;\n+\t}\n+}\n+\n+\n static inline void\n virtqueue_enqueue_xmit(struct virtnet_tx *txvq, struct rte_mbuf *cookie,\n \t\t\tuint16_t needed, int use_indirect, int can_push,\n@@ -736,6 +885,9 @@ virtio_dev_tx_queue_setup_finish(struct rte_eth_dev *dev,\n \tif (hw->use_inorder_tx)\n \t\tvq->vq_ring.desc[vq->vq_nentries - 1].next = 0;\n \n+\tif (vtpci_packed_queue(hw))\n+\t\tvq->vq_ring.avail_wrap_counter = 1;\n+\n \tVIRTQUEUE_DUMP(vq);\n \n \treturn 0;\n@@ -1346,6 +1498,97 @@ virtio_recv_mergeable_pkts(void *rx_queue,\n \treturn nb_rx;\n }\n \n+uint16_t\n+virtio_xmit_pkts_packed(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)\n+{\n+\tstruct virtnet_tx *txvq = tx_queue;\n+\tstruct virtqueue *vq = txvq->vq;\n+\tstruct virtio_hw *hw = vq->hw;\n+\tuint16_t hdr_size = hw->vtnet_hdr_size;\n+\tuint16_t nb_tx = 0;\n+\tint error;\n+\n+\tif (unlikely(hw->started == 0 && tx_pkts != hw->inject_pkts))\n+\t\treturn nb_tx;\n+\n+\tif (unlikely(nb_pkts < 1))\n+\t\treturn nb_pkts;\n+\n+\tPMD_TX_LOG(DEBUG, \"%d packets to xmit\", nb_pkts);\n+\n+\tvirtio_rmb();\n+\tif (likely(nb_pkts > vq->vq_nentries - vq->vq_free_thresh))\n+\t\tvirtio_xmit_cleanup_packed(vq);\n+\n+\tfor (nb_tx = 0; nb_tx < nb_pkts; nb_tx++) {\n+\t\tstruct rte_mbuf *txm = tx_pkts[nb_tx];\n+\t\tint can_push = 0, use_indirect = 0, slots, need;\n+\n+\t\t/* Do VLAN tag insertion */\n+\t\tif (unlikely(txm->ol_flags & PKT_TX_VLAN_PKT)) {\n+\t\t\terror = rte_vlan_insert(&txm);\n+\t\t\tif (unlikely(error)) {\n+\t\t\t\trte_pktmbuf_free(txm);\n+\t\t\t\tcontinue;\n+\t\t\t}\n+\t\t}\n+\n+\t\t/* optimize ring usage */\n+\t\tif ((vtpci_with_feature(hw, VIRTIO_F_ANY_LAYOUT) ||\n+\t\t      vtpci_with_feature(hw, VIRTIO_F_VERSION_1)) &&\n+\t\t    rte_mbuf_refcnt_read(txm) == 1 &&\n+\t\t    RTE_MBUF_DIRECT(txm) &&\n+\t\t    txm->nb_segs == 1 &&\n+\t\t    rte_pktmbuf_headroom(txm) >= hdr_size &&\n+\t\t    rte_is_aligned(rte_pktmbuf_mtod(txm, char *),\n+\t\t\t\t   __alignof__(struct virtio_net_hdr_mrg_rxbuf)))\n+\t\t\tcan_push = 1;\n+\t\telse if (vtpci_with_feature(hw, VIRTIO_RING_F_INDIRECT_DESC) &&\n+\t\t\t txm->nb_segs < VIRTIO_MAX_TX_INDIRECT)\n+\t\t\tuse_indirect = 1;\n+\n+\t\t/* How many main ring entries are needed to this Tx?\n+\t\t * any_layout => number of segments\n+\t\t * indirect   => 1\n+\t\t * default    => number of segments + 1\n+\t\t */\n+\t\tslots = use_indirect ? 1 : (txm->nb_segs + !can_push);\n+\t\tneed = slots - vq->vq_free_cnt;\n+\n+\t\t/* Positive value indicates it need free vring descriptors */\n+\t\tif (unlikely(need > 0)) {\n+\t\t\tvirtio_rmb();\n+\t\t\tneed = RTE_MIN(need, (int)nb_pkts);\n+\n+\t\t\tvirtio_xmit_cleanup_packed(vq);\n+\t\t\tneed = slots - vq->vq_free_cnt;\n+\t\t\tif (unlikely(need > 0)) {\n+\t\t\t\tPMD_TX_LOG(ERR,\n+\t\t\t\t\t   \"No free tx descriptors to transmit\");\n+\t\t\t\tbreak;\n+\t\t\t}\n+\t\t}\n+\n+\t\t/* Enqueue Packet buffers */\n+\t\tvirtqueue_enqueue_xmit_packed(txvq, txm, slots, use_indirect,\n+\t\t\tcan_push, 0);\n+\n+\t\ttxvq->stats.bytes += txm->pkt_len;\n+\t\tvirtio_update_packet_stats(&txvq->stats, txm);\n+\t}\n+\n+\ttxvq->stats.packets += nb_tx;\n+\n+\tif (likely(nb_tx)) {\n+\t\tif (unlikely(virtqueue_kick_prepare_packed(vq))) {\n+\t\t\tvirtqueue_notify(vq);\n+\t\t\tPMD_TX_LOG(DEBUG, \"Notified backend after xmit\");\n+\t\t}\n+\t}\n+\n+\treturn nb_tx;\n+}\n+\n uint16_t\n virtio_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)\n {\ndiff --git a/drivers/net/virtio/virtqueue.h b/drivers/net/virtio/virtqueue.h\nindex eb220563f..ffa2d8f92 100644\n--- a/drivers/net/virtio/virtqueue.h\n+++ b/drivers/net/virtio/virtqueue.h\n@@ -241,8 +241,12 @@ struct virtio_net_hdr_mrg_rxbuf {\n #define VIRTIO_MAX_TX_INDIRECT 8\n struct virtio_tx_region {\n \tstruct virtio_net_hdr_mrg_rxbuf tx_hdr;\n-\tstruct vring_desc tx_indir[VIRTIO_MAX_TX_INDIRECT]\n-\t\t\t   __attribute__((__aligned__(16)));\n+\tunion {\n+\t\tstruct vring_desc tx_indir[VIRTIO_MAX_TX_INDIRECT]\n+\t\t\t__attribute__((__aligned__(16)));\n+\t\tstruct vring_desc_packed tx_indir_pq[VIRTIO_MAX_TX_INDIRECT]\n+\t\t\t__attribute__((__aligned__(16)));\n+\t};\n };\n \n static inline uint16_t\n@@ -360,6 +364,15 @@ virtqueue_kick_prepare(struct virtqueue *vq)\n \treturn !(vq->vq_ring.used->flags & VRING_USED_F_NO_NOTIFY);\n }\n \n+static inline int\n+virtqueue_kick_prepare_packed(struct virtqueue *vq)\n+{\n+\tuint16_t flags;\n+\n+\tflags = vq->vq_ring.device_event->desc_event_flags & RING_EVENT_FLAGS_DESC;\n+\treturn (flags != RING_EVENT_FLAGS_DISABLE);\n+}\n+\n static inline void\n virtqueue_notify(struct virtqueue *vq)\n {\n",
    "prefixes": [
        "v6",
        "06/11"
    ]
}