From patchwork Wed Jul 8 07:26:45 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Patchwork-Submitter: "Jiang, Cheng1" X-Patchwork-Id: 73506 X-Patchwork-Delegate: maxime.coquelin@redhat.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 61799A00BE; Wed, 8 Jul 2020 09:32:17 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 1BFCE1D9C4; Wed, 8 Jul 2020 09:32:15 +0200 (CEST) Received: from mga02.intel.com (mga02.intel.com [134.134.136.20]) by dpdk.org (Postfix) with ESMTP id 8D74D1D9C3 for ; Wed, 8 Jul 2020 09:32:12 +0200 (CEST) IronPort-SDR: DMZTrqWzDbM8Qt9M0QfevPHcH5z2u00PRfgXuszNr55ea03NfJaDjzAluJ24veLdE4xIg1PS11 K49YGgIbMZLQ== X-IronPort-AV: E=McAfee;i="6000,8403,9675"; a="135986603" X-IronPort-AV: E=Sophos;i="5.75,327,1589266800"; d="scan'208";a="135986603" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga004.jf.intel.com ([10.7.209.38]) by orsmga101.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 08 Jul 2020 00:32:11 -0700 IronPort-SDR: lFThSv2AT6fDkvu/h8TqC2vmmQSAHW+4L7a4GgPsYUSQnL9Akiob1yrrAMvoK3hPzt7EQ9ybaR KLn4raCqJaOg== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.75,327,1589266800"; d="scan'208";a="427748829" Received: from dpdk_jiangcheng.sh.intel.com ([10.67.119.112]) by orsmga004.jf.intel.com with ESMTP; 08 Jul 2020 00:32:09 -0700 From: Cheng Jiang To: maxime.coquelin@redhat.com, chenbo.xia@intel.com, zhihong.wang@intel.com Cc: dev@dpdk.org, patrick.fu@intel.com, cunming.liang@intel.com, Cheng Jiang Date: Wed, 8 Jul 2020 07:26:45 +0000 Message-Id: <20200708072645.39031-1-Cheng1.jiang@intel.com> X-Mailer: git-send-email 2.27.0 In-Reply-To: <20200622025914.85175-1-Cheng1.jiang@intel.com> References: <20200622025914.85175-1-Cheng1.jiang@intel.com> MIME-Version: 1.0 Subject: [dpdk-dev] [RFC v2] example/vhost: add support for vhost async data path X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch makes vhost-vswitch be able to use vhost asynchronous api for enqueue operations. Demonstrated how the application leverage IOAT DMA channel with vhost async api. Since this is an early preview patch, the performance has not been fully optimized and it's not suggested to use this patch as a tool for benchmark. We introduce two parameters to enable DMA acceleration for Tx operations of queues: –async_vhost_driver Async vhost-user net driver which demonstrates how to use the async vhost APIs will be used when this option is given. It is disabled by default. -dmas This parameter is used to specify the assigned DMA device of a queue. This patch depends on following patch set: http://patches.dpdk.org/cover/73359/ Signed-off-by: Cheng Jiang --- v2: * updated some variable names based on the latest async vhost patch * fixed a bug in virtio_xmit function * fixed a hardcode * fixed a typo --- examples/vhost/main.c | 248 +++++++++++++++++++++++++++++++++++++++++- examples/vhost/main.h | 1 + 2 files changed, 245 insertions(+), 4 deletions(-) -- 2.27.0 diff --git a/examples/vhost/main.c b/examples/vhost/main.c index 312829e8b..72135a3df 100644 --- a/examples/vhost/main.c +++ b/examples/vhost/main.c @@ -24,11 +24,15 @@ #include #include #include +#include +#include +#include +#include #include "main.h" #ifndef MAX_QUEUES -#define MAX_QUEUES 128 +#define MAX_QUEUES 512 #endif /* the maximum number of external ports supported */ @@ -58,6 +62,12 @@ /* Maximum long option length for option parsing. */ #define MAX_LONG_OPT_SZ 64 +#define IOAT_RING_SIZE 4096 + +#define MAX_ENQUEUED_SIZE 2048 + +#define MAX_VHOST_DEVICE 1024 + /* mask of enabled ports */ static uint32_t enabled_port_mask = 0; @@ -96,6 +106,20 @@ static int dequeue_zero_copy; static int builtin_net_driver; +static int async_vhost_driver; + +struct dma_info { + struct rte_pci_addr addr; + uint16_t dev_id; + bool is_valid; +}; + +struct dma_info_input { + struct dma_info dmas[RTE_MAX_QUEUES_PER_PORT * 2]; + uint16_t nr; +}; + +static struct dma_info_input dma_bind[MAX_VHOST_DEVICE]; /* Specify timeout (in useconds) between retries on RX. */ static uint32_t burst_rx_delay_time = BURST_RX_WAIT_US; /* Specify the number of retries on RX. */ @@ -141,6 +165,61 @@ static struct rte_eth_conf vmdq_conf_default = { }, }; +static int +ioat_transfer_data_cb(int vid, uint16_t queue_id, struct rte_vhost_async_desc *descs, + struct rte_vhost_async_status *opaque_data, uint16_t count) +{ + int ret; + uint16_t i_desc; + + struct rte_vhost_iov_iter *src = NULL; + struct rte_vhost_iov_iter *dst = NULL; + unsigned long i_seg; + + int dev_id = dma_bind[vid].dmas[queue_id * 2 + VIRTIO_RXQ].dev_id; + if (likely(!opaque_data)) { + for (i_desc = 0; i_desc < count; i_desc++) { + src = descs[i_desc].src; + dst = descs[i_desc].dst; + i_seg = 0; + while (i_seg < src->nr_segs) { + ret = rte_ioat_enqueue_copy(dev_id, + (uintptr_t)(src->iov[i_seg].iov_base) + + src->offset, + (uintptr_t)(dst->iov[i_seg].iov_base) + + dst->offset, + src->iov[i_seg].iov_len, + 0, + 0, + 0); + if (ret != 1) + break; + i_seg++; + } + } + } else { + /* Opaque data is not supported */ + return -1; + } + /* ring the doorbell */ + rte_ioat_do_copies(dev_id); + return i_desc; +} + +static int +ioat_check_completed_copies_cb(int vid, uint16_t queue_id, + struct rte_vhost_async_status *opaque_data, + uint16_t max_packets __rte_unused) +{ + if (!opaque_data) { + uintptr_t dump[255]; + return rte_ioat_completed_copies(dma_bind[vid].dmas[queue_id * 2 + + VIRTIO_RXQ].dev_id, 255, dump, dump); + } else { + /* Opaque data is not supported */ + return -1; + } +} static unsigned lcore_ids[RTE_MAX_LCORE]; static uint16_t ports[RTE_MAX_ETHPORTS]; @@ -186,6 +265,94 @@ struct mbuf_table lcore_tx_queue[RTE_MAX_LCORE]; * Builds up the correct configuration for VMDQ VLAN pool map * according to the pool & queue limits. */ + +static inline int +open_dma(const char *value, void *dma_bind_info) +{ + struct dma_info_input *dma_info = dma_bind_info; + char *input = strndup(value, strlen(value) + 1); + char *addrs = input; + char *ptrs[2]; + char *start, *end, *substr; + int64_t qid, vring_id; + struct rte_ioat_rawdev_config config; + struct rte_rawdev_info info = { .dev_private = &config }; + char name[32]; + int dev_id; + int ret = 0; + + while (isblank(*addrs)) + addrs++; + if (*addrs == '\0') { + ret = -1; + goto out; + } + + /* process DMA devices within bracket. */ + addrs++; + substr = strtok(addrs, ";]"); + if (!substr) { + ret = -1; + goto out; + } + + do { + rte_strsplit(substr, strlen(substr), ptrs, 2, '@'); + + start = strstr(ptrs[0], "txq"); + if (start == NULL) { + ret = -1; + goto out; + } + + start += 3; + qid = strtol(start, &end, 0); + if (end == start) { + ret = -1; + goto out; + } + + vring_id = qid * 2 + VIRTIO_RXQ; + if (rte_pci_addr_parse(ptrs[1], + &dma_info->dmas[vring_id].addr) < 0) { + ret = -1; + goto out; + } + + rte_pci_device_name(&dma_info->dmas[vring_id].addr, + name, sizeof(name)); + dev_id = rte_rawdev_get_dev_id(name); + if (dev_id == (uint16_t)(-ENODEV) || + dev_id == (uint16_t)(-EINVAL)) { + ret = -1; + goto out; + } + + if (rte_rawdev_info_get(dev_id, &info) < 0 || + strstr(info.driver_name, "ioat") == NULL) { + ret = -1; + goto out; + } + + dma_info->dmas[vring_id].dev_id = dev_id; + dma_info->dmas[vring_id].is_valid = true; + config.ring_size = IOAT_RING_SIZE; + if (rte_rawdev_configure(dev_id, &info) < 0) { + ret = -1; + goto out; + } + rte_rawdev_start(dev_id); + + dma_info->nr++; + + substr = strtok(NULL, ";]"); + } while (substr); + +out: + free(input); + return ret; +} + static inline int get_eth_conf(struct rte_eth_conf *eth_conf, uint32_t num_devices) { @@ -488,6 +655,8 @@ us_vhost_parse_args(int argc, char **argv) {"client", no_argument, &client_mode, 1}, {"dequeue-zero-copy", no_argument, &dequeue_zero_copy, 1}, {"builtin-net-driver", no_argument, &builtin_net_driver, 1}, + {"async_vhost_driver", no_argument, &async_vhost_driver, 1}, + {"dmas", required_argument, NULL, 0}, {NULL, 0, 0, 0}, }; @@ -623,13 +792,25 @@ us_vhost_parse_args(int argc, char **argv) "socket-file", MAX_LONG_OPT_SZ)) { if (us_vhost_parse_socket_path(optarg) == -1) { RTE_LOG(INFO, VHOST_CONFIG, - "Invalid argument for socket name (Max %d characters)\n", - PATH_MAX); + "Invalid argument for socket name (Max %d characters)\n", + PATH_MAX); us_vhost_usage(prgname); return -1; } } + if (!strncmp(long_option[option_index].name, + "dmas", MAX_LONG_OPT_SZ)) { + if (open_dma(optarg, &(dma_bind[0])) == -1) { + if (*optarg == -1) { + RTE_LOG(INFO, VHOST_CONFIG, + "Wrong DMA args\n"); + us_vhost_usage(prgname); + return -1; + } + } + } + break; /* Invalid option - print options. */ @@ -785,9 +966,26 @@ virtio_xmit(struct vhost_dev *dst_vdev, struct vhost_dev *src_vdev, struct rte_mbuf *m) { uint16_t ret; + struct rte_mbuf *m_cpl[1]; if (builtin_net_driver) { ret = vs_enqueue_pkts(dst_vdev, VIRTIO_RXQ, &m, 1); + } else if (async_vhost_driver) { + ret = rte_vhost_submit_enqueue_burst(dst_vdev->vid, VIRTIO_RXQ, + &m, 1); + + if (likely(ret)) { + dst_vdev->nr_async_pkts++; + rte_mbuf_refcnt_update(m, 1); + } + + while (likely(dst_vdev->nr_async_pkts)) { + if (rte_vhost_poll_enqueue_completed(dst_vdev->vid, + VIRTIO_RXQ, m_cpl, 1)) { + dst_vdev->nr_async_pkts--; + rte_pktmbuf_free(*m_cpl); + } + } } else { ret = rte_vhost_enqueue_burst(dst_vdev->vid, VIRTIO_RXQ, &m, 1); } @@ -1036,6 +1234,19 @@ drain_mbuf_table(struct mbuf_table *tx_q) } } +static __rte_always_inline void +complete_async_pkts(struct vhost_dev *vdev, uint16_t qid) +{ + struct rte_mbuf *p_cpl[MAX_PKT_BURST]; + uint16_t complete_count; + + complete_count = rte_vhost_poll_enqueue_completed(vdev->vid, + qid, p_cpl, MAX_PKT_BURST); + vdev->nr_async_pkts -= complete_count; + if (complete_count) + free_pkts(p_cpl, complete_count); +} + static __rte_always_inline void drain_eth_rx(struct vhost_dev *vdev) { @@ -1044,6 +1255,10 @@ drain_eth_rx(struct vhost_dev *vdev) rx_count = rte_eth_rx_burst(ports[0], vdev->vmdq_rx_q, pkts, MAX_PKT_BURST); + + while (likely(vdev->nr_async_pkts)) + complete_async_pkts(vdev, VIRTIO_RXQ); + if (!rx_count) return; @@ -1068,16 +1283,22 @@ drain_eth_rx(struct vhost_dev *vdev) if (builtin_net_driver) { enqueue_count = vs_enqueue_pkts(vdev, VIRTIO_RXQ, pkts, rx_count); + } else if (async_vhost_driver) { + enqueue_count = rte_vhost_submit_enqueue_burst(vdev->vid, + VIRTIO_RXQ, pkts, rx_count); + vdev->nr_async_pkts += enqueue_count; } else { enqueue_count = rte_vhost_enqueue_burst(vdev->vid, VIRTIO_RXQ, pkts, rx_count); } + if (enable_stats) { rte_atomic64_add(&vdev->stats.rx_total_atomic, rx_count); rte_atomic64_add(&vdev->stats.rx_atomic, enqueue_count); } - free_pkts(pkts, rx_count); + if (!async_vhost_driver) + free_pkts(pkts, rx_count); } static __rte_always_inline void @@ -1224,6 +1445,9 @@ destroy_device(int vid) "(%d) device has been removed from data core\n", vdev->vid); + if (async_vhost_driver) + rte_vhost_async_channel_unregister(vid, VIRTIO_RXQ); + rte_free(vdev); } @@ -1238,6 +1462,12 @@ new_device(int vid) uint32_t device_num_min = num_devices; struct vhost_dev *vdev; + struct rte_vhost_async_channel_ops channel_ops = { + .transfer_data = ioat_transfer_data_cb, + .check_completed_copies = ioat_check_completed_copies_cb + }; + struct rte_vhost_async_features f; + vdev = rte_zmalloc("vhost device", sizeof(*vdev), RTE_CACHE_LINE_SIZE); if (vdev == NULL) { RTE_LOG(INFO, VHOST_DATA, @@ -1278,6 +1508,13 @@ new_device(int vid) "(%d) device has been added to data core %d\n", vid, vdev->coreid); + if (async_vhost_driver) { + f.async_inorder = 1; + f.async_threshold = 256; + return rte_vhost_async_channel_register(vid, VIRTIO_RXQ, + f.intval, &channel_ops); + } + return 0; } @@ -1519,6 +1756,9 @@ main(int argc, char *argv[]) /* Register vhost user driver to handle vhost messages. */ for (i = 0; i < nb_sockets; i++) { char *file = socket_files + i * PATH_MAX; + if (async_vhost_driver) + flags = flags | RTE_VHOST_USER_ASYNC_COPY; + ret = rte_vhost_driver_register(file, flags); if (ret != 0) { unregister_drivers(i); diff --git a/examples/vhost/main.h b/examples/vhost/main.h index 7cba0edbf..4317b6ae8 100644 --- a/examples/vhost/main.h +++ b/examples/vhost/main.h @@ -51,6 +51,7 @@ struct vhost_dev { uint64_t features; size_t hdr_len; uint16_t nr_vrings; + uint16_t nr_async_pkts; struct rte_vhost_memory *mem; struct device_statistics stats; TAILQ_ENTRY(vhost_dev) global_vdev_entry;