From patchwork Thu May 21 07:49:40 2015 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ouyang Changchun X-Patchwork-Id: 4814 Return-Path: X-Original-To: patchwork@dpdk.org Delivered-To: patchwork@dpdk.org Received: from [92.243.14.124] (localhost [IPv6:::1]) by dpdk.org (Postfix) with ESMTP id 2827CC374; Thu, 21 May 2015 09:50:08 +0200 (CEST) Received: from mga14.intel.com (mga14.intel.com [192.55.52.115]) by dpdk.org (Postfix) with ESMTP id 21D0CC332 for ; Thu, 21 May 2015 09:50:04 +0200 (CEST) Received: from fmsmga001.fm.intel.com ([10.253.24.23]) by fmsmga103.fm.intel.com with ESMTP; 21 May 2015 00:50:03 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.13,467,1427785200"; d="scan'208";a="713523672" Received: from shvmail01.sh.intel.com ([10.239.29.42]) by fmsmga001.fm.intel.com with ESMTP; 21 May 2015 00:50:03 -0700 Received: from shecgisg004.sh.intel.com (shecgisg004.sh.intel.com [10.239.29.89]) by shvmail01.sh.intel.com with ESMTP id t4L7nw30026672; Thu, 21 May 2015 15:49:58 +0800 Received: from shecgisg004.sh.intel.com (localhost [127.0.0.1]) by shecgisg004.sh.intel.com (8.13.6/8.13.6/SuSE Linux 0.8) with ESMTP id t4L7ntw2015370; Thu, 21 May 2015 15:49:57 +0800 Received: (from couyang@localhost) by shecgisg004.sh.intel.com (8.13.6/8.13.6/Submit) id t4L7ntjR015366; Thu, 21 May 2015 15:49:55 +0800 From: Ouyang Changchun To: dev@dpdk.org Date: Thu, 21 May 2015 15:49:40 +0800 Message-Id: <1432194581-15301-6-git-send-email-changchun.ouyang@intel.com> X-Mailer: git-send-email 1.7.12.2 In-Reply-To: <1432194581-15301-1-git-send-email-changchun.ouyang@intel.com> References: <1432194581-15301-1-git-send-email-changchun.ouyang@intel.com> Subject: [dpdk-dev] [PATCH 5/6] vhost: Support multiple queues X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: patches and discussions about DPDK List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Sample vhost leverage the VMDq+RSS in HW to receive packets and distribute them into different queue in the pool according to 5 tuples. And enable multiple queues mode in vhost/virtio layer. HW queue numbers in pool exactly same with the queue number in virtio device, e.g. rxq = 4, the queue number is 4, it means 4 HW queues in each VMDq pool, and 4 queues in each virtio device/port, one maps to each. ========================================= ==================| |==================| vport0 | | vport1 | --- --- --- ---| |--- --- --- ---| q0 | q1 | q2 | q3 | |q0 | q1 | q2 | q3 | /\= =/\= =/\= =/\=| |/\= =/\= =/\= =/\=| || || || || || || || || || || || || || || || || ||= =||= =||= =||=| =||== ||== ||== ||=| q0 | q1 | q2 | q3 | |q0 | q1 | q2 | q3 | ------------------| |------------------| VMDq pool0 | | VMDq pool1 | ==================| |==================| In RX side, it firstly polls each queue of the pool and gets the packets from it and enqueue them into its corresponding queue in virtio device/port. In TX side, it dequeue packets from each queue of virtio device/port and send to either physical port or another virtio device according to its destination MAC address. Signed-off-by: Changchun Ouyang --- examples/vhost/main.c | 132 ++++++++++++++++++++++++++++++-------------------- 1 file changed, 79 insertions(+), 53 deletions(-) diff --git a/examples/vhost/main.c b/examples/vhost/main.c index 16d4463..0a33e57 100644 --- a/examples/vhost/main.c +++ b/examples/vhost/main.c @@ -998,8 +998,9 @@ link_vmdq(struct vhost_dev *vdev, struct rte_mbuf *m) /* Enable stripping of the vlan tag as we handle routing. */ if (vlan_strip) - rte_eth_dev_set_vlan_strip_on_queue(ports[0], - (uint16_t)vdev->vmdq_rx_q, 1); + for (i = 0; i < (int)rxq; i++) + rte_eth_dev_set_vlan_strip_on_queue(ports[0], + (uint16_t)(vdev->vmdq_rx_q + i), 1); /* Set device as ready for RX. */ vdev->ready = DEVICE_RX; @@ -1014,7 +1015,7 @@ link_vmdq(struct vhost_dev *vdev, struct rte_mbuf *m) static inline void unlink_vmdq(struct vhost_dev *vdev) { - unsigned i = 0; + unsigned i = 0, j = 0; unsigned rx_count; struct rte_mbuf *pkts_burst[MAX_PKT_BURST]; @@ -1027,15 +1028,19 @@ unlink_vmdq(struct vhost_dev *vdev) vdev->vlan_tag = 0; /*Clear out the receive buffers*/ - rx_count = rte_eth_rx_burst(ports[0], - (uint16_t)vdev->vmdq_rx_q, pkts_burst, MAX_PKT_BURST); + for (i = 0; i < rxq; i++) { + rx_count = rte_eth_rx_burst(ports[0], + (uint16_t)vdev->vmdq_rx_q + i, + pkts_burst, MAX_PKT_BURST); - while (rx_count) { - for (i = 0; i < rx_count; i++) - rte_pktmbuf_free(pkts_burst[i]); + while (rx_count) { + for (j = 0; j < rx_count; j++) + rte_pktmbuf_free(pkts_burst[j]); - rx_count = rte_eth_rx_burst(ports[0], - (uint16_t)vdev->vmdq_rx_q, pkts_burst, MAX_PKT_BURST); + rx_count = rte_eth_rx_burst(ports[0], + (uint16_t)vdev->vmdq_rx_q + i, + pkts_burst, MAX_PKT_BURST); + } } vdev->ready = DEVICE_MAC_LEARNING; @@ -1047,7 +1052,7 @@ unlink_vmdq(struct vhost_dev *vdev) * the packet on that devices RX queue. If not then return. */ static inline int __attribute__((always_inline)) -virtio_tx_local(struct vhost_dev *vdev, struct rte_mbuf *m) +virtio_tx_local(struct vhost_dev *vdev, struct rte_mbuf *m, uint32_t q_idx) { struct virtio_net_data_ll *dev_ll; struct ether_hdr *pkt_hdr; @@ -1062,7 +1067,7 @@ virtio_tx_local(struct vhost_dev *vdev, struct rte_mbuf *m) while (dev_ll != NULL) { if ((dev_ll->vdev->ready == DEVICE_RX) && ether_addr_cmp(&(pkt_hdr->d_addr), - &dev_ll->vdev->mac_address)) { + &dev_ll->vdev->mac_address)) { /* Drop the packet if the TX packet is destined for the TX device. */ if (dev_ll->vdev->dev->device_fh == dev->device_fh) { @@ -1080,7 +1085,9 @@ virtio_tx_local(struct vhost_dev *vdev, struct rte_mbuf *m) LOG_DEBUG(VHOST_DATA, "(%"PRIu64") Device is marked for removal\n", tdev->device_fh); } else { /*send the packet to the local virtio device*/ - ret = rte_vhost_enqueue_burst(tdev, VIRTIO_RXQ, &m, 1); + ret = rte_vhost_enqueue_burst(tdev, + VIRTIO_RXQ + q_idx * VIRTIO_QNUM, + &m, 1); if (enable_stats) { rte_atomic64_add( &dev_statistics[tdev->device_fh].rx_total_atomic, @@ -1157,7 +1164,8 @@ find_local_dest(struct virtio_net *dev, struct rte_mbuf *m, * or the physical port. */ static inline void __attribute__((always_inline)) -virtio_tx_route(struct vhost_dev *vdev, struct rte_mbuf *m, uint16_t vlan_tag) +virtio_tx_route(struct vhost_dev *vdev, struct rte_mbuf *m, + uint16_t vlan_tag, uint32_t q_idx) { struct mbuf_table *tx_q; struct rte_mbuf **m_table; @@ -1167,7 +1175,8 @@ virtio_tx_route(struct vhost_dev *vdev, struct rte_mbuf *m, uint16_t vlan_tag) struct ether_hdr *nh; /*check if destination is local VM*/ - if ((vm2vm_mode == VM2VM_SOFTWARE) && (virtio_tx_local(vdev, m) == 0)) { + if ((vm2vm_mode == VM2VM_SOFTWARE) && + (virtio_tx_local(vdev, m, q_idx) == 0)) { rte_pktmbuf_free(m); return; } @@ -1331,49 +1340,60 @@ switch_worker(__attribute__((unused)) void *arg) } if (likely(vdev->ready == DEVICE_RX)) { /*Handle guest RX*/ - rx_count = rte_eth_rx_burst(ports[0], - vdev->vmdq_rx_q, pkts_burst, MAX_PKT_BURST); + for (i = 0; i < rxq; i ++) { + rx_count = rte_eth_rx_burst(ports[0], + vdev->vmdq_rx_q + i, pkts_burst, MAX_PKT_BURST); - if (rx_count) { - /* - * Retry is enabled and the queue is full then we wait and retry to avoid packet loss - * Here MAX_PKT_BURST must be less than virtio queue size - */ - if (enable_retry && unlikely(rx_count > rte_vring_available_entries(dev, VIRTIO_RXQ))) { - for (retry = 0; retry < burst_rx_retry_num; retry++) { - rte_delay_us(burst_rx_delay_time); - if (rx_count <= rte_vring_available_entries(dev, VIRTIO_RXQ)) - break; + if (rx_count) { + /* + * Retry is enabled and the queue is full then we wait and retry to avoid packet loss + * Here MAX_PKT_BURST must be less than virtio queue size + */ + if (enable_retry && unlikely(rx_count > rte_vring_available_entries(dev, + VIRTIO_RXQ + i * VIRTIO_QNUM))) { + for (retry = 0; retry < burst_rx_retry_num; retry++) { + rte_delay_us(burst_rx_delay_time); + if (rx_count <= rte_vring_available_entries(dev, + VIRTIO_RXQ + i * VIRTIO_QNUM)) + break; + } + } + ret_count = rte_vhost_enqueue_burst(dev, VIRTIO_RXQ + i * VIRTIO_QNUM, + pkts_burst, rx_count); + if (enable_stats) { + rte_atomic64_add( + &dev_statistics[dev_ll->vdev->dev->device_fh].rx_total_atomic, + rx_count); + rte_atomic64_add( + &dev_statistics[dev_ll->vdev->dev->device_fh].rx_atomic, ret_count); + } + while (likely(rx_count)) { + rx_count--; + rte_pktmbuf_free(pkts_burst[rx_count]); } } - ret_count = rte_vhost_enqueue_burst(dev, VIRTIO_RXQ, pkts_burst, rx_count); - if (enable_stats) { - rte_atomic64_add( - &dev_statistics[dev_ll->vdev->dev->device_fh].rx_total_atomic, - rx_count); - rte_atomic64_add( - &dev_statistics[dev_ll->vdev->dev->device_fh].rx_atomic, ret_count); - } - while (likely(rx_count)) { - rx_count--; - rte_pktmbuf_free(pkts_burst[rx_count]); - } - } } if (likely(!vdev->remove)) { /* Handle guest TX*/ - tx_count = rte_vhost_dequeue_burst(dev, VIRTIO_TXQ, mbuf_pool, pkts_burst, MAX_PKT_BURST); - /* If this is the first received packet we need to learn the MAC and setup VMDQ */ - if (unlikely(vdev->ready == DEVICE_MAC_LEARNING) && tx_count) { - if (vdev->remove || (link_vmdq(vdev, pkts_burst[0]) == -1)) { - while (tx_count) - rte_pktmbuf_free(pkts_burst[--tx_count]); + for (i = 0; i < rxq; i++) { + tx_count = rte_vhost_dequeue_burst(dev, VIRTIO_TXQ + i * 2, + mbuf_pool, pkts_burst, MAX_PKT_BURST); + /* + * If this is the first received packet we need to learn + * the MAC and setup VMDQ + */ + if (unlikely(vdev->ready == DEVICE_MAC_LEARNING) && tx_count) { + if (vdev->remove || (link_vmdq(vdev, pkts_burst[0]) == -1)) { + while (tx_count) + rte_pktmbuf_free(pkts_burst[--tx_count]); + } } + while (tx_count) + virtio_tx_route(vdev, pkts_burst[--tx_count], + (uint16_t)dev->device_fh, i); } - while (tx_count) - virtio_tx_route(vdev, pkts_burst[--tx_count], (uint16_t)dev->device_fh); } /*move to the next device in the list*/ @@ -2677,12 +2697,12 @@ new_device (struct virtio_net *dev) } } - /* Add device to main ll */ ll_dev = get_data_ll_free_entry(&ll_root_free); if (ll_dev == NULL) { - RTE_LOG(INFO, VHOST_DATA, "(%"PRIu64") No free entry found in linked list. Device limit " - "of %d devices per core has been reached\n", + RTE_LOG(INFO, VHOST_DATA, + "(%"PRIu64") No free entry found in linked list." + "Device limit of %d devices per core has been reached\n", dev->device_fh, num_devices); if (vdev->regions_hpa) rte_free(vdev->regions_hpa); @@ -2691,8 +2711,12 @@ new_device (struct virtio_net *dev) } ll_dev->vdev = vdev; add_data_ll_entry(&ll_root_used, ll_dev); - vdev->vmdq_rx_q - = dev->device_fh * queues_per_pool + vmdq_queue_base; + vdev->vmdq_rx_q = dev->device_fh * rxq + vmdq_queue_base; + + if ((rxq > 1) && (queues_per_pool != rxq)) { + RTE_LOG(INFO, VHOST_DATA, "(%"PRIu64") rxq: %d != queues_per_pool: %d \n", + dev->device_fh, rxq, queues_per_pool); + } if (zero_copy) { uint32_t index = vdev->vmdq_rx_q; @@ -2938,6 +2962,8 @@ main(int argc, char *argv[]) if (ret < 0) rte_exit(EXIT_FAILURE, "Invalid argument\n"); + rte_vhost_q_num_set(rxq); + for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id ++) if (rte_lcore_is_enabled(lcore_id)) lcore_ids[core_id ++] = lcore_id;