From patchwork Wed Mar 17 17:09:21 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Maxime Coquelin X-Patchwork-Id: 89409 X-Patchwork-Delegate: maxime.coquelin@redhat.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 1D7D9A0561; Wed, 17 Mar 2021 18:09:50 +0100 (CET) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 40B52140EB9; Wed, 17 Mar 2021 18:09:44 +0100 (CET) Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [63.128.21.124]) by mails.dpdk.org (Postfix) with ESMTP id 78398140EB4 for ; Wed, 17 Mar 2021 18:09:42 +0100 (CET) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1616000982; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=13j5KElrZilxYmitE7uv3O48TN4CEyh1Q+rl3n44rx0=; b=a1kDiKZvPYyhcde/E1rNuSGOW/DK1T13kqjb7TC7WkGpMn5uvO4CkV3efT3di4ne+nMXQr vboQ9ddd/kZYmws+GF/Ri0A+Av1Mpbw5aPeNSiJ7R9v7ntjeK6nmTKe4wZcHAX/VZY873p W4oSCZb30zTyg+T4ENdgutsCtZmSRyk= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-324-XQ9tUUXjNV-aEa902N1ZWA-1; Wed, 17 Mar 2021 13:09:40 -0400 X-MC-Unique: XQ9tUUXjNV-aEa902N1ZWA-1 Received: from smtp.corp.redhat.com (int-mx01.intmail.prod.int.phx2.redhat.com [10.5.11.11]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 16E55107ACCD; Wed, 17 Mar 2021 17:09:39 +0000 (UTC) Received: from max-t490s.redhat.com (unknown [10.36.110.14]) by smtp.corp.redhat.com (Postfix) with ESMTP id 5D14E19CB6; Wed, 17 Mar 2021 17:09:37 +0000 (UTC) From: Maxime Coquelin To: dev@dpdk.org, chenbo.xia@intel.com, amorenoz@redhat.com, david.marchand@redhat.com, olivier.matz@6wind.com, bnemeth@redhat.com Cc: Maxime Coquelin Date: Wed, 17 Mar 2021 18:09:21 +0100 Message-Id: <20210317170922.25046-3-maxime.coquelin@redhat.com> In-Reply-To: <20210317170922.25046-1-maxime.coquelin@redhat.com> References: <20210317170922.25046-1-maxime.coquelin@redhat.com> MIME-Version: 1.0 X-Scanned-By: MIMEDefang 2.79 on 10.5.11.11 Authentication-Results: relay.mimecast.com; auth=pass smtp.auth=CUSA124A263 smtp.mailfrom=maxime.coquelin@redhat.com X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Subject: [dpdk-dev] [PATCH v3 2/3] vhost: move dirty logging cache out of the virtqueue X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the per-virtqueue's dirty logging cache out of the virtqueue struct, by allocating it dynamically only when live-migration is enabled. It saves 8 cachelines in vhost_virtqueue struct. Signed-off-by: Maxime Coquelin Reviewed-by: Chenbo Xia --- lib/librte_vhost/vhost.c | 13 +++++++++++++ lib/librte_vhost/vhost.h | 2 +- lib/librte_vhost/vhost_user.c | 21 +++++++++++++++++++++ 3 files changed, 35 insertions(+), 1 deletion(-) diff --git a/lib/librte_vhost/vhost.c b/lib/librte_vhost/vhost.c index 5a7c0c6cff..a8032e3ba1 100644 --- a/lib/librte_vhost/vhost.c +++ b/lib/librte_vhost/vhost.c @@ -145,6 +145,10 @@ __vhost_log_cache_sync(struct virtio_net *dev, struct vhost_virtqueue *vq) if (unlikely(!dev->log_base)) return; + /* No cache, nothing to sync */ + if (unlikely(!vq->log_cache)) + return; + rte_atomic_thread_fence(__ATOMIC_RELEASE); log_base = (unsigned long *)(uintptr_t)dev->log_base; @@ -177,6 +181,14 @@ vhost_log_cache_page(struct virtio_net *dev, struct vhost_virtqueue *vq, uint32_t offset = page / (sizeof(unsigned long) << 3); int i; + if (unlikely(!vq->log_cache)) { + /* No logging cache allocated, write dirty log map directly */ + rte_atomic_thread_fence(__ATOMIC_RELEASE); + vhost_log_page((uint8_t *)(uintptr_t)dev->log_base, page); + + return; + } + for (i = 0; i < vq->log_cache_nb_elem; i++) { struct log_cache_entry *elem = vq->log_cache + i; @@ -354,6 +366,7 @@ free_vq(struct virtio_net *dev, struct vhost_virtqueue *vq) } rte_free(vq->batch_copy_elems); rte_mempool_free(vq->iotlb_pool); + rte_free(vq->log_cache); rte_free(vq); } diff --git a/lib/librte_vhost/vhost.h b/lib/librte_vhost/vhost.h index 717f410548..3a71dfeed9 100644 --- a/lib/librte_vhost/vhost.h +++ b/lib/librte_vhost/vhost.h @@ -183,7 +183,7 @@ struct vhost_virtqueue { bool used_wrap_counter; bool avail_wrap_counter; - struct log_cache_entry log_cache[VHOST_LOG_CACHE_NR]; + struct log_cache_entry *log_cache; uint16_t log_cache_nb_elem; rte_rwlock_t iotlb_lock; diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c index a60bb945ad..4d9e76e49e 100644 --- a/lib/librte_vhost/vhost_user.c +++ b/lib/librte_vhost/vhost_user.c @@ -2022,6 +2022,9 @@ vhost_user_get_vring_base(struct virtio_net **pdev, rte_free(vq->batch_copy_elems); vq->batch_copy_elems = NULL; + rte_free(vq->log_cache); + vq->log_cache = NULL; + msg->size = sizeof(msg->payload.state); msg->fd_num = 0; @@ -2121,6 +2124,7 @@ vhost_user_set_log_base(struct virtio_net **pdev, struct VhostUserMsg *msg, int fd = msg->fds[0]; uint64_t size, off; void *addr; + uint32_t i; if (validate_msg_fds(msg, 1) != 0) return RTE_VHOST_MSG_RESULT_ERR; @@ -2174,6 +2178,23 @@ vhost_user_set_log_base(struct virtio_net **pdev, struct VhostUserMsg *msg, dev->log_base = dev->log_addr + off; dev->log_size = size; + for (i = 0; i < dev->nr_vring; i++) { + struct vhost_virtqueue *vq = dev->virtqueue[i]; + + rte_free(vq->log_cache); + vq->log_cache = NULL; + vq->log_cache_nb_elem = 0; + vq->log_cache = rte_zmalloc("vq log cache", + sizeof(struct log_cache_entry) * VHOST_LOG_CACHE_NR, + 0); + /* + * If log cache alloc fail, don't fail migration, but no + * caching will be done, which will impact performance + */ + if (!vq->log_cache) + VHOST_LOG_CONFIG(ERR, "Failed to allocate VQ logging cache\n"); + } + /* * The spec is not clear about it (yet), but QEMU doesn't expect * any payload in the reply.