[v3] vhost: add external message handling callbacks to the public API
diff mbox series

Message ID 20190117153226.72792-1-dariusz.stojaczyk@intel.com
State Accepted, archived
Delegated to: Maxime Coquelin
Headers show
Series
  • [v3] vhost: add external message handling callbacks to the public API
Related show

Checks

Context Check Description
ci/checkpatch success coding style OK
ci/mellanox-Performance-Testing success Performance Testing PASS
ci/intel-Performance-Testing success Performance Testing PASS
ci/Intel-compilation success Compilation OK

Commit Message

Stojaczyk, Dariusz Jan. 17, 2019, 3:32 p.m. UTC
External message callbacks are used e.g. by vhost crypto
to parse crypto-specific vhost-user messages.

We are now publishing the API to register those callbacks,
so that other backends outside of DPDK can use them as well.

Signed-off-by: Darek Stojaczyk <dariusz.stojaczyk@intel.com>
Reviewed-by: Tiwei Bie <tiwei.bie@intel.com>
---
 lib/librte_vhost/rte_vhost.h           |  69 ++++++++++++++
 lib/librte_vhost/rte_vhost_version.map |   1 +
 lib/librte_vhost/vhost.c               |  13 +++
 lib/librte_vhost/vhost.h               |  54 +----------
 lib/librte_vhost/vhost_crypto.c        |  12 +--
 lib/librte_vhost/vhost_user.c          | 123 +++++++++++++------------
 6 files changed, 153 insertions(+), 119 deletions(-)

Comments

Maxime Coquelin Jan. 21, 2019, 8:25 a.m. UTC | #1
On 1/17/19 4:32 PM, Darek Stojaczyk wrote:
> External message callbacks are used e.g. by vhost crypto
> to parse crypto-specific vhost-user messages.
> 
> We are now publishing the API to register those callbacks,
> so that other backends outside of DPDK can use them as well.
> 
> Signed-off-by: Darek Stojaczyk <dariusz.stojaczyk@intel.com>
> Reviewed-by: Tiwei Bie <tiwei.bie@intel.com>
> ---
>   lib/librte_vhost/rte_vhost.h           |  69 ++++++++++++++
>   lib/librte_vhost/rte_vhost_version.map |   1 +
>   lib/librte_vhost/vhost.c               |  13 +++
>   lib/librte_vhost/vhost.h               |  54 +----------
>   lib/librte_vhost/vhost_crypto.c        |  12 +--
>   lib/librte_vhost/vhost_user.c          | 123 +++++++++++++------------
>   6 files changed, 153 insertions(+), 119 deletions(-)
> 


Reviewed-by: Maxime Coquelin <maxime.coquelin@redhat.com>

I'll apply it however only for v19.05 because API changes should
be submitted early in the release cycle (before -rc1, -rc2 at he
latest).

Thanks!
Maxime
Maxime Coquelin Feb. 8, 2019, 6:04 p.m. UTC | #2
On 1/17/19 4:32 PM, Darek Stojaczyk wrote:
> External message callbacks are used e.g. by vhost crypto
> to parse crypto-specific vhost-user messages.
> 
> We are now publishing the API to register those callbacks,
> so that other backends outside of DPDK can use them as well.
> 
> Signed-off-by: Darek Stojaczyk <dariusz.stojaczyk@intel.com>
> Reviewed-by: Tiwei Bie <tiwei.bie@intel.com>
> ---
>   lib/librte_vhost/rte_vhost.h           |  69 ++++++++++++++
>   lib/librte_vhost/rte_vhost_version.map |   1 +
>   lib/librte_vhost/vhost.c               |  13 +++
>   lib/librte_vhost/vhost.h               |  54 +----------
>   lib/librte_vhost/vhost_crypto.c        |  12 +--
>   lib/librte_vhost/vhost_user.c          | 123 +++++++++++++------------
>   6 files changed, 153 insertions(+), 119 deletions(-)
> 

I fixed build issue with postcopy build flag enabled.

Applied to dpdk-next-virtio/master.

Thanks,
Maxime

Patch
diff mbox series

diff --git a/lib/librte_vhost/rte_vhost.h b/lib/librte_vhost/rte_vhost.h
index d280ac420..2753670a2 100644
--- a/lib/librte_vhost/rte_vhost.h
+++ b/lib/librte_vhost/rte_vhost.h
@@ -111,6 +111,59 @@  struct rte_vhost_vring {
 	uint16_t		size;
 };
 
+/**
+ * Possible results of the vhost user message handling callbacks
+ */
+enum rte_vhost_msg_result {
+	/* Message handling failed */
+	RTE_VHOST_MSG_RESULT_ERR = -1,
+	/* Message handling successful */
+	RTE_VHOST_MSG_RESULT_OK =  0,
+	/* Message handling successful and reply prepared */
+	RTE_VHOST_MSG_RESULT_REPLY =  1,
+};
+
+/**
+ * Function prototype for the vhost backend to handler specific vhost user
+ * messages prior to the master message handling
+ *
+ * @param vid
+ *  vhost device id
+ * @param msg
+ *  Message pointer.
+ * @param skip_master
+ *  If the handler requires skipping the master message handling, this variable
+ *  shall be written 1, otherwise 0.
+ * @return
+ *  VH_RESULT_OK on success, VH_RESULT_REPLY on success with reply,
+ *  VH_RESULT_ERR on failure
+ */
+typedef enum rte_vhost_msg_result (*rte_vhost_msg_pre_handle)(int vid,
+		void *msg, uint32_t *skip_master);
+
+/**
+ * Function prototype for the vhost backend to handler specific vhost user
+ * messages after the master message handling is done
+ *
+ * @param vid
+ *  vhost device id
+ * @param msg
+ *  Message pointer.
+ * @return
+ *  VH_RESULT_OK on success, VH_RESULT_REPLY on success with reply,
+ *  VH_RESULT_ERR on failure
+ */
+typedef enum rte_vhost_msg_result (*rte_vhost_msg_post_handle)(int vid,
+		void *msg);
+
+/**
+ * Optional vhost user message handlers.
+ */
+struct rte_vhost_user_extern_ops {
+	rte_vhost_msg_pre_handle pre_msg_handle;
+	rte_vhost_msg_post_handle post_msg_handle;
+};
+
 /**
  * Device and vring operations.
  */
@@ -640,6 +693,22 @@  int __rte_experimental
 rte_vhost_set_vring_base(int vid, uint16_t queue_id,
 		uint16_t last_avail_idx, uint16_t last_used_idx);
 
+/**
+ * Register external message handling callbacks
+ *
+ * @param vid
+ *  vhost device ID
+ * @param ops
+ *  virtio external callbacks to register
+ * @param ctx
+ *  additional context passed to the callbacks
+ * @return
+ *  0 on success, -1 on failure
+ */
+int __rte_experimental
+rte_vhost_extern_callback_register(int vid,
+		struct rte_vhost_user_extern_ops const * const ops, void *ctx);
+
 /**
  * Get vdpa device id for vhost device.
  *
diff --git a/lib/librte_vhost/rte_vhost_version.map b/lib/librte_vhost/rte_vhost_version.map
index dd3b4c1cb..5223dbc5d 100644
--- a/lib/librte_vhost/rte_vhost_version.map
+++ b/lib/librte_vhost/rte_vhost_version.map
@@ -86,4 +86,5 @@  EXPERIMENTAL {
 	rte_vhost_host_notifier_ctrl;
 	rte_vdpa_relay_vring_avail;
 	rte_vdpa_relay_vring_used;
+	rte_vhost_extern_callback_register;
 };
diff --git a/lib/librte_vhost/vhost.c b/lib/librte_vhost/vhost.c
index b32babee4..00ec58e01 100644
--- a/lib/librte_vhost/vhost.c
+++ b/lib/librte_vhost/vhost.c
@@ -820,3 +820,16 @@  int rte_vhost_set_vring_base(int vid, uint16_t queue_id,
 
 	return 0;
 }
+
+int rte_vhost_extern_callback_register(int vid,
+		struct rte_vhost_user_extern_ops const * const ops, void *ctx)
+{
+	struct virtio_net *dev = get_device(vid);
+
+	if (!dev)
+		return -1;
+
+	dev->extern_ops = *ops;
+	dev->extern_data = ctx;
+	return 0;
+}
diff --git a/lib/librte_vhost/vhost.h b/lib/librte_vhost/vhost.h
index f267f898c..fc31796bf 100644
--- a/lib/librte_vhost/vhost.h
+++ b/lib/librte_vhost/vhost.h
@@ -286,56 +286,6 @@  struct guest_page {
 	uint64_t size;
 };
 
-/* The possible results of a message handling function */
-enum vh_result {
-	/* Message handling failed */
-	VH_RESULT_ERR   = -1,
-	/* Message handling successful */
-	VH_RESULT_OK    =  0,
-	/* Message handling successful and reply prepared */
-	VH_RESULT_REPLY =  1,
-};
-
-/**
- * function prototype for the vhost backend to handler specific vhost user
- * messages prior to the master message handling
- *
- * @param vid
- *  vhost device id
- * @param msg
- *  Message pointer.
- * @param skip_master
- *  If the handler requires skipping the master message handling, this variable
- *  shall be written 1, otherwise 0.
- * @return
- *  VH_RESULT_OK on success, VH_RESULT_REPLY on success with reply,
- *  VH_RESULT_ERR on failure
- */
-typedef enum vh_result (*vhost_msg_pre_handle)(int vid, void *msg,
-		uint32_t *skip_master);
-
-/**
- * function prototype for the vhost backend to handler specific vhost user
- * messages after the master message handling is done
- *
- * @param vid
- *  vhost device id
- * @param msg
- *  Message pointer.
- * @return
- *  VH_RESULT_OK on success, VH_RESULT_REPLY on success with reply,
- *  VH_RESULT_ERR on failure
- */
-typedef enum vh_result (*vhost_msg_post_handle)(int vid, void *msg);
-
-/**
- * pre and post vhost user message handlers
- */
-struct vhost_user_extern_ops {
-	vhost_msg_pre_handle pre_msg_handle;
-	vhost_msg_post_handle post_msg_handle;
-};
-
 /**
  * Device structure contains all configuration information relating
  * to the device.
@@ -379,10 +329,10 @@  struct virtio_net {
 	 */
 	int			vdpa_dev_id;
 
-	/* private data for virtio device */
+	/* context data for the external message handlers */
 	void			*extern_data;
 	/* pre and post vhost user message handlers for the device */
-	struct vhost_user_extern_ops extern_ops;
+	struct rte_vhost_user_extern_ops extern_ops;
 } __rte_cache_aligned;
 
 static __rte_always_inline bool
diff --git a/lib/librte_vhost/vhost_crypto.c b/lib/librte_vhost/vhost_crypto.c
index ebf26f030..0f437c4a1 100644
--- a/lib/librte_vhost/vhost_crypto.c
+++ b/lib/librte_vhost/vhost_crypto.c
@@ -434,33 +434,33 @@  vhost_crypto_close_sess(struct vhost_crypto *vcrypto, uint64_t session_id)
 	return 0;
 }
 
-static enum vh_result
+static enum rte_vhost_msg_result
 vhost_crypto_msg_post_handler(int vid, void *msg)
 {
 	struct virtio_net *dev = get_device(vid);
 	struct vhost_crypto *vcrypto;
 	VhostUserMsg *vmsg = msg;
-	enum vh_result ret = VH_RESULT_OK;
+	enum rte_vhost_msg_result ret = RTE_VHOST_MSG_RESULT_OK;
 
 	if (dev == NULL) {
 		VC_LOG_ERR("Invalid vid %i", vid);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	vcrypto = dev->extern_data;
 	if (vcrypto == NULL) {
 		VC_LOG_ERR("Cannot find required data, is it initialized?");
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	if (vmsg->request.master == VHOST_USER_CRYPTO_CREATE_SESS) {
 		vhost_crypto_create_sess(vcrypto,
 				&vmsg->payload.crypto_session);
 		vmsg->fd_num = 0;
-		ret = VH_RESULT_REPLY;
+		ret = RTE_VHOST_MSG_RESULT_REPLY;
 	} else if (vmsg->request.master == VHOST_USER_CRYPTO_CLOSE_SESS) {
 		if (vhost_crypto_close_sess(vcrypto, vmsg->payload.u64))
-			ret = VH_RESULT_ERR;
+			ret = RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	return ret;
diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c
index 1843e032f..bf7bf1745 100644
--- a/lib/librte_vhost/vhost_user.c
+++ b/lib/librte_vhost/vhost_user.c
@@ -150,7 +150,7 @@  vhost_user_set_owner(struct virtio_net **pdev __rte_unused,
 			struct VhostUserMsg *msg __rte_unused,
 			int main_fd __rte_unused)
 {
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -163,7 +163,7 @@  vhost_user_reset_owner(struct virtio_net **pdev,
 
 	cleanup_device(dev, 0);
 	reset_device(dev);
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 /*
@@ -182,7 +182,7 @@  vhost_user_get_features(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	msg->size = sizeof(msg->payload.u64);
 	msg->fd_num = 0;
 
-	return VH_RESULT_REPLY;
+	return RTE_VHOST_MSG_RESULT_REPLY;
 }
 
 /*
@@ -201,7 +201,7 @@  vhost_user_get_queue_num(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	msg->size = sizeof(msg->payload.u64);
 	msg->fd_num = 0;
 
-	return VH_RESULT_REPLY;
+	return RTE_VHOST_MSG_RESULT_REPLY;
 }
 
 /*
@@ -222,12 +222,12 @@  vhost_user_set_features(struct virtio_net **pdev, struct VhostUserMsg *msg,
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"(%d) received invalid negotiated features.\n",
 			dev->vid);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	if (dev->flags & VIRTIO_DEV_RUNNING) {
 		if (dev->features == features)
-			return VH_RESULT_OK;
+			return RTE_VHOST_MSG_RESULT_OK;
 
 		/*
 		 * Error out if master tries to change features while device is
@@ -238,7 +238,7 @@  vhost_user_set_features(struct virtio_net **pdev, struct VhostUserMsg *msg,
 			RTE_LOG(ERR, VHOST_CONFIG,
 				"(%d) features changed while device is running.\n",
 				dev->vid);
-			return VH_RESULT_ERR;
+			return RTE_VHOST_MSG_RESULT_ERR;
 		}
 
 		if (dev->notify_ops->features_changed)
@@ -283,7 +283,7 @@  vhost_user_set_features(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	if (vdpa_dev && vdpa_dev->ops->set_features)
 		vdpa_dev->ops->set_features(dev->vid);
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 /*
@@ -307,7 +307,7 @@  vhost_user_set_vring_num(struct virtio_net **pdev,
 	if ((vq->size & (vq->size - 1)) || vq->size > 32768) {
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"invalid virtqueue size %u\n", vq->size);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	if (dev->dequeue_zero_copy) {
@@ -333,7 +333,7 @@  vhost_user_set_vring_num(struct virtio_net **pdev,
 		if (!vq->shadow_used_packed) {
 			RTE_LOG(ERR, VHOST_CONFIG,
 					"failed to allocate memory for shadow used ring.\n");
-			return VH_RESULT_ERR;
+			return RTE_VHOST_MSG_RESULT_ERR;
 		}
 
 	} else {
@@ -343,7 +343,7 @@  vhost_user_set_vring_num(struct virtio_net **pdev,
 		if (!vq->shadow_used_split) {
 			RTE_LOG(ERR, VHOST_CONFIG,
 					"failed to allocate memory for shadow used ring.\n");
-			return VH_RESULT_ERR;
+			return RTE_VHOST_MSG_RESULT_ERR;
 		}
 	}
 
@@ -353,10 +353,10 @@  vhost_user_set_vring_num(struct virtio_net **pdev,
 	if (!vq->batch_copy_elems) {
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"failed to allocate memory for batching copy.\n");
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 /*
@@ -669,7 +669,7 @@  vhost_user_set_vring_addr(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	struct vhost_vring_addr *addr = &msg->payload.addr;
 
 	if (dev->mem == NULL)
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 
 	/* addr->index refers to the queue index. The txq 1, rxq is 0. */
 	vq = dev->virtqueue[msg->payload.addr.index];
@@ -686,12 +686,12 @@  vhost_user_set_vring_addr(struct virtio_net **pdev, struct VhostUserMsg *msg,
 				(1ULL << VHOST_USER_F_PROTOCOL_FEATURES))) {
 		dev = translate_ring_addresses(dev, msg->payload.addr.index);
 		if (!dev)
-			return VH_RESULT_ERR;
+			return RTE_VHOST_MSG_RESULT_ERR;
 
 		*pdev = dev;
 	}
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 /*
@@ -725,7 +725,7 @@  vhost_user_set_vring_base(struct virtio_net **pdev,
 		vq->last_avail_idx = msg->payload.state.num;
 	}
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -867,7 +867,7 @@  vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	if (memory->nregions > VHOST_MEMORY_MAX_NREGIONS) {
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"too many memory regions (%u)\n", memory->nregions);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	if (dev->mem && !vhost_memory_changed(memory, dev->mem)) {
@@ -877,7 +877,7 @@  vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
 		for (i = 0; i < memory->nregions; i++)
 			close(msg->fds[i]);
 
-		return VH_RESULT_OK;
+		return RTE_VHOST_MSG_RESULT_OK;
 	}
 
 	if (dev->mem) {
@@ -901,7 +901,7 @@  vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
 				"(%d) failed to allocate memory "
 				"for dev->guest_pages\n",
 				dev->vid);
-			return VH_RESULT_ERR;
+			return RTE_VHOST_MSG_RESULT_ERR;
 		}
 	}
 
@@ -911,7 +911,7 @@  vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"(%d) failed to allocate memory for dev->mem\n",
 			dev->vid);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 	dev->mem->nregions = memory->nregions;
 
@@ -1082,13 +1082,13 @@  vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
 
 	dump_guest_pages(dev);
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 
 err_mmap:
 	free_mem_region(dev);
 	rte_free(dev->mem);
 	dev->mem = NULL;
-	return VH_RESULT_ERR;
+	return RTE_VHOST_MSG_RESULT_ERR;
 }
 
 static bool
@@ -1152,7 +1152,7 @@  vhost_user_set_vring_call(struct virtio_net **pdev, struct VhostUserMsg *msg,
 
 	vq->callfd = file.fd;
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int vhost_user_set_vring_err(struct virtio_net **pdev __rte_unused,
@@ -1163,7 +1163,7 @@  static int vhost_user_set_vring_err(struct virtio_net **pdev __rte_unused,
 		close(msg->fds[0]);
 	RTE_LOG(INFO, VHOST_CONFIG, "not implemented\n");
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -1185,7 +1185,7 @@  vhost_user_set_vring_kick(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	/* Interpret ring addresses only when ring is started. */
 	dev = translate_ring_addresses(dev, file.index);
 	if (!dev)
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 
 	*pdev = dev;
 
@@ -1203,7 +1203,7 @@  vhost_user_set_vring_kick(struct virtio_net **pdev, struct VhostUserMsg *msg,
 		close(vq->kickfd);
 	vq->kickfd = file.fd;
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static void
@@ -1287,7 +1287,7 @@  vhost_user_get_vring_base(struct virtio_net **pdev,
 	msg->size = sizeof(msg->payload.state);
 	msg->fd_num = 0;
 
-	return VH_RESULT_REPLY;
+	return RTE_VHOST_MSG_RESULT_REPLY;
 }
 
 /*
@@ -1320,7 +1320,7 @@  vhost_user_set_vring_enable(struct virtio_net **pdev,
 
 	dev->virtqueue[index]->enabled = enable;
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -1347,7 +1347,7 @@  vhost_user_get_protocol_features(struct virtio_net **pdev,
 	msg->size = sizeof(msg->payload.u64);
 	msg->fd_num = 0;
 
-	return VH_RESULT_REPLY;
+	return RTE_VHOST_MSG_RESULT_REPLY;
 }
 
 static int
@@ -1365,12 +1365,12 @@  vhost_user_set_protocol_features(struct virtio_net **pdev,
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"(%d) received invalid protocol features.\n",
 			dev->vid);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	dev->protocol_features = protocol_features;
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -1384,14 +1384,14 @@  vhost_user_set_log_base(struct virtio_net **pdev, struct VhostUserMsg *msg,
 
 	if (fd < 0) {
 		RTE_LOG(ERR, VHOST_CONFIG, "invalid log fd: %d\n", fd);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	if (msg->size != sizeof(VhostUserLog)) {
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"invalid log base msg size: %"PRId32" != %d\n",
 			msg->size, (int)sizeof(VhostUserLog));
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	size = msg->payload.log.mmap_size;
@@ -1402,7 +1402,7 @@  vhost_user_set_log_base(struct virtio_net **pdev, struct VhostUserMsg *msg,
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"log offset %#"PRIx64" exceeds log size %#"PRIx64"\n",
 			off, size);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	RTE_LOG(INFO, VHOST_CONFIG,
@@ -1417,7 +1417,7 @@  vhost_user_set_log_base(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	close(fd);
 	if (addr == MAP_FAILED) {
 		RTE_LOG(ERR, VHOST_CONFIG, "mmap log base failed!\n");
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	/*
@@ -1438,7 +1438,7 @@  vhost_user_set_log_base(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	msg->size = 0;
 	msg->fd_num = 0;
 
-	return VH_RESULT_REPLY;
+	return RTE_VHOST_MSG_RESULT_REPLY;
 }
 
 static int vhost_user_set_log_fd(struct virtio_net **pdev __rte_unused,
@@ -1448,7 +1448,7 @@  static int vhost_user_set_log_fd(struct virtio_net **pdev __rte_unused,
 	close(msg->fds[0]);
 	RTE_LOG(INFO, VHOST_CONFIG, "not implemented.\n");
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 /*
@@ -1487,7 +1487,7 @@  vhost_user_send_rarp(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	if (vdpa_dev && vdpa_dev->ops->migration_done)
 		vdpa_dev->ops->migration_done(dev->vid);
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -1500,12 +1500,12 @@  vhost_user_net_set_mtu(struct virtio_net **pdev, struct VhostUserMsg *msg,
 		RTE_LOG(ERR, VHOST_CONFIG, "Invalid MTU size (%"PRIu64")\n",
 				msg->payload.u64);
 
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	dev->mtu = msg->payload.u64;
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -1519,12 +1519,12 @@  vhost_user_set_req_fd(struct virtio_net **pdev, struct VhostUserMsg *msg,
 		RTE_LOG(ERR, VHOST_CONFIG,
 				"Invalid file descriptor for slave channel (%d)\n",
 				fd);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 	dev->slave_req_fd = fd;
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -1590,7 +1590,7 @@  vhost_user_iotlb_msg(struct virtio_net **pdev, struct VhostUserMsg *msg,
 		len = imsg->size;
 		vva = qva_to_vva(dev, imsg->uaddr, &len);
 		if (!vva)
-			return VH_RESULT_ERR;
+			return RTE_VHOST_MSG_RESULT_ERR;
 
 		for (i = 0; i < dev->nr_vring; i++) {
 			struct vhost_virtqueue *vq = dev->virtqueue[i];
@@ -1616,10 +1616,10 @@  vhost_user_iotlb_msg(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	default:
 		RTE_LOG(ERR, VHOST_CONFIG, "Invalid IOTLB message type (%d)\n",
 				imsg->type);
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -1655,7 +1655,7 @@  vhost_user_set_postcopy_advise(struct virtio_net **pdev,
 	dev->postcopy_ufd = -1;
 	msg->fd_num = 0;
 
-	return VH_RESULT_ERR;
+	return RTE_VHOST_MSG_RESULT_ERR;
 #endif
 }
 
@@ -1669,11 +1669,11 @@  vhost_user_set_postcopy_listen(struct virtio_net **pdev,
 	if (dev->mem && dev->mem->nregions) {
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"Regions already registered at postcopy-listen\n");
-		return VH_RESULT_ERR;
+		return RTE_VHOST_MSG_RESULT_ERR;
 	}
 	dev->postcopy_listening = 1;
 
-	return VH_RESULT_OK;
+	return RTE_VHOST_MSG_RESULT_OK;
 }
 
 static int
@@ -1692,7 +1692,7 @@  vhost_user_postcopy_end(struct virtio_net **pdev, struct VhostUserMsg *msg,
 	msg->size = sizeof(msg->payload.u64);
 	msg->fd_num = 0;
 
-	return VH_RESULT_REPLY;
+	return RTE_VHOST_MSG_RESULT_REPLY;
 }
 
 typedef int (*vhost_message_handler_t)(struct virtio_net **pdev,
@@ -1961,9 +1961,9 @@  vhost_user_msg_handler(int vid, int fd)
 	if (dev->extern_ops.pre_msg_handle) {
 		ret = (*dev->extern_ops.pre_msg_handle)(dev->vid,
 				(void *)&msg, &skip_master);
-		if (ret == VH_RESULT_ERR)
+		if (ret == RTE_VHOST_MSG_RESULT_ERR)
 			goto skip_to_reply;
-		else if (ret == VH_RESULT_REPLY)
+		else if (ret == RTE_VHOST_MSG_RESULT_REPLY)
 			send_vhost_reply(fd, &msg);
 
 		if (skip_master)
@@ -1977,17 +1977,17 @@  vhost_user_msg_handler(int vid, int fd)
 		ret = vhost_message_handlers[request](&dev, &msg, fd);
 
 		switch (ret) {
-		case VH_RESULT_ERR:
+		case RTE_VHOST_MSG_RESULT_ERR:
 			RTE_LOG(ERR, VHOST_CONFIG,
 				"Processing %s failed.\n",
 				vhost_message_str[request]);
 			break;
-		case VH_RESULT_OK:
+		case RTE_VHOST_MSG_RESULT_OK:
 			RTE_LOG(DEBUG, VHOST_CONFIG,
 				"Processing %s succeeded.\n",
 				vhost_message_str[request]);
 			break;
-		case VH_RESULT_REPLY:
+		case RTE_VHOST_MSG_RESULT_REPLY:
 			RTE_LOG(DEBUG, VHOST_CONFIG,
 				"Processing %s succeeded and needs reply.\n",
 				vhost_message_str[request]);
@@ -1997,16 +1997,17 @@  vhost_user_msg_handler(int vid, int fd)
 	} else {
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"Requested invalid message type %d.\n", request);
-		ret = VH_RESULT_ERR;
+		ret = RTE_VHOST_MSG_RESULT_ERR;
 	}
 
 skip_to_post_handle:
-	if (ret != VH_RESULT_ERR && dev->extern_ops.post_msg_handle) {
+	if (ret != RTE_VHOST_MSG_RESULT_ERR &&
+			dev->extern_ops.post_msg_handle) {
 		ret = (*dev->extern_ops.post_msg_handle)(
 				dev->vid, (void *)&msg);
-		if (ret == VH_RESULT_ERR)
+		if (ret == RTE_VHOST_MSG_RESULT_ERR)
 			goto skip_to_reply;
-		else if (ret == VH_RESULT_REPLY)
+		else if (ret == RTE_VHOST_MSG_RESULT_REPLY)
 			send_vhost_reply(fd, &msg);
 	}
 
@@ -2020,11 +2021,11 @@  vhost_user_msg_handler(int vid, int fd)
 	 * VHOST_USER_NEED_REPLY was cleared in send_vhost_reply().
 	 */
 	if (msg.flags & VHOST_USER_NEED_REPLY) {
-		msg.payload.u64 = ret == VH_RESULT_ERR;
+		msg.payload.u64 = ret == RTE_VHOST_MSG_RESULT_ERR;
 		msg.size = sizeof(msg.payload.u64);
 		msg.fd_num = 0;
 		send_vhost_reply(fd, &msg);
-	} else if (ret == VH_RESULT_ERR) {
+	} else if (ret == RTE_VHOST_MSG_RESULT_ERR) {
 		RTE_LOG(ERR, VHOST_CONFIG,
 			"vhost message handling failed.\n");
 		return -1;