From patchwork Thu Feb 29 13:29:17 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Srujana Challa X-Patchwork-Id: 137494 X-Patchwork-Delegate: maxime.coquelin@redhat.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 6474643C1D; Thu, 29 Feb 2024 14:29:32 +0100 (CET) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 2710F42D80; Thu, 29 Feb 2024 14:29:31 +0100 (CET) Received: from mx0b-0016f401.pphosted.com (mx0a-0016f401.pphosted.com [67.231.148.174]) by mails.dpdk.org (Postfix) with ESMTP id F0FAB402B4 for ; Thu, 29 Feb 2024 14:29:28 +0100 (CET) Received: from pps.filterd (m0045849.ppops.net [127.0.0.1]) by mx0a-0016f401.pphosted.com (8.17.1.24/8.17.1.24) with ESMTP id 41T83C0b022376; Thu, 29 Feb 2024 05:29:28 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=marvell.com; h= from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding:content-type; s= pfpt0220; bh=UaqhdiRMPzbuHbmxnxvDjc8ouRkOcEfE3i42GRDWkuQ=; b=T3O eRah1HdnJHw6gnGjIhBNO0QQlFQG/70iVvxudjNOsJS/qu4Ff3oqSc0ZTGW2FQ/p TtNOZPx9aBTEGJXjxDfl+wzHQnCQjgHGriwpZ/X1ctHKNBzG4vOAxiCQS9kmBcs7 BN35OVhd5YvjPzTkgPr/dzBSCLkqWtg6pI6YC9fBWITPAv4QRNCGp5eWGURZlQLm Of5k4NN1pFqa0mhOYxX8RL5YBSc2HkOXu1F93R9IDkhcp+hINBHKGoET4exEtLHa gixR6N9dSgzuMuIYgx9Y8fp02mit4dAcXrucyfPt2o7SJHvjfORKrM/7hye/FYmN dzrLNPClOpEGzNuMTzw== Received: from dc6wp-exch02.marvell.com ([4.21.29.225]) by mx0a-0016f401.pphosted.com (PPS) with ESMTPS id 3wjp2y91j8-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Thu, 29 Feb 2024 05:29:27 -0800 (PST) Received: from DC6WP-EXCH02.marvell.com (10.76.176.209) by DC6WP-EXCH02.marvell.com (10.76.176.209) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.1258.12; Thu, 29 Feb 2024 05:29:26 -0800 Received: from maili.marvell.com (10.69.176.80) by DC6WP-EXCH02.marvell.com (10.76.176.209) with Microsoft SMTP Server id 15.2.1258.12 via Frontend Transport; Thu, 29 Feb 2024 05:29:26 -0800 Received: from localhost.localdomain (unknown [10.28.36.175]) by maili.marvell.com (Postfix) with ESMTP id DC8303F7252; Thu, 29 Feb 2024 05:29:23 -0800 (PST) From: Srujana Challa To: , , CC: , , , Subject: [PATCH v2 1/3] net/virtio_user: avoid cq descriptor buffer address accessing Date: Thu, 29 Feb 2024 18:59:17 +0530 Message-ID: <20240229132919.2186118-2-schalla@marvell.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20240229132919.2186118-1-schalla@marvell.com> References: <20240229132919.2186118-1-schalla@marvell.com> MIME-Version: 1.0 X-Proofpoint-GUID: EpgNGOEFpKWhri3nZL3OtwFDpXFWpaIb X-Proofpoint-ORIG-GUID: EpgNGOEFpKWhri3nZL3OtwFDpXFWpaIb X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.272,Aquarius:18.0.1011,Hydra:6.0.619,FMLib:17.11.176.26 definitions=2024-02-29_02,2024-02-29_01,2023-05-22_02 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org This patch makes changes to avoid descriptor buffer address accessing while processing shadow control queue. So that Virtio-user can work with having IOVA as descriptor buffer address. Signed-off-by: Srujana Challa --- .../net/virtio/virtio_user/virtio_user_dev.c | 68 +++++++++---------- 1 file changed, 33 insertions(+), 35 deletions(-) diff --git a/drivers/net/virtio/virtio_user/virtio_user_dev.c b/drivers/net/virtio/virtio_user/virtio_user_dev.c index d395fc1676..bf3da4340f 100644 --- a/drivers/net/virtio/virtio_user/virtio_user_dev.c +++ b/drivers/net/virtio/virtio_user/virtio_user_dev.c @@ -885,11 +885,11 @@ static uint32_t virtio_user_handle_ctrl_msg_split(struct virtio_user_dev *dev, struct vring *vring, uint16_t idx_hdr) { - struct virtio_net_ctrl_hdr *hdr; virtio_net_ctrl_ack status = ~0; - uint16_t i, idx_data, idx_status; + uint16_t i, idx_data; uint32_t n_descs = 0; int dlen[CVQ_MAX_DATA_DESCS], nb_dlen = 0; + struct virtio_pmd_ctrl *ctrl; /* locate desc for header, data, and status */ idx_data = vring->desc[idx_hdr].next; @@ -902,34 +902,33 @@ virtio_user_handle_ctrl_msg_split(struct virtio_user_dev *dev, struct vring *vri n_descs++; } - /* locate desc for status */ - idx_status = i; n_descs++; - hdr = (void *)(uintptr_t)vring->desc[idx_hdr].addr; - if (hdr->class == VIRTIO_NET_CTRL_MQ && - hdr->cmd == VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET) { - uint16_t queues; + /* Access control command via VA from CVQ */ + ctrl = (struct virtio_pmd_ctrl *)dev->hw.cvq->hdr_mz->addr; + if (ctrl->hdr.class == VIRTIO_NET_CTRL_MQ && + ctrl->hdr.cmd == VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET) { + uint16_t *queues; - queues = *(uint16_t *)(uintptr_t)vring->desc[idx_data].addr; - status = virtio_user_handle_mq(dev, queues); - } else if (hdr->class == VIRTIO_NET_CTRL_MQ && hdr->cmd == VIRTIO_NET_CTRL_MQ_RSS_CONFIG) { + queues = (uint16_t *)ctrl->data; + status = virtio_user_handle_mq(dev, *queues); + } else if (ctrl->hdr.class == VIRTIO_NET_CTRL_MQ && + ctrl->hdr.cmd == VIRTIO_NET_CTRL_MQ_RSS_CONFIG) { struct virtio_net_ctrl_rss *rss; - rss = (struct virtio_net_ctrl_rss *)(uintptr_t)vring->desc[idx_data].addr; + rss = (struct virtio_net_ctrl_rss *)ctrl->data; status = virtio_user_handle_mq(dev, rss->max_tx_vq); - } else if (hdr->class == VIRTIO_NET_CTRL_RX || - hdr->class == VIRTIO_NET_CTRL_MAC || - hdr->class == VIRTIO_NET_CTRL_VLAN) { + } else if (ctrl->hdr.class == VIRTIO_NET_CTRL_RX || + ctrl->hdr.class == VIRTIO_NET_CTRL_MAC || + ctrl->hdr.class == VIRTIO_NET_CTRL_VLAN) { status = 0; } if (!status && dev->scvq) - status = virtio_send_command(&dev->scvq->cq, - (struct virtio_pmd_ctrl *)hdr, dlen, nb_dlen); + status = virtio_send_command(&dev->scvq->cq, ctrl, dlen, nb_dlen); /* Update status */ - *(virtio_net_ctrl_ack *)(uintptr_t)vring->desc[idx_status].addr = status; + ctrl->status = status; return n_descs; } @@ -948,7 +947,7 @@ virtio_user_handle_ctrl_msg_packed(struct virtio_user_dev *dev, struct vring_packed *vring, uint16_t idx_hdr) { - struct virtio_net_ctrl_hdr *hdr; + struct virtio_pmd_ctrl *ctrl; virtio_net_ctrl_ack status = ~0; uint16_t idx_data, idx_status; /* initialize to one, header is first */ @@ -971,32 +970,31 @@ virtio_user_handle_ctrl_msg_packed(struct virtio_user_dev *dev, n_descs++; } - hdr = (void *)(uintptr_t)vring->desc[idx_hdr].addr; - if (hdr->class == VIRTIO_NET_CTRL_MQ && - hdr->cmd == VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET) { - uint16_t queues; + /* Access control command via VA from CVQ */ + ctrl = (struct virtio_pmd_ctrl *)dev->hw.cvq->hdr_mz->addr; + if (ctrl->hdr.class == VIRTIO_NET_CTRL_MQ && + ctrl->hdr.cmd == VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET) { + uint16_t *queues; - queues = *(uint16_t *)(uintptr_t) - vring->desc[idx_data].addr; - status = virtio_user_handle_mq(dev, queues); - } else if (hdr->class == VIRTIO_NET_CTRL_MQ && hdr->cmd == VIRTIO_NET_CTRL_MQ_RSS_CONFIG) { + queues = (uint16_t *)ctrl->data; + status = virtio_user_handle_mq(dev, *queues); + } else if (ctrl->hdr.class == VIRTIO_NET_CTRL_MQ && + ctrl->hdr.cmd == VIRTIO_NET_CTRL_MQ_RSS_CONFIG) { struct virtio_net_ctrl_rss *rss; - rss = (struct virtio_net_ctrl_rss *)(uintptr_t)vring->desc[idx_data].addr; + rss = (struct virtio_net_ctrl_rss *)ctrl->data; status = virtio_user_handle_mq(dev, rss->max_tx_vq); - } else if (hdr->class == VIRTIO_NET_CTRL_RX || - hdr->class == VIRTIO_NET_CTRL_MAC || - hdr->class == VIRTIO_NET_CTRL_VLAN) { + } else if (ctrl->hdr.class == VIRTIO_NET_CTRL_RX || + ctrl->hdr.class == VIRTIO_NET_CTRL_MAC || + ctrl->hdr.class == VIRTIO_NET_CTRL_VLAN) { status = 0; } if (!status && dev->scvq) - status = virtio_send_command(&dev->scvq->cq, - (struct virtio_pmd_ctrl *)hdr, dlen, nb_dlen); + status = virtio_send_command(&dev->scvq->cq, ctrl, dlen, nb_dlen); /* Update status */ - *(virtio_net_ctrl_ack *)(uintptr_t) - vring->desc[idx_status].addr = status; + ctrl->status = status; /* Update used descriptor */ vring->desc[idx_hdr].id = vring->desc[idx_status].id;