get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/112365/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 112365,
    "url": "https://patches.dpdk.org/api/patches/112365/?format=api",
    "web_url": "https://patches.dpdk.org/project/dpdk/patch/20220606112109.208873-25-lizh@nvidia.com/",
    "project": {
        "id": 1,
        "url": "https://patches.dpdk.org/api/projects/1/?format=api",
        "name": "DPDK",
        "link_name": "dpdk",
        "list_id": "dev.dpdk.org",
        "list_email": "dev@dpdk.org",
        "web_url": "http://core.dpdk.org",
        "scm_url": "git://dpdk.org/dpdk",
        "webscm_url": "http://git.dpdk.org/dpdk",
        "list_archive_url": "https://inbox.dpdk.org/dev",
        "list_archive_url_format": "https://inbox.dpdk.org/dev/{}",
        "commit_url_format": ""
    },
    "msgid": "<20220606112109.208873-25-lizh@nvidia.com>",
    "list_archive_url": "https://inbox.dpdk.org/dev/20220606112109.208873-25-lizh@nvidia.com",
    "date": "2022-06-06T11:21:01",
    "name": "[13/16] vdpa/mlx5: add virtq LM log task",
    "commit_ref": null,
    "pull_url": null,
    "state": "superseded",
    "archived": true,
    "hash": "ab6e45298eb400233e109fca67571c2f5ef959aa",
    "submitter": {
        "id": 1967,
        "url": "https://patches.dpdk.org/api/people/1967/?format=api",
        "name": "Li Zhang",
        "email": "lizh@nvidia.com"
    },
    "delegate": {
        "id": 1,
        "url": "https://patches.dpdk.org/api/users/1/?format=api",
        "username": "tmonjalo",
        "first_name": "Thomas",
        "last_name": "Monjalon",
        "email": "thomas@monjalon.net"
    },
    "mbox": "https://patches.dpdk.org/project/dpdk/patch/20220606112109.208873-25-lizh@nvidia.com/mbox/",
    "series": [
        {
            "id": 23339,
            "url": "https://patches.dpdk.org/api/series/23339/?format=api",
            "web_url": "https://patches.dpdk.org/project/dpdk/list/?series=23339",
            "date": "2022-06-06T11:20:37",
            "name": "Add vDPA multi-threads optiomization",
            "version": 1,
            "mbox": "https://patches.dpdk.org/series/23339/mbox/"
        }
    ],
    "comments": "https://patches.dpdk.org/api/patches/112365/comments/",
    "check": "success",
    "checks": "https://patches.dpdk.org/api/patches/112365/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<dev-bounces@dpdk.org>",
        "X-Original-To": "patchwork@inbox.dpdk.org",
        "Delivered-To": "patchwork@inbox.dpdk.org",
        "Received": [
            "from mails.dpdk.org (mails.dpdk.org [217.70.189.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id C1B9EA0543;\n\tMon,  6 Jun 2022 13:24:39 +0200 (CEST)",
            "from [217.70.189.124] (localhost [127.0.0.1])\n\tby mails.dpdk.org (Postfix) with ESMTP id 9DDC842BD1;\n\tMon,  6 Jun 2022 13:22:50 +0200 (CEST)",
            "from NAM12-MW2-obe.outbound.protection.outlook.com\n (mail-mw2nam12on2073.outbound.protection.outlook.com [40.107.244.73])\n by mails.dpdk.org (Postfix) with ESMTP id EED6642B84\n for <dev@dpdk.org>; Mon,  6 Jun 2022 13:22:48 +0200 (CEST)",
            "from DM3PR03CA0017.namprd03.prod.outlook.com (2603:10b6:0:50::27) by\n CH2PR12MB5564.namprd12.prod.outlook.com (2603:10b6:610:65::24) with\n Microsoft\n SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id\n 15.20.5314.13; Mon, 6 Jun 2022 11:22:47 +0000",
            "from DM6NAM11FT060.eop-nam11.prod.protection.outlook.com\n (2603:10b6:0:50:cafe::6e) by DM3PR03CA0017.outlook.office365.com\n (2603:10b6:0:50::27) with Microsoft SMTP Server (version=TLS1_2,\n cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.5314.12 via Frontend\n Transport; Mon, 6 Jun 2022 11:22:47 +0000",
            "from mail.nvidia.com (12.22.5.238) by\n DM6NAM11FT060.mail.protection.outlook.com (10.13.173.63) with Microsoft SMTP\n Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384) id\n 15.20.5314.12 via Frontend Transport; Mon, 6 Jun 2022 11:22:47 +0000",
            "from rnnvmail201.nvidia.com (10.129.68.8) by DRHQMAIL105.nvidia.com\n (10.27.9.14) with Microsoft SMTP Server (TLS) id 15.0.1497.32;\n Mon, 6 Jun 2022 11:22:46 +0000",
            "from nvidia.com (10.126.231.35) by rnnvmail201.nvidia.com\n (10.129.68.8) with Microsoft SMTP Server (version=TLS1_2,\n cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.22; Mon, 6 Jun 2022\n 04:22:43 -0700"
        ],
        "ARC-Seal": "i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none;\n b=iP9jdXKEtBAnCekt2W64nfkvKtjdloKV+aZ9LVn+owVLXXPHvgh6oZJW5LGpXhuk9mJDXsa2ddHa7AOSJKQrYrb47fGMof/0DGjWqqcE3dBxQosmvKhYjz0yK16HaOAMPLyeVWyb4V6Akg0Ktf+HKHEmrsTYLnG+0Hq53HOZDyW1R5nxx/n0zQIE56apdcZikYYMJt84R2X2XuN+5Qnv8km8WM5TqjvaXrjI2dr8XYsXSHZlsgh4YtJ05OdbuWoA6wxjZGrsNVdR1wtTE4T4+HMKoKV5x/7ih4vEAE1zlZpS75iXhO2NqdvXp0aKZb/LMXn7exYM3owrW+gyl+EbjQ==",
        "ARC-Message-Signature": "i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com;\n s=arcselector9901;\n h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1;\n bh=TFlSU0NM+Vl7ACH+AnzFBZIOnfKwLgjvtRy8gMC45iw=;\n b=WSyhnJzg1vsGyorUzQjx5dAfyLhqi5Kv5jUY8ZivY33Z+aolB2H2tIG5j96pM22bluqosPzyTuBmiWlDH0Rn0Yc1eNGXYpbjJTNx27d4uJiK/UgbiTuseu58KSMAeBG2+5hdcSr7s5maVew6iYVLC/RvA7xsLrWhtIpqz/+mHsSafkbGU45Ja7dp8SvzM5HujSXFWBNo8dzOU3XvCq+gzxjm9pv4VbV/YhxVhPoU/Lle0XBEPNpt0TDKPTDnZT5VeRiSS9DTWRRgVHZem2pHO9dB/O9+zfoQ1+K1X0zguFUw5HO5Xk4AkVqTvgYe8Qe1qX0YGfCel+4JtkScnFWZTg==",
        "ARC-Authentication-Results": "i=1; mx.microsoft.com 1; spf=pass (sender ip is\n 12.22.5.238) smtp.rcpttodomain=monjalon.net smtp.mailfrom=nvidia.com;\n dmarc=pass (p=reject sp=reject pct=100) action=none header.from=nvidia.com;\n dkim=none (message not signed); arc=none",
        "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed; d=Nvidia.com;\n s=selector2;\n h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck;\n bh=TFlSU0NM+Vl7ACH+AnzFBZIOnfKwLgjvtRy8gMC45iw=;\n b=JTV8pimO9IBmsnSG7/rKE29hth8F8ycvYr5F0j3OcTi2gPFR8uW54ampUv/XkJnqMkW1qZBP+HzWH4do3ETJ5QbI2xQvZkQWMdyRotL2niJsP5eGhW8xyE3FShxVdeULqQVzd6hm0kIPM4FhkMtC/bqpCZg3qy3kxNslm2wDwwuOKmrqI6EK3F7X7+2KEg/f32U1EjtfR2ZrwTMgpzIN/vqZZiFui5MimsDnc22D/94hdFp7ss2cHDgBBAvUNDsbW65YqEK0mOAbeHZEXVrLXzxgyiYsG3s4w9+giNmH+sWe6zDyPUOqOt9FiICmzVCQ4BebWFYrLlZJeZ0s3j6IWA==",
        "X-MS-Exchange-Authentication-Results": "spf=pass (sender IP is 12.22.5.238)\n smtp.mailfrom=nvidia.com; dkim=none (message not signed)\n header.d=none;dmarc=pass action=none header.from=nvidia.com;",
        "Received-SPF": "Pass (protection.outlook.com: domain of nvidia.com designates\n 12.22.5.238 as permitted sender) receiver=protection.outlook.com;\n client-ip=12.22.5.238; helo=mail.nvidia.com; pr=C",
        "From": "Li Zhang <lizh@nvidia.com>",
        "To": "<orika@nvidia.com>, <viacheslavo@nvidia.com>, <matan@nvidia.com>,\n <shahafs@nvidia.com>",
        "CC": "<dev@dpdk.org>, <thomas@monjalon.net>, <rasland@nvidia.com>,\n <roniba@nvidia.com>",
        "Subject": "[PATCH 13/16] vdpa/mlx5: add virtq LM log task",
        "Date": "Mon, 6 Jun 2022 14:21:01 +0300",
        "Message-ID": "<20220606112109.208873-25-lizh@nvidia.com>",
        "X-Mailer": "git-send-email 2.31.1",
        "In-Reply-To": "<20220606112109.208873-1-lizh@nvidia.com>",
        "References": "<20220408075606.33056-1-lizh@nvidia.com>\n <20220606112109.208873-1-lizh@nvidia.com>",
        "MIME-Version": "1.0",
        "Content-Transfer-Encoding": "8bit",
        "Content-Type": "text/plain",
        "X-Originating-IP": "[10.126.231.35]",
        "X-ClientProxiedBy": "rnnvmail202.nvidia.com (10.129.68.7) To\n rnnvmail201.nvidia.com (10.129.68.8)",
        "X-EOPAttributedMessage": "0",
        "X-MS-PublicTrafficType": "Email",
        "X-MS-Office365-Filtering-Correlation-Id": "bd5507fa-eba6-4da0-2707-08da47aee25f",
        "X-MS-TrafficTypeDiagnostic": "CH2PR12MB5564:EE_",
        "X-LD-Processed": "43083d15-7273-40c1-b7db-39efd9ccc17a,ExtAddr",
        "X-Microsoft-Antispam-PRVS": "\n <CH2PR12MB55648B1C7EB57804C90656E1BFA29@CH2PR12MB5564.namprd12.prod.outlook.com>",
        "X-MS-Exchange-SenderADCheck": "1",
        "X-MS-Exchange-AntiSpam-Relay": "0",
        "X-Microsoft-Antispam": "BCL:0;",
        "X-Microsoft-Antispam-Message-Info": "\n X3C04lkMQ3Vq2KnSX/hL7x5nfepTLM+sdYHMRtKZwQvFOKprU4dGZ9svQWa/2SpQZkW7jT4PoitWAFC0jlkSApd+5PQe029IRNO4vOfuszm+jL7fXVHrAKd177ftxvj/kPQxrW4Zb/Ko1raq/DzWroLg73irU/mBxy3VSWf/OsWZU+CsC5Gfag42nJY2PTRcGNgVHTR4uxUC4qghPaN4HTVUU/nuf0pAFKCDdpble+ItrtJZP3KTxOE5Vg42LnnhIOxvk0b2Iv+WOjyojXgbq1XFm9FRArTIPaXwze1YVhUfQ4wS6m/auOjmeP57tcNdOvobz3DMMTxEYZv4RCoT5VHJnHHtG1LHPZdrlMvT0P/t1w91x2AOitt8/5ffGCngJsx6uStOKXA2+JF4iYNj35H0rSi6ffLZcT/KEl1+pX49TN89uI3Ms8VZeY8oi44KjnBu4m4fS8H2AzZONDnyIljRCK51R3gOJKGQ6VxK1nHKsUZheau4SYNQBG4ZgdKofbVBRH+lb1+SZMeoZw1Ze5M7OmX9swet+cE38obSWi4QVrlmMKz0R1kZxi0hUTS70cz7aWFOA6n/7gUUj6wP1XoF65c3GqmhanTVKXWetqVdobe/XErH634t/r8Io4c3AzUVoHohvzRVdjiFYKMMU1XknXeIpahAHPSgLeaG+u35kMyeyIRhbLtiNCHfG/8qc8ReKWmn19SU1zKn0TV7rw==",
        "X-Forefront-Antispam-Report": "CIP:12.22.5.238; CTRY:US; LANG:en; SCL:1; SRV:;\n IPV:NLI; SFV:NSPM; H:mail.nvidia.com; PTR:InfoNoRecords; CAT:NONE;\n SFS:(13230001)(4636009)(46966006)(36840700001)(40470700004)(40460700003)(336012)(82310400005)(70586007)(70206006)(16526019)(8936002)(186003)(5660300002)(508600001)(110136005)(4326008)(2616005)(8676002)(1076003)(107886003)(7696005)(54906003)(47076005)(426003)(81166007)(26005)(55016003)(356005)(2906002)(6636002)(6666004)(316002)(36756003)(36860700001)(83380400001)(86362001)(6286002)(36900700001);\n DIR:OUT; SFP:1101;",
        "X-OriginatorOrg": "Nvidia.com",
        "X-MS-Exchange-CrossTenant-OriginalArrivalTime": "06 Jun 2022 11:22:47.2708 (UTC)",
        "X-MS-Exchange-CrossTenant-Network-Message-Id": "\n bd5507fa-eba6-4da0-2707-08da47aee25f",
        "X-MS-Exchange-CrossTenant-Id": "43083d15-7273-40c1-b7db-39efd9ccc17a",
        "X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp": "\n TenantId=43083d15-7273-40c1-b7db-39efd9ccc17a; Ip=[12.22.5.238];\n Helo=[mail.nvidia.com]",
        "X-MS-Exchange-CrossTenant-AuthSource": "\n DM6NAM11FT060.eop-nam11.prod.protection.outlook.com",
        "X-MS-Exchange-CrossTenant-AuthAs": "Anonymous",
        "X-MS-Exchange-CrossTenant-FromEntityHeader": "HybridOnPrem",
        "X-MS-Exchange-Transport-CrossTenantHeadersStamped": "CH2PR12MB5564",
        "X-BeenThere": "dev@dpdk.org",
        "X-Mailman-Version": "2.1.29",
        "Precedence": "list",
        "List-Id": "DPDK patches and discussions <dev.dpdk.org>",
        "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n <mailto:dev-request@dpdk.org?subject=unsubscribe>",
        "List-Archive": "<http://mails.dpdk.org/archives/dev/>",
        "List-Post": "<mailto:dev@dpdk.org>",
        "List-Help": "<mailto:dev-request@dpdk.org?subject=help>",
        "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n <mailto:dev-request@dpdk.org?subject=subscribe>",
        "Errors-To": "dev-bounces@dpdk.org"
    },
    "content": "Split the virtqs LM log between the configuration threads.\nThis accelerates the LM process and reduces its time by 20%.\n\nSigned-off-by: Li Zhang <lizh@nvidia.com>\n---\n drivers/vdpa/mlx5/mlx5_vdpa.h         |  3 +\n drivers/vdpa/mlx5/mlx5_vdpa_cthread.c | 34 +++++++++++\n drivers/vdpa/mlx5/mlx5_vdpa_lm.c      | 85 +++++++++++++++++++++------\n 3 files changed, 105 insertions(+), 17 deletions(-)",
    "diff": "diff --git a/drivers/vdpa/mlx5/mlx5_vdpa.h b/drivers/vdpa/mlx5/mlx5_vdpa.h\nindex 35221f5ddc..e08931719f 100644\n--- a/drivers/vdpa/mlx5/mlx5_vdpa.h\n+++ b/drivers/vdpa/mlx5/mlx5_vdpa.h\n@@ -72,6 +72,8 @@ enum {\n \tMLX5_VDPA_NOTIFIER_STATE_ERR\n };\n \n+#define MLX5_VDPA_USED_RING_LEN(size) \\\n+\t((size) * sizeof(struct vring_used_elem) + sizeof(uint16_t) * 3)\n #define MLX5_VDPA_MAX_C_THRD 256\n #define MLX5_VDPA_MAX_TASKS_PER_THRD 4096\n #define MLX5_VDPA_TASKS_PER_DEV 64\n@@ -81,6 +83,7 @@ enum {\n enum mlx5_vdpa_task_type {\n \tMLX5_VDPA_TASK_REG_MR = 1,\n \tMLX5_VDPA_TASK_SETUP_VIRTQ,\n+\tMLX5_VDPA_TASK_STOP_VIRTQ,\n };\n \n /* Generic task information and size must be multiple of 4B. */\ndiff --git a/drivers/vdpa/mlx5/mlx5_vdpa_cthread.c b/drivers/vdpa/mlx5/mlx5_vdpa_cthread.c\nindex 1389d369ae..98369f0887 100644\n--- a/drivers/vdpa/mlx5/mlx5_vdpa_cthread.c\n+++ b/drivers/vdpa/mlx5/mlx5_vdpa_cthread.c\n@@ -104,6 +104,7 @@ mlx5_vdpa_c_thread_handle(void *arg)\n \tstruct mlx5_vdpa_priv *priv;\n \tstruct mlx5_vdpa_task task;\n \tstruct rte_ring *rng;\n+\tuint64_t features;\n \tuint32_t thrd_idx;\n \tuint32_t task_num;\n \tint ret;\n@@ -153,6 +154,39 @@ mlx5_vdpa_c_thread_handle(void *arg)\n \t\t\t}\n \t\t\tpthread_mutex_unlock(&virtq->virtq_lock);\n \t\t\tbreak;\n+\t\tcase MLX5_VDPA_TASK_STOP_VIRTQ:\n+\t\t\tvirtq = &priv->virtqs[task.idx];\n+\t\t\tpthread_mutex_lock(&virtq->virtq_lock);\n+\t\t\tret = mlx5_vdpa_virtq_stop(priv,\n+\t\t\t\t\ttask.idx);\n+\t\t\tif (ret) {\n+\t\t\t\tDRV_LOG(ERR,\n+\t\t\t\t\"Failed to stop virtq %d.\",\n+\t\t\t\ttask.idx);\n+\t\t\t\t__atomic_fetch_add(\n+\t\t\t\t\ttask.err_cnt, 1,\n+\t\t\t\t\t__ATOMIC_RELAXED);\n+\t\t\t\tpthread_mutex_unlock(&virtq->virtq_lock);\n+\t\t\t\tbreak;\n+\t\t\t}\n+\t\t\tret = rte_vhost_get_negotiated_features(\n+\t\t\t\tpriv->vid, &features);\n+\t\t\tif (ret) {\n+\t\t\t\tDRV_LOG(ERR,\n+\t\t\"Failed to get negotiated features virtq %d.\",\n+\t\t\t\ttask.idx);\n+\t\t\t\t__atomic_fetch_add(\n+\t\t\t\t\ttask.err_cnt, 1,\n+\t\t\t\t\t__ATOMIC_RELAXED);\n+\t\t\t\tpthread_mutex_unlock(&virtq->virtq_lock);\n+\t\t\t\tbreak;\n+\t\t\t}\n+\t\t\tif (RTE_VHOST_NEED_LOG(features))\n+\t\t\t\trte_vhost_log_used_vring(\n+\t\t\t\tpriv->vid, task.idx, 0,\n+\t\t\t    MLX5_VDPA_USED_RING_LEN(virtq->vq_size));\n+\t\t\tpthread_mutex_unlock(&virtq->virtq_lock);\n+\t\t\tbreak;\n \t\tdefault:\n \t\t\tDRV_LOG(ERR, \"Invalid vdpa task type %d.\",\n \t\t\ttask.type);\ndiff --git a/drivers/vdpa/mlx5/mlx5_vdpa_lm.c b/drivers/vdpa/mlx5/mlx5_vdpa_lm.c\nindex efebf364d0..c2e78218ca 100644\n--- a/drivers/vdpa/mlx5/mlx5_vdpa_lm.c\n+++ b/drivers/vdpa/mlx5/mlx5_vdpa_lm.c\n@@ -89,39 +89,90 @@ mlx5_vdpa_dirty_bitmap_set(struct mlx5_vdpa_priv *priv, uint64_t log_base,\n \treturn -1;\n }\n \n-#define MLX5_VDPA_USED_RING_LEN(size) \\\n-\t((size) * sizeof(struct vring_used_elem) + sizeof(uint16_t) * 3)\n-\n int\n mlx5_vdpa_lm_log(struct mlx5_vdpa_priv *priv)\n {\n+\tuint32_t remaining_cnt = 0, err_cnt = 0, task_num = 0;\n+\tuint32_t i, thrd_idx, data[1];\n \tstruct mlx5_vdpa_virtq *virtq;\n \tuint64_t features;\n-\tint ret = rte_vhost_get_negotiated_features(priv->vid, &features);\n-\tint i;\n+\tint ret;\n \n+\tret = rte_vhost_get_negotiated_features(priv->vid, &features);\n \tif (ret) {\n \t\tDRV_LOG(ERR, \"Failed to get negotiated features.\");\n \t\treturn -1;\n \t}\n-\tif (!RTE_VHOST_NEED_LOG(features))\n-\t\treturn 0;\n-\tfor (i = 0; i < priv->nr_virtqs; ++i) {\n-\t\tvirtq = &priv->virtqs[i];\n-\t\tif (!priv->virtqs[i].virtq) {\n-\t\t\tDRV_LOG(DEBUG, \"virtq %d is invalid for LM log.\", i);\n-\t\t} else {\n+\tif (priv->use_c_thread && priv->nr_virtqs) {\n+\t\tuint32_t main_task_idx[priv->nr_virtqs];\n+\n+\t\tfor (i = 0; i < priv->nr_virtqs; i++) {\n+\t\t\tvirtq = &priv->virtqs[i];\n+\t\t\tif (!virtq->configured)\n+\t\t\t\tcontinue;\n+\t\t\tthrd_idx = i % (conf_thread_mng.max_thrds + 1);\n+\t\t\tif (!thrd_idx) {\n+\t\t\t\tmain_task_idx[task_num] = i;\n+\t\t\t\ttask_num++;\n+\t\t\t\tcontinue;\n+\t\t\t}\n+\t\t\tthrd_idx = priv->last_c_thrd_idx + 1;\n+\t\t\tif (thrd_idx >= conf_thread_mng.max_thrds)\n+\t\t\t\tthrd_idx = 0;\n+\t\t\tpriv->last_c_thrd_idx = thrd_idx;\n+\t\t\tdata[0] = i;\n+\t\t\tif (mlx5_vdpa_task_add(priv, thrd_idx,\n+\t\t\t\tMLX5_VDPA_TASK_STOP_VIRTQ,\n+\t\t\t\t&remaining_cnt, &err_cnt,\n+\t\t\t\t(void **)&data, 1)) {\n+\t\t\t\tDRV_LOG(ERR, \"Fail to add \"\n+\t\t\t\t\t\"task stop virtq (%d).\", i);\n+\t\t\t\tmain_task_idx[task_num] = i;\n+\t\t\t\ttask_num++;\n+\t\t\t}\n+\t\t}\n+\t\tfor (i = 0; i < task_num; i++) {\n+\t\t\tvirtq = &priv->virtqs[main_task_idx[i]];\n \t\t\tpthread_mutex_lock(&virtq->virtq_lock);\n-\t\t\tret = mlx5_vdpa_virtq_stop(priv, i);\n+\t\t\tret = mlx5_vdpa_virtq_stop(priv,\n+\t\t\t\t\tmain_task_idx[i]);\n+\t\t\tif (ret) {\n+\t\t\t\tpthread_mutex_unlock(&virtq->virtq_lock);\n+\t\t\t\tDRV_LOG(ERR,\n+\t\t\t\t\"Failed to stop virtq %d.\", i);\n+\t\t\t\treturn -1;\n+\t\t\t}\n+\t\t\tif (RTE_VHOST_NEED_LOG(features))\n+\t\t\t\trte_vhost_log_used_vring(priv->vid, i, 0,\n+\t\t\t\tMLX5_VDPA_USED_RING_LEN(virtq->vq_size));\n \t\t\tpthread_mutex_unlock(&virtq->virtq_lock);\n+\t\t}\n+\t\tif (mlx5_vdpa_c_thread_wait_bulk_tasks_done(&remaining_cnt,\n+\t\t\t&err_cnt, 2000)) {\n+\t\t\tDRV_LOG(ERR,\n+\t\t\t\"Failed to wait virt-queue setup tasks ready.\");\n+\t\t\treturn -1;\n+\t\t}\n+\t} else {\n+\t\tfor (i = 0; i < priv->nr_virtqs; i++) {\n+\t\t\tvirtq = &priv->virtqs[i];\n+\t\t\tpthread_mutex_lock(&virtq->virtq_lock);\n+\t\t\tif (!virtq->configured) {\n+\t\t\t\tpthread_mutex_unlock(&virtq->virtq_lock);\n+\t\t\t\tcontinue;\n+\t\t\t}\n+\t\t\tret = mlx5_vdpa_virtq_stop(priv, i);\n \t\t\tif (ret) {\n-\t\t\t\tDRV_LOG(ERR, \"Failed to stop virtq %d for LM \"\n-\t\t\t\t\t\"log.\", i);\n+\t\t\t\tpthread_mutex_unlock(&virtq->virtq_lock);\n+\t\t\t\tDRV_LOG(ERR,\n+\t\t\t\t\"Failed to stop virtq %d for LM log.\", i);\n \t\t\t\treturn -1;\n \t\t\t}\n+\t\t\tif (RTE_VHOST_NEED_LOG(features))\n+\t\t\t\trte_vhost_log_used_vring(priv->vid, i, 0,\n+\t\t\t\tMLX5_VDPA_USED_RING_LEN(virtq->vq_size));\n+\t\t\tpthread_mutex_unlock(&virtq->virtq_lock);\n \t\t}\n-\t\trte_vhost_log_used_vring(priv->vid, i, 0,\n-\t\t\t      MLX5_VDPA_USED_RING_LEN(priv->virtqs[i].vq_size));\n \t}\n \treturn 0;\n }\n",
    "prefixes": [
        "13/16"
    ]
}