Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/138883/?format=api
http://patches.dpdk.org/api/patches/138883/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/patch/1711579078-10624-8-git-send-email-roretzla@linux.microsoft.com/", "project": { "id": 1, "url": "http://patches.dpdk.org/api/projects/1/?format=api", "name": "DPDK", "link_name": "dpdk", "list_id": "dev.dpdk.org", "list_email": "dev@dpdk.org", "web_url": "http://core.dpdk.org", "scm_url": "git://dpdk.org/dpdk", "webscm_url": "http://git.dpdk.org/dpdk", "list_archive_url": "https://inbox.dpdk.org/dev", "list_archive_url_format": "https://inbox.dpdk.org/dev/{}", "commit_url_format": "" }, "msgid": "<1711579078-10624-8-git-send-email-roretzla@linux.microsoft.com>", "list_archive_url": "https://inbox.dpdk.org/dev/1711579078-10624-8-git-send-email-roretzla@linux.microsoft.com", "date": "2024-03-27T22:37:20", "name": "[v3,07/45] net/bnxt: use rte stdatomic API", "commit_ref": null, "pull_url": null, "state": "superseded", "archived": false, "hash": "180c64cda9d7a660a3945cee05922a23dce6e884", "submitter": { "id": 2077, "url": "http://patches.dpdk.org/api/people/2077/?format=api", "name": "Tyler Retzlaff", "email": "roretzla@linux.microsoft.com" }, "delegate": { "id": 1, "url": "http://patches.dpdk.org/api/users/1/?format=api", "username": "tmonjalo", "first_name": "Thomas", "last_name": "Monjalon", "email": "thomas@monjalon.net" }, "mbox": "http://patches.dpdk.org/project/dpdk/patch/1711579078-10624-8-git-send-email-roretzla@linux.microsoft.com/mbox/", "series": [ { "id": 31633, "url": "http://patches.dpdk.org/api/series/31633/?format=api", "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=31633", "date": "2024-03-27T22:37:13", "name": "use stdatomic API", "version": 3, "mbox": "http://patches.dpdk.org/series/31633/mbox/" } ], "comments": "http://patches.dpdk.org/api/patches/138883/comments/", "check": "warning", "checks": "http://patches.dpdk.org/api/patches/138883/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "<dev-bounces@dpdk.org>", "X-Original-To": "patchwork@inbox.dpdk.org", "Delivered-To": "patchwork@inbox.dpdk.org", "Received": [ "from mails.dpdk.org (mails.dpdk.org [217.70.189.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id 17C9843D55;\n\tWed, 27 Mar 2024 23:38:50 +0100 (CET)", "from mails.dpdk.org (localhost [127.0.0.1])\n\tby mails.dpdk.org (Postfix) with ESMTP id 1A7D442D6B;\n\tWed, 27 Mar 2024 23:38:12 +0100 (CET)", "from linux.microsoft.com (linux.microsoft.com [13.77.154.182])\n by mails.dpdk.org (Postfix) with ESMTP id D2EF5402B2\n for <dev@dpdk.org>; Wed, 27 Mar 2024 23:38:01 +0100 (CET)", "by linux.microsoft.com (Postfix, from userid 1086)\n id 49F3A20E6951; Wed, 27 Mar 2024 15:37:59 -0700 (PDT)" ], "DKIM-Filter": "OpenDKIM Filter v2.11.0 linux.microsoft.com 49F3A20E6951", "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.microsoft.com;\n s=default; t=1711579080;\n bh=HvBk5DLutzFAhS96eefNxwDSx1g207PTA5iAcwgJy+E=;\n h=From:To:Cc:Subject:Date:In-Reply-To:References:From;\n b=lXqw8b2dsRuSwMvPlnq+H1Kv1Y0wj3FilWuKNDH3KNDis724WpuaXfPvdYbtPfl1g\n BG6VVM8eT2K6J8vxa77BsgPMrdWJ0nw6uH0Jj9eSB1L9wS11GWmLyXNWc/fRdaGdVM\n SrfigmfbzP4yRRaU5Fl47vLsjQeuvGqphsPC5zJk=", "From": "Tyler Retzlaff <roretzla@linux.microsoft.com>", "To": "dev@dpdk.org", "Cc": "=?utf-8?q?Mattias_R=C3=B6nnblom?= <mattias.ronnblom@ericsson.com>,\n\t=?utf-8?q?Morten_Br=C3=B8rup?= <mb@smartsharesystems.com>,\n Abdullah Sevincer <abdullah.sevincer@intel.com>,\n Ajit Khaparde <ajit.khaparde@broadcom.com>, Alok Prasad <palok@marvell.com>,\n Anatoly Burakov <anatoly.burakov@intel.com>,\n Andrew Rybchenko <andrew.rybchenko@oktetlabs.ru>,\n Anoob Joseph <anoobj@marvell.com>,\n Bruce Richardson <bruce.richardson@intel.com>,\n Byron Marohn <byron.marohn@intel.com>, Chenbo Xia <chenbox@nvidia.com>,\n Chengwen Feng <fengchengwen@huawei.com>,\n Ciara Loftus <ciara.loftus@intel.com>, Ciara Power <ciara.power@intel.com>,\n Dariusz Sosnowski <dsosnowski@nvidia.com>, David Hunt <david.hunt@intel.com>,\n Devendra Singh Rawat <dsinghrawat@marvell.com>,\n Erik Gabriel Carrillo <erik.g.carrillo@intel.com>,\n Guoyang Zhou <zhouguoyang@huawei.com>, Harman Kalra <hkalra@marvell.com>,\n Harry van Haaren <harry.van.haaren@intel.com>,\n Honnappa Nagarahalli <honnappa.nagarahalli@arm.com>,\n Jakub Grajciar <jgrajcia@cisco.com>, Jerin Jacob <jerinj@marvell.com>,\n Jeroen de Borst <jeroendb@google.com>, Jian Wang <jianwang@trustnetic.com>,\n Jiawen Wu <jiawenwu@trustnetic.com>, Jie Hai <haijie1@huawei.com>,\n Jingjing Wu <jingjing.wu@intel.com>, Joshua Washington <joshwash@google.com>,\n Joyce Kong <joyce.kong@arm.com>, Junfeng Guo <junfeng.guo@intel.com>,\n Kevin Laatz <kevin.laatz@intel.com>,\n Konstantin Ananyev <konstantin.v.ananyev@yandex.ru>,\n Liang Ma <liangma@liangbit.com>, Long Li <longli@microsoft.com>,\n Maciej Czekaj <mczekaj@marvell.com>, Matan Azrad <matan@nvidia.com>,\n Maxime Coquelin <maxime.coquelin@redhat.com>,\n Nicolas Chautru <nicolas.chautru@intel.com>, Ori Kam <orika@nvidia.com>,\n Pavan Nikhilesh <pbhagavatula@marvell.com>,\n Peter Mccarthy <peter.mccarthy@intel.com>,\n Rahul Lakkireddy <rahul.lakkireddy@chelsio.com>,\n Reshma Pattan <reshma.pattan@intel.com>, Rosen Xu <rosen.xu@intel.com>,\n Ruifeng Wang <ruifeng.wang@arm.com>, Rushil Gupta <rushilg@google.com>,\n Sameh Gobriel <sameh.gobriel@intel.com>,\n Sivaprasad Tummala <sivaprasad.tummala@amd.com>,\n Somnath Kotur <somnath.kotur@broadcom.com>,\n Stephen Hemminger <stephen@networkplumber.org>,\n Suanming Mou <suanmingm@nvidia.com>, Sunil Kumar Kori <skori@marvell.com>,\n Sunil Uttarwar <sunilprakashrao.uttarwar@amd.com>,\n Tetsuya Mukawa <mtetsuyah@gmail.com>, Vamsi Attunuru <vattunuru@marvell.com>,\n Viacheslav Ovsiienko <viacheslavo@nvidia.com>,\n Vladimir Medvedkin <vladimir.medvedkin@intel.com>,\n Xiaoyun Wang <cloud.wangxiaoyun@huawei.com>,\n Yipeng Wang <yipeng1.wang@intel.com>, Yisen Zhuang <yisen.zhuang@huawei.com>,\n Yuying Zhang <Yuying.Zhang@intel.com>, Yuying Zhang <yuying.zhang@intel.com>,\n Ziyang Xuan <xuanziyang2@huawei.com>,\n Tyler Retzlaff <roretzla@linux.microsoft.com>", "Subject": "[PATCH v3 07/45] net/bnxt: use rte stdatomic API", "Date": "Wed, 27 Mar 2024 15:37:20 -0700", "Message-Id": "<1711579078-10624-8-git-send-email-roretzla@linux.microsoft.com>", "X-Mailer": "git-send-email 1.8.3.1", "In-Reply-To": "<1711579078-10624-1-git-send-email-roretzla@linux.microsoft.com>", "References": "<1710967892-7046-1-git-send-email-roretzla@linux.microsoft.com>\n <1711579078-10624-1-git-send-email-roretzla@linux.microsoft.com>", "X-BeenThere": "dev@dpdk.org", "X-Mailman-Version": "2.1.29", "Precedence": "list", "List-Id": "DPDK patches and discussions <dev.dpdk.org>", "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n <mailto:dev-request@dpdk.org?subject=unsubscribe>", "List-Archive": "<http://mails.dpdk.org/archives/dev/>", "List-Post": "<mailto:dev@dpdk.org>", "List-Help": "<mailto:dev-request@dpdk.org?subject=help>", "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n <mailto:dev-request@dpdk.org?subject=subscribe>", "Errors-To": "dev-bounces@dpdk.org" }, "content": "Replace the use of gcc builtin __atomic_xxx intrinsics with\ncorresponding rte_atomic_xxx optional rte stdatomic API.\n\nSigned-off-by: Tyler Retzlaff <roretzla@linux.microsoft.com>\nAcked-by: Stephen Hemminger <stephen@networkplumber.org>\n---\n drivers/net/bnxt/bnxt_cpr.h | 4 ++--\n drivers/net/bnxt/bnxt_rxq.h | 2 +-\n drivers/net/bnxt/bnxt_rxr.c | 13 ++++++++-----\n drivers/net/bnxt/bnxt_rxtx_vec_neon.c | 2 +-\n drivers/net/bnxt/bnxt_stats.c | 4 ++--\n 5 files changed, 14 insertions(+), 11 deletions(-)", "diff": "diff --git a/drivers/net/bnxt/bnxt_cpr.h b/drivers/net/bnxt/bnxt_cpr.h\nindex c7b3480..43f06fd 100644\n--- a/drivers/net/bnxt/bnxt_cpr.h\n+++ b/drivers/net/bnxt/bnxt_cpr.h\n@@ -107,7 +107,7 @@ struct bnxt_cp_ring_info {\n \n /**\n * Check validity of a completion ring entry. If the entry is valid, include a\n- * C11 __ATOMIC_ACQUIRE fence to ensure that subsequent loads of fields in the\n+ * C11 rte_memory_order_acquire fence to ensure that subsequent loads of fields in the\n * completion are not hoisted by the compiler or by the CPU to come before the\n * loading of the \"valid\" field.\n *\n@@ -130,7 +130,7 @@ struct bnxt_cp_ring_info {\n \texpected = !(raw_cons & ring_size);\n \tvalid = !!(rte_le_to_cpu_32(c->info3_v) & CMPL_BASE_V);\n \tif (valid == expected) {\n-\t\trte_atomic_thread_fence(__ATOMIC_ACQUIRE);\n+\t\trte_atomic_thread_fence(rte_memory_order_acquire);\n \t\treturn true;\n \t}\n \treturn false;\ndiff --git a/drivers/net/bnxt/bnxt_rxq.h b/drivers/net/bnxt/bnxt_rxq.h\nindex 77bc382..36e0ac3 100644\n--- a/drivers/net/bnxt/bnxt_rxq.h\n+++ b/drivers/net/bnxt/bnxt_rxq.h\n@@ -40,7 +40,7 @@ struct bnxt_rx_queue {\n \tstruct bnxt_rx_ring_info\t*rx_ring;\n \tstruct bnxt_cp_ring_info\t*cp_ring;\n \tstruct rte_mbuf\t\t\tfake_mbuf;\n-\tuint64_t\t\t\trx_mbuf_alloc_fail;\n+\tRTE_ATOMIC(uint64_t)\t\trx_mbuf_alloc_fail;\n \tuint8_t\t\t\t\tneed_realloc;\n \tconst struct rte_memzone *mz;\n };\ndiff --git a/drivers/net/bnxt/bnxt_rxr.c b/drivers/net/bnxt/bnxt_rxr.c\nindex 3542975..ca5d2c6 100644\n--- a/drivers/net/bnxt/bnxt_rxr.c\n+++ b/drivers/net/bnxt/bnxt_rxr.c\n@@ -49,7 +49,8 @@ static inline int bnxt_alloc_rx_data(struct bnxt_rx_queue *rxq,\n \trx_buf = &rxr->rx_buf_ring[prod];\n \tmbuf = __bnxt_alloc_rx_data(rxq->mb_pool);\n \tif (!mbuf) {\n-\t\t__atomic_fetch_add(&rxq->rx_mbuf_alloc_fail, 1, __ATOMIC_RELAXED);\n+\t\trte_atomic_fetch_add_explicit(&rxq->rx_mbuf_alloc_fail, 1,\n+\t\t rte_memory_order_relaxed);\n \t\t/* If buff has failed already, setting this again won't hurt */\n \t\trxq->need_realloc = 1;\n \t\treturn -ENOMEM;\n@@ -86,7 +87,8 @@ static inline int bnxt_alloc_ag_data(struct bnxt_rx_queue *rxq,\n \n \tmbuf = __bnxt_alloc_rx_data(rxq->mb_pool);\n \tif (!mbuf) {\n-\t\t__atomic_fetch_add(&rxq->rx_mbuf_alloc_fail, 1, __ATOMIC_RELAXED);\n+\t\trte_atomic_fetch_add_explicit(&rxq->rx_mbuf_alloc_fail, 1,\n+\t\t rte_memory_order_relaxed);\n \t\t/* If buff has failed already, setting this again won't hurt */\n \t\trxq->need_realloc = 1;\n \t\treturn -ENOMEM;\n@@ -465,7 +467,8 @@ static inline struct rte_mbuf *bnxt_tpa_end(\n \tstruct rte_mbuf *new_data = __bnxt_alloc_rx_data(rxq->mb_pool);\n \tRTE_ASSERT(new_data != NULL);\n \tif (!new_data) {\n-\t\t__atomic_fetch_add(&rxq->rx_mbuf_alloc_fail, 1, __ATOMIC_RELAXED);\n+\t\trte_atomic_fetch_add_explicit(&rxq->rx_mbuf_alloc_fail, 1,\n+\t\t rte_memory_order_relaxed);\n \t\treturn NULL;\n \t}\n \ttpa_info->mbuf = new_data;\n@@ -1677,8 +1680,8 @@ int bnxt_init_one_rx_ring(struct bnxt_rx_queue *rxq)\n \t\t\t\trxr->tpa_info[i].mbuf =\n \t\t\t\t\t__bnxt_alloc_rx_data(rxq->mb_pool);\n \t\t\t\tif (!rxr->tpa_info[i].mbuf) {\n-\t\t\t\t\t__atomic_fetch_add(&rxq->rx_mbuf_alloc_fail, 1,\n-\t\t\t\t\t\t\t__ATOMIC_RELAXED);\n+\t\t\t\t\trte_atomic_fetch_add_explicit(&rxq->rx_mbuf_alloc_fail, 1,\n+\t\t\t\t\t\t\trte_memory_order_relaxed);\n \t\t\t\t\treturn -ENOMEM;\n \t\t\t\t}\n \t\t\t}\ndiff --git a/drivers/net/bnxt/bnxt_rxtx_vec_neon.c b/drivers/net/bnxt/bnxt_rxtx_vec_neon.c\nindex 775400f..04864e0 100644\n--- a/drivers/net/bnxt/bnxt_rxtx_vec_neon.c\n+++ b/drivers/net/bnxt/bnxt_rxtx_vec_neon.c\n@@ -240,7 +240,7 @@\n \t\trxcmp1[0] = vld1q_u32((void *)&cpr->cp_desc_ring[cons + 1]);\n \n \t\t/* Use acquire fence to order loads of descriptor words. */\n-\t\trte_atomic_thread_fence(__ATOMIC_ACQUIRE);\n+\t\trte_atomic_thread_fence(rte_memory_order_acquire);\n \t\t/* Reload lower 64b of descriptors to make it ordered after info3_v. */\n \t\trxcmp1[3] = vreinterpretq_u32_u64(vld1q_lane_u64\n \t\t\t\t((void *)&cpr->cp_desc_ring[cons + 7],\ndiff --git a/drivers/net/bnxt/bnxt_stats.c b/drivers/net/bnxt/bnxt_stats.c\nindex 6a6feab..479f819 100644\n--- a/drivers/net/bnxt/bnxt_stats.c\n+++ b/drivers/net/bnxt/bnxt_stats.c\n@@ -663,7 +663,7 @@ static int bnxt_stats_get_ext(struct rte_eth_dev *eth_dev,\n \n \t\tbnxt_fill_rte_eth_stats_ext(bnxt_stats, &ring_stats, i, true);\n \t\tbnxt_stats->rx_nombuf +=\n-\t\t\t\t__atomic_load_n(&rxq->rx_mbuf_alloc_fail, __ATOMIC_RELAXED);\n+\t\t rte_atomic_load_explicit(&rxq->rx_mbuf_alloc_fail, rte_memory_order_relaxed);\n \t}\n \n \tnum_q_stats = RTE_MIN(bp->tx_cp_nr_rings,\n@@ -724,7 +724,7 @@ int bnxt_stats_get_op(struct rte_eth_dev *eth_dev,\n \n \t\tbnxt_fill_rte_eth_stats(bnxt_stats, &ring_stats, i, true);\n \t\tbnxt_stats->rx_nombuf +=\n-\t\t\t\t__atomic_load_n(&rxq->rx_mbuf_alloc_fail, __ATOMIC_RELAXED);\n+\t\t rte_atomic_load_explicit(&rxq->rx_mbuf_alloc_fail, rte_memory_order_relaxed);\n \t}\n \n \tnum_q_stats = RTE_MIN(bp->tx_cp_nr_rings,\n", "prefixes": [ "v3", "07/45" ] }{ "id": 138883, "url": "