get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/80584/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 80584,
    "url": "http://patches.dpdk.org/api/patches/80584/?format=api",
    "web_url": "http://patches.dpdk.org/project/dpdk/patch/20201013152610.2589-9-stephen@networkplumber.org/",
    "project": {
        "id": 1,
        "url": "http://patches.dpdk.org/api/projects/1/?format=api",
        "name": "DPDK",
        "link_name": "dpdk",
        "list_id": "dev.dpdk.org",
        "list_email": "dev@dpdk.org",
        "web_url": "http://core.dpdk.org",
        "scm_url": "git://dpdk.org/dpdk",
        "webscm_url": "http://git.dpdk.org/dpdk",
        "list_archive_url": "https://inbox.dpdk.org/dev",
        "list_archive_url_format": "https://inbox.dpdk.org/dev/{}",
        "commit_url_format": ""
    },
    "msgid": "<20201013152610.2589-9-stephen@networkplumber.org>",
    "list_archive_url": "https://inbox.dpdk.org/dev/20201013152610.2589-9-stephen@networkplumber.org",
    "date": "2020-10-13T15:26:00",
    "name": "[v5,08/18] app/test: replace uses of master/slave",
    "commit_ref": null,
    "pull_url": null,
    "state": "superseded",
    "archived": true,
    "hash": "6bf8bea99e0ee12ac8c9a784c067e462f79e8fe5",
    "submitter": {
        "id": 27,
        "url": "http://patches.dpdk.org/api/people/27/?format=api",
        "name": "Stephen Hemminger",
        "email": "stephen@networkplumber.org"
    },
    "delegate": {
        "id": 1,
        "url": "http://patches.dpdk.org/api/users/1/?format=api",
        "username": "tmonjalo",
        "first_name": "Thomas",
        "last_name": "Monjalon",
        "email": "thomas@monjalon.net"
    },
    "mbox": "http://patches.dpdk.org/project/dpdk/patch/20201013152610.2589-9-stephen@networkplumber.org/mbox/",
    "series": [
        {
            "id": 12924,
            "url": "http://patches.dpdk.org/api/series/12924/?format=api",
            "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=12924",
            "date": "2020-10-13T15:25:52",
            "name": "Replace terms master/slave",
            "version": 5,
            "mbox": "http://patches.dpdk.org/series/12924/mbox/"
        }
    ],
    "comments": "http://patches.dpdk.org/api/patches/80584/comments/",
    "check": "warning",
    "checks": "http://patches.dpdk.org/api/patches/80584/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<dev-bounces@dpdk.org>",
        "X-Original-To": "patchwork@inbox.dpdk.org",
        "Delivered-To": "patchwork@inbox.dpdk.org",
        "Received": [
            "from dpdk.org (dpdk.org [92.243.14.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id D7254A04B7;\n\tTue, 13 Oct 2020 17:29:54 +0200 (CEST)",
            "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id B80BF1DD40;\n\tTue, 13 Oct 2020 17:27:14 +0200 (CEST)",
            "from mail-pj1-f52.google.com (mail-pj1-f52.google.com\n [209.85.216.52]) by dpdk.org (Postfix) with ESMTP id 4ECE71DCEA\n for <dev@dpdk.org>; Tue, 13 Oct 2020 17:27:12 +0200 (CEST)",
            "by mail-pj1-f52.google.com with SMTP id h4so51849pjk.0\n for <dev@dpdk.org>; Tue, 13 Oct 2020 08:27:12 -0700 (PDT)",
            "from hermes.local (204-195-22-127.wavecable.com. [204.195.22.127])\n by smtp.gmail.com with ESMTPSA id z12sm23815233pfr.197.2020.10.13.08.27.02\n (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256);\n Tue, 13 Oct 2020 08:27:03 -0700 (PDT)"
        ],
        "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed;\n d=networkplumber-org.20150623.gappssmtp.com; s=20150623;\n h=from:to:cc:subject:date:message-id:in-reply-to:references\n :mime-version:content-transfer-encoding;\n bh=CcG3tawCggmhnqz4yaVLn/03wW41aYXT7XcjGMTiBU8=;\n b=z8gUmhot1nXBnIJ94w0kXp1AQC+u7iKrC8dF2BgPo0Uf4zekDncN/p/r0sXUVYRXzx\n FMNt93LSwKSBDOofacTiEe2CLpAWZecazJ1fywz3b2GPUz0VzxsYDpn0tU0G8JJeZDrH\n FWfm2QLU+6y78HGJO32yu+SHFPtVFwYwVeW5UOkQmjwXnFTpg/azo7LSyDING3bGX+ce\n +kqHan+XeapSliAOB/jZ65WleItta8OfLTsMFDj7NxtsiO/GGDkG2Ptv0Pqwp4tWDff1\n qKxwmHv2rSkLlAmeMJPGaD3TP3BtjwQ5uZhhFz5mAleX8vi4otFPMRVQznEI1yYF2HpK\n Hq7w==",
        "X-Google-DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed;\n d=1e100.net; s=20161025;\n h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to\n :references:mime-version:content-transfer-encoding;\n bh=CcG3tawCggmhnqz4yaVLn/03wW41aYXT7XcjGMTiBU8=;\n b=W8FlpqC3kIEvYEOoy6xcp7Ld3rQIWX8fiYiB0NPvbwsYnFvB0xfSratNrW+kb2w4t6\n oTcvVEkB8x3j7i1aMXw0WSf/X9S6eBzx4NXeATaImzSlwzspyhDsLuGx49QuDCH5c5aw\n xEyxe9ini+b5SV6JdkYHYTUE1yP1UWQGKaKt8lvvJgQFo9UR8he86WyjabjkJh60Wzoy\n 59soNNW4z1DrmmaL0jUePAaFPciY4Lhi+VhJ8ht8PH9f5q2OuhtELovkU1I/snz2cVOR\n MxLxiHyjQNMDHcjnQIlgl/L0kMbdvASVULKKg7cOltJpbAG4I1PRxk2jKKGdvIJnOWJy\n Oniw==",
        "X-Gm-Message-State": "AOAM530zhs5y8dx7bsZHS61sFXryg+6WnX8fI5LRYmlMT+8O9eLnnXgT\n KvJu10FksnL/MSSe0737Bxr9cLjm/O4+na/1",
        "X-Google-Smtp-Source": "\n ABdhPJwzXwS0/Fra425b1m3RyliyxzUMfZe+snbMSAAO5AJBEj06s6AVHMxwGln5CWBBEYNJ3G4V8w==",
        "X-Received": "by 2002:a17:902:c408:b029:d4:c99a:19e8 with SMTP id\n k8-20020a170902c408b02900d4c99a19e8mr92180plk.8.1602602827770;\n Tue, 13 Oct 2020 08:27:07 -0700 (PDT)",
        "From": "Stephen Hemminger <stephen@networkplumber.org>",
        "To": "dev@dpdk.org",
        "Cc": "Stephen Hemminger <stephen@networkplumber.org>,\n Anatoly Burakov <anatoly.burakov@intel.com>,\n Fiona Trahe <fiona.trahe@intel.com>,\n Ashish Gupta <ashish.gupta@marvell.com>,\n Declan Doherty <declan.doherty@intel.com>,\n David Hunt <david.hunt@intel.com>, Byron Marohn <byron.marohn@intel.com>,\n Yipeng Wang <yipeng1.wang@intel.com>,\n Olivier Matz <olivier.matz@6wind.com>,\n Andrew Rybchenko <andrew.rybchenko@oktetlabs.ru>,\n Bruce Richardson <bruce.richardson@intel.com>,\n Vladimir Medvedkin <vladimir.medvedkin@intel.com>,\n Honnappa Nagarahalli <honnappa.nagarahalli@arm.com>,\n Konstantin Ananyev <konstantin.ananyev@intel.com>,\n Sameh Gobriel <sameh.gobriel@intel.com>,\n Ferruh Yigit <ferruh.yigit@intel.com>,\n Reshma Pattan <reshma.pattan@intel.com>,\n Harry van Haaren <harry.van.haaren@intel.com>,\n Gage Eads <gage.eads@intel.com>, Joyce Kong <joyce.kong@arm.com>,\n Robert Sanford <rsanford@akamai.com>,\n Erik Gabriel Carrillo <erik.g.carrillo@intel.com>,\n Jerin Jacob <jerinj@marvell.com>, Sunil Kumar Kori <skori@marvell.com>",
        "Date": "Tue, 13 Oct 2020 08:26:00 -0700",
        "Message-Id": "<20201013152610.2589-9-stephen@networkplumber.org>",
        "X-Mailer": "git-send-email 2.27.0",
        "In-Reply-To": "<20201013152610.2589-1-stephen@networkplumber.org>",
        "References": "<20200914182002.6750-1-stephen@networkplumber.org>\n <20201013152610.2589-1-stephen@networkplumber.org>",
        "MIME-Version": "1.0",
        "Content-Transfer-Encoding": "8bit",
        "Subject": "[dpdk-dev] [PATCH v5 08/18] app/test: replace uses of master/slave",
        "X-BeenThere": "dev@dpdk.org",
        "X-Mailman-Version": "2.1.15",
        "Precedence": "list",
        "List-Id": "DPDK patches and discussions <dev.dpdk.org>",
        "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n <mailto:dev-request@dpdk.org?subject=unsubscribe>",
        "List-Archive": "<http://mails.dpdk.org/archives/dev/>",
        "List-Post": "<mailto:dev@dpdk.org>",
        "List-Help": "<mailto:dev-request@dpdk.org?subject=help>",
        "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n <mailto:dev-request@dpdk.org?subject=subscribe>",
        "Errors-To": "dev-bounces@dpdk.org",
        "Sender": "\"dev\" <dev-bounces@dpdk.org>"
    },
    "content": "Replace master lcore with main lcore and slave lcore with worker lcore.\nMostly automatic replacement.\n\nAcked-by: Anatoly Burakov <anatoly.burakov@intel.com>\nSigned-off-by: Stephen Hemminger <stephen@networkplumber.org>\n---\n app/test/autotest_test_funcs.py         |   2 +-\n app/test/meson.build                    |   2 +-\n app/test/test.c                         |   2 +-\n app/test/test_atomic.c                  |  26 +++---\n app/test/test_barrier.c                 |   2 +-\n app/test/test_compressdev_test_buffer.h |   2 +-\n app/test/test_cryptodev.c               |  16 ++--\n app/test/test_distributor.c             |   8 +-\n app/test/test_distributor_perf.c        |  10 +--\n app/test/test_eal_flags.c               |  32 +++----\n app/test/test_efd.c                     |   2 +-\n app/test/test_efd_perf.c                |   2 +-\n app/test/test_func_reentrancy.c         |  20 ++---\n app/test/test_hash_multiwriter.c        |   4 +-\n app/test/test_hash_readwrite.c          |  39 +++++----\n app/test/test_kni.c                     |  16 ++--\n app/test/test_lpm_perf.c                |   4 +-\n app/test/test_malloc.c                  |  12 +--\n app/test/test_mbuf.c                    |  41 +++++----\n app/test/test_mcslock.c                 |  28 +++---\n app/test/test_mempool_perf.c            |  10 +--\n app/test/test_mp_secondary.c            |   2 +-\n app/test/test_pdump.c                   |   2 +-\n app/test/test_per_lcore.c               |  14 +--\n app/test/test_pmd_perf.c                |  20 ++---\n app/test/test_rcu_qsbr.c                |   2 +-\n app/test/test_rcu_qsbr_perf.c           |   2 +-\n app/test/test_ring_perf.c               |  14 +--\n app/test/test_ring_stress_impl.h        |  10 +--\n app/test/test_rwlock.c                  |  28 +++---\n app/test/test_service_cores.c           |  14 +--\n app/test/test_spinlock.c                |  34 ++++----\n app/test/test_stack.c                   |   2 +-\n app/test/test_stack_perf.c              |   6 +-\n app/test/test_ticketlock.c              |  36 ++++----\n app/test/test_timer.c                   | 109 ++++++++++++------------\n app/test/test_timer_racecond.c          |  28 +++---\n app/test/test_timer_secondary.c         |   8 +-\n app/test/test_trace_perf.c              |   4 +-\n 39 files changed, 307 insertions(+), 308 deletions(-)",
    "diff": "diff --git a/app/test/autotest_test_funcs.py b/app/test/autotest_test_funcs.py\nindex 775dfd1dc5d9..0811066cb054 100644\n--- a/app/test/autotest_test_funcs.py\n+++ b/app/test/autotest_test_funcs.py\n@@ -103,7 +103,7 @@ def rwlock_autotest(child, test_name):\n         index = child.expect([\"Test OK\",\n                               \"Test Failed\",\n                               \"Hello from core ([0-9]*) !\",\n-                              \"Global write lock taken on master \"\n+                              \"Global write lock taken on main \"\n                               \"core ([0-9]*)\",\n                               pexpect.TIMEOUT], timeout=10)\n         # ok\ndiff --git a/app/test/meson.build b/app/test/meson.build\nindex dedf29dd7f26..6e07302d8fdc 100644\n--- a/app/test/meson.build\n+++ b/app/test/meson.build\n@@ -189,7 +189,7 @@ fast_tests = [\n         ['cycles_autotest', true],\n         ['debug_autotest', true],\n         ['eal_flags_c_opt_autotest', false],\n-        ['eal_flags_master_opt_autotest', false],\n+        ['eal_flags_main_opt_autotest', false],\n         ['eal_flags_n_opt_autotest', false],\n         ['eal_flags_hpet_autotest', false],\n         ['eal_flags_no_huge_autotest', false],\ndiff --git a/app/test/test.c b/app/test/test.c\nindex 94d26ab1f67c..3b3d172c91fa 100644\n--- a/app/test/test.c\n+++ b/app/test/test.c\n@@ -58,7 +58,7 @@ do_recursive_call(void)\n #endif\n #endif\n \t\t\t{ \"test_missing_c_flag\", no_action },\n-\t\t\t{ \"test_master_lcore_flag\", no_action },\n+\t\t\t{ \"test_main_lcore_flag\", no_action },\n \t\t\t{ \"test_invalid_n_flag\", no_action },\n \t\t\t{ \"test_no_hpet_flag\", no_action },\n \t\t\t{ \"test_whitelist_flag\", no_action },\ndiff --git a/app/test/test_atomic.c b/app/test/test_atomic.c\nindex 214452e54399..f10f555af8b4 100644\n--- a/app/test/test_atomic.c\n+++ b/app/test/test_atomic.c\n@@ -456,7 +456,7 @@ test_atomic(void)\n \n \tprintf(\"usual inc/dec/add/sub functions\\n\");\n \n-\trte_eal_mp_remote_launch(test_atomic_usual, NULL, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(test_atomic_usual, NULL, SKIP_MAIN);\n \trte_atomic32_set(&synchro, 1);\n \trte_eal_mp_wait_lcore();\n \trte_atomic32_set(&synchro, 0);\n@@ -482,7 +482,7 @@ test_atomic(void)\n \trte_atomic32_set(&a32, 0);\n \trte_atomic16_set(&a16, 0);\n \trte_atomic64_set(&count, 0);\n-\trte_eal_mp_remote_launch(test_atomic_tas, NULL, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(test_atomic_tas, NULL, SKIP_MAIN);\n \trte_atomic32_set(&synchro, 1);\n \trte_eal_mp_wait_lcore();\n \trte_atomic32_set(&synchro, 0);\n@@ -499,7 +499,7 @@ test_atomic(void)\n \trte_atomic16_set(&a16, 0);\n \trte_atomic64_set(&count, 0);\n \trte_eal_mp_remote_launch(test_atomic_addsub_and_return, NULL,\n-\t\t\t\t SKIP_MASTER);\n+\t\t\t\t SKIP_MAIN);\n \trte_atomic32_set(&synchro, 1);\n \trte_eal_mp_wait_lcore();\n \trte_atomic32_set(&synchro, 0);\n@@ -510,8 +510,8 @@ test_atomic(void)\n \t}\n \n \t/*\n-\t * Set a64, a32 and a16 with the same value of minus \"number of slave\n-\t * lcores\", launch all slave lcores to atomically increase by one and\n+\t * Set a64, a32 and a16 with the same value of minus \"number of worker\n+\t * lcores\", launch all worker lcores to atomically increase by one and\n \t * test them respectively.\n \t * Each lcore should have only one chance to increase a64 by one and\n \t * then check if it is equal to 0, but there should be only one lcore\n@@ -519,7 +519,7 @@ test_atomic(void)\n \t * Then a variable of \"count\", initialized to zero, is increased by\n \t * one if a64, a32 or a16 is 0 after being increased and tested\n \t * atomically.\n-\t * We can check if \"count\" is finally equal to 3 to see if all slave\n+\t * We can check if \"count\" is finally equal to 3 to see if all worker\n \t * lcores performed \"atomic inc and test\" right.\n \t */\n \tprintf(\"inc and test\\n\");\n@@ -533,7 +533,7 @@ test_atomic(void)\n \trte_atomic64_set(&a64, (int64_t)(1 - (int64_t)rte_lcore_count()));\n \trte_atomic32_set(&a32, (int32_t)(1 - (int32_t)rte_lcore_count()));\n \trte_atomic16_set(&a16, (int16_t)(1 - (int16_t)rte_lcore_count()));\n-\trte_eal_mp_remote_launch(test_atomic_inc_and_test, NULL, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(test_atomic_inc_and_test, NULL, SKIP_MAIN);\n \trte_atomic32_set(&synchro, 1);\n \trte_eal_mp_wait_lcore();\n \trte_atomic32_clear(&synchro);\n@@ -544,7 +544,7 @@ test_atomic(void)\n \t}\n \n \t/*\n-\t * Same as above, but this time we set the values to \"number of slave\n+\t * Same as above, but this time we set the values to \"number of worker\n \t * lcores\", and decrement instead of increment.\n \t */\n \tprintf(\"dec and test\\n\");\n@@ -555,7 +555,7 @@ test_atomic(void)\n \trte_atomic64_set(&a64, (int64_t)(rte_lcore_count() - 1));\n \trte_atomic32_set(&a32, (int32_t)(rte_lcore_count() - 1));\n \trte_atomic16_set(&a16, (int16_t)(rte_lcore_count() - 1));\n-\trte_eal_mp_remote_launch(test_atomic_dec_and_test, NULL, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(test_atomic_dec_and_test, NULL, SKIP_MAIN);\n \trte_atomic32_set(&synchro, 1);\n \trte_eal_mp_wait_lcore();\n \trte_atomic32_clear(&synchro);\n@@ -569,10 +569,10 @@ test_atomic(void)\n \t/*\n \t * This case tests the functionality of rte_atomic128_cmp_exchange\n \t * API. It calls rte_atomic128_cmp_exchange with four kinds of memory\n-\t * models successively on each slave core. Once each 128-bit atomic\n+\t * models successively on each worker core. Once each 128-bit atomic\n \t * compare and swap operation is successful, it updates the global\n \t * 128-bit counter by 2 for the first 64-bit and 1 for the second\n-\t * 64-bit. Each slave core iterates this test N times.\n+\t * 64-bit. Each worker core iterates this test N times.\n \t * At the end of test, verify whether the first 64-bits of the 128-bit\n \t * counter and the second 64bits is differ by the total iterations. If\n \t * it is, the test passes.\n@@ -585,7 +585,7 @@ test_atomic(void)\n \tcount128.val[1] = 0;\n \n \trte_eal_mp_remote_launch(test_atomic128_cmp_exchange, NULL,\n-\t\t\t\t SKIP_MASTER);\n+\t\t\t\t SKIP_MAIN);\n \trte_atomic32_set(&synchro, 1);\n \trte_eal_mp_wait_lcore();\n \trte_atomic32_clear(&synchro);\n@@ -619,7 +619,7 @@ test_atomic(void)\n \ttoken64 = ((uint64_t)get_crc8(&t.u8[0], sizeof(token64) - 1) << 56)\n \t\t| (t.u64 & 0x00ffffffffffffff);\n \n-\trte_eal_mp_remote_launch(test_atomic_exchange, NULL, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(test_atomic_exchange, NULL, SKIP_MAIN);\n \trte_atomic32_set(&synchro, 1);\n \trte_eal_mp_wait_lcore();\n \trte_atomic32_clear(&synchro);\ndiff --git a/app/test/test_barrier.c b/app/test/test_barrier.c\nindex 43b5f6232c6d..c27f8a0742f2 100644\n--- a/app/test/test_barrier.c\n+++ b/app/test/test_barrier.c\n@@ -236,7 +236,7 @@ plock_test(uint64_t iter, enum plock_use_type utype)\n \n \t/* test phase - start and wait for completion on each active lcore */\n \n-\trte_eal_mp_remote_launch(plock_test1_lcore, lpt, CALL_MASTER);\n+\trte_eal_mp_remote_launch(plock_test1_lcore, lpt, CALL_MAIN);\n \trte_eal_mp_wait_lcore();\n \n \t/* validation phase - make sure that shared and local data match */\ndiff --git a/app/test/test_compressdev_test_buffer.h b/app/test/test_compressdev_test_buffer.h\nindex 4a16ade150aa..b09a7e99688f 100644\n--- a/app/test/test_compressdev_test_buffer.h\n+++ b/app/test/test_compressdev_test_buffer.h\n@@ -190,7 +190,7 @@ static const char test_buf_shakespeare[] =\n \t\"\tservitude: I will no longer endure it, though yet I\\n\"\n \t\"\tknow no wise remedy how to avoid it.\\n\"\n \t\"\\n\"\n-\t\"ADAM\tYonder comes my master, your brother.\\n\"\n+\t\"ADAM\tYonder comes my main, your brother.\\n\"\n \t\"\\n\"\n \t\"ORLANDO\tGo apart, Adam, and thou shalt hear how he will\\n\";\n \ndiff --git a/app/test/test_cryptodev.c b/app/test/test_cryptodev.c\nindex 153d00119491..61de6a783ac5 100644\n--- a/app/test/test_cryptodev.c\n+++ b/app/test/test_cryptodev.c\n@@ -479,29 +479,29 @@ testsuite_setup(void)\n \tchar vdev_args[VDEV_ARGS_SIZE] = {\"\"};\n \tchar temp_str[VDEV_ARGS_SIZE] = {\"mode=multi-core,\"\n \t\t\"ordering=enable,name=cryptodev_test_scheduler,corelist=\"};\n-\tuint16_t slave_core_count = 0;\n+\tuint16_t worker_core_count = 0;\n \tuint16_t socket_id = 0;\n \n \tif (gbl_driver_id == rte_cryptodev_driver_id_get(\n \t\t\tRTE_STR(CRYPTODEV_NAME_SCHEDULER_PMD))) {\n \n-\t\t/* Identify the Slave Cores\n-\t\t * Use 2 slave cores for the device args\n+\t\t/* Identify the Worker Cores\n+\t\t * Use 2 worker cores for the device args\n \t\t */\n-\t\tRTE_LCORE_FOREACH_SLAVE(i) {\n-\t\t\tif (slave_core_count > 1)\n+\t\tRTE_LCORE_FOREACH_WORKER(i) {\n+\t\t\tif (worker_core_count > 1)\n \t\t\t\tbreak;\n \t\t\tsnprintf(vdev_args, sizeof(vdev_args),\n \t\t\t\t\t\"%s%d\", temp_str, i);\n \t\t\tstrcpy(temp_str, vdev_args);\n \t\t\tstrlcat(temp_str, \";\", sizeof(temp_str));\n-\t\t\tslave_core_count++;\n+\t\t\tworker_core_count++;\n \t\t\tsocket_id = rte_lcore_to_socket_id(i);\n \t\t}\n-\t\tif (slave_core_count != 2) {\n+\t\tif (worker_core_count != 2) {\n \t\t\tRTE_LOG(ERR, USER1,\n \t\t\t\t\"Cryptodev scheduler test require at least \"\n-\t\t\t\t\"two slave cores to run. \"\n+\t\t\t\t\"two worker cores to run. \"\n \t\t\t\t\"Please use the correct coremask.\\n\");\n \t\t\treturn TEST_FAILED;\n \t\t}\ndiff --git a/app/test/test_distributor.c b/app/test/test_distributor.c\nindex ba1f81cf8d19..19aa9f90e501 100644\n--- a/app/test/test_distributor.c\n+++ b/app/test/test_distributor.c\n@@ -654,13 +654,13 @@ test_distributor(void)\n \t\t\t\t\tsizeof(worker_params.name));\n \n \t\trte_eal_mp_remote_launch(handle_work,\n-\t\t\t\t&worker_params, SKIP_MASTER);\n+\t\t\t\t&worker_params, SKIP_MAIN);\n \t\tif (sanity_test(&worker_params, p) < 0)\n \t\t\tgoto err;\n \t\tquit_workers(&worker_params, p);\n \n \t\trte_eal_mp_remote_launch(handle_work_with_free_mbufs,\n-\t\t\t\t&worker_params, SKIP_MASTER);\n+\t\t\t\t&worker_params, SKIP_MAIN);\n \t\tif (sanity_test_with_mbuf_alloc(&worker_params, p) < 0)\n \t\t\tgoto err;\n \t\tquit_workers(&worker_params, p);\n@@ -668,7 +668,7 @@ test_distributor(void)\n \t\tif (rte_lcore_count() > 2) {\n \t\t\trte_eal_mp_remote_launch(handle_work_for_shutdown_test,\n \t\t\t\t\t&worker_params,\n-\t\t\t\t\tSKIP_MASTER);\n+\t\t\t\t\tSKIP_MAIN);\n \t\t\tif (sanity_test_with_worker_shutdown(&worker_params,\n \t\t\t\t\tp) < 0)\n \t\t\t\tgoto err;\n@@ -676,7 +676,7 @@ test_distributor(void)\n \n \t\t\trte_eal_mp_remote_launch(handle_work_for_shutdown_test,\n \t\t\t\t\t&worker_params,\n-\t\t\t\t\tSKIP_MASTER);\n+\t\t\t\t\tSKIP_MAIN);\n \t\t\tif (test_flush_with_worker_shutdown(&worker_params,\n \t\t\t\t\tp) < 0)\n \t\t\t\tgoto err;\ndiff --git a/app/test/test_distributor_perf.c b/app/test/test_distributor_perf.c\nindex f153bcf9bd87..b25f79a3486c 100644\n--- a/app/test/test_distributor_perf.c\n+++ b/app/test/test_distributor_perf.c\n@@ -54,10 +54,10 @@ time_cache_line_switch(void)\n \t/* allocate a full cache line for data, we use only first byte of it */\n \tuint64_t data[RTE_CACHE_LINE_SIZE*3 / sizeof(uint64_t)];\n \n-\tunsigned i, slaveid = rte_get_next_lcore(rte_lcore_id(), 0, 0);\n+\tunsigned int i, workerid = rte_get_next_lcore(rte_lcore_id(), 0, 0);\n \tvolatile uint64_t *pdata = &data[0];\n \t*pdata = 1;\n-\trte_eal_remote_launch((lcore_function_t *)flip_bit, &data[0], slaveid);\n+\trte_eal_remote_launch((lcore_function_t *)flip_bit, &data[0], workerid);\n \twhile (*pdata)\n \t\trte_pause();\n \n@@ -72,7 +72,7 @@ time_cache_line_switch(void)\n \twhile (*pdata)\n \t\trte_pause();\n \t*pdata = 2;\n-\trte_eal_wait_lcore(slaveid);\n+\trte_eal_wait_lcore(workerid);\n \tprintf(\"==== Cache line switch test ===\\n\");\n \tprintf(\"Time for %u iterations = %\"PRIu64\" ticks\\n\", (1<<ITER_POWER_CL),\n \t\t\tend_time-start_time);\n@@ -251,13 +251,13 @@ test_distributor_perf(void)\n \t}\n \n \tprintf(\"=== Performance test of distributor (single mode) ===\\n\");\n-\trte_eal_mp_remote_launch(handle_work, ds, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(handle_work, ds, SKIP_MAIN);\n \tif (perf_test(ds, p) < 0)\n \t\treturn -1;\n \tquit_workers(ds, p);\n \n \tprintf(\"=== Performance test of distributor (burst mode) ===\\n\");\n-\trte_eal_mp_remote_launch(handle_work, db, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(handle_work, db, SKIP_MAIN);\n \tif (perf_test(db, p) < 0)\n \t\treturn -1;\n \tquit_workers(db, p);\ndiff --git a/app/test/test_eal_flags.c b/app/test/test_eal_flags.c\nindex b019656b29c3..a12836d46d61 100644\n--- a/app/test/test_eal_flags.c\n+++ b/app/test/test_eal_flags.c\n@@ -599,10 +599,10 @@ test_missing_c_flag(void)\n }\n \n /*\n- * Test --master-lcore option with matching coremask\n+ * Test --main-lcore option with matching coremask\n  */\n static int\n-test_master_lcore_flag(void)\n+test_main_lcore_flag(void)\n {\n #ifdef RTE_EXEC_ENV_FREEBSD\n \t/* BSD target doesn't support prefixes at this point */\n@@ -619,34 +619,34 @@ test_master_lcore_flag(void)\n \tif (!rte_lcore_is_enabled(0) || !rte_lcore_is_enabled(1))\n \t\treturn TEST_SKIPPED;\n \n-\t/* --master-lcore flag but no value */\n+\t/* --main-lcore flag but no value */\n \tconst char *argv1[] = { prgname, prefix, mp_flag,\n-\t\t\t\t\"-c\", \"3\", \"--master-lcore\"};\n-\t/* --master-lcore flag with invalid value */\n+\t\t\t\t\"-c\", \"3\", \"--main-lcore\"};\n+\t/* --main-lcore flag with invalid value */\n \tconst char *argv2[] = { prgname, prefix, mp_flag,\n-\t\t\t\t\"-c\", \"3\", \"--master-lcore\", \"-1\"};\n+\t\t\t\t\"-c\", \"3\", \"--main-lcore\", \"-1\"};\n \tconst char *argv3[] = { prgname, prefix, mp_flag,\n-\t\t\t\t\"-c\", \"3\", \"--master-lcore\", \"X\"};\n-\t/* master lcore not in coremask */\n+\t\t\t\t\"-c\", \"3\", \"--main-lcore\", \"X\"};\n+\t/* main lcore not in coremask */\n \tconst char *argv4[] = { prgname, prefix, mp_flag,\n-\t\t\t\t\"-c\", \"3\", \"--master-lcore\", \"2\"};\n+\t\t\t\t\"-c\", \"3\", \"--main-lcore\", \"2\"};\n \t/* valid value */\n \tconst char *argv5[] = { prgname, prefix, mp_flag,\n-\t\t\t\t\"-c\", \"3\", \"--master-lcore\", \"1\"};\n+\t\t\t\t\"-c\", \"3\", \"--main-lcore\", \"1\"};\n \t/* valid value set before coremask */\n \tconst char *argv6[] = { prgname, prefix, mp_flag,\n-\t\t\t\t\"--master-lcore\", \"1\", \"-c\", \"3\"};\n+\t\t\t\t\"--main-lcore\", \"1\", \"-c\", \"3\"};\n \n \tif (launch_proc(argv1) == 0\n \t\t\t|| launch_proc(argv2) == 0\n \t\t\t|| launch_proc(argv3) == 0\n \t\t\t|| launch_proc(argv4) == 0) {\n-\t\tprintf(\"Error - process ran without error with wrong --master-lcore\\n\");\n+\t\tprintf(\"Error - process ran without error with wrong --main-lcore\\n\");\n \t\treturn -1;\n \t}\n \tif (launch_proc(argv5) != 0\n \t\t\t|| launch_proc(argv6) != 0) {\n-\t\tprintf(\"Error - process did not run ok with valid --master-lcore\\n\");\n+\t\tprintf(\"Error - process did not run ok with valid --main-lcore\\n\");\n \t\treturn -1;\n \t}\n \treturn 0;\n@@ -1468,9 +1468,9 @@ test_eal_flags(void)\n \t\treturn ret;\n \t}\n \n-\tret = test_master_lcore_flag();\n+\tret = test_main_lcore_flag();\n \tif (ret < 0) {\n-\t\tprintf(\"Error in test_master_lcore_flag()\\n\");\n+\t\tprintf(\"Error in test_main_lcore_flag()\\n\");\n \t\treturn ret;\n \t}\n \n@@ -1542,7 +1542,7 @@ REGISTER_TEST_COMMAND(eal_flags_autotest, test_eal_flags);\n \n /* subtests used in meson for CI */\n REGISTER_TEST_COMMAND(eal_flags_c_opt_autotest, test_missing_c_flag);\n-REGISTER_TEST_COMMAND(eal_flags_master_opt_autotest, test_master_lcore_flag);\n+REGISTER_TEST_COMMAND(eal_flags_main_opt_autotest, test_main_lcore_flag);\n REGISTER_TEST_COMMAND(eal_flags_n_opt_autotest, test_invalid_n_flag);\n REGISTER_TEST_COMMAND(eal_flags_hpet_autotest, test_no_hpet_flag);\n REGISTER_TEST_COMMAND(eal_flags_no_huge_autotest, test_no_huge_flag);\ndiff --git a/app/test/test_efd.c b/app/test/test_efd.c\nindex e9df46967d61..180dc4748ed0 100644\n--- a/app/test/test_efd.c\n+++ b/app/test/test_efd.c\n@@ -95,7 +95,7 @@ static inline uint8_t efd_get_all_sockets_bitmask(void)\n {\n \tuint8_t all_cpu_sockets_bitmask = 0;\n \tunsigned int i;\n-\tunsigned int next_lcore = rte_get_master_lcore();\n+\tunsigned int next_lcore = rte_get_main_lcore();\n \tconst int val_true = 1, val_false = 0;\n \tfor (i = 0; i < rte_lcore_count(); i++) {\n \t\tall_cpu_sockets_bitmask |= 1 << rte_lcore_to_socket_id(next_lcore);\ndiff --git a/app/test/test_efd_perf.c b/app/test/test_efd_perf.c\nindex d47622d5cafd..1c4770447526 100644\n--- a/app/test/test_efd_perf.c\n+++ b/app/test/test_efd_perf.c\n@@ -33,7 +33,7 @@ static inline uint8_t efd_get_all_sockets_bitmask(void)\n {\n \tuint8_t all_cpu_sockets_bitmask = 0;\n \tunsigned int i;\n-\tunsigned int next_lcore = rte_get_master_lcore();\n+\tunsigned int next_lcore = rte_get_main_lcore();\n \tconst int val_true = 1, val_false = 0;\n \tfor (i = 0; i < rte_lcore_count(); i++) {\n \t\tall_cpu_sockets_bitmask |= 1 << rte_lcore_to_socket_id(next_lcore);\ndiff --git a/app/test/test_func_reentrancy.c b/app/test/test_func_reentrancy.c\nindex f5ddd03d7670..f407d82d5de7 100644\n--- a/app/test/test_func_reentrancy.c\n+++ b/app/test/test_func_reentrancy.c\n@@ -57,8 +57,8 @@ typedef void (*case_clean_t)(unsigned lcore_id);\n static rte_atomic32_t obj_count = RTE_ATOMIC32_INIT(0);\n static rte_atomic32_t synchro = RTE_ATOMIC32_INIT(0);\n \n-#define WAIT_SYNCHRO_FOR_SLAVES()   do{ \\\n-\tif (lcore_self != rte_get_master_lcore())                  \\\n+#define WAIT_SYNCHRO_FOR_WORKERS()   do { \\\n+\tif (lcore_self != rte_get_main_lcore())                  \\\n \t\twhile (rte_atomic32_read(&synchro) == 0);        \\\n } while(0)\n \n@@ -70,7 +70,7 @@ test_eal_init_once(__rte_unused void *arg)\n {\n \tunsigned lcore_self =  rte_lcore_id();\n \n-\tWAIT_SYNCHRO_FOR_SLAVES();\n+\tWAIT_SYNCHRO_FOR_WORKERS();\n \n \trte_atomic32_set(&obj_count, 1); /* silent the check in the caller */\n \tif (rte_eal_init(0, NULL) != -1)\n@@ -106,7 +106,7 @@ ring_create_lookup(__rte_unused void *arg)\n \tchar ring_name[MAX_STRING_SIZE];\n \tint i;\n \n-\tWAIT_SYNCHRO_FOR_SLAVES();\n+\tWAIT_SYNCHRO_FOR_WORKERS();\n \n \t/* create the same ring simultaneously on all threads */\n \tfor (i = 0; i < MAX_ITER_ONCE; i++) {\n@@ -166,7 +166,7 @@ mempool_create_lookup(__rte_unused void *arg)\n \tchar mempool_name[MAX_STRING_SIZE];\n \tint i;\n \n-\tWAIT_SYNCHRO_FOR_SLAVES();\n+\tWAIT_SYNCHRO_FOR_WORKERS();\n \n \t/* create the same mempool simultaneously on all threads */\n \tfor (i = 0; i < MAX_ITER_ONCE; i++) {\n@@ -232,7 +232,7 @@ hash_create_free(__rte_unused void *arg)\n \t\t.socket_id = 0,\n \t};\n \n-\tWAIT_SYNCHRO_FOR_SLAVES();\n+\tWAIT_SYNCHRO_FOR_WORKERS();\n \n \t/* create the same hash simultaneously on all threads */\n \thash_params.name = \"fr_test_once\";\n@@ -296,7 +296,7 @@ fbk_create_free(__rte_unused void *arg)\n \t\t.init_val = RTE_FBK_HASH_INIT_VAL_DEFAULT,\n \t};\n \n-\tWAIT_SYNCHRO_FOR_SLAVES();\n+\tWAIT_SYNCHRO_FOR_WORKERS();\n \n \t/* create the same fbk hash table simultaneously on all threads */\n \tfbk_params.name = \"fr_test_once\";\n@@ -359,7 +359,7 @@ lpm_create_free(__rte_unused void *arg)\n \tchar lpm_name[MAX_STRING_SIZE];\n \tint i;\n \n-\tWAIT_SYNCHRO_FOR_SLAVES();\n+\tWAIT_SYNCHRO_FOR_WORKERS();\n \n \t/* create the same lpm simultaneously on all threads */\n \tfor (i = 0; i < MAX_ITER_ONCE; i++) {\n@@ -430,7 +430,7 @@ launch_test(struct test_case *pt_case)\n \trte_atomic32_set(&obj_count, 0);\n \trte_atomic32_set(&synchro, 0);\n \n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (cores == 1)\n \t\t\tbreak;\n \t\tcores--;\n@@ -443,7 +443,7 @@ launch_test(struct test_case *pt_case)\n \t\tret = -1;\n \n \tcores = cores_save;\n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (cores == 1)\n \t\t\tbreak;\n \t\tcores--;\ndiff --git a/app/test/test_hash_multiwriter.c b/app/test/test_hash_multiwriter.c\nindex 46ab45f82a99..afa3c7b93d85 100644\n--- a/app/test/test_hash_multiwriter.c\n+++ b/app/test/test_hash_multiwriter.c\n@@ -64,7 +64,7 @@ test_hash_multiwriter_worker(void *arg)\n \n \t/*\n \t * Calculate offset for entries based on the position of the\n-\t * logical core, from the master core (not counting not enabled cores)\n+\t * logical core, from the main core (not counting not enabled cores)\n \t */\n \toffset = pos_core * tbl_multiwriter_test_params.nb_tsx_insertion;\n \n@@ -194,7 +194,7 @@ test_hash_multiwriter(void)\n \n \t/* Fire all threads. */\n \trte_eal_mp_remote_launch(test_hash_multiwriter_worker,\n-\t\t\t\t enabled_core_ids, CALL_MASTER);\n+\t\t\t\t enabled_core_ids, CALL_MAIN);\n \trte_eal_mp_wait_lcore();\n \n \tcount = rte_hash_count(handle);\ndiff --git a/app/test/test_hash_readwrite.c b/app/test/test_hash_readwrite.c\nindex 73333dff1392..4860768a6491 100644\n--- a/app/test/test_hash_readwrite.c\n+++ b/app/test/test_hash_readwrite.c\n@@ -25,7 +25,7 @@\n #define NUM_TEST 3\n unsigned int core_cnt[NUM_TEST] = {2, 4, 8};\n \n-unsigned int slave_core_ids[RTE_MAX_LCORE];\n+unsigned int worker_core_ids[RTE_MAX_LCORE];\n struct perf {\n \tuint32_t single_read;\n \tuint32_t single_write;\n@@ -65,7 +65,7 @@ test_hash_readwrite_worker(__rte_unused void *arg)\n \tret = rte_malloc(NULL, sizeof(int) *\n \t\t\t\ttbl_rw_test_param.num_insert, 0);\n \tfor (i = 0; i < rte_lcore_count(); i++) {\n-\t\tif (slave_core_ids[i] == lcore_id)\n+\t\tif (worker_core_ids[i] == lcore_id)\n \t\t\tbreak;\n \t}\n \toffset = tbl_rw_test_param.num_insert * i;\n@@ -206,7 +206,7 @@ test_hash_readwrite_functional(int use_htm, int use_rw_lf, int use_ext)\n \tuint32_t duplicated_keys = 0;\n \tuint32_t lost_keys = 0;\n \tint use_jhash = 1;\n-\tint slave_cnt = rte_lcore_count() - 1;\n+\tint worker_cnt = rte_lcore_count() - 1;\n \tuint32_t tot_insert = 0;\n \n \trte_atomic64_init(&gcycles);\n@@ -224,11 +224,10 @@ test_hash_readwrite_functional(int use_htm, int use_rw_lf, int use_ext)\n \t\ttot_insert = TOTAL_INSERT;\n \n \ttbl_rw_test_param.num_insert =\n-\t\ttot_insert / slave_cnt;\n+\t\ttot_insert / worker_cnt;\n \n \ttbl_rw_test_param.rounded_tot_insert =\n-\t\ttbl_rw_test_param.num_insert\n-\t\t* slave_cnt;\n+\t\ttbl_rw_test_param.num_insert * worker_cnt;\n \n \tprintf(\"\\nHTM = %d, RW-LF = %d, EXT-Table = %d\\n\",\n \t\tuse_htm, use_rw_lf, use_ext);\n@@ -236,7 +235,7 @@ test_hash_readwrite_functional(int use_htm, int use_rw_lf, int use_ext)\n \n \t/* Fire all threads. */\n \trte_eal_mp_remote_launch(test_hash_readwrite_worker,\n-\t\t\t\t NULL, SKIP_MASTER);\n+\t\t\t\t NULL, SKIP_MAIN);\n \trte_eal_mp_wait_lcore();\n \n \twhile (rte_hash_iterate(tbl_rw_test_param.h, &next_key,\n@@ -330,7 +329,7 @@ test_rw_writer(void *arg)\n \tuint64_t offset;\n \n \tfor (i = 0; i < rte_lcore_count(); i++) {\n-\t\tif (slave_core_ids[i] == lcore_id)\n+\t\tif (worker_core_ids[i] == lcore_id)\n \t\t\tbreak;\n \t}\n \n@@ -433,8 +432,8 @@ test_hash_readwrite_perf(struct perf *perf_results, int use_htm,\n \tperf_results->single_read = end / i;\n \n \tfor (n = 0; n < NUM_TEST; n++) {\n-\t\tunsigned int tot_slave_lcore = rte_lcore_count() - 1;\n-\t\tif (tot_slave_lcore < core_cnt[n] * 2)\n+\t\tunsigned int tot_worker_lcore = rte_lcore_count() - 1;\n+\t\tif (tot_worker_lcore < core_cnt[n] * 2)\n \t\t\tgoto finish;\n \n \t\trte_atomic64_clear(&greads);\n@@ -467,7 +466,7 @@ test_hash_readwrite_perf(struct perf *perf_results, int use_htm,\n \t\tfor (i = 0; i < core_cnt[n]; i++)\n \t\t\trte_eal_remote_launch(test_rw_reader,\n \t\t\t\t\t(void *)(uintptr_t)read_cnt,\n-\t\t\t\t\tslave_core_ids[i]);\n+\t\t\t\t\tworker_core_ids[i]);\n \n \t\trte_eal_mp_wait_lcore();\n \n@@ -476,7 +475,7 @@ test_hash_readwrite_perf(struct perf *perf_results, int use_htm,\n \t\tfor (; i < core_cnt[n] * 2; i++)\n \t\t\trte_eal_remote_launch(test_rw_writer,\n \t\t\t\t\t(void *)((uintptr_t)start_coreid),\n-\t\t\t\t\tslave_core_ids[i]);\n+\t\t\t\t\tworker_core_ids[i]);\n \n \t\trte_eal_mp_wait_lcore();\n \n@@ -521,20 +520,20 @@ test_hash_readwrite_perf(struct perf *perf_results, int use_htm,\n \t\t\tfor (i = core_cnt[n]; i < core_cnt[n] * 2; i++)\n \t\t\t\trte_eal_remote_launch(test_rw_writer,\n \t\t\t\t\t(void *)((uintptr_t)start_coreid),\n-\t\t\t\t\tslave_core_ids[i]);\n+\t\t\t\t\tworker_core_ids[i]);\n \t\t\tfor (i = 0; i < core_cnt[n]; i++)\n \t\t\t\trte_eal_remote_launch(test_rw_reader,\n \t\t\t\t\t(void *)(uintptr_t)read_cnt,\n-\t\t\t\t\tslave_core_ids[i]);\n+\t\t\t\t\tworker_core_ids[i]);\n \t\t} else {\n \t\t\tfor (i = 0; i < core_cnt[n]; i++)\n \t\t\t\trte_eal_remote_launch(test_rw_reader,\n \t\t\t\t\t(void *)(uintptr_t)read_cnt,\n-\t\t\t\t\tslave_core_ids[i]);\n+\t\t\t\t\tworker_core_ids[i]);\n \t\t\tfor (; i < core_cnt[n] * 2; i++)\n \t\t\t\trte_eal_remote_launch(test_rw_writer,\n \t\t\t\t\t(void *)((uintptr_t)start_coreid),\n-\t\t\t\t\tslave_core_ids[i]);\n+\t\t\t\t\tworker_core_ids[i]);\n \t\t}\n \n \t\trte_eal_mp_wait_lcore();\n@@ -626,8 +625,8 @@ test_hash_rw_perf_main(void)\n \t\treturn TEST_SKIPPED;\n \t}\n \n-\tRTE_LCORE_FOREACH_SLAVE(core_id) {\n-\t\tslave_core_ids[i] = core_id;\n+\tRTE_LCORE_FOREACH_WORKER(core_id) {\n+\t\tworker_core_ids[i] = core_id;\n \t\ti++;\n \t}\n \n@@ -710,8 +709,8 @@ test_hash_rw_func_main(void)\n \t\treturn TEST_SKIPPED;\n \t}\n \n-\tRTE_LCORE_FOREACH_SLAVE(core_id) {\n-\t\tslave_core_ids[i] = core_id;\n+\tRTE_LCORE_FOREACH_WORKER(core_id) {\n+\t\tworker_core_ids[i] = core_id;\n \t\ti++;\n \t}\n \ndiff --git a/app/test/test_kni.c b/app/test/test_kni.c\nindex e47ab36e0231..d4e31e4ae13d 100644\n--- a/app/test/test_kni.c\n+++ b/app/test/test_kni.c\n@@ -85,7 +85,7 @@ static struct rte_kni_ops kni_ops = {\n \t.config_promiscusity = NULL,\n };\n \n-static unsigned lcore_master, lcore_ingress, lcore_egress;\n+static unsigned int lcore_main, lcore_ingress, lcore_egress;\n static struct rte_kni *test_kni_ctx;\n static struct test_kni_stats stats;\n \n@@ -202,7 +202,7 @@ test_kni_link_change(void)\n  * supported by KNI kernel module. The ingress lcore will allocate mbufs and\n  * transmit them to kernel space; while the egress lcore will receive the mbufs\n  * from kernel space and free them.\n- * On the master lcore, several commands will be run to check handling the\n+ * On the main lcore, several commands will be run to check handling the\n  * kernel requests. And it will finally set the flag to exit the KNI\n  * transmitting/receiving to/from the kernel space.\n  *\n@@ -217,7 +217,7 @@ test_kni_loop(__rte_unused void *arg)\n \tconst unsigned lcore_id = rte_lcore_id();\n \tstruct rte_mbuf *pkts_burst[PKT_BURST_SZ];\n \n-\tif (lcore_id == lcore_master) {\n+\tif (lcore_id == lcore_main) {\n \t\trte_delay_ms(KNI_TIMEOUT_MS);\n \t\t/* tests of handling kernel request */\n \t\tif (system(IFCONFIG TEST_KNI_PORT\" up\") == -1)\n@@ -276,12 +276,12 @@ test_kni_allocate_lcores(void)\n {\n \tunsigned i, count = 0;\n \n-\tlcore_master = rte_get_master_lcore();\n-\tprintf(\"master lcore: %u\\n\", lcore_master);\n+\tlcore_main = rte_get_main_lcore();\n+\tprintf(\"main lcore: %u\\n\", lcore_main);\n \tfor (i = 0; i < RTE_MAX_LCORE; i++) {\n \t\tif (count >=2 )\n \t\t\tbreak;\n-\t\tif (rte_lcore_is_enabled(i) && i != lcore_master) {\n+\t\tif (rte_lcore_is_enabled(i) && i != lcore_main) {\n \t\t\tcount ++;\n \t\t\tif (count == 1)\n \t\t\t\tlcore_ingress = i;\n@@ -487,8 +487,8 @@ test_kni_processing(uint16_t port_id, struct rte_mempool *mp)\n \tif (ret != 0)\n \t\tgoto fail_kni;\n \n-\trte_eal_mp_remote_launch(test_kni_loop, NULL, CALL_MASTER);\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\trte_eal_mp_remote_launch(test_kni_loop, NULL, CALL_MAIN);\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\tif (rte_eal_wait_lcore(i) < 0) {\n \t\t\tret = -1;\n \t\t\tgoto fail_kni;\ndiff --git a/app/test/test_lpm_perf.c b/app/test/test_lpm_perf.c\nindex 58076795ec3a..c5a238b9d1e8 100644\n--- a/app/test/test_lpm_perf.c\n+++ b/app/test/test_lpm_perf.c\n@@ -498,7 +498,7 @@ test_lpm_rcu_perf_multi_writer(void)\n \t}\n \n \tnum_cores = 0;\n-\tRTE_LCORE_FOREACH_SLAVE(core_id) {\n+\tRTE_LCORE_FOREACH_WORKER(core_id) {\n \t\tenabled_core_ids[num_cores] = core_id;\n \t\tnum_cores++;\n \t}\n@@ -651,7 +651,7 @@ test_lpm_rcu_perf(void)\n \t}\n \n \tnum_cores = 0;\n-\tRTE_LCORE_FOREACH_SLAVE(core_id) {\n+\tRTE_LCORE_FOREACH_WORKER(core_id) {\n \t\tenabled_core_ids[num_cores] = core_id;\n \t\tnum_cores++;\n \t}\ndiff --git a/app/test/test_malloc.c b/app/test/test_malloc.c\nindex 71b3cfdde5cf..758e6194a852 100644\n--- a/app/test/test_malloc.c\n+++ b/app/test/test_malloc.c\n@@ -1007,11 +1007,11 @@ test_malloc(void)\n \telse printf(\"test_realloc() passed\\n\");\n \n \t/*----------------------------*/\n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\trte_eal_remote_launch(test_align_overlap_per_lcore, NULL, lcore_id);\n \t}\n \n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (rte_eal_wait_lcore(lcore_id) < 0)\n \t\t\tret = -1;\n \t}\n@@ -1022,11 +1022,11 @@ test_malloc(void)\n \telse printf(\"test_align_overlap_per_lcore() passed\\n\");\n \n \t/*----------------------------*/\n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\trte_eal_remote_launch(test_reordered_free_per_lcore, NULL, lcore_id);\n \t}\n \n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (rte_eal_wait_lcore(lcore_id) < 0)\n \t\t\tret = -1;\n \t}\n@@ -1037,11 +1037,11 @@ test_malloc(void)\n \telse printf(\"test_reordered_free_per_lcore() passed\\n\");\n \n \t/*----------------------------*/\n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\trte_eal_remote_launch(test_random_alloc_free, NULL, lcore_id);\n \t}\n \n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (rte_eal_wait_lcore(lcore_id) < 0)\n \t\t\tret = -1;\n \t}\ndiff --git a/app/test/test_mbuf.c b/app/test/test_mbuf.c\nindex 06e44f0a79f8..80d1850da9d1 100644\n--- a/app/test/test_mbuf.c\n+++ b/app/test/test_mbuf.c\n@@ -72,7 +72,7 @@\n \n #ifdef RTE_MBUF_REFCNT_ATOMIC\n \n-static volatile uint32_t refcnt_stop_slaves;\n+static volatile uint32_t refcnt_stop_workers;\n static unsigned refcnt_lcore[RTE_MAX_LCORE];\n \n #endif\n@@ -1000,7 +1000,7 @@ test_pktmbuf_free_segment(struct rte_mempool *pktmbuf_pool)\n #ifdef RTE_MBUF_REFCNT_ATOMIC\n \n static int\n-test_refcnt_slave(void *arg)\n+test_refcnt_worker(void *arg)\n {\n \tunsigned lcore, free;\n \tvoid *mp = 0;\n@@ -1010,7 +1010,7 @@ test_refcnt_slave(void *arg)\n \tprintf(\"%s started at lcore %u\\n\", __func__, lcore);\n \n \tfree = 0;\n-\twhile (refcnt_stop_slaves == 0) {\n+\twhile (refcnt_stop_workers == 0) {\n \t\tif (rte_ring_dequeue(refcnt_mbuf_ring, &mp) == 0) {\n \t\t\tfree++;\n \t\t\trte_pktmbuf_free(mp);\n@@ -1038,7 +1038,7 @@ test_refcnt_iter(unsigned int lcore, unsigned int iter,\n \t/* For each mbuf in the pool:\n \t * - allocate mbuf,\n \t * - increment it's reference up to N+1,\n-\t * - enqueue it N times into the ring for slave cores to free.\n+\t * - enqueue it N times into the ring for worker cores to free.\n \t */\n \tfor (i = 0, n = rte_mempool_avail_count(refcnt_pool);\n \t    i != n && (m = rte_pktmbuf_alloc(refcnt_pool)) != NULL;\n@@ -1062,7 +1062,7 @@ test_refcnt_iter(unsigned int lcore, unsigned int iter,\n \t\trte_panic(\"(lcore=%u, iter=%u): was able to allocate only \"\n \t\t          \"%u from %u mbufs\\n\", lcore, iter, i, n);\n \n-\t/* wait till slave lcores  will consume all mbufs */\n+\t/* wait till worker lcores  will consume all mbufs */\n \twhile (!rte_ring_empty(refcnt_mbuf_ring))\n \t\t;\n \n@@ -1083,7 +1083,7 @@ test_refcnt_iter(unsigned int lcore, unsigned int iter,\n }\n \n static int\n-test_refcnt_master(struct rte_mempool *refcnt_pool,\n+test_refcnt_main(struct rte_mempool *refcnt_pool,\n \t\t   struct rte_ring *refcnt_mbuf_ring)\n {\n \tunsigned i, lcore;\n@@ -1094,7 +1094,7 @@ test_refcnt_master(struct rte_mempool *refcnt_pool,\n \tfor (i = 0; i != REFCNT_MAX_ITER; i++)\n \t\ttest_refcnt_iter(lcore, i, refcnt_pool, refcnt_mbuf_ring);\n \n-\trefcnt_stop_slaves = 1;\n+\trefcnt_stop_workers = 1;\n \trte_wmb();\n \n \tprintf(\"%s finished at lcore %u\\n\", __func__, lcore);\n@@ -1107,7 +1107,7 @@ static int\n test_refcnt_mbuf(void)\n {\n #ifdef RTE_MBUF_REFCNT_ATOMIC\n-\tunsigned int master, slave, tref;\n+\tunsigned int main_lcore, worker, tref;\n \tint ret = -1;\n \tstruct rte_mempool *refcnt_pool = NULL;\n \tstruct rte_ring *refcnt_mbuf_ring = NULL;\n@@ -1126,39 +1126,38 @@ test_refcnt_mbuf(void)\n \t\t\t\t\t      SOCKET_ID_ANY);\n \tif (refcnt_pool == NULL) {\n \t\tprintf(\"%s: cannot allocate \" MAKE_STRING(refcnt_pool) \"\\n\",\n-\t\t    __func__);\n+\t\t       __func__);\n \t\treturn -1;\n \t}\n \n \trefcnt_mbuf_ring = rte_ring_create(\"refcnt_mbuf_ring\",\n-\t\t\trte_align32pow2(REFCNT_RING_SIZE), SOCKET_ID_ANY,\n-\t\t\t\t\tRING_F_SP_ENQ);\n+\t\t\t\t\t   rte_align32pow2(REFCNT_RING_SIZE), SOCKET_ID_ANY,\n+\t\t\t\t\t   RING_F_SP_ENQ);\n \tif (refcnt_mbuf_ring == NULL) {\n \t\tprintf(\"%s: cannot allocate \" MAKE_STRING(refcnt_mbuf_ring)\n-\t\t    \"\\n\", __func__);\n+\t\t       \"\\n\", __func__);\n \t\tgoto err;\n \t}\n \n-\trefcnt_stop_slaves = 0;\n+\trefcnt_stop_workers = 0;\n \tmemset(refcnt_lcore, 0, sizeof (refcnt_lcore));\n \n-\trte_eal_mp_remote_launch(test_refcnt_slave, refcnt_mbuf_ring,\n-\t\t\t\t SKIP_MASTER);\n+\trte_eal_mp_remote_launch(test_refcnt_worker, refcnt_mbuf_ring, SKIP_MAIN);\n \n-\ttest_refcnt_master(refcnt_pool, refcnt_mbuf_ring);\n+\ttest_refcnt_main(refcnt_pool, refcnt_mbuf_ring);\n \n \trte_eal_mp_wait_lcore();\n \n \t/* check that we porcessed all references */\n \ttref = 0;\n-\tmaster = rte_get_master_lcore();\n+\tmain_lcore = rte_get_main_lcore();\n \n-\tRTE_LCORE_FOREACH_SLAVE(slave)\n-\t\ttref += refcnt_lcore[slave];\n+\tRTE_LCORE_FOREACH_WORKER(worker)\n+\t\ttref += refcnt_lcore[worker];\n \n-\tif (tref != refcnt_lcore[master])\n+\tif (tref != refcnt_lcore[main_lcore])\n \t\trte_panic(\"referenced mbufs: %u, freed mbufs: %u\\n\",\n-\t\t          tref, refcnt_lcore[master]);\n+\t\t\t  tref, refcnt_lcore[main_lcore]);\n \n \trte_mempool_dump(stdout, refcnt_pool);\n \trte_ring_dump(stdout, refcnt_mbuf_ring);\ndiff --git a/app/test/test_mcslock.c b/app/test/test_mcslock.c\nindex ddccaafa9242..fbca78707d2e 100644\n--- a/app/test/test_mcslock.c\n+++ b/app/test/test_mcslock.c\n@@ -28,7 +28,7 @@\n  * These tests are derived from spin lock test cases.\n  *\n  * - The functional test takes all of these locks and launches the\n- *   ''test_mcslock_per_core()'' function on each core (except the master).\n+ *   ''test_mcslock_per_core()'' function on each core (except the main).\n  *\n  *   - The function takes the global lock, display something, then releases\n  *     the global lock on each core.\n@@ -123,9 +123,9 @@ test_mcslock_perf(void)\n \tprintf(\"\\nTest with lock on %u cores...\\n\", (rte_lcore_count()));\n \n \trte_atomic32_set(&synchro, 0);\n-\trte_eal_mp_remote_launch(load_loop_fn, &lock, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(load_loop_fn, &lock, SKIP_MAIN);\n \n-\t/* start synchro and launch test on master */\n+\t/* start synchro and launch test on main */\n \trte_atomic32_set(&synchro, 1);\n \tload_loop_fn(&lock);\n \n@@ -154,8 +154,8 @@ test_mcslock_try(__rte_unused void *arg)\n \trte_mcslock_t ml_me     = RTE_PER_LCORE(_ml_me);\n \trte_mcslock_t ml_try_me = RTE_PER_LCORE(_ml_try_me);\n \n-\t/* Locked ml_try in the master lcore, so it should fail\n-\t * when trying to lock it in the slave lcore.\n+\t/* Locked ml_try in the main lcore, so it should fail\n+\t * when trying to lock it in the worker lcore.\n \t */\n \tif (rte_mcslock_trylock(&p_ml_try, &ml_try_me) == 0) {\n \t\trte_mcslock_lock(&p_ml, &ml_me);\n@@ -185,20 +185,20 @@ test_mcslock(void)\n \t * Test mcs lock & unlock on each core\n \t */\n \n-\t/* slave cores should be waiting: print it */\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\t/* worker cores should be waiting: print it */\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\tprintf(\"lcore %d state: %d\\n\", i,\n \t\t\t\t(int) rte_eal_get_lcore_state(i));\n \t}\n \n \trte_mcslock_lock(&p_ml, &ml_me);\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_eal_remote_launch(test_mcslock_per_core, NULL, i);\n \t}\n \n-\t/* slave cores should be busy: print it */\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\t/* worker cores should be busy: print it */\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\tprintf(\"lcore %d state: %d\\n\", i,\n \t\t\t\t(int) rte_eal_get_lcore_state(i));\n \t}\n@@ -210,19 +210,19 @@ test_mcslock(void)\n \t/*\n \t * Test if it could return immediately from try-locking a locked object.\n \t * Here it will lock the mcs lock object first, then launch all the\n-\t * slave lcores to trylock the same mcs lock object.\n-\t * All the slave lcores should give up try-locking a locked object and\n+\t * worker lcores to trylock the same mcs lock object.\n+\t * All the worker lcores should give up try-locking a locked object and\n \t * return immediately, and then increase the \"count\" initialized with\n \t * zero by one per times.\n \t * We can check if the \"count\" is finally equal to the number of all\n-\t * slave lcores to see if the behavior of try-locking a locked\n+\t * worker lcores to see if the behavior of try-locking a locked\n \t * mcslock object is correct.\n \t */\n \tif (rte_mcslock_trylock(&p_ml_try, &ml_try_me) == 0)\n \t\treturn -1;\n \n \tcount = 0;\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_eal_remote_launch(test_mcslock_try, NULL, i);\n \t}\n \trte_eal_mp_wait_lcore();\ndiff --git a/app/test/test_mempool_perf.c b/app/test/test_mempool_perf.c\nindex 60bda8aadbe8..d7d0aaa3340c 100644\n--- a/app/test/test_mempool_perf.c\n+++ b/app/test/test_mempool_perf.c\n@@ -143,8 +143,8 @@ per_lcore_mempool_test(void *arg)\n \n \tstats[lcore_id].enq_count = 0;\n \n-\t/* wait synchro for slaves */\n-\tif (lcore_id != rte_get_master_lcore())\n+\t/* wait synchro for workers */\n+\tif (lcore_id != rte_get_main_lcore())\n \t\twhile (rte_atomic32_read(&synchro) == 0);\n \n \tstart_cycles = rte_get_timer_cycles();\n@@ -214,7 +214,7 @@ launch_cores(struct rte_mempool *mp, unsigned int cores)\n \t\treturn -1;\n \t}\n \n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (cores == 1)\n \t\t\tbreak;\n \t\tcores--;\n@@ -222,13 +222,13 @@ launch_cores(struct rte_mempool *mp, unsigned int cores)\n \t\t\t\t      mp, lcore_id);\n \t}\n \n-\t/* start synchro and launch test on master */\n+\t/* start synchro and launch test on main */\n \trte_atomic32_set(&synchro, 1);\n \n \tret = per_lcore_mempool_test(mp);\n \n \tcores = cores_save;\n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (cores == 1)\n \t\t\tbreak;\n \t\tcores--;\ndiff --git a/app/test/test_mp_secondary.c b/app/test/test_mp_secondary.c\nindex 3a655c340041..4fa0dfe9c0b8 100644\n--- a/app/test/test_mp_secondary.c\n+++ b/app/test/test_mp_secondary.c\n@@ -94,7 +94,7 @@ run_secondary_instances(void)\n #endif\n \n \tsnprintf(coremask, sizeof(coremask), \"%x\", \\\n-\t\t\t(1 << rte_get_master_lcore()));\n+\t\t\t(1 << rte_get_main_lcore()));\n \n \tret |= launch_proc(argv1);\n \tprintf(\"### Testing rte_mp_disable() reject:\\n\");\ndiff --git a/app/test/test_pdump.c b/app/test/test_pdump.c\nindex 6a1180bcb78e..21fcc1bc4cba 100644\n--- a/app/test/test_pdump.c\n+++ b/app/test/test_pdump.c\n@@ -184,7 +184,7 @@ run_pdump_server_tests(void)\n \t};\n \n \tsnprintf(coremask, sizeof(coremask), \"%x\",\n-\t\t (1 << rte_get_master_lcore()));\n+\t\t (1 << rte_get_main_lcore()));\n \n \tret = test_pdump_init();\n \tret |= launch_p(argv1);\ndiff --git a/app/test/test_per_lcore.c b/app/test/test_per_lcore.c\nindex fcd00212f1eb..129578d1a30f 100644\n--- a/app/test/test_per_lcore.c\n+++ b/app/test/test_per_lcore.c\n@@ -73,31 +73,31 @@ test_per_lcore(void)\n \tunsigned lcore_id;\n \tint ret;\n \n-\trte_eal_mp_remote_launch(assign_vars, NULL, SKIP_MASTER);\n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\trte_eal_mp_remote_launch(assign_vars, NULL, SKIP_MAIN);\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (rte_eal_wait_lcore(lcore_id) < 0)\n \t\t\treturn -1;\n \t}\n \n-\trte_eal_mp_remote_launch(display_vars, NULL, SKIP_MASTER);\n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\trte_eal_mp_remote_launch(display_vars, NULL, SKIP_MAIN);\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (rte_eal_wait_lcore(lcore_id) < 0)\n \t\t\treturn -1;\n \t}\n \n \t/* test if it could do remote launch twice at the same time or not */\n-\tret = rte_eal_mp_remote_launch(test_per_lcore_delay, NULL, SKIP_MASTER);\n+\tret = rte_eal_mp_remote_launch(test_per_lcore_delay, NULL, SKIP_MAIN);\n \tif (ret < 0) {\n \t\tprintf(\"It fails to do remote launch but it should able to do\\n\");\n \t\treturn -1;\n \t}\n \t/* it should not be able to launch a lcore which is running */\n-\tret = rte_eal_mp_remote_launch(test_per_lcore_delay, NULL, SKIP_MASTER);\n+\tret = rte_eal_mp_remote_launch(test_per_lcore_delay, NULL, SKIP_MAIN);\n \tif (ret == 0) {\n \t\tprintf(\"It does remote launch successfully but it should not at this time\\n\");\n \t\treturn -1;\n \t}\n-\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\tif (rte_eal_wait_lcore(lcore_id) < 0)\n \t\t\treturn -1;\n \t}\ndiff --git a/app/test/test_pmd_perf.c b/app/test/test_pmd_perf.c\nindex d1240b76f955..890c2d7a3055 100644\n--- a/app/test/test_pmd_perf.c\n+++ b/app/test/test_pmd_perf.c\n@@ -275,7 +275,7 @@ alloc_lcore(uint16_t socketid)\n \tfor (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {\n \t\tif (LCORE_AVAIL != lcore_conf[lcore_id].status ||\n \t\t    lcore_conf[lcore_id].socketid != socketid ||\n-\t\t    lcore_id == rte_get_master_lcore())\n+\t\t    lcore_id == rte_get_main_lcore())\n \t\t\tcontinue;\n \t\tlcore_conf[lcore_id].status = LCORE_USED;\n \t\tlcore_conf[lcore_id].nb_ports = 0;\n@@ -661,7 +661,7 @@ exec_burst(uint32_t flags, int lcore)\n static int\n test_pmd_perf(void)\n {\n-\tuint16_t nb_ports, num, nb_lcores, slave_id = (uint16_t)-1;\n+\tuint16_t nb_ports, num, nb_lcores, worker_id = (uint16_t)-1;\n \tuint16_t nb_rxd = MAX_TRAFFIC_BURST;\n \tuint16_t nb_txd = MAX_TRAFFIC_BURST;\n \tuint16_t portid;\n@@ -699,13 +699,13 @@ test_pmd_perf(void)\n \tRTE_ETH_FOREACH_DEV(portid) {\n \t\tif (socketid == -1) {\n \t\t\tsocketid = rte_eth_dev_socket_id(portid);\n-\t\t\tslave_id = alloc_lcore(socketid);\n-\t\t\tif (slave_id == (uint16_t)-1) {\n+\t\t\tworker_id = alloc_lcore(socketid);\n+\t\t\tif (worker_id == (uint16_t)-1) {\n \t\t\t\tprintf(\"No avail lcore to run test\\n\");\n \t\t\t\treturn -1;\n \t\t\t}\n \t\t\tprintf(\"Performance test runs on lcore %u socket %u\\n\",\n-\t\t\t       slave_id, socketid);\n+\t\t\t       worker_id, socketid);\n \t\t}\n \n \t\tif (socketid != rte_eth_dev_socket_id(portid)) {\n@@ -762,8 +762,8 @@ test_pmd_perf(void)\n \t\t\t\t \"rte_eth_promiscuous_enable: err=%s, port=%d\\n\",\n \t\t\t\t rte_strerror(-ret), portid);\n \n-\t\tlcore_conf[slave_id].portlist[num++] = portid;\n-\t\tlcore_conf[slave_id].nb_ports++;\n+\t\tlcore_conf[worker_id].portlist[num++] = portid;\n+\t\tlcore_conf[worker_id].nb_ports++;\n \t}\n \tcheck_all_ports_link_status(nb_ports, RTE_PORT_ALL);\n \n@@ -788,13 +788,13 @@ test_pmd_perf(void)\n \t\tif (NULL == do_measure)\n \t\t\tdo_measure = measure_rxtx;\n \n-\t\trte_eal_remote_launch(main_loop, NULL, slave_id);\n+\t\trte_eal_remote_launch(main_loop, NULL, worker_id);\n \n-\t\tif (rte_eal_wait_lcore(slave_id) < 0)\n+\t\tif (rte_eal_wait_lcore(worker_id) < 0)\n \t\t\treturn -1;\n \t} else if (sc_flag == SC_BURST_POLL_FIRST ||\n \t\t   sc_flag == SC_BURST_XMIT_FIRST)\n-\t\tif (exec_burst(sc_flag, slave_id) < 0)\n+\t\tif (exec_burst(sc_flag, worker_id) < 0)\n \t\t\treturn -1;\n \n \t/* port tear down */\ndiff --git a/app/test/test_rcu_qsbr.c b/app/test/test_rcu_qsbr.c\nindex 0a9e5ecd1a44..7ae66e4dfb76 100644\n--- a/app/test/test_rcu_qsbr.c\n+++ b/app/test/test_rcu_qsbr.c\n@@ -1327,7 +1327,7 @@ test_rcu_qsbr_main(void)\n \t}\n \n \tnum_cores = 0;\n-\tRTE_LCORE_FOREACH_SLAVE(core_id) {\n+\tRTE_LCORE_FOREACH_WORKER(core_id) {\n \t\tenabled_core_ids[num_cores] = core_id;\n \t\tnum_cores++;\n \t}\ndiff --git a/app/test/test_rcu_qsbr_perf.c b/app/test/test_rcu_qsbr_perf.c\nindex d35a6d089784..3017e71120ad 100644\n--- a/app/test/test_rcu_qsbr_perf.c\n+++ b/app/test/test_rcu_qsbr_perf.c\n@@ -625,7 +625,7 @@ test_rcu_qsbr_main(void)\n \trte_atomic64_init(&check_cycles);\n \n \tnum_cores = 0;\n-\tRTE_LCORE_FOREACH_SLAVE(core_id) {\n+\tRTE_LCORE_FOREACH_WORKER(core_id) {\n \t\tenabled_core_ids[num_cores] = core_id;\n \t\tnum_cores++;\n \t}\ndiff --git a/app/test/test_ring_perf.c b/app/test/test_ring_perf.c\nindex ac9bf5608daa..e63e25a867f0 100644\n--- a/app/test/test_ring_perf.c\n+++ b/app/test/test_ring_perf.c\n@@ -297,7 +297,7 @@ run_on_core_pair(struct lcore_pair *cores, struct rte_ring *r, const int esize)\n \t\tlcore_count = 0;\n \t\tparam1.size = param2.size = bulk_sizes[i];\n \t\tparam1.r = param2.r = r;\n-\t\tif (cores->c1 == rte_get_master_lcore()) {\n+\t\tif (cores->c1 == rte_get_main_lcore()) {\n \t\t\trte_eal_remote_launch(f2, &param2, cores->c2);\n \t\t\tf1(&param1);\n \t\t\trte_eal_wait_lcore(cores->c2);\n@@ -340,8 +340,8 @@ load_loop_fn_helper(struct thread_params *p, const int esize)\n \tif (burst == NULL)\n \t\treturn -1;\n \n-\t/* wait synchro for slaves */\n-\tif (lcore != rte_get_master_lcore())\n+\t/* wait synchro for workers */\n+\tif (lcore != rte_get_main_lcore())\n \t\twhile (rte_atomic32_read(&synchro) == 0)\n \t\t\trte_pause();\n \n@@ -397,12 +397,12 @@ run_on_all_cores(struct rte_ring *r, const int esize)\n \t\tparam.size = bulk_sizes[i];\n \t\tparam.r = r;\n \n-\t\t/* clear synchro and start slaves */\n+\t\t/* clear synchro and start workers */\n \t\trte_atomic32_set(&synchro, 0);\n-\t\tif (rte_eal_mp_remote_launch(lcore_f, &param, SKIP_MASTER) < 0)\n+\t\tif (rte_eal_mp_remote_launch(lcore_f, &param, SKIP_MAIN) < 0)\n \t\t\treturn -1;\n \n-\t\t/* start synchro and launch test on master */\n+\t\t/* start synchro and launch test on main */\n \t\trte_atomic32_set(&synchro, 1);\n \t\tlcore_f(&param);\n \n@@ -553,7 +553,7 @@ test_ring_perf_esize(const int esize)\n \t\t\tgoto test_fail;\n \t}\n \n-\tprintf(\"\\n### Testing using all slave nodes ###\\n\");\n+\tprintf(\"\\n### Testing using all worker nodes ###\\n\");\n \tif (run_on_all_cores(r, esize) < 0)\n \t\tgoto test_fail;\n \ndiff --git a/app/test/test_ring_stress_impl.h b/app/test/test_ring_stress_impl.h\nindex 222d62bc4f4d..3b9a480eb9cc 100644\n--- a/app/test/test_ring_stress_impl.h\n+++ b/app/test/test_ring_stress_impl.h\n@@ -6,7 +6,7 @@\n \n /**\n  * Stress test for ring enqueue/dequeue operations.\n- * Performs the following pattern on each slave worker:\n+ * Performs the following pattern on each worker:\n  * dequeue/read-write data from the dequeued objects/enqueue.\n  * Serves as both functional and performance test of ring\n  * enqueue/dequeue operations under high contention\n@@ -348,8 +348,8 @@ test_mt1(int (*test)(void *))\n \n \tmemset(arg, 0, sizeof(arg));\n \n-\t/* launch on all slaves */\n-\tRTE_LCORE_FOREACH_SLAVE(lc) {\n+\t/* launch on all workers */\n+\tRTE_LCORE_FOREACH_WORKER(lc) {\n \t\targ[lc].rng = r;\n \t\targ[lc].stats = init_stat;\n \t\trte_eal_remote_launch(test, &arg[lc], lc);\n@@ -365,12 +365,12 @@ test_mt1(int (*test)(void *))\n \twrk_cmd = WRK_CMD_STOP;\n \trte_smp_wmb();\n \n-\t/* wait for slaves and collect stats. */\n+\t/* wait for workers and collect stats. */\n \tmc = rte_lcore_id();\n \targ[mc].stats = init_stat;\n \n \trc = 0;\n-\tRTE_LCORE_FOREACH_SLAVE(lc) {\n+\tRTE_LCORE_FOREACH_WORKER(lc) {\n \t\trc |= rte_eal_wait_lcore(lc);\n \t\tlcore_stat_aggr(&arg[mc].stats, &arg[lc].stats);\n \t\tif (verbose != 0)\ndiff --git a/app/test/test_rwlock.c b/app/test/test_rwlock.c\nindex 61bee7d7c296..701187f39884 100644\n--- a/app/test/test_rwlock.c\n+++ b/app/test/test_rwlock.c\n@@ -99,8 +99,8 @@ load_loop_fn(__rte_unused void *arg)\n \tuint64_t lcount = 0;\n \tconst unsigned int lcore = rte_lcore_id();\n \n-\t/* wait synchro for slaves */\n-\tif (lcore != rte_get_master_lcore())\n+\t/* wait synchro for workers */\n+\tif (lcore != rte_get_main_lcore())\n \t\twhile (rte_atomic32_read(&synchro) == 0)\n \t\t\t;\n \n@@ -134,12 +134,12 @@ test_rwlock_perf(void)\n \n \tprintf(\"\\nRwlock Perf Test on %u cores...\\n\", rte_lcore_count());\n \n-\t/* clear synchro and start slaves */\n+\t/* clear synchro and start workers */\n \trte_atomic32_set(&synchro, 0);\n-\tif (rte_eal_mp_remote_launch(load_loop_fn, NULL, SKIP_MASTER) < 0)\n+\tif (rte_eal_mp_remote_launch(load_loop_fn, NULL, SKIP_MAIN) < 0)\n \t\treturn -1;\n \n-\t/* start synchro and launch test on master */\n+\t/* start synchro and launch test on main */\n \trte_atomic32_set(&synchro, 1);\n \tload_loop_fn(NULL);\n \n@@ -161,7 +161,7 @@ test_rwlock_perf(void)\n  * - There is a global rwlock and a table of rwlocks (one per lcore).\n  *\n  * - The test function takes all of these locks and launches the\n- *   ``test_rwlock_per_core()`` function on each core (except the master).\n+ *   ``test_rwlock_per_core()`` function on each core (except the main).\n  *\n  *   - The function takes the global write lock, display something,\n  *     then releases the global lock.\n@@ -187,21 +187,21 @@ rwlock_test1(void)\n \n \trte_rwlock_write_lock(&sl);\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_rwlock_write_lock(&sl_tab[i]);\n \t\trte_eal_remote_launch(test_rwlock_per_core, NULL, i);\n \t}\n \n \trte_rwlock_write_unlock(&sl);\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_rwlock_write_unlock(&sl_tab[i]);\n \t\trte_delay_ms(100);\n \t}\n \n \trte_rwlock_write_lock(&sl);\n \t/* this message should be the last message of test */\n-\tprintf(\"Global write lock taken on master core %u\\n\", rte_lcore_id());\n+\tprintf(\"Global write lock taken on main core %u\\n\", rte_lcore_id());\n \trte_rwlock_write_unlock(&sl);\n \n \trte_eal_mp_wait_lcore();\n@@ -462,26 +462,26 @@ try_rwlock_test_rda(void)\n \ttry_test_reset();\n \n \t/* start read test on all avaialble lcores */\n-\trte_eal_mp_remote_launch(try_read_lcore, NULL, CALL_MASTER);\n+\trte_eal_mp_remote_launch(try_read_lcore, NULL, CALL_MAIN);\n \trte_eal_mp_wait_lcore();\n \n \treturn process_try_lcore_stats();\n }\n \n-/* all slave lcores grab RDLOCK, master one grabs WRLOCK */\n+/* all worker lcores grab RDLOCK, main one grabs WRLOCK */\n static int\n try_rwlock_test_rds_wrm(void)\n {\n \ttry_test_reset();\n \n-\trte_eal_mp_remote_launch(try_read_lcore, NULL, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(try_read_lcore, NULL, SKIP_MAIN);\n \ttry_write_lcore(NULL);\n \trte_eal_mp_wait_lcore();\n \n \treturn process_try_lcore_stats();\n }\n \n-/* master and even slave lcores grab RDLOCK, odd lcores grab WRLOCK */\n+/* main and even worker lcores grab RDLOCK, odd lcores grab WRLOCK */\n static int\n try_rwlock_test_rde_wro(void)\n {\n@@ -489,7 +489,7 @@ try_rwlock_test_rde_wro(void)\n \n \ttry_test_reset();\n \n-\tmlc = rte_get_master_lcore();\n+\tmlc = rte_get_main_lcore();\n \n \tRTE_LCORE_FOREACH(lc) {\n \t\tif (lc != mlc) {\ndiff --git a/app/test/test_service_cores.c b/app/test/test_service_cores.c\nindex 5d92bea8af79..b387b468b75a 100644\n--- a/app/test/test_service_cores.c\n+++ b/app/test/test_service_cores.c\n@@ -30,7 +30,7 @@ static int\n testsuite_setup(void)\n {\n \tslcore_id = rte_get_next_lcore(/* start core */ -1,\n-\t\t\t\t       /* skip master */ 1,\n+\t\t\t\t       /* skip main */ 1,\n \t\t\t\t       /* wrap */ 0);\n \n \treturn TEST_SUCCESS;\n@@ -552,12 +552,12 @@ service_lcore_add_del(void)\n \tTEST_ASSERT_EQUAL(1, rte_service_lcore_count(),\n \t\t\t\"Service core count not equal to one\");\n \tuint32_t slcore_1 = rte_get_next_lcore(/* start core */ -1,\n-\t\t\t\t\t       /* skip master */ 1,\n+\t\t\t\t\t       /* skip main */ 1,\n \t\t\t\t\t       /* wrap */ 0);\n \tTEST_ASSERT_EQUAL(0, rte_service_lcore_add(slcore_1),\n \t\t\t\"Service core add did not return zero\");\n \tuint32_t slcore_2 = rte_get_next_lcore(/* start core */ slcore_1,\n-\t\t\t\t\t       /* skip master */ 1,\n+\t\t\t\t\t       /* skip main */ 1,\n \t\t\t\t\t       /* wrap */ 0);\n \tTEST_ASSERT_EQUAL(0, rte_service_lcore_add(slcore_2),\n \t\t\t\"Service core add did not return zero\");\n@@ -603,12 +603,12 @@ service_threaded_test(int mt_safe)\n \n \t/* add next 2 cores */\n \tuint32_t slcore_1 = rte_get_next_lcore(/* start core */ -1,\n-\t\t\t\t\t       /* skip master */ 1,\n+\t\t\t\t\t       /* skip main */ 1,\n \t\t\t\t\t       /* wrap */ 0);\n \tTEST_ASSERT_EQUAL(0, rte_service_lcore_add(slcore_1),\n \t\t\t\"mt safe lcore add fail\");\n \tuint32_t slcore_2 = rte_get_next_lcore(/* start core */ slcore_1,\n-\t\t\t\t\t       /* skip master */ 1,\n+\t\t\t\t\t       /* skip main */ 1,\n \t\t\t\t\t       /* wrap */ 0);\n \tTEST_ASSERT_EQUAL(0, rte_service_lcore_add(slcore_2),\n \t\t\t\"mt safe lcore add fail\");\n@@ -942,10 +942,10 @@ service_active_two_cores(void)\n \tint i;\n \n \tuint32_t lcore = rte_get_next_lcore(/* start core */ -1,\n-\t\t\t\t\t    /* skip master */ 1,\n+\t\t\t\t\t    /* skip main */ 1,\n \t\t\t\t\t    /* wrap */ 0);\n \tuint32_t slcore = rte_get_next_lcore(/* start core */ lcore,\n-\t\t\t\t\t     /* skip master */ 1,\n+\t\t\t\t\t     /* skip main */ 1,\n \t\t\t\t\t     /* wrap */ 0);\n \n \t/* start the service on the second available lcore */\ndiff --git a/app/test/test_spinlock.c b/app/test/test_spinlock.c\nindex 842990ed3b30..054fb43a9fe5 100644\n--- a/app/test/test_spinlock.c\n+++ b/app/test/test_spinlock.c\n@@ -28,7 +28,7 @@\n  * - There is a global spinlock and a table of spinlocks (one per lcore).\n  *\n  * - The test function takes all of these locks and launches the\n- *   ``test_spinlock_per_core()`` function on each core (except the master).\n+ *   ``test_spinlock_per_core()`` function on each core (except the main).\n  *\n  *   - The function takes the global lock, display something, then releases\n  *     the global lock.\n@@ -109,8 +109,8 @@ load_loop_fn(void *func_param)\n \tconst int use_lock = *(int*)func_param;\n \tconst unsigned lcore = rte_lcore_id();\n \n-\t/* wait synchro for slaves */\n-\tif (lcore != rte_get_master_lcore())\n+\t/* wait synchro for workers */\n+\tif (lcore != rte_get_main_lcore())\n \t\twhile (rte_atomic32_read(&synchro) == 0);\n \n \tbegin = rte_get_timer_cycles();\n@@ -149,11 +149,11 @@ test_spinlock_perf(void)\n \n \tprintf(\"\\nTest with lock on %u cores...\\n\", rte_lcore_count());\n \n-\t/* Clear synchro and start slaves */\n+\t/* Clear synchro and start workers */\n \trte_atomic32_set(&synchro, 0);\n-\trte_eal_mp_remote_launch(load_loop_fn, &lock, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(load_loop_fn, &lock, SKIP_MAIN);\n \n-\t/* start synchro and launch test on master */\n+\t/* start synchro and launch test on main */\n \trte_atomic32_set(&synchro, 1);\n \tload_loop_fn(&lock);\n \n@@ -200,8 +200,8 @@ test_spinlock(void)\n \tint ret = 0;\n \tint i;\n \n-\t/* slave cores should be waiting: print it */\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\t/* worker cores should be waiting: print it */\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\tprintf(\"lcore %d state: %d\\n\", i,\n \t\t       (int) rte_eal_get_lcore_state(i));\n \t}\n@@ -214,19 +214,19 @@ test_spinlock(void)\n \n \trte_spinlock_lock(&sl);\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_spinlock_lock(&sl_tab[i]);\n \t\trte_eal_remote_launch(test_spinlock_per_core, NULL, i);\n \t}\n \n-\t/* slave cores should be busy: print it */\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\t/* worker cores should be busy: print it */\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\tprintf(\"lcore %d state: %d\\n\", i,\n \t\t       (int) rte_eal_get_lcore_state(i));\n \t}\n \trte_spinlock_unlock(&sl);\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_spinlock_unlock(&sl_tab[i]);\n \t\trte_delay_ms(10);\n \t}\n@@ -245,7 +245,7 @@ test_spinlock(void)\n \t} else\n \t\trte_spinlock_recursive_unlock(&slr);\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_eal_remote_launch(test_spinlock_recursive_per_core, NULL, i);\n \t}\n \trte_spinlock_recursive_unlock(&slr);\n@@ -253,12 +253,12 @@ test_spinlock(void)\n \n \t/*\n \t * Test if it could return immediately from try-locking a locked object.\n-\t * Here it will lock the spinlock object first, then launch all the slave\n+\t * Here it will lock the spinlock object first, then launch all the worker\n \t * lcores to trylock the same spinlock object.\n-\t * All the slave lcores should give up try-locking a locked object and\n+\t * All the worker lcores should give up try-locking a locked object and\n \t * return immediately, and then increase the \"count\" initialized with zero\n \t * by one per times.\n-\t * We can check if the \"count\" is finally equal to the number of all slave\n+\t * We can check if the \"count\" is finally equal to the number of all worker\n \t * lcores to see if the behavior of try-locking a locked spinlock object\n \t * is correct.\n \t */\n@@ -266,7 +266,7 @@ test_spinlock(void)\n \t\treturn -1;\n \t}\n \tcount = 0;\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_eal_remote_launch(test_spinlock_try, NULL, i);\n \t}\n \trte_eal_mp_wait_lcore();\ndiff --git a/app/test/test_stack.c b/app/test/test_stack.c\nindex 463460ccc65f..02422a32d64f 100644\n--- a/app/test/test_stack.c\n+++ b/app/test/test_stack.c\n@@ -328,7 +328,7 @@ test_stack_multithreaded(uint32_t flags)\n \n \tthread_test_args.s = s;\n \n-\tif (rte_eal_mp_remote_launch(stack_thread_push_pop, NULL, CALL_MASTER))\n+\tif (rte_eal_mp_remote_launch(stack_thread_push_pop, NULL, CALL_MAIN))\n \t\trte_panic(\"Failed to launch tests\\n\");\n \n \tRTE_LCORE_FOREACH(lcore_id) {\ndiff --git a/app/test/test_stack_perf.c b/app/test/test_stack_perf.c\nindex 3ab7267b1b72..3590625c49e7 100644\n--- a/app/test/test_stack_perf.c\n+++ b/app/test/test_stack_perf.c\n@@ -180,7 +180,7 @@ run_on_core_pair(struct lcore_pair *cores, struct rte_stack *s,\n \t\targs[0].sz = args[1].sz = bulk_sizes[i];\n \t\targs[0].s = args[1].s = s;\n \n-\t\tif (cores->c1 == rte_get_master_lcore()) {\n+\t\tif (cores->c1 == rte_get_main_lcore()) {\n \t\t\trte_eal_remote_launch(fn, &args[1], cores->c2);\n \t\t\tfn(&args[0]);\n \t\t\trte_eal_wait_lcore(cores->c2);\n@@ -210,7 +210,7 @@ run_on_n_cores(struct rte_stack *s, lcore_function_t fn, int n)\n \n \t\trte_atomic32_set(&lcore_barrier, n);\n \n-\t\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\t\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\t\tif (++cnt >= n)\n \t\t\t\tbreak;\n \n@@ -235,7 +235,7 @@ run_on_n_cores(struct rte_stack *s, lcore_function_t fn, int n)\n \t\tavg = args[rte_lcore_id()].avg;\n \n \t\tcnt = 0;\n-\t\tRTE_LCORE_FOREACH_SLAVE(lcore_id) {\n+\t\tRTE_LCORE_FOREACH_WORKER(lcore_id) {\n \t\t\tif (++cnt >= n)\n \t\t\t\tbreak;\n \t\t\tavg += args[lcore_id].avg;\ndiff --git a/app/test/test_ticketlock.c b/app/test/test_ticketlock.c\nindex 66ab3d1a0248..7aab8665bc77 100644\n--- a/app/test/test_ticketlock.c\n+++ b/app/test/test_ticketlock.c\n@@ -28,7 +28,7 @@\n  * - There is a global ticketlock and a table of ticketlocks (one per lcore).\n  *\n  * - The test function takes all of these locks and launches the\n- *   ``test_ticketlock_per_core()`` function on each core (except the master).\n+ *   ``test_ticketlock_per_core()`` function on each core (except the main).\n  *\n  *   - The function takes the global lock, display something, then releases\n  *     the global lock.\n@@ -110,8 +110,8 @@ load_loop_fn(void *func_param)\n \tconst int use_lock = *(int *)func_param;\n \tconst unsigned int lcore = rte_lcore_id();\n \n-\t/* wait synchro for slaves */\n-\tif (lcore != rte_get_master_lcore())\n+\t/* wait synchro for workers */\n+\tif (lcore != rte_get_main_lcore())\n \t\twhile (rte_atomic32_read(&synchro) == 0)\n \t\t\t;\n \n@@ -154,11 +154,11 @@ test_ticketlock_perf(void)\n \tlcount = 0;\n \tprintf(\"\\nTest with lock on %u cores...\\n\", rte_lcore_count());\n \n-\t/* Clear synchro and start slaves */\n+\t/* Clear synchro and start workers */\n \trte_atomic32_set(&synchro, 0);\n-\trte_eal_mp_remote_launch(load_loop_fn, &lock, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(load_loop_fn, &lock, SKIP_MAIN);\n \n-\t/* start synchro and launch test on master */\n+\t/* start synchro and launch test on main */\n \trte_atomic32_set(&synchro, 1);\n \tload_loop_fn(&lock);\n \n@@ -208,8 +208,8 @@ test_ticketlock(void)\n \tint ret = 0;\n \tint i;\n \n-\t/* slave cores should be waiting: print it */\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\t/* worker cores should be waiting: print it */\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\tprintf(\"lcore %d state: %d\\n\", i,\n \t\t       (int) rte_eal_get_lcore_state(i));\n \t}\n@@ -217,25 +217,25 @@ test_ticketlock(void)\n \trte_ticketlock_init(&tl);\n \trte_ticketlock_init(&tl_try);\n \trte_ticketlock_recursive_init(&tlr);\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_ticketlock_init(&tl_tab[i]);\n \t}\n \n \trte_ticketlock_lock(&tl);\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_ticketlock_lock(&tl_tab[i]);\n \t\trte_eal_remote_launch(test_ticketlock_per_core, NULL, i);\n \t}\n \n-\t/* slave cores should be busy: print it */\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\t/* worker cores should be busy: print it */\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\tprintf(\"lcore %d state: %d\\n\", i,\n \t\t       (int) rte_eal_get_lcore_state(i));\n \t}\n \trte_ticketlock_unlock(&tl);\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_ticketlock_unlock(&tl_tab[i]);\n \t\trte_delay_ms(10);\n \t}\n@@ -254,7 +254,7 @@ test_ticketlock(void)\n \t} else\n \t\trte_ticketlock_recursive_unlock(&tlr);\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_eal_remote_launch(test_ticketlock_recursive_per_core,\n \t\t\t\t\tNULL, i);\n \t}\n@@ -264,19 +264,19 @@ test_ticketlock(void)\n \t/*\n \t * Test if it could return immediately from try-locking a locked object.\n \t * Here it will lock the ticketlock object first, then launch all the\n-\t * slave lcores to trylock the same ticketlock object.\n-\t * All the slave lcores should give up try-locking a locked object and\n+\t * worker lcores to trylock the same ticketlock object.\n+\t * All the worker lcores should give up try-locking a locked object and\n \t * return immediately, and then increase the \"count\" initialized with\n \t * zero by one per times.\n \t * We can check if the \"count\" is finally equal to the number of all\n-\t * slave lcores to see if the behavior of try-locking a locked\n+\t * worker lcores to see if the behavior of try-locking a locked\n \t * ticketlock object is correct.\n \t */\n \tif (rte_ticketlock_trylock(&tl_try) == 0)\n \t\treturn -1;\n \n \tcount = 0;\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n \t\trte_eal_remote_launch(test_ticketlock_try, NULL, i);\n \t}\n \trte_eal_mp_wait_lcore();\ndiff --git a/app/test/test_timer.c b/app/test/test_timer.c\nindex 5933f56ed544..a10b2fe9daf4 100644\n--- a/app/test/test_timer.c\n+++ b/app/test/test_timer.c\n@@ -37,7 +37,7 @@\n  *    - All cores then simultaneously are set to schedule all the timers at\n  *      the same time, so conflicts should occur.\n  *    - Then there is a delay while we wait for the timers to expire\n- *    - Then the master lcore calls timer_manage() and we check that all\n+ *    - Then the main lcore calls timer_manage() and we check that all\n  *      timers have had their callbacks called exactly once - no more no less.\n  *    - Then we repeat the process, except after setting up the timers, we have\n  *      all cores randomly reschedule them.\n@@ -58,7 +58,7 @@\n  *\n  *    - timer0\n  *\n- *      - At initialization, timer0 is loaded by the master core, on master core\n+ *      - At initialization, timer0 is loaded by the main core, on main core\n  *        in \"single\" mode (time = 1 second).\n  *      - In the first 19 callbacks, timer0 is reloaded on the same core,\n  *        then, it is explicitly stopped at the 20th call.\n@@ -66,21 +66,21 @@\n  *\n  *    - timer1\n  *\n- *      - At initialization, timer1 is loaded by the master core, on the\n- *        master core in \"single\" mode (time = 2 seconds).\n+ *      - At initialization, timer1 is loaded by the main core, on the\n+ *        main core in \"single\" mode (time = 2 seconds).\n  *      - In the first 9 callbacks, timer1 is reloaded on another\n  *        core. After the 10th callback, timer1 is not reloaded anymore.\n  *\n  *    - timer2\n  *\n- *      - At initialization, timer2 is loaded by the master core, on the\n- *        master core in \"periodical\" mode (time = 1 second).\n+ *      - At initialization, timer2 is loaded by the main core, on the\n+ *        main core in \"periodical\" mode (time = 1 second).\n  *      - In the callback, when t=25s, it stops timer3 and reloads timer0\n  *        on the current core.\n  *\n  *    - timer3\n  *\n- *      - At initialization, timer3 is loaded by the master core, on\n+ *      - At initialization, timer3 is loaded by the main core, on\n  *        another core in \"periodical\" mode (time = 1 second).\n  *      - It is stopped at t=25s by timer2.\n  */\n@@ -201,68 +201,69 @@ timer_stress_main_loop(__rte_unused void *arg)\n \treturn 0;\n }\n \n-/* Need to synchronize slave lcores through multiple steps. */\n-enum { SLAVE_WAITING = 1, SLAVE_RUN_SIGNAL, SLAVE_RUNNING, SLAVE_FINISHED };\n-static rte_atomic16_t slave_state[RTE_MAX_LCORE];\n+/* Need to synchronize worker lcores through multiple steps. */\n+enum { WORKER_WAITING = 1, WORKER_RUN_SIGNAL, WORKER_RUNNING, WORKER_FINISHED };\n+static rte_atomic16_t lcore_state[RTE_MAX_LCORE];\n \n static void\n-master_init_slaves(void)\n+main_init_workers(void)\n {\n \tunsigned i;\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n-\t\trte_atomic16_set(&slave_state[i], SLAVE_WAITING);\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n+\t\trte_atomic16_set(&lcore_state[i], WORKER_WAITING);\n \t}\n }\n \n static void\n-master_start_slaves(void)\n+main_start_workers(void)\n {\n \tunsigned i;\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n-\t\trte_atomic16_set(&slave_state[i], SLAVE_RUN_SIGNAL);\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n+\t\trte_atomic16_set(&lcore_state[i], WORKER_RUN_SIGNAL);\n \t}\n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n-\t\twhile (rte_atomic16_read(&slave_state[i]) != SLAVE_RUNNING)\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n+\t\twhile (rte_atomic16_read(&lcore_state[i]) != WORKER_RUNNING)\n \t\t\trte_pause();\n \t}\n }\n \n static void\n-master_wait_for_slaves(void)\n+main_wait_for_workers(void)\n {\n \tunsigned i;\n \n-\tRTE_LCORE_FOREACH_SLAVE(i) {\n-\t\twhile (rte_atomic16_read(&slave_state[i]) != SLAVE_FINISHED)\n+\tRTE_LCORE_FOREACH_WORKER(i) {\n+\t\twhile (rte_atomic16_read(&lcore_state[i]) != WORKER_FINISHED)\n \t\t\trte_pause();\n \t}\n }\n \n static void\n-slave_wait_to_start(void)\n+worker_wait_to_start(void)\n {\n \tunsigned lcore_id = rte_lcore_id();\n \n-\twhile (rte_atomic16_read(&slave_state[lcore_id]) != SLAVE_RUN_SIGNAL)\n+\twhile (rte_atomic16_read(&lcore_state[lcore_id]) != WORKER_RUN_SIGNAL)\n \t\trte_pause();\n-\trte_atomic16_set(&slave_state[lcore_id], SLAVE_RUNNING);\n+\trte_atomic16_set(&lcore_state[lcore_id], WORKER_RUNNING);\n }\n \n static void\n-slave_finish(void)\n+worker_finish(void)\n {\n \tunsigned lcore_id = rte_lcore_id();\n \n-\trte_atomic16_set(&slave_state[lcore_id], SLAVE_FINISHED);\n+\trte_atomic16_set(&lcore_state[lcore_id], WORKER_FINISHED);\n }\n \n \n static volatile int cb_count = 0;\n \n /* callback for second stress test. will only be called\n- * on master lcore */\n+ * on main lcore\n+ */\n static void\n timer_stress2_cb(struct rte_timer *tim __rte_unused, void *arg __rte_unused)\n {\n@@ -277,36 +278,36 @@ timer_stress2_main_loop(__rte_unused void *arg)\n \tstatic struct rte_timer *timers;\n \tint i, ret;\n \tuint64_t delay = rte_get_timer_hz() / 20;\n-\tunsigned lcore_id = rte_lcore_id();\n-\tunsigned master = rte_get_master_lcore();\n+\tunsigned int lcore_id = rte_lcore_id();\n+\tunsigned int main_lcore = rte_get_main_lcore();\n \tint32_t my_collisions = 0;\n \tstatic rte_atomic32_t collisions;\n \n-\tif (lcore_id == master) {\n+\tif (lcore_id == main_lcore) {\n \t\tcb_count = 0;\n \t\ttest_failed = 0;\n \t\trte_atomic32_set(&collisions, 0);\n-\t\tmaster_init_slaves();\n+\t\tmain_init_workers();\n \t\ttimers = rte_malloc(NULL, sizeof(*timers) * NB_STRESS2_TIMERS, 0);\n \t\tif (timers == NULL) {\n \t\t\tprintf(\"Test Failed\\n\");\n \t\t\tprintf(\"- Cannot allocate memory for timers\\n\" );\n \t\t\ttest_failed = 1;\n-\t\t\tmaster_start_slaves();\n+\t\t\tmain_start_workers();\n \t\t\tgoto cleanup;\n \t\t}\n \t\tfor (i = 0; i < NB_STRESS2_TIMERS; i++)\n \t\t\trte_timer_init(&timers[i]);\n-\t\tmaster_start_slaves();\n+\t\tmain_start_workers();\n \t} else {\n-\t\tslave_wait_to_start();\n+\t\tworker_wait_to_start();\n \t\tif (test_failed)\n \t\t\tgoto cleanup;\n \t}\n \n-\t/* have all cores schedule all timers on master lcore */\n+\t/* have all cores schedule all timers on main lcore */\n \tfor (i = 0; i < NB_STRESS2_TIMERS; i++) {\n-\t\tret = rte_timer_reset(&timers[i], delay, SINGLE, master,\n+\t\tret = rte_timer_reset(&timers[i], delay, SINGLE, main_lcore,\n \t\t\t\ttimer_stress2_cb, NULL);\n \t\t/* there will be collisions when multiple cores simultaneously\n \t\t * configure the same timers */\n@@ -320,14 +321,14 @@ timer_stress2_main_loop(__rte_unused void *arg)\n \trte_delay_ms(100);\n \n \t/* all cores rendezvous */\n-\tif (lcore_id == master) {\n-\t\tmaster_wait_for_slaves();\n+\tif (lcore_id == main_lcore) {\n+\t\tmain_wait_for_workers();\n \t} else {\n-\t\tslave_finish();\n+\t\tworker_finish();\n \t}\n \n \t/* now check that we get the right number of callbacks */\n-\tif (lcore_id == master) {\n+\tif (lcore_id == main_lcore) {\n \t\tmy_collisions = rte_atomic32_read(&collisions);\n \t\tif (my_collisions != 0)\n \t\t\tprintf(\"- %d timer reset collisions (OK)\\n\", my_collisions);\n@@ -338,23 +339,23 @@ timer_stress2_main_loop(__rte_unused void *arg)\n \t\t\tprintf(\"- Expected %d callbacks, got %d\\n\", NB_STRESS2_TIMERS,\n \t\t\t\t\tcb_count);\n \t\t\ttest_failed = 1;\n-\t\t\tmaster_start_slaves();\n+\t\t\tmain_start_workers();\n \t\t\tgoto cleanup;\n \t\t}\n \t\tcb_count = 0;\n \n \t\t/* proceed */\n-\t\tmaster_start_slaves();\n+\t\tmain_start_workers();\n \t} else {\n \t\t/* proceed */\n-\t\tslave_wait_to_start();\n+\t\tworker_wait_to_start();\n \t\tif (test_failed)\n \t\t\tgoto cleanup;\n \t}\n \n \t/* now test again, just stop and restart timers at random after init*/\n \tfor (i = 0; i < NB_STRESS2_TIMERS; i++)\n-\t\trte_timer_reset(&timers[i], delay, SINGLE, master,\n+\t\trte_timer_reset(&timers[i], delay, SINGLE, main_lcore,\n \t\t\t\ttimer_stress2_cb, NULL);\n \n \t/* pick random timer to reset, stopping them first half the time */\n@@ -362,7 +363,7 @@ timer_stress2_main_loop(__rte_unused void *arg)\n \t\tint r = rand() % NB_STRESS2_TIMERS;\n \t\tif (i % 2)\n \t\t\trte_timer_stop(&timers[r]);\n-\t\trte_timer_reset(&timers[r], delay, SINGLE, master,\n+\t\trte_timer_reset(&timers[r], delay, SINGLE, main_lcore,\n \t\t\t\ttimer_stress2_cb, NULL);\n \t}\n \n@@ -370,8 +371,8 @@ timer_stress2_main_loop(__rte_unused void *arg)\n \trte_delay_ms(100);\n \n \t/* now check that we get the right number of callbacks */\n-\tif (lcore_id == master) {\n-\t\tmaster_wait_for_slaves();\n+\tif (lcore_id == main_lcore) {\n+\t\tmain_wait_for_workers();\n \n \t\trte_timer_manage();\n \t\tif (cb_count != NB_STRESS2_TIMERS) {\n@@ -386,14 +387,14 @@ timer_stress2_main_loop(__rte_unused void *arg)\n \t}\n \n cleanup:\n-\tif (lcore_id == master) {\n-\t\tmaster_wait_for_slaves();\n+\tif (lcore_id == main_lcore) {\n+\t\tmain_wait_for_workers();\n \t\tif (timers != NULL) {\n \t\t\trte_free(timers);\n \t\t\ttimers = NULL;\n \t\t}\n \t} else {\n-\t\tslave_finish();\n+\t\tworker_finish();\n \t}\n \n \treturn 0;\n@@ -465,7 +466,7 @@ timer_basic_main_loop(__rte_unused void *arg)\n \tint64_t diff = 0;\n \n \t/* launch all timers on core 0 */\n-\tif (lcore_id == rte_get_master_lcore()) {\n+\tif (lcore_id == rte_get_main_lcore()) {\n \t\tmytimer_reset(&mytiminfo[0], hz/4, SINGLE, lcore_id,\n \t\t\t      timer_basic_cb);\n \t\tmytimer_reset(&mytiminfo[1], hz/2, SINGLE, lcore_id,\n@@ -563,7 +564,7 @@ test_timer(void)\n \n \t/* start other cores */\n \tprintf(\"Start timer stress tests\\n\");\n-\trte_eal_mp_remote_launch(timer_stress_main_loop, NULL, CALL_MASTER);\n+\trte_eal_mp_remote_launch(timer_stress_main_loop, NULL, CALL_MAIN);\n \trte_eal_mp_wait_lcore();\n \n \t/* stop timer 0 used for stress test */\n@@ -572,7 +573,7 @@ test_timer(void)\n \t/* run a second, slightly different set of stress tests */\n \tprintf(\"\\nStart timer stress tests 2\\n\");\n \ttest_failed = 0;\n-\trte_eal_mp_remote_launch(timer_stress2_main_loop, NULL, CALL_MASTER);\n+\trte_eal_mp_remote_launch(timer_stress2_main_loop, NULL, CALL_MAIN);\n \trte_eal_mp_wait_lcore();\n \tif (test_failed)\n \t\treturn TEST_FAILED;\n@@ -584,7 +585,7 @@ test_timer(void)\n \n \t/* start other cores */\n \tprintf(\"\\nStart timer basic tests\\n\");\n-\trte_eal_mp_remote_launch(timer_basic_main_loop, NULL, CALL_MASTER);\n+\trte_eal_mp_remote_launch(timer_basic_main_loop, NULL, CALL_MAIN);\n \trte_eal_mp_wait_lcore();\n \n \t/* stop all timers */\ndiff --git a/app/test/test_timer_racecond.c b/app/test/test_timer_racecond.c\nindex 28af9ca765ea..133d3324ee66 100644\n--- a/app/test/test_timer_racecond.c\n+++ b/app/test/test_timer_racecond.c\n@@ -54,10 +54,10 @@\n #define N_TIMERS    50\n \n static struct rte_timer timer[N_TIMERS];\n-static unsigned timer_lcore_id[N_TIMERS];\n+static unsigned int timer_lcore_id[N_TIMERS];\n \n-static unsigned master;\n-static volatile unsigned stop_slaves;\n+static unsigned int main_lcore;\n+static volatile unsigned int stop_workers;\n \n static int reload_timer(struct rte_timer *tim);\n \n@@ -90,7 +90,7 @@ reload_timer(struct rte_timer *tim)\n \t    (tim - timer);\n \tint ret;\n \n-\tret = rte_timer_reset(tim, ticks, PERIODICAL, master, timer_cb, NULL);\n+\tret = rte_timer_reset(tim, ticks, PERIODICAL, main_lcore, timer_cb, NULL);\n \tif (ret != 0) {\n \t\trte_log(RTE_LOG_DEBUG, timer_logtype_test,\n \t\t\t\"- core %u failed to reset timer %\" PRIuPTR \" (OK)\\n\",\n@@ -101,7 +101,7 @@ reload_timer(struct rte_timer *tim)\n }\n \n static int\n-slave_main_loop(__rte_unused void *arg)\n+worker_main_loop(__rte_unused void *arg)\n {\n \tunsigned lcore_id = rte_lcore_id();\n \tunsigned i;\n@@ -110,7 +110,7 @@ slave_main_loop(__rte_unused void *arg)\n \n \tprintf(\"Starting main loop on core %u\\n\", lcore_id);\n \n-\twhile (!stop_slaves) {\n+\twhile (!stop_workers) {\n \t\t/* Wait until the timer manager is running.\n \t\t * We know it's running when we see timer[0] NOT pending.\n \t\t */\n@@ -147,7 +147,7 @@ test_timer_racecond(void)\n \tunsigned lcore_id;\n \tunsigned i;\n \n-\tmaster = lcore_id = rte_lcore_id();\n+\tmain_lcore = lcore_id = rte_lcore_id();\n \thz = rte_get_timer_hz();\n \n \t/* init and start timers */\n@@ -156,8 +156,8 @@ test_timer_racecond(void)\n \t\tret = reload_timer(&timer[i]);\n \t\tTEST_ASSERT(ret == 0, \"reload_timer failed\");\n \n-\t\t/* Distribute timers to slaves.\n-\t\t * Note that we assign timer[0] to the master.\n+\t\t/* Distribute timers to workers.\n+\t\t * Note that we assign timer[0] to the main.\n \t\t */\n \t\ttimer_lcore_id[i] = lcore_id;\n \t\tlcore_id = rte_get_next_lcore(lcore_id, 1, 1);\n@@ -167,11 +167,11 @@ test_timer_racecond(void)\n \tcur_time = rte_get_timer_cycles();\n \tend_time = cur_time + (hz * TEST_DURATION_S);\n \n-\t/* start slave cores */\n-\tstop_slaves = 0;\n+\t/* start worker cores */\n+\tstop_workers = 0;\n \tprintf(\"Start timer manage race condition test (%u seconds)\\n\",\n \t\t\tTEST_DURATION_S);\n-\trte_eal_mp_remote_launch(slave_main_loop, NULL, SKIP_MASTER);\n+\trte_eal_mp_remote_launch(worker_main_loop, NULL, SKIP_MAIN);\n \n \twhile (diff >= 0) {\n \t\t/* run the timers */\n@@ -184,9 +184,9 @@ test_timer_racecond(void)\n \t\tdiff = end_time - cur_time;\n \t}\n \n-\t/* stop slave cores */\n+\t/* stop worker cores */\n \tprintf(\"Stopping timer manage race condition test\\n\");\n-\tstop_slaves = 1;\n+\tstop_workers = 1;\n \trte_eal_mp_wait_lcore();\n \n \t/* stop timers */\ndiff --git a/app/test/test_timer_secondary.c b/app/test/test_timer_secondary.c\nindex 7a3bc873b359..1e8f1d45499a 100644\n--- a/app/test/test_timer_secondary.c\n+++ b/app/test/test_timer_secondary.c\n@@ -26,7 +26,7 @@\n #define launch_proc(ARGV) process_dup(ARGV, RTE_DIM(ARGV), __func__)\n \n struct test_info {\n-\tunsigned int mstr_lcore;\n+\tunsigned int main_lcore;\n \tunsigned int mgr_lcore;\n \tunsigned int sec_lcore;\n \tuint32_t timer_data_id;\n@@ -137,12 +137,12 @@ test_timer_secondary(void)\n \t\tTEST_ASSERT_SUCCESS(ret, \"Failed to allocate timer data \"\n \t\t\t\t    \"instance\");\n \n-\t\tunsigned int *mstr_lcorep = &test_info->mstr_lcore;\n+\t\tunsigned int *main_lcorep = &test_info->main_lcore;\n \t\tunsigned int *mgr_lcorep = &test_info->mgr_lcore;\n \t\tunsigned int *sec_lcorep = &test_info->sec_lcore;\n \n-\t\t*mstr_lcorep = rte_get_master_lcore();\n-\t\t*mgr_lcorep = rte_get_next_lcore(*mstr_lcorep, 1, 1);\n+\t\t*main_lcorep = rte_get_main_lcore();\n+\t\t*mgr_lcorep = rte_get_next_lcore(*main_lcorep, 1, 1);\n \t\t*sec_lcorep = rte_get_next_lcore(*mgr_lcorep, 1, 1);\n \n \t\tret = rte_eal_remote_launch(timer_manage_loop,\ndiff --git a/app/test/test_trace_perf.c b/app/test/test_trace_perf.c\nindex 50c7381b77e7..e1ad8e6f555c 100644\n--- a/app/test/test_trace_perf.c\n+++ b/app/test/test_trace_perf.c\n@@ -132,7 +132,7 @@ run_test(const char *str, lcore_function_t f, struct test_data *data, size_t sz)\n \n \tmemset(data, 0, sz);\n \tdata->nb_workers = rte_lcore_count() - 1;\n-\tRTE_LCORE_FOREACH_SLAVE(id)\n+\tRTE_LCORE_FOREACH_WORKER(id)\n \t\trte_eal_remote_launch(f, &data->ldata[worker++], id);\n \n \twait_till_workers_are_ready(data);\n@@ -140,7 +140,7 @@ run_test(const char *str, lcore_function_t f, struct test_data *data, size_t sz)\n \tmeasure_perf(str, data);\n \tsignal_workers_to_finish(data);\n \n-\tRTE_LCORE_FOREACH_SLAVE(id)\n+\tRTE_LCORE_FOREACH_WORKER(id)\n \t\trte_eal_wait_lcore(id);\n }\n \n",
    "prefixes": [
        "v5",
        "08/18"
    ]
}