@@ -257,7 +257,7 @@ int rte_dpaa_portal_init(void *arg)
BUS_INIT_FUNC_TRACE();
if ((size_t)arg == 1 || lcore == LCORE_ID_ANY)
- lcore = rte_get_master_lcore();
+ lcore = rte_get_main_lcore();
else
if (lcore >= RTE_MAX_LCORE)
return -1;
@@ -51,7 +51,8 @@ pci_uio_map_secondary(struct rte_pci_device *dev)
void *mapaddr = pci_map_resource(uio_res->maps[i].addr,
fd, (off_t)uio_res->maps[i].offset,
(size_t)uio_res->maps[i].size, 0);
- /* fd is not needed in slave process, close it */
+
+ /* fd is not needed in secondary process, close it */
close(fd);
if (mapaddr != uio_res->maps[i].addr) {
RTE_LOG(ERR, EAL,
@@ -85,7 +85,7 @@ vmbus_uio_map_secondary(struct rte_vmbus_device *dev)
return -1;
}
- /* fd is not needed in slave process, close it */
+ /* fd is not needed in secondary process, close it */
close(fd);
dev->primary = uio_res->primary;
@@ -501,8 +501,8 @@ wait_workers_to_join(int lcore, const rte_atomic32_t *count)
static int
-launch_workers_and_wait(int (*master_worker)(void *),
- int (*slave_workers)(void *), uint32_t total_events,
+launch_workers_and_wait(int (*main_worker)(void *),
+ int (*workers)(void *), uint32_t total_events,
uint8_t nb_workers, uint8_t sched_type)
{
uint8_t port = 0;
@@ -537,9 +537,9 @@ launch_workers_and_wait(int (*master_worker)(void *),
w_lcore = rte_get_next_lcore(
/* start core */ -1,
- /* skip master */ 1,
+ /* skip main */ 1,
/* wrap */ 0);
- rte_eal_remote_launch(master_worker, ¶m[0], w_lcore);
+ rte_eal_remote_launch(main_worker, ¶m[0], w_lcore);
for (port = 1; port < nb_workers; port++) {
param[port].total_events = &atomic_total_events;
@@ -548,7 +548,7 @@ launch_workers_and_wait(int (*master_worker)(void *),
param[port].dequeue_tmo_ticks = dequeue_tmo_ticks;
rte_smp_wmb();
w_lcore = rte_get_next_lcore(w_lcore, 1, 0);
- rte_eal_remote_launch(slave_workers, ¶m[port], w_lcore);
+ rte_eal_remote_launch(workers, ¶m[port], w_lcore);
}
ret = wait_workers_to_join(w_lcore, &atomic_total_events);
@@ -601,8 +601,8 @@ wait_workers_to_join(int lcore, const rte_atomic32_t *count)
static inline int
-launch_workers_and_wait(int (*master_worker)(void *),
- int (*slave_workers)(void *), uint32_t total_events,
+launch_workers_and_wait(int (*main_worker)(void *),
+ int (*worker)(void *), uint32_t total_events,
uint8_t nb_workers, uint8_t sched_type)
{
uint8_t port = 0;
@@ -637,9 +637,9 @@ launch_workers_and_wait(int (*master_worker)(void *),
w_lcore = rte_get_next_lcore(
/* start core */ -1,
- /* skip master */ 1,
+ /* skip main */ 1,
/* wrap */ 0);
- rte_eal_remote_launch(master_worker, ¶m[0], w_lcore);
+ rte_eal_remote_launch(main_worker, ¶m[0], w_lcore);
for (port = 1; port < nb_workers; port++) {
param[port].total_events = &atomic_total_events;
@@ -648,7 +648,7 @@ launch_workers_and_wait(int (*master_worker)(void *),
param[port].dequeue_tmo_ticks = dequeue_tmo_ticks;
rte_smp_wmb();
w_lcore = rte_get_next_lcore(w_lcore, 1, 0);
- rte_eal_remote_launch(slave_workers, ¶m[port], w_lcore);
+ rte_eal_remote_launch(worker, ¶m[port], w_lcore);
}
ret = wait_workers_to_join(w_lcore, &atomic_total_events);
@@ -3108,7 +3108,7 @@ worker_loopback(struct test *t, uint8_t disable_implicit_release)
p_lcore = rte_get_next_lcore(
/* start core */ -1,
- /* skip master */ 1,
+ /* skip main */ 1,
/* wrap */ 0);
w_lcore = rte_get_next_lcore(p_lcore, 1, 0);
@@ -429,7 +429,7 @@ int bnxt_alloc_rxtx_nq_ring(struct bnxt *bp)
if (!BNXT_HAS_NQ(bp) || bp->rxtx_nq_ring)
return 0;
- socket_id = rte_lcore_to_socket_id(rte_get_master_lcore());
+ socket_id = rte_lcore_to_socket_id(rte_get_main_lcore());
nqr = rte_zmalloc_socket("nqr",
sizeof(struct bnxt_cp_ring_info),
@@ -820,7 +820,7 @@ int bnxt_alloc_async_ring_struct(struct bnxt *bp)
if (BNXT_NUM_ASYNC_CPR(bp) == 0)
return 0;
- socket_id = rte_lcore_to_socket_id(rte_get_master_lcore());
+ socket_id = rte_lcore_to_socket_id(rte_get_main_lcore());
cpr = rte_zmalloc_socket("cpr",
sizeof(struct bnxt_cp_ring_info),
@@ -814,7 +814,7 @@ mrvl_flush_bpool(struct rte_eth_dev *dev)
unsigned int core_id = rte_lcore_id();
if (core_id == LCORE_ID_ANY)
- core_id = rte_get_master_lcore();
+ core_id = rte_get_main_lcore();
hif = mrvl_get_hif(priv, core_id);
@@ -1623,7 +1623,7 @@ mrvl_fill_bpool(struct mrvl_rxq *rxq, int num)
core_id = rte_lcore_id();
if (core_id == LCORE_ID_ANY)
- core_id = rte_get_master_lcore();
+ core_id = rte_get_main_lcore();
hif = mrvl_get_hif(rxq->priv, core_id);
if (!hif)
@@ -1773,7 +1773,7 @@ mrvl_rx_queue_release(void *rxq)
unsigned int core_id = rte_lcore_id();
if (core_id == LCORE_ID_ANY)
- core_id = rte_get_master_lcore();
+ core_id = rte_get_main_lcore();
if (!q)
return;
@@ -143,7 +143,7 @@ void *osal_dma_alloc_coherent(struct ecore_dev *p_dev,
snprintf(mz_name, sizeof(mz_name), "%lx",
(unsigned long)rte_get_timer_cycles());
if (core_id == (unsigned int)LCORE_ID_ANY)
- core_id = rte_get_master_lcore();
+ core_id = rte_get_main_lcore();
socket_id = rte_lcore_to_socket_id(core_id);
mz = rte_memzone_reserve_aligned(mz_name, size, socket_id,
RTE_MEMZONE_IOVA_CONTIG, RTE_CACHE_LINE_SIZE);
@@ -182,7 +182,7 @@ void *osal_dma_alloc_coherent_aligned(struct ecore_dev *p_dev,
snprintf(mz_name, sizeof(mz_name), "%lx",
(unsigned long)rte_get_timer_cycles());
if (core_id == (unsigned int)LCORE_ID_ANY)
- core_id = rte_get_master_lcore();
+ core_id = rte_get_main_lcore();
socket_id = rte_lcore_to_socket_id(core_id);
mz = rte_memzone_reserve_aligned(mz_name, size, socket_id,
RTE_MEMZONE_IOVA_CONTIG, align);
@@ -547,7 +547,7 @@ TAILQ_HEAD(pipeline_list, pipeline);
#endif
/**
- * Master thead: data plane thread context
+ * Main thead: data plane thread context
*/
struct softnic_thread {
struct rte_ring *msgq_req;
@@ -18,14 +18,14 @@
#include "rte_eth_softnic_internals.h"
/**
- * Master thread: data plane thread init
+ * Main thread: data plane thread init
*/
void
softnic_thread_free(struct pmd_internals *softnic)
{
uint32_t i;
- RTE_LCORE_FOREACH_SLAVE(i) {
+ RTE_LCORE_FOREACH_WORKER(i) {
struct softnic_thread *t = &softnic->thread[i];
/* MSGQs */
@@ -78,7 +78,7 @@ softnic_thread_init(struct pmd_internals *softnic)
return -1;
}
- /* Master thread records */
+ /* Main thread records */
t->msgq_req = msgq_req;
t->msgq_rsp = msgq_rsp;
t->service_id = UINT32_MAX;
@@ -99,7 +99,7 @@ softnic_thread_init(struct pmd_internals *softnic)
static inline int
thread_is_valid(struct pmd_internals *softnic, uint32_t thread_id)
{
- if (thread_id == rte_get_master_lcore())
+ if (thread_id == rte_get_main_lcore())
return 0; /* FALSE */
if (softnic->params.sc && rte_lcore_has_role(thread_id, ROLE_SERVICE))
@@ -209,7 +209,7 @@ pipeline_is_running(struct pipeline *p)
}
/**
- * Master thread & data plane threads: message passing
+ * Main thread & data plane threads: message passing
*/
enum thread_req_type {
THREAD_REQ_PIPELINE_ENABLE = 0,
@@ -243,7 +243,7 @@ struct thread_msg_rsp {
};
/**
- * Master thread
+ * Main thread
*/
static struct thread_msg_req *
thread_msg_alloc(void)
@@ -587,7 +587,7 @@ thread_msg_handle(struct softnic_thread_data *t)
}
/**
- * Master thread & data plane threads: message passing
+ * Main thread & data plane threads: message passing
*/
enum pipeline_req_type {
/* Port IN */
@@ -753,7 +753,7 @@ struct pipeline_msg_rsp {
};
/**
- * Master thread
+ * Main thread
*/
static struct pipeline_msg_req *
pipeline_msg_alloc(void)