[dpdk-dev,2/8] app/eventdev: modify app setup to support ethdev
Checks
Commit Message
Modify app setup to accommodate event port and queue setup based on the
number of ethernet ports.
Signed-off-by: Pavan Nikhilesh <pbhagavatula@caviumnetworks.com>
---
app/test-eventdev/test_perf_atq.c | 18 ++++++++++++++----
app/test-eventdev/test_perf_common.c | 27 +++++++++++++++++++++------
app/test-eventdev/test_perf_common.h | 1 +
app/test-eventdev/test_perf_queue.c | 21 ++++++++++++++++-----
4 files changed, 52 insertions(+), 15 deletions(-)
Comments
-----Original Message-----
> Date: Wed, 18 Oct 2017 18:09:02 +0530
> From: Pavan Nikhilesh <pbhagavatula@caviumnetworks.com>
> To: bruce.richardson@intel.com, harry.van.haaren@intel.com,
> gage.eads@intel.com, hemant.agrawal@nxp.com, nipun.gupta@nxp.com,
> nikhil.rao@intel.com, santosh.shukla@caviumnetworks.com,
> jerin.jacob@caviumnetworks.com
> Cc: dev@dpdk.org, Pavan Nikhilesh <pbhagavatula@caviumnetworks.com>
> Subject: [PATCH 2/8] app/eventdev: modify app setup to support ethdev
> X-Mailer: git-send-email 2.7.4
>
> Modify app setup to accommodate event port and queue setup based on the
> number of ethernet ports.
>
> Signed-off-by: Pavan Nikhilesh <pbhagavatula@caviumnetworks.com>
> ---
> app/test-eventdev/test_perf_atq.c | 18 ++++++++++++++----
> app/test-eventdev/test_perf_common.c | 27 +++++++++++++++++++++------
> app/test-eventdev/test_perf_common.h | 1 +
> app/test-eventdev/test_perf_queue.c | 21 ++++++++++++++++-----
> 4 files changed, 52 insertions(+), 15 deletions(-)
>
> diff --git a/app/test-eventdev/test_perf_atq.c b/app/test-eventdev/test_perf_atq.c
> index 0e9f2db..f6fd5d9 100644
> --- a/app/test-eventdev/test_perf_atq.c
> +++ b/app/test-eventdev/test_perf_atq.c
> @@ -185,10 +185,20 @@ perf_atq_eventdev_setup(struct evt_test *test, struct evt_options *opt)
> {
> int ret;
> uint8_t queue;
> + uint8_t nb_queues;
unnecessary space between uint8_t and nb_queues.
> + uint8_t nb_ports;
> + RTE_SET_USED(test);
test being used down. So RTE_SET_USED may not need here.
> +
> + nb_ports = evt_nr_active_lcores(opt->wlcores);
> + nb_ports += opt->prod_type == EVT_PROD_TYPE_ETH_RX_ADPTR ? 0 :
> + evt_nr_active_lcores(opt->plcores);
> +
> + nb_queues = opt->prod_type == EVT_PROD_TYPE_ETH_RX_ADPTR ?
> + rte_eth_dev_count() : atq_nb_event_queues(opt);
>
> const struct rte_event_dev_config config = {
> - .nb_event_queues = atq_nb_event_queues(opt),
> - .nb_event_ports = perf_nb_event_ports(opt),
> + .nb_event_queues = nb_queues,
> + .nb_event_ports = nb_ports,
> .nb_events_limit = 4096,
> .nb_event_queue_flows = opt->nb_flows,
> .nb_event_port_dequeue_depth = 128,
> @@ -208,7 +218,7 @@ perf_atq_eventdev_setup(struct evt_test *test, struct evt_options *opt)
> .nb_atomic_order_sequences = opt->nb_flows,
> };
> /* queue configurations */
> - for (queue = 0; queue < atq_nb_event_queues(opt); queue++) {
> + for (queue = 0; queue < nb_queues; queue++) {
> ret = rte_event_queue_setup(opt->dev_id, queue, &q_conf);
> if (ret) {
> evt_err("failed to setup queue=%d", queue);
> @@ -217,7 +227,7 @@ perf_atq_eventdev_setup(struct evt_test *test, struct evt_options *opt)
> }
>
> ret = perf_event_dev_port_setup(test, opt, 1 /* stride */,
> - atq_nb_event_queues(opt));
> + nb_queues);
> if (ret)
> return ret;
>
> diff --git a/app/test-eventdev/test_perf_common.c b/app/test-eventdev/test_perf_common.c
> index 9d2865e..f7961be 100644
> --- a/app/test-eventdev/test_perf_common.c
> +++ b/app/test-eventdev/test_perf_common.c
> @@ -88,6 +88,18 @@ perf_producer(void *arg)
> return 0;
> }
>
> +static int
> +perf_producer_wrapper(void *arg)
> +{
> + RTE_SET_USED(arg);
arg has been used in next line.
> + struct prod_data *p = arg;
> + struct test_perf *t = p->t;
> + /* Launch the producer function only in case of synthetic producer. */
> + if (t->opt->prod_type == EVT_PROD_TYPE_SYNT)
> + return perf_producer(arg);
> + return 0;
> +}
> +
Other than that, it looks good to me
Acked-by: Jerin Jacob <jerin.jacob@caviumnetworks.com>
@@ -185,10 +185,20 @@ perf_atq_eventdev_setup(struct evt_test *test, struct evt_options *opt)
{
int ret;
uint8_t queue;
+ uint8_t nb_queues;
+ uint8_t nb_ports;
+ RTE_SET_USED(test);
+
+ nb_ports = evt_nr_active_lcores(opt->wlcores);
+ nb_ports += opt->prod_type == EVT_PROD_TYPE_ETH_RX_ADPTR ? 0 :
+ evt_nr_active_lcores(opt->plcores);
+
+ nb_queues = opt->prod_type == EVT_PROD_TYPE_ETH_RX_ADPTR ?
+ rte_eth_dev_count() : atq_nb_event_queues(opt);
const struct rte_event_dev_config config = {
- .nb_event_queues = atq_nb_event_queues(opt),
- .nb_event_ports = perf_nb_event_ports(opt),
+ .nb_event_queues = nb_queues,
+ .nb_event_ports = nb_ports,
.nb_events_limit = 4096,
.nb_event_queue_flows = opt->nb_flows,
.nb_event_port_dequeue_depth = 128,
@@ -208,7 +218,7 @@ perf_atq_eventdev_setup(struct evt_test *test, struct evt_options *opt)
.nb_atomic_order_sequences = opt->nb_flows,
};
/* queue configurations */
- for (queue = 0; queue < atq_nb_event_queues(opt); queue++) {
+ for (queue = 0; queue < nb_queues; queue++) {
ret = rte_event_queue_setup(opt->dev_id, queue, &q_conf);
if (ret) {
evt_err("failed to setup queue=%d", queue);
@@ -217,7 +227,7 @@ perf_atq_eventdev_setup(struct evt_test *test, struct evt_options *opt)
}
ret = perf_event_dev_port_setup(test, opt, 1 /* stride */,
- atq_nb_event_queues(opt));
+ nb_queues);
if (ret)
return ret;
@@ -88,6 +88,18 @@ perf_producer(void *arg)
return 0;
}
+static int
+perf_producer_wrapper(void *arg)
+{
+ RTE_SET_USED(arg);
+ struct prod_data *p = arg;
+ struct test_perf *t = p->t;
+ /* Launch the producer function only in case of synthetic producer. */
+ if (t->opt->prod_type == EVT_PROD_TYPE_SYNT)
+ return perf_producer(arg);
+ return 0;
+}
+
static inline uint64_t
processed_pkts(struct test_perf *t)
{
@@ -142,8 +154,8 @@ perf_launch_lcores(struct evt_test *test, struct evt_options *opt,
if (!(opt->plcores[lcore_id]))
continue;
- ret = rte_eal_remote_launch(perf_producer, &t->prod[port_idx],
- lcore_id);
+ ret = rte_eal_remote_launch(perf_producer_wrapper,
+ &t->prod[port_idx], lcore_id);
if (ret) {
evt_err("failed to launch perf_producer %d", lcore_id);
return ret;
@@ -193,14 +205,17 @@ perf_launch_lcores(struct evt_test *test, struct evt_options *opt,
fflush(stdout);
if (remaining <= 0) {
- t->done = true;
t->result = EVT_TEST_SUCCESS;
- rte_smp_wmb();
- break;
+ if (opt->prod_type == EVT_PROD_TYPE_SYNT) {
+ t->done = true;
+ rte_smp_wmb();
+ break;
+ }
}
}
- if (new_cycles - dead_lock_cycles > dead_lock_sample) {
+ if (new_cycles - dead_lock_cycles > dead_lock_sample &&
+ opt->prod_type == EVT_PROD_TYPE_SYNT) {
remaining = t->outstand_pkts - processed_pkts(t);
if (dead_lock_remaining == remaining) {
rte_event_dev_dump(opt->dev_id, stdout);
@@ -38,6 +38,7 @@
#include <unistd.h>
#include <rte_cycles.h>
+#include <rte_ethdev.h>
#include <rte_eventdev.h>
#include <rte_lcore.h>
#include <rte_malloc.h>
@@ -182,10 +182,21 @@ perf_queue_eventdev_setup(struct evt_test *test, struct evt_options *opt)
uint8_t queue;
int nb_stages = opt->nb_stages;
int ret;
+ int nb_ports;
+ int nb_queues;
+ RTE_SET_USED(test);
+
+ nb_ports = evt_nr_active_lcores(opt->wlcores);
+ nb_ports += opt->prod_type == EVT_PROD_TYPE_ETH_RX_ADPTR ? 0 :
+ evt_nr_active_lcores(opt->plcores);
+
+ nb_queues = opt->prod_type == EVT_PROD_TYPE_ETH_RX_ADPTR ?
+ rte_eth_dev_count() * nb_stages :
+ perf_queue_nb_event_queues(opt);
const struct rte_event_dev_config config = {
- .nb_event_queues = perf_queue_nb_event_queues(opt),
- .nb_event_ports = perf_nb_event_ports(opt),
+ .nb_event_queues = nb_queues,
+ .nb_event_ports = nb_ports,
.nb_events_limit = 4096,
.nb_event_queue_flows = opt->nb_flows,
.nb_event_port_dequeue_depth = 128,
@@ -204,8 +215,8 @@ perf_queue_eventdev_setup(struct evt_test *test, struct evt_options *opt)
.nb_atomic_order_sequences = opt->nb_flows,
};
/* queue configurations */
- for (queue = 0; queue < perf_queue_nb_event_queues(opt); queue++) {
- q_conf.event_queue_cfg =
+ for (queue = 0; queue < nb_queues; queue++) {
+ q_conf.schedule_type =
(opt->sched_type_list[queue % nb_stages]);
if (opt->q_priority) {
@@ -228,7 +239,7 @@ perf_queue_eventdev_setup(struct evt_test *test, struct evt_options *opt)
}
ret = perf_event_dev_port_setup(test, opt, nb_stages /* stride */,
- perf_queue_nb_event_queues(opt));
+ nb_queues);
if (ret)
return ret;