get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/77526/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 77526,
    "url": "http://patches.dpdk.org/api/patches/77526/?format=api",
    "web_url": "http://patches.dpdk.org/project/dpdk/patch/1599855987-25976-20-git-send-email-timothy.mcdaniel@intel.com/",
    "project": {
        "id": 1,
        "url": "http://patches.dpdk.org/api/projects/1/?format=api",
        "name": "DPDK",
        "link_name": "dpdk",
        "list_id": "dev.dpdk.org",
        "list_email": "dev@dpdk.org",
        "web_url": "http://core.dpdk.org",
        "scm_url": "git://dpdk.org/dpdk",
        "webscm_url": "http://git.dpdk.org/dpdk",
        "list_archive_url": "https://inbox.dpdk.org/dev",
        "list_archive_url_format": "https://inbox.dpdk.org/dev/{}",
        "commit_url_format": ""
    },
    "msgid": "<1599855987-25976-20-git-send-email-timothy.mcdaniel@intel.com>",
    "list_archive_url": "https://inbox.dpdk.org/dev/1599855987-25976-20-git-send-email-timothy.mcdaniel@intel.com",
    "date": "2020-09-11T20:26:24",
    "name": "[19/22] event/dlb2: add PMD self-tests",
    "commit_ref": null,
    "pull_url": null,
    "state": "superseded",
    "archived": true,
    "hash": "c26d8c344280a178ba56e9d84cc59398e6f27d42",
    "submitter": {
        "id": 826,
        "url": "http://patches.dpdk.org/api/people/826/?format=api",
        "name": "Timothy McDaniel",
        "email": "timothy.mcdaniel@intel.com"
    },
    "delegate": {
        "id": 310,
        "url": "http://patches.dpdk.org/api/users/310/?format=api",
        "username": "jerin",
        "first_name": "Jerin",
        "last_name": "Jacob",
        "email": "jerinj@marvell.com"
    },
    "mbox": "http://patches.dpdk.org/project/dpdk/patch/1599855987-25976-20-git-send-email-timothy.mcdaniel@intel.com/mbox/",
    "series": [
        {
            "id": 12164,
            "url": "http://patches.dpdk.org/api/series/12164/?format=api",
            "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=12164",
            "date": "2020-09-11T20:26:05",
            "name": "Add DLB2 PMD",
            "version": 1,
            "mbox": "http://patches.dpdk.org/series/12164/mbox/"
        }
    ],
    "comments": "http://patches.dpdk.org/api/patches/77526/comments/",
    "check": "success",
    "checks": "http://patches.dpdk.org/api/patches/77526/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<dev-bounces@dpdk.org>",
        "X-Original-To": "patchwork@inbox.dpdk.org",
        "Delivered-To": "patchwork@inbox.dpdk.org",
        "Received": [
            "from dpdk.org (dpdk.org [92.243.14.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id 623F8A04C1;\n\tFri, 11 Sep 2020 22:33:35 +0200 (CEST)",
            "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id 534371C296;\n\tFri, 11 Sep 2020 22:30:42 +0200 (CEST)",
            "from mga04.intel.com (mga04.intel.com [192.55.52.120])\n by dpdk.org (Postfix) with ESMTP id 2AD671C194\n for <dev@dpdk.org>; Fri, 11 Sep 2020 22:30:10 +0200 (CEST)",
            "from orsmga005.jf.intel.com ([10.7.209.41])\n by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 11 Sep 2020 13:30:09 -0700",
            "from txasoft-yocto.an.intel.com ([10.123.72.192])\n by orsmga005.jf.intel.com with ESMTP; 11 Sep 2020 13:30:08 -0700"
        ],
        "IronPort-SDR": [
            "\n RWxV0+L2pi8Ia8HhoY6WQ/2ezr3S8O7fb4u7xp20Ez/ubWQgqnqJKDR4AonKjNzAKEHwA+vTfr\n Bu9pktzzfpBA==",
            "\n /T37Q4ayUZiJV7n5DNfQGIn9ufhDle7oK6ZrybEsJ5IVMSeB+QM7MpL9aNR0G78XDkIAInNCm5\n VRyeylb5IoNA=="
        ],
        "X-IronPort-AV": [
            "E=McAfee;i=\"6000,8403,9741\"; a=\"156244390\"",
            "E=Sophos;i=\"5.76,417,1592895600\"; d=\"scan'208\";a=\"156244390\"",
            "E=Sophos;i=\"5.76,417,1592895600\"; d=\"scan'208\";a=\"481453689\""
        ],
        "X-Amp-Result": "SKIPPED(no attachment in message)",
        "X-Amp-File-Uploaded": "False",
        "X-ExtLoop1": "1",
        "From": "Timothy McDaniel <timothy.mcdaniel@intel.com>",
        "To": "Jerin Jacob <jerinj@marvell.com>",
        "Cc": "dev@dpdk.org, erik.g.carrillo@intel.com, gage.eads@intel.com,\n harry.van.haaren@intel.com",
        "Date": "Fri, 11 Sep 2020 15:26:24 -0500",
        "Message-Id": "<1599855987-25976-20-git-send-email-timothy.mcdaniel@intel.com>",
        "X-Mailer": "git-send-email 1.7.10",
        "In-Reply-To": "<1599855987-25976-1-git-send-email-timothy.mcdaniel@intel.com>",
        "References": "<1599855987-25976-1-git-send-email-timothy.mcdaniel@intel.com>",
        "Subject": "[dpdk-dev] [PATCH 19/22] event/dlb2: add PMD self-tests",
        "X-BeenThere": "dev@dpdk.org",
        "X-Mailman-Version": "2.1.15",
        "Precedence": "list",
        "List-Id": "DPDK patches and discussions <dev.dpdk.org>",
        "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n <mailto:dev-request@dpdk.org?subject=unsubscribe>",
        "List-Archive": "<http://mails.dpdk.org/archives/dev/>",
        "List-Post": "<mailto:dev@dpdk.org>",
        "List-Help": "<mailto:dev-request@dpdk.org?subject=help>",
        "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n <mailto:dev-request@dpdk.org?subject=subscribe>",
        "Errors-To": "dev-bounces@dpdk.org",
        "Sender": "\"dev\" <dev-bounces@dpdk.org>"
    },
    "content": "Add a variety of self-tests for both ldb and directed\nports/queues, as well as configure, start, stop, link, etc...\n\nSigned-off-by: Timothy McDaniel <timothy.mcdaniel@intel.com>\n---\n app/test/test_eventdev.c           |    9 +\n drivers/event/dlb2/dlb2.c          |    1 +\n drivers/event/dlb2/dlb2_selftest.c | 1570 ++++++++++++++++++++++++++++++++++++\n drivers/event/dlb2/meson.build     |    3 +-\n 4 files changed, 1582 insertions(+), 1 deletion(-)\n create mode 100644 drivers/event/dlb2/dlb2_selftest.c",
    "diff": "diff --git a/app/test/test_eventdev.c b/app/test/test_eventdev.c\nindex 43ccb1c..b8d8df8 100644\n--- a/app/test/test_eventdev.c\n+++ b/app/test/test_eventdev.c\n@@ -1030,6 +1030,13 @@ test_eventdev_selftest_dpaa2(void)\n \treturn test_eventdev_selftest_impl(\"event_dpaa2\", \"\");\n }\n \n+static int\n+test_eventdev_selftest_dlb2(void)\n+{\n+\treturn test_eventdev_selftest_impl(\"dlb2_event\", \"\");\n+}\n+\n+\n REGISTER_TEST_COMMAND(eventdev_common_autotest, test_eventdev_common);\n REGISTER_TEST_COMMAND(eventdev_selftest_sw, test_eventdev_selftest_sw);\n REGISTER_TEST_COMMAND(eventdev_selftest_octeontx,\n@@ -1037,3 +1044,5 @@ REGISTER_TEST_COMMAND(eventdev_selftest_octeontx,\n REGISTER_TEST_COMMAND(eventdev_selftest_octeontx2,\n \t\ttest_eventdev_selftest_octeontx2);\n REGISTER_TEST_COMMAND(eventdev_selftest_dpaa2, test_eventdev_selftest_dpaa2);\n+REGISTER_TEST_COMMAND(eventdev_selftest_dlb2, test_eventdev_selftest_dlb2);\n+\ndiff --git a/drivers/event/dlb2/dlb2.c b/drivers/event/dlb2/dlb2.c\nindex 43b85d7..620a0a5 100644\n--- a/drivers/event/dlb2/dlb2.c\n+++ b/drivers/event/dlb2/dlb2.c\n@@ -3780,6 +3780,7 @@ dlb2_entry_points_init(struct rte_eventdev *dev)\n \t\t.xstats_get_names = dlb2_eventdev_xstats_get_names,\n \t\t.xstats_get_by_name = dlb2_eventdev_xstats_get_by_name,\n \t\t.xstats_reset\t    = dlb2_eventdev_xstats_reset,\n+\t\t.dev_selftest     = test_dlb2_eventdev,\n \t};\n \n \t/* Expose PMD's eventdev interface */\ndiff --git a/drivers/event/dlb2/dlb2_selftest.c b/drivers/event/dlb2/dlb2_selftest.c\nnew file mode 100644\nindex 0000000..86d6344\n--- /dev/null\n+++ b/drivers/event/dlb2/dlb2_selftest.c\n@@ -0,0 +1,1570 @@\n+/* SPDX-License-Identifier: BSD-3-Clause\n+ * Copyright(c) 2016-2020 Intel Corporation\n+ */\n+\n+#include <stdio.h>\n+#include <string.h>\n+#include <stdint.h>\n+#include <errno.h>\n+#include <unistd.h>\n+#include <sys/queue.h>\n+\n+#include <rte_memory.h>\n+#include <rte_memzone.h>\n+#include <rte_launch.h>\n+#include <rte_eal.h>\n+#include <rte_per_lcore.h>\n+#include <rte_lcore.h>\n+#include <rte_debug.h>\n+#include <rte_ethdev.h>\n+#include <rte_cycles.h>\n+#include <rte_eventdev.h>\n+#include <rte_pause.h>\n+\n+#include \"dlb2_priv.h\"\n+#include \"rte_pmd_dlb2.h\"\n+\n+#define MAX_PORTS 32\n+#define MAX_QIDS 32\n+#define DEFAULT_NUM_SEQ_NUMS 64\n+\n+static struct rte_mempool *eventdev_func_mempool;\n+static int evdev;\n+\n+struct test {\n+\tstruct rte_mempool *mbuf_pool;\n+\tint nb_qids;\n+};\n+\n+/* initialization and config */\n+static inline int\n+init(struct test *t, int nb_queues, int nb_ports)\n+{\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_dev_info info;\n+\tint ret;\n+\n+\tmemset(t, 0, sizeof(*t));\n+\n+\tt->mbuf_pool = eventdev_func_mempool;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = nb_queues;\n+\tconfig.nb_event_ports = nb_ports;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0)\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\n+\treturn ret;\n+}\n+\n+static inline int\n+create_ports(int num_ports)\n+{\n+\tint i;\n+\n+\tif (num_ports > MAX_PORTS)\n+\t\treturn -1;\n+\n+\tfor (i = 0; i < num_ports; i++) {\n+\t\tstruct rte_event_port_conf conf;\n+\n+\t\tif (rte_event_port_default_conf_get(evdev, i, &conf)) {\n+\t\t\tprintf(\"%d: Error querying default port conf\\n\",\n+\t\t\t       __LINE__);\n+\t\t\treturn -1;\n+\t\t}\n+\n+\t\tif (rte_event_port_setup(evdev, i, &conf) < 0) {\n+\t\t\tprintf(\"%d: Error setting up port %d\\n\", __LINE__, i);\n+\t\t\treturn -1;\n+\t\t}\n+\t}\n+\n+\treturn 0;\n+}\n+\n+static inline int\n+create_lb_qids(struct test *t, int num_qids, uint32_t flags)\n+{\n+\tint i;\n+\n+\tfor (i = t->nb_qids; i < t->nb_qids + num_qids; i++) {\n+\t\tstruct rte_event_queue_conf conf;\n+\n+\t\tif (rte_event_queue_default_conf_get(evdev, i, &conf)) {\n+\t\t\tprintf(\"%d: Error querying default queue conf\\n\",\n+\t\t\t       __LINE__);\n+\t\t\treturn -1;\n+\t\t}\n+\n+\t\tconf.schedule_type = flags;\n+\n+\t\tif (conf.schedule_type == RTE_SCHED_TYPE_PARALLEL)\n+\t\t\tconf.nb_atomic_order_sequences = 0;\n+\t\telse\n+\t\t\tconf.nb_atomic_order_sequences = DEFAULT_NUM_SEQ_NUMS;\n+\n+\t\tif (rte_event_queue_setup(evdev, i, &conf) < 0) {\n+\t\t\tprintf(\"%d: error creating qid %d\\n\", __LINE__, i);\n+\t\t\treturn -1;\n+\t\t}\n+\t}\n+\n+\tt->nb_qids += num_qids;\n+\tif (t->nb_qids > MAX_QIDS)\n+\t\treturn -1;\n+\n+\treturn 0;\n+}\n+\n+static inline int\n+create_atomic_qids(struct test *t, int num_qids)\n+{\n+\treturn create_lb_qids(t, num_qids, RTE_SCHED_TYPE_ATOMIC);\n+}\n+\n+static inline int\n+create_ordered_qids(struct test *t, int num_qids)\n+{\n+\treturn create_lb_qids(t, num_qids, RTE_SCHED_TYPE_ORDERED);\n+}\n+\n+static inline int\n+create_unordered_qids(struct test *t, int num_qids)\n+{\n+\treturn create_lb_qids(t, num_qids, RTE_SCHED_TYPE_PARALLEL);\n+}\n+\n+static inline int\n+create_directed_qids(struct test *t, int num_qids, const uint8_t ports[])\n+{\n+\tstatic struct rte_event_queue_conf conf;\n+\tint i;\n+\n+\tconf.priority = RTE_EVENT_DEV_PRIORITY_NORMAL;\n+\tconf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tfor (i = t->nb_qids; i < t->nb_qids + num_qids; i++) {\n+\t\tuint8_t queue_id;\n+\n+\t\tif (rte_event_queue_setup(evdev, i, &conf) < 0) {\n+\t\t\tprintf(\"%d: error creating qid %d\\n\", __LINE__, i);\n+\t\t\treturn -1;\n+\t\t}\n+\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, ports[i - t->nb_qids],\n+\t\t\t\t\t&queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: error creating link for qid %d\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\treturn -1;\n+\t\t}\n+\t}\n+\n+\tt->nb_qids += num_qids;\n+\tif (t->nb_qids > MAX_QIDS)\n+\t\treturn -1;\n+\n+\treturn 0;\n+}\n+\n+/* destruction */\n+static inline void\n+cleanup(struct test *t __rte_unused)\n+{\n+\tint ret = 0;\n+\n+\trte_event_dev_stop(evdev);\n+\tret = rte_event_dev_close(evdev);\n+\n+\tif (ret)\n+\t\tprintf(\"%d: rte_event_dev_close failed, ret = %d\\n\",\n+\t\t\t__LINE__, ret);\n+};\n+\n+static inline int\n+enqueue_timeout(uint8_t port_id, struct rte_event *ev, uint64_t tmo_us)\n+{\n+\tconst uint64_t start = rte_get_timer_cycles();\n+\tconst uint64_t ticks = (tmo_us * rte_get_timer_hz()) / 1E6;\n+\n+\twhile ((rte_get_timer_cycles() - start) < ticks) {\n+\t\tif (rte_event_enqueue_burst(evdev, port_id, ev, 1) == 1)\n+\t\t\treturn 0;\n+\n+\t\tif (rte_errno != -ENOSPC) {\n+\t\t\tprintf(\"enqueue_burst returned rte_errno %d\\n\",\n+\t\t\t       rte_errno);\n+\t\t\treturn -1;\n+\t\t}\n+\t}\n+\tprintf(\"%s time out\\n\", __func__);\n+\treturn -1;\n+}\n+\n+static void\n+flush(uint8_t id __rte_unused, struct rte_event event, void *arg __rte_unused)\n+{\n+\trte_pktmbuf_free(event.mbuf);\n+}\n+\n+static int\n+test_stop_flush(struct test *t) /* test to check we can properly flush events */\n+{\n+\tstruct rte_event ev;\n+\tuint32_t dequeue_depth;\n+\tunsigned int i, count;\n+\tuint8_t queue_id;\n+\n+\tev.op = RTE_EVENT_OP_NEW;\n+\n+\tif (init(t, 2, 1) < 0 ||\n+\t    create_ports(1) < 0 ||\n+\t    create_atomic_qids(t, 2) < 0) {\n+\t\tprintf(\"%d: Error initializing device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tif (rte_event_port_link(evdev, 0, NULL, NULL, 0) != 2) {\n+\t\tprintf(\"%d: Error linking queues to the port\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: Error with start call\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Unlink queue 1 so the PMD's stop callback has to cleanup an unlinked\n+\t * queue.\n+\t */\n+\tqueue_id = 1;\n+\n+\tif (rte_event_port_unlink(evdev, 0, &queue_id, 1) != 1) {\n+\t\tprintf(\"%d: Error unlinking queue 1 from port\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcount = rte_mempool_avail_count(t->mbuf_pool);\n+\n+\tif (rte_event_port_attr_get(evdev,\n+\t\t\t\t    0,\n+\t\t\t\t    RTE_EVENT_PORT_ATTR_DEQ_DEPTH,\n+\t\t\t\t    &dequeue_depth)) {\n+\t\tprintf(\"%d: Error retrieveing dequeue depth\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Send QEs to queue 0 */\n+\tfor (i = 0; i < dequeue_depth + 1; i++) {\n+\t\tev.mbuf = rte_pktmbuf_alloc(t->mbuf_pool);\n+\t\tev.queue_id = 0;\n+\t\tev.sched_type = RTE_SCHED_TYPE_ATOMIC;\n+\n+\t\tif (enqueue_timeout(0, &ev, 1000)) {\n+\t\t\tprintf(\"%d: Error enqueuing events\\n\", __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Send QEs to queue 1 */\n+\tfor (i = 0; i < dequeue_depth + 1; i++) {\n+\t\tev.mbuf = rte_pktmbuf_alloc(t->mbuf_pool);\n+\t\tev.queue_id = 1;\n+\t\tev.sched_type = RTE_SCHED_TYPE_ATOMIC;\n+\n+\t\tif (enqueue_timeout(0, &ev, 1000)) {\n+\t\t\tprintf(\"%d: Error enqueuing events\\n\", __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Now the DLB is scheduling events from the port to the IQ, and at\n+\t * least one event should be remaining in each queue.\n+\t */\n+\n+\tif (rte_event_dev_stop_flush_callback_register(evdev, flush, NULL)) {\n+\t\tprintf(\"%d: Error installing the flush callback\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcleanup(t);\n+\n+\tif (count != rte_mempool_avail_count(t->mbuf_pool)) {\n+\t\tprintf(\"%d: Error executing the flush callback\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_dev_stop_flush_callback_register(evdev, NULL, NULL)) {\n+\t\tprintf(\"%d: Error uninstalling the flush callback\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\treturn 0;\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_single_link(void)\n+{\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tuint8_t queue_id;\n+\tint ret;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 2;\n+\tconfig.nb_event_ports = 2;\n+\tconfig.nb_single_link_event_port_queues = 1;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Create a directed port */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.event_port_cfg = RTE_EVENT_PORT_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Attempt to create another directed port */\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) == 0) {\n+\t\tprintf(\"%d: port 1 setup expected to fail\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.event_port_cfg = 0;\n+\n+\t/* Create a load-balanced port */\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Create a directed queue */\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_conf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Attempt to create another directed queue */\n+\tif (rte_event_queue_setup(evdev, 1, &queue_conf) == 0) {\n+\t\tprintf(\"%d: queue 1 setup expected to fail\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Create a load-balanced queue */\n+\tqueue_conf.event_queue_cfg = 0;\n+\n+\tif (rte_event_queue_setup(evdev, 1, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 1 setup expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Attempt to link directed and load-balanced resources */\n+\tqueue_id = 1;\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) == 1) {\n+\t\tprintf(\"%d: port 0 link expected to fail\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_id = 0;\n+\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) == 1) {\n+\t\tprintf(\"%d: port 1 link expected to fail\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link ports to queues */\n+\tqueue_id = 0;\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_id = 1;\n+\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 1 link expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tret = rte_event_dev_close(evdev);\n+\tif (ret)\n+\t\tprintf(\"%d: rte_event_dev_close failed, ret = %d\\n\",\n+\t\t\t__LINE__, ret);\n+\n+\treturn 0;\n+\n+err:\n+\tret = rte_event_dev_close(evdev);\n+\tif (ret)\n+\t\tprintf(\"%d: rte_event_dev_close failed, ret = %d\\n\",\n+\t\t\t__LINE__, ret);\n+\n+\treturn -1;\n+}\n+\n+#define NUM_LDB_PORTS 64\n+#define NUM_LDB_QUEUES 32\n+\n+static int\n+test_info_get(void)\n+{\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_dev_info info;\n+\tint ret;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tif (info.max_event_ports != NUM_LDB_PORTS) {\n+\t\tprintf(\"%d: Got %u ports, expected %u\\n\",\n+\t\t       __LINE__, info.max_event_ports, NUM_LDB_PORTS);\n+\t\tgoto err;\n+\t}\n+\n+\tif (info.max_event_queues != NUM_LDB_QUEUES) {\n+\t\tprintf(\"%d: Got %u queues, expected %u\\n\",\n+\t\t       __LINE__, info.max_event_queues, NUM_LDB_QUEUES);\n+\t\tgoto err;\n+\t}\n+\n+\tconfig.nb_event_ports = info.max_event_ports;\n+\tconfig.nb_event_queues = NUM_LDB_QUEUES + info.max_event_ports / 2;\n+\tconfig.nb_single_link_event_port_queues = info.max_event_ports / 2;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* The DLB2 PMD only reports load-balanced ports and queues in its\n+\t * info_get function. Confirm that these values don't include the\n+\t * directed port or queue counts.\n+\t */\n+\n+\tif (info.max_event_ports != NUM_LDB_PORTS) {\n+\t\tprintf(\"%d: Got %u ports, expected %u\\n\",\n+\t\t       __LINE__, info.max_event_ports, NUM_LDB_PORTS);\n+\t\tgoto err;\n+\t}\n+\n+\tif (info.max_event_queues != NUM_LDB_QUEUES) {\n+\t\tprintf(\"%d: Got %u queues, expected %u\\n\",\n+\t\t       __LINE__, info.max_event_queues, NUM_LDB_QUEUES);\n+\t\tgoto err;\n+\t}\n+\n+\tret = rte_event_dev_close(evdev);\n+\tif (ret) {\n+\t\tprintf(\"%d: rte_event_dev_close failed, ret = %d\\n\",\n+\t\t\t__LINE__, ret);\n+\t\treturn -1;\n+\t}\n+\treturn 0;\n+\n+err:\n+\tret = rte_event_dev_close(evdev);\n+\tif (ret)\n+\t\tprintf(\"%d: rte_event_dev_close failed, ret = %d\\n\",\n+\t\t\t__LINE__, ret);\n+\n+\treturn -1;\n+}\n+\n+static int\n+test_reconfiguration_link(struct test *t)\n+{\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tuint8_t queue_id;\n+\tint ret, i;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 2;\n+\tconfig.nb_event_ports = 2;\n+\tconfig.nb_single_link_event_port_queues = 0;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 2 LDB ports and 2 LDB queues */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tfor (i = 0; i < 2; i++) {\n+\t\tif (rte_event_port_setup(evdev, i, &port_conf) < 0) {\n+\t\t\tprintf(\"%d: port %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tfor (i = 0; i < 2; i++) {\n+\t\tif (rte_event_queue_setup(evdev, i, &queue_conf) < 0) {\n+\t\t\tprintf(\"%d: queue %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Link P0->Q0 and P1->Q1 */\n+\tfor (i = 0; i < 2; i++) {\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, i, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: port %d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Stop the device */\n+\trte_event_dev_stop(evdev);\n+\n+\t/* Reconfigure device */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error re-configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure P1 and Q1, leave P0 and Q0 to be configured by the PMD. */\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, 1, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 and Q1 */\n+\tfor (i = 0; i < 2; i++) {\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: P0->Q%d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Link P1->Q0 and Q1 */\n+\tfor (i = 0; i < 2; i++) {\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: P1->Q%d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Stop the device */\n+\trte_event_dev_stop(evdev);\n+\n+\t/* Configure device with 2 DIR ports and 2 DIR queues */\n+\tconfig.nb_single_link_event_port_queues = 2;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tport_conf.event_port_cfg = RTE_EVENT_PORT_CFG_SINGLE_LINK;\n+\n+\tfor (i = 0; i < 2; i++) {\n+\t\tif (rte_event_port_setup(evdev, i, &port_conf) < 0) {\n+\t\t\tprintf(\"%d: port %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tqueue_conf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tfor (i = 0; i < 2; i++) {\n+\t\tif (rte_event_queue_setup(evdev, i, &queue_conf) < 0) {\n+\t\t\tprintf(\"%d: queue %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Link P0->Q0 and P1->Q1 */\n+\tfor (i = 0; i < 2; i++) {\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, i, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: port %d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Stop the device */\n+\trte_event_dev_stop(evdev);\n+\n+\t/* Reconfigure device */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error re-configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure P1 and Q0, leave P0 and Q1 to be configured by the PMD. */\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q1 */\n+\tqueue_id = 1;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: P0->Q%d link expected to succeed\\n\",\n+\t\t       __LINE__, i);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P1->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: P1->Q%d link expected to succeed\\n\",\n+\t\t       __LINE__, i);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\trte_event_dev_stop(evdev);\n+\n+\tconfig.nb_event_queues = 5;\n+\tconfig.nb_event_ports = 5;\n+\tconfig.nb_single_link_event_port_queues = 1;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error re-configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tfor (i = 0; i < config.nb_event_queues - 1; i++) {\n+\t\tport_conf.event_port_cfg = 0;\n+\t\tqueue_conf.event_queue_cfg = 0;\n+\n+\t\tif (rte_event_port_setup(evdev, i, &port_conf) < 0) {\n+\t\t\tprintf(\"%d: port %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\n+\t\tif (rte_event_queue_setup(evdev, i, &queue_conf) < 0) {\n+\t\t\tprintf(\"%d: queue %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, i, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: P%d->Q%d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tport_conf.event_port_cfg = RTE_EVENT_PORT_CFG_SINGLE_LINK;\n+\tqueue_conf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_port_setup(evdev, i, &port_conf) < 0) {\n+\t\tprintf(\"%d: port %d setup expected to succeed\\n\",\n+\t\t       __LINE__, i);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, i, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue %d setup expected to succeed\\n\",\n+\t\t       __LINE__, i);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_id = i;\n+\n+\tif (rte_event_port_link(evdev, i, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: P%d->Q%d link expected to succeed\\n\",\n+\t\t       __LINE__, i, i);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Stop the device */\n+\trte_event_dev_stop(evdev);\n+\n+\tconfig.nb_event_ports += 1;\n+\n+\t/* Reconfigure device with 1 more load-balanced port */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error re-configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tport_conf.event_port_cfg = 0;\n+\n+\t/* Configure the new port */\n+\tif (rte_event_port_setup(evdev, config.nb_event_ports - 1,\n+\t\t\t\t &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_load_balanced_traffic(struct test *t)\n+{\n+\tuint64_t timeout;\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tstruct rte_event ev;\n+\tuint8_t queue_id;\n+\tint ret;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 1;\n+\tconfig.nb_event_ports = 1;\n+\tconfig.nb_single_link_event_port_queues = 0;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 1 LDB port and queue */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Enqueue 1 NEW event */\n+\tev.op = RTE_EVENT_OP_NEW;\n+\tev.sched_type = RTE_SCHED_TYPE_ATOMIC;\n+\tev.queue_id = 0;\n+\tev.priority = 0;\n+\tev.u64 = 0;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue and enqueue 1 FORWARD event */\n+\ttimeout = 0xFFFFFFFFF;\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_FORWARD;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue and enqueue 1 RELEASE operation */\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_directed_traffic(struct test *t)\n+{\n+\tuint64_t timeout;\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tstruct rte_event ev;\n+\tuint8_t queue_id;\n+\tint ret;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 1;\n+\tconfig.nb_event_ports = 1;\n+\tconfig.nb_single_link_event_port_queues = 1;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 1 DIR port and queue */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.event_port_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_conf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Enqueue 1 NEW event */\n+\tev.op = RTE_EVENT_OP_NEW;\n+\tev.queue_id = 0;\n+\tev.priority = 0;\n+\tev.u64 = 0;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue and enqueue 1 FORWARD event */\n+\ttimeout = 0xFFFFFFFFF;\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (ev.queue_id != 0) {\n+\t\tprintf(\"%d: invalid dequeued event queue ID (%d)\\n\",\n+\t\t       __LINE__, ev.queue_id);\n+\t\tgoto err;\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_FORWARD;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue and enqueue 1 RELEASE operation */\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_deferred_sched(struct test *t)\n+{\n+\tuint64_t timeout;\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tconst int num_events = 128;\n+\tstruct rte_event ev;\n+\tuint8_t queue_id;\n+\tint ret, i;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 1;\n+\tconfig.nb_event_ports = 2;\n+\tconfig.nb_single_link_event_port_queues = 0;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 2 LDB ports and 1 queue */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tret = rte_pmd_dlb2_set_token_pop_mode(evdev, 0, DEFERRED_POP);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error setting deferred scheduling\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tret = rte_pmd_dlb2_set_token_pop_mode(evdev, 1, DEFERRED_POP);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error setting deferred scheduling\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.dequeue_depth = 1;\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_conf.schedule_type = RTE_SCHED_TYPE_PARALLEL;\n+\tqueue_conf.nb_atomic_order_sequences = 0;\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 and P1->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 1 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Enqueue 128 NEW events */\n+\tev.op = RTE_EVENT_OP_NEW;\n+\tev.sched_type = RTE_SCHED_TYPE_PARALLEL;\n+\tev.queue_id = 0;\n+\tev.priority = 0;\n+\tev.u64 = 0;\n+\n+\tfor (i = 0; i < num_events; i++) {\n+\t\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Dequeue one event from port 0 */\n+\ttimeout = 0xFFFFFFFFF;\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue (and release) all other events from port 1. Deferred\n+\t * scheduling ensures no other events are scheduled to port 0 without a\n+\t * subsequent rte_event_dequeue_burst() call.\n+\t */\n+\tfor (i = 0; i < num_events - 1; i++) {\n+\t\tif (rte_event_dequeue_burst(evdev, 1, &ev, 1, timeout) != 1) {\n+\t\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\n+\t\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\t\tif (rte_event_enqueue_burst(evdev, 1, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: RELEASE enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_delayed_pop(struct test *t)\n+{\n+\tuint64_t timeout;\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tint ret, i, num_events;\n+\tstruct rte_event ev;\n+\tuint8_t queue_id;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 1;\n+\tconfig.nb_event_ports = 1;\n+\tconfig.nb_single_link_event_port_queues = 0;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 1 LDB port and queue */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tret = rte_pmd_dlb2_set_token_pop_mode(evdev, 0, DELAYED_POP);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error setting deferred scheduling\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.event_port_cfg = RTE_EVENT_PORT_CFG_DISABLE_IMPL_REL;\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tnum_events = 2 * port_conf.dequeue_depth;\n+\n+\t/* Enqueue 2 * dequeue_depth NEW events */\n+\tev.op = RTE_EVENT_OP_NEW;\n+\tev.sched_type = RTE_SCHED_TYPE_ATOMIC;\n+\tev.queue_id = 0;\n+\tev.priority = 0;\n+\tev.u64 = 0;\n+\n+\tfor (i = 0; i < num_events; i++) {\n+\t\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Dequeue dequeue_depth events but only release dequeue_depth - 2.\n+\t * Delayed pop won't perform the pop and no more events will be\n+\t * scheduled.\n+\t */\n+\ttimeout = 0xFFFFFFFFF;\n+\n+\tfor (i = 0; i < port_conf.dequeue_depth; i++) {\n+\t\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\tfor (i = 0; i < port_conf.dequeue_depth - 2; i++) {\n+\t\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: RELEASE enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\ttimeout = 0x10000;\n+\n+\tret = rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout);\n+\tif (ret != 0) {\n+\t\tprintf(\"%d: event dequeue expected to fail (ret = %d)\\n\",\n+\t\t       __LINE__, ret);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Release one more event. This will trigger the token pop, and\n+\t * dequeue_depth - 1 more events will be scheduled to the device.\n+\t */\n+\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: RELEASE enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\ttimeout = 0xFFFFFFFFF;\n+\n+\tfor (i = 0; i < port_conf.dequeue_depth - 1; i++) {\n+\t\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\ttimeout = 0x10000;\n+\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 0) {\n+\t\tprintf(\"%d: event dequeue expected to fail\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+do_selftest(void)\n+{\n+\tstruct test t;\n+\tint ret;\n+\n+\t/* Only create mbuf pool once, reuse for each test run */\n+\tif (!eventdev_func_mempool) {\n+\t\teventdev_func_mempool =\n+\t\t\trte_pktmbuf_pool_create(\"EVENTDEV_DLB2_ST_POOL\",\n+\t\t\t\t\t\t(1 << 12), /* 4k buffers */\n+\t\t\t\t\t\t32 /*MBUF_CACHE_SIZE*/,\n+\t\t\t\t\t\t0,\n+\t\t\t\t\t\t512, /* use very small mbufs */\n+\t\t\t\t\t\trte_socket_id());\n+\t\tif (!eventdev_func_mempool) {\n+\t\t\tprintf(\"ERROR creating mempool\\n\");\n+\t\t\tgoto test_fail;\n+\t\t}\n+\t}\n+\tt.mbuf_pool = eventdev_func_mempool;\n+\n+\tprintf(\"*** Running Stop Flush test...\\n\");\n+\tret = test_stop_flush(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Stop Flush test FAILED.\\n\");\n+\t\treturn ret;\n+\t}\n+\n+\tprintf(\"*** Running Single Link test...\\n\");\n+\tret = test_single_link();\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Single Link test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Info Get test...\\n\");\n+\tret = test_info_get();\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Stop Flush test FAILED.\\n\");\n+\t\treturn ret;\n+\t}\n+\n+\tprintf(\"*** Running Reconfiguration Link test...\\n\");\n+\tret = test_reconfiguration_link(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Reconfiguration Link test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Load-Balanced Traffic test...\\n\");\n+\tret = test_load_balanced_traffic(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Load-Balanced Traffic test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Directed Traffic test...\\n\");\n+\tret = test_directed_traffic(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Directed Traffic test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Deferred Scheduling test...\\n\");\n+\tret = test_deferred_sched(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Deferred Scheduling test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Delayed Pop test...\\n\");\n+\tret = test_delayed_pop(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Delayed Pop test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\treturn 0;\n+\n+test_fail:\n+\treturn -1;\n+}\n+\n+int\n+test_dlb2_eventdev(void)\n+{\n+\tconst char *dlb2_eventdev_name = \"dlb2_event\";\n+\tuint8_t num_evdevs = rte_event_dev_count();\n+\tint i, ret = 0;\n+\tint found = 0, skipped = 0, passed = 0, failed = 0;\n+\tstruct rte_event_dev_info info;\n+\n+\tfor (i = 0; found + skipped < num_evdevs && i < RTE_EVENT_MAX_DEVS;\n+\t     i++) {\n+\t\tret = rte_event_dev_info_get(i, &info);\n+\t\tif (ret < 0)\n+\t\t\tcontinue;\n+\n+\t\t/* skip non-dlb2 event devices */\n+\t\tif (strncmp(info.driver_name, dlb2_eventdev_name,\n+\t\t\t    sizeof(*info.driver_name)) != 0) {\n+\t\t\tskipped++;\n+\t\t\tcontinue;\n+\t\t}\n+\n+\t\tevdev = rte_event_dev_get_dev_id(info.driver_name);\n+\t\tif (evdev < 0) {\n+\t\t\tprintf(\"Could not get dev_id for eventdev with name %s, i=%d\\n\",\n+\t\t\t       info.driver_name, i);\n+\t\t\tskipped++;\n+\t\t\tcontinue;\n+\t\t}\n+\t\tfound++;\n+\t\tprintf(\"Running selftest on eventdev %s\\n\", info.driver_name);\n+\t\tret = do_selftest();\n+\t\tif (ret == 0) {\n+\t\t\tpassed++;\n+\t\t\tprintf(\"Selftest passed for eventdev %s\\n\",\n+\t\t\t       info.driver_name);\n+\t\t} else {\n+\t\t\tfailed++;\n+\t\t\tprintf(\"Selftest failed for eventdev %s, err=%d\\n\",\n+\t\t\t       info.driver_name, ret);\n+\t\t}\n+\t}\n+\n+\tprintf(\"Ran selftest on %d eventdevs, %d skipped, %d passed, %d failed\\n\",\n+\t       found, skipped, passed, failed);\n+\treturn ret;\n+}\ndiff --git a/drivers/event/dlb2/meson.build b/drivers/event/dlb2/meson.build\nindex 4549a75..07a25dc 100644\n--- a/drivers/event/dlb2/meson.build\n+++ b/drivers/event/dlb2/meson.build\n@@ -8,7 +8,8 @@ sources = files('dlb2.c',\n \t\t'pf/dlb2_main.c',\n \t\t'pf/dlb2_pf.c',\n \t\t'pf/base/dlb2_resource.c',\n-\t\t'rte_pmd_dlb2.c'\n+\t\t'rte_pmd_dlb2.c',\n+\t\t'dlb2_selftest.c'\n )\n \n deps += ['mbuf', 'mempool', 'ring', 'bus_vdev', 'pci', 'bus_pci']\n",
    "prefixes": [
        "19/22"
    ]
}