get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/77502/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 77502,
    "url": "http://patches.dpdk.org/api/patches/77502/?format=api",
    "web_url": "http://patches.dpdk.org/project/dpdk/patch/1599851920-16802-20-git-send-email-timothy.mcdaniel@intel.com/",
    "project": {
        "id": 1,
        "url": "http://patches.dpdk.org/api/projects/1/?format=api",
        "name": "DPDK",
        "link_name": "dpdk",
        "list_id": "dev.dpdk.org",
        "list_email": "dev@dpdk.org",
        "web_url": "http://core.dpdk.org",
        "scm_url": "git://dpdk.org/dpdk",
        "webscm_url": "http://git.dpdk.org/dpdk",
        "list_archive_url": "https://inbox.dpdk.org/dev",
        "list_archive_url_format": "https://inbox.dpdk.org/dev/{}",
        "commit_url_format": ""
    },
    "msgid": "<1599851920-16802-20-git-send-email-timothy.mcdaniel@intel.com>",
    "list_archive_url": "https://inbox.dpdk.org/dev/1599851920-16802-20-git-send-email-timothy.mcdaniel@intel.com",
    "date": "2020-09-11T19:18:37",
    "name": "[v4,19/22] event/dlb: add PMD self-tests",
    "commit_ref": null,
    "pull_url": null,
    "state": "superseded",
    "archived": true,
    "hash": "fdaef18305e0f3b046573f20886d463978f74b5e",
    "submitter": {
        "id": 826,
        "url": "http://patches.dpdk.org/api/people/826/?format=api",
        "name": "Timothy McDaniel",
        "email": "timothy.mcdaniel@intel.com"
    },
    "delegate": {
        "id": 310,
        "url": "http://patches.dpdk.org/api/users/310/?format=api",
        "username": "jerin",
        "first_name": "Jerin",
        "last_name": "Jacob",
        "email": "jerinj@marvell.com"
    },
    "mbox": "http://patches.dpdk.org/project/dpdk/patch/1599851920-16802-20-git-send-email-timothy.mcdaniel@intel.com/mbox/",
    "series": [
        {
            "id": 12163,
            "url": "http://patches.dpdk.org/api/series/12163/?format=api",
            "web_url": "http://patches.dpdk.org/project/dpdk/list/?series=12163",
            "date": "2020-09-11T19:18:18",
            "name": "Add DLB PMD",
            "version": 4,
            "mbox": "http://patches.dpdk.org/series/12163/mbox/"
        }
    ],
    "comments": "http://patches.dpdk.org/api/patches/77502/comments/",
    "check": "success",
    "checks": "http://patches.dpdk.org/api/patches/77502/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<dev-bounces@dpdk.org>",
        "X-Original-To": "patchwork@inbox.dpdk.org",
        "Delivered-To": "patchwork@inbox.dpdk.org",
        "Received": [
            "from dpdk.org (dpdk.org [92.243.14.124])\n\tby inbox.dpdk.org (Postfix) with ESMTP id 6D054A04C1;\n\tFri, 11 Sep 2020 21:25:12 +0200 (CEST)",
            "from [92.243.14.124] (localhost [127.0.0.1])\n\tby dpdk.org (Postfix) with ESMTP id 768A11C231;\n\tFri, 11 Sep 2020 21:22:33 +0200 (CEST)",
            "from mga18.intel.com (mga18.intel.com [134.134.136.126])\n by dpdk.org (Postfix) with ESMTP id C25851C1C3\n for <dev@dpdk.org>; Fri, 11 Sep 2020 21:22:17 +0200 (CEST)",
            "from fmsmga005.fm.intel.com ([10.253.24.32])\n by orsmga106.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384;\n 11 Sep 2020 12:22:16 -0700",
            "from txasoft-yocto.an.intel.com ([10.123.72.192])\n by fmsmga005.fm.intel.com with ESMTP; 11 Sep 2020 12:22:16 -0700"
        ],
        "IronPort-SDR": [
            "\n G/yjg1nuAdkBPXk9NvVjC/Qe8Fmx88pTfjwwRXumsbe5U/IyM8PmEU48HuCbun+KUPRNUSpz2l\n THAh8iiHmPfg==",
            "\n 01jXHoqRKzbFLZD3Yu2ipwMiURGqTawKHWUrCYH5QUeXReZB5sOMEvkXtMlbi0GkFzwVQywEUX\n n9OhgXEBziLQ=="
        ],
        "X-IronPort-AV": [
            "E=McAfee;i=\"6000,8403,9741\"; a=\"146570564\"",
            "E=Sophos;i=\"5.76,416,1592895600\"; d=\"scan'208\";a=\"146570564\"",
            "E=Sophos;i=\"5.76,416,1592895600\"; d=\"scan'208\";a=\"506375719\""
        ],
        "X-Amp-Result": "SKIPPED(no attachment in message)",
        "X-Amp-File-Uploaded": "False",
        "X-ExtLoop1": "1",
        "From": "Timothy McDaniel <timothy.mcdaniel@intel.com>",
        "To": "Jerin Jacob <jerinj@marvell.com>",
        "Cc": "dev@dpdk.org, erik.g.carrillo@intel.com, gage.eads@intel.com,\n harry.van.haaren@intel.com",
        "Date": "Fri, 11 Sep 2020 14:18:37 -0500",
        "Message-Id": "<1599851920-16802-20-git-send-email-timothy.mcdaniel@intel.com>",
        "X-Mailer": "git-send-email 1.7.10",
        "In-Reply-To": "<1599851920-16802-1-git-send-email-timothy.mcdaniel@intel.com>",
        "References": "<1599851920-16802-1-git-send-email-timothy.mcdaniel@intel.com>",
        "Subject": "[dpdk-dev] [PATCH v4 19/22] event/dlb: add PMD self-tests",
        "X-BeenThere": "dev@dpdk.org",
        "X-Mailman-Version": "2.1.15",
        "Precedence": "list",
        "List-Id": "DPDK patches and discussions <dev.dpdk.org>",
        "List-Unsubscribe": "<https://mails.dpdk.org/options/dev>,\n <mailto:dev-request@dpdk.org?subject=unsubscribe>",
        "List-Archive": "<http://mails.dpdk.org/archives/dev/>",
        "List-Post": "<mailto:dev@dpdk.org>",
        "List-Help": "<mailto:dev-request@dpdk.org?subject=help>",
        "List-Subscribe": "<https://mails.dpdk.org/listinfo/dev>,\n <mailto:dev-request@dpdk.org?subject=subscribe>",
        "Errors-To": "dev-bounces@dpdk.org",
        "Sender": "\"dev\" <dev-bounces@dpdk.org>"
    },
    "content": "Add a variety of self-tests for both ldb and directed\nports/queues, as well as configure, start, stop, link, etc...\n\nSigned-off-by: Timothy McDaniel <timothy.mcdaniel@intel.com>\n---\n app/test/test_eventdev.c         |    8 +\n drivers/event/dlb/dlb.c          |    1 +\n drivers/event/dlb/dlb_selftest.c | 1563 ++++++++++++++++++++++++++++++++++++++\n drivers/event/dlb/meson.build    |    1 +\n 4 files changed, 1573 insertions(+)\n create mode 100644 drivers/event/dlb/dlb_selftest.c",
    "diff": "diff --git a/app/test/test_eventdev.c b/app/test/test_eventdev.c\nindex 62019c1..2f6ad49 100644\n--- a/app/test/test_eventdev.c\n+++ b/app/test/test_eventdev.c\n@@ -1030,6 +1030,13 @@ test_eventdev_selftest_dpaa2(void)\n \treturn test_eventdev_selftest_impl(\"event_dpaa2\", \"\");\n }\n \n+static int\n+test_eventdev_selftest_dlb(void)\n+{\n+\treturn test_eventdev_selftest_impl(\"dlb_event\", \"\");\n+}\n+\n+\n REGISTER_TEST_COMMAND(eventdev_common_autotest, test_eventdev_common);\n REGISTER_TEST_COMMAND(eventdev_selftest_sw, test_eventdev_selftest_sw);\n REGISTER_TEST_COMMAND(eventdev_selftest_octeontx,\n@@ -1037,3 +1044,4 @@ REGISTER_TEST_COMMAND(eventdev_selftest_octeontx,\n REGISTER_TEST_COMMAND(eventdev_selftest_octeontx2,\n \t\ttest_eventdev_selftest_octeontx2);\n REGISTER_TEST_COMMAND(eventdev_selftest_dpaa2, test_eventdev_selftest_dpaa2);\n+REGISTER_TEST_COMMAND(eventdev_selftest_dlb, test_eventdev_selftest_dlb);\ndiff --git a/drivers/event/dlb/dlb.c b/drivers/event/dlb/dlb.c\nindex 1166aa3..eabc123 100644\n--- a/drivers/event/dlb/dlb.c\n+++ b/drivers/event/dlb/dlb.c\n@@ -3896,6 +3896,7 @@ dlb_entry_points_init(struct rte_eventdev *dev)\n \t\t.xstats_get_names = dlb_eventdev_xstats_get_names,\n \t\t.xstats_get_by_name = dlb_eventdev_xstats_get_by_name,\n \t\t.xstats_reset\t    = dlb_eventdev_xstats_reset,\n+\t\t.dev_selftest     = test_dlb_eventdev,\n \t};\n \n \t/* Expose PMD's eventdev interface */\ndiff --git a/drivers/event/dlb/dlb_selftest.c b/drivers/event/dlb/dlb_selftest.c\nnew file mode 100644\nindex 0000000..ebdc54c\n--- /dev/null\n+++ b/drivers/event/dlb/dlb_selftest.c\n@@ -0,0 +1,1563 @@\n+/* SPDX-License-Identifier: BSD-3-Clause\n+ * Copyright(c) 2016-2020 Intel Corporation\n+ */\n+\n+#include <stdio.h>\n+#include <string.h>\n+#include <stdint.h>\n+#include <errno.h>\n+#include <unistd.h>\n+#include <sys/queue.h>\n+\n+#include <rte_memory.h>\n+#include <rte_memzone.h>\n+#include <rte_launch.h>\n+#include <rte_eal.h>\n+#include <rte_lcore.h>\n+#include <rte_debug.h>\n+#include <rte_cycles.h>\n+#include <rte_eventdev.h>\n+#include <rte_mempool.h>\n+#include <rte_mbuf.h>\n+\n+#include \"dlb_priv.h\"\n+#include \"rte_pmd_dlb.h\"\n+\n+#define MAX_PORTS 32\n+#define MAX_QIDS 32\n+#define DEFAULT_NUM_SEQ_NUMS 32\n+\n+static struct rte_mempool *eventdev_func_mempool;\n+static int evdev;\n+\n+struct test {\n+\tstruct rte_mempool *mbuf_pool;\n+\tint nb_qids;\n+};\n+\n+/* initialization and config */\n+static inline int\n+init(struct test *t, int nb_queues, int nb_ports)\n+{\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_dev_info info;\n+\tint ret;\n+\n+\tmemset(t, 0, sizeof(*t));\n+\n+\tt->mbuf_pool = eventdev_func_mempool;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = nb_queues;\n+\tconfig.nb_event_ports = nb_ports;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0)\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\n+\treturn ret;\n+}\n+\n+static inline int\n+create_ports(int num_ports)\n+{\n+\tint i;\n+\n+\tif (num_ports > MAX_PORTS)\n+\t\treturn -1;\n+\n+\tfor (i = 0; i < num_ports; i++) {\n+\t\tstruct rte_event_port_conf conf;\n+\n+\t\tif (rte_event_port_default_conf_get(evdev, i, &conf)) {\n+\t\t\tprintf(\"%d: Error querying default port conf\\n\",\n+\t\t\t       __LINE__);\n+\t\t\treturn -1;\n+\t\t}\n+\n+\t\tif (rte_event_port_setup(evdev, i, &conf) < 0) {\n+\t\t\tprintf(\"%d: Error setting up port %d\\n\", i, __LINE__);\n+\t\t\treturn -1;\n+\t\t}\n+\t}\n+\n+\treturn 0;\n+}\n+\n+static inline int\n+create_lb_qids(struct test *t, int num_qids, uint32_t flags)\n+{\n+\tint i;\n+\n+\tfor (i = t->nb_qids; i < t->nb_qids + num_qids; i++) {\n+\t\tstruct rte_event_queue_conf conf;\n+\n+\t\tif (rte_event_queue_default_conf_get(evdev, i, &conf)) {\n+\t\t\tprintf(\"%d: Error querying default queue conf\\n\",\n+\t\t\t       __LINE__);\n+\t\t\treturn -1;\n+\t\t}\n+\n+\t\tconf.schedule_type = flags;\n+\n+\t\tif (conf.schedule_type == RTE_SCHED_TYPE_PARALLEL)\n+\t\t\tconf.nb_atomic_order_sequences = 0;\n+\t\telse\n+\t\t\tconf.nb_atomic_order_sequences = DEFAULT_NUM_SEQ_NUMS;\n+\n+\t\tif (rte_event_queue_setup(evdev, i, &conf) < 0) {\n+\t\t\tprintf(\"%d: error creating qid %d\\n\", __LINE__, i);\n+\t\t\treturn -1;\n+\t\t}\n+\t}\n+\n+\tt->nb_qids += num_qids;\n+\tif (t->nb_qids > MAX_QIDS)\n+\t\treturn -1;\n+\n+\treturn 0;\n+}\n+\n+static inline int\n+create_atomic_qids(struct test *t, int num_qids)\n+{\n+\treturn create_lb_qids(t, num_qids, RTE_SCHED_TYPE_ATOMIC);\n+}\n+\n+static inline int\n+create_ordered_qids(struct test *t, int num_qids)\n+{\n+\treturn create_lb_qids(t, num_qids, RTE_SCHED_TYPE_ORDERED);\n+}\n+\n+static inline int\n+create_unordered_qids(struct test *t, int num_qids)\n+{\n+\treturn create_lb_qids(t, num_qids, RTE_SCHED_TYPE_PARALLEL);\n+}\n+\n+/* destruction */\n+static inline int\n+cleanup(struct test *t __rte_unused)\n+{\n+\tint ret;\n+\n+\trte_event_dev_stop(evdev);\n+\tret = rte_event_dev_close(evdev);\n+\tif (ret)\n+\t\treturn -1;\n+\n+\treturn 0;\n+};\n+\n+static inline int\n+enqueue_timeout(uint8_t port_id, struct rte_event *ev, uint64_t tmo_us)\n+{\n+\tconst uint64_t start = rte_get_timer_cycles();\n+\tconst uint64_t ticks = (tmo_us * rte_get_timer_hz()) / 1E6;\n+\n+\twhile ((rte_get_timer_cycles() - start) < ticks) {\n+\t\tif (rte_event_enqueue_burst(evdev, port_id, ev, 1) == 1)\n+\t\t\treturn 0;\n+\n+\t\tif (rte_errno != -ENOSPC)\n+\t\t\treturn -1;\n+\t}\n+\n+\treturn -1;\n+}\n+\n+static void\n+flush(uint8_t id __rte_unused, struct rte_event event, void *arg __rte_unused)\n+{\n+\trte_pktmbuf_free(event.mbuf);\n+}\n+\n+static int\n+test_stop_flush(struct test *t) /* test to check we can properly flush events */\n+{\n+\tstruct rte_event ev;\n+\tuint32_t dequeue_depth;\n+\tunsigned int i, count;\n+\tuint8_t queue_id;\n+\n+\tev.op = RTE_EVENT_OP_NEW;\n+\n+\tif (init(t, 2, 1) < 0 ||\n+\t    create_ports(1) < 0 ||\n+\t    create_atomic_qids(t, 2) < 0) {\n+\t\tprintf(\"%d: Error initializing device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tif (rte_event_port_link(evdev, 0, NULL, NULL, 0) != 2) {\n+\t\tprintf(\"%d: Error linking queues to the port\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: Error with start call\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Unlink queue 1 so the PMD's stop callback has to cleanup an unlinked\n+\t * queue.\n+\t */\n+\tqueue_id = 1;\n+\n+\tif (rte_event_port_unlink(evdev, 0, &queue_id, 1) != 1) {\n+\t\tprintf(\"%d: Error unlinking queue 1 from port\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcount = rte_mempool_avail_count(t->mbuf_pool);\n+\n+\tif (rte_event_port_attr_get(evdev,\n+\t\t\t\t    0,\n+\t\t\t\t    RTE_EVENT_PORT_ATTR_DEQ_DEPTH,\n+\t\t\t\t    &dequeue_depth)) {\n+\t\tprintf(\"%d: Error retrieveing dequeue depth\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Send QEs to queue 0 */\n+\tfor (i = 0; i < dequeue_depth + 1; i++) {\n+\t\tev.mbuf = rte_pktmbuf_alloc(t->mbuf_pool);\n+\t\tev.queue_id = 0;\n+\t\tev.sched_type = RTE_SCHED_TYPE_ATOMIC;\n+\n+\t\tif (enqueue_timeout(0, &ev, 1000)) {\n+\t\t\tprintf(\"%d: Error enqueuing events\\n\", __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Send QEs to queue 1 */\n+\tfor (i = 0; i < dequeue_depth + 1; i++) {\n+\t\tev.mbuf = rte_pktmbuf_alloc(t->mbuf_pool);\n+\t\tev.queue_id = 1;\n+\t\tev.sched_type = RTE_SCHED_TYPE_ATOMIC;\n+\n+\t\tif (enqueue_timeout(0, &ev, 1000)) {\n+\t\t\tprintf(\"%d: Error enqueuing events\\n\", __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Now the DLB is scheduling events from the port to the IQ, and at\n+\t * least one event should be remaining in each queue.\n+\t */\n+\n+\tif (rte_event_dev_stop_flush_callback_register(evdev, flush, NULL)) {\n+\t\tprintf(\"%d: Error installing the flush callback\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcleanup(t);\n+\n+\tif (count != rte_mempool_avail_count(t->mbuf_pool)) {\n+\t\tprintf(\"%d: Error executing the flush callback\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_dev_stop_flush_callback_register(evdev, NULL, NULL)) {\n+\t\tprintf(\"%d: Error uninstalling the flush callback\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\treturn 0;\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_single_link(void)\n+{\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tuint8_t queue_id;\n+\tint ret;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 2;\n+\tconfig.nb_event_ports = 2;\n+\tconfig.nb_single_link_event_port_queues = 1;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Create a directed port */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.event_port_cfg = RTE_EVENT_PORT_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Attempt to create another directed port */\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) == 0) {\n+\t\tprintf(\"%d: port 1 setup expected to fail\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.event_port_cfg = 0;\n+\n+\t/* Create a load-balanced port */\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Create a directed queue */\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_conf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Attempt to create another directed queue */\n+\tif (rte_event_queue_setup(evdev, 1, &queue_conf) == 0) {\n+\t\tprintf(\"%d: queue 1 setup expected to fail\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Create a load-balanced queue */\n+\tqueue_conf.event_queue_cfg = 0;\n+\n+\tif (rte_event_queue_setup(evdev, 1, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 1 setup expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Attempt to link directed and load-balanced resources */\n+\tqueue_id = 1;\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) == 1) {\n+\t\tprintf(\"%d: port 0 link expected to fail\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_id = 0;\n+\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) == 1) {\n+\t\tprintf(\"%d: port 1 link expected to fail\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link ports to queues */\n+\tqueue_id = 0;\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_id = 1;\n+\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 1 link expected to succeed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tret = rte_event_dev_close(evdev);\n+\tif (ret) {\n+\t\tprintf(\"rte_event_dev_close err %d\\n\", ret);\n+\t\tgoto err;\n+\t}\n+\n+\treturn 0;\n+\n+err:\n+\trte_event_dev_close(evdev);\n+\treturn -1;\n+}\n+\n+#define NUM_LDB_PORTS 64\n+#define NUM_LDB_QUEUES 128\n+\n+static int\n+test_info_get(void)\n+{\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_dev_info info;\n+\tint ret;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tif (info.max_event_ports != NUM_LDB_PORTS) {\n+\t\tprintf(\"%d: Got %u ports, expected %u\\n\",\n+\t\t       __LINE__, info.max_event_ports, NUM_LDB_PORTS);\n+\t\tgoto err;\n+\t}\n+\n+\tif (info.max_event_queues != NUM_LDB_QUEUES) {\n+\t\tprintf(\"%d: Got %u queues, expected %u\\n\",\n+\t\t       __LINE__, info.max_event_queues, NUM_LDB_QUEUES);\n+\t\tgoto err;\n+\t}\n+\n+\tconfig.nb_event_ports = info.max_event_ports;\n+\tconfig.nb_event_queues = NUM_LDB_QUEUES + info.max_event_ports / 2;\n+\tconfig.nb_single_link_event_port_queues = info.max_event_ports / 2;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* The DLB PMD only reports load-balanced ports and queues in its\n+\t * info_get function. Confirm that these values don't include the\n+\t * directed port or queue counts.\n+\t */\n+\n+\tif (info.max_event_ports != NUM_LDB_PORTS) {\n+\t\tprintf(\"%d: Got %u ports, expected %u\\n\",\n+\t\t       __LINE__, info.max_event_ports, NUM_LDB_PORTS);\n+\t\tgoto err;\n+\t}\n+\n+\tif (info.max_event_queues != NUM_LDB_QUEUES) {\n+\t\tprintf(\"%d: Got %u queues, expected %u\\n\",\n+\t\t       __LINE__, info.max_event_queues, NUM_LDB_QUEUES);\n+\t\tgoto err;\n+\t}\n+\n+\tret = rte_event_dev_close(evdev);\n+\tif (ret) {\n+\t\tprintf(\"rte_event_dev_close err %d\\n\", ret);\n+\t\tgoto err;\n+\t}\n+\n+\treturn 0;\n+\n+err:\n+\trte_event_dev_close(evdev);\n+\treturn -1;\n+}\n+\n+static int\n+test_reconfiguration_link(struct test *t)\n+{\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tuint8_t queue_id;\n+\tint ret, i;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 2;\n+\tconfig.nb_event_ports = 2;\n+\tconfig.nb_single_link_event_port_queues = 0;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 2 LDB ports and 2 LDB queues */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tfor (i = 0; i < 2; i++) {\n+\t\tif (rte_event_port_setup(evdev, i, &port_conf) < 0) {\n+\t\t\tprintf(\"%d: port %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tfor (i = 0; i < 2; i++) {\n+\t\tif (rte_event_queue_setup(evdev, i, &queue_conf) < 0) {\n+\t\t\tprintf(\"%d: queue %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Link P0->Q0 and P1->Q1 */\n+\tfor (i = 0; i < 2; i++) {\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, i, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: port %d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Stop the device */\n+\trte_event_dev_stop(evdev);\n+\n+\t/* Reconfigure device */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error re-configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure P1 and Q1, leave P0 and Q0 to be configured by the PMD. */\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, 1, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 and Q1 */\n+\tfor (i = 0; i < 2; i++) {\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: P0->Q%d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Link P1->Q0 and Q1 */\n+\tfor (i = 0; i < 2; i++) {\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: P1->Q%d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Stop the device */\n+\trte_event_dev_stop(evdev);\n+\n+\t/* Configure device with 2 DIR ports and 2 DIR queues */\n+\tconfig.nb_single_link_event_port_queues = 2;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tport_conf.event_port_cfg = RTE_EVENT_PORT_CFG_SINGLE_LINK;\n+\n+\tfor (i = 0; i < 2; i++) {\n+\t\tif (rte_event_port_setup(evdev, i, &port_conf) < 0) {\n+\t\t\tprintf(\"%d: port %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tqueue_conf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tfor (i = 0; i < 2; i++) {\n+\t\tif (rte_event_queue_setup(evdev, i, &queue_conf) < 0) {\n+\t\t\tprintf(\"%d: queue %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Link P0->Q0 and P1->Q1 */\n+\tfor (i = 0; i < 2; i++) {\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, i, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: port %d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Stop the device */\n+\trte_event_dev_stop(evdev);\n+\n+\t/* Reconfigure device */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error re-configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure P1 and Q0, leave P0 and Q1 to be configured by the PMD. */\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q1 */\n+\tqueue_id = 1;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: P0->Q%d link expected to succeed\\n\",\n+\t\t       __LINE__, i);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P1->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: P1->Q%d link expected to succeed\\n\",\n+\t\t       __LINE__, i);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\trte_event_dev_stop(evdev);\n+\n+\tconfig.nb_event_queues = 5;\n+\tconfig.nb_event_ports = 5;\n+\tconfig.nb_single_link_event_port_queues = 1;\n+\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error re-configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tfor (i = 0; i < config.nb_event_queues - 1; i++) {\n+\t\tport_conf.event_port_cfg = 0;\n+\t\tqueue_conf.event_queue_cfg = 0;\n+\n+\t\tif (rte_event_port_setup(evdev, i, &port_conf) < 0) {\n+\t\t\tprintf(\"%d: port %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\n+\t\tif (rte_event_queue_setup(evdev, i, &queue_conf) < 0) {\n+\t\t\tprintf(\"%d: queue %d setup expected to succeed\\n\",\n+\t\t\t       __LINE__, i);\n+\t\t\tgoto err;\n+\t\t}\n+\n+\t\tqueue_id = i;\n+\n+\t\tif (rte_event_port_link(evdev, i, &queue_id, NULL, 1) != 1) {\n+\t\t\tprintf(\"%d: P%d->Q%d link expected to succeed\\n\",\n+\t\t\t       __LINE__, i, i);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tport_conf.event_port_cfg = RTE_EVENT_PORT_CFG_SINGLE_LINK;\n+\tqueue_conf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_port_setup(evdev, i, &port_conf) < 0) {\n+\t\tprintf(\"%d: port %d setup expected to succeed\\n\",\n+\t\t       __LINE__, i);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, i, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue %d setup expected to succeed\\n\",\n+\t\t       __LINE__, i);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_id = i;\n+\n+\tif (rte_event_port_link(evdev, i, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: P%d->Q%d link expected to succeed\\n\",\n+\t\t       __LINE__, i, i);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Stop the device */\n+\trte_event_dev_stop(evdev);\n+\n+\tconfig.nb_event_ports += 1;\n+\n+\t/* Reconfigure device with 1 more load-balanced port */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error re-configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tport_conf.event_port_cfg = 0;\n+\n+\t/* Configure the new port */\n+\tif (rte_event_port_setup(evdev, config.nb_event_ports - 1,\n+\t\t\t\t &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_load_balanced_traffic(struct test *t)\n+{\n+\tuint64_t timeout;\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tstruct rte_event ev;\n+\tuint8_t queue_id;\n+\tint ret;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 1;\n+\tconfig.nb_event_ports = 1;\n+\tconfig.nb_single_link_event_port_queues = 0;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 1 LDB port and queue */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Enqueue 1 NEW event */\n+\tev.op = RTE_EVENT_OP_NEW;\n+\tev.sched_type = RTE_SCHED_TYPE_ATOMIC;\n+\tev.queue_id = 0;\n+\tev.priority = 0;\n+\tev.u64 = 0;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue and enqueue 1 FORWARD event */\n+\ttimeout = 0xFFFFFFFFF;\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_FORWARD;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue and enqueue 1 RELEASE operation */\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_directed_traffic(struct test *t)\n+{\n+\tuint64_t timeout;\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tstruct rte_event ev;\n+\tuint8_t queue_id;\n+\tint ret;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 1;\n+\tconfig.nb_event_ports = 1;\n+\tconfig.nb_single_link_event_port_queues = 1;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 1 DIR port and queue */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.event_port_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_conf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK;\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Enqueue 1 NEW event */\n+\tev.op = RTE_EVENT_OP_NEW;\n+\tev.queue_id = 0;\n+\tev.priority = 0;\n+\tev.u64 = 0;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue and enqueue 1 FORWARD event */\n+\ttimeout = 0xFFFFFFFFF;\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (ev.queue_id != 0) {\n+\t\tprintf(\"%d: invalid dequeued event queue ID (%d)\\n\",\n+\t\t       __LINE__, ev.queue_id);\n+\t\tgoto err;\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_FORWARD;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue and enqueue 1 RELEASE operation */\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_deferred_sched(struct test *t)\n+{\n+\tuint64_t timeout;\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tconst int num_events = 128;\n+\tstruct rte_event ev;\n+\tuint8_t queue_id;\n+\tint ret, i;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 1;\n+\tconfig.nb_event_ports = 2;\n+\tconfig.nb_single_link_event_port_queues = 0;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 2 LDB ports and 1 queue */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tret = rte_pmd_dlb_set_token_pop_mode(evdev, 0, DEFERRED_POP);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error setting deferred scheduling\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tret = rte_pmd_dlb_set_token_pop_mode(evdev, 1, DEFERRED_POP);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error setting deferred scheduling\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.dequeue_depth = 1;\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_port_setup(evdev, 1, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 1 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tqueue_conf.schedule_type = RTE_SCHED_TYPE_PARALLEL;\n+\tqueue_conf.nb_atomic_order_sequences = 0;\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 and P1->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_port_link(evdev, 1, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 1 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Enqueue 128 NEW events */\n+\tev.op = RTE_EVENT_OP_NEW;\n+\tev.sched_type = RTE_SCHED_TYPE_PARALLEL;\n+\tev.queue_id = 0;\n+\tev.priority = 0;\n+\tev.u64 = 0;\n+\n+\tfor (i = 0; i < num_events; i++) {\n+\t\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Dequeue two events from port 0 (dequeue_depth * 2 due to the\n+\t * reserved token scheme)\n+\t */\n+\ttimeout = 0xFFFFFFFFF;\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Dequeue (and release) all other events from port 1. Deferred\n+\t * scheduling ensures no other events are scheduled to port 0 without a\n+\t * subsequent rte_event_dequeue_burst() call.\n+\t */\n+\tfor (i = 0; i < num_events - 2; i++) {\n+\t\tif (rte_event_dequeue_burst(evdev, 1, &ev, 1, timeout) != 1) {\n+\t\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\n+\t\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\t\tif (rte_event_enqueue_burst(evdev, 1, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: RELEASE enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+test_delayed_pop(struct test *t)\n+{\n+\tuint64_t timeout;\n+\tstruct rte_event_dev_config config = {0};\n+\tstruct rte_event_queue_conf queue_conf;\n+\tstruct rte_event_port_conf port_conf;\n+\tstruct rte_event_dev_info info;\n+\tint ret, i, num_events;\n+\tstruct rte_event ev;\n+\tuint8_t queue_id;\n+\n+\tif (rte_event_dev_info_get(evdev, &info)) {\n+\t\tprintf(\"%d: Error querying device info\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tconfig.nb_event_queues = 1;\n+\tconfig.nb_event_ports = 1;\n+\tconfig.nb_single_link_event_port_queues = 0;\n+\tconfig.nb_event_queue_flows = info.max_event_queue_flows;\n+\tconfig.nb_events_limit = info.max_num_events;\n+\tconfig.nb_event_port_dequeue_depth = info.max_event_port_dequeue_depth;\n+\tconfig.nb_event_port_enqueue_depth = info.max_event_port_enqueue_depth;\n+\tconfig.dequeue_timeout_ns = info.max_dequeue_timeout_ns;\n+\tconfig.event_dev_cfg = RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT;\n+\n+\t/* Configure the device with 1 LDB port and queue */\n+\tret = rte_event_dev_configure(evdev, &config);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error configuring device\\n\", __LINE__);\n+\t\treturn -1;\n+\t}\n+\n+\tret = rte_pmd_dlb_set_token_pop_mode(evdev, 0, DELAYED_POP);\n+\tif (ret < 0) {\n+\t\tprintf(\"%d: Error setting deferred scheduling\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Configure the ports and queues */\n+\tif (rte_event_port_default_conf_get(evdev, 0, &port_conf)) {\n+\t\tprintf(\"%d: Error querying default port conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tport_conf.dequeue_depth = 16;\n+\tport_conf.event_port_cfg = RTE_EVENT_PORT_CFG_DISABLE_IMPL_REL;\n+\n+\tif (rte_event_port_setup(evdev, 0, &port_conf) < 0) {\n+\t\tprintf(\"%d: port 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) {\n+\t\tprintf(\"%d: Error querying default queue conf\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tif (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) {\n+\t\tprintf(\"%d: queue 0 setup expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Link P0->Q0 */\n+\tqueue_id = 0;\n+\n+\tif (rte_event_port_link(evdev, 0, &queue_id, NULL, 1) != 1) {\n+\t\tprintf(\"%d: port 0 link expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Start the device */\n+\tif (rte_event_dev_start(evdev) < 0) {\n+\t\tprintf(\"%d: device start failed\\n\", __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\tnum_events = 2 * port_conf.dequeue_depth;\n+\n+\t/* Enqueue 2 * dequeue_depth NEW events. Due to the PMD's reserved\n+\t * token scheme, the port will initially behave as though its\n+\t * dequeue_depth is twice the requested size.\n+\t */\n+\tev.op = RTE_EVENT_OP_NEW;\n+\tev.sched_type = RTE_SCHED_TYPE_PARALLEL;\n+\tev.queue_id = 0;\n+\tev.priority = 0;\n+\tev.u64 = 0;\n+\n+\tfor (i = 0; i < num_events; i++) {\n+\t\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Flush these events out of the CQ */\n+\ttimeout = 0xFFFFFFFFF;\n+\n+\tfor (i = 0; i < num_events; i++) {\n+\t\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\tfor (i = 0; i < num_events; i++) {\n+\t\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: RELEASE enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Enqueue 2 * dequeue_depth NEW events again */\n+\tev.op = RTE_EVENT_OP_NEW;\n+\tev.sched_type = RTE_SCHED_TYPE_ATOMIC;\n+\tev.queue_id = 0;\n+\tev.priority = 0;\n+\tev.u64 = 0;\n+\n+\tfor (i = 0; i < num_events; i++) {\n+\t\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: NEW enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\t/* Dequeue dequeue_depth events but only release dequeue_depth - 1.\n+\t * Delayed pop won't perform the pop and no more events will be\n+\t * scheduled.\n+\t */\n+\tfor (i = 0; i < port_conf.dequeue_depth; i++) {\n+\t\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\tfor (i = 0; i < port_conf.dequeue_depth - 1; i++) {\n+\t\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\t\tprintf(\"%d: RELEASE enqueue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\ttimeout = 0x10000;\n+\n+\tret = rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout);\n+\tif (ret != 0) {\n+\t\tprintf(\"%d: event dequeue expected to fail (ret = %d)\\n\",\n+\t\t       __LINE__, ret);\n+\t\tgoto err;\n+\t}\n+\n+\t/* Release one more event. This will trigger the token pop, and\n+\t * another batch of events will be scheduled to the device.\n+\t */\n+\tev.op = RTE_EVENT_OP_RELEASE;\n+\n+\tif (rte_event_enqueue_burst(evdev, 0, &ev, 1) != 1) {\n+\t\tprintf(\"%d: RELEASE enqueue expected to succeed\\n\",\n+\t\t       __LINE__);\n+\t\tgoto err;\n+\t}\n+\n+\ttimeout = 0xFFFFFFFFF;\n+\n+\tfor (i = 0; i < port_conf.dequeue_depth; i++) {\n+\t\tif (rte_event_dequeue_burst(evdev, 0, &ev, 1, timeout) != 1) {\n+\t\t\tprintf(\"%d: event dequeue expected to succeed\\n\",\n+\t\t\t       __LINE__);\n+\t\t\tgoto err;\n+\t\t}\n+\t}\n+\n+\tcleanup(t);\n+\treturn 0;\n+\n+err:\n+\tcleanup(t);\n+\treturn -1;\n+}\n+\n+static int\n+do_selftest(void)\n+{\n+\tstruct test t;\n+\tint ret;\n+\n+\t/* Only create mbuf pool once, reuse for each test run */\n+\tif (!eventdev_func_mempool) {\n+\t\teventdev_func_mempool =\n+\t\t\trte_pktmbuf_pool_create(\"EVENTDEV_DLB_SA_MBUF_POOL\",\n+\t\t\t\t\t\t(1 << 12), /* 4k buffers */\n+\t\t\t\t\t\t32 /*MBUF_CACHE_SIZE*/,\n+\t\t\t\t\t\t0,\n+\t\t\t\t\t\t512, /* use very small mbufs */\n+\t\t\t\t\t\trte_socket_id());\n+\t\tif (!eventdev_func_mempool) {\n+\t\t\tprintf(\"ERROR creating mempool\\n\");\n+\t\t\tgoto test_fail;\n+\t\t}\n+\t}\n+\tt.mbuf_pool = eventdev_func_mempool;\n+\n+\tprintf(\"*** Running Stop Flush test...\\n\");\n+\tret = test_stop_flush(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Stop Flush test FAILED.\\n\");\n+\t\treturn ret;\n+\t}\n+\n+\tprintf(\"*** Running Single Link test...\\n\");\n+\tret = test_single_link();\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Single Link test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Info Get test...\\n\");\n+\tret = test_info_get();\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Stop Flush test FAILED.\\n\");\n+\t\treturn ret;\n+\t}\n+\n+\tprintf(\"*** Running Reconfiguration Link test...\\n\");\n+\tret = test_reconfiguration_link(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Reconfiguration Link test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Load-Balanced Traffic test...\\n\");\n+\tret = test_load_balanced_traffic(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Load-Balanced Traffic test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Directed Traffic test...\\n\");\n+\tret = test_directed_traffic(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Directed Traffic test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Deferred Scheduling test...\\n\");\n+\tret = test_deferred_sched(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Deferred Scheduling test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\tprintf(\"*** Running Delayed Pop test...\\n\");\n+\tret = test_delayed_pop(&t);\n+\tif (ret != 0) {\n+\t\tprintf(\"ERROR - Delayed Pop test FAILED.\\n\");\n+\n+\t\tgoto test_fail;\n+\t}\n+\n+\treturn 0;\n+\n+test_fail:\n+\treturn -1;\n+}\n+\n+int\n+test_dlb_eventdev(void)\n+{\n+\tconst char *dlb_eventdev_name = \"dlb_event\";\n+\tuint8_t num_evdevs = rte_event_dev_count();\n+\tint i, ret = 0;\n+\tint found = 0, skipped = 0, passed = 0, failed = 0;\n+\tstruct rte_event_dev_info info;\n+\n+\tfor (i = 0; found + skipped < num_evdevs && i < RTE_EVENT_MAX_DEVS;\n+\t     i++) {\n+\t\tret = rte_event_dev_info_get(i, &info);\n+\t\tif (ret < 0)\n+\t\t\tcontinue;\n+\n+\t\t/* skip non-dlb event devices */\n+\t\tif (strncmp(info.driver_name, dlb_eventdev_name,\n+\t\t\t    sizeof(*info.driver_name)) != 0) {\n+\t\t\tskipped++;\n+\t\t\tcontinue;\n+\t\t}\n+\n+\t\tevdev = rte_event_dev_get_dev_id(info.driver_name);\n+\t\tif (evdev < 0) {\n+\t\t\tprintf(\"Could not get dev_id for eventdev with name %s, i=%d\\n\",\n+\t\t\t       info.driver_name, i);\n+\t\t\tskipped++;\n+\t\t\tcontinue;\n+\t\t}\n+\t\tfound++;\n+\t\tprintf(\"Running selftest on eventdev %s\\n\", info.driver_name);\n+\t\tret = do_selftest();\n+\t\tif (ret == 0) {\n+\t\t\tpassed++;\n+\t\t\tprintf(\"Selftest passed for eventdev %s\\n\",\n+\t\t\t       info.driver_name);\n+\t\t} else {\n+\t\t\tfailed++;\n+\t\t\tprintf(\"Selftest failed for eventdev %s, err=%d\\n\",\n+\t\t\t       info.driver_name, ret);\n+\t\t}\n+\t}\n+\n+\tprintf(\"Ran selftest on %d eventdevs, %d skipped, %d passed, %d failed\\n\",\n+\t       found, skipped, passed, failed);\n+\treturn ret;\n+}\ndiff --git a/drivers/event/dlb/meson.build b/drivers/event/dlb/meson.build\nindex d7aeb3b..9e8ef96 100644\n--- a/drivers/event/dlb/meson.build\n+++ b/drivers/event/dlb/meson.build\n@@ -8,6 +8,7 @@ sources = files('dlb.c',\n \t\t'pf/dlb_pf.c',\n \t\t'pf/base/dlb_resource.c',\n \t\t'rte_pmd_dlb.c',\n+\t\t'dlb_selftest.c'\n )\n \n deps += ['mbuf', 'mempool', 'ring', 'pci', 'bus_pci']\n",
    "prefixes": [
        "v4",
        "19/22"
    ]
}