From patchwork Wed Sep 28 18:44:19 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Sevincer, Abdullah" X-Patchwork-Id: 117073 X-Patchwork-Delegate: jerinj@marvell.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id E6640A00C4; Wed, 28 Sep 2022 20:44:26 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 83CE84113C; Wed, 28 Sep 2022 20:44:26 +0200 (CEST) Received: from mga05.intel.com (mga05.intel.com [192.55.52.43]) by mails.dpdk.org (Postfix) with ESMTP id D880E410FA; Wed, 28 Sep 2022 20:44:23 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1664390664; x=1695926664; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=j81YtsdMPDTBC4mma36yZEOqS7TG4sGjKnRNxMJz7lM=; b=R61Ipw7ktaS5Q0tJJ4SAfGc86dchFrsvngWPYTbx4mxQjL4bzRHToKMn jRO842tCrA7vJkQ6vzoytL/D22ODo8ygGEJc/2uv4+6oR9p7X9JRVdd2w 2hzGznGzVboXQjob7CMa7X2fN01vuA4iZ+ryvMvp9+8C/CxFFSMG+N1LM KYPEwwbCpVEeuHkpp8QZnMejkCo37nM4b9U64ku7/zYmtzLBmrg1OBlpV uCdLXFVI/UxpV4Ei8R1MvJfXANT9rwAI1iqZAGPck4Q+C4Lz+mJD0Ui4h wCPkxYUSI/0g0dXUZna8K9upS8owaErGyI9smNRBbr3jyJuV7RhjmFSzB w==; X-IronPort-AV: E=McAfee;i="6500,9779,10484"; a="387969051" X-IronPort-AV: E=Sophos;i="5.93,352,1654585200"; d="scan'208";a="387969051" Received: from fmsmga006.fm.intel.com ([10.253.24.20]) by fmsmga105.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 28 Sep 2022 11:44:22 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6500,9779,10484"; a="867077701" X-IronPort-AV: E=Sophos;i="5.93,352,1654585200"; d="scan'208";a="867077701" Received: from txanpdk02.an.intel.com ([10.123.117.76]) by fmsmga006.fm.intel.com with ESMTP; 28 Sep 2022 11:44:22 -0700 From: Abdullah Sevincer To: dev@dpdk.org Cc: jerinj@marvell.com, Abdullah Sevincer , stable@dpdk.org Subject: [PATCH v5] event/dlb2: fix max enqueue and dequeue cli override Date: Wed, 28 Sep 2022 13:44:19 -0500 Message-Id: <20220928184419.1618909-1-abdullah.sevincer@intel.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20220927150537.1464936-1-abdullah.sevincer@intel.com> References: <20220927150537.1464936-1-abdullah.sevincer@intel.com> MIME-Version: 1.0 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org This patch addresses an issue of enqueuing more than max_enq_depth and not able to dequeuing events equal to max_cq_depth in a single call of rte_event_enqueue_burst and rte_event_dequeue_burst. Apply fix for restricting enqueue of events to max_enq_depth so that in a single rte_event_enqueue_burst() call at most max_enq_depth events are enqueued. Also set per port and domain history list sizes based on cq_depth. This results in dequeuing correct number of events as set by max_cq_depth. Fixes: f3cad285bb88 ("event/dlb2: add infos get and configure") Cc: stable@dpdk.org Signed-off-by: Abdullah Sevincer --- drivers/event/dlb2/dlb2.c | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/drivers/event/dlb2/dlb2.c b/drivers/event/dlb2/dlb2.c index 759578378f..dbb8284135 100644 --- a/drivers/event/dlb2/dlb2.c +++ b/drivers/event/dlb2/dlb2.c @@ -813,7 +813,7 @@ dlb2_hw_create_sched_domain(struct dlb2_eventdev *dlb2, cfg->num_ldb_queues; cfg->num_hist_list_entries = resources_asked->num_ldb_ports * - DLB2_NUM_HIST_LIST_ENTRIES_PER_LDB_PORT; + evdev_dlb2_default_info.max_event_port_dequeue_depth; if (device_version == DLB2_HW_V2_5) { DLB2_LOG_DBG("sched domain create - ldb_qs=%d, ldb_ports=%d, dir_ports=%d, atomic_inflights=%d, hist_list_entries=%d, credits=%d\n", @@ -1538,7 +1538,7 @@ dlb2_hw_create_ldb_port(struct dlb2_eventdev *dlb2, cfg.cq_depth = rte_align32pow2(dequeue_depth); cfg.cq_depth_threshold = 1; - cfg.cq_history_list_size = DLB2_NUM_HIST_LIST_ENTRIES_PER_LDB_PORT; + cfg.cq_history_list_size = cfg.cq_depth; cfg.cos_id = ev_port->cos_id; cfg.cos_strict = 0;/* best effots */ @@ -2966,6 +2966,7 @@ __dlb2_event_enqueue_burst(void *event_port, struct dlb2_port *qm_port = &ev_port->qm_port; struct process_local_port_data *port_data; int retries = ev_port->enq_retries; + int num_tx; int i; RTE_ASSERT(ev_port->enq_configured); @@ -2974,8 +2975,8 @@ __dlb2_event_enqueue_burst(void *event_port, i = 0; port_data = &dlb2_port[qm_port->id][PORT_TYPE(qm_port)]; - - while (i < num) { + num_tx = RTE_MIN(num, ev_port->conf.enqueue_depth); + while (i < num_tx) { uint8_t sched_types[DLB2_NUM_QES_PER_CACHE_LINE]; uint8_t queue_ids[DLB2_NUM_QES_PER_CACHE_LINE]; int pop_offs = 0;