event/octeontx: validate events requested against available
diff mbox series

Message ID 20200728182224.1359-1-pbhagavatula@marvell.com
State Accepted, archived
Delegated to: Jerin Jacob
Headers show
Series
  • event/octeontx: validate events requested against available
Related show

Checks

Context Check Description
ci/iol-mellanox-Performance success Performance Testing PASS
ci/iol-testing success Testing PASS
ci/iol-intel-Performance success Performance Testing PASS
ci/Intel-compilation success Compilation OK
ci/travis-robot success Travis build: passed
ci/checkpatch warning coding style issues

Commit Message

Pavan Nikhilesh Bhagavatula July 28, 2020, 6:22 p.m. UTC
From: Pavan Nikhilesh <pbhagavatula@marvell.com>

Validate events configured in ssopf against the total number of
events configured across all the RX/TIM event adapters.

Events available to ssopf can be reconfigured by passing the required
amount to kernel bootargs and are only limited by DRAM size.
Example:
	ssopf.max_events= 2097152

Cc: stable@dpdk.org

Signed-off-by: Pavan Nikhilesh <pbhagavatula@marvell.com>
---
 doc/guides/eventdevs/octeontx.rst    | 23 +++++--
 drivers/event/octeontx/ssovf_evdev.c | 99 +++++++++++++++++++++++++---
 drivers/event/octeontx/ssovf_evdev.h |  6 ++
 drivers/event/octeontx/timvf_evdev.c | 68 +++++++++++++++++--
 drivers/event/octeontx/timvf_evdev.h |  2 +
 5 files changed, 176 insertions(+), 22 deletions(-)

--
2.17.1

Comments

Jerin Jacob Oct. 4, 2020, 10:40 a.m. UTC | #1
On Tue, Jul 28, 2020 at 11:52 PM <pbhagavatula@marvell.com> wrote:
>
> From: Pavan Nikhilesh <pbhagavatula@marvell.com>
>
> Validate events configured in ssopf against the total number of
> events configured across all the RX/TIM event adapters.
>
> Events available to ssopf can be reconfigured by passing the required
> amount to kernel bootargs and are only limited by DRAM size.
> Example:
>         ssopf.max_events= 2097152
>
> Cc: stable@dpdk.org
>
> Signed-off-by: Pavan Nikhilesh <pbhagavatula@marvell.com>


Applied to dpdk-next-eventdev/for-main. Thanks.



> ---
>  doc/guides/eventdevs/octeontx.rst    | 23 +++++--
>  drivers/event/octeontx/ssovf_evdev.c | 99 +++++++++++++++++++++++++---
>  drivers/event/octeontx/ssovf_evdev.h |  6 ++
>  drivers/event/octeontx/timvf_evdev.c | 68 +++++++++++++++++--
>  drivers/event/octeontx/timvf_evdev.h |  2 +
>  5 files changed, 176 insertions(+), 22 deletions(-)
>
> diff --git a/doc/guides/eventdevs/octeontx.rst b/doc/guides/eventdevs/octeontx.rst
> index 9a3646db0..21d251341 100644
> --- a/doc/guides/eventdevs/octeontx.rst
> +++ b/doc/guides/eventdevs/octeontx.rst
> @@ -140,9 +140,22 @@ follows:
>  When timvf is used as Event timer adapter event schedule type
>  ``RTE_SCHED_TYPE_PARALLEL`` is not supported.
>
> -Max mempool size
> -~~~~~~~~~~~~~~~~
> +Max number of events
> +~~~~~~~~~~~~~~~~~~~~
>
> -Max mempool size when using OCTEON TX Eventdev (SSO) should be limited to 128K.
> -When running dpdk-test-eventdev on OCTEON TX the application can limit the
> -number of mbufs by using the option ``--pool_sz 131072``
> +Max number of events in OCTEON TX Eventdev (SSO) are only limited by DRAM size
> +and they can be configured by passing limits to kernel bootargs as follows:
> +
> +.. code-block:: console
> +
> +        ssopf.max_events=4194304
> +
> +The same can be verified by looking at the following sysfs entry:
> +
> +.. code-block:: console
> +
> +        # cat /sys/module/ssopf/parameters/max_events
> +        4194304
> +
> +The maximum number of events that can be added to SSO by the event adapters such
> +as (Rx/Timer) should be limited to the above configured value.
> diff --git a/drivers/event/octeontx/ssovf_evdev.c b/drivers/event/octeontx/ssovf_evdev.c
> index 4fc4e8f7e..33cb50204 100644
> --- a/drivers/event/octeontx/ssovf_evdev.c
> +++ b/drivers/event/octeontx/ssovf_evdev.c
> @@ -384,22 +384,78 @@ ssovf_eth_rx_adapter_queue_add(const struct rte_eventdev *dev,
>                 const struct rte_eth_dev *eth_dev, int32_t rx_queue_id,
>                 const struct rte_event_eth_rx_adapter_queue_conf *queue_conf)
>  {
> -       int ret = 0;
>         const struct octeontx_nic *nic = eth_dev->data->dev_private;
>         struct ssovf_evdev *edev = ssovf_pmd_priv(dev);
> +       uint16_t free_idx = UINT16_MAX;
> +       struct octeontx_rxq *rxq;
>         pki_mod_qos_t pki_qos;
> -       RTE_SET_USED(dev);
> +       uint8_t found = false;
> +       int i, ret = 0;
> +       void *old_ptr;
>
>         ret = strncmp(eth_dev->data->name, "eth_octeontx", 12);
>         if (ret)
>                 return -EINVAL;
>
> -       if (rx_queue_id >= 0)
> -               return -EINVAL;
> -
>         if (queue_conf->ev.sched_type == RTE_SCHED_TYPE_PARALLEL)
>                 return -ENOTSUP;
>
> +       /* eth_octeontx only supports one rq. */
> +       rx_queue_id = rx_queue_id == -1 ? 0 : rx_queue_id;
> +       rxq = eth_dev->data->rx_queues[rx_queue_id];
> +       /* Add rxq pool to list of used pools and reduce available events. */
> +       for (i = 0; i < edev->rxq_pools; i++) {
> +               if (edev->rxq_pool_array[i] == (uintptr_t)rxq->pool) {
> +                       edev->rxq_pool_rcnt[i]++;
> +                       found = true;
> +                       break;
> +               } else if (free_idx == UINT16_MAX &&
> +                          edev->rxq_pool_array[i] == 0) {
> +                       free_idx = i;
> +               }
> +       }
> +
> +       if (!found) {
> +               uint16_t idx;
> +
> +               if (edev->available_events < rxq->pool->size) {
> +                       ssovf_log_err(
> +                               "Max available events %"PRIu32" requested events in rxq pool %"PRIu32"",
> +                               edev->available_events, rxq->pool->size);
> +                       return -ENOMEM;
> +               }
> +
> +               if (free_idx != UINT16_MAX) {
> +                       idx = free_idx;
> +               } else {
> +                       old_ptr = edev->rxq_pool_array;
> +                       edev->rxq_pools++;
> +                       edev->rxq_pool_array = rte_realloc(
> +                               edev->rxq_pool_array,
> +                               sizeof(uint64_t) * edev->rxq_pools, 0);
> +                       if (edev->rxq_pool_array == NULL) {
> +                               edev->rxq_pools--;
> +                               edev->rxq_pool_array = old_ptr;
> +                               return -ENOMEM;
> +                       }
> +
> +                       old_ptr = edev->rxq_pool_rcnt;
> +                       edev->rxq_pool_rcnt = rte_realloc(
> +                               edev->rxq_pool_rcnt,
> +                               sizeof(uint8_t) * edev->rxq_pools, 0);
> +                       if (edev->rxq_pool_rcnt == NULL) {
> +                               edev->rxq_pools--;
> +                               edev->rxq_pool_rcnt = old_ptr;
> +                               return -ENOMEM;
> +                       }
> +                       idx = edev->rxq_pools - 1;
> +               }
> +
> +               edev->rxq_pool_array[idx] = (uintptr_t)rxq->pool;
> +               edev->rxq_pool_rcnt[idx] = 1;
> +               edev->available_events -= rxq->pool->size;
> +       }
> +
>         memset(&pki_qos, 0, sizeof(pki_mod_qos_t));
>
>         pki_qos.port_type = 0;
> @@ -432,10 +488,28 @@ static int
>  ssovf_eth_rx_adapter_queue_del(const struct rte_eventdev *dev,
>                 const struct rte_eth_dev *eth_dev, int32_t rx_queue_id)
>  {
> -       int ret = 0;
>         const struct octeontx_nic *nic = eth_dev->data->dev_private;
> +       struct ssovf_evdev *edev = ssovf_pmd_priv(dev);
> +       struct octeontx_rxq *rxq;
>         pki_del_qos_t pki_qos;
> -       RTE_SET_USED(dev);
> +       uint8_t found = false;
> +       int i, ret = 0;
> +
> +       rx_queue_id = rx_queue_id == -1 ? 0 : rx_queue_id;
> +       rxq = eth_dev->data->rx_queues[rx_queue_id];
> +       for (i = 0; i < edev->rxq_pools; i++) {
> +               if (edev->rxq_pool_array[i] == (uintptr_t)rxq->pool) {
> +                       found = true;
> +                       break;
> +               }
> +       }
> +
> +       if (found) {
> +               edev->rxq_pool_rcnt[i]--;
> +               if (edev->rxq_pool_rcnt[i] == 0)
> +                       edev->rxq_pool_array[i] = 0;
> +               edev->available_events += rxq->pool->size;
> +       }
>
>         ret = strncmp(eth_dev->data->name, "eth_octeontx", 12);
>         if (ret)
> @@ -754,6 +828,8 @@ ssovf_vdev_probe(struct rte_vdev_device *vdev)
>         }
>         eventdev->dev_ops = &ssovf_ops;
>
> +       timvf_set_eventdevice(eventdev);
> +
>         /* For secondary processes, the primary has done all the work */
>         if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
>                 ssovf_fastpath_fns_set(eventdev);
> @@ -781,9 +857,12 @@ ssovf_vdev_probe(struct rte_vdev_device *vdev)
>         edev->min_deq_timeout_ns = info.min_deq_timeout_ns;
>         edev->max_deq_timeout_ns = info.max_deq_timeout_ns;
>         edev->max_num_events =  info.max_num_events;
> -       ssovf_log_dbg("min_deq_tmo=%"PRId64" max_deq_tmo=%"PRId64" max_evts=%d",
> -                       info.min_deq_timeout_ns, info.max_deq_timeout_ns,
> -                       info.max_num_events);
> +       edev->available_events = info.max_num_events;
> +
> +       ssovf_log_dbg("min_deq_tmo=%" PRId64 " max_deq_tmo=%" PRId64
> +                     " max_evts=%d",
> +                     info.min_deq_timeout_ns, info.max_deq_timeout_ns,
> +                     info.max_num_events);
>
>         if (!edev->max_event_ports || !edev->max_event_queues) {
>                 ssovf_log_err("Not enough eventdev resource queues=%d ports=%d",
> diff --git a/drivers/event/octeontx/ssovf_evdev.h b/drivers/event/octeontx/ssovf_evdev.h
> index aa5acf246..90d760a54 100644
> --- a/drivers/event/octeontx/ssovf_evdev.h
> +++ b/drivers/event/octeontx/ssovf_evdev.h
> @@ -146,6 +146,12 @@ struct ssovf_evdev {
>         uint32_t min_deq_timeout_ns;
>         uint32_t max_deq_timeout_ns;
>         int32_t max_num_events;
> +       uint32_t available_events;
> +       uint16_t rxq_pools;
> +       uint64_t *rxq_pool_array;
> +       uint8_t *rxq_pool_rcnt;
> +       uint16_t tim_ring_cnt;
> +       uint16_t *tim_ring_ids;
>  } __rte_cache_aligned;
>
>  /* Event port aka HWS */
> diff --git a/drivers/event/octeontx/timvf_evdev.c b/drivers/event/octeontx/timvf_evdev.c
> index c61aacacc..8af4d6e37 100644
> --- a/drivers/event/octeontx/timvf_evdev.c
> +++ b/drivers/event/octeontx/timvf_evdev.c
> @@ -2,10 +2,13 @@
>   * Copyright(c) 2017 Cavium, Inc
>   */
>
> +#include "ssovf_evdev.h"
>  #include "timvf_evdev.h"
>
>  RTE_LOG_REGISTER(otx_logtype_timvf, pmd.event.octeontx.timer, NOTICE);
>
> +static struct rte_eventdev *event_dev;
> +
>  struct __rte_packed timvf_mbox_dev_info {
>         uint64_t ring_active[4];
>         uint64_t clk_freq;
> @@ -222,19 +225,21 @@ timvf_ring_stop(const struct rte_event_timer_adapter *adptr)
>  static int
>  timvf_ring_create(struct rte_event_timer_adapter *adptr)
>  {
> -       char pool_name[25];
> -       int ret;
> -       uint8_t tim_ring_id;
> -       uint64_t nb_timers;
>         struct rte_event_timer_adapter_conf *rcfg = &adptr->data->conf;
> +       uint16_t free_idx = UINT16_MAX;
> +       unsigned int mp_flags = 0;
> +       struct ssovf_evdev *edev;
>         struct timvf_ring *timr;
>         const char *mempool_ops;
> -       unsigned int mp_flags = 0;
> +       uint8_t tim_ring_id;
> +       char pool_name[25];
> +       int i, ret;
>
>         tim_ring_id = timvf_get_ring();
>         if (tim_ring_id == UINT8_MAX)
>                 return -ENODEV;
>
> +       edev = ssovf_pmd_priv(event_dev);
>         timr = rte_zmalloc("octeontx_timvf_priv",
>                         sizeof(struct timvf_ring), 0);
>         if (timr == NULL)
> @@ -256,10 +261,42 @@ timvf_ring_create(struct rte_event_timer_adapter *adptr)
>         timr->nb_bkts = (timr->max_tout / timr->tck_nsec);
>         timr->vbar0 = timvf_bar(timr->tim_ring_id, 0);
>         timr->bkt_pos = (uint8_t *)timr->vbar0 + TIM_VRING_REL;
> -       nb_timers = rcfg->nb_timers;
> +       timr->nb_timers = rcfg->nb_timers;
>         timr->get_target_bkt = bkt_mod;
>
> -       timr->nb_chunks = nb_timers / nb_chunk_slots;
> +       if (edev->available_events < timr->nb_timers) {
> +               timvf_log_err(
> +                       "Max available events %"PRIu32" requested timer events %"PRIu64"",
> +                       edev->available_events, timr->nb_timers);
> +               return -ENOMEM;
> +       }
> +
> +       for (i = 0; i < edev->tim_ring_cnt; i++) {
> +               if (edev->tim_ring_ids[i] == UINT16_MAX)
> +                       free_idx = i;
> +       }
> +
> +       if (free_idx == UINT16_MAX) {
> +               void *old_ptr;
> +
> +               edev->tim_ring_cnt++;
> +               old_ptr = edev->tim_ring_ids;
> +               edev->tim_ring_ids =
> +                       rte_realloc(edev->tim_ring_ids,
> +                                   sizeof(uint16_t) * edev->tim_ring_cnt, 0);
> +               if (edev->tim_ring_ids == NULL) {
> +                       edev->tim_ring_ids = old_ptr;
> +                       edev->tim_ring_cnt--;
> +                       return -ENOMEM;
> +               }
> +
> +               edev->available_events -= timr->nb_timers;
> +       } else {
> +               edev->tim_ring_ids[free_idx] = tim_ring_id;
> +               edev->available_events -= timr->nb_timers;
> +       }
> +
> +       timr->nb_chunks = timr->nb_timers / nb_chunk_slots;
>
>         /* Try to optimize the bucket parameters. */
>         if ((rcfg->flags & RTE_EVENT_TIMER_ADAPTER_F_ADJUST_RES)
> @@ -328,6 +365,17 @@ static int
>  timvf_ring_free(struct rte_event_timer_adapter *adptr)
>  {
>         struct timvf_ring *timr = adptr->data->adapter_priv;
> +       struct ssovf_evdev *edev;
> +       int i;
> +
> +       edev = ssovf_pmd_priv(event_dev);
> +       for (i = 0; i < edev->tim_ring_cnt; i++) {
> +               if (edev->tim_ring_ids[i] == timr->tim_ring_id) {
> +                       edev->available_events += timr->nb_timers;
> +                       edev->tim_ring_ids[i] = UINT16_MAX;
> +                       break;
> +               }
> +       }
>
>         rte_mempool_free(timr->chunk_pool);
>         rte_free(timr->bkt);
> @@ -396,3 +444,9 @@ timvf_timer_adapter_caps_get(const struct rte_eventdev *dev, uint64_t flags,
>         *ops = &timvf_ops;
>         return 0;
>  }
> +
> +void
> +timvf_set_eventdevice(struct rte_eventdev *dev)
> +{
> +       event_dev = dev;
> +}
> diff --git a/drivers/event/octeontx/timvf_evdev.h b/drivers/event/octeontx/timvf_evdev.h
> index d0e5921db..2977063d6 100644
> --- a/drivers/event/octeontx/timvf_evdev.h
> +++ b/drivers/event/octeontx/timvf_evdev.h
> @@ -175,6 +175,7 @@ struct timvf_ring {
>         void *bkt_pos;
>         uint64_t max_tout;
>         uint64_t nb_chunks;
> +       uint64_t nb_timers;
>         enum timvf_clk_src clk_src;
>         uint16_t tim_ring_id;
>  } __rte_cache_aligned;
> @@ -217,5 +218,6 @@ uint16_t timvf_timer_arm_tmo_brst_stats(
>                 struct rte_event_timer **tim, const uint64_t timeout_tick,
>                 const uint16_t nb_timers);
>  void timvf_set_chunk_refill(struct timvf_ring * const timr, uint8_t use_fpa);
> +void timvf_set_eventdevice(struct rte_eventdev *dev);
>
>  #endif /* __TIMVF_EVDEV_H__ */
> --
> 2.17.1
>

Patch
diff mbox series

diff --git a/doc/guides/eventdevs/octeontx.rst b/doc/guides/eventdevs/octeontx.rst
index 9a3646db0..21d251341 100644
--- a/doc/guides/eventdevs/octeontx.rst
+++ b/doc/guides/eventdevs/octeontx.rst
@@ -140,9 +140,22 @@  follows:
 When timvf is used as Event timer adapter event schedule type
 ``RTE_SCHED_TYPE_PARALLEL`` is not supported.

-Max mempool size
-~~~~~~~~~~~~~~~~
+Max number of events
+~~~~~~~~~~~~~~~~~~~~

-Max mempool size when using OCTEON TX Eventdev (SSO) should be limited to 128K.
-When running dpdk-test-eventdev on OCTEON TX the application can limit the
-number of mbufs by using the option ``--pool_sz 131072``
+Max number of events in OCTEON TX Eventdev (SSO) are only limited by DRAM size
+and they can be configured by passing limits to kernel bootargs as follows:
+
+.. code-block:: console
+
+        ssopf.max_events=4194304
+
+The same can be verified by looking at the following sysfs entry:
+
+.. code-block:: console
+
+        # cat /sys/module/ssopf/parameters/max_events
+        4194304
+
+The maximum number of events that can be added to SSO by the event adapters such
+as (Rx/Timer) should be limited to the above configured value.
diff --git a/drivers/event/octeontx/ssovf_evdev.c b/drivers/event/octeontx/ssovf_evdev.c
index 4fc4e8f7e..33cb50204 100644
--- a/drivers/event/octeontx/ssovf_evdev.c
+++ b/drivers/event/octeontx/ssovf_evdev.c
@@ -384,22 +384,78 @@  ssovf_eth_rx_adapter_queue_add(const struct rte_eventdev *dev,
 		const struct rte_eth_dev *eth_dev, int32_t rx_queue_id,
 		const struct rte_event_eth_rx_adapter_queue_conf *queue_conf)
 {
-	int ret = 0;
 	const struct octeontx_nic *nic = eth_dev->data->dev_private;
 	struct ssovf_evdev *edev = ssovf_pmd_priv(dev);
+	uint16_t free_idx = UINT16_MAX;
+	struct octeontx_rxq *rxq;
 	pki_mod_qos_t pki_qos;
-	RTE_SET_USED(dev);
+	uint8_t found = false;
+	int i, ret = 0;
+	void *old_ptr;

 	ret = strncmp(eth_dev->data->name, "eth_octeontx", 12);
 	if (ret)
 		return -EINVAL;

-	if (rx_queue_id >= 0)
-		return -EINVAL;
-
 	if (queue_conf->ev.sched_type == RTE_SCHED_TYPE_PARALLEL)
 		return -ENOTSUP;

+	/* eth_octeontx only supports one rq. */
+	rx_queue_id = rx_queue_id == -1 ? 0 : rx_queue_id;
+	rxq = eth_dev->data->rx_queues[rx_queue_id];
+	/* Add rxq pool to list of used pools and reduce available events. */
+	for (i = 0; i < edev->rxq_pools; i++) {
+		if (edev->rxq_pool_array[i] == (uintptr_t)rxq->pool) {
+			edev->rxq_pool_rcnt[i]++;
+			found = true;
+			break;
+		} else if (free_idx == UINT16_MAX &&
+			   edev->rxq_pool_array[i] == 0) {
+			free_idx = i;
+		}
+	}
+
+	if (!found) {
+		uint16_t idx;
+
+		if (edev->available_events < rxq->pool->size) {
+			ssovf_log_err(
+				"Max available events %"PRIu32" requested events in rxq pool %"PRIu32"",
+				edev->available_events, rxq->pool->size);
+			return -ENOMEM;
+		}
+
+		if (free_idx != UINT16_MAX) {
+			idx = free_idx;
+		} else {
+			old_ptr = edev->rxq_pool_array;
+			edev->rxq_pools++;
+			edev->rxq_pool_array = rte_realloc(
+				edev->rxq_pool_array,
+				sizeof(uint64_t) * edev->rxq_pools, 0);
+			if (edev->rxq_pool_array == NULL) {
+				edev->rxq_pools--;
+				edev->rxq_pool_array = old_ptr;
+				return -ENOMEM;
+			}
+
+			old_ptr = edev->rxq_pool_rcnt;
+			edev->rxq_pool_rcnt = rte_realloc(
+				edev->rxq_pool_rcnt,
+				sizeof(uint8_t) * edev->rxq_pools, 0);
+			if (edev->rxq_pool_rcnt == NULL) {
+				edev->rxq_pools--;
+				edev->rxq_pool_rcnt = old_ptr;
+				return -ENOMEM;
+			}
+			idx = edev->rxq_pools - 1;
+		}
+
+		edev->rxq_pool_array[idx] = (uintptr_t)rxq->pool;
+		edev->rxq_pool_rcnt[idx] = 1;
+		edev->available_events -= rxq->pool->size;
+	}
+
 	memset(&pki_qos, 0, sizeof(pki_mod_qos_t));

 	pki_qos.port_type = 0;
@@ -432,10 +488,28 @@  static int
 ssovf_eth_rx_adapter_queue_del(const struct rte_eventdev *dev,
 		const struct rte_eth_dev *eth_dev, int32_t rx_queue_id)
 {
-	int ret = 0;
 	const struct octeontx_nic *nic = eth_dev->data->dev_private;
+	struct ssovf_evdev *edev = ssovf_pmd_priv(dev);
+	struct octeontx_rxq *rxq;
 	pki_del_qos_t pki_qos;
-	RTE_SET_USED(dev);
+	uint8_t found = false;
+	int i, ret = 0;
+
+	rx_queue_id = rx_queue_id == -1 ? 0 : rx_queue_id;
+	rxq = eth_dev->data->rx_queues[rx_queue_id];
+	for (i = 0; i < edev->rxq_pools; i++) {
+		if (edev->rxq_pool_array[i] == (uintptr_t)rxq->pool) {
+			found = true;
+			break;
+		}
+	}
+
+	if (found) {
+		edev->rxq_pool_rcnt[i]--;
+		if (edev->rxq_pool_rcnt[i] == 0)
+			edev->rxq_pool_array[i] = 0;
+		edev->available_events += rxq->pool->size;
+	}

 	ret = strncmp(eth_dev->data->name, "eth_octeontx", 12);
 	if (ret)
@@ -754,6 +828,8 @@  ssovf_vdev_probe(struct rte_vdev_device *vdev)
 	}
 	eventdev->dev_ops = &ssovf_ops;

+	timvf_set_eventdevice(eventdev);
+
 	/* For secondary processes, the primary has done all the work */
 	if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
 		ssovf_fastpath_fns_set(eventdev);
@@ -781,9 +857,12 @@  ssovf_vdev_probe(struct rte_vdev_device *vdev)
 	edev->min_deq_timeout_ns = info.min_deq_timeout_ns;
 	edev->max_deq_timeout_ns = info.max_deq_timeout_ns;
 	edev->max_num_events =  info.max_num_events;
-	ssovf_log_dbg("min_deq_tmo=%"PRId64" max_deq_tmo=%"PRId64" max_evts=%d",
-			info.min_deq_timeout_ns, info.max_deq_timeout_ns,
-			info.max_num_events);
+	edev->available_events = info.max_num_events;
+
+	ssovf_log_dbg("min_deq_tmo=%" PRId64 " max_deq_tmo=%" PRId64
+		      " max_evts=%d",
+		      info.min_deq_timeout_ns, info.max_deq_timeout_ns,
+		      info.max_num_events);

 	if (!edev->max_event_ports || !edev->max_event_queues) {
 		ssovf_log_err("Not enough eventdev resource queues=%d ports=%d",
diff --git a/drivers/event/octeontx/ssovf_evdev.h b/drivers/event/octeontx/ssovf_evdev.h
index aa5acf246..90d760a54 100644
--- a/drivers/event/octeontx/ssovf_evdev.h
+++ b/drivers/event/octeontx/ssovf_evdev.h
@@ -146,6 +146,12 @@  struct ssovf_evdev {
 	uint32_t min_deq_timeout_ns;
 	uint32_t max_deq_timeout_ns;
 	int32_t max_num_events;
+	uint32_t available_events;
+	uint16_t rxq_pools;
+	uint64_t *rxq_pool_array;
+	uint8_t *rxq_pool_rcnt;
+	uint16_t tim_ring_cnt;
+	uint16_t *tim_ring_ids;
 } __rte_cache_aligned;

 /* Event port aka HWS */
diff --git a/drivers/event/octeontx/timvf_evdev.c b/drivers/event/octeontx/timvf_evdev.c
index c61aacacc..8af4d6e37 100644
--- a/drivers/event/octeontx/timvf_evdev.c
+++ b/drivers/event/octeontx/timvf_evdev.c
@@ -2,10 +2,13 @@ 
  * Copyright(c) 2017 Cavium, Inc
  */

+#include "ssovf_evdev.h"
 #include "timvf_evdev.h"

 RTE_LOG_REGISTER(otx_logtype_timvf, pmd.event.octeontx.timer, NOTICE);

+static struct rte_eventdev *event_dev;
+
 struct __rte_packed timvf_mbox_dev_info {
 	uint64_t ring_active[4];
 	uint64_t clk_freq;
@@ -222,19 +225,21 @@  timvf_ring_stop(const struct rte_event_timer_adapter *adptr)
 static int
 timvf_ring_create(struct rte_event_timer_adapter *adptr)
 {
-	char pool_name[25];
-	int ret;
-	uint8_t tim_ring_id;
-	uint64_t nb_timers;
 	struct rte_event_timer_adapter_conf *rcfg = &adptr->data->conf;
+	uint16_t free_idx = UINT16_MAX;
+	unsigned int mp_flags = 0;
+	struct ssovf_evdev *edev;
 	struct timvf_ring *timr;
 	const char *mempool_ops;
-	unsigned int mp_flags = 0;
+	uint8_t tim_ring_id;
+	char pool_name[25];
+	int i, ret;

 	tim_ring_id = timvf_get_ring();
 	if (tim_ring_id == UINT8_MAX)
 		return -ENODEV;

+	edev = ssovf_pmd_priv(event_dev);
 	timr = rte_zmalloc("octeontx_timvf_priv",
 			sizeof(struct timvf_ring), 0);
 	if (timr == NULL)
@@ -256,10 +261,42 @@  timvf_ring_create(struct rte_event_timer_adapter *adptr)
 	timr->nb_bkts = (timr->max_tout / timr->tck_nsec);
 	timr->vbar0 = timvf_bar(timr->tim_ring_id, 0);
 	timr->bkt_pos = (uint8_t *)timr->vbar0 + TIM_VRING_REL;
-	nb_timers = rcfg->nb_timers;
+	timr->nb_timers = rcfg->nb_timers;
 	timr->get_target_bkt = bkt_mod;

-	timr->nb_chunks = nb_timers / nb_chunk_slots;
+	if (edev->available_events < timr->nb_timers) {
+		timvf_log_err(
+			"Max available events %"PRIu32" requested timer events %"PRIu64"",
+			edev->available_events, timr->nb_timers);
+		return -ENOMEM;
+	}
+
+	for (i = 0; i < edev->tim_ring_cnt; i++) {
+		if (edev->tim_ring_ids[i] == UINT16_MAX)
+			free_idx = i;
+	}
+
+	if (free_idx == UINT16_MAX) {
+		void *old_ptr;
+
+		edev->tim_ring_cnt++;
+		old_ptr = edev->tim_ring_ids;
+		edev->tim_ring_ids =
+			rte_realloc(edev->tim_ring_ids,
+				    sizeof(uint16_t) * edev->tim_ring_cnt, 0);
+		if (edev->tim_ring_ids == NULL) {
+			edev->tim_ring_ids = old_ptr;
+			edev->tim_ring_cnt--;
+			return -ENOMEM;
+		}
+
+		edev->available_events -= timr->nb_timers;
+	} else {
+		edev->tim_ring_ids[free_idx] = tim_ring_id;
+		edev->available_events -= timr->nb_timers;
+	}
+
+	timr->nb_chunks = timr->nb_timers / nb_chunk_slots;

 	/* Try to optimize the bucket parameters. */
 	if ((rcfg->flags & RTE_EVENT_TIMER_ADAPTER_F_ADJUST_RES)
@@ -328,6 +365,17 @@  static int
 timvf_ring_free(struct rte_event_timer_adapter *adptr)
 {
 	struct timvf_ring *timr = adptr->data->adapter_priv;
+	struct ssovf_evdev *edev;
+	int i;
+
+	edev = ssovf_pmd_priv(event_dev);
+	for (i = 0; i < edev->tim_ring_cnt; i++) {
+		if (edev->tim_ring_ids[i] == timr->tim_ring_id) {
+			edev->available_events += timr->nb_timers;
+			edev->tim_ring_ids[i] = UINT16_MAX;
+			break;
+		}
+	}

 	rte_mempool_free(timr->chunk_pool);
 	rte_free(timr->bkt);
@@ -396,3 +444,9 @@  timvf_timer_adapter_caps_get(const struct rte_eventdev *dev, uint64_t flags,
 	*ops = &timvf_ops;
 	return 0;
 }
+
+void
+timvf_set_eventdevice(struct rte_eventdev *dev)
+{
+	event_dev = dev;
+}
diff --git a/drivers/event/octeontx/timvf_evdev.h b/drivers/event/octeontx/timvf_evdev.h
index d0e5921db..2977063d6 100644
--- a/drivers/event/octeontx/timvf_evdev.h
+++ b/drivers/event/octeontx/timvf_evdev.h
@@ -175,6 +175,7 @@  struct timvf_ring {
 	void *bkt_pos;
 	uint64_t max_tout;
 	uint64_t nb_chunks;
+	uint64_t nb_timers;
 	enum timvf_clk_src clk_src;
 	uint16_t tim_ring_id;
 } __rte_cache_aligned;
@@ -217,5 +218,6 @@  uint16_t timvf_timer_arm_tmo_brst_stats(
 		struct rte_event_timer **tim, const uint64_t timeout_tick,
 		const uint16_t nb_timers);
 void timvf_set_chunk_refill(struct timvf_ring * const timr, uint8_t use_fpa);
+void timvf_set_eventdevice(struct rte_eventdev *dev);

 #endif /* __TIMVF_EVDEV_H__ */