@@ -514,7 +514,6 @@ cn10k_sso_fp_fns_set(struct rte_eventdev *event_dev)
cn10k_sso_fp_blk_fns_set(event_dev);
cn10k_sso_fp_tmplt_fns_set(event_dev);
- event_dev->enqueue = cn10k_sso_hws_enq;
event_dev->enqueue_burst = cn10k_sso_hws_enq_burst;
event_dev->enqueue_new_burst = cn10k_sso_hws_enq_new_burst;
event_dev->enqueue_forward_burst = cn10k_sso_hws_enq_fwd_burst;
@@ -848,7 +847,6 @@ eventdev_fops_update(struct rte_eventdev *event_dev)
struct rte_event_fp_ops *fp_op =
rte_event_fp_ops + event_dev->data->dev_id;
- fp_op->dequeue = event_dev->dequeue;
fp_op->dequeue_burst = event_dev->dequeue_burst;
}
@@ -107,32 +107,6 @@ sso_lmt_aw_wait_fc(struct cn10k_sso_hws *ws, int64_t req)
}
}
-uint16_t __rte_hot
-cn10k_sso_hws_enq(void *port, const struct rte_event *ev)
-{
- struct cn10k_sso_hws *ws = port;
-
- switch (ev->op) {
- case RTE_EVENT_OP_NEW:
- return cn10k_sso_hws_new_event(ws, ev);
- case RTE_EVENT_OP_FORWARD:
- cn10k_sso_hws_forward_event(ws, ev);
- break;
- case RTE_EVENT_OP_RELEASE:
- if (ws->swtag_req) {
- cnxk_sso_hws_desched(ev->u64, ws->base);
- ws->swtag_req = 0;
- break;
- }
- cnxk_sso_hws_swtag_flush(ws->base);
- break;
- default:
- return 0;
- }
-
- return 1;
-}
-
#define VECTOR_SIZE_BITS 0xFFFFFFFFFFF80000ULL
#define VECTOR_GET_LINE_OFFSET(line) (19 + (3 * line))
@@ -384,8 +358,29 @@ uint16_t __rte_hot
cn10k_sso_hws_enq_burst(void *port, const struct rte_event ev[],
uint16_t nb_events)
{
+ struct cn10k_sso_hws *ws = port;
+
RTE_SET_USED(nb_events);
- return cn10k_sso_hws_enq(port, ev);
+
+ switch (ev->op) {
+ case RTE_EVENT_OP_NEW:
+ return cn10k_sso_hws_new_event(ws, ev);
+ case RTE_EVENT_OP_FORWARD:
+ cn10k_sso_hws_forward_event(ws, ev);
+ break;
+ case RTE_EVENT_OP_RELEASE:
+ if (ws->swtag_req) {
+ cnxk_sso_hws_desched(ev->u64, ws->base);
+ ws->swtag_req = 0;
+ break;
+ }
+ cnxk_sso_hws_swtag_flush(ws->base);
+ break;
+ default:
+ return 0;
+ }
+
+ return 1;
}
uint16_t __rte_hot
@@ -366,7 +366,6 @@ cn10k_sso_hws_get_work_empty(struct cn10k_sso_hws *ws, struct rte_event *ev,
}
/* CN10K Fastpath functions. */
-uint16_t __rte_hot cn10k_sso_hws_enq(void *port, const struct rte_event *ev);
uint16_t __rte_hot cn10k_sso_hws_enq_burst(void *port,
const struct rte_event ev[],
uint16_t nb_events);
@@ -381,8 +380,6 @@ int __rte_hot cn10k_sso_hws_preschedule_modify(void *port,
enum rte_event_dev_preschedule_type type);
#define R(name, flags) \
- uint16_t __rte_hot cn10k_sso_hws_deq_##name( \
- void *port, struct rte_event *ev, uint64_t timeout_ticks); \
uint16_t __rte_hot cn10k_sso_hws_deq_burst_##name( \
void *port, struct rte_event ev[], uint16_t nb_events, \
uint64_t timeout_ticks); \
@@ -421,8 +418,6 @@ int __rte_hot cn10k_sso_hws_preschedule_modify(void *port,
uint16_t __rte_hot cn10k_sso_hws_deq_tmo_ca_seg_burst_##name( \
void *port, struct rte_event ev[], uint16_t nb_events, \
uint64_t timeout_ticks); \
- uint16_t __rte_hot cn10k_sso_hws_reas_deq_##name( \
- void *port, struct rte_event *ev, uint64_t timeout_ticks); \
uint16_t __rte_hot cn10k_sso_hws_reas_deq_burst_##name( \
void *port, struct rte_event ev[], uint16_t nb_events, \
uint64_t timeout_ticks); \
@@ -519,12 +514,8 @@ NIX_RX_FASTPATH_MODES
return fn(port, ev, timeout_ticks); \
}
-uint16_t __rte_hot cn10k_sso_hws_deq_all_offload(void *port, struct rte_event *ev,
- uint64_t timeout_ticks);
uint16_t __rte_hot cn10k_sso_hws_deq_burst_all_offload(void *port, struct rte_event ev[],
uint16_t nb_events, uint64_t timeout_ticks);
-uint16_t __rte_hot cn10k_sso_hws_deq_all_offload_tst(void *port, struct rte_event *ev,
- uint64_t timeout_ticks);
uint16_t __rte_hot cn10k_sso_hws_deq_burst_all_offload_tst(void *port, struct rte_event ev[],
uint16_t nb_events,
uint64_t timeout_ticks);
@@ -917,7 +917,6 @@ eventdev_fops_tstamp_update(struct rte_eventdev *event_dev)
struct rte_event_fp_ops *fp_op =
rte_event_fp_ops + event_dev->data->dev_id;
- fp_op->dequeue = event_dev->dequeue;
fp_op->dequeue_burst = event_dev->dequeue_burst;
}
@@ -8,10 +8,13 @@
#include "cn9k_cryptodev_ops.h"
uint16_t __rte_hot
-cn9k_sso_hws_enq(void *port, const struct rte_event *ev)
+cn9k_sso_hws_enq_burst(void *port, const struct rte_event ev[],
+ uint16_t nb_events)
{
struct cn9k_sso_hws *ws = port;
+ RTE_SET_USED(nb_events);
+
switch (ev->op) {
case RTE_EVENT_OP_NEW:
return cn9k_sso_hws_new_event(ws, ev);
@@ -33,14 +36,6 @@ cn9k_sso_hws_enq(void *port, const struct rte_event *ev)
return 1;
}
-uint16_t __rte_hot
-cn9k_sso_hws_enq_burst(void *port, const struct rte_event ev[],
- uint16_t nb_events)
-{
- RTE_SET_USED(nb_events);
- return cn9k_sso_hws_enq(port, ev);
-}
-
uint16_t __rte_hot
cn9k_sso_hws_enq_new_burst(void *port, const struct rte_event ev[],
uint16_t nb_events)
@@ -80,11 +75,14 @@ cn9k_sso_hws_profile_switch(void *port, uint8_t profile)
/* Dual ws ops. */
uint16_t __rte_hot
-cn9k_sso_hws_dual_enq(void *port, const struct rte_event *ev)
+cn9k_sso_hws_dual_enq_burst(void *port, const struct rte_event ev[],
+ uint16_t nb_events)
{
struct cn9k_sso_hws_dual *dws = port;
uint64_t base;
+ RTE_SET_USED(nb_events);
+
base = dws->base[!dws->vws];
switch (ev->op) {
case RTE_EVENT_OP_NEW:
@@ -107,14 +105,6 @@ cn9k_sso_hws_dual_enq(void *port, const struct rte_event *ev)
return 1;
}
-uint16_t __rte_hot
-cn9k_sso_hws_dual_enq_burst(void *port, const struct rte_event ev[],
- uint16_t nb_events)
-{
- RTE_SET_USED(nb_events);
- return cn9k_sso_hws_dual_enq(port, ev);
-}
-
uint16_t __rte_hot
cn9k_sso_hws_dual_enq_new_burst(void *port, const struct rte_event ev[],
uint16_t nb_events)
@@ -359,7 +359,6 @@ cn9k_sso_hws_get_work_empty(uint64_t base, struct rte_event *ev,
}
/* CN9K Fastpath functions. */
-uint16_t __rte_hot cn9k_sso_hws_enq(void *port, const struct rte_event *ev);
uint16_t __rte_hot cn9k_sso_hws_enq_burst(void *port,
const struct rte_event ev[],
uint16_t nb_events);
@@ -371,8 +370,6 @@ uint16_t __rte_hot cn9k_sso_hws_enq_fwd_burst(void *port,
uint16_t nb_events);
int __rte_hot cn9k_sso_hws_profile_switch(void *port, uint8_t profile);
-uint16_t __rte_hot cn9k_sso_hws_dual_enq(void *port,
- const struct rte_event *ev);
uint16_t __rte_hot cn9k_sso_hws_dual_enq_burst(void *port,
const struct rte_event ev[],
uint16_t nb_events);
@@ -451,8 +448,6 @@ NIX_RX_FASTPATH_MODES
#define SSO_DEQ_TMO_SEG(fn, flags) SSO_DEQ_TMO(fn, flags | NIX_RX_MULTI_SEG_F)
#define R(name, flags) \
- uint16_t __rte_hot cn9k_sso_hws_dual_deq_##name( \
- void *port, struct rte_event *ev, uint64_t timeout_ticks); \
uint16_t __rte_hot cn9k_sso_hws_dual_deq_burst_##name( \
void *port, struct rte_event ev[], uint16_t nb_events, \
uint64_t timeout_ticks); \
@@ -542,12 +537,6 @@ NIX_RX_FASTPATH_MODES
return fn(port, ev, timeout_ticks); \
}
-uint16_t __rte_hot cn9k_sso_hws_deq_all_offload(void *port, struct rte_event *ev,
- uint64_t timeout_ticks);
-
-uint16_t __rte_hot cn9k_sso_hws_deq_dual_all_offload(void *port, struct rte_event *ev,
- uint64_t timeout_ticks);
-
uint16_t __rte_hot cn9k_sso_hws_deq_burst_all_offload(void *port, struct rte_event ev[],
uint16_t nb_events, uint64_t timeout_ticks);
@@ -555,12 +544,6 @@ uint16_t __rte_hot cn9k_sso_hws_deq_dual_burst_all_offload(void *port, struct rt
uint16_t nb_events,
uint64_t timeout_ticks);
-uint16_t __rte_hot cn9k_sso_hws_deq_all_offload_tst(void *port, struct rte_event *ev,
- uint64_t timeout_ticks);
-
-uint16_t __rte_hot cn9k_sso_hws_deq_dual_all_offload_tst(void *port, struct rte_event *ev,
- uint64_t timeout_ticks);
-
uint16_t __rte_hot cn9k_sso_hws_deq_burst_all_offload_tst(void *port, struct rte_event ev[],
uint16_t nb_events,
uint64_t timeout_ticks);