@@ -250,6 +250,10 @@ struct rte_event_crypto_adapter_conf {
* max_nb crypto ops. This isn't treated as a requirement; batching
* may cause the adapter to process more than max_nb crypto ops.
*/
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**
@@ -112,6 +112,10 @@ struct rte_event_eth_rx_adapter_conf {
* max_nb_rx mbufs. This isn't treated as a requirement; batching may
* cause the adapter to process more than max_nb_rx mbufs.
*/
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**
@@ -171,6 +175,10 @@ struct rte_event_eth_rx_adapter_queue_conf {
* The event adapter sets ev.event_type to RTE_EVENT_TYPE_ETHDEV in the
* enqueued event.
*/
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**
@@ -97,6 +97,10 @@ struct rte_event_eth_tx_adapter_conf {
* max_nb_tx mbufs. This isn't treated as a requirement; batching may
* cause the adapter to process more than max_nb_tx mbufs.
*/
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**
@@ -171,6 +171,10 @@ struct rte_event_timer_adapter_conf {
/**< Total number of timers per adapter */
uint64_t flags;
/**< Timer adapter config flags (RTE_EVENT_TIMER_ADAPTER_F_*) */
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**
@@ -268,6 +272,10 @@ struct rte_event_timer_adapter_info {
/**< Event timer adapter capabilities */
int16_t event_dev_port_id;
/**< Event device port ID, if applicable */
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**
@@ -387,6 +387,10 @@ struct rte_event_dev_info {
*/
uint32_t event_dev_cap;
/**< Event device capabilities(RTE_EVENT_DEV_CAP_)*/
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**
@@ -494,6 +498,10 @@ struct rte_event_dev_config {
*/
uint32_t event_dev_cfg;
/**< Event device config flags(RTE_EVENT_DEV_CFG_)*/
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**
@@ -574,6 +582,10 @@ struct rte_event_queue_conf {
* event device supported priority value.
* Valid when the device has RTE_EVENT_DEV_CAP_QUEUE_QOS capability
*/
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**
@@ -705,6 +717,10 @@ struct rte_event_port_conf {
* RTE_EVENT_OP_FORWARD. Must be false when the device is not
* RTE_EVENT_DEV_CAP_IMPLICIT_RELEASE_DISABLE capable.
*/
+ uint64_t reserved_64s[4];
+ /**< Reserved for future fields. */
+ void *reserved_ptrs[4];
+ /**< Reserved for future fields. */
};
/**