@@ -168,7 +168,7 @@ test_op_forward_mode(uint8_t session_less)
{
struct rte_crypto_sym_xform cipher_xform;
struct rte_cryptodev_sym_session *sess;
- union rte_event_crypto_metadata m_data;
+ struct rte_event_crypto_metadata m_data;
struct rte_crypto_sym_op *sym_op;
struct rte_crypto_op *op;
struct rte_mbuf *m;
@@ -368,7 +368,7 @@ test_op_new_mode(uint8_t session_less)
{
struct rte_crypto_sym_xform cipher_xform;
struct rte_cryptodev_sym_session *sess;
- union rte_event_crypto_metadata m_data;
+ struct rte_event_crypto_metadata m_data;
struct rte_crypto_sym_op *sym_op;
struct rte_crypto_op *op;
struct rte_mbuf *m;
@@ -406,7 +406,7 @@ test_op_new_mode(uint8_t session_less)
if (cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_SESSION_PRIVATE_DATA) {
/* Fill in private user data information */
rte_memcpy(&m_data.response_info, &response_info,
- sizeof(m_data));
+ sizeof(response_info));
rte_cryptodev_sym_session_set_user_data(sess,
&m_data, sizeof(m_data));
}
@@ -426,7 +426,7 @@ test_op_new_mode(uint8_t session_less)
op->private_data_offset = len;
/* Fill in private data information */
rte_memcpy(&m_data.response_info, &response_info,
- sizeof(m_data));
+ sizeof(response_info));
rte_memcpy((uint8_t *)op + len, &m_data, sizeof(m_data));
}
@@ -519,7 +519,7 @@ configure_cryptodev(void)
DEFAULT_NUM_XFORMS *
sizeof(struct rte_crypto_sym_xform) +
MAXIMUM_IV_LENGTH +
- sizeof(union rte_event_crypto_metadata),
+ sizeof(struct rte_event_crypto_metadata),
rte_socket_id());
if (params.op_mpool == NULL) {
RTE_LOG(ERR, USER1, "Can't create CRYPTO_OP_POOL\n");
@@ -549,12 +549,12 @@ configure_cryptodev(void)
* to include the session headers & private data
*/
session_size = rte_cryptodev_sym_get_private_session_size(TEST_CDEV_ID);
- session_size += sizeof(union rte_event_crypto_metadata);
+ session_size += sizeof(struct rte_event_crypto_metadata);
params.session_mpool = rte_cryptodev_sym_session_pool_create(
"CRYPTO_ADAPTER_SESSION_MP",
MAX_NB_SESSIONS, 0, 0,
- sizeof(union rte_event_crypto_metadata),
+ sizeof(struct rte_event_crypto_metadata),
SOCKET_ID_ANY);
TEST_ASSERT_NOT_NULL(params.session_mpool,
"session mempool allocation failed\n");
@@ -710,17 +710,17 @@ submit_request_to_sso(struct ssows *ws, uintptr_t req,
ssovf_store_pair(add_work, req, ws->grps[rsp_info->queue_id]);
}
-static inline union rte_event_crypto_metadata *
+static inline struct rte_event_crypto_metadata *
get_event_crypto_mdata(struct rte_crypto_op *op)
{
- union rte_event_crypto_metadata *ec_mdata;
+ struct rte_event_crypto_metadata *ec_mdata;
if (op->sess_type == RTE_CRYPTO_OP_WITH_SESSION)
ec_mdata = rte_cryptodev_sym_session_get_user_data(
op->sym->session);
else if (op->sess_type == RTE_CRYPTO_OP_SESSIONLESS &&
op->private_data_offset)
- ec_mdata = (union rte_event_crypto_metadata *)
+ ec_mdata = (struct rte_event_crypto_metadata *)
((uint8_t *)op + op->private_data_offset);
else
return NULL;
@@ -731,7 +731,7 @@ get_event_crypto_mdata(struct rte_crypto_op *op)
uint16_t __rte_hot
otx_crypto_adapter_enqueue(void *port, struct rte_crypto_op *op)
{
- union rte_event_crypto_metadata *ec_mdata;
+ struct rte_event_crypto_metadata *ec_mdata;
struct cpt_instance *instance;
struct cpt_request_info *req;
struct rte_event *rsp_info;
@@ -453,7 +453,7 @@ otx2_ca_enqueue_req(const struct otx2_cpt_qp *qp,
struct rte_crypto_op *op,
uint64_t cpt_inst_w7)
{
- union rte_event_crypto_metadata *m_data;
+ struct rte_event_crypto_metadata *m_data;
union cpt_inst_s inst;
uint64_t lmt_status;
@@ -468,7 +468,7 @@ otx2_ca_enqueue_req(const struct otx2_cpt_qp *qp,
}
} else if (op->sess_type == RTE_CRYPTO_OP_SESSIONLESS &&
op->private_data_offset) {
- m_data = (union rte_event_crypto_metadata *)
+ m_data = (struct rte_event_crypto_metadata *)
((uint8_t *)op +
op->private_data_offset);
} else {
@@ -16,7 +16,7 @@
static inline uint16_t
otx2_ca_enq(uintptr_t tag_op, const struct rte_event *ev)
{
- union rte_event_crypto_metadata *m_data;
+ struct rte_event_crypto_metadata *m_data;
struct rte_crypto_op *crypto_op;
struct rte_cryptodev *cdev;
struct otx2_cpt_qp *qp;
@@ -37,7 +37,7 @@ otx2_ca_enq(uintptr_t tag_op, const struct rte_event *ev)
qp_id = m_data->request_info.queue_pair_id;
} else if (crypto_op->sess_type == RTE_CRYPTO_OP_SESSIONLESS &&
crypto_op->private_data_offset) {
- m_data = (union rte_event_crypto_metadata *)
+ m_data = (struct rte_event_crypto_metadata *)
((uint8_t *)crypto_op +
crypto_op->private_data_offset);
cdev_id = m_data->request_info.cdev_id;
@@ -333,7 +333,7 @@ eca_enq_to_cryptodev(struct rte_event_crypto_adapter *adapter,
struct rte_event *ev, unsigned int cnt)
{
struct rte_event_crypto_adapter_stats *stats = &adapter->crypto_stats;
- union rte_event_crypto_metadata *m_data = NULL;
+ struct rte_event_crypto_metadata *m_data = NULL;
struct crypto_queue_pair_info *qp_info = NULL;
struct rte_crypto_op *crypto_op;
unsigned int i, n;
@@ -371,7 +371,7 @@ eca_enq_to_cryptodev(struct rte_event_crypto_adapter *adapter,
len++;
} else if (crypto_op->sess_type == RTE_CRYPTO_OP_SESSIONLESS &&
crypto_op->private_data_offset) {
- m_data = (union rte_event_crypto_metadata *)
+ m_data = (struct rte_event_crypto_metadata *)
((uint8_t *)crypto_op +
crypto_op->private_data_offset);
cdev_id = m_data->request_info.cdev_id;
@@ -504,7 +504,7 @@ eca_ops_enqueue_burst(struct rte_event_crypto_adapter *adapter,
struct rte_crypto_op **ops, uint16_t num)
{
struct rte_event_crypto_adapter_stats *stats = &adapter->crypto_stats;
- union rte_event_crypto_metadata *m_data = NULL;
+ struct rte_event_crypto_metadata *m_data = NULL;
uint8_t event_dev_id = adapter->eventdev_id;
uint8_t event_port_id = adapter->event_port_id;
struct rte_event events[BATCH_SIZE];
@@ -523,7 +523,7 @@ eca_ops_enqueue_burst(struct rte_event_crypto_adapter *adapter,
ops[i]->sym->session);
} else if (ops[i]->sess_type == RTE_CRYPTO_OP_SESSIONLESS &&
ops[i]->private_data_offset) {
- m_data = (union rte_event_crypto_metadata *)
+ m_data = (struct rte_event_crypto_metadata *)
((uint8_t *)ops[i] +
ops[i]->private_data_offset);
}
@@ -200,11 +200,6 @@ enum rte_event_crypto_adapter_mode {
* provide event request information to the adapter.
*/
struct rte_event_crypto_request {
- uint8_t resv[8];
- /**< Overlaps with first 8 bytes of struct rte_event
- * that encode the response event information. Application
- * is expected to fill in struct rte_event response_info.
- */
uint16_t cdev_id;
/**< cryptodev ID to be used */
uint16_t queue_pair_id;
@@ -223,16 +218,16 @@ struct rte_event_crypto_request {
* operation. If the transfer is done by SW, event response information
* will be used by the adapter.
*/
-union rte_event_crypto_metadata {
- struct rte_event_crypto_request request_info;
- /**< Request information to be filled in by application
- * for RTE_EVENT_CRYPTO_ADAPTER_OP_FORWARD mode.
- */
+struct rte_event_crypto_metadata {
struct rte_event response_info;
/**< Response information to be filled in by application
* for RTE_EVENT_CRYPTO_ADAPTER_OP_NEW and
* RTE_EVENT_CRYPTO_ADAPTER_OP_FORWARD mode.
*/
+ struct rte_event_crypto_request request_info;
+ /**< Request information to be filled in by application
+ * for RTE_EVENT_CRYPTO_ADAPTER_OP_FORWARD mode.
+ */
};
/**