@@ -271,23 +271,23 @@ ci_tx_free_bufs_vec(struct ci_tx_queue *txq, ci_desc_done_fn desc_done, bool ctx
return txq->tx_rs_thresh;
}
-#define IETH_FREE_BUFS_LOOP(txq, swr, start) do { \
+#define IETH_FREE_BUFS_LOOP(swr, nb_desc, start, end) do { \
uint16_t i = start; \
- if (txq->tx_tail < i) { \
- for (; i < txq->nb_tx_desc; i++) { \
+ if (end < i) { \
+ for (; i < nb_desc; i++) { \
rte_pktmbuf_free_seg(swr[i].mbuf); \
swr[i].mbuf = NULL; \
} \
i = 0; \
} \
- for (; i < txq->tx_tail; i++) { \
+ for (; i < end; i++) { \
rte_pktmbuf_free_seg(swr[i].mbuf); \
swr[i].mbuf = NULL; \
} \
} while (0)
static inline void
-ci_txq_release_all_mbufs(struct ci_tx_queue *txq)
+ci_txq_release_all_mbufs(struct ci_tx_queue *txq, bool use_ctx)
{
if (unlikely(!txq || !txq->sw_ring))
return;
@@ -306,15 +306,14 @@ ci_txq_release_all_mbufs(struct ci_tx_queue *txq)
* vPMD tx will not set sw_ring's mbuf to NULL after free,
* so need to free remains more carefully.
*/
- const uint16_t start = txq->tx_next_dd - txq->tx_rs_thresh + 1;
-
- if (txq->vector_sw_ring) {
- struct ci_tx_entry_vec *swr = txq->sw_ring_vec;
- IETH_FREE_BUFS_LOOP(txq, swr, start);
- } else {
- struct ci_tx_entry *swr = txq->sw_ring;
- IETH_FREE_BUFS_LOOP(txq, swr, start);
- }
+ const uint16_t start = (txq->tx_next_dd - txq->tx_rs_thresh + 1) >> use_ctx;
+ const uint16_t nb_desc = txq->nb_tx_desc >> use_ctx;
+ const uint16_t end = txq->tx_tail >> use_ctx;
+
+ if (txq->vector_sw_ring)
+ IETH_FREE_BUFS_LOOP(txq->sw_ring_vec, nb_desc, start, end);
+ else
+ IETH_FREE_BUFS_LOOP(txq->sw_ring, nb_desc, start, end);
}
#endif /* _COMMON_INTEL_TX_H_ */
@@ -1933,7 +1933,7 @@ i40e_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id)
return err;
}
- ci_txq_release_all_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, false);
i40e_reset_tx_queue(txq);
dev->data->tx_queue_state[tx_queue_id] = RTE_ETH_QUEUE_STATE_STOPPED;
@@ -2608,7 +2608,7 @@ i40e_tx_queue_release(void *txq)
return;
}
- ci_txq_release_all_mbufs(q);
+ ci_txq_release_all_mbufs(q, false);
rte_free(q->sw_ring);
rte_memzone_free(q->mz);
rte_free(q);
@@ -3071,7 +3071,7 @@ i40e_dev_clear_queues(struct rte_eth_dev *dev)
for (i = 0; i < dev->data->nb_tx_queues; i++) {
if (!dev->data->tx_queues[i])
continue;
- ci_txq_release_all_mbufs(dev->data->tx_queues[i]);
+ ci_txq_release_all_mbufs(dev->data->tx_queues[i], false);
i40e_reset_tx_queue(dev->data->tx_queues[i]);
}
@@ -387,24 +387,6 @@ release_rxq_mbufs(struct iavf_rx_queue *rxq)
rxq->rx_nb_avail = 0;
}
-static inline void
-release_txq_mbufs(struct ci_tx_queue *txq)
-{
- uint16_t i;
-
- if (!txq || !txq->sw_ring) {
- PMD_DRV_LOG(DEBUG, "Pointer to rxq or sw_ring is NULL");
- return;
- }
-
- for (i = 0; i < txq->nb_tx_desc; i++) {
- if (txq->sw_ring[i].mbuf) {
- rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf);
- txq->sw_ring[i].mbuf = NULL;
- }
- }
-}
-
static const
struct iavf_rxq_ops iavf_rxq_release_mbufs_ops[] = {
[IAVF_REL_MBUFS_DEFAULT].release_mbufs = release_rxq_mbufs,
@@ -413,18 +395,6 @@ struct iavf_rxq_ops iavf_rxq_release_mbufs_ops[] = {
#endif
};
-static const
-struct iavf_txq_ops iavf_txq_release_mbufs_ops[] = {
- [IAVF_REL_MBUFS_DEFAULT].release_mbufs = release_txq_mbufs,
-#ifdef RTE_ARCH_X86
- [IAVF_REL_MBUFS_SSE_VEC].release_mbufs = iavf_tx_queue_release_mbufs_sse,
-#ifdef CC_AVX512_SUPPORT
- [IAVF_REL_MBUFS_AVX512_VEC].release_mbufs = iavf_tx_queue_release_mbufs_avx512,
-#endif
-#endif
-
-};
-
static inline void
iavf_rxd_to_pkt_fields_by_comms_ovs(__rte_unused struct iavf_rx_queue *rxq,
struct rte_mbuf *mb,
@@ -889,7 +859,6 @@ iavf_dev_tx_queue_setup(struct rte_eth_dev *dev,
txq->q_set = true;
dev->data->tx_queues[queue_idx] = txq;
txq->qtx_tail = hw->hw_addr + IAVF_QTX_TAIL1(queue_idx);
- txq->rel_mbufs_type = IAVF_REL_MBUFS_DEFAULT;
if (check_tx_vec_allow(txq) == false) {
struct iavf_adapter *ad =
@@ -1068,7 +1037,7 @@ iavf_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id)
}
txq = dev->data->tx_queues[tx_queue_id];
- iavf_txq_release_mbufs_ops[txq->rel_mbufs_type].release_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, txq->use_ctx);
reset_tx_queue(txq);
dev->data->tx_queue_state[tx_queue_id] = RTE_ETH_QUEUE_STATE_STOPPED;
@@ -1097,7 +1066,7 @@ iavf_dev_tx_queue_release(struct rte_eth_dev *dev, uint16_t qid)
if (!q)
return;
- iavf_txq_release_mbufs_ops[q->rel_mbufs_type].release_mbufs(q);
+ ci_txq_release_all_mbufs(q, q->use_ctx);
rte_free(q->sw_ring);
rte_memzone_free(q->mz);
rte_free(q);
@@ -1114,7 +1083,7 @@ iavf_reset_queues(struct rte_eth_dev *dev)
txq = dev->data->tx_queues[i];
if (!txq)
continue;
- iavf_txq_release_mbufs_ops[txq->rel_mbufs_type].release_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, txq->use_ctx);
reset_tx_queue(txq);
dev->data->tx_queue_state[i] = RTE_ETH_QUEUE_STATE_STOPPED;
}
@@ -2357,31 +2357,11 @@ iavf_xmit_pkts_vec_avx512(void *tx_queue, struct rte_mbuf **tx_pkts,
return iavf_xmit_pkts_vec_avx512_cmn(tx_queue, tx_pkts, nb_pkts, false);
}
-void __rte_cold
-iavf_tx_queue_release_mbufs_avx512(struct ci_tx_queue *txq)
-{
- unsigned int i;
- const uint16_t max_desc = (uint16_t)(txq->nb_tx_desc - 1);
- const uint16_t end_desc = txq->tx_tail >> txq->use_ctx; /* next empty slot */
- const uint16_t wrap_point = txq->nb_tx_desc >> txq->use_ctx; /* end of SW ring */
- struct ci_tx_entry_vec *swr = (void *)txq->sw_ring;
-
- if (!txq->sw_ring || txq->nb_tx_free == max_desc)
- return;
-
- i = (txq->tx_next_dd - txq->tx_rs_thresh + 1) >> txq->use_ctx;
- while (i != end_desc) {
- rte_pktmbuf_free_seg(swr[i].mbuf);
- swr[i].mbuf = NULL;
- if (++i == wrap_point)
- i = 0;
- }
-}
-
int __rte_cold
iavf_txq_vec_setup_avx512(struct ci_tx_queue *txq)
{
- txq->rel_mbufs_type = IAVF_REL_MBUFS_AVX512_VEC;
+ txq->vector_tx = true;
+ txq->vector_sw_ring = true;
return 0;
}
@@ -60,24 +60,6 @@ _iavf_rx_queue_release_mbufs_vec(struct iavf_rx_queue *rxq)
memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc);
}
-static inline void
-_iavf_tx_queue_release_mbufs_vec(struct ci_tx_queue *txq)
-{
- unsigned i;
- const uint16_t max_desc = (uint16_t)(txq->nb_tx_desc - 1);
-
- if (!txq->sw_ring || txq->nb_tx_free == max_desc)
- return;
-
- i = txq->tx_next_dd - txq->tx_rs_thresh + 1;
- while (i != txq->tx_tail) {
- rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf);
- txq->sw_ring[i].mbuf = NULL;
- if (++i == txq->nb_tx_desc)
- i = 0;
- }
-}
-
static inline int
iavf_rxq_vec_setup_default(struct iavf_rx_queue *rxq)
{
@@ -1458,16 +1458,11 @@ iavf_rx_queue_release_mbufs_sse(struct iavf_rx_queue *rxq)
_iavf_rx_queue_release_mbufs_vec(rxq);
}
-void __rte_cold
-iavf_tx_queue_release_mbufs_sse(struct ci_tx_queue *txq)
-{
- _iavf_tx_queue_release_mbufs_vec(txq);
-}
-
int __rte_cold
iavf_txq_vec_setup(struct ci_tx_queue *txq)
{
- txq->rel_mbufs_type = IAVF_REL_MBUFS_SSE_VEC;
+ txq->vector_tx = true;
+ txq->vector_sw_ring = false;
return 0;
}
@@ -501,7 +501,7 @@ ice_dcf_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id)
}
txq = dev->data->tx_queues[tx_queue_id];
- ci_txq_release_all_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, false);
reset_tx_queue(txq);
dev->data->tx_queue_state[tx_queue_id] = RTE_ETH_QUEUE_STATE_STOPPED;
@@ -651,7 +651,7 @@ ice_dcf_stop_queues(struct rte_eth_dev *dev)
txq = dev->data->tx_queues[i];
if (!txq)
continue;
- ci_txq_release_all_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, false);
reset_tx_queue(txq);
dev->data->tx_queue_state[i] = RTE_ETH_QUEUE_STATE_STOPPED;
}
@@ -1089,7 +1089,7 @@ ice_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id)
return -EINVAL;
}
- ci_txq_release_all_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, false);
ice_reset_tx_queue(txq);
dev->data->tx_queue_state[tx_queue_id] = RTE_ETH_QUEUE_STATE_STOPPED;
@@ -1152,7 +1152,7 @@ ice_fdir_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id)
return -EINVAL;
}
- ci_txq_release_all_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, false);
txq->qtx_tail = NULL;
return 0;
@@ -1531,7 +1531,7 @@ ice_tx_queue_release(void *txq)
return;
}
- ci_txq_release_all_mbufs(q);
+ ci_txq_release_all_mbufs(q, false);
rte_free(q->sw_ring);
rte_memzone_free(q->mz);
rte_free(q);
@@ -2457,7 +2457,7 @@ static void __rte_cold
ixgbe_tx_queue_release(struct ci_tx_queue *txq)
{
if (txq != NULL && txq->ops != NULL) {
- ci_txq_release_all_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, false);
txq->ops->free_swring(txq);
rte_memzone_free(txq->mz);
rte_free(txq);
@@ -3364,7 +3364,7 @@ ixgbe_dev_clear_queues(struct rte_eth_dev *dev)
struct ci_tx_queue *txq = dev->data->tx_queues[i];
if (txq != NULL) {
- ci_txq_release_all_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, false);
txq->ops->reset(txq);
dev->data->tx_queue_state[i] = RTE_ETH_QUEUE_STATE_STOPPED;
}
@@ -5639,7 +5639,7 @@ ixgbe_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id)
}
if (txq->ops != NULL) {
- ci_txq_release_all_mbufs(txq);
+ ci_txq_release_all_mbufs(txq, false);
txq->ops->reset(txq);
}
dev->data->tx_queue_state[tx_queue_id] = RTE_ETH_QUEUE_STATE_STOPPED;