[RFC,3/3] net/mlx5: support enhanced multi-packet write on Windows
Checks
Commit Message
Add support for enhanced multi-packet write on Windows.
Enhanced multi-packet write allows the Tx burst function to pack up
multiple packets in a single descriptor session to save PCI bandwidth
and improve performance.
The feature can be controlled by the txq_mpw_en PMD argument:
txq_mpw_en=1 - PMD will first attempt to use "enhanced multi packet write"
if the feature is not supported by the HW the legacy "multi packet write"
will be used.
if both are unsupported the multi packet write feature is disabled.
txq_mpw_en=0 - multi packet write is disabled.
txq_mpw_en unset(default) - enhanced multi packet write
will be activated if supported.
if unsupported the multi packet write feature is disabled.
Signed-off-by: Tal Shnaiderman <talshn@nvidia.com>
---
drivers/common/mlx5/mlx5_devx_cmds.c | 6 ++++++
drivers/common/mlx5/mlx5_devx_cmds.h | 2 ++
drivers/net/mlx5/windows/mlx5_os.c | 14 ++++++++++++--
3 files changed, 20 insertions(+), 2 deletions(-)
@@ -1119,6 +1119,12 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
attr->rss_ind_tbl_cap = MLX5_GET
(per_protocol_networking_offload_caps,
hcattr, rss_ind_tbl_cap);
+ attr->multi_pkt_send_wqe = MLX5_GET
+ (per_protocol_networking_offload_caps,
+ hcattr, multi_pkt_send_wqe);
+ attr->enhanced_multi_pkt_send_wqe = MLX5_GET
+ (per_protocol_networking_offload_caps,
+ hcattr, enhanced_multi_pkt_send_wqe);
/* Query HCA attribute for ROCE. */
if (attr->roce) {
hcattr = mlx5_devx_get_hca_cap(ctx, in, out, &rc,
@@ -259,6 +259,8 @@ struct mlx5_hca_attr {
uint32_t striding_rq:1;
uint32_t ext_stride_num_range:1;
uint32_t cqe_compression_128:1;
+ uint32_t multi_pkt_send_wqe:1;
+ uint32_t enhanced_multi_pkt_send_wqe:1;
};
/* LAG Context. */
@@ -173,8 +173,6 @@ mlx5_os_capabilities_prepare(struct mlx5_dev_ctx_shared *sh)
sh->dev_cap.max_qp = 1 << hca_attr->log_max_qp;
sh->dev_cap.max_qp_wr = 1 << hca_attr->log_max_qp_sz;
sh->dev_cap.dv_flow_en = 1;
- sh->dev_cap.mps = MLX5_MPW_DISABLED;
- DRV_LOG(DEBUG, "MPW isn't supported.");
DRV_LOG(DEBUG, "MPLS over GRE/UDP tunnel offloading is no supported.");
sh->dev_cap.hw_csum = hca_attr->csum_cap;
DRV_LOG(DEBUG, "Checksum offloading is %ssupported.",
@@ -224,6 +222,18 @@ mlx5_os_capabilities_prepare(struct mlx5_dev_ctx_shared *sh)
DRV_LOG(DEBUG, "Maximum Rx indirection table size is %u",
sh->dev_cap.ind_table_max_size);
}
+ if (hca_attr->enhanced_multi_pkt_send_wqe) {
+ sh->dev_cap.mps = MLX5_MPW_ENHANCED;
+ DRV_LOG(DEBUG, "Enhanced MPW is supported.");
+ }
+ else if (hca_attr->multi_pkt_send_wqe &&
+ sh->dev_cap.mps != MLX5_ARG_UNSET) {
+ sh->dev_cap.mps = MLX5_MPW;
+ DRV_LOG(DEBUG, "MPW is supported.");
+ } else {
+ sh->dev_cap.mps = MLX5_MPW_DISABLED;
+ DRV_LOG(DEBUG, "MPW isn't supported.");
+ }
sh->dev_cap.swp = mlx5_get_supported_sw_parsing_offloads(hca_attr);
sh->dev_cap.tunnel_en = mlx5_get_supported_tunneling_offloads(hca_attr);
if (sh->dev_cap.tunnel_en) {