From patchwork Tue Jun 19 21:30:47 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ajit Khaparde X-Patchwork-Id: 41293 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@dpdk.org Delivered-To: patchwork@dpdk.org Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id C2B861B470; Tue, 19 Jun 2018 23:32:00 +0200 (CEST) Received: from rnd-relay.smtp.broadcom.com (rnd-relay.smtp.broadcom.com [192.19.229.170]) by dpdk.org (Postfix) with ESMTP id E27D41B057 for ; Tue, 19 Jun 2018 23:31:09 +0200 (CEST) Received: from nis-sj1-27.broadcom.com (nis-sj1-27.lvn.broadcom.net [10.75.144.136]) by rnd-relay.smtp.broadcom.com (Postfix) with ESMTP id 5AA4130C078; Tue, 19 Jun 2018 14:31:07 -0700 (PDT) DKIM-Filter: OpenDKIM Filter v2.10.3 rnd-relay.smtp.broadcom.com 5AA4130C078 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=broadcom.com; s=dkimrelay; t=1529443867; bh=nEy+cIHiSi2v++Xzs6K6Pk1cJUxBuVzEIej0Yuqk/eE=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=L1cz9c1cFewQfmL/TqEjoxN82XXah+YtPkxvHRwl16E6DhQIuq5h0N6CfQAqKvg32 5VZDMGQBz9hZviA2PSDX+z8I0KZpBuHDc7GeciFDMIvpkhVZglxGRPZG8z5mEnqO0k IRSSQe9ga0s6JZ5zVGCdDxCCQkEzUoRB9nRjXhIY= Received: from C02VPB22HTD6.dhcp.broadcom.net (c02vpb22htd6.dhcp.broadcom.net [10.136.50.120]) by nis-sj1-27.broadcom.com (Postfix) with ESMTP id 8BD7EAC0799; Tue, 19 Jun 2018 14:31:07 -0700 (PDT) From: Ajit Khaparde To: dev@dpdk.org Cc: ferruh.yigit@intel.com, Scott Branden Date: Tue, 19 Jun 2018 14:30:47 -0700 Message-Id: <20180619213058.12273-21-ajit.khaparde@broadcom.com> X-Mailer: git-send-email 2.15.1 (Apple Git-101) In-Reply-To: <20180619213058.12273-1-ajit.khaparde@broadcom.com> References: <20180619213058.12273-1-ajit.khaparde@broadcom.com> Subject: [dpdk-dev] [PATCH 20/31] net/bnxt: code cleanup style of bnxt ethdev X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" From: Scott Branden Cleanup alignment, brackets, debug string style of bnxt_ethdev Signed-off-by: Scott Branden Reviewed-by: Randy Schacher Reviewed-by: Ajit Khaparde --- drivers/net/bnxt/bnxt_ethdev.c | 204 ++++++++++++++++++++++------------------- 1 file changed, 112 insertions(+), 92 deletions(-) diff --git a/drivers/net/bnxt/bnxt_ethdev.c b/drivers/net/bnxt/bnxt_ethdev.c index d66a29758..6516aeedd 100644 --- a/drivers/net/bnxt/bnxt_ethdev.c +++ b/drivers/net/bnxt/bnxt_ethdev.c @@ -315,8 +315,9 @@ static int bnxt_init_chip(struct bnxt *bp) intr_vector = bp->eth_dev->data->nb_rx_queues; PMD_DRV_LOG(DEBUG, "intr_vector = %d\n", intr_vector); if (intr_vector > bp->rx_cp_nr_rings) { - PMD_DRV_LOG(ERR, "At most %d intr queues supported", - bp->rx_cp_nr_rings); + PMD_DRV_LOG(ERR, + "At most %d intr queues supported\n", + bp->rx_cp_nr_rings); return -ENOTSUP; } if (rte_intr_efd_enable(intr_handle, intr_vector)) @@ -329,14 +330,15 @@ static int bnxt_init_chip(struct bnxt *bp) bp->eth_dev->data->nb_rx_queues * sizeof(int), 0); if (intr_handle->intr_vec == NULL) { - PMD_DRV_LOG(ERR, "Failed to allocate %d rx_queues" - " intr_vec", bp->eth_dev->data->nb_rx_queues); + PMD_DRV_LOG(ERR, + "Failed to allocate %d rx_queues intr_vec\n", + bp->eth_dev->data->nb_rx_queues); return -ENOMEM; } - PMD_DRV_LOG(DEBUG, "intr_handle->intr_vec = %p " - "intr_handle->nb_efd = %d intr_handle->max_intr = %d\n", - intr_handle->intr_vec, intr_handle->nb_efd, - intr_handle->max_intr); + PMD_DRV_LOG(DEBUG, + "intr_handle->intr_vec = %p intr_handle->nb_efd = %d intr_handle->max_intr = %d\n", + intr_handle->intr_vec, intr_handle->nb_efd, + intr_handle->max_intr); } for (queue_id = 0; queue_id < bp->eth_dev->data->nb_rx_queues; @@ -404,7 +406,7 @@ static int bnxt_init_nic(struct bnxt *bp) */ static void bnxt_dev_info_get_op(struct rte_eth_dev *eth_dev, - struct rte_eth_dev_info *dev_info) + struct rte_eth_dev_info *dev_info) { struct bnxt *bp = (struct bnxt *)eth_dev->data->dev_private; uint16_t max_vnics, i, j, vpool, vrxq; @@ -706,15 +708,22 @@ static void bnxt_mac_addr_remove_op(struct rte_eth_dev *eth_dev, while (filter) { temp_filter = STAILQ_NEXT(filter, next); if (filter->mac_index == index) { - STAILQ_REMOVE(&vnic->filter, filter, - bnxt_filter_info, next); + STAILQ_REMOVE(&vnic->filter, + filter, + bnxt_filter_info, + next); + bnxt_hwrm_clear_l2_filter(bp, filter); filter->mac_index = INVALID_MAC_INDEX; - memset(&filter->l2_addr, 0, + + memset(&filter->l2_addr, + 0, ETHER_ADDR_LEN); - STAILQ_INSERT_TAIL( - &bp->free_filter_list, - filter, next); + + STAILQ_INSERT_TAIL + (&bp->free_filter_list, + filter, + next); } filter = temp_filter; } @@ -785,9 +794,10 @@ int bnxt_link_update_op(struct rte_eth_dev *eth_dev, int wait_to_complete) out: /* Timed out or success */ if (new.link_status != eth_dev->data->dev_link.link_status || - new.link_speed != eth_dev->data->dev_link.link_speed) { - memcpy(ð_dev->data->dev_link, &new, - sizeof(struct rte_eth_link)); + new.link_speed != eth_dev->data->dev_link.link_speed) { + memcpy(ð_dev->data->dev_link, + &new, + sizeof(struct rte_eth_link)); _rte_eth_dev_callback_process(eth_dev, RTE_ETH_EVENT_INTR_LSC, @@ -856,8 +866,8 @@ static void bnxt_allmulticast_disable_op(struct rte_eth_dev *eth_dev) } static int bnxt_reta_update_op(struct rte_eth_dev *eth_dev, - struct rte_eth_rss_reta_entry64 *reta_conf, - uint16_t reta_size) + struct rte_eth_rss_reta_entry64 *reta_conf, + uint16_t reta_size) { struct bnxt *bp = (struct bnxt *)eth_dev->data->dev_private; struct rte_eth_conf *dev_conf = &bp->eth_dev->data->dev_conf; @@ -868,9 +878,9 @@ static int bnxt_reta_update_op(struct rte_eth_dev *eth_dev, return -EINVAL; if (reta_size != HW_HASH_INDEX_SIZE) { - PMD_DRV_LOG(ERR, "The configured hash table lookup size " - "(%d) must equal the size supported by the hardware " - "(%d)\n", reta_size, HW_HASH_INDEX_SIZE); + PMD_DRV_LOG(ERR, + "Configured hash table lookup size (%d) != (%d)\n", + reta_size, HW_HASH_INDEX_SIZE); return -EINVAL; } /* Update the RSS VNIC(s) */ @@ -900,9 +910,9 @@ static int bnxt_reta_query_op(struct rte_eth_dev *eth_dev, return -EINVAL; if (reta_size != HW_HASH_INDEX_SIZE) { - PMD_DRV_LOG(ERR, "The configured hash table lookup size " - "(%d) must equal the size supported by the hardware " - "(%d)\n", reta_size, HW_HASH_INDEX_SIZE); + PMD_DRV_LOG(ERR, + "Configured hash table lookup size (%d) != (%d)\n", + reta_size, HW_HASH_INDEX_SIZE); return -EINVAL; } /* EW - need to revisit here copying from uint64_t to uint16_t */ @@ -1021,8 +1031,8 @@ static int bnxt_rss_hash_conf_get_op(struct rte_eth_dev *eth_dev, } if (hash_types) { PMD_DRV_LOG(ERR, - "Unknwon RSS config from firmware (%08x), RSS disabled", - vnic->hash_type); + "Unknown RSS config (%08x), RSS disabled\n", + vnic->hash_type); return -ENOTSUP; } } else { @@ -1032,7 +1042,7 @@ static int bnxt_rss_hash_conf_get_op(struct rte_eth_dev *eth_dev, } static int bnxt_flow_ctrl_get_op(struct rte_eth_dev *dev, - struct rte_eth_fc_conf *fc_conf) + struct rte_eth_fc_conf *fc_conf) { struct bnxt *bp = (struct bnxt *)dev->data->dev_private; struct rte_eth_link link_info; @@ -1064,7 +1074,7 @@ static int bnxt_flow_ctrl_get_op(struct rte_eth_dev *dev, } static int bnxt_flow_ctrl_set_op(struct rte_eth_dev *dev, - struct rte_eth_fc_conf *fc_conf) + struct rte_eth_fc_conf *fc_conf) { struct bnxt *bp = (struct bnxt *)dev->data->dev_private; @@ -1120,7 +1130,7 @@ static int bnxt_flow_ctrl_set_op(struct rte_eth_dev *dev, /* Add UDP tunneling port */ static int bnxt_udp_tunnel_port_add_op(struct rte_eth_dev *eth_dev, - struct rte_eth_udp_tunnel *udp_tunnel) + struct rte_eth_udp_tunnel *udp_tunnel) { struct bnxt *bp = (struct bnxt *)eth_dev->data->dev_private; uint16_t tunnel_type = 0; @@ -1168,7 +1178,7 @@ bnxt_udp_tunnel_port_add_op(struct rte_eth_dev *eth_dev, static int bnxt_udp_tunnel_port_del_op(struct rte_eth_dev *eth_dev, - struct rte_eth_udp_tunnel *udp_tunnel) + struct rte_eth_udp_tunnel *udp_tunnel) { struct bnxt *bp = (struct bnxt *)eth_dev->data->dev_private; uint16_t tunnel_type = 0; @@ -1256,9 +1266,10 @@ static int bnxt_del_vlan_filter(struct bnxt *bp, uint16_t vlan_id) STAILQ_REMOVE(&vnic->filter, filter, bnxt_filter_info, next); bnxt_hwrm_clear_l2_filter(bp, filter); - STAILQ_INSERT_TAIL( - &bp->free_filter_list, - filter, next); + STAILQ_INSERT_TAIL + (&bp->free_filter_list, + filter, + next); /* * Need to examine to see if the MAC @@ -1281,9 +1292,10 @@ static int bnxt_del_vlan_filter(struct bnxt *bp, uint16_t vlan_id) memcpy(new_filter->l2_addr, filter->l2_addr, ETHER_ADDR_LEN); /* MAC only filter */ - rc = bnxt_hwrm_set_l2_filter(bp, - vnic->fw_vnic_id, - new_filter); + rc = bnxt_hwrm_set_l2_filter + (bp, + vnic->fw_vnic_id, + new_filter); if (rc) goto exit; PMD_DRV_LOG(INFO, @@ -1335,9 +1347,10 @@ static int bnxt_add_vlan_filter(struct bnxt *bp, uint16_t vlan_id) bnxt_filter_info, next); bnxt_hwrm_clear_l2_filter(bp, filter); filter->l2_ovlan = 0; - STAILQ_INSERT_TAIL( - &bp->free_filter_list, - filter, next); + STAILQ_INSERT_TAIL + (&bp->free_filter_list, + filter, + next); } new_filter = bnxt_alloc_filter(bp); if (!new_filter) { @@ -1405,6 +1418,7 @@ bnxt_vlan_offload_set_op(struct rte_eth_dev *dev, int mask) /* Enable or disable VLAN stripping */ for (i = 0; i < bp->nr_vnics; i++) { struct bnxt_vnic_info *vnic = &bp->vnic_info[i]; + if (rx_offloads & DEV_RX_OFFLOAD_VLAN_STRIP) vnic->vlan_strip = true; else @@ -1460,8 +1474,8 @@ bnxt_set_default_mac_addr_op(struct rte_eth_dev *dev, struct ether_addr *addr) static int bnxt_dev_set_mc_addr_list_op(struct rte_eth_dev *eth_dev, - struct ether_addr *mc_addr_set, - uint32_t nb_mc_addr) + struct ether_addr *mc_addr_set, + uint32_t nb_mc_addr) { struct bnxt *bp = (struct bnxt *)eth_dev->data->dev_private; char *mc_addr_list = (char *)mc_addr_set; @@ -1497,8 +1511,9 @@ bnxt_fw_version_get(struct rte_eth_dev *dev, char *fw_version, size_t fw_size) uint8_t fw_updt = (bp->fw_ver >> 8) & 0xff; int ret; - ret = snprintf(fw_version, fw_size, "%d.%d.%d", - fw_major, fw_minor, fw_updt); + ret = snprintf(fw_version, fw_size, + "%d.%d.%d", + fw_major, fw_minor, fw_updt); ret += 1; /* add the size of '\0' */ if (fw_size < (uint32_t)ret) @@ -1508,8 +1523,9 @@ bnxt_fw_version_get(struct rte_eth_dev *dev, char *fw_version, size_t fw_size) } static void -bnxt_rxq_info_get_op(struct rte_eth_dev *dev, uint16_t queue_id, - struct rte_eth_rxq_info *qinfo) +bnxt_rxq_info_get_op(struct rte_eth_dev *dev, + uint16_t queue_id, + struct rte_eth_rxq_info *qinfo) { struct bnxt_rx_queue *rxq; @@ -1525,8 +1541,9 @@ bnxt_rxq_info_get_op(struct rte_eth_dev *dev, uint16_t queue_id, } static void -bnxt_txq_info_get_op(struct rte_eth_dev *dev, uint16_t queue_id, - struct rte_eth_txq_info *qinfo) +bnxt_txq_info_get_op(struct rte_eth_dev *dev, + uint16_t queue_id, + struct rte_eth_txq_info *qinfo) { struct bnxt_tx_queue *txq; @@ -1561,7 +1578,6 @@ static int bnxt_mtu_set_op(struct rte_eth_dev *eth_dev, uint16_t new_mtu) return -EINVAL; } - if (new_mtu > ETHER_MTU) { bp->flags |= BNXT_FLAG_JUMBO; bp->eth_dev->data->dev_conf.rxmode.offloads |= @@ -1655,17 +1671,16 @@ bnxt_rx_queue_count_op(struct rte_eth_dev *dev, uint16_t rx_queue_id) valid = FLIP_VALID(cons, cpr->cp_ring_struct->ring_mask, valid); cmp_type = CMP_TYPE(rxcmp); if (cmp_type == RX_TPA_END_CMPL_TYPE_RX_TPA_END) { - cmp = (rte_le_to_cpu_32( - ((struct rx_tpa_end_cmpl *) - (rxcmp))->agg_bufs_v1) & - RX_TPA_END_CMPL_AGG_BUFS_MASK) >> - RX_TPA_END_CMPL_AGG_BUFS_SFT; + cmp = (rte_le_to_cpu_32 + (((struct rx_tpa_end_cmpl *) + (rxcmp))->agg_bufs_v1) + & RX_TPA_END_CMPL_AGG_BUFS_MASK) + >> RX_TPA_END_CMPL_AGG_BUFS_SFT; desc++; } else if (cmp_type == 0x11) { desc++; - cmp = (rxcmp->agg_bufs_v1 & - RX_PKT_CMPL_AGG_BUFS_MASK) >> - RX_PKT_CMPL_AGG_BUFS_SFT; + cmp = (rxcmp->agg_bufs_v1 & RX_PKT_CMPL_AGG_BUFS_MASK) + >> RX_PKT_CMPL_AGG_BUFS_SFT; } else { cmp = 1; } @@ -1710,7 +1725,6 @@ bnxt_rx_descriptor_status_op(void *rx_queue, uint16_t offset) if (rx_buf->mbuf == NULL) return RTE_ETH_RX_DESC_UNAVAIL; - return RTE_ETH_RX_DESC_AVAIL; } @@ -2882,16 +2896,20 @@ bnxt_get_eeprom_length_op(struct rte_eth_dev *dev) static int bnxt_get_eeprom_op(struct rte_eth_dev *dev, - struct rte_dev_eeprom_info *in_eeprom) + struct rte_dev_eeprom_info *in_eeprom) { struct bnxt *bp = (struct bnxt *)dev->data->dev_private; uint32_t index; uint32_t offset; - PMD_DRV_LOG(INFO, "%04x:%02x:%02x:%02x in_eeprom->offset = %d " - "len = %d\n", bp->pdev->addr.domain, - bp->pdev->addr.bus, bp->pdev->addr.devid, - bp->pdev->addr.function, in_eeprom->offset, in_eeprom->length); + PMD_DRV_LOG(INFO, + "%04x:%02x:%02x:%02x in_eeprom->offset = %d len = %d\n", + bp->pdev->addr.domain, + bp->pdev->addr.bus, + bp->pdev->addr.devid, + bp->pdev->addr.function, + in_eeprom->offset, + in_eeprom->length); if (in_eeprom->offset == 0) /* special offset value to get directory */ return bnxt_get_nvram_directory(bp, in_eeprom->length, @@ -2953,16 +2971,17 @@ static bool bnxt_dir_type_is_executable(uint16_t dir_type) static int bnxt_set_eeprom_op(struct rte_eth_dev *dev, - struct rte_dev_eeprom_info *in_eeprom) + struct rte_dev_eeprom_info *in_eeprom) { struct bnxt *bp = (struct bnxt *)dev->data->dev_private; uint8_t index, dir_op; uint16_t type, ext, ordinal, attr; - PMD_DRV_LOG(INFO, "%04x:%02x:%02x:%02x in_eeprom->offset = %d " - "len = %d\n", bp->pdev->addr.domain, - bp->pdev->addr.bus, bp->pdev->addr.devid, - bp->pdev->addr.function, in_eeprom->offset, in_eeprom->length); + PMD_DRV_LOG(INFO, + "%04x:%02x:%02x:%02x in_eeprom->offset = %d len = %d\n", + bp->pdev->addr.domain, bp->pdev->addr.bus, + bp->pdev->addr.devid, bp->pdev->addr.function, + in_eeprom->offset, in_eeprom->length); if (!BNXT_PF(bp)) { PMD_DRV_LOG(ERR, "NVM write not supported from a VF\n"); @@ -3195,14 +3214,14 @@ bnxt_dev_init(struct rte_eth_dev *eth_dev) pci_dev->addr.function, "rx_port_stats"); mz_name[RTE_MEMZONE_NAMESIZE - 1] = 0; mz = rte_memzone_lookup(mz_name); - total_alloc_len = RTE_CACHE_LINE_ROUNDUP( - sizeof(struct rx_port_stats) + 512); + total_alloc_len = RTE_CACHE_LINE_ROUNDUP + (sizeof(struct rx_port_stats) + 512); if (!mz) { mz = rte_memzone_reserve(mz_name, total_alloc_len, - SOCKET_ID_ANY, - RTE_MEMZONE_2MB | - RTE_MEMZONE_SIZE_HINT_ONLY | - RTE_MEMZONE_IOVA_CONTIG); + SOCKET_ID_ANY, + RTE_MEMZONE_2MB | + RTE_MEMZONE_SIZE_HINT_ONLY | + RTE_MEMZONE_IOVA_CONTIG); if (mz == NULL) return -ENOMEM; } @@ -3216,7 +3235,7 @@ bnxt_dev_init(struct rte_eth_dev *eth_dev) mz_phys_addr = rte_mem_virt2iova(mz->addr); if (mz_phys_addr == 0) { PMD_DRV_LOG(ERR, - "unable to map address to physical memory\n"); + "unable to map addr to phys memory\n"); return -ENOMEM; } } @@ -3231,15 +3250,15 @@ bnxt_dev_init(struct rte_eth_dev *eth_dev) pci_dev->addr.function, "tx_port_stats"); mz_name[RTE_MEMZONE_NAMESIZE - 1] = 0; mz = rte_memzone_lookup(mz_name); - total_alloc_len = RTE_CACHE_LINE_ROUNDUP( - sizeof(struct tx_port_stats) + 512); + total_alloc_len = RTE_CACHE_LINE_ROUNDUP + (sizeof(struct tx_port_stats) + 512); if (!mz) { mz = rte_memzone_reserve(mz_name, - total_alloc_len, - SOCKET_ID_ANY, - RTE_MEMZONE_2MB | - RTE_MEMZONE_SIZE_HINT_ONLY | - RTE_MEMZONE_IOVA_CONTIG); + total_alloc_len, + SOCKET_ID_ANY, + RTE_MEMZONE_2MB | + RTE_MEMZONE_SIZE_HINT_ONLY | + RTE_MEMZONE_IOVA_CONTIG); if (mz == NULL) return -ENOMEM; } @@ -3253,7 +3272,7 @@ bnxt_dev_init(struct rte_eth_dev *eth_dev) mz_phys_addr = rte_mem_virt2iova(mz->addr); if (mz_phys_addr == 0) { PMD_DRV_LOG(ERR, - "unable to map address to physical memory\n"); + "unable to map to phys memory\n"); return -ENOMEM; } } @@ -3298,10 +3317,11 @@ bnxt_dev_init(struct rte_eth_dev *eth_dev) goto error_free; } eth_dev->data->mac_addrs = rte_zmalloc("bnxt_mac_addr_tbl", - ETHER_ADDR_LEN * bp->max_l2_ctx, 0); + ETHER_ADDR_LEN * bp->max_l2_ctx, + 0); if (eth_dev->data->mac_addrs == NULL) { PMD_DRV_LOG(ERR, - "Failed to alloc %u bytes needed to store MAC addr tbl", + "Failed to alloc %u bytes to store MAC addr tbl\n", ETHER_ADDR_LEN * bp->max_l2_ctx); rc = -ENOMEM; goto error_free; @@ -3328,7 +3348,8 @@ bnxt_dev_init(struct rte_eth_dev *eth_dev) } bp->grp_info = rte_zmalloc("bnxt_grp_info", - sizeof(*bp->grp_info) * bp->max_ring_grps, 0); + sizeof(*bp->grp_info) * bp->max_ring_grps, + 0); if (!bp->grp_info) { PMD_DRV_LOG(ERR, "Failed to alloc %zu bytes to store group info table\n", @@ -3339,8 +3360,8 @@ bnxt_dev_init(struct rte_eth_dev *eth_dev) /* Forward all requests if firmware is new enough */ if (((bp->fw_ver >= ((20 << 24) | (6 << 16) | (100 << 8))) && - (bp->fw_ver < ((20 << 24) | (7 << 16)))) || - ((bp->fw_ver >= ((20 << 24) | (8 << 16))))) { + (bp->fw_ver < ((20 << 24) | (7 << 16)))) || + (bp->fw_ver >= ((20 << 24) | (8 << 16)))) { memset(bp->pf.vf_req_fwd, 0xff, sizeof(bp->pf.vf_req_fwd)); } else { PMD_DRV_LOG(WARNING, @@ -3363,8 +3384,7 @@ bnxt_dev_init(struct rte_eth_dev *eth_dev) ALLOW_FUNC(HWRM_VNIC_TPA_CFG); rc = bnxt_hwrm_func_driver_register(bp); if (rc) { - PMD_DRV_LOG(ERR, - "Failed to register driver"); + PMD_DRV_LOG(ERR, "Failed to register driver\n"); rc = -EBUSY; goto error_free; } @@ -3477,7 +3497,7 @@ bnxt_dev_uninit(struct rte_eth_dev *eth_dev) } static int bnxt_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, - struct rte_pci_device *pci_dev) + struct rte_pci_device *pci_dev) { return rte_eth_dev_pci_generic_probe(pci_dev, sizeof(struct bnxt), bnxt_dev_init);