/* SPDX-License-Identifier: BSD-3-Clause * Copyright (c) 2013-2015 Brocade Communications Systems, Inc. * Copyright (c) 2015-2018 Cavium Inc. * All rights reserved. * www.cavium.com */ #include "bnx2x.h" #include "bnx2x_rxtx.h" #include #include #include int bnx2x_logtype_init; int bnx2x_logtype_driver; /* * The set of PCI devices this driver supports */ #define BROADCOM_PCI_VENDOR_ID 0x14E4 static const struct rte_pci_id pci_id_bnx2x_map[] = { { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57800) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57711) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_OBS) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_4_10) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_2_20) }, #ifdef RTE_LIBRTE_BNX2X_MF_SUPPORT { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810_MF) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811_MF) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_MF) }, #endif { .vendor_id = 0, } }; static const struct rte_pci_id pci_id_bnx2xvf_map[] = { { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57800_VF) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810_VF) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811_VF) }, { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_VF) }, { .vendor_id = 0, } }; struct rte_bnx2x_xstats_name_off { char name[RTE_ETH_XSTATS_NAME_SIZE]; uint32_t offset_hi; uint32_t offset_lo; }; static const struct rte_bnx2x_xstats_name_off bnx2x_xstats_strings[] = { {"rx_buffer_drops", offsetof(struct bnx2x_eth_stats, brb_drop_hi), offsetof(struct bnx2x_eth_stats, brb_drop_lo)}, {"rx_buffer_truncates", offsetof(struct bnx2x_eth_stats, brb_truncate_hi), offsetof(struct bnx2x_eth_stats, brb_truncate_lo)}, {"rx_buffer_truncate_discard", offsetof(struct bnx2x_eth_stats, brb_truncate_discard), offsetof(struct bnx2x_eth_stats, brb_truncate_discard)}, {"mac_filter_discard", offsetof(struct bnx2x_eth_stats, mac_filter_discard), offsetof(struct bnx2x_eth_stats, mac_filter_discard)}, {"no_match_vlan_tag_discard", offsetof(struct bnx2x_eth_stats, mf_tag_discard), offsetof(struct bnx2x_eth_stats, mf_tag_discard)}, {"tx_pause", offsetof(struct bnx2x_eth_stats, pause_frames_sent_hi), offsetof(struct bnx2x_eth_stats, pause_frames_sent_lo)}, {"rx_pause", offsetof(struct bnx2x_eth_stats, pause_frames_received_hi), offsetof(struct bnx2x_eth_stats, pause_frames_received_lo)}, {"tx_priority_flow_control", offsetof(struct bnx2x_eth_stats, pfc_frames_sent_hi), offsetof(struct bnx2x_eth_stats, pfc_frames_sent_lo)}, {"rx_priority_flow_control", offsetof(struct bnx2x_eth_stats, pfc_frames_received_hi), offsetof(struct bnx2x_eth_stats, pfc_frames_received_lo)} }; static int bnx2x_link_update(struct rte_eth_dev *dev) { struct bnx2x_softc *sc = dev->data->dev_private; struct rte_eth_link link; PMD_INIT_FUNC_TRACE(sc); bnx2x_link_status_update(sc); memset(&link, 0, sizeof(link)); mb(); link.link_speed = sc->link_vars.line_speed; switch (sc->link_vars.duplex) { case DUPLEX_FULL: link.link_duplex = ETH_LINK_FULL_DUPLEX; break; case DUPLEX_HALF: link.link_duplex = ETH_LINK_HALF_DUPLEX; break; } link.link_autoneg = !(dev->data->dev_conf.link_speeds & ETH_LINK_SPEED_FIXED); link.link_status = sc->link_vars.link_up; return rte_eth_linkstatus_set(dev, &link); } static void bnx2x_interrupt_action(struct rte_eth_dev *dev, int intr_cxt) { struct bnx2x_softc *sc = dev->data->dev_private; uint32_t link_status; bnx2x_intr_legacy(sc); if ((atomic_load_acq_long(&sc->periodic_flags) == PERIODIC_GO) && !intr_cxt) bnx2x_periodic_callout(sc); link_status = REG_RD(sc, sc->link_params.shmem_base + offsetof(struct shmem_region, port_mb[sc->link_params.port].link_status)); if ((link_status & LINK_STATUS_LINK_UP) != dev->data->dev_link.link_status) bnx2x_link_update(dev); } static void bnx2x_interrupt_handler(void *param) { struct rte_eth_dev *dev = (struct rte_eth_dev *)param; struct bnx2x_softc *sc = dev->data->dev_private; PMD_DEBUG_PERIODIC_LOG(INFO, sc, "Interrupt handled"); bnx2x_interrupt_action(dev, 1); rte_intr_enable(&sc->pci_dev->intr_handle); } static void bnx2x_periodic_start(void *param) { struct rte_eth_dev *dev = (struct rte_eth_dev *)param; struct bnx2x_softc *sc = dev->data->dev_private; int ret = 0; atomic_store_rel_long(&sc->periodic_flags, PERIODIC_GO); bnx2x_interrupt_action(dev, 0); if (IS_PF(sc)) { ret = rte_eal_alarm_set(BNX2X_SP_TIMER_PERIOD, bnx2x_periodic_start, (void *)dev); if (ret) { PMD_DRV_LOG(ERR, sc, "Unable to start periodic" " timer rc %d", ret); assert(false && "Unable to start periodic timer"); } } } void bnx2x_periodic_stop(void *param) { struct rte_eth_dev *dev = (struct rte_eth_dev *)param; struct bnx2x_softc *sc = dev->data->dev_private; atomic_store_rel_long(&sc->periodic_flags, PERIODIC_STOP); rte_eal_alarm_cancel(bnx2x_periodic_start, (void *)dev); PMD_DRV_LOG(DEBUG, sc, "Periodic poll stopped"); } /* * Devops - helper functions can be called from user application */ static int bnx2x_dev_configure(struct rte_eth_dev *dev) { struct bnx2x_softc *sc = dev->data->dev_private; struct rte_eth_rxmode *rxmode = &dev->data->dev_conf.rxmode; int mp_ncpus = sysconf(_SC_NPROCESSORS_CONF); PMD_INIT_FUNC_TRACE(sc); if (rxmode->offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) { sc->mtu = dev->data->dev_conf.rxmode.max_rx_pkt_len; dev->data->mtu = sc->mtu; } if (dev->data->nb_tx_queues > dev->data->nb_rx_queues) { PMD_DRV_LOG(ERR, sc, "The number of TX queues is greater than number of RX queues"); return -EINVAL; } sc->num_queues = MAX(dev->data->nb_rx_queues, dev->data->nb_tx_queues); if (sc->num_queues > mp_ncpus) { PMD_DRV_LOG(ERR, sc, "The number of queues is more than number of CPUs"); return -EINVAL; } PMD_DRV_LOG(DEBUG, sc, "num_queues=%d, mtu=%d", sc->num_queues, sc->mtu); /* allocate ilt */ if (bnx2x_alloc_ilt_mem(sc) != 0) { PMD_DRV_LOG(ERR, sc, "bnx2x_alloc_ilt_mem was failed"); return -ENXIO; } /* allocate the host hardware/software hsi structures */ if (bnx2x_alloc_hsi_mem(sc) != 0) { PMD_DRV_LOG(ERR, sc, "bnx2x_alloc_hsi_mem was failed"); bnx2x_free_ilt_mem(sc); return -ENXIO; } bnx2x_dev_rxtx_init_dummy(dev); return 0; } static int bnx2x_dev_start(struct rte_eth_dev *dev) { struct bnx2x_softc *sc = dev->data->dev_private; int ret = 0; PMD_INIT_FUNC_TRACE(sc); /* start the periodic callout */ if (atomic_load_acq_long(&sc->periodic_flags) == PERIODIC_STOP) { bnx2x_periodic_start(dev); PMD_DRV_LOG(DEBUG, sc, "Periodic poll re-started"); } ret = bnx2x_init(sc); if (ret) { PMD_DRV_LOG(DEBUG, sc, "bnx2x_init failed (%d)", ret); return -1; } if (IS_PF(sc)) { rte_intr_callback_register(&sc->pci_dev->intr_handle, bnx2x_interrupt_handler, (void *)dev); if (rte_intr_enable(&sc->pci_dev->intr_handle)) PMD_DRV_LOG(ERR, sc, "rte_intr_enable failed"); } bnx2x_dev_rxtx_init(dev); bnx2x_print_device_info(sc); return ret; } static void bnx2x_dev_stop(struct rte_eth_dev *dev) { struct bnx2x_softc *sc = dev->data->dev_private; int ret = 0; PMD_INIT_FUNC_TRACE(sc); bnx2x_dev_rxtx_init_dummy(dev); if (IS_PF(sc)) { rte_intr_disable(&sc->pci_dev->intr_handle); rte_intr_callback_unregister(&sc->pci_dev->intr_handle, bnx2x_interrupt_handler, (void *)dev); } /* stop the periodic callout */ bnx2x_periodic_stop(dev); ret = bnx2x_nic_unload(sc, UNLOAD_NORMAL, FALSE); if (ret) { PMD_DRV_LOG(DEBUG, sc, "bnx2x_nic_unload failed (%d)", ret); return; } return; } static void bnx2x_dev_close(struct rte_eth_dev *dev) { struct bnx2x_softc *sc = dev->data->dev_private; PMD_INIT_FUNC_TRACE(sc); if (IS_VF(sc)) bnx2x_vf_close(sc); bnx2x_dev_clear_queues(dev); memset(&(dev->data->dev_link), 0 , sizeof(struct rte_eth_link)); /* free the host hardware/software hsi structures */ bnx2x_free_hsi_mem(sc); /* free ilt */ bnx2x_free_ilt_mem(sc); } static void bnx2x_promisc_enable(struct rte_eth_dev *dev) { struct bnx2x_softc *sc = dev->data->dev_private; PMD_INIT_FUNC_TRACE(sc); sc->rx_mode = BNX2X_RX_MODE_PROMISC; if (rte_eth_allmulticast_get(dev->data->port_id) == 1) sc->rx_mode = BNX2X_RX_MODE_ALLMULTI_PROMISC; bnx2x_set_rx_mode(sc); } static void bnx2x_promisc_disable(struct rte_eth_dev *dev) { struct bnx2x_softc *sc = dev->data->dev_private; PMD_INIT_FUNC_TRACE(sc); sc->rx_mode = BNX2X_RX_MODE_NORMAL; if (rte_eth_allmulticast_get(dev->data->port_id) == 1) sc->rx_mode = BNX2X_RX_MODE_ALLMULTI; bnx2x_set_rx_mode(sc); } static void bnx2x_dev_allmulticast_enable(struct rte_eth_dev *dev) { struct bnx2x_softc *sc = dev->data->dev_private; PMD_INIT_FUNC_TRACE(sc); sc->rx_mode = BNX2X_RX_MODE_ALLMULTI; if (rte_eth_promiscuous_get(dev->data->port_id) == 1) sc->rx_mode = BNX2X_RX_MODE_ALLMULTI_PROMISC; bnx2x_set_rx_mode(sc); } static void bnx2x_dev_allmulticast_disable(struct rte_eth_dev *dev) { struct bnx2x_softc *sc = dev->data->dev_private; PMD_INIT_FUNC_TRACE(sc); sc->rx_mode = BNX2X_RX_MODE_NORMAL; if (rte_eth_promiscuous_get(dev->data->port_id) == 1) sc->rx_mode = BNX2X_RX_MODE_PROMISC; bnx2x_set_rx_mode(sc); } static int bnx2x_dev_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete) { struct bnx2x_softc *sc = dev->data->dev_private; PMD_INIT_FUNC_TRACE(sc); return bnx2x_link_update(dev); } static int bnx2xvf_dev_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete) { struct bnx2x_softc *sc = dev->data->dev_private; int ret = 0; ret = bnx2x_link_update(dev); bnx2x_check_bull(sc); if (sc->old_bulletin.valid_bitmap & (1 << CHANNEL_DOWN)) { PMD_DRV_LOG(ERR, sc, "PF indicated channel is down." "VF device is no longer operational"); dev->data->dev_link.link_status = ETH_LINK_DOWN; } return ret; } static int bnx2x_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats) { struct bnx2x_softc *sc = dev->data->dev_private; uint32_t brb_truncate_discard; uint64_t brb_drops; uint64_t brb_truncates; PMD_INIT_FUNC_TRACE(sc); bnx2x_stats_handle(sc, STATS_EVENT_UPDATE); memset(stats, 0, sizeof (struct rte_eth_stats)); stats->ipackets = HILO_U64(sc->eth_stats.total_unicast_packets_received_hi, sc->eth_stats.total_unicast_packets_received_lo) + HILO_U64(sc->eth_stats.total_multicast_packets_received_hi, sc->eth_stats.total_multicast_packets_received_lo) + HILO_U64(sc->eth_stats.total_broadcast_packets_received_hi, sc->eth_stats.total_broadcast_packets_received_lo); stats->opackets = HILO_U64(sc->eth_stats.total_unicast_packets_transmitted_hi, sc->eth_stats.total_unicast_packets_transmitted_lo) + HILO_U64(sc->eth_stats.total_multicast_packets_transmitted_hi, sc->eth_stats.total_multicast_packets_transmitted_lo) + HILO_U64(sc->eth_stats.total_broadcast_packets_transmitted_hi, sc->eth_stats.total_broadcast_packets_transmitted_lo); stats->ibytes = HILO_U64(sc->eth_stats.total_bytes_received_hi, sc->eth_stats.total_bytes_received_lo); stats->obytes = HILO_U64(sc->eth_stats.total_bytes_transmitted_hi, sc->eth_stats.total_bytes_transmitted_lo); stats->ierrors = HILO_U64(sc->eth_stats.error_bytes_received_hi, sc->eth_stats.error_bytes_received_lo); stats->oerrors = 0; stats->rx_nombuf = HILO_U64(sc->eth_stats.no_buff_discard_hi, sc->eth_stats.no_buff_discard_lo); brb_drops = HILO_U64(sc->eth_stats.brb_drop_hi, sc->eth_stats.brb_drop_lo); brb_truncates = HILO_U64(sc->eth_stats.brb_truncate_hi, sc->eth_stats.brb_truncate_lo); brb_truncate_discard = sc->eth_stats.brb_truncate_discard; stats->imissed = brb_drops + brb_truncates + brb_truncate_discard + stats->rx_nombuf; return 0; } static int bnx2x_get_xstats_names(__rte_unused struct rte_eth_dev *dev, struct rte_eth_xstat_name *xstats_names, __rte_unused unsigned limit) { unsigned int i, stat_cnt = RTE_DIM(bnx2x_xstats_strings); if (xstats_names != NULL) for (i = 0; i < stat_cnt; i++) snprintf(xstats_names[i].name, sizeof(xstats_names[i].name), "%s", bnx2x_xstats_strings[i].name); return stat_cnt; } static int bnx2x_dev_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats, unsigned int n) { struct bnx2x_softc *sc = dev->data->dev_private; unsigned int num = RTE_DIM(bnx2x_xstats_strings); if (n < num) return num; bnx2x_stats_handle(sc, STATS_EVENT_UPDATE); for (num = 0; num < n; num++) { if (bnx2x_xstats_strings[num].offset_hi != bnx2x_xstats_strings[num].offset_lo) xstats[num].value = HILO_U64( *(uint32_t *)((char *)&sc->eth_stats + bnx2x_xstats_strings[num].offset_hi), *(uint32_t *)((char *)&sc->eth_stats + bnx2x_xstats_strings[num].offset_lo)); else xstats[num].value = *(uint64_t *)((char *)&sc->eth_stats + bnx2x_xstats_strings[num].offset_lo); xstats[num].id = num; } return num; } static void bnx2x_dev_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) { struct bnx2x_softc *sc = dev->data->dev_private; dev_info->max_rx_queues = sc->max_rx_queues; dev_info->max_tx_queues = sc->max_tx_queues; dev_info->min_rx_bufsize = BNX2X_MIN_RX_BUF_SIZE; dev_info->max_rx_pktlen = BNX2X_MAX_RX_PKT_LEN; dev_info->max_mac_addrs = BNX2X_MAX_MAC_ADDRS; dev_info->speed_capa = ETH_LINK_SPEED_10G | ETH_LINK_SPEED_20G; dev_info->rx_offload_capa = DEV_RX_OFFLOAD_JUMBO_FRAME; } static int bnx2x_mac_addr_add(struct rte_eth_dev *dev, struct ether_addr *mac_addr, uint32_t index, uint32_t pool) { struct bnx2x_softc *sc = dev->data->dev_private; if (sc->mac_ops.mac_addr_add) { sc->mac_ops.mac_addr_add(dev, mac_addr, index, pool); return 0; } return -ENOTSUP; } static void bnx2x_mac_addr_remove(struct rte_eth_dev *dev, uint32_t index) { struct bnx2x_softc *sc = dev->data->dev_private; if (sc->mac_ops.mac_addr_remove) sc->mac_ops.mac_addr_remove(dev, index); } static const struct eth_dev_ops bnx2x_eth_dev_ops = { .dev_configure = bnx2x_dev_configure, .dev_start = bnx2x_dev_start, .dev_stop = bnx2x_dev_stop, .dev_close = bnx2x_dev_close, .promiscuous_enable = bnx2x_promisc_enable, .promiscuous_disable = bnx2x_promisc_disable, .allmulticast_enable = bnx2x_dev_allmulticast_enable, .allmulticast_disable = bnx2x_dev_allmulticast_disable, .link_update = bnx2x_dev_link_update, .stats_get = bnx2x_dev_stats_get, .xstats_get = bnx2x_dev_xstats_get, .xstats_get_names = bnx2x_get_xstats_names, .dev_infos_get = bnx2x_dev_infos_get, .rx_queue_setup = bnx2x_dev_rx_queue_setup, .rx_queue_release = bnx2x_dev_rx_queue_release, .tx_queue_setup = bnx2x_dev_tx_queue_setup, .tx_queue_release = bnx2x_dev_tx_queue_release, .mac_addr_add = bnx2x_mac_addr_add, .mac_addr_remove = bnx2x_mac_addr_remove, }; /* * dev_ops for virtual function */ static const struct eth_dev_ops bnx2xvf_eth_dev_ops = { .dev_configure = bnx2x_dev_configure, .dev_start = bnx2x_dev_start, .dev_stop = bnx2x_dev_stop, .dev_close = bnx2x_dev_close, .promiscuous_enable = bnx2x_promisc_enable, .promiscuous_disable = bnx2x_promisc_disable, .allmulticast_enable = bnx2x_dev_allmulticast_enable, .allmulticast_disable = bnx2x_dev_allmulticast_disable, .link_update = bnx2xvf_dev_link_update, .stats_get = bnx2x_dev_stats_get, .xstats_get = bnx2x_dev_xstats_get, .xstats_get_names = bnx2x_get_xstats_names, .dev_infos_get = bnx2x_dev_infos_get, .rx_queue_setup = bnx2x_dev_rx_queue_setup, .rx_queue_release = bnx2x_dev_rx_queue_release, .tx_queue_setup = bnx2x_dev_tx_queue_setup, .tx_queue_release = bnx2x_dev_tx_queue_release, .mac_addr_add = bnx2x_mac_addr_add, .mac_addr_remove = bnx2x_mac_addr_remove, }; static int bnx2x_common_dev_init(struct rte_eth_dev *eth_dev, int is_vf) { int ret = 0; struct rte_pci_device *pci_dev; struct rte_pci_addr pci_addr; struct bnx2x_softc *sc; static bool adapter_info = true; /* Extract key data structures */ sc = eth_dev->data->dev_private; pci_dev = RTE_DEV_TO_PCI(eth_dev->device); pci_addr = pci_dev->addr; snprintf(sc->devinfo.name, NAME_SIZE, PCI_SHORT_PRI_FMT ":dpdk-port-%u", pci_addr.bus, pci_addr.devid, pci_addr.function, eth_dev->data->port_id); PMD_INIT_FUNC_TRACE(sc); eth_dev->dev_ops = is_vf ? &bnx2xvf_eth_dev_ops : &bnx2x_eth_dev_ops; rte_eth_copy_pci_info(eth_dev, pci_dev); sc->pcie_bus = pci_dev->addr.bus; sc->pcie_device = pci_dev->addr.devid; sc->devinfo.vendor_id = pci_dev->id.vendor_id; sc->devinfo.device_id = pci_dev->id.device_id; sc->devinfo.subvendor_id = pci_dev->id.subsystem_vendor_id; sc->devinfo.subdevice_id = pci_dev->id.subsystem_device_id; if (is_vf) sc->flags = BNX2X_IS_VF_FLAG; sc->pcie_func = pci_dev->addr.function; sc->bar[BAR0].base_addr = (void *)pci_dev->mem_resource[0].addr; if (is_vf) sc->bar[BAR1].base_addr = (void *) ((uintptr_t)pci_dev->mem_resource[0].addr + PXP_VF_ADDR_DB_START); else sc->bar[BAR1].base_addr = pci_dev->mem_resource[2].addr; assert(sc->bar[BAR0].base_addr); assert(sc->bar[BAR1].base_addr); bnx2x_load_firmware(sc); assert(sc->firmware); if (eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf & ETH_RSS_NONFRAG_IPV4_UDP) sc->udp_rss = 1; sc->rx_budget = BNX2X_RX_BUDGET; sc->hc_rx_ticks = BNX2X_RX_TICKS; sc->hc_tx_ticks = BNX2X_TX_TICKS; sc->interrupt_mode = INTR_MODE_SINGLE_MSIX; sc->rx_mode = BNX2X_RX_MODE_NORMAL; sc->pci_dev = pci_dev; ret = bnx2x_attach(sc); if (ret) { PMD_DRV_LOG(ERR, sc, "bnx2x_attach failed (%d)", ret); return ret; } /* Print important adapter info for the user. */ if (adapter_info) { bnx2x_print_adapter_info(sc); adapter_info = false; } /* schedule periodic poll for slowpath link events */ if (IS_PF(sc)) { PMD_DRV_LOG(DEBUG, sc, "Scheduling periodic poll for slowpath link events"); ret = rte_eal_alarm_set(BNX2X_SP_TIMER_PERIOD, bnx2x_periodic_start, (void *)eth_dev); if (ret) { PMD_DRV_LOG(ERR, sc, "Unable to start periodic" " timer rc %d", ret); return -EINVAL; } } eth_dev->data->mac_addrs = (struct ether_addr *)sc->link_params.mac_addr; if (IS_VF(sc)) { rte_spinlock_init(&sc->vf2pf_lock); ret = bnx2x_dma_alloc(sc, sizeof(struct bnx2x_vf_mbx_msg), &sc->vf2pf_mbox_mapping, "vf2pf_mbox", RTE_CACHE_LINE_SIZE); if (ret) goto out; sc->vf2pf_mbox = (struct bnx2x_vf_mbx_msg *) sc->vf2pf_mbox_mapping.vaddr; ret = bnx2x_dma_alloc(sc, sizeof(struct bnx2x_vf_bulletin), &sc->pf2vf_bulletin_mapping, "vf2pf_bull", RTE_CACHE_LINE_SIZE); if (ret) goto out; sc->pf2vf_bulletin = (struct bnx2x_vf_bulletin *) sc->pf2vf_bulletin_mapping.vaddr; ret = bnx2x_vf_get_resources(sc, sc->max_tx_queues, sc->max_rx_queues); if (ret) goto out; } return 0; out: bnx2x_periodic_stop(eth_dev); return ret; } static int eth_bnx2x_dev_init(struct rte_eth_dev *eth_dev) { struct bnx2x_softc *sc = eth_dev->data->dev_private; PMD_INIT_FUNC_TRACE(sc); return bnx2x_common_dev_init(eth_dev, 0); } static int eth_bnx2xvf_dev_init(struct rte_eth_dev *eth_dev) { struct bnx2x_softc *sc = eth_dev->data->dev_private; PMD_INIT_FUNC_TRACE(sc); return bnx2x_common_dev_init(eth_dev, 1); } static struct rte_pci_driver rte_bnx2x_pmd; static struct rte_pci_driver rte_bnx2xvf_pmd; static int eth_bnx2x_pci_probe(struct rte_pci_driver *pci_drv, struct rte_pci_device *pci_dev) { if (pci_drv == &rte_bnx2x_pmd) return rte_eth_dev_pci_generic_probe(pci_dev, sizeof(struct bnx2x_softc), eth_bnx2x_dev_init); else if (pci_drv == &rte_bnx2xvf_pmd) return rte_eth_dev_pci_generic_probe(pci_dev, sizeof(struct bnx2x_softc), eth_bnx2xvf_dev_init); else return -EINVAL; } static int eth_bnx2x_pci_remove(struct rte_pci_device *pci_dev) { return rte_eth_dev_pci_generic_remove(pci_dev, NULL); } static struct rte_pci_driver rte_bnx2x_pmd = { .id_table = pci_id_bnx2x_map, .drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC, .probe = eth_bnx2x_pci_probe, .remove = eth_bnx2x_pci_remove, }; /* * virtual function driver struct */ static struct rte_pci_driver rte_bnx2xvf_pmd = { .id_table = pci_id_bnx2xvf_map, .drv_flags = RTE_PCI_DRV_NEED_MAPPING, .probe = eth_bnx2x_pci_probe, .remove = eth_bnx2x_pci_remove, }; RTE_PMD_REGISTER_PCI(net_bnx2x, rte_bnx2x_pmd); RTE_PMD_REGISTER_PCI_TABLE(net_bnx2x, pci_id_bnx2x_map); RTE_PMD_REGISTER_KMOD_DEP(net_bnx2x, "* igb_uio | uio_pci_generic | vfio-pci"); RTE_PMD_REGISTER_PCI(net_bnx2xvf, rte_bnx2xvf_pmd); RTE_PMD_REGISTER_PCI_TABLE(net_bnx2xvf, pci_id_bnx2xvf_map); RTE_PMD_REGISTER_KMOD_DEP(net_bnx2xvf, "* igb_uio | vfio-pci"); RTE_INIT(bnx2x_init_log) { bnx2x_logtype_init = rte_log_register("pmd.net.bnx2x.init"); if (bnx2x_logtype_init >= 0) rte_log_set_level(bnx2x_logtype_init, RTE_LOG_NOTICE); bnx2x_logtype_driver = rte_log_register("pmd.net.bnx2x.driver"); if (bnx2x_logtype_driver >= 0) rte_log_set_level(bnx2x_logtype_driver, RTE_LOG_NOTICE); }