/* SPDX-License-Identifier: BSD-3-Clause * Copyright(c) 2010-2017 Intel Corporation */ #include #include #include #include #include #include #include #include "rte_eth_bond.h" #include "rte_eth_bond_private.h" #include "rte_eth_bond_8023ad_private.h" int check_for_bonded_ethdev(const struct rte_eth_dev *eth_dev) { /* Check valid pointer */ if (eth_dev == NULL || eth_dev->device == NULL || eth_dev->device->driver == NULL || eth_dev->device->driver->name == NULL) return -1; /* return 0 if driver name matches */ return eth_dev->device->driver->name != pmd_bond_drv.driver.name; } int valid_bonded_port_id(uint16_t port_id) { RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -1); return check_for_bonded_ethdev(&rte_eth_devices[port_id]); } int check_for_master_bonded_ethdev(const struct rte_eth_dev *eth_dev) { int i; struct bond_dev_private *internals; if (check_for_bonded_ethdev(eth_dev) != 0) return 0; internals = eth_dev->data->dev_private; /* Check if any of slave devices is a bonded device */ for (i = 0; i < internals->slave_count; i++) if (valid_bonded_port_id(internals->slaves[i].port_id) == 0) return 1; return 0; } int valid_slave_port_id(uint16_t port_id, uint8_t mode) { RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -1); /* Verify that port_id refers to a non bonded port */ if (check_for_bonded_ethdev(&rte_eth_devices[port_id]) == 0 && mode == BONDING_MODE_8023AD) { RTE_BOND_LOG(ERR, "Cannot add slave to bonded device in 802.3ad" " mode as slave is also a bonded device, only " "physical devices can be support in this mode."); return -1; } return 0; } void activate_slave(struct rte_eth_dev *eth_dev, uint16_t port_id) { struct bond_dev_private *internals = eth_dev->data->dev_private; uint16_t active_count = internals->active_slave_count; if (internals->mode == BONDING_MODE_8023AD) bond_mode_8023ad_activate_slave(eth_dev, port_id); if (internals->mode == BONDING_MODE_TLB || internals->mode == BONDING_MODE_ALB) { internals->tlb_slaves_order[active_count] = port_id; } RTE_ASSERT(internals->active_slave_count < (RTE_DIM(internals->active_slaves) - 1)); internals->active_slaves[internals->active_slave_count] = port_id; internals->active_slave_count++; if (internals->mode == BONDING_MODE_TLB) bond_tlb_activate_slave(internals); if (internals->mode == BONDING_MODE_ALB) bond_mode_alb_client_list_upd(eth_dev); } void deactivate_slave(struct rte_eth_dev *eth_dev, uint16_t port_id) { uint16_t slave_pos; struct bond_dev_private *internals = eth_dev->data->dev_private; uint16_t active_count = internals->active_slave_count; if (internals->mode == BONDING_MODE_8023AD) { bond_mode_8023ad_stop(eth_dev); bond_mode_8023ad_deactivate_slave(eth_dev, port_id); } else if (internals->mode == BONDING_MODE_TLB || internals->mode == BONDING_MODE_ALB) bond_tlb_disable(internals); slave_pos = find_slave_by_id(internals->active_slaves, active_count, port_id); /* If slave was not at the end of the list * shift active slaves up active array list */ if (slave_pos < active_count) { active_count--; memmove(internals->active_slaves + slave_pos, internals->active_slaves + slave_pos + 1, (active_count - slave_pos) * sizeof(internals->active_slaves[0])); } RTE_ASSERT(active_count < RTE_DIM(internals->active_slaves)); internals->active_slave_count = active_count; /* Resetting active_slave when reaches to max * no of slaves in active list */ if (internals->active_slave >= active_count) internals->active_slave = 0; if (eth_dev->data->dev_started) { if (internals->mode == BONDING_MODE_8023AD) { bond_mode_8023ad_start(eth_dev); } else if (internals->mode == BONDING_MODE_TLB) { bond_tlb_enable(internals); } else if (internals->mode == BONDING_MODE_ALB) { bond_tlb_enable(internals); bond_mode_alb_client_list_upd(eth_dev); } } } int rte_eth_bond_create(const char *name, uint8_t mode, uint8_t socket_id) { struct bond_dev_private *internals; char devargs[52]; uint16_t port_id; int ret; if (name == NULL) { RTE_BOND_LOG(ERR, "Invalid name specified"); return -EINVAL; } ret = snprintf(devargs, sizeof(devargs), "driver=net_bonding,mode=%d,socket_id=%d", mode, socket_id); if (ret < 0 || ret >= (int)sizeof(devargs)) return -ENOMEM; ret = rte_vdev_init(name, devargs); if (ret) return -ENOMEM; ret = rte_eth_dev_get_port_by_name(name, &port_id); RTE_ASSERT(!ret); /* * To make bond_ethdev_configure() happy we need to free the * internals->kvlist here. * * Also see comment in bond_ethdev_configure(). */ internals = rte_eth_devices[port_id].data->dev_private; rte_kvargs_free(internals->kvlist); internals->kvlist = NULL; return port_id; } int rte_eth_bond_free(const char *name) { return rte_vdev_uninit(name); } static int slave_vlan_filter_set(uint16_t bonded_port_id, uint16_t slave_port_id) { struct rte_eth_dev *bonded_eth_dev; struct bond_dev_private *internals; int found; int res = 0; uint64_t slab = 0; uint32_t pos = 0; uint16_t first; bonded_eth_dev = &rte_eth_devices[bonded_port_id]; if ((bonded_eth_dev->data->dev_conf.rxmode.offloads & DEV_RX_OFFLOAD_VLAN_FILTER) == 0) return 0; internals = bonded_eth_dev->data->dev_private; found = rte_bitmap_scan(internals->vlan_filter_bmp, &pos, &slab); first = pos; if (!found) return 0; do { uint32_t i; uint64_t mask; for (i = 0, mask = 1; i < RTE_BITMAP_SLAB_BIT_SIZE; i ++, mask <<= 1) { if (unlikely(slab & mask)) { uint16_t vlan_id = pos + i; res = rte_eth_dev_vlan_filter(slave_port_id, vlan_id, 1); } } found = rte_bitmap_scan(internals->vlan_filter_bmp, &pos, &slab); } while (found && first != pos && res == 0); return res; } static int slave_rte_flow_prepare(uint16_t slave_id, struct bond_dev_private *internals) { struct rte_flow *flow; struct rte_flow_error ferror; uint16_t slave_port_id = internals->slaves[slave_id].port_id; if (internals->flow_isolated_valid != 0) { rte_eth_dev_stop(slave_port_id); if (rte_flow_isolate(slave_port_id, internals->flow_isolated, &ferror)) { RTE_BOND_LOG(ERR, "rte_flow_isolate failed for slave" " %d: %s", slave_id, ferror.message ? ferror.message : "(no stated reason)"); return -1; } } TAILQ_FOREACH(flow, &internals->flow_list, next) { flow->flows[slave_id] = rte_flow_create(slave_port_id, flow->rule.attr, flow->rule.pattern, flow->rule.actions, &ferror); if (flow->flows[slave_id] == NULL) { RTE_BOND_LOG(ERR, "Cannot create flow for slave" " %d: %s", slave_id, ferror.message ? ferror.message : "(no stated reason)"); /* Destroy successful bond flows from the slave */ TAILQ_FOREACH(flow, &internals->flow_list, next) { if (flow->flows[slave_id] != NULL) { rte_flow_destroy(slave_port_id, flow->flows[slave_id], &ferror); flow->flows[slave_id] = NULL; } } return -1; } } return 0; } static void eth_bond_slave_inherit_dev_info_rx_first(struct bond_dev_private *internals, const struct rte_eth_dev_info *di) { struct rte_eth_rxconf *rxconf_i = &internals->default_rxconf; internals->reta_size = di->reta_size; /* Inherit Rx offload capabilities from the first slave device */ internals->rx_offload_capa = di->rx_offload_capa; internals->rx_queue_offload_capa = di->rx_queue_offload_capa; internals->flow_type_rss_offloads = di->flow_type_rss_offloads; /* Inherit maximum Rx packet size from the first slave device */ internals->candidate_max_rx_pktlen = di->max_rx_pktlen; /* Inherit default Rx queue settings from the first slave device */ memcpy(rxconf_i, &di->default_rxconf, sizeof(*rxconf_i)); /* * Turn off descriptor prefetch and writeback by default for all * slave devices. Applications may tweak this setting if need be. */ rxconf_i->rx_thresh.pthresh = 0; rxconf_i->rx_thresh.hthresh = 0; rxconf_i->rx_thresh.wthresh = 0; /* Setting this to zero should effectively enable default values */ rxconf_i->rx_free_thresh = 0; /* Disable deferred start by default for all slave devices */ rxconf_i->rx_deferred_start = 0; } static void eth_bond_slave_inherit_dev_info_tx_first(struct bond_dev_private *internals, const struct rte_eth_dev_info *di) { struct rte_eth_txconf *txconf_i = &internals->default_txconf; /* Inherit Tx offload capabilities from the first slave device */ internals->tx_offload_capa = di->tx_offload_capa; internals->tx_queue_offload_capa = di->tx_queue_offload_capa; /* Inherit default Tx queue settings from the first slave device */ memcpy(txconf_i, &di->default_txconf, sizeof(*txconf_i)); /* * Turn off descriptor prefetch and writeback by default for all * slave devices. Applications may tweak this setting if need be. */ txconf_i->tx_thresh.pthresh = 0; txconf_i->tx_thresh.hthresh = 0; txconf_i->tx_thresh.wthresh = 0; /* * Setting these parameters to zero assumes that default * values will be configured implicitly by slave devices. */ txconf_i->tx_free_thresh = 0; txconf_i->tx_rs_thresh = 0; /* Disable deferred start by default for all slave devices */ txconf_i->tx_deferred_start = 0; } static void eth_bond_slave_inherit_dev_info_rx_next(struct bond_dev_private *internals, const struct rte_eth_dev_info *di) { struct rte_eth_rxconf *rxconf_i = &internals->default_rxconf; const struct rte_eth_rxconf *rxconf = &di->default_rxconf; internals->rx_offload_capa &= di->rx_offload_capa; internals->rx_queue_offload_capa &= di->rx_queue_offload_capa; internals->flow_type_rss_offloads &= di->flow_type_rss_offloads; /* * If at least one slave device suggests enabling this * setting by default, enable it for all slave devices * since disabling it may not be necessarily supported. */ if (rxconf->rx_drop_en == 1) rxconf_i->rx_drop_en = 1; /* * Adding a new slave device may cause some of previously inherited * offloads to be withdrawn from the internal rx_queue_offload_capa * value. Thus, the new internal value of default Rx queue offloads * has to be masked by rx_queue_offload_capa to make sure that only * commonly supported offloads are preserved from both the previous * value and the value being inhereted from the new slave device. */ rxconf_i->offloads = (rxconf_i->offloads | rxconf->offloads) & internals->rx_queue_offload_capa; /* * RETA size is GCD of all slaves RETA sizes, so, if all sizes will be * the power of 2, the lower one is GCD */ if (internals->reta_size > di->reta_size) internals->reta_size = di->reta_size; if (!internals->max_rx_pktlen && di->max_rx_pktlen < internals->candidate_max_rx_pktlen) internals->candidate_max_rx_pktlen = di->max_rx_pktlen; } static void eth_bond_slave_inherit_dev_info_tx_next(struct bond_dev_private *internals, const struct rte_eth_dev_info *di) { struct rte_eth_txconf *txconf_i = &internals->default_txconf; const struct rte_eth_txconf *txconf = &di->default_txconf; internals->tx_offload_capa &= di->tx_offload_capa; internals->tx_queue_offload_capa &= di->tx_queue_offload_capa; /* * Adding a new slave device may cause some of previously inherited * offloads to be withdrawn from the internal tx_queue_offload_capa * value. Thus, the new internal value of default Tx queue offloads * has to be masked by tx_queue_offload_capa to make sure that only * commonly supported offloads are preserved from both the previous * value and the value being inhereted from the new slave device. */ txconf_i->offloads = (txconf_i->offloads | txconf->offloads) & internals->tx_queue_offload_capa; } static void eth_bond_slave_inherit_desc_lim_first(struct rte_eth_desc_lim *bond_desc_lim, const struct rte_eth_desc_lim *slave_desc_lim) { memcpy(bond_desc_lim, slave_desc_lim, sizeof(*bond_desc_lim)); } static int eth_bond_slave_inherit_desc_lim_next(struct rte_eth_desc_lim *bond_desc_lim, const struct rte_eth_desc_lim *slave_desc_lim) { bond_desc_lim->nb_max = RTE_MIN(bond_desc_lim->nb_max, slave_desc_lim->nb_max); bond_desc_lim->nb_min = RTE_MAX(bond_desc_lim->nb_min, slave_desc_lim->nb_min); bond_desc_lim->nb_align = RTE_MAX(bond_desc_lim->nb_align, slave_desc_lim->nb_align); if (bond_desc_lim->nb_min > bond_desc_lim->nb_max || bond_desc_lim->nb_align > bond_desc_lim->nb_max) { RTE_BOND_LOG(ERR, "Failed to inherit descriptor limits"); return -EINVAL; } /* Treat maximum number of segments equal to 0 as unspecified */ if (slave_desc_lim->nb_seg_max != 0 && (bond_desc_lim->nb_seg_max == 0 || slave_desc_lim->nb_seg_max < bond_desc_lim->nb_seg_max)) bond_desc_lim->nb_seg_max = slave_desc_lim->nb_seg_max; if (slave_desc_lim->nb_mtu_seg_max != 0 && (bond_desc_lim->nb_mtu_seg_max == 0 || slave_desc_lim->nb_mtu_seg_max < bond_desc_lim->nb_mtu_seg_max)) bond_desc_lim->nb_mtu_seg_max = slave_desc_lim->nb_mtu_seg_max; return 0; } static int __eth_bond_slave_add_lock_free(uint16_t bonded_port_id, uint16_t slave_port_id) { struct rte_eth_dev *bonded_eth_dev, *slave_eth_dev; struct bond_dev_private *internals; struct rte_eth_link link_props; struct rte_eth_dev_info dev_info; bonded_eth_dev = &rte_eth_devices[bonded_port_id]; internals = bonded_eth_dev->data->dev_private; if (valid_slave_port_id(slave_port_id, internals->mode) != 0) return -1; slave_eth_dev = &rte_eth_devices[slave_port_id]; if (slave_eth_dev->data->dev_flags & RTE_ETH_DEV_BONDED_SLAVE) { RTE_BOND_LOG(ERR, "Slave device is already a slave of a bonded device"); return -1; } rte_eth_dev_info_get(slave_port_id, &dev_info); if (dev_info.max_rx_pktlen < internals->max_rx_pktlen) { RTE_BOND_LOG(ERR, "Slave (port %u) max_rx_pktlen too small", slave_port_id); return -1; } slave_add(internals, slave_eth_dev); /* We need to store slaves reta_size to be able to synchronize RETA for all * slave devices even if its sizes are different. */ internals->slaves[internals->slave_count].reta_size = dev_info.reta_size; if (internals->slave_count < 1) { /* if MAC is not user defined then use MAC of first slave add to * bonded device */ if (!internals->user_defined_mac) { if (mac_address_set(bonded_eth_dev, slave_eth_dev->data->mac_addrs)) { RTE_BOND_LOG(ERR, "Failed to set MAC address"); return -1; } } /* Make primary slave */ internals->primary_port = slave_port_id; internals->current_primary_port = slave_port_id; /* Inherit queues settings from first slave */ internals->nb_rx_queues = slave_eth_dev->data->nb_rx_queues; internals->nb_tx_queues = slave_eth_dev->data->nb_tx_queues; eth_bond_slave_inherit_dev_info_rx_first(internals, &dev_info); eth_bond_slave_inherit_dev_info_tx_first(internals, &dev_info); eth_bond_slave_inherit_desc_lim_first(&internals->rx_desc_lim, &dev_info.rx_desc_lim); eth_bond_slave_inherit_desc_lim_first(&internals->tx_desc_lim, &dev_info.tx_desc_lim); } else { int ret; eth_bond_slave_inherit_dev_info_rx_next(internals, &dev_info); eth_bond_slave_inherit_dev_info_tx_next(internals, &dev_info); ret = eth_bond_slave_inherit_desc_lim_next( &internals->rx_desc_lim, &dev_info.rx_desc_lim); if (ret != 0) return ret; ret = eth_bond_slave_inherit_desc_lim_next( &internals->tx_desc_lim, &dev_info.tx_desc_lim); if (ret != 0) return ret; } bonded_eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf &= internals->flow_type_rss_offloads; if (slave_rte_flow_prepare(internals->slave_count, internals) != 0) { RTE_BOND_LOG(ERR, "Failed to prepare new slave flows: port=%d", slave_port_id); return -1; } /* Add additional MAC addresses to the slave */ if (slave_add_mac_addresses(bonded_eth_dev, slave_port_id) != 0) { RTE_BOND_LOG(ERR, "Failed to add mac address(es) to slave %hu", slave_port_id); return -1; } internals->slave_count++; if (bonded_eth_dev->data->dev_started) { if (slave_configure(bonded_eth_dev, slave_eth_dev) != 0) { internals->slave_count--; RTE_BOND_LOG(ERR, "rte_bond_slaves_configure: port=%d", slave_port_id); return -1; } } /* Add slave details to bonded device */ slave_eth_dev->data->dev_flags |= RTE_ETH_DEV_BONDED_SLAVE; /* Update all slave devices MACs */ mac_address_slaves_update(bonded_eth_dev); /* Register link status change callback with bonded device pointer as * argument*/ rte_eth_dev_callback_register(slave_port_id, RTE_ETH_EVENT_INTR_LSC, bond_ethdev_lsc_event_callback, &bonded_eth_dev->data->port_id); /* If bonded device is started then we can add the slave to our active * slave array */ if (bonded_eth_dev->data->dev_started) { rte_eth_link_get_nowait(slave_port_id, &link_props); if (link_props.link_status == ETH_LINK_UP) { if (internals->active_slave_count == 0 && !internals->user_defined_primary_port) bond_ethdev_primary_set(internals, slave_port_id); } } slave_vlan_filter_set(bonded_port_id, slave_port_id); return 0; } int rte_eth_bond_slave_add(uint16_t bonded_port_id, uint16_t slave_port_id) { struct rte_eth_dev *bonded_eth_dev; struct bond_dev_private *internals; int retval; /* Verify that port id's are valid bonded and slave ports */ if (valid_bonded_port_id(bonded_port_id) != 0) return -1; bonded_eth_dev = &rte_eth_devices[bonded_port_id]; internals = bonded_eth_dev->data->dev_private; rte_spinlock_lock(&internals->lock); retval = __eth_bond_slave_add_lock_free(bonded_port_id, slave_port_id); rte_spinlock_unlock(&internals->lock); return retval; } static int __eth_bond_slave_remove_lock_free(uint16_t bonded_port_id, uint16_t slave_port_id) { struct rte_eth_dev *bonded_eth_dev; struct bond_dev_private *internals; struct rte_eth_dev *slave_eth_dev; struct rte_flow_error flow_error; struct rte_flow *flow; int i, slave_idx; bonded_eth_dev = &rte_eth_devices[bonded_port_id]; internals = bonded_eth_dev->data->dev_private; if (valid_slave_port_id(slave_port_id, internals->mode) < 0) return -1; /* first remove from active slave list */ slave_idx = find_slave_by_id(internals->active_slaves, internals->active_slave_count, slave_port_id); if (slave_idx < internals->active_slave_count) deactivate_slave(bonded_eth_dev, slave_port_id); slave_idx = -1; /* now find in slave list */ for (i = 0; i < internals->slave_count; i++) if (internals->slaves[i].port_id == slave_port_id) { slave_idx = i; break; } if (slave_idx < 0) { RTE_BOND_LOG(ERR, "Couldn't find slave in port list, slave count %d", internals->slave_count); return -1; } /* Un-register link status change callback with bonded device pointer as * argument*/ rte_eth_dev_callback_unregister(slave_port_id, RTE_ETH_EVENT_INTR_LSC, bond_ethdev_lsc_event_callback, &rte_eth_devices[bonded_port_id].data->port_id); /* Restore original MAC address of slave device */ rte_eth_dev_default_mac_addr_set(slave_port_id, &(internals->slaves[slave_idx].persisted_mac_addr)); /* remove additional MAC addresses from the slave */ slave_remove_mac_addresses(bonded_eth_dev, slave_port_id); /* * Remove bond device flows from slave device. * Note: don't restore flow isolate mode. */ TAILQ_FOREACH(flow, &internals->flow_list, next) { if (flow->flows[slave_idx] != NULL) { rte_flow_destroy(slave_port_id, flow->flows[slave_idx], &flow_error); flow->flows[slave_idx] = NULL; } } slave_eth_dev = &rte_eth_devices[slave_port_id]; slave_remove(internals, slave_eth_dev); slave_eth_dev->data->dev_flags &= (~RTE_ETH_DEV_BONDED_SLAVE); /* first slave in the active list will be the primary by default, * otherwise use first device in list */ if (internals->current_primary_port == slave_port_id) { if (internals->active_slave_count > 0) internals->current_primary_port = internals->active_slaves[0]; else if (internals->slave_count > 0) internals->current_primary_port = internals->slaves[0].port_id; else internals->primary_port = 0; } if (internals->active_slave_count < 1) { /* if no slaves are any longer attached to bonded device and MAC is not * user defined then clear MAC of bonded device as it will be reset * when a new slave is added */ if (internals->slave_count < 1 && !internals->user_defined_mac) memset(rte_eth_devices[bonded_port_id].data->mac_addrs, 0, sizeof(*(rte_eth_devices[bonded_port_id].data->mac_addrs))); } if (internals->slave_count == 0) { internals->rx_offload_capa = 0; internals->tx_offload_capa = 0; internals->rx_queue_offload_capa = 0; internals->tx_queue_offload_capa = 0; internals->flow_type_rss_offloads = ETH_RSS_PROTO_MASK; internals->reta_size = 0; internals->candidate_max_rx_pktlen = 0; internals->max_rx_pktlen = 0; } return 0; } int rte_eth_bond_slave_remove(uint16_t bonded_port_id, uint16_t slave_port_id) { struct rte_eth_dev *bonded_eth_dev; struct bond_dev_private *internals; int retval; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; bonded_eth_dev = &rte_eth_devices[bonded_port_id]; internals = bonded_eth_dev->data->dev_private; rte_spinlock_lock(&internals->lock); retval = __eth_bond_slave_remove_lock_free(bonded_port_id, slave_port_id); rte_spinlock_unlock(&internals->lock); return retval; } int rte_eth_bond_mode_set(uint16_t bonded_port_id, uint8_t mode) { struct rte_eth_dev *bonded_eth_dev; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; bonded_eth_dev = &rte_eth_devices[bonded_port_id]; if (check_for_master_bonded_ethdev(bonded_eth_dev) != 0 && mode == BONDING_MODE_8023AD) return -1; return bond_ethdev_mode_set(bonded_eth_dev, mode); } int rte_eth_bond_mode_get(uint16_t bonded_port_id) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; return internals->mode; } int rte_eth_bond_primary_set(uint16_t bonded_port_id, uint16_t slave_port_id) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; if (valid_slave_port_id(slave_port_id, internals->mode) != 0) return -1; internals->user_defined_primary_port = 1; internals->primary_port = slave_port_id; bond_ethdev_primary_set(internals, slave_port_id); return 0; } int rte_eth_bond_primary_get(uint16_t bonded_port_id) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; if (internals->slave_count < 1) return -1; return internals->current_primary_port; } int rte_eth_bond_slaves_get(uint16_t bonded_port_id, uint16_t slaves[], uint16_t len) { struct bond_dev_private *internals; uint16_t i; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; if (slaves == NULL) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; if (internals->slave_count > len) return -1; for (i = 0; i < internals->slave_count; i++) slaves[i] = internals->slaves[i].port_id; return internals->slave_count; } int rte_eth_bond_active_slaves_get(uint16_t bonded_port_id, uint16_t slaves[], uint16_t len) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; if (slaves == NULL) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; if (internals->active_slave_count > len) return -1; memcpy(slaves, internals->active_slaves, internals->active_slave_count * sizeof(internals->active_slaves[0])); return internals->active_slave_count; } int rte_eth_bond_mac_address_set(uint16_t bonded_port_id, struct ether_addr *mac_addr) { struct rte_eth_dev *bonded_eth_dev; struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; bonded_eth_dev = &rte_eth_devices[bonded_port_id]; internals = bonded_eth_dev->data->dev_private; /* Set MAC Address of Bonded Device */ if (mac_address_set(bonded_eth_dev, mac_addr)) return -1; internals->user_defined_mac = 1; /* Update all slave devices MACs*/ if (internals->slave_count > 0) return mac_address_slaves_update(bonded_eth_dev); return 0; } int rte_eth_bond_mac_address_reset(uint16_t bonded_port_id) { struct rte_eth_dev *bonded_eth_dev; struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; bonded_eth_dev = &rte_eth_devices[bonded_port_id]; internals = bonded_eth_dev->data->dev_private; internals->user_defined_mac = 0; if (internals->slave_count > 0) { int slave_port; /* Get the primary slave location based on the primary port * number as, while slave_add(), we will keep the primary * slave based on slave_count,but not based on the primary port. */ for (slave_port = 0; slave_port < internals->slave_count; slave_port++) { if (internals->slaves[slave_port].port_id == internals->primary_port) break; } /* Set MAC Address of Bonded Device */ if (mac_address_set(bonded_eth_dev, &internals->slaves[slave_port].persisted_mac_addr) != 0) { RTE_BOND_LOG(ERR, "Failed to set MAC address on bonded device"); return -1; } /* Update all slave devices MAC addresses */ return mac_address_slaves_update(bonded_eth_dev); } /* No need to update anything as no slaves present */ return 0; } int rte_eth_bond_xmit_policy_set(uint16_t bonded_port_id, uint8_t policy) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; switch (policy) { case BALANCE_XMIT_POLICY_LAYER2: internals->balance_xmit_policy = policy; internals->burst_xmit_hash = burst_xmit_l2_hash; break; case BALANCE_XMIT_POLICY_LAYER23: internals->balance_xmit_policy = policy; internals->burst_xmit_hash = burst_xmit_l23_hash; break; case BALANCE_XMIT_POLICY_LAYER34: internals->balance_xmit_policy = policy; internals->burst_xmit_hash = burst_xmit_l34_hash; break; default: return -1; } return 0; } int rte_eth_bond_xmit_policy_get(uint16_t bonded_port_id) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; return internals->balance_xmit_policy; } int rte_eth_bond_link_monitoring_set(uint16_t bonded_port_id, uint32_t internal_ms) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; internals->link_status_polling_interval_ms = internal_ms; return 0; } int rte_eth_bond_link_monitoring_get(uint16_t bonded_port_id) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; return internals->link_status_polling_interval_ms; } int rte_eth_bond_link_down_prop_delay_set(uint16_t bonded_port_id, uint32_t delay_ms) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; internals->link_down_delay_ms = delay_ms; return 0; } int rte_eth_bond_link_down_prop_delay_get(uint16_t bonded_port_id) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; return internals->link_down_delay_ms; } int rte_eth_bond_link_up_prop_delay_set(uint16_t bonded_port_id, uint32_t delay_ms) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; internals->link_up_delay_ms = delay_ms; return 0; } int rte_eth_bond_link_up_prop_delay_get(uint16_t bonded_port_id) { struct bond_dev_private *internals; if (valid_bonded_port_id(bonded_port_id) != 0) return -1; internals = rte_eth_devices[bonded_port_id].data->dev_private; return internals->link_up_delay_ms; }