diff options
Diffstat (limited to 'src/vnet/interface_api.c')
-rw-r--r-- | src/vnet/interface_api.c | 958 |
1 files changed, 958 insertions, 0 deletions
diff --git a/src/vnet/interface_api.c b/src/vnet/interface_api.c new file mode 100644 index 00000000..6374a880 --- /dev/null +++ b/src/vnet/interface_api.c @@ -0,0 +1,958 @@ +/* + *------------------------------------------------------------------ + * interface_api.c - vnet interface api + * + * Copyright (c) 2016 Cisco and/or its affiliates. + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at: + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + *------------------------------------------------------------------ + */ + +#include <vnet/vnet.h> +#include <vlibmemory/api.h> + +#include <vnet/interface.h> +#include <vnet/api_errno.h> +#include <vnet/ethernet/ethernet.h> +#include <vnet/ip/ip.h> +#include <vnet/fib/fib_table.h> +#include <vnet/mfib/mfib_table.h> +#include <vnet/l2/l2_vtr.h> +#include <vnet/vnet_msg_enum.h> +#include <vnet/fib/fib_api.h> +#include <vnet/mfib/mfib_table.h> + +#define vl_typedefs /* define message structures */ +#include <vnet/vnet_all_api_h.h> +#undef vl_typedefs + +#define vl_endianfun /* define message structures */ +#include <vnet/vnet_all_api_h.h> +#undef vl_endianfun + +/* instantiate all the print functions we know about */ +#define vl_print(handle, ...) vlib_cli_output (handle, __VA_ARGS__) +#define vl_printfun +#include <vnet/vnet_all_api_h.h> +#undef vl_printfun + +#include <vlibapi/api_helper_macros.h> +vpe_api_main_t vpe_api_main; + +#define foreach_vpe_api_msg \ +_(SW_INTERFACE_SET_FLAGS, sw_interface_set_flags) \ +_(SW_INTERFACE_SET_MTU, sw_interface_set_mtu) \ +_(WANT_INTERFACE_EVENTS, want_interface_events) \ +_(SW_INTERFACE_DUMP, sw_interface_dump) \ +_(SW_INTERFACE_ADD_DEL_ADDRESS, sw_interface_add_del_address) \ +_(SW_INTERFACE_SET_TABLE, sw_interface_set_table) \ +_(SW_INTERFACE_GET_TABLE, sw_interface_get_table) \ +_(SW_INTERFACE_SET_UNNUMBERED, sw_interface_set_unnumbered) \ +_(SW_INTERFACE_CLEAR_STATS, sw_interface_clear_stats) \ +_(SW_INTERFACE_TAG_ADD_DEL, sw_interface_tag_add_del) \ +_(SW_INTERFACE_SET_MAC_ADDRESS, sw_interface_set_mac_address) + +static void +vl_api_sw_interface_set_flags_t_handler (vl_api_sw_interface_set_flags_t * mp) +{ + vl_api_sw_interface_set_flags_reply_t *rmp; + vnet_main_t *vnm = vnet_get_main (); + int rv = 0; + clib_error_t *error; + u16 flags; + + VALIDATE_SW_IF_INDEX (mp); + + flags = mp->admin_up_down ? VNET_SW_INTERFACE_FLAG_ADMIN_UP : 0; + + error = vnet_sw_interface_set_flags (vnm, ntohl (mp->sw_if_index), flags); + if (error) + { + rv = -1; + clib_error_report (error); + } + + BAD_SW_IF_INDEX_LABEL; + REPLY_MACRO (VL_API_SW_INTERFACE_SET_FLAGS_REPLY); +} + +static void +vl_api_sw_interface_set_mtu_t_handler (vl_api_sw_interface_set_mtu_t * mp) +{ + vl_api_sw_interface_set_mtu_reply_t *rmp; + vnet_main_t *vnm = vnet_get_main (); + u32 flags = ETHERNET_INTERFACE_FLAG_MTU; + u32 sw_if_index = ntohl (mp->sw_if_index); + u16 mtu = ntohs (mp->mtu); + ethernet_main_t *em = ðernet_main; + int rv = 0; + + VALIDATE_SW_IF_INDEX (mp); + + vnet_sw_interface_t *si = vnet_get_sw_interface (vnm, sw_if_index); + if (si->type != VNET_SW_INTERFACE_TYPE_HARDWARE) + { + rv = VNET_API_ERROR_INVALID_VALUE; + goto bad_sw_if_index; + } + + vnet_hw_interface_t *hi = vnet_get_hw_interface (vnm, si->hw_if_index); + ethernet_interface_t *eif = ethernet_get_interface (em, si->hw_if_index); + + if (!eif) + { + rv = VNET_API_ERROR_FEATURE_DISABLED; + goto bad_sw_if_index; + } + + if (mtu < hi->min_supported_packet_bytes) + { + rv = VNET_API_ERROR_INVALID_VALUE; + goto bad_sw_if_index; + } + + if (mtu > hi->max_supported_packet_bytes) + { + rv = VNET_API_ERROR_INVALID_VALUE; + goto bad_sw_if_index; + } + + if (hi->max_packet_bytes != mtu) + { + hi->max_packet_bytes = mtu; + ethernet_set_flags (vnm, si->hw_if_index, flags); + } + + BAD_SW_IF_INDEX_LABEL; + REPLY_MACRO (VL_API_SW_INTERFACE_SET_MTU_REPLY); +} + +static void +send_sw_interface_details (vpe_api_main_t * am, + unix_shared_memory_queue_t * q, + vnet_sw_interface_t * swif, + u8 * interface_name, u32 context) +{ + vnet_hw_interface_t *hi = + vnet_get_sup_hw_interface (am->vnet_main, swif->sw_if_index); + + vl_api_sw_interface_details_t *mp = vl_msg_api_alloc (sizeof (*mp)); + memset (mp, 0, sizeof (*mp)); + mp->_vl_msg_id = ntohs (VL_API_SW_INTERFACE_DETAILS); + mp->sw_if_index = ntohl (swif->sw_if_index); + mp->sup_sw_if_index = ntohl (swif->sup_sw_if_index); + mp->admin_up_down = (swif->flags & VNET_SW_INTERFACE_FLAG_ADMIN_UP) ? 1 : 0; + mp->link_up_down = (hi->flags & VNET_HW_INTERFACE_FLAG_LINK_UP) ? 1 : 0; + mp->link_duplex = ((hi->flags & VNET_HW_INTERFACE_FLAG_DUPLEX_MASK) >> + VNET_HW_INTERFACE_FLAG_DUPLEX_SHIFT); + mp->link_speed = ((hi->flags & VNET_HW_INTERFACE_FLAG_SPEED_MASK) >> + VNET_HW_INTERFACE_FLAG_SPEED_SHIFT); + mp->link_mtu = ntohs (hi->max_packet_bytes); + mp->context = context; + + strncpy ((char *) mp->interface_name, + (char *) interface_name, ARRAY_LEN (mp->interface_name) - 1); + + /* Send the L2 address for ethernet physical intfcs */ + if (swif->sup_sw_if_index == swif->sw_if_index + && hi->hw_class_index == ethernet_hw_interface_class.index) + { + ethernet_main_t *em = ethernet_get_main (am->vlib_main); + ethernet_interface_t *ei; + + ei = pool_elt_at_index (em->interfaces, hi->hw_instance); + ASSERT (sizeof (mp->l2_address) >= sizeof (ei->address)); + clib_memcpy (mp->l2_address, ei->address, sizeof (ei->address)); + mp->l2_address_length = ntohl (sizeof (ei->address)); + } + else if (swif->sup_sw_if_index != swif->sw_if_index) + { + vnet_sub_interface_t *sub = &swif->sub; + mp->sub_id = ntohl (sub->id); + mp->sub_dot1ad = sub->eth.flags.dot1ad; + mp->sub_number_of_tags = + sub->eth.flags.one_tag + sub->eth.flags.two_tags * 2; + mp->sub_outer_vlan_id = ntohs (sub->eth.outer_vlan_id); + mp->sub_inner_vlan_id = ntohs (sub->eth.inner_vlan_id); + mp->sub_exact_match = sub->eth.flags.exact_match; + mp->sub_default = sub->eth.flags.default_sub; + mp->sub_outer_vlan_id_any = sub->eth.flags.outer_vlan_id_any; + mp->sub_inner_vlan_id_any = sub->eth.flags.inner_vlan_id_any; + + /* vlan tag rewrite data */ + u32 vtr_op = L2_VTR_DISABLED; + u32 vtr_push_dot1q = 0, vtr_tag1 = 0, vtr_tag2 = 0; + + if (l2vtr_get (am->vlib_main, am->vnet_main, swif->sw_if_index, + &vtr_op, &vtr_push_dot1q, &vtr_tag1, &vtr_tag2) != 0) + { + // error - default to disabled + mp->vtr_op = ntohl (L2_VTR_DISABLED); + clib_warning ("cannot get vlan tag rewrite for sw_if_index %d", + swif->sw_if_index); + } + else + { + mp->vtr_op = ntohl (vtr_op); + mp->vtr_push_dot1q = ntohl (vtr_push_dot1q); + mp->vtr_tag1 = ntohl (vtr_tag1); + mp->vtr_tag2 = ntohl (vtr_tag2); + } + } + + /* pbb tag rewrite data */ + ethernet_header_t eth_hdr; + u32 vtr_op = L2_VTR_DISABLED; + u16 outer_tag = 0; + u16 b_vlanid = 0; + u32 i_sid = 0; + memset (ð_hdr, 0, sizeof (eth_hdr)); + + if (!l2pbb_get (am->vlib_main, am->vnet_main, swif->sw_if_index, + &vtr_op, &outer_tag, ð_hdr, &b_vlanid, &i_sid)) + { + mp->sub_dot1ah = 1; + clib_memcpy (mp->b_dmac, eth_hdr.dst_address, + sizeof (eth_hdr.dst_address)); + clib_memcpy (mp->b_smac, eth_hdr.src_address, + sizeof (eth_hdr.src_address)); + mp->b_vlanid = b_vlanid; + mp->i_sid = i_sid; + } + + u8 *tag = vnet_get_sw_interface_tag (vnet_get_main (), swif->sw_if_index); + if (tag) + strncpy ((char *) mp->tag, (char *) tag, ARRAY_LEN (mp->tag) - 1); + + vl_msg_api_send_shmem (q, (u8 *) & mp); +} + +static void +vl_api_sw_interface_dump_t_handler (vl_api_sw_interface_dump_t * mp) +{ + vpe_api_main_t *am = &vpe_api_main; + vnet_sw_interface_t *swif; + vnet_interface_main_t *im = &am->vnet_main->interface_main; + + unix_shared_memory_queue_t *q = + vl_api_client_index_to_input_queue (mp->client_index); + if (q == 0) + return; + + u8 *filter = 0, *name = 0; + if (mp->name_filter_valid) + { + mp->name_filter[ARRAY_LEN (mp->name_filter) - 1] = 0; + filter = format (0, "%s%c", mp->name_filter, 0); + } + + char *strcasestr (char *, char *); /* lnx hdr file botch */ + /* *INDENT-OFF* */ + pool_foreach (swif, im->sw_interfaces, + ({ + if (!vnet_swif_is_api_visible (swif)) + continue; + vec_reset_length(name); + name = format (name, "%U%c", format_vnet_sw_interface_name, am->vnet_main, + swif, 0); + + if (filter && !strcasestr((char *) name, (char *) filter)) + continue; + + send_sw_interface_details (am, q, swif, name, mp->context); + })); + /* *INDENT-ON* */ + + vec_free (name); + vec_free (filter); +} + +static void + vl_api_sw_interface_add_del_address_t_handler + (vl_api_sw_interface_add_del_address_t * mp) +{ + vlib_main_t *vm = vlib_get_main (); + vnet_main_t *vnm = vnet_get_main (); + vl_api_sw_interface_add_del_address_reply_t *rmp; + int rv = 0; + u32 is_del; + clib_error_t *error = 0; + + VALIDATE_SW_IF_INDEX (mp); + + is_del = mp->is_add == 0; + vnm->api_errno = 0; + + if (mp->del_all) + ip_del_all_interface_addresses (vm, ntohl (mp->sw_if_index)); + else if (mp->is_ipv6) + error = ip6_add_del_interface_address (vm, ntohl (mp->sw_if_index), + (void *) mp->address, + mp->address_length, is_del); + else + error = ip4_add_del_interface_address (vm, ntohl (mp->sw_if_index), + (void *) mp->address, + mp->address_length, is_del); + + if (error) + { + rv = vnm->api_errno; + clib_error_report (error); + goto done; + } + + BAD_SW_IF_INDEX_LABEL; + +done: + REPLY_MACRO (VL_API_SW_INTERFACE_ADD_DEL_ADDRESS_REPLY); +} + +void stats_dslock_with_hint (int hint, int tag) __attribute__ ((weak)); +void +stats_dslock_with_hint (int hint, int tag) +{ +} + +void stats_dsunlock (void) __attribute__ ((weak)); +void +stats_dsunlock (void) +{ +} + +static void +vl_api_sw_interface_set_table_t_handler (vl_api_sw_interface_set_table_t * mp) +{ + vl_api_sw_interface_set_table_reply_t *rmp; + u32 sw_if_index = ntohl (mp->sw_if_index); + u32 table_id = ntohl (mp->vrf_id); + int rv = 0; + + VALIDATE_SW_IF_INDEX (mp); + + stats_dslock_with_hint (1 /* release hint */ , 4 /* tag */ ); + + if (mp->is_ipv6) + rv = ip_table_bind (FIB_PROTOCOL_IP6, sw_if_index, table_id, 1); + else + rv = ip_table_bind (FIB_PROTOCOL_IP4, sw_if_index, table_id, 1); + + stats_dsunlock (); + + BAD_SW_IF_INDEX_LABEL; + + REPLY_MACRO (VL_API_SW_INTERFACE_SET_TABLE_REPLY); +} + +int +ip_table_bind (fib_protocol_t fproto, + u32 sw_if_index, u32 table_id, u8 is_api) +{ + CLIB_UNUSED (ip_interface_address_t * ia); + u32 fib_index, mfib_index; + fib_source_t src; + mfib_source_t msrc; + + if (is_api) + { + src = FIB_SOURCE_API; + msrc = MFIB_SOURCE_API; + } + else + { + src = FIB_SOURCE_CLI; + msrc = MFIB_SOURCE_CLI; + } + + /* + * This is temporary whilst I do the song and dance with the CSIT version + */ + if (0 != table_id) + { + fib_index = fib_table_find_or_create_and_lock (fproto, table_id, src); + mfib_index = + mfib_table_find_or_create_and_lock (fproto, table_id, msrc); + } + else + { + fib_index = 0; + mfib_index = 0; + } + + /* + * This if table does not exist = error is what we want in the end. + */ + /* fib_index = fib_table_find (fproto, table_id); */ + /* mfib_index = mfib_table_find (fproto, table_id); */ + + /* if (~0 == fib_index || ~0 == mfib_index) */ + /* { */ + /* return (VNET_API_ERROR_NO_SUCH_FIB); */ + /* } */ + + if (FIB_PROTOCOL_IP6 == fproto) + { + /* + * If the interface already has in IP address, then a change int + * VRF is not allowed. The IP address applied must first be removed. + * We do not do that automatically here, since VPP has no knowledge + * of whether thoses subnets are valid in the destination VRF. + */ + /* *INDENT-OFF* */ + foreach_ip_interface_address (&ip6_main.lookup_main, + ia, sw_if_index, + 1 /* honor unnumbered */ , + ({ + return (VNET_API_ERROR_ADDRESS_FOUND_FOR_INTERFACE); + })); + /* *INDENT-ON* */ + + vec_validate (ip6_main.fib_index_by_sw_if_index, sw_if_index); + vec_validate (ip6_main.mfib_index_by_sw_if_index, sw_if_index); + + /* + * tell those that are interested that the binding is changing. + */ + ip6_table_bind_callback_t *cb; + vec_foreach (cb, ip6_main.table_bind_callbacks) + cb->function (&ip6_main, cb->function_opaque, + sw_if_index, + fib_index, + ip6_main.fib_index_by_sw_if_index[sw_if_index]); + + if (0 == table_id) + { + /* reset back to default */ + if (0 != ip6_main.fib_index_by_sw_if_index[sw_if_index]) + fib_table_unlock (ip6_main.fib_index_by_sw_if_index[sw_if_index], + FIB_PROTOCOL_IP6, src); + if (0 != ip6_main.mfib_index_by_sw_if_index[sw_if_index]) + mfib_table_unlock (ip6_main.mfib_index_by_sw_if_index + [sw_if_index], FIB_PROTOCOL_IP6, msrc); + + } + else + { + /* we need to lock the table now it's inuse */ + fib_table_lock (fib_index, FIB_PROTOCOL_IP6, src); + mfib_table_lock (mfib_index, FIB_PROTOCOL_IP6, msrc); + } + + ip6_main.fib_index_by_sw_if_index[sw_if_index] = fib_index; + ip6_main.mfib_index_by_sw_if_index[sw_if_index] = mfib_index; + } + else + { + /* + * If the interface already has in IP address, then a change int + * VRF is not allowed. The IP address applied must first be removed. + * We do not do that automatically here, since VPP has no knowledge + * of whether thoses subnets are valid in the destination VRF. + */ + /* *INDENT-OFF* */ + foreach_ip_interface_address (&ip4_main.lookup_main, + ia, sw_if_index, + 1 /* honor unnumbered */ , + ({ + return (VNET_API_ERROR_ADDRESS_FOUND_FOR_INTERFACE); + })); + /* *INDENT-ON* */ + + vec_validate (ip4_main.fib_index_by_sw_if_index, sw_if_index); + vec_validate (ip4_main.mfib_index_by_sw_if_index, sw_if_index); + + /* + * tell those that are interested that the binding is changing. + */ + ip4_table_bind_callback_t *cb; + vec_foreach (cb, ip4_main.table_bind_callbacks) + cb->function (&ip4_main, cb->function_opaque, + sw_if_index, + fib_index, + ip4_main.fib_index_by_sw_if_index[sw_if_index]); + + if (0 == table_id) + { + /* reset back to default */ + if (0 != ip4_main.fib_index_by_sw_if_index[sw_if_index]) + fib_table_unlock (ip4_main.fib_index_by_sw_if_index[sw_if_index], + FIB_PROTOCOL_IP4, src); + if (0 != ip4_main.mfib_index_by_sw_if_index[sw_if_index]) + mfib_table_unlock (ip4_main.mfib_index_by_sw_if_index + [sw_if_index], FIB_PROTOCOL_IP4, msrc); + + } + else + { + /* we need to lock the table now it's inuse */ + fib_index = fib_table_find_or_create_and_lock (FIB_PROTOCOL_IP4, + table_id, src); + + mfib_index = mfib_table_find_or_create_and_lock (FIB_PROTOCOL_IP4, + table_id, msrc); + } + + ip4_main.fib_index_by_sw_if_index[sw_if_index] = fib_index; + ip4_main.mfib_index_by_sw_if_index[sw_if_index] = mfib_index; + } + + /* + * Temporary. undo the locks from the find and create at the staart + */ + if (0 != table_id) + { + fib_table_unlock (fib_index, fproto, src); + mfib_table_unlock (mfib_index, fproto, msrc); + } + + return (0); +} + +static void +send_sw_interface_get_table_reply (unix_shared_memory_queue_t * q, + u32 context, int retval, u32 vrf_id) +{ + vl_api_sw_interface_get_table_reply_t *mp; + + mp = vl_msg_api_alloc (sizeof (*mp)); + memset (mp, 0, sizeof (*mp)); + mp->_vl_msg_id = ntohs (VL_API_SW_INTERFACE_GET_TABLE_REPLY); + mp->context = context; + mp->retval = htonl (retval); + mp->vrf_id = htonl (vrf_id); + + vl_msg_api_send_shmem (q, (u8 *) & mp); +} + +static void +vl_api_sw_interface_get_table_t_handler (vl_api_sw_interface_get_table_t * mp) +{ + unix_shared_memory_queue_t *q; + fib_table_t *fib_table = 0; + u32 sw_if_index = ~0; + u32 fib_index = ~0; + u32 table_id = ~0; + fib_protocol_t fib_proto = FIB_PROTOCOL_IP4; + int rv = 0; + + q = vl_api_client_index_to_input_queue (mp->client_index); + if (q == 0) + return; + + VALIDATE_SW_IF_INDEX (mp); + + sw_if_index = ntohl (mp->sw_if_index); + + if (mp->is_ipv6) + fib_proto = FIB_PROTOCOL_IP6; + + fib_index = fib_table_get_index_for_sw_if_index (fib_proto, sw_if_index); + if (fib_index != ~0) + { + fib_table = fib_table_get (fib_index, fib_proto); + table_id = fib_table->ft_table_id; + } + + BAD_SW_IF_INDEX_LABEL; + + send_sw_interface_get_table_reply (q, mp->context, rv, table_id); +} + +static void vl_api_sw_interface_set_unnumbered_t_handler + (vl_api_sw_interface_set_unnumbered_t * mp) +{ + vl_api_sw_interface_set_unnumbered_reply_t *rmp; + int rv = 0; + vnet_main_t *vnm = vnet_get_main (); + u32 sw_if_index = ntohl (mp->sw_if_index); + u32 unnumbered_sw_if_index = ntohl (mp->unnumbered_sw_if_index); + u32 was_unnum; + + /* + * The API message field names are backwards from + * the underlying data structure names. + * It's not worth changing them now. + */ + if (!vnet_sw_interface_is_api_valid (vnm, unnumbered_sw_if_index)) + { + rv = VNET_API_ERROR_INVALID_SW_IF_INDEX; + goto done; + } + + /* Only check the "use loop0" field when setting the binding */ + if (mp->is_add && !vnet_sw_interface_is_api_valid (vnm, sw_if_index)) + { + rv = VNET_API_ERROR_INVALID_SW_IF_INDEX_2; + goto done; + } + + vnet_sw_interface_t *si = + vnet_get_sw_interface (vnm, unnumbered_sw_if_index); + was_unnum = (si->flags & VNET_SW_INTERFACE_FLAG_UNNUMBERED); + + if (mp->is_add) + { + si->flags |= VNET_SW_INTERFACE_FLAG_UNNUMBERED; + si->unnumbered_sw_if_index = sw_if_index; + + ip4_main.lookup_main.if_address_pool_index_by_sw_if_index + [unnumbered_sw_if_index] = + ip4_main. + lookup_main.if_address_pool_index_by_sw_if_index[sw_if_index]; + ip6_main. + lookup_main.if_address_pool_index_by_sw_if_index + [unnumbered_sw_if_index] = + ip6_main. + lookup_main.if_address_pool_index_by_sw_if_index[sw_if_index]; + } + else + { + si->flags &= ~(VNET_SW_INTERFACE_FLAG_UNNUMBERED); + si->unnumbered_sw_if_index = (u32) ~ 0; + + ip4_main.lookup_main.if_address_pool_index_by_sw_if_index + [unnumbered_sw_if_index] = ~0; + ip6_main.lookup_main.if_address_pool_index_by_sw_if_index + [unnumbered_sw_if_index] = ~0; + } + + if (was_unnum != (si->flags & VNET_SW_INTERFACE_FLAG_UNNUMBERED)) + { + ip4_sw_interface_enable_disable (unnumbered_sw_if_index, mp->is_add); + ip6_sw_interface_enable_disable (unnumbered_sw_if_index, mp->is_add); + } + +done: + REPLY_MACRO (VL_API_SW_INTERFACE_SET_UNNUMBERED_REPLY); +} + +static void +vl_api_sw_interface_clear_stats_t_handler (vl_api_sw_interface_clear_stats_t * + mp) +{ + vl_api_sw_interface_clear_stats_reply_t *rmp; + + vnet_main_t *vnm = vnet_get_main (); + vnet_interface_main_t *im = &vnm->interface_main; + vlib_simple_counter_main_t *sm; + vlib_combined_counter_main_t *cm; + static vnet_main_t **my_vnet_mains; + int i, j, n_counters; + int rv = 0; + + if (mp->sw_if_index != ~0) + VALIDATE_SW_IF_INDEX (mp); + + vec_reset_length (my_vnet_mains); + + for (i = 0; i < vec_len (vnet_mains); i++) + { + if (vnet_mains[i]) + vec_add1 (my_vnet_mains, vnet_mains[i]); + } + + if (vec_len (vnet_mains) == 0) + vec_add1 (my_vnet_mains, vnm); + + n_counters = vec_len (im->combined_sw_if_counters); + + for (j = 0; j < n_counters; j++) + { + for (i = 0; i < vec_len (my_vnet_mains); i++) + { + im = &my_vnet_mains[i]->interface_main; + cm = im->combined_sw_if_counters + j; + if (mp->sw_if_index == (u32) ~ 0) + vlib_clear_combined_counters (cm); + else + vlib_zero_combined_counter (cm, ntohl (mp->sw_if_index)); + } + } + + n_counters = vec_len (im->sw_if_counters); + + for (j = 0; j < n_counters; j++) + { + for (i = 0; i < vec_len (my_vnet_mains); i++) + { + im = &my_vnet_mains[i]->interface_main; + sm = im->sw_if_counters + j; + if (mp->sw_if_index == (u32) ~ 0) + vlib_clear_simple_counters (sm); + else + vlib_zero_simple_counter (sm, ntohl (mp->sw_if_index)); + } + } + + BAD_SW_IF_INDEX_LABEL; + + REPLY_MACRO (VL_API_SW_INTERFACE_CLEAR_STATS_REPLY); +} + +#define API_LINK_STATE_EVENT 1 +#define API_ADMIN_UP_DOWN_EVENT 2 + +static int +event_data_cmp (void *a1, void *a2) +{ + uword *e1 = a1; + uword *e2 = a2; + + return (word) e1[0] - (word) e2[0]; +} + +static void +send_sw_interface_event (vpe_api_main_t * am, + vpe_client_registration_t * reg, + unix_shared_memory_queue_t * q, + vnet_sw_interface_t * swif) +{ + vl_api_sw_interface_event_t *mp; + vnet_main_t *vnm = am->vnet_main; + + vnet_hw_interface_t *hi = vnet_get_sup_hw_interface (vnm, + swif->sw_if_index); + mp = vl_msg_api_alloc (sizeof (*mp)); + memset (mp, 0, sizeof (*mp)); + mp->_vl_msg_id = ntohs (VL_API_SW_INTERFACE_EVENT); + mp->sw_if_index = ntohl (swif->sw_if_index); + mp->client_index = reg->client_index; + mp->pid = reg->client_pid; + + mp->admin_up_down = (swif->flags & VNET_SW_INTERFACE_FLAG_ADMIN_UP) ? 1 : 0; + mp->link_up_down = (hi->flags & VNET_HW_INTERFACE_FLAG_LINK_UP) ? 1 : 0; + vl_msg_api_send_shmem (q, (u8 *) & mp); +} + +static uword +link_state_process (vlib_main_t * vm, + vlib_node_runtime_t * rt, vlib_frame_t * f) +{ + vpe_api_main_t *vam = &vpe_api_main; + vnet_main_t *vnm = vam->vnet_main; + vnet_sw_interface_t *swif; + uword *event_data = 0; + vpe_client_registration_t *reg; + int i; + u32 prev_sw_if_index; + unix_shared_memory_queue_t *q; + + vam->link_state_process_up = 1; + + while (1) + { + vlib_process_wait_for_event (vm); + + /* Unified list of changed link or admin state sw_if_indices */ + vlib_process_get_events_with_type + (vm, &event_data, API_LINK_STATE_EVENT); + vlib_process_get_events_with_type + (vm, &event_data, API_ADMIN_UP_DOWN_EVENT); + + /* Sort, so we can eliminate duplicates */ + vec_sort_with_function (event_data, event_data_cmp); + + prev_sw_if_index = ~0; + + for (i = 0; i < vec_len (event_data); i++) + { + /* Only one message per swif */ + if (prev_sw_if_index == event_data[i]) + continue; + prev_sw_if_index = event_data[i]; + + /* *INDENT-OFF* */ + pool_foreach(reg, vam->interface_events_registrations, + ({ + q = vl_api_client_index_to_input_queue (reg->client_index); + if (q) + { + /* sw_interface may be deleted already */ + if (!pool_is_free_index (vnm->interface_main.sw_interfaces, + event_data[i])) + { + swif = vnet_get_sw_interface (vnm, event_data[i]); + send_sw_interface_event (vam, reg, q, swif); + } + } + })); + /* *INDENT-ON* */ + } + vec_reset_length (event_data); + } + + return 0; +} + +static clib_error_t *link_up_down_function (vnet_main_t * vm, u32 hw_if_index, + u32 flags); +static clib_error_t *admin_up_down_function (vnet_main_t * vm, + u32 hw_if_index, u32 flags); + +/* *INDENT-OFF* */ +VLIB_REGISTER_NODE (link_state_process_node,static) = { + .function = link_state_process, + .type = VLIB_NODE_TYPE_PROCESS, + .name = "vpe-link-state-process", +}; +/* *INDENT-ON* */ + +VNET_SW_INTERFACE_ADMIN_UP_DOWN_FUNCTION (admin_up_down_function); +VNET_HW_INTERFACE_LINK_UP_DOWN_FUNCTION (link_up_down_function); + +static clib_error_t * +link_up_down_function (vnet_main_t * vm, u32 hw_if_index, u32 flags) +{ + vpe_api_main_t *vam = &vpe_api_main; + vnet_hw_interface_t *hi = vnet_get_hw_interface (vm, hw_if_index); + + if (vam->link_state_process_up) + vlib_process_signal_event (vam->vlib_main, + link_state_process_node.index, + API_LINK_STATE_EVENT, hi->sw_if_index); + return 0; +} + +static clib_error_t * +admin_up_down_function (vnet_main_t * vm, u32 sw_if_index, u32 flags) +{ + vpe_api_main_t *vam = &vpe_api_main; + + /* + * Note: it's perfectly fair to set a subif admin up / admin down. + * Note the subtle distinction between this routine and the previous + * routine. + */ + if (vam->link_state_process_up) + vlib_process_signal_event (vam->vlib_main, + link_state_process_node.index, + API_ADMIN_UP_DOWN_EVENT, sw_if_index); + return 0; +} + +static void vl_api_sw_interface_tag_add_del_t_handler + (vl_api_sw_interface_tag_add_del_t * mp) +{ + vnet_main_t *vnm = vnet_get_main (); + vl_api_sw_interface_tag_add_del_reply_t *rmp; + int rv = 0; + u8 *tag; + u32 sw_if_index = ntohl (mp->sw_if_index); + + VALIDATE_SW_IF_INDEX (mp); + + if (mp->is_add) + { + if (mp->tag[0] == 0) + { + rv = VNET_API_ERROR_INVALID_VALUE; + goto out; + } + + mp->tag[ARRAY_LEN (mp->tag) - 1] = 0; + tag = format (0, "%s%c", mp->tag, 0); + vnet_set_sw_interface_tag (vnm, tag, sw_if_index); + } + else + vnet_clear_sw_interface_tag (vnm, sw_if_index); + + BAD_SW_IF_INDEX_LABEL; +out: + REPLY_MACRO (VL_API_SW_INTERFACE_TAG_ADD_DEL_REPLY); +} + +static void vl_api_sw_interface_set_mac_address_t_handler + (vl_api_sw_interface_set_mac_address_t * mp) +{ + vl_api_sw_interface_set_mac_address_reply_t *rmp; + vnet_main_t *vnm = vnet_get_main (); + u32 sw_if_index = ntohl (mp->sw_if_index); + vnet_sw_interface_t *si; + u64 mac; + clib_error_t *error; + int rv = 0; + + VALIDATE_SW_IF_INDEX (mp); + + mac = ((u64) mp->mac_address[0] << (8 * 0) + | (u64) mp->mac_address[1] << (8 * 1) + | (u64) mp->mac_address[2] << (8 * 2) + | (u64) mp->mac_address[3] << (8 * 3) + | (u64) mp->mac_address[4] << (8 * 4) + | (u64) mp->mac_address[5] << (8 * 5)); + + si = vnet_get_sw_interface (vnm, sw_if_index); + error = vnet_hw_interface_change_mac_address (vnm, si->hw_if_index, mac); + if (error) + { + rv = VNET_API_ERROR_UNIMPLEMENTED; + clib_error_report (error); + goto out; + } + + BAD_SW_IF_INDEX_LABEL; +out: + REPLY_MACRO (VL_API_SW_INTERFACE_SET_MAC_ADDRESS_REPLY); +} + +/* + * vpe_api_hookup + * Add vpe's API message handlers to the table. + * vlib has alread mapped shared memory and + * added the client registration handlers. + * See .../vlib-api/vlibmemory/memclnt_vlib.c:memclnt_process() + */ +#define vl_msg_name_crc_list +#include <vnet/interface.api.h> +#undef vl_msg_name_crc_list + +static void +setup_message_id_table (api_main_t * am) +{ +#define _(id,n,crc) vl_msg_api_add_msg_name_crc (am, #n "_" #crc, id); + foreach_vl_msg_name_crc_interface; +#undef _ +} + +pub_sub_handler (interface_events, INTERFACE_EVENTS); + +static clib_error_t * +interface_api_hookup (vlib_main_t * vm) +{ + api_main_t *am = &api_main; + +#define _(N,n) \ + vl_msg_api_set_handlers(VL_API_##N, #n, \ + vl_api_##n##_t_handler, \ + vl_noop_handler, \ + vl_api_##n##_t_endian, \ + vl_api_##n##_t_print, \ + sizeof(vl_api_##n##_t), 1); + foreach_vpe_api_msg; +#undef _ + + /* + * Set up the (msg_name, crc, message-id) table + */ + setup_message_id_table (am); + + return 0; +} + +VLIB_API_INIT_FUNCTION (interface_api_hookup); + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ |