/* * Copyright (c) 2016 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include #include #include #include #include #include #include #include /* * per-source type vft */ static fib_entry_src_vft_t fib_entry_src_bh_vft[FIB_SOURCE_BH_MAX]; /** * Get the VFT for a given source. This is a combination of the source * enum and the interposer flags */ const fib_entry_src_vft_t* fib_entry_src_get_vft (const fib_entry_src_t *esrc) { fib_source_behaviour_t bh; bh = fib_source_get_behaviour(esrc->fes_src); if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_INTERPOSE) { return (&fib_entry_src_bh_vft[FIB_SOURCE_BH_INTERPOSE]); } ASSERT(bh < FIB_SOURCE_BH_MAX); return (&fib_entry_src_bh_vft[bh]); } static void fib_entry_src_copy_default (const fib_entry_src_t *orig_src, const fib_entry_t *fib_entry, fib_entry_src_t *copy_src) { clib_memcpy(©_src->u, &orig_src->u, sizeof(copy_src->u)); } void fib_entry_src_behaviour_register (fib_source_behaviour_t bh, const fib_entry_src_vft_t *vft) { fib_entry_src_bh_vft[bh] = *vft; if (NULL == fib_entry_src_bh_vft[bh].fesv_copy) { fib_entry_src_bh_vft[bh].fesv_copy = fib_entry_src_copy_default; } } static int fib_entry_src_cmp_for_sort (void * v1, void * v2) { fib_entry_src_t *esrc1 = v1, *esrc2 = v2; return (fib_source_get_prio(esrc1->fes_src) - fib_source_get_prio(esrc2->fes_src)); } static void fib_entry_src_action_init (fib_entry_t *fib_entry, fib_source_t source, fib_entry_flag_t flags) { fib_entry_src_t esrc = { .fes_pl = FIB_NODE_INDEX_INVALID, .fes_flags = FIB_ENTRY_SRC_FLAG_NONE, .fes_src = source, .fes_entry_flags = flags, }; FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, &esrc, fesv_init, (&esrc)); vec_add1(fib_entry->fe_srcs, esrc); vec_sort_with_function(fib_entry->fe_srcs, fib_entry_src_cmp_for_sort); } static fib_entry_src_t * fib_entry_src_find_i (const fib_entry_t *fib_entry, fib_source_t source, u32 *index) { fib_entry_src_t *esrc; int ii; ii = 0; vec_foreach(esrc, fib_entry->fe_srcs) { if (esrc->fes_src == source) { if (NULL != index) { *index = ii; } return (esrc); } else { ii++; } } return (NULL); } fib_entry_src_t * fib_entry_src_find (const fib_entry_t *fib_entry, fib_source_t source) { return (fib_entry_src_find_i(fib_entry, source, NULL)); } int fib_entry_is_sourced (fib_node_index_t fib_entry_index, fib_source_t source) { fib_entry_t *fib_entry; fib_entry = fib_entry_get(fib_entry_index); return (NULL != fib_entry_src_find(fib_entry, source)); } int fib_entry_is_marked (fib_node_index_t fib_entry_index, fib_source_t source) { fib_entry_t *fib_entry; fib_entry_src_t *esrc; fib_entry = fib_entry_get(fib_entry_index); esrc = fib_entry_src_find(fib_entry, source); if (NULL == esrc) { return (0); } else { return (!!(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_STALE)); } } void fib_entry_mark (fib_node_index_t fib_entry_index, fib_source_t source) { fib_entry_t *fib_entry; fib_entry_src_t *esrc; fib_entry = fib_entry_get(fib_entry_index); esrc = fib_entry_src_find(fib_entry, source); if (NULL != esrc) { esrc->fes_flags |= FIB_ENTRY_SRC_FLAG_STALE; } } static fib_entry_src_t * fib_entry_src_find_or_create (fib_entry_t *fib_entry, fib_source_t source, fib_entry_flag_t flags) { fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); if (NULL == esrc) { fib_entry_src_action_init(fib_entry, source, flags); } return (fib_entry_src_find(fib_entry, source)); } static void fib_entry_src_action_deinit (fib_entry_t *fib_entry, fib_source_t source) { fib_entry_src_t *esrc; u32 index = ~0; esrc = fib_entry_src_find_i(fib_entry, source, &index); ASSERT(NULL != esrc); FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_deinit, (esrc)); fib_path_ext_list_flush(&esrc->fes_path_exts); vec_del1(fib_entry->fe_srcs, index); vec_sort_with_function(fib_entry->fe_srcs, fib_entry_src_cmp_for_sort); } fib_entry_src_cover_res_t fib_entry_src_action_cover_change (fib_entry_t *fib_entry, fib_entry_src_t *esrc) { FIB_ENTRY_SRC_VFT_INVOKE_AND_RETURN(esrc, fesv_cover_change, (esrc, fib_entry)); fib_entry_src_cover_res_t res = { .install = !0, .bw_reason = FIB_NODE_BW_REASON_FLAG_NONE, }; return (res); } fib_entry_src_cover_res_t fib_entry_src_action_cover_update (fib_entry_t *fib_entry, fib_entry_src_t *esrc) { FIB_ENTRY_SRC_VFT_INVOKE_AND_RETURN(esrc, fesv_cover_update, (esrc, fib_entry)); fib_entry_src_cover_res_t res = { .install = !0, .bw_reason = FIB_NODE_BW_REASON_FLAG_NONE, }; return (res); } typedef struct fib_entry_src_collect_forwarding_ctx_t_ { load_balance_path_t *next_hops; const fib_entry_t *fib_entry; i32 start_source_index, end_source_index; fib_forward_chain_type_t fct; int n_recursive_constrained; u16 preference; dpo_proto_t payload_proto; } fib_entry_src_collect_forwarding_ctx_t; /** * @brief Determine whether this FIB entry should use a load-balance MAP * to support PIC edge fast convergence */ static load_balance_flags_t fib_entry_calc_lb_flags (fib_entry_src_collect_forwarding_ctx_t *ctx, const fib_entry_src_t *esrc) { /** * We'll use a LB map if the path-list has multiple recursive paths. * recursive paths implies BGP, and hence scale. */ if (ctx->n_recursive_constrained > 1 && fib_path_list_is_popular(esrc->fes_pl)) { return (LOAD_BALANCE_FLAG_USES_MAP); } return (LOAD_BALANCE_FLAG_NONE); } static int fib_entry_src_valid_out_label (mpls_label_t label) { return ((MPLS_LABEL_IS_REAL(label) || MPLS_LABEL_POP == label || MPLS_IETF_IPV4_EXPLICIT_NULL_LABEL == label || MPLS_IETF_IPV6_EXPLICIT_NULL_LABEL == label || MPLS_IETF_IMPLICIT_NULL_LABEL == label)); } static dpo_proto_t fib_prefix_get_payload_proto (const fib_prefix_t *pfx) { switch (pfx->fp_proto) { case FIB_PROTOCOL_IP4: return (DPO_PROTO_IP4); case FIB_PROTOCOL_IP6: return (DPO_PROTO_IP6); case FIB_PROTOCOL_MPLS: return (pfx->fp_payload_proto); } ASSERT(0); return (DPO_PROTO_IP4); } static void fib_entry_src_get_path_forwarding (fib_node_index_t path_index, fib_entry_src_collect_forwarding_ctx_t *ctx) { load_balance_path_t *nh; /* * no extension => no out-going label for this path. that's OK * in the case of an IP or EOS chain, but not for non-EOS */ switch (ctx->fct) { case FIB_FORW_CHAIN_TYPE_UNICAST_IP4: case FIB_FORW_CHAIN_TYPE_UNICAST_IP6: case FIB_FORW_CHAIN_TYPE_MCAST_IP4: case FIB_FORW_CHAIN_TYPE_MCAST_IP6: case FIB_FORW_CHAIN_TYPE_BIER: /* * EOS traffic with no label to stack, we need the IP Adj */ vec_add2(ctx->next_hops, nh, 1); nh->path_index = path_index; nh->path_weight = fib_path_get_weight(path_index); fib_path_contribute_forwarding(path_index, ctx->fct, ctx->payload_proto, &nh->path_dpo); break; case FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS: if (fib_path_is_exclusive(path_index) || fib_path_is_deag(path_index)) { vec_add2(ctx->next_hops, nh, 1); nh->path_index = path_index; nh->path_weight = fib_path_get_weight(path_index); fib_path_contribute_forwarding(path_index, FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS, ctx->payload_proto, &nh->path_dpo); } break; case FIB_FORW_CHAIN_TYPE_MPLS_EOS: { /* * no label. we need a chain based on the payload. fixup. */ vec_add2(ctx->next_hops, nh, 1); nh->path_index = path_index; nh->path_weight = fib_path_get_weight(path_index); fib_path_contribute_forwarding(path_index, ctx->fct, ctx->payload_proto, &nh->path_dpo); fib_path_stack_mpls_disp(path_index, ctx->payload_proto, FIB_MPLS_LSP_MODE_PIPE, &nh->path_dpo); break; } case FIB_FORW_CHAIN_TYPE_ETHERNET: case FIB_FORW_CHAIN_TYPE_NSH: ASSERT(0); break; } } static fib_path_list_walk_rc_t fib_entry_src_collect_forwarding (fib_node_index_t pl_index, fib_node_index_t path_index, void *arg) { fib_entry_src_collect_forwarding_ctx_t *ctx; const fib_entry_src_t *esrc; fib_path_ext_t *path_ext; u32 n_nhs; ctx = arg; n_nhs = vec_len(ctx->next_hops); /* * walk the paths and extension of the best non-interpose source */ esrc = &ctx->fib_entry->fe_srcs[ctx->end_source_index]; /* * if the path is not resolved, don't include it. */ if (!fib_path_is_resolved(path_index)) { return (FIB_PATH_LIST_WALK_CONTINUE); } if (fib_path_is_recursive_constrained(path_index)) { ctx->n_recursive_constrained += 1; } if (0xffff == ctx->preference) { /* * not set a preference yet, so the first path we encounter * sets the preference we are collecting. */ ctx->preference = fib_path_get_preference(path_index); } else if (ctx->preference != fib_path_get_preference(path_index)) { /* * this path does not belong to the same preference as the * previous paths encountered. we are done now. */ return (FIB_PATH_LIST_WALK_STOP); } /* * get the matching path-extension for the path being visited. */ path_ext = fib_path_ext_list_find_by_path_index(&esrc->fes_path_exts, path_index); if (NULL != path_ext) { switch (path_ext->fpe_type) { case FIB_PATH_EXT_MPLS: if (fib_entry_src_valid_out_label(path_ext->fpe_label_stack[0].fml_value)) { /* * found a matching extension. stack it to obtain the forwarding * info for this path. */ ctx->next_hops = fib_path_ext_stack(path_ext, ctx->payload_proto, ctx->fct, ctx->next_hops); } else { fib_entry_src_get_path_forwarding(path_index, ctx); } break; case FIB_PATH_EXT_ADJ: if (FIB_PATH_EXT_ADJ_FLAG_REFINES_COVER & path_ext->fpe_adj_flags) { fib_entry_src_get_path_forwarding(path_index, ctx); } /* * else * the path does not refine the cover, meaning that * the adjacency does/does not match the sub-net on the link. * So this path does not contribute forwarding. */ break; } } else { fib_entry_src_get_path_forwarding(path_index, ctx); } /* * a this point 'ctx' has the DPO the path contributed, plus * any labels from path extensions. * check if there are any interpose sources that want to contribute */ if (n_nhs < vec_len(ctx->next_hops)) { /* * the path contributed a new choice. */ const fib_entry_src_vft_t *vft; /* * roll up the sources that are interposes */ i32 index; for (index = ctx->end_source_index; index >= ctx->start_source_index; index--) { const dpo_id_t *interposer; esrc = &ctx->fib_entry->fe_srcs[index]; vft = fib_entry_src_get_vft(esrc); if (!(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_CONTRIBUTING) || !(esrc->fes_entry_flags & FIB_ENTRY_FLAG_INTERPOSE)) continue; ASSERT(vft->fesv_contribute_interpose); interposer = vft->fesv_contribute_interpose(esrc, ctx->fib_entry); if (NULL != interposer) { dpo_id_t clone = DPO_INVALID; dpo_mk_interpose(interposer, &ctx->next_hops[n_nhs].path_dpo, &clone); dpo_copy(&ctx->next_hops[n_nhs].path_dpo, &clone); dpo_reset(&clone); } } } return (FIB_PATH_LIST_WALK_CONTINUE); } void fib_entry_src_mk_lb (fib_entry_t *fib_entry, fib_source_t source, fib_forward_chain_type_t fct, dpo_id_t *dpo_lb) { const fib_entry_src_t *esrc; dpo_proto_t lb_proto; u32 start, end; /* * The source passed here is the 'best', i.e. the one the client * wants. however, if it's an interpose then it does not contribute * the forwarding, the next best source that is not an interpose does. * So roll down the sources, to find the best non-interpose */ vec_foreach_index (start, fib_entry->fe_srcs) { if (source == fib_entry->fe_srcs[start].fes_src) break; } for (end = start; end < vec_len (fib_entry->fe_srcs); end++) { if (!(fib_entry->fe_srcs[end].fes_entry_flags & FIB_ENTRY_FLAG_INTERPOSE) && (fib_entry->fe_srcs[end].fes_flags & FIB_ENTRY_SRC_FLAG_CONTRIBUTING)) break; } if (end == vec_len(fib_entry->fe_srcs)) { /* didn't find any contributing non-interpose sources */ end = start; } esrc = &fib_entry->fe_srcs[end]; /* * If the entry has path extensions then we construct a load-balance * by stacking the extensions on the forwarding chains of the paths. * Otherwise we use the load-balance of the path-list */ fib_entry_src_collect_forwarding_ctx_t ctx = { .fib_entry = fib_entry, .next_hops = NULL, .n_recursive_constrained = 0, .fct = fct, .preference = 0xffff, .start_source_index = start, .end_source_index = end, .payload_proto = fib_prefix_get_payload_proto(&fib_entry->fe_prefix), }; /* * As an optimisation we allocate the vector of next-hops to be sized * equal to the maximum number of paths we will need, which is also the * most likely number we will need, since in most cases the paths are 'up'. */ vec_validate(ctx.next_hops, fib_path_list_get_n_paths(esrc->fes_pl)); vec_reset_length(ctx.next_hops); lb_proto = fib_forw_chain_type_to_dpo_proto(fct); fib_path_list_walk(esrc->fes_pl, fib_entry_src_collect_forwarding, &ctx); if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_EXCLUSIVE) { /* * the client provided the DPO that the entry should link to. * all entries must link to a LB, so if it is an LB already * then we can use it. */ if ((1 == vec_len(ctx.next_hops)) && (DPO_LOAD_BALANCE == ctx.next_hops[0].path_dpo.dpoi_type)) { dpo_copy(dpo_lb, &ctx.next_hops[0].path_dpo); dpo_reset(&ctx.next_hops[0].path_dpo); return; } } if (!dpo_id_is_valid(dpo_lb)) { /* * first time create */ if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_MULTICAST) { dpo_set(dpo_lb, DPO_REPLICATE, lb_proto, MPLS_IS_REPLICATE | replicate_create(0, lb_proto)); } else { fib_protocol_t flow_hash_proto; flow_hash_config_t fhc; /* * if the protocol for the LB we are building does not match that * of the fib_entry (i.e. we are build the [n]EOS LB for an IPv[46] * then the fib_index is not an index that relates to the table * type we need. So get the default flow-hash config instead. */ flow_hash_proto = dpo_proto_to_fib(lb_proto); if (fib_entry->fe_prefix.fp_proto != flow_hash_proto) { fhc = fib_table_get_default_flow_hash_config(flow_hash_proto); } else { fhc = fib_table_get_flow_hash_config(fib_entry->fe_fib_index, flow_hash_proto); } dpo_set(dpo_lb, DPO_LOAD_BALANCE, lb_proto, load_balance_create(0, lb_proto, fhc)); } } if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_MULTICAST) { /* * MPLS multicast */ replicate_multipath_update(dpo_lb, ctx.next_hops); } else { load_balance_multipath_update(dpo_lb, ctx.next_hops, fib_entry_calc_lb_flags(&ctx, esrc)); vec_free(ctx.next_hops); /* * if this entry is sourced by the uRPF-exempt source then we * append the always present local0 interface (index 0) to the * uRPF list so it is not empty. that way packets pass the loose check. */ index_t ui = fib_path_list_get_urpf(esrc->fes_pl); if ((fib_entry_is_sourced(fib_entry_get_index(fib_entry), FIB_SOURCE_URPF_EXEMPT) || (esrc->fes_entry_flags & FIB_ENTRY_FLAG_LOOSE_URPF_EXEMPT))&& (0 == fib_urpf_check_size(ui))) { /* * The uRPF list we get from the path-list is shared by all * other users of the list, but the uRPF exemption applies * only to this prefix. So we need our own list. */ ui = fib_urpf_list_alloc_and_lock(); fib_urpf_list_append(ui, 0); fib_urpf_list_bake(ui); load_balance_set_urpf(dpo_lb->dpoi_index, ui); fib_urpf_list_unlock(ui); } else { load_balance_set_urpf(dpo_lb->dpoi_index, ui); } load_balance_set_fib_entry_flags(dpo_lb->dpoi_index, fib_entry_get_flags_i(fib_entry)); } } void fib_entry_src_action_install (fib_entry_t *fib_entry, fib_source_t source) { /* * Install the forwarding chain for the given source into the forwarding * tables */ fib_forward_chain_type_t fct; int insert; fct = fib_entry_get_default_chain_type(fib_entry); /* * Every entry has its own load-balance object. All changes to the entry's * forwarding result in an inplace modify of the load-balance. This means * the load-balance object only needs to be added to the forwarding * DB once, when it is created. */ insert = !dpo_id_is_valid(&fib_entry->fe_lb); fib_entry_src_mk_lb(fib_entry, source, fct, &fib_entry->fe_lb); ASSERT(dpo_id_is_valid(&fib_entry->fe_lb)); FIB_ENTRY_DBG(fib_entry, "install: %d", fib_entry->fe_lb); /* * insert the adj into the data-plane forwarding trie */ if (insert) { fib_table_fwding_dpo_update(fib_entry->fe_fib_index, &fib_entry->fe_prefix, &fib_entry->fe_lb); } /* * if any of the other chain types are already created they will need * updating too */ fib_entry_delegate_type_t fdt; fib_entry_delegate_t *fed; FOR_EACH_DELEGATE_CHAIN(fib_entry, fdt, fed, { fib_entry_src_mk_lb(fib_entry, source, fib_entry_delegate_type_to_chain_type(fdt), &fed->fd_dpo); }); } void fib_entry_src_action_uninstall (fib_entry_t *fib_entry) { /* * uninstall the forwarding chain from the forwarding tables */ FIB_ENTRY_DBG(fib_entry, "uninstall"); if (dpo_id_is_valid(&fib_entry->fe_lb)) { fib_table_fwding_dpo_remove( fib_entry->fe_fib_index, &fib_entry->fe_prefix, &fib_entry->fe_lb); vlib_worker_wait_one_loop(); dpo_reset(&fib_entry->fe_lb); } } static void fib_entry_recursive_loop_detect_i (fib_node_index_t path_list_index) { fib_node_index_t *entries = NULL; fib_path_list_recursive_loop_detect(path_list_index, &entries); vec_free(entries); } /* * fib_entry_src_action_copy * * copy a source data from another entry to this one */ static fib_entry_t * fib_entry_src_action_copy (fib_entry_t *fib_entry, const fib_entry_src_t *orig_src) { fib_entry_src_t *esrc; esrc = fib_entry_src_find_or_create(fib_entry, orig_src->fes_src, orig_src->fes_entry_flags); FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_copy, (orig_src, fib_entry, esrc)); fib_path_list_unlock(esrc->fes_pl); /* * copy over all the data ... */ esrc->fes_flags = orig_src->fes_flags; esrc->fes_pl = orig_src->fes_pl; /* * ... then update */ esrc->fes_ref_count = 1; esrc->fes_flags |= FIB_ENTRY_SRC_FLAG_INHERITED; esrc->fes_flags &= ~(FIB_ENTRY_SRC_FLAG_ACTIVE | FIB_ENTRY_SRC_FLAG_CONTRIBUTING); esrc->fes_entry_flags &= ~FIB_ENTRY_FLAG_COVERED_INHERIT; /* * the source owns a lock on the entry */ fib_path_list_lock(esrc->fes_pl); fib_entry_lock(fib_entry_get_index(fib_entry)); return (fib_entry); } /* * fib_entry_src_action_update * * copy a source data from another entry to this one */ static fib_entry_src_t * fib_entry_src_action_update_from_cover (fib_entry_t *fib_entry, const fib_entry_src_t *orig_src) { fib_entry_src_t *esrc; esrc = fib_entry_src_find_or_create(fib_entry, orig_src->fes_src, orig_src->fes_entry_flags); /* * the source owns a lock on the entry */ fib_path_list_unlock(esrc->fes_pl); esrc->fes_pl = orig_src->fes_pl; fib_path_list_lock(esrc->fes_pl); return (esrc); } static fib_table_walk_rc_t fib_entry_src_covered_inherit_add_i (fib_entry_t *fib_entry, const fib_entry_src_t *cover_src) { fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, cover_src->fes_src); if (cover_src == esrc) { return (FIB_TABLE_WALK_CONTINUE); } if (NULL != esrc) { /* * the covered entry already has this source. */ if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT) { /* * the covered source is itself a COVERED_INHERIT, i.e. * it also pushes this source down the sub-tree. * We consider this more specific covered to be the owner * of the sub-tree from this point down. */ return (FIB_TABLE_WALK_SUB_TREE_STOP); } if (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED) { /* * The covered's source data has been inherited, presumably * from this cover, i.e. this is a modify. */ fib_source_t best_source; best_source = fib_entry_get_best_source( fib_entry_get_index(fib_entry)); fib_entry_src_action_update_from_cover(fib_entry, cover_src); fib_entry_source_change(fib_entry, best_source, cover_src->fes_src); } else { /* * The covered's source was not inherited and it is also * not inheriting. Nevertheless, it still owns the sub-tree from * this point down. */ return (FIB_TABLE_WALK_SUB_TREE_STOP); } } else { /* * The covered does not have this source - add it. */ fib_source_t best_source; best_source = fib_entry_get_best_source( fib_entry_get_index(fib_entry)); fib_entry_src_action_copy(fib_entry, cover_src); fib_entry_source_change(fib_entry, best_source, cover_src->fes_src); } return (FIB_TABLE_WALK_CONTINUE); } static fib_table_walk_rc_t fib_entry_src_covered_inherit_walk_add (fib_node_index_t fei, void *ctx) { return (fib_entry_src_covered_inherit_add_i(fib_entry_get(fei), ctx)); } static fib_table_walk_rc_t fib_entry_src_covered_inherit_walk_remove (fib_node_index_t fei, void *ctx) { fib_entry_src_t *cover_src, *esrc; fib_entry_t *fib_entry; fib_entry = fib_entry_get(fei); cover_src = ctx; esrc = fib_entry_src_find(fib_entry, cover_src->fes_src); if (cover_src == esrc) { return (FIB_TABLE_WALK_CONTINUE); } if (NULL != esrc) { /* * the covered entry already has this source. */ if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT) { /* * the covered source is itself a COVERED_INHERIT, i.e. * it also pushes this source down the sub-tree. * We consider this more specific covered to be the owner * of the sub-tree from this point down. */ return (FIB_TABLE_WALK_SUB_TREE_STOP); } if (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED) { /* * The covered's source data has been inherited, presumably * from this cover */ fib_entry_src_flag_t remaining; remaining = fib_entry_special_remove(fei, cover_src->fes_src); ASSERT(FIB_ENTRY_SRC_FLAG_ADDED == remaining); } else { /* * The covered's source was not inherited and it is also * not inheriting. Nevertheless, it still owns the sub-tree from * this point down. */ return (FIB_TABLE_WALK_SUB_TREE_STOP); } } else { /* * The covered does not have this source - that's an error, * since it should have inherited, but there is nothing we can do * about it now. */ } return (FIB_TABLE_WALK_CONTINUE); } void fib_entry_src_inherit (const fib_entry_t *cover, fib_entry_t *covered) { CLIB_UNUSED(fib_source_t source); const fib_entry_src_t *src; FOR_EACH_SRC_ADDED(cover, src, source, ({ if ((src->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT) || (src->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED)) { fib_entry_src_covered_inherit_add_i(covered, src); } })) } static void fib_entry_src_covered_inherit_add (fib_entry_t *fib_entry, fib_source_t source) { fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE); if ((esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT) || (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED)) { fib_table_sub_tree_walk(fib_entry->fe_fib_index, fib_entry->fe_prefix.fp_proto, &fib_entry->fe_prefix, fib_entry_src_covered_inherit_walk_add, esrc); } } static void fib_entry_src_covered_inherit_remove (fib_entry_t *fib_entry, fib_entry_src_t *esrc) { ASSERT(!(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE)); if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT) { fib_table_sub_tree_walk(fib_entry->fe_fib_index, fib_entry->fe_prefix.fp_proto, &fib_entry->fe_prefix, fib_entry_src_covered_inherit_walk_remove, esrc); } } void fib_entry_src_action_activate (fib_entry_t *fib_entry, fib_source_t source) { int houston_we_are_go_for_install; const fib_entry_src_vft_t *vft; fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); ASSERT(!(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE)); ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ADDED); esrc->fes_flags |= (FIB_ENTRY_SRC_FLAG_ACTIVE | FIB_ENTRY_SRC_FLAG_CONTRIBUTING); vft = fib_entry_src_get_vft(esrc); if (NULL != vft->fesv_activate) { houston_we_are_go_for_install = vft->fesv_activate(esrc, fib_entry); } else { /* * the source is not providing an activate function, we'll assume * therefore it has no objection to installing the entry */ houston_we_are_go_for_install = !0; } /* * link to the path-list provided by the source, and go check * if that forms any loops in the graph. */ fib_entry->fe_parent = esrc->fes_pl; fib_entry->fe_sibling = fib_path_list_child_add(fib_entry->fe_parent, FIB_NODE_TYPE_ENTRY, fib_entry_get_index(fib_entry)); fib_entry_recursive_loop_detect_i(fib_entry->fe_parent); FIB_ENTRY_DBG(fib_entry, "activate: %d", fib_entry->fe_parent); /* * If this source should push its state to covered prefixs, do that now. */ fib_entry_src_covered_inherit_add(fib_entry, source); if (0 != houston_we_are_go_for_install) { fib_entry_src_action_install(fib_entry, source); } else { fib_entry_src_action_uninstall(fib_entry); } } void fib_entry_src_action_deactivate (fib_entry_t *fib_entry, fib_source_t source) { fib_node_index_t path_list_index; fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE); FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_deactivate, (esrc, fib_entry)); esrc->fes_flags &= ~(FIB_ENTRY_SRC_FLAG_ACTIVE | FIB_ENTRY_SRC_FLAG_CONTRIBUTING); FIB_ENTRY_DBG(fib_entry, "deactivate: %d", fib_entry->fe_parent); /* * If this source should pull its state from covered prefixs, do that now. * If this source also has the INHERITED flag set then it has a cover * that wants to push down forwarding. We only want the covereds to see * one update. */ fib_entry_src_covered_inherit_remove(fib_entry, esrc); /* * un-link from an old path-list. Check for any loops this will clear */ path_list_index = fib_entry->fe_parent; fib_entry->fe_parent = FIB_NODE_INDEX_INVALID; fib_entry_recursive_loop_detect_i(path_list_index); /* * this will unlock the path-list, so it may be invalid thereafter. */ fib_path_list_child_remove(path_list_index, fib_entry->fe_sibling); fib_entry->fe_sibling = FIB_NODE_INDEX_INVALID; } static void fib_entry_src_action_fwd_update (const fib_entry_t *fib_entry, fib_source_t source) { fib_entry_src_t *esrc; vec_foreach(esrc, fib_entry->fe_srcs) { FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_fwd_update, (esrc, fib_entry, source)); } } void fib_entry_src_action_reactivate (fib_entry_t *fib_entry, fib_source_t source) { fib_node_index_t path_list_index; const fib_entry_src_vft_t *vft; fib_entry_src_t *esrc; int remain_installed; esrc = fib_entry_src_find(fib_entry, source); ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE); FIB_ENTRY_DBG(fib_entry, "reactivate: %d to %d", fib_entry->fe_parent, esrc->fes_pl); /* * call the source to reactive and get the go/no-go to remain installed */ vft = fib_entry_src_get_vft(esrc); if (NULL != vft->fesv_reactivate) { remain_installed = vft->fesv_reactivate(esrc, fib_entry); } else { remain_installed = 1; } if (fib_entry->fe_parent != esrc->fes_pl) { /* * un-link from an old path-list. Check for any loops this will clear */ path_list_index = fib_entry->fe_parent; fib_entry->fe_parent = FIB_NODE_INDEX_INVALID; /* * temporary lock so it doesn't get deleted when this entry is no * longer a child. */ fib_path_list_lock(path_list_index); /* * this entry is no longer a child. after unlinking check if any loops * were broken */ fib_path_list_child_remove(path_list_index, fib_entry->fe_sibling); fib_entry_recursive_loop_detect_i(path_list_index); /* * link to the path-list provided by the source, and go check * if that forms any loops in the graph. */ fib_entry->fe_parent = esrc->fes_pl; fib_entry->fe_sibling = fib_path_list_child_add(fib_entry->fe_parent, FIB_NODE_TYPE_ENTRY, fib_entry_get_index(fib_entry)); fib_entry_recursive_loop_detect_i(fib_entry->fe_parent); fib_path_list_unlock(path_list_index); /* * If this source should push its state to covered prefixs, do that now. */ fib_entry_src_covered_inherit_add(fib_entry, source); } if (!remain_installed) { fib_entry_src_action_uninstall(fib_entry); } else { fib_entry_src_action_install(fib_entry, source); } fib_entry_src_action_fwd_update(fib_entry, source); } fib_entry_t * fib_entry_src_action_installed (fib_entry_t *fib_entry, fib_source_t source) { fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_installed, (esrc, fib_entry)); fib_entry_src_action_fwd_update(fib_entry, source); return (fib_entry); } /* * fib_entry_src_action_add * * Adding a source can result in a new fib_entry being created, which * can inturn mean the pool is realloc'd and thus the entry passed as * an argument it also realloc'd * @return the original entry */ fib_entry_t * fib_entry_src_action_add (fib_entry_t *fib_entry, fib_source_t source, fib_entry_flag_t flags, const dpo_id_t *dpo) { fib_entry_src_t *esrc; esrc = fib_entry_src_find_or_create(fib_entry, source, flags); ASSERT(esrc->fes_ref_count < 255); esrc->fes_ref_count++; if (flags != esrc->fes_entry_flags) { FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_flags_change, (esrc, fib_entry, flags)); } esrc->fes_entry_flags = flags; if (1 != esrc->fes_ref_count) { /* * we only want to add the source on the 0->1 transition */ return (fib_entry); } FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_add, (esrc, fib_entry, flags, fib_entry_get_dpo_proto(fib_entry), dpo)); esrc->fes_flags |= FIB_ENTRY_SRC_FLAG_ADDED; fib_path_list_lock(esrc->fes_pl); /* * the source owns a lock on the entry */ fib_entry_lock(fib_entry_get_index(fib_entry)); return (fib_entry); } /* * fib_entry_src_action_update * * Adding a source can result in a new fib_entry being created, which * can inturn mean the pool is realloc'd and thus the entry passed as * an argument it also realloc'd * @return the original entry */ fib_entry_t * fib_entry_src_action_update (fib_entry_t *fib_entry, fib_source_t source, fib_entry_flag_t flags, const dpo_id_t *dpo) { fib_node_index_t old_path_list_index; fib_entry_src_t *esrc; esrc = fib_entry_src_find_or_create(fib_entry, source, flags); if (NULL == esrc) { return (fib_entry_src_action_add(fib_entry, source, flags, dpo)); } old_path_list_index = esrc->fes_pl; esrc->fes_entry_flags = flags; FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_add, (esrc, fib_entry, flags, fib_entry_get_dpo_proto(fib_entry), dpo)); esrc->fes_flags |= FIB_ENTRY_SRC_FLAG_ADDED; fib_path_list_lock(esrc->fes_pl); fib_path_list_unlock(old_path_list_index); return (fib_entry); } fib_entry_src_flag_t fib_entry_src_action_remove_or_update_inherit (fib_entry_t *fib_entry, fib_source_t source) { fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); if (NULL == esrc) return (FIB_ENTRY_SRC_FLAG_ACTIVE); if ((esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT) && (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED)) { fib_entry_src_t *cover_src; fib_node_index_t coveri; fib_entry_t *cover; /* * this source was pushing inherited state, but so is its * cover. Now that this source is going away, we need to * pull the covers forwarding and use it to update the covereds. * Go grab the path-list from the cover, rather than start a walk from * the cover, so we don't recursively update this entry. */ coveri = fib_table_get_less_specific(fib_entry->fe_fib_index, &fib_entry->fe_prefix); /* * only the default route has itself as its own cover, but the * default route cannot have inherited from something else. */ ASSERT(coveri != fib_entry_get_index(fib_entry)); cover = fib_entry_get(coveri); cover_src = fib_entry_src_find(cover, source); ASSERT(NULL != cover_src); esrc = fib_entry_src_action_update_from_cover(fib_entry, cover_src); esrc->fes_entry_flags &= ~FIB_ENTRY_FLAG_COVERED_INHERIT; /* * Now push the new state from the cover down to the covereds */ fib_entry_src_covered_inherit_add(fib_entry, source); return (esrc->fes_flags); } else { return (fib_entry_src_action_remove(fib_entry, source)); } } fib_entry_src_flag_t fib_entry_src_action_remove (fib_entry_t *fib_entry, fib_source_t source) { fib_node_index_t old_path_list; fib_entry_src_flag_t sflags; fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); if (NULL == esrc) return (FIB_ENTRY_SRC_FLAG_ACTIVE); esrc->fes_ref_count--; sflags = esrc->fes_flags; if (0 != esrc->fes_ref_count) { /* * only remove the source on the 1->0 transisition */ return (sflags); } if (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE) { fib_entry_src_action_deactivate(fib_entry, source); } else if (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_CONTRIBUTING) { FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_deactivate, (esrc, fib_entry)); esrc->fes_flags &= ~FIB_ENTRY_SRC_FLAG_CONTRIBUTING; } old_path_list = esrc->fes_pl; FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_remove, (esrc)); fib_path_list_unlock(old_path_list); fib_entry_unlock(fib_entry_get_index(fib_entry)); sflags &= ~FIB_ENTRY_SRC_FLAG_ADDED; fib_entry_src_action_deinit(fib_entry, source); return (sflags); } /* * fib_route_attached_cross_table * * Return true the the route is attached via an interface that * is not in the same table as the route */ static int fib_route_attached_cross_table (const fib_entry_t *fib_entry, const fib_route_path_t *rpath) { const fib_prefix_t *pfx = &fib_entry->fe_prefix; switch (pfx->fp_proto) { case FIB_PROTOCOL_MPLS: /* MPLS routes are never imported/exported */ return (0); case FIB_PROTOCOL_IP6: /* Ignore link local addresses these also can't be imported/exported */ if (ip6_address_is_link_local_unicast (&pfx->fp_addr.ip6)) { return (0); } break; case FIB_PROTOCOL_IP4: break; } /* * an attached path and entry's fib index not equal to interface's index */ if (fib_route_path_is_attached(rpath) && fib_entry->fe_fib_index != fib_table_get_index_for_sw_if_index(fib_entry_get_proto(fib_entry), rpath->frp_sw_if_index)) { return (!0); } return (0); } fib_path_list_flags_t fib_entry_src_flags_2_path_list_flags (fib_entry_flag_t eflags) { fib_path_list_flags_t plf = FIB_PATH_LIST_FLAG_NONE; if (eflags & FIB_ENTRY_FLAG_DROP) { plf |= FIB_PATH_LIST_FLAG_DROP; } if (eflags & FIB_ENTRY_FLAG_EXCLUSIVE) { plf |= FIB_PATH_LIST_FLAG_EXCLUSIVE; } if (eflags & FIB_ENTRY_FLAG_LOCAL) { plf |= FIB_PATH_LIST_FLAG_LOCAL; } return (plf); } static void fib_entry_flags_update (const fib_entry_t *fib_entry, const fib_route_path_t *rpaths, fib_path_list_flags_t *pl_flags, fib_entry_src_t *esrc) { const fib_route_path_t *rpath; vec_foreach(rpath, rpaths) { if ((esrc->fes_src == FIB_SOURCE_API) || (esrc->fes_src == FIB_SOURCE_CLI)) { if (fib_route_path_is_attached(rpath)) { esrc->fes_entry_flags |= FIB_ENTRY_FLAG_ATTACHED; } else { esrc->fes_entry_flags &= ~FIB_ENTRY_FLAG_ATTACHED; } if (rpath->frp_flags & FIB_ROUTE_PATH_DEAG) { esrc->fes_entry_flags |= FIB_ENTRY_FLAG_LOOSE_URPF_EXEMPT; } if (rpath->frp_flags & FIB_ROUTE_PATH_DROP) { esrc->fes_entry_flags |= FIB_ENTRY_FLAG_NO_ATTACHED_EXPORT; } } if (fib_route_attached_cross_table(fib_entry, rpath) && !(esrc->fes_entry_flags & FIB_ENTRY_FLAG_NO_ATTACHED_EXPORT)) { esrc->fes_entry_flags |= FIB_ENTRY_FLAG_IMPORT; } else { esrc->fes_entry_flags &= ~FIB_ENTRY_FLAG_IMPORT; } } } /* * fib_entry_src_action_add * * Adding a source can result in a new fib_entry being created, which * can inturn mean the pool is realloc'd and thus the entry passed as * an argument it also realloc'd * @return the entry */ fib_entry_t* fib_entry_src_action_path_add (fib_entry_t *fib_entry, fib_source_t source, fib_entry_flag_t flags, const fib_route_path_t *rpaths) { fib_node_index_t old_path_list; fib_path_list_flags_t pl_flags; fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); if (NULL == esrc) { const dpo_id_t *dpo; if (flags == FIB_ENTRY_FLAG_EXCLUSIVE) { dpo = &rpaths->dpo; } else { dpo = drop_dpo_get(fib_entry_get_dpo_proto(fib_entry)); } fib_entry = fib_entry_src_action_add(fib_entry, source, flags, dpo); esrc = fib_entry_src_find(fib_entry, source); } /* * we are no doubt modifying a path-list. If the path-list * is shared, and hence not modifiable, then the index returned * will be for a different path-list. This FIB entry to needs * to maintain its lock appropriately. */ old_path_list = esrc->fes_pl; ASSERT(FIB_ENTRY_SRC_VFT_EXISTS(esrc, fesv_path_add)); pl_flags = fib_entry_src_flags_2_path_list_flags(fib_entry_get_flags_i(fib_entry)); fib_entry_flags_update(fib_entry, rpaths, &pl_flags, esrc); FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_path_add, (esrc, fib_entry, pl_flags, rpaths)); fib_path_list_lock(esrc->fes_pl); fib_path_list_unlock(old_path_list); return (fib_entry); } /* * fib_entry_src_action_swap * * The source is providing new paths to replace the old ones. * Adding a source can result in a new fib_entry being created, which * can inturn mean the pool is realloc'd and thus the entry passed as * an argument it also realloc'd * @return the entry */ fib_entry_t* fib_entry_src_action_path_swap (fib_entry_t *fib_entry, fib_source_t source, fib_entry_flag_t flags, const fib_route_path_t *rpaths) { fib_node_index_t old_path_list; fib_path_list_flags_t pl_flags; fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); if (NULL == esrc) { const dpo_id_t *dpo; if (flags == FIB_ENTRY_FLAG_EXCLUSIVE) { dpo = &rpaths->dpo; } else { dpo = drop_dpo_get(fib_entry_get_dpo_proto(fib_entry)); } fib_entry = fib_entry_src_action_add(fib_entry, source, flags, dpo); esrc = fib_entry_src_find(fib_entry, source); } else { if (flags != esrc->fes_entry_flags) { FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_flags_change, (esrc, fib_entry, flags)); } esrc->fes_entry_flags = flags; } /* * swapping paths may create a new path-list (or may use an existing shared) * but we are certainly getting a different one. This FIB entry to needs * to maintain its lock appropriately. */ old_path_list = esrc->fes_pl; ASSERT(FIB_ENTRY_SRC_VFT_EXISTS(esrc, fesv_path_swap)); pl_flags = fib_entry_src_flags_2_path_list_flags(flags); fib_entry_flags_update(fib_entry, rpaths, &pl_flags, esrc); FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_path_swap, (esrc, fib_entry, pl_flags, rpaths)); fib_path_list_lock(esrc->fes_pl); fib_path_list_unlock(old_path_list); return (fib_entry); } fib_entry_src_flag_t fib_entry_src_action_path_remove (fib_entry_t *fib_entry, fib_source_t source, const fib_route_path_t *rpaths) { fib_path_list_flags_t pl_flags; fib_node_index_t old_path_list; fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); ASSERT(NULL != esrc); ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ADDED); /* * we no doubt modifying a path-list. If the path-list * is shared, and hence not modifiable, then the index returned * will be for a different path-list. This FIB entry to needs * to maintain its lock appropriately. */ old_path_list = esrc->fes_pl; ASSERT(FIB_ENTRY_SRC_VFT_EXISTS(esrc, fesv_path_remove)); pl_flags = fib_entry_src_flags_2_path_list_flags(fib_entry_get_flags_i(fib_entry)); fib_entry_flags_update(fib_entry, rpaths, &pl_flags, esrc); FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_path_remove, (esrc, pl_flags, rpaths)); /* * lock the new path-list, unlock the old if it had one */ fib_path_list_unlock(old_path_list); if (FIB_NODE_INDEX_INVALID != esrc->fes_pl) { fib_path_list_lock(esrc->fes_pl); return (FIB_ENTRY_SRC_FLAG_ADDED); } else { /* * no more paths left from this source */ fib_entry_src_action_remove_or_update_inherit(fib_entry, source); return (FIB_ENTRY_SRC_FLAG_NONE); } } u8* fib_entry_src_format (fib_entry_t *fib_entry, fib_source_t source, u8* s) { fib_entry_src_t *esrc; esrc = fib_entry_src_find(fib_entry, source); FIB_ENTRY_SRC_VFT_INVOKE_AND_RETURN(esrc, fesv_format, (esrc, s)); return (s); } adj_index_t fib_entry_get_adj_for_source (fib_node_index_t fib_entry_index, fib_source_t source) { fib_entry_t *fib_entry; fib_entry_src_t *esrc; if (FIB_NODE_INDEX_INVALID == fib_entry_index) return (ADJ_INDEX_INVALID); fib_entry = fib_entry_get(fib_entry_index); esrc = fib_entry_src_find(fib_entry, source); if (NULL != esrc) { if (FIB_NODE_INDEX_INVALID != esrc->fes_pl) { return (fib_path_list_get_adj( esrc->fes_pl, fib_entry_get_default_chain_type(fib_entry))); } } return (ADJ_INDEX_INVALID); } const int fib_entry_get_dpo_for_source (fib_node_index_t fib_entry_index, fib_source_t source, dpo_id_t *dpo) { fib_entry_t *fib_entry; fib_entry_src_t *esrc; if (FIB_NODE_INDEX_INVALID == fib_entry_index) return (0); fib_entry = fib_entry_get(fib_entry_index); esrc = fib_entry_src_find(fib_entry, source); if (NULL != esrc) { if (FIB_NODE_INDEX_INVALID != esrc->fes_pl) { fib_path_list_contribute_forwarding( esrc->fes_pl, fib_entry_get_default_chain_type(fib_entry), FIB_PATH_LIST_FWD_FLAG_NONE, dpo); return (dpo_id_is_valid(dpo)); } } return (0); } fib_node_index_t fib_entry_get_path_list_for_source (fib_node_index_t fib_entry_index, fib_source_t source) { fib_entry_t *fib_entry; fib_entry_src_t *esrc; if (FIB_NODE_INDEX_INVALID == fib_entry_index) return FIB_NODE_INDEX_INVALID; fib_entry = fib_entry_get(fib_entry_index); esrc = fib_entry_src_find(fib_entry, source); if (esrc) return esrc->fes_pl; return FIB_NODE_INDEX_INVALID; } u32 fib_entry_get_resolving_interface_for_source (fib_node_index_t entry_index, fib_source_t source) { fib_entry_t *fib_entry; fib_entry_src_t *esrc; fib_entry = fib_entry_get(entry_index); esrc = fib_entry_src_find(fib_entry, source); if (NULL != esrc) { if (FIB_NODE_INDEX_INVALID != esrc->fes_pl) { return (fib_path_list_get_resolving_interface(esrc->fes_pl)); } } return (~0); } fib_entry_flag_t fib_entry_get_flags_for_source (fib_node_index_t entry_index, fib_source_t source) { fib_entry_t *fib_entry; fib_entry_src_t *esrc; fib_entry = fib_entry_get(entry_index); esrc = fib_entry_src_find(fib_entry, source); if (NULL != esrc) { return (esrc->fes_entry_flags); } return (FIB_ENTRY_FLAG_NONE); } fib_source_t fib_entry_get_source_i (const fib_entry_t *fib_entry) { /* the vector of sources is deliberately arranged in priority order */ if (0 == vec_len(fib_entry->fe_srcs)) return (FIB_SOURCE_INVALID); return (vec_elt(fib_entry->fe_srcs, 0).fes_src); } fib_entry_flag_t fib_entry_get_flags_i (const fib_entry_t *fib_entry) { /* the vector of sources is deliberately arranged in priority order */ if (0 == vec_len(fib_entry->fe_srcs)) return (FIB_ENTRY_FLAG_NONE); return (vec_elt(fib_entry->fe_srcs, 0).fes_entry_flags); } void fib_entry_set_source_data (fib_node_index_t fib_entry_index, fib_source_t source, const void *data) { fib_entry_t *fib_entry; fib_entry_src_t *esrc; fib_entry = fib_entry_get(fib_entry_index); esrc = fib_entry_src_find(fib_entry, source); if (NULL != esrc) { FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_set_data, (esrc, fib_entry, data)); } } const void* fib_entry_get_source_data (fib_node_index_t fib_entry_index, fib_source_t source) { fib_entry_t *fib_entry; fib_entry_src_t *esrc; fib_entry = fib_entry_get(fib_entry_index); esrc = fib_entry_src_find(fib_entry, source); if (NULL != esrc) { FIB_ENTRY_SRC_VFT_INVOKE_AND_RETURN(esrc, fesv_get_data, (esrc, fib_entry)); } return (NULL); } void fib_entry_src_module_init (void) { fib_entry_src_rr_register(); fib_entry_src_interface_register(); fib_entry_src_interpose_register(); fib_entry_src_drop_register(); fib_entry_src_simple_register(); fib_entry_src_api_register(); fib_entry_src_adj_register(); fib_entry_src_mpls_register(); fib_entry_src_lisp_register(); }