aboutsummaryrefslogtreecommitdiffstats
path: root/src/vnet/vxlan-gpe/vxlan_gpe.c
diff options
context:
space:
mode:
authorsharath reddy <sharathkumarboyanapally@gmail.com>2017-12-11 11:31:31 +0530
committerJohn Lo <loj@cisco.com>2017-12-13 05:19:13 +0000
commit6f8273aec1fa3f3d597b090871cc7ac301d7e014 (patch)
tree767554f021b58dffb1ac2008fd70b99249d5874d /src/vnet/vxlan-gpe/vxlan_gpe.c
parent129e80e32abb0999b069a2175fbb64635f311ed0 (diff)
VPP-275 Coding standards cleanup - vnet/vnet/vxlan-gpe
Change-Id: Ifabb8d22d20bc1031664d5f004e74cd363759ab6 Signed-off-by: sharath reddy <sharathkumarboyanapally@gmail.com>
Diffstat (limited to 'src/vnet/vxlan-gpe/vxlan_gpe.c')
-rw-r--r--src/vnet/vxlan-gpe/vxlan_gpe.c902
1 files changed, 470 insertions, 432 deletions
diff --git a/src/vnet/vxlan-gpe/vxlan_gpe.c b/src/vnet/vxlan-gpe/vxlan_gpe.c
index b13a7343ddd..939b5fafd7d 100644
--- a/src/vnet/vxlan-gpe/vxlan_gpe.c
+++ b/src/vnet/vxlan-gpe/vxlan_gpe.c
@@ -53,15 +53,16 @@ vxlan_gpe_main_t vxlan_gpe_main;
* @return *s formatted string
*
*/
-u8 * format_vxlan_gpe_tunnel (u8 * s, va_list * args)
+u8 *
+format_vxlan_gpe_tunnel (u8 * s, va_list * args)
{
- vxlan_gpe_tunnel_t * t = va_arg (*args, vxlan_gpe_tunnel_t *);
- vxlan_gpe_main_t * ngm = &vxlan_gpe_main;
+ vxlan_gpe_tunnel_t *t = va_arg (*args, vxlan_gpe_tunnel_t *);
+ vxlan_gpe_main_t *ngm = &vxlan_gpe_main;
s = format (s, "[%d] local: %U remote: %U ",
- t - ngm->tunnels,
- format_ip46_address, &t->local, IP46_TYPE_ANY,
- format_ip46_address, &t->remote, IP46_TYPE_ANY);
+ t - ngm->tunnels,
+ format_ip46_address, &t->local, IP46_TYPE_ANY,
+ format_ip46_address, &t->remote, IP46_TYPE_ANY);
s = format (s, " vxlan VNI %d ", t->vni);
@@ -85,10 +86,9 @@ u8 * format_vxlan_gpe_tunnel (u8 * s, va_list * args)
if (ip46_address_is_multicast (&t->remote))
s = format (s, "mcast_sw_if_index %d ", t->mcast_sw_if_index);
-
+
s = format (s, " fibs: (encap %d, decap %d)",
- t->encap_fib_index,
- t->decap_fib_index);
+ t->encap_fib_index, t->decap_fib_index);
return s;
}
@@ -102,15 +102,16 @@ u8 * format_vxlan_gpe_tunnel (u8 * s, va_list * args)
* @return *s formatted string
*
*/
-static u8 * format_vxlan_gpe_name (u8 * s, va_list * args)
+static u8 *
+format_vxlan_gpe_name (u8 * s, va_list * args)
{
u32 dev_instance = va_arg (*args, u32);
return format (s, "vxlan_gpe_tunnel%d", dev_instance);
}
-static uword dummy_interface_tx (vlib_main_t * vm,
- vlib_node_runtime_t * node,
- vlib_frame_t * frame)
+static uword
+dummy_interface_tx (vlib_main_t * vm,
+ vlib_node_runtime_t * node, vlib_frame_t * frame)
{
clib_warning ("you shouldn't be here, leaking buffers...");
return frame->n_vectors;
@@ -127,7 +128,8 @@ static uword dummy_interface_tx (vlib_main_t * vm,
*
*/
static clib_error_t *
-vxlan_gpe_interface_admin_up_down (vnet_main_t * vnm, u32 hw_if_index, u32 flags)
+vxlan_gpe_interface_admin_up_down (vnet_main_t * vnm, u32 hw_if_index,
+ u32 flags)
{
u32 hw_flags = (flags & VNET_SW_INTERFACE_FLAG_ADMIN_UP) ?
VNET_HW_INTERFACE_FLAG_LINK_UP : 0;
@@ -136,6 +138,7 @@ vxlan_gpe_interface_admin_up_down (vnet_main_t * vnm, u32 hw_if_index, u32 flags
return 0;
}
+/* *INDENT-OFF* */
VNET_DEVICE_CLASS (vxlan_gpe_device_class,static) = {
.name = "VXLAN_GPE",
.format_device_name = format_vxlan_gpe_name,
@@ -143,6 +146,7 @@ VNET_DEVICE_CLASS (vxlan_gpe_device_class,static) = {
.tx_function = dummy_interface_tx,
.admin_up_down_function = vxlan_gpe_interface_admin_up_down,
};
+/* *INDENT-ON* */
/**
@@ -154,38 +158,42 @@ VNET_DEVICE_CLASS (vxlan_gpe_device_class,static) = {
* @return *s
*
*/
-static u8 * format_vxlan_gpe_header_with_length (u8 * s, va_list * args)
+static u8 *
+format_vxlan_gpe_header_with_length (u8 * s, va_list * args)
{
u32 dev_instance = va_arg (*args, u32);
s = format (s, "unimplemented dev %u", dev_instance);
return s;
}
+/* *INDENT-OFF* */
VNET_HW_INTERFACE_CLASS (vxlan_gpe_hw_class) = {
.name = "VXLAN_GPE",
.format_header = format_vxlan_gpe_header_with_length,
.build_rewrite = default_build_rewrite,
};
+/* *INDENT-ON* */
static void
-vxlan_gpe_tunnel_restack_dpo(vxlan_gpe_tunnel_t * t)
+vxlan_gpe_tunnel_restack_dpo (vxlan_gpe_tunnel_t * t)
{
- dpo_id_t dpo = DPO_INVALID;
- u32 encap_index = vxlan_gpe_encap_node.index;
- fib_forward_chain_type_t forw_type = ip46_address_is_ip4(&t->remote) ?
- FIB_FORW_CHAIN_TYPE_UNICAST_IP4 : FIB_FORW_CHAIN_TYPE_UNICAST_IP6;
-
- fib_entry_contribute_forwarding (t->fib_entry_index, forw_type, &dpo);
- dpo_stack_from_node (encap_index, &t->next_dpo, &dpo);
- dpo_reset(&dpo);
+ dpo_id_t dpo = DPO_INVALID;
+ u32 encap_index = vxlan_gpe_encap_node.index;
+ fib_forward_chain_type_t forw_type = ip46_address_is_ip4 (&t->remote) ?
+ FIB_FORW_CHAIN_TYPE_UNICAST_IP4 : FIB_FORW_CHAIN_TYPE_UNICAST_IP6;
+
+ fib_entry_contribute_forwarding (t->fib_entry_index, forw_type, &dpo);
+ dpo_stack_from_node (encap_index, &t->next_dpo, &dpo);
+ dpo_reset (&dpo);
}
static vxlan_gpe_tunnel_t *
-vxlan_gpe_tunnel_from_fib_node (fib_node_t *node)
+vxlan_gpe_tunnel_from_fib_node (fib_node_t * node)
{
- ASSERT(FIB_NODE_TYPE_VXLAN_GPE_TUNNEL == node->fn_type);
- return ((vxlan_gpe_tunnel_t*) (((char*)node) -
- STRUCT_OFFSET_OF(vxlan_gpe_tunnel_t, node)));
+ ASSERT (FIB_NODE_TYPE_VXLAN_GPE_TUNNEL == node->fn_type);
+ return ((vxlan_gpe_tunnel_t *) (((char *) node) -
+ STRUCT_OFFSET_OF (vxlan_gpe_tunnel_t,
+ node)));
}
/**
@@ -193,38 +201,37 @@ vxlan_gpe_tunnel_from_fib_node (fib_node_t *node)
* Here we will restack the new dpo of VXLAN_GPE DIP to encap node.
*/
static fib_node_back_walk_rc_t
-vxlan_gpe_tunnel_back_walk (fib_node_t *node,
- fib_node_back_walk_ctx_t *ctx)
+vxlan_gpe_tunnel_back_walk (fib_node_t * node, fib_node_back_walk_ctx_t * ctx)
{
- vxlan_gpe_tunnel_restack_dpo(vxlan_gpe_tunnel_from_fib_node(node));
- return (FIB_NODE_BACK_WALK_CONTINUE);
+ vxlan_gpe_tunnel_restack_dpo (vxlan_gpe_tunnel_from_fib_node (node));
+ return (FIB_NODE_BACK_WALK_CONTINUE);
}
/**
* Function definition to get a FIB node from its index
*/
-static fib_node_t*
+static fib_node_t *
vxlan_gpe_tunnel_fib_node_get (fib_node_index_t index)
{
- vxlan_gpe_tunnel_t * t;
- vxlan_gpe_main_t * ngm = &vxlan_gpe_main;
+ vxlan_gpe_tunnel_t *t;
+ vxlan_gpe_main_t *ngm = &vxlan_gpe_main;
- t = pool_elt_at_index(ngm->tunnels, index);
+ t = pool_elt_at_index (ngm->tunnels, index);
- return (&t->node);
+ return (&t->node);
}
/**
* Function definition to inform the FIB node that its last lock has gone.
*/
static void
-vxlan_gpe_tunnel_last_lock_gone (fib_node_t *node)
+vxlan_gpe_tunnel_last_lock_gone (fib_node_t * node)
{
- /*
- * The VXLAN_GPE tunnel is a root of the graph. As such
- * it never has children and thus is never locked.
- */
- ASSERT(0);
+ /*
+ * The VXLAN_GPE tunnel is a root of the graph. As such
+ * it never has children and thus is never locked.
+ */
+ ASSERT (0);
}
/*
@@ -232,9 +239,9 @@ vxlan_gpe_tunnel_last_lock_gone (fib_node_t *node)
* for participation in the FIB object graph.
*/
const static fib_node_vft_t vxlan_gpe_vft = {
- .fnv_get = vxlan_gpe_tunnel_fib_node_get,
- .fnv_last_lock = vxlan_gpe_tunnel_last_lock_gone,
- .fnv_back_walk = vxlan_gpe_tunnel_back_walk,
+ .fnv_get = vxlan_gpe_tunnel_fib_node_get,
+ .fnv_last_lock = vxlan_gpe_tunnel_last_lock_gone,
+ .fnv_back_walk = vxlan_gpe_tunnel_back_walk,
};
#define foreach_gpe_copy_field \
@@ -265,18 +272,19 @@ _(decap_fib_index)
* @return rc
*
*/
-int vxlan4_gpe_rewrite (vxlan_gpe_tunnel_t * t, u32 extension_size,
- u8 protocol_override, uword encap_next_node)
+int
+vxlan4_gpe_rewrite (vxlan_gpe_tunnel_t * t, u32 extension_size,
+ u8 protocol_override, uword encap_next_node)
{
u8 *rw = 0;
- ip4_header_t * ip0;
- ip4_vxlan_gpe_header_t * h0;
+ ip4_header_t *ip0;
+ ip4_vxlan_gpe_header_t *h0;
int len;
len = sizeof (*h0) + extension_size;
- vec_free(t->rewrite);
- vec_validate_aligned (rw, len-1, CLIB_CACHE_LINE_BYTES);
+ vec_free (t->rewrite);
+ vec_validate_aligned (rw, len - 1, CLIB_CACHE_LINE_BYTES);
h0 = (ip4_vxlan_gpe_header_t *) rw;
@@ -299,15 +307,15 @@ int vxlan4_gpe_rewrite (vxlan_gpe_tunnel_t * t, u32 extension_size,
h0->vxlan.flags = VXLAN_GPE_FLAGS_I | VXLAN_GPE_FLAGS_P;
h0->vxlan.ver_res = VXLAN_GPE_VERSION;
if (protocol_override)
- {
+ {
h0->vxlan.protocol = protocol_override;
- }
+ }
else
- {
+ {
h0->vxlan.protocol = t->protocol;
- }
- t->rewrite_size = sizeof(ip4_vxlan_gpe_header_t) + extension_size;
- h0->vxlan.vni_res = clib_host_to_net_u32 (t->vni<<8);
+ }
+ t->rewrite_size = sizeof (ip4_vxlan_gpe_header_t) + extension_size;
+ h0->vxlan.vni_res = clib_host_to_net_u32 (t->vni << 8);
t->rewrite = rw;
t->encap_next_node = encap_next_node;
@@ -322,24 +330,26 @@ int vxlan4_gpe_rewrite (vxlan_gpe_tunnel_t * t, u32 extension_size,
* @return rc
*
*/
-int vxlan6_gpe_rewrite (vxlan_gpe_tunnel_t * t, u32 extension_size,
- u8 protocol_override, uword encap_next_node)
+int
+vxlan6_gpe_rewrite (vxlan_gpe_tunnel_t * t, u32 extension_size,
+ u8 protocol_override, uword encap_next_node)
{
u8 *rw = 0;
- ip6_header_t * ip0;
- ip6_vxlan_gpe_header_t * h0;
+ ip6_header_t *ip0;
+ ip6_vxlan_gpe_header_t *h0;
int len;
len = sizeof (*h0) + extension_size;
- vec_free(t->rewrite);
- vec_validate_aligned (rw, len-1, CLIB_CACHE_LINE_BYTES);
+ vec_free (t->rewrite);
+ vec_validate_aligned (rw, len - 1, CLIB_CACHE_LINE_BYTES);
h0 = (ip6_vxlan_gpe_header_t *) rw;
/* Fixed portion of the (outer) ip4 header */
ip0 = &h0->ip6;
- ip0->ip_version_traffic_class_and_flow_label = clib_host_to_net_u32(6 << 28);
+ ip0->ip_version_traffic_class_and_flow_label =
+ clib_host_to_net_u32 (6 << 28);
ip0->hop_limit = 255;
ip0->protocol = IP_PROTOCOL_UDP;
@@ -356,15 +366,15 @@ int vxlan6_gpe_rewrite (vxlan_gpe_tunnel_t * t, u32 extension_size,
h0->vxlan.flags = VXLAN_GPE_FLAGS_I | VXLAN_GPE_FLAGS_P;
h0->vxlan.ver_res = VXLAN_GPE_VERSION;
if (protocol_override)
- {
+ {
h0->vxlan.protocol = t->protocol;
- }
+ }
else
- {
+ {
h0->vxlan.protocol = protocol_override;
- }
- t->rewrite_size = sizeof(ip4_vxlan_gpe_header_t) + extension_size;
- h0->vxlan.vni_res = clib_host_to_net_u32 (t->vni<<8);
+ }
+ t->rewrite_size = sizeof (ip4_vxlan_gpe_header_t) + extension_size;
+ h0->vxlan.vni_res = clib_host_to_net_u32 (t->vni << 8);
t->rewrite = rw;
t->encap_next_node = encap_next_node;
@@ -372,51 +382,54 @@ int vxlan6_gpe_rewrite (vxlan_gpe_tunnel_t * t, u32 extension_size,
}
static void
-hash_set_key_copy (uword ** h, void * key, uword v) {
- size_t ksz = hash_header(*h)->user;
- void * copy = clib_mem_alloc (ksz);
- clib_memcpy (copy, key, ksz);
- hash_set_mem (*h, copy, v);
+hash_set_key_copy (uword ** h, void *key, uword v)
+{
+ size_t ksz = hash_header (*h)->user;
+ void *copy = clib_mem_alloc (ksz);
+ clib_memcpy (copy, key, ksz);
+ hash_set_mem (*h, copy, v);
}
static void
-hash_unset_key_free (uword ** h, void * key) {
- hash_pair_t * hp = hash_get_pair_mem (*h, key);
- ASSERT (hp);
- key = uword_to_pointer (hp->key, void *);
- hash_unset_mem (*h, key);
- clib_mem_free (key);
+hash_unset_key_free (uword ** h, void *key)
+{
+ hash_pair_t *hp = hash_get_pair_mem (*h, key);
+ ASSERT (hp);
+ key = uword_to_pointer (hp->key, void *);
+ hash_unset_mem (*h, key);
+ clib_mem_free (key);
}
static uword
-vtep_addr_ref(ip46_address_t *ip)
+vtep_addr_ref (ip46_address_t * ip)
{
- uword *vtep = ip46_address_is_ip4(ip) ?
- hash_get (vxlan_gpe_main.vtep4, ip->ip4.as_u32) :
- hash_get_mem (vxlan_gpe_main.vtep6, &ip->ip6);
- if (vtep)
- return ++(*vtep);
- ip46_address_is_ip4(ip) ?
- hash_set (vxlan_gpe_main.vtep4, ip->ip4.as_u32, 1) :
- hash_set_key_copy (&vxlan_gpe_main.vtep6, &ip->ip6, 1);
- return 1;
+ uword *vtep = ip46_address_is_ip4 (ip) ?
+ hash_get (vxlan_gpe_main.vtep4, ip->ip4.as_u32) :
+ hash_get_mem (vxlan_gpe_main.vtep6, &ip->ip6);
+ if (vtep)
+ return ++(*vtep);
+ ip46_address_is_ip4 (ip) ?
+ hash_set (vxlan_gpe_main.vtep4, ip->ip4.as_u32, 1) :
+ hash_set_key_copy (&vxlan_gpe_main.vtep6, &ip->ip6, 1);
+ return 1;
}
static uword
-vtep_addr_unref(ip46_address_t *ip)
+vtep_addr_unref (ip46_address_t * ip)
{
- uword *vtep = ip46_address_is_ip4(ip) ?
- hash_get (vxlan_gpe_main.vtep4, ip->ip4.as_u32) :
- hash_get_mem (vxlan_gpe_main.vtep6, &ip->ip6);
- ASSERT(vtep);
- if (--(*vtep) != 0)
- return *vtep;
- ip46_address_is_ip4(ip) ?
- hash_unset (vxlan_gpe_main.vtep4, ip->ip4.as_u32) :
- hash_unset_key_free (&vxlan_gpe_main.vtep6, &ip->ip6);
- return 0;
+ uword *vtep = ip46_address_is_ip4 (ip) ?
+ hash_get (vxlan_gpe_main.vtep4, ip->ip4.as_u32) :
+ hash_get_mem (vxlan_gpe_main.vtep6, &ip->ip6);
+ ASSERT (vtep);
+ if (--(*vtep) != 0)
+ return *vtep;
+ ip46_address_is_ip4 (ip) ?
+ hash_unset (vxlan_gpe_main.vtep4, ip->ip4.as_u32) :
+ hash_unset_key_free (&vxlan_gpe_main.vtep6, &ip->ip6);
+ return 0;
}
+/* *INDENT-OFF* */
typedef CLIB_PACKED(union {
struct {
fib_node_index_t mfib_entry_index;
@@ -424,43 +437,44 @@ typedef CLIB_PACKED(union {
};
u64 as_u64;
}) mcast_shared_t;
+/* *INDENT-ON* */
static inline mcast_shared_t
-mcast_shared_get(ip46_address_t * ip)
+mcast_shared_get (ip46_address_t * ip)
{
- ASSERT(ip46_address_is_multicast(ip));
- uword * p = hash_get_mem (vxlan_gpe_main.mcast_shared, ip);
- ASSERT(p);
- return (mcast_shared_t) { .as_u64 = *p };
+ ASSERT (ip46_address_is_multicast (ip));
+ uword *p = hash_get_mem (vxlan_gpe_main.mcast_shared, ip);
+ ASSERT (p);
+ return (mcast_shared_t)
+ {
+ .as_u64 = *p};
}
static inline void
-mcast_shared_add(ip46_address_t *remote,
- fib_node_index_t mfei,
- adj_index_t ai)
+mcast_shared_add (ip46_address_t * remote,
+ fib_node_index_t mfei, adj_index_t ai)
{
- mcast_shared_t new_ep = {
- .mcast_adj_index = ai,
- .mfib_entry_index = mfei,
- };
+ mcast_shared_t new_ep = {
+ .mcast_adj_index = ai,
+ .mfib_entry_index = mfei,
+ };
- hash_set_key_copy (&vxlan_gpe_main.mcast_shared, remote, new_ep.as_u64);
+ hash_set_key_copy (&vxlan_gpe_main.mcast_shared, remote, new_ep.as_u64);
}
static inline void
-mcast_shared_remove(ip46_address_t *remote)
+mcast_shared_remove (ip46_address_t * remote)
{
- mcast_shared_t ep = mcast_shared_get(remote);
+ mcast_shared_t ep = mcast_shared_get (remote);
- adj_unlock(ep.mcast_adj_index);
- mfib_table_entry_delete_index(ep.mfib_entry_index,
- MFIB_SOURCE_VXLAN_GPE);
+ adj_unlock (ep.mcast_adj_index);
+ mfib_table_entry_delete_index (ep.mfib_entry_index, MFIB_SOURCE_VXLAN_GPE);
- hash_unset_key_free (&vxlan_gpe_main.mcast_shared, remote);
+ hash_unset_key_free (&vxlan_gpe_main.mcast_shared, remote);
}
static inline fib_protocol_t
-fib_ip_proto(bool is_ip6)
+fib_ip_proto (bool is_ip6)
{
return (is_ip6) ? FIB_PROTOCOL_IP6 : FIB_PROTOCOL_IP4;
}
@@ -475,13 +489,13 @@ fib_ip_proto(bool is_ip6)
*
*/
int vnet_vxlan_gpe_add_del_tunnel
-(vnet_vxlan_gpe_add_del_tunnel_args_t *a, u32 * sw_if_indexp)
+ (vnet_vxlan_gpe_add_del_tunnel_args_t * a, u32 * sw_if_indexp)
{
- vxlan_gpe_main_t * ngm = &vxlan_gpe_main;
+ vxlan_gpe_main_t *ngm = &vxlan_gpe_main;
vxlan_gpe_tunnel_t *t = 0;
- vnet_main_t * vnm = ngm->vnet_main;
- vnet_hw_interface_t * hi;
- uword * p;
+ vnet_main_t *vnm = ngm->vnet_main;
+ vnet_hw_interface_t *hi;
+ uword *p;
u32 hw_if_index = ~0;
u32 sw_if_index = ~0;
int rv;
@@ -490,32 +504,32 @@ int vnet_vxlan_gpe_add_del_tunnel
u32 is_ip6 = a->is_ip6;
if (!is_ip6)
- {
- key4.local = a->local.ip4.as_u32;
- key4.remote = a->remote.ip4.as_u32;
- key4.vni = clib_host_to_net_u32 (a->vni << 8);
- key4.pad = 0;
+ {
+ key4.local = a->local.ip4.as_u32;
+ key4.remote = a->remote.ip4.as_u32;
+ key4.vni = clib_host_to_net_u32 (a->vni << 8);
+ key4.pad = 0;
- p = hash_get_mem(ngm->vxlan4_gpe_tunnel_by_key, &key4);
- }
+ p = hash_get_mem (ngm->vxlan4_gpe_tunnel_by_key, &key4);
+ }
else
- {
- key6.local.as_u64[0] = a->local.ip6.as_u64[0];
- key6.local.as_u64[1] = a->local.ip6.as_u64[1];
- key6.remote.as_u64[0] = a->remote.ip6.as_u64[0];
- key6.remote.as_u64[1] = a->remote.ip6.as_u64[1];
- key6.vni = clib_host_to_net_u32 (a->vni << 8);
+ {
+ key6.local.as_u64[0] = a->local.ip6.as_u64[0];
+ key6.local.as_u64[1] = a->local.ip6.as_u64[1];
+ key6.remote.as_u64[0] = a->remote.ip6.as_u64[0];
+ key6.remote.as_u64[1] = a->remote.ip6.as_u64[1];
+ key6.vni = clib_host_to_net_u32 (a->vni << 8);
- p = hash_get_mem(ngm->vxlan6_gpe_tunnel_by_key, &key6);
- }
+ p = hash_get_mem (ngm->vxlan6_gpe_tunnel_by_key, &key6);
+ }
if (a->is_add)
{
- l2input_main_t * l2im = &l2input_main;
+ l2input_main_t *l2im = &l2input_main;
/* adding a tunnel: tunnel must not already exist */
if (p)
- return VNET_API_ERROR_TUNNEL_EXIST;
+ return VNET_API_ERROR_TUNNEL_EXIST;
pool_get_aligned (ngm->tunnels, t, CLIB_CACHE_LINE_BYTES);
memset (t, 0, sizeof (*t));
@@ -523,72 +537,81 @@ int vnet_vxlan_gpe_add_del_tunnel
/* copy from arg structure */
#define _(x) t->x = a->x;
foreach_gpe_copy_field;
- if (!a->is_ip6) foreach_copy_ipv4
- else foreach_copy_ipv6
+ if (!a->is_ip6)
+ foreach_copy_ipv4
+ else
+ foreach_copy_ipv6
#undef _
+ if (!a->is_ip6)
+ t->flags |= VXLAN_GPE_TUNNEL_IS_IPV4;
- if (!a->is_ip6) t->flags |= VXLAN_GPE_TUNNEL_IS_IPV4;
-
- if (!a->is_ip6) {
- rv = vxlan4_gpe_rewrite (t, 0, 0, VXLAN_GPE_ENCAP_NEXT_IP4_LOOKUP);
- } else {
- rv = vxlan6_gpe_rewrite (t, 0, 0, VXLAN_GPE_ENCAP_NEXT_IP6_LOOKUP);
- }
+ if (!a->is_ip6)
+ {
+ rv = vxlan4_gpe_rewrite (t, 0, 0, VXLAN_GPE_ENCAP_NEXT_IP4_LOOKUP);
+ }
+ else
+ {
+ rv = vxlan6_gpe_rewrite (t, 0, 0, VXLAN_GPE_ENCAP_NEXT_IP6_LOOKUP);
+ }
if (rv)
- {
- pool_put (ngm->tunnels, t);
- return rv;
- }
+ {
+ pool_put (ngm->tunnels, t);
+ return rv;
+ }
if (!is_ip6)
- {
- key4_copy = clib_mem_alloc (sizeof (*key4_copy));
- clib_memcpy (key4_copy, &key4, sizeof (*key4_copy));
- hash_set_mem (ngm->vxlan4_gpe_tunnel_by_key, key4_copy,
- t - ngm->tunnels);
- }
+ {
+ key4_copy = clib_mem_alloc (sizeof (*key4_copy));
+ clib_memcpy (key4_copy, &key4, sizeof (*key4_copy));
+ hash_set_mem (ngm->vxlan4_gpe_tunnel_by_key, key4_copy,
+ t - ngm->tunnels);
+ }
else
- {
- key6_copy = clib_mem_alloc (sizeof (*key6_copy));
- clib_memcpy (key6_copy, &key6, sizeof (*key6_copy));
- hash_set_mem (ngm->vxlan6_gpe_tunnel_by_key, key6_copy,
- t - ngm->tunnels);
- }
+ {
+ key6_copy = clib_mem_alloc (sizeof (*key6_copy));
+ clib_memcpy (key6_copy, &key6, sizeof (*key6_copy));
+ hash_set_mem (ngm->vxlan6_gpe_tunnel_by_key, key6_copy,
+ t - ngm->tunnels);
+ }
if (vec_len (ngm->free_vxlan_gpe_tunnel_hw_if_indices) > 0)
- {
- vnet_interface_main_t * im = &vnm->interface_main;
- hw_if_index = ngm->free_vxlan_gpe_tunnel_hw_if_indices
- [vec_len (ngm->free_vxlan_gpe_tunnel_hw_if_indices)-1];
- _vec_len (ngm->free_vxlan_gpe_tunnel_hw_if_indices) -= 1;
-
- hi = vnet_get_hw_interface (vnm, hw_if_index);
- hi->dev_instance = t - ngm->tunnels;
- hi->hw_instance = hi->dev_instance;
+ {
+ vnet_interface_main_t *im = &vnm->interface_main;
+ hw_if_index = ngm->free_vxlan_gpe_tunnel_hw_if_indices
+ [vec_len (ngm->free_vxlan_gpe_tunnel_hw_if_indices) - 1];
+ _vec_len (ngm->free_vxlan_gpe_tunnel_hw_if_indices) -= 1;
+
+ hi = vnet_get_hw_interface (vnm, hw_if_index);
+ hi->dev_instance = t - ngm->tunnels;
+ hi->hw_instance = hi->dev_instance;
/* clear old stats of freed tunnel before reuse */
sw_if_index = hi->sw_if_index;
- vnet_interface_counter_lock(im);
+ vnet_interface_counter_lock (im);
vlib_zero_combined_counter
- (&im->combined_sw_if_counters[VNET_INTERFACE_COUNTER_TX], sw_if_index);
- vlib_zero_combined_counter
- (&im->combined_sw_if_counters[VNET_INTERFACE_COUNTER_RX], sw_if_index);
- vlib_zero_simple_counter
- (&im->sw_if_counters[VNET_INTERFACE_COUNTER_DROP], sw_if_index);
- vnet_interface_counter_unlock(im);
- }
+ (&im->combined_sw_if_counters[VNET_INTERFACE_COUNTER_TX],
+ sw_if_index);
+ vlib_zero_combined_counter (&im->combined_sw_if_counters
+ [VNET_INTERFACE_COUNTER_RX],
+ sw_if_index);
+ vlib_zero_simple_counter (&im->sw_if_counters
+ [VNET_INTERFACE_COUNTER_DROP],
+ sw_if_index);
+ vnet_interface_counter_unlock (im);
+ }
else
- {
- hw_if_index = vnet_register_interface
- (vnm, vxlan_gpe_device_class.index, t - ngm->tunnels,
- vxlan_gpe_hw_class.index, t - ngm->tunnels);
- hi = vnet_get_hw_interface (vnm, hw_if_index);
- hi->output_node_index = vxlan_gpe_encap_node.index;
- }
+ {
+ hw_if_index = vnet_register_interface
+ (vnm, vxlan_gpe_device_class.index, t - ngm->tunnels,
+ vxlan_gpe_hw_class.index, t - ngm->tunnels);
+ hi = vnet_get_hw_interface (vnm, hw_if_index);
+ hi->output_node_index = vxlan_gpe_encap_node.index;
+ }
t->hw_if_index = hw_if_index;
t->sw_if_index = sw_if_index = hi->sw_if_index;
- vec_validate_init_empty (ngm->tunnel_index_by_sw_if_index, sw_if_index, ~0);
+ vec_validate_init_empty (ngm->tunnel_index_by_sw_if_index, sw_if_index,
+ ~0);
ngm->tunnel_index_by_sw_if_index[sw_if_index] = t - ngm->tunnels;
/* setup l2 input config with l2 feature and bd 0 to drop packet */
@@ -596,159 +619,160 @@ int vnet_vxlan_gpe_add_del_tunnel
l2im->configs[sw_if_index].feature_bitmap = L2INPUT_FEAT_DROP;
l2im->configs[sw_if_index].bd_index = 0;
- vnet_sw_interface_t * si = vnet_get_sw_interface (vnm, sw_if_index);
+ vnet_sw_interface_t *si = vnet_get_sw_interface (vnm, sw_if_index);
si->flags &= ~VNET_SW_INTERFACE_FLAG_HIDDEN;
vnet_sw_interface_set_flags (vnm, hi->sw_if_index,
- VNET_SW_INTERFACE_FLAG_ADMIN_UP);
- fib_node_init(&t->node, FIB_NODE_TYPE_VXLAN_GPE_TUNNEL);
+ VNET_SW_INTERFACE_FLAG_ADMIN_UP);
+ fib_node_init (&t->node, FIB_NODE_TYPE_VXLAN_GPE_TUNNEL);
fib_prefix_t tun_remote_pfx;
u32 encap_index = vxlan_gpe_encap_node.index;
vnet_flood_class_t flood_class = VNET_FLOOD_CLASS_TUNNEL_NORMAL;
- fib_prefix_from_ip46_addr(&t->remote, &tun_remote_pfx);
- if (!ip46_address_is_multicast(&t->remote))
- {
- /* Unicast tunnel -
- * source the FIB entry for the tunnel's destination
- * and become a child thereof. The tunnel will then get poked
- * when the forwarding for the entry updates, and the tunnel can
- * re-stack accordingly
- */
- vtep_addr_ref(&t->local);
- t->fib_entry_index = fib_table_entry_special_add
- (t->encap_fib_index, &tun_remote_pfx, FIB_SOURCE_RR,
+ fib_prefix_from_ip46_addr (&t->remote, &tun_remote_pfx);
+ if (!ip46_address_is_multicast (&t->remote))
+ {
+ /* Unicast tunnel -
+ * source the FIB entry for the tunnel's destination
+ * and become a child thereof. The tunnel will then get poked
+ * when the forwarding for the entry updates, and the tunnel can
+ * re-stack accordingly
+ */
+ vtep_addr_ref (&t->local);
+ t->fib_entry_index = fib_table_entry_special_add
+ (t->encap_fib_index, &tun_remote_pfx, FIB_SOURCE_RR,
FIB_ENTRY_FLAG_NONE);
- t->sibling_index = fib_entry_child_add
- (t->fib_entry_index, FIB_NODE_TYPE_VXLAN_GPE_TUNNEL, t - ngm->tunnels);
- vxlan_gpe_tunnel_restack_dpo(t);
+ t->sibling_index = fib_entry_child_add
+ (t->fib_entry_index, FIB_NODE_TYPE_VXLAN_GPE_TUNNEL,
+ t - ngm->tunnels);
+ vxlan_gpe_tunnel_restack_dpo (t);
}
else
- {
- /* Multicast tunnel -
+ {
+ /* Multicast tunnel -
* as the same mcast group can be used for mutiple mcast tunnels
* with different VNIs, create the output fib adjecency only if
* it does not already exist
*/
- fib_protocol_t fp = fib_ip_proto(is_ip6);
-
- if (vtep_addr_ref(&t->remote) == 1)
- {
- fib_node_index_t mfei;
- adj_index_t ai;
- fib_route_path_t path = {
- .frp_proto = fib_proto_to_dpo(fp),
- .frp_addr = zero_addr,
- .frp_sw_if_index = 0xffffffff,
- .frp_fib_index = ~0,
- .frp_weight = 0,
- .frp_flags = FIB_ROUTE_PATH_LOCAL,
- };
- const mfib_prefix_t mpfx = {
- .fp_proto = fp,
- .fp_len = (is_ip6 ? 128 : 32),
- .fp_grp_addr = tun_remote_pfx.fp_addr,
- };
-
- /*
- * Setup the (*,G) to receive traffic on the mcast group
- * - the forwarding interface is for-us
- * - the accepting interface is that from the API
- */
- mfib_table_entry_path_update(t->encap_fib_index,
- &mpfx,
- MFIB_SOURCE_VXLAN_GPE,
- &path,
- MFIB_ITF_FLAG_FORWARD);
-
- path.frp_sw_if_index = a->mcast_sw_if_index;
- path.frp_flags = FIB_ROUTE_PATH_FLAG_NONE;
- mfei = mfib_table_entry_path_update(t->encap_fib_index,
- &mpfx,
- MFIB_SOURCE_VXLAN_GPE,
- &path,
- MFIB_ITF_FLAG_ACCEPT);
-
- /*
- * Create the mcast adjacency to send traffic to the group
- */
- ai = adj_mcast_add_or_lock(fp,
- fib_proto_to_link(fp),
- a->mcast_sw_if_index);
-
- /*
- * create a new end-point
- */
- mcast_shared_add(&t->remote, mfei, ai);
- }
-
- dpo_id_t dpo = DPO_INVALID;
- mcast_shared_t ep = mcast_shared_get(&t->remote);
-
- /* Stack shared mcast remote mac addr rewrite on encap */
- dpo_set (&dpo, DPO_ADJACENCY_MCAST,
- fib_proto_to_dpo(fp),
- ep.mcast_adj_index);
-
- dpo_stack_from_node (encap_index, &t->next_dpo, &dpo);
- dpo_reset (&dpo);
+ fib_protocol_t fp = fib_ip_proto (is_ip6);
+
+ if (vtep_addr_ref (&t->remote) == 1)
+ {
+ fib_node_index_t mfei;
+ adj_index_t ai;
+ fib_route_path_t path = {
+ .frp_proto = fib_proto_to_dpo (fp),
+ .frp_addr = zero_addr,
+ .frp_sw_if_index = 0xffffffff,
+ .frp_fib_index = ~0,
+ .frp_weight = 0,
+ .frp_flags = FIB_ROUTE_PATH_LOCAL,
+ };
+ const mfib_prefix_t mpfx = {
+ .fp_proto = fp,
+ .fp_len = (is_ip6 ? 128 : 32),
+ .fp_grp_addr = tun_remote_pfx.fp_addr,
+ };
+
+ /*
+ * Setup the (*,G) to receive traffic on the mcast group
+ * - the forwarding interface is for-us
+ * - the accepting interface is that from the API
+ */
+ mfib_table_entry_path_update (t->encap_fib_index,
+ &mpfx,
+ MFIB_SOURCE_VXLAN_GPE,
+ &path, MFIB_ITF_FLAG_FORWARD);
+
+ path.frp_sw_if_index = a->mcast_sw_if_index;
+ path.frp_flags = FIB_ROUTE_PATH_FLAG_NONE;
+ mfei = mfib_table_entry_path_update (t->encap_fib_index,
+ &mpfx,
+ MFIB_SOURCE_VXLAN_GPE,
+ &path,
+ MFIB_ITF_FLAG_ACCEPT);
+
+ /*
+ * Create the mcast adjacency to send traffic to the group
+ */
+ ai = adj_mcast_add_or_lock (fp,
+ fib_proto_to_link (fp),
+ a->mcast_sw_if_index);
+
+ /*
+ * create a new end-point
+ */
+ mcast_shared_add (&t->remote, mfei, ai);
+ }
+
+ dpo_id_t dpo = DPO_INVALID;
+ mcast_shared_t ep = mcast_shared_get (&t->remote);
+
+ /* Stack shared mcast remote mac addr rewrite on encap */
+ dpo_set (&dpo, DPO_ADJACENCY_MCAST,
+ fib_proto_to_dpo (fp), ep.mcast_adj_index);
+
+ dpo_stack_from_node (encap_index, &t->next_dpo, &dpo);
+ dpo_reset (&dpo);
flood_class = VNET_FLOOD_CLASS_TUNNEL_MASTER;
}
/* Set vxlan tunnel output node */
hi->output_node_index = encap_index;
- vnet_get_sw_interface (vnet_get_main(), sw_if_index)->flood_class = flood_class;
+ vnet_get_sw_interface (vnet_get_main (), sw_if_index)->flood_class =
+ flood_class;
}
else
{
/* deleting a tunnel: tunnel must exist */
if (!p)
- return VNET_API_ERROR_NO_SUCH_ENTRY;
+ return VNET_API_ERROR_NO_SUCH_ENTRY;
t = pool_elt_at_index (ngm->tunnels, p[0]);
sw_if_index = t->sw_if_index;
- vnet_sw_interface_set_flags (vnm, t->sw_if_index, 0 /* down */);
- vnet_sw_interface_t * si = vnet_get_sw_interface (vnm, t->sw_if_index);
+ vnet_sw_interface_set_flags (vnm, t->sw_if_index, 0 /* down */ );
+ vnet_sw_interface_t *si = vnet_get_sw_interface (vnm, t->sw_if_index);
si->flags |= VNET_SW_INTERFACE_FLAG_HIDDEN;
- set_int_l2_mode(ngm->vlib_main, vnm, MODE_L3, t->sw_if_index, 0, 0, 0, 0);
+ set_int_l2_mode (ngm->vlib_main, vnm, MODE_L3, t->sw_if_index, 0, 0, 0,
+ 0);
vec_add1 (ngm->free_vxlan_gpe_tunnel_hw_if_indices, t->hw_if_index);
ngm->tunnel_index_by_sw_if_index[t->sw_if_index] = ~0;
if (!is_ip6)
- hash_unset (ngm->vxlan4_gpe_tunnel_by_key, key4.as_u64);
+ hash_unset (ngm->vxlan4_gpe_tunnel_by_key, key4.as_u64);
else
hash_unset_key_free (&ngm->vxlan6_gpe_tunnel_by_key, &key6);
- if (!ip46_address_is_multicast(&t->remote))
- {
- vtep_addr_unref(&t->local);
- fib_entry_child_remove(t->fib_entry_index, t->sibling_index);
- fib_table_entry_delete_index(t->fib_entry_index, FIB_SOURCE_RR);
- }
- else if (vtep_addr_unref(&t->remote) == 0)
- {
- mcast_shared_remove(&t->remote);
- }
-
- fib_node_deinit(&t->node);
+ if (!ip46_address_is_multicast (&t->remote))
+ {
+ vtep_addr_unref (&t->local);
+ fib_entry_child_remove (t->fib_entry_index, t->sibling_index);
+ fib_table_entry_delete_index (t->fib_entry_index, FIB_SOURCE_RR);
+ }
+ else if (vtep_addr_unref (&t->remote) == 0)
+ {
+ mcast_shared_remove (&t->remote);
+ }
+
+ fib_node_deinit (&t->node);
vec_free (t->rewrite);
pool_put (ngm->tunnels, t);
}
if (sw_if_indexp)
- *sw_if_indexp = sw_if_index;
+ *sw_if_indexp = sw_if_index;
return 0;
}
static clib_error_t *
vxlan_gpe_add_del_tunnel_command_fn (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input,
+ vlib_cli_command_t * cmd)
{
- unformat_input_t _line_input, * line_input = &_line_input;
+ unformat_input_t _line_input, *line_input = &_line_input;
u8 is_add = 1;
ip46_address_t local, remote;
u8 local_set = 0;
@@ -764,100 +788,103 @@ vxlan_gpe_add_del_tunnel_command_fn (vlib_main_t * vm,
u8 vni_set = 0;
int rv;
u32 tmp;
- vnet_vxlan_gpe_add_del_tunnel_args_t _a, * a = &_a;
+ vnet_vxlan_gpe_add_del_tunnel_args_t _a, *a = &_a;
u32 sw_if_index;
clib_error_t *error = NULL;
/* Get a line of input. */
- if (! unformat_user (input, unformat_line_input, line_input))
+ if (!unformat_user (input, unformat_line_input, line_input))
return 0;
- while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT) {
- if (unformat (line_input, "del"))
- is_add = 0;
- else if (unformat (line_input, "local %U",
- unformat_ip4_address, &local.ip4))
- {
- local_set = 1;
- ipv4_set = 1;
- }
- else if (unformat (line_input, "remote %U",
- unformat_ip4_address, &remote.ip4))
- {
- remote_set = 1;
- ipv4_set = 1;
- }
- else if (unformat (line_input, "local %U",
- unformat_ip6_address, &local.ip6))
- {
- local_set = 1;
- ipv6_set = 1;
- }
- else if (unformat (line_input, "remote %U",
- unformat_ip6_address, &remote.ip6))
+ while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
{
- remote_set = 1;
- ipv6_set = 1;
+ if (unformat (line_input, "del"))
+ is_add = 0;
+ else if (unformat (line_input, "local %U",
+ unformat_ip4_address, &local.ip4))
+ {
+ local_set = 1;
+ ipv4_set = 1;
+ }
+ else if (unformat (line_input, "remote %U",
+ unformat_ip4_address, &remote.ip4))
+ {
+ remote_set = 1;
+ ipv4_set = 1;
+ }
+ else if (unformat (line_input, "local %U",
+ unformat_ip6_address, &local.ip6))
+ {
+ local_set = 1;
+ ipv6_set = 1;
+ }
+ else if (unformat (line_input, "remote %U",
+ unformat_ip6_address, &remote.ip6))
+ {
+ remote_set = 1;
+ ipv6_set = 1;
+ }
+ else if (unformat (line_input, "group %U %U",
+ unformat_ip4_address, &remote.ip4,
+ unformat_vnet_sw_interface,
+ vnet_get_main (), &mcast_sw_if_index))
+ {
+ grp_set = remote_set = 1;
+ ipv4_set = 1;
+ }
+ else if (unformat (line_input, "group %U %U",
+ unformat_ip6_address, &remote.ip6,
+ unformat_vnet_sw_interface,
+ vnet_get_main (), &mcast_sw_if_index))
+ {
+ grp_set = remote_set = 1;
+ ipv6_set = 1;
+ }
+ else if (unformat (line_input, "encap-vrf-id %d", &tmp))
+ {
+ if (ipv6_set)
+ encap_fib_index = fib_table_find (FIB_PROTOCOL_IP6, tmp);
+ else
+ encap_fib_index = fib_table_find (FIB_PROTOCOL_IP4, tmp);
+
+ if (encap_fib_index == ~0)
+ {
+ error =
+ clib_error_return (0, "nonexistent encap fib id %d", tmp);
+ goto done;
+ }
+ }
+ else if (unformat (line_input, "decap-vrf-id %d", &tmp))
+ {
+ if (ipv6_set)
+ decap_fib_index = fib_table_find (FIB_PROTOCOL_IP6, tmp);
+ else
+ decap_fib_index = fib_table_find (FIB_PROTOCOL_IP4, tmp);
+
+ if (decap_fib_index == ~0)
+ {
+ error =
+ clib_error_return (0, "nonexistent decap fib id %d", tmp);
+ goto done;
+ }
+ }
+ else if (unformat (line_input, "vni %d", &vni))
+ vni_set = 1;
+ else if (unformat (line_input, "next-ip4"))
+ protocol = VXLAN_GPE_PROTOCOL_IP4;
+ else if (unformat (line_input, "next-ip6"))
+ protocol = VXLAN_GPE_PROTOCOL_IP6;
+ else if (unformat (line_input, "next-ethernet"))
+ protocol = VXLAN_GPE_PROTOCOL_ETHERNET;
+ else if (unformat (line_input, "next-nsh"))
+ protocol = VXLAN_GPE_PROTOCOL_NSH;
+ else
+ {
+ error = clib_error_return (0, "parse error: '%U'",
+ format_unformat_error, line_input);
+ goto done;
+ }
}
- else if (unformat (line_input, "group %U %U",
- unformat_ip4_address, &remote.ip4,
- unformat_vnet_sw_interface,
- vnet_get_main(), &mcast_sw_if_index))
- {
- grp_set = remote_set = 1;
- ipv4_set = 1;
- }
- else if (unformat (line_input, "group %U %U",
- unformat_ip6_address, &remote.ip6,
- unformat_vnet_sw_interface,
- vnet_get_main(), &mcast_sw_if_index))
- {
- grp_set = remote_set = 1;
- ipv6_set = 1;
- }
- else if (unformat (line_input, "encap-vrf-id %d", &tmp))
- {
- if (ipv6_set)
- encap_fib_index = fib_table_find (FIB_PROTOCOL_IP6, tmp);
- else
- encap_fib_index = fib_table_find (FIB_PROTOCOL_IP4, tmp);
-
- if (encap_fib_index == ~0)
- {
- error = clib_error_return (0, "nonexistent encap fib id %d", tmp);
- goto done;
- }
- }
- else if (unformat (line_input, "decap-vrf-id %d", &tmp))
- {
- if (ipv6_set)
- decap_fib_index = fib_table_find (FIB_PROTOCOL_IP6, tmp);
- else
- decap_fib_index = fib_table_find (FIB_PROTOCOL_IP4, tmp);
-
- if (decap_fib_index == ~0)
- {
- error = clib_error_return (0, "nonexistent decap fib id %d", tmp);
- goto done;
- }
- }
- else if (unformat (line_input, "vni %d", &vni))
- vni_set = 1;
- else if (unformat(line_input, "next-ip4"))
- protocol = VXLAN_GPE_PROTOCOL_IP4;
- else if (unformat(line_input, "next-ip6"))
- protocol = VXLAN_GPE_PROTOCOL_IP6;
- else if (unformat(line_input, "next-ethernet"))
- protocol = VXLAN_GPE_PROTOCOL_ETHERNET;
- else if (unformat(line_input, "next-nsh"))
- protocol = VXLAN_GPE_PROTOCOL_NSH;
- else
- {
- error = clib_error_return (0, "parse error: '%U'",
- format_unformat_error, line_input);
- goto done;
- }
- }
if (local_set == 0)
{
@@ -871,13 +898,13 @@ vxlan_gpe_add_del_tunnel_command_fn (vlib_main_t * vm,
goto done;
}
- if (grp_set && !ip46_address_is_multicast(&remote))
+ if (grp_set && !ip46_address_is_multicast (&remote))
{
error = clib_error_return (0, "tunnel group address not multicast");
goto done;
}
- if (grp_set == 0 && ip46_address_is_multicast(&remote))
+ if (grp_set == 0 && ip46_address_is_multicast (&remote))
{
error = clib_error_return (0, "remote address must be unicast");
goto done;
@@ -894,8 +921,9 @@ vxlan_gpe_add_del_tunnel_command_fn (vlib_main_t * vm,
goto done;
}
- if ((ipv4_set && memcmp(&local.ip4, &remote.ip4, sizeof(local.ip4)) == 0) ||
- (ipv6_set && memcmp(&local.ip6, &remote.ip6, sizeof(local.ip6)) == 0))
+ if ((ipv4_set && memcmp (&local.ip4, &remote.ip4, sizeof (local.ip4)) == 0)
+ || (ipv6_set
+ && memcmp (&local.ip6, &remote.ip6, sizeof (local.ip6)) == 0))
{
error = clib_error_return (0, "src and remote addresses are identical");
goto done;
@@ -914,16 +942,18 @@ vxlan_gpe_add_del_tunnel_command_fn (vlib_main_t * vm,
#define _(x) a->x = x;
foreach_gpe_copy_field;
- if (ipv4_set) foreach_copy_ipv4
- else foreach_copy_ipv6
+ if (ipv4_set)
+ foreach_copy_ipv4
+ else
+ foreach_copy_ipv6
#undef _
+ rv = vnet_vxlan_gpe_add_del_tunnel (a, &sw_if_index);
- rv = vnet_vxlan_gpe_add_del_tunnel (a, &sw_if_index);
-
- switch(rv)
+ switch (rv)
{
case 0:
- vlib_cli_output(vm, "%U\n", format_vnet_sw_if_index_name, vnet_get_main(), sw_if_index);
+ vlib_cli_output (vm, "%U\n", format_vnet_sw_if_index_name,
+ vnet_get_main (), sw_if_index);
break;
case VNET_API_ERROR_INVALID_DECAP_NEXT:
error = clib_error_return (0, "invalid decap-next...");
@@ -939,7 +969,7 @@ vxlan_gpe_add_del_tunnel_command_fn (vlib_main_t * vm,
default:
error = clib_error_return
- (0, "vnet_vxlan_gpe_add_del_tunnel returned %d", rv);
+ (0, "vnet_vxlan_gpe_add_del_tunnel returned %d", rv);
goto done;
}
@@ -992,19 +1022,21 @@ VLIB_CLI_COMMAND (create_vxlan_gpe_tunnel_command, static) = {
*/
static clib_error_t *
show_vxlan_gpe_tunnel_command_fn (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input,
+ vlib_cli_command_t * cmd)
{
- vxlan_gpe_main_t * ngm = &vxlan_gpe_main;
- vxlan_gpe_tunnel_t * t;
+ vxlan_gpe_main_t *ngm = &vxlan_gpe_main;
+ vxlan_gpe_tunnel_t *t;
if (pool_elts (ngm->tunnels) == 0)
vlib_cli_output (vm, "No vxlan-gpe tunnels configured.");
+ /* *INDENT-OFF* */
pool_foreach (t, ngm->tunnels,
({
vlib_cli_output (vm, "%U", format_vxlan_gpe_tunnel, t);
}));
+ /* *INDENT-ON* */
return 0;
}
@@ -1025,9 +1057,8 @@ VLIB_CLI_COMMAND (show_vxlan_gpe_tunnel_command, static) = {
};
/* *INDENT-ON* */
-void vnet_int_vxlan_gpe_bypass_mode (u32 sw_if_index,
- u8 is_ip6,
- u8 is_enable)
+void
+vnet_int_vxlan_gpe_bypass_mode (u32 sw_if_index, u8 is_ip6, u8 is_enable)
{
if (is_ip6)
vnet_feature_enable_disable ("ip6-unicast", "ip6-vxlan-gpe-bypass",
@@ -1040,28 +1071,28 @@ void vnet_int_vxlan_gpe_bypass_mode (u32 sw_if_index,
static clib_error_t *
set_ip_vxlan_gpe_bypass (u32 is_ip6,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input, vlib_cli_command_t * cmd)
{
- unformat_input_t _line_input, * line_input = &_line_input;
- vnet_main_t * vnm = vnet_get_main();
- clib_error_t * error = 0;
+ unformat_input_t _line_input, *line_input = &_line_input;
+ vnet_main_t *vnm = vnet_get_main ();
+ clib_error_t *error = 0;
u32 sw_if_index, is_enable;
sw_if_index = ~0;
is_enable = 1;
- if (! unformat_user (input, unformat_line_input, line_input))
+ if (!unformat_user (input, unformat_line_input, line_input))
return 0;
while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
{
- if (unformat_user (line_input, unformat_vnet_sw_interface, vnm, &sw_if_index))
- ;
+ if (unformat_user
+ (line_input, unformat_vnet_sw_interface, vnm, &sw_if_index))
+ ;
else if (unformat (line_input, "del"))
- is_enable = 0;
+ is_enable = 0;
else
- {
+ {
error = unformat_parse_error (line_input);
goto done;
}
@@ -1076,7 +1107,7 @@ set_ip_vxlan_gpe_bypass (u32 is_ip6,
vnet_int_vxlan_gpe_bypass_mode (sw_if_index, is_ip6, is_enable);
- done:
+done:
unformat_free (line_input);
return error;
@@ -1084,8 +1115,7 @@ set_ip_vxlan_gpe_bypass (u32 is_ip6,
static clib_error_t *
set_ip4_vxlan_gpe_bypass (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input, vlib_cli_command_t * cmd)
{
return set_ip_vxlan_gpe_bypass (0, input, cmd);
}
@@ -1142,8 +1172,7 @@ VLIB_CLI_COMMAND (set_interface_ip_vxlan_gpe_bypass_command, static) = {
static clib_error_t *
set_ip6_vxlan_gpe_bypass (vlib_main_t * vm,
- unformat_input_t * input,
- vlib_cli_command_t * cmd)
+ unformat_input_t * input, vlib_cli_command_t * cmd)
{
return set_ip_vxlan_gpe_bypass (1, input, cmd);
}
@@ -1222,41 +1251,50 @@ VNET_FEATURE_INIT (ip6_vxlan_gpe_bypass, static) =
* @return error
*
*/
-clib_error_t *vxlan_gpe_init (vlib_main_t *vm)
+clib_error_t *
+vxlan_gpe_init (vlib_main_t * vm)
{
vxlan_gpe_main_t *ngm = &vxlan_gpe_main;
- ngm->vnet_main = vnet_get_main();
+ ngm->vnet_main = vnet_get_main ();
ngm->vlib_main = vm;
ngm->vxlan4_gpe_tunnel_by_key
- = hash_create_mem (0, sizeof(vxlan4_gpe_tunnel_key_t), sizeof (uword));
+ = hash_create_mem (0, sizeof (vxlan4_gpe_tunnel_key_t), sizeof (uword));
ngm->vxlan6_gpe_tunnel_by_key
- = hash_create_mem (0, sizeof(vxlan6_gpe_tunnel_key_t), sizeof (uword));
+ = hash_create_mem (0, sizeof (vxlan6_gpe_tunnel_key_t), sizeof (uword));
- ngm->mcast_shared = hash_create_mem(0,
- sizeof(ip46_address_t),
- sizeof(mcast_shared_t));
+ ngm->mcast_shared = hash_create_mem (0,
+ sizeof (ip46_address_t),
+ sizeof (mcast_shared_t));
udp_register_dst_port (vm, UDP_DST_PORT_VXLAN_GPE,
- vxlan4_gpe_input_node.index, 1 /* is_ip4 */);
+ vxlan4_gpe_input_node.index, 1 /* is_ip4 */ );
udp_register_dst_port (vm, UDP_DST_PORT_VXLAN6_GPE,
- vxlan6_gpe_input_node.index, 0 /* is_ip4 */);
+ vxlan6_gpe_input_node.index, 0 /* is_ip4 */ );
/* Register the list of standard decap protocols supported */
vxlan_gpe_register_decap_protocol (VXLAN_GPE_PROTOCOL_IP4,
- VXLAN_GPE_INPUT_NEXT_IP4_INPUT);
+ VXLAN_GPE_INPUT_NEXT_IP4_INPUT);
vxlan_gpe_register_decap_protocol (VXLAN_GPE_PROTOCOL_IP6,
- VXLAN_GPE_INPUT_NEXT_IP6_INPUT);
+ VXLAN_GPE_INPUT_NEXT_IP6_INPUT);
vxlan_gpe_register_decap_protocol (VXLAN_GPE_PROTOCOL_ETHERNET,
- VXLAN_GPE_INPUT_NEXT_ETHERNET_INPUT);
+ VXLAN_GPE_INPUT_NEXT_ETHERNET_INPUT);
- fib_node_register_type(FIB_NODE_TYPE_VXLAN_GPE_TUNNEL, &vxlan_gpe_vft);
+ fib_node_register_type (FIB_NODE_TYPE_VXLAN_GPE_TUNNEL, &vxlan_gpe_vft);
return 0;
}
-VLIB_INIT_FUNCTION(vxlan_gpe_init);
+VLIB_INIT_FUNCTION (vxlan_gpe_init);
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */