aboutsummaryrefslogtreecommitdiffstats
path: root/src/vnet/vxlan-gbp
diff options
context:
space:
mode:
authorFilip Tehlar <ftehlar@cisco.com>2019-03-05 03:01:43 -0800
committerDamjan Marion <dmarion@me.com>2019-03-06 11:40:53 +0000
commite1714d3ba2fa91db5fcede631e2acafca899604c (patch)
tree43a634884f0d64083036a9a2b020ee847861d9f3 /src/vnet/vxlan-gbp
parentf6266b56ecd6bca3074886b37d48cf5c20f37fa3 (diff)
vxlan*: migrate old MULTIARCH macros to VLIB_NODE_FN
Change-Id: Ide23bb3d82024118214902850821a8184fe65dfc Signed-off-by: Filip Tehlar <ftehlar@cisco.com>
Diffstat (limited to 'src/vnet/vxlan-gbp')
-rw-r--r--src/vnet/vxlan-gbp/decap.c41
-rw-r--r--src/vnet/vxlan-gbp/encap.c20
2 files changed, 24 insertions, 37 deletions
diff --git a/src/vnet/vxlan-gbp/decap.c b/src/vnet/vxlan-gbp/decap.c
index 1a8ece702a6..39dac4169b4 100644
--- a/src/vnet/vxlan-gbp/decap.c
+++ b/src/vnet/vxlan-gbp/decap.c
@@ -19,9 +19,6 @@
#include <vnet/pg/pg.h>
#include <vnet/vxlan-gbp/vxlan_gbp.h>
-vlib_node_registration_t vxlan4_gbp_input_node;
-vlib_node_registration_t vxlan6_gbp_input_node;
-
typedef struct
{
u32 next_index;
@@ -494,16 +491,16 @@ vxlan_gbp_input (vlib_main_t * vm,
return from_frame->n_vectors;
}
-static uword
-vxlan4_gbp_input (vlib_main_t * vm,
- vlib_node_runtime_t * node, vlib_frame_t * from_frame)
+VLIB_NODE_FN (vxlan4_gbp_input_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * from_frame)
{
return vxlan_gbp_input (vm, node, from_frame, /* is_ip4 */ 1);
}
-static uword
-vxlan6_gbp_input (vlib_main_t * vm,
- vlib_node_runtime_t * node, vlib_frame_t * from_frame)
+VLIB_NODE_FN (vxlan6_gbp_input_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * from_frame)
{
return vxlan_gbp_input (vm, node, from_frame, /* is_ip4 */ 0);
}
@@ -518,7 +515,6 @@ static char *vxlan_gbp_error_strings[] = {
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (vxlan4_gbp_input_node) =
{
- .function = vxlan4_gbp_input,
.name = "vxlan4-gbp-input",
.vector_size = sizeof (u32),
.n_errors = VXLAN_GBP_N_ERROR,
@@ -531,11 +527,9 @@ VLIB_REGISTER_NODE (vxlan4_gbp_input_node) =
#undef _
},
};
-VLIB_NODE_FUNCTION_MULTIARCH (vxlan4_gbp_input_node, vxlan4_gbp_input)
VLIB_REGISTER_NODE (vxlan6_gbp_input_node) =
{
- .function = vxlan6_gbp_input,
.name = "vxlan6-gbp-input",
.vector_size = sizeof (u32),
.n_errors = VXLAN_GBP_N_ERROR,
@@ -548,7 +542,6 @@ VLIB_REGISTER_NODE (vxlan6_gbp_input_node) =
},
.format_trace = format_vxlan_gbp_rx_trace,
};
-VLIB_NODE_FUNCTION_MULTIARCH (vxlan6_gbp_input_node, vxlan6_gbp_input)
/* *INDENT-ON* */
typedef enum
@@ -957,9 +950,9 @@ ip_vxlan_gbp_bypass_inline (vlib_main_t * vm,
return frame->n_vectors;
}
-static uword
-ip4_vxlan_gbp_bypass (vlib_main_t * vm,
- vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (ip4_vxlan_gbp_bypass_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * frame)
{
return ip_vxlan_gbp_bypass_inline (vm, node, frame, /* is_ip4 */ 1);
}
@@ -967,7 +960,6 @@ ip4_vxlan_gbp_bypass (vlib_main_t * vm,
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (ip4_vxlan_gbp_bypass_node) =
{
- .function = ip4_vxlan_gbp_bypass,
.name = "ip4-vxlan-gbp-bypass",
.vector_size = sizeof (u32),
.n_next_nodes = IP_VXLAN_GBP_BYPASS_N_NEXT,
@@ -978,10 +970,9 @@ VLIB_REGISTER_NODE (ip4_vxlan_gbp_bypass_node) =
.format_buffer = format_ip4_header,
.format_trace = format_ip4_forward_next_trace,
};
-
-VLIB_NODE_FUNCTION_MULTIARCH (ip4_vxlan_gbp_bypass_node, ip4_vxlan_gbp_bypass)
/* *INDENT-ON* */
+#ifndef CLIB_MARCH_VARIANT
/* Dummy init function to get us linked in. */
clib_error_t *
ip4_vxlan_gbp_bypass_init (vlib_main_t * vm)
@@ -990,10 +981,11 @@ ip4_vxlan_gbp_bypass_init (vlib_main_t * vm)
}
VLIB_INIT_FUNCTION (ip4_vxlan_gbp_bypass_init);
+#endif /* CLIB_MARCH_VARIANT */
-static uword
-ip6_vxlan_gbp_bypass (vlib_main_t * vm,
- vlib_node_runtime_t * node, vlib_frame_t * frame)
+VLIB_NODE_FN (ip6_vxlan_gbp_bypass_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * frame)
{
return ip_vxlan_gbp_bypass_inline (vm, node, frame, /* is_ip4 */ 0);
}
@@ -1001,7 +993,6 @@ ip6_vxlan_gbp_bypass (vlib_main_t * vm,
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (ip6_vxlan_gbp_bypass_node) =
{
- .function = ip6_vxlan_gbp_bypass,
.name = "ip6-vxlan-gbp-bypass",
.vector_size = sizeof (u32),
.n_next_nodes = IP_VXLAN_GBP_BYPASS_N_NEXT,
@@ -1012,10 +1003,9 @@ VLIB_REGISTER_NODE (ip6_vxlan_gbp_bypass_node) =
.format_buffer = format_ip6_header,
.format_trace = format_ip6_forward_next_trace,
};
-
-VLIB_NODE_FUNCTION_MULTIARCH (ip6_vxlan_gbp_bypass_node, ip6_vxlan_gbp_bypass)
/* *INDENT-ON* */
+#ifndef CLIB_MARCH_VARIANT
/* Dummy init function to get us linked in. */
clib_error_t *
ip6_vxlan_gbp_bypass_init (vlib_main_t * vm)
@@ -1024,6 +1014,7 @@ ip6_vxlan_gbp_bypass_init (vlib_main_t * vm)
}
VLIB_INIT_FUNCTION (ip6_vxlan_gbp_bypass_init);
+#endif /* CLIB_MARCH_VARIANT */
/*
* fd.io coding-style-patch-verification: ON
diff --git a/src/vnet/vxlan-gbp/encap.c b/src/vnet/vxlan-gbp/encap.c
index f8fc9b4b998..2fe3fa8d437 100644
--- a/src/vnet/vxlan-gbp/encap.c
+++ b/src/vnet/vxlan-gbp/encap.c
@@ -53,6 +53,7 @@ typedef struct
u8 flags;
} vxlan_gbp_encap_trace_t;
+#ifndef CLIB_MARCH_VARIANT
u8 *
format_vxlan_gbp_encap_trace (u8 * s, va_list * args)
{
@@ -66,6 +67,7 @@ format_vxlan_gbp_encap_trace (u8 * s, va_list * args)
t->tunnel_index, t->vni, t->sclass, t->flags);
return s;
}
+#endif /* CLIB_MARCH_VARIANT */
always_inline uword
vxlan_gbp_encap_inline (vlib_main_t * vm,
@@ -488,9 +490,9 @@ vxlan_gbp_encap_inline (vlib_main_t * vm,
return from_frame->n_vectors;
}
-static uword
-vxlan4_gbp_encap (vlib_main_t * vm,
- vlib_node_runtime_t * node, vlib_frame_t * from_frame)
+VLIB_NODE_FN (vxlan4_gbp_encap_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * from_frame)
{
/* Disable chksum offload as setup overhead in tx node is not worthwhile
for ip4 header checksum only, unless udp checksum is also required */
@@ -498,9 +500,9 @@ vxlan4_gbp_encap (vlib_main_t * vm,
/* csum_offload */ 0);
}
-static uword
-vxlan6_gbp_encap (vlib_main_t * vm,
- vlib_node_runtime_t * node, vlib_frame_t * from_frame)
+VLIB_NODE_FN (vxlan6_gbp_encap_node) (vlib_main_t * vm,
+ vlib_node_runtime_t * node,
+ vlib_frame_t * from_frame)
{
/* Enable checksum offload for ip6 as udp checksum is mandatory, */
return vxlan_gbp_encap_inline (vm, node, from_frame, /* is_ip4 */ 0,
@@ -510,7 +512,6 @@ vxlan6_gbp_encap (vlib_main_t * vm,
/* *INDENT-OFF* */
VLIB_REGISTER_NODE (vxlan4_gbp_encap_node) =
{
- .function = vxlan4_gbp_encap,
.name = "vxlan4-gbp-encap",
.vector_size = sizeof (u32),
.format_trace = format_vxlan_gbp_encap_trace,
@@ -523,11 +524,8 @@ VLIB_REGISTER_NODE (vxlan4_gbp_encap_node) =
},
};
-VLIB_NODE_FUNCTION_MULTIARCH (vxlan4_gbp_encap_node, vxlan4_gbp_encap)
-
VLIB_REGISTER_NODE (vxlan6_gbp_encap_node) =
{
- .function = vxlan6_gbp_encap,
.name = "vxlan6-gbp-encap",
.vector_size = sizeof (u32),
.format_trace = format_vxlan_gbp_encap_trace,
@@ -539,8 +537,6 @@ VLIB_REGISTER_NODE (vxlan6_gbp_encap_node) =
[VXLAN_GBP_ENCAP_NEXT_DROP] = "error-drop",
},
};
-
-VLIB_NODE_FUNCTION_MULTIARCH (vxlan6_gbp_encap_node, vxlan6_gbp_encap)
/* *INDENT-ON* */
/*