aboutsummaryrefslogtreecommitdiffstats
path: root/vnet/vnet/map/ip6_map.c
diff options
context:
space:
mode:
authorKeith Burns (alagalah) <alagalah@gmail.com>2016-08-07 08:43:18 -0700
committerDave Barach <openvpp@barachs.net>2016-08-08 15:32:42 +0000
commit06e3d075760be43dbd61385d3c7861cd5fe44af2 (patch)
treed2d5477cc0127c858a1f4b43021a4b0c635c5d00 /vnet/vnet/map/ip6_map.c
parentba868bb7898edc46ad5f8cd4925af6c8b92e7c89 (diff)
VPP-263 - Coding standards cleanup - vnet/vnet/map
Change-Id: I401124320a21e30b5cdf3684d36890e51a2fcb6c Signed-off-by: Keith Burns (alagalah) <alagalah@gmail.com>
Diffstat (limited to 'vnet/vnet/map/ip6_map.c')
-rw-r--r--vnet/vnet/map/ip6_map.c1697
1 files changed, 981 insertions, 716 deletions
diff --git a/vnet/vnet/map/ip6_map.c b/vnet/vnet/map/ip6_map.c
index 178b7c3fb9e..b51e8358d41 100644
--- a/vnet/vnet/map/ip6_map.c
+++ b/vnet/vnet/map/ip6_map.c
@@ -16,7 +16,8 @@
#include "../ip/ip_frag.h"
-enum ip6_map_next_e {
+enum ip6_map_next_e
+{
IP6_MAP_NEXT_IP4_LOOKUP,
#ifdef MAP_SKIP_IP6_LOOKUP
IP6_MAP_NEXT_IP4_REWRITE,
@@ -31,20 +32,23 @@ enum ip6_map_next_e {
IP6_MAP_N_NEXT,
};
-enum ip6_map_ip6_reass_next_e {
+enum ip6_map_ip6_reass_next_e
+{
IP6_MAP_IP6_REASS_NEXT_IP6_MAP,
IP6_MAP_IP6_REASS_NEXT_DROP,
IP6_MAP_IP6_REASS_N_NEXT,
};
-enum ip6_map_ip4_reass_next_e {
+enum ip6_map_ip4_reass_next_e
+{
IP6_MAP_IP4_REASS_NEXT_IP4_LOOKUP,
IP6_MAP_IP4_REASS_NEXT_IP4_FRAGMENT,
IP6_MAP_IP4_REASS_NEXT_DROP,
IP6_MAP_IP4_REASS_N_NEXT,
};
-enum ip6_icmp_relay_next_e {
+enum ip6_icmp_relay_next_e
+{
IP6_ICMP_RELAY_NEXT_IP4_LOOKUP,
IP6_ICMP_RELAY_NEXT_DROP,
IP6_ICMP_RELAY_N_NEXT,
@@ -54,91 +58,113 @@ vlib_node_registration_t ip6_map_ip4_reass_node;
vlib_node_registration_t ip6_map_ip6_reass_node;
static vlib_node_registration_t ip6_map_icmp_relay_node;
-typedef struct {
+typedef struct
+{
u32 map_domain_index;
u16 port;
u8 cached;
} map_ip6_map_ip4_reass_trace_t;
u8 *
-format_ip6_map_ip4_reass_trace (u8 *s, va_list *args)
+format_ip6_map_ip4_reass_trace (u8 * s, va_list * args)
{
- CLIB_UNUSED(vlib_main_t *vm) = va_arg (*args, vlib_main_t *);
- CLIB_UNUSED(vlib_node_t *node) = va_arg (*args, vlib_node_t *);
- map_ip6_map_ip4_reass_trace_t *t = va_arg (*args, map_ip6_map_ip4_reass_trace_t *);
- return format(s, "MAP domain index: %d L4 port: %u Status: %s", t->map_domain_index,
- t->port, t->cached?"cached":"forwarded");
+ CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
+ CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
+ map_ip6_map_ip4_reass_trace_t *t =
+ va_arg (*args, map_ip6_map_ip4_reass_trace_t *);
+ return format (s, "MAP domain index: %d L4 port: %u Status: %s",
+ t->map_domain_index, t->port,
+ t->cached ? "cached" : "forwarded");
}
-typedef struct {
+typedef struct
+{
u16 offset;
u16 frag_len;
u8 out;
} map_ip6_map_ip6_reass_trace_t;
u8 *
-format_ip6_map_ip6_reass_trace (u8 *s, va_list *args)
+format_ip6_map_ip6_reass_trace (u8 * s, va_list * args)
{
- CLIB_UNUSED(vlib_main_t *vm) = va_arg (*args, vlib_main_t *);
- CLIB_UNUSED(vlib_node_t *node) = va_arg (*args, vlib_node_t *);
- map_ip6_map_ip6_reass_trace_t *t = va_arg (*args, map_ip6_map_ip6_reass_trace_t *);
- return format(s, "Offset: %d Fragment length: %d Status: %s", t->offset, t->frag_len, t->out?"out":"in");
+ CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
+ CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
+ map_ip6_map_ip6_reass_trace_t *t =
+ va_arg (*args, map_ip6_map_ip6_reass_trace_t *);
+ return format (s, "Offset: %d Fragment length: %d Status: %s", t->offset,
+ t->frag_len, t->out ? "out" : "in");
}
/*
* ip6_map_sec_check
*/
static_always_inline bool
-ip6_map_sec_check (map_domain_t *d, u16 port, ip4_header_t *ip4, ip6_header_t *ip6)
+ip6_map_sec_check (map_domain_t * d, u16 port, ip4_header_t * ip4,
+ ip6_header_t * ip6)
{
- u16 sp4 = clib_net_to_host_u16(port);
- u32 sa4 = clib_net_to_host_u32(ip4->src_address.as_u32);
- u64 sal6 = map_get_pfx(d, sa4, sp4);
- u64 sar6 = map_get_sfx(d, sa4, sp4);
-
- if (PREDICT_FALSE(sal6 != clib_net_to_host_u64(ip6->src_address.as_u64[0]) ||
- sar6 != clib_net_to_host_u64(ip6->src_address.as_u64[1])))
+ u16 sp4 = clib_net_to_host_u16 (port);
+ u32 sa4 = clib_net_to_host_u32 (ip4->src_address.as_u32);
+ u64 sal6 = map_get_pfx (d, sa4, sp4);
+ u64 sar6 = map_get_sfx (d, sa4, sp4);
+
+ if (PREDICT_FALSE
+ (sal6 != clib_net_to_host_u64 (ip6->src_address.as_u64[0])
+ || sar6 != clib_net_to_host_u64 (ip6->src_address.as_u64[1])))
return (false);
return (true);
}
static_always_inline void
-ip6_map_security_check (map_domain_t *d, ip4_header_t *ip4, ip6_header_t *ip6, u32 *next, u8 *error)
+ip6_map_security_check (map_domain_t * d, ip4_header_t * ip4,
+ ip6_header_t * ip6, u32 * next, u8 * error)
{
map_main_t *mm = &map_main;
- if (d->ea_bits_len || d->rules) {
- if (d->psid_length > 0) {
- if (!ip4_is_fragment(ip4)) {
- u16 port = ip4_map_get_port(ip4, MAP_SENDER);
- if (port) {
- if (mm->sec_check)
- *error = ip6_map_sec_check(d, port, ip4, ip6) ? MAP_ERROR_NONE : MAP_ERROR_DECAP_SEC_CHECK;
- } else {
- *error = MAP_ERROR_BAD_PROTOCOL;
+ if (d->ea_bits_len || d->rules)
+ {
+ if (d->psid_length > 0)
+ {
+ if (!ip4_is_fragment (ip4))
+ {
+ u16 port = ip4_map_get_port (ip4, MAP_SENDER);
+ if (port)
+ {
+ if (mm->sec_check)
+ *error =
+ ip6_map_sec_check (d, port, ip4,
+ ip6) ? MAP_ERROR_NONE :
+ MAP_ERROR_DECAP_SEC_CHECK;
+ }
+ else
+ {
+ *error = MAP_ERROR_BAD_PROTOCOL;
+ }
+ }
+ else
+ {
+ *next = mm->sec_check_frag ? IP6_MAP_NEXT_IP4_REASS : *next;
+ }
}
- } else {
- *next = mm->sec_check_frag ? IP6_MAP_NEXT_IP4_REASS : *next;
- }
}
- }
}
static_always_inline bool
-ip6_map_ip4_lookup_bypass (vlib_buffer_t *p0, ip4_header_t *ip)
+ip6_map_ip4_lookup_bypass (vlib_buffer_t * p0, ip4_header_t * ip)
{
#ifdef MAP_SKIP_IP6_LOOKUP
map_main_t *mm = &map_main;
u32 adj_index0 = mm->adj4_index;
- if (adj_index0 > 0) {
- ip_lookup_main_t *lm4 = &ip4_main.lookup_main;
- ip_adjacency_t *adj = ip_get_adjacency(lm4, mm->adj4_index);
- if (adj->n_adj > 1) {
- u32 hash_c0 = ip4_compute_flow_hash(ip, IP_FLOW_HASH_DEFAULT);
- adj_index0 += (hash_c0 & (adj->n_adj - 1));
+ if (adj_index0 > 0)
+ {
+ ip_lookup_main_t *lm4 = &ip4_main.lookup_main;
+ ip_adjacency_t *adj = ip_get_adjacency (lm4, mm->adj4_index);
+ if (adj->n_adj > 1)
+ {
+ u32 hash_c0 = ip4_compute_flow_hash (ip, IP_FLOW_HASH_DEFAULT);
+ adj_index0 += (hash_c0 & (adj->n_adj - 1));
+ }
+ vnet_buffer (p0)->ip.adj_index[VLIB_TX] = adj_index0;
+ return (true);
}
- vnet_buffer(p0)->ip.adj_index[VLIB_TX] = adj_index0;
- return (true);
- }
#endif
return (false);
}
@@ -147,376 +173,503 @@ ip6_map_ip4_lookup_bypass (vlib_buffer_t *p0, ip4_header_t *ip)
* ip6_map
*/
static uword
-ip6_map (vlib_main_t *vm,
- vlib_node_runtime_t *node,
- vlib_frame_t *frame)
+ip6_map (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame)
{
u32 n_left_from, *from, next_index, *to_next, n_left_to_next;
- vlib_node_runtime_t *error_node = vlib_node_get_runtime(vm, ip6_map_node.index);
+ vlib_node_runtime_t *error_node =
+ vlib_node_get_runtime (vm, ip6_map_node.index);
map_main_t *mm = &map_main;
vlib_combined_counter_main_t *cm = mm->domain_counters;
- u32 cpu_index = os_get_cpu_number();
+ u32 cpu_index = os_get_cpu_number ();
- from = vlib_frame_vector_args(frame);
+ from = vlib_frame_vector_args (frame);
n_left_from = frame->n_vectors;
next_index = node->cached_next_index;
- while (n_left_from > 0) {
- vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next);
-
- /* Dual loop */
- while (n_left_from >= 4 && n_left_to_next >= 2) {
- u32 pi0, pi1;
- vlib_buffer_t *p0, *p1;
- u8 error0 = MAP_ERROR_NONE;
- u8 error1 = MAP_ERROR_NONE;
- map_domain_t *d0 = 0, *d1 = 0;
- ip4_header_t *ip40, *ip41;
- ip6_header_t *ip60, *ip61;
- u16 port0 = 0, port1 = 0;
- u32 map_domain_index0 = ~0, map_domain_index1 = ~0;
- u32 next0 = IP6_MAP_NEXT_IP4_LOOKUP;
- u32 next1 = IP6_MAP_NEXT_IP4_LOOKUP;
-
- /* Prefetch next iteration. */
- {
- vlib_buffer_t *p2, *p3;
-
- p2 = vlib_get_buffer(vm, from[2]);
- p3 = vlib_get_buffer(vm, from[3]);
-
- vlib_prefetch_buffer_header(p2, LOAD);
- vlib_prefetch_buffer_header(p3, LOAD);
-
- /* IPv6 + IPv4 header + 8 bytes of ULP */
- CLIB_PREFETCH(p2->data, 68, LOAD);
- CLIB_PREFETCH(p3->data, 68, LOAD);
- }
-
- pi0 = to_next[0] = from[0];
- pi1 = to_next[1] = from[1];
- from += 2;
- n_left_from -= 2;
- to_next +=2;
- n_left_to_next -= 2;
-
- p0 = vlib_get_buffer(vm, pi0);
- p1 = vlib_get_buffer(vm, pi1);
- ip60 = vlib_buffer_get_current(p0);
- ip61 = vlib_buffer_get_current(p1);
- vlib_buffer_advance(p0, sizeof(ip6_header_t));
- vlib_buffer_advance(p1, sizeof(ip6_header_t));
- ip40 = vlib_buffer_get_current(p0);
- ip41 = vlib_buffer_get_current(p1);
-
- /*
- * Encapsulated IPv4 packet
- * - IPv4 fragmented -> Pass to virtual reassembly unless security check disabled
- * - Lookup/Rewrite or Fragment node in case of packet > MTU
- * Fragmented IPv6 packet
- * ICMP IPv6 packet
- * - Error -> Pass to ICMPv6/ICMPv4 relay
- * - Info -> Pass to IPv6 local
- * Anything else -> drop
- */
- if (PREDICT_TRUE(ip60->protocol == IP_PROTOCOL_IP_IN_IP && clib_net_to_host_u16(ip60->payload_length) > 20)) {
- d0 = ip6_map_get_domain(vnet_buffer(p0)->ip.adj_index[VLIB_TX], (ip4_address_t *)&ip40->src_address.as_u32,
- &map_domain_index0, &error0);
- } else if (ip60->protocol == IP_PROTOCOL_ICMP6 &&
- clib_net_to_host_u16(ip60->payload_length) > sizeof(icmp46_header_t)) {
- icmp46_header_t *icmp = (void *)(ip60 + 1);
- next0 = (icmp->type == ICMP6_echo_request || icmp->type == ICMP6_echo_reply) ?
- IP6_MAP_NEXT_IP6_LOCAL : IP6_MAP_NEXT_IP6_ICMP_RELAY;
- } else if (ip60->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION) {
- next0 = IP6_MAP_NEXT_IP6_REASS;
- } else {
- error0 = MAP_ERROR_BAD_PROTOCOL;
- }
- if (PREDICT_TRUE(ip61->protocol == IP_PROTOCOL_IP_IN_IP && clib_net_to_host_u16(ip61->payload_length) > 20)) {
- d1 = ip6_map_get_domain(vnet_buffer(p1)->ip.adj_index[VLIB_TX], (ip4_address_t *)&ip41->src_address.as_u32,
- &map_domain_index1, &error1);
- } else if (ip61->protocol == IP_PROTOCOL_ICMP6 &&
- clib_net_to_host_u16(ip61->payload_length) > sizeof(icmp46_header_t)) {
- icmp46_header_t *icmp = (void *)(ip61 + 1);
- next1 = (icmp->type == ICMP6_echo_request || icmp->type == ICMP6_echo_reply) ?
- IP6_MAP_NEXT_IP6_LOCAL : IP6_MAP_NEXT_IP6_ICMP_RELAY;
- } else if (ip61->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION) {
- next1 = IP6_MAP_NEXT_IP6_REASS;
- } else {
- error1 = MAP_ERROR_BAD_PROTOCOL;
- }
-
- if (d0) {
- /* MAP inbound security check */
- ip6_map_security_check(d0, ip40, ip60, &next0, &error0);
-
- if (PREDICT_TRUE(error0 == MAP_ERROR_NONE &&
- next0 == IP6_MAP_NEXT_IP4_LOOKUP)) {
- if (PREDICT_FALSE(d0->mtu && (clib_host_to_net_u16(ip40->length) > d0->mtu))) {
- vnet_buffer(p0)->ip_frag.header_offset = 0;
- vnet_buffer(p0)->ip_frag.flags = 0;
- vnet_buffer(p0)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP;
- vnet_buffer(p0)->ip_frag.mtu = d0->mtu;
- next0 = IP6_MAP_NEXT_IP4_FRAGMENT;
- } else {
- next0 = ip6_map_ip4_lookup_bypass(p0, ip40) ? IP6_MAP_NEXT_IP4_REWRITE : next0;
- }
- vlib_increment_combined_counter(cm + MAP_DOMAIN_COUNTER_RX, cpu_index, map_domain_index0, 1,
- clib_net_to_host_u16(ip40->length));
- }
- }
- if (d1) {
- /* MAP inbound security check */
- ip6_map_security_check(d1, ip41, ip61, &next1, &error1);
-
- if (PREDICT_TRUE(error1 == MAP_ERROR_NONE &&
- next1 == IP6_MAP_NEXT_IP4_LOOKUP)) {
- if (PREDICT_FALSE(d1->mtu && (clib_host_to_net_u16(ip41->length) > d1->mtu))) {
- vnet_buffer(p1)->ip_frag.header_offset = 0;
- vnet_buffer(p1)->ip_frag.flags = 0;
- vnet_buffer(p1)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP;
- vnet_buffer(p1)->ip_frag.mtu = d1->mtu;
- next1 = IP6_MAP_NEXT_IP4_FRAGMENT;
- } else {
- next1 = ip6_map_ip4_lookup_bypass(p1, ip41) ? IP6_MAP_NEXT_IP4_REWRITE : next1;
+ while (n_left_from > 0)
+ {
+ vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+
+ /* Dual loop */
+ while (n_left_from >= 4 && n_left_to_next >= 2)
+ {
+ u32 pi0, pi1;
+ vlib_buffer_t *p0, *p1;
+ u8 error0 = MAP_ERROR_NONE;
+ u8 error1 = MAP_ERROR_NONE;
+ map_domain_t *d0 = 0, *d1 = 0;
+ ip4_header_t *ip40, *ip41;
+ ip6_header_t *ip60, *ip61;
+ u16 port0 = 0, port1 = 0;
+ u32 map_domain_index0 = ~0, map_domain_index1 = ~0;
+ u32 next0 = IP6_MAP_NEXT_IP4_LOOKUP;
+ u32 next1 = IP6_MAP_NEXT_IP4_LOOKUP;
+
+ /* Prefetch next iteration. */
+ {
+ vlib_buffer_t *p2, *p3;
+
+ p2 = vlib_get_buffer (vm, from[2]);
+ p3 = vlib_get_buffer (vm, from[3]);
+
+ vlib_prefetch_buffer_header (p2, LOAD);
+ vlib_prefetch_buffer_header (p3, LOAD);
+
+ /* IPv6 + IPv4 header + 8 bytes of ULP */
+ CLIB_PREFETCH (p2->data, 68, LOAD);
+ CLIB_PREFETCH (p3->data, 68, LOAD);
}
- vlib_increment_combined_counter(cm + MAP_DOMAIN_COUNTER_RX, cpu_index, map_domain_index1, 1,
- clib_net_to_host_u16(ip41->length));
+
+ pi0 = to_next[0] = from[0];
+ pi1 = to_next[1] = from[1];
+ from += 2;
+ n_left_from -= 2;
+ to_next += 2;
+ n_left_to_next -= 2;
+
+ p0 = vlib_get_buffer (vm, pi0);
+ p1 = vlib_get_buffer (vm, pi1);
+ ip60 = vlib_buffer_get_current (p0);
+ ip61 = vlib_buffer_get_current (p1);
+ vlib_buffer_advance (p0, sizeof (ip6_header_t));
+ vlib_buffer_advance (p1, sizeof (ip6_header_t));
+ ip40 = vlib_buffer_get_current (p0);
+ ip41 = vlib_buffer_get_current (p1);
+
+ /*
+ * Encapsulated IPv4 packet
+ * - IPv4 fragmented -> Pass to virtual reassembly unless security check disabled
+ * - Lookup/Rewrite or Fragment node in case of packet > MTU
+ * Fragmented IPv6 packet
+ * ICMP IPv6 packet
+ * - Error -> Pass to ICMPv6/ICMPv4 relay
+ * - Info -> Pass to IPv6 local
+ * Anything else -> drop
+ */
+ if (PREDICT_TRUE
+ (ip60->protocol == IP_PROTOCOL_IP_IN_IP
+ && clib_net_to_host_u16 (ip60->payload_length) > 20))
+ {
+ d0 =
+ ip6_map_get_domain (vnet_buffer (p0)->ip.adj_index[VLIB_TX],
+ (ip4_address_t *) & ip40->src_address.
+ as_u32, &map_domain_index0, &error0);
+ }
+ else if (ip60->protocol == IP_PROTOCOL_ICMP6 &&
+ clib_net_to_host_u16 (ip60->payload_length) >
+ sizeof (icmp46_header_t))
+ {
+ icmp46_header_t *icmp = (void *) (ip60 + 1);
+ next0 = (icmp->type == ICMP6_echo_request
+ || icmp->type ==
+ ICMP6_echo_reply) ? IP6_MAP_NEXT_IP6_LOCAL :
+ IP6_MAP_NEXT_IP6_ICMP_RELAY;
+ }
+ else if (ip60->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION)
+ {
+ next0 = IP6_MAP_NEXT_IP6_REASS;
+ }
+ else
+ {
+ error0 = MAP_ERROR_BAD_PROTOCOL;
+ }
+ if (PREDICT_TRUE
+ (ip61->protocol == IP_PROTOCOL_IP_IN_IP
+ && clib_net_to_host_u16 (ip61->payload_length) > 20))
+ {
+ d1 =
+ ip6_map_get_domain (vnet_buffer (p1)->ip.adj_index[VLIB_TX],
+ (ip4_address_t *) & ip41->src_address.
+ as_u32, &map_domain_index1, &error1);
+ }
+ else if (ip61->protocol == IP_PROTOCOL_ICMP6 &&
+ clib_net_to_host_u16 (ip61->payload_length) >
+ sizeof (icmp46_header_t))
+ {
+ icmp46_header_t *icmp = (void *) (ip61 + 1);
+ next1 = (icmp->type == ICMP6_echo_request
+ || icmp->type ==
+ ICMP6_echo_reply) ? IP6_MAP_NEXT_IP6_LOCAL :
+ IP6_MAP_NEXT_IP6_ICMP_RELAY;
+ }
+ else if (ip61->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION)
+ {
+ next1 = IP6_MAP_NEXT_IP6_REASS;
+ }
+ else
+ {
+ error1 = MAP_ERROR_BAD_PROTOCOL;
+ }
+
+ if (d0)
+ {
+ /* MAP inbound security check */
+ ip6_map_security_check (d0, ip40, ip60, &next0, &error0);
+
+ if (PREDICT_TRUE (error0 == MAP_ERROR_NONE &&
+ next0 == IP6_MAP_NEXT_IP4_LOOKUP))
+ {
+ if (PREDICT_FALSE
+ (d0->mtu
+ && (clib_host_to_net_u16 (ip40->length) > d0->mtu)))
+ {
+ vnet_buffer (p0)->ip_frag.header_offset = 0;
+ vnet_buffer (p0)->ip_frag.flags = 0;
+ vnet_buffer (p0)->ip_frag.next_index =
+ IP4_FRAG_NEXT_IP4_LOOKUP;
+ vnet_buffer (p0)->ip_frag.mtu = d0->mtu;
+ next0 = IP6_MAP_NEXT_IP4_FRAGMENT;
+ }
+ else
+ {
+ next0 =
+ ip6_map_ip4_lookup_bypass (p0,
+ ip40) ?
+ IP6_MAP_NEXT_IP4_REWRITE : next0;
+ }
+ vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_RX,
+ cpu_index,
+ map_domain_index0, 1,
+ clib_net_to_host_u16
+ (ip40->length));
+ }
+ }
+ if (d1)
+ {
+ /* MAP inbound security check */
+ ip6_map_security_check (d1, ip41, ip61, &next1, &error1);
+
+ if (PREDICT_TRUE (error1 == MAP_ERROR_NONE &&
+ next1 == IP6_MAP_NEXT_IP4_LOOKUP))
+ {
+ if (PREDICT_FALSE
+ (d1->mtu
+ && (clib_host_to_net_u16 (ip41->length) > d1->mtu)))
+ {
+ vnet_buffer (p1)->ip_frag.header_offset = 0;
+ vnet_buffer (p1)->ip_frag.flags = 0;
+ vnet_buffer (p1)->ip_frag.next_index =
+ IP4_FRAG_NEXT_IP4_LOOKUP;
+ vnet_buffer (p1)->ip_frag.mtu = d1->mtu;
+ next1 = IP6_MAP_NEXT_IP4_FRAGMENT;
+ }
+ else
+ {
+ next1 =
+ ip6_map_ip4_lookup_bypass (p1,
+ ip41) ?
+ IP6_MAP_NEXT_IP4_REWRITE : next1;
+ }
+ vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_RX,
+ cpu_index,
+ map_domain_index1, 1,
+ clib_net_to_host_u16
+ (ip41->length));
+ }
+ }
+
+ if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED))
+ {
+ map_trace_t *tr = vlib_add_trace (vm, node, p0, sizeof (*tr));
+ tr->map_domain_index = map_domain_index0;
+ tr->port = port0;
+ }
+
+ if (PREDICT_FALSE (p1->flags & VLIB_BUFFER_IS_TRACED))
+ {
+ map_trace_t *tr = vlib_add_trace (vm, node, p1, sizeof (*tr));
+ tr->map_domain_index = map_domain_index1;
+ tr->port = port1;
+ }
+
+ if (error0 == MAP_ERROR_DECAP_SEC_CHECK && mm->icmp6_enabled)
+ {
+ /* Set ICMP parameters */
+ vlib_buffer_advance (p0, -sizeof (ip6_header_t));
+ icmp6_error_set_vnet_buffer (p0, ICMP6_destination_unreachable,
+ ICMP6_destination_unreachable_source_address_failed_policy,
+ 0);
+ next0 = IP6_MAP_NEXT_ICMP;
+ }
+ else
+ {
+ next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_NEXT_DROP;
+ }
+
+ if (error1 == MAP_ERROR_DECAP_SEC_CHECK && mm->icmp6_enabled)
+ {
+ /* Set ICMP parameters */
+ vlib_buffer_advance (p1, -sizeof (ip6_header_t));
+ icmp6_error_set_vnet_buffer (p1, ICMP6_destination_unreachable,
+ ICMP6_destination_unreachable_source_address_failed_policy,
+ 0);
+ next1 = IP6_MAP_NEXT_ICMP;
+ }
+ else
+ {
+ next1 = (error1 == MAP_ERROR_NONE) ? next1 : IP6_MAP_NEXT_DROP;
+ }
+
+ /* Reset packet */
+ if (next0 == IP6_MAP_NEXT_IP6_LOCAL)
+ vlib_buffer_advance (p0, -sizeof (ip6_header_t));
+ if (next1 == IP6_MAP_NEXT_IP6_LOCAL)
+ vlib_buffer_advance (p1, -sizeof (ip6_header_t));
+
+ p0->error = error_node->errors[error0];
+ p1->error = error_node->errors[error1];
+ vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next,
+ n_left_to_next, pi0, pi1, next0,
+ next1);
}
- }
-
- if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) {
- map_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr));
- tr->map_domain_index = map_domain_index0;
- tr->port = port0;
- }
-
- if (PREDICT_FALSE(p1->flags & VLIB_BUFFER_IS_TRACED)) {
- map_trace_t *tr = vlib_add_trace(vm, node, p1, sizeof(*tr));
- tr->map_domain_index = map_domain_index1;
- tr->port = port1;
- }
-
- if (error0 == MAP_ERROR_DECAP_SEC_CHECK && mm->icmp6_enabled) {
- /* Set ICMP parameters */
- vlib_buffer_advance(p0, -sizeof(ip6_header_t));
- icmp6_error_set_vnet_buffer(p0, ICMP6_destination_unreachable,
- ICMP6_destination_unreachable_source_address_failed_policy, 0);
- next0 = IP6_MAP_NEXT_ICMP;
- } else {
- next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_NEXT_DROP;
- }
-
- if (error1 == MAP_ERROR_DECAP_SEC_CHECK && mm->icmp6_enabled) {
- /* Set ICMP parameters */
- vlib_buffer_advance(p1, -sizeof(ip6_header_t));
- icmp6_error_set_vnet_buffer(p1, ICMP6_destination_unreachable,
- ICMP6_destination_unreachable_source_address_failed_policy, 0);
- next1 = IP6_MAP_NEXT_ICMP;
- } else {
- next1 = (error1 == MAP_ERROR_NONE) ? next1 : IP6_MAP_NEXT_DROP;
- }
-
- /* Reset packet */
- if (next0 == IP6_MAP_NEXT_IP6_LOCAL)
- vlib_buffer_advance(p0, -sizeof(ip6_header_t));
- if (next1 == IP6_MAP_NEXT_IP6_LOCAL)
- vlib_buffer_advance(p1, -sizeof(ip6_header_t));
-
- p0->error = error_node->errors[error0];
- p1->error = error_node->errors[error1];
- vlib_validate_buffer_enqueue_x2(vm, node, next_index, to_next, n_left_to_next, pi0, pi1, next0, next1);
- }
- /* Single loop */
- while (n_left_from > 0 && n_left_to_next > 0) {
- u32 pi0;
- vlib_buffer_t *p0;
- u8 error0 = MAP_ERROR_NONE;
- map_domain_t *d0 = 0;
- ip4_header_t *ip40;
- ip6_header_t *ip60;
- i32 port0 = 0;
- u32 map_domain_index0 = ~0;
- u32 next0 = IP6_MAP_NEXT_IP4_LOOKUP;
-
- pi0 = to_next[0] = from[0];
- from += 1;
- n_left_from -= 1;
- to_next +=1;
- n_left_to_next -= 1;
-
- p0 = vlib_get_buffer(vm, pi0);
- ip60 = vlib_buffer_get_current(p0);
- vlib_buffer_advance(p0, sizeof(ip6_header_t));
- ip40 = vlib_buffer_get_current(p0);
-
- /*
- * Encapsulated IPv4 packet
- * - IPv4 fragmented -> Pass to virtual reassembly unless security check disabled
- * - Lookup/Rewrite or Fragment node in case of packet > MTU
- * Fragmented IPv6 packet
- * ICMP IPv6 packet
- * - Error -> Pass to ICMPv6/ICMPv4 relay
- * - Info -> Pass to IPv6 local
- * Anything else -> drop
- */
- if (PREDICT_TRUE(ip60->protocol == IP_PROTOCOL_IP_IN_IP && clib_net_to_host_u16(ip60->payload_length) > 20)) {
- d0 = ip6_map_get_domain(vnet_buffer(p0)->ip.adj_index[VLIB_TX], (ip4_address_t *)&ip40->src_address.as_u32,
- &map_domain_index0, &error0);
- } else if (ip60->protocol == IP_PROTOCOL_ICMP6 &&
- clib_net_to_host_u16(ip60->payload_length) > sizeof(icmp46_header_t)) {
- icmp46_header_t *icmp = (void *)(ip60 + 1);
- next0 = (icmp->type == ICMP6_echo_request || icmp->type == ICMP6_echo_reply) ?
- IP6_MAP_NEXT_IP6_LOCAL : IP6_MAP_NEXT_IP6_ICMP_RELAY;
- } else if (ip60->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION &&
- (((ip6_frag_hdr_t *)(ip60+1))->next_hdr == IP_PROTOCOL_IP_IN_IP)) {
- next0 = IP6_MAP_NEXT_IP6_REASS;
- } else {
- error0 = MAP_ERROR_BAD_PROTOCOL;
- }
-
- if (d0) {
- /* MAP inbound security check */
- ip6_map_security_check(d0, ip40, ip60, &next0, &error0);
-
- if (PREDICT_TRUE(error0 == MAP_ERROR_NONE &&
- next0 == IP6_MAP_NEXT_IP4_LOOKUP)) {
- if (PREDICT_FALSE(d0->mtu && (clib_host_to_net_u16(ip40->length) > d0->mtu))) {
- vnet_buffer(p0)->ip_frag.header_offset = 0;
- vnet_buffer(p0)->ip_frag.flags = 0;
- vnet_buffer(p0)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP;
- vnet_buffer(p0)->ip_frag.mtu = d0->mtu;
- next0 = IP6_MAP_NEXT_IP4_FRAGMENT;
- } else {
- next0 = ip6_map_ip4_lookup_bypass(p0, ip40) ? IP6_MAP_NEXT_IP4_REWRITE : next0;
- }
- vlib_increment_combined_counter(cm + MAP_DOMAIN_COUNTER_RX, cpu_index, map_domain_index0, 1,
- clib_net_to_host_u16(ip40->length));
+ /* Single loop */
+ while (n_left_from > 0 && n_left_to_next > 0)
+ {
+ u32 pi0;
+ vlib_buffer_t *p0;
+ u8 error0 = MAP_ERROR_NONE;
+ map_domain_t *d0 = 0;
+ ip4_header_t *ip40;
+ ip6_header_t *ip60;
+ i32 port0 = 0;
+ u32 map_domain_index0 = ~0;
+ u32 next0 = IP6_MAP_NEXT_IP4_LOOKUP;
+
+ pi0 = to_next[0] = from[0];
+ from += 1;
+ n_left_from -= 1;
+ to_next += 1;
+ n_left_to_next -= 1;
+
+ p0 = vlib_get_buffer (vm, pi0);
+ ip60 = vlib_buffer_get_current (p0);
+ vlib_buffer_advance (p0, sizeof (ip6_header_t));
+ ip40 = vlib_buffer_get_current (p0);
+
+ /*
+ * Encapsulated IPv4 packet
+ * - IPv4 fragmented -> Pass to virtual reassembly unless security check disabled
+ * - Lookup/Rewrite or Fragment node in case of packet > MTU
+ * Fragmented IPv6 packet
+ * ICMP IPv6 packet
+ * - Error -> Pass to ICMPv6/ICMPv4 relay
+ * - Info -> Pass to IPv6 local
+ * Anything else -> drop
+ */
+ if (PREDICT_TRUE
+ (ip60->protocol == IP_PROTOCOL_IP_IN_IP
+ && clib_net_to_host_u16 (ip60->payload_length) > 20))
+ {
+ d0 =
+ ip6_map_get_domain (vnet_buffer (p0)->ip.adj_index[VLIB_TX],
+ (ip4_address_t *) & ip40->src_address.
+ as_u32, &map_domain_index0, &error0);
+ }
+ else if (ip60->protocol == IP_PROTOCOL_ICMP6 &&
+ clib_net_to_host_u16 (ip60->payload_length) >
+ sizeof (icmp46_header_t))
+ {
+ icmp46_header_t *icmp = (void *) (ip60 + 1);
+ next0 = (icmp->type == ICMP6_echo_request
+ || icmp->type ==
+ ICMP6_echo_reply) ? IP6_MAP_NEXT_IP6_LOCAL :
+ IP6_MAP_NEXT_IP6_ICMP_RELAY;
+ }
+ else if (ip60->protocol == IP_PROTOCOL_IPV6_FRAGMENTATION &&
+ (((ip6_frag_hdr_t *) (ip60 + 1))->next_hdr ==
+ IP_PROTOCOL_IP_IN_IP))
+ {
+ next0 = IP6_MAP_NEXT_IP6_REASS;
+ }
+ else
+ {
+ error0 = MAP_ERROR_BAD_PROTOCOL;
+ }
+
+ if (d0)
+ {
+ /* MAP inbound security check */
+ ip6_map_security_check (d0, ip40, ip60, &next0, &error0);
+
+ if (PREDICT_TRUE (error0 == MAP_ERROR_NONE &&
+ next0 == IP6_MAP_NEXT_IP4_LOOKUP))
+ {
+ if (PREDICT_FALSE
+ (d0->mtu
+ && (clib_host_to_net_u16 (ip40->length) > d0->mtu)))
+ {
+ vnet_buffer (p0)->ip_frag.header_offset = 0;
+ vnet_buffer (p0)->ip_frag.flags = 0;
+ vnet_buffer (p0)->ip_frag.next_index =
+ IP4_FRAG_NEXT_IP4_LOOKUP;
+ vnet_buffer (p0)->ip_frag.mtu = d0->mtu;
+ next0 = IP6_MAP_NEXT_IP4_FRAGMENT;
+ }
+ else
+ {
+ next0 =
+ ip6_map_ip4_lookup_bypass (p0,
+ ip40) ?
+ IP6_MAP_NEXT_IP4_REWRITE : next0;
+ }
+ vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_RX,
+ cpu_index,
+ map_domain_index0, 1,
+ clib_net_to_host_u16
+ (ip40->length));
+ }
+ }
+
+ if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED))
+ {
+ map_trace_t *tr = vlib_add_trace (vm, node, p0, sizeof (*tr));
+ tr->map_domain_index = map_domain_index0;
+ tr->port = (u16) port0;
+ }
+
+ if (mm->icmp6_enabled &&
+ (error0 == MAP_ERROR_DECAP_SEC_CHECK
+ || error0 == MAP_ERROR_NO_DOMAIN))
+ {
+ /* Set ICMP parameters */
+ vlib_buffer_advance (p0, -sizeof (ip6_header_t));
+ icmp6_error_set_vnet_buffer (p0, ICMP6_destination_unreachable,
+ ICMP6_destination_unreachable_source_address_failed_policy,
+ 0);
+ next0 = IP6_MAP_NEXT_ICMP;
+ }
+ else
+ {
+ next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_NEXT_DROP;
+ }
+
+ /* Reset packet */
+ if (next0 == IP6_MAP_NEXT_IP6_LOCAL)
+ vlib_buffer_advance (p0, -sizeof (ip6_header_t));
+
+ p0->error = error_node->errors[error0];
+ vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
+ n_left_to_next, pi0, next0);
}
- }
-
- if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) {
- map_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr));
- tr->map_domain_index = map_domain_index0;
- tr->port = (u16)port0;
- }
-
- if (mm->icmp6_enabled &&
- (error0 == MAP_ERROR_DECAP_SEC_CHECK || error0 == MAP_ERROR_NO_DOMAIN)) {
- /* Set ICMP parameters */
- vlib_buffer_advance(p0, -sizeof(ip6_header_t));
- icmp6_error_set_vnet_buffer(p0, ICMP6_destination_unreachable,
- ICMP6_destination_unreachable_source_address_failed_policy, 0);
- next0 = IP6_MAP_NEXT_ICMP;
- } else {
- next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_NEXT_DROP;
- }
-
- /* Reset packet */
- if (next0 == IP6_MAP_NEXT_IP6_LOCAL)
- vlib_buffer_advance(p0, -sizeof(ip6_header_t));
-
- p0->error = error_node->errors[error0];
- vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, next0);
+ vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
- vlib_put_next_frame(vm, node, next_index, n_left_to_next);
- }
return frame->n_vectors;
}
static_always_inline void
-ip6_map_ip6_reass_prepare(vlib_main_t *vm, vlib_node_runtime_t *node, map_ip6_reass_t *r,
- u32 **fragments_ready, u32 **fragments_to_drop)
+ip6_map_ip6_reass_prepare (vlib_main_t * vm, vlib_node_runtime_t * node,
+ map_ip6_reass_t * r, u32 ** fragments_ready,
+ u32 ** fragments_to_drop)
{
ip4_header_t *ip40;
ip6_header_t *ip60;
ip6_frag_hdr_t *frag0;
vlib_buffer_t *p0;
- if(!r->ip4_header.ip_version_and_header_length)
+ if (!r->ip4_header.ip_version_and_header_length)
return;
//The IP header is here, we need to check for packets
//that can be forwarded
int i;
- for (i=0; i<MAP_IP6_REASS_MAX_FRAGMENTS_PER_REASSEMBLY; i++) {
- if (r->fragments[i].pi == ~0 ||
- ((!r->fragments[i].next_data_len) && (r->fragments[i].next_data_offset != (0xffff))))
- continue;
-
- p0 = vlib_get_buffer(vm, r->fragments[i].pi);
- ip60 = vlib_buffer_get_current(p0);
- frag0 = (ip6_frag_hdr_t *)(ip60 + 1);
- ip40 = (ip4_header_t *)(frag0 + 1);
-
- if (ip6_frag_hdr_offset(frag0)) {
- //Not first fragment, add the IPv4 header
- clib_memcpy(ip40, &r->ip4_header, 20);
- }
+ for (i = 0; i < MAP_IP6_REASS_MAX_FRAGMENTS_PER_REASSEMBLY; i++)
+ {
+ if (r->fragments[i].pi == ~0 ||
+ ((!r->fragments[i].next_data_len)
+ && (r->fragments[i].next_data_offset != (0xffff))))
+ continue;
+
+ p0 = vlib_get_buffer (vm, r->fragments[i].pi);
+ ip60 = vlib_buffer_get_current (p0);
+ frag0 = (ip6_frag_hdr_t *) (ip60 + 1);
+ ip40 = (ip4_header_t *) (frag0 + 1);
+
+ if (ip6_frag_hdr_offset (frag0))
+ {
+ //Not first fragment, add the IPv4 header
+ clib_memcpy (ip40, &r->ip4_header, 20);
+ }
#ifdef MAP_IP6_REASS_COUNT_BYTES
- r->forwarded += clib_net_to_host_u16(ip60->payload_length) - sizeof(*frag0);
+ r->forwarded +=
+ clib_net_to_host_u16 (ip60->payload_length) - sizeof (*frag0);
#endif
- if (ip6_frag_hdr_more(frag0)) {
- //Not last fragment, we copy end of next
- clib_memcpy(u8_ptr_add(ip60, p0->current_length), r->fragments[i].next_data, 20);
- p0->current_length += 20;
- ip60->payload_length = u16_net_add(ip60->payload_length, 20);
- }
-
- if (!ip4_is_fragment(ip40)) {
- ip40->fragment_id = frag_id_6to4(frag0->identification);
- ip40->flags_and_fragment_offset = clib_host_to_net_u16(ip6_frag_hdr_offset(frag0));
- } else {
- ip40->flags_and_fragment_offset = clib_host_to_net_u16(ip4_get_fragment_offset(ip40) + ip6_frag_hdr_offset(frag0));
- }
-
- if (ip6_frag_hdr_more(frag0))
- ip40->flags_and_fragment_offset |= clib_host_to_net_u16(IP4_HEADER_FLAG_MORE_FRAGMENTS);
+ if (ip6_frag_hdr_more (frag0))
+ {
+ //Not last fragment, we copy end of next
+ clib_memcpy (u8_ptr_add (ip60, p0->current_length),
+ r->fragments[i].next_data, 20);
+ p0->current_length += 20;
+ ip60->payload_length = u16_net_add (ip60->payload_length, 20);
+ }
- ip40->length = clib_host_to_net_u16(p0->current_length - sizeof(*ip60) - sizeof(*frag0));
- ip40->checksum = ip4_header_checksum(ip40);
+ if (!ip4_is_fragment (ip40))
+ {
+ ip40->fragment_id = frag_id_6to4 (frag0->identification);
+ ip40->flags_and_fragment_offset =
+ clib_host_to_net_u16 (ip6_frag_hdr_offset (frag0));
+ }
+ else
+ {
+ ip40->flags_and_fragment_offset =
+ clib_host_to_net_u16 (ip4_get_fragment_offset (ip40) +
+ ip6_frag_hdr_offset (frag0));
+ }
- if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) {
- map_ip6_map_ip6_reass_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr));
- tr->offset = ip4_get_fragment_offset(ip40);
- tr->frag_len = clib_net_to_host_u16(ip40->length) - sizeof(*ip40);
- tr->out = 1;
- }
+ if (ip6_frag_hdr_more (frag0))
+ ip40->flags_and_fragment_offset |=
+ clib_host_to_net_u16 (IP4_HEADER_FLAG_MORE_FRAGMENTS);
+
+ ip40->length =
+ clib_host_to_net_u16 (p0->current_length - sizeof (*ip60) -
+ sizeof (*frag0));
+ ip40->checksum = ip4_header_checksum (ip40);
+
+ if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED))
+ {
+ map_ip6_map_ip6_reass_trace_t *tr =
+ vlib_add_trace (vm, node, p0, sizeof (*tr));
+ tr->offset = ip4_get_fragment_offset (ip40);
+ tr->frag_len = clib_net_to_host_u16 (ip40->length) - sizeof (*ip40);
+ tr->out = 1;
+ }
- vec_add1(*fragments_ready, r->fragments[i].pi);
- r->fragments[i].pi = ~0;
- r->fragments[i].next_data_len = 0;
- r->fragments[i].next_data_offset = 0;
- map_main.ip6_reass_buffered_counter--;
+ vec_add1 (*fragments_ready, r->fragments[i].pi);
+ r->fragments[i].pi = ~0;
+ r->fragments[i].next_data_len = 0;
+ r->fragments[i].next_data_offset = 0;
+ map_main.ip6_reass_buffered_counter--;
- //TODO: Best solution would be that ip6_map handles extension headers
- // and ignores atomic fragment. But in the meantime, let's just copy the header.
+ //TODO: Best solution would be that ip6_map handles extension headers
+ // and ignores atomic fragment. But in the meantime, let's just copy the header.
- u8 protocol = frag0->next_hdr;
- memmove(u8_ptr_add(ip40, - sizeof(*ip60)), ip60, sizeof(*ip60));
- ((ip6_header_t *)u8_ptr_add(ip40, - sizeof(*ip60)))->protocol = protocol;
- vlib_buffer_advance(p0, sizeof(*frag0));
- }
+ u8 protocol = frag0->next_hdr;
+ memmove (u8_ptr_add (ip40, -sizeof (*ip60)), ip60, sizeof (*ip60));
+ ((ip6_header_t *) u8_ptr_add (ip40, -sizeof (*ip60)))->protocol =
+ protocol;
+ vlib_buffer_advance (p0, sizeof (*frag0));
+ }
}
void
-map_ip6_drop_pi(u32 pi)
+map_ip6_drop_pi (u32 pi)
{
- vlib_main_t *vm = vlib_get_main();
- vlib_node_runtime_t *n = vlib_node_get_runtime(vm, ip6_map_ip6_reass_node.index);
- vlib_set_next_frame_buffer(vm, n, IP6_MAP_IP6_REASS_NEXT_DROP, pi);
+ vlib_main_t *vm = vlib_get_main ();
+ vlib_node_runtime_t *n =
+ vlib_node_get_runtime (vm, ip6_map_ip6_reass_node.index);
+ vlib_set_next_frame_buffer (vm, n, IP6_MAP_IP6_REASS_NEXT_DROP, pi);
}
void
-map_ip4_drop_pi(u32 pi)
+map_ip4_drop_pi (u32 pi)
{
- vlib_main_t *vm = vlib_get_main();
- vlib_node_runtime_t *n = vlib_node_get_runtime(vm, ip6_map_ip4_reass_node.index);
- vlib_set_next_frame_buffer(vm, n, IP6_MAP_IP4_REASS_NEXT_DROP, pi);
+ vlib_main_t *vm = vlib_get_main ();
+ vlib_node_runtime_t *n =
+ vlib_node_get_runtime (vm, ip6_map_ip4_reass_node.index);
+ vlib_set_next_frame_buffer (vm, n, IP6_MAP_IP4_REASS_NEXT_DROP, pi);
}
/*
@@ -527,109 +680,139 @@ map_ip4_drop_pi(u32 pi)
* have been forwarded.
*/
static uword
-ip6_map_ip6_reass (vlib_main_t *vm,
- vlib_node_runtime_t *node,
- vlib_frame_t *frame)
+ip6_map_ip6_reass (vlib_main_t * vm,
+ vlib_node_runtime_t * node, vlib_frame_t * frame)
{
u32 n_left_from, *from, next_index, *to_next, n_left_to_next;
- vlib_node_runtime_t *error_node = vlib_node_get_runtime(vm, ip6_map_ip6_reass_node.index);
+ vlib_node_runtime_t *error_node =
+ vlib_node_get_runtime (vm, ip6_map_ip6_reass_node.index);
u32 *fragments_to_drop = NULL;
u32 *fragments_ready = NULL;
- from = vlib_frame_vector_args(frame);
+ from = vlib_frame_vector_args (frame);
n_left_from = frame->n_vectors;
next_index = node->cached_next_index;
- while (n_left_from > 0) {
- vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next);
-
- /* Single loop */
- while (n_left_from > 0 && n_left_to_next > 0) {
- u32 pi0;
- vlib_buffer_t *p0;
- u8 error0 = MAP_ERROR_NONE;
- ip6_header_t *ip60;
- ip6_frag_hdr_t *frag0;
- u16 offset;
- u16 next_offset;
- u16 frag_len;
-
- pi0 = to_next[0] = from[0];
- from += 1;
- n_left_from -= 1;
- to_next +=1;
- n_left_to_next -= 1;
-
- p0 = vlib_get_buffer(vm, pi0);
- ip60 = vlib_buffer_get_current(p0);
- frag0 = (ip6_frag_hdr_t *)(ip60 + 1);
- offset = clib_host_to_net_u16(frag0->fragment_offset_and_more) & (~7);
- frag_len = clib_net_to_host_u16(ip60->payload_length) - sizeof(*frag0);
- next_offset = ip6_frag_hdr_more(frag0) ? (offset + frag_len) : (0xffff);
-
- //FIXME: Support other extension headers, maybe
-
- if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) {
- map_ip6_map_ip6_reass_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr));
- tr->offset = offset;
- tr->frag_len = frag_len;
- tr->out = 0;
- }
-
- map_ip6_reass_lock();
- map_ip6_reass_t *r = map_ip6_reass_get(&ip60->src_address, &ip60->dst_address,
- frag0->identification, frag0->next_hdr, &fragments_to_drop);
- //FIXME: Use better error codes
- if (PREDICT_FALSE(!r)) {
- // Could not create a caching entry
- error0 = MAP_ERROR_FRAGMENT_MEMORY;
- } else if (PREDICT_FALSE((frag_len <= 20 &&
- (ip6_frag_hdr_more(frag0) || (!offset))))) {
- //Very small fragment are restricted to the last one and
- //can't be the first one
- error0 = MAP_ERROR_FRAGMENT_MALFORMED;
- } else if (map_ip6_reass_add_fragment(r, pi0, offset, next_offset, (u8 *)(frag0 + 1), frag_len)) {
- map_ip6_reass_free(r, &fragments_to_drop);
- error0 = MAP_ERROR_FRAGMENT_MEMORY;
- } else {
+ while (n_left_from > 0)
+ {
+ vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+
+ /* Single loop */
+ while (n_left_from > 0 && n_left_to_next > 0)
+ {
+ u32 pi0;
+ vlib_buffer_t *p0;
+ u8 error0 = MAP_ERROR_NONE;
+ ip6_header_t *ip60;
+ ip6_frag_hdr_t *frag0;
+ u16 offset;
+ u16 next_offset;
+ u16 frag_len;
+
+ pi0 = to_next[0] = from[0];
+ from += 1;
+ n_left_from -= 1;
+ to_next += 1;
+ n_left_to_next -= 1;
+
+ p0 = vlib_get_buffer (vm, pi0);
+ ip60 = vlib_buffer_get_current (p0);
+ frag0 = (ip6_frag_hdr_t *) (ip60 + 1);
+ offset =
+ clib_host_to_net_u16 (frag0->fragment_offset_and_more) & (~7);
+ frag_len =
+ clib_net_to_host_u16 (ip60->payload_length) - sizeof (*frag0);
+ next_offset =
+ ip6_frag_hdr_more (frag0) ? (offset + frag_len) : (0xffff);
+
+ //FIXME: Support other extension headers, maybe
+
+ if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED))
+ {
+ map_ip6_map_ip6_reass_trace_t *tr =
+ vlib_add_trace (vm, node, p0, sizeof (*tr));
+ tr->offset = offset;
+ tr->frag_len = frag_len;
+ tr->out = 0;
+ }
+
+ map_ip6_reass_lock ();
+ map_ip6_reass_t *r =
+ map_ip6_reass_get (&ip60->src_address, &ip60->dst_address,
+ frag0->identification, frag0->next_hdr,
+ &fragments_to_drop);
+ //FIXME: Use better error codes
+ if (PREDICT_FALSE (!r))
+ {
+ // Could not create a caching entry
+ error0 = MAP_ERROR_FRAGMENT_MEMORY;
+ }
+ else if (PREDICT_FALSE ((frag_len <= 20 &&
+ (ip6_frag_hdr_more (frag0) || (!offset)))))
+ {
+ //Very small fragment are restricted to the last one and
+ //can't be the first one
+ error0 = MAP_ERROR_FRAGMENT_MALFORMED;
+ }
+ else
+ if (map_ip6_reass_add_fragment
+ (r, pi0, offset, next_offset, (u8 *) (frag0 + 1), frag_len))
+ {
+ map_ip6_reass_free (r, &fragments_to_drop);
+ error0 = MAP_ERROR_FRAGMENT_MEMORY;
+ }
+ else
+ {
#ifdef MAP_IP6_REASS_COUNT_BYTES
- if (!ip6_frag_hdr_more(frag0))
- r->expected_total = offset + frag_len;
+ if (!ip6_frag_hdr_more (frag0))
+ r->expected_total = offset + frag_len;
#endif
- ip6_map_ip6_reass_prepare(vm, node, r, &fragments_ready, &fragments_to_drop);
+ ip6_map_ip6_reass_prepare (vm, node, r, &fragments_ready,
+ &fragments_to_drop);
#ifdef MAP_IP6_REASS_COUNT_BYTES
- if(r->forwarded >= r->expected_total)
- map_ip6_reass_free(r, &fragments_to_drop);
+ if (r->forwarded >= r->expected_total)
+ map_ip6_reass_free (r, &fragments_to_drop);
#endif
- }
- map_ip6_reass_unlock();
-
- if (error0 == MAP_ERROR_NONE) {
- if (frag_len > 20) {
- //Dequeue the packet
- n_left_to_next++;
- to_next--;
- } else {
- //All data from that packet was copied no need to keep it, but this is not an error
- p0->error = error_node->errors[MAP_ERROR_NONE];
- vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, IP6_MAP_IP6_REASS_NEXT_DROP);
- }
- } else {
- p0->error = error_node->errors[error0];
- vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, IP6_MAP_IP6_REASS_NEXT_DROP);
- }
+ }
+ map_ip6_reass_unlock ();
+
+ if (error0 == MAP_ERROR_NONE)
+ {
+ if (frag_len > 20)
+ {
+ //Dequeue the packet
+ n_left_to_next++;
+ to_next--;
+ }
+ else
+ {
+ //All data from that packet was copied no need to keep it, but this is not an error
+ p0->error = error_node->errors[MAP_ERROR_NONE];
+ vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
+ to_next, n_left_to_next,
+ pi0,
+ IP6_MAP_IP6_REASS_NEXT_DROP);
+ }
+ }
+ else
+ {
+ p0->error = error_node->errors[error0];
+ vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
+ n_left_to_next, pi0,
+ IP6_MAP_IP6_REASS_NEXT_DROP);
+ }
+ }
+ vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
- vlib_put_next_frame(vm, node, next_index, n_left_to_next);
- }
-
- map_send_all_to_node(vm, fragments_ready, node,
- &error_node->errors[MAP_ERROR_NONE],
- IP6_MAP_IP6_REASS_NEXT_IP6_MAP);
- map_send_all_to_node(vm, fragments_to_drop, node,
- &error_node->errors[MAP_ERROR_FRAGMENT_DROPPED],
- IP6_MAP_IP6_REASS_NEXT_DROP);
-
- vec_free(fragments_to_drop);
- vec_free(fragments_ready);
+
+ map_send_all_to_node (vm, fragments_ready, node,
+ &error_node->errors[MAP_ERROR_NONE],
+ IP6_MAP_IP6_REASS_NEXT_IP6_MAP);
+ map_send_all_to_node (vm, fragments_to_drop, node,
+ &error_node->errors[MAP_ERROR_FRAGMENT_DROPPED],
+ IP6_MAP_IP6_REASS_NEXT_DROP);
+
+ vec_free (fragments_to_drop);
+ vec_free (fragments_ready);
return frame->n_vectors;
}
@@ -637,147 +820,195 @@ ip6_map_ip6_reass (vlib_main_t *vm,
* ip6_ip4_virt_reass
*/
static uword
-ip6_map_ip4_reass (vlib_main_t *vm,
- vlib_node_runtime_t *node,
- vlib_frame_t *frame)
+ip6_map_ip4_reass (vlib_main_t * vm,
+ vlib_node_runtime_t * node, vlib_frame_t * frame)
{
u32 n_left_from, *from, next_index, *to_next, n_left_to_next;
- vlib_node_runtime_t *error_node = vlib_node_get_runtime(vm, ip6_map_ip4_reass_node.index);
+ vlib_node_runtime_t *error_node =
+ vlib_node_get_runtime (vm, ip6_map_ip4_reass_node.index);
map_main_t *mm = &map_main;
vlib_combined_counter_main_t *cm = mm->domain_counters;
- u32 cpu_index = os_get_cpu_number();
+ u32 cpu_index = os_get_cpu_number ();
u32 *fragments_to_drop = NULL;
u32 *fragments_to_loopback = NULL;
- from = vlib_frame_vector_args(frame);
+ from = vlib_frame_vector_args (frame);
n_left_from = frame->n_vectors;
next_index = node->cached_next_index;
- while (n_left_from > 0) {
- vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next);
-
- /* Single loop */
- while (n_left_from > 0 && n_left_to_next > 0) {
- u32 pi0;
- vlib_buffer_t *p0;
- u8 error0 = MAP_ERROR_NONE;
- map_domain_t *d0;
- ip4_header_t *ip40;
- ip6_header_t *ip60;
- i32 port0 = 0;
- u32 map_domain_index0 = ~0;
- u32 next0 = IP6_MAP_IP4_REASS_NEXT_IP4_LOOKUP;
- u8 cached = 0;
-
- pi0 = to_next[0] = from[0];
- from += 1;
- n_left_from -= 1;
- to_next +=1;
- n_left_to_next -= 1;
-
- p0 = vlib_get_buffer(vm, pi0);
- ip40 = vlib_buffer_get_current(p0);
- ip60 = ((ip6_header_t *)ip40) - 1;
-
- d0 = ip6_map_get_domain(vnet_buffer(p0)->ip.adj_index[VLIB_TX], (ip4_address_t *)&ip40->src_address.as_u32,
- &map_domain_index0, &error0);
-
- map_ip4_reass_lock();
- //This node only deals with fragmented ip4
- map_ip4_reass_t *r = map_ip4_reass_get(ip40->src_address.as_u32, ip40->dst_address.as_u32,
- ip40->fragment_id, ip40->protocol, &fragments_to_drop);
- if (PREDICT_FALSE(!r)) {
- // Could not create a caching entry
- error0 = MAP_ERROR_FRAGMENT_MEMORY;
- } else if (PREDICT_TRUE(ip4_get_fragment_offset(ip40))) {
- // This is a fragment
- if (r->port >= 0) {
- // We know the port already
- port0 = r->port;
- } else if (map_ip4_reass_add_fragment(r, pi0)) {
- // Not enough space for caching
- error0 = MAP_ERROR_FRAGMENT_MEMORY;
- map_ip4_reass_free(r, &fragments_to_drop);
- } else {
- cached = 1;
- }
- } else if ((port0 = ip4_get_port(ip40, MAP_SENDER, p0->current_length)) < 0) {
- // Could not find port from first fragment. Stop reassembling.
- error0 = MAP_ERROR_BAD_PROTOCOL;
- port0 = 0;
- map_ip4_reass_free(r, &fragments_to_drop);
- } else {
- // Found port. Remember it and loopback saved fragments
- r->port = port0;
- map_ip4_reass_get_fragments(r, &fragments_to_loopback);
- }
+ while (n_left_from > 0)
+ {
+ vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+
+ /* Single loop */
+ while (n_left_from > 0 && n_left_to_next > 0)
+ {
+ u32 pi0;
+ vlib_buffer_t *p0;
+ u8 error0 = MAP_ERROR_NONE;
+ map_domain_t *d0;
+ ip4_header_t *ip40;
+ ip6_header_t *ip60;
+ i32 port0 = 0;
+ u32 map_domain_index0 = ~0;
+ u32 next0 = IP6_MAP_IP4_REASS_NEXT_IP4_LOOKUP;
+ u8 cached = 0;
+
+ pi0 = to_next[0] = from[0];
+ from += 1;
+ n_left_from -= 1;
+ to_next += 1;
+ n_left_to_next -= 1;
+
+ p0 = vlib_get_buffer (vm, pi0);
+ ip40 = vlib_buffer_get_current (p0);
+ ip60 = ((ip6_header_t *) ip40) - 1;
+
+ d0 =
+ ip6_map_get_domain (vnet_buffer (p0)->ip.adj_index[VLIB_TX],
+ (ip4_address_t *) & ip40->src_address.as_u32,
+ &map_domain_index0, &error0);
+
+ map_ip4_reass_lock ();
+ //This node only deals with fragmented ip4
+ map_ip4_reass_t *r =
+ map_ip4_reass_get (ip40->src_address.as_u32,
+ ip40->dst_address.as_u32,
+ ip40->fragment_id, ip40->protocol,
+ &fragments_to_drop);
+ if (PREDICT_FALSE (!r))
+ {
+ // Could not create a caching entry
+ error0 = MAP_ERROR_FRAGMENT_MEMORY;
+ }
+ else if (PREDICT_TRUE (ip4_get_fragment_offset (ip40)))
+ {
+ // This is a fragment
+ if (r->port >= 0)
+ {
+ // We know the port already
+ port0 = r->port;
+ }
+ else if (map_ip4_reass_add_fragment (r, pi0))
+ {
+ // Not enough space for caching
+ error0 = MAP_ERROR_FRAGMENT_MEMORY;
+ map_ip4_reass_free (r, &fragments_to_drop);
+ }
+ else
+ {
+ cached = 1;
+ }
+ }
+ else
+ if ((port0 =
+ ip4_get_port (ip40, MAP_SENDER, p0->current_length)) < 0)
+ {
+ // Could not find port from first fragment. Stop reassembling.
+ error0 = MAP_ERROR_BAD_PROTOCOL;
+ port0 = 0;
+ map_ip4_reass_free (r, &fragments_to_drop);
+ }
+ else
+ {
+ // Found port. Remember it and loopback saved fragments
+ r->port = port0;
+ map_ip4_reass_get_fragments (r, &fragments_to_loopback);
+ }
#ifdef MAP_IP4_REASS_COUNT_BYTES
- if (!cached && r) {
- r->forwarded += clib_host_to_net_u16(ip40->length) - 20;
- if (!ip4_get_fragment_more(ip40))
- r->expected_total = ip4_get_fragment_offset(ip40) * 8 + clib_host_to_net_u16(ip40->length) - 20;
- if(r->forwarded >= r->expected_total)
- map_ip4_reass_free(r, &fragments_to_drop);
- }
+ if (!cached && r)
+ {
+ r->forwarded += clib_host_to_net_u16 (ip40->length) - 20;
+ if (!ip4_get_fragment_more (ip40))
+ r->expected_total =
+ ip4_get_fragment_offset (ip40) * 8 +
+ clib_host_to_net_u16 (ip40->length) - 20;
+ if (r->forwarded >= r->expected_total)
+ map_ip4_reass_free (r, &fragments_to_drop);
+ }
#endif
- map_ip4_reass_unlock();
-
- if(PREDICT_TRUE(error0 == MAP_ERROR_NONE))
- error0 = ip6_map_sec_check(d0, port0, ip40, ip60) ? MAP_ERROR_NONE : MAP_ERROR_DECAP_SEC_CHECK;
-
- if (PREDICT_FALSE(d0->mtu && (clib_host_to_net_u16(ip40->length) > d0->mtu) &&
- error0 == MAP_ERROR_NONE && !cached)) {
- vnet_buffer(p0)->ip_frag.header_offset = 0;
- vnet_buffer(p0)->ip_frag.flags = 0;
- vnet_buffer(p0)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP;
- vnet_buffer(p0)->ip_frag.mtu = d0->mtu;
- next0 = IP6_MAP_IP4_REASS_NEXT_IP4_FRAGMENT;
- }
-
- if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) {
- map_ip6_map_ip4_reass_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr));
- tr->map_domain_index = map_domain_index0;
- tr->port = port0;
- tr->cached = cached;
- }
-
- if (cached) {
- //Dequeue the packet
- n_left_to_next++;
- to_next--;
- } else {
- if (error0 == MAP_ERROR_NONE)
- vlib_increment_combined_counter(cm + MAP_DOMAIN_COUNTER_RX, cpu_index, map_domain_index0, 1,
- clib_net_to_host_u16(ip40->length));
- next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_MAP_IP4_REASS_NEXT_DROP;
- p0->error = error_node->errors[error0];
- vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, next0);
- }
-
- //Loopback when we reach the end of the inpu vector
- if(n_left_from == 0 && vec_len(fragments_to_loopback)) {
- from = vlib_frame_vector_args(frame);
- u32 len = vec_len(fragments_to_loopback);
- if(len <= VLIB_FRAME_SIZE) {
- clib_memcpy(from, fragments_to_loopback, sizeof(u32)*len);
- n_left_from = len;
- vec_reset_length(fragments_to_loopback);
- } else {
- clib_memcpy(from, fragments_to_loopback + (len - VLIB_FRAME_SIZE), sizeof(u32)*VLIB_FRAME_SIZE);
- n_left_from = VLIB_FRAME_SIZE;
- _vec_len(fragments_to_loopback) = len - VLIB_FRAME_SIZE;
- }
- }
+ map_ip4_reass_unlock ();
+
+ if (PREDICT_TRUE (error0 == MAP_ERROR_NONE))
+ error0 =
+ ip6_map_sec_check (d0, port0, ip40,
+ ip60) ? MAP_ERROR_NONE :
+ MAP_ERROR_DECAP_SEC_CHECK;
+
+ if (PREDICT_FALSE
+ (d0->mtu && (clib_host_to_net_u16 (ip40->length) > d0->mtu)
+ && error0 == MAP_ERROR_NONE && !cached))
+ {
+ vnet_buffer (p0)->ip_frag.header_offset = 0;
+ vnet_buffer (p0)->ip_frag.flags = 0;
+ vnet_buffer (p0)->ip_frag.next_index = IP4_FRAG_NEXT_IP4_LOOKUP;
+ vnet_buffer (p0)->ip_frag.mtu = d0->mtu;
+ next0 = IP6_MAP_IP4_REASS_NEXT_IP4_FRAGMENT;
+ }
+
+ if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED))
+ {
+ map_ip6_map_ip4_reass_trace_t *tr =
+ vlib_add_trace (vm, node, p0, sizeof (*tr));
+ tr->map_domain_index = map_domain_index0;
+ tr->port = port0;
+ tr->cached = cached;
+ }
+
+ if (cached)
+ {
+ //Dequeue the packet
+ n_left_to_next++;
+ to_next--;
+ }
+ else
+ {
+ if (error0 == MAP_ERROR_NONE)
+ vlib_increment_combined_counter (cm + MAP_DOMAIN_COUNTER_RX,
+ cpu_index, map_domain_index0,
+ 1,
+ clib_net_to_host_u16 (ip40->
+ length));
+ next0 =
+ (error0 ==
+ MAP_ERROR_NONE) ? next0 : IP6_MAP_IP4_REASS_NEXT_DROP;
+ p0->error = error_node->errors[error0];
+ vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
+ n_left_to_next, pi0, next0);
+ }
+
+ //Loopback when we reach the end of the inpu vector
+ if (n_left_from == 0 && vec_len (fragments_to_loopback))
+ {
+ from = vlib_frame_vector_args (frame);
+ u32 len = vec_len (fragments_to_loopback);
+ if (len <= VLIB_FRAME_SIZE)
+ {
+ clib_memcpy (from, fragments_to_loopback,
+ sizeof (u32) * len);
+ n_left_from = len;
+ vec_reset_length (fragments_to_loopback);
+ }
+ else
+ {
+ clib_memcpy (from,
+ fragments_to_loopback + (len -
+ VLIB_FRAME_SIZE),
+ sizeof (u32) * VLIB_FRAME_SIZE);
+ n_left_from = VLIB_FRAME_SIZE;
+ _vec_len (fragments_to_loopback) = len - VLIB_FRAME_SIZE;
+ }
+ }
+ }
+ vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
- vlib_put_next_frame(vm, node, next_index, n_left_to_next);
- }
- map_send_all_to_node(vm, fragments_to_drop, node,
- &error_node->errors[MAP_ERROR_FRAGMENT_DROPPED],
- IP6_MAP_IP4_REASS_NEXT_DROP);
-
- vec_free(fragments_to_drop);
- vec_free(fragments_to_loopback);
+ map_send_all_to_node (vm, fragments_to_drop, node,
+ &error_node->errors[MAP_ERROR_FRAGMENT_DROPPED],
+ IP6_MAP_IP4_REASS_NEXT_DROP);
+
+ vec_free (fragments_to_drop);
+ vec_free (fragments_to_loopback);
return frame->n_vectors;
}
@@ -785,142 +1016,160 @@ ip6_map_ip4_reass (vlib_main_t *vm,
* ip6_icmp_relay
*/
static uword
-ip6_map_icmp_relay (vlib_main_t *vm,
- vlib_node_runtime_t *node,
- vlib_frame_t *frame)
+ip6_map_icmp_relay (vlib_main_t * vm,
+ vlib_node_runtime_t * node, vlib_frame_t * frame)
{
u32 n_left_from, *from, next_index, *to_next, n_left_to_next;
- vlib_node_runtime_t *error_node = vlib_node_get_runtime(vm, ip6_map_icmp_relay_node.index);
+ vlib_node_runtime_t *error_node =
+ vlib_node_get_runtime (vm, ip6_map_icmp_relay_node.index);
map_main_t *mm = &map_main;
- u32 cpu_index = os_get_cpu_number();
+ u32 cpu_index = os_get_cpu_number ();
u16 *fragment_ids, *fid;
- from = vlib_frame_vector_args(frame);
+ from = vlib_frame_vector_args (frame);
n_left_from = frame->n_vectors;
next_index = node->cached_next_index;
/* Get random fragment IDs for replies. */
- fid = fragment_ids = clib_random_buffer_get_data (&vm->random_buffer, n_left_from * sizeof (fragment_ids[0]));
-
- while (n_left_from > 0) {
- vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next);
-
- /* Single loop */
- while (n_left_from > 0 && n_left_to_next > 0) {
- u32 pi0;
- vlib_buffer_t *p0;
- u8 error0 = MAP_ERROR_NONE;
- ip6_header_t *ip60;
- u32 next0 = IP6_ICMP_RELAY_NEXT_IP4_LOOKUP;
- u32 mtu;
-
- pi0 = to_next[0] = from[0];
- from += 1;
- n_left_from -= 1;
- to_next +=1;
- n_left_to_next -= 1;
-
- p0 = vlib_get_buffer(vm, pi0);
- ip60 = vlib_buffer_get_current(p0);
- u16 tlen = clib_net_to_host_u16(ip60->payload_length);
-
- /*
- * In:
- * IPv6 header (40)
- * ICMPv6 header (8)
- * IPv6 header (40)
- * Original IPv4 header / packet
- * Out:
- * New IPv4 header
- * New ICMP header
- * Original IPv4 header / packet
- */
-
- /* Need at least ICMP(8) + IPv6(40) + IPv4(20) + L4 header(8) */
- if (tlen < 76) {
- error0 = MAP_ERROR_ICMP_RELAY;
- goto error;
- }
-
- icmp46_header_t *icmp60 = (icmp46_header_t *)(ip60 + 1);
- ip6_header_t *inner_ip60 = (ip6_header_t *)(icmp60 + 2);
-
- if (inner_ip60->protocol != IP_PROTOCOL_IP_IN_IP) {
- error0 = MAP_ERROR_ICMP_RELAY;
- goto error;
- }
-
- ip4_header_t *inner_ip40 = (ip4_header_t *)(inner_ip60 + 1);
- vlib_buffer_advance(p0, 60); /* sizeof ( IPv6 + ICMP + IPv6 - IPv4 - ICMP ) */
- ip4_header_t *new_ip40 = vlib_buffer_get_current(p0);
- icmp46_header_t *new_icmp40 = (icmp46_header_t *)(new_ip40 + 1);
-
- /*
- * Relay according to RFC2473, section 8.3
- */
- switch (icmp60->type) {
- case ICMP6_destination_unreachable:
- case ICMP6_time_exceeded:
- case ICMP6_parameter_problem:
- /* Type 3 - destination unreachable, Code 1 - host unreachable */
- new_icmp40->type = ICMP4_destination_unreachable;
- new_icmp40->code = ICMP4_destination_unreachable_destination_unreachable_host;
- break;
-
- case ICMP6_packet_too_big:
- /* Type 3 - destination unreachable, Code 4 - packet too big */
- /* Potential TODO: Adjust domain tunnel MTU based on the value received here */
- mtu = clib_net_to_host_u32(*((u32 *)(icmp60 + 1)));
-
- /* Check DF flag */
- if (!(inner_ip40->flags_and_fragment_offset & clib_host_to_net_u16(IP4_HEADER_FLAG_DONT_FRAGMENT))) {
- error0 = MAP_ERROR_ICMP_RELAY;
- goto error;
+ fid = fragment_ids =
+ clib_random_buffer_get_data (&vm->random_buffer,
+ n_left_from * sizeof (fragment_ids[0]));
+
+ while (n_left_from > 0)
+ {
+ vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+
+ /* Single loop */
+ while (n_left_from > 0 && n_left_to_next > 0)
+ {
+ u32 pi0;
+ vlib_buffer_t *p0;
+ u8 error0 = MAP_ERROR_NONE;
+ ip6_header_t *ip60;
+ u32 next0 = IP6_ICMP_RELAY_NEXT_IP4_LOOKUP;
+ u32 mtu;
+
+ pi0 = to_next[0] = from[0];
+ from += 1;
+ n_left_from -= 1;
+ to_next += 1;
+ n_left_to_next -= 1;
+
+ p0 = vlib_get_buffer (vm, pi0);
+ ip60 = vlib_buffer_get_current (p0);
+ u16 tlen = clib_net_to_host_u16 (ip60->payload_length);
+
+ /*
+ * In:
+ * IPv6 header (40)
+ * ICMPv6 header (8)
+ * IPv6 header (40)
+ * Original IPv4 header / packet
+ * Out:
+ * New IPv4 header
+ * New ICMP header
+ * Original IPv4 header / packet
+ */
+
+ /* Need at least ICMP(8) + IPv6(40) + IPv4(20) + L4 header(8) */
+ if (tlen < 76)
+ {
+ error0 = MAP_ERROR_ICMP_RELAY;
+ goto error;
+ }
+
+ icmp46_header_t *icmp60 = (icmp46_header_t *) (ip60 + 1);
+ ip6_header_t *inner_ip60 = (ip6_header_t *) (icmp60 + 2);
+
+ if (inner_ip60->protocol != IP_PROTOCOL_IP_IN_IP)
+ {
+ error0 = MAP_ERROR_ICMP_RELAY;
+ goto error;
+ }
+
+ ip4_header_t *inner_ip40 = (ip4_header_t *) (inner_ip60 + 1);
+ vlib_buffer_advance (p0, 60); /* sizeof ( IPv6 + ICMP + IPv6 - IPv4 - ICMP ) */
+ ip4_header_t *new_ip40 = vlib_buffer_get_current (p0);
+ icmp46_header_t *new_icmp40 = (icmp46_header_t *) (new_ip40 + 1);
+
+ /*
+ * Relay according to RFC2473, section 8.3
+ */
+ switch (icmp60->type)
+ {
+ case ICMP6_destination_unreachable:
+ case ICMP6_time_exceeded:
+ case ICMP6_parameter_problem:
+ /* Type 3 - destination unreachable, Code 1 - host unreachable */
+ new_icmp40->type = ICMP4_destination_unreachable;
+ new_icmp40->code =
+ ICMP4_destination_unreachable_destination_unreachable_host;
+ break;
+
+ case ICMP6_packet_too_big:
+ /* Type 3 - destination unreachable, Code 4 - packet too big */
+ /* Potential TODO: Adjust domain tunnel MTU based on the value received here */
+ mtu = clib_net_to_host_u32 (*((u32 *) (icmp60 + 1)));
+
+ /* Check DF flag */
+ if (!
+ (inner_ip40->
+ flags_and_fragment_offset &
+ clib_host_to_net_u16 (IP4_HEADER_FLAG_DONT_FRAGMENT)))
+ {
+ error0 = MAP_ERROR_ICMP_RELAY;
+ goto error;
+ }
+
+ new_icmp40->type = ICMP4_destination_unreachable;
+ new_icmp40->code =
+ ICMP4_destination_unreachable_fragmentation_needed_and_dont_fragment_set;
+ *((u32 *) (new_icmp40 + 1)) =
+ clib_host_to_net_u32 (mtu < 1280 ? 1280 : mtu);
+ break;
+
+ default:
+ error0 = MAP_ERROR_ICMP_RELAY;
+ break;
+ }
+
+ /*
+ * Ensure the total ICMP packet is no longer than 576 bytes (RFC1812)
+ */
+ new_ip40->ip_version_and_header_length = 0x45;
+ new_ip40->tos = 0;
+ u16 nlen = (tlen - 20) > 576 ? 576 : tlen - 20;
+ new_ip40->length = clib_host_to_net_u16 (nlen);
+ new_ip40->fragment_id = fid[0];
+ fid++;
+ new_ip40->ttl = 64;
+ new_ip40->protocol = IP_PROTOCOL_ICMP;
+ new_ip40->src_address = mm->icmp4_src_address;
+ new_ip40->dst_address = inner_ip40->src_address;
+ new_ip40->checksum = ip4_header_checksum (new_ip40);
+
+ new_icmp40->checksum = 0;
+ ip_csum_t sum = ip_incremental_checksum (0, new_icmp40, nlen - 20);
+ new_icmp40->checksum = ~ip_csum_fold (sum);
+
+ vlib_increment_simple_counter (&mm->icmp_relayed, cpu_index, 0, 1);
+
+ error:
+ if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED))
+ {
+ map_trace_t *tr = vlib_add_trace (vm, node, p0, sizeof (*tr));
+ tr->map_domain_index = 0;
+ tr->port = 0;
+ }
+
+ next0 =
+ (error0 == MAP_ERROR_NONE) ? next0 : IP6_ICMP_RELAY_NEXT_DROP;
+ p0->error = error_node->errors[error0];
+ vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
+ n_left_to_next, pi0, next0);
}
-
- new_icmp40->type = ICMP4_destination_unreachable;
- new_icmp40->code = ICMP4_destination_unreachable_fragmentation_needed_and_dont_fragment_set;
- *((u32 *)(new_icmp40 + 1)) = clib_host_to_net_u32(mtu < 1280 ? 1280 : mtu);
- break;
-
- default:
- error0 = MAP_ERROR_ICMP_RELAY;
- break;
- }
-
- /*
- * Ensure the total ICMP packet is no longer than 576 bytes (RFC1812)
- */
- new_ip40->ip_version_and_header_length = 0x45;
- new_ip40->tos = 0;
- u16 nlen = (tlen - 20) > 576 ? 576 : tlen - 20;
- new_ip40->length = clib_host_to_net_u16(nlen);
- new_ip40->fragment_id = fid[0]; fid++;
- new_ip40->ttl = 64;
- new_ip40->protocol = IP_PROTOCOL_ICMP;
- new_ip40->src_address = mm->icmp4_src_address;
- new_ip40->dst_address = inner_ip40->src_address;
- new_ip40->checksum = ip4_header_checksum(new_ip40);
-
- new_icmp40->checksum = 0;
- ip_csum_t sum = ip_incremental_checksum(0, new_icmp40, nlen - 20);
- new_icmp40->checksum = ~ip_csum_fold(sum);
-
- vlib_increment_simple_counter(&mm->icmp_relayed, cpu_index, 0, 1);
-
- error:
- if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED)) {
- map_trace_t *tr = vlib_add_trace(vm, node, p0, sizeof(*tr));
- tr->map_domain_index = 0;
- tr->port = 0;
- }
-
- next0 = (error0 == MAP_ERROR_NONE) ? next0 : IP6_ICMP_RELAY_NEXT_DROP;
- p0->error = error_node->errors[error0];
- vlib_validate_buffer_enqueue_x1(vm, node, next_index, to_next, n_left_to_next, pi0, next0);
+ vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
- vlib_put_next_frame(vm, node, next_index, n_left_to_next);
- }
return frame->n_vectors;
@@ -932,6 +1181,7 @@ static char *map_error_strings[] = {
#undef _
};
+/* *INDENT-OFF* */
VLIB_REGISTER_NODE(ip6_map_node) = {
.function = ip6_map,
.name = "ip6-map",
@@ -957,7 +1207,9 @@ VLIB_REGISTER_NODE(ip6_map_node) = {
[IP6_MAP_NEXT_ICMP] = "ip6-icmp-error",
},
};
+/* *INDENT-ON* */
+/* *INDENT-OFF* */
VLIB_REGISTER_NODE(ip6_map_ip6_reass_node) = {
.function = ip6_map_ip6_reass,
.name = "ip6-map-ip6-reass",
@@ -972,7 +1224,9 @@ VLIB_REGISTER_NODE(ip6_map_ip6_reass_node) = {
[IP6_MAP_IP6_REASS_NEXT_DROP] = "error-drop",
},
};
+/* *INDENT-ON* */
+/* *INDENT-OFF* */
VLIB_REGISTER_NODE(ip6_map_ip4_reass_node) = {
.function = ip6_map_ip4_reass,
.name = "ip6-map-ip4-reass",
@@ -988,7 +1242,9 @@ VLIB_REGISTER_NODE(ip6_map_ip4_reass_node) = {
[IP6_MAP_IP4_REASS_NEXT_DROP] = "error-drop",
},
};
+/* *INDENT-ON* */
+/* *INDENT-OFF* */
VLIB_REGISTER_NODE(ip6_map_icmp_relay_node, static) = {
.function = ip6_map_icmp_relay,
.name = "ip6-map-icmp-relay",
@@ -1003,3 +1259,12 @@ VLIB_REGISTER_NODE(ip6_map_icmp_relay_node, static) = {
[IP6_ICMP_RELAY_NEXT_DROP] = "error-drop",
},
};
+/* *INDENT-ON* */
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */