/* * Copyright (c) 2016 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include <lb/lb.h> #include <vnet/gre/packet.h> #include <lb/lbhash.h> #define foreach_lb_error \ _(NONE, "no error") \ _(PROTO_NOT_SUPPORTED, "protocol not supported") typedef enum { #define _(sym,str) LB_ERROR_##sym, foreach_lb_error #undef _ LB_N_ERROR, } lb_error_t; static char *lb_error_strings[] = { #define _(sym,string) string, foreach_lb_error #undef _ }; typedef struct { u32 vip_index; u32 as_index; } lb_trace_t; u8 * format_lb_trace (u8 * s, va_list * args) { lb_main_t *lbm = &lb_main; CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); lb_trace_t *t = va_arg (*args, lb_trace_t *); if (pool_is_free_index(lbm->vips, t->vip_index)) { s = format(s, "lb vip[%d]: This VIP was freed since capture\n"); } else { s = format(s, "lb vip[%d]: %U\n", t->vip_index, format_lb_vip, &lbm->vips[t->vip_index]); } if (pool_is_free_index(lbm->ass, t->as_index)) { s = format(s, "lb as[%d]: This AS was freed since capture\n"); } else { s = format(s, "lb as[%d]: %U\n", t->as_index, format_lb_as, &lbm->ass[t->as_index]); } return s; } lb_hash_t *lb_get_sticky_table(u32 thread_index) { lb_main_t *lbm = &lb_main; lb_hash_t *sticky_ht = lbm->per_cpu[thread_index].sticky_ht; //Check if size changed if (PREDICT_FALSE(sticky_ht && (lbm->per_cpu_sticky_buckets != lb_hash_nbuckets(sticky_ht)))) { //Dereference everything in there lb_hash_bucket_t *b; u32 i; lb_hash_foreach_entry(sticky_ht, b, i) { vlib_refcount_add(&lbm->as_refcount, thread_index, b->value[i], -1); vlib_refcount_add(&lbm->as_refcount, thread_index, 0, 1); } lb_hash_free(sticky_ht); sticky_ht = NULL; } //Create if necessary if (PREDICT_FALSE(sticky_ht == NULL)) { lbm->per_cpu[thread_index].sticky_ht = lb_hash_alloc(lbm->per_cpu_sticky_buckets, lbm->flow_timeout); sticky_ht = lbm->per_cpu[thread_index].sticky_ht; clib_warning("Regenerated sticky table %p", sticky_ht); } ASSERT(sticky_ht); //Update timeout sticky_ht->timeout = lbm->flow_timeout; return sticky_ht; } u64 lb_node_get_other_ports4(ip4_header_t *ip40) { return 0; } u64 lb_node_get_other_ports6(ip6_header_t *ip60) { return 0; } static_always_inline u32 lb_node_get_hash(vlib_buffer_t *p, u8 is_input_v4) { u32 hash; if (is_input_v4) { ip4_header_t *ip40; u64 ports; ip40 = vlib_buffer_get_current (p); if (PREDICT_TRUE (ip40->protocol == IP_PROTOCOL_TCP || ip40->protocol == IP_PROTOCOL_UDP)) ports = ((u64)((udp_header_t *)(ip40 + 1))->src_port << 16) | ((u64)((udp_header_t *)(ip40 + 1))->dst_port); else ports = lb_node_get_other_ports4(ip40); hash = lb_hash_hash(*((u64 *)&ip40->address_pair), ports, 0, 0, 0); } else { ip6_header_t *ip60; ip60 = vlib_buffer_get_current (p); u64 ports; if (PREDICT_TRUE (ip60->protocol == IP_PROTOCOL_TCP || ip60->protocol == IP_PROTOCOL_UDP)) ports = ((u64)((udp_header_t *)(ip60 + 1))->src_port << 16) | ((u64)((udp_header_t *)(ip60 + 1))->dst_port); else ports = lb_node_get_other_ports6(ip60); hash = lb_hash_hash(ip60->src_address.as_u64[0], ip60->src_address.as_u64[1], ip60->dst_address.as_u64[0], ip60->dst_address.as_u64[1], ports); } return hash; } static_always_inline uword lb_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame, u8 is_input_v4, //Compile-time parameter stating that is input is v4 (or v6) lb_encap_type_t encap_type) //Compile-time parameter stating that is GRE4 or GRE6 or L3DSR { lb_main_t *lbm = &lb_main; u32 n_left_from, *from, next_index, *to_next, n_left_to_next; u32 thread_index = vlib_get_thread_index(); u32 lb_time = lb_hash_time_now(vm); lb_hash_t *sticky_ht = lb_get_sticky_table(thread_index); from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; next_index = node->cached_next_index; u32 nexthash0 = 0; if (PREDICT_TRUE(n_left_from > 0)) nexthash0 = lb_node_get_hash(vlib_get_buffer (vm, from[0]), is_input_v4); while (n_left_from > 0) { vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); while (n_left_from > 0 && n_left_to_next > 0) { u32 pi0; vlib_buffer_t *p0; lb_vip_t *vip0; u32 asindex0; u16 len0; u32 available_index0; u8 counter = 0; u32 hash0 = nexthash0; if (PREDICT_TRUE(n_left_from > 1)) { vlib_buffer_t *p1 = vlib_get_buffer (vm, from[1]); //Compute next hash and prefetch bucket nexthash0 = lb_node_get_hash(p1, is_input_v4); lb_hash_prefetch_bucket(sticky_ht, nexthash0); //Prefetch for encap, next CLIB_PREFETCH (vlib_buffer_get_current(p1) - 64, 64, STORE); } if (PREDICT_TRUE(n_left_from > 2)) { vlib_buffer_t *p2; p2 = vlib_get_buffer(vm, from[2]); /* prefetch packet header and data */ vlib_prefetch_buffer_header(p2, STORE); CLIB_PREFETCH (vlib_buffer_get_current(p2), 64, STORE); } pi0 = to_next[0] = from[0]; from += 1; n_left_from -= 1; to_next += 1; n_left_to_next -= 1; p0 = vlib_get_buffer (vm, pi0); vip0 = pool_elt_at_index (lbm->vips, vnet_buffer (p0)->ip.adj_index[VLIB_TX]); if (is_input_v4) { ip4_header_t *ip40; ip40 = vlib_buffer_get_current (p0); len0 = clib_net_to_host_u16(ip40->length); } else { ip6_header_t *ip60; ip60 = vlib_buffer_get_current (p0); len0 = clib_net_to_host_u16(ip60->payload_length) + sizeof(ip6_header_t); } lb_hash_get(sticky_ht, hash0, vnet_buffer (p0)->ip.adj_index[VLIB_TX], lb_time, &available_index0, &asindex0); if (PREDICT_TRUE(asindex0 != ~0)) { //Found an existing entry counter = LB_VIP_COUNTER_NEXT_PACKET; } else if (PREDICT_TRUE(available_index0 != ~0)) { //There is an available slot for a new flow asindex0 = vip0->new_flow_table[hash0 & vip0->new_flow_table_mask].as_index; counter = LB_VIP_COUNTER_FIRST_PACKET; counter = (asindex0 == 0)?LB_VIP_COUNTER_NO_SERVER:counter; //TODO: There are race conditions with as0 and vip0 manipulation. //Configuration may be changed, vectors resized, etc... //Dereference previously used vlib_refcount_add(&lbm->as_refcount, thread_index, lb_hash_available_value(sticky_ht, hash0, available_index0), -1); vlib_refcount_add(&lbm->as_refcount, thread_index, asindex0, 1); //Add sticky entry //Note that when there is no AS configured, an entry is configured anyway. //But no configured AS is not something that should happen lb_hash_put(sticky_ht, hash0, asindex0, vnet_buffer (p0)->ip.adj_index[VLIB_TX], available_index0, lb_time); } else { //Could not store new entry in the table asindex0 = vip0->new_flow_table[hash0 & vip0->new_flow_table_mask].as_index; counter = LB_VIP_COUNTER_UNTRACKED_PACKET; } vlib_increment_simple_counter(&lbm->vip_counters[counter], thread_index, vnet_buffer (p0)->ip.adj_index[VLIB_TX], 1); //Now let's encap if ( (encap_type == LB_ENCAP_TYPE_GRE4) || (encap_type == LB_ENCAP_TYPE_GRE6) ) { gre_header_t *gre0; if (encap_type == LB_ENCAP_TYPE_GRE4) /* encap GRE4*/ { ip4_header_t *ip40; vlib_buffer_advance(p0, - sizeof(ip4_header_t) - sizeof(gre_header_t)); ip40 = vlib_buffer_get_current(p0); gre0 = (gre_header_t *)(ip40 + 1); ip40->src_address = lbm->ip4_src_address; ip40->dst_address = lbm->ass[asindex0].address.ip4; ip40->ip_version_and_header_length = 0x45; ip40->ttl = 128; ip40->fragment_id = 0; ip40->flags_and_fragment_offset = 0; ip40->length = clib_host_to_net_u16(len0 + sizeof(gre_header_t) + sizeof(ip4_header_t)); ip40->protocol = IP_PROTOCOL_GRE; ip40->checksum = ip4_header_checksum (ip40); } else /* encap GRE6*/ { ip6_header_t *ip60; vlib_buffer_advance(p0, - sizeof(ip6_header_t) - sizeof(gre_header_t)); ip60 = vlib_buffer_get_current(p0); gre0 = (gre_header_t *)(ip60 + 1); ip60->dst_address = lbm->ass[asindex0].address.ip6; ip60->src_address = lbm->ip6_src_address; ip60->hop_limit = 128; ip60->ip_version_traffic_class_and_flow_label = clib_host_to_net_u32 (0x6<<28); ip60->payload_length = clib_host_to_net_u16(len0 + sizeof(gre_header_t)); ip60->protocol = IP_PROTOCOL_GRE; } gre0->flags_and_version = 0; gre0->protocol = (is_input_v4)? clib_host_to_net_u16(0x0800): clib_host_to_net_u16(0x86DD); } else if (encap_type == LB_ENCAP_TYPE_L3DSR) /* encap L3DSR*/ { ip4_header_t *ip40; tcp_header_t *th0; ip40 = vlib_buffer_get_current(p0); ip40->dst_address = lbm->ass[asindex0].address.ip4; /* Get and rewrite DSCP bit */ ip40->tos = (u8)((vip0->dscp & 0x3F)<<2); ip40->checksum = ip4_header_checksum (ip40); /* Recomputing L4 checksum after dst-IP modifying */ th0 = ip4_next_header(ip40); th0->checksum = 0; th0->checksum = ip4_tcp_udp_compute_checksum(vm, p0, ip40); } if (PREDICT_FALSE (p0->flags & VLIB_BUFFER_IS_TRACED)) { lb_trace_t *tr = vlib_add_trace (vm, node, p0, sizeof (*tr)); tr->as_index = asindex0; tr->vip_index = vnet_buffer (p0)->ip.adj_index[VLIB_TX]; } //Enqueue to next //Note that this is going to error if asindex0 == 0 vnet_buffer (p0)->ip.adj_index[VLIB_TX] = lbm->ass[asindex0].dpo.dpoi_index; vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, n_left_to_next, pi0, lbm->ass[asindex0].dpo.dpoi_next_node); } vlib_put_next_frame (vm, node, next_index, n_left_to_next); } return frame->n_vectors; } static uword lb6_gre6_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { return lb_node_fn(vm, node, frame, 0, LB_ENCAP_TYPE_GRE6); } static uword lb6_gre4_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { return lb_node_fn(vm, node, frame, 0, LB_ENCAP_TYPE_GRE4); } static uword lb4_gre6_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { return lb_node_fn(vm, node, frame, 1, LB_ENCAP_TYPE_GRE6); } static uword lb4_gre4_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { return lb_node_fn(vm, node, frame, 1, LB_ENCAP_TYPE_GRE4); } static uword lb4_l3dsr_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { return lb_node_fn(vm, node, frame, 1, LB_ENCAP_TYPE_L3DSR); } VLIB_REGISTER_NODE (lb6_gre6_node) = { .function = lb6_gre6_node_fn, .name = "lb6-gre6", .vector_size = sizeof (u32), .format_trace = format_lb_trace, .n_errors = LB_N_ERROR, .error_strings = lb_error_strings, .n_next_nodes = LB_N_NEXT, .next_nodes = { [LB_NEXT_DROP] = "error-drop" }, }; VLIB_REGISTER_NODE (lb6_gre4_node) = { .function = lb6_gre4_node_fn, .name = "lb6-gre4", .vector_size = sizeof (u32), .format_trace = format_lb_trace, .n_errors = LB_N_ERROR, .error_strings = lb_error_strings, .n_next_nodes = LB_N_NEXT, .next_nodes = { [LB_NEXT_DROP] = "error-drop" }, }; VLIB_REGISTER_NODE (lb4_gre6_node) = { .function = lb4_gre6_node_fn, .name = "lb4-gre6", .vector_size = sizeof (u32), .format_trace = format_lb_trace, .n_errors = LB_N_ERROR, .error_strings = lb_error_strings, .n_next_nodes = LB_N_NEXT, .next_nodes = { [LB_NEXT_DROP] = "error-drop" }, }; VLIB_REGISTER_NODE (lb4_gre4_node) = { .function = lb4_gre4_node_fn, .name = "lb4-gre4", .vector_size = sizeof (u32), .format_trace = format_lb_trace, .n_errors = LB_N_ERROR, .error_strings = lb_error_strings, .n_next_nodes = LB_N_NEXT, .next_nodes = { [LB_NEXT_DROP] = "error-drop" }, }; VLIB_REGISTER_NODE (lb4_l3dsr_node) = { .function = lb4_l3dsr_node_fn, .name = "lb4-l3dsr", .vector_size = sizeof (u32), .format_trace = format_lb_trace, .n_errors = LB_N_ERROR, .error_strings = lb_error_strings, .n_next_nodes = LB_N_NEXT, .next_nodes = { [LB_NEXT_DROP] = "error-drop" }, };