From 9466e4fed6faaf9c161767e2bc42c9cb0cf2c7ec Mon Sep 17 00:00:00 2001 From: Klement Sekera Date: Wed, 17 Jun 2020 13:46:41 +0000 Subject: nat: replace speculative buffer enqueue model Replace speculative buffer enqueue coding model with vlib_get_buffers(...)/vlib_buffer_enqueue_to_next(...). Type: improvement Change-Id: I7dbfac2234a7bd754c599857eb1d5b601da5bc7c Signed-off-by: Klement Sekera --- src/plugins/nat/in2out.c | 1158 +++++++++++++++++------------------ src/plugins/nat/in2out_ed.c | 821 ++++++++++++------------- src/plugins/nat/nat_inlines.h | 157 +++-- src/plugins/nat/out2in.c | 1334 ++++++++++++++++++++--------------------- src/plugins/nat/out2in_ed.c | 915 ++++++++++++++-------------- 5 files changed, 2107 insertions(+), 2278 deletions(-) (limited to 'src/plugins') diff --git a/src/plugins/nat/in2out.c b/src/plugins/nat/in2out.c index f904d02b344..06a2fa1737a 100644 --- a/src/plugins/nat/in2out.c +++ b/src/plugins/nat/in2out.c @@ -874,8 +874,7 @@ snat_in2out_node_fn_inline (vlib_main_t * vm, vlib_frame_t * frame, int is_slow_path, int is_output_feature) { - u32 n_left_from, *from, *to_next; - snat_in2out_next_t next_index; + u32 n_left_from, *from; u32 pkts_processed = 0; snat_main_t *sm = &snat_main; f64 now = vlib_time_now (vm); @@ -889,728 +888,699 @@ snat_in2out_node_fn_inline (vlib_main_t * vm, from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; - next_index = node->cached_next_index; - while (n_left_from > 0) + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; + u16 nexts[VLIB_FRAME_SIZE], *next = nexts; + vlib_get_buffers (vm, from, b, n_left_from); + + while (n_left_from >= 2) { - u32 n_left_to_next; + vlib_buffer_t *b0, *b1; + u32 next0, next1; + u32 sw_if_index0, sw_if_index1; + ip4_header_t *ip0, *ip1; + ip_csum_t sum0, sum1; + u32 new_addr0, old_addr0, new_addr1, old_addr1; + u16 old_port0, new_port0, old_port1, new_port1; + udp_header_t *udp0, *udp1; + tcp_header_t *tcp0, *tcp1; + icmp46_header_t *icmp0, *icmp1; + u32 rx_fib_index0, rx_fib_index1; + u32 proto0, proto1; + snat_session_t *s0 = 0, *s1 = 0; + clib_bihash_kv_8_8_t kv0, value0, kv1, value1; + u32 iph_offset0 = 0, iph_offset1 = 0; + + b0 = *b; + b++; + b1 = *b; + b++; + + /* Prefetch next iteration. */ + if (PREDICT_TRUE (n_left_from >= 4)) + { + vlib_buffer_t *p2, *p3; - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + p2 = *b; + p3 = *(b + 1); - while (n_left_from >= 4 && n_left_to_next >= 2) - { - u32 bi0, bi1; - vlib_buffer_t *b0, *b1; - u32 next0, next1; - u32 sw_if_index0, sw_if_index1; - ip4_header_t *ip0, *ip1; - ip_csum_t sum0, sum1; - u32 new_addr0, old_addr0, new_addr1, old_addr1; - u16 old_port0, new_port0, old_port1, new_port1; - udp_header_t *udp0, *udp1; - tcp_header_t *tcp0, *tcp1; - icmp46_header_t *icmp0, *icmp1; - u32 rx_fib_index0, rx_fib_index1; - u32 proto0, proto1; - snat_session_t *s0 = 0, *s1 = 0; - clib_bihash_kv_8_8_t kv0, value0, kv1, value1; - u32 iph_offset0 = 0, iph_offset1 = 0; - - /* Prefetch next iteration. */ - { - vlib_buffer_t *p2, *p3; - - p2 = vlib_get_buffer (vm, from[2]); - p3 = vlib_get_buffer (vm, from[3]); - - vlib_prefetch_buffer_header (p2, LOAD); - vlib_prefetch_buffer_header (p3, LOAD); - - CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, LOAD); - CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, LOAD); - } - - /* speculatively enqueue b0 and b1 to the current next frame */ - to_next[0] = bi0 = from[0]; - to_next[1] = bi1 = from[1]; - from += 2; - to_next += 2; - n_left_from -= 2; - n_left_to_next -= 2; + vlib_prefetch_buffer_header (p2, LOAD); + vlib_prefetch_buffer_header (p3, LOAD); - b0 = vlib_get_buffer (vm, bi0); - b1 = vlib_get_buffer (vm, bi1); + CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, LOAD); + CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, LOAD); + } - if (is_output_feature) - iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length; + if (is_output_feature) + iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length; - ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) + - iph_offset0); + ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) + + iph_offset0); - udp0 = ip4_next_header (ip0); - tcp0 = (tcp_header_t *) udp0; - icmp0 = (icmp46_header_t *) udp0; + udp0 = ip4_next_header (ip0); + tcp0 = (tcp_header_t *) udp0; + icmp0 = (icmp46_header_t *) udp0; - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - rx_fib_index0 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, - sw_if_index0); + sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + rx_fib_index0 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, + sw_if_index0); - next0 = next1 = SNAT_IN2OUT_NEXT_LOOKUP; + next0 = next1 = SNAT_IN2OUT_NEXT_LOOKUP; - if (PREDICT_FALSE (ip0->ttl == 1)) - { - vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next0 = SNAT_IN2OUT_NEXT_ICMP_ERROR; - goto trace00; - } + if (PREDICT_FALSE (ip0->ttl == 1)) + { + vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next0 = SNAT_IN2OUT_NEXT_ICMP_ERROR; + goto trace00; + } - proto0 = ip_proto_to_nat_proto (ip0->protocol); + proto0 = ip_proto_to_nat_proto (ip0->protocol); - /* Next configured feature, probably ip4-lookup */ - if (is_slow_path) + /* Next configured feature, probably ip4-lookup */ + if (is_slow_path) + { + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) { - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) + if (nat_in2out_sm_unknown_proto (sm, b0, ip0, rx_fib_index0)) { - if (nat_in2out_sm_unknown_proto - (sm, b0, ip0, rx_fib_index0)) - { - next0 = SNAT_IN2OUT_NEXT_DROP; - b0->error = - node->errors[SNAT_IN2OUT_ERROR_UNSUPPORTED_PROTOCOL]; - } - other_packets++; - goto trace00; + next0 = SNAT_IN2OUT_NEXT_DROP; + b0->error = + node->errors[SNAT_IN2OUT_ERROR_UNSUPPORTED_PROTOCOL]; } + other_packets++; + goto trace00; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) - { - next0 = icmp_in2out_slow_path - (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, - node, next0, now, thread_index, &s0); - icmp_packets++; - goto trace00; - } + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + { + next0 = icmp_in2out_slow_path + (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, + node, next0, now, thread_index, &s0); + icmp_packets++; + goto trace00; } - else + } + else + { + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) { - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) - { - next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; - goto trace00; - } + next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; + goto trace00; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) - { - next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; - goto trace00; - } + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + { + next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; + goto trace00; } + } - init_nat_k (&kv0, ip0->src_address, - vnet_buffer (b0)->ip.reass.l4_src_port, rx_fib_index0, - proto0); - if (PREDICT_FALSE - (clib_bihash_search_8_8 - (&sm->per_thread_data[thread_index].in2out, &kv0, - &value0) != 0)) + init_nat_k (&kv0, ip0->src_address, + vnet_buffer (b0)->ip.reass.l4_src_port, rx_fib_index0, + proto0); + if (PREDICT_FALSE + (clib_bihash_search_8_8 + (&sm->per_thread_data[thread_index].in2out, &kv0, &value0) != 0)) + { + if (is_slow_path) { - if (is_slow_path) + if (is_output_feature) { - if (is_output_feature) - { - if (PREDICT_FALSE - (nat_not_translate_output_feature - (sm, ip0, proto0, - vnet_buffer (b0)->ip.reass.l4_src_port, - vnet_buffer (b0)->ip.reass.l4_dst_port, - thread_index, sw_if_index0))) - goto trace00; - - /* - * Send DHCP packets to the ipv4 stack, or we won't - * be able to use dhcp client on the outside interface - */ - if (PREDICT_FALSE - (proto0 == NAT_PROTOCOL_UDP - && (vnet_buffer (b0)->ip.reass.l4_dst_port == - clib_host_to_net_u16 - (UDP_DST_PORT_dhcp_to_server)) - && ip0->dst_address.as_u32 == 0xffffffff)) - goto trace00; - } - else - { - if (PREDICT_FALSE - (snat_not_translate - (sm, node, sw_if_index0, ip0, proto0, - rx_fib_index0, thread_index))) - goto trace00; - } - - next0 = slow_path (sm, b0, ip0, - ip0->src_address, - vnet_buffer (b0)->ip.reass.l4_src_port, - rx_fib_index0, - proto0, - &s0, node, next0, thread_index, now); - if (PREDICT_FALSE (next0 == SNAT_IN2OUT_NEXT_DROP)) + if (PREDICT_FALSE + (nat_not_translate_output_feature + (sm, ip0, proto0, + vnet_buffer (b0)->ip.reass.l4_src_port, + vnet_buffer (b0)->ip.reass.l4_dst_port, + thread_index, sw_if_index0))) goto trace00; - if (PREDICT_FALSE (!s0)) + /* + * Send DHCP packets to the ipv4 stack, or we won't + * be able to use dhcp client on the outside interface + */ + if (PREDICT_FALSE + (proto0 == NAT_PROTOCOL_UDP + && (vnet_buffer (b0)->ip.reass.l4_dst_port == + clib_host_to_net_u16 + (UDP_DST_PORT_dhcp_to_server)) + && ip0->dst_address.as_u32 == 0xffffffff)) goto trace00; } else { - next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; - goto trace00; + if (PREDICT_FALSE + (snat_not_translate + (sm, node, sw_if_index0, ip0, proto0, + rx_fib_index0, thread_index))) + goto trace00; } + + next0 = slow_path (sm, b0, ip0, + ip0->src_address, + vnet_buffer (b0)->ip.reass.l4_src_port, + rx_fib_index0, + proto0, &s0, node, next0, thread_index, now); + if (PREDICT_FALSE (next0 == SNAT_IN2OUT_NEXT_DROP)) + goto trace00; + + if (PREDICT_FALSE (!s0)) + goto trace00; } else - s0 = - pool_elt_at_index (sm->per_thread_data[thread_index].sessions, - value0.value); + { + next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; + goto trace00; + } + } + else + s0 = + pool_elt_at_index (sm->per_thread_data[thread_index].sessions, + value0.value); - b0->flags |= VNET_BUFFER_F_IS_NATED; + b0->flags |= VNET_BUFFER_F_IS_NATED; - old_addr0 = ip0->src_address.as_u32; - ip0->src_address = s0->out2in.addr; - new_addr0 = ip0->src_address.as_u32; - if (!is_output_feature) - vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index; + old_addr0 = ip0->src_address.as_u32; + ip0->src_address = s0->out2in.addr; + new_addr0 = ip0->src_address.as_u32; + if (!is_output_feature) + vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index; - sum0 = ip0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - src_address /* changed member */ ); - ip0->checksum = ip_csum_fold (sum0); + sum0 = ip0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, src_address /* changed member */ ); + ip0->checksum = ip_csum_fold (sum0); - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; - new_port0 = udp0->src_port = s0->out2in.port; - sum0 = tcp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - sum0 = ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - mss_clamping (sm->mss_clamping, tcp0, &sum0); - tcp0->checksum = ip_csum_fold (sum0); - } - tcp_packets++; + old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; + new_port0 = udp0->src_port = s0->out2in.port; + sum0 = tcp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, + dst_address /* changed member */ ); + sum0 = ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + mss_clamping (sm->mss_clamping, tcp0, &sum0); + tcp0->checksum = ip_csum_fold (sum0); } - else + tcp_packets++; + } + else + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + udp0->src_port = s0->out2in.port; + if (PREDICT_FALSE (udp0->checksum)) { - udp0->src_port = s0->out2in.port; - if (PREDICT_FALSE (udp0->checksum)) - { - old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; - new_port0 = udp0->src_port; - sum0 = udp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, dst_address /* changed member */ - ); - sum0 = - ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - udp0->checksum = ip_csum_fold (sum0); - } + old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; + new_port0 = udp0->src_port; + sum0 = udp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, dst_address /* changed member */ + ); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + udp0->checksum = ip_csum_fold (sum0); } - udp_packets++; } + udp_packets++; + } - /* Accounting */ - nat44_session_update_counters (s0, now, - vlib_buffer_length_in_chain (vm, b0), - thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s0, thread_index); - trace00: + /* Accounting */ + nat44_session_update_counters (s0, now, + vlib_buffer_length_in_chain (vm, b0), + thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s0, thread_index); + trace00: - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) - { - snat_in2out_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->is_slow_path = is_slow_path; - t->sw_if_index = sw_if_index0; - t->next_index = next0; - t->session_index = ~0; - if (s0) - t->session_index = - s0 - sm->per_thread_data[thread_index].sessions; - } + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + snat_in2out_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); + t->is_slow_path = is_slow_path; + t->sw_if_index = sw_if_index0; + t->next_index = next0; + t->session_index = ~0; + if (s0) + t->session_index = + s0 - sm->per_thread_data[thread_index].sessions; + } - pkts_processed += next0 == SNAT_IN2OUT_NEXT_LOOKUP; + pkts_processed += next0 == SNAT_IN2OUT_NEXT_LOOKUP; - if (is_output_feature) - iph_offset1 = vnet_buffer (b1)->ip.reass.save_rewrite_length; + if (is_output_feature) + iph_offset1 = vnet_buffer (b1)->ip.reass.save_rewrite_length; - ip1 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b1) + - iph_offset1); + ip1 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b1) + + iph_offset1); - udp1 = ip4_next_header (ip1); - tcp1 = (tcp_header_t *) udp1; - icmp1 = (icmp46_header_t *) udp1; + udp1 = ip4_next_header (ip1); + tcp1 = (tcp_header_t *) udp1; + icmp1 = (icmp46_header_t *) udp1; - sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX]; - rx_fib_index1 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, - sw_if_index1); + sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX]; + rx_fib_index1 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, + sw_if_index1); - if (PREDICT_FALSE (ip1->ttl == 1)) - { - vnet_buffer (b1)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b1, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next1 = SNAT_IN2OUT_NEXT_ICMP_ERROR; - goto trace01; - } + if (PREDICT_FALSE (ip1->ttl == 1)) + { + vnet_buffer (b1)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b1, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next1 = SNAT_IN2OUT_NEXT_ICMP_ERROR; + goto trace01; + } - proto1 = ip_proto_to_nat_proto (ip1->protocol); + proto1 = ip_proto_to_nat_proto (ip1->protocol); - /* Next configured feature, probably ip4-lookup */ - if (is_slow_path) + /* Next configured feature, probably ip4-lookup */ + if (is_slow_path) + { + if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_OTHER)) { - if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_OTHER)) + if (nat_in2out_sm_unknown_proto (sm, b1, ip1, rx_fib_index1)) { - if (nat_in2out_sm_unknown_proto - (sm, b1, ip1, rx_fib_index1)) - { - next1 = SNAT_IN2OUT_NEXT_DROP; - b1->error = - node->errors[SNAT_IN2OUT_ERROR_UNSUPPORTED_PROTOCOL]; - } - other_packets++; - goto trace01; + next1 = SNAT_IN2OUT_NEXT_DROP; + b1->error = + node->errors[SNAT_IN2OUT_ERROR_UNSUPPORTED_PROTOCOL]; } + other_packets++; + goto trace01; + } - if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_ICMP)) - { - next1 = icmp_in2out_slow_path - (sm, b1, ip1, icmp1, sw_if_index1, rx_fib_index1, node, - next1, now, thread_index, &s1); - icmp_packets++; - goto trace01; - } + if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_ICMP)) + { + next1 = icmp_in2out_slow_path + (sm, b1, ip1, icmp1, sw_if_index1, rx_fib_index1, node, + next1, now, thread_index, &s1); + icmp_packets++; + goto trace01; } - else + } + else + { + if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_OTHER)) { - if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_OTHER)) - { - next1 = SNAT_IN2OUT_NEXT_SLOW_PATH; - goto trace01; - } + next1 = SNAT_IN2OUT_NEXT_SLOW_PATH; + goto trace01; + } - if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_ICMP)) - { - next1 = SNAT_IN2OUT_NEXT_SLOW_PATH; - goto trace01; - } + if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_ICMP)) + { + next1 = SNAT_IN2OUT_NEXT_SLOW_PATH; + goto trace01; } + } - init_nat_k (&kv1, ip1->src_address, - vnet_buffer (b1)->ip.reass.l4_src_port, rx_fib_index1, - proto1); - if (PREDICT_FALSE - (clib_bihash_search_8_8 - (&sm->per_thread_data[thread_index].in2out, &kv1, - &value1) != 0)) + init_nat_k (&kv1, ip1->src_address, + vnet_buffer (b1)->ip.reass.l4_src_port, rx_fib_index1, + proto1); + if (PREDICT_FALSE + (clib_bihash_search_8_8 + (&sm->per_thread_data[thread_index].in2out, &kv1, &value1) != 0)) + { + if (is_slow_path) { - if (is_slow_path) + if (is_output_feature) { - if (is_output_feature) - { - if (PREDICT_FALSE - (nat_not_translate_output_feature - (sm, ip1, proto1, - vnet_buffer (b1)->ip.reass.l4_src_port, - vnet_buffer (b1)->ip.reass.l4_dst_port, - thread_index, sw_if_index1))) - goto trace01; - - /* - * Send DHCP packets to the ipv4 stack, or we won't - * be able to use dhcp client on the outside interface - */ - if (PREDICT_FALSE - (proto1 == NAT_PROTOCOL_UDP - && (vnet_buffer (b1)->ip.reass.l4_dst_port == - clib_host_to_net_u16 - (UDP_DST_PORT_dhcp_to_server)) - && ip1->dst_address.as_u32 == 0xffffffff)) - goto trace01; - } - else - { - if (PREDICT_FALSE - (snat_not_translate - (sm, node, sw_if_index1, ip1, proto1, - rx_fib_index1, thread_index))) - goto trace01; - } - - next1 = - slow_path (sm, b1, ip1, ip1->src_address, - vnet_buffer (b1)->ip.reass.l4_src_port, - rx_fib_index1, proto1, &s1, node, next1, - thread_index, now); - if (PREDICT_FALSE (next1 == SNAT_IN2OUT_NEXT_DROP)) + if (PREDICT_FALSE + (nat_not_translate_output_feature + (sm, ip1, proto1, + vnet_buffer (b1)->ip.reass.l4_src_port, + vnet_buffer (b1)->ip.reass.l4_dst_port, + thread_index, sw_if_index1))) goto trace01; - if (PREDICT_FALSE (!s1)) + /* + * Send DHCP packets to the ipv4 stack, or we won't + * be able to use dhcp client on the outside interface + */ + if (PREDICT_FALSE + (proto1 == NAT_PROTOCOL_UDP + && (vnet_buffer (b1)->ip.reass.l4_dst_port == + clib_host_to_net_u16 + (UDP_DST_PORT_dhcp_to_server)) + && ip1->dst_address.as_u32 == 0xffffffff)) goto trace01; } else { - next1 = SNAT_IN2OUT_NEXT_SLOW_PATH; - goto trace01; + if (PREDICT_FALSE + (snat_not_translate + (sm, node, sw_if_index1, ip1, proto1, + rx_fib_index1, thread_index))) + goto trace01; } + + next1 = + slow_path (sm, b1, ip1, ip1->src_address, + vnet_buffer (b1)->ip.reass.l4_src_port, + rx_fib_index1, proto1, &s1, node, next1, + thread_index, now); + if (PREDICT_FALSE (next1 == SNAT_IN2OUT_NEXT_DROP)) + goto trace01; + + if (PREDICT_FALSE (!s1)) + goto trace01; } else - s1 = - pool_elt_at_index (sm->per_thread_data[thread_index].sessions, - value1.value); + { + next1 = SNAT_IN2OUT_NEXT_SLOW_PATH; + goto trace01; + } + } + else + s1 = + pool_elt_at_index (sm->per_thread_data[thread_index].sessions, + value1.value); - b1->flags |= VNET_BUFFER_F_IS_NATED; + b1->flags |= VNET_BUFFER_F_IS_NATED; - old_addr1 = ip1->src_address.as_u32; - ip1->src_address = s1->out2in.addr; - new_addr1 = ip1->src_address.as_u32; - if (!is_output_feature) - vnet_buffer (b1)->sw_if_index[VLIB_TX] = s1->out2in.fib_index; + old_addr1 = ip1->src_address.as_u32; + ip1->src_address = s1->out2in.addr; + new_addr1 = ip1->src_address.as_u32; + if (!is_output_feature) + vnet_buffer (b1)->sw_if_index[VLIB_TX] = s1->out2in.fib_index; - sum1 = ip1->checksum; - sum1 = ip_csum_update (sum1, old_addr1, new_addr1, - ip4_header_t, - src_address /* changed member */ ); - ip1->checksum = ip_csum_fold (sum1); + sum1 = ip1->checksum; + sum1 = ip_csum_update (sum1, old_addr1, new_addr1, + ip4_header_t, src_address /* changed member */ ); + ip1->checksum = ip_csum_fold (sum1); - if (PREDICT_TRUE (proto1 == NAT_PROTOCOL_TCP)) + if (PREDICT_TRUE (proto1 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b1)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - old_port1 = vnet_buffer (b1)->ip.reass.l4_src_port; - new_port1 = udp1->src_port = s1->out2in.port; - sum1 = tcp1->checksum; - sum1 = ip_csum_update (sum1, old_addr1, new_addr1, - ip4_header_t, - dst_address /* changed member */ ); - sum1 = ip_csum_update (sum1, old_port1, new_port1, - ip4_header_t /* cheat */ , - length /* changed member */ ); - mss_clamping (sm->mss_clamping, tcp1, &sum1); - tcp1->checksum = ip_csum_fold (sum1); - } - tcp_packets++; + old_port1 = vnet_buffer (b1)->ip.reass.l4_src_port; + new_port1 = udp1->src_port = s1->out2in.port; + sum1 = tcp1->checksum; + sum1 = ip_csum_update (sum1, old_addr1, new_addr1, + ip4_header_t, + dst_address /* changed member */ ); + sum1 = ip_csum_update (sum1, old_port1, new_port1, + ip4_header_t /* cheat */ , + length /* changed member */ ); + mss_clamping (sm->mss_clamping, tcp1, &sum1); + tcp1->checksum = ip_csum_fold (sum1); } - else + tcp_packets++; + } + else + { + if (!vnet_buffer (b1)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + udp1->src_port = s1->out2in.port; + if (PREDICT_FALSE (udp1->checksum)) { - udp1->src_port = s1->out2in.port; - if (PREDICT_FALSE (udp1->checksum)) - { - old_port1 = vnet_buffer (b1)->ip.reass.l4_src_port; - new_port1 = udp1->src_port; - sum1 = udp1->checksum; - sum1 = ip_csum_update (sum1, old_addr1, new_addr1, ip4_header_t, dst_address /* changed member */ - ); - sum1 = - ip_csum_update (sum1, old_port1, new_port1, - ip4_header_t /* cheat */ , - length /* changed member */ ); - udp1->checksum = ip_csum_fold (sum1); - } + old_port1 = vnet_buffer (b1)->ip.reass.l4_src_port; + new_port1 = udp1->src_port; + sum1 = udp1->checksum; + sum1 = ip_csum_update (sum1, old_addr1, new_addr1, ip4_header_t, dst_address /* changed member */ + ); + sum1 = + ip_csum_update (sum1, old_port1, new_port1, + ip4_header_t /* cheat */ , + length /* changed member */ ); + udp1->checksum = ip_csum_fold (sum1); } - udp_packets++; } - - /* Accounting */ - nat44_session_update_counters (s1, now, - vlib_buffer_length_in_chain (vm, b1), - thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s1, thread_index); - trace01: - - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b1->flags & VLIB_BUFFER_IS_TRACED))) - { - snat_in2out_trace_t *t = - vlib_add_trace (vm, node, b1, sizeof (*t)); - t->sw_if_index = sw_if_index1; - t->next_index = next1; - t->session_index = ~0; - if (s1) - t->session_index = - s1 - sm->per_thread_data[thread_index].sessions; - } - - pkts_processed += next1 == SNAT_IN2OUT_NEXT_LOOKUP; - - /* verify speculative enqueues, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x2 (vm, node, next_index, - to_next, n_left_to_next, - bi0, bi1, next0, next1); + udp_packets++; } - while (n_left_from > 0 && n_left_to_next > 0) - { - u32 bi0; - vlib_buffer_t *b0; - u32 next0; - u32 sw_if_index0; - ip4_header_t *ip0; - ip_csum_t sum0; - u32 new_addr0, old_addr0; - u16 old_port0, new_port0; - udp_header_t *udp0; - tcp_header_t *tcp0; - icmp46_header_t *icmp0; - u32 rx_fib_index0; - u32 proto0; - snat_session_t *s0 = 0; - clib_bihash_kv_8_8_t kv0, value0; - u32 iph_offset0 = 0; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; + /* Accounting */ + nat44_session_update_counters (s1, now, + vlib_buffer_length_in_chain (vm, b1), + thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s1, thread_index); + trace01: - b0 = vlib_get_buffer (vm, bi0); - next0 = SNAT_IN2OUT_NEXT_LOOKUP; + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b1->flags & VLIB_BUFFER_IS_TRACED))) + { + snat_in2out_trace_t *t = vlib_add_trace (vm, node, b1, sizeof (*t)); + t->sw_if_index = sw_if_index1; + t->next_index = next1; + t->session_index = ~0; + if (s1) + t->session_index = + s1 - sm->per_thread_data[thread_index].sessions; + } - if (is_output_feature) - iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length; + pkts_processed += next1 == SNAT_IN2OUT_NEXT_LOOKUP; - ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) + - iph_offset0); + n_left_from -= 2; + next[0] = next0; + next[1] = next1; + next += 2; + } - udp0 = ip4_next_header (ip0); - tcp0 = (tcp_header_t *) udp0; - icmp0 = (icmp46_header_t *) udp0; + while (n_left_from > 0) + { + vlib_buffer_t *b0; + u32 next0; + u32 sw_if_index0; + ip4_header_t *ip0; + ip_csum_t sum0; + u32 new_addr0, old_addr0; + u16 old_port0, new_port0; + udp_header_t *udp0; + tcp_header_t *tcp0; + icmp46_header_t *icmp0; + u32 rx_fib_index0; + u32 proto0; + snat_session_t *s0 = 0; + clib_bihash_kv_8_8_t kv0, value0; + u32 iph_offset0 = 0; + + b0 = *b; + b++; + next0 = SNAT_IN2OUT_NEXT_LOOKUP; + + if (is_output_feature) + iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length; + + ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) + + iph_offset0); + + udp0 = ip4_next_header (ip0); + tcp0 = (tcp_header_t *) udp0; + icmp0 = (icmp46_header_t *) udp0; + + sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + rx_fib_index0 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, + sw_if_index0); + + if (PREDICT_FALSE (ip0->ttl == 1)) + { + vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next0 = SNAT_IN2OUT_NEXT_ICMP_ERROR; + goto trace0; + } - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - rx_fib_index0 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, - sw_if_index0); + proto0 = ip_proto_to_nat_proto (ip0->protocol); - if (PREDICT_FALSE (ip0->ttl == 1)) + /* Next configured feature, probably ip4-lookup */ + if (is_slow_path) + { + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) { - vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next0 = SNAT_IN2OUT_NEXT_ICMP_ERROR; + if (nat_in2out_sm_unknown_proto (sm, b0, ip0, rx_fib_index0)) + { + next0 = SNAT_IN2OUT_NEXT_DROP; + b0->error = + node->errors[SNAT_IN2OUT_ERROR_UNSUPPORTED_PROTOCOL]; + } + other_packets++; goto trace0; } - proto0 = ip_proto_to_nat_proto (ip0->protocol); - - /* Next configured feature, probably ip4-lookup */ - if (is_slow_path) + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) { - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) - { - if (nat_in2out_sm_unknown_proto - (sm, b0, ip0, rx_fib_index0)) - { - next0 = SNAT_IN2OUT_NEXT_DROP; - b0->error = - node->errors[SNAT_IN2OUT_ERROR_UNSUPPORTED_PROTOCOL]; - } - other_packets++; - goto trace0; - } - - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) - { - next0 = icmp_in2out_slow_path - (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, node, - next0, now, thread_index, &s0); - icmp_packets++; - goto trace0; - } + next0 = icmp_in2out_slow_path + (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, node, + next0, now, thread_index, &s0); + icmp_packets++; + goto trace0; } - else + } + else + { + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) { - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) - { - next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; - goto trace0; - } + next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; + goto trace0; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) - { - next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; - goto trace0; - } + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + { + next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; + goto trace0; } + } - init_nat_k (&kv0, ip0->src_address, - vnet_buffer (b0)->ip.reass.l4_src_port, rx_fib_index0, - proto0); + init_nat_k (&kv0, ip0->src_address, + vnet_buffer (b0)->ip.reass.l4_src_port, rx_fib_index0, + proto0); - if (clib_bihash_search_8_8 - (&sm->per_thread_data[thread_index].in2out, &kv0, &value0)) + if (clib_bihash_search_8_8 + (&sm->per_thread_data[thread_index].in2out, &kv0, &value0)) + { + if (is_slow_path) { - if (is_slow_path) + if (is_output_feature) { - if (is_output_feature) - { - if (PREDICT_FALSE - (nat_not_translate_output_feature - (sm, ip0, proto0, - vnet_buffer (b0)->ip.reass.l4_src_port, - vnet_buffer (b0)->ip.reass.l4_dst_port, - thread_index, sw_if_index0))) - goto trace0; - - /* - * Send DHCP packets to the ipv4 stack, or we won't - * be able to use dhcp client on the outside interface - */ - if (PREDICT_FALSE - (proto0 == NAT_PROTOCOL_UDP - && (vnet_buffer (b0)->ip.reass.l4_dst_port == - clib_host_to_net_u16 - (UDP_DST_PORT_dhcp_to_server)) - && ip0->dst_address.as_u32 == 0xffffffff)) - goto trace0; - } - else - { - if (PREDICT_FALSE - (snat_not_translate - (sm, node, sw_if_index0, ip0, proto0, - rx_fib_index0, thread_index))) - goto trace0; - } - - next0 = slow_path (sm, b0, ip0, - ip0->src_address, - vnet_buffer (b0)->ip.reass.l4_src_port, - rx_fib_index0, - proto0, - &s0, node, next0, thread_index, now); - - if (PREDICT_FALSE (next0 == SNAT_IN2OUT_NEXT_DROP)) + if (PREDICT_FALSE + (nat_not_translate_output_feature + (sm, ip0, proto0, + vnet_buffer (b0)->ip.reass.l4_src_port, + vnet_buffer (b0)->ip.reass.l4_dst_port, + thread_index, sw_if_index0))) goto trace0; - if (PREDICT_FALSE (!s0)) + /* + * Send DHCP packets to the ipv4 stack, or we won't + * be able to use dhcp client on the outside interface + */ + if (PREDICT_FALSE + (proto0 == NAT_PROTOCOL_UDP + && (vnet_buffer (b0)->ip.reass.l4_dst_port == + clib_host_to_net_u16 + (UDP_DST_PORT_dhcp_to_server)) + && ip0->dst_address.as_u32 == 0xffffffff)) goto trace0; } else { - next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; - goto trace0; + if (PREDICT_FALSE + (snat_not_translate + (sm, node, sw_if_index0, ip0, proto0, rx_fib_index0, + thread_index))) + goto trace0; } + + next0 = + slow_path (sm, b0, ip0, ip0->src_address, + vnet_buffer (b0)->ip.reass.l4_src_port, + rx_fib_index0, proto0, &s0, node, next0, + thread_index, now); + + if (PREDICT_FALSE (next0 == SNAT_IN2OUT_NEXT_DROP)) + goto trace0; + + if (PREDICT_FALSE (!s0)) + goto trace0; } else - s0 = - pool_elt_at_index (sm->per_thread_data[thread_index].sessions, - value0.value); + { + next0 = SNAT_IN2OUT_NEXT_SLOW_PATH; + goto trace0; + } + } + else + s0 = + pool_elt_at_index (sm->per_thread_data[thread_index].sessions, + value0.value); - b0->flags |= VNET_BUFFER_F_IS_NATED; + b0->flags |= VNET_BUFFER_F_IS_NATED; - old_addr0 = ip0->src_address.as_u32; - ip0->src_address = s0->out2in.addr; - new_addr0 = ip0->src_address.as_u32; - if (!is_output_feature) - vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index; + old_addr0 = ip0->src_address.as_u32; + ip0->src_address = s0->out2in.addr; + new_addr0 = ip0->src_address.as_u32; + if (!is_output_feature) + vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index; - sum0 = ip0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - src_address /* changed member */ ); - ip0->checksum = ip_csum_fold (sum0); + sum0 = ip0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, src_address /* changed member */ ); + ip0->checksum = ip_csum_fold (sum0); - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; - new_port0 = udp0->src_port = s0->out2in.port; - sum0 = tcp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - sum0 = ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - mss_clamping (sm->mss_clamping, tcp0, &sum0); - tcp0->checksum = ip_csum_fold (sum0); - } - tcp_packets++; + old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; + new_port0 = udp0->src_port = s0->out2in.port; + sum0 = tcp0->checksum; + sum0 = + ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address /* changed member */ ); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + mss_clamping (sm->mss_clamping, tcp0, &sum0); + tcp0->checksum = ip_csum_fold (sum0); } - else + tcp_packets++; + } + else + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + udp0->src_port = s0->out2in.port; + if (PREDICT_FALSE (udp0->checksum)) { - udp0->src_port = s0->out2in.port; - if (PREDICT_FALSE (udp0->checksum)) - { - old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; - new_port0 = udp0->src_port; - sum0 = udp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, dst_address /* changed member */ - ); - sum0 = - ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - udp0->checksum = ip_csum_fold (sum0); - } + old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; + new_port0 = udp0->src_port; + sum0 = udp0->checksum; + sum0 = + ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address /* changed member */ ); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + udp0->checksum = ip_csum_fold (sum0); } - udp_packets++; - } - - /* Accounting */ - nat44_session_update_counters (s0, now, - vlib_buffer_length_in_chain (vm, b0), - thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s0, thread_index); - - trace0: - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) - { - snat_in2out_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->is_slow_path = is_slow_path; - t->sw_if_index = sw_if_index0; - t->next_index = next0; - t->session_index = ~0; - if (s0) - t->session_index = - s0 - sm->per_thread_data[thread_index].sessions; } + udp_packets++; + } - pkts_processed += next0 == SNAT_IN2OUT_NEXT_LOOKUP; + /* Accounting */ + nat44_session_update_counters (s0, now, + vlib_buffer_length_in_chain (vm, b0), + thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s0, thread_index); - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + trace0: + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + snat_in2out_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); + t->is_slow_path = is_slow_path; + t->sw_if_index = sw_if_index0; + t->next_index = next0; + t->session_index = ~0; + if (s0) + t->session_index = + s0 - sm->per_thread_data[thread_index].sessions; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + pkts_processed += next0 == SNAT_IN2OUT_NEXT_LOOKUP; + + n_left_from--; + next[0] = next0; + next++; } + vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, + frame->n_vectors); + vlib_node_increment_counter (vm, stats_node_index, SNAT_IN2OUT_ERROR_IN2OUT_PACKETS, pkts_processed); diff --git a/src/plugins/nat/in2out_ed.c b/src/plugins/nat/in2out_ed.c index e5f29e4b445..c59c2d4536c 100644 --- a/src/plugins/nat/in2out_ed.c +++ b/src/plugins/nat/in2out_ed.c @@ -909,8 +909,7 @@ nat44_ed_in2out_fast_path_node_fn_inline (vlib_main_t * vm, vlib_frame_t * frame, int is_output_feature) { - u32 n_left_from, *from, *to_next, pkts_processed = 0, stats_node_index; - nat_next_t next_index; + u32 n_left_from, *from, pkts_processed = 0, stats_node_index; snat_main_t *sm = &snat_main; f64 now = vlib_time_now (vm); u32 thread_index = vm->thread_index; @@ -925,190 +924,147 @@ nat44_ed_in2out_fast_path_node_fn_inline (vlib_main_t * vm, from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; - next_index = node->cached_next_index; + + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; + u16 nexts[VLIB_FRAME_SIZE], *next = nexts; + vlib_get_buffers (vm, from, b, n_left_from); while (n_left_from > 0) { - u32 n_left_to_next; - - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); - - while (n_left_from > 0 && n_left_to_next > 0) + vlib_buffer_t *b0; + u32 sw_if_index0, rx_fib_index0, proto0, new_addr0, old_addr0, + iph_offset0 = 0; + u16 old_port0, new_port0; + ip4_header_t *ip0; + udp_header_t *udp0; + tcp_header_t *tcp0; + snat_session_t *s0 = 0; + clib_bihash_kv_16_8_t kv0, value0; + ip_csum_t sum0; + + b0 = *b; + + if (is_output_feature) { - u32 bi0; - vlib_buffer_t *b0; - u32 next0, sw_if_index0, rx_fib_index0, iph_offset0 = 0, proto0, - new_addr0, old_addr0; - u16 old_port0, new_port0; - ip4_header_t *ip0; - udp_header_t *udp0; - tcp_header_t *tcp0; - snat_session_t *s0 = 0; - clib_bihash_kv_16_8_t kv0, value0; - ip_csum_t sum0; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); + vnet_feature_next (&vnet_buffer2 (b0)->nat.arc_next, b0); + iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length; + } - if (is_output_feature) - { - vnet_feature_next (&vnet_buffer2 (b0)->nat.arc_next, b0); - iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length; - } + next[0] = vnet_buffer2 (b0)->nat.arc_next; - next0 = vnet_buffer2 (b0)->nat.arc_next; + ip0 = + (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) + iph_offset0); - ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) + - iph_offset0); + sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + rx_fib_index0 = + fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, sw_if_index0); - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - rx_fib_index0 = - fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, - sw_if_index0); + if (PREDICT_FALSE (ip0->ttl == 1)) + { + vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next[0] = NAT_NEXT_ICMP_ERROR; + goto trace0; + } - if (PREDICT_FALSE (ip0->ttl == 1)) - { - vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next0 = NAT_NEXT_ICMP_ERROR; - goto trace0; - } + udp0 = ip4_next_header (ip0); + tcp0 = (tcp_header_t *) udp0; + proto0 = ip_proto_to_nat_proto (ip0->protocol); - udp0 = ip4_next_header (ip0); - tcp0 = (tcp_header_t *) udp0; - proto0 = ip_proto_to_nat_proto (ip0->protocol); + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) + { + next[0] = def_slow; + goto trace0; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) - { - next0 = def_slow; - goto trace0; - } + if (is_output_feature) + { + if (PREDICT_FALSE + (nat_not_translate_output_feature_fwd + (sm, ip0, thread_index, now, vm, b0))) + goto trace0; + } - if (is_output_feature) - { - if (PREDICT_FALSE (nat_not_translate_output_feature_fwd - (sm, ip0, thread_index, now, vm, b0))) - goto trace0; - } + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + { + next[0] = def_slow; + goto trace0; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) - { - next0 = def_slow; - goto trace0; - } + init_ed_k (&kv0, ip0->src_address, + vnet_buffer (b0)->ip.reass.l4_src_port, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, + ip0->protocol); - init_ed_k (&kv0, ip0->src_address, - vnet_buffer (b0)->ip.reass.l4_src_port, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, - ip0->protocol); + // lookup for session + if (clib_bihash_search_16_8 (&tsm->in2out_ed, &kv0, &value0)) + { + // session does not exist go slow path + next[0] = def_slow; + goto trace0; + } + ASSERT (thread_index == ed_value_get_thread_index (&value0)); + s0 = + pool_elt_at_index (tsm->sessions, + ed_value_get_session_index (&value0)); - // lookup for session - if (clib_bihash_search_16_8 (&tsm->in2out_ed, &kv0, &value0)) + if (s0->tcp_closed_timestamp) + { + if (now >= s0->tcp_closed_timestamp) { - // session does not exist go slow path - next0 = def_slow; - goto trace0; + // session is closed, go slow path + next[0] = def_slow; } - ASSERT (thread_index == ed_value_get_thread_index (&value0)); - s0 = - pool_elt_at_index (tsm->sessions, - ed_value_get_session_index (&value0)); - - if (s0->tcp_closed_timestamp) + else { - if (now >= s0->tcp_closed_timestamp) - { - // session is closed, go slow path - next0 = def_slow; - } - else - { - // session in transitory timeout, drop - b0->error = node->errors[NAT_IN2OUT_ED_ERROR_TCP_CLOSED]; - next0 = NAT_NEXT_DROP; - } - goto trace0; + // session in transitory timeout, drop + b0->error = node->errors[NAT_IN2OUT_ED_ERROR_TCP_CLOSED]; + next[0] = NAT_NEXT_DROP; } + goto trace0; + } - // drop if session expired - u64 sess_timeout_time; - sess_timeout_time = s0->last_heard + - (f64) nat44_session_get_timeout (sm, s0); - if (now >= sess_timeout_time) - { - nat_free_session_data (sm, s0, thread_index, 0); - nat_ed_session_delete (sm, s0, thread_index, 1); - // session is closed, go slow path - next0 = def_slow; - goto trace0; - } + // drop if session expired + u64 sess_timeout_time; + sess_timeout_time = + s0->last_heard + (f64) nat44_session_get_timeout (sm, s0); + if (now >= sess_timeout_time) + { + nat_free_session_data (sm, s0, thread_index, 0); + nat_ed_session_delete (sm, s0, thread_index, 1); + // session is closed, go slow path + next[0] = def_slow; + goto trace0; + } - b0->flags |= VNET_BUFFER_F_IS_NATED; + b0->flags |= VNET_BUFFER_F_IS_NATED; - if (!is_output_feature) - vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index; + if (!is_output_feature) + vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index; - old_addr0 = ip0->src_address.as_u32; - new_addr0 = ip0->src_address.as_u32 = s0->out2in.addr.as_u32; - sum0 = ip0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, - src_address); - if (PREDICT_FALSE (is_twice_nat_session (s0))) - sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32, - s0->ext_host_addr.as_u32, ip4_header_t, - dst_address); - ip0->checksum = ip_csum_fold (sum0); + old_addr0 = ip0->src_address.as_u32; + new_addr0 = ip0->src_address.as_u32 = s0->out2in.addr.as_u32; + sum0 = ip0->checksum; + sum0 = + ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + src_address); + if (PREDICT_FALSE (is_twice_nat_session (s0))) + sum0 = + ip_csum_update (sum0, ip0->dst_address.as_u32, + s0->ext_host_addr.as_u32, ip4_header_t, + dst_address); + ip0->checksum = ip_csum_fold (sum0); - old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; + old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) - { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - new_port0 = udp0->src_port = s0->out2in.port; - sum0 = tcp0->checksum; - sum0 = - ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, dst_address); - sum0 = - ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t, length); - if (PREDICT_FALSE (is_twice_nat_session (s0))) - { - sum0 = - ip_csum_update (sum0, ip0->dst_address.as_u32, - s0->ext_host_addr.as_u32, - ip4_header_t, dst_address); - sum0 = - ip_csum_update (sum0, - vnet_buffer (b0)->ip. - reass.l4_dst_port, s0->ext_host_port, - ip4_header_t, length); - tcp0->dst_port = s0->ext_host_port; - ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; - } - mss_clamping (sm->mss_clamping, tcp0, &sum0); - tcp0->checksum = ip_csum_fold (sum0); - } - tcp_packets++; - if (nat44_set_tcp_session_state_i2o - (sm, now, s0, b0, thread_index)) - goto trace0; - } - else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment - && udp0->checksum) + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { new_port0 = udp0->src_port = s0->out2in.port; - sum0 = udp0->checksum; + sum0 = tcp0->checksum; sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, dst_address); @@ -1117,66 +1073,97 @@ nat44_ed_in2out_fast_path_node_fn_inline (vlib_main_t * vm, length); if (PREDICT_FALSE (is_twice_nat_session (s0))) { - sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32, - s0->ext_host_addr.as_u32, - ip4_header_t, dst_address); + sum0 = + ip_csum_update (sum0, ip0->dst_address.as_u32, + s0->ext_host_addr.as_u32, ip4_header_t, + dst_address); sum0 = ip_csum_update (sum0, vnet_buffer (b0)->ip.reass.l4_dst_port, s0->ext_host_port, ip4_header_t, length); - udp0->dst_port = s0->ext_host_port; + tcp0->dst_port = s0->ext_host_port; ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; } - udp0->checksum = ip_csum_fold (sum0); - udp_packets++; + mss_clamping (sm->mss_clamping, tcp0, &sum0); + tcp0->checksum = ip_csum_fold (sum0); } - else + tcp_packets++; + if (nat44_set_tcp_session_state_i2o (sm, now, s0, b0, thread_index)) + goto trace0; + } + else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment + && udp0->checksum) + { + new_port0 = udp0->src_port = s0->out2in.port; + sum0 = udp0->checksum; + sum0 = + ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, length); + if (PREDICT_FALSE (is_twice_nat_session (s0))) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + sum0 = + ip_csum_update (sum0, ip0->dst_address.as_u32, + s0->ext_host_addr.as_u32, ip4_header_t, + dst_address); + sum0 = + ip_csum_update (sum0, vnet_buffer (b0)->ip.reass.l4_dst_port, + s0->ext_host_port, ip4_header_t, length); + udp0->dst_port = s0->ext_host_port; + ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; + } + udp0->checksum = ip_csum_fold (sum0); + udp_packets++; + } + else + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + { + new_port0 = udp0->src_port = s0->out2in.port; + if (PREDICT_FALSE (is_twice_nat_session (s0))) { - new_port0 = udp0->src_port = s0->out2in.port; - if (PREDICT_FALSE (is_twice_nat_session (s0))) - { - udp0->dst_port = s0->ext_host_port; - ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; - } - udp_packets++; + udp0->dst_port = s0->ext_host_port; + ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; } + udp_packets++; } + } - /* Accounting */ - nat44_session_update_counters (s0, now, - vlib_buffer_length_in_chain - (vm, b0), thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s0, thread_index); - - trace0: - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) - { - nat_in2out_ed_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - t->next_index = next0; - t->is_slow_path = 0; - - if (s0) - t->session_index = s0 - tsm->sessions; - else - t->session_index = ~0; - } + /* Accounting */ + nat44_session_update_counters (s0, now, + vlib_buffer_length_in_chain (vm, b0), + thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s0, thread_index); - pkts_processed += next0 == vnet_buffer2 (b0)->nat.arc_next; - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + trace0: + if (PREDICT_FALSE + ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + nat_in2out_ed_trace_t *t = + vlib_add_trace (vm, node, b0, sizeof (*t)); + t->sw_if_index = sw_if_index0; + t->next_index = next[0]; + t->is_slow_path = 0; + + if (s0) + t->session_index = s0 - tsm->sessions; + else + t->session_index = ~0; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + pkts_processed += next[0] == vnet_buffer2 (b0)->nat.arc_next; + + n_left_from--; + next++; + b++; } + vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, + frame->n_vectors); + vlib_node_increment_counter (vm, stats_node_index, NAT_IN2OUT_ED_ERROR_IN2OUT_PACKETS, pkts_processed); @@ -1199,8 +1186,7 @@ nat44_ed_in2out_slow_path_node_fn_inline (vlib_main_t * vm, vlib_frame_t * frame, int is_output_feature) { - u32 n_left_from, *from, *to_next, pkts_processed = 0, stats_node_index; - nat_next_t next_index; + u32 n_left_from, *from, pkts_processed = 0, stats_node_index; snat_main_t *sm = &snat_main; f64 now = vlib_time_now (vm); u32 thread_index = vm->thread_index; @@ -1211,281 +1197,262 @@ nat44_ed_in2out_slow_path_node_fn_inline (vlib_main_t * vm, from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; - next_index = node->cached_next_index; + + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; + u16 nexts[VLIB_FRAME_SIZE], *next = nexts; + vlib_get_buffers (vm, from, b, n_left_from); while (n_left_from > 0) { - u32 n_left_to_next; + vlib_buffer_t *b0; + u32 sw_if_index0, rx_fib_index0, proto0, new_addr0, old_addr0, + iph_offset0 = 0; + u16 old_port0, new_port0; + ip4_header_t *ip0; + udp_header_t *udp0; + tcp_header_t *tcp0; + icmp46_header_t *icmp0; + snat_session_t *s0 = 0; + clib_bihash_kv_16_8_t kv0, value0; + ip_csum_t sum0; - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + b0 = *b; - while (n_left_from > 0 && n_left_to_next > 0) - { - u32 bi0; - vlib_buffer_t *b0; - u32 next0, sw_if_index0, rx_fib_index0, iph_offset0 = 0, proto0, - new_addr0, old_addr0; - u16 old_port0, new_port0; - ip4_header_t *ip0; - udp_header_t *udp0; - tcp_header_t *tcp0; - icmp46_header_t *icmp0; - snat_session_t *s0 = 0; - clib_bihash_kv_16_8_t kv0, value0; - ip_csum_t sum0; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); + if (is_output_feature) + iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length; - if (is_output_feature) - iph_offset0 = vnet_buffer (b0)->ip.reass.save_rewrite_length; + next[0] = vnet_buffer2 (b0)->nat.arc_next; - next0 = vnet_buffer2 (b0)->nat.arc_next; + ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) + + iph_offset0); - ip0 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b0) + - iph_offset0); + sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + rx_fib_index0 = + fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, sw_if_index0); - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - rx_fib_index0 = - fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, - sw_if_index0); + if (PREDICT_FALSE (ip0->ttl == 1)) + { + vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next[0] = NAT_NEXT_ICMP_ERROR; + goto trace0; + } - if (PREDICT_FALSE (ip0->ttl == 1)) - { - vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next0 = NAT_NEXT_ICMP_ERROR; - goto trace0; - } + udp0 = ip4_next_header (ip0); + tcp0 = (tcp_header_t *) udp0; + icmp0 = (icmp46_header_t *) udp0; + proto0 = ip_proto_to_nat_proto (ip0->protocol); - udp0 = ip4_next_header (ip0); - tcp0 = (tcp_header_t *) udp0; - icmp0 = (icmp46_header_t *) udp0; - proto0 = ip_proto_to_nat_proto (ip0->protocol); + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) + { + s0 = nat44_ed_in2out_unknown_proto (sm, b0, ip0, + rx_fib_index0, + thread_index, now, vm, node); + if (!s0) + next[0] = NAT_NEXT_DROP; - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) - { - s0 = nat44_ed_in2out_unknown_proto (sm, b0, ip0, - rx_fib_index0, - thread_index, now, - vm, node); - if (!s0) - next0 = NAT_NEXT_DROP; - - other_packets++; - goto trace0; - } + other_packets++; + goto trace0; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + { + next[0] = icmp_in2out_ed_slow_path + (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, + node, next[0], now, thread_index, &s0); + icmp_packets++; + goto trace0; + } + + init_ed_k (&kv0, ip0->src_address, + vnet_buffer (b0)->ip.reass.l4_src_port, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, + ip0->protocol); + if (!clib_bihash_search_16_8 (&tsm->in2out_ed, &kv0, &value0)) + { + ASSERT (thread_index == ed_value_get_thread_index (&value0)); + s0 = + pool_elt_at_index (tsm->sessions, + ed_value_get_session_index (&value0)); + + if (s0->tcp_closed_timestamp && now >= s0->tcp_closed_timestamp) { - next0 = icmp_in2out_ed_slow_path - (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, - node, next0, now, thread_index, &s0); - icmp_packets++; - goto trace0; + nat_free_session_data (sm, s0, thread_index, 0); + nat_ed_session_delete (sm, s0, thread_index, 1); + s0 = NULL; } + } - init_ed_k (&kv0, ip0->src_address, - vnet_buffer (b0)->ip.reass.l4_src_port, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, - ip0->protocol); - if (!clib_bihash_search_16_8 (&tsm->in2out_ed, &kv0, &value0)) + if (!s0) + { + if (is_output_feature) { - ASSERT (thread_index == ed_value_get_thread_index (&value0)); - s0 = - pool_elt_at_index (tsm->sessions, - ed_value_get_session_index (&value0)); + if (PREDICT_FALSE + (nat44_ed_not_translate_output_feature + (sm, ip0, vnet_buffer (b0)->ip.reass.l4_src_port, + vnet_buffer (b0)->ip.reass.l4_dst_port, thread_index, + sw_if_index0, vnet_buffer (b0)->sw_if_index[VLIB_TX]))) + goto trace0; - if (s0->tcp_closed_timestamp && now >= s0->tcp_closed_timestamp) - { - nat_free_session_data (sm, s0, thread_index, 0); - nat_ed_session_delete (sm, s0, thread_index, 1); - s0 = NULL; - } + /* + * Send DHCP packets to the ipv4 stack, or we won't + * be able to use dhcp client on the outside interface + */ + if (PREDICT_FALSE + (proto0 == NAT_PROTOCOL_UDP + && (vnet_buffer (b0)->ip.reass.l4_dst_port == + clib_host_to_net_u16 (UDP_DST_PORT_dhcp_to_server)) + && ip0->dst_address.as_u32 == 0xffffffff)) + goto trace0; } - - if (!s0) + else { - if (is_output_feature) - { - if (PREDICT_FALSE - (nat44_ed_not_translate_output_feature - (sm, ip0, vnet_buffer (b0)->ip.reass.l4_src_port, - vnet_buffer (b0)->ip.reass.l4_dst_port, thread_index, - sw_if_index0, - vnet_buffer (b0)->sw_if_index[VLIB_TX]))) - goto trace0; - - /* - * Send DHCP packets to the ipv4 stack, or we won't - * be able to use dhcp client on the outside interface - */ - if (PREDICT_FALSE - (proto0 == NAT_PROTOCOL_UDP - && (vnet_buffer (b0)->ip.reass.l4_dst_port == - clib_host_to_net_u16 (UDP_DST_PORT_dhcp_to_server)) - && ip0->dst_address.as_u32 == 0xffffffff)) - goto trace0; - } - else - { - if (PREDICT_FALSE - (nat44_ed_not_translate - (sm, node, sw_if_index0, ip0, proto0, rx_fib_index0, - thread_index))) - goto trace0; - } + if (PREDICT_FALSE + (nat44_ed_not_translate + (sm, node, sw_if_index0, ip0, proto0, rx_fib_index0, + thread_index))) + goto trace0; + } - next0 = - slow_path_ed (sm, b0, ip0->src_address, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_src_port, - vnet_buffer (b0)->ip.reass.l4_dst_port, - ip0->protocol, rx_fib_index0, &s0, node, next0, - thread_index, now); + next[0] = + slow_path_ed (sm, b0, ip0->src_address, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_src_port, + vnet_buffer (b0)->ip.reass.l4_dst_port, + ip0->protocol, rx_fib_index0, &s0, node, next[0], + thread_index, now); - if (PREDICT_FALSE (next0 == NAT_NEXT_DROP)) - goto trace0; + if (PREDICT_FALSE (next[0] == NAT_NEXT_DROP)) + goto trace0; - if (PREDICT_FALSE (!s0)) - goto trace0; + if (PREDICT_FALSE (!s0)) + goto trace0; - } + } - b0->flags |= VNET_BUFFER_F_IS_NATED; + b0->flags |= VNET_BUFFER_F_IS_NATED; - if (!is_output_feature) - vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index; + if (!is_output_feature) + vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->out2in.fib_index; - old_addr0 = ip0->src_address.as_u32; - new_addr0 = ip0->src_address.as_u32 = s0->out2in.addr.as_u32; - sum0 = ip0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, - src_address); - if (PREDICT_FALSE (is_twice_nat_session (s0))) - sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32, - s0->ext_host_addr.as_u32, ip4_header_t, - dst_address); - ip0->checksum = ip_csum_fold (sum0); + old_addr0 = ip0->src_address.as_u32; + new_addr0 = ip0->src_address.as_u32 = s0->out2in.addr.as_u32; + sum0 = ip0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + src_address); + if (PREDICT_FALSE (is_twice_nat_session (s0))) + sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32, + s0->ext_host_addr.as_u32, ip4_header_t, + dst_address); + ip0->checksum = ip_csum_fold (sum0); - old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; + old_port0 = vnet_buffer (b0)->ip.reass.l4_src_port; - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) - { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - new_port0 = udp0->src_port = s0->out2in.port; - sum0 = tcp0->checksum; - sum0 = - ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, dst_address); - sum0 = - ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t, length); - if (PREDICT_FALSE (is_twice_nat_session (s0))) - { - sum0 = - ip_csum_update (sum0, ip0->dst_address.as_u32, - s0->ext_host_addr.as_u32, - ip4_header_t, dst_address); - sum0 = - ip_csum_update (sum0, - vnet_buffer (b0)->ip. - reass.l4_dst_port, s0->ext_host_port, - ip4_header_t, length); - tcp0->dst_port = s0->ext_host_port; - ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; - } - mss_clamping (sm->mss_clamping, tcp0, &sum0); - tcp0->checksum = ip_csum_fold (sum0); - } - tcp_packets++; - if (nat44_set_tcp_session_state_i2o - (sm, now, s0, b0, thread_index)) - goto trace0; - } - else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment - && udp0->checksum) + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { new_port0 = udp0->src_port = s0->out2in.port; - sum0 = udp0->checksum; + sum0 = tcp0->checksum; sum0 = - ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, - dst_address); + ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, dst_address); sum0 = - ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, - length); + ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t, length); if (PREDICT_FALSE (is_twice_nat_session (s0))) { - sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32, - s0->ext_host_addr.as_u32, - ip4_header_t, dst_address); + sum0 = + ip_csum_update (sum0, ip0->dst_address.as_u32, + s0->ext_host_addr.as_u32, + ip4_header_t, dst_address); sum0 = ip_csum_update (sum0, vnet_buffer (b0)->ip.reass.l4_dst_port, s0->ext_host_port, ip4_header_t, length); - udp0->dst_port = s0->ext_host_port; + tcp0->dst_port = s0->ext_host_port; ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; } - udp0->checksum = ip_csum_fold (sum0); - udp_packets++; + mss_clamping (sm->mss_clamping, tcp0, &sum0); + tcp0->checksum = ip_csum_fold (sum0); } - else + tcp_packets++; + if (nat44_set_tcp_session_state_i2o (sm, now, s0, b0, thread_index)) + goto trace0; + } + else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment + && udp0->checksum) + { + new_port0 = udp0->src_port = s0->out2in.port; + sum0 = udp0->checksum; + sum0 = + ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, length); + if (PREDICT_FALSE (is_twice_nat_session (s0))) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - new_port0 = udp0->src_port = s0->out2in.port; - if (PREDICT_FALSE (is_twice_nat_session (s0))) - { - udp0->dst_port = s0->ext_host_port; - ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; - } - udp_packets++; - } + sum0 = ip_csum_update (sum0, ip0->dst_address.as_u32, + s0->ext_host_addr.as_u32, + ip4_header_t, dst_address); + sum0 = + ip_csum_update (sum0, + vnet_buffer (b0)->ip.reass.l4_dst_port, + s0->ext_host_port, ip4_header_t, length); + udp0->dst_port = s0->ext_host_port; + ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; } - - /* Accounting */ - nat44_session_update_counters (s0, now, - vlib_buffer_length_in_chain - (vm, b0), thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s0, thread_index); - - trace0: - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) + udp0->checksum = ip_csum_fold (sum0); + udp_packets++; + } + else + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - nat_in2out_ed_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - t->next_index = next0; - t->is_slow_path = 1; - - if (s0) - t->session_index = s0 - tsm->sessions; - else - t->session_index = ~0; + new_port0 = udp0->src_port = s0->out2in.port; + if (PREDICT_FALSE (is_twice_nat_session (s0))) + { + udp0->dst_port = s0->ext_host_port; + ip0->dst_address.as_u32 = s0->ext_host_addr.as_u32; + } + udp_packets++; } + } - pkts_processed += next0 == vnet_buffer2 (b0)->nat.arc_next; + /* Accounting */ + nat44_session_update_counters (s0, now, + vlib_buffer_length_in_chain + (vm, b0), thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s0, thread_index); - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + trace0: + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + nat_in2out_ed_trace_t *t = + vlib_add_trace (vm, node, b0, sizeof (*t)); + t->sw_if_index = sw_if_index0; + t->next_index = next[0]; + t->is_slow_path = 1; + + if (s0) + t->session_index = s0 - tsm->sessions; + else + t->session_index = ~0; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + pkts_processed += next[0] == vnet_buffer2 (b0)->nat.arc_next; + + n_left_from--; + next++; + b++; } + vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, + frame->n_vectors); + vlib_node_increment_counter (vm, stats_node_index, NAT_IN2OUT_ED_ERROR_IN2OUT_PACKETS, pkts_processed); diff --git a/src/plugins/nat/nat_inlines.h b/src/plugins/nat/nat_inlines.h index 0254de34d49..01c866a0787 100644 --- a/src/plugins/nat/nat_inlines.h +++ b/src/plugins/nat/nat_inlines.h @@ -105,120 +105,99 @@ nat_pre_node_fn_inline (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame, u32 def_next) { - u32 n_left_from, *from, *to_next; - u16 next_index; + u32 n_left_from, *from; from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; - next_index = node->cached_next_index; - while (n_left_from > 0) + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; + u16 nexts[VLIB_FRAME_SIZE], *next = nexts; + vlib_get_buffers (vm, from, b, n_left_from); + + while (n_left_from >= 2) { - u32 n_left_to_next; + u32 next0, next1; + u32 arc_next0, arc_next1; + vlib_buffer_t *b0, *b1; - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + b0 = *b; + b++; + b1 = *b; + b++; - while (n_left_from >= 4 && n_left_to_next >= 2) + /* Prefetch next iteration. */ + if (PREDICT_TRUE (n_left_from >= 4)) { - u32 next0, next1; - u32 arc_next0, arc_next1; - u32 bi0, bi1; - vlib_buffer_t *b0, *b1; - - /* Prefetch next iteration. */ - { - vlib_buffer_t *p2, *p3; - - p2 = vlib_get_buffer (vm, from[2]); - p3 = vlib_get_buffer (vm, from[3]); - - vlib_prefetch_buffer_header (p2, LOAD); - vlib_prefetch_buffer_header (p3, LOAD); - - CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, LOAD); - CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, LOAD); - } + vlib_buffer_t *p2, *p3; - /* speculatively enqueue b0 and b1 to the current next frame */ - to_next[0] = bi0 = from[0]; - to_next[1] = bi1 = from[1]; - from += 2; - to_next += 2; - n_left_from -= 2; - n_left_to_next -= 2; + p2 = *b; + p3 = *(b + 1); - b0 = vlib_get_buffer (vm, bi0); - b1 = vlib_get_buffer (vm, bi1); + vlib_prefetch_buffer_header (p2, LOAD); + vlib_prefetch_buffer_header (p3, LOAD); - next0 = def_next; - next1 = def_next; + CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, LOAD); + CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, LOAD); + } - vnet_feature_next (&arc_next0, b0); - vnet_feature_next (&arc_next1, b1); + next0 = def_next; + next1 = def_next; - vnet_buffer2 (b0)->nat.arc_next = arc_next0; - vnet_buffer2 (b1)->nat.arc_next = arc_next1; + vnet_feature_next (&arc_next0, b0); + vnet_feature_next (&arc_next1, b1); - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) - { - if (b0->flags & VLIB_BUFFER_IS_TRACED) - { - nat_pre_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->next_index = next0; - t->arc_next_index = arc_next0; - } - if (b1->flags & VLIB_BUFFER_IS_TRACED) - { - nat_pre_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->next_index = next1; - t->arc_next_index = arc_next1; - } - } + vnet_buffer2 (b0)->nat.arc_next = arc_next0; + vnet_buffer2 (b1)->nat.arc_next = arc_next1; - /* verify speculative enqueues, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x2 (vm, node, next_index, - to_next, n_left_to_next, - bi0, bi1, next0, next1); - } - - while (n_left_from > 0 && n_left_to_next > 0) + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) { - u32 next0; - u32 arc_next0; - u32 bi0; - vlib_buffer_t *b0; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); - next0 = def_next; - vnet_feature_next (&arc_next0, b0); - vnet_buffer2 (b0)->nat.arc_next = arc_next0; - - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) + if (b0->flags & VLIB_BUFFER_IS_TRACED) { nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->next_index = next0; t->arc_next_index = arc_next0; } + if (b1->flags & VLIB_BUFFER_IS_TRACED) + { + nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); + t->next_index = next1; + t->arc_next_index = arc_next1; + } + } + + n_left_from -= 2; + next[0] = next0; + next[1] = next1; + next += 2; + } - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + while (n_left_from > 0) + { + u32 next0; + u32 arc_next0; + vlib_buffer_t *b0; + + b0 = *b; + b++; + + next0 = def_next; + vnet_feature_next (&arc_next0, b0); + vnet_buffer2 (b0)->nat.arc_next = arc_next0; + + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); + t->next_index = next0; + t->arc_next_index = arc_next0; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + n_left_from--; + next[0] = next0; + next++; } + vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, + frame->n_vectors); return frame->n_vectors; } diff --git a/src/plugins/nat/out2in.c b/src/plugins/nat/out2in.c index 071e16d4416..4e6ceb72f6a 100644 --- a/src/plugins/nat/out2in.c +++ b/src/plugins/nat/out2in.c @@ -724,8 +724,7 @@ VLIB_NODE_FN (snat_out2in_node) (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { - u32 n_left_from, *from, *to_next; - snat_out2in_next_t next_index; + u32 n_left_from, *from; u32 pkts_processed = 0; snat_main_t *sm = &snat_main; f64 now = vlib_time_now (vm); @@ -736,658 +735,626 @@ VLIB_NODE_FN (snat_out2in_node) (vlib_main_t * vm, from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; - next_index = node->cached_next_index; - while (n_left_from > 0) + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; + u16 nexts[VLIB_FRAME_SIZE], *next = nexts; + vlib_get_buffers (vm, from, b, n_left_from); + + while (n_left_from >= 2) { - u32 n_left_to_next; + vlib_buffer_t *b0, *b1; + u32 next0 = SNAT_OUT2IN_NEXT_LOOKUP; + u32 next1 = SNAT_OUT2IN_NEXT_LOOKUP; + u32 sw_if_index0, sw_if_index1; + ip4_header_t *ip0, *ip1; + ip_csum_t sum0, sum1; + u32 new_addr0, old_addr0; + u16 new_port0, old_port0; + u32 new_addr1, old_addr1; + u16 new_port1, old_port1; + udp_header_t *udp0, *udp1; + tcp_header_t *tcp0, *tcp1; + icmp46_header_t *icmp0, *icmp1; + u32 rx_fib_index0, rx_fib_index1; + u32 proto0, proto1; + snat_session_t *s0 = 0, *s1 = 0; + clib_bihash_kv_8_8_t kv0, kv1, value0, value1; + u8 identity_nat0, identity_nat1; + ip4_address_t sm_addr0, sm_addr1; + u16 sm_port0, sm_port1; + u32 sm_fib_index0, sm_fib_index1; + + b0 = *b; + b++; + b1 = *b; + b++; + + /* Prefetch next iteration. */ + if (PREDICT_TRUE (n_left_from >= 4)) + { + vlib_buffer_t *p2, *p3; + + p2 = *b; + p3 = *(b + 1); + + vlib_prefetch_buffer_header (p2, LOAD); + vlib_prefetch_buffer_header (p3, LOAD); + + CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, LOAD); + CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, LOAD); + } - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + vnet_buffer (b0)->snat.flags = 0; + vnet_buffer (b1)->snat.flags = 0; - while (n_left_from >= 4 && n_left_to_next >= 2) + ip0 = vlib_buffer_get_current (b0); + udp0 = ip4_next_header (ip0); + tcp0 = (tcp_header_t *) udp0; + icmp0 = (icmp46_header_t *) udp0; + + sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + rx_fib_index0 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, + sw_if_index0); + + if (PREDICT_FALSE (ip0->ttl == 1)) { - u32 bi0, bi1; - vlib_buffer_t *b0, *b1; - u32 next0 = SNAT_OUT2IN_NEXT_LOOKUP; - u32 next1 = SNAT_OUT2IN_NEXT_LOOKUP; - u32 sw_if_index0, sw_if_index1; - ip4_header_t *ip0, *ip1; - ip_csum_t sum0, sum1; - u32 new_addr0, old_addr0; - u16 new_port0, old_port0; - u32 new_addr1, old_addr1; - u16 new_port1, old_port1; - udp_header_t *udp0, *udp1; - tcp_header_t *tcp0, *tcp1; - icmp46_header_t *icmp0, *icmp1; - u32 rx_fib_index0, rx_fib_index1; - u32 proto0, proto1; - snat_session_t *s0 = 0, *s1 = 0; - clib_bihash_kv_8_8_t kv0, kv1, value0, value1; - u8 identity_nat0, identity_nat1; - ip4_address_t sm_addr0, sm_addr1; - u16 sm_port0, sm_port1; - u32 sm_fib_index0, sm_fib_index1; - - /* Prefetch next iteration. */ - { - vlib_buffer_t *p2, *p3; - - p2 = vlib_get_buffer (vm, from[2]); - p3 = vlib_get_buffer (vm, from[3]); - - vlib_prefetch_buffer_header (p2, LOAD); - vlib_prefetch_buffer_header (p3, LOAD); - - CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, LOAD); - CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, LOAD); - } - - /* speculatively enqueue b0 and b1 to the current next frame */ - to_next[0] = bi0 = from[0]; - to_next[1] = bi1 = from[1]; - from += 2; - to_next += 2; - n_left_from -= 2; - n_left_to_next -= 2; - - b0 = vlib_get_buffer (vm, bi0); - b1 = vlib_get_buffer (vm, bi1); - - vnet_buffer (b0)->snat.flags = 0; - vnet_buffer (b1)->snat.flags = 0; - - ip0 = vlib_buffer_get_current (b0); - udp0 = ip4_next_header (ip0); - tcp0 = (tcp_header_t *) udp0; - icmp0 = (icmp46_header_t *) udp0; - - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - rx_fib_index0 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, - sw_if_index0); - - if (PREDICT_FALSE (ip0->ttl == 1)) - { - vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next0 = SNAT_OUT2IN_NEXT_ICMP_ERROR; - goto trace0; - } + vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next0 = SNAT_OUT2IN_NEXT_ICMP_ERROR; + goto trace0; + } - proto0 = ip_proto_to_nat_proto (ip0->protocol); + proto0 = ip_proto_to_nat_proto (ip0->protocol); - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) + { + if (nat_out2in_sm_unknown_proto (sm, b0, ip0, rx_fib_index0)) { - if (nat_out2in_sm_unknown_proto (sm, b0, ip0, rx_fib_index0)) + if (!sm->forwarding_enabled) { - if (!sm->forwarding_enabled) - { - b0->error = - node->errors[SNAT_OUT2IN_ERROR_UNSUPPORTED_PROTOCOL]; - next0 = SNAT_OUT2IN_NEXT_DROP; - } + b0->error = + node->errors[SNAT_OUT2IN_ERROR_UNSUPPORTED_PROTOCOL]; + next0 = SNAT_OUT2IN_NEXT_DROP; } - other_packets++; - goto trace0; } + other_packets++; + goto trace0; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) - { - next0 = icmp_out2in_slow_path - (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, node, - next0, now, thread_index, &s0); - icmp_packets++; - goto trace0; - } + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + { + next0 = icmp_out2in_slow_path + (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, node, + next0, now, thread_index, &s0); + icmp_packets++; + goto trace0; + } - init_nat_k (&kv0, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, - proto0); - if (clib_bihash_search_8_8 - (&sm->per_thread_data[thread_index].out2in, &kv0, &value0)) + init_nat_k (&kv0, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, + proto0); + if (clib_bihash_search_8_8 + (&sm->per_thread_data[thread_index].out2in, &kv0, &value0)) + { + /* Try to match static mapping by external address and port, + destination address and port in packet */ + if (snat_static_mapping_match + (sm, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, + proto0, &sm_addr0, &sm_port0, &sm_fib_index0, 1, 0, 0, 0, + 0, &identity_nat0)) { - /* Try to match static mapping by external address and port, - destination address and port in packet */ - if (snat_static_mapping_match - (sm, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, - proto0, &sm_addr0, &sm_port0, &sm_fib_index0, 1, 0, 0, 0, - 0, &identity_nat0)) + /* + * Send DHCP packets to the ipv4 stack, or we won't + * be able to use dhcp client on the outside interface + */ + if (PREDICT_FALSE + (proto0 == NAT_PROTOCOL_UDP + && (vnet_buffer (b0)->ip.reass.l4_dst_port == + clib_host_to_net_u16 (UDP_DST_PORT_dhcp_to_client)))) { - /* - * Send DHCP packets to the ipv4 stack, or we won't - * be able to use dhcp client on the outside interface - */ - if (PREDICT_FALSE - (proto0 == NAT_PROTOCOL_UDP - && (vnet_buffer (b0)->ip.reass.l4_dst_port == - clib_host_to_net_u16 - (UDP_DST_PORT_dhcp_to_client)))) - { - vnet_feature_next (&next0, b0); - goto trace0; - } - - if (!sm->forwarding_enabled) - { - b0->error = - node->errors[SNAT_OUT2IN_ERROR_NO_TRANSLATION]; - next0 = SNAT_OUT2IN_NEXT_DROP; - } + vnet_feature_next (&next0, b0); goto trace0; } - if (PREDICT_FALSE (identity_nat0)) - goto trace0; - - /* Create session initiated by host from external network */ - s0 = create_session_for_static_mapping (sm, b0, - sm_addr0, sm_port0, - sm_fib_index0, - ip0->dst_address, - vnet_buffer (b0)-> - ip.reass.l4_dst_port, - rx_fib_index0, proto0, - node, thread_index, - now); - if (!s0) + if (!sm->forwarding_enabled) { + b0->error = node->errors[SNAT_OUT2IN_ERROR_NO_TRANSLATION]; next0 = SNAT_OUT2IN_NEXT_DROP; - goto trace0; } + goto trace0; } - else - s0 = pool_elt_at_index (tsm->sessions, value0.value); - old_addr0 = ip0->dst_address.as_u32; - ip0->dst_address = s0->in2out.addr; - new_addr0 = ip0->dst_address.as_u32; - vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->in2out.fib_index; + if (PREDICT_FALSE (identity_nat0)) + goto trace0; + + /* Create session initiated by host from external network */ + s0 = create_session_for_static_mapping (sm, b0, + sm_addr0, sm_port0, + sm_fib_index0, + ip0->dst_address, + vnet_buffer (b0)->ip. + reass.l4_dst_port, + rx_fib_index0, proto0, node, + thread_index, now); + if (!s0) + { + next0 = SNAT_OUT2IN_NEXT_DROP; + goto trace0; + } + } + else + s0 = pool_elt_at_index (tsm->sessions, value0.value); + + old_addr0 = ip0->dst_address.as_u32; + ip0->dst_address = s0->in2out.addr; + new_addr0 = ip0->dst_address.as_u32; + vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->in2out.fib_index; - sum0 = ip0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - ip0->checksum = ip_csum_fold (sum0); + sum0 = ip0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, dst_address /* changed member */ ); + ip0->checksum = ip_csum_fold (sum0); - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; - new_port0 = udp0->dst_port = s0->in2out.port; - sum0 = tcp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - - sum0 = ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - tcp0->checksum = ip_csum_fold (sum0); - } - tcp_packets++; + old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; + new_port0 = udp0->dst_port = s0->in2out.port; + sum0 = tcp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, + dst_address /* changed member */ ); + + sum0 = ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + tcp0->checksum = ip_csum_fold (sum0); } - else + tcp_packets++; + } + else + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; + new_port0 = udp0->dst_port = s0->in2out.port; + if (PREDICT_FALSE (udp0->checksum)) { - old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; - new_port0 = udp0->dst_port = s0->in2out.port; - if (PREDICT_FALSE (udp0->checksum)) - { - sum0 = udp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, dst_address /* changed member */ - ); - sum0 = - ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - udp0->checksum = ip_csum_fold (sum0); - } + sum0 = udp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, dst_address /* changed member */ + ); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + udp0->checksum = ip_csum_fold (sum0); } - udp_packets++; } + udp_packets++; + } - /* Accounting */ - nat44_session_update_counters (s0, now, - vlib_buffer_length_in_chain (vm, b0), - thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s0, thread_index); - trace0: + /* Accounting */ + nat44_session_update_counters (s0, now, + vlib_buffer_length_in_chain (vm, b0), + thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s0, thread_index); + trace0: - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) - { - snat_out2in_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - t->next_index = next0; - t->session_index = ~0; - if (s0) - t->session_index = - s0 - sm->per_thread_data[thread_index].sessions; - } + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + snat_out2in_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); + t->sw_if_index = sw_if_index0; + t->next_index = next0; + t->session_index = ~0; + if (s0) + t->session_index = + s0 - sm->per_thread_data[thread_index].sessions; + } - pkts_processed += next0 == SNAT_OUT2IN_NEXT_LOOKUP; + pkts_processed += next0 == SNAT_OUT2IN_NEXT_LOOKUP; - ip1 = vlib_buffer_get_current (b1); - udp1 = ip4_next_header (ip1); - tcp1 = (tcp_header_t *) udp1; - icmp1 = (icmp46_header_t *) udp1; + ip1 = vlib_buffer_get_current (b1); + udp1 = ip4_next_header (ip1); + tcp1 = (tcp_header_t *) udp1; + icmp1 = (icmp46_header_t *) udp1; - sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX]; - rx_fib_index1 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, - sw_if_index1); + sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX]; + rx_fib_index1 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, + sw_if_index1); - if (PREDICT_FALSE (ip1->ttl == 1)) - { - vnet_buffer (b1)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b1, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next1 = SNAT_OUT2IN_NEXT_ICMP_ERROR; - goto trace1; - } + if (PREDICT_FALSE (ip1->ttl == 1)) + { + vnet_buffer (b1)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b1, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next1 = SNAT_OUT2IN_NEXT_ICMP_ERROR; + goto trace1; + } - proto1 = ip_proto_to_nat_proto (ip1->protocol); + proto1 = ip_proto_to_nat_proto (ip1->protocol); - if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_OTHER)) + if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_OTHER)) + { + if (nat_out2in_sm_unknown_proto (sm, b1, ip1, rx_fib_index1)) { - if (nat_out2in_sm_unknown_proto (sm, b1, ip1, rx_fib_index1)) + if (!sm->forwarding_enabled) { - if (!sm->forwarding_enabled) - { - b1->error = - node->errors[SNAT_OUT2IN_ERROR_UNSUPPORTED_PROTOCOL]; - next1 = SNAT_OUT2IN_NEXT_DROP; - } + b1->error = + node->errors[SNAT_OUT2IN_ERROR_UNSUPPORTED_PROTOCOL]; + next1 = SNAT_OUT2IN_NEXT_DROP; } - other_packets++; - goto trace1; } + other_packets++; + goto trace1; + } - if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_ICMP)) - { - next1 = icmp_out2in_slow_path - (sm, b1, ip1, icmp1, sw_if_index1, rx_fib_index1, node, - next1, now, thread_index, &s1); - icmp_packets++; - goto trace1; - } + if (PREDICT_FALSE (proto1 == NAT_PROTOCOL_ICMP)) + { + next1 = icmp_out2in_slow_path + (sm, b1, ip1, icmp1, sw_if_index1, rx_fib_index1, node, + next1, now, thread_index, &s1); + icmp_packets++; + goto trace1; + } - init_nat_k (&kv1, ip1->dst_address, - vnet_buffer (b1)->ip.reass.l4_dst_port, rx_fib_index1, - proto1); + init_nat_k (&kv1, ip1->dst_address, + vnet_buffer (b1)->ip.reass.l4_dst_port, rx_fib_index1, + proto1); - if (clib_bihash_search_8_8 - (&sm->per_thread_data[thread_index].out2in, &kv1, &value1)) + if (clib_bihash_search_8_8 + (&sm->per_thread_data[thread_index].out2in, &kv1, &value1)) + { + /* Try to match static mapping by external address and port, + destination address and port in packet */ + if (snat_static_mapping_match + (sm, ip1->dst_address, + vnet_buffer (b1)->ip.reass.l4_dst_port, proto1, + rx_fib_index1, &sm_addr1, &sm_port1, &sm_fib_index1, 1, 0, + 0, 0, 0, &identity_nat1)) { - /* Try to match static mapping by external address and port, - destination address and port in packet */ - if (snat_static_mapping_match - (sm, ip1->dst_address, - vnet_buffer (b1)->ip.reass.l4_dst_port, proto1, - rx_fib_index1, &sm_addr1, &sm_port1, &sm_fib_index1, 1, 0, - 0, 0, 0, &identity_nat1)) + /* + * Send DHCP packets to the ipv4 stack, or we won't + * be able to use dhcp client on the outside interface + */ + if (PREDICT_FALSE + (proto1 == NAT_PROTOCOL_UDP + && (vnet_buffer (b1)->ip.reass.l4_dst_port == + clib_host_to_net_u16 (UDP_DST_PORT_dhcp_to_client)))) { - /* - * Send DHCP packets to the ipv4 stack, or we won't - * be able to use dhcp client on the outside interface - */ - if (PREDICT_FALSE - (proto1 == NAT_PROTOCOL_UDP - && (vnet_buffer (b1)->ip.reass.l4_dst_port == - clib_host_to_net_u16 - (UDP_DST_PORT_dhcp_to_client)))) - { - vnet_feature_next (&next1, b1); - goto trace1; - } - - if (!sm->forwarding_enabled) - { - b1->error = - node->errors[SNAT_OUT2IN_ERROR_NO_TRANSLATION]; - next1 = SNAT_OUT2IN_NEXT_DROP; - } + vnet_feature_next (&next1, b1); goto trace1; } - if (PREDICT_FALSE (identity_nat1)) - goto trace1; - - /* Create session initiated by host from external network */ - s1 = - create_session_for_static_mapping (sm, b1, sm_addr1, sm_port1, - sm_fib_index1, - ip1->dst_address, - vnet_buffer (b1)->ip. - reass.l4_dst_port, - rx_fib_index1, proto1, - node, thread_index, now); - if (!s1) + if (!sm->forwarding_enabled) { + b1->error = node->errors[SNAT_OUT2IN_ERROR_NO_TRANSLATION]; next1 = SNAT_OUT2IN_NEXT_DROP; - goto trace1; } + goto trace1; } - else - s1 = - pool_elt_at_index (sm->per_thread_data[thread_index].sessions, - value1.value); - - old_addr1 = ip1->dst_address.as_u32; - ip1->dst_address = s1->in2out.addr; - new_addr1 = ip1->dst_address.as_u32; - vnet_buffer (b1)->sw_if_index[VLIB_TX] = s1->in2out.fib_index; - - sum1 = ip1->checksum; - sum1 = ip_csum_update (sum1, old_addr1, new_addr1, - ip4_header_t, - dst_address /* changed member */ ); - ip1->checksum = ip_csum_fold (sum1); - - if (PREDICT_TRUE (proto1 == NAT_PROTOCOL_TCP)) + + if (PREDICT_FALSE (identity_nat1)) + goto trace1; + + /* Create session initiated by host from external network */ + s1 = + create_session_for_static_mapping (sm, b1, sm_addr1, sm_port1, + sm_fib_index1, + ip1->dst_address, + vnet_buffer (b1)->ip. + reass.l4_dst_port, + rx_fib_index1, proto1, node, + thread_index, now); + if (!s1) { - if (!vnet_buffer (b1)->ip.reass.is_non_first_fragment) - { - old_port1 = vnet_buffer (b1)->ip.reass.l4_dst_port; - new_port1 = udp1->dst_port = s1->in2out.port; - - sum1 = tcp1->checksum; - sum1 = ip_csum_update (sum1, old_addr1, new_addr1, - ip4_header_t, - dst_address /* changed member */ ); - - sum1 = ip_csum_update (sum1, old_port1, new_port1, - ip4_header_t /* cheat */ , - length /* changed member */ ); - tcp1->checksum = ip_csum_fold (sum1); - } - tcp_packets++; + next1 = SNAT_OUT2IN_NEXT_DROP; + goto trace1; } - else + } + else + s1 = + pool_elt_at_index (sm->per_thread_data[thread_index].sessions, + value1.value); + + old_addr1 = ip1->dst_address.as_u32; + ip1->dst_address = s1->in2out.addr; + new_addr1 = ip1->dst_address.as_u32; + vnet_buffer (b1)->sw_if_index[VLIB_TX] = s1->in2out.fib_index; + + sum1 = ip1->checksum; + sum1 = ip_csum_update (sum1, old_addr1, new_addr1, + ip4_header_t, dst_address /* changed member */ ); + ip1->checksum = ip_csum_fold (sum1); + + if (PREDICT_TRUE (proto1 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b1)->ip.reass.is_non_first_fragment) + { + old_port1 = vnet_buffer (b1)->ip.reass.l4_dst_port; + new_port1 = udp1->dst_port = s1->in2out.port; + + sum1 = tcp1->checksum; + sum1 = ip_csum_update (sum1, old_addr1, new_addr1, + ip4_header_t, + dst_address /* changed member */ ); + + sum1 = ip_csum_update (sum1, old_port1, new_port1, + ip4_header_t /* cheat */ , + length /* changed member */ ); + tcp1->checksum = ip_csum_fold (sum1); + } + tcp_packets++; + } + else + { + if (!vnet_buffer (b1)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b1)->ip.reass.is_non_first_fragment) + old_port1 = vnet_buffer (b1)->ip.reass.l4_dst_port; + new_port1 = udp1->dst_port = s1->in2out.port; + if (PREDICT_FALSE (udp1->checksum)) { - old_port1 = vnet_buffer (b1)->ip.reass.l4_dst_port; - new_port1 = udp1->dst_port = s1->in2out.port; - if (PREDICT_FALSE (udp1->checksum)) - { - - sum1 = udp1->checksum; - sum1 = - ip_csum_update (sum1, old_addr1, new_addr1, - ip4_header_t, - dst_address /* changed member */ ); - sum1 = - ip_csum_update (sum1, old_port1, new_port1, - ip4_header_t /* cheat */ , - length /* changed member */ ); - udp1->checksum = ip_csum_fold (sum1); - } + + sum1 = udp1->checksum; + sum1 = + ip_csum_update (sum1, old_addr1, new_addr1, + ip4_header_t, + dst_address /* changed member */ ); + sum1 = + ip_csum_update (sum1, old_port1, new_port1, + ip4_header_t /* cheat */ , + length /* changed member */ ); + udp1->checksum = ip_csum_fold (sum1); } - udp_packets++; } + udp_packets++; + } - /* Accounting */ - nat44_session_update_counters (s1, now, - vlib_buffer_length_in_chain (vm, b1), - thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s1, thread_index); - trace1: + /* Accounting */ + nat44_session_update_counters (s1, now, + vlib_buffer_length_in_chain (vm, b1), + thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s1, thread_index); + trace1: - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b1->flags & VLIB_BUFFER_IS_TRACED))) - { - snat_out2in_trace_t *t = - vlib_add_trace (vm, node, b1, sizeof (*t)); - t->sw_if_index = sw_if_index1; - t->next_index = next1; - t->session_index = ~0; - if (s1) - t->session_index = - s1 - sm->per_thread_data[thread_index].sessions; - } + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b1->flags & VLIB_BUFFER_IS_TRACED))) + { + snat_out2in_trace_t *t = vlib_add_trace (vm, node, b1, sizeof (*t)); + t->sw_if_index = sw_if_index1; + t->next_index = next1; + t->session_index = ~0; + if (s1) + t->session_index = + s1 - sm->per_thread_data[thread_index].sessions; + } - pkts_processed += next1 == SNAT_OUT2IN_NEXT_LOOKUP; + pkts_processed += next1 == SNAT_OUT2IN_NEXT_LOOKUP; - /* verify speculative enqueues, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x2 (vm, node, next_index, - to_next, n_left_to_next, - bi0, bi1, next0, next1); - } + n_left_from -= 2; + next[0] = next0; + next[1] = next1; + next += 2; + } - while (n_left_from > 0 && n_left_to_next > 0) + while (n_left_from > 0) + { + vlib_buffer_t *b0; + u32 next0 = SNAT_OUT2IN_NEXT_LOOKUP; + u32 sw_if_index0; + ip4_header_t *ip0; + ip_csum_t sum0; + u32 new_addr0, old_addr0; + u16 new_port0, old_port0; + udp_header_t *udp0; + tcp_header_t *tcp0; + icmp46_header_t *icmp0; + u32 rx_fib_index0; + u32 proto0; + snat_session_t *s0 = 0; + clib_bihash_kv_8_8_t kv0, value0; + u8 identity_nat0; + ip4_address_t sm_addr0; + u16 sm_port0; + u32 sm_fib_index0; + + b0 = *b; + ++b; + + vnet_buffer (b0)->snat.flags = 0; + + ip0 = vlib_buffer_get_current (b0); + udp0 = ip4_next_header (ip0); + tcp0 = (tcp_header_t *) udp0; + icmp0 = (icmp46_header_t *) udp0; + + sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + rx_fib_index0 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, + sw_if_index0); + + proto0 = ip_proto_to_nat_proto (ip0->protocol); + + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) { - u32 bi0; - vlib_buffer_t *b0; - u32 next0 = SNAT_OUT2IN_NEXT_LOOKUP; - u32 sw_if_index0; - ip4_header_t *ip0; - ip_csum_t sum0; - u32 new_addr0, old_addr0; - u16 new_port0, old_port0; - udp_header_t *udp0; - tcp_header_t *tcp0; - icmp46_header_t *icmp0; - u32 rx_fib_index0; - u32 proto0; - snat_session_t *s0 = 0; - clib_bihash_kv_8_8_t kv0, value0; - u8 identity_nat0; - ip4_address_t sm_addr0; - u16 sm_port0; - u32 sm_fib_index0; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); - - vnet_buffer (b0)->snat.flags = 0; - - ip0 = vlib_buffer_get_current (b0); - udp0 = ip4_next_header (ip0); - tcp0 = (tcp_header_t *) udp0; - icmp0 = (icmp46_header_t *) udp0; - - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - rx_fib_index0 = vec_elt (sm->ip4_main->fib_index_by_sw_if_index, - sw_if_index0); - - proto0 = ip_proto_to_nat_proto (ip0->protocol); - - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) + if (nat_out2in_sm_unknown_proto (sm, b0, ip0, rx_fib_index0)) { - if (nat_out2in_sm_unknown_proto (sm, b0, ip0, rx_fib_index0)) + if (!sm->forwarding_enabled) { - if (!sm->forwarding_enabled) - { - b0->error = - node->errors[SNAT_OUT2IN_ERROR_UNSUPPORTED_PROTOCOL]; - next0 = SNAT_OUT2IN_NEXT_DROP; - } + b0->error = + node->errors[SNAT_OUT2IN_ERROR_UNSUPPORTED_PROTOCOL]; + next0 = SNAT_OUT2IN_NEXT_DROP; } - other_packets++; - goto trace00; } + other_packets++; + goto trace00; + } - if (PREDICT_FALSE (ip0->ttl == 1)) - { - vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next0 = SNAT_OUT2IN_NEXT_ICMP_ERROR; - goto trace00; - } + if (PREDICT_FALSE (ip0->ttl == 1)) + { + vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next0 = SNAT_OUT2IN_NEXT_ICMP_ERROR; + goto trace00; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) - { - next0 = icmp_out2in_slow_path - (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, node, - next0, now, thread_index, &s0); - icmp_packets++; - goto trace00; - } + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + { + next0 = icmp_out2in_slow_path + (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, node, + next0, now, thread_index, &s0); + icmp_packets++; + goto trace00; + } - init_nat_k (&kv0, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, - proto0); + init_nat_k (&kv0, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, + proto0); - if (clib_bihash_search_8_8 - (&sm->per_thread_data[thread_index].out2in, &kv0, &value0)) + if (clib_bihash_search_8_8 + (&sm->per_thread_data[thread_index].out2in, &kv0, &value0)) + { + /* Try to match static mapping by external address and port, + destination address and port in packet */ + if (snat_static_mapping_match + (sm, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, + proto0, &sm_addr0, &sm_port0, &sm_fib_index0, 1, 0, 0, 0, + 0, &identity_nat0)) { - /* Try to match static mapping by external address and port, - destination address and port in packet */ - if (snat_static_mapping_match - (sm, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, - proto0, &sm_addr0, &sm_port0, &sm_fib_index0, 1, 0, 0, 0, - 0, &identity_nat0)) + /* + * Send DHCP packets to the ipv4 stack, or we won't + * be able to use dhcp client on the outside interface + */ + if (PREDICT_FALSE + (proto0 == NAT_PROTOCOL_UDP + && (vnet_buffer (b0)->ip.reass.l4_dst_port == + clib_host_to_net_u16 (UDP_DST_PORT_dhcp_to_client)))) { - /* - * Send DHCP packets to the ipv4 stack, or we won't - * be able to use dhcp client on the outside interface - */ - if (PREDICT_FALSE - (proto0 == NAT_PROTOCOL_UDP - && (vnet_buffer (b0)->ip.reass.l4_dst_port == - clib_host_to_net_u16 - (UDP_DST_PORT_dhcp_to_client)))) - { - vnet_feature_next (&next0, b0); - goto trace00; - } - - if (!sm->forwarding_enabled) - { - b0->error = - node->errors[SNAT_OUT2IN_ERROR_NO_TRANSLATION]; - next0 = SNAT_OUT2IN_NEXT_DROP; - } + vnet_feature_next (&next0, b0); goto trace00; } - if (PREDICT_FALSE (identity_nat0)) - goto trace00; - - /* Create session initiated by host from external network */ - s0 = create_session_for_static_mapping (sm, b0, - sm_addr0, sm_port0, - sm_fib_index0, - ip0->dst_address, - vnet_buffer (b0)-> - ip.reass.l4_dst_port, - rx_fib_index0, proto0, - node, thread_index, - now); - if (!s0) + if (!sm->forwarding_enabled) { + b0->error = node->errors[SNAT_OUT2IN_ERROR_NO_TRANSLATION]; next0 = SNAT_OUT2IN_NEXT_DROP; - goto trace00; } + goto trace00; } - else - s0 = - pool_elt_at_index (sm->per_thread_data[thread_index].sessions, - value0.value); - old_addr0 = ip0->dst_address.as_u32; - ip0->dst_address = s0->in2out.addr; - new_addr0 = ip0->dst_address.as_u32; - vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->in2out.fib_index; + if (PREDICT_FALSE (identity_nat0)) + goto trace00; - sum0 = ip0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - ip0->checksum = ip_csum_fold (sum0); + /* Create session initiated by host from external network */ + s0 = create_session_for_static_mapping (sm, b0, + sm_addr0, sm_port0, + sm_fib_index0, + ip0->dst_address, + vnet_buffer (b0)->ip. + reass.l4_dst_port, + rx_fib_index0, proto0, node, + thread_index, now); + if (!s0) + { + next0 = SNAT_OUT2IN_NEXT_DROP; + goto trace00; + } + } + else + s0 = + pool_elt_at_index (sm->per_thread_data[thread_index].sessions, + value0.value); - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + old_addr0 = ip0->dst_address.as_u32; + ip0->dst_address = s0->in2out.addr; + new_addr0 = ip0->dst_address.as_u32; + vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->in2out.fib_index; + + sum0 = ip0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, dst_address /* changed member */ ); + ip0->checksum = ip_csum_fold (sum0); + + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; - new_port0 = udp0->dst_port = s0->in2out.port; - - sum0 = tcp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - - sum0 = ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - tcp0->checksum = ip_csum_fold (sum0); - } - tcp_packets++; + old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; + new_port0 = udp0->dst_port = s0->in2out.port; + + sum0 = tcp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, + dst_address /* changed member */ ); + + sum0 = ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + tcp0->checksum = ip_csum_fold (sum0); } - else + tcp_packets++; + } + else + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; + new_port0 = udp0->dst_port = s0->in2out.port; + if (PREDICT_FALSE (udp0->checksum)) { - old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; - new_port0 = udp0->dst_port = s0->in2out.port; - if (PREDICT_FALSE (udp0->checksum)) - { - sum0 = udp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, dst_address /* changed member */ - ); - sum0 = - ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - udp0->checksum = ip_csum_fold (sum0); - } + sum0 = udp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, dst_address /* changed member */ + ); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + udp0->checksum = ip_csum_fold (sum0); } - udp_packets++; - } - - /* Accounting */ - nat44_session_update_counters (s0, now, - vlib_buffer_length_in_chain (vm, b0), - thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s0, thread_index); - trace00: - - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) - { - snat_out2in_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - t->next_index = next0; - t->session_index = ~0; - if (s0) - t->session_index = - s0 - sm->per_thread_data[thread_index].sessions; } + udp_packets++; + } - pkts_processed += next0 == SNAT_OUT2IN_NEXT_LOOKUP; + /* Accounting */ + nat44_session_update_counters (s0, now, + vlib_buffer_length_in_chain (vm, b0), + thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s0, thread_index); + trace00: - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + snat_out2in_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); + t->sw_if_index = sw_if_index0; + t->next_index = next0; + t->session_index = ~0; + if (s0) + t->session_index = + s0 - sm->per_thread_data[thread_index].sessions; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + pkts_processed += next0 == SNAT_OUT2IN_NEXT_LOOKUP; + + n_left_from--; + next[0] = next0; + next++; } + vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, + frame->n_vectors); + vlib_node_increment_counter (vm, sm->out2in_node_index, SNAT_OUT2IN_ERROR_OUT2IN_PACKETS, pkts_processed); @@ -1433,173 +1400,156 @@ VLIB_NODE_FN (snat_out2in_fast_node) (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { - u32 n_left_from, *from, *to_next; - snat_out2in_next_t next_index; + u32 n_left_from, *from; u32 pkts_processed = 0; snat_main_t *sm = &snat_main; from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; - next_index = node->cached_next_index; + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; + u16 nexts[VLIB_FRAME_SIZE], *next = nexts; + vlib_get_buffers (vm, from, b, n_left_from); while (n_left_from > 0) { - u32 n_left_to_next; + vlib_buffer_t *b0; + u32 next0 = SNAT_OUT2IN_NEXT_DROP; + u32 sw_if_index0; + ip4_header_t *ip0; + ip_csum_t sum0; + u32 new_addr0, old_addr0; + u16 new_port0, old_port0; + udp_header_t *udp0; + tcp_header_t *tcp0; + icmp46_header_t *icmp0; + u32 proto0; + u32 rx_fib_index0; + ip4_address_t sm_addr0; + u16 sm_port0; + u32 sm_fib_index0; + + b0 = *b; + b++; + + ip0 = vlib_buffer_get_current (b0); + udp0 = ip4_next_header (ip0); + tcp0 = (tcp_header_t *) udp0; + icmp0 = (icmp46_header_t *) udp0; + + sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + rx_fib_index0 = ip4_fib_table_get_index_for_sw_if_index (sw_if_index0); + + vnet_feature_next (&next0, b0); + + if (PREDICT_FALSE (ip0->ttl == 1)) + { + vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next0 = SNAT_OUT2IN_NEXT_ICMP_ERROR; + goto trace00; + } - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + proto0 = ip_proto_to_nat_proto (ip0->protocol); - while (n_left_from > 0 && n_left_to_next > 0) + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) + goto trace00; + + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) { - u32 bi0; - vlib_buffer_t *b0; - u32 next0 = SNAT_OUT2IN_NEXT_DROP; - u32 sw_if_index0; - ip4_header_t *ip0; - ip_csum_t sum0; - u32 new_addr0, old_addr0; - u16 new_port0, old_port0; - udp_header_t *udp0; - tcp_header_t *tcp0; - icmp46_header_t *icmp0; - u32 proto0; - u32 rx_fib_index0; - ip4_address_t sm_addr0; - u16 sm_port0; - u32 sm_fib_index0; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); - - ip0 = vlib_buffer_get_current (b0); - udp0 = ip4_next_header (ip0); - tcp0 = (tcp_header_t *) udp0; - icmp0 = (icmp46_header_t *) udp0; - - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - rx_fib_index0 = - ip4_fib_table_get_index_for_sw_if_index (sw_if_index0); - - vnet_feature_next (&next0, b0); - - if (PREDICT_FALSE (ip0->ttl == 1)) - { - vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next0 = SNAT_OUT2IN_NEXT_ICMP_ERROR; - goto trace00; - } + next0 = icmp_out2in (sm, b0, ip0, icmp0, sw_if_index0, + rx_fib_index0, node, next0, ~0, 0, 0); + goto trace00; + } - proto0 = ip_proto_to_nat_proto (ip0->protocol); + if (snat_static_mapping_match + (sm, ip0->dst_address, udp0->dst_port, rx_fib_index0, proto0, + &sm_addr0, &sm_port0, &sm_fib_index0, 1, 0, 0, 0, 0, 0)) + { + b0->error = node->errors[SNAT_OUT2IN_ERROR_NO_TRANSLATION]; + goto trace00; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) - goto trace00; + new_addr0 = sm_addr0.as_u32; + new_port0 = sm_port0; + vnet_buffer (b0)->sw_if_index[VLIB_TX] = sm_fib_index0; + old_addr0 = ip0->dst_address.as_u32; + ip0->dst_address.as_u32 = new_addr0; - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) - { - next0 = icmp_out2in (sm, b0, ip0, icmp0, sw_if_index0, - rx_fib_index0, node, next0, ~0, 0, 0); - goto trace00; - } + sum0 = ip0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, dst_address /* changed member */ ); + ip0->checksum = ip_csum_fold (sum0); - if (snat_static_mapping_match - (sm, ip0->dst_address, udp0->dst_port, rx_fib_index0, proto0, - &sm_addr0, &sm_port0, &sm_fib_index0, 1, 0, 0, 0, 0, 0)) + if (PREDICT_FALSE (new_port0 != udp0->dst_port)) + { + old_port0 = udp0->dst_port; + udp0->dst_port = new_port0; + + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) { - b0->error = node->errors[SNAT_OUT2IN_ERROR_NO_TRANSLATION]; - goto trace00; + sum0 = tcp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, + dst_address /* changed member */ ); + sum0 = ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + tcp0->checksum = ip_csum_fold (sum0); } - - new_addr0 = sm_addr0.as_u32; - new_port0 = sm_port0; - vnet_buffer (b0)->sw_if_index[VLIB_TX] = sm_fib_index0; - old_addr0 = ip0->dst_address.as_u32; - ip0->dst_address.as_u32 = new_addr0; - - sum0 = ip0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - ip0->checksum = ip_csum_fold (sum0); - - if (PREDICT_FALSE (new_port0 != udp0->dst_port)) + else if (udp0->checksum) { - old_port0 = udp0->dst_port; - udp0->dst_port = new_port0; - - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) - { - sum0 = tcp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - sum0 = ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - tcp0->checksum = ip_csum_fold (sum0); - } - else if (udp0->checksum) - { - sum0 = udp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - sum0 = ip_csum_update (sum0, old_port0, new_port0, - ip4_header_t /* cheat */ , - length /* changed member */ ); - udp0->checksum = ip_csum_fold (sum0); - } + sum0 = udp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, + dst_address /* changed member */ ); + sum0 = ip_csum_update (sum0, old_port0, new_port0, + ip4_header_t /* cheat */ , + length /* changed member */ ); + udp0->checksum = ip_csum_fold (sum0); } - else + } + else + { + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) { - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) - { - sum0 = tcp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - tcp0->checksum = ip_csum_fold (sum0); - } - else if (udp0->checksum) - { - sum0 = udp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, - ip4_header_t, - dst_address /* changed member */ ); - udp0->checksum = ip_csum_fold (sum0); - } + sum0 = tcp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, + dst_address /* changed member */ ); + tcp0->checksum = ip_csum_fold (sum0); } - - trace00: - - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) + else if (udp0->checksum) { - snat_out2in_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - t->next_index = next0; + sum0 = udp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, + ip4_header_t, + dst_address /* changed member */ ); + udp0->checksum = ip_csum_fold (sum0); } + } - pkts_processed += next0 != SNAT_OUT2IN_NEXT_DROP; + trace00: - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + snat_out2in_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); + t->sw_if_index = sw_if_index0; + t->next_index = next0; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + pkts_processed += next0 != SNAT_OUT2IN_NEXT_DROP; + + n_left_from--; + next[0] = next0; + next++; } + vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, + frame->n_vectors); + vlib_node_increment_counter (vm, sm->out2in_fast_node_index, SNAT_OUT2IN_ERROR_OUT2IN_PACKETS, pkts_processed); diff --git a/src/plugins/nat/out2in_ed.c b/src/plugins/nat/out2in_ed.c index 3d081e15313..8286e66ffda 100644 --- a/src/plugins/nat/out2in_ed.c +++ b/src/plugins/nat/out2in_ed.c @@ -675,8 +675,7 @@ nat44_ed_out2in_fast_path_node_fn_inline (vlib_main_t * vm, vlib_frame_t * frame, int is_multi_worker) { - u32 n_left_from, *from, *to_next, pkts_processed = 0, stats_node_index; - nat_next_t next_index; + u32 n_left_from, *from, pkts_processed = 0, stats_node_index; snat_main_t *sm = &snat_main; f64 now = vlib_time_now (vm); u32 thread_index = vm->thread_index; @@ -688,205 +687,158 @@ nat44_ed_out2in_fast_path_node_fn_inline (vlib_main_t * vm, from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; - next_index = node->cached_next_index; + + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; + u16 nexts[VLIB_FRAME_SIZE], *next = nexts; + vlib_get_buffers (vm, from, b, n_left_from); while (n_left_from > 0) { - u32 n_left_to_next; + vlib_buffer_t *b0; + u32 sw_if_index0, rx_fib_index0, proto0, old_addr0, new_addr0; + u16 old_port0, new_port0; + ip4_header_t *ip0; + udp_header_t *udp0; + tcp_header_t *tcp0; + snat_session_t *s0 = 0; + clib_bihash_kv_16_8_t kv0, value0; + ip_csum_t sum0; + + b0 = *b; + next[0] = vnet_buffer2 (b0)->nat.arc_next; + + vnet_buffer (b0)->snat.flags = 0; + ip0 = vlib_buffer_get_current (b0); + + sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + rx_fib_index0 = + fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, sw_if_index0); + + if (PREDICT_FALSE (ip0->ttl == 1)) + { + vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next[0] = NAT_NEXT_ICMP_ERROR; + goto trace0; + } - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + udp0 = ip4_next_header (ip0); + tcp0 = (tcp_header_t *) udp0; + proto0 = ip_proto_to_nat_proto (ip0->protocol); - while (n_left_from > 0 && n_left_to_next > 0) + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) { - u32 bi0; - vlib_buffer_t *b0; - u32 next0, sw_if_index0, rx_fib_index0, proto0, old_addr0, - new_addr0; - u16 old_port0, new_port0; - ip4_header_t *ip0; - udp_header_t *udp0; - tcp_header_t *tcp0; - snat_session_t *s0 = 0; - clib_bihash_kv_16_8_t kv0, value0; - ip_csum_t sum0; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); - next0 = vnet_buffer2 (b0)->nat.arc_next; - - vnet_buffer (b0)->snat.flags = 0; - ip0 = vlib_buffer_get_current (b0); - - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - rx_fib_index0 = - fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, - sw_if_index0); - - if (PREDICT_FALSE (ip0->ttl == 1)) - { - vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next0 = NAT_NEXT_ICMP_ERROR; - goto trace0; - } + next[0] = NAT_NEXT_OUT2IN_ED_SLOW_PATH; + goto trace0; + } - udp0 = ip4_next_header (ip0); - tcp0 = (tcp_header_t *) udp0; - proto0 = ip_proto_to_nat_proto (ip0->protocol); + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + { + next[0] = NAT_NEXT_OUT2IN_ED_SLOW_PATH; + goto trace0; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) + init_ed_k (&kv0, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_dst_port, ip0->src_address, + vnet_buffer (b0)->ip.reass.l4_src_port, rx_fib_index0, + ip0->protocol); + + /* there is a stashed index in vnet_buffer2 from handoff node, + * see if we can use it */ + if (is_multi_worker + && + PREDICT_TRUE (!pool_is_free_index + (tsm->sessions, + vnet_buffer2 (b0)->nat.ed_out2in_nat_session_index))) + { + s0 = pool_elt_at_index (tsm->sessions, + vnet_buffer2 (b0)-> + nat.ed_out2in_nat_session_index); + if (PREDICT_TRUE + (s0->out2in.addr.as_u32 == ip0->dst_address.as_u32 + && s0->out2in.port == vnet_buffer (b0)->ip.reass.l4_dst_port + && s0->nat_proto == ip_proto_to_nat_proto (ip0->protocol) + && s0->out2in.fib_index == rx_fib_index0 + && s0->ext_host_addr.as_u32 == ip0->src_address.as_u32 + && s0->ext_host_port == + vnet_buffer (b0)->ip.reass.l4_src_port)) { - next0 = NAT_NEXT_OUT2IN_ED_SLOW_PATH; - goto trace0; + /* yes, this is the droid we're looking for */ + goto skip_lookup; } + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) - { - next0 = NAT_NEXT_OUT2IN_ED_SLOW_PATH; - goto trace0; - } + if (clib_bihash_search_16_8 (&sm->out2in_ed, &kv0, &value0)) + { + next[0] = NAT_NEXT_OUT2IN_ED_SLOW_PATH; + goto trace0; + } + ASSERT (thread_index == ed_value_get_thread_index (&value0)); + s0 = + pool_elt_at_index (tsm->sessions, + ed_value_get_session_index (&value0)); - init_ed_k (&kv0, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_dst_port, ip0->src_address, - vnet_buffer (b0)->ip.reass.l4_src_port, rx_fib_index0, - ip0->protocol); - - /* there is a stashed index in vnet_buffer2 from handoff node, - * see if we can use it */ - if (is_multi_worker && PREDICT_TRUE - (!pool_is_free_index - (tsm->sessions, - vnet_buffer2 (b0)->nat.ed_out2in_nat_session_index))) + skip_lookup: + if (s0->tcp_closed_timestamp) + { + if (now >= s0->tcp_closed_timestamp) { - s0 = pool_elt_at_index (tsm->sessions, - vnet_buffer2 (b0)-> - nat.ed_out2in_nat_session_index); - if (PREDICT_TRUE - (s0->out2in.addr.as_u32 == ip0->dst_address.as_u32 - && s0->out2in.port == - vnet_buffer (b0)->ip.reass.l4_dst_port - && s0->nat_proto == ip_proto_to_nat_proto (ip0->protocol) - && s0->out2in.fib_index == rx_fib_index0 - && s0->ext_host_addr.as_u32 == ip0->src_address.as_u32 - && s0->ext_host_port == - vnet_buffer (b0)->ip.reass.l4_src_port)) - { - /* yes, this is the droid we're looking for */ - goto skip_lookup; - } + // session is closed, go slow path + next[0] = NAT_NEXT_OUT2IN_ED_SLOW_PATH; } - - if (clib_bihash_search_16_8 (&sm->out2in_ed, &kv0, &value0)) + else { - next0 = NAT_NEXT_OUT2IN_ED_SLOW_PATH; - goto trace0; + // session in transitory timeout, drop + b0->error = node->errors[NAT_OUT2IN_ED_ERROR_TCP_CLOSED]; + next[0] = NAT_NEXT_DROP; } - ASSERT (thread_index == ed_value_get_thread_index (&value0)); - s0 = - pool_elt_at_index (tsm->sessions, - ed_value_get_session_index (&value0)); + goto trace0; + } - skip_lookup: - if (s0->tcp_closed_timestamp) - { - if (now >= s0->tcp_closed_timestamp) - { - // session is closed, go slow path - next0 = NAT_NEXT_OUT2IN_ED_SLOW_PATH; - } - else - { - // session in transitory timeout, drop - b0->error = node->errors[NAT_OUT2IN_ED_ERROR_TCP_CLOSED]; - next0 = NAT_NEXT_DROP; - } - goto trace0; - } + // drop if session expired + u64 sess_timeout_time; + sess_timeout_time = + s0->last_heard + (f64) nat44_session_get_timeout (sm, s0); + if (now >= sess_timeout_time) + { + // session is closed, go slow path + nat_free_session_data (sm, s0, thread_index, 0); + nat_ed_session_delete (sm, s0, thread_index, 1); + next[0] = NAT_NEXT_OUT2IN_ED_SLOW_PATH; + goto trace0; + } + // - // drop if session expired - u64 sess_timeout_time; - sess_timeout_time = s0->last_heard + - (f64) nat44_session_get_timeout (sm, s0); - if (now >= sess_timeout_time) - { - // session is closed, go slow path - nat_free_session_data (sm, s0, thread_index, 0); - nat_ed_session_delete (sm, s0, thread_index, 1); - next0 = NAT_NEXT_OUT2IN_ED_SLOW_PATH; - goto trace0; - } - // + old_addr0 = ip0->dst_address.as_u32; + new_addr0 = ip0->dst_address.as_u32 = s0->in2out.addr.as_u32; + vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->in2out.fib_index; - old_addr0 = ip0->dst_address.as_u32; - new_addr0 = ip0->dst_address.as_u32 = s0->in2out.addr.as_u32; - vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->in2out.fib_index; + sum0 = ip0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address); + if (PREDICT_FALSE (is_twice_nat_session (s0))) + sum0 = ip_csum_update (sum0, ip0->src_address.as_u32, + s0->ext_host_nat_addr.as_u32, ip4_header_t, + src_address); + ip0->checksum = ip_csum_fold (sum0); - sum0 = ip0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, - dst_address); - if (PREDICT_FALSE (is_twice_nat_session (s0))) - sum0 = ip_csum_update (sum0, ip0->src_address.as_u32, - s0->ext_host_nat_addr.as_u32, ip4_header_t, - src_address); - ip0->checksum = ip_csum_fold (sum0); + old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; - old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; - - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) - { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - new_port0 = udp0->dst_port = s0->in2out.port; - sum0 = tcp0->checksum; - sum0 = - ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, - dst_address); - sum0 = - ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, - length); - if (is_twice_nat_session (s0)) - { - sum0 = ip_csum_update (sum0, ip0->src_address.as_u32, - s0->ext_host_nat_addr.as_u32, - ip4_header_t, dst_address); - sum0 = - ip_csum_update (sum0, - vnet_buffer (b0)->ip. - reass.l4_src_port, - s0->ext_host_nat_port, ip4_header_t, - length); - tcp0->src_port = s0->ext_host_nat_port; - ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; - } - tcp0->checksum = ip_csum_fold (sum0); - } - tcp_packets++; - if (nat44_set_tcp_session_state_o2i - (sm, now, s0, - vnet_buffer (b0)->ip.reass.icmp_type_or_tcp_flags, - vnet_buffer (b0)->ip.reass.tcp_ack_number, - vnet_buffer (b0)->ip.reass.tcp_seq_number, thread_index)) - goto trace0; - } - else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment - && udp0->checksum) + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { new_port0 = udp0->dst_port = s0->in2out.port; - sum0 = udp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, - dst_address); - sum0 = ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, - length); - if (PREDICT_FALSE (is_twice_nat_session (s0))) + sum0 = tcp0->checksum; + sum0 = + ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, + length); + if (is_twice_nat_session (s0)) { sum0 = ip_csum_update (sum0, ip0->src_address.as_u32, s0->ext_host_nat_addr.as_u32, @@ -896,59 +848,91 @@ nat44_ed_out2in_fast_path_node_fn_inline (vlib_main_t * vm, vnet_buffer (b0)->ip.reass.l4_src_port, s0->ext_host_nat_port, ip4_header_t, length); - udp0->src_port = s0->ext_host_nat_port; + tcp0->src_port = s0->ext_host_nat_port; ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; } - udp0->checksum = ip_csum_fold (sum0); - udp_packets++; + tcp0->checksum = ip_csum_fold (sum0); } - else + tcp_packets++; + if (nat44_set_tcp_session_state_o2i + (sm, now, s0, + vnet_buffer (b0)->ip.reass.icmp_type_or_tcp_flags, + vnet_buffer (b0)->ip.reass.tcp_ack_number, + vnet_buffer (b0)->ip.reass.tcp_seq_number, thread_index)) + goto trace0; + } + else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment + && udp0->checksum) + { + new_port0 = udp0->dst_port = s0->in2out.port; + sum0 = udp0->checksum; + sum0 = + ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, length); + if (PREDICT_FALSE (is_twice_nat_session (s0))) + { + sum0 = + ip_csum_update (sum0, ip0->src_address.as_u32, + s0->ext_host_nat_addr.as_u32, ip4_header_t, + dst_address); + sum0 = + ip_csum_update (sum0, vnet_buffer (b0)->ip.reass.l4_src_port, + s0->ext_host_nat_port, ip4_header_t, length); + udp0->src_port = s0->ext_host_nat_port; + ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; + } + udp0->checksum = ip_csum_fold (sum0); + udp_packets++; + } + else + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + new_port0 = udp0->dst_port = s0->in2out.port; + if (PREDICT_FALSE (is_twice_nat_session (s0))) { - new_port0 = udp0->dst_port = s0->in2out.port; - if (PREDICT_FALSE (is_twice_nat_session (s0))) - { - udp0->src_port = s0->ext_host_nat_port; - ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; - } + udp0->src_port = s0->ext_host_nat_port; + ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; } - udp_packets++; } + udp_packets++; + } - /* Accounting */ - nat44_session_update_counters (s0, now, - vlib_buffer_length_in_chain (vm, b0), - thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s0, thread_index); - - trace0: - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) - { - nat44_ed_out2in_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - t->next_index = next0; - t->is_slow_path = 0; - - if (s0) - t->session_index = s0 - tsm->sessions; - else - t->session_index = ~0; - } + /* Accounting */ + nat44_session_update_counters (s0, now, + vlib_buffer_length_in_chain (vm, b0), + thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s0, thread_index); - pkts_processed += next0 == vnet_buffer2 (b0)->nat.arc_next; - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + trace0: + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + nat44_ed_out2in_trace_t *t = + vlib_add_trace (vm, node, b0, sizeof (*t)); + t->sw_if_index = sw_if_index0; + t->next_index = next[0]; + t->is_slow_path = 0; + + if (s0) + t->session_index = s0 - tsm->sessions; + else + t->session_index = ~0; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + pkts_processed += next[0] == vnet_buffer2 (b0)->nat.arc_next; + + n_left_from--; + next++; + b++; } + vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, + frame->n_vectors); + vlib_node_increment_counter (vm, stats_node_index, NAT_OUT2IN_ED_ERROR_OUT2IN_PACKETS, pkts_processed); @@ -972,8 +956,7 @@ nat44_ed_out2in_slow_path_node_fn_inline (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { - u32 n_left_from, *from, *to_next, pkts_processed = 0, stats_node_index; - nat_next_t next_index; + u32 n_left_from, *from, pkts_processed = 0, stats_node_index; snat_main_t *sm = &snat_main; f64 now = vlib_time_now (vm); u32 thread_index = vm->thread_index; @@ -985,256 +968,205 @@ nat44_ed_out2in_slow_path_node_fn_inline (vlib_main_t * vm, from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; - next_index = node->cached_next_index; + + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; + u16 nexts[VLIB_FRAME_SIZE], *next = nexts; + vlib_get_buffers (vm, from, b, n_left_from); while (n_left_from > 0) { - u32 n_left_to_next; + vlib_buffer_t *b0; + u32 sw_if_index0, rx_fib_index0, proto0, old_addr0, new_addr0; + u16 old_port0, new_port0; + ip4_header_t *ip0; + udp_header_t *udp0; + tcp_header_t *tcp0; + icmp46_header_t *icmp0; + snat_session_t *s0 = 0; + clib_bihash_kv_16_8_t kv0, value0; + ip_csum_t sum0; + lb_nat_type_t lb_nat0; + twice_nat_type_t twice_nat0; + u8 identity_nat0; + ip4_address_t sm_addr; + u16 sm_port; + u32 sm_fib_index; + + b0 = *b; + next[0] = vnet_buffer2 (b0)->nat.arc_next; + + vnet_buffer (b0)->snat.flags = 0; + ip0 = vlib_buffer_get_current (b0); + + sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + rx_fib_index0 = + fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, sw_if_index0); + + if (PREDICT_FALSE (ip0->ttl == 1)) + { + vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; + icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, + ICMP4_time_exceeded_ttl_exceeded_in_transit, + 0); + next[0] = NAT_NEXT_ICMP_ERROR; + goto trace0; + } - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + udp0 = ip4_next_header (ip0); + tcp0 = (tcp_header_t *) udp0; + icmp0 = (icmp46_header_t *) udp0; + proto0 = ip_proto_to_nat_proto (ip0->protocol); - while (n_left_from > 0 && n_left_to_next > 0) + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) { - u32 bi0; - vlib_buffer_t *b0; - u32 next0, sw_if_index0, rx_fib_index0, proto0, old_addr0, - new_addr0; - u16 old_port0, new_port0; - ip4_header_t *ip0; - udp_header_t *udp0; - tcp_header_t *tcp0; - icmp46_header_t *icmp0; - snat_session_t *s0 = 0; - clib_bihash_kv_16_8_t kv0, value0; - ip_csum_t sum0; - lb_nat_type_t lb_nat0; - twice_nat_type_t twice_nat0; - u8 identity_nat0; - ip4_address_t sm_addr; - u16 sm_port; - u32 sm_fib_index; - - /* speculatively enqueue b0 to the current next frame */ - bi0 = from[0]; - to_next[0] = bi0; - from += 1; - to_next += 1; - n_left_from -= 1; - n_left_to_next -= 1; - - b0 = vlib_get_buffer (vm, bi0); - next0 = vnet_buffer2 (b0)->nat.arc_next; - - vnet_buffer (b0)->snat.flags = 0; - ip0 = vlib_buffer_get_current (b0); - - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; - rx_fib_index0 = - fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4, - sw_if_index0); - - if (PREDICT_FALSE (ip0->ttl == 1)) + s0 = + nat44_ed_out2in_unknown_proto (sm, b0, ip0, rx_fib_index0, + thread_index, now, vm, node); + if (!sm->forwarding_enabled) { - vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0; - icmp4_error_set_vnet_buffer (b0, ICMP4_time_exceeded, - ICMP4_time_exceeded_ttl_exceeded_in_transit, - 0); - next0 = NAT_NEXT_ICMP_ERROR; - goto trace0; + if (!s0) + next[0] = NAT_NEXT_DROP; } + other_packets++; + goto trace0; + } - udp0 = ip4_next_header (ip0); - tcp0 = (tcp_header_t *) udp0; - icmp0 = (icmp46_header_t *) udp0; - proto0 = ip_proto_to_nat_proto (ip0->protocol); + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + { + next[0] = icmp_out2in_ed_slow_path + (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, node, + next[0], now, thread_index, &s0); + icmp_packets++; + goto trace0; + } - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_OTHER)) - { - s0 = - nat44_ed_out2in_unknown_proto (sm, b0, ip0, rx_fib_index0, - thread_index, now, vm, node); - if (!sm->forwarding_enabled) - { - if (!s0) - next0 = NAT_NEXT_DROP; - } - other_packets++; - goto trace0; - } + init_ed_k (&kv0, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_dst_port, ip0->src_address, + vnet_buffer (b0)->ip.reass.l4_src_port, rx_fib_index0, + ip0->protocol); - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_ICMP)) + s0 = NULL; + if (!clib_bihash_search_16_8 (&sm->out2in_ed, &kv0, &value0)) + { + ASSERT (thread_index == ed_value_get_thread_index (&value0)); + s0 = + pool_elt_at_index (tsm->sessions, + ed_value_get_session_index (&value0)); + + if (s0->tcp_closed_timestamp && now >= s0->tcp_closed_timestamp) { - next0 = icmp_out2in_ed_slow_path - (sm, b0, ip0, icmp0, sw_if_index0, rx_fib_index0, node, - next0, now, thread_index, &s0); - icmp_packets++; - goto trace0; + nat_free_session_data (sm, s0, thread_index, 0); + nat_ed_session_delete (sm, s0, thread_index, 1); + s0 = NULL; } + } - init_ed_k (&kv0, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_dst_port, ip0->src_address, - vnet_buffer (b0)->ip.reass.l4_src_port, rx_fib_index0, - ip0->protocol); - - s0 = NULL; - if (!clib_bihash_search_16_8 (&sm->out2in_ed, &kv0, &value0)) + if (!s0) + { + /* Try to match static mapping by external address and port, + destination address and port in packet */ + + if (snat_static_mapping_match + (sm, ip0->dst_address, + vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, + proto0, &sm_addr, &sm_port, &sm_fib_index, 1, 0, + &twice_nat0, &lb_nat0, &ip0->src_address, &identity_nat0)) { - ASSERT (thread_index == ed_value_get_thread_index (&value0)); - s0 = - pool_elt_at_index (tsm->sessions, - ed_value_get_session_index (&value0)); - - if (s0->tcp_closed_timestamp && now >= s0->tcp_closed_timestamp) + /* + * Send DHCP packets to the ipv4 stack, or we won't + * be able to use dhcp client on the outside interface + */ + if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_UDP + && (vnet_buffer (b0)->ip.reass.l4_dst_port == + clib_host_to_net_u16 + (UDP_DST_PORT_dhcp_to_client)))) { - nat_free_session_data (sm, s0, thread_index, 0); - nat_ed_session_delete (sm, s0, thread_index, 1); - s0 = NULL; + goto trace0; } - } - if (!s0) - { - /* Try to match static mapping by external address and port, - destination address and port in packet */ - - if (snat_static_mapping_match - (sm, ip0->dst_address, - vnet_buffer (b0)->ip.reass.l4_dst_port, rx_fib_index0, - proto0, &sm_addr, &sm_port, &sm_fib_index, 1, 0, - &twice_nat0, &lb_nat0, &ip0->src_address, &identity_nat0)) + if (!sm->forwarding_enabled) + { + b0->error = + node->errors[NAT_OUT2IN_ED_ERROR_NO_TRANSLATION]; + next[0] = NAT_NEXT_DROP; + } + else { - /* - * Send DHCP packets to the ipv4 stack, or we won't - * be able to use dhcp client on the outside interface - */ - if (PREDICT_FALSE (proto0 == NAT_PROTOCOL_UDP - && (vnet_buffer (b0)->ip. - reass.l4_dst_port == - clib_host_to_net_u16 - (UDP_DST_PORT_dhcp_to_client)))) + if (next_src_nat + (sm, ip0, vnet_buffer (b0)->ip.reass.l4_src_port, + vnet_buffer (b0)->ip.reass.l4_dst_port, + thread_index, rx_fib_index0)) { + next[0] = NAT_NEXT_IN2OUT_ED_FAST_PATH; goto trace0; } - - if (!sm->forwarding_enabled) - { - b0->error = - node->errors[NAT_OUT2IN_ED_ERROR_NO_TRANSLATION]; - next0 = NAT_NEXT_DROP; - } + if (sm->num_workers > 1) + create_bypass_for_fwd_worker (sm, b0, ip0, rx_fib_index0); else - { - if (next_src_nat - (sm, ip0, vnet_buffer (b0)->ip.reass.l4_src_port, - vnet_buffer (b0)->ip.reass.l4_dst_port, - thread_index, rx_fib_index0)) - { - next0 = NAT_NEXT_IN2OUT_ED_FAST_PATH; - goto trace0; - } - if (sm->num_workers > 1) - create_bypass_for_fwd_worker (sm, b0, ip0, - rx_fib_index0); - else - create_bypass_for_fwd (sm, b0, ip0, rx_fib_index0, - thread_index); - } - goto trace0; + create_bypass_for_fwd (sm, b0, ip0, rx_fib_index0, + thread_index); } + goto trace0; + } - if (PREDICT_FALSE (identity_nat0)) - goto trace0; + if (PREDICT_FALSE (identity_nat0)) + goto trace0; - if ((proto0 == NAT_PROTOCOL_TCP) - && !tcp_flags_is_init (vnet_buffer (b0)->ip. - reass.icmp_type_or_tcp_flags)) - { - b0->error = node->errors[NAT_OUT2IN_ED_ERROR_NON_SYN]; - next0 = NAT_NEXT_DROP; - goto trace0; - } + if ((proto0 == NAT_PROTOCOL_TCP) + && !tcp_flags_is_init (vnet_buffer (b0)->ip. + reass.icmp_type_or_tcp_flags)) + { + b0->error = node->errors[NAT_OUT2IN_ED_ERROR_NON_SYN]; + next[0] = NAT_NEXT_DROP; + goto trace0; + } - /* Create session initiated by host from external network */ - s0 = create_session_for_static_mapping_ed (sm, b0, - sm_addr, sm_port, - sm_fib_index, - ip0->dst_address, - vnet_buffer (b0)-> - ip.reass.l4_dst_port, - rx_fib_index0, - proto0, node, - rx_fib_index0, - thread_index, - twice_nat0, lb_nat0, - now); - if (!s0) - { - next0 = NAT_NEXT_DROP; - goto trace0; - } + /* Create session initiated by host from external network */ + s0 = create_session_for_static_mapping_ed (sm, b0, + sm_addr, sm_port, + sm_fib_index, + ip0->dst_address, + vnet_buffer (b0)-> + ip.reass.l4_dst_port, + rx_fib_index0, proto0, + node, rx_fib_index0, + thread_index, twice_nat0, + lb_nat0, now); + if (!s0) + { + next[0] = NAT_NEXT_DROP; + goto trace0; } + } - old_addr0 = ip0->dst_address.as_u32; - new_addr0 = ip0->dst_address.as_u32 = s0->in2out.addr.as_u32; - vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->in2out.fib_index; + old_addr0 = ip0->dst_address.as_u32; + new_addr0 = ip0->dst_address.as_u32 = s0->in2out.addr.as_u32; + vnet_buffer (b0)->sw_if_index[VLIB_TX] = s0->in2out.fib_index; - sum0 = ip0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, - dst_address); - if (PREDICT_FALSE (is_twice_nat_session (s0))) - sum0 = ip_csum_update (sum0, ip0->src_address.as_u32, - s0->ext_host_nat_addr.as_u32, ip4_header_t, - src_address); - ip0->checksum = ip_csum_fold (sum0); + sum0 = ip0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address); + if (PREDICT_FALSE (is_twice_nat_session (s0))) + sum0 = ip_csum_update (sum0, ip0->src_address.as_u32, + s0->ext_host_nat_addr.as_u32, ip4_header_t, + src_address); + ip0->checksum = ip_csum_fold (sum0); - old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; + old_port0 = vnet_buffer (b0)->ip.reass.l4_dst_port; - if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) - { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) - { - new_port0 = udp0->dst_port = s0->in2out.port; - sum0 = tcp0->checksum; - sum0 = - ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, - dst_address); - sum0 = - ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, - length); - if (is_twice_nat_session (s0)) - { - sum0 = ip_csum_update (sum0, ip0->src_address.as_u32, - s0->ext_host_nat_addr.as_u32, - ip4_header_t, dst_address); - sum0 = - ip_csum_update (sum0, - vnet_buffer (b0)->ip. - reass.l4_src_port, - s0->ext_host_nat_port, ip4_header_t, - length); - tcp0->src_port = s0->ext_host_nat_port; - ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; - } - tcp0->checksum = ip_csum_fold (sum0); - } - tcp_packets++; - if (nat44_set_tcp_session_state_o2i - (sm, now, s0, - vnet_buffer (b0)->ip.reass.icmp_type_or_tcp_flags, - vnet_buffer (b0)->ip.reass.tcp_ack_number, - vnet_buffer (b0)->ip.reass.tcp_seq_number, thread_index)) - goto trace0; - } - else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment - && udp0->checksum) + if (PREDICT_TRUE (proto0 == NAT_PROTOCOL_TCP)) + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) { new_port0 = udp0->dst_port = s0->in2out.port; - sum0 = udp0->checksum; - sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, - dst_address); - sum0 = ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, - length); - if (PREDICT_FALSE (is_twice_nat_session (s0))) + sum0 = tcp0->checksum; + sum0 = + ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address); + sum0 = + ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, + length); + if (is_twice_nat_session (s0)) { sum0 = ip_csum_update (sum0, ip0->src_address.as_u32, s0->ext_host_nat_addr.as_u32, @@ -1244,59 +1176,90 @@ nat44_ed_out2in_slow_path_node_fn_inline (vlib_main_t * vm, vnet_buffer (b0)->ip.reass.l4_src_port, s0->ext_host_nat_port, ip4_header_t, length); - udp0->src_port = s0->ext_host_nat_port; + tcp0->src_port = s0->ext_host_nat_port; ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; } - udp0->checksum = ip_csum_fold (sum0); - udp_packets++; + tcp0->checksum = ip_csum_fold (sum0); } - else + tcp_packets++; + if (nat44_set_tcp_session_state_o2i + (sm, now, s0, + vnet_buffer (b0)->ip.reass.icmp_type_or_tcp_flags, + vnet_buffer (b0)->ip.reass.tcp_ack_number, + vnet_buffer (b0)->ip.reass.tcp_seq_number, thread_index)) + goto trace0; + } + else if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment + && udp0->checksum) + { + new_port0 = udp0->dst_port = s0->in2out.port; + sum0 = udp0->checksum; + sum0 = ip_csum_update (sum0, old_addr0, new_addr0, ip4_header_t, + dst_address); + sum0 = ip_csum_update (sum0, old_port0, new_port0, ip4_header_t, + length); + if (PREDICT_FALSE (is_twice_nat_session (s0))) { - if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + sum0 = ip_csum_update (sum0, ip0->src_address.as_u32, + s0->ext_host_nat_addr.as_u32, + ip4_header_t, dst_address); + sum0 = + ip_csum_update (sum0, + vnet_buffer (b0)->ip.reass.l4_src_port, + s0->ext_host_nat_port, ip4_header_t, length); + udp0->src_port = s0->ext_host_nat_port; + ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; + } + udp0->checksum = ip_csum_fold (sum0); + udp_packets++; + } + else + { + if (!vnet_buffer (b0)->ip.reass.is_non_first_fragment) + { + new_port0 = udp0->dst_port = s0->in2out.port; + if (PREDICT_FALSE (is_twice_nat_session (s0))) { - new_port0 = udp0->dst_port = s0->in2out.port; - if (PREDICT_FALSE (is_twice_nat_session (s0))) - { - udp0->src_port = s0->ext_host_nat_port; - ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; - } + udp0->src_port = s0->ext_host_nat_port; + ip0->src_address.as_u32 = s0->ext_host_nat_addr.as_u32; } - udp_packets++; } + udp_packets++; + } - /* Accounting */ - nat44_session_update_counters (s0, now, - vlib_buffer_length_in_chain (vm, b0), - thread_index); - /* Per-user LRU list maintenance */ - nat44_session_update_lru (sm, s0, thread_index); - - trace0: - if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) - && (b0->flags & VLIB_BUFFER_IS_TRACED))) - { - nat44_ed_out2in_trace_t *t = - vlib_add_trace (vm, node, b0, sizeof (*t)); - t->sw_if_index = sw_if_index0; - t->next_index = next0; - t->is_slow_path = 1; - - if (s0) - t->session_index = s0 - tsm->sessions; - else - t->session_index = ~0; - } + /* Accounting */ + nat44_session_update_counters (s0, now, + vlib_buffer_length_in_chain (vm, b0), + thread_index); + /* Per-user LRU list maintenance */ + nat44_session_update_lru (sm, s0, thread_index); - pkts_processed += next0 == vnet_buffer2 (b0)->nat.arc_next; - /* verify speculative enqueue, maybe switch current next frame */ - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, - to_next, n_left_to_next, - bi0, next0); + trace0: + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + nat44_ed_out2in_trace_t *t = + vlib_add_trace (vm, node, b0, sizeof (*t)); + t->sw_if_index = sw_if_index0; + t->next_index = next[0]; + t->is_slow_path = 1; + + if (s0) + t->session_index = s0 - tsm->sessions; + else + t->session_index = ~0; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + pkts_processed += next[0] == vnet_buffer2 (b0)->nat.arc_next; + + n_left_from--; + next++; + b++; } + vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, + frame->n_vectors); + vlib_node_increment_counter (vm, stats_node_index, NAT_OUT2IN_ED_ERROR_OUT2IN_PACKETS, pkts_processed); -- cgit 1.2.3-korg