From 77eb28f4d355884e0f6be2f4323ccce3c452dff8 Mon Sep 17 00:00:00 2001 From: Neale Ranns Date: Mon, 4 Mar 2019 14:13:14 +0000 Subject: IPSEC: tunnel encap/decap dual loop speedups baseline: ipsec0-tx 1.27e1 ipsec-if-input 8.19e1 this change: ipsec0-tx 6.17e0 ipsec-if-input 6.39e1 this also fixes the double tunnel TX counts by removing the duplicate from the TX node. Change-Id: Ie4608acda08dc653b6fb9e2c85185d83625efd40 Signed-off-by: Neale Ranns --- src/vnet/ipsec/ipsec.h | 1 - src/vnet/ipsec/ipsec_if.c | 157 ++++++++++++++++-------------------- src/vnet/ipsec/ipsec_if_in.c | 186 +++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 255 insertions(+), 89 deletions(-) (limited to 'src/vnet') diff --git a/src/vnet/ipsec/ipsec.h b/src/vnet/ipsec/ipsec.h index 3279fab6701..78d000030c9 100644 --- a/src/vnet/ipsec/ipsec.h +++ b/src/vnet/ipsec/ipsec.h @@ -140,7 +140,6 @@ typedef struct uword *sa_index_by_sa_id; uword *ipsec_if_pool_index_by_key; uword *ipsec_if_real_dev_by_show_dev; - u32 *ipsec_if_by_sw_if_index; /* node indices */ u32 error_drop_node_index; diff --git a/src/vnet/ipsec/ipsec_if.c b/src/vnet/ipsec/ipsec_if.c index d03d1a9ffe3..5150e95d6de 100644 --- a/src/vnet/ipsec/ipsec_if.c +++ b/src/vnet/ipsec/ipsec_if.c @@ -70,101 +70,88 @@ format_ipsec_if_tx_trace (u8 * s, va_list * args) return s; } -always_inline ipsec_tunnel_if_t * -ipsec_tun_get_by_sw_if_index (u32 sw_if_index) +static void +ipsec_output_trace (vlib_main_t * vm, + vlib_node_runtime_t * node, + vlib_frame_t * frame, const ipsec_tunnel_if_t * t0) { ipsec_main_t *im = &ipsec_main; - u32 ti; + u32 *from, n_left; - ti = im->ipsec_if_by_sw_if_index[sw_if_index]; + n_left = frame->n_vectors; + from = vlib_frame_vector_args (frame); - return (pool_elt_at_index (im->tunnel_interfaces, ti)); -} - -static uword -ipsec_if_tx_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, - vlib_frame_t * from_frame) -{ - ipsec_main_t *im = &ipsec_main; - vnet_main_t *vnm = im->vnet_main; - vnet_interface_main_t *vim = &vnm->interface_main; - u32 *from, *to_next = 0, next_index; - u32 n_left_from, sw_if_index0, last_sw_if_index = ~0; - u32 thread_index = vm->thread_index; - u32 n_bytes = 0, n_packets = 0; - - from = vlib_frame_vector_args (from_frame); - n_left_from = from_frame->n_vectors; - next_index = node->cached_next_index; - - while (n_left_from > 0) + while (n_left > 0) { - u32 n_left_to_next; + vlib_buffer_t *b0; - vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + b0 = vlib_get_buffer (vm, from[0]); - while (n_left_from > 0 && n_left_to_next > 0) + if (b0->flags & VLIB_BUFFER_IS_TRACED) { - const ipsec_tunnel_if_t *t0; - u32 bi0, next0, len0; - vlib_buffer_t *b0; - - bi0 = to_next[0] = from[0]; - from += 1; - n_left_from -= 1; - to_next += 1; - n_left_to_next -= 1; - b0 = vlib_get_buffer (vm, bi0); - sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_TX]; - t0 = ipsec_tun_get_by_sw_if_index (sw_if_index0); - vnet_buffer (b0)->ipsec.sad_index = t0->output_sa_index; - - /* 0, tx-node next[0] was added by vlib_node_add_next_with_slot */ - next0 = 0; - - len0 = vlib_buffer_length_in_chain (vm, b0); - - if (PREDICT_TRUE (sw_if_index0 == last_sw_if_index)) - { - n_packets++; - n_bytes += len0; - } - else - { - vlib_increment_combined_counter (vim->combined_sw_if_counters + - VNET_INTERFACE_COUNTER_TX, - thread_index, sw_if_index0, - n_packets, n_bytes); - last_sw_if_index = sw_if_index0; - n_packets = 1; - n_bytes = len0; - } - - if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) - { - ipsec_if_tx_trace_t *tr = - vlib_add_trace (vm, node, b0, sizeof (*tr)); - ipsec_sa_t *sa0 = - pool_elt_at_index (im->sad, t0->output_sa_index); - tr->spi = sa0->spi; - tr->seq = sa0->seq; - } - - vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, - n_left_to_next, bi0, next0); + ipsec_if_tx_trace_t *tr = + vlib_add_trace (vm, node, b0, sizeof (*tr)); + ipsec_sa_t *sa0 = pool_elt_at_index (im->sad, t0->output_sa_index); + tr->spi = sa0->spi; + tr->seq = sa0->seq; } - vlib_put_next_frame (vm, node, next_index, n_left_to_next); + + from += 1; + n_left -= 1; } +} + +VNET_DEVICE_CLASS_TX_FN (ipsec_device_class) (vlib_main_t * vm, + vlib_node_runtime_t * node, + vlib_frame_t * frame) +{ + ipsec_main_t *im = &ipsec_main; + u32 *from, n_left; + vnet_interface_output_runtime_t *rd = (void *) node->runtime_data; + const ipsec_tunnel_if_t *t0; + vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b; + u16 nexts[VLIB_FRAME_SIZE]; - if (last_sw_if_index != ~0) + from = vlib_frame_vector_args (frame); + t0 = pool_elt_at_index (im->tunnel_interfaces, rd->dev_instance); + n_left = frame->n_vectors; + b = bufs; + + /* All going to encrypt */ + clib_memset (nexts, 0, sizeof (nexts)); + + if (node->flags & VLIB_NODE_FLAG_TRACE) + ipsec_output_trace (vm, node, frame, t0); + + vlib_get_buffers (vm, from, bufs, n_left); + + while (n_left >= 8) { - vlib_increment_combined_counter (vim->combined_sw_if_counters + - VNET_INTERFACE_COUNTER_TX, - thread_index, - last_sw_if_index, n_packets, n_bytes); + /* Prefetch the buffer header for the N+2 loop iteration */ + vlib_prefetch_buffer_header (b[4], STORE); + vlib_prefetch_buffer_header (b[5], STORE); + vlib_prefetch_buffer_header (b[6], STORE); + vlib_prefetch_buffer_header (b[7], STORE); + + vnet_buffer (b[0])->ipsec.sad_index = t0->output_sa_index; + vnet_buffer (b[1])->ipsec.sad_index = t0->output_sa_index; + vnet_buffer (b[2])->ipsec.sad_index = t0->output_sa_index; + vnet_buffer (b[3])->ipsec.sad_index = t0->output_sa_index; + + n_left -= 4; + b += 4; } + while (n_left > 0) + { + vnet_buffer (b[0])->ipsec.sad_index = t0->output_sa_index; - return from_frame->n_vectors; + n_left -= 1; + b += 1; + } + + vlib_buffer_enqueue_to_next (vm, node, from, nexts, frame->n_vectors); + + return frame->n_vectors; } @@ -224,12 +211,11 @@ ipsec_admin_up_down_function (vnet_main_t * vnm, u32 hw_if_index, u32 flags) /* *INDENT-OFF* */ -VNET_DEVICE_CLASS (ipsec_device_class, static) = +VNET_DEVICE_CLASS (ipsec_device_class) = { .name = "IPSec", .format_device_name = format_ipsec_name, .format_tx_trace = format_ipsec_if_tx_trace, - .tx_function = ipsec_if_tx_node_fn, .tx_function_n_errors = IPSEC_IF_TX_N_ERROR, .tx_function_error_strings = ipsec_if_tx_error_strings, .admin_up_down_function = ipsec_admin_up_down_function, @@ -384,10 +370,6 @@ ipsec_add_del_tunnel_if_internal (vnet_main_t * vnm, t->hw_if_index = hw_if_index; t->sw_if_index = hi->sw_if_index; - vec_validate_init_empty (im->ipsec_if_by_sw_if_index, - t->sw_if_index, ~0); - im->ipsec_if_by_sw_if_index[t->sw_if_index] = t - im->tunnel_interfaces; - vnet_feature_enable_disable ("interface-output", "ipsec-if-output", hi->sw_if_index, 1, 0, 0); @@ -414,7 +396,6 @@ ipsec_add_del_tunnel_if_internal (vnet_main_t * vnm, hash_unset (im->ipsec_if_pool_index_by_key, key); hash_unset (im->ipsec_if_real_dev_by_show_dev, t->show_instance); - im->ipsec_if_by_sw_if_index[t->sw_if_index] = ~0; pool_put (im->tunnel_interfaces, t); diff --git a/src/vnet/ipsec/ipsec_if_in.c b/src/vnet/ipsec/ipsec_if_in.c index c7f5735a64e..9979446b71a 100644 --- a/src/vnet/ipsec/ipsec_if_in.c +++ b/src/vnet/ipsec/ipsec_if_in.c @@ -90,6 +90,192 @@ VLIB_NODE_FN (ipsec_if_input_node) (vlib_main_t * vm, vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + while (n_left_from >= 4 && n_left_to_next >= 2) + { + u32 bi0, bi1, next0, next1, sw_if_index0, sw_if_index1; + const esp_header_t *esp0, *esp1; + const ip4_header_t *ip0, *ip1; + vlib_buffer_t *b0, *b1; + uword *p0, *p1; + u32 len0, len1; + u64 key0, key1; + + /* Prefetch next iteration. */ + { + vlib_buffer_t *p2, *p3; + + p2 = vlib_get_buffer (vm, from[2]); + p3 = vlib_get_buffer (vm, from[3]); + + vlib_prefetch_buffer_header (p2, STORE); + vlib_prefetch_buffer_header (p3, STORE); + + CLIB_PREFETCH (p2->data, sizeof (ip0[0]), STORE); + CLIB_PREFETCH (p3->data, sizeof (ip0[0]), STORE); + } + + bi0 = to_next[0] = from[0]; + bi1 = to_next[1] = from[1]; + + from += 2; + n_left_from -= 2; + to_next += 2; + n_left_to_next -= 2; + next0 = next1 = IPSEC_INPUT_NEXT_DROP; + + b0 = vlib_get_buffer (vm, bi0); + b1 = vlib_get_buffer (vm, bi1); + ip0 = vlib_buffer_get_current (b0); + ip1 = vlib_buffer_get_current (b1); + esp0 = (const esp_header_t *) ((u8 *) ip0 + ip4_header_bytes (ip0)); + esp1 = (const esp_header_t *) ((u8 *) ip1 + ip4_header_bytes (ip1)); + + key0 = (u64) ip0->src_address.as_u32 << 32 | (u64) esp0->spi; + key1 = (u64) ip1->src_address.as_u32 << 32 | (u64) esp1->spi; + + p0 = hash_get (im->ipsec_if_pool_index_by_key, key0); + p1 = hash_get (im->ipsec_if_pool_index_by_key, key1); + + /* stats for the tunnel include all the data after the IP header + just like a norml IP-IP tunnel */ + vlib_buffer_advance (b0, ip4_header_bytes (ip0)); + vlib_buffer_advance (b1, ip4_header_bytes (ip1)); + len0 = vlib_buffer_length_in_chain (vm, b0); + len1 = vlib_buffer_length_in_chain (vm, b1); + + if (PREDICT_TRUE (NULL != p0)) + { + const ipsec_tunnel_if_t *t0; + + t0 = pool_elt_at_index (im->tunnel_interfaces, p0[0]); + vnet_buffer (b0)->ipsec.sad_index = t0->input_sa_index; + + if (PREDICT_TRUE (t0->hw_if_index != ~0)) + { + vnet_buffer (b0)->ipsec.flags = 0; + sw_if_index0 = t0->sw_if_index; + vnet_buffer (b0)->sw_if_index[VLIB_RX] = sw_if_index0; + + if (PREDICT_FALSE + (!(t0->flags & VNET_HW_INTERFACE_FLAG_LINK_UP))) + { + vlib_increment_combined_counter + (drop_counter, thread_index, sw_if_index0, 1, len0); + b0->error = node->errors[IPSEC_IF_INPUT_ERROR_DISABLED]; + n_disabled++; + goto pkt1; + } + + if (PREDICT_TRUE (sw_if_index0 == last_sw_if_index)) + { + n_packets++; + n_bytes += len0; + } + else + { + if (last_t) + { + vlib_increment_combined_counter + (rx_counter, thread_index, sw_if_index0, + n_packets, n_bytes); + } + + last_sw_if_index = sw_if_index0; + last_t = t0; + n_packets = 1; + n_bytes = len0; + } + } + else + { + vnet_buffer (b0)->ipsec.flags = IPSEC_FLAG_IPSEC_GRE_TUNNEL; + } + + next0 = im->esp4_decrypt_next_index; + } + else + { + b0->error = node->errors[IPSEC_IF_INPUT_ERROR_NO_TUNNEL]; + n_no_tunnel++; + } + + pkt1: + if (PREDICT_TRUE (NULL != p1)) + { + const ipsec_tunnel_if_t *t1; + + t1 = pool_elt_at_index (im->tunnel_interfaces, p1[0]); + vnet_buffer (b1)->ipsec.sad_index = t1->input_sa_index; + + if (PREDICT_TRUE (t1->hw_if_index != ~0)) + { + vnet_buffer (b1)->ipsec.flags = 0; + sw_if_index1 = t1->sw_if_index; + vnet_buffer (b1)->sw_if_index[VLIB_RX] = sw_if_index1; + + if (PREDICT_FALSE + (!(t1->flags & VNET_HW_INTERFACE_FLAG_LINK_UP))) + { + vlib_increment_combined_counter + (drop_counter, thread_index, sw_if_index1, 1, len1); + b1->error = node->errors[IPSEC_IF_INPUT_ERROR_DISABLED]; + n_disabled++; + goto trace1; + } + + if (PREDICT_TRUE (sw_if_index1 == last_sw_if_index)) + { + n_packets++; + n_bytes += len1; + } + else + { + if (last_t) + { + vlib_increment_combined_counter + (rx_counter, thread_index, sw_if_index1, + n_packets, n_bytes); + } + + last_sw_if_index = sw_if_index1; + last_t = t1; + n_packets = 1; + n_bytes = len1; + } + } + else + { + vnet_buffer (b1)->ipsec.flags = IPSEC_FLAG_IPSEC_GRE_TUNNEL; + } + + next1 = im->esp4_decrypt_next_index; + } + else + { + b1->error = node->errors[IPSEC_IF_INPUT_ERROR_NO_TUNNEL]; + n_no_tunnel++; + } + + trace1: + if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) + { + ipsec_if_input_trace_t *tr = + vlib_add_trace (vm, node, b0, sizeof (*tr)); + tr->spi = clib_host_to_net_u32 (esp0->spi); + tr->seq = clib_host_to_net_u32 (esp0->seq); + } + if (PREDICT_FALSE (b1->flags & VLIB_BUFFER_IS_TRACED)) + { + ipsec_if_input_trace_t *tr = + vlib_add_trace (vm, node, b1, sizeof (*tr)); + tr->spi = clib_host_to_net_u32 (esp1->spi); + tr->seq = clib_host_to_net_u32 (esp1->seq); + } + + vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next, + n_left_to_next, + bi0, bi1, next0, next1); + } while (n_left_from > 0 && n_left_to_next > 0) { u32 bi0, next0, sw_if_index0; -- cgit 1.2.3-korg