aboutsummaryrefslogtreecommitdiffstats
path: root/tests/vpp
AgeCommit message (Expand)AuthorFilesLines
2018-10-12Tag l2bdscale as L2BDSCALE, not L2BDBASEVratko Polak24-24/+24
2018-10-10CSIT-1297: Modify the bootstrap.sh to run only selected mrr suitesTibor Frank10-10/+10
2018-10-10Add VXLAN scale perf testsJan Gelety20-60/+2228
2018-10-02CSIT-1325 - FIX: AVF tests failing on unbindPeter Mikus4-4/+4
2018-10-01CSIT-1327 Migrate from vpp-dkms-dpdk to vpp-ext-depsPeter Mikus1-1/+1
2018-09-28CSIT-1321 Add IP4base tests using AVF driverPeter Mikus4-0/+484
2018-09-21Fix autogen regenerator and few suitesVratko Polak12-200/+152
2018-09-20FIX: 40Ge2P1Vic1385-Eth-L2Bdbasemaclrn-Mrr typoPeter Mikus1-1/+1
2018-09-20Add dot1q-l2bdbase-2memif-[ndrpdr|mrr] perf tests for 2-node topologyJan Gelety8-0/+1242
2018-09-20Add dot1q-l2bdbase-2vhost-[ndrpdr|mrr] perf tests for 2-node topologyJan Gelety8-94/+696
2018-09-20CSIT-1205 Create AVF driver test - Add Tests L2BD L2XCPeter Mikus8-0/+969
2018-09-20Create basic 2-node topology performance tests for Intel-X553 NICYulong Pei10-0/+1280
2018-09-18Add dot1q-ip6base-[ndrpdr|mrr] perf tests for 2-node topologyJan Gelety12-4/+1090
2018-09-17Add dot1q-l2[bd|xc]base-[ndrpdr|mrr] perf tests for 2-node topologyJan Gelety16-128/+1231
2018-09-14Add dot1q-ip4base-[ndrpdr|mrr] perf tests for 2-node topologyJan Gelety14-13/+1120
2018-09-10Create l2bdscale tests for 2-node topologyJan Gelety38-397/+1983
2018-09-07CSIT-1272: Fix VTS testsTibor Frank6-6/+6
2018-09-07Create ip6scale tests for 2-node topologyJan Gelety24-162/+1734
2018-09-06Fix autogen violationsVratko Polak4-76/+124
2018-09-05Create ip4scale tests for 2-node topologyJan Gelety24-138/+1722
2018-09-05CSIT-1205 Create AVF driver testPeter Mikus4-0/+482
2018-09-05Add eth-ip4base-2memif[ndrpdr|mrr] perf tests for 2-node topologyJan Gelety10-0/+1468
2018-08-28Create ip4-feature perf tests for 2-node topologyJan Gelety144-846/+4337
2018-08-27FIX: Correction of ACL perf test suite documentationJan Gelety72-905/+840
2018-08-24VPP link bonding perf tests with two links between DUTsJan Gelety12-4/+687
2018-08-22CSIT-1164 Create VM vhost tests for 2-node topologyPeter Mikus36-0/+4946
2018-08-21FIX: Use correct interface type in x710_1lbvpplacp perf testsJan Gelety4-4/+4
2018-08-20Refactor VHOST codePeter Mikus99-1368/+622
2018-08-17CSIT-1245 - Wrong configuration of srv6proxy-ndrpdr-stat and srv6proxy-masq-n...Peter Mikus2-2/+2
2018-08-15FIX: Set default values for VTS teardown argumentsVratko Polak6-6/+18
2018-08-10CSIT-1156 Create container memif tests for 2-node topologyPeter Mikus28-34/+1210
2018-08-10CSIT-1193 De-duplicate bootstrap scripts into onePeter Mikus1-0/+9
2018-08-06CSIT_VIRL phased migration, prio P0 - apply SKIP_VPP_PATCH tagJan Gelety9-14/+16
2018-08-03Fix ommissions from recent ChangesVratko Polak2-0/+153
2018-07-27FIX: Interface up handlingPeter Mikus8-16/+0
2018-07-25Add MTU handling into perf testsPeter Mikus45-45/+11
2018-07-24CSIT-1097: Migrate L2 to NDRPDR and edit MRRVratko Polak168-7320/+13484
2018-07-24CSIT-1202: Unify min_rateVratko Polak3-3/+3
2018-07-24CSIT-1097: Unify max NIC bandwidth in ip4Peter Mikus4-15/+12
2018-07-24CSIT-1097: Migrate Vts suite to NDRPDRVratko Polak11-287/+1020
2018-07-24CSIT-1097: Migrate ip4 to NDRPDR and edit MRRVratko Polak152-6875/+12460
2018-07-24CSIT-1097: Migrate vm_vhost to NDRPDR and edit MRRVratko Polak126-11852/+10058
2018-07-20CSIT-1097: Migrate ip4_tun to NDRPDR and edit MRRVratko Polak19-2380/+1358
2018-07-20CSIT-1097: Migrate memif to NDRPDR and edit MRRVratko Polak26-1325/+1849
2018-07-18CSIT-1097: Crypto NDRPDR and missing MRR suitesVratko Polak34-3613/+2873
2018-07-18CSIT-1097: Ip6 NDRPDR and edited MRR suitesVratko Polak38-3775/+2413
2018-07-18CSIT-1201 Create L2patch perf testsPeter Mikus10-0/+1298
2018-07-17CSIT-1186: Add MRR and NDRPDR ip6_tunnels testsVratko Polak7-1192/+566
2018-07-17CSIT-1186: Add srv6 NDRPDR suitesVratko Polak13-620/+616
2018-07-16Add L2xc/L2bd/Ip4/Ip6 2-node tests for x710Peter Mikus16-0/+2107
"p">) { *port = (key >> 16) & (u16) ~ 0; } if (fib_index) { *fib_index = key >> 3 & ((1 << 13) - 1); } if (proto) { *proto = key & 0x7; } } always_inline void init_nat_k (clib_bihash_kv_8_8_t * kv, ip4_address_t addr, u16 port, u32 fib_index, nat_protocol_t proto) { kv->key = calc_nat_key (addr, port, fib_index, proto); kv->value = ~0ULL; } always_inline void init_nat_kv (clib_bihash_kv_8_8_t * kv, ip4_address_t addr, u16 port, u32 fib_index, nat_protocol_t proto, u64 value) { init_nat_k (kv, addr, port, fib_index, proto); kv->value = value; } always_inline void init_nat_i2o_k (clib_bihash_kv_8_8_t * kv, snat_session_t * s) { return init_nat_k (kv, s->in2out.addr, s->in2out.port, s->in2out.fib_index, s->nat_proto); } always_inline void init_nat_i2o_kv (clib_bihash_kv_8_8_t * kv, snat_session_t * s, u64 value) { init_nat_k (kv, s->in2out.addr, s->in2out.port, s->in2out.fib_index, s->nat_proto); kv->value = value; } always_inline void init_nat_o2i_k (clib_bihash_kv_8_8_t * kv, snat_session_t * s) { return init_nat_k (kv, s->out2in.addr, s->out2in.port, s->out2in.fib_index, s->nat_proto); } always_inline void init_nat_o2i_kv (clib_bihash_kv_8_8_t * kv, snat_session_t * s, u64 value) { init_nat_k (kv, s->out2in.addr, s->out2in.port, s->out2in.fib_index, s->nat_proto); kv->value = value; } static inline uword nat_pre_node_fn_inline (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame, u32 def_next) { u32 n_left_from, *from; from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs; u16 nexts[VLIB_FRAME_SIZE], *next = nexts; vlib_get_buffers (vm, from, b, n_left_from); while (n_left_from >= 2) { u32 next0, next1; u32 arc_next0, arc_next1; vlib_buffer_t *b0, *b1; b0 = *b; b++; b1 = *b; b++; /* Prefetch next iteration. */ if (PREDICT_TRUE (n_left_from >= 4)) { vlib_buffer_t *p2, *p3; p2 = *b; p3 = *(b + 1); vlib_prefetch_buffer_header (p2, LOAD); vlib_prefetch_buffer_header (p3, LOAD); CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, LOAD); CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, LOAD); } next0 = def_next; next1 = def_next; vnet_feature_next (&arc_next0, b0); vnet_feature_next (&arc_next1, b1); vnet_buffer2 (b0)->nat.arc_next = arc_next0; vnet_buffer2 (b1)->nat.arc_next = arc_next1; if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) { if (b0->flags & VLIB_BUFFER_IS_TRACED) { nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->next_index = next0; t->arc_next_index = arc_next0; } if (b1->flags & VLIB_BUFFER_IS_TRACED) { nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->next_index = next1; t->arc_next_index = arc_next1; } } n_left_from -= 2; next[0] = next0; next[1] = next1; next += 2; } while (n_left_from > 0) { u32 next0; u32 arc_next0; vlib_buffer_t *b0; b0 = *b; b++; next0 = def_next; vnet_feature_next (&arc_next0, b0); vnet_buffer2 (b0)->nat.arc_next = arc_next0; if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) && (b0->flags & VLIB_BUFFER_IS_TRACED))) { nat_pre_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->next_index = next0; t->arc_next_index = arc_next0; } n_left_from--; next[0] = next0; next++; } vlib_buffer_enqueue_to_next (vm, node, from, (u16 *) nexts, frame->n_vectors); return frame->n_vectors; } always_inline u8 is_interface_addr (snat_main_t * sm, vlib_node_runtime_t * node, u32 sw_if_index0, u32 ip4_addr) { snat_runtime_t *rt = (snat_runtime_t *) node->runtime_data; ip4_address_t *first_int_addr; if (PREDICT_FALSE (rt->cached_sw_if_index != sw_if_index0)) { first_int_addr = ip4_interface_first_address (sm->ip4_main, sw_if_index0, 0 /* just want the address */ ); rt->cached_sw_if_index = sw_if_index0; if (first_int_addr) rt->cached_ip4_address = first_int_addr->as_u32; else rt->cached_ip4_address = 0; } if (PREDICT_FALSE (ip4_addr == rt->cached_ip4_address)) return 1; else return 0; } always_inline void user_session_increment (snat_main_t * sm, snat_user_t * u, u8 is_static) { if (u->nsessions + u->nstaticsessions < sm->max_translations_per_user) { if (is_static) u->nstaticsessions++; else u->nsessions++; } } always_inline void nat44_delete_user_with_no_session (snat_main_t * sm, snat_user_t * u, u32 thread_index) { clib_bihash_kv_8_8_t kv; snat_user_key_t u_key; snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data, thread_index); if (u->nstaticsessions == 0 && u->nsessions == 0) { u_key.addr.as_u32 = u->addr.as_u32; u_key.fib_index = u->fib_index; kv.key = u_key.as_u64; pool_put_index (tsm->list_pool, u->sessions_per_user_list_head_index); pool_put (tsm->users, u); clib_bihash_add_del_8_8 (&tsm->user_hash, &kv, 0); vlib_set_simple_counter (&sm->total_users, thread_index, 0, pool_elts (tsm->users)); } } always_inline void nat44_delete_session (snat_main_t * sm, snat_session_t * ses, u32 thread_index) { snat_main_per_thread_data_t *tsm = vec_elt_at_index (sm->per_thread_data, thread_index); clib_bihash_kv_8_8_t kv, value; snat_user_t *u; const snat_user_key_t u_key = { .addr = ses->in2out.addr, .fib_index = ses->in2out.fib_index }; const u8 u_static = snat_is_session_static (ses); clib_dlist_remove (tsm->list_pool, ses->per_user_index); pool_put_index (tsm->list_pool, ses->per_user_index); if (sm->endpoint_dependent) { clib_dlist_remove (tsm->lru_pool, ses->lru_index); pool_put_index (tsm->lru_pool, ses->lru_index); } pool_put (tsm->sessions, ses); vlib_set_simple_counter (&sm->total_sessions, thread_index, 0, pool_elts (tsm->sessions)); kv.key = u_key.as_u64; if (!clib_bihash_search_8_8 (&tsm->user_hash, &kv, &value)) { u = pool_elt_at_index (tsm->users, value.value); if (u_static) u->nstaticsessions--; else u->nsessions--; nat44_delete_user_with_no_session (sm, u, thread_index); } } always_inline void nat44_set_tcp_session_state_i2o (snat_main_t * sm, f64 now, snat_session_t * ses, vlib_buffer_t * b, u32 thread_index) { snat_main_per_thread_data_t *tsm = &sm->per_thread_data[thread_index]; u8 tcp_flags = vnet_buffer (b)->ip.reass.icmp_type_or_tcp_flags; u32 tcp_ack_number = vnet_buffer (b)->ip.reass.tcp_ack_number; u32 tcp_seq_number = vnet_buffer (b)->ip.reass.tcp_seq_number; if ((ses->state == 0) && (tcp_flags & TCP_FLAG_RST)) ses->state = NAT44_SES_RST; if ((ses->state == NAT44_SES_RST) && !(tcp_flags & TCP_FLAG_RST)) ses->state = 0; if ((tcp_flags & TCP_FLAG_ACK) && (ses->state & NAT44_SES_I2O_SYN) && (ses->state & NAT44_SES_O2I_SYN)) ses->state = 0; if (tcp_flags & TCP_FLAG_SYN) ses->state |= NAT44_SES_I2O_SYN; if (tcp_flags & TCP_FLAG_FIN) { ses->i2o_fin_seq = clib_net_to_host_u32 (tcp_seq_number); ses->state |= NAT44_SES_I2O_FIN; } if ((tcp_flags & TCP_FLAG_ACK) && (ses->state & NAT44_SES_O2I_FIN)) { if (clib_net_to_host_u32 (tcp_ack_number) > ses->o2i_fin_seq) { ses->state |= NAT44_SES_O2I_FIN_ACK; if (nat44_is_ses_closed (ses)) { // if session is now closed, save the timestamp ses->tcp_closed_timestamp = now + sm->timeouts.tcp.transitory; ses->last_lru_update = now; } } } // move the session to proper LRU if (ses->state) { ses->lru_head_index = tsm->tcp_trans_lru_head_index; } else { ses->lru_head_index = tsm->tcp_estab_lru_head_index; } clib_dlist_remove (tsm->lru_pool, ses->lru_index); clib_dlist_addtail (tsm->lru_pool, ses->lru_head_index, ses->lru_index); } always_inline void nat44_set_tcp_session_state_o2i (snat_main_t * sm, f64 now, snat_session_t * ses, u8 tcp_flags, u32 tcp_ack_number, u32 tcp_seq_number, u32 thread_index) { snat_main_per_thread_data_t *tsm = &sm->per_thread_data[thread_index]; if ((ses->state == 0) && (tcp_flags & TCP_FLAG_RST)) ses->state = NAT44_SES_RST; if ((ses->state == NAT44_SES_RST) && !(tcp_flags & TCP_FLAG_RST)) ses->state = 0; if ((tcp_flags & TCP_FLAG_ACK) && (ses->state & NAT44_SES_I2O_SYN) && (ses->state & NAT44_SES_O2I_SYN)) ses->state = 0; if (tcp_flags & TCP_FLAG_SYN) ses->state |= NAT44_SES_O2I_SYN; if (tcp_flags & TCP_FLAG_FIN) { ses->o2i_fin_seq = clib_net_to_host_u32 (tcp_seq_number); ses->state |= NAT44_SES_O2I_FIN; } if ((tcp_flags & TCP_FLAG_ACK) && (ses->state & NAT44_SES_I2O_FIN)) { if (clib_net_to_host_u32 (tcp_ack_number) > ses->i2o_fin_seq) ses->state |= NAT44_SES_I2O_FIN_ACK; if (nat44_is_ses_closed (ses)) { // if session is now closed, save the timestamp ses->tcp_closed_timestamp = now + sm->timeouts.tcp.transitory; ses->last_lru_update = now; } } // move the session to proper LRU if (ses->state) { ses->lru_head_index = tsm->tcp_trans_lru_head_index; } else { ses->lru_head_index = tsm->tcp_estab_lru_head_index; } clib_dlist_remove (tsm->lru_pool, ses->lru_index); clib_dlist_addtail (tsm->lru_pool, ses->lru_head_index, ses->lru_index); } always_inline u32 nat44_session_get_timeout (snat_main_t * sm, snat_session_t * s) { switch (s->nat_proto) { case NAT_PROTOCOL_ICMP: return sm->timeouts.icmp; case NAT_PROTOCOL_UDP: return sm->timeouts.udp; case NAT_PROTOCOL_TCP: { if (s->state) return sm->timeouts.tcp.transitory; else return sm->timeouts.tcp.established; } default: return sm->timeouts.udp; } return 0; } always_inline void nat44_session_update_counters (snat_session_t * s, f64 now, uword bytes, u32 thread_index) { s->last_heard = now; s->total_pkts++; s->total_bytes += bytes; nat_ha_sref (&s->out2in.addr, s->out2in.port, &s->ext_host_addr, s->ext_host_port, s->nat_proto, s->out2in.fib_index, s->total_pkts, s->total_bytes, thread_index, &s->ha_last_refreshed, now); } /** \brief Per-user LRU list maintenance */ always_inline void nat44_session_update_lru (snat_main_t * sm, snat_session_t * s, u32 thread_index) { /* don't update too often - timeout is in magnitude of seconds anyway */ if (s->last_heard > s->last_lru_update + 1) { if (!sm->endpoint_dependent) { clib_dlist_remove (sm->per_thread_data[thread_index].list_pool, s->per_user_index); clib_dlist_addtail (sm->per_thread_data[thread_index].list_pool, s->per_user_list_head_index, s->per_user_index); } else { clib_dlist_remove (sm->per_thread_data[thread_index].lru_pool, s->lru_index); clib_dlist_addtail (sm->per_thread_data[thread_index].lru_pool, s->lru_head_index, s->lru_index); } s->last_lru_update = s->last_heard; } } always_inline void init_ed_k (clib_bihash_kv_16_8_t * kv, ip4_address_t l_addr, u16 l_port, ip4_address_t r_addr, u16 r_port, u32 fib_index, u8 proto) { kv->key[0] = (u64) r_addr.as_u32 << 32 | l_addr.as_u32; kv->key[1] = (u64) r_port << 48 | (u64) l_port << 32 | fib_index << 8 | proto; } always_inline void init_ed_kv (clib_bihash_kv_16_8_t * kv, ip4_address_t l_addr, u16 l_port, ip4_address_t r_addr, u16 r_port, u32 fib_index, u8 proto, u32 thread_index, u32 session_index) { init_ed_k (kv, l_addr, l_port, r_addr, r_port, fib_index, proto); kv->value = (u64) thread_index << 32 | session_index; } always_inline u32 ed_value_get_thread_index (clib_bihash_kv_16_8_t * value) { return value->value >> 32; } always_inline u32 ed_value_get_session_index (clib_bihash_kv_16_8_t * value) { return value->value & ~(u32) 0; } always_inline void split_ed_kv (clib_bihash_kv_16_8_t * kv, ip4_address_t * l_addr, ip4_address_t * r_addr, u8 * proto, u32 * fib_index, u16 * l_port, u16 * r_port) { if (l_addr) { l_addr->as_u32 = kv->key[0] & (u32) ~ 0; } if (r_addr) { r_addr->as_u32 = kv->key[0] >> 32; } if (r_port) { *r_port = kv->key[1] >> 48; } if (l_port) { *l_port = (kv->key[1] >> 32) & (u16) ~ 0; } if (fib_index) { *fib_index = (kv->key[1] >> 8) & ((1 << 24) - 1); } if (proto) { *proto = kv->key[1] & (u8) ~ 0; } } static_always_inline int get_icmp_i2o_ed_key (vlib_buffer_t * b, ip4_header_t * ip0, u32 rx_fib_index, u32 thread_index, u32 session_index, nat_protocol_t * nat_proto, u16 * l_port, u16 * r_port, clib_bihash_kv_16_8_t * kv) { u8 proto; u16 _l_port, _r_port; ip4_address_t *l_addr, *r_addr; icmp46_header_t *icmp0; icmp_echo_header_t *echo0, *inner_echo0 = 0; ip4_header_t *inner_ip0 = 0; void *l4_header = 0; icmp46_header_t *inner_icmp0; icmp0 = (icmp46_header_t *) ip4_next_header (ip0); echo0 = (icmp_echo_header_t *) (icmp0 + 1); if (!icmp_type_is_error_message (vnet_buffer (b)->ip.reass.icmp_type_or_tcp_flags)) { proto = IP_PROTOCOL_ICMP; l_addr = &ip0->src_address; r_addr = &ip0->dst_address; _l_port = vnet_buffer (b)->ip.reass.l4_src_port; _r_port = 0; } else { inner_ip0 = (ip4_header_t *) (echo0 + 1); l4_header = ip4_next_header (inner_ip0); proto = inner_ip0->protocol; r_addr = &inner_ip0->src_address; l_addr = &inner_ip0->dst_address; switch (ip_proto_to_nat_proto (inner_ip0->protocol)) { case NAT_PROTOCOL_ICMP: inner_icmp0 = (icmp46_header_t *) l4_header; inner_echo0 = (icmp_echo_header_t *) (inner_icmp0 + 1); _r_port = 0; _l_port = inner_echo0->identifier; break; case NAT_PROTOCOL_UDP: case NAT_PROTOCOL_TCP: _l_port = ((tcp_udp_header_t *) l4_header)->dst_port; _r_port = ((tcp_udp_header_t *) l4_header)->src_port; break; default: return NAT_IN2OUT_ED_ERROR_UNSUPPORTED_PROTOCOL; } } init_ed_kv (kv, *l_addr, _l_port, *r_addr, _r_port, rx_fib_index, proto, thread_index, session_index); if (nat_proto) { *nat_proto = ip_proto_to_nat_proto (proto); } if (l_port) { *l_port = _l_port; } if (r_port) { *r_port = _r_port; } return 0; } static_always_inline int get_icmp_o2i_ed_key (vlib_buffer_t * b, ip4_header_t * ip0, u32 rx_fib_index, u32 thread_index, u32 session_index, nat_protocol_t * nat_proto, u16 * l_port, u16 * r_port, clib_bihash_kv_16_8_t * kv) { icmp46_header_t *icmp0; u8 proto; ip4_address_t *l_addr, *r_addr; u16 _l_port, _r_port; icmp_echo_header_t *echo0, *inner_echo0 = 0; ip4_header_t *inner_ip0; void *l4_header = 0; icmp46_header_t *inner_icmp0; icmp0 = (icmp46_header_t *) ip4_next_header (ip0); echo0 = (icmp_echo_header_t *) (icmp0 + 1); if (!icmp_type_is_error_message (vnet_buffer (b)->ip.reass.icmp_type_or_tcp_flags)) { proto = IP_PROTOCOL_ICMP; l_addr = &ip0->dst_address; r_addr = &ip0->src_address; _l_port = vnet_buffer (b)->ip.reass.l4_src_port; _r_port = 0; } else { inner_ip0 = (ip4_header_t *) (echo0 + 1); l4_header = ip4_next_header (inner_ip0); proto = inner_ip0->protocol; l_addr = &inner_ip0->src_address; r_addr = &inner_ip0->dst_address; switch (ip_proto_to_nat_proto (inner_ip0->protocol)) { case NAT_PROTOCOL_ICMP: inner_icmp0 = (icmp46_header_t *) l4_header; inner_echo0 = (icmp_echo_header_t *) (inner_icmp0 + 1); _l_port = inner_echo0->identifier; _r_port = 0; break; case NAT_PROTOCOL_UDP: case NAT_PROTOCOL_TCP: _l_port = ((tcp_udp_header_t *) l4_header)->src_port; _r_port = ((tcp_udp_header_t *) l4_header)->dst_port; break; default: return -1; } } init_ed_kv (kv, *l_addr, _l_port, *r_addr, _r_port, rx_fib_index, proto, thread_index, session_index); if (nat_proto) { *nat_proto = ip_proto_to_nat_proto (proto); } if (l_port) { *l_port = _l_port; } if (r_port) { *r_port = _r_port; } return 0; } /** * @brief Check if packet should be translated * * Packets aimed at outside interface and external address with active session * should be translated. * * @param sm NAT main * @param rt NAT runtime data * @param sw_if_index0 index of the inside interface * @param ip0 IPv4 header * @param proto0 NAT protocol * @param rx_fib_index0 RX FIB index * * @returns 0 if packet should be translated otherwise 1 */ static inline int snat_not_translate_fast (snat_main_t * sm, vlib_node_runtime_t * node, u32 sw_if_index0, ip4_header_t * ip0, u32 proto0, u32 rx_fib_index0) { if (sm->out2in_dpo) return 0; fib_node_index_t fei = FIB_NODE_INDEX_INVALID; nat_outside_fib_t *outside_fib; fib_prefix_t pfx = { .fp_proto = FIB_PROTOCOL_IP4, .fp_len = 32, .fp_addr = { .ip4.as_u32 = ip0->dst_address.as_u32, } , }; /* Don't NAT packet aimed at the intfc address */ if (PREDICT_FALSE (is_interface_addr (sm, node, sw_if_index0, ip0->dst_address.as_u32))) return 1; fei = fib_table_lookup (rx_fib_index0, &pfx); if (FIB_NODE_INDEX_INVALID != fei) { u32 sw_if_index = fib_entry_get_resolving_interface (fei); if (sw_if_index == ~0) { vec_foreach (outside_fib, sm->outside_fibs) { fei = fib_table_lookup (outside_fib->fib_index, &pfx); if (FIB_NODE_INDEX_INVALID != fei) { sw_if_index = fib_entry_get_resolving_interface (fei); if (sw_if_index != ~0) break; } } } if (sw_if_index == ~0) return 1; snat_interface_t *i; /* *INDENT-OFF* */ pool_foreach (i, sm->interfaces, ({ /* NAT packet aimed at outside interface */ if ((nat_interface_is_outside (i)) && (sw_if_index == i->sw_if_index)) return 0; })); /* *INDENT-ON* */ } return 1; } static_always_inline u16 snat_random_port (u16 min, u16 max) { snat_main_t *sm = &snat_main; u32 rwide; u16 r; rwide = random_u32 (&sm->random_seed); r = rwide & 0xFFFF; if (r >= min && r <= max) return r; return min + (rwide % (max - min + 1)); } #endif /* __included_nat_inlines_h__ */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */