aboutsummaryrefslogtreecommitdiffstats
path: root/tests/vpp/perf/ip4/10ge2p1x710-ethip4-ip4scale600k-rnd-ndrpdr.robot
blob: 0cba4c1ffc3ae191b4831457f95b4b8e6dc4b0fe (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
/*
 * Copyright (c) 2015 Cisco and/or its affiliates.
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at:
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

#ifndef __IP_PUNT_DROP_H__
#define __IP_PUNT_DROP_H__

#include <vnet/ip/ip.h>
#include <vnet/policer/policer.h>
#include <vnet/policer/police_inlines.h>

/**
 * IP4 punt policer configuration
 *   we police the punt rate to prevent overloading the host
 */
typedef struct ip_punt_policer_t_
{
  u32 policer_index;
} ip_punt_policer_t;

typedef enum ip_punt_policer_next_t_
{
  IP_PUNT_POLICER_NEXT_DROP,
  IP_PUNT_POLICER_N_NEXT,
} ip_punt_policer_next_t;

typedef struct ip_punt_policer_trace_t_
{
  u32 policer_index;
  u32 next;
} ip_punt_policer_trace_t;

#define foreach_ip_punt_policer_error          \
_(DROP, "ip punt policer drop")

typedef enum
{
#define _(sym,str) IP_PUNT_POLICER_ERROR_##sym,
  foreach_ip_punt_policer_error
#undef _
    IP4_PUNT_POLICER_N_ERROR,
} ip_punt_policer_error_t;

extern u8 *format_ip_punt_policer_trace (u8 * s, va_list * args);

/**
 * IP punt policing node function
 */
always_inline uword
ip_punt_policer (vlib_main_t * vm,
		 vlib_node_runtime_t * node,
		 vlib_frame_t * frame, u8 arc_index, u32 policer_index)
{
  u32 *from, *to_next, n_left_from, n_left_to_next, next_index;
  u64 time_in_policer_periods;
  vnet_feature_main_t *fm = &feature_main;
  vnet_feature_config_main_t *cm = &fm->feature_config_mains[arc_index];

  time_in_policer_periods =
    clib_cpu_time_now () >> POLICER_TICKS_PER_PERIOD_SHIFT;

  from = vlib_frame_vector_args (frame);
  n_left_from = frame->n_vectors;
  next_index = node->cached_next_index;

  while (n_left_from > 0)
    {
      vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);

      while (n_left_from >= 4 && n_left_to_next >= 2)
	{
	  vlib_buffer_t *b0, *b1;
	  u32 next0, next1;
	  u8 act0, act1;
	  u32 bi0, bi1;

	  next0 = next1 = 0;
	  bi0 = to_next[0] = from[0];
	  bi1 = to_next[1] = from[1];

	  from += 2;
	  n_left_from -= 2;
	  to_next += 2;
	  n_left_to_next -= 2;

	  b0 = vlib_get_buffer (vm, bi0);
	  b1 = vlib_get_buffer (vm, bi1);

	  vnet_get_config_data (&cm->config_main,
				&b0->current_config_index, &next0, 0);
	  vnet_get_config_data (&cm->config_main,
				&b1->current_config_index, &next1, 0);

	  act0 = vnet_policer_police (vm, b0,
				      policer_index,
				      time_in_policer_periods,
				      POLICE_CONFORM);
	  act1 = vnet_policer_police (vm, b1,
				      policer_index,
				      time_in_policer_periods,
				      POLICE_CONFORM);

	  if (PREDICT_FALSE (act0 == SSE2_QOS_ACTION_DROP))
	    {
	      next0 = IP_PUNT_POLICER_NEXT_DROP;
	      b0->error = node->errors[IP_PUNT_POLICER_ERROR_DROP];
	    }
	  if (PREDICT_FALSE (act1 == SSE2_QOS_ACTION_DROP))
	    {
	      next1 = IP_PUNT_POLICER_NEXT_DROP;
	      b1->error = node->errors[IP_PUNT_POLICER_ERROR_DROP];
	    }

	  if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED))
	    {
	      ip_punt_policer_trace_t *t =
		vlib_add_trace (vm, node
{ ip_punt_policer_trace_t *t = vlib_add_trace (vm, node, b1, sizeof (*t)); t->next = next1; t->policer_index = policer_index; } vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next, n_left_to_next, bi0, bi1, next0, next1); } while (n_left_from > 0 && n_left_to_next > 0) { vlib_buffer_t *b0; u32 next0; u32 bi0; u8 act0; next0 = 0; bi0 = to_next[0] = from[0]; from += 1; n_left_from -= 1; to_next += 1; n_left_to_next -= 1; b0 = vlib_get_buffer (vm, bi0); vnet_get_config_data (&cm->config_main, &b0->current_config_index, &next0, 0); act0 = vnet_policer_police (vm, b0, policer_index, time_in_policer_periods, POLICE_CONFORM); if (PREDICT_FALSE (act0 == SSE2_QOS_ACTION_DROP)) { next0 = IP_PUNT_POLICER_NEXT_DROP; b0->error = node->errors[IP_PUNT_POLICER_ERROR_DROP]; } if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) { ip_punt_policer_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->next = next0; t->policer_index = policer_index; } vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, n_left_to_next, bi0, next0); } vlib_put_next_frame (vm, node, next_index, n_left_to_next); } return frame->n_vectors; } /** * IP4 punt redirect per-rx interface configuration * redirect punted traffic to another location. */ typedef struct ip_punt_redirect_rx_t_ { /** * The next-hop to send redirected packets to */ ip46_address_t nh; /** * the TX interface to send redirected packets */ u32 tx_sw_if_index; /** * redirect forwarding adjacency */ adj_index_t adj_index; } ip_punt_redirect_rx_t; /** * IP punt redirect configuration */ typedef struct ip_punt_redirect_t_ { /** * any RX interface redirect */ ip_punt_redirect_rx_t any_rx_sw_if_index; /** * per-RX interface configuration */ ip_punt_redirect_rx_t *redirect_by_rx_sw_if_index; } ip_punt_redirect_t; /** * IP punt redirect next nodes */ typedef enum ip_punt_redirect_next_t_ { IP_PUNT_REDIRECT_NEXT_DROP, IP_PUNT_REDIRECT_NEXT_TX, IP_PUNT_REDIRECT_NEXT_ARP, IP_PUNT_REDIRECT_N_NEXT, } ip_punt_redirect_next_t; /** * IP Punt redirect trace */ typedef struct ip4_punt_redirect_trace_t_ { ip_punt_redirect_rx_t redirect; u32 next; } ip_punt_redirect_trace_t; /** * Add a punt redirect entry */ extern void ip_punt_redirect_add (ip_punt_redirect_t * cfg, u32 rx_sw_if_index, ip_punt_redirect_rx_t * redirect, fib_protocol_t fproto, vnet_link_t linkt); extern void ip_punt_redirect_del (ip_punt_redirect_t * cfg, u32 rx_sw_if_index); extern u8 *format_ip_punt_redirect (u8 * s, va_list * args); extern u8 *format_ip_punt_redirect_trace (u8 * s, va_list * args); always_inline u32 ip_punt_redirect_tx_via_adj (vlib_buffer_t * b0, adj_index_t ai) { ip_adjacency_t *adj = adj_get (ai); u32 next0; vnet_buffer (b0)->ip.adj_index[VLIB_TX] = ai; switch (adj->lookup_next_index) { case IP_LOOKUP_NEXT_ARP: next0 = IP_PUNT_REDIRECT_NEXT_ARP; break; case IP_LOOKUP_NEXT_REWRITE: next0 = IP_PUNT_REDIRECT_NEXT_TX; break; default: next0 = IP_PUNT_REDIRECT_NEXT_DROP; break; } return (next0); } always_inline uword ip_punt_redirect (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame, u8 arc_index, ip_punt_redirect_t * redirect) { u32 *from, *to_next, n_left_from, n_left_to_next, next_index; vnet_feature_main_t *fm = &feature_main; vnet_feature_config_main_t *cm = &fm->feature_config_mains[arc_index]; from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; next_index = node->cached_next_index; while (n_left_from > 0) { vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); while (n_left_from > 0 && n_left_to_next > 0) { u32 rx_sw_if_index0; ip_punt_redirect_rx_t *rrx0; vlib_buffer_t *b0; u32 next0; u32 bi0; rrx0 = NULL; next0 = 0; bi0 = to_next[0] = from[0]; from += 1; n_left_from -= 1; to_next += 1; n_left_to_next -= 1; b0 = vlib_get_buffer (vm, bi0); vnet_get_config_data (&cm->config_main, &b0->current_config_index, &next0, 0); rx_sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX]; if (vec_len (redirect->redirect_by_rx_sw_if_index) > rx_sw_if_index0) { rrx0 = &redirect->redirect_by_rx_sw_if_index[rx_sw_if_index0]; if (~0 != rrx0->tx_sw_if_index) { next0 = ip_punt_redirect_tx_via_adj (b0, rrx0->adj_index); } else if (~0 != redirect->any_rx_sw_if_index.tx_sw_if_index) { rrx0 = &redirect->any_rx_sw_if_index; next0 = ip_punt_redirect_tx_via_adj (b0, rrx0->adj_index); } } else if (~0 != redirect->any_rx_sw_if_index.tx_sw_if_index) { rrx0 = &redirect->any_rx_sw_if_index; next0 = ip_punt_redirect_tx_via_adj (b0, rrx0->adj_index); } if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_IS_TRACED)) { ip_punt_redirect_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->next = next0; if (rrx0) t->redirect = *rrx0; } vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, n_left_to_next, bi0, next0); } vlib_put_next_frame (vm, node, next_index, n_left_to_next); } return frame->n_vectors; } always_inline uword ip_drop_or_punt (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame, u8 arc_index) { u32 *from, *to_next, n_left_from, n_left_to_next, next_index; from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; next_index = node->cached_next_index; while (n_left_from > 0) { vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); while (n_left_from >= 8 && n_left_to_next >= 4) { vlib_buffer_t *b0, *b1, *b2, *b3; u32 next0, next1, next2, next3; u32 bi0, bi1, bi2, bi3; next0 = next1 = next2 = next3 = 0; /* Prefetch next iteration. */ { vlib_buffer_t *p4, *p5, *p6, *p7; p4 = vlib_get_buffer (vm, from[4]); p5 = vlib_get_buffer (vm, from[5]); p6 = vlib_get_buffer (vm, from[6]); p7 = vlib_get_buffer (vm, from[7]); vlib_prefetch_buffer_header (p4, LOAD); vlib_prefetch_buffer_header (p5, LOAD); vlib_prefetch_buffer_header (p6, LOAD); vlib_prefetch_buffer_header (p7, LOAD); } bi0 = to_next[0] = from[0]; bi1 = to_next[1] = from[1]; bi2 = to_next[2] = from[2]; bi3 = to_next[3] = from[3]; from += 4; n_left_from -= 4; to_next += 4; n_left_to_next -= 4; b0 = vlib_get_buffer (vm, bi0); b1 = vlib_get_buffer (vm, bi1); b2 = vlib_get_buffer (vm, bi2); b3 = vlib_get_buffer (vm, bi3); /* punt and drop features are not associated with a given interface * so the special index 0 is used */ vnet_feature_arc_start (arc_index, 0, &next0, b0); vnet_feature_arc_start (arc_index, 0, &next1, b1); vnet_feature_arc_start (arc_index, 0, &next2, b2); vnet_feature_arc_start (arc_index, 0, &next3, b3); vlib_validate_buffer_enqueue_x4 (vm, node, next_index, to_next, n_left_to_next, bi0, bi1, bi2, bi3, next0, next1, next2, next3); } while (n_left_from > 0 && n_left_to_next > 0) { vlib_buffer_t *b0; u32 next0; u32 bi0; next0 = 0; bi0 = to_next[0] = from[0]; from += 1; n_left_from -= 1; to_next += 1; n_left_to_next -= 1; b0 = vlib_get_buffer (vm, bi0); vnet_feature_arc_start (arc_index, 0, &next0, b0); vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, n_left_to_next, bi0, next0); } vlib_put_next_frame (vm, node, next_index, n_left_to_next); } return frame->n_vectors; } #endif /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */