aboutsummaryrefslogtreecommitdiffstats
path: root/lib/src/protocol/tcp.c
blob: 2afc4f6f42bf304bacefcd4db491928ceb939353 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
/*
 * Copyright (c) 2017-2019 Cisco and/or its affiliates.
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at:
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

#include <string.h>
#include "tcp.h"

#include "../error.h"
#include "../ops.h"

#define TCP_DEFAULT_SRC_PORT           0x8000
#define TCP_DEFAULT_DST_PORT           0x0080
#define TCP_DEFAULT_WINDOW_SIZE        0	// In [2, 65535]
#define TCP_DEFAULT_HLEN               20
#define TCP_DEFAULT_DATA_OFFSET_RES    (TCP_DEFAULT_HLEN >> 2) << 4
#define TCP_DEFAULT_CWR                0
#define TCP_DEFAULT_ECE                0
#define TCP_DEFAULT_URG                0
#define TCP_DEFAULT_ACK                0
#define TCP_DEFAULT_PSH                0
#define TCP_DEFAULT_RST                0
#define TCP_DEFAULT_SYN                1
#define TCP_DEFAULT_FIN                0

DECLARE_get_interest_locator (tcp, UNEXPECTED);
DECLARE_set_interest_locator (tcp, UNEXPECTED);
DECLARE_get_interest_name (tcp, UNEXPECTED);
DECLARE_set_interest_name (tcp, UNEXPECTED);
DECLARE_get_data_locator (tcp, UNEXPECTED);
DECLARE_set_data_locator (tcp, UNEXPECTED);
DECLARE_get_data_name (tcp, UNEXPECTED);
DECLARE_set_data_name (tcp, UNEXPECTED);
DECLARE_get_length (tcp, UNEXPECTED);
DECLARE_get_payload_length (tcp, UNEXPECTED);
DECLARE_set_payload_length (tcp, UNEXPECTED);

int
tcp_init_packet_header (hicn_type_t type, hicn_protocol_t * h)
{
  h->tcp = (_tcp_header_t)
  {
  .sport = htons (TCP_DEFAULT_SRC_PORT),.dport =
      htons (TCP_DEFAULT_DST_PORT),.seq = 0,.seq_ack =
      0,.data_offset_and_reserved = TCP_DEFAULT_DATA_OFFSET_RES,.flags =
      TCP_DEFAULT_CWR << 7 | TCP_DEFAULT_ECE << 6 | TCP_DEFAULT_URG << 5 |
      TCP_DEFAULT_ACK << 4 | TCP_DEFAULT_PSH << 3 | TCP_DEFAULT_RST << 2 |
      TCP_DEFAULT_SYN << 1 | TCP_DEFAULT_FIN << 0,.window =
      htons (TCP_DEFAULT_WINDOW_SIZE),.csum = 0,.urg_ptr = 65000,};

  return CHILD_OPS (init_packet_header, type, h);
}

int
tcp_get_interest_name_suffix (hicn_type_t type, const hicn_protocol_t * h,
			      hicn_name_suffix_t * suffix)
{
  *suffix = ntohl (h->tcp.name_suffix);
  return HICN_LIB_ERROR_NONE;
}

int
tcp_set_interest_name_suffix (hicn_type_t type, hicn_protocol_t * h,
			      const hicn_name_suffix_t * suffix)
{
  h->tcp.name_suffix = htonl (*suffix);

  return HICN_LIB_ERROR_NONE;
}

int
tcp_reset_interest_for_hash (hicn_type_t type, hicn_protocol_t * h)
{
  memset (&(h->tcp), 0, 4);
  memset (&(h->tcp.seq_ack), 0, 12);

  return CHILD_OPS (reset_interest_for_hash, type, h);
}


int
tcp_get_data_name_suffix (hicn_type_t type, const hicn_protocol_t * h,
			  hicn_name_suffix_t * suffix)
{
  *suffix = ntohl (h->tcp.name_suffix);
  return HICN_LIB_ERROR_NONE;
}

int
tcp_set_data_name_suffix (hicn_type_t type, hicn_protocol_t * h,
			  const hicn_name_suffix_t * suffix)
{
  h->tcp.name_suffix = htonl (*suffix);
  return HICN_LIB_ERROR_NONE;
}

int
tcp_get_data_pathlabel (hicn_type_t type, const hicn_protocol_t * h,
			u32 * pathlabel)
{
  *pathlabel = h->tcp.seq_ack;
  return HICN_LIB_ERROR_NONE;
}

int
tcp_set_data_pathlabel (hicn_type_t type, hicn_protocol_t * h,
			const u32 pathlabel)
{
  h->tcp.seq_ack = pathlabel;
  return HICN_LIB_ERROR_NONE;
}

int
tcp_update_data_pathlabel (hicn_type_t type, hicn_protocol_t * h,
			   const hicn_faceid_t face_id)
{
  hicn_pathlabel_t pl =
    (hicn_pathlabel_t) ((h->tcp.pathlabel & HICN_PATH_LABEL_MASK) >> (32 -
								      HICN_PATH_LABEL_SIZE));
  hicn_pathlabel_t new_pl;

  update_pathlabel (pl, face_id, &new_pl);
  h->tcp.pathlabel = new_pl;

  return HICN_LIB_ERROR_NONE;
}

int
tcp_reset_data_for_hash (hicn_type_t type, hicn_protocol_t * h)
{
  memset (&(h->tcp), 0, 4);
  memset (&(h->tcp.seq_ack), 0, 12);

  return CHILD_OPS (reset_data_for_hash, type, h);
}


int
tcp_get_lifetime (hicn_type_t type, const hicn_protocol_t * h,
		  hicn_lifetime_t * lifetime)
{
  *lifetime =
    ntohs (h->tcp.urg_ptr) << (h->tcp.data_offset_and_reserved & 0xF);
  return HICN_LIB_ERROR_NONE;
}

int
tcp_set_lifetime (hicn_type_t type, hicn_protocol_t * h,
		  const hicn_lifetime_t lifetime)
{
  u8 multiplier = 0;
  u32 lifetime_scaled = lifetime;

  if (PREDICT_FALSE (lifetime >= HICN_MAX_LIFETIME))
    {
      h->tcp.urg_ptr = htons (HICN_MAX_LIFETIME_SCALED);
      h->tcp.data_offset_and_reserved =
	(h->
	 tcp.data_offset_and_reserved & ~0xF) | HICN_MAX_LIFETIME_MULTIPLIER;
      return HICN_LIB_ERROR_NONE;
    }

  while (lifetime_scaled > HICN_MAX_LIFETIME_SCALED
	 && multiplier <= HICN_MAX_LIFETIME_MULTIPLIER)
    {
      multiplier++;
      lifetime_scaled = lifetime_scaled >> 1;
    }

  h->tcp.urg_ptr = htons (lifetime_scaled);
  h->tcp.data_offset_and_reserved =
    (h->tcp.data_offset_and_reserved & ~0xF) | multiplier;

  return HICN_LIB_ERROR_NONE;
}

int
tcp_update_checksums (hicn_type_t type, hicn_protocol_t * h, u16 partial_csum,
		      size_t payload_length)
{
  h->tcp.csum = 0;

  if (PREDICT_TRUE (partial_csum != 0))
    {
      partial_csum = ~partial_csum;
    }

  h->tcp.csum = csum (h, TCP_HDRLEN + payload_length, partial_csum);

  return CHILD_OPS (update_checksums, type, h, 0, payload_length);
}

int
tcp_verify_checksums (hicn_type_t type, hicn_protocol_t * h, u16 partial_csum,
		      size_t payload_length)
{
  if (csum (h, TCP_HDRLEN + payload_length, ~partial_csum) != 0)
    return HICN_LIB_ERROR_CORRUPTED_PACKET;
  return CHILD_OPS (verify_checksums, type, h, 0, payload_length);
}

#define TCP_OFFSET_MASK                13
#define TCP_OFFSET_DATA_OFFSET         12
#define TCP_OFFSET_IN_BITS_DATA_OFFSET 0
#define TCP_OFFSET_IN_BITS_RESERVED    4
#define TCP_OFFSET_IN_BITS_NS          7

#define TCP_DEFAULT_SRC_PORT           0x8000
#define TCP_DEFAULT_DST_PORT           0x0080
#define TCP_DEFAULT_WINDOW_SIZE        0	// In [2, 65535]
#define TCP_DEFAULT_DATA_OFFSET        5	// Size of the TCP header in words (= 4 bytes). Must be greater or equal than 5.
#define TCP_DEFAULT_CWR                0
#define TCP_DEFAULT_ECE                0
#define TCP_DEFAULT_URG                0
#define TCP_DEFAULT_ACK                0
#define TCP_DEFAULT_PSH                0
#define TCP_DEFAULT_RST                0
#define TCP_DEFAULT_SYN                1
#define TCP_DEFAULT_FIN                0

int
tcp_rewrite_interest (hicn_type_t type, hicn_protocol_t * h,
		      const ip46_address_t * addr_new,
		      ip46_address_t * addr_old)
{
  u16 *tcp_checksum = &(h->tcp.csum);

  /*
   * Padding fields are set to zero so we can apply checksum on the
   * whole struct by interpreting it as IPv6 in all cases
   *
   * v4 code would be:
   * csum = ip_csum_sub_even (*tcp_checksum, h->ipv4.saddr.as_u32);
   * csum = ip_csum_add_even (csum, h->ipv4.saddr.as_u32);
   */
  u16 csum = ip_csum_sub_even (*tcp_checksum, h->ipv6.saddr.as_u64[0]);
  csum = ip_csum_sub_even (csum, h->ipv6.saddr.as_u64[1]);
  csum = ip_csum_add_even (csum, h->ipv6.saddr.as_u64[0]);
  csum = ip_csum_add_even (csum, h->ipv6.saddr.as_u64[1]);

  *tcp_checksum = ip_csum_fold (csum);

  return HICN_LIB_ERROR_NONE;
}

int
tcp_rewrite_data (hicn_type_t type, hicn_protocol_t * h,
		  const ip46_address_t * addr_new, ip46_address_t * addr_old,
		  const hicn_faceid_t face_id)
{
  u16 *tcp_checksum = &(h->tcp.csum);

  /*
   * Padding fields are set to zero so we can apply checksum on the
   * whole struct by interpreting it as IPv6 in all cases
   *
   * v4 code would be:
   * csum = ip_csum_sub_even (*tcp_checksum, h->ipv4.saddr.as_u32);
   * csum = ip_csum_add_even (csum, h->ipv4.saddr.as_u32);
   */
  u16 csum = ip_csum_sub_even (*tcp_checksum, addr_old->ip6.as_u64[0]);
  csum = ip_csum_sub_even (*tcp_checksum, addr_old->ip6.as_u64[1]);
  csum = ip_csum_add_even (csum, addr_new->ip6.as_u64[0]);
  csum = ip_csum_add_even (csum, addr_new->ip6.as_u64[1]);

  csum = ip_csum_sub_even (csum, h->tcp.pathlabel);
  tcp_update_data_pathlabel (type, h, face_id);
  csum = ip_csum_add_even (csum, h->tcp.pathlabel);

  *tcp_checksum = ip_csum_fold (csum);

  return HICN_LIB_ERROR_NONE;
}

int
tcp_get_current_header_length (hicn_type_t type, const hicn_protocol_t * h,
			       size_t * header_length)
{
  *header_length = TCP_HDRLEN;
  return HICN_LIB_ERROR_NONE;
}

int
tcp_get_header_length (hicn_type_t type, const hicn_protocol_t * h,
		       size_t * header_length)
{
  size_t child_header_length = 0;
  int rc = CHILD_OPS (get_header_length, type, h, &child_header_length);
  if (rc < 0)
    return rc;

  *header_length = TCP_HDRLEN + child_header_length;
  return HICN_LIB_ERROR_NONE;
}

int
tcp_get_signature_size (hicn_type_t type, const hicn_protocol_t * h,
			size_t * signature_size)
{
  return CHILD_OPS (get_signature_size, type, h, signature_size);
}

int
tcp_set_signature_size (hicn_type_t type, hicn_protocol_t * h,
			size_t signature_size)
{
  return CHILD_OPS (set_signature_size, type, h, signature_size);
}

int
tcp_set_signature_timestamp(hicn_type_t type, hicn_protocol_t * h,
       uint64_t signature_timestamp)
{
  return CHILD_OPS (set_signature_timestamp, type, h, signature_timestamp);
}

int
tcp_get_signature_timestamp (hicn_type_t type, const hicn_protocol_t * h,
       uint64_t * signature_timestamp)
{
  return CHILD_OPS (get_signature_timestamp, type, h, signature_timestamp);
}

int
tcp_set_validation_algorithm (hicn_type_t type, hicn_protocol_t * h,
       uint8_t validation_algorithm)
{
  return CHILD_OPS (set_validation_algorithm, type, h, validation_algorithm);
}

int
tcp_get_validation_algorithm (hicn_type_t type, const hicn_protocol_t * h,
       uint8_t * validation_algorithm)
{
  return CHILD_OPS (get_validation_algorithm, type, h, validation_algorithm);
}

int
tcp_set_key_id (hicn_type_t type, hicn_protocol_t * h,
       uint8_t *key_id)
{
  return CHILD_OPS (set_key_id, type, h, key_id);
}

int
tcp_get_key_id (hicn_type_t type, hicn_protocol_t * h,
       uint8_t **key_id, uint8_t *key_id_size)
{
  return CHILD_OPS (get_key_id, type, h, key_id, key_id_size);
}

DECLARE_HICN_OPS (tcp);

/*
 * fd.io coding-style-patch-verification: ON
 *
 * Local Variables:
 * eval: (c-set-style "gnu")
 * End:
 */
>my_address) { e2e_option->id.as_u64[0] = my_address->as_u64[0]; e2e_option->id.as_u64[1] = my_address->as_u64[1]; } } /* Following functions are for the caching of ioam header * to enable reattaching it for a complete request-response * message exchange */ inline static void ioam_cache_entry_free (ioam_cache_entry_t * entry) { ioam_cache_main_t *cm = &ioam_cache_main; if (entry) { vec_free (entry->ioam_rewrite_string); memset (entry, 0, sizeof (*entry)); pool_put (cm->ioam_rewrite_pool, entry); } } inline static ioam_cache_entry_t * ioam_cache_entry_cleanup (u32 pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_rewrite_pool, pool_index); ioam_cache_entry_free (entry); return (0); } inline static ioam_cache_entry_t * ioam_cache_lookup (ip6_header_t * ip0, u16 src_port, u16 dst_port, u32 seq_no) { ioam_cache_main_t *cm = &ioam_cache_main; u32 flow_hash = ip6_compute_flow_hash_ext (ip0, ip0->protocol, src_port, dst_port, IP_FLOW_HASH_DEFAULT | IP_FLOW_HASH_REVERSE_SRC_DST); clib_bihash_kv_8_8_t kv, value; kv.key = (u64) flow_hash << 32 | seq_no; kv.value = 0; value.key = 0; value.value = 0; if (clib_bihash_search_8_8 (&cm->ioam_rewrite_cache_table, &kv, &value) >= 0) { ioam_cache_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_rewrite_pool, value.value); /* match */ if (ip6_address_compare (&ip0->src_address, &entry->dst_address) == 0 && ip6_address_compare (&ip0->dst_address, &entry->src_address) == 0 && entry->src_port == dst_port && entry->dst_port == src_port && entry->seq_no == seq_no) { /* If lookup is successful remove it from the hash */ clib_bihash_add_del_8_8 (&cm->ioam_rewrite_cache_table, &kv, 0); return (entry); } else return (0); } return (0); } /* * Caches ioam hbh header * Extends the hbh header with option to contain IP6 address of the node * that caches it */ inline static int ioam_cache_add (vlib_buffer_t * b0, ip6_header_t * ip0, u16 src_port, u16 dst_port, ip6_hop_by_hop_header_t * hbh0, u32 seq_no) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_entry_t *entry = 0; u32 rewrite_len = 0, e2e_id_offset = 0; u32 pool_index = 0; ioam_e2e_id_option_t *e2e = 0; pool_get_aligned (cm->ioam_rewrite_pool, entry, CLIB_CACHE_LINE_BYTES); memset (entry, 0, sizeof (*entry)); pool_index = entry - cm->ioam_rewrite_pool; clib_memcpy (entry->dst_address.as_u64, ip0->dst_address.as_u64, sizeof (ip6_address_t)); clib_memcpy (entry->src_address.as_u64, ip0->src_address.as_u64, sizeof (ip6_address_t)); entry->src_port = src_port; entry->dst_port = dst_port; entry->seq_no = seq_no; rewrite_len = ((hbh0->length + 1) << 3); vec_validate (entry->ioam_rewrite_string, rewrite_len - 1); e2e = ip6_ioam_find_hbh_option (hbh0, HBH_OPTION_TYPE_IOAM_EDGE_TO_EDGE_ID); if (e2e) { entry->next_hop.as_u64[0] = e2e->id.as_u64[0]; entry->next_hop.as_u64[1] = e2e->id.as_u64[1]; } else { return (-1); } e2e_id_offset = (u8 *) e2e - (u8 *) hbh0; /* setup e2e id option to insert v6 address of the node caching it */ clib_memcpy (entry->ioam_rewrite_string, hbh0, rewrite_len); hbh0 = (ip6_hop_by_hop_header_t *) entry->ioam_rewrite_string; /* suffix rewrite string with e2e ID option */ e2e = (ioam_e2e_id_option_t *) (entry->ioam_rewrite_string + e2e_id_offset); ioam_e2e_id_rewrite_handler (e2e, b0); entry->my_address_offset = (u8 *) (&e2e->id) - (u8 *) hbh0; /* add it to hash, replacing and freeing any collision for now */ u32 flow_hash = ip6_compute_flow_hash_ext (ip0, hbh0->protocol, src_port, dst_port, IP_FLOW_HASH_DEFAULT); clib_bihash_kv_8_8_t kv, value; kv.key = (u64) flow_hash << 32 | seq_no; kv.value = 0; if (clib_bihash_search_8_8 (&cm->ioam_rewrite_cache_table, &kv, &value) >= 0) { /* replace */ ioam_cache_entry_cleanup (value.value); } kv.value = pool_index; clib_bihash_add_del_8_8 (&cm->ioam_rewrite_cache_table, &kv, 1); return (0); } /* Creates SR rewrite string * This is appended with ioam header on the server facing * node. * This SR header is necessary to attract packets towards * selected Anycast server. */ inline static void ioam_cache_sr_rewrite_template_create (void) { ioam_cache_main_t *cm = &ioam_cache_main; ip6_address_t *segments = 0; ip6_address_t *this_seg = 0; /* This nodes address and the original dest will be * filled when the packet is processed */ vec_add2 (segments, this_seg, 1); memset (this_seg, 0xfe, sizeof (ip6_address_t)); cm->sr_rewrite_template = ip6_sr_compute_rewrite_string_insert (segments); vec_free (segments); } inline static int ioam_cache_table_init (vlib_main_t * vm) { ioam_cache_main_t *cm = &ioam_cache_main; pool_alloc_aligned (cm->ioam_rewrite_pool, MAX_CACHE_ENTRIES, CLIB_CACHE_LINE_BYTES); cm->lookup_table_nbuckets = IOAM_CACHE_TABLE_DEFAULT_HASH_NUM_BUCKETS; cm->lookup_table_nbuckets = 1 << max_log2 (cm->lookup_table_nbuckets); cm->lookup_table_size = IOAM_CACHE_TABLE_DEFAULT_HASH_MEMORY_SIZE; clib_bihash_init_8_8 (&cm->ioam_rewrite_cache_table, "ioam rewrite cache table", cm->lookup_table_nbuckets, cm->lookup_table_size); /* Create SR rewrite template */ ioam_cache_sr_rewrite_template_create (); return (1); } inline static int ioam_cache_table_destroy (vlib_main_t * vm) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_entry_t *entry = 0; /* free pool and hash table */ clib_bihash_free_8_8 (&cm->ioam_rewrite_cache_table); pool_foreach (entry, cm->ioam_rewrite_pool, ( { ioam_cache_entry_free (entry); })); pool_free (cm->ioam_rewrite_pool); cm->ioam_rewrite_pool = 0; vec_free (cm->sr_rewrite_template); cm->sr_rewrite_template = 0; return (0); } inline static u8 * format_ioam_cache_entry (u8 * s, va_list * args) { ioam_cache_entry_t *e = va_arg (*args, ioam_cache_entry_t *); ioam_cache_main_t *cm = &ioam_cache_main; int rewrite_len = vec_len (e->ioam_rewrite_string); s = format (s, "%d: %U:%d to %U:%d seq_no %lu\n", (e - cm->ioam_rewrite_pool), format_ip6_address, &e->src_address, e->src_port, format_ip6_address, &e->dst_address, e->dst_port, e->seq_no); if (rewrite_len) { s = format (s, " %U", format_ip6_hop_by_hop_ext_hdr, (ip6_hop_by_hop_header_t *) e->ioam_rewrite_string, rewrite_len - 1); } return s; } void ioam_cache_ts_timer_node_enable (vlib_main_t * vm, u8 enable); #define IOAM_CACHE_TS_TIMEOUT 1.0 //SYN timeout 1 sec #define IOAM_CACHE_TS_TICK 100e-3 /* Timer delays as multiples of 100ms */ #define IOAM_CACHE_TS_TIMEOUT_TICKS IOAM_CACHE_TS_TICK*9 #define TIMER_HANDLE_INVALID ((u32) ~0) void expired_cache_ts_timer_callback (u32 * expired_timers); /* * Following functions are to manage M-Anycast server selection * cache * There is a per worker thread pool to create a cache entry * for a TCP SYN received. TCP SYN-ACK contians ioam header * with HBH_OPTION_TYPE_IOAM_E2E_CACHE_ID option to point to the * entry. */ inline static int ioam_cache_ts_table_init (vlib_main_t * vm) { ioam_cache_main_t *cm = &ioam_cache_main; int no_of_threads = vec_len (vlib_worker_threads); int i; vec_validate_aligned (cm->ioam_ts_pool, no_of_threads - 1, CLIB_CACHE_LINE_BYTES); vec_validate_aligned (cm->ts_stats, no_of_threads - 1, CLIB_CACHE_LINE_BYTES); vec_validate_aligned (cm->timer_wheels, no_of_threads - 1, CLIB_CACHE_LINE_BYTES); cm->lookup_table_nbuckets = IOAM_CACHE_TABLE_DEFAULT_HASH_NUM_BUCKETS; cm->lookup_table_nbuckets = 1 << max_log2 (cm->lookup_table_nbuckets); cm->lookup_table_size = IOAM_CACHE_TABLE_DEFAULT_HASH_MEMORY_SIZE; for (i = 0; i < no_of_threads; i++) { pool_alloc_aligned (cm->ioam_ts_pool[i], MAX_CACHE_TS_ENTRIES, CLIB_CACHE_LINE_BYTES); memset (&cm->ts_stats[i], 0, sizeof (ioam_cache_ts_pool_stats_t)); tw_timer_wheel_init_16t_2w_512sl (&cm->timer_wheels[i], expired_cache_ts_timer_callback, IOAM_CACHE_TS_TICK /* timer period 100ms */ , 10e4); cm->timer_wheels[i].last_run_time = vlib_time_now (vm); } ioam_cache_ts_timer_node_enable (vm, 1); return (1); } always_inline void ioam_cache_ts_timer_set (ioam_cache_main_t * cm, ioam_cache_ts_entry_t * entry, u32 interval) { entry->timer_handle = tw_timer_start_16t_2w_512sl (&cm->timer_wheels[entry->pool_id], entry->pool_index, 1, interval); } always_inline void ioam_cache_ts_timer_reset (ioam_cache_main_t * cm, ioam_cache_ts_entry_t * entry) { tw_timer_stop_16t_2w_512sl (&cm->timer_wheels[entry->pool_id], entry->timer_handle); entry->timer_handle = TIMER_HANDLE_INVALID; } inline static void ioam_cache_ts_entry_free (u32 thread_id, ioam_cache_ts_entry_t * entry, u32 node_index) { ioam_cache_main_t *cm = &ioam_cache_main; vlib_main_t *vm = cm->vlib_main; vlib_frame_t *nf = 0; u32 *to_next; if (entry) { if (entry->hbh != 0) { nf = vlib_get_frame_to_node (vm, node_index); nf->n_vectors = 0; to_next = vlib_frame_vector_args (nf); nf->n_vectors = 1; to_next[0] = entry->buffer_index; vlib_put_frame_to_node (vm, node_index, nf); } pool_put (cm->ioam_ts_pool[thread_id], entry); cm->ts_stats[thread_id].inuse--; memset (entry, 0, sizeof (*entry)); } } inline static int ioam_cache_ts_table_destroy (vlib_main_t * vm) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; int no_of_threads = vec_len (vlib_worker_threads); int i; /* free pool and hash table */ for (i = 0; i < no_of_threads; i++) { pool_foreach (entry, cm->ioam_ts_pool[i], ( { ioam_cache_ts_entry_free (i, entry, cm->error_node_index); } )); pool_free (cm->ioam_ts_pool[i]); cm->ioam_ts_pool = 0; tw_timer_wheel_free_16t_2w_512sl (&cm->timer_wheels[i]); } vec_free (cm->ioam_ts_pool); return (0); } inline static int ioam_cache_ts_entry_cleanup (u32 thread_id, u32 pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_ts_pool[thread_id], pool_index); ioam_cache_ts_entry_free (thread_id, entry, cm->error_node_index); return (0); } /* * Caches buffer for ioam SR tunnel select for Anycast service */ inline static int ioam_cache_ts_add (ip6_header_t * ip0, u16 src_port, u16 dst_port, u32 seq_no, u8 max_responses, u64 now, u32 thread_id, u32 * pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; if (cm->ts_stats[thread_id].inuse == MAX_CACHE_TS_ENTRIES) { cm->ts_stats[thread_id].add_failed++; return (-1); } pool_get_aligned (cm->ioam_ts_pool[thread_id], entry, CLIB_CACHE_LINE_BYTES); memset (entry, 0, sizeof (*entry)); *pool_index = entry - cm->ioam_ts_pool[thread_id]; clib_memcpy (entry->dst_address.as_u64, ip0->dst_address.as_u64, sizeof (ip6_address_t)); clib_memcpy (entry->src_address.as_u64, ip0->src_address.as_u64, sizeof (ip6_address_t)); entry->src_port = src_port; entry->dst_port = dst_port; entry->seq_no = seq_no; entry->response_received = 0; entry->max_responses = max_responses; entry->created_at = now; entry->hbh = 0; entry->buffer_index = 0; entry->pool_id = thread_id; entry->pool_index = *pool_index; ioam_cache_ts_timer_set (cm, entry, IOAM_CACHE_TS_TIMEOUT); cm->ts_stats[thread_id].inuse++; return (0); } inline static void ioam_cache_ts_send (u32 thread_id, i32 pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_ts_pool[thread_id], pool_index); if (!pool_is_free (cm->ioam_ts_pool[thread_id], entry) && entry) { /* send and free pool entry */ ioam_cache_ts_entry_free (thread_id, entry, cm->ip6_hbh_pop_node_index); } } inline static void ioam_cache_ts_check_and_send (u32 thread_id, i32 pool_index) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; entry = pool_elt_at_index (cm->ioam_ts_pool[thread_id], pool_index); if (entry && entry->hbh) { if (entry->response_received == entry->max_responses || entry->created_at + IOAM_CACHE_TS_TIMEOUT <= vlib_time_now (cm->vlib_main)) { ioam_cache_ts_timer_reset (cm, entry); ioam_cache_ts_send (thread_id, pool_index); } } } inline static int ioam_cache_ts_update (u32 thread_id, i32 pool_index, u32 buffer_index, ip6_hop_by_hop_header_t * hbh) { ioam_cache_main_t *cm = &ioam_cache_main; ioam_cache_ts_entry_t *entry = 0; vlib_main_t *vm = cm->vlib_main; vlib_frame_t *nf = 0; u32 *to_next; entry = pool_elt_at_index (cm->ioam_ts_pool[thread_id], pool_index); if (!pool_is_free (cm->ioam_ts_pool[thread_id], entry) && entry) { /* drop existing buffer */ if (entry->hbh != 0) { nf = vlib_get_frame_to_node (vm, cm->error_node_index); nf->n_vectors = 0; to_next = vlib_frame_vector_args (nf); nf->n_vectors = 1; to_next[0] = entry->buffer_index; vlib_put_frame_to_node (vm, cm->error_node_index, nf); } /* update */ entry->buffer_index = buffer_index; entry->hbh = hbh; /* check and send */ ioam_cache_ts_check_and_send (thread_id, pool_index); return (0); } return (-1); } /* * looks up the entry based on the e2e option pool index * result = 0 found the entry * result < 0 indicates failture to find an entry */ inline static int ioam_cache_ts_lookup (ip6_header_t * ip0, u8 protocol, u16 src_port, u16 dst_port, u32 seq_no, ip6_hop_by_hop_header_t ** hbh, u32 * pool_index, u8 * thread_id, u8 response_seen) { ioam_cache_main_t *cm = &ioam_cache_main; ip6_hop_by_hop_header_t *hbh0 = 0; ioam_e2e_cache_option_t *e2e = 0; hbh0 = (ip6_hop_by_hop_header_t *) (ip0 + 1); e2e = (ioam_e2e_cache_option_t *) ((u8 *) hbh0 + cm->rewrite_pool_index_offset); if ((u8 *) e2e < ((u8 *) hbh0 + ((hbh0->length + 1) << 3)) && e2e->hdr.type == HBH_OPTION_TYPE_IOAM_E2E_CACHE_ID) { ioam_cache_ts_entry_t *entry = 0; *pool_index = e2e->pool_index; *thread_id = e2e->pool_id; entry = pool_elt_at_index (cm->ioam_ts_pool[*thread_id], *pool_index); /* match */ if (entry && ip6_address_compare (&ip0->src_address, &entry->dst_address) == 0 && ip6_address_compare (&ip0->dst_address, &entry->src_address) == 0 && entry->src_port == dst_port && entry->dst_port == src_port && entry->seq_no == seq_no) { *hbh = entry->hbh; entry->response_received += response_seen; return (0); } else if (entry) { return (-1); } } return (-1); } inline static u8 * format_ioam_cache_ts_entry (u8 * s, va_list * args) { ioam_cache_ts_entry_t *e = va_arg (*args, ioam_cache_ts_entry_t *); u32 thread_id = va_arg (*args, u32); ioam_cache_main_t *cm = &ioam_cache_main; ioam_e2e_id_option_t *e2e = 0; vlib_main_t *vm = cm->vlib_main; clib_time_t *ct = &vm->clib_time; if (!e) goto end; if (e->hbh) { e2e = ip6_ioam_find_hbh_option (e->hbh, HBH_OPTION_TYPE_IOAM_EDGE_TO_EDGE_ID); s = format (s, "%d: %U:%d to %U:%d seq_no %u buffer %u %U \n\t\tCreated at %U Received %d\n", (e - cm->ioam_ts_pool[thread_id]), format_ip6_address, &e->src_address, e->src_port, format_ip6_address, &e->dst_address, e->dst_port, e->seq_no, e->buffer_index, format_ip6_address, e2e ? &e2e->id : 0, format_time_interval, "h:m:s:u", (e->created_at - vm->cpu_time_main_loop_start) * ct->seconds_per_clock, e->response_received); } else { s = format (s, "%d: %U:%d to %U:%d seq_no %u Buffer %u \n\t\tCreated at %U Received %d\n", (e - cm->ioam_ts_pool[thread_id]), format_ip6_address, &e->src_address, e->src_port, format_ip6_address, &e->dst_address, e->dst_port, e->seq_no, e->buffer_index, format_time_interval, "h:m:s:u", (e->created_at - vm->cpu_time_main_loop_start) * ct->seconds_per_clock, e->response_received); } end: return s; } /* * Get extended rewrite string for iOAM data in v6 * This makes space for an e2e options to carry cache pool info * and manycast server address. * It set the rewrite string per configs in ioam ip6 + new option * for cache along with offset to the option to populate cache * pool id and index */ static inline int ip6_ioam_ts_cache_set_rewrite (void) { ip6_hop_by_hop_ioam_main_t *hm = &ip6_hop_by_hop_ioam_main; ioam_cache_main_t *cm = &ioam_cache_main; ip6_hop_by_hop_header_t *hbh; u32 rewrite_len = 0; ioam_e2e_cache_option_t *e2e = 0; ioam_e2e_id_option_t *e2e_id = 0; vec_free (cm->rewrite); ip6_ioam_set_rewrite (&(cm->rewrite), hm->has_trace_option, hm->has_pot_option, hm->has_seqno_option); hbh = (ip6_hop_by_hop_header_t *) cm->rewrite; rewrite_len = ((hbh->length + 1) << 3); vec_validate (cm->rewrite, rewrite_len - 1 + IOAM_E2E_CACHE_OPTION_RND + IOAM_E2E_ID_OPTION_RND); hbh = (ip6_hop_by_hop_header_t *) cm->rewrite; /* setup e2e id option to insert pool id and index of the node caching it */ hbh->length += IOAM_E2E_CACHE_HBH_EXT_LEN + IOAM_E2E_ID_HBH_EXT_LEN; cm->rewrite_pool_index_offset = rewrite_len; e2e = (ioam_e2e_cache_option_t *) (cm->rewrite + rewrite_len); e2e->hdr.type = HBH_OPTION_TYPE_IOAM_E2E_CACHE_ID | HBH_OPTION_TYPE_SKIP_UNKNOWN; e2e->hdr.length = sizeof (ioam_e2e_cache_option_t) - sizeof (ip6_hop_by_hop_option_t); e2e->e2e_type = 2; e2e_id = (ioam_e2e_id_option_t *) ((u8 *) e2e + sizeof (ioam_e2e_cache_option_t)); e2e_id->hdr.type = HBH_OPTION_TYPE_IOAM_EDGE_TO_EDGE_ID | HBH_OPTION_TYPE_SKIP_UNKNOWN; e2e_id->hdr.length = sizeof (ioam_e2e_id_option_t) - sizeof (ip6_hop_by_hop_option_t); e2e_id->e2e_type = 1; return (0); } static inline int ip6_ioam_ts_cache_cleanup_rewrite (void) { ioam_cache_main_t *cm = &ioam_cache_main; vec_free (cm->rewrite); cm->rewrite = 0; cm->rewrite_pool_index_offset = 0; return (0); } #endif /* __included_ioam_cache_h__ */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */