/* * l2_node.c - l2 ipfix-per-packet graph node * * Copyright (c) * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include #include #include /** * @file l2 flow record generator graph node */ typedef struct { /** interface handle */ u32 rx_sw_if_index; u32 tx_sw_if_index; /** src and dst L2 addresses */ u8 src_mac[6]; u8 dst_mac[6]; /** Ethertype */ u16 ethertype; /** packet timestamp */ u64 timestamp; /** size of the buffer */ u16 buffer_size; } flowperpkt_l2_trace_t; /* packet trace format function */ static u8 * format_flowperpkt_l2_trace (u8 * s, va_list * args) { CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); flowperpkt_l2_trace_t *t = va_arg (*args, flowperpkt_l2_trace_t *); s = format (s, "FLOWPERPKT-L2: rx_sw_if_index %d, tx_sw_if_index %d, src %U dst %U ethertype %0x2, timestamp %lld, size %d", t->rx_sw_if_index, t->tx_sw_if_index, format_ethernet_address, &t->src_mac, format_ethernet_address, &t->dst_mac, t->ethertype, t->timestamp, t->buffer_size); return s; } vlib_node_registration_t flowperpkt_l2_node; /* No counters at the moment */ #define foreach_flowperpkt_l2_error typedef enum { #define _(sym,str) FLOWPERPKT_ERROR_##sym, foreach_flowperpkt_l2_error #undef _ FLOWPERPKT_N_ERROR, } flowperpkt_l2_error_t; static char *flowperpkt_l2_error_strings[] = { #define _(sym,string) string, foreach_flowperpkt_l2_error #undef _ }; typedef enum { FLOWPERPKT_L2_NEXT_DROP, FLOWPERPKT_L2_NEXT_IP4_LOOKUP, FLOWPERPKT_L2_N_NEXT, } flowperpkt_l2_next_t; /** * @brief add an entry to the flow record under construction * @param vm vlib_main_t * current worker thread main structure pointer * @param fm flowperpkt_main_t * flow-per-packet main structure pointer * @param sw_if_index u32 interface handle * @param tos u8 ToS bits from the packet * @param timestamp u64 timestamp, nanoseconds since 1/1/70 * @param length u16 ip length of the packet * @param do_flush int 1 = flush all cached records, 0 = construct a record */ static inline void add_to_flow_record_l2 (vlib_main_t * vm, vlib_node_runtime_t * node, flowperpkt_main_t * fm, u32 rx_sw_if_index, u32 tx_sw_if_index, u8 * src_mac, u8 * dst_mac, u16 ethertype, u64 timestamp, u16 length, int do_flush) { u32 my_cpu_number = vm->cpu_index; flow_report_main_t *frm = &flow_report_main; ip4_header_t *ip; udp_header_t *udp; ip4_ipfix_template_packet_t *tp; ipfix_message_header_t *h; ipfix_set_header_t *s; vlib_frame_t *f; vlib_buffer_t *b0; u16 offset; u32 bi0; vlib_buffer_free_list_t *fl; /* Find or allocate a buffer */ b0 = fm->l2_buffers_per_worker[my_cpu_number]; /* Need to allocate a buffer? */ if (PREDICT_FALSE (b0 == 0)) { /* Nothing to flush */ if (do_flush) return; /* $$$$ drop counter? */ if (vlib_buffer_alloc (vm, &bi0, 1) != 1) return; /* Initialize the buffer */ b0 = fm->l2_buffers_per_worker[my_cpu_number] = vlib_get_buffer (vm, bi0); fl = vlib_buffer_get_free_list (vm, VLIB_BUFFER_DEFAULT_FREE_LIST_INDEX); vlib_buffer_init_for_free_list (b0, fl); VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b0); offset = 0; } else { /* use the current buffer */ bi0 = vlib_get_buffer_index (vm, b0); offset = fm->l2_next_record_offset_per_worker[my_cpu_number]; } /* Find or allocate a frame */ f = fm->l2_frames_per_worker[my_cpu_number]; if (PREDICT_FALSE (f == 0)) { u32 *to_next; f = vlib_get_frame_to_node (vm, ip4_lookup_node.index); fm->l2_frames_per_worker[my_cpu_number] = f; /* Enqueue the buffer */ to_next = vlib_frame_vector_args (f); to_next[0] = bi0; f->n_vectors = 1; } /* Fresh packet, construct header */ if (PREDICT_FALSE (offset == 0)) { flow_report_stream_t *stream; stream = &frm->streams[0]; b0->current_data = 0; b0->current_length = sizeof (*ip) + sizeof (*udp) + sizeof (*h) + sizeof (*s); b0->flags |= (VLIB_BUFFER_TOTAL_LENGTH_VALID | VLIB_BUFFER_FLOW_REPORT); vnet_buffer (b0)->sw_if_index[VLIB_RX] = 0; vnet_buffer (b0)->sw_if_index[VLIB_TX] = frm->fib_index; tp = vlib_buffer_get_current (b0); ip = (ip4_header_t *) & tp->ip4; udp = (udp_header_t *) (ip + 1); h = (ipfix_message_header_t *) (udp + 1); s = (ipfix_set_header_t *) (h + 1); ip->ip_version_and_header_length = 0x45; ip->ttl = 254; ip->protocol = IP_PROTOCOL_UDP; ip->flags_and_fragment_offset = 0; ip->src_address.as_u32 = frm->src_address.as_u32; ip->dst_address.as_u32 = frm->ipfix_collector.as_u32; udp->src_port = clib_host_to_net_u16 (UDP_DST_PORT_ipfix); udp->dst_port = clib_host_to_net_u16 (UDP_DST_PORT_ipfix); udp->checksum = 0; /* FIXUP: message header export_time */ h->export_time = (u32) (((f64) frm->unix_time_0) + (vlib_time_now (frm->vlib_main) - frm->vlib_time_0)); h->export_time = clib_host_to_net_u32 (h->export_time); h->domain_id = clib_host_to_net_u32 (stream->domain_id); /* FIXUP: message header sequence_number */ h->sequence_number = stream->sequence_number++; h->sequence_number = clib_host_to_net_u32 (h->sequence_number); offset = (u32) (((u8 *) (s + 1)) - (u8 *) tp); } /* Add data, unless we're flushing stale data */ if (PREDICT_TRUE (do_flush == 0)) { /* Add data */ /* Ingress interface */ { u32 ingress_interface = clib_host_to_net_u32 (rx_sw_if_index); clib_memcpy (b0->data + offset, &ingress_interface, sizeof (ingress_interface)); offset += sizeof (ingress_interface); } /* Egress interface */ { u32 egress_interface = clib_host_to_net_u32 (tx_sw_if_index); clib_memcpy (b0->data + offset, &egress_interface, sizeof (egress_interface)); offset += sizeof (egress_interface); } /* src mac address */ { clib_memcpy (b0->data + offset, src_mac, 6); offset += 6; } /* dst mac address */ { clib_memcpy (b0->data + offset, dst_mac, 6); offset += 6; } /* ethertype */ b0->data[offset++] = ethertype >> 8; b0->data[offset++] = ethertype & 0xFF; /* Timestamp */ clib_memcpy (b0->data + offset, ×tamp, sizeof (f64)); offset += sizeof (f64); /* pkt size */ { u16 pkt_size = clib_host_to_net_u16 (length); clib_memcpy (b0->data + offset, &pkt_size, sizeof (pkt_size)); offset += sizeof (pkt_size); } b0->current_length += /* 2*sw_if_index + 2*mac + ethertype + timestamp + length = 32 */ 2 * sizeof (u32) + 12 + sizeof (u16) + sizeof (f64) + sizeof (u16); } /* Time to flush the buffer? */ if (PREDICT_FALSE (do_flush || (offset + 2 * sizeof (u32) + 12 + sizeof (u16) + +sizeof (f64) + sizeof (u16)) > frm->path_mtu)) { tp = vlib_buffer_get_current (b0); ip = (ip4_header_t *) & tp->ip4; udp = (udp_header_t *) (ip + 1); h = (ipfix_message_header_t *) (udp + 1); s = (ipfix_set_header_t *) (h + 1); s->set_id_length = ipfix_set_id_length (fm->l2_report_id, b0->current_length - (sizeof (*ip) + sizeof (*udp) + sizeof (*h))); h->version_length = version_length (b0->current_length - (sizeof (*ip) + sizeof (*udp))); ip->length = clib_host_to_net_u16 (b0->current_length); ip->checksum = ip4_header_checksum (ip); udp->length = clib_host_to_net_u16 (b0->current_length - sizeof (*ip)); if (frm->udp_checksum) { /* RFC 7011 section 10.3.2. */ udp->checksum = ip4_tcp_udp_compute_checksum (vm, b0, ip); if (udp->checksum == 0) udp->checksum = 0xffff; } ASSERT (ip->checksum == ip4_header_checksum (ip)); if (PREDICT_FALSE (vlib_get_trace_count (vm, node) > 0)) { vlib_trace_buffer (vm, node, FLOWPERPKT_L2_NEXT_IP4_LOOKUP, b0, 0 /* follow chain */ ); flowperpkt_l2_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); memset (t, 0, sizeof (*t)); t->rx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX]; t->tx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_TX]; t->buffer_size = b0->current_length; } vlib_put_frame_to_node (vm, ip4_lookup_node.index, fm->l2_frames_per_worker[my_cpu_number]); fm->l2_frames_per_worker[my_cpu_number] = 0; fm->l2_buffers_per_worker[my_cpu_number] = 0; offset = 0; } fm->l2_next_record_offset_per_worker[my_cpu_number] = offset; } void flowperpkt_flush_callback_l2 (void) { vlib_main_t *vm = vlib_get_main (); flowperpkt_main_t *fm = &flowperpkt_main; vlib_node_runtime_t *node; node = vlib_node_get_runtime (vm, flowperpkt_l2_node.index); add_to_flow_record_l2 (vm, node, fm, 0 /* rx_sw_if_index */ , 0 /* tx_sw_if_index */ , 0 /* src mac */ , 0 /* dst mac */ , 0 /* ethertype */ , 0ULL /* timestamp */ , 0 /* length */ , 1 /* do_flush */ ); } static uword flowperpkt_l2_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame) { u32 n_left_from, *from, *to_next; flowperpkt_l2_next_t next_index; flowperpkt_main_t *fm = &flowperpkt_main; u64 now; now = (u64) ((vlib_time_now (vm) - fm->vlib_time_0) * 1e9); now += fm->nanosecond_time_0; from = vlib_frame_vector_args (frame); n_left_from = frame->n_vectors; next_index = node->cached_next_index; while (n_left_from > 0) { u32 n_left_to_next; vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); while (n_left_from >= 4 && n_left_to_next >= 2) { u32 next0 = FLOWPERPKT_L2_NEXT_DROP; u32 next1 = FLOWPERPKT_L2_NEXT_DROP; ethernet_header_t *eh0, *eh1; u16 len0, len1; u32 bi0, bi1; vlib_buffer_t *b0, *b1; /* Prefetch next iteration. */ { vlib_buffer_t *p2, *p3; p2 = vlib_get_buffer (vm, from[2]); p3 = vlib_get_buffer (vm, from[3]); vlib_prefetch_buffer_header (p2, LOAD); vlib_prefetch_buffer_header (p3, LOAD); CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, STORE); CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, STORE); } /* speculatively enqueue b0 and b1 to the current next frame */ to_next[0] = bi0 = from[0]; to_next[1] = bi1 = from[1]; from += 2; to_next += 2; n_left_from -= 2; n_left_to_next -= 2; b0 = vlib_get_buffer (vm, bi0); b1 = vlib_get_buffer (vm, bi1); vnet_feature_next (vnet_buffer (b0)->sw_if_index[VLIB_TX], &next0, b0); vnet_feature_next (vnet_buffer (b1)->sw_if_index[VLIB_TX], &next1, b1); eh0 = vlib_buffer_get_current (b0); len0 = vlib_buffer_length_in_chain (vm, b0); if (PREDICT_TRUE ((b0->flags & VLIB_BUFFER_FLOW_REPORT) == 0)) add_to_flow_record_l2 (vm, node, fm, vnet_buffer (b0)->sw_if_index[VLIB_RX], vnet_buffer (b0)->sw_if_index[VLIB_TX], eh0->src_address, eh0->dst_address, eh0->type, now, len0, 0 /* flush */ ); eh1 = vlib_buffer_get_current (b0); len1 = vlib_buffer_length_in_chain (vm, b0); if (PREDICT_TRUE ((b1->flags & VLIB_BUFFER_FLOW_REPORT) == 0)) add_to_flow_record_l2 (vm, node, fm, vnet_buffer (b1)->sw_if_index[VLIB_RX], vnet_buffer (b1)->sw_if_index[VLIB_TX], eh1->src_address, eh1->dst_address, eh1->type, now, len1, 0 /* flush */ ); if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) { if (b0->flags & VLIB_BUFFER_IS_TRACED) { flowperpkt_l2_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->rx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX]; t->tx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_TX]; clib_memcpy (t->src_mac, eh0->src_address, 6); clib_memcpy (t->dst_mac, eh0->dst_address, 6); t->ethertype = clib_net_to_host_u16 (eh0->type); t->timestamp = now; t->buffer_size = len0; } if (b1->flags & VLIB_BUFFER_IS_TRACED) { flowperpkt_l2_trace_t *t = vlib_add_trace (vm, node, b1, sizeof (*t)); t->rx_sw_if_index = vnet_buffer (b1)->sw_if_index[VLIB_RX]; t->tx_sw_if_index = vnet_buffer (b1)->sw_if_index[VLIB_TX]; clib_memcpy (t->src_mac, eh1->src_address, 6); clib_memcpy (t->dst_mac, eh1->dst_address, 6); t->ethertype = clib_net_to_host_u16 (eh1->type); t->timestamp = now; t->buffer_size = len1; } } /* verify speculative enqueues, maybe switch current next frame */ vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next, n_left_to_next, bi0, bi1, next0, next1); } while (n_left_from > 0 && n_left_to_next > 0) { u32 bi0; vlib_buffer_t *b0; u32 next0 = FLOWPERPKT_L2_NEXT_DROP; ethernet_header_t *eh0; u16 len0; /* speculatively enqueue b0 to the current next frame */ bi0 = from[0]; to_next[0] = bi0; from += 1; to_next += 1; n_left_from -= 1; n_left_to_next -= 1; b0 = vlib_get_buffer (vm, bi0); vnet_feature_next (vnet_buffer (b0)->sw_if_index[VLIB_TX], &next0, b0); eh0 = vlib_buffer_get_current (b0); len0 = vlib_buffer_length_in_chain (vm, b0); if (PREDICT_TRUE ((b0->flags & VLIB_BUFFER_FLOW_REPORT) == 0)) add_to_flow_record_l2 (vm, node, fm, vnet_buffer (b0)->sw_if_index[VLIB_RX], vnet_buffer (b0)->sw_if_index[VLIB_TX], eh0->src_address, eh0->dst_address, eh0->type, now, len0, 0 /* flush */ ); if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) && (b0->flags & VLIB_BUFFER_IS_TRACED))) { flowperpkt_l2_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->rx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX]; t->tx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_TX]; clib_memcpy (t->src_mac, eh0->src_address, 6); clib_memcpy (t->dst_mac, eh0->dst_address, 6); t->ethertype = clib_net_to_host_u16 (eh0->type); t->timestamp = now; t->buffer_size = len0; } /* verify speculative enqueue, maybe switch current next frame */ vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next, n_left_to_next, bi0, next0); } vlib_put_next_frame (vm, node, next_index, n_left_to_next); } return frame->n_vectors; } /** * @brief IPFIX l2 flow-per-packet graph node * @node flowperpkt-l2 * * This is the IPFIX flow-record-per-packet node. * * @param vm vlib_main_t corresponding to the current thread. * @param node vlib_node_runtime_t data for this node. * @param frame vlib_frame_t whose contents should be dispatched. * * @par Graph mechanics: buffer metadata, next index usage * * Uses: * - vnet_buffer(b)->ip.save_rewrite_length * - tells the node the length of the rewrite which was applied in * ip4/6_rewrite_inline, allows the code to find the IP header without * having to parse L2 headers, or make stupid assumptions about their * length. * - vnet_buffer(b)->flags & VLIB_BUFFER_FLOW_REPORT * - Used to suppress flow record generation for flow record packets. * * Sets: * - vnet_buffer(b)->flags & VLIB_BUFFER_FLOW_REPORT * - To suppress flow record generation for flow record packets * * Next Index: * - Next configured output feature on the interface, usually * "interface-output." Generated flow records head for ip4-lookup */ /* *INDENT-OFF* */ VLIB_REGISTER_NODE (flowperpkt_l2_node) = { .function = flowperpkt_l2_node_fn, .name = "flowperpkt-l2", .vector_size = sizeof (u32), .format_trace = format_flowperpkt_l2_trace, .type = VLIB_NODE_TYPE_INTERNAL, .n_errors = ARRAY_LEN(flowperpkt_l2_error_strings), .error_strings = flowperpkt_l2_error_strings, .n_next_nodes = FLOWPERPKT_L2_N_NEXT, /* edit / add dispositions here */ .next_nodes = { [FLOWPERPKT_L2_NEXT_DROP] = "error-drop", [FLOWPERPKT_L2_NEXT_IP4_LOOKUP] = "ip4-lookup", }, }; /* *INDENT-ON* */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */