diff options
author | Dave Barach <dave@barachs.net> | 2016-12-20 17:59:27 -0500 |
---|---|---|
committer | Damjan Marion <dmarion.lists@gmail.com> | 2016-12-22 19:04:16 +0000 |
commit | 504b83009a00d70c777c04014967494642074d52 (patch) | |
tree | 8159edc93c0027b0dac200875edf1f07481bbbff /plugins/flowperpkt-plugin | |
parent | abea9664983b723023e692d77fd4fe3ca29c1e34 (diff) |
Add L2 flow-record generation
Change-Id: I32a9513f2b66d35e0abe2aeb6eb86a4729ba7f74
Signed-off-by: Dave Barach <dave@barachs.net>
Diffstat (limited to 'plugins/flowperpkt-plugin')
-rw-r--r-- | plugins/flowperpkt-plugin/Makefile.am | 1 | ||||
-rw-r--r-- | plugins/flowperpkt-plugin/flowperpkt/flowperpkt.api | 2 | ||||
-rw-r--r-- | plugins/flowperpkt-plugin/flowperpkt/flowperpkt.c | 299 | ||||
-rw-r--r-- | plugins/flowperpkt-plugin/flowperpkt/flowperpkt.h | 32 | ||||
-rw-r--r-- | plugins/flowperpkt-plugin/flowperpkt/flowperpkt_test.c | 5 | ||||
-rw-r--r-- | plugins/flowperpkt-plugin/flowperpkt/l2_node.c | 544 | ||||
-rw-r--r-- | plugins/flowperpkt-plugin/flowperpkt/node.c | 159 |
7 files changed, 881 insertions, 161 deletions
diff --git a/plugins/flowperpkt-plugin/Makefile.am b/plugins/flowperpkt-plugin/Makefile.am index 1df758ed6f4..9354e26faa2 100644 --- a/plugins/flowperpkt-plugin/Makefile.am +++ b/plugins/flowperpkt-plugin/Makefile.am @@ -24,6 +24,7 @@ vppplugins_LTLIBRARIES = flowperpkt_plugin.la vppapitestplugins_LTLIBRARIES = flowperpkt_test_plugin.la flowperpkt_plugin_la_SOURCES = flowperpkt/flowperpkt.c \ + flowperpkt/l2_node.c \ flowperpkt/node.c \ flowperpkt/flowperpkt_plugin.api.h flowperpkt_plugin_la_LDFLAGS = -module diff --git a/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.api b/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.api index 8933b5852f4..fa878f21ed3 100644 --- a/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.api +++ b/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.api @@ -22,7 +22,7 @@ manual_print define flowperpkt_tx_interface_add_del /* Enable / disable the feature */ u8 is_add; - u8 is_ipv6; + u8 which; /* 0 = ipv4, 1 = l2, 2 = ipv6 (not yet implemented) */ /* Interface handle */ u32 sw_if_index; diff --git a/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.c b/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.c index 96c930ae599..fb71d5b0ffc 100644 --- a/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.c +++ b/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.c @@ -57,13 +57,21 @@ flowperpkt_main_t flowperpkt_main; #include <flowperpkt/flowperpkt_all_api_h.h> #undef vl_api_version -/* Define the per-interface configurable feature */ +/* Define the per-interface configurable features */ /* *INDENT-OFF* */ -VNET_FEATURE_INIT (flow_perpacket, static) = { +VNET_FEATURE_INIT (flow_perpacket_ipv4, static) = +{ .arc_name = "ip4-output", - .node_name = "flowperpkt", + .node_name = "flowperpkt-ipv4", .runs_before = VNET_FEATURES ("interface-output"), }; + +VNET_FEATURE_INIT (flow_perpacket_l2, static) = +{ + .arc_name = "interface-output", + .node_name = "flowperpkt-l2", + .runs_before = VNET_FEATURES ("interface-tx"), +}; /* *INDENT-ON* */ /* @@ -120,11 +128,12 @@ bad_sw_if_index: \ * @param collector_port u16 the collector port we should use, host byte order * @returns u8 * vector containing the indicated IPFIX template packet */ -u8 * -flowperpkt_template_rewrite (flow_report_main_t * frm, - flow_report_t * fr, - ip4_address_t * collector_address, - ip4_address_t * src_address, u16 collector_port) +static inline u8 * +flowperpkt_template_rewrite_inline (flow_report_main_t * frm, + flow_report_t * fr, + ip4_address_t * collector_address, + ip4_address_t * src_address, + u16 collector_port, int variant) { ip4_header_t *ip; udp_header_t *udp; @@ -137,32 +146,64 @@ flowperpkt_template_rewrite (flow_report_main_t * frm, ip4_ipfix_template_packet_t *tp; u32 field_count = 0; flow_report_stream_t *stream; + flowperpkt_main_t *fm = &flowperpkt_main; stream = &frm->streams[fr->stream_index]; - /* - * Supported Fields: - * - * ingressInterface, TLV type 10, u32 - * egressInterface, TLV type 14, u32 - * sourceIpv4Address, TLV type 8, u32 - * destinationIPv4Address, TLV type 12, u32 - * ipClassOfService, TLV type 5, u8 - * flowStartNanoseconds, TLV type 156, dateTimeNanoseconds (f64) - * Implementation: f64 nanoseconds since VPP started - * warning: wireshark doesn't really understand this TLV - * dataLinkFrameSize, TLV type 312, u16 - * warning: wireshark doesn't understand this TLV at all - */ - - /* Currently 7 fields */ - field_count += 7; - - /* allocate rewrite space */ - vec_validate_aligned (rewrite, - sizeof (ip4_ipfix_template_packet_t) - + field_count * sizeof (ipfix_field_specifier_t) - 1, - CLIB_CACHE_LINE_BYTES); + if (variant == FLOW_VARIANT_IPV4) + { + /* + * ip4 Supported Fields: + * + * ingressInterface, TLV type 10, u32 + * egressInterface, TLV type 14, u32 + * sourceIpv4Address, TLV type 8, u32 + * destinationIPv4Address, TLV type 12, u32 + * ipClassOfService, TLV type 5, u8 + * flowStartNanoseconds, TLV type 156, dateTimeNanoseconds (f64) + * Implementation: f64 nanoseconds since VPP started + * warning: wireshark doesn't really understand this TLV + * dataLinkFrameSize, TLV type 312, u16 + * warning: wireshark doesn't understand this TLV at all + */ + + /* Currently 7 fields */ + field_count += 7; + + /* allocate rewrite space */ + vec_validate_aligned + (rewrite, + sizeof (ip4_ipfix_template_packet_t) + + field_count * sizeof (ipfix_field_specifier_t) - 1, + CLIB_CACHE_LINE_BYTES); + } + else if (variant == FLOW_VARIANT_L2) + { + /* + * L2 Supported Fields: + * + * ingressInterface, TLV type 10, u32 + * egressInterface, TLV type 14, u32 + * sourceMacAddress, TLV type 56, u8[6] we hope + * destinationMacAddress, TLV type 57, u8[6] we hope + * ethernetType, TLV type 256, u16 + * flowStartNanoseconds, TLV type 156, dateTimeNanoseconds (f64) + * Implementation: f64 nanoseconds since VPP started + * warning: wireshark doesn't really understand this TLV + * dataLinkFrameSize, TLV type 312, u16 + * warning: wireshark doesn't understand this TLV at all + */ + + /* Currently 7 fields */ + field_count += 7; + + /* allocate rewrite space */ + vec_validate_aligned + (rewrite, + sizeof (ip4_ipfix_template_packet_t) + + field_count * sizeof (ipfix_field_specifier_t) - 1, + CLIB_CACHE_LINE_BYTES); + } tp = (ip4_ipfix_template_packet_t *) rewrite; ip = (ip4_header_t *) & tp->ip4; @@ -186,28 +227,66 @@ flowperpkt_template_rewrite (flow_report_main_t * frm, h->domain_id = clib_host_to_net_u32 (stream->domain_id); /* Add TLVs to the template */ - f->e_id_length = ipfix_e_id_length (0 /* enterprise */ , ingressInterface, - 4); - f++; - f->e_id_length = ipfix_e_id_length (0 /* enterprise */ , egressInterface, - 4); - f++; - f->e_id_length = ipfix_e_id_length (0 /* enterprise */ , sourceIPv4Address, - 4); - f++; - f->e_id_length = ipfix_e_id_length (0 /* enterprise */ , - destinationIPv4Address, 4); - f++; - f->e_id_length = ipfix_e_id_length (0 /* enterprise */ , ipClassOfService, - 1); - f++; - f->e_id_length = - ipfix_e_id_length (0 /* enterprise */ , flowStartNanoseconds, - 8); - f++; - f->e_id_length = ipfix_e_id_length (0 /* enterprise */ , dataLinkFrameSize, - 2); - f++; + if (variant == FLOW_VARIANT_IPV4) + { + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , ingressInterface, + 4); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , egressInterface, + 4); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , sourceIPv4Address, + 4); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , destinationIPv4Address, 4); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , ipClassOfService, + 1); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , flowStartNanoseconds, + 8); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , dataLinkFrameSize, + 2); + f++; + } + else if (variant == FLOW_VARIANT_L2) + { + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , ingressInterface, + 4); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , egressInterface, + 4); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , sourceMacAddress, + 6); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , destinationMacAddress, 6); + f++; + f->e_id_length = ipfix_e_id_length (0 /* enterprise */ , ethernetType, + 2); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , flowStartNanoseconds, + 8); + f++; + f->e_id_length = + ipfix_e_id_length (0 /* enterprise */ , dataLinkFrameSize, + 2); + f++; + } + /* Extend in the obvious way, right here... */ /* Back to the template packet... */ @@ -218,6 +297,11 @@ flowperpkt_template_rewrite (flow_report_main_t * frm, /* Field count in this template */ t->id_count = ipfix_id_count (fr->template_id, f - first_field); + if (variant == FLOW_VARIANT_IPV4) + fm->ipv4_report_id = fr->template_id; + else if (variant == FLOW_VARIANT_L2) + fm->l2_report_id = fr->template_id; + /* set length in octets */ s->set_id_length = ipfix_set_id_length (2 /* set_id */ , (u8 *) f - (u8 *) s); @@ -231,6 +315,31 @@ flowperpkt_template_rewrite (flow_report_main_t * frm, return rewrite; } +u8 * +flowperpkt_template_rewrite_ipv4 (flow_report_main_t * frm, + flow_report_t * fr, + ip4_address_t * collector_address, + ip4_address_t * src_address, + u16 collector_port) +{ + return flowperpkt_template_rewrite_inline + (frm, fr, collector_address, src_address, collector_port, + FLOW_VARIANT_IPV4); +} + +u8 * +flowperpkt_template_rewrite_l2 (flow_report_main_t * frm, + flow_report_t * fr, + ip4_address_t * collector_address, + ip4_address_t * src_address, + u16 collector_port) +{ + return flowperpkt_template_rewrite_inline + (frm, fr, collector_address, src_address, collector_port, + FLOW_VARIANT_L2); +} + + /** * @brief Flush accumulated data * @param frm flow_report_main_t * @@ -242,11 +351,21 @@ flowperpkt_template_rewrite (flow_report_main_t * frm, * will be sent. */ vlib_frame_t * -flowperpkt_data_callback (flow_report_main_t * frm, - flow_report_t * fr, - vlib_frame_t * f, u32 * to_next, u32 node_index) +flowperpkt_data_callback_ipv4 (flow_report_main_t * frm, + flow_report_t * fr, + vlib_frame_t * f, u32 * to_next, + u32 node_index) +{ + flowperpkt_flush_callback_ipv4 (); + return f; +} + +vlib_frame_t * +flowperpkt_data_callback_l2 (flow_report_main_t * frm, + flow_report_t * fr, + vlib_frame_t * f, u32 * to_next, u32 node_index) { - flowperpkt_flush_callback (); + flowperpkt_flush_callback_l2 (); return f; } @@ -259,21 +378,21 @@ flowperpkt_data_callback (flow_report_main_t * frm, */ static int flowperpkt_tx_interface_add_del_feature - (flowperpkt_main_t * fm, u32 sw_if_index, int is_add) + (flowperpkt_main_t * fm, u32 sw_if_index, int which, int is_add) { flow_report_main_t *frm = &flow_report_main; vnet_flow_report_add_del_args_t _a, *a = &_a; int rv; - if (!fm->report_created) + if (which == FLOW_VARIANT_IPV4 && !fm->ipv4_report_created) { memset (a, 0, sizeof (*a)); - a->rewrite_callback = flowperpkt_template_rewrite; - a->flow_data_callback = flowperpkt_data_callback; + a->rewrite_callback = flowperpkt_template_rewrite_ipv4; + a->flow_data_callback = flowperpkt_data_callback_ipv4; a->is_add = 1; a->domain_id = 1; /*$$$$ config parameter */ a->src_port = 4739; /*$$$$ config parameter */ - fm->report_created = 1; + fm->ipv4_report_created = 1; rv = vnet_flow_report_add_del (frm, a); if (rv) @@ -282,9 +401,30 @@ static int flowperpkt_tx_interface_add_del_feature return -1; } } + else if (which == FLOW_VARIANT_L2 && !fm->l2_report_created) + { + memset (a, 0, sizeof (*a)); + a->rewrite_callback = flowperpkt_template_rewrite_l2; + a->flow_data_callback = flowperpkt_data_callback_l2; + a->is_add = 1; + a->domain_id = 1; /*$$$$ config parameter */ + a->src_port = 4739; /*$$$$ config parameter */ + fm->l2_report_created = 1; - vnet_feature_enable_disable ("ip4-output", "flowperpkt", sw_if_index, - is_add, 0, 0); + rv = vnet_flow_report_add_del (frm, a); + if (rv) + { + clib_warning ("vnet_flow_report_add_del returned %d", rv); + return -1; + } + } + + if (which == FLOW_VARIANT_IPV4) + vnet_feature_enable_disable ("ip4-output", "flowperpkt-ipv4", + sw_if_index, is_add, 0, 0); + else if (which == FLOW_VARIANT_L2) + vnet_feature_enable_disable ("interface-output", "flowperpkt-l2", + sw_if_index, is_add, 0, 0); return 0; } @@ -303,8 +443,15 @@ void vl_api_flowperpkt_tx_interface_add_del_t_handler VALIDATE_SW_IF_INDEX (mp); - rv = flowperpkt_tx_interface_add_del_feature (fm, sw_if_index, mp->is_add); + if (mp->which != FLOW_VARIANT_IPV4 && mp->which != FLOW_VARIANT_L2) + { + rv = VNET_API_ERROR_UNIMPLEMENTED; + goto out; + } + rv = flowperpkt_tx_interface_add_del_feature (fm, sw_if_index, mp->which, + mp->is_add); +out: BAD_SW_IF_INDEX_LABEL; REPLY_MACRO (VL_API_FLOWPERPKT_TX_INTERFACE_ADD_DEL_REPLY); @@ -322,9 +469,9 @@ static void *vl_api_flowperpkt_tx_interface_add_del_t_print u8 *s; s = format (0, "SCRIPT: flowperpkt_tx_interface_add_del "); - s = format (s, "sw_if_index %d is_add %d is_ipv6 %d ", + s = format (s, "sw_if_index %d is_add %d which %d ", clib_host_to_net_u32 (mp->sw_if_index), - (int) mp->is_add, (int) mp->is_ipv6); + (int) mp->is_add, (int) mp->which); FINISH; } @@ -365,6 +512,7 @@ flowperpkt_tx_interface_add_del_feature_command_fn (vlib_main_t * vm, flowperpkt_main_t *fm = &flowperpkt_main; u32 sw_if_index = ~0; int is_add = 1; + u8 which = FLOW_VARIANT_IPV4; int rv; @@ -373,8 +521,9 @@ flowperpkt_tx_interface_add_del_feature_command_fn (vlib_main_t * vm, if (unformat (input, "disable")) is_add = 0; else if (unformat (input, "%U", unformat_vnet_sw_interface, - fm->vnet_main, &sw_if_index)) - ; + fm->vnet_main, &sw_if_index)); + else if (unformat (input, "l2")) + which = FLOW_VARIANT_L2; else break; } @@ -382,7 +531,8 @@ flowperpkt_tx_interface_add_del_feature_command_fn (vlib_main_t * vm, if (sw_if_index == ~0) return clib_error_return (0, "Please specify an interface..."); - rv = flowperpkt_tx_interface_add_del_feature (fm, sw_if_index, is_add); + rv = + flowperpkt_tx_interface_add_del_feature (fm, sw_if_index, which, is_add); switch (rv) { case 0: @@ -496,9 +646,12 @@ flowperpkt_init (vlib_main_t * vm) num_threads = 1 /* main thread */ + tm->n_eal_threads; /* Allocate per worker thread vectors */ - vec_validate (fm->buffers_per_worker, num_threads - 1); - vec_validate (fm->frames_per_worker, num_threads - 1); - vec_validate (fm->next_record_offset_per_worker, num_threads - 1); + vec_validate (fm->ipv4_buffers_per_worker, num_threads - 1); + vec_validate (fm->l2_buffers_per_worker, num_threads - 1); + vec_validate (fm->ipv4_frames_per_worker, num_threads - 1); + vec_validate (fm->l2_frames_per_worker, num_threads - 1); + vec_validate (fm->ipv4_next_record_offset_per_worker, num_threads - 1); + vec_validate (fm->l2_next_record_offset_per_worker, num_threads - 1); /* Set up time reference pair */ fm->vlib_time_0 = vlib_time_now (vm); diff --git a/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.h b/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.h index 31e685eb68d..20f6939dda5 100644 --- a/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.h +++ b/plugins/flowperpkt-plugin/flowperpkt/flowperpkt.h @@ -35,15 +35,25 @@ typedef struct /** API message ID base */ u16 msg_id_base; - /** Has the report been created? */ - int report_created; + /** Have the reports [templates] been created? */ + int ipv4_report_created; + int l2_report_created; + + /** stream/template IDs */ + u16 ipv4_report_id; + u16 l2_report_id; /** ipfix buffers under construction, per-worker thread */ - vlib_buffer_t **buffers_per_worker; + vlib_buffer_t **ipv4_buffers_per_worker; + vlib_buffer_t **l2_buffers_per_worker; + /** frames containing ipfix buffers, per-worker thread */ - vlib_frame_t **frames_per_worker; + vlib_frame_t **ipv4_frames_per_worker; + vlib_frame_t **l2_frames_per_worker; + /** next record offset, per worker thread */ - u16 *next_record_offset_per_worker; + u16 *ipv4_next_record_offset_per_worker; + u16 *l2_next_record_offset_per_worker; /** Time reference pair */ u64 nanosecond_time_0; @@ -55,11 +65,19 @@ typedef struct vnet_main_t *vnet_main; } flowperpkt_main_t; +typedef enum +{ + FLOW_VARIANT_IPV4, + FLOW_VARIANT_L2, + FLOW_N_VARIANTS, +} flowperpkt_variant_t; + extern flowperpkt_main_t flowperpkt_main; -vlib_node_registration_t flowperpkt_node; +extern vlib_node_registration_t flowperpkt_ipv4_node; -void flowperpkt_flush_callback (void); +void flowperpkt_flush_callback_ipv4 (void); +void flowperpkt_flush_callback_l2 (void); #endif /* __included_flowperpkt_h__ */ diff --git a/plugins/flowperpkt-plugin/flowperpkt/flowperpkt_test.c b/plugins/flowperpkt-plugin/flowperpkt/flowperpkt_test.c index 3c1cd227f57..716818ffe0a 100644 --- a/plugins/flowperpkt-plugin/flowperpkt/flowperpkt_test.c +++ b/plugins/flowperpkt-plugin/flowperpkt/flowperpkt_test.c @@ -132,6 +132,7 @@ api_flowperpkt_tx_interface_add_del (vat_main_t * vam) unformat_input_t *i = vam->input; f64 timeout; int enable_disable = 1; + u8 which = 0; /* ipv4 by default */ u32 sw_if_index = ~0; vl_api_flowperpkt_tx_interface_add_del_t *mp; @@ -144,6 +145,8 @@ api_flowperpkt_tx_interface_add_del (vat_main_t * vam) ; else if (unformat (i, "disable")) enable_disable = 0; + else if (unformat (i, "l2")) + which = 1; else break; } @@ -158,7 +161,7 @@ api_flowperpkt_tx_interface_add_del (vat_main_t * vam) M (FLOWPERPKT_TX_INTERFACE_ADD_DEL, flowperpkt_tx_interface_add_del); mp->sw_if_index = ntohl (sw_if_index); mp->is_add = enable_disable; - mp->is_ipv6 = 0; /* $$$$ */ + mp->which = which; /* send it... */ S; diff --git a/plugins/flowperpkt-plugin/flowperpkt/l2_node.c b/plugins/flowperpkt-plugin/flowperpkt/l2_node.c new file mode 100644 index 00000000000..ba87d431b88 --- /dev/null +++ b/plugins/flowperpkt-plugin/flowperpkt/l2_node.c @@ -0,0 +1,544 @@ +/* + * l2_node.c - l2 ipfix-per-packet graph node + * + * Copyright (c) <current-year> <your-organization> + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at: + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +#include <vlib/vlib.h> +#include <vnet/vnet.h> +#include <vnet/pg/pg.h> +#include <vppinfra/error.h> +#include <flowperpkt/flowperpkt.h> + +/** + * @file l2 flow record generator graph node + */ + +typedef struct +{ + /** interface handle */ + u32 rx_sw_if_index; + u32 tx_sw_if_index; + /** src and dst L2 addresses */ + u8 src_mac[6]; + u8 dst_mac[6]; + /** Ethertype */ + u16 ethertype; + /** packet timestamp */ + u64 timestamp; + /** size of the buffer */ + u16 buffer_size; +} flowperpkt_l2_trace_t; + +/* packet trace format function */ +static u8 * +format_flowperpkt_l2_trace (u8 * s, va_list * args) +{ + CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); + CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); + flowperpkt_l2_trace_t *t = va_arg (*args, flowperpkt_l2_trace_t *); + + s = format (s, + "FLOWPERPKT-L2: rx_sw_if_index %d, tx_sw_if_index %d, src %U dst %U ethertype %0x2, timestamp %lld, size %d", + t->rx_sw_if_index, t->tx_sw_if_index, + format_ethernet_address, &t->src_mac, + format_ethernet_address, &t->dst_mac, + t->ethertype, t->timestamp, t->buffer_size); + return s; +} + +vlib_node_registration_t flowperpkt_l2_node; + +/* No counters at the moment */ +#define foreach_flowperpkt_l2_error + +typedef enum +{ +#define _(sym,str) FLOWPERPKT_ERROR_##sym, + foreach_flowperpkt_l2_error +#undef _ + FLOWPERPKT_N_ERROR, +} flowperpkt_l2_error_t; + +static char *flowperpkt_l2_error_strings[] = { +#define _(sym,string) string, + foreach_flowperpkt_l2_error +#undef _ +}; + +typedef enum +{ + FLOWPERPKT_L2_NEXT_DROP, + FLOWPERPKT_L2_N_NEXT, +} flowperpkt_l2_next_t; + +/** + * @brief add an entry to the flow record under construction + * @param vm vlib_main_t * current worker thread main structure pointer + * @param fm flowperpkt_main_t * flow-per-packet main structure pointer + * @param sw_if_index u32 interface handle + * @param tos u8 ToS bits from the packet + * @param timestamp u64 timestamp, nanoseconds since 1/1/70 + * @param length u16 ip length of the packet + * @param do_flush int 1 = flush all cached records, 0 = construct a record + */ + +static inline void +add_to_flow_record_l2 (vlib_main_t * vm, + flowperpkt_main_t * fm, + u32 rx_sw_if_index, u32 tx_sw_if_index, + u8 * src_mac, u8 * dst_mac, + u16 ethertype, u64 timestamp, u16 length, int do_flush) +{ + u32 my_cpu_number = vm->cpu_index; + flow_report_main_t *frm = &flow_report_main; + ip4_header_t *ip; + udp_header_t *udp; + ip4_ipfix_template_packet_t *tp; + ipfix_message_header_t *h; + ipfix_set_header_t *s; + vlib_frame_t *f; + vlib_buffer_t *b0; + u16 offset; + u32 bi0; + vlib_buffer_free_list_t *fl; + + /* Find or allocate a buffer */ + b0 = fm->l2_buffers_per_worker[my_cpu_number]; + + /* Need to allocate a buffer? */ + if (PREDICT_FALSE (b0 == 0)) + { + /* Nothing to flush */ + if (do_flush) + return; + + /* $$$$ drop counter? */ + if (vlib_buffer_alloc (vm, &bi0, 1) != 1) + return; + + /* Initialize the buffer */ + b0 = fm->l2_buffers_per_worker[my_cpu_number] = + vlib_get_buffer (vm, bi0); + fl = + vlib_buffer_get_free_list (vm, VLIB_BUFFER_DEFAULT_FREE_LIST_INDEX); + vlib_buffer_init_for_free_list (b0, fl); + VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b0); + offset = 0; + } + else + { + /* use the current buffer */ + bi0 = vlib_get_buffer_index (vm, b0); + offset = fm->l2_next_record_offset_per_worker[my_cpu_number]; + } + + /* Find or allocate a frame */ + f = fm->l2_frames_per_worker[my_cpu_number]; + if (PREDICT_FALSE (f == 0)) + { + u32 *to_next; + f = vlib_get_frame_to_node (vm, ip4_lookup_node.index); + fm->l2_frames_per_worker[my_cpu_number] = f; + + /* Enqueue the buffer */ + to_next = vlib_frame_vector_args (f); + to_next[0] = bi0; + f->n_vectors = 1; + } + + /* Fresh packet, construct header */ + if (PREDICT_FALSE (offset == 0)) + { + flow_report_stream_t *stream; + + stream = &frm->streams[0]; + + b0->current_data = 0; + b0->current_length = sizeof (*ip) + sizeof (*udp) + sizeof (*h) + + sizeof (*s); + b0->flags |= (VLIB_BUFFER_TOTAL_LENGTH_VALID | VLIB_BUFFER_FLOW_REPORT); + vnet_buffer (b0)->sw_if_index[VLIB_RX] = 0; + vnet_buffer (b0)->sw_if_index[VLIB_TX] = frm->fib_index; + + tp = vlib_buffer_get_current (b0); + ip = (ip4_header_t *) & tp->ip4; + udp = (udp_header_t *) (ip + 1); + h = (ipfix_message_header_t *) (udp + 1); + s = (ipfix_set_header_t *) (h + 1); + + ip->ip_version_and_header_length = 0x45; + ip->ttl = 254; + ip->protocol = IP_PROTOCOL_UDP; + ip->flags_and_fragment_offset = 0; + ip->src_address.as_u32 = frm->src_address.as_u32; + ip->dst_address.as_u32 = frm->ipfix_collector.as_u32; + udp->src_port = clib_host_to_net_u16 (UDP_DST_PORT_ipfix); + udp->dst_port = clib_host_to_net_u16 (UDP_DST_PORT_ipfix); + udp->checksum = 0; + + /* FIXUP: message header export_time */ + h->export_time = (u32) + (((f64) frm->unix_time_0) + + (vlib_time_now (frm->vlib_main) - frm->vlib_time_0)); + h->export_time = clib_host_to_net_u32 (h->export_time); + h->domain_id = clib_host_to_net_u32 (stream->domain_id); + + /* FIXUP: message header sequence_number */ + h->sequence_number = stream->sequence_number++; + h->sequence_number = clib_host_to_net_u32 (h->sequence_number); + + offset = (u32) (((u8 *) (s + 1)) - (u8 *) tp); + } + + /* Add data, unless we're flushing stale data */ + if (PREDICT_TRUE (do_flush == 0)) + { + + /* Add data */ + /* Ingress interface */ + { + u32 ingress_interface = clib_host_to_net_u32 (rx_sw_if_index); + clib_memcpy (b0->data + offset, &ingress_interface, + sizeof (ingress_interface)); + offset += sizeof (ingress_interface); + } + /* Egress interface */ + { + u32 egress_interface = clib_host_to_net_u32 (tx_sw_if_index); + clib_memcpy (b0->data + offset, &egress_interface, + sizeof (egress_interface)); + offset += sizeof (egress_interface); + } + /* src mac address */ + { + clib_memcpy (b0->data + offset, src_mac, 6); + offset += 6; + } + /* dst mac address */ + { + clib_memcpy (b0->data + offset, dst_mac, 6); + offset += 6; + } + + /* ethertype */ + b0->data[offset++] = ethertype >> 8; + b0->data[offset++] = ethertype & 0xFF; + + /* Timestamp */ + clib_memcpy (b0->data + offset, ×tamp, sizeof (f64)); + offset += sizeof (f64); + + /* pkt size */ + { + u16 pkt_size = clib_host_to_net_u16 (length); + clib_memcpy (b0->data + offset, &pkt_size, sizeof (pkt_size)); + offset += sizeof (pkt_size); + } + + b0->current_length += + /* 2*sw_if_index + 2*mac + ethertype + timestamp + length = 32 */ + 2 * sizeof (u32) + 12 + sizeof (u16) + sizeof (f64) + sizeof (u16); + + } + /* Time to flush the buffer? */ + if (PREDICT_FALSE + (do_flush || (offset + 2 * sizeof (u32) + 12 + sizeof (u16) + + +sizeof (f64) + sizeof (u16)) > frm->path_mtu)) + { + tp = vlib_buffer_get_current (b0); + ip = (ip4_header_t *) & tp->ip4; + udp = (udp_header_t *) (ip + 1); + h = (ipfix_message_header_t *) (udp + 1); + s = (ipfix_set_header_t *) (h + 1); + + s->set_id_length = ipfix_set_id_length (fm->l2_report_id, + b0->current_length - + (sizeof (*ip) + sizeof (*udp) + + sizeof (*h))); + h->version_length = version_length (b0->current_length - + (sizeof (*ip) + sizeof (*udp))); + + ip->length = clib_host_to_net_u16 (b0->current_length); + + ip->checksum = ip4_header_checksum (ip); + udp->length = clib_host_to_net_u16 (b0->current_length - sizeof (*ip)); + + if (frm->udp_checksum) + { + /* RFC 7011 section 10.3.2. */ + udp->checksum = ip4_tcp_udp_compute_checksum (vm, b0, ip); + if (udp->checksum == 0) + udp->checksum = 0xffff; + } + + ASSERT (ip->checksum == ip4_header_checksum (ip)); + + vlib_put_frame_to_node (vm, ip4_lookup_node.index, + fm->l2_frames_per_worker[my_cpu_number]); + fm->l2_frames_per_worker[my_cpu_number] = 0; + fm->l2_buffers_per_worker[my_cpu_number] = 0; + offset = 0; + } + + fm->l2_next_record_offset_per_worker[my_cpu_number] = offset; +} + +void +flowperpkt_flush_callback_l2 (void) +{ + vlib_main_t *vm = vlib_get_main (); + flowperpkt_main_t *fm = &flowperpkt_main; + + add_to_flow_record_l2 (vm, fm, 0 /* rx_sw_if_index */ , + 0 /* tx_sw_if_index */ , + 0 /* src mac */ , + 0 /* dst mac */ , + 0 /* ethertype */ , + 0ULL /* timestamp */ , + 0 /* length */ , + 1 /* do_flush */ ); +} + + +static uword +flowperpkt_l2_node_fn (vlib_main_t * vm, + vlib_node_runtime_t * node, vlib_frame_t * frame) +{ + u32 n_left_from, *from, *to_next; + flowperpkt_l2_next_t next_index; + flowperpkt_main_t *fm = &flowperpkt_main; + u64 now; + + now = (u64) ((vlib_time_now (vm) - fm->vlib_time_0) * 1e9); + now += fm->nanosecond_time_0; + + from = vlib_frame_vector_args (frame); + n_left_from = frame->n_vectors; + next_index = node->cached_next_index; + + while (n_left_from > 0) + { + u32 n_left_to_next; + + vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next); + + while (n_left_from >= 4 && n_left_to_next >= 2) + { + u32 next0 = FLOWPERPKT_L2_NEXT_DROP; + u32 next1 = FLOWPERPKT_L2_NEXT_DROP; + ethernet_header_t *eh0, *eh1; + u16 len0, len1; + u32 bi0, bi1; + vlib_buffer_t *b0, *b1; + + /* Prefetch next iteration. */ + { + vlib_buffer_t *p2, *p3; + + p2 = vlib_get_buffer (vm, from[2]); + p3 = vlib_get_buffer (vm, from[3]); + + vlib_prefetch_buffer_header (p2, LOAD); + vlib_prefetch_buffer_header (p3, LOAD); + + CLIB_PREFETCH (p2->data, CLIB_CACHE_LINE_BYTES, STORE); + CLIB_PREFETCH (p3->data, CLIB_CACHE_LINE_BYTES, STORE); + } + + /* speculatively enqueue b0 and b1 to the current next frame */ + to_next[0] = bi0 = from[0]; + to_next[1] = bi1 = from[1]; + from += 2; + to_next += 2; + n_left_from -= 2; + n_left_to_next -= 2; + + b0 = vlib_get_buffer (vm, bi0); + b1 = vlib_get_buffer (vm, bi1); + + vnet_feature_next (vnet_buffer (b0)->sw_if_index[VLIB_TX], + &next0, b0); + vnet_feature_next (vnet_buffer (b1)->sw_if_index[VLIB_TX], + &next1, b1); + + eh0 = vlib_buffer_get_current (b0); + len0 = vlib_buffer_length_in_chain (vm, b0); + + if (PREDICT_TRUE ((b0->flags & VLIB_BUFFER_FLOW_REPORT) == 0)) + add_to_flow_record_l2 (vm, fm, + vnet_buffer (b0)->sw_if_index[VLIB_RX], + vnet_buffer (b0)->sw_if_index[VLIB_TX], + eh0->src_address, + eh0->dst_address, + eh0->type, now, len0, 0 /* flush */ ); + + eh1 = vlib_buffer_get_current (b0); + len1 = vlib_buffer_length_in_chain (vm, b0); + + if (PREDICT_TRUE ((b1->flags & VLIB_BUFFER_FLOW_REPORT) == 0)) + add_to_flow_record_l2 (vm, fm, + vnet_buffer (b1)->sw_if_index[VLIB_RX], + vnet_buffer (b1)->sw_if_index[VLIB_TX], + eh1->src_address, + eh1->dst_address, + eh1->type, now, len1, 0 /* flush */ ); + + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) + { + if (b0->flags & VLIB_BUFFER_IS_TRACED) + { + flowperpkt_l2_trace_t *t = + vlib_add_trace (vm, node, b0, sizeof (*t)); + t->rx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + t->tx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_TX]; + clib_memcpy (t->src_mac, eh0->src_address, 6); + clib_memcpy (t->dst_mac, eh0->dst_address, 6); + t->ethertype = clib_net_to_host_u16 (eh0->type); + t->timestamp = now; + t->buffer_size = len0; + } + if (b1->flags & VLIB_BUFFER_IS_TRACED) + { + flowperpkt_l2_trace_t *t = + vlib_add_trace (vm, node, b1, sizeof (*t)); + t->rx_sw_if_index = vnet_buffer (b1)->sw_if_index[VLIB_RX]; + t->tx_sw_if_index = vnet_buffer (b1)->sw_if_index[VLIB_TX]; + clib_memcpy (t->src_mac, eh1->src_address, 6); + clib_memcpy (t->dst_mac, eh1->dst_address, 6); + t->ethertype = clib_net_to_host_u16 (eh1->type); + t->timestamp = now; + t->buffer_size = len1; + } + } + + /* verify speculative enqueues, maybe switch current next frame */ + vlib_validate_buffer_enqueue_x2 (vm, node, next_index, + to_next, n_left_to_next, + bi0, bi1, next0, next1); + } + + while (n_left_from > 0 && n_left_to_next > 0) + { + u32 bi0; + vlib_buffer_t *b0; + u32 next0 = FLOWPERPKT_L2_NEXT_DROP; + ethernet_header_t *eh0; + u16 len0; + + /* speculatively enqueue b0 to the current next frame */ + bi0 = from[0]; + to_next[0] = bi0; + from += 1; + to_next += 1; + n_left_from -= 1; + n_left_to_next -= 1; + + b0 = vlib_get_buffer (vm, bi0); + + vnet_feature_next (vnet_buffer (b0)->sw_if_index[VLIB_TX], + &next0, b0); + + eh0 = vlib_buffer_get_current (b0); + len0 = vlib_buffer_length_in_chain (vm, b0); + + if (PREDICT_TRUE ((b0->flags & VLIB_BUFFER_FLOW_REPORT) == 0)) + add_to_flow_record_l2 (vm, fm, + vnet_buffer (b0)->sw_if_index[VLIB_RX], + vnet_buffer (b0)->sw_if_index[VLIB_TX], + eh0->src_address, + eh0->dst_address, + eh0->type, now, len0, 0 /* flush */ ); + + if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) + && (b0->flags & VLIB_BUFFER_IS_TRACED))) + { + flowperpkt_l2_trace_t *t = + vlib_add_trace (vm, node, b0, sizeof (*t)); + t->rx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX]; + t->tx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_TX]; + clib_memcpy (t->src_mac, eh0->src_address, 6); + clib_memcpy (t->dst_mac, eh0->dst_address, 6); + t->ethertype = clib_net_to_host_u16 (eh0->type); + t->timestamp = now; + t->buffer_size = len0; + } + + /* verify speculative enqueue, maybe switch current next frame */ + vlib_validate_buffer_enqueue_x1 (vm, node, next_index, + to_next, n_left_to_next, + bi0, next0); + } + + vlib_put_next_frame (vm, node, next_index, n_left_to_next); + } + return frame->n_vectors; +} + +/** + * @brief IPFIX l2 flow-per-packet graph node + * @node flowperpkt-l2 + * + * This is the IPFIX flow-record-per-packet node. + * + * @param vm vlib_main_t corresponding to the current thread. + * @param node vlib_node_runtime_t data for this node. + * @param frame vlib_frame_t whose contents should be dispatched. + * + * @par Graph mechanics: buffer metadata, next index usage + * + * <em>Uses:</em> + * - <code>vnet_buffer(b)->ip.save_rewrite_length</code> + * - tells the node the length of the rewrite which was applied in + * ip4/6_rewrite_inline, allows the code to find the IP header without + * having to parse L2 headers, or make stupid assumptions about their + * length. + * - <code>vnet_buffer(b)->flags & VLIB_BUFFER_FLOW_REPORT</code> + * - Used to suppress flow record generation for flow record packets. + * + * <em>Sets:</em> + * - <code>vnet_buffer(b)->flags & VLIB_BUFFER_FLOW_REPORT</code> + * - To suppress flow record generation for flow record packets + * + * <em>Next Index:</em> + * - Next configured output feature on the interface, usually + * "interface-output." Generated flow records head for ip4-lookup + */ + +/* *INDENT-OFF* */ +VLIB_REGISTER_NODE (flowperpkt_l2_node) = { + .function = flowperpkt_l2_node_fn, + .name = "flowperpkt-l2", + .vector_size = sizeof (u32), + .format_trace = format_flowperpkt_l2_trace, + .type = VLIB_NODE_TYPE_INTERNAL, + + .n_errors = ARRAY_LEN(flowperpkt_l2_error_strings), + .error_strings = flowperpkt_l2_error_strings, + + .n_next_nodes = FLOWPERPKT_L2_N_NEXT, + + /* edit / add dispositions here */ + .next_nodes = { + [FLOWPERPKT_L2_NEXT_DROP] = "error-drop", + }, +}; +/* *INDENT-ON* */ + +/* + * fd.io coding-style-patch-verification: ON + * + * Local Variables: + * eval: (c-set-style "gnu") + * End: + */ diff --git a/plugins/flowperpkt-plugin/flowperpkt/node.c b/plugins/flowperpkt-plugin/flowperpkt/node.c index b1806fc4a82..460aa3be387 100644 --- a/plugins/flowperpkt-plugin/flowperpkt/node.c +++ b/plugins/flowperpkt-plugin/flowperpkt/node.c @@ -1,5 +1,5 @@ /* - * node.c - skeleton vpp engine plug-in dual-loop node skeleton + * node.c - ipv4 ipfix-per-packet graph node * * Copyright (c) <current-year> <your-organization> * Licensed under the Apache License, Version 2.0 (the "License"); @@ -21,7 +21,7 @@ #include <flowperpkt/flowperpkt.h> /** - * @file flow record generator graph node + * @file ipv4 flow record generator graph node */ typedef struct @@ -37,18 +37,18 @@ typedef struct u64 timestamp; /** size of the buffer */ u16 buffer_size; -} flowperpkt_trace_t; +} flowperpkt_ipv4_trace_t; /* packet trace format function */ static u8 * -format_flowperpkt_trace (u8 * s, va_list * args) +format_flowperpkt_ipv4_trace (u8 * s, va_list * args) { CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *); CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *); - flowperpkt_trace_t *t = va_arg (*args, flowperpkt_trace_t *); + flowperpkt_ipv4_trace_t *t = va_arg (*args, flowperpkt_ipv4_trace_t *); s = format (s, - "FLOWPERPKT: rx_sw_if_index %d, tx_sw_if_index %d, src %U dst %U tos %0x2, timestamp %lld, size %d", + "FLOWPERPKT-V4: rx_sw_if_index %d, tx_sw_if_index %d, src %U dst %U tos %0x2, timestamp %lld, size %d", t->rx_sw_if_index, t->tx_sw_if_index, format_ip4_address, &t->src_address, format_ip4_address, &t->dst_address, @@ -56,30 +56,30 @@ format_flowperpkt_trace (u8 * s, va_list * args) return s; } -vlib_node_registration_t flowperpkt_node; +vlib_node_registration_t flowperpkt_ipv4_node; -#define foreach_flowperpkt_error \ -_(SWAPPED, "Mac swap packets processed") +/* No counters at the moment */ +#define foreach_flowperpkt_ipv4_error typedef enum { #define _(sym,str) FLOWPERPKT_ERROR_##sym, - foreach_flowperpkt_error + foreach_flowperpkt_ipv4_error #undef _ FLOWPERPKT_N_ERROR, -} flowperpkt_error_t; +} flowperpkt_ipv4_error_t; -static char *flowperpkt_error_strings[] = { +static char *flowperpkt_ipv4_error_strings[] = { #define _(sym,string) string, - foreach_flowperpkt_error + foreach_flowperpkt_ipv4_error #undef _ }; typedef enum { - FLOWPERPKT_NEXT_DROP, - FLOWPERPKT_N_NEXT, -} flowperpkt_next_t; + FLOWPERPKT_IPV4_NEXT_DROP, + FLOWPERPKT_IPV4_N_NEXT, +} flowperpkt_ipv4_next_t; /** * @brief add an entry to the flow record under construction @@ -93,11 +93,11 @@ typedef enum */ static inline void -add_to_flow_record (vlib_main_t * vm, - flowperpkt_main_t * fm, - u32 rx_sw_if_index, u32 tx_sw_if_index, - u32 src_address, u32 dst_address, - u8 tos, u64 timestamp, u16 length, int do_flush) +add_to_flow_record_ipv4 (vlib_main_t * vm, + flowperpkt_main_t * fm, + u32 rx_sw_if_index, u32 tx_sw_if_index, + u32 src_address, u32 dst_address, + u8 tos, u64 timestamp, u16 length, int do_flush) { u32 my_cpu_number = vm->cpu_index; flow_report_main_t *frm = &flow_report_main; @@ -113,7 +113,7 @@ add_to_flow_record (vlib_main_t * vm, vlib_buffer_free_list_t *fl; /* Find or allocate a buffer */ - b0 = fm->buffers_per_worker[my_cpu_number]; + b0 = fm->ipv4_buffers_per_worker[my_cpu_number]; /* Need to allocate a buffer? */ if (PREDICT_FALSE (b0 == 0)) @@ -127,7 +127,8 @@ add_to_flow_record (vlib_main_t * vm, return; /* Initialize the buffer */ - b0 = fm->buffers_per_worker[my_cpu_number] = vlib_get_buffer (vm, bi0); + b0 = fm->ipv4_buffers_per_worker[my_cpu_number] = + vlib_get_buffer (vm, bi0); fl = vlib_buffer_get_free_list (vm, VLIB_BUFFER_DEFAULT_FREE_LIST_INDEX); vlib_buffer_init_for_free_list (b0, fl); @@ -138,16 +139,16 @@ add_to_flow_record (vlib_main_t * vm, { /* use the current buffer */ bi0 = vlib_get_buffer_index (vm, b0); - offset = fm->next_record_offset_per_worker[my_cpu_number]; + offset = fm->ipv4_next_record_offset_per_worker[my_cpu_number]; } /* Find or allocate a frame */ - f = fm->frames_per_worker[my_cpu_number]; + f = fm->ipv4_frames_per_worker[my_cpu_number]; if (PREDICT_FALSE (f == 0)) { u32 *to_next; f = vlib_get_frame_to_node (vm, ip4_lookup_node.index); - fm->frames_per_worker[my_cpu_number] = f; + fm->ipv4_frames_per_worker[my_cpu_number] = f; /* Enqueue the buffer */ to_next = vlib_frame_vector_args (f); @@ -250,8 +251,8 @@ add_to_flow_record (vlib_main_t * vm, } /* Time to flush the buffer? */ if (PREDICT_FALSE - (do_flush || (offset + sizeof (u32) + sizeof (u8) - + sizeof (f64)) > frm->path_mtu)) + (do_flush || (offset + 4 * sizeof (u32) + sizeof (u8) + + sizeof (f64) + sizeof (u16)) > frm->path_mtu)) { tp = vlib_buffer_get_current (b0); ip = (ip4_header_t *) & tp->ip4; @@ -259,7 +260,7 @@ add_to_flow_record (vlib_main_t * vm, h = (ipfix_message_header_t *) (udp + 1); s = (ipfix_set_header_t *) (h + 1); - s->set_id_length = ipfix_set_id_length (256, + s->set_id_length = ipfix_set_id_length (fm->ipv4_report_id, b0->current_length - (sizeof (*ip) + sizeof (*udp) + sizeof (*h))); @@ -282,38 +283,38 @@ add_to_flow_record (vlib_main_t * vm, ASSERT (ip->checksum == ip4_header_checksum (ip)); vlib_put_frame_to_node (vm, ip4_lookup_node.index, - fm->frames_per_worker[my_cpu_number]); - fm->frames_per_worker[my_cpu_number] = 0; - fm->buffers_per_worker[my_cpu_number] = 0; + fm->ipv4_frames_per_worker[my_cpu_number]); + fm->ipv4_frames_per_worker[my_cpu_number] = 0; + fm->ipv4_buffers_per_worker[my_cpu_number] = 0; offset = 0; } - fm->next_record_offset_per_worker[my_cpu_number] = offset; + fm->ipv4_next_record_offset_per_worker[my_cpu_number] = offset; } void -flowperpkt_flush_callback (void) +flowperpkt_flush_callback_ipv4 (void) { vlib_main_t *vm = vlib_get_main (); flowperpkt_main_t *fm = &flowperpkt_main; - add_to_flow_record (vm, fm, 0 /* rx_sw_if_index */ , - 0 /* tx_sw_if_index */ , - 0 /* src_address */ , - 0 /* dst_address */ , - 0 /* ToS */ , - 0ULL /* timestamp */ , - 0 /* length */ , - 1 /* do_flush */ ); + add_to_flow_record_ipv4 (vm, fm, 0 /* rx_sw_if_index */ , + 0 /* tx_sw_if_index */ , + 0 /* src_address */ , + 0 /* dst_address */ , + 0 /* ToS */ , + 0ULL /* timestamp */ , + 0 /* length */ , + 1 /* do_flush */ ); } static uword -flowperpkt_node_fn (vlib_main_t * vm, - vlib_node_runtime_t * node, vlib_frame_t * frame) +flowperpkt_ipv4_node_fn (vlib_main_t * vm, + vlib_node_runtime_t * node, vlib_frame_t * frame) { u32 n_left_from, *from, *to_next; - flowperpkt_next_t next_index; + flowperpkt_ipv4_next_t next_index; flowperpkt_main_t *fm = &flowperpkt_main; u64 now; @@ -332,8 +333,8 @@ flowperpkt_node_fn (vlib_main_t * vm, while (n_left_from >= 4 && n_left_to_next >= 2) { - u32 next0 = FLOWPERPKT_NEXT_DROP; - u32 next1 = FLOWPERPKT_NEXT_DROP; + u32 next0 = FLOWPERPKT_IPV4_NEXT_DROP; + u32 next1 = FLOWPERPKT_IPV4_NEXT_DROP; ip4_header_t *ip0, *ip1; u16 len0, len1; u32 bi0, bi1; @@ -375,30 +376,30 @@ flowperpkt_node_fn (vlib_main_t * vm, len0 = vlib_buffer_length_in_chain (vm, b0); if (PREDICT_TRUE ((b0->flags & VLIB_BUFFER_FLOW_REPORT) == 0)) - add_to_flow_record (vm, fm, - vnet_buffer (b0)->sw_if_index[VLIB_RX], - vnet_buffer (b0)->sw_if_index[VLIB_TX], - ip0->src_address.as_u32, - ip0->dst_address.as_u32, - ip0->tos, now, len0, 0 /* flush */ ); + add_to_flow_record_ipv4 (vm, fm, + vnet_buffer (b0)->sw_if_index[VLIB_RX], + vnet_buffer (b0)->sw_if_index[VLIB_TX], + ip0->src_address.as_u32, + ip0->dst_address.as_u32, + ip0->tos, now, len0, 0 /* flush */ ); ip1 = (ip4_header_t *) ((u8 *) vlib_buffer_get_current (b1) + vnet_buffer (b1)->ip.save_rewrite_length); len1 = vlib_buffer_length_in_chain (vm, b1); if (PREDICT_TRUE ((b1->flags & VLIB_BUFFER_FLOW_REPORT) == 0)) - add_to_flow_record (vm, fm, - vnet_buffer (b1)->sw_if_index[VLIB_RX], - vnet_buffer (b1)->sw_if_index[VLIB_TX], - ip1->src_address.as_u32, - ip1->dst_address.as_u32, - ip1->tos, now, len1, 0 /* flush */ ); + add_to_flow_record_ipv4 (vm, fm, + vnet_buffer (b1)->sw_if_index[VLIB_RX], + vnet_buffer (b1)->sw_if_index[VLIB_TX], + ip1->src_address.as_u32, + ip1->dst_address.as_u32, + ip1->tos, now, len1, 0 /* flush */ ); if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE))) { if (b0->flags & VLIB_BUFFER_IS_TRACED) { - flowperpkt_trace_t *t = + flowperpkt_ipv4_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->rx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX]; t->tx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_TX]; @@ -410,7 +411,7 @@ flowperpkt_node_fn (vlib_main_t * vm, } if (b1->flags & VLIB_BUFFER_IS_TRACED) { - flowperpkt_trace_t *t = + flowperpkt_ipv4_trace_t *t = vlib_add_trace (vm, node, b1, sizeof (*t)); t->rx_sw_if_index = vnet_buffer (b1)->sw_if_index[VLIB_RX]; t->tx_sw_if_index = vnet_buffer (b1)->sw_if_index[VLIB_TX]; @@ -432,7 +433,7 @@ flowperpkt_node_fn (vlib_main_t * vm, { u32 bi0; vlib_buffer_t *b0; - u32 next0 = FLOWPERPKT_NEXT_DROP; + u32 next0 = FLOWPERPKT_IPV4_NEXT_DROP; ip4_header_t *ip0; u16 len0; @@ -461,17 +462,17 @@ flowperpkt_node_fn (vlib_main_t * vm, len0 = vlib_buffer_length_in_chain (vm, b0); if (PREDICT_TRUE ((b0->flags & VLIB_BUFFER_FLOW_REPORT) == 0)) - add_to_flow_record (vm, fm, - vnet_buffer (b0)->sw_if_index[VLIB_RX], - vnet_buffer (b0)->sw_if_index[VLIB_TX], - ip0->src_address.as_u32, - ip0->dst_address.as_u32, - ip0->tos, now, len0, 0 /* flush */ ); + add_to_flow_record_ipv4 (vm, fm, + vnet_buffer (b0)->sw_if_index[VLIB_RX], + vnet_buffer (b0)->sw_if_index[VLIB_TX], + ip0->src_address.as_u32, + ip0->dst_address.as_u32, + ip0->tos, now, len0, 0 /* flush */ ); if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE) && (b0->flags & VLIB_BUFFER_IS_TRACED))) { - flowperpkt_trace_t *t = + flowperpkt_ipv4_trace_t *t = vlib_add_trace (vm, node, b0, sizeof (*t)); t->rx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX]; t->tx_sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_TX]; @@ -494,8 +495,8 @@ flowperpkt_node_fn (vlib_main_t * vm, } /** - * @brief IPFIX flow-per-packet graph node - * @node flowperpkt + * @brief IPFIX ipv4 flow-per-packet graph node + * @node flowperpkt-ipv4 * * This is the IPFIX flow-record-per-packet node. * @@ -524,21 +525,21 @@ flowperpkt_node_fn (vlib_main_t * vm, */ /* *INDENT-OFF* */ -VLIB_REGISTER_NODE (flowperpkt_node) = { - .function = flowperpkt_node_fn, - .name = "flowperpkt", +VLIB_REGISTER_NODE (flowperpkt_ipv4_node) = { + .function = flowperpkt_ipv4_node_fn, + .name = "flowperpkt-ipv4", .vector_size = sizeof (u32), - .format_trace = format_flowperpkt_trace, + .format_trace = format_flowperpkt_ipv4_trace, .type = VLIB_NODE_TYPE_INTERNAL, - .n_errors = ARRAY_LEN(flowperpkt_error_strings), - .error_strings = flowperpkt_error_strings, + .n_errors = ARRAY_LEN(flowperpkt_ipv4_error_strings), + .error_strings = flowperpkt_ipv4_error_strings, - .n_next_nodes = FLOWPERPKT_N_NEXT, + .n_next_nodes = FLOWPERPKT_IPV4_N_NEXT, /* edit / add dispositions here */ .next_nodes = { - [FLOWPERPKT_NEXT_DROP] = "error-drop", + [FLOWPERPKT_IPV4_NEXT_DROP] = "error-drop", }, }; /* *INDENT-ON* */ |