aboutsummaryrefslogtreecommitdiffstats
path: root/src
diff options
context:
space:
mode:
authorKeith Burns (alagalah) <alagalah@gmail.com>2017-08-06 08:26:29 -0700
committerDave Wallace <dwallacelf@gmail.com>2017-09-07 22:12:31 +0000
commit8a19f12a0cfe6d611f6e266931af691fb69a74ad (patch)
treebc47ff7146c235c1df17594f4529929cf25ab8b8 /src
parent49d66f1f42cbc310e4fa0dc526b9fdb91d0ca220 (diff)
Allow individual stats API and introduce stats.api
- want_interface_simple_stats - want_interface_combined_stats - want_ip4|6_fib|nbr_stats Change-Id: I4e97461def508958b3e429c3fe8859b36fef2d18 Signed-off-by: Keith Burns (alagalah) <alagalah@gmail.com>
Diffstat (limited to 'src')
-rw-r--r--src/uri.am4
-rw-r--r--src/vlibapi/api_helper_macros.h1
-rw-r--r--src/vpp.am12
-rw-r--r--src/vpp/api/api.c59
-rw-r--r--src/vpp/api/vpe.api112
-rw-r--r--src/vpp/api/vpe_all_api_h.h3
-rw-r--r--src/vpp/stats/stats.api222
-rw-r--r--src/vpp/stats/stats.c1265
-rw-r--r--src/vpp/stats/stats.h45
9 files changed, 1131 insertions, 592 deletions
diff --git a/src/uri.am b/src/uri.am
index ae6feb6f35b..660f897dd71 100644
--- a/src/uri.am
+++ b/src/uri.am
@@ -1,4 +1,4 @@
-# Copyright (c) 2016 Cisco and/or its affiliates.
+# Copyright (c) 2017 Cisco and/or its affiliates.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at:
@@ -58,7 +58,7 @@ vcl_test_server_SOURCES = uri/vcl_test_server.c
vcl_test_server_LDADD = libvppcom.la
vcl_test_client_SOURCES = uri/vcl_test_client.c
-vcl_test_client_LDADD = libvppcom.la
+vcl_test_client_LDADD = libvppcom.la
sock_test_server_SOURCES = uri/sock_test_server.c
sock_test_client_SOURCES = uri/sock_test_client.c
diff --git a/src/vlibapi/api_helper_macros.h b/src/vlibapi/api_helper_macros.h
index 2e29d4d7983..b8a9978a0a0 100644
--- a/src/vlibapi/api_helper_macros.h
+++ b/src/vlibapi/api_helper_macros.h
@@ -217,7 +217,6 @@ _(from_netconf_client) \
_(oam_events) \
_(bfd_events)
-/* WARNING: replicated in vpp/stats.h */
typedef struct
{
u32 client_index; /* in memclnt registration pool */
diff --git a/src/vpp.am b/src/vpp.am
index 10a4e311252..9a07cefe2c0 100644
--- a/src/vpp.am
+++ b/src/vpp.am
@@ -42,9 +42,11 @@ bin_vpp_CFLAGS = @APICLI@
nobase_include_HEADERS += \
vpp/api/vpe_all_api_h.h \
vpp/api/vpe_msg_enum.h \
+ vpp/stats/stats.api.h \
vpp/api/vpe.api.h
API_FILES += vpp/api/vpe.api
+API_FILES += vpp/stats/stats.api
BUILT_SOURCES += .version
@@ -116,13 +118,13 @@ endif
noinst_PROGRAMS += bin/summary_stats_client
bin_summary_stats_client_SOURCES = \
- vpp/api/summary_stats_client.c
+ vpp/api/summary_stats_client.c
bin_summary_stats_client_LDADD = \
- libvlibmemoryclient.la \
- libsvm.la \
- libvppinfra.la \
- -lpthread -lm -lrt
+ libvlibmemoryclient.la \
+ libsvm.la \
+ libvppinfra.la \
+ -lpthread -lm -lrt
bin_PROGRAMS += bin/vpp_get_metrics
diff --git a/src/vpp/api/api.c b/src/vpp/api/api.c
index e8bc22ae563..f9c3129cc78 100644
--- a/src/vpp/api/api.c
+++ b/src/vpp/api/api.c
@@ -110,7 +110,6 @@ _(CREATE_VLAN_SUBIF, create_vlan_subif) \
_(CREATE_SUBIF, create_subif) \
_(PROXY_ARP_ADD_DEL, proxy_arp_add_del) \
_(PROXY_ARP_INTFC_ENABLE_DISABLE, proxy_arp_intfc_enable_disable) \
-_(VNET_GET_SUMMARY_STATS, vnet_get_summary_stats) \
_(RESET_FIB, reset_fib) \
_(CREATE_LOOPBACK, create_loopback) \
_(CREATE_LOOPBACK_INSTANCE, create_loopback_instance) \
@@ -745,64 +744,6 @@ vl_api_oam_add_del_t_handler (vl_api_oam_add_del_t * mp)
REPLY_MACRO (VL_API_OAM_ADD_DEL_REPLY);
}
-static void
-vl_api_vnet_get_summary_stats_t_handler (vl_api_vnet_get_summary_stats_t * mp)
-{
- stats_main_t *sm = &stats_main;
- vnet_interface_main_t *im = sm->interface_main;
- vl_api_vnet_get_summary_stats_reply_t *rmp;
- vlib_combined_counter_main_t *cm;
- vlib_counter_t v;
- int i, which;
- u64 total_pkts[VLIB_N_RX_TX];
- u64 total_bytes[VLIB_N_RX_TX];
-
- unix_shared_memory_queue_t *q =
- vl_api_client_index_to_input_queue (mp->client_index);
-
- if (!q)
- return;
-
- rmp = vl_msg_api_alloc (sizeof (*rmp));
- rmp->_vl_msg_id = ntohs (VL_API_VNET_GET_SUMMARY_STATS_REPLY);
- rmp->context = mp->context;
- rmp->retval = 0;
-
- memset (total_pkts, 0, sizeof (total_pkts));
- memset (total_bytes, 0, sizeof (total_bytes));
-
- vnet_interface_counter_lock (im);
-
- vec_foreach (cm, im->combined_sw_if_counters)
- {
- which = cm - im->combined_sw_if_counters;
-
- for (i = 0; i < vlib_combined_counter_n_counters (cm); i++)
- {
- vlib_get_combined_counter (cm, i, &v);
- total_pkts[which] += v.packets;
- total_bytes[which] += v.bytes;
- }
- }
- vnet_interface_counter_unlock (im);
-
- rmp->total_pkts[VLIB_RX] = clib_host_to_net_u64 (total_pkts[VLIB_RX]);
- rmp->total_bytes[VLIB_RX] = clib_host_to_net_u64 (total_bytes[VLIB_RX]);
- rmp->total_pkts[VLIB_TX] = clib_host_to_net_u64 (total_pkts[VLIB_TX]);
- rmp->total_bytes[VLIB_TX] = clib_host_to_net_u64 (total_bytes[VLIB_TX]);
- rmp->vector_rate =
- clib_host_to_net_u64 (vlib_last_vector_length_per_node (sm->vlib_main));
-
- vl_msg_api_send_shmem (q, (u8 *) & rmp);
-}
-
-/* *INDENT-OFF* */
-typedef CLIB_PACKED (struct {
- ip4_address_t address;
- u32 address_length: 6;
- u32 index:26;
-}) ip4_route_t;
-/* *INDENT-ON* */
static int
ip4_reset_fib_t_handler (vl_api_reset_fib_t * mp)
diff --git a/src/vpp/api/vpe.api b/src/vpp/api/vpe.api
index eda95ce57f7..d68beae13d3 100644
--- a/src/vpp/api/vpe.api
+++ b/src/vpp/api/vpe.api
@@ -47,6 +47,7 @@
* DHCP APIs: see ... /src/vnet/dhcp/{dhcpk.api, dhcp_api.c}
* COP APIs: see ... /src/vnet/cop/{cop.api, cop_api.c}
* POLICER APIs: see ... /src/vnet/policer/{policer.api, policer_api.c}
+ * STATS APIs: see .../src/vpp/stats/{stats.api, stats.c}
*/
/** \brief Create a new subinterface with the given vlan id
@@ -136,117 +137,6 @@ autoreply define reset_vrf
u32 vrf_id;
};
-/** \brief Want Stats, register for stats updates
- @param client_index - opaque cookie to identify the sender
- @param context - sender context, to match reply w/ request
- @param enable_disable - 1 = enable stats, 0 = disable
- @param pid - pid of process requesting stats updates
-*/
-autoreply define want_stats
-{
- u32 client_index;
- u32 context;
- u32 enable_disable;
- u32 pid;
-};
-
-typeonly manual_print manual_endian define ip4_fib_counter
-{
- u32 address;
- u8 address_length;
- u64 packets;
- u64 bytes;
-};
-
-manual_print manual_endian define vnet_ip4_fib_counters
-{
- u32 vrf_id;
- u32 count;
- vl_api_ip4_fib_counter_t c[count];
-};
-
-typeonly manual_print manual_endian define ip4_nbr_counter
-{
- u32 address;
- u8 link_type;
- u64 packets;
- u64 bytes;
-};
-
-/**
- * @brief Per-neighbour (i.e. per-adjacency) coutners
- * @param count The size of the array of counters
- * @param sw_if_index The interface the adjacency is on
- * @param begin Flag to indicate this is the first set of stats for this
- * interface. If this flag is not set the it is a continuation of
- * stats for this interface
- * @param c counters
- */
-manual_print manual_endian define vnet_ip4_nbr_counters
-{
- u32 count;
- u32 sw_if_index;
- u8 begin;
- vl_api_ip4_nbr_counter_t c[count];
-};
-
-typeonly manual_print manual_endian define ip6_fib_counter
-{
- u64 address[2];
- u8 address_length;
- u64 packets;
- u64 bytes;
-};
-
-manual_print manual_endian define vnet_ip6_fib_counters
-{
- u32 vrf_id;
- u32 count;
- vl_api_ip6_fib_counter_t c[count];
-};
-
-typeonly manual_print manual_endian define ip6_nbr_counter
-{
- u64 address[2];
- u8 link_type;
- u64 packets;
- u64 bytes;
-};
-
-manual_print manual_endian define vnet_ip6_nbr_counters
-{
- u32 count;
- u32 sw_if_index;
- u8 begin;
- vl_api_ip6_nbr_counter_t c[count];
-};
-
-/** \brief Request for a single block of summary stats
- @param client_index - opaque cookie to identify the sender
- @param context - sender context, to match reply w/ request
-*/
-define vnet_get_summary_stats
-{
- u32 client_index;
- u32 context;
-};
-
-/** \brief Reply for vnet_get_summary_stats request
- @param context - sender context, to match reply w/ request
- @param retval - return code for request
- @param total_pkts -
- @param total_bytes -
- @param vector_rate -
-*/
-define vnet_get_summary_stats_reply
-{
- u32 context;
- i32 retval;
- u64 total_pkts[2];
- u64 total_bytes[2];
- f64 vector_rate;
-};
-
/** \brief OAM event structure
@param dst_address[] -
@param state
diff --git a/src/vpp/api/vpe_all_api_h.h b/src/vpp/api/vpe_all_api_h.h
index 397cd8074b3..d35a0535c72 100644
--- a/src/vpp/api/vpe_all_api_h.h
+++ b/src/vpp/api/vpe_all_api_h.h
@@ -28,6 +28,9 @@
/* Include the current layer (third) vpp API definition layer */
#include <vpp/api/vpe.api.h>
+/* Include stats APIs */
+#include <vpp/stats/stats.api.h>
+
/*
* fd.io coding-style-patch-verification: ON
*
diff --git a/src/vpp/stats/stats.api b/src/vpp/stats/stats.api
new file mode 100644
index 00000000000..7f745859610
--- /dev/null
+++ b/src/vpp/stats/stats.api
@@ -0,0 +1,222 @@
+/*
+ * Copyright (c) 2015-2016 Cisco and/or its affiliates.
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at:
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+/** \file
+
+ This file defines the stats API
+*/
+
+
+/** \brief Want Stats, enable/disable ALL stats updates
+ @param client_index - opaque cookie to identify the sender
+ @param context - sender context, to match reply w/ request
+ @param enable_disable - 1 = enable stats, 0 = disable
+ @param pid - pid of process requesting stats updates
+*/
+autoreply define want_stats
+{
+ u32 client_index;
+ u32 context;
+ u32 enable_disable;
+ u32 pid;
+};
+
+/** \brief Want Interface Simple Stats, register for detailed interface stats
+ @param client_index - opaque cookie to identify the sender
+ @param context - sender context, to match reply w/ request
+ @param enable_disable - 1 = enable stats, 0 = disable
+ @param pid - pid of process requesting stats updates
+*/
+autoreply define want_interface_simple_stats
+{
+ u32 client_index;
+ u32 context;
+ u32 enable_disable;
+ u32 pid;
+};
+
+/** \brief Want Interface Combined Stats, register for continuous stats
+ @param client_index - opaque cookie to identify the sender
+ @param context - sender context, to match reply w/ request
+ @param enable_disable - 1 = enable stats, 0 = disable
+ @param pid - pid of process requesting stats updates
+*/
+autoreply define want_interface_combined_stats
+{
+ u32 client_index;
+ u32 context;
+ u32 enable_disable;
+ u32 pid;
+};
+
+/** \brief Want IP4 FIB Stats, register for continuous stats
+ @param client_index - opaque cookie to identify the sender
+ @param context - sender context, to match reply w/ request
+ @param enable_disable - 1 = enable stats, 0 = disable
+ @param pid - pid of process requesting stats updates
+*/
+autoreply define want_ip4_fib_stats
+{
+ u32 client_index;
+ u32 context;
+ u32 enable_disable;
+ u32 pid;
+};
+
+/** \brief Want IP6 FIB Stats, register for continuous stats
+ @param client_index - opaque cookie to identify the sender
+ @param context - sender context, to match reply w/ request
+ @param enable_disable - 1 = enable stats, 0 = disable
+ @param pid - pid of process requesting stats updates
+*/
+autoreply define want_ip6_fib_stats
+{
+ u32 client_index;
+ u32 context;
+ u32 enable_disable;
+ u32 pid;
+};
+
+/** \brief Want IP4 NBR Stats, register for continuous stats
+ @param client_index - opaque cookie to identify the sender
+ @param context - sender context, to match reply w/ request
+ @param enable_disable - 1 = enable stats, 0 = disable
+ @param pid - pid of process requesting stats updates
+*/
+autoreply define want_ip4_nbr_stats
+{
+ u32 client_index;
+ u32 context;
+ u32 enable_disable;
+ u32 pid;
+};
+
+/** \brief Want IP6 NBR Stats, register for continuous stats
+ @param client_index - opaque cookie to identify the sender
+ @param context - sender context, to match reply w/ request
+ @param enable_disable - 1 = enable stats, 0 = disable
+ @param pid - pid of process requesting stats updates
+*/
+autoreply define want_ip6_nbr_stats
+{
+ u32 client_index;
+ u32 context;
+ u32 enable_disable;
+ u32 pid;
+};
+
+typeonly manual_print manual_endian define ip4_fib_counter
+{
+ u32 address;
+ u8 address_length;
+ u64 packets;
+ u64 bytes;
+};
+
+manual_print manual_endian define vnet_ip4_fib_counters
+{
+ u32 vrf_id;
+ u32 count;
+ vl_api_ip4_fib_counter_t c[count];
+};
+
+typeonly manual_print manual_endian define ip4_nbr_counter
+{
+ u32 address;
+ u8 link_type;
+ u64 packets;
+ u64 bytes;
+};
+
+/**
+ * @brief Per-neighbour (i.e. per-adjacency) coutners
+ * @param count The size of the array of counters
+ * @param sw_if_index The interface the adjacency is on
+ * @param begin Flag to indicate this is the first set of stats for this
+ * interface. If this flag is not set the it is a continuation of
+ * stats for this interface
+ * @param c counters
+ */
+manual_print manual_endian define vnet_ip4_nbr_counters
+{
+ u32 count;
+ u32 sw_if_index;
+ u8 begin;
+ vl_api_ip4_nbr_counter_t c[count];
+};
+
+typeonly manual_print manual_endian define ip6_fib_counter
+{
+ u64 address[2];
+ u8 address_length;
+ u64 packets;
+ u64 bytes;
+};
+
+manual_print manual_endian define vnet_ip6_fib_counters
+{
+ u32 vrf_id;
+ u32 count;
+ vl_api_ip6_fib_counter_t c[count];
+};
+
+typeonly manual_print manual_endian define ip6_nbr_counter
+{
+ u64 address[2];
+ u8 link_type;
+ u64 packets;
+ u64 bytes;
+};
+
+manual_print manual_endian define vnet_ip6_nbr_counters
+{
+ u32 count;
+ u32 sw_if_index;
+ u8 begin;
+ vl_api_ip6_nbr_counter_t c[count];
+};
+
+
+/** \brief Request for a single block of summary stats
+ @param client_index - opaque cookie to identify the sender
+ @param context - sender context, to match reply w/ request
+*/
+define vnet_get_summary_stats
+{
+ u32 client_index;
+ u32 context;
+};
+
+/** \brief Reply for vnet_get_summary_stats request
+ @param context - sender context, to match reply w/ request
+ @param retval - return code for request
+ @param total_pkts -
+ @param total_bytes -
+ @param vector_rate -
+*/
+define vnet_get_summary_stats_reply
+{
+ u32 context;
+ i32 retval;
+ u64 total_pkts[2];
+ u64 total_bytes[2];
+ f64 vector_rate;
+};
+
+/*
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
diff --git a/src/vpp/stats/stats.c b/src/vpp/stats/stats.c
index b0fac73dbe7..c78a8aef3cd 100644
--- a/src/vpp/stats/stats.c
+++ b/src/vpp/stats/stats.c
@@ -14,10 +14,8 @@
*/
#include <vpp/stats/stats.h>
#include <signal.h>
-#include <vlib/threads.h>
-#include <vnet/fib/fib_entry.h>
-#include <vnet/fib/fib_table.h>
#include <vnet/fib/ip4_fib.h>
+#include <vnet/fib/fib_entry.h>
#include <vnet/dpo/load_balance.h>
#define STATS_DEBUG 0
@@ -48,11 +46,19 @@ stats_main_t stats_main;
#define foreach_stats_msg \
_(WANT_STATS, want_stats) \
_(VNET_INTERFACE_SIMPLE_COUNTERS, vnet_interface_simple_counters) \
+_(WANT_INTERFACE_SIMPLE_STATS, want_interface_simple_stats) \
_(VNET_INTERFACE_COMBINED_COUNTERS, vnet_interface_combined_counters) \
+_(WANT_INTERFACE_COMBINED_STATS, want_interface_combined_stats) \
_(VNET_IP4_FIB_COUNTERS, vnet_ip4_fib_counters) \
+_(WANT_IP4_FIB_STATS, want_ip4_fib_stats) \
_(VNET_IP6_FIB_COUNTERS, vnet_ip6_fib_counters) \
+_(WANT_IP6_FIB_STATS, want_ip6_fib_stats) \
_(VNET_IP4_NBR_COUNTERS, vnet_ip4_nbr_counters) \
-_(VNET_IP6_NBR_COUNTERS, vnet_ip6_nbr_counters)
+_(WANT_IP4_NBR_STATS, want_ip4_nbr_stats) \
+_(VNET_IP6_NBR_COUNTERS, vnet_ip6_nbr_counters) \
+_(WANT_IP6_NBR_STATS, want_ip6_nbr_stats) \
+_(VNET_GET_SUMMARY_STATS, vnet_get_summary_stats)
+
/* These constants ensure msg sizes <= 1024, aka ring allocation */
#define SIMPLE_COUNTER_BATCH_SIZE 126
@@ -64,6 +70,110 @@ _(VNET_IP6_NBR_COUNTERS, vnet_ip6_nbr_counters)
#define STATS_RELEASE_DELAY_NS (1000 * 1000 * 5)
/* ns/us us/ms */
+u8 *
+format_vnet_interface_combined_counters (u8 * s, va_list * args)
+{
+ stats_main_t *sm = &stats_main;
+ vl_api_vnet_interface_combined_counters_t *mp =
+ va_arg (*args, vl_api_vnet_interface_combined_counters_t *);
+
+ char *counter_name;
+ u32 count, sw_if_index;
+ int i;
+ count = ntohl (mp->count);
+ sw_if_index = ntohl (mp->first_sw_if_index);
+
+ vlib_counter_t *vp;
+ u64 packets, bytes;
+ vp = (vlib_counter_t *) mp->data;
+
+ switch (mp->vnet_counter_type)
+ {
+ case VNET_INTERFACE_COUNTER_RX:
+ counter_name = "rx";
+ break;
+ case VNET_INTERFACE_COUNTER_TX:
+ counter_name = "tx";
+ break;
+ default:
+ counter_name = "bogus";
+ break;
+ }
+ for (i = 0; i < count; i++)
+ {
+ packets = clib_mem_unaligned (&vp->packets, u64);
+ packets = clib_net_to_host_u64 (packets);
+ bytes = clib_mem_unaligned (&vp->bytes, u64);
+ bytes = clib_net_to_host_u64 (bytes);
+ vp++;
+ s = format (s, "%U.%s.packets %lld\n",
+ format_vnet_sw_if_index_name,
+ sm->vnet_main, sw_if_index, counter_name, packets);
+ s = format (s, "%U.%s.bytes %lld\n",
+ format_vnet_sw_if_index_name,
+ sm->vnet_main, sw_if_index, counter_name, bytes);
+ sw_if_index++;
+ }
+ return s;
+}
+
+u8 *
+format_vnet_interface_simple_counters (u8 * s, va_list * args)
+{
+ stats_main_t *sm = &stats_main;
+ vl_api_vnet_interface_simple_counters_t *mp =
+ va_arg (*args, vl_api_vnet_interface_simple_counters_t *);
+ char *counter_name;
+ u32 count, sw_if_index;
+ count = ntohl (mp->count);
+ sw_if_index = ntohl (mp->first_sw_if_index);
+ u64 *vp, v;
+ vp = (u64 *) mp->data;
+ int i;
+
+ switch (mp->vnet_counter_type)
+ {
+ case VNET_INTERFACE_COUNTER_DROP:
+ counter_name = "drop";
+ break;
+ case VNET_INTERFACE_COUNTER_PUNT:
+ counter_name = "punt";
+ break;
+ case VNET_INTERFACE_COUNTER_IP4:
+ counter_name = "ip4";
+ break;
+ case VNET_INTERFACE_COUNTER_IP6:
+ counter_name = "ip6";
+ break;
+ case VNET_INTERFACE_COUNTER_RX_NO_BUF:
+ counter_name = "rx-no-buff";
+ break;
+ case VNET_INTERFACE_COUNTER_RX_MISS:
+ counter_name = "rx-miss";
+ break;
+ case VNET_INTERFACE_COUNTER_RX_ERROR:
+ counter_name = "rx-error (fifo-full)";
+ break;
+ case VNET_INTERFACE_COUNTER_TX_ERROR:
+ counter_name = "tx-error (fifo-full)";
+ break;
+ default:
+ counter_name = "bogus";
+ break;
+ }
+ for (i = 0; i < count; i++)
+ {
+ v = clib_mem_unaligned (vp, u64);
+ v = clib_net_to_host_u64 (v);
+ vp++;
+ s = format (s, "%U.%s %lld\n", format_vnet_sw_if_index_name,
+ sm->vnet_main, sw_if_index, counter_name, v);
+ sw_if_index++;
+ }
+
+ return s;
+}
+
void
dslock (stats_main_t * sm, int release_hint, int tag)
{
@@ -231,14 +341,6 @@ do_combined_interface_counters (stats_main_t * sm)
vnet_interface_counter_unlock (im);
}
-/* from .../vnet/vnet/ip/lookup.c. Yuck */
-typedef CLIB_PACKED (struct
- {
- ip4_address_t address;
-u32 address_length: 6;
-u32 index: 26;
- }) ip4_route_t;
-
static void
ip46_fib_stats_delay (stats_main_t * sm, u32 sec, u32 nsec)
{
@@ -573,154 +675,171 @@ do_ip4_fibs (stats_main_t * sm)
api_main_t *am = sm->api_main;
vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
unix_shared_memory_queue_t *q = shmem_hdr->vl_input_queue;
- static ip4_route_t *routes;
ip4_route_t *r;
fib_table_t *fib;
ip4_fib_t *v4_fib;
- static uword *results;
+ do_ip46_fibs_t *do_fibs;
vl_api_vnet_ip4_fib_counters_t *mp = 0;
u32 items_this_message;
vl_api_ip4_fib_counter_t *ctrp = 0;
u32 start_at_fib_index = 0;
- int i;
+ int i, j, k;
+
+ do_fibs = &sm->do_ip46_fibs;
again:
+ vec_reset_length (do_fibs->fibs);
/* *INDENT-OFF* */
pool_foreach (fib, im4->fibs,
- ({
- /* We may have bailed out due to control-plane activity */
- while ((fib - im4->fibs) < start_at_fib_index)
- continue;
+ ({vec_add1(do_fibs->fibs,fib);}));
- v4_fib = pool_elt_at_index (im4->v4_fibs, fib->ft_index);
+ /* *INDENT-ON* */
- if (mp == 0)
- {
- items_this_message = IP4_FIB_COUNTER_BATCH_SIZE;
- mp = vl_msg_api_alloc_as_if_client
- (sizeof (*mp) +
- items_this_message * sizeof (vl_api_ip4_fib_counter_t));
- mp->_vl_msg_id = ntohs (VL_API_VNET_IP4_FIB_COUNTERS);
- mp->count = 0;
- mp->vrf_id = ntohl (fib->ft_table_id);
- ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
- }
- else
- {
- /* happens if the last FIB was empty... */
- ASSERT (mp->count == 0);
- mp->vrf_id = ntohl (fib->ft_table_id);
- }
+ for (j = 0; j < vec_len (do_fibs->fibs); j++)
+ {
+ fib = do_fibs->fibs[j];
+ /* We may have bailed out due to control-plane activity */
+ while ((fib - im4->fibs) < start_at_fib_index)
+ continue;
- dslock (sm, 0 /* release hint */ , 1 /* tag */ );
+ v4_fib = pool_elt_at_index (im4->v4_fibs, fib->ft_index);
- vec_reset_length (routes);
- vec_reset_length (results);
+ if (mp == 0)
+ {
+ items_this_message = IP4_FIB_COUNTER_BATCH_SIZE;
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) +
+ items_this_message * sizeof (vl_api_ip4_fib_counter_t));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_IP4_FIB_COUNTERS);
+ mp->count = 0;
+ mp->vrf_id = ntohl (fib->ft_table_id);
+ ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
+ }
+ else
+ {
+ /* happens if the last FIB was empty... */
+ ASSERT (mp->count == 0);
+ mp->vrf_id = ntohl (fib->ft_table_id);
+ }
- for (i = 0; i < ARRAY_LEN (v4_fib->fib_entry_by_dst_address); i++)
- {
- uword *hash = v4_fib->fib_entry_by_dst_address[i];
- hash_pair_t *p;
- ip4_route_t x;
-
- x.address_length = i;
-
- hash_foreach_pair (p, hash,
- ({
- x.address.data_u32 = p->key;
- x.index = p->value[0];
-
- vec_add1 (routes, x);
- if (sm->data_structure_lock->release_hint)
- {
- start_at_fib_index = fib - im4->fibs;
- dsunlock (sm);
- ip46_fib_stats_delay (sm, 0 /* sec */,
- STATS_RELEASE_DELAY_NS);
- mp->count = 0;
- ctrp = (vl_api_ip4_fib_counter_t *)mp->c;
- goto again;
- }
- }));
- }
+ dslock (sm, 0 /* release hint */ , 1 /* tag */ );
- vec_foreach (r, routes)
- {
- vlib_counter_t c;
- const dpo_id_t *dpo_id;
-
- dpo_id = fib_entry_contribute_ip_forwarding(r->index);
- vlib_get_combined_counter (&load_balance_main.lbm_to_counters,
- (u32)dpo_id->dpoi_index, &c);
- /*
- * If it has actually
- * seen at least one packet, send it.
- */
- if (c.packets > 0)
- {
+ vec_reset_length (do_fibs->ip4routes);
+ vec_reset_length (do_fibs->results);
- /* already in net byte order */
- ctrp->address = r->address.as_u32;
- ctrp->address_length = r->address_length;
- ctrp->packets = clib_host_to_net_u64 (c.packets);
- ctrp->bytes = clib_host_to_net_u64 (c.bytes);
- mp->count++;
- ctrp++;
+ for (i = 0; i < ARRAY_LEN (v4_fib->fib_entry_by_dst_address); i++)
+ {
+ uword *hash = v4_fib->fib_entry_by_dst_address[i];
+ hash_pair_t *p;
+ ip4_route_t x;
+
+ vec_reset_length (do_fibs->pvec);
+
+ x.address_length = i;
+
+ hash_foreach_pair (p, hash, (
+ {
+ vec_add1 (do_fibs->pvec, p);}
+ ));
+ for (k = 0; k < vec_len (do_fibs->pvec); k++)
+ {
+ p = do_fibs->pvec[k];
+ x.address.data_u32 = p->key;
+ x.index = p->value[0];
+
+ vec_add1 (do_fibs->ip4routes, x);
+ if (sm->data_structure_lock->release_hint)
+ {
+ start_at_fib_index = fib - im4->fibs;
+ dsunlock (sm);
+ ip46_fib_stats_delay (sm, 0 /* sec */ ,
+ STATS_RELEASE_DELAY_NS);
+ mp->count = 0;
+ ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
+ goto again;
+ }
+ }
+ }
- if (mp->count == items_this_message)
- {
- mp->count = htonl (items_this_message);
- /*
- * If the main thread's input queue is stuffed,
- * drop the data structure lock (which the main thread
- * may want), and take a pause.
- */
- unix_shared_memory_queue_lock (q);
- if (unix_shared_memory_queue_is_full (q))
- {
- dsunlock (sm);
- vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
- unix_shared_memory_queue_unlock (q);
- mp = 0;
- ip46_fib_stats_delay (sm, 0 /* sec */ ,
- STATS_RELEASE_DELAY_NS);
- goto again;
- }
- vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
- unix_shared_memory_queue_unlock (q);
-
- items_this_message = IP4_FIB_COUNTER_BATCH_SIZE;
- mp = vl_msg_api_alloc_as_if_client
- (sizeof (*mp) +
- items_this_message * sizeof (vl_api_ip4_fib_counter_t));
- mp->_vl_msg_id = ntohs (VL_API_VNET_IP4_FIB_COUNTERS);
- mp->count = 0;
- mp->vrf_id = ntohl (fib->ft_table_id);
- ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
- }
- } /* for each (mp or single) adj */
- if (sm->data_structure_lock->release_hint)
- {
- start_at_fib_index = fib - im4->fibs;
- dsunlock (sm);
- ip46_fib_stats_delay (sm, 0 /* sec */ , STATS_RELEASE_DELAY_NS);
- mp->count = 0;
- ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
- goto again;
- }
+ vec_foreach (r, do_fibs->ip4routes)
+ {
+ vlib_counter_t c;
+ const dpo_id_t *dpo_id;
+ u32 index;
+
+ dpo_id = fib_entry_contribute_ip_forwarding (r->index);
+ index = (u32) dpo_id->dpoi_index;
+
+ vlib_get_combined_counter (&load_balance_main.lbm_to_counters,
+ index, &c);
+ /*
+ * If it has actually
+ * seen at least one packet, send it.
+ */
+ if (c.packets > 0)
+ {
+
+ /* already in net byte order */
+ ctrp->address = r->address.as_u32;
+ ctrp->address_length = r->address_length;
+ ctrp->packets = clib_host_to_net_u64 (c.packets);
+ ctrp->bytes = clib_host_to_net_u64 (c.bytes);
+ mp->count++;
+ ctrp++;
+
+ if (mp->count == items_this_message)
+ {
+ mp->count = htonl (items_this_message);
+ /*
+ * If the main thread's input queue is stuffed,
+ * drop the data structure lock (which the main thread
+ * may want), and take a pause.
+ */
+ unix_shared_memory_queue_lock (q);
+ if (unix_shared_memory_queue_is_full (q))
+ {
+ dsunlock (sm);
+ vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
+ unix_shared_memory_queue_unlock (q);
+ mp = 0;
+ ip46_fib_stats_delay (sm, 0 /* sec */ ,
+ STATS_RELEASE_DELAY_NS);
+ goto again;
+ }
+ vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
+ unix_shared_memory_queue_unlock (q);
+
+ items_this_message = IP4_FIB_COUNTER_BATCH_SIZE;
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) +
+ items_this_message * sizeof (vl_api_ip4_fib_counter_t));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_IP4_FIB_COUNTERS);
+ mp->count = 0;
+ mp->vrf_id = ntohl (fib->ft_table_id);
+ ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
+ }
+ } /* for each (mp or single) adj */
+ if (sm->data_structure_lock->release_hint)
+ {
+ start_at_fib_index = fib - im4->fibs;
+ dsunlock (sm);
+ ip46_fib_stats_delay (sm, 0 /* sec */ , STATS_RELEASE_DELAY_NS);
+ mp->count = 0;
+ ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
+ goto again;
+ }
} /* vec_foreach (routes) */
- dsunlock (sm);
+ dsunlock (sm);
- /* Flush any data from this fib */
- if (mp->count)
- {
- mp->count = htonl (mp->count);
- vl_msg_api_send_shmem (q, (u8 *) & mp);
- mp = 0;
- }
- }));
- /* *INDENT-ON* */
+ /* Flush any data from this fib */
+ if (mp->count)
+ {
+ mp->count = htonl (mp->count);
+ vl_msg_api_send_shmem (q, (u8 *) & mp);
+ mp = 0;
+ }
+ }
/* If e.g. the last FIB had no reportable routes, free the buffer */
if (mp)
@@ -729,13 +848,6 @@ again:
typedef struct
{
- ip6_address_t address;
- u32 address_length;
- u32 index;
-} ip6_route_t;
-
-typedef struct
-{
u32 fib_index;
ip6_route_t **routep;
stats_main_t *sm;
@@ -769,137 +881,144 @@ do_ip6_fibs (stats_main_t * sm)
api_main_t *am = sm->api_main;
vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
unix_shared_memory_queue_t *q = shmem_hdr->vl_input_queue;
- static ip6_route_t *routes;
ip6_route_t *r;
fib_table_t *fib;
- static uword *results;
+ do_ip46_fibs_t *do_fibs;
vl_api_vnet_ip6_fib_counters_t *mp = 0;
u32 items_this_message;
vl_api_ip6_fib_counter_t *ctrp = 0;
u32 start_at_fib_index = 0;
BVT (clib_bihash) * h = &im6->ip6_table[IP6_FIB_TABLE_FWDING].ip6_hash;
add_routes_in_fib_arg_t _a, *a = &_a;
+ int i;
+ do_fibs = &sm->do_ip46_fibs;
again:
+ vec_reset_length (do_fibs->fibs);
/* *INDENT-OFF* */
pool_foreach (fib, im6->fibs,
- ({
- /* We may have bailed out due to control-plane activity */
- while ((fib - im6->fibs) < start_at_fib_index)
- continue;
-
- if (mp == 0)
- {
- items_this_message = IP6_FIB_COUNTER_BATCH_SIZE;
- mp = vl_msg_api_alloc_as_if_client
- (sizeof (*mp) +
- items_this_message * sizeof (vl_api_ip6_fib_counter_t));
- mp->_vl_msg_id = ntohs (VL_API_VNET_IP6_FIB_COUNTERS);
- mp->count = 0;
- mp->vrf_id = ntohl (fib->ft_table_id);
- ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
- }
+ ({vec_add1(do_fibs->fibs,fib);}));
+ /* *INDENT-ON* */
- dslock (sm, 0 /* release hint */ , 1 /* tag */ );
- vec_reset_length (routes);
- vec_reset_length (results);
+ for (i = 0; i < vec_len (do_fibs->fibs); i++)
+ {
+ fib = do_fibs->fibs[i];
+ /* We may have bailed out due to control-plane activity */
+ while ((fib - im6->fibs) < start_at_fib_index)
+ continue;
- a->fib_index = fib - im6->fibs;
- a->routep = &routes;
- a->sm = sm;
+ if (mp == 0)
+ {
+ items_this_message = IP6_FIB_COUNTER_BATCH_SIZE;
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) +
+ items_this_message * sizeof (vl_api_ip6_fib_counter_t));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_IP6_FIB_COUNTERS);
+ mp->count = 0;
+ mp->vrf_id = ntohl (fib->ft_table_id);
+ ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
+ }
- if (clib_setjmp (&sm->jmp_buf, 0) == 0)
- {
- start_at_fib_index = fib - im6->fibs;
- BV (clib_bihash_foreach_key_value_pair) (h, add_routes_in_fib, a);
- }
- else
- {
- dsunlock (sm);
- ip46_fib_stats_delay (sm, 0 /* sec */ ,
- STATS_RELEASE_DELAY_NS);
- mp->count = 0;
- ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
- goto again;
- }
+ dslock (sm, 0 /* release hint */ , 1 /* tag */ );
- vec_foreach (r, routes)
- {
- vlib_counter_t c;
-
- vlib_get_combined_counter (&load_balance_main.lbm_to_counters,
- r->index, &c);
- /*
- * If it has actually
- * seen at least one packet, send it.
- */
- if (c.packets > 0)
- {
- /* already in net byte order */
- ctrp->address[0] = r->address.as_u64[0];
- ctrp->address[1] = r->address.as_u64[1];
- ctrp->address_length = (u8) r->address_length;
- ctrp->packets = clib_host_to_net_u64 (c.packets);
- ctrp->bytes = clib_host_to_net_u64 (c.bytes);
- mp->count++;
- ctrp++;
-
- if (mp->count == items_this_message)
- {
- mp->count = htonl (items_this_message);
- /*
- * If the main thread's input queue is stuffed,
- * drop the data structure lock (which the main thread
- * may want), and take a pause.
- */
- unix_shared_memory_queue_lock (q);
- if (unix_shared_memory_queue_is_full (q))
- {
- dsunlock (sm);
- vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
- unix_shared_memory_queue_unlock (q);
- mp = 0;
- ip46_fib_stats_delay (sm, 0 /* sec */ ,
- STATS_RELEASE_DELAY_NS);
- goto again;
- }
- vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
- unix_shared_memory_queue_unlock (q);
-
- items_this_message = IP6_FIB_COUNTER_BATCH_SIZE;
- mp = vl_msg_api_alloc_as_if_client
- (sizeof (*mp) +
- items_this_message * sizeof (vl_api_ip6_fib_counter_t));
- mp->_vl_msg_id = ntohs (VL_API_VNET_IP6_FIB_COUNTERS);
- mp->count = 0;
- mp->vrf_id = ntohl (fib->ft_table_id);
- ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
- }
- }
+ vec_reset_length (do_fibs->ip6routes);
+ vec_reset_length (do_fibs->results);
- if (sm->data_structure_lock->release_hint)
- {
- start_at_fib_index = fib - im6->fibs;
- dsunlock (sm);
- ip46_fib_stats_delay (sm, 0 /* sec */ , STATS_RELEASE_DELAY_NS);
- mp->count = 0;
- ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
- goto again;
- }
- } /* vec_foreach (routes) */
+ a->fib_index = fib - im6->fibs;
+ a->routep = &do_fibs->ip6routes;
+ a->sm = sm;
- dsunlock (sm);
+ if (clib_setjmp (&sm->jmp_buf, 0) == 0)
+ {
+ start_at_fib_index = fib - im6->fibs;
+ BV (clib_bihash_foreach_key_value_pair) (h, add_routes_in_fib, a);
+ }
+ else
+ {
+ dsunlock (sm);
+ ip46_fib_stats_delay (sm, 0 /* sec */ ,
+ STATS_RELEASE_DELAY_NS);
+ mp->count = 0;
+ ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
+ goto again;
+ }
- /* Flush any data from this fib */
- if (mp->count)
+ vec_foreach (r, do_fibs->ip6routes)
{
- mp->count = htonl (mp->count);
- vl_msg_api_send_shmem (q, (u8 *) & mp);
- mp = 0;
- }
- }));
- /* *INDENT-ON* */
+ vlib_counter_t c;
+
+ vlib_get_combined_counter (&load_balance_main.lbm_to_counters,
+ r->index, &c);
+ /*
+ * If it has actually
+ * seen at least one packet, send it.
+ */
+ if (c.packets > 0)
+ {
+ /* already in net byte order */
+ ctrp->address[0] = r->address.as_u64[0];
+ ctrp->address[1] = r->address.as_u64[1];
+ ctrp->address_length = (u8) r->address_length;
+ ctrp->packets = clib_host_to_net_u64 (c.packets);
+ ctrp->bytes = clib_host_to_net_u64 (c.bytes);
+ mp->count++;
+ ctrp++;
+
+ if (mp->count == items_this_message)
+ {
+ mp->count = htonl (items_this_message);
+ /*
+ * If the main thread's input queue is stuffed,
+ * drop the data structure lock (which the main thread
+ * may want), and take a pause.
+ */
+ unix_shared_memory_queue_lock (q);
+ if (unix_shared_memory_queue_is_full (q))
+ {
+ dsunlock (sm);
+ vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
+ unix_shared_memory_queue_unlock (q);
+ mp = 0;
+ ip46_fib_stats_delay (sm, 0 /* sec */ ,
+ STATS_RELEASE_DELAY_NS);
+ goto again;
+ }
+ vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
+ unix_shared_memory_queue_unlock (q);
+
+ items_this_message = IP6_FIB_COUNTER_BATCH_SIZE;
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) +
+ items_this_message * sizeof (vl_api_ip6_fib_counter_t));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_IP6_FIB_COUNTERS);
+ mp->count = 0;
+ mp->vrf_id = ntohl (fib->ft_table_id);
+ ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
+ }
+ }
+
+ if (sm->data_structure_lock->release_hint)
+ {
+ start_at_fib_index = fib - im6->fibs;
+ dsunlock (sm);
+ ip46_fib_stats_delay (sm, 0 /* sec */ , STATS_RELEASE_DELAY_NS);
+ mp->count = 0;
+ ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
+ goto again;
+ }
+ } /* vec_foreach (routes) */
+
+ dsunlock (sm);
+
+ /* Flush any data from this fib */
+ if (mp->count)
+ {
+ mp->count = htonl (mp->count);
+ vl_msg_api_send_shmem (q, (u8 *) & mp);
+ mp = 0;
+ }
+ }
/* If e.g. the last FIB had no reportable routes, free the buffer */
if (mp)
@@ -946,84 +1065,45 @@ static void
vl_api_vnet_interface_simple_counters_t_handler
(vl_api_vnet_interface_simple_counters_t * mp)
{
- vpe_client_registration_t *reg;
+ vpe_client_stats_registration_t *reg;
stats_main_t *sm = &stats_main;
unix_shared_memory_queue_t *q, *q_prev = NULL;
vl_api_vnet_interface_simple_counters_t *mp_copy = NULL;
u32 mp_size;
-
-#if STATS_DEBUG > 0
- char *counter_name;
- u32 count, sw_if_index;
int i;
-#endif
mp_size = sizeof (*mp) + (ntohl (mp->count) * sizeof (u64));
/* *INDENT-OFF* */
+ vec_reset_length(sm->regs);
pool_foreach(reg, sm->stats_registrations,
({
- q = vl_api_client_index_to_input_queue (reg->client_index);
- if (q)
- {
- if (q_prev && (q_prev->cursize < q_prev->maxsize))
- {
- mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
- clib_memcpy(mp_copy, mp, mp_size);
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- mp = mp_copy;
- }
- q_prev = q;
- }
+ vec_add1(sm->regs,reg);
}));
/* *INDENT-ON* */
-
-#if STATS_DEBUG > 0
- count = ntohl (mp->count);
- sw_if_index = ntohl (mp->first_sw_if_index);
- u64 *vp, v;
- vp = (u64 *) mp->data;
-
- switch (mp->vnet_counter_type)
- {
- case VNET_INTERFACE_COUNTER_DROP:
- counter_name = "drop";
- break;
- case VNET_INTERFACE_COUNTER_PUNT:
- counter_name = "punt";
- break;
- case VNET_INTERFACE_COUNTER_IP4:
- counter_name = "ip4";
- break;
- case VNET_INTERFACE_COUNTER_IP6:
- counter_name = "ip6";
- break;
- case VNET_INTERFACE_COUNTER_RX_NO_BUF:
- counter_name = "rx-no-buff";
- break;
- case VNET_INTERFACE_COUNTER_RX_MISS:
- , counter_name = "rx-miss";
- break;
- case VNET_INTERFACE_COUNTER_RX_ERROR:
- , counter_name = "rx-error (fifo-full)";
- break;
- case VNET_INTERFACE_COUNTER_TX_ERROR:
- , counter_name = "tx-error (fifo-full)";
- break;
- default:
- counter_name = "bogus";
- break;
- }
- for (i = 0; i < count; i++)
+ for (i = 0; i < vec_len (sm->regs); i++)
{
- v = clib_mem_unaligned (vp, u64);
- v = clib_net_to_host_u64 (v);
- vp++;
- fformat (stdout, "%U.%s %lld\n", format_vnet_sw_if_index_name,
- sm->vnet_main, sw_if_index, counter_name, v);
- sw_if_index++;
+ reg = sm->regs[i];
+ if (reg->stats_registrations & INTERFACE_SIMPLE_COUNTERS)
+ {
+ q = vl_api_client_index_to_input_queue (reg->client.client_index);
+ if (q)
+ {
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ mp_copy = vl_msg_api_alloc_as_if_client (mp_size);
+ clib_memcpy (mp_copy, mp, mp_size);
+ vl_msg_api_send_shmem (q_prev, (u8 *) & mp);
+ mp = mp_copy;
+ }
+ q_prev = q;
+ }
+ }
}
+#if STATS_DEBUG > 0
+ fformat (stdout, "%U\n", format_vnet_simple_counters, mp);
#endif
+
if (q_prev && (q_prev->cursize < q_prev->maxsize))
{
vl_msg_api_send_shmem (q_prev, (u8 *) & mp);
@@ -1038,75 +1118,39 @@ static void
vl_api_vnet_interface_combined_counters_t_handler
(vl_api_vnet_interface_combined_counters_t * mp)
{
- vpe_client_registration_t *reg;
+ vpe_client_stats_registration_t *reg;
stats_main_t *sm = &stats_main;
unix_shared_memory_queue_t *q, *q_prev = NULL;
vl_api_vnet_interface_combined_counters_t *mp_copy = NULL;
u32 mp_size;
-#if STATS_DEBUG > 0
- char *counter_name;
- u32 count, sw_if_index;
- int i;
-#endif
-
mp_size = sizeof (*mp) + (ntohl (mp->count) * sizeof (vlib_counter_t));
/* *INDENT-OFF* */
pool_foreach(reg, sm->stats_registrations,
({
- q = vl_api_client_index_to_input_queue (reg->client_index);
- if (q)
- {
- if (q_prev && (q_prev->cursize < q_prev->maxsize))
- {
- mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
- clib_memcpy(mp_copy, mp, mp_size);
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- mp = mp_copy;
- }
- q_prev = q;
- }
+ if (reg->stats_registrations & INTERFACE_COMBINED_COUNTERS)
+ {
+ q = vl_api_client_index_to_input_queue (reg->client.client_index);
+ if (q)
+ {
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
+ clib_memcpy(mp_copy, mp, mp_size);
+ vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
+ mp = mp_copy;
+ }
+ q_prev = q;
+ }
+ }
}));
/* *INDENT-ON* */
#if STATS_DEBUG > 0
- count = ntohl (mp->count);
- sw_if_index = ntohl (mp->first_sw_if_index);
-
- vlib_counter_t *vp;
- u64 packets, bytes;
- vp = (vlib_counter_t *) mp->data;
-
- switch (mp->vnet_counter_type)
- {
- case VNET_INTERFACE_COUNTER_RX:
- counter_name = "rx";
- break;
- case VNET_INTERFACE_COUNTER_TX:
- counter_name = "tx";
- break;
- default:
- counter_name = "bogus";
- break;
- }
- for (i = 0; i < count; i++)
- {
- packets = clib_mem_unaligned (&vp->packets, u64);
- packets = clib_net_to_host_u64 (packets);
- bytes = clib_mem_unaligned (&vp->bytes, u64);
- bytes = clib_net_to_host_u64 (bytes);
- vp++;
- fformat (stdout, "%U.%s.packets %lld\n",
- format_vnet_sw_if_index_name,
- sm->vnet_main, sw_if_index, counter_name, packets);
- fformat (stdout, "%U.%s.bytes %lld\n",
- format_vnet_sw_if_index_name,
- sm->vnet_main, sw_if_index, counter_name, bytes);
- sw_if_index++;
- }
-
+ fformat (stdout, "%U\n", format_vnet_combined_counters, mp);
#endif
+
if (q_prev && (q_prev->cursize < q_prev->maxsize))
{
vl_msg_api_send_shmem (q_prev, (u8 *) & mp);
@@ -1120,7 +1164,7 @@ static void
static void
vl_api_vnet_ip4_fib_counters_t_handler (vl_api_vnet_ip4_fib_counters_t * mp)
{
- vpe_client_registration_t *reg;
+ vpe_client_stats_registration_t *reg;
stats_main_t *sm = &stats_main;
unix_shared_memory_queue_t *q, *q_prev = NULL;
vl_api_vnet_ip4_fib_counters_t *mp_copy = NULL;
@@ -1132,17 +1176,20 @@ vl_api_vnet_ip4_fib_counters_t_handler (vl_api_vnet_ip4_fib_counters_t * mp)
/* *INDENT-OFF* */
pool_foreach(reg, sm->stats_registrations,
({
- q = vl_api_client_index_to_input_queue (reg->client_index);
- if (q)
+ if (reg->stats_registrations & IP4_FIB_COUNTERS)
{
- if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ q = vl_api_client_index_to_input_queue (reg->client.client_index);
+ if (q)
{
- mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
- clib_memcpy(mp_copy, mp, mp_size);
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- mp = mp_copy;
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
+ clib_memcpy(mp_copy, mp, mp_size);
+ vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
+ mp = mp_copy;
+ }
+ q_prev = q;
}
- q_prev = q;
}
}));
/* *INDENT-ON* */
@@ -1159,7 +1206,7 @@ vl_api_vnet_ip4_fib_counters_t_handler (vl_api_vnet_ip4_fib_counters_t * mp)
static void
vl_api_vnet_ip4_nbr_counters_t_handler (vl_api_vnet_ip4_nbr_counters_t * mp)
{
- vpe_client_registration_t *reg;
+ vpe_client_stats_registration_t *reg;
stats_main_t *sm = &stats_main;
unix_shared_memory_queue_t *q, *q_prev = NULL;
vl_api_vnet_ip4_nbr_counters_t *mp_copy = NULL;
@@ -1171,17 +1218,20 @@ vl_api_vnet_ip4_nbr_counters_t_handler (vl_api_vnet_ip4_nbr_counters_t * mp)
/* *INDENT-OFF* */
pool_foreach(reg, sm->stats_registrations,
({
- q = vl_api_client_index_to_input_queue (reg->client_index);
- if (q)
+ if (reg->stats_registrations & IP4_NBR_COUNTERS)
{
- if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ q = vl_api_client_index_to_input_queue (reg->client.client_index);
+ if (q)
{
- mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
- clib_memcpy(mp_copy, mp, mp_size);
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- mp = mp_copy;
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
+ clib_memcpy(mp_copy, mp, mp_size);
+ vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
+ mp = mp_copy;
+ }
+ q_prev = q;
}
- q_prev = q;
}
}));
/* *INDENT-ON* */
@@ -1198,7 +1248,7 @@ vl_api_vnet_ip4_nbr_counters_t_handler (vl_api_vnet_ip4_nbr_counters_t * mp)
static void
vl_api_vnet_ip6_fib_counters_t_handler (vl_api_vnet_ip6_fib_counters_t * mp)
{
- vpe_client_registration_t *reg;
+ vpe_client_stats_registration_t *reg;
stats_main_t *sm = &stats_main;
unix_shared_memory_queue_t *q, *q_prev = NULL;
vl_api_vnet_ip6_fib_counters_t *mp_copy = NULL;
@@ -1210,19 +1260,22 @@ vl_api_vnet_ip6_fib_counters_t_handler (vl_api_vnet_ip6_fib_counters_t * mp)
/* *INDENT-OFF* */
pool_foreach(reg, sm->stats_registrations,
({
- q = vl_api_client_index_to_input_queue (reg->client_index);
- if (q)
+ if (reg->stats_registrations & IP6_FIB_COUNTERS)
{
- if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ q = vl_api_client_index_to_input_queue (reg->client.client_index);
+ if (q)
{
- mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
- clib_memcpy(mp_copy, mp, mp_size);
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- mp = mp_copy;
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
+ clib_memcpy(mp_copy, mp, mp_size);
+ vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
+ mp = mp_copy;
+ }
+ q_prev = q;
}
- q_prev = q;
}
- }));
+ }));
/* *INDENT-ON* */
if (q_prev && (q_prev->cursize < q_prev->maxsize))
{
@@ -1237,7 +1290,7 @@ vl_api_vnet_ip6_fib_counters_t_handler (vl_api_vnet_ip6_fib_counters_t * mp)
static void
vl_api_vnet_ip6_nbr_counters_t_handler (vl_api_vnet_ip6_nbr_counters_t * mp)
{
- vpe_client_registration_t *reg;
+ vpe_client_stats_registration_t *reg;
stats_main_t *sm = &stats_main;
unix_shared_memory_queue_t *q, *q_prev = NULL;
vl_api_vnet_ip6_nbr_counters_t *mp_copy = NULL;
@@ -1249,17 +1302,20 @@ vl_api_vnet_ip6_nbr_counters_t_handler (vl_api_vnet_ip6_nbr_counters_t * mp)
/* *INDENT-OFF* */
pool_foreach(reg, sm->stats_registrations,
({
- q = vl_api_client_index_to_input_queue (reg->client_index);
- if (q)
+ if (reg->stats_registrations & IP6_NBR_COUNTERS)
{
- if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ q = vl_api_client_index_to_input_queue (reg->client.client_index);
+ if (q)
{
- mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
- clib_memcpy(mp_copy, mp, mp_size);
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- mp = mp_copy;
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
+ clib_memcpy(mp_copy, mp, mp_size);
+ vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
+ mp = mp_copy;
+ }
+ q_prev = q;
}
- q_prev = q;
}
}));
/* *INDENT-ON* */
@@ -1277,7 +1333,7 @@ static void
vl_api_want_stats_t_handler (vl_api_want_stats_t * mp)
{
stats_main_t *sm = &stats_main;
- vpe_client_registration_t *rp;
+ vpe_client_stats_registration_t *rp;
vl_api_want_stats_reply_t *rmp;
uword *p;
i32 retval = 0;
@@ -1307,9 +1363,16 @@ vl_api_want_stats_t_handler (vl_api_want_stats_t * mp)
goto reply;
}
pool_get (sm->stats_registrations, rp);
- rp->client_index = mp->client_index;
- rp->client_pid = mp->pid;
- hash_set (sm->stats_registration_hash, rp->client_index,
+ rp->client.client_index = mp->client_index;
+ rp->client.client_pid = mp->pid;
+ rp->stats_registrations |= INTERFACE_SIMPLE_COUNTERS;
+ rp->stats_registrations |= INTERFACE_COMBINED_COUNTERS;
+ rp->stats_registrations |= IP4_FIB_COUNTERS;
+ rp->stats_registrations |= IP4_NBR_COUNTERS;
+ rp->stats_registrations |= IP6_FIB_COUNTERS;
+ rp->stats_registrations |= IP6_NBR_COUNTERS;
+
+ hash_set (sm->stats_registration_hash, rp->client.client_index,
rp - sm->stats_registrations);
reply:
@@ -1331,10 +1394,386 @@ reply:
vl_msg_api_send_shmem (q, (u8 *) & rmp);
}
+static void
+ vl_api_want_interface_simple_stats_t_handler
+ (vl_api_want_interface_simple_stats_t * mp)
+{
+ stats_main_t *sm = &stats_main;
+ vpe_client_stats_registration_t *rp;
+ vl_api_want_interface_simple_stats_reply_t *rmp;
+ uword *p;
+ i32 retval = 0;
+ unix_shared_memory_queue_t *q;
+
+ p = hash_get (sm->stats_registration_hash, mp->client_index);
+
+ /* Disable case */
+ if (mp->enable_disable == 0)
+ {
+ if (!p) // No client to disable
+ {
+ clib_warning ("pid %d: already disabled for stats...", mp->pid);
+ retval = 0;
+ goto reply;
+ }
+
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+ if (!rp->stats_registrations & INTERFACE_SIMPLE_COUNTERS) // Client but doesn't want this.
+ {
+ clib_warning
+ ("pid %d: already disabled for interface simple stats...",
+ mp->pid);
+ retval = 0;
+ goto reply;
+ }
+ else
+ {
+ rp->stats_registrations &= ~(INTERFACE_SIMPLE_COUNTERS); // Clear flag
+ if (rp->stats_registrations == 0) // Client isn't listening to anything else
+ {
+ pool_put (sm->stats_registrations, rp);
+ hash_unset (sm->stats_registration_hash, mp->client_index);
+ }
+ goto reply;
+ }
+ }
+ /* Enable case */
+ /* Get client from pool */
+ if (p)
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+
+ if (!p || !rp) // Doesn't exist, make a new entry
+ {
+ pool_get (sm->stats_registrations, rp);
+ rp->client.client_index = mp->client_index;
+ rp->client.client_pid = mp->pid;
+ }
+ rp->stats_registrations |= INTERFACE_SIMPLE_COUNTERS;
+ hash_set (sm->stats_registration_hash, rp->client.client_index,
+ rp - sm->stats_registrations);
+
+reply:
+ if (pool_elts (sm->stats_registrations)) // Someone wants something, somewhere so enable globally for now.
+ sm->enable_poller = 1;
+ else
+ sm->enable_poller = 0;
+
+ q = vl_api_client_index_to_input_queue (mp->client_index);
+
+ if (!q)
+ return;
+
+ rmp = vl_msg_api_alloc (sizeof (*rmp));
+ rmp->_vl_msg_id = ntohs (VL_API_WANT_INTERFACE_SIMPLE_STATS_REPLY);
+ rmp->context = mp->context;
+ rmp->retval = retval;
+
+ vl_msg_api_send_shmem (q, (u8 *) & rmp);
+}
+
+static void
+ vl_api_want_interface_combined_stats_t_handler
+ (vl_api_want_interface_combined_stats_t * mp)
+{
+ stats_main_t *sm = &stats_main;
+ vpe_client_stats_registration_t *rp;
+ vl_api_want_interface_combined_stats_reply_t *rmp;
+ uword *p;
+ i32 retval = 0;
+ unix_shared_memory_queue_t *q;
+
+ p = hash_get (sm->stats_registration_hash, mp->client_index);
+
+ /* Disable case */
+ if (mp->enable_disable == 0)
+ {
+ if (!p) // No client to disable
+ {
+ clib_warning ("pid %d: already disabled for stats...", mp->pid);
+ retval = 0;
+ goto reply;
+ }
+
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+ if (!rp->stats_registrations & INTERFACE_COMBINED_COUNTERS) // Client but doesn't want this.
+ {
+ clib_warning
+ ("pid %d: already disabled for interface COMBINED stats...",
+ mp->pid);
+ retval = 0;
+ goto reply;
+ }
+ else
+ {
+ rp->stats_registrations &= ~(INTERFACE_COMBINED_COUNTERS); // Clear flag
+ if (rp->stats_registrations == 0) // Client isn't listening to anything else
+ {
+ pool_put (sm->stats_registrations, rp);
+ hash_unset (sm->stats_registration_hash, mp->client_index);
+ }
+ goto reply;
+ }
+ }
+ /* Enable case */
+ /* Get client from pool */
+ if (p)
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+
+ if (!p || !rp) // Doesn't exist, make a new entry
+ {
+ pool_get (sm->stats_registrations, rp);
+ rp->client.client_index = mp->client_index;
+ rp->client.client_pid = mp->pid;
+ }
+ rp->stats_registrations |= INTERFACE_COMBINED_COUNTERS;
+ hash_set (sm->stats_registration_hash, rp->client.client_index,
+ rp - sm->stats_registrations);
+
+reply:
+ if (pool_elts (sm->stats_registrations)) // Someone wants something, somewhere so enable globally for now.
+ sm->enable_poller = 1;
+ else
+ sm->enable_poller = 0;
+
+ q = vl_api_client_index_to_input_queue (mp->client_index);
+
+ if (!q)
+ return;
+
+ rmp = vl_msg_api_alloc (sizeof (*rmp));
+ rmp->_vl_msg_id = ntohs (VL_API_WANT_INTERFACE_COMBINED_STATS_REPLY);
+ rmp->context = mp->context;
+ rmp->retval = retval;
+
+ vl_msg_api_send_shmem (q, (u8 *) & rmp);
+}
+
+static void
+vl_api_want_ip4_fib_stats_t_handler (vl_api_want_ip4_fib_stats_t * mp)
+{
+ stats_main_t *sm = &stats_main;
+ vpe_client_stats_registration_t *rp;
+ vl_api_want_ip4_fib_stats_reply_t *rmp;
+ uword *p;
+ i32 retval = 0;
+ unix_shared_memory_queue_t *q;
+
+ p = hash_get (sm->stats_registration_hash, mp->client_index);
+
+ /* Disable case */
+ /*
+ $$$ FIXME: need std return codes. Still undecided if enabling already
+ enabled (and similar for disabled) is really a -'ve error condition or
+ if 0 is sufficient
+ */
+ if (mp->enable_disable == 0)
+ {
+ if (!p) // No client to disable
+ {
+ clib_warning ("pid %d: already disabled for stats...", mp->pid);
+ retval = -3;
+ goto reply;
+ }
+
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+ if (!rp->stats_registrations & IP4_FIB_COUNTERS) // Client but doesn't want this.
+ {
+ clib_warning ("pid %d: already disabled for interface ip4 fib...",
+ mp->pid);
+ retval = -2;
+ goto reply;
+ }
+ else
+ {
+ rp->stats_registrations &= ~(IP4_FIB_COUNTERS); // Clear flag
+ if (rp->stats_registrations == 0) // Client isn't listening to anything else
+ {
+ pool_put (sm->stats_registrations, rp);
+ hash_unset (sm->stats_registration_hash, mp->client_index);
+ }
+ goto reply;
+ }
+ }
+ /* Enable case */
+ /* Get client from pool */
+ if (p)
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+
+ if (!p || !rp) // Doesn't exist, make a new entry
+ {
+ pool_get (sm->stats_registrations, rp);
+ rp->client.client_index = mp->client_index;
+ rp->client.client_pid = mp->pid;
+ }
+ rp->stats_registrations |= IP4_FIB_COUNTERS;
+ hash_set (sm->stats_registration_hash, rp->client.client_index,
+ rp - sm->stats_registrations);
+
+reply:
+ if (pool_elts (sm->stats_registrations)) // Someone wants something, somewhere so enable globally for now.
+ sm->enable_poller = 1;
+ else
+ sm->enable_poller = 0;
+
+ q = vl_api_client_index_to_input_queue (mp->client_index);
+
+ if (!q)
+ return;
+
+ rmp = vl_msg_api_alloc (sizeof (*rmp));
+ rmp->_vl_msg_id = ntohs (VL_API_WANT_IP4_FIB_STATS_REPLY);
+ rmp->context = mp->context;
+ rmp->retval = retval;
+
+ vl_msg_api_send_shmem (q, (u8 *) & rmp);
+}
+
+static void
+vl_api_want_ip6_fib_stats_t_handler (vl_api_want_ip6_fib_stats_t * mp)
+{
+ stats_main_t *sm = &stats_main;
+ vpe_client_stats_registration_t *rp;
+ vl_api_want_ip6_fib_stats_reply_t *rmp;
+ uword *p;
+ i32 retval = 0;
+ unix_shared_memory_queue_t *q;
+
+ p = hash_get (sm->stats_registration_hash, mp->client_index);
+
+ /* Disable case */
+ /*
+ $$$ FIXME: need std return codes. Still undecided if enabling already
+ enabled (and similar for disabled) is really a -'ve error condition or
+ if 0 is sufficient
+ */
+ if (mp->enable_disable == 0)
+ {
+ if (!p) // No client to disable
+ {
+ clib_warning ("pid %d: already disabled for stats...", mp->pid);
+ retval = -3;
+ goto reply;
+ }
+
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+ if (!rp->stats_registrations & IP6_FIB_COUNTERS) // Client but doesn't want this.
+ {
+ clib_warning ("pid %d: already disabled for interface ip6 fib...",
+ mp->pid);
+ retval = -2;
+ goto reply;
+ }
+ else
+ {
+ rp->stats_registrations &= ~(IP6_FIB_COUNTERS); // Clear flag
+ if (rp->stats_registrations == 0) // Client isn't listening to anything else
+ {
+ pool_put (sm->stats_registrations, rp);
+ hash_unset (sm->stats_registration_hash, mp->client_index);
+ }
+ goto reply;
+ }
+ }
+ /* Enable case */
+ /* Get client from pool */
+ if (p)
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+
+ if (!p || !rp) // Doesn't exist, make a new entry
+ {
+ pool_get (sm->stats_registrations, rp);
+ rp->client.client_index = mp->client_index;
+ rp->client.client_pid = mp->pid;
+ }
+ rp->stats_registrations |= IP6_FIB_COUNTERS;
+ hash_set (sm->stats_registration_hash, rp->client.client_index,
+ rp - sm->stats_registrations);
+
+reply:
+ if (pool_elts (sm->stats_registrations)) // Someone wants something, somewhere so enable globally for now.
+ sm->enable_poller = 1;
+ else
+ sm->enable_poller = 0;
+
+ q = vl_api_client_index_to_input_queue (mp->client_index);
+
+ if (!q)
+ return;
+
+ rmp = vl_msg_api_alloc (sizeof (*rmp));
+ rmp->_vl_msg_id = ntohs (VL_API_WANT_IP6_FIB_STATS_REPLY);
+ rmp->context = mp->context;
+ rmp->retval = retval;
+
+ vl_msg_api_send_shmem (q, (u8 *) & rmp);
+}
+
+/* FIXME - NBR stats broken - this will be fixed in subsequent patch */
+static void
+vl_api_want_ip4_nbr_stats_t_handler (vl_api_want_ip4_nbr_stats_t * mp)
+{
+}
+
+static void
+vl_api_want_ip6_nbr_stats_t_handler (vl_api_want_ip6_nbr_stats_t * mp)
+{
+}
+
+static void
+vl_api_vnet_get_summary_stats_t_handler (vl_api_vnet_get_summary_stats_t * mp)
+{
+ stats_main_t *sm = &stats_main;
+ vnet_interface_main_t *im = sm->interface_main;
+ vl_api_vnet_get_summary_stats_reply_t *rmp;
+ vlib_combined_counter_main_t *cm;
+ vlib_counter_t v;
+ int i, which;
+ u64 total_pkts[VLIB_N_RX_TX];
+ u64 total_bytes[VLIB_N_RX_TX];
+
+ unix_shared_memory_queue_t *q =
+ vl_api_client_index_to_input_queue (mp->client_index);
+
+ if (!q)
+ return;
+
+ rmp = vl_msg_api_alloc (sizeof (*rmp));
+ rmp->_vl_msg_id = ntohs (VL_API_VNET_GET_SUMMARY_STATS_REPLY);
+ rmp->context = mp->context;
+ rmp->retval = 0;
+
+ memset (total_pkts, 0, sizeof (total_pkts));
+ memset (total_bytes, 0, sizeof (total_bytes));
+
+ vnet_interface_counter_lock (im);
+
+ vec_foreach (cm, im->combined_sw_if_counters)
+ {
+ which = cm - im->combined_sw_if_counters;
+
+ for (i = 0; i < vlib_combined_counter_n_counters (cm); i++)
+ {
+ vlib_get_combined_counter (cm, i, &v);
+ total_pkts[which] += v.packets;
+ total_bytes[which] += v.bytes;
+ }
+ }
+ vnet_interface_counter_unlock (im);
+
+ rmp->total_pkts[VLIB_RX] = clib_host_to_net_u64 (total_pkts[VLIB_RX]);
+ rmp->total_bytes[VLIB_RX] = clib_host_to_net_u64 (total_bytes[VLIB_RX]);
+ rmp->total_pkts[VLIB_TX] = clib_host_to_net_u64 (total_pkts[VLIB_TX]);
+ rmp->total_bytes[VLIB_TX] = clib_host_to_net_u64 (total_bytes[VLIB_TX]);
+ rmp->vector_rate =
+ clib_host_to_net_u64 (vlib_last_vector_length_per_node (sm->vlib_main));
+
+ vl_msg_api_send_shmem (q, (u8 *) & rmp);
+}
+
int
stats_memclnt_delete_callback (u32 client_index)
{
- vpe_client_registration_t *rp;
+ vpe_client_stats_registration_t *rp;
stats_main_t *sm = &stats_main;
uword *p;
diff --git a/src/vpp/stats/stats.h b/src/vpp/stats/stats.h
index 024dc78e447..167b09bd11c 100644
--- a/src/vpp/stats/stats.h
+++ b/src/vpp/stats/stats.h
@@ -21,6 +21,7 @@
#include <vnet/interface.h>
#include <pthread.h>
#include <vlib/threads.h>
+#include <vnet/fib/fib_table.h>
#include <vlib/unix/unix.h>
#include <vlibmemory/api.h>
#include <vlibmemory/unix_shared_memory_queue.h>
@@ -37,6 +38,44 @@ typedef struct
typedef struct
{
+ vpe_client_registration_t client;
+ u8 stats_registrations;
+#define INTERFACE_SIMPLE_COUNTERS (1 << 0)
+#define INTERFACE_COMBINED_COUNTERS (1 << 1)
+#define IP4_FIB_COUNTERS (1 << 2)
+#define IP4_NBR_COUNTERS (1 << 3)
+#define IP6_FIB_COUNTERS (1 << 4)
+#define IP6_NBR_COUNTERS (1 << 5)
+
+} vpe_client_stats_registration_t;
+
+/* from .../vnet/vnet/ip/lookup.c. Yuck */
+typedef CLIB_PACKED (struct
+ {
+ ip4_address_t address;
+u32 address_length: 6;
+u32 index: 26;
+ }) ip4_route_t;
+
+typedef struct
+{
+ ip6_address_t address;
+ u32 address_length;
+ u32 index;
+} ip6_route_t;
+
+
+typedef struct
+{
+ ip4_route_t *ip4routes;
+ ip6_route_t *ip6routes;
+ fib_table_t **fibs;
+ hash_pair_t **pvec;
+ uword *results;
+} do_ip46_fibs_t;
+
+typedef struct
+{
void *mheap;
pthread_t thread_self;
pthread_t thread_handle;
@@ -45,7 +84,8 @@ typedef struct
u32 enable_poller;
uword *stats_registration_hash;
- vpe_client_registration_t *stats_registrations;
+ vpe_client_stats_registration_t *stats_registrations;
+ vpe_client_stats_registration_t **regs;
/* control-plane data structure lock */
data_structure_lock_t *data_structure_lock;
@@ -53,6 +93,9 @@ typedef struct
/* bail out of FIB walk if set */
clib_longjmp_t jmp_buf;
+ /* Vectors for Distribution funcs: do_ip4_fibs and do_ip6_fibs. */
+ do_ip46_fibs_t do_ip46_fibs;
+
/* convenience */
vlib_main_t *vlib_main;
vnet_main_t *vnet_main;