summaryrefslogtreecommitdiffstats
path: root/vpp/stats
diff options
context:
space:
mode:
authorDave Barach <dave@barachs.net>2016-08-05 18:09:54 -0400
committerDave Barach <dave@barachs.net>2016-08-05 18:10:06 -0400
commita8d77eda738855480c3028664028c7a448d52424 (patch)
treec3e30344155ab5aef053e2dd57bc89bb1349bda3 /vpp/stats
parent310f7fe7a60467569f9d62d97847bc36dd5eb4ca (diff)
VPP-237 Coding standards cleanup
Change-Id: I239143f920285050ad85244ece6893236195534f Signed-off-by: Dave Barach <dave@barachs.net>
Diffstat (limited to 'vpp/stats')
-rw-r--r--vpp/stats/stats.c1626
-rw-r--r--vpp/stats/stats.h67
2 files changed, 902 insertions, 791 deletions
diff --git a/vpp/stats/stats.c b/vpp/stats/stats.c
index 2d366fa5..823e0641 100644
--- a/vpp/stats/stats.c
+++ b/vpp/stats/stats.c
@@ -27,12 +27,12 @@ stats_main_t stats_main;
#define f64_endian(a)
#define f64_print(a,b)
-#define vl_typedefs /* define message structures */
-#include <vpp-api/vpe_all_api_h.h>
+#define vl_typedefs /* define message structures */
+#include <vpp-api/vpe_all_api_h.h>
#undef vl_typedefs
-#define vl_endianfun /* define message structures */
-#include <vpp-api/vpe_all_api_h.h>
+#define vl_endianfun /* define message structures */
+#include <vpp-api/vpe_all_api_h.h>
#undef vl_endianfun
/* instantiate all the print functions we know about */
@@ -58,383 +58,420 @@ _(VNET_IP6_FIB_COUNTERS, vnet_ip6_fib_counters)
#define STATS_RELEASE_DELAY_NS (1000 * 1000 * 5)
/* ns/us us/ms */
-void dslock (stats_main_t *sm, int release_hint, int tag)
+void
+dslock (stats_main_t * sm, int release_hint, int tag)
{
- u32 thread_id;
- data_structure_lock_t *l = sm->data_structure_lock;
+ u32 thread_id;
+ data_structure_lock_t *l = sm->data_structure_lock;
- if(PREDICT_FALSE(l == 0))
- return;
+ if (PREDICT_FALSE (l == 0))
+ return;
- thread_id = os_get_cpu_number();
- if (l->lock && l->thread_id == thread_id) {
- l->count++;
- return;
+ thread_id = os_get_cpu_number ();
+ if (l->lock && l->thread_id == thread_id)
+ {
+ l->count++;
+ return;
}
- if (release_hint)
- l->release_hint++;
+ if (release_hint)
+ l->release_hint++;
- while (__sync_lock_test_and_set (&l->lock, 1))
- /* zzzz */ ;
- l->tag = tag;
- l->thread_id = thread_id;
- l->count = 1;
+ while (__sync_lock_test_and_set (&l->lock, 1))
+ /* zzzz */ ;
+ l->tag = tag;
+ l->thread_id = thread_id;
+ l->count = 1;
}
-void dsunlock (stats_main_t *sm)
+void
+dsunlock (stats_main_t * sm)
{
- u32 thread_id;
- data_structure_lock_t *l = sm->data_structure_lock;
-
- if(PREDICT_FALSE(l == 0))
- return;
-
- thread_id = os_get_cpu_number();
- ASSERT (l->lock && l->thread_id == thread_id);
- l->count--;
- if (l->count == 0) {
- l->tag = -l->tag;
- l->release_hint = 0;
- CLIB_MEMORY_BARRIER();
- l->lock = 0;
+ u32 thread_id;
+ data_structure_lock_t *l = sm->data_structure_lock;
+
+ if (PREDICT_FALSE (l == 0))
+ return;
+
+ thread_id = os_get_cpu_number ();
+ ASSERT (l->lock && l->thread_id == thread_id);
+ l->count--;
+ if (l->count == 0)
+ {
+ l->tag = -l->tag;
+ l->release_hint = 0;
+ CLIB_MEMORY_BARRIER ();
+ l->lock = 0;
}
}
-static void do_simple_interface_counters (stats_main_t * sm)
+static void
+do_simple_interface_counters (stats_main_t * sm)
{
- vl_api_vnet_interface_counters_t * mp = 0;
- vnet_interface_main_t * im = sm->interface_main;
- api_main_t * am = sm->api_main;
- vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
- unix_shared_memory_queue_t * q = shmem_hdr->vl_input_queue;
- vlib_simple_counter_main_t * cm;
- u32 items_this_message = 0;
- u64 v, *vp = 0;
- int i;
-
- /*
- * Prevent interface registration from expanding / moving the vectors...
- * That tends never to happen, so we can hold this lock for a while.
- */
- vnet_interface_counter_lock (im);
-
- vec_foreach (cm, im->sw_if_counters) {
-
- for (i = 0; i < vec_len (cm->maxi); i++) {
- if (mp == 0) {
- items_this_message = clib_min (SIMPLE_COUNTER_BATCH_SIZE,
- vec_len (cm->maxi) - i);
-
- mp = vl_msg_api_alloc_as_if_client
- (sizeof (*mp) + items_this_message * sizeof (v));
- mp->_vl_msg_id = ntohs (VL_API_VNET_INTERFACE_COUNTERS);
- mp->vnet_counter_type = cm - im->sw_if_counters;
- mp->is_combined = 0;
- mp->first_sw_if_index = htonl (i);
- mp->count = 0;
- vp = (u64 *) mp->data;
- }
- v = vlib_get_simple_counter (cm, i);
- clib_mem_unaligned (vp, u64) = clib_host_to_net_u64 (v);
- vp++;
- mp->count++;
- if (mp->count == items_this_message) {
- mp->count = htonl (items_this_message);
- /* Send to the main thread... */
- vl_msg_api_send_shmem (q, (u8 *)&mp);
- mp = 0;
- }
- }
- ASSERT (mp == 0);
- }
- vnet_interface_counter_unlock (im);
+ vl_api_vnet_interface_counters_t *mp = 0;
+ vnet_interface_main_t *im = sm->interface_main;
+ api_main_t *am = sm->api_main;
+ vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
+ unix_shared_memory_queue_t *q = shmem_hdr->vl_input_queue;
+ vlib_simple_counter_main_t *cm;
+ u32 items_this_message = 0;
+ u64 v, *vp = 0;
+ int i;
+
+ /*
+ * Prevent interface registration from expanding / moving the vectors...
+ * That tends never to happen, so we can hold this lock for a while.
+ */
+ vnet_interface_counter_lock (im);
+
+ vec_foreach (cm, im->sw_if_counters)
+ {
+
+ for (i = 0; i < vec_len (cm->maxi); i++)
+ {
+ if (mp == 0)
+ {
+ items_this_message = clib_min (SIMPLE_COUNTER_BATCH_SIZE,
+ vec_len (cm->maxi) - i);
+
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) + items_this_message * sizeof (v));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_INTERFACE_COUNTERS);
+ mp->vnet_counter_type = cm - im->sw_if_counters;
+ mp->is_combined = 0;
+ mp->first_sw_if_index = htonl (i);
+ mp->count = 0;
+ vp = (u64 *) mp->data;
+ }
+ v = vlib_get_simple_counter (cm, i);
+ clib_mem_unaligned (vp, u64) = clib_host_to_net_u64 (v);
+ vp++;
+ mp->count++;
+ if (mp->count == items_this_message)
+ {
+ mp->count = htonl (items_this_message);
+ /* Send to the main thread... */
+ vl_msg_api_send_shmem (q, (u8 *) & mp);
+ mp = 0;
+ }
+ }
+ ASSERT (mp == 0);
+ }
+ vnet_interface_counter_unlock (im);
}
-static void do_combined_interface_counters (stats_main_t * sm)
+static void
+do_combined_interface_counters (stats_main_t * sm)
{
- vl_api_vnet_interface_counters_t * mp = 0;
- vnet_interface_main_t * im = sm->interface_main;
- api_main_t * am = sm->api_main;
- vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
- unix_shared_memory_queue_t * q = shmem_hdr->vl_input_queue;
- vlib_combined_counter_main_t * cm;
- u32 items_this_message = 0;
- vlib_counter_t v, *vp = 0;
- int i;
-
- vnet_interface_counter_lock (im);
-
- vec_foreach (cm, im->combined_sw_if_counters) {
-
- for (i = 0; i < vec_len (cm->maxi); i++) {
- if (mp == 0) {
- items_this_message = clib_min (COMBINED_COUNTER_BATCH_SIZE,
- vec_len (cm->maxi) - i);
-
- mp = vl_msg_api_alloc_as_if_client
- (sizeof (*mp) + items_this_message * sizeof (v));
- mp->_vl_msg_id = ntohs (VL_API_VNET_INTERFACE_COUNTERS);
- mp->vnet_counter_type = cm - im->combined_sw_if_counters;
- mp->is_combined = 1;
- mp->first_sw_if_index = htonl (i);
- mp->count = 0;
- vp = (vlib_counter_t *)mp->data;
- }
- vlib_get_combined_counter (cm, i, &v);
- clib_mem_unaligned (&vp->packets, u64)
- = clib_host_to_net_u64 (v.packets);
- clib_mem_unaligned (&vp->bytes, u64)
- = clib_host_to_net_u64 (v.bytes);
- vp++;
- mp->count++;
- if (mp->count == items_this_message) {
- mp->count = htonl (items_this_message);
- /* Send to the main thread... */
- vl_msg_api_send_shmem (q, (u8 *)&mp);
- mp = 0;
- }
- }
- ASSERT (mp == 0);
- }
- vnet_interface_counter_unlock (im);
+ vl_api_vnet_interface_counters_t *mp = 0;
+ vnet_interface_main_t *im = sm->interface_main;
+ api_main_t *am = sm->api_main;
+ vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
+ unix_shared_memory_queue_t *q = shmem_hdr->vl_input_queue;
+ vlib_combined_counter_main_t *cm;
+ u32 items_this_message = 0;
+ vlib_counter_t v, *vp = 0;
+ int i;
+
+ vnet_interface_counter_lock (im);
+
+ vec_foreach (cm, im->combined_sw_if_counters)
+ {
+
+ for (i = 0; i < vec_len (cm->maxi); i++)
+ {
+ if (mp == 0)
+ {
+ items_this_message = clib_min (COMBINED_COUNTER_BATCH_SIZE,
+ vec_len (cm->maxi) - i);
+
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) + items_this_message * sizeof (v));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_INTERFACE_COUNTERS);
+ mp->vnet_counter_type = cm - im->combined_sw_if_counters;
+ mp->is_combined = 1;
+ mp->first_sw_if_index = htonl (i);
+ mp->count = 0;
+ vp = (vlib_counter_t *) mp->data;
+ }
+ vlib_get_combined_counter (cm, i, &v);
+ clib_mem_unaligned (&vp->packets, u64)
+ = clib_host_to_net_u64 (v.packets);
+ clib_mem_unaligned (&vp->bytes, u64) = clib_host_to_net_u64 (v.bytes);
+ vp++;
+ mp->count++;
+ if (mp->count == items_this_message)
+ {
+ mp->count = htonl (items_this_message);
+ /* Send to the main thread... */
+ vl_msg_api_send_shmem (q, (u8 *) & mp);
+ mp = 0;
+ }
+ }
+ ASSERT (mp == 0);
+ }
+ vnet_interface_counter_unlock (im);
}
/* from .../vnet/vnet/ip/lookup.c. Yuck */
-typedef CLIB_PACKED (struct {
- ip4_address_t address;
-
- u32 address_length : 6;
-
- u32 index : 26;
-}) ip4_route_t;
-
-static void ip46_fib_stats_delay (stats_main_t * sm, u32 sec, u32 nsec)
+typedef CLIB_PACKED (struct
+ {
+ ip4_address_t address;
+u32 address_length: 6;
+u32 index: 26;
+ }) ip4_route_t;
+
+static void
+ip46_fib_stats_delay (stats_main_t * sm, u32 sec, u32 nsec)
{
- struct timespec _req, *req = &_req;
- struct timespec _rem, *rem = &_rem;
-
- req->tv_sec = sec;
- req->tv_nsec = nsec;
- while (1) {
- if (nanosleep (req, rem) == 0)
- break;
- *req = *rem;
- if (errno == EINTR)
- continue;
- clib_unix_warning ("nanosleep");
- break;
+ struct timespec _req, *req = &_req;
+ struct timespec _rem, *rem = &_rem;
+
+ req->tv_sec = sec;
+ req->tv_nsec = nsec;
+ while (1)
+ {
+ if (nanosleep (req, rem) == 0)
+ break;
+ *req = *rem;
+ if (errno == EINTR)
+ continue;
+ clib_unix_warning ("nanosleep");
+ break;
}
}
-static void do_ip4_fibs (stats_main_t * sm)
+static void
+do_ip4_fibs (stats_main_t * sm)
{
- ip4_main_t * im4 = &ip4_main;
- api_main_t * am = sm->api_main;
- vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
- unix_shared_memory_queue_t * q = shmem_hdr->vl_input_queue;
- static ip4_route_t * routes;
- ip4_route_t * r;
- ip4_fib_t * fib;
- ip_lookup_main_t * lm = &im4->lookup_main;
- static uword * results;
- vl_api_vnet_ip4_fib_counters_t * mp = 0;
- u32 items_this_message;
- vl_api_ip4_fib_counter_t *ctrp = 0;
- u32 start_at_fib_index = 0;
- int i;
+ ip4_main_t *im4 = &ip4_main;
+ api_main_t *am = sm->api_main;
+ vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
+ unix_shared_memory_queue_t *q = shmem_hdr->vl_input_queue;
+ static ip4_route_t *routes;
+ ip4_route_t *r;
+ ip4_fib_t *fib;
+ ip_lookup_main_t *lm = &im4->lookup_main;
+ static uword *results;
+ vl_api_vnet_ip4_fib_counters_t *mp = 0;
+ u32 items_this_message;
+ vl_api_ip4_fib_counter_t *ctrp = 0;
+ u32 start_at_fib_index = 0;
+ int i;
again:
- vec_foreach (fib, im4->fibs) {
- /* We may have bailed out due to control-plane activity */
- while ((fib - im4->fibs) < start_at_fib_index)
- continue;
-
- if (mp == 0) {
- items_this_message = IP4_FIB_COUNTER_BATCH_SIZE;
- mp = vl_msg_api_alloc_as_if_client
- (sizeof(*mp) +
- items_this_message*sizeof(vl_api_ip4_fib_counter_t));
- mp->_vl_msg_id = ntohs (VL_API_VNET_IP4_FIB_COUNTERS);
- mp->count = 0;
- mp->vrf_id = ntohl(fib->table_id);
- ctrp = (vl_api_ip4_fib_counter_t *)mp->c;
- } else {
- /* happens if the last FIB was empty... */
- ASSERT(mp->count == 0);
- mp->vrf_id = ntohl(fib->table_id);
- }
-
- dslock (sm, 0 /* release hint */, 1 /* tag */);
-
- vec_reset_length (routes);
- vec_reset_length (results);
-
- for (i = 0; i < ARRAY_LEN (fib->adj_index_by_dst_address); i++) {
- uword * hash = fib->adj_index_by_dst_address[i];
- hash_pair_t * p;
- ip4_route_t x;
-
- x.address_length = i;
-
- hash_foreach_pair (p, hash,
- ({
- x.address.data_u32 = p->key;
- if (lm->fib_result_n_words > 1) {
- x.index = vec_len (results);
- vec_add (results, p->value, lm->fib_result_n_words);
- }
- else
- x.index = p->value[0];
-
- vec_add1 (routes, x);
- if (sm->data_structure_lock->release_hint) {
- start_at_fib_index = fib - im4->fibs;
- dsunlock (sm);
- ip46_fib_stats_delay (sm, 0 /* sec */,
- STATS_RELEASE_DELAY_NS);
- mp->count = 0;
- ctrp = (vl_api_ip4_fib_counter_t *)mp->c;
- goto again;
- }
- }));
- }
-
- vec_foreach (r, routes) {
- vlib_counter_t c, sum;
- uword i, j, n_left, n_nhs, adj_index, * result = 0;
- ip_adjacency_t * adj;
- ip_multipath_next_hop_t * nhs, tmp_nhs[1];
-
- adj_index = r->index;
- if (lm->fib_result_n_words > 1) {
- result = vec_elt_at_index (results, adj_index);
- adj_index = result[0];
- }
+ vec_foreach (fib, im4->fibs)
+ {
+ /* We may have bailed out due to control-plane activity */
+ while ((fib - im4->fibs) < start_at_fib_index)
+ continue;
+
+ if (mp == 0)
+ {
+ items_this_message = IP4_FIB_COUNTER_BATCH_SIZE;
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) +
+ items_this_message * sizeof (vl_api_ip4_fib_counter_t));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_IP4_FIB_COUNTERS);
+ mp->count = 0;
+ mp->vrf_id = ntohl (fib->table_id);
+ ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
+ }
+ else
+ {
+ /* happens if the last FIB was empty... */
+ ASSERT (mp->count == 0);
+ mp->vrf_id = ntohl (fib->table_id);
+ }
+
+ dslock (sm, 0 /* release hint */ , 1 /* tag */ );
+
+ vec_reset_length (routes);
+ vec_reset_length (results);
+
+ for (i = 0; i < ARRAY_LEN (fib->adj_index_by_dst_address); i++)
+ {
+ uword *hash = fib->adj_index_by_dst_address[i];
+ hash_pair_t *p;
+ ip4_route_t x;
+
+ x.address_length = i;
+
+ /* *INDENT-OFF* */
+ hash_foreach_pair (p, hash,
+ ({
+ x.address.data_u32 = p->key;
+ if (lm->fib_result_n_words > 1)
+ {
+ x.index = vec_len (results);
+ vec_add (results, p->value, lm->fib_result_n_words);
+ }
+ else
+ x.index = p->value[0];
+
+ vec_add1 (routes, x);
+ if (sm->data_structure_lock->release_hint)
+ {
+ start_at_fib_index = fib - im4->fibs;
+ dsunlock (sm);
+ ip46_fib_stats_delay (sm, 0 /* sec */,
+ STATS_RELEASE_DELAY_NS);
+ mp->count = 0;
+ ctrp = (vl_api_ip4_fib_counter_t *)mp->c;
+ goto again;
+ }
+ }));
+ /* *INDENT-ON* */
+ }
- adj = ip_get_adjacency (lm, adj_index);
- if (adj->n_adj == 1) {
- nhs = &tmp_nhs[0];
- nhs[0].next_hop_adj_index = ~0; /* not used */
- nhs[0].weight = 1;
- n_nhs = 1;
- } else {
- ip_multipath_adjacency_t * madj;
- madj = vec_elt_at_index (lm->multipath_adjacencies,
- adj->heap_handle);
- nhs = heap_elt_at_index
- (lm->next_hop_heap,
- madj->normalized_next_hops.heap_offset);
- n_nhs = madj->normalized_next_hops.count;
- }
-
- n_left = nhs[0].weight;
- vlib_counter_zero (&sum);
- for (i = j = 0; i < adj->n_adj; i++) {
- n_left -= 1;
- vlib_get_combined_counter (&lm->adjacency_counters,
- adj_index + i, &c);
- vlib_counter_add (&sum, &c);
- /*
- * If we're done with this adj and it has actually
- * seen at least one packet, send it.
- */
- if (n_left == 0 && sum.packets > 0) {
-
- /* already in net byte order */
- ctrp->address = r->address.as_u32;
- ctrp->address_length = r->address_length;
- ctrp->packets = clib_host_to_net_u64 (sum.packets);
- ctrp->bytes = clib_host_to_net_u64(sum.bytes);
- mp->count++;
- ctrp++;
-
- if (mp->count == items_this_message) {
- mp->count = htonl (items_this_message);
- /*
- * If the main thread's input queue is stuffed,
- * drop the data structure lock (which the main thread
- * may want), and take a pause.
- */
- unix_shared_memory_queue_lock (q);
- if (unix_shared_memory_queue_is_full (q)) {
- dsunlock (sm);
- vl_msg_api_send_shmem_nolock (q, (u8 *)&mp);
- unix_shared_memory_queue_unlock (q);
- mp = 0;
- ip46_fib_stats_delay (sm, 0 /* sec */,
- STATS_RELEASE_DELAY_NS);
- goto again;
- }
- vl_msg_api_send_shmem_nolock (q, (u8 *)&mp);
- unix_shared_memory_queue_unlock (q);
-
- items_this_message = IP4_FIB_COUNTER_BATCH_SIZE;
- mp = vl_msg_api_alloc_as_if_client
- (sizeof(*mp) +
- items_this_message*
- sizeof(vl_api_ip4_fib_counter_t));
- mp->_vl_msg_id = ntohs (VL_API_VNET_IP4_FIB_COUNTERS);
- mp->count = 0;
- mp->vrf_id = ntohl(fib->table_id);
- ctrp = (vl_api_ip4_fib_counter_t *)mp->c;
- }
-
- j++;
- if (j < n_nhs) {
- n_left = nhs[j].weight;
- vlib_counter_zero (&sum);
+ vec_foreach (r, routes)
+ {
+ vlib_counter_t c, sum;
+ uword i, j, n_left, n_nhs, adj_index, *result = 0;
+ ip_adjacency_t *adj;
+ ip_multipath_next_hop_t *nhs, tmp_nhs[1];
+
+ adj_index = r->index;
+ if (lm->fib_result_n_words > 1)
+ {
+ result = vec_elt_at_index (results, adj_index);
+ adj_index = result[0];
+ }
+
+ adj = ip_get_adjacency (lm, adj_index);
+ if (adj->n_adj == 1)
+ {
+ nhs = &tmp_nhs[0];
+ nhs[0].next_hop_adj_index = ~0; /* not used */
+ nhs[0].weight = 1;
+ n_nhs = 1;
+ }
+ else
+ {
+ ip_multipath_adjacency_t *madj;
+ madj = vec_elt_at_index (lm->multipath_adjacencies,
+ adj->heap_handle);
+ nhs = heap_elt_at_index
+ (lm->next_hop_heap, madj->normalized_next_hops.heap_offset);
+ n_nhs = madj->normalized_next_hops.count;
+ }
+
+ n_left = nhs[0].weight;
+ vlib_counter_zero (&sum);
+ for (i = j = 0; i < adj->n_adj; i++)
+ {
+ n_left -= 1;
+ vlib_get_combined_counter (&lm->adjacency_counters,
+ adj_index + i, &c);
+ vlib_counter_add (&sum, &c);
+ /*
+ * If we're done with this adj and it has actually
+ * seen at least one packet, send it.
+ */
+ if (n_left == 0 && sum.packets > 0)
+ {
+
+ /* already in net byte order */
+ ctrp->address = r->address.as_u32;
+ ctrp->address_length = r->address_length;
+ ctrp->packets = clib_host_to_net_u64 (sum.packets);
+ ctrp->bytes = clib_host_to_net_u64 (sum.bytes);
+ mp->count++;
+ ctrp++;
+
+ if (mp->count == items_this_message)
+ {
+ mp->count = htonl (items_this_message);
+ /*
+ * If the main thread's input queue is stuffed,
+ * drop the data structure lock (which the main thread
+ * may want), and take a pause.
+ */
+ unix_shared_memory_queue_lock (q);
+ if (unix_shared_memory_queue_is_full (q))
+ {
+ dsunlock (sm);
+ vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
+ unix_shared_memory_queue_unlock (q);
+ mp = 0;
+ ip46_fib_stats_delay (sm, 0 /* sec */ ,
+ STATS_RELEASE_DELAY_NS);
+ goto again;
}
+ vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
+ unix_shared_memory_queue_unlock (q);
+
+ items_this_message = IP4_FIB_COUNTER_BATCH_SIZE;
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) +
+ items_this_message * sizeof (vl_api_ip4_fib_counter_t));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_IP4_FIB_COUNTERS);
+ mp->count = 0;
+ mp->vrf_id = ntohl (fib->table_id);
+ ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
}
- } /* for each (mp or single) adj */
- if (sm->data_structure_lock->release_hint) {
- start_at_fib_index = fib - im4->fibs;
- dsunlock (sm);
- ip46_fib_stats_delay (sm, 0 /* sec */, STATS_RELEASE_DELAY_NS);
- mp->count = 0;
- ctrp = (vl_api_ip4_fib_counter_t *)mp->c;
- goto again;
- }
- } /* vec_foreach (routes) */
-
- dsunlock(sm);
-
- /* Flush any data from this fib */
- if (mp->count) {
- mp->count = htonl (mp->count);
- vl_msg_api_send_shmem (q, (u8 *)&mp);
- mp = 0;
- }
- } /* vec_foreach (fib) */
- /* If e.g. the last FIB had no reportable routes, free the buffer */
- if (mp)
- vl_msg_api_free (mp);
+
+ j++;
+ if (j < n_nhs)
+ {
+ n_left = nhs[j].weight;
+ vlib_counter_zero (&sum);
+ }
+ }
+ } /* for each (mp or single) adj */
+ if (sm->data_structure_lock->release_hint)
+ {
+ start_at_fib_index = fib - im4->fibs;
+ dsunlock (sm);
+ ip46_fib_stats_delay (sm, 0 /* sec */ , STATS_RELEASE_DELAY_NS);
+ mp->count = 0;
+ ctrp = (vl_api_ip4_fib_counter_t *) mp->c;
+ goto again;
+ }
+ } /* vec_foreach (routes) */
+
+ dsunlock (sm);
+
+ /* Flush any data from this fib */
+ if (mp->count)
+ {
+ mp->count = htonl (mp->count);
+ vl_msg_api_send_shmem (q, (u8 *) & mp);
+ mp = 0;
+ }
+ } /* vec_foreach (fib) */
+ /* If e.g. the last FIB had no reportable routes, free the buffer */
+ if (mp)
+ vl_msg_api_free (mp);
}
-typedef struct {
+typedef struct
+{
ip6_address_t address;
u32 address_length;
u32 index;
} ip6_route_t;
-typedef struct {
+typedef struct
+{
u32 fib_index;
- ip6_route_t ** routep;
- stats_main_t * sm;
+ ip6_route_t **routep;
+ stats_main_t *sm;
} add_routes_in_fib_arg_t;
-static void add_routes_in_fib (BVT(clib_bihash_kv) * kvp, void *arg)
+static void
+add_routes_in_fib (BVT (clib_bihash_kv) * kvp, void *arg)
{
- add_routes_in_fib_arg_t * ap = arg;
- stats_main_t * sm = ap->sm;
+ add_routes_in_fib_arg_t *ap = arg;
+ stats_main_t *sm = ap->sm;
if (sm->data_structure_lock->release_hint)
clib_longjmp (&sm->jmp_buf, 1);
- if (kvp->key[2]>>32 == ap->fib_index)
+ if (kvp->key[2] >> 32 == ap->fib_index)
{
ip6_address_t *addr;
- ip6_route_t * r;
+ ip6_route_t *r;
addr = (ip6_address_t *) kvp;
vec_add2 (*ap->routep, r, 1);
r->address = addr[0];
@@ -443,457 +480,509 @@ static void add_routes_in_fib (BVT(clib_bihash_kv) * kvp, void *arg)
}
}
-static void do_ip6_fibs (stats_main_t * sm)
+static void
+do_ip6_fibs (stats_main_t * sm)
{
- ip6_main_t * im6 = &ip6_main;
- api_main_t * am = sm->api_main;
- vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
- unix_shared_memory_queue_t * q = shmem_hdr->vl_input_queue;
- static ip6_route_t * routes;
- ip6_route_t * r;
- ip6_fib_t * fib;
- ip_lookup_main_t * lm = &im6->lookup_main;
- static uword * results;
- vl_api_vnet_ip6_fib_counters_t * mp = 0;
- u32 items_this_message;
- vl_api_ip6_fib_counter_t *ctrp = 0;
- u32 start_at_fib_index = 0;
- BVT(clib_bihash) * h = &im6->ip6_lookup_table;
- add_routes_in_fib_arg_t _a, *a=&_a;
+ ip6_main_t *im6 = &ip6_main;
+ api_main_t *am = sm->api_main;
+ vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr;
+ unix_shared_memory_queue_t *q = shmem_hdr->vl_input_queue;
+ static ip6_route_t *routes;
+ ip6_route_t *r;
+ ip6_fib_t *fib;
+ ip_lookup_main_t *lm = &im6->lookup_main;
+ static uword *results;
+ vl_api_vnet_ip6_fib_counters_t *mp = 0;
+ u32 items_this_message;
+ vl_api_ip6_fib_counter_t *ctrp = 0;
+ u32 start_at_fib_index = 0;
+ BVT (clib_bihash) * h = &im6->ip6_lookup_table;
+ add_routes_in_fib_arg_t _a, *a = &_a;
again:
- vec_foreach (fib, im6->fibs) {
- /* We may have bailed out due to control-plane activity */
- while ((fib - im6->fibs) < start_at_fib_index)
- continue;
-
- if (mp == 0) {
- items_this_message = IP6_FIB_COUNTER_BATCH_SIZE;
- mp = vl_msg_api_alloc_as_if_client
- (sizeof(*mp) +
- items_this_message*sizeof(vl_api_ip6_fib_counter_t));
- mp->_vl_msg_id = ntohs (VL_API_VNET_IP6_FIB_COUNTERS);
- mp->count = 0;
- mp->vrf_id = ntohl(fib->table_id);
- ctrp = (vl_api_ip6_fib_counter_t *)mp->c;
- }
-
- dslock (sm, 0 /* release hint */, 1 /* tag */);
-
- vec_reset_length (routes);
- vec_reset_length (results);
-
- a->fib_index = fib - im6->fibs;
- a->routep = &routes;
- a->sm = sm;
-
- if (clib_setjmp(&sm->jmp_buf, 0) == 0)
- {
- start_at_fib_index = fib - im6->fibs;
- BV(clib_bihash_foreach_key_value_pair)(h, add_routes_in_fib, a);
- }
- else
- {
- dsunlock (sm);
- ip46_fib_stats_delay (sm, 0 /* sec */,
- STATS_RELEASE_DELAY_NS);
- mp->count = 0;
- ctrp = (vl_api_ip6_fib_counter_t *)mp->c;
- goto again;
- }
-
- vec_foreach (r, routes) {
- vlib_counter_t c, sum;
- uword i, j, n_left, n_nhs, adj_index, * result = 0;
- ip_adjacency_t * adj;
- ip_multipath_next_hop_t * nhs, tmp_nhs[1];
-
- adj_index = r->index;
- if (lm->fib_result_n_words > 1) {
- result = vec_elt_at_index (results, adj_index);
- adj_index = result[0];
- }
-
- adj = ip_get_adjacency (lm, adj_index);
- if (adj->n_adj == 1) {
- nhs = &tmp_nhs[0];
- nhs[0].next_hop_adj_index = ~0; /* not used */
- nhs[0].weight = 1;
- n_nhs = 1;
- } else {
- ip_multipath_adjacency_t * madj;
- madj = vec_elt_at_index (lm->multipath_adjacencies,
- adj->heap_handle);
- nhs = heap_elt_at_index
- (lm->next_hop_heap,
- madj->normalized_next_hops.heap_offset);
- n_nhs = madj->normalized_next_hops.count;
- }
-
- n_left = nhs[0].weight;
- vlib_counter_zero (&sum);
- for (i = j = 0; i < adj->n_adj; i++) {
- n_left -= 1;
- vlib_get_combined_counter (&lm->adjacency_counters,
- adj_index + i, &c);
- vlib_counter_add (&sum, &c);
- if (n_left == 0 && sum.packets > 0) {
-
- /* already in net byte order */
- ctrp->address[0] = r->address.as_u64[0];
- ctrp->address[1] = r->address.as_u64[1];
- ctrp->address_length = (u8) r->address_length;
- ctrp->packets = clib_host_to_net_u64 (sum.packets);
- ctrp->bytes = clib_host_to_net_u64(sum.bytes);
- mp->count++;
- ctrp++;
-
- if (mp->count == items_this_message) {
- mp->count = htonl (items_this_message);
- /*
- * If the main thread's input queue is stuffed,
- * drop the data structure lock (which the main thread
- * may want), and take a pause.
- */
- unix_shared_memory_queue_lock (q);
- if (unix_shared_memory_queue_is_full (q)) {
- dsunlock (sm);
- vl_msg_api_send_shmem_nolock (q, (u8 *)&mp);
- unix_shared_memory_queue_unlock (q);
- mp = 0;
- ip46_fib_stats_delay (sm, 0 /* sec */,
- STATS_RELEASE_DELAY_NS);
- goto again;
- }
- vl_msg_api_send_shmem_nolock (q, (u8 *)&mp);
- unix_shared_memory_queue_unlock (q);
-
- items_this_message = IP6_FIB_COUNTER_BATCH_SIZE;
- mp = vl_msg_api_alloc_as_if_client
- (sizeof(*mp) +
- items_this_message*
- sizeof(vl_api_ip6_fib_counter_t));
- mp->_vl_msg_id = ntohs (VL_API_VNET_IP6_FIB_COUNTERS);
- mp->count = 0;
- mp->vrf_id = ntohl(fib->table_id);
- ctrp = (vl_api_ip6_fib_counter_t *)mp->c;
- }
-
- j++;
- if (j < n_nhs) {
- n_left = nhs[j].weight;
- vlib_counter_zero (&sum);
+ vec_foreach (fib, im6->fibs)
+ {
+ /* We may have bailed out due to control-plane activity */
+ while ((fib - im6->fibs) < start_at_fib_index)
+ continue;
+
+ if (mp == 0)
+ {
+ items_this_message = IP6_FIB_COUNTER_BATCH_SIZE;
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) +
+ items_this_message * sizeof (vl_api_ip6_fib_counter_t));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_IP6_FIB_COUNTERS);
+ mp->count = 0;
+ mp->vrf_id = ntohl (fib->table_id);
+ ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
+ }
+
+ dslock (sm, 0 /* release hint */ , 1 /* tag */ );
+
+ vec_reset_length (routes);
+ vec_reset_length (results);
+
+ a->fib_index = fib - im6->fibs;
+ a->routep = &routes;
+ a->sm = sm;
+
+ if (clib_setjmp (&sm->jmp_buf, 0) == 0)
+ {
+ start_at_fib_index = fib - im6->fibs;
+ BV (clib_bihash_foreach_key_value_pair) (h, add_routes_in_fib, a);
+ }
+ else
+ {
+ dsunlock (sm);
+ ip46_fib_stats_delay (sm, 0 /* sec */ ,
+ STATS_RELEASE_DELAY_NS);
+ mp->count = 0;
+ ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
+ goto again;
+ }
+
+ vec_foreach (r, routes)
+ {
+ vlib_counter_t c, sum;
+ uword i, j, n_left, n_nhs, adj_index, *result = 0;
+ ip_adjacency_t *adj;
+ ip_multipath_next_hop_t *nhs, tmp_nhs[1];
+
+ adj_index = r->index;
+ if (lm->fib_result_n_words > 1)
+ {
+ result = vec_elt_at_index (results, adj_index);
+ adj_index = result[0];
+ }
+
+ adj = ip_get_adjacency (lm, adj_index);
+ if (adj->n_adj == 1)
+ {
+ nhs = &tmp_nhs[0];
+ nhs[0].next_hop_adj_index = ~0; /* not used */
+ nhs[0].weight = 1;
+ n_nhs = 1;
+ }
+ else
+ {
+ ip_multipath_adjacency_t *madj;
+ madj = vec_elt_at_index (lm->multipath_adjacencies,
+ adj->heap_handle);
+ nhs = heap_elt_at_index
+ (lm->next_hop_heap, madj->normalized_next_hops.heap_offset);
+ n_nhs = madj->normalized_next_hops.count;
+ }
+
+ n_left = nhs[0].weight;
+ vlib_counter_zero (&sum);
+ for (i = j = 0; i < adj->n_adj; i++)
+ {
+ n_left -= 1;
+ vlib_get_combined_counter (&lm->adjacency_counters,
+ adj_index + i, &c);
+ vlib_counter_add (&sum, &c);
+ if (n_left == 0 && sum.packets > 0)
+ {
+
+ /* already in net byte order */
+ ctrp->address[0] = r->address.as_u64[0];
+ ctrp->address[1] = r->address.as_u64[1];
+ ctrp->address_length = (u8) r->address_length;
+ ctrp->packets = clib_host_to_net_u64 (sum.packets);
+ ctrp->bytes = clib_host_to_net_u64 (sum.bytes);
+ mp->count++;
+ ctrp++;
+
+ if (mp->count == items_this_message)
+ {
+ mp->count = htonl (items_this_message);
+ /*
+ * If the main thread's input queue is stuffed,
+ * drop the data structure lock (which the main thread
+ * may want), and take a pause.
+ */
+ unix_shared_memory_queue_lock (q);
+ if (unix_shared_memory_queue_is_full (q))
+ {
+ dsunlock (sm);
+ vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
+ unix_shared_memory_queue_unlock (q);
+ mp = 0;
+ ip46_fib_stats_delay (sm, 0 /* sec */ ,
+ STATS_RELEASE_DELAY_NS);
+ goto again;
}
+ vl_msg_api_send_shmem_nolock (q, (u8 *) & mp);
+ unix_shared_memory_queue_unlock (q);
+
+ items_this_message = IP6_FIB_COUNTER_BATCH_SIZE;
+ mp = vl_msg_api_alloc_as_if_client
+ (sizeof (*mp) +
+ items_this_message * sizeof (vl_api_ip6_fib_counter_t));
+ mp->_vl_msg_id = ntohs (VL_API_VNET_IP6_FIB_COUNTERS);
+ mp->count = 0;
+ mp->vrf_id = ntohl (fib->table_id);
+ ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
}
- } /* for each (mp or single) adj */
- if (sm->data_structure_lock->release_hint) {
- start_at_fib_index = fib - im6->fibs;
- dsunlock (sm);
- ip46_fib_stats_delay (sm, 0 /* sec */, STATS_RELEASE_DELAY_NS);
- mp->count = 0;
- ctrp = (vl_api_ip6_fib_counter_t *)mp->c;
- goto again;
- }
- } /* vec_foreach (routes) */
-
- dsunlock(sm);
-
- /* Flush any data from this fib */
- if (mp->count) {
- mp->count = htonl (mp->count);
- vl_msg_api_send_shmem (q, (u8 *)&mp);
- mp = 0;
- }
- } /* vec_foreach (fib) */
- /* If e.g. the last FIB had no reportable routes, free the buffer */
- if (mp)
- vl_msg_api_free (mp);
+
+ j++;
+ if (j < n_nhs)
+ {
+ n_left = nhs[j].weight;
+ vlib_counter_zero (&sum);
+ }
+ }
+ } /* for each (mp or single) adj */
+ if (sm->data_structure_lock->release_hint)
+ {
+ start_at_fib_index = fib - im6->fibs;
+ dsunlock (sm);
+ ip46_fib_stats_delay (sm, 0 /* sec */ , STATS_RELEASE_DELAY_NS);
+ mp->count = 0;
+ ctrp = (vl_api_ip6_fib_counter_t *) mp->c;
+ goto again;
+ }
+ } /* vec_foreach (routes) */
+
+ dsunlock (sm);
+
+ /* Flush any data from this fib */
+ if (mp->count)
+ {
+ mp->count = htonl (mp->count);
+ vl_msg_api_send_shmem (q, (u8 *) & mp);
+ mp = 0;
+ }
+ } /* vec_foreach (fib) */
+ /* If e.g. the last FIB had no reportable routes, free the buffer */
+ if (mp)
+ vl_msg_api_free (mp);
}
-static void stats_thread_fn (void *arg)
+static void
+stats_thread_fn (void *arg)
{
- stats_main_t *sm = &stats_main;
- vlib_worker_thread_t *w = (vlib_worker_thread_t *)arg;
- vlib_thread_main_t *tm = vlib_get_thread_main();
-
- /* stats thread wants no signals. */
- {
- sigset_t s;
- sigfillset (&s);
- pthread_sigmask (SIG_SETMASK, &s, 0);
- }
+ stats_main_t *sm = &stats_main;
+ vlib_worker_thread_t *w = (vlib_worker_thread_t *) arg;
+ vlib_thread_main_t *tm = vlib_get_thread_main ();
- if (vec_len(tm->thread_prefix))
- vlib_set_thread_name((char *)
- format(0, "%v_stats%c", tm->thread_prefix, '\0'));
+ /* stats thread wants no signals. */
+ {
+ sigset_t s;
+ sigfillset (&s);
+ pthread_sigmask (SIG_SETMASK, &s, 0);
+ }
- clib_mem_set_heap (w->thread_mheap);
+ if (vec_len (tm->thread_prefix))
+ vlib_set_thread_name ((char *)
+ format (0, "%v_stats%c", tm->thread_prefix, '\0'));
- while (1) {
- /* 10 second poll interval */
- ip46_fib_stats_delay (sm, 10 /* secs */, 0 /* nsec */);
+ clib_mem_set_heap (w->thread_mheap);
- if (! (sm->enable_poller))
- continue;
- do_simple_interface_counters (sm);
- do_combined_interface_counters (sm);
- do_ip4_fibs (sm);
- do_ip6_fibs (sm);
+ while (1)
+ {
+ /* 10 second poll interval */
+ ip46_fib_stats_delay (sm, 10 /* secs */ , 0 /* nsec */ );
+
+ if (!(sm->enable_poller))
+ continue;
+ do_simple_interface_counters (sm);
+ do_combined_interface_counters (sm);
+ do_ip4_fibs (sm);
+ do_ip6_fibs (sm);
}
}
-static void vl_api_vnet_interface_counters_t_handler (
- vl_api_vnet_interface_counters_t *mp)
+static void
+vl_api_vnet_interface_counters_t_handler (vl_api_vnet_interface_counters_t *
+ mp)
{
- vpe_client_registration_t *reg;
- stats_main_t * sm = &stats_main;
- unix_shared_memory_queue_t *q, *q_prev = NULL;
- vl_api_vnet_interface_counters_t *mp_copy = NULL;
- u32 mp_size;
+ vpe_client_registration_t *reg;
+ stats_main_t *sm = &stats_main;
+ unix_shared_memory_queue_t *q, *q_prev = NULL;
+ vl_api_vnet_interface_counters_t *mp_copy = NULL;
+ u32 mp_size;
#if STATS_DEBUG > 0
- char *counter_name;
- u32 count, sw_if_index;
- int i;
+ char *counter_name;
+ u32 count, sw_if_index;
+ int i;
#endif
-
- mp_size = sizeof (*mp) + (ntohl(mp->count) *
- (mp->is_combined ? sizeof (vlib_counter_t) : sizeof (u64)));
-
- pool_foreach(reg, sm->stats_registrations,
- ({
- q = vl_api_client_index_to_input_queue (reg->client_index);
- if (q) {
- if (q_prev && (q_prev->cursize < q_prev->maxsize)) {
- mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
- clib_memcpy(mp_copy, mp, mp_size);
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- mp = mp_copy;
- }
- q_prev = q;
- }
- }));
+
+ mp_size = sizeof (*mp) + (ntohl (mp->count) *
+ (mp->is_combined ? sizeof (vlib_counter_t) :
+ sizeof (u64)));
+
+ /* *INDENT-OFF* */
+ pool_foreach(reg, sm->stats_registrations,
+ ({
+ q = vl_api_client_index_to_input_queue (reg->client_index);
+ if (q)
+ {
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
+ clib_memcpy(mp_copy, mp, mp_size);
+ vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
+ mp = mp_copy;
+ }
+ q_prev = q;
+ }
+ }));
+ /* *INDENT-ON* */
#if STATS_DEBUG > 0
- count = ntohl (mp->count);
- sw_if_index = ntohl (mp->first_sw_if_index);
- if (mp->is_combined == 0) {
- u64 * vp, v;
- vp = (u64 *) mp->data;
-
- switch (mp->vnet_counter_type) {
- case VNET_INTERFACE_COUNTER_DROP:
- counter_name = "drop";
- break;
- case VNET_INTERFACE_COUNTER_PUNT:
- counter_name = "punt";
- break;
- case VNET_INTERFACE_COUNTER_IP4:
- counter_name = "ip4";
- break;
- case VNET_INTERFACE_COUNTER_IP6:
- counter_name = "ip6";
- break;
- case VNET_INTERFACE_COUNTER_RX_NO_BUF:
- counter_name = "rx-no-buff";
- break;
- case VNET_INTERFACE_COUNTER_RX_MISS:,
- counter_name = "rx-miss";
- break;
- case VNET_INTERFACE_COUNTER_RX_ERROR:,
- counter_name = "rx-error (fifo-full)";
- break;
- case VNET_INTERFACE_COUNTER_TX_ERROR:,
- counter_name = "tx-error (fifo-full)";
- break;
- default:
- counter_name = "bogus";
- break;
- }
- for (i = 0; i < count; i++) {
- v = clib_mem_unaligned (vp, u64);
- v = clib_net_to_host_u64 (v);
- vp++;
- fformat (stdout, "%U.%s %lld\n", format_vnet_sw_if_index_name,
- sm->vnet_main, sw_if_index, counter_name, v);
- sw_if_index++;
- }
- } else {
- vlib_counter_t *vp;
- u64 packets, bytes;
- vp = (vlib_counter_t *) mp->data;
-
- switch (mp->vnet_counter_type) {
- case VNET_INTERFACE_COUNTER_RX:
- counter_name = "rx";
- break;
- case VNET_INTERFACE_COUNTER_TX:
- counter_name = "tx";
- break;
- default:
- counter_name = "bogus";
- break;
- }
- for (i = 0; i < count; i++) {
- packets = clib_mem_unaligned (&vp->packets, u64);
- packets = clib_net_to_host_u64 (packets);
- bytes = clib_mem_unaligned (&vp->bytes, u64);
- bytes = clib_net_to_host_u64 (bytes);
- vp++;
- fformat (stdout, "%U.%s.packets %lld\n",
- format_vnet_sw_if_index_name,
- sm->vnet_main, sw_if_index, counter_name, packets);
- fformat (stdout, "%U.%s.bytes %lld\n",
- format_vnet_sw_if_index_name,
- sm->vnet_main, sw_if_index, counter_name, bytes);
- sw_if_index++;
- }
+ count = ntohl (mp->count);
+ sw_if_index = ntohl (mp->first_sw_if_index);
+ if (mp->is_combined == 0)
+ {
+ u64 *vp, v;
+ vp = (u64 *) mp->data;
+
+ switch (mp->vnet_counter_type)
+ {
+ case VNET_INTERFACE_COUNTER_DROP:
+ counter_name = "drop";
+ break;
+ case VNET_INTERFACE_COUNTER_PUNT:
+ counter_name = "punt";
+ break;
+ case VNET_INTERFACE_COUNTER_IP4:
+ counter_name = "ip4";
+ break;
+ case VNET_INTERFACE_COUNTER_IP6:
+ counter_name = "ip6";
+ break;
+ case VNET_INTERFACE_COUNTER_RX_NO_BUF:
+ counter_name = "rx-no-buff";
+ break;
+ case VNET_INTERFACE_COUNTER_RX_MISS:
+ , counter_name = "rx-miss";
+ break;
+ case VNET_INTERFACE_COUNTER_RX_ERROR:
+ , counter_name = "rx-error (fifo-full)";
+ break;
+ case VNET_INTERFACE_COUNTER_TX_ERROR:
+ , counter_name = "tx-error (fifo-full)";
+ break;
+ default:
+ counter_name = "bogus";
+ break;
+ }
+ for (i = 0; i < count; i++)
+ {
+ v = clib_mem_unaligned (vp, u64);
+ v = clib_net_to_host_u64 (v);
+ vp++;
+ fformat (stdout, "%U.%s %lld\n", format_vnet_sw_if_index_name,
+ sm->vnet_main, sw_if_index, counter_name, v);
+ sw_if_index++;
+ }
+ }
+ else
+ {
+ vlib_counter_t *vp;
+ u64 packets, bytes;
+ vp = (vlib_counter_t *) mp->data;
+
+ switch (mp->vnet_counter_type)
+ {
+ case VNET_INTERFACE_COUNTER_RX:
+ counter_name = "rx";
+ break;
+ case VNET_INTERFACE_COUNTER_TX:
+ counter_name = "tx";
+ break;
+ default:
+ counter_name = "bogus";
+ break;
+ }
+ for (i = 0; i < count; i++)
+ {
+ packets = clib_mem_unaligned (&vp->packets, u64);
+ packets = clib_net_to_host_u64 (packets);
+ bytes = clib_mem_unaligned (&vp->bytes, u64);
+ bytes = clib_net_to_host_u64 (bytes);
+ vp++;
+ fformat (stdout, "%U.%s.packets %lld\n",
+ format_vnet_sw_if_index_name,
+ sm->vnet_main, sw_if_index, counter_name, packets);
+ fformat (stdout, "%U.%s.bytes %lld\n",
+ format_vnet_sw_if_index_name,
+ sm->vnet_main, sw_if_index, counter_name, bytes);
+ sw_if_index++;
+ }
}
#endif
- if (q_prev &&
- (q_prev->cursize < q_prev->maxsize)) {
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- } else {
- vl_msg_api_free (mp);
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ vl_msg_api_send_shmem (q_prev, (u8 *) & mp);
+ }
+ else
+ {
+ vl_msg_api_free (mp);
}
}
-static void vl_api_vnet_ip4_fib_counters_t_handler (
- vl_api_vnet_ip4_fib_counters_t *mp)
+static void
+vl_api_vnet_ip4_fib_counters_t_handler (vl_api_vnet_ip4_fib_counters_t * mp)
{
- vpe_client_registration_t *reg;
- stats_main_t * sm = &stats_main;
- unix_shared_memory_queue_t *q, *q_prev = NULL;
- vl_api_vnet_ip4_fib_counters_t *mp_copy = NULL;
- u32 mp_size;
-
- mp_size = sizeof(*mp_copy) +
- ntohl(mp->count) * sizeof(vl_api_ip4_fib_counter_t);
-
- pool_foreach(reg, sm->stats_registrations,
- ({
- q = vl_api_client_index_to_input_queue (reg->client_index);
- if (q) {
- if (q_prev && (q_prev->cursize < q_prev->maxsize)) {
- mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
- clib_memcpy(mp_copy, mp, mp_size);
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- mp = mp_copy;
- }
- q_prev = q;
- }
- }));
- if (q_prev &&
- (q_prev->cursize < q_prev->maxsize)) {
+ vpe_client_registration_t *reg;
+ stats_main_t *sm = &stats_main;
+ unix_shared_memory_queue_t *q, *q_prev = NULL;
+ vl_api_vnet_ip4_fib_counters_t *mp_copy = NULL;
+ u32 mp_size;
+
+ mp_size = sizeof (*mp_copy) +
+ ntohl (mp->count) * sizeof (vl_api_ip4_fib_counter_t);
+
+ /* *INDENT-OFF* */
+ pool_foreach(reg, sm->stats_registrations,
+ ({
+ q = vl_api_client_index_to_input_queue (reg->client_index);
+ if (q)
+ {
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
+ clib_memcpy(mp_copy, mp, mp_size);
vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- } else {
- vl_msg_api_free (mp);
+ mp = mp_copy;
+ }
+ q_prev = q;
+ }
+ }));
+ /* *INDENT-ON* */
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ vl_msg_api_send_shmem (q_prev, (u8 *) & mp);
+ }
+ else
+ {
+ vl_msg_api_free (mp);
}
}
-static void vl_api_vnet_ip6_fib_counters_t_handler (
- vl_api_vnet_ip6_fib_counters_t *mp)
+static void
+vl_api_vnet_ip6_fib_counters_t_handler (vl_api_vnet_ip6_fib_counters_t * mp)
{
- vpe_client_registration_t *reg;
- stats_main_t * sm = &stats_main;
- unix_shared_memory_queue_t *q, *q_prev = NULL;
- vl_api_vnet_ip6_fib_counters_t *mp_copy = NULL;
- u32 mp_size;
-
- mp_size = sizeof(*mp_copy) +
- ntohl(mp->count) * sizeof(vl_api_ip6_fib_counter_t);
-
- pool_foreach(reg, sm->stats_registrations,
- ({
- q = vl_api_client_index_to_input_queue (reg->client_index);
- if (q) {
- if (q_prev && (q_prev->cursize < q_prev->maxsize)) {
- mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
- clib_memcpy(mp_copy, mp, mp_size);
- vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- mp = mp_copy;
- }
- q_prev = q;
- }
- }));
- if (q_prev &&
- (q_prev->cursize < q_prev->maxsize)) {
+ vpe_client_registration_t *reg;
+ stats_main_t *sm = &stats_main;
+ unix_shared_memory_queue_t *q, *q_prev = NULL;
+ vl_api_vnet_ip6_fib_counters_t *mp_copy = NULL;
+ u32 mp_size;
+
+ mp_size = sizeof (*mp_copy) +
+ ntohl (mp->count) * sizeof (vl_api_ip6_fib_counter_t);
+
+ /* *INDENT-OFF* */
+ pool_foreach(reg, sm->stats_registrations,
+ ({
+ q = vl_api_client_index_to_input_queue (reg->client_index);
+ if (q)
+ {
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ mp_copy = vl_msg_api_alloc_as_if_client(mp_size);
+ clib_memcpy(mp_copy, mp, mp_size);
vl_msg_api_send_shmem (q_prev, (u8 *)&mp);
- } else {
- vl_msg_api_free (mp);
+ mp = mp_copy;
+ }
+ q_prev = q;
+ }
+ }));
+ /* *INDENT-ON* */
+ if (q_prev && (q_prev->cursize < q_prev->maxsize))
+ {
+ vl_msg_api_send_shmem (q_prev, (u8 *) & mp);
+ }
+ else
+ {
+ vl_msg_api_free (mp);
}
}
-static void vl_api_want_stats_reply_t_handler (vl_api_want_stats_reply_t *mp)
-{ clib_warning ("BUG"); }
+static void
+vl_api_want_stats_reply_t_handler (vl_api_want_stats_reply_t * mp)
+{
+ clib_warning ("BUG");
+}
-static void vl_api_want_stats_t_handler (
- vl_api_want_stats_t *mp)
+static void
+vl_api_want_stats_t_handler (vl_api_want_stats_t * mp)
{
- stats_main_t *sm = &stats_main;
- vpe_client_registration_t *rp;
- vl_api_want_stats_reply_t *rmp;
- uword *p;
- i32 retval = 0;
- unix_shared_memory_queue_t *q;
-
- p = hash_get (sm->stats_registration_hash, mp->client_index);
- if (p) {
- if (mp->enable_disable) {
- clib_warning ("pid %d: already enabled...", mp->pid);
- retval = -2;
- goto reply;
- } else {
- rp = pool_elt_at_index (sm->stats_registrations, p[0]);
- pool_put (sm->stats_registrations, rp);
- hash_unset (sm->stats_registration_hash, mp->client_index);
- goto reply;
- }
+ stats_main_t *sm = &stats_main;
+ vpe_client_registration_t *rp;
+ vl_api_want_stats_reply_t *rmp;
+ uword *p;
+ i32 retval = 0;
+ unix_shared_memory_queue_t *q;
+
+ p = hash_get (sm->stats_registration_hash, mp->client_index);
+ if (p)
+ {
+ if (mp->enable_disable)
+ {
+ clib_warning ("pid %d: already enabled...", mp->pid);
+ retval = -2;
+ goto reply;
+ }
+ else
+ {
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+ pool_put (sm->stats_registrations, rp);
+ hash_unset (sm->stats_registration_hash, mp->client_index);
+ goto reply;
+ }
}
- if (mp->enable_disable == 0) {
- clib_warning ("pid %d: already disabled...", mp->pid);
- retval = -3;
- goto reply;
+ if (mp->enable_disable == 0)
+ {
+ clib_warning ("pid %d: already disabled...", mp->pid);
+ retval = -3;
+ goto reply;
}
- pool_get (sm->stats_registrations, rp);
- rp->client_index = mp->client_index;
- rp->client_pid = mp->pid;
- hash_set (sm->stats_registration_hash, rp->client_index,
- rp - sm->stats_registrations);
+ pool_get (sm->stats_registrations, rp);
+ rp->client_index = mp->client_index;
+ rp->client_pid = mp->pid;
+ hash_set (sm->stats_registration_hash, rp->client_index,
+ rp - sm->stats_registrations);
reply:
- if (pool_elts(sm->stats_registrations))
- sm->enable_poller = 1;
- else
- sm->enable_poller = 0;
-
- q = vl_api_client_index_to_input_queue (mp->client_index);
+ if (pool_elts (sm->stats_registrations))
+ sm->enable_poller = 1;
+ else
+ sm->enable_poller = 0;
- if (!q)
- return;
+ q = vl_api_client_index_to_input_queue (mp->client_index);
- rmp = vl_msg_api_alloc (sizeof (*rmp));
- rmp->_vl_msg_id = ntohs(VL_API_WANT_STATS_REPLY);
- rmp->context = mp->context;
- rmp->retval = retval;
+ if (!q)
+ return;
- vl_msg_api_send_shmem (q, (u8 *)&rmp);
+ rmp = vl_msg_api_alloc (sizeof (*rmp));
+ rmp->_vl_msg_id = ntohs (VL_API_WANT_STATS_REPLY);
+ rmp->context = mp->context;
+ rmp->retval = retval;
+
+ vl_msg_api_send_shmem (q, (u8 *) & rmp);
}
-int stats_memclnt_delete_callback (u32 client_index)
+int
+stats_memclnt_delete_callback (u32 client_index)
{
- vpe_client_registration_t *rp;
- stats_main_t *sm = &stats_main;
- uword * p;
-
- p = hash_get (sm->stats_registration_hash, client_index);
- if (p) {
- rp = pool_elt_at_index (sm->stats_registrations, p[0]);
- pool_put (sm->stats_registrations, rp);
- hash_unset (sm->stats_registration_hash, client_index);
+ vpe_client_registration_t *rp;
+ stats_main_t *sm = &stats_main;
+ uword *p;
+
+ p = hash_get (sm->stats_registration_hash, client_index);
+ if (p)
+ {
+ rp = pool_elt_at_index (sm->stats_registrations, p[0]);
+ pool_put (sm->stats_registrations, rp);
+ hash_unset (sm->stats_registration_hash, client_index);
}
- return 0;
+ return 0;
}
#define vl_api_vnet_ip4_fib_counters_t_endian vl_noop_handler
@@ -901,21 +990,22 @@ int stats_memclnt_delete_callback (u32 client_index)
#define vl_api_vnet_ip6_fib_counters_t_endian vl_noop_handler
#define vl_api_vnet_ip6_fib_counters_t_print vl_noop_handler
-static clib_error_t * stats_init (vlib_main_t * vm)
+static clib_error_t *
+stats_init (vlib_main_t * vm)
{
- stats_main_t * sm = &stats_main;
- api_main_t * am = &api_main;
- void *vlib_worker_thread_bootstrap_fn (void *arg);
-
- sm->vlib_main = vm;
- sm->vnet_main = vnet_get_main();
- sm->interface_main = &vnet_get_main()->interface_main;
- sm->api_main = am;
- sm->stats_poll_interval_in_seconds = 10;
- sm->data_structure_lock =
- clib_mem_alloc_aligned (sizeof(data_structure_lock_t),
- CLIB_CACHE_LINE_BYTES);
- memset(sm->data_structure_lock, 0, sizeof (*sm->data_structure_lock));
+ stats_main_t *sm = &stats_main;
+ api_main_t *am = &api_main;
+ void *vlib_worker_thread_bootstrap_fn (void *arg);
+
+ sm->vlib_main = vm;
+ sm->vnet_main = vnet_get_main ();
+ sm->interface_main = &vnet_get_main ()->interface_main;
+ sm->api_main = am;
+ sm->stats_poll_interval_in_seconds = 10;
+ sm->data_structure_lock =
+ clib_mem_alloc_aligned (sizeof (data_structure_lock_t),
+ CLIB_CACHE_LINE_BYTES);
+ memset (sm->data_structure_lock, 0, sizeof (*sm->data_structure_lock));
#define _(N,n) \
vl_msg_api_set_handlers(VL_API_##N, #n, \
@@ -923,25 +1013,35 @@ static clib_error_t * stats_init (vlib_main_t * vm)
vl_noop_handler, \
vl_api_##n##_t_endian, \
vl_api_##n##_t_print, \
- sizeof(vl_api_##n##_t), 0 /* do NOT trace! */);
- foreach_stats_msg;
+ sizeof(vl_api_##n##_t), 0 /* do NOT trace! */);
+ foreach_stats_msg;
#undef _
- /* tell the msg infra not to free these messages... */
- am->message_bounce [VL_API_VNET_INTERFACE_COUNTERS] = 1;
- am->message_bounce [VL_API_VNET_IP4_FIB_COUNTERS] = 1;
- am->message_bounce [VL_API_VNET_IP6_FIB_COUNTERS] = 1;
+ /* tell the msg infra not to free these messages... */
+ am->message_bounce[VL_API_VNET_INTERFACE_COUNTERS] = 1;
+ am->message_bounce[VL_API_VNET_IP4_FIB_COUNTERS] = 1;
+ am->message_bounce[VL_API_VNET_IP6_FIB_COUNTERS] = 1;
- return 0;
+ return 0;
}
VLIB_INIT_FUNCTION (stats_init);
+/* *INDENT-OFF* */
VLIB_REGISTER_THREAD (stats_thread_reg, static) = {
- .name = "stats",
- .function = stats_thread_fn,
- .fixed_count = 1,
- .count = 1,
- .no_data_structure_clone = 1,
- .use_pthreads = 1,
+ .name = "stats",
+ .function = stats_thread_fn,
+ .fixed_count = 1,
+ .count = 1,
+ .no_data_structure_clone = 1,
+ .use_pthreads = 1,
};
+/* *INDENT-ON* */
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */
diff --git a/vpp/stats/stats.h b/vpp/stats/stats.h
index d6046a42..d6b4724a 100644
--- a/vpp/stats/stats.h
+++ b/vpp/stats/stats.h
@@ -25,46 +25,57 @@
#include <vlibmemory/api.h>
#include <vlibmemory/unix_shared_memory_queue.h>
-typedef struct {
- u32 client_index; /* in memclnt registration pool */
- u32 client_pid;
+typedef struct
+{
+ u32 client_index; /* in memclnt registration pool */
+ u32 client_pid;
} vpe_client_registration_t;
-typedef struct {
- volatile u32 lock;
- volatile u32 release_hint;
- u32 thread_id;
- u32 count;
- int tag;
+typedef struct
+{
+ volatile u32 lock;
+ volatile u32 release_hint;
+ u32 thread_id;
+ u32 count;
+ int tag;
} data_structure_lock_t;
-typedef struct {
- void *mheap;
- pthread_t thread_self;
- pthread_t thread_handle;
+typedef struct
+{
+ void *mheap;
+ pthread_t thread_self;
+ pthread_t thread_handle;
- u32 stats_poll_interval_in_seconds;
- u32 enable_poller;
+ u32 stats_poll_interval_in_seconds;
+ u32 enable_poller;
- uword *stats_registration_hash;
- vpe_client_registration_t *stats_registrations;
+ uword *stats_registration_hash;
+ vpe_client_registration_t *stats_registrations;
- /* control-plane data structure lock */
- data_structure_lock_t * data_structure_lock;
+ /* control-plane data structure lock */
+ data_structure_lock_t *data_structure_lock;
- /* bail out of FIB walk if set */
- clib_longjmp_t jmp_buf;
+ /* bail out of FIB walk if set */
+ clib_longjmp_t jmp_buf;
- /* convenience */
- vlib_main_t * vlib_main;
- vnet_main_t * vnet_main;
- vnet_interface_main_t * interface_main;
- api_main_t * api_main;
+ /* convenience */
+ vlib_main_t *vlib_main;
+ vnet_main_t *vnet_main;
+ vnet_interface_main_t *interface_main;
+ api_main_t *api_main;
} stats_main_t;
stats_main_t stats_main;
-void dslock (stats_main_t *sm, int release_hint, int tag);
-void dsunlock (stats_main_t *sm);
+void dslock (stats_main_t * sm, int release_hint, int tag);
+void dsunlock (stats_main_t * sm);
#endif /* __included_stats_h__ */
+
+/*
+ * fd.io coding-style-patch-verification: ON
+ *
+ * Local Variables:
+ * eval: (c-set-style "gnu")
+ * End:
+ */