summaryrefslogtreecommitdiffstats
path: root/src/plugins/perfmon/perfmon.c
diff options
context:
space:
mode:
authorZachary Leaf <zachary.leaf@arm.com>2022-05-23 06:22:27 -0500
committerDamjan Marion <dmarion@0xa5.net>2022-07-12 15:29:23 +0000
commitc7d43a5eb19f2acab900274432cfd0e136d6cb44 (patch)
tree3ad3bd5191a1199430db40cec0b286ab531993be /src/plugins/perfmon/perfmon.c
parent20ac58e5c5ef59bc860270037aa7a3b0966a4ec2 (diff)
perfmon: make less arch dependent
In preparation for enabling perfmon on Arm platforms, move some Intel /arch specific logic into the /intel directory and update the CMake to split the common code from arch specific files. Since the dispatch_wrapper code is very different on Arm/Intel, each arch can provide their own implementation + conduct any additional arch specific config e.g. on Intel, all indexes from the mmap pages are cached. The new method intel_config_dispatch_wrapper conducts this config and returns a pointer to the dispatch wrapper to use. Similarly, is_bundle_supported() looks very different on Arm/Intel, so each implementation is to provide their own arch specific checks. Two new callbacks/function ptrs are added in PERFMON_REGISTER_SOURCE to support this - .bundle_support and .config_dispatch_wrapper. Type: refactor Signed-off-by: Zachary Leaf <zachary.leaf@arm.com> Change-Id: Idd121ddcfd1cc80a57c949cecd64eb2db0ac8be3
Diffstat (limited to 'src/plugins/perfmon/perfmon.c')
-rw-r--r--src/plugins/perfmon/perfmon.c100
1 files changed, 8 insertions, 92 deletions
diff --git a/src/plugins/perfmon/perfmon.c b/src/plugins/perfmon/perfmon.c
index cc978888f02..0643384957e 100644
--- a/src/plugins/perfmon/perfmon.c
+++ b/src/plugins/perfmon/perfmon.c
@@ -213,33 +213,6 @@ error:
return err;
}
-static_always_inline u32
-perfmon_mmap_read_index (const struct perf_event_mmap_page *mmap_page)
-{
- u32 idx;
- u32 seq;
-
- /* See documentation in /usr/include/linux/perf_event.h, for more details
- * but the 2 main important things are:
- * 1) if seq != mmap_page->lock, it means the kernel is currently updating
- * the user page and we need to read it again
- * 2) if idx == 0, it means the perf event is currently turned off and we
- * just need to read the kernel-updated 'offset', otherwise we must also
- * add the current hw value (hence rdmpc) */
- do
- {
- seq = mmap_page->lock;
- CLIB_COMPILER_BARRIER ();
-
- idx = mmap_page->index;
-
- CLIB_COMPILER_BARRIER ();
- }
- while (mmap_page->lock != seq);
-
- return idx;
-}
-
clib_error_t *
perfmon_start (vlib_main_t *vm, perfmon_bundle_t *b)
{
@@ -266,27 +239,17 @@ perfmon_start (vlib_main_t *vm, perfmon_bundle_t *b)
}
if (b->active_type == PERFMON_BUNDLE_TYPE_NODE)
{
- for (int i = 0; i < vec_len (pm->thread_runtimes); i++)
+ vlib_node_function_t *dispatch_wrapper = NULL;
+ err = b->src->config_dispatch_wrapper (b, &dispatch_wrapper);
+ if (err || !dispatch_wrapper)
{
- perfmon_thread_runtime_t *tr;
- tr = vec_elt_at_index (pm->thread_runtimes, i);
-
- for (int j = 0; j < b->n_events; j++)
- {
- tr->indexes[j] = perfmon_mmap_read_index (tr->mmap_pages[j]);
-
- /* if a zero index is returned generate error */
- if (!tr->indexes[j])
- {
- perfmon_reset (vm);
- return clib_error_return (0, "invalid rdpmc index");
- }
- }
+ perfmon_reset (vm);
+ return err;
}
for (int i = 0; i < vlib_get_n_threads (); i++)
- vlib_node_set_dispatch_wrapper (
- vlib_get_main_by_index (i), perfmon_dispatch_wrappers[b->n_events]);
+ vlib_node_set_dispatch_wrapper (vlib_get_main_by_index (i),
+ dispatch_wrapper);
}
pm->sample_time = vlib_time_now (vm);
pm->is_running = 1;
@@ -324,53 +287,6 @@ perfmon_stop (vlib_main_t *vm)
return 0;
}
-static_always_inline u8
-is_enough_counters (perfmon_bundle_t *b)
-{
- u8 bl[PERFMON_EVENT_TYPE_MAX];
- u8 cpu[PERFMON_EVENT_TYPE_MAX];
-
- clib_memset (&bl, 0, sizeof (bl));
- clib_memset (&cpu, 0, sizeof (cpu));
-
- /* how many does this uarch support */
- if (!clib_get_pmu_counter_count (&cpu[PERFMON_EVENT_TYPE_FIXED],
- &cpu[PERFMON_EVENT_TYPE_GENERAL]))
- return 0;
-
- /* how many does the bundle require */
- for (u16 i = 0; i < b->n_events; i++)
- {
- /* if source allows us to identify events, otherwise assume general */
- if (b->src->get_event_type)
- bl[b->src->get_event_type (b->events[i])]++;
- else
- bl[PERFMON_EVENT_TYPE_GENERAL]++;
- }
-
- /* consciously ignoring pseudo events here */
- return cpu[PERFMON_EVENT_TYPE_GENERAL] >= bl[PERFMON_EVENT_TYPE_GENERAL] &&
- cpu[PERFMON_EVENT_TYPE_FIXED] >= bl[PERFMON_EVENT_TYPE_FIXED];
-}
-
-static_always_inline u8
-is_bundle_supported (perfmon_bundle_t *b)
-{
- perfmon_cpu_supports_t *supports = b->cpu_supports;
-
- if (!is_enough_counters (b))
- return 0;
-
- if (!b->cpu_supports)
- return 1;
-
- for (int i = 0; i < b->n_cpu_supports; ++i)
- if (supports[i].cpu_supports ())
- return 1;
-
- return 0;
-}
-
static clib_error_t *
perfmon_init (vlib_main_t *vm)
{
@@ -413,7 +329,7 @@ perfmon_init (vlib_main_t *vm)
}
b->src = (perfmon_source_t *) p[0];
- if (!is_bundle_supported (b))
+ if (b->src->bundle_support && !b->src->bundle_support (b))
{
log_debug ("skipping bundle '%s' - not supported", b->name);
b = b->next;