perf/x86: Hybrid PMU support for hardware cache event
authorKan Liang <kan.liang@linux.intel.com>
Mon, 12 Apr 2021 14:30:48 +0000 (07:30 -0700)
committerPeter Zijlstra <peterz@infradead.org>
Mon, 19 Apr 2021 18:03:25 +0000 (20:03 +0200)
The hardware cache events are different among hybrid PMUs. Each hybrid
PMU should have its own hw cache event table.

Suggested-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Signed-off-by: Kan Liang <kan.liang@linux.intel.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Link: https://lkml.kernel.org/r/1618237865-33448-9-git-send-email-kan.liang@linux.intel.com
arch/x86/events/core.c
arch/x86/events/perf_event.h

index 1aeb31c..e8cb892 100644 (file)
@@ -376,8 +376,7 @@ set_ext_hw_attr(struct hw_perf_event *hwc, struct perf_event *event)
                return -EINVAL;
        cache_result = array_index_nospec(cache_result, PERF_COUNT_HW_CACHE_RESULT_MAX);
 
-       val = hw_cache_event_ids[cache_type][cache_op][cache_result];
-
+       val = hybrid_var(event->pmu, hw_cache_event_ids)[cache_type][cache_op][cache_result];
        if (val == 0)
                return -ENOENT;
 
@@ -385,7 +384,7 @@ set_ext_hw_attr(struct hw_perf_event *hwc, struct perf_event *event)
                return -EINVAL;
 
        hwc->config |= val;
-       attr->config1 = hw_cache_extra_regs[cache_type][cache_op][cache_result];
+       attr->config1 = hybrid_var(event->pmu, hw_cache_extra_regs)[cache_type][cache_op][cache_result];
        return x86_pmu_extra_regs(val, event);
 }
 
index 2688e45..b65cf46 100644 (file)
@@ -639,6 +639,15 @@ struct x86_hybrid_pmu {
        int                             num_counters;
        int                             num_counters_fixed;
        struct event_constraint         unconstrained;
+
+       u64                             hw_cache_event_ids
+                                       [PERF_COUNT_HW_CACHE_MAX]
+                                       [PERF_COUNT_HW_CACHE_OP_MAX]
+                                       [PERF_COUNT_HW_CACHE_RESULT_MAX];
+       u64                             hw_cache_extra_regs
+                                       [PERF_COUNT_HW_CACHE_MAX]
+                                       [PERF_COUNT_HW_CACHE_OP_MAX]
+                                       [PERF_COUNT_HW_CACHE_RESULT_MAX];
 };
 
 static __always_inline struct x86_hybrid_pmu *hybrid_pmu(struct pmu *pmu)