| /linux/arch/xtensa/kernel/ |
| H A D | perf_event.c | 115 unsigned int cache_type, cache_op, cache_result; in xtensa_pmu_cache_event() local 120 cache_result = (config >> 16) & 0xff; in xtensa_pmu_cache_event() 124 cache_result >= C(RESULT_MAX)) in xtensa_pmu_cache_event() 127 ret = xtensa_cache_ctl[cache_type][cache_op][cache_result]; in xtensa_pmu_cache_event()
|
| /linux/arch/arc/kernel/ |
| H A D | perf_event.c | 303 unsigned int cache_type, cache_op, cache_result; in arc_pmu_cache_event() local 308 cache_result = (config >> 16) & 0xff; in arc_pmu_cache_event() 313 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in arc_pmu_cache_event() 316 ret = arc_pmu_cache_map[cache_type][cache_op][cache_result]; in arc_pmu_cache_event() 322 cache_type, cache_op, cache_result, ret, in arc_pmu_cache_event()
|
| /linux/arch/loongarch/kernel/ |
| H A D | perf_event.c | 603 unsigned int cache_type, cache_op, cache_result; in loongarch_pmu_map_cache_event() local 614 cache_result = (config >> 16) & 0xff; in loongarch_pmu_map_cache_event() 615 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in loongarch_pmu_map_cache_event() 621 [cache_result]); in loongarch_pmu_map_cache_event()
|
| /linux/tools/perf/pmu-events/ |
| H A D | make_legacy_cache.py | 66 cache_id: int, cache_op: int, cache_result: int,
|
| /linux/drivers/perf/ |
| H A D | arm_pmu.c | 131 unsigned int cache_type, cache_op, cache_result, ret; in armpmu_map_cache_event() local 141 cache_result = (config >> 16) & 0xff; in armpmu_map_cache_event() 142 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in armpmu_map_cache_event() 148 ret = (int)(*cache_map)[cache_type][cache_op][cache_result]; in armpmu_map_cache_event()
|
| H A D | riscv_pmu_sbi.c | 612 unsigned int cache_type, cache_op, cache_result, ret; in pmu_event_find_cache() local 622 cache_result = (config >> 16) & 0xff; in pmu_event_find_cache() 623 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in pmu_event_find_cache() 626 ret = pmu_cache_event_map[cache_type][cache_op][cache_result].event_idx; in pmu_event_find_cache()
|
| /linux/arch/riscv/kvm/ |
| H A D | vcpu_pmu.c | 98 unsigned int cache_type, cache_op, cache_result; in kvm_pmu_get_perf_event_cache_config() local 105 cache_result = sbi_event_code & SBI_PMU_EVENT_CACHE_RESULT_ID_CODE_MASK; in kvm_pmu_get_perf_event_cache_config() 109 cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in kvm_pmu_get_perf_event_cache_config() 112 config = cache_type | (cache_op << 8) | (cache_result << 16); in kvm_pmu_get_perf_event_cache_config()
|
| /linux/tools/perf/util/ |
| H A D | parse-events.c | 367 int len, cache_type = -1, cache_op = -1, cache_result = -1; in parse_events__decode_legacy_cache() local 384 cache_result = parse_aliases(str, evsel__hw_cache_result, in parse_events__decode_legacy_cache() 386 if (cache_result >= 0) in parse_events__decode_legacy_cache() 398 } else if (cache_result < 0) { in parse_events__decode_legacy_cache() 399 cache_result = parse_aliases(str, evsel__hw_cache_result, in parse_events__decode_legacy_cache() 413 if (cache_result == -1) in parse_events__decode_legacy_cache() 414 cache_result = PERF_COUNT_HW_CACHE_RESULT_ACCESS; in parse_events__decode_legacy_cache() 416 *config = cache_type | (cache_op << 8) | (cache_result << 16); in parse_events__decode_legacy_cache() 1022 int cache_type, cache_op, cache_result; in config_term_pmu() local 1028 cache_result = (term->val.num >> 16) & 0xFF; in config_term_pmu() [all …]
|
| H A D | pmu.c | 1553 int cache_result = (term->val.num >> 16) & 0xFF; in pmu_config_term() local 1557 assert(cache_result < PERF_COUNT_HW_CACHE_RESULT_MAX); in pmu_config_term()
|
| /linux/arch/csky/kernel/ |
| H A D | perf_event.c | 952 unsigned int cache_type, cache_op, cache_result; in csky_pmu_cache_event() local 956 cache_result = (config >> 16) & 0xff; in csky_pmu_cache_event() 962 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in csky_pmu_cache_event() 965 return csky_pmu_cache_map[cache_type][cache_op][cache_result]; in csky_pmu_cache_event()
|
| /linux/arch/x86/events/ |
| H A D | core.c | 381 unsigned int cache_type, cache_op, cache_result; in set_ext_hw_attr() local 396 cache_result = (config >> 16) & 0xff; in set_ext_hw_attr() 397 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in set_ext_hw_attr() 399 cache_result = array_index_nospec(cache_result, PERF_COUNT_HW_CACHE_RESULT_MAX); in set_ext_hw_attr() 401 val = hybrid_var(event->pmu, hw_cache_event_ids)[cache_type][cache_op][cache_result]; in set_ext_hw_attr() 409 attr->config1 = hybrid_var(event->pmu, hw_cache_extra_regs)[cache_type][cache_op][cache_result]; in set_ext_hw_attr()
|
| /linux/arch/mips/kernel/ |
| H A D | perf_event_mipsxx.c | 733 unsigned int cache_type, cache_op, cache_result; in mipspmu_map_cache_event() local 744 cache_result = (config >> 16) & 0xff; in mipspmu_map_cache_event() 745 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in mipspmu_map_cache_event() 751 [cache_result]); in mipspmu_map_cache_event()
|
| /linux/arch/sparc/kernel/ |
| H A D | perf_event.c | 1198 unsigned int cache_type, cache_op, cache_result; in sparc_map_cache_event() local 1212 cache_result = (config >> 16) & 0xff; in sparc_map_cache_event() 1213 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in sparc_map_cache_event() 1216 pmap = &((*sparc_pmu->cache_map)[cache_type][cache_op][cache_result]); in sparc_map_cache_event()
|