Home
last modified time | relevance | path

Searched refs:cache_result (Results 1 – 13 of 13) sorted by relevance

/linux/arch/xtensa/kernel/
H A Dperf_event.c115 unsigned int cache_type, cache_op, cache_result; in xtensa_pmu_cache_event() local
120 cache_result = (config >> 16) & 0xff; in xtensa_pmu_cache_event()
124 cache_result >= C(RESULT_MAX)) in xtensa_pmu_cache_event()
127 ret = xtensa_cache_ctl[cache_type][cache_op][cache_result]; in xtensa_pmu_cache_event()
/linux/arch/arc/kernel/
H A Dperf_event.c303 unsigned int cache_type, cache_op, cache_result; in arc_pmu_cache_event() local
308 cache_result = (config >> 16) & 0xff; in arc_pmu_cache_event()
313 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in arc_pmu_cache_event()
316 ret = arc_pmu_cache_map[cache_type][cache_op][cache_result]; in arc_pmu_cache_event()
322 cache_type, cache_op, cache_result, ret, in arc_pmu_cache_event()
/linux/arch/loongarch/kernel/
H A Dperf_event.c603 unsigned int cache_type, cache_op, cache_result; in loongarch_pmu_map_cache_event() local
614 cache_result = (config >> 16) & 0xff; in loongarch_pmu_map_cache_event()
615 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in loongarch_pmu_map_cache_event()
621 [cache_result]); in loongarch_pmu_map_cache_event()
/linux/tools/perf/pmu-events/
H A Dmake_legacy_cache.py66 cache_id: int, cache_op: int, cache_result: int,
/linux/drivers/perf/
H A Darm_pmu.c131 unsigned int cache_type, cache_op, cache_result, ret; in armpmu_map_cache_event() local
141 cache_result = (config >> 16) & 0xff; in armpmu_map_cache_event()
142 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in armpmu_map_cache_event()
148 ret = (int)(*cache_map)[cache_type][cache_op][cache_result]; in armpmu_map_cache_event()
H A Driscv_pmu_sbi.c612 unsigned int cache_type, cache_op, cache_result, ret; in pmu_event_find_cache() local
622 cache_result = (config >> 16) & 0xff; in pmu_event_find_cache()
623 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in pmu_event_find_cache()
626 ret = pmu_cache_event_map[cache_type][cache_op][cache_result].event_idx; in pmu_event_find_cache()
/linux/arch/riscv/kvm/
H A Dvcpu_pmu.c98 unsigned int cache_type, cache_op, cache_result; in kvm_pmu_get_perf_event_cache_config() local
105 cache_result = sbi_event_code & SBI_PMU_EVENT_CACHE_RESULT_ID_CODE_MASK; in kvm_pmu_get_perf_event_cache_config()
109 cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in kvm_pmu_get_perf_event_cache_config()
112 config = cache_type | (cache_op << 8) | (cache_result << 16); in kvm_pmu_get_perf_event_cache_config()
/linux/tools/perf/util/
H A Dparse-events.c367 int len, cache_type = -1, cache_op = -1, cache_result = -1; in parse_events__decode_legacy_cache() local
384 cache_result = parse_aliases(str, evsel__hw_cache_result, in parse_events__decode_legacy_cache()
386 if (cache_result >= 0) in parse_events__decode_legacy_cache()
398 } else if (cache_result < 0) { in parse_events__decode_legacy_cache()
399 cache_result = parse_aliases(str, evsel__hw_cache_result, in parse_events__decode_legacy_cache()
413 if (cache_result == -1) in parse_events__decode_legacy_cache()
414 cache_result = PERF_COUNT_HW_CACHE_RESULT_ACCESS; in parse_events__decode_legacy_cache()
416 *config = cache_type | (cache_op << 8) | (cache_result << 16); in parse_events__decode_legacy_cache()
1022 int cache_type, cache_op, cache_result; in config_term_pmu() local
1028 cache_result = (term->val.num >> 16) & 0xFF; in config_term_pmu()
[all …]
H A Dpmu.c1553 int cache_result = (term->val.num >> 16) & 0xFF; in pmu_config_term() local
1557 assert(cache_result < PERF_COUNT_HW_CACHE_RESULT_MAX); in pmu_config_term()
/linux/arch/csky/kernel/
H A Dperf_event.c952 unsigned int cache_type, cache_op, cache_result; in csky_pmu_cache_event() local
956 cache_result = (config >> 16) & 0xff; in csky_pmu_cache_event()
962 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in csky_pmu_cache_event()
965 return csky_pmu_cache_map[cache_type][cache_op][cache_result]; in csky_pmu_cache_event()
/linux/arch/x86/events/
H A Dcore.c381 unsigned int cache_type, cache_op, cache_result; in set_ext_hw_attr() local
396 cache_result = (config >> 16) & 0xff; in set_ext_hw_attr()
397 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in set_ext_hw_attr()
399 cache_result = array_index_nospec(cache_result, PERF_COUNT_HW_CACHE_RESULT_MAX); in set_ext_hw_attr()
401 val = hybrid_var(event->pmu, hw_cache_event_ids)[cache_type][cache_op][cache_result]; in set_ext_hw_attr()
409 attr->config1 = hybrid_var(event->pmu, hw_cache_extra_regs)[cache_type][cache_op][cache_result]; in set_ext_hw_attr()
/linux/arch/mips/kernel/
H A Dperf_event_mipsxx.c733 unsigned int cache_type, cache_op, cache_result; in mipspmu_map_cache_event() local
744 cache_result = (config >> 16) & 0xff; in mipspmu_map_cache_event()
745 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in mipspmu_map_cache_event()
751 [cache_result]); in mipspmu_map_cache_event()
/linux/arch/sparc/kernel/
H A Dperf_event.c1198 unsigned int cache_type, cache_op, cache_result; in sparc_map_cache_event() local
1212 cache_result = (config >> 16) & 0xff; in sparc_map_cache_event()
1213 if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX) in sparc_map_cache_event()
1216 pmap = &((*sparc_pmu->cache_map)[cache_type][cache_op][cache_result]); in sparc_map_cache_event()