| /linux/arch/x86/kvm/ |
| H A D | pmu.c | 450 static bool filter_contains_match(u64 *events, u64 nevents, u64 eventsel) in filter_contains_match() argument 452 u64 event_select = eventsel & kvm_pmu_ops.EVENTSEL_EVENT; in filter_contains_match() 453 u64 umask = eventsel & ARCH_PERFMON_EVENTSEL_UMASK; in filter_contains_match() 484 u64 eventsel) in is_gp_event_allowed() argument 486 if (filter_contains_match(f->includes, f->nr_includes, eventsel) && in is_gp_event_allowed() 487 !filter_contains_match(f->excludes, f->nr_excludes, eventsel)) in is_gp_event_allowed() 518 return is_gp_event_allowed(filter, pmc->eventsel); in pmc_is_event_allowed() 531 pmc->eventsel_hw |= pmc->eventsel & in kvm_mediated_pmu_refresh_event_filter() 545 u64 eventsel = pmc->eventsel; in reprogram_counter() local 546 u64 new_config = eventsel; in reprogram_counter() [all …]
|
| H A D | pmu.h | 193 return pmc->eventsel & ARCH_PERFMON_EVENTSEL_ENABLE; in pmc_is_locally_enabled()
|
| /linux/tools/testing/selftests/kvm/include/x86/ |
| H A D | pmu.h | 21 #define RAW_EVENT(eventsel, umask) (((eventsel & 0xf00UL) << 24) | \ argument 22 ((eventsel) & 0xff) | \
|
| /linux/arch/arm64/kvm/ |
| H A D | pmu-emul.c | 665 static int kvm_map_pmu_event(struct kvm *kvm, unsigned int eventsel) in kvm_map_pmu_event() argument 674 return pmu->map_pmuv3_event(eventsel); in kvm_map_pmu_event() 676 return eventsel; in kvm_map_pmu_event() 689 int eventsel; in kvm_pmu_create_perf_event() local 696 eventsel = ARMV8_PMUV3_PERFCTR_CPU_CYCLES; in kvm_pmu_create_perf_event() 698 eventsel = evtreg & kvm_pmu_event_mask(vcpu->kvm); in kvm_pmu_create_perf_event() 704 if (eventsel == ARMV8_PMUV3_PERFCTR_SW_INCR || in kvm_pmu_create_perf_event() 705 eventsel == ARMV8_PMUV3_PERFCTR_CHAIN) in kvm_pmu_create_perf_event() 713 !test_bit(eventsel, vcpu->kvm->arch.pmu_filter)) in kvm_pmu_create_perf_event() 720 eventsel = kvm_map_pmu_event(vcpu->kvm, eventsel); in kvm_pmu_create_perf_event() [all …]
|
| /linux/arch/x86/kvm/vmx/ |
| H A D | pmu_intel.c | 360 msr_info->data = pmc->eventsel; in intel_pmu_get_msr() 432 if (data != pmc->eventsel) { in intel_pmu_set_msr() 433 pmc->eventsel = data; in intel_pmu_set_msr() 468 u64 eventsel; in intel_get_fixed_pmc_eventsel() local 477 eventsel = perf_get_hw_event_config(fixed_pmc_perf_ids[index]); in intel_get_fixed_pmc_eventsel() 478 WARN_ON_ONCE(!eventsel && index < kvm_pmu_cap.num_counters_fixed); in intel_get_fixed_pmc_eventsel() 479 return eventsel; in intel_get_fixed_pmc_eventsel() 613 pmu->fixed_counters[i].eventsel = intel_get_fixed_pmc_eventsel(i); in intel_pmu_init()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | pmu.c | 143 msr_info->data = pmc->eventsel; in amd_pmu_get_msr() 167 if (data != pmc->eventsel) { in amd_pmu_set_msr() 168 pmc->eventsel = data; in amd_pmu_set_msr()
|
| /linux/arch/x86/events/amd/ |
| H A D | core.c | 317 static inline int amd_pmu_addr_offset(int index, bool eventsel) in amd_pmu_addr_offset() argument 324 if (eventsel) in amd_pmu_addr_offset() 337 if (eventsel) in amd_pmu_addr_offset() 1324 .eventsel = MSR_K7_EVNTSEL0, 1426 x86_pmu.eventsel = MSR_F15H_PERF_CTL; in amd_core_pmu_init()
|
| /linux/drivers/perf/ |
| H A D | apple_m1_cpu_pmu.c | 568 static int m1_pmu_map_pmuv3_event(unsigned int eventsel) in m1_pmu_map_pmuv3_event() argument 572 if (eventsel < ARMV8_PMUV3_MAX_COMMON_EVENTS) in m1_pmu_map_pmuv3_event() 573 m1_event = m1_pmu_pmceid_map[eventsel]; in m1_pmu_map_pmuv3_event()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | df_v3_6.c | 412 uint32_t eventsel, instance, unitmask; in df_v3_6_pmc_get_ctrl_settings() local 424 eventsel = DF_V3_6_GET_EVENT(config) & 0x3f; in df_v3_6_pmc_get_ctrl_settings() 432 *lo_val = (unitmask << 8) | (instance_10 << 6) | eventsel; in df_v3_6_pmc_get_ctrl_settings()
|
| /linux/include/linux/perf/ |
| H A D | arm_pmu.h | 110 int (*map_pmuv3_event)(unsigned int eventsel);
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | pmu_counters_test.c | 292 uint64_t eventsel = ARCH_PERFMON_EVENTSEL_OS | in guest_test_arch_event() local 301 MSR_P6_EVNTSEL0 + i, eventsel); in guest_test_arch_event()
|
| /linux/arch/x86/events/intel/ |
| H A D | knc.c | 301 .eventsel = MSR_KNC_EVNTSEL0,
|
| H A D | p4.c | 1346 .eventsel = MSR_P4_BPU_CCCR0,
|
| H A D | core.c | 6294 .eventsel = MSR_ARCH_PERFMON_EVENTSEL0, 6348 .eventsel = MSR_ARCH_PERFMON_EVENTSEL0, 7376 static inline int intel_pmu_v6_addr_offset(int index, bool eventsel) in intel_pmu_v6_addr_offset() argument 8517 x86_pmu.eventsel = MSR_IA32_PMC_V6_GP0_CFG_A; in intel_pmu_init()
|
| /linux/arch/x86/events/ |
| H A D | perf_event.h | 845 unsigned eventsel; member 848 int (*addr_offset)(int index, bool eventsel); 1216 return x86_pmu.eventsel + (x86_pmu.addr_offset ? in x86_pmu_config_addr()
|
| /linux/arch/x86/events/zhaoxin/ |
| H A D | core.c | 468 .eventsel = MSR_ARCH_PERFMON_EVENTSEL0,
|
| /linux/arch/x86/include/asm/ |
| H A D | kvm_host.h | 539 u64 eventsel; member
|