| /linux/arch/arm64/kvm/hyp/include/hyp/ ! |
| H A D | switch.h | 168 if (!cpus_have_final_cap(ARM64_SVE)) in __deactivate_cptr_traps_nvhe() 170 if (!cpus_have_final_cap(ARM64_SME)) in __deactivate_cptr_traps_nvhe() 180 if (cpus_have_final_cap(ARM64_SVE)) in __deactivate_cptr_traps_vhe() 182 if (cpus_have_final_cap(ARM64_SME)) in __deactivate_cptr_traps_vhe() 214 if (!cpus_have_final_cap(ARM64_HAS_FGT)) in __activate_traps_hfgxtr() 226 if (!cpus_have_final_cap(ARM64_HAS_FGT2)) in __activate_traps_hfgxtr() 246 if (!cpus_have_final_cap(ARM64_HAS_FGT)) in __deactivate_traps_hfgxtr() 258 if (!cpus_have_final_cap(ARM64_HAS_FGT2)) in __deactivate_traps_hfgxtr() 318 if (cpus_have_final_cap(ARM64_HAS_HCX)) { in __activate_traps_common() 344 if (cpus_have_final_cap(ARM64_HAS_HCX)) in __deactivate_traps_common() [all …]
|
| H A D | sysreg-sr.h | 83 if (!cpus_have_final_cap(ARM64_HAS_S1PIE)) in ctxt_has_s1pie() 94 if (!cpus_have_final_cap(ARM64_HAS_TCR2)) in ctxt_has_tcrx() 116 if (!cpus_have_final_cap(ARM64_HAS_RAS_EXTN)) in ctxt_has_ras() 127 if (!cpus_have_final_cap(ARM64_HAS_SCTLR2)) in ctxt_has_sctlr2() 187 if (!cpus_have_final_cap(ARM64_HAS_RAS_EXTN)) in __sysreg_save_el2_return_state() 218 !cpus_have_final_cap(ARM64_WORKAROUND_SPECULATIVE_AT)) { in __sysreg_restore_el1_state() 265 cpus_have_final_cap(ARM64_WORKAROUND_SPECULATIVE_AT) && in __sysreg_restore_el1_state() 330 if (!cpus_have_final_cap(ARM64_HAS_RAS_EXTN)) in __sysreg_restore_el2_return_state()
|
| H A D | fault.h | 69 if (cpus_have_final_cap(ARM64_WORKAROUND_834220) && in __hpfar_valid()
|
| /linux/arch/arm64/include/asm/ ! |
| H A D | virt.h | 145 return cpus_have_final_cap(ARM64_HAS_VIRT_HOST_EXTN); in has_vhe() 153 return cpus_have_final_cap(ARM64_KVM_PROTECTED_MODE); in is_protected_kvm_enabled() 161 return cpus_have_final_cap(ARM64_KVM_HVHE); in has_hvhe()
|
| H A D | cpufeature.h | 519 static __always_inline bool cpus_have_final_cap(int num) in cpus_have_final_cap() function 825 return cpus_have_final_cap(ARM64_BTI); in system_supports_bti() 841 return cpus_have_final_cap(ARM64_HAS_LPA2); in system_supports_lpa2() 856 return cpus_have_final_cap(ARM64_HAFT); in system_supports_haft() 871 return cpus_have_final_cap(ARM64_HAS_PMUV3); in system_supports_pmuv3()
|
| H A D | kvm_mmu.h | 225 if (cpus_have_final_cap(ARM64_HAS_STAGE2_FWB)) in __clean_dcache_guest_page() 390 return cpus_have_final_cap(ARM64_HAS_STAGE2_FWB) && in kvm_supports_cacheable_pfnmap() 391 cpus_have_final_cap(ARM64_HAS_CACHE_DIC); in kvm_supports_cacheable_pfnmap()
|
| H A D | kvm_emulate.h | 115 if (!cpus_have_final_cap(ARM64_HAS_STAGE2_FWB)) in vcpu_reset_hcr() 198 return (!cpus_have_final_cap(ARM64_HAS_HCR_NV1) || in vcpu_el2_e2h_is_set() 663 if (cpus_have_final_cap(ARM64_HAS_HCX)) { in vcpu_set_hcrx()
|
| H A D | module.h | 48 return cpus_have_final_cap(ARM64_WORKAROUND_843419) && in is_forbidden_offset_for_adrp()
|
| H A D | kvm_nested.h | 13 cpus_have_final_cap(ARM64_HAS_NESTED_VIRT) && in vcpu_has_nv() 125 return cpus_have_final_cap(ARM64_HAS_XNX) && in kvm_has_xnx()
|
| H A D | kvm_pgtable.h | 72 ({ cpus_have_final_cap(ARM64_KVM_HVHE) ? 2 : 3; }) 74 ({ cpus_have_final_cap(ARM64_KVM_HVHE) ? 0 : 1; })
|
| H A D | barrier.h | 81 if (cpus_have_final_cap(ARM64_WORKAROUND_TSB_FLUSH_FAILURE)) \
|
| H A D | kvm_host.h | 1119 ((cpus_have_final_cap(ARM64_HAS_ADDRESS_AUTH) || \ 1120 cpus_have_final_cap(ARM64_HAS_GENERIC_AUTH)) && \ 1151 if (unlikely(cpus_have_final_cap(ARM64_HAS_NESTED_VIRT) && in ___ctxt_sys_reg() 1404 return cpus_have_final_cap(ARM64_SPECTRE_V3A); in kvm_system_needs_idmapped_vectors()
|
| /linux/arch/arm64/kvm/ ! |
| H A D | arm.c | 424 r = cpus_have_final_cap(ARM64_HAS_32BIT_EL1); in kvm_vm_ioctl_check_extension() 427 r = cpus_have_final_cap(ARM64_HAS_NESTED_VIRT); in kvm_vm_ioctl_check_extension() 430 r = cpus_have_final_cap(ARM64_HAS_HCR_NV1); in kvm_vm_ioctl_check_extension() 442 r = cpus_have_final_cap(ARM64_HAS_RAS_EXTN); in kvm_vm_ioctl_check_extension() 1496 if (!cpus_have_final_cap(ARM64_HAS_32BIT_EL1)) in system_supported_vcpu_features() 1510 if (!cpus_have_final_cap(ARM64_HAS_NESTED_VIRT)) in system_supported_vcpu_features() 1666 if (!cpus_have_final_cap(ARM64_HAS_STAGE2_FWB)) in kvm_arch_vcpu_ioctl_vcpu_init() 2077 if (cpus_have_final_cap(ARM64_KVM_HVHE)) { in cpu_prepare_hyp_mode() 2102 if (cpus_have_final_cap(ARM64_KVM_HVHE)) in cpu_prepare_hyp_mode() 2768 cpus_have_final_cap(ARM64_HAS_ADDRESS_AUTH)) in init_hyp_mode() [all …]
|
| H A D | arch_timer.c | 264 return (cpus_have_final_cap(ARM64_HAS_WFXT) && in vcpu_has_wfit_active() 788 if (cpus_have_final_cap(ARM64_HAS_ECV) && is_hyp_ctxt(vcpu)) { in timer_set_traps() 951 if (!cpus_have_final_cap(ARM64_HAS_ECV)) { in kvm_timer_sync_nested() 1396 cpus_have_final_cap(ARM64_WORKAROUND_QCOM_ORYON_CNTVOFF)) { in kvm_timer_handle_errata() 1591 if (cpus_have_final_cap(ARM64_HAS_ECV_CNTPOFF)) in kvm_timer_init_vhe()
|
| H A D | at.c | 577 if (cpus_have_final_cap(ARM64_HAS_TCR2)) { in __mmu_config_save() 579 if (cpus_have_final_cap(ARM64_HAS_S1PIE)) { in __mmu_config_save() 605 if (cpus_have_final_cap(ARM64_HAS_TCR2)) { in __mmu_config_restore() 607 if (cpus_have_final_cap(ARM64_HAS_S1PIE)) { in __mmu_config_restore() 1757 if (cpus_have_final_cap(ARM64_HAS_LSE_ATOMICS)) in __kvm_at_swap_desc()
|
| /linux/arch/arm64/kvm/hyp/nvhe/ ! |
| H A D | tlb.c | 74 if (cpus_have_final_cap(ARM64_WORKAROUND_SPECULATIVE_AT)) { in enter_vmid_context() 138 if (cpus_have_final_cap(ARM64_WORKAROUND_SPECULATIVE_AT)) { in exit_vmid_context()
|
| H A D | switch.c | 62 if (cpus_have_final_cap(ARM64_WORKAROUND_SPECULATIVE_AT)) { in __activate_traps() 83 if (cpus_have_final_cap(ARM64_WORKAROUND_SPECULATIVE_AT)) { in __deactivate_traps()
|
| H A D | pkvm.c | 38 if (cpus_have_final_cap(ARM64_HAS_RAS_EXTN)) { in pkvm_vcpu_reset_hcr() 45 if (cpus_have_final_cap(ARM64_HAS_STAGE2_FWB)) in pkvm_vcpu_reset_hcr() 48 if (cpus_have_final_cap(ARM64_HAS_EVT) && in pkvm_vcpu_reset_hcr() 49 !cpus_have_final_cap(ARM64_MISMATCHED_CACHE_TYPE) && in pkvm_vcpu_reset_hcr()
|
| H A D | sys_regs.c | 64 return (cpus_have_final_cap(ARM64_HAS_ADDRESS_AUTH) || in vm_has_ptrauth() 65 cpus_have_final_cap(ARM64_HAS_GENERIC_AUTH)) && in vm_has_ptrauth()
|
| /linux/arch/arm64/kvm/hyp/vhe/ ! |
| H A D | tlb.c | 33 if (cpus_have_final_cap(ARM64_WORKAROUND_SPECULATIVE_AT)) { in enter_vmid_context() 83 if (cpus_have_final_cap(ARM64_WORKAROUND_SPECULATIVE_AT)) { in exit_vmid_context()
|
| /linux/arch/arm64/kvm/vgic/ ! |
| H A D | vgic-v5.c | 18 if (!cpus_have_final_cap(ARM64_HAS_GICV5_LEGACY)) in vgic_v5_probe()
|
| /linux/arch/arm64/kvm/hyp/ ! |
| H A D | pgtable.c | 345 if (!cpus_have_final_cap(ARM64_KVM_HVHE)) in hyp_set_prot_attr() 359 if (cpus_have_final_cap(ARM64_KVM_HVHE)) { in hyp_set_prot_attr() 387 if (cpus_have_final_cap(ARM64_KVM_HVHE)) { in kvm_pgtable_hyp_pte_prot() 637 if (!cpus_have_final_cap(ARM64_WORKAROUND_AMPERE_AC03_CPU_38)) in kvm_get_vtcr() 690 if (!cpus_have_final_cap(ARM64_HAS_XNX) && px != ux) in stage2_set_xn_attr() 889 return system_supports_tlb_range() && cpus_have_final_cap(ARM64_HAS_STAGE2_FWB); in stage2_unmap_defer_tlb_flush() 1169 need_flush = !cpus_have_final_cap(ARM64_HAS_STAGE2_FWB); in stage2_unmap_walker() 1400 if (cpus_have_final_cap(ARM64_HAS_STAGE2_FWB)) in kvm_pgtable_stage2_flush()
|
| /linux/arch/arm64/kernel/ ! |
| H A D | sys_compat.c | 34 if (cpus_have_final_cap(ARM64_WORKAROUND_1542419)) { in __do_compat_cache_op()
|
| H A D | module-plts.c | 203 if (!cpus_have_final_cap(ARM64_WORKAROUND_843419)) in count_plts() 238 if (cpus_have_final_cap(ARM64_WORKAROUND_843419)) { in count_plts()
|
| /linux/include/kvm/ ! |
| H A D | arm_arch_timer.h | 161 return (has_vhe() && cpus_have_final_cap(ARM64_HAS_ECV_CNTPOFF)); in has_cntpoff()
|