Lines Matching full:intercept
51 vmcb->control.intercept &= ~(1ULL << INTERCEPT_VMRUN); in prepare_no_vmrun_int()
72 vmcb->control.intercept |= 1 << INTERCEPT_RSM; in prepare_rsm_intercept()
95 vmcb->control.intercept &= ~(1 << INTERCEPT_RSM); in finished_rsm_intercept()
211 report_fail("dr%u read intercept", i); in test_dr_intercept()
247 report_fail("dr%u write intercept", i); in test_dr_intercept()
265 * Per Appendix C "SVM Intercept Exit Codes" of AMD64 Architecture in dr_intercept_finished()
290 vmcb->control.intercept |= (1ULL << INTERCEPT_RDTSC); in prepare_next_rip()
314 vmcb->control.intercept |= (1ULL << INTERCEPT_MSR_PROT); in prepare_msr_intercept()
345 /* Check that a read intercept occurred for MSR at msr_index */ in test_msr_intercept()
347 report_fail("MSR 0x%lx read intercept", msr_index); in test_msr_intercept()
357 /* Check that a write intercept occurred for MSR with msr_value */ in test_msr_intercept()
359 report_fail("MSR 0x%lx write intercept", msr_index); in test_msr_intercept()
409 * Test whether the intercept was for RDMSR/WRMSR. in msr_intercept_finished()
543 vmcb->control.intercept |= (1ULL << INTERCEPT_IOIO_PROT); in prepare_ioio()
668 vmcb->control.intercept |= (1ULL << INTERCEPT_SELECTIVE_CR0); in sel_cr0_bug_prepare()
687 * are not in guest-mode anymore so we can't trigger an intercept. in sel_cr0_bug_test()
987 vmcb->control.intercept |= (1ULL << INTERCEPT_INTR); in pending_event_prepare()
1011 vmcb->control.intercept &= ~(1ULL << INTERCEPT_INTR); in pending_event_finished()
1216 vmcb->control.intercept |= (1ULL << INTERCEPT_INTR); in interrupt_finished()
1223 report_fail("VMEXIT not due to intr intercept. Exit reason 0x%x", in interrupt_finished()
1230 vmcb->control.intercept &= ~(1ULL << INTERCEPT_INTR); in interrupt_finished()
1292 vmcb->control.intercept |= (1ULL << INTERCEPT_NMI); in nmi_finished()
1297 report_fail("VMEXIT not due to NMI intercept. Exit reason 0x%x", in nmi_finished()
1302 report_pass("NMI intercept while running guest"); in nmi_finished()
1385 vmcb->control.intercept |= (1ULL << INTERCEPT_NMI); in nmi_hlt_finished()
1390 report_fail("VMEXIT not due to NMI intercept. Exit reason 0x%x", in nmi_hlt_finished()
1395 report_pass("NMI intercept while running guest"); in nmi_hlt_finished()
1421 vmcb->control.intercept &= ~(1ULL << INTERCEPT_NMI); in vnmi_prepare()
1445 vmcb->control.intercept |= (1ULL << INTERCEPT_NMI); in vnmi_finished()
1620 vmcb->control.intercept |= (1ULL << INTERCEPT_VINTR); in virq_inject_finished()
1635 vmcb->control.intercept &= ~(1ULL << INTERCEPT_VINTR); in virq_inject_finished()
1658 vmcb->control.intercept |= (1ULL << INTERCEPT_VINTR); in virq_inject_finished()
1731 * and we do intercept them in L1
1738 * In our intercept handler we detect the bug by checking that RIP is that of
1759 vmcb->control.intercept |= (1ULL << INTERCEPT_INTR); in reg_corruption_prepare()
1866 vmcb->control.intercept |= (1ULL << INTERCEPT_INIT); in init_intercept_prepare()
1879 report_fail("VMEXIT not due to init intercept. Exit reason 0x%x", in init_intercept_finished()
2363 vmcb->control.intercept = saved_intercept | 1ULL << type; \
2373 * If the MSR or IOIO intercept table extends to a physical address that
2390 u64 saved_intercept = vmcb->control.intercept; in test_msrpm_iopm_bitmap_addrs()
2431 vmcb->control.intercept = saved_intercept; in test_msrpm_iopm_bitmap_addrs()
2627 u32 intercept_saved = vmcb->control.intercept; in svm_vmload_vmsave()
2632 * Disabling intercept for VMLOAD and VMSAVE doesn't cause in svm_vmload_vmsave()
2635 vmcb->control.intercept &= ~(1ULL << INTERCEPT_VMLOAD); in svm_vmload_vmsave()
2636 vmcb->control.intercept &= ~(1ULL << INTERCEPT_VMSAVE); in svm_vmload_vmsave()
2639 "VMLOAD/VMSAVE intercept: Expected VMMCALL #VMEXIT"); in svm_vmload_vmsave()
2642 * Enabling intercept for VMLOAD and VMSAVE causes respective in svm_vmload_vmsave()
2645 vmcb->control.intercept |= (1ULL << INTERCEPT_VMLOAD); in svm_vmload_vmsave()
2648 "VMLOAD/VMSAVE intercept: Expected VMLOAD #VMEXIT"); in svm_vmload_vmsave()
2649 vmcb->control.intercept &= ~(1ULL << INTERCEPT_VMLOAD); in svm_vmload_vmsave()
2650 vmcb->control.intercept |= (1ULL << INTERCEPT_VMSAVE); in svm_vmload_vmsave()
2653 "VMLOAD/VMSAVE intercept: Expected VMSAVE #VMEXIT"); in svm_vmload_vmsave()
2654 vmcb->control.intercept &= ~(1ULL << INTERCEPT_VMSAVE); in svm_vmload_vmsave()
2657 "VMLOAD/VMSAVE intercept: Expected VMMCALL #VMEXIT"); in svm_vmload_vmsave()
2659 vmcb->control.intercept |= (1ULL << INTERCEPT_VMLOAD); in svm_vmload_vmsave()
2662 "VMLOAD/VMSAVE intercept: Expected VMLOAD #VMEXIT"); in svm_vmload_vmsave()
2663 vmcb->control.intercept &= ~(1ULL << INTERCEPT_VMLOAD); in svm_vmload_vmsave()
2666 "VMLOAD/VMSAVE intercept: Expected VMMCALL #VMEXIT"); in svm_vmload_vmsave()
2668 vmcb->control.intercept |= (1ULL << INTERCEPT_VMSAVE); in svm_vmload_vmsave()
2671 "VMLOAD/VMSAVE intercept: Expected VMSAVE #VMEXIT"); in svm_vmload_vmsave()
2672 vmcb->control.intercept &= ~(1ULL << INTERCEPT_VMSAVE); in svm_vmload_vmsave()
2675 "VMLOAD/VMSAVE intercept: Expected VMMCALL #VMEXIT"); in svm_vmload_vmsave()
2677 vmcb->control.intercept = intercept_saved; in svm_vmload_vmsave()
2790 vmcb->control.intercept |= (1 << INTERCEPT_PAUSE); in pause_filter_test()
3100 vmcb->control.intercept |= (1 << INTERCEPT_INTR); in svm_intr_intercept_mix_if()
3133 vmcb->control.intercept |= (1 << INTERCEPT_INTR); in svm_intr_intercept_mix_gif()
3163 vmcb->control.intercept |= (1 << INTERCEPT_INTR); in svm_intr_intercept_mix_gif2()
3192 vmcb->control.intercept |= (1 << INTERCEPT_NMI); in svm_intr_intercept_mix_nmi()
3216 vmcb->control.intercept |= (1 << INTERCEPT_SMI); in svm_intr_intercept_mix_smi()
3312 vmcb->control.intercept |= (1ULL << INTERCEPT_SHUTDOWN); in svm_shutdown_intercept_test()
3327 { "vmrun intercept check", default_supported, prepare_no_vmrun_int,
3333 { "cr3 read intercept", default_supported,
3339 { "cr3 read intercept emulate", smp_supported,
3342 { "dr intercept check", default_supported, prepare_dr_intercept,
3348 { "msr intercept check", default_supported, prepare_msr_intercept,