| /linux/arch/x86/include/asm/ ! |
| H A D | traps.h | 52 if (regs->flags & X86_EFLAGS_IF) in cond_local_irq_enable() 58 if (regs->flags & X86_EFLAGS_IF) in cond_local_irq_disable()
|
| H A D | irqflags.h | 59 return !(flags & X86_EFLAGS_IF); in native_irqs_disabled_flags() 146 return !(flags & X86_EFLAGS_IF); in arch_irqs_disabled_flags()
|
| H A D | ptrace.h | 314 return !(regs->flags & X86_EFLAGS_IF); in regs_irqs_disabled()
|
| H A D | efi.h | 32 #define ARCH_EFI_IRQ_FLAGS_MASK X86_EFLAGS_IF
|
| /linux/tools/testing/selftests/x86/ ! |
| H A D | entry_from_vm86.c | 281 v86.regs.eflags &= ~X86_EFLAGS_IF; in main() 291 v86.regs.eax = X86_EFLAGS_IF; in main() 296 v86.regs.eax = X86_EFLAGS_IF; in main()
|
| H A D | syscall_nt.c | 32 set_eflags(X86_EFLAGS_IF | X86_EFLAGS_FIXED); in do_it()
|
| /linux/arch/x86/kernel/ ! |
| H A D | vm86_32.c | 381 if (flags & X86_EFLAGS_IF) in set_vflags_long() 391 if (flags & X86_EFLAGS_IF) in set_vflags_short() 402 flags |= X86_EFLAGS_IF; in get_vflags() 694 if (vmpi->force_return_for_pic && (VEFLAGS & (X86_EFLAGS_IF | X86_EFLAGS_VIF))) { in handle_vm86_fault()
|
| H A D | process_32.c | 122 regs->flags = X86_EFLAGS_IF; in start_thread()
|
| H A D | traps.c | 425 if (regs->flags & X86_EFLAGS_IF) in handle_bug() 473 if (regs->flags & X86_EFLAGS_IF) in handle_bug()
|
| H A D | process_64.c | 578 regs->flags = X86_EFLAGS_IF | X86_EFLAGS_FIXED; in start_thread_common()
|
| H A D | kvm.c | 289 if (unlikely(!(regs->flags & X86_EFLAGS_IF))) in __kvm_handle_async_pf()
|
| /linux/arch/x86/kernel/kprobes/ ! |
| H A D | core.c | 489 regs->flags &= ~(X86_EFLAGS_IF); in kprobe_emulate_ifmodifiers() 492 regs->flags |= X86_EFLAGS_IF; in kprobe_emulate_ifmodifiers() 840 = (regs->flags & X86_EFLAGS_IF); in set_current_kprobe() 897 regs->flags &= ~X86_EFLAGS_IF; in setup_singlestep()
|
| /linux/arch/x86/mm/ ! |
| H A D | kmmio.c | 299 ctx->saved_flags = (regs->flags & (X86_EFLAGS_TF | X86_EFLAGS_IF)); in kmmio_handler() 310 regs->flags &= ~X86_EFLAGS_IF; in kmmio_handler()
|
| H A D | fault.c | 1274 if (WARN_ON_ONCE(!(regs->flags & X86_EFLAGS_IF))) { in do_user_addr_fault()
|
| /linux/arch/x86/include/uapi/asm/ ! |
| H A D | processor-flags.h | 26 #define X86_EFLAGS_IF _BITUL(X86_EFLAGS_IF_BIT) macro
|
| /linux/arch/x86/entry/ ! |
| H A D | syscall_32.c | 368 regs->flags |= X86_EFLAGS_IF; in do_SYSENTER_32()
|
| H A D | entry_64.S | 267 testl $X86_EFLAGS_IF, %eax
|
| /linux/arch/x86/kvm/ ! |
| H A D | emulate.c | 280 unsigned long flags = (ctxt->eflags & EFLAGS_MASK) | X86_EFLAGS_IF; \ 1869 change_mask |= X86_EFLAGS_IF; in emulate_popf() 1874 change_mask |= X86_EFLAGS_IF; in emulate_popf() 1877 change_mask |= (X86_EFLAGS_IOPL | X86_EFLAGS_IF); in emulate_popf() 2018 ctxt->eflags &= ~(X86_EFLAGS_IF | X86_EFLAGS_TF | X86_EFLAGS_AC); in __emulate_int_real() 2086 X86_EFLAGS_IF | X86_EFLAGS_DF | X86_EFLAGS_OF | in emulate_iret_real() 2399 ctxt->eflags &= ~(X86_EFLAGS_VM | X86_EFLAGS_IF); in em_syscall() 2436 ctxt->eflags &= ~(X86_EFLAGS_VM | X86_EFLAGS_IF); in em_sysenter() 3566 ctxt->eflags &= ~X86_EFLAGS_IF; in em_cli() 3576 ctxt->eflags |= X86_EFLAGS_IF; in em_sti()
|
| H A D | x86.h | 646 .flags = X86_EFLAGS_IF, in kvm_machine_check()
|
| H A D | x86.c | 9630 if (unlikely((ctxt->eflags & ~rflags) & X86_EFLAGS_IF)) in x86_emulate_instruction()
|
| /linux/arch/x86/kvm/svm/ ! |
| H A D | svm.c | 1476 : kvm_get_rflags(vcpu) & X86_EFLAGS_IF; in svm_get_if_flag() 3887 ? !(svm->vmcb01.ptr->save.rflags & X86_EFLAGS_IF) in svm_interrupt_blocked() 3888 : !(kvm_get_rflags(vcpu) & X86_EFLAGS_IF)) in svm_interrupt_blocked()
|
| H A D | nested.c | 155 if (!(svm->vmcb01.ptr->save.rflags & X86_EFLAGS_IF)) in recalc_intercepts()
|
| /linux/arch/x86/kernel/cpu/ ! |
| H A D | common.c | 2295 X86_EFLAGS_IF|X86_EFLAGS_DF|X86_EFLAGS_OF| in idt_syscall_init()
|
| /linux/arch/x86/kvm/vmx/ ! |
| H A D | vmx.c | 1692 return vmx_get_rflags(vcpu) & X86_EFLAGS_IF; in vmx_get_if_flag() 5294 return !(vmx_get_rflags(vcpu) & X86_EFLAGS_IF) || in __vmx_interrupt_blocked()
|
| H A D | nested.c | 3878 (vmcs12->guest_rflags & X86_EFLAGS_IF))) { in nested_vmx_run()
|