| /linux/arch/x86/include/asm/ |
| H A D | smap.h | 19 ALTERNATIVE "", "clac", X86_FEATURE_SMAP 22 ALTERNATIVE "", "stac", X86_FEATURE_SMAP 80 ALTERNATIVE(ANNOTATE_IGNORE_ALTERNATIVE "\n\t" in smap_save() 91 ALTERNATIVE(ANNOTATE_IGNORE_ALTERNATIVE "\n\t" in smap_restore() 99 ALTERNATIVE("", "clac", X86_FEATURE_SMAP) 101 ALTERNATIVE("", "stac", X86_FEATURE_SMAP) 104 ALTERNATIVE("", ANNOTATE_IGNORE_ALTERNATIVE "\n\t" "clac", X86_FEATURE_SMAP) 106 ALTERNATIVE("", ANNOTATE_IGNORE_ALTERNATIVE "\n\t" "stac", X86_FEATURE_SMAP)
|
| H A D | alternative.h | 217 #define ALTERNATIVE(oldinstr, newinstr, ft_flags) \ macro 223 ALTERNATIVE(ALTERNATIVE(oldinstr, newinstr1, ft_flags1), newinstr2, ft_flags2) 231 ALTERNATIVE(ALTERNATIVE_2(oldinstr, newinstr1, ft_flags1, newinstr2, ft_flags2), \ 247 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) : : : "memory") 261 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 266 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 280 asm_inline volatile(ALTERNATIVE("call %c[old]", "call %c[new]", ft_flags) \ 375 .macro ALTERNATIVE oldinstr, newinstr, ft_flags
|
| H A D | nospec-branch.h | 299 ALTERNATIVE "", \ 330 ALTERNATIVE "", __CLEAR_CPU_BUFFERS, X86_FEATURE_CLEAR_CPU_BUF 334 ALTERNATIVE "", "call clear_bhb_loop", X86_FEATURE_CLEAR_BHB_LOOP 338 ALTERNATIVE "", "call clear_bhb_loop", X86_FEATURE_CLEAR_BHB_VMEXIT 403 ALTERNATIVE("", \ 528 asm volatile(ALTERNATIVE("", "wrmsr", %c[feature]) in alternative_msr_write() 540 asm_inline volatile(ALTERNATIVE("", "call write_ibpb", X86_FEATURE_IBPB) in indirect_branch_prediction_barrier()
|
| H A D | barrier.h | 15 #define mb() asm volatile(ALTERNATIVE("lock addl $0,-4(%%esp)", "mfence", \ 17 #define rmb() asm volatile(ALTERNATIVE("lock addl $0,-4(%%esp)", "lfence", \ 19 #define wmb() asm volatile(ALTERNATIVE("lock addl $0,-4(%%esp)", "sfence", \
|
| H A D | uaccess_64.h | 29 asm_inline (ALTERNATIVE("", "and " __percpu_arg([mask]) ", %[addr]", in __untagged_addr() 128 ALTERNATIVE("rep movsb", in copy_user_generic() 190 ALTERNATIVE("rep stosb", in __clear_user()
|
| H A D | arch_hweight.h | 19 asm_inline (ALTERNATIVE("call __sw_hweight32", in __arch_hweight32() 48 asm_inline (ALTERNATIVE("call __sw_hweight64", in __arch_hweight64()
|
| /linux/arch/riscv/include/asm/ |
| H A D | errata_list.h | 18 ALTERNATIVE(__stringify(RISCV_PTR do_trap_insn_fault), \ 24 ALTERNATIVE(__stringify(RISCV_PTR do_page_fault), \ 31 asm(ALTERNATIVE("sfence.vma x0, %0", "sfence.vma", SIFIVE_VENDOR_ID, \ 36 asm(ALTERNATIVE("sfence.vma %0", "sfence.vma", SIFIVE_VENDOR_ID, \ 41 asm(ALTERNATIVE("sfence.vma %0, %1", "sfence.vma", SIFIVE_VENDOR_ID, \ 46 asm(ALTERNATIVE( \ 81 asm volatile(ALTERNATIVE( \ 102 asm volatile(ALTERNATIVE( \
|
| H A D | cpufeature-macros.h | 21 asm goto(ALTERNATIVE("j %l[l_no]", "nop", %[vendor], %[ext], 1) in __riscv_has_extension_likely() 35 asm goto(ALTERNATIVE("nop", "j %l[l_yes]", %[vendor], %[ext], 1) in __riscv_has_extension_unlikely()
|
| /linux/arch/parisc/include/asm/ |
| H A D | cache.h | 53 ALTERNATIVE(ALT_COND_NO_SMP, INSN_PxTLB) \ 56 ALTERNATIVE(ALT_COND_NO_SMP, INSN_PxTLB) \ 57 ALTERNATIVE(ALT_COND_NO_SPLIT_TLB, INSN_NOP) \ 61 ALTERNATIVE(ALT_COND_NO_DCACHE, INSN_NOP) \ 62 ALTERNATIVE(ALT_COND_NO_IOC_FDC, INSN_NOP) \ 65 ALTERNATIVE(ALT_COND_NO_DCACHE, INSN_NOP) \ 66 ALTERNATIVE(ALT_COND_NO_IOC_FDC, INSN_NOP) :::"memory")
|
| H A D | alternative.h | 36 #define ALTERNATIVE(cond, replacement) "!0:" \ macro 47 #define ALTERNATIVE(from, to, cond, replacement)\ macro
|
| /linux/arch/x86/entry/ |
| H A D | calling.h | 168 ALTERNATIVE "", "SET_NOFLUSH_BIT \reg", X86_FEATURE_PCID 174 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 187 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID 217 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 223 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 231 ALTERNATIVE "jmp .Ldone_\@", "", X86_FEATURE_PTI 250 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 260 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID 307 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_KERNEL_IBRS 336 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_KERNEL_IBRS [all …]
|
| H A D | entry_64.S | 130 ALTERNATIVE "testb %al, %al; jz swapgs_restore_regs_and_return_to_usermode", \ 255 ALTERNATIVE "jmp swapgs_restore_regs_and_return_to_usermode", \ 299 ALTERNATIVE "call error_entry; movq %rax, %rsp", \ 563 ALTERNATIVE "", "jmp xenpv_restore_regs_and_return_to_usermode", X86_FEATURE_XENPV 566 ALTERNATIVE "", "jmp .Lpti_restore_regs_and_return_to_usermode", X86_FEATURE_PTI 752 2: ALTERNATIVE "", "mfence", X86_BUG_SWAPGS_FENCE 765 ALTERNATIVE "", "ZAP_GS", X86_BUG_NULL_SEG 899 ALTERNATIVE "jmp .Lparanoid_entry_checkgs", "", X86_FEATURE_FSGSBASE 985 ALTERNATIVE "jmp .Lparanoid_exit_checkgs", "", X86_FEATURE_FSGSBASE 1421 ALTERNATIVE "jmp nmi_no_fsgsbase", "", X86_FEATURE_FSGSBASE
|
| /linux/arch/s390/include/asm/ |
| H A D | alternative.h | 131 #define ALTERNATIVE(oldinstr, altinstr, feature) \ macro 164 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) : : : "memory") 172 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, feature) \ 177 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) \ 209 .macro ALTERNATIVE oldinstr, newinstr, feature
|
| H A D | preempt.h | 35 ALTERNATIVE("llgt %[count],%[offzero](%%r0)\n", in preempt_count() 91 ALTERNATIVE("asi %[offzero](%%r0),%[val]\n", in __preempt_count_add() 122 ALTERNATIVE("alsi %[offzero](%%r0),%[val]\n", in __preempt_count_dec_and_test()
|
| H A D | lowcore.h | 226 ALTERNATIVE(" lghi %[lc],0", in get_lowcore() 244 ALTERNATIVE "lghi \reg,0", \ 250 ALTERNATIVE "stmg \start, \end, \savearea", \
|
| H A D | spinlock.h | 27 ALTERNATIVE(" ly %[lockval],%[offzero](%%r0)\n", in spinlock_lockval() 100 ALTERNATIVE("nop", ".insn rre,0xb2fa0000,7,0", ALT_FACILITY(49)) /* NIAI 7 */ in arch_spin_unlock()
|
| /linux/arch/arm64/include/asm/ |
| H A D | arch_timer.h | 69 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0", in arch_timer_read_cntpct_el0() 81 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0", in arch_timer_read_cntvct_el0() 183 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0", in __arch_counter_get_cntpct() 204 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0", in __arch_counter_get_cntvct()
|
| H A D | alternative-macros.h | 219 #define ALTERNATIVE(oldinstr, newinstr, ...) \ macro 234 ALTERNATIVE("b %l[l_no]", "nop", %[cpucap]) 255 ALTERNATIVE("nop", "b %l[l_yes]", %[cpucap]) in alternative_has_cap_unlikely()
|
| H A D | io.h | 57 asm volatile(ALTERNATIVE("ldrb %w0, [%1]", in __raw_readb() 69 asm volatile(ALTERNATIVE("ldrh %w0, [%1]", in __raw_readw() 80 asm volatile(ALTERNATIVE("ldr %w0, [%1]", in __raw_readl() 91 asm volatile(ALTERNATIVE("ldr %0, [%1]", in __raw_readq()
|
| /linux/arch/s390/kernel/ |
| H A D | entry.S | 38 ALTERNATIVE "nop", ".insn s,0xb2010000,\address", ALT_FACILITY(193) 42 ALTERNATIVE "nop", ".insn s,0xb2000000,\address", ALT_FACILITY(193) 53 ALTERNATIVE "brcl 0,0", __stringify(mvc __PT_LAST_BREAK(8,\reg),__LC_LAST_BREAK(\lowcore)),\ 98 ALTERNATIVE "nop", ".insn rrf,0xb2e80000,0,0,12,0", ALT_SPEC(82) 102 ALTERNATIVE "nop", ".insn rrf,0xb2e80000,0,0,13,0", ALT_SPEC(82) 106 ALTERNATIVE "TSTMSK \tif_ptr,\tif_mask; jz .+8; .insn rrf,0xb2e80000,0,0,13,0", \ 112 ALTERNATIVE "jz .+8; .insn rrf,0xb2e80000,0,0,12,0", \ 168 ALTERNATIVE "nop", "lpp _LPP_OFFSET(%r13)", ALT_FACILITY(40)
|
| /linux/arch/parisc/kernel/ |
| H A D | pacache.S | 106 ALTERNATIVE(88b, fitdone, ALT_COND_NO_SPLIT_TLB, INSN_NOP) 244 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_ICACHE, INSN_NOP) 305 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) 548 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 549 ALTERNATIVE(1b, 1b+4, ALT_COND_NO_SMP, INSN_PxTLB) 677 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 743 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 775 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) 792 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 824 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) [all …]
|
| /linux/tools/include/asm/ |
| H A D | alternative.h | 7 .macro ALTERNATIVE oldinstr, newinstr, feature 15 #define ALTERNATIVE # macro
|
| /linux/arch/x86/entry/vdso/vdso32/ |
| H A D | system_call.S | 27 ALTERNATIVE "int $0x80; ret", "", X86_FEATURE_SYSFAST32 67 ALTERNATIVE SYSENTER_SEQUENCE, SYSCALL_SEQUENCE, X86_FEATURE_SYSCALL32
|
| /linux/arch/s390/kernel/vdso/ |
| H A D | vgetrandom-chacha.S | 50 ALTERNATIVE __stringify(VL BEPERM,16,,%r1), "brcl 0,0", ALT_FACILITY(148) 133 ALTERNATIVE \ 172 ALTERNATIVE "nopr", "br %r14", ALT_FACILITY(148)
|
| /linux/arch/loongarch/include/asm/ |
| H A D | alternative.h | 71 #define ALTERNATIVE(oldinstr, newinstr, feature) \ macro 104 (asm volatile (ALTERNATIVE(oldinstr, newinstr, feature) : : : "memory"))
|