| /linux/arch/x86/entry/vdso/vdso64/ |
| H A D | vsgx.S | 34 push %rbx 35 .cfi_rel_offset %rbx, -8 48 mov $SGX_ENCLAVE_RUN_RESERVED_START, %rbx 50 cmpq $0, (%rcx, %rbx) 52 add $8, %rbx 53 cmpq $SGX_ENCLAVE_RUN_RESERVED_END, %rbx 57 mov SGX_ENCLAVE_RUN_TCS(%rcx), %rbx 66 mov SGX_ENCLAVE_OFFSET_OF_RUN(%rbp), %rbx 69 movl $EEXIT, SGX_ENCLAVE_RUN_LEAF(%rbx) 73 cmpq $0, SGX_ENCLAVE_RUN_USER_HANDLER(%rbx) [all …]
|
| /linux/lib/crypto/x86/ |
| H A D | sha256-ni-asm.S | 280 push %rbx 308 mov OFFSETOF_BYTECOUNT(CTX), %rbx 309 lea (%rbx, LEN64, 1), COUNT 331 movdqu MSG0_A, 0*16(%rsp,%rbx) 332 movdqu MSG1_A, 1*16(%rsp,%rbx) 333 movdqu MSG2_A, 2*16(%rsp,%rbx) 334 movdqu MSG3_A, 3*16(%rsp,%rbx) 344 movdqu MSG0_B, 0*16(%rsp,%rbx) 345 movdqu MSG1_B, 1*16(%rsp,%rbx) 346 movdqu MSG2_B, 2*16(%rsp,%rbx) [all …]
|
| /linux/arch/x86/kernel/acpi/ |
| H A D | wakeup_64.S | 39 movq saved_rbx(%rip), %rbx 60 movq %rbx, pt_regs_bx(%rax) 78 movq %rbx, saved_rbx(%rip) 94 movq saved_context_cr4(%rax), %rbx 95 movq %rbx, %cr4 96 movq saved_context_cr3(%rax), %rbx 97 movq %rbx, %cr3 98 movq saved_context_cr2(%rax), %rbx 99 movq %rbx, %cr2 100 movq saved_context_cr0(%rax), %rbx [all …]
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | sync_regs_test.c | 53 REG_COMPARE(rbx); in compare_regs() 286 run->s.regs.regs.rbx = 0xBAD1DEA; in KVM_ONE_VCPU_TEST() 294 TEST_ASSERT(run->s.regs.regs.rbx == 0xBAD1DEA + 1, in KVM_ONE_VCPU_TEST() 296 run->s.regs.regs.rbx); in KVM_ONE_VCPU_TEST() 320 run->s.regs.regs.rbx = 0xDEADBEEF; in KVM_ONE_VCPU_TEST() 323 TEST_ASSERT(run->s.regs.regs.rbx != 0xDEADBEEF, in KVM_ONE_VCPU_TEST() 325 run->s.regs.regs.rbx); in KVM_ONE_VCPU_TEST() 344 run->s.regs.regs.rbx = 0xAAAA; in KVM_ONE_VCPU_TEST() 346 regs.rbx = 0xBAC0; in KVM_ONE_VCPU_TEST() 350 TEST_ASSERT(run->s.regs.regs.rbx == 0xAAAA, in KVM_ONE_VCPU_TEST() [all …]
|
| /linux/arch/x86/lib/ |
| H A D | csum-copy_64.S | 41 movq %rbx, 0*8(%rsp) 70 movq (%rdi), %rbx 95 adcq %rbx, %rax 107 movq %rbx, (%rsi) 142 movq (%rdi), %rbx 143 adcq %rbx, %rax 146 movq %rbx, (%rsi) 198 movq 0*8(%rsp), %rbx 221 addq %rbx, %rax 231 addq %rbx, %rax [all …]
|
| H A D | copy_page_64.S | 27 movq %rbx, (%rsp) 35 movq 0x8*1(%rsi), %rbx 46 movq %rbx, 0x8*1(%rdi) 65 movq 0x8*1(%rsi), %rbx 74 movq %rbx, 0x8*1(%rdi) 86 movq (%rsp), %rbx
|
| H A D | putuser.S | 38 mov %rcx, %rbx 39 sar $63, %rbx 40 or %rbx, %rcx
|
| /linux/arch/x86/kernel/cpu/sgx/ |
| H A D | encls.h | 85 #define __encls_ret_2(rax, rbx, rcx) \ argument 87 __encls_ret_N(rax, "b"(rbx), "c"(rcx)); \ 90 #define __encls_ret_3(rax, rbx, rcx, rdx) \ argument 92 __encls_ret_N(rax, "b"(rbx), "c"(rcx), "d"(rdx)); \ 124 #define __encls_2(rax, rbx, rcx) \ argument 127 __encls_N(rax, ign_rbx_out, "b"(rbx), "c"(rcx)); \ 206 unsigned long rbx = SGX_PAGE_TYPE_VA; in __epa() local 208 return __encls_2(EPA, rbx, addr); in __epa()
|
| /linux/arch/x86/boot/startup/ |
| H A D | la57toggle.S | 32 pushq %rbx 35 movq %rsp, %rbx 36 shrq $32, %rbx 51 shlq $32, %rbx 52 orq %rbx, %rsp 55 popq %rbx
|
| /linux/arch/x86/platform/efi/ |
| H A D | efi_thunk_64.S | 32 push %rbx 46 movq 0x20(%rax), %rbx 59 leaq 2f(%rip), %rbx 61 subq %rax, %rbx 85 pop %rbx
|
| /linux/arch/x86/hyperv/ |
| H A D | hv_trampoline.S | 89 movq %cs:(%eax), %rbx 90 movq %rbx, %cr3 94 movq %cs:(%eax), %rbx 96 jmp *%rbx
|
| H A D | mshv_vtl_asm.S | 36 pushq %rbx 42 mov MSHV_VTL_CPU_CONTEXT_rbx(%rax), %rbx 82 mov %rbx, MSHV_VTL_CPU_CONTEXT_rbx(%rax) 96 pop %rbx
|
| /linux/arch/x86/power/ |
| H A D | hibernate_asm_64.S | 48 movq pt_regs_bx(%rax), %rbx 79 movq %rbx, pt_regs_bx(%rax) 110 movq mmu_cr4_features(%rip), %rbx 127 movq %rbx, %rcx 132 movq %rbx, %cr4; # turn PGE back on
|
| /linux/arch/x86/kernel/ |
| H A D | relocate_kernel_64.S | 73 pushq %rbx 327 popq %rbx 347 movq (%rbx), %rcx 348 addq $8, %rbx 358 movq %rcx, %rbx 359 andq $0xfffffffffffff000, %rbx 489 1: rolq $4, %rbx 510 movq \r, %rbx 557 pushq %rbx 611 popq %rbx
|
| /linux/tools/testing/selftests/sgx/ |
| H A D | call.S | 23 push %rbx 25 .cfi_rel_offset %rbx, 0 33 pop %rbx
|
| H A D | test_encl_bootstrap.S | 48 sub %rax, %rbx 50 add %rbx, %rax 55 lea -1(%rbx), %rax 69 pop %rbx
|
| /linux/arch/x86/boot/compressed/ |
| H A D | head_64.S | 331 addq %rbp, %rbx 334 leaq rva(boot_stack_end)(%rbx), %rsp 408 leaq rva(top_pgtable)(%rbx), %rsi 420 leaq rva(_bss-8)(%rbx), %rdi 432 leaq rva(gdt64)(%rbx), %rax 433 leaq rva(gdt)(%rbx), %rdx 440 leaq rva(.Lrelocated)(%rbx), %rax
|
| /linux/arch/x86/virt/vmx/tdx/ |
| H A D | tdxcall.S | 66 pushq %rbx 76 movq TDX_MODULE_rbx(%rsi), %rbx 134 movq %rbx, TDX_MODULE_rbx(%rsi) 187 popq %rbx
|
| /linux/arch/x86/kernel/cpu/ |
| H A D | vmware.c | 75 unsigned long out0, rbx, rcx, rdx, rsi, rdi; in vmware_hypercall_slow() local 80 : "=a" (out0), "=b" (rbx), "=c" (rcx), in vmware_hypercall_slow() 92 : "=a" (out0), "=b" (rbx), "=c" (rcx), in vmware_hypercall_slow() 104 : "=a" (out0), "=b" (rbx), "=c" (rcx), in vmware_hypercall_slow() 118 *out1 = rbx; in vmware_hypercall_slow() 520 args.rbx = in1; in vmware_tdx_hypercall() 534 *out1 = args.rbx; in vmware_tdx_hypercall()
|
| /linux/arch/x86/um/ |
| H A D | setjmp_64.S | 10 # %rbx 27 movq %rbx,(%rdi) 46 movq (%rdi),%rbx
|
| /linux/arch/x86/purgatory/ |
| H A D | entry64.S | 39 movq rbx(%rip), %rbx 65 rbx: .quad 0x0 label
|
| /linux/arch/x86/crypto/ |
| H A D | blowfish-x86_64-asm_64.S | 25 #define RX1 %rbx 282 pushq %rbx; 305 popq %rbx; 318 pushq %rbx; 350 popq %rbx;
|
| H A D | twofish-x86_64-asm_64-3way.S | 29 #define RAB1 %rbx 233 pushq %rbx; 259 popq %rbx; 267 popq %rbx; 281 pushq %rbx; 302 popq %rbx;
|
| /linux/tools/testing/selftests/x86/ |
| H A D | thunks.S | 19 pushq %rbx 55 popq %rbx
|
| /linux/arch/x86/entry/ |
| H A D | entry_64.S | 184 pushq %rbx 195 movq TASK_stack_canary(%rsi), %rbx 196 movq %rbx, PER_CPU_VAR(__stack_chk_guard) 213 popq %rbx 243 movq %rbx, %rdx /* fn */ 910 SAVE_AND_SET_GSBASE scratch_reg=%rax save_reg=%rbx 988 wrgsbase %rbx 1423 wrgsbase %rbx
|