/arch/x86/um/ |
D | stub_64.S | 10 movq $(STUB_DATA >> 32), %rbx 11 salq $32, %rbx 13 or %rcx, %rbx 14 movq %rax, (%rbx) 19 mov $(STUB_DATA >> 32), %rbx 20 sal $32, %rbx 22 or %rax, %rbx 24 mov %rbx, %rsp 32 mov %rax, 8(%rbx) 37 mov %rsp, 8(%rbx) [all …]
|
D | setjmp_64.S | 9 # %rbx 26 movq %rbx,(%rdi) 45 movq (%rdi),%rbx
|
/arch/x86/platform/efi/ |
D | efi_thunk_64.S | 29 push %rbx 50 leaq efi_exit32(%rip), %rbx 51 subq %rax, %rbx 54 leaq __efi64_thunk(%rip), %rbx 55 subq %rax, %rbx 56 call *%rbx 59 pop %rbx 86 leaq 1f(%rip), %rbx 87 movq %rbx, func_rt_ptr(%rip) 97 pop %rbx [all …]
|
/arch/x86/kernel/acpi/ |
D | wakeup_64.S | 29 movq saved_rbx, %rbx 51 movq %rbx, pt_regs_bx(%rax) 69 movq %rbx, saved_rbx 84 movq saved_context_cr4(%rax), %rbx 85 movq %rbx, %cr4 86 movq saved_context_cr3(%rax), %rbx 87 movq %rbx, %cr3 88 movq saved_context_cr2(%rax), %rbx 89 movq %rbx, %cr2 90 movq saved_context_cr0(%rax), %rbx [all …]
|
/arch/x86/lib/ |
D | copy_page_64.S | 25 movq %rbx, (%rsp) 26 CFI_REL_OFFSET rbx, 0 35 movq 0x8*1(%rsi), %rbx 46 movq %rbx, 0x8*1(%rdi) 65 movq 0x8*1(%rsi), %rbx 74 movq %rbx, 0x8*1(%rdi) 86 movq (%rsp), %rbx 87 CFI_RESTORE rbx
|
D | csum-copy_64.S | 57 movq %rbx, 2*8(%rsp) 58 CFI_REL_OFFSET rbx, 2*8 89 movq (%rdi), %rbx 109 adcq %rbx, %rax 121 movq %rbx, (%rsi) 158 movq (%rdi), %rbx 159 adcq %rbx, %rax 162 movq %rbx, (%rsi) 211 movq 2*8(%rsp), %rbx 212 CFI_RESTORE rbx
|
D | msr-reg.S | 17 pushq_cfi %rbx 39 popq_cfi %rbx
|
D | cmpxchg16b_emu.S | 43 movq %rbx, PER_CPU_VAR((%rsi))
|
/arch/x86/boot/compressed/ |
D | efi_thunk_64.S | 27 push %rbx 57 leaq 1f(%rip), %rbx 58 movq %rbx, func_rt_ptr(%rip) 77 pop %rbx 79 pop %rbx 81 pop %rbx 96 pop %rbx
|
D | head_64.S | 340 leaq z_extract_offset(%rbp), %rbx 343 leaq boot_stack_end(%rbx), %rsp 355 leaq (_bss-8)(%rbx), %rdi 366 leaq relocated(%rbx), %rax 404 addq %rbx, (%rdx)
|
/arch/x86/kernel/ |
D | relocate_kernel_64.S | 53 pushq %rbx 209 popq %rbx 221 movq (%rbx), %rcx 222 addq $8, %rbx 232 movq %rcx, %rbx 233 andq $0xfffffffffffff000, %rbx
|
D | head_64.S | 105 leaq early_level4_pgt(%rip), %rbx 110 leaq (4096 + _KERNPG_TABLE)(%rbx), %rdx 111 movq %rdx, 0(%rbx,%rax,8) 112 movq %rdx, 8(%rbx,%rax,8) 118 movq %rdx, 4096(%rbx,%rax,8) 121 movq %rdx, 4096(%rbx,%rax,8) 123 addq $8192, %rbx 134 movq %rax, (%rbx,%rdi,8)
|
D | mcount_64.S | 83 movq %rbx, RBX(%rsp) 117 movq RBX(%rsp), %rbx
|
/arch/x86/purgatory/ |
D | entry64.S | 41 movq rbx(%rip), %rbx 66 rbx: .quad 0x0 label
|
/arch/x86/crypto/sha-mb/ |
D | sha1_mb_mgr_flush_avx2.S | 73 #define unused_lanes %rbx 74 #define lane_data %rbx 75 #define tmp2 %rbx 119 mov %rbx, _GPR_SAVE(%rsp) 234 mov _GPR_SAVE(%rsp), %rbx 255 push %rbx 299 pop %rbx 305 pop %rbx
|
D | sha1_mb_mgr_submit_avx2.S | 82 unused_lanes = %rbx 109 mov %rbx, (%rsp) 207 mov (%rsp), %rbx
|
/arch/x86/crypto/ |
D | blowfish-x86_64-asm_64.S | 40 #define RX1 %rbx 307 pushq %rbx; 335 popq %rbx; 342 popq %rbx; 354 pushq %rbx; 375 popq %rbx;
|
D | salsa20-x86_64-asm_64.S | 35 # rbx_stack = rbx 36 movq %rbx,40(%rsp) 150 mov %r15,%rbx 220 xor %r15,%rbx 222 lea (%rbp,%rbx),%r15 228 lea (%rbx,%rcx),%r15 352 lea (%rbx,%r15),%rbp 368 xor %rbp,%rbx 370 lea (%r13,%rbx),%rbp 436 xor %r15,%rbx [all …]
|
D | twofish-x86_64-asm_64-3way.S | 43 #define RAB1 %rbx 231 pushq %rbx; 255 popq %rbx; 266 popq %rbx; 286 pushq %rbx; 305 popq %rbx;
|
D | camellia-x86_64-asm_64.S | 59 #define RAB1 %rbx 434 pushq %rbx; 465 popq %rbx; 472 popq %rbx; 487 movq %rbx, RXOR; 512 movq RXOR, %rbx;
|
D | crc32c-pcl-intel-asm_64.S | 90 #define tmp %rbx 96 pushq %rbx 311 popq %rbx
|
/arch/x86/power/ |
D | hibernate_asm_64.S | 31 movq %rbx, pt_regs_bx(%rax) 72 movq restore_cr3(%rip), %rbx 111 movq %rbx, %cr3 128 movq pt_regs_bx(%rax), %rbx
|
/arch/x86/net/ |
D | bpf_jit.S | 76 mov %rbx, %rdi; /* arg1 == skb */ \ 110 mov %rbx, %rdi; /* arg1 == skb */ \ 154 mov - MAX_BPF_STACK(%rbp),%rbx
|
/arch/x86/include/uapi/asm/ |
D | ptrace.h | 49 unsigned long rbx; member
|
/arch/x86/include/asm/ |
D | calling.h | 169 movq_cfi rbx, 5*8 183 movq_cfi_restore 5*8, rbx
|