/arch/um/sys-x86_64/ |
D | stub.S | 10 movq $(STUB_DATA >> 32), %rbx 11 salq $32, %rbx 13 or %rcx, %rbx 14 movq %rax, (%rbx) 19 mov $(STUB_DATA >> 32), %rbx 20 sal $32, %rbx 22 or %rax, %rbx 24 mov %rbx, %rsp 32 mov %rax, 8(%rbx) 37 mov %rsp, 8(%rbx) [all …]
|
D | setjmp.S | 9 # %rbx 26 movq %rbx,(%rdi) 45 movq (%rdi),%rbx
|
/arch/x86/kernel/acpi/ |
D | wakeup_64.S | 29 movq saved_rbx, %rbx 51 movq %rbx, pt_regs_bx(%rax) 69 movq %rbx, saved_rbx 84 movq saved_context_cr4(%rax), %rbx 85 movq %rbx, %cr4 86 movq saved_context_cr3(%rax), %rbx 87 movq %rbx, %cr3 88 movq saved_context_cr2(%rax), %rbx 89 movq %rbx, %cr2 90 movq saved_context_cr0(%rax), %rbx [all …]
|
/arch/x86/lib/ |
D | copy_page_64.S | 24 movq %rbx,(%rsp) 25 CFI_REL_OFFSET rbx, 0 37 movq 8 (%rsi), %rbx 48 movq %rbx, 8 (%rdi) 67 movq 8 (%rsi), %rbx 76 movq %rbx, 8 (%rdi) 89 movq (%rsp),%rbx 90 CFI_RESTORE rbx
|
D | csum-copy_64.S | 65 movq %rbx,2*8(%rsp) 66 CFI_REL_OFFSET rbx, 2*8 97 movq (%rdi),%rbx 117 adcq %rbx,%rax 129 movq %rbx,(%rsi) 166 movq (%rdi),%rbx 167 adcq %rbx,%rax 170 movq %rbx,(%rsi) 219 movq 2*8(%rsp),%rbx 220 CFI_RESTORE rbx
|
D | memcpy_64.S | 36 pushq %rbx 38 CFI_REL_OFFSET rbx, 0 105 popq %rbx 107 CFI_RESTORE rbx
|
/arch/x86/boot/compressed/ |
D | head_64.S | 227 movq %rbp, %rbx 230 movq %rbp, %rbx 235 subq %rax, %rbx 237 addq %rax, %rbx 240 addq %rax, %rbx 242 addq $(32768 + 18 + 4095), %rbx 243 andq $~4095, %rbx 249 leaq _end_before_pgt(%rbx), %r9 261 leaq relocated(%rbx), %rax 271 leaq _edata(%rbx), %rdi [all …]
|
/arch/x86/kernel/ |
D | relocate_kernel_64.S | 210 movq (%rbx), %rcx 211 addq $8, %rbx 221 movq %rcx, %rbx 222 andq $0xfffffffffffff000, %rbx 251 xorq %rbx, %rbx
|
D | head_64.S | 107 leaq level3_ident_pgt(%rip), %rbx 108 movq %rdx, 0(%rbx, %rax, 8) 114 leaq level2_spare_pgt(%rip), %rbx 115 movq %rdx, 0(%rbx, %rax, 8)
|
D | entry_64.S | 310 CFI_REL_OFFSET rbx, RBX+\offset 362 movq_cfi rbx, RBX+16 387 movq_cfi rbx, RBX+8 731 movq_cfi_restore RBX+8, rbx 1463 movq_cfi rbx, RBX+8
|
/arch/x86/power/ |
D | hibernate_asm_64.S | 31 movq %rbx, pt_regs_bx(%rax) 72 movq restore_cr3(%rip), %rbx 111 movq %rbx, %cr3 128 movq pt_regs_bx(%rax), %rbx
|
/arch/x86/include/asm/ |
D | calling.h | 127 movq %rbx, 5*8(%rsp) 128 CFI_REL_OFFSET rbx, 5*8 152 movq 5*8(%rsp), %rbx 153 CFI_RESTORE rbx
|
D | sigcontext.h | 244 unsigned long rbx; member
|
D | ptrace.h | 74 unsigned long rbx; member
|
D | kvm.h | 82 __u64 rax, rbx, rcx, rdx; member
|
/arch/x86/crypto/ |
D | salsa20-x86_64-asm_64.S | 36 # rbx_stack = rbx 37 movq %rbx,40(%rsp) 151 mov %r15,%rbx 221 xor %r15,%rbx 223 lea (%rbp,%rbx),%r15 229 lea (%rbx,%rcx),%r15 353 lea (%rbx,%r15),%rbp 369 xor %rbp,%rbx 371 lea (%r13,%rbx),%rbp 437 xor %r15,%rbx [all …]
|
D | aes-x86_64-asm_64.S | 27 #define R2 %rbx
|
D | twofish-x86_64-asm_64.S | 46 #define R1 %rbx
|
/arch/x86/kvm/ |
D | svm.c | 1804 [rbx]"i"(offsetof(struct vcpu_svm, vcpu.arch.regs[VCPU_REGS_RBX])), in svm_vcpu_run()
|
D | x86.c | 3171 regs->rbx = kvm_register_read(vcpu, VCPU_REGS_RBX); in kvm_arch_vcpu_ioctl_get_regs() 3208 kvm_register_write(vcpu, VCPU_REGS_RBX, regs->rbx); in kvm_arch_vcpu_ioctl_set_regs()
|
D | vmx.c | 3417 [rbx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RBX])), in vmx_vcpu_run()
|