/arch/x86/lib/ |
D | memmove_64.S | 31 mov %rdi, %rax 36 cmp %rdi, %rsi 40 cmp %rdi, %r8 69 movq %r11, 0*8(%rdi) 70 movq %r10, 1*8(%rdi) 71 movq %r9, 2*8(%rdi) 72 movq %r8, 3*8(%rdi) 73 leaq 4*8(%rdi), %rdi 84 lea -8(%rdi, %rdx), %r10 98 movq %rdi, %r10 [all …]
|
D | memcpy_64.S | 29 movq %rdi, %rax 49 movq %rdi, %rax 59 movq %rdi, %rax 83 movq %r8, 0*8(%rdi) 84 movq %r9, 1*8(%rdi) 85 movq %r10, 2*8(%rdi) 86 movq %r11, 3*8(%rdi) 87 leaq 4*8(%rdi), %rdi 97 addq %rdx, %rdi 111 movq %r8, -1*8(%rdi) [all …]
|
D | copy_page_64.S | 45 movq %rax, 0x8*0(%rdi) 46 movq %rbx, 0x8*1(%rdi) 47 movq %rdx, 0x8*2(%rdi) 48 movq %r8, 0x8*3(%rdi) 49 movq %r9, 0x8*4(%rdi) 50 movq %r10, 0x8*5(%rdi) 51 movq %r11, 0x8*6(%rdi) 52 movq %r12, 0x8*7(%rdi) 55 leaq 64 (%rdi), %rdi 73 movq %rax, 0x8*0(%rdi) [all …]
|
D | memset_64.S | 21 movq %rdi,%r9 50 movq %rdi,%r9 62 movq %rdi,%r10 83 movq %rax,(%rdi) 84 movq %rax,8(%rdi) 85 movq %rax,16(%rdi) 86 movq %rax,24(%rdi) 87 movq %rax,32(%rdi) 88 movq %rax,40(%rdi) 89 movq %rax,48(%rdi) [all …]
|
D | copy_user_nocache_64.S | 30 101: movb %al,(%rdi) 32 incq %rdi 64 5: movnti %r8,(%rdi) 65 6: movnti %r9,1*8(%rdi) 66 7: movnti %r10,2*8(%rdi) 67 8: movnti %r11,3*8(%rdi) 72 13: movnti %r8,4*8(%rdi) 73 14: movnti %r9,5*8(%rdi) 74 15: movnti %r10,6*8(%rdi) 75 16: movnti %r11,7*8(%rdi) [all …]
|
D | csum-copy_64.S | 89 movq (%rdi), %rbx 91 movq 8(%rdi), %r8 93 movq 16(%rdi), %r11 95 movq 24(%rdi), %rdx 98 movq 32(%rdi), %r10 100 movq 40(%rdi), %rbp 102 movq 48(%rdi), %r14 104 movq 56(%rdi), %r13 107 prefetcht0 5*64(%rdi) 140 leaq 64(%rdi), %rdi [all …]
|
D | copy_user_64.S | 57 101: movb %al,(%rdi) 59 incq %rdi 77 movq %rdi,%rcx 146 5: movq %r8,(%rdi) 147 6: movq %r9,1*8(%rdi) 148 7: movq %r10,2*8(%rdi) 149 8: movq %r11,3*8(%rdi) 154 13: movq %r8,4*8(%rdi) 155 14: movq %r9,5*8(%rdi) 156 15: movq %r10,6*8(%rdi) [all …]
|
D | rwsem.S | 67 pushq_cfi %rdi; CFI_REL_OFFSET rdi, 0; \ 82 popq_cfi %rdi; CFI_RESTORE rdi 92 movq %rax,%rdi 104 movq %rax,%rdi 117 movq %rax,%rdi 129 movq %rax,%rdi
|
D | msr-reg.S | 19 movq %rdi, %r10 /* Save pointer */ 21 movl (%rdi), %eax 22 movl 4(%rdi), %ecx 23 movl 8(%rdi), %edx 24 movl 12(%rdi), %ebx 25 movl 20(%rdi), %ebp 26 movl 24(%rdi), %esi 27 movl 28(%rdi), %edi
|
D | clear_page_64.S | 34 #define PUT(x) movq %rax,x*8(%rdi) 35 movq %rax,(%rdi) 43 leaq 64(%rdi),%rdi
|
/arch/x86/um/ |
D | setjmp_64.S | 26 movq %rbx,(%rdi) 27 movq %rsp,8(%rdi) # Post-return %rsp! 29 movq %rbp,16(%rdi) 30 movq %r12,24(%rdi) 31 movq %r13,32(%rdi) 32 movq %r14,40(%rdi) 33 movq %r15,48(%rdi) 34 movq %rsi,56(%rdi) # Return address 45 movq (%rdi),%rbx 46 movq 8(%rdi),%rsp [all …]
|
/arch/x86/crypto/ |
D | twofish-x86_64-asm_64.S | 81 mov s1(%r11,%rdi,4),%r8d;\ 83 mov s2(%r11,%rdi,4),%r9d;\ 86 xor s2(%r11,%rdi,4),%r8d;\ 89 xor s3(%r11,%rdi,4),%r9d;\ 91 xor s3(%r11,%rdi,4),%r8d;\ 93 xor (%r11,%rdi,4), %r9d;\ 96 xor (%r11,%rdi,4), %r8d;\ 98 xor s1(%r11,%rdi,4),%r9d;\ 119 mov s1(%r11,%rdi,4),%r8d;\ 121 mov s2(%r11,%rdi,4),%r9d;\ [all …]
|
D | salsa20-x86_64-asm_64.S | 10 mov %rdi,%r8 14 mov %rdx,%rdi 81 movq %rdi,128(%rsp) 83 leaq 192(%rsp),%rdi 89 leaq 192(%rsp),%rdi 96 movq %rdi,136(%rsp) 102 movq 56(%rsp),%rdi 104 mov %rdi,%rdx 106 shr $32,%rdi 170 lea (%rdi,%r15),%rbp [all …]
|
D | ghash-clmulni-intel_asm.S | 97 movups (%rdi), DATA 103 movups DATA, (%rdi) 115 movups (%rdi), DATA 129 movups DATA, (%rdi)
|
/arch/x86/kernel/ |
D | head_64.S | 104 leaq _text(%rip), %rdi 107 movq %rdi, %rax 115 movq %rdi, %rax 124 movq %rdi, %rax 125 shrq $PMD_SHIFT, %rdi 129 subq %rdi, %rcx 133 andq $(PTRS_PER_PMD - 1), %rdi 134 movq %rax, (%rbx,%rdi,8) 135 incq %rdi 146 leaq level2_kernel_pgt(%rip), %rdi [all …]
|
D | mcount_64.S | 38 movq RIP(%rsp), %rdi 39 subq $MCOUNT_INSN_SIZE, %rdi 153 movq RIP(%rsp), %rdi 159 subq $MCOUNT_INSN_SIZE, %rdi 175 leaq SS+16(%rsp), %rdi 178 leaq 8(%rbp), %rdi 197 movq %rbp, %rdi 201 movq %rax, %rdi 205 jmp *%rdi
|
D | relocate_kernel_64.S | 89 movq %rdi, CP_PA_BACKUP_PAGES_MAP(%r11) 184 movq CP_PA_BACKUP_PAGES_MAP(%r8), %rdi 214 movq %rdi, %rcx /* Put the page_list in %rcx */ 226 movq %rcx, %rdi 227 andq $0xfffffffffffff000, %rdi 245 movq %rdi, %rdx 248 movq %r10, %rdi 252 movq %rax, %rdi 257 movq %rdx, %rdi
|
D | entry_64.S | 217 CFI_REL_OFFSET rdi, RDI+\offset-ARGOFFSET 245 movq_cfi rdi, (RDI-RBP) 261 leaq -RBP(%rsp),%rdi /* arg1 for handler */ 289 movq %rdi, RDI+8(%rsp) 328 call schedule_tail # rdi: 'prev' task parameter 349 movq %rbp, %rdi 461 pushq_cfi %rdi 463 popq_cfi %rdi 500 leaq -REST_SKIP(%rsp), %rdi 510 FIXUP_TOP_OF_STACK %rdi [all …]
|
/arch/x86/purgatory/ |
D | entry64.S | 45 movq rdi(%rip), %rdi 70 rdi: .quad 0x0 label
|
/arch/x86/kernel/acpi/ |
D | wakeup_64.S | 30 movq saved_rdi, %rdi 50 movq %rdi, pt_regs_di(%rax) 70 movq %rdi, saved_rdi 97 movq pt_regs_di(%rax), %rdi
|
/arch/x86/boot/compressed/ |
D | head_64.S | 267 movq %rax, %rdi 287 movq efi_config(%rip), %rdi 355 leaq (_bss-8)(%rbx), %rdi 372 movq %rdi, efi64_config(%rip) /* Handle */ 390 leaq _bss(%rip), %rdi 392 subq %rdi, %rcx 415 movq %rsi, %rdi /* real mode address */
|
/arch/x86/include/asm/ |
D | ftrace.h | 16 movq %rdi, RDI(%rsp) 27 movq RDI(%rsp), %rdi
|
D | calling.h | 91 movq_cfi rdi, 8*8 141 movq_cfi_restore 8*8, rdi 157 movq \offset+64(%rsp), %rdi
|
/arch/x86/xen/ |
D | xen-asm.S | 126 push %rdi 136 pop %rdi
|
/arch/x86/power/ |
D | hibernate_asm_64.S | 30 movq %rdi, pt_regs_di(%rax) 87 movq pbe_orig_address(%rdx), %rdi 127 movq pt_regs_di(%rax), %rdi
|