/arch/x86/lib/ |
D | memmove_64.S | 36 cmp %rdi, %rsi 38 mov %rsi, %r8 63 movq 0*8(%rsi), %r11 64 movq 1*8(%rsi), %r10 65 movq 2*8(%rsi), %r9 66 movq 3*8(%rsi), %r8 67 leaq 4*8(%rsi), %rsi 83 movq -8(%rsi, %rdx), %r11 97 movq (%rsi), %r11 99 leaq -8(%rsi, %rdx), %rsi [all …]
|
D | copy_page_64.S | 34 movq 0x8*0(%rsi), %rax 35 movq 0x8*1(%rsi), %rbx 36 movq 0x8*2(%rsi), %rdx 37 movq 0x8*3(%rsi), %r8 38 movq 0x8*4(%rsi), %r9 39 movq 0x8*5(%rsi), %r10 40 movq 0x8*6(%rsi), %r11 41 movq 0x8*7(%rsi), %r12 43 prefetcht0 5*64(%rsi) 54 leaq 64 (%rsi), %rsi [all …]
|
D | memcpy_64.S | 77 movq 0*8(%rsi), %r8 78 movq 1*8(%rsi), %r9 79 movq 2*8(%rsi), %r10 80 movq 3*8(%rsi), %r11 81 leaq 4*8(%rsi), %rsi 96 addq %rdx, %rsi 106 movq -1*8(%rsi), %r8 107 movq -2*8(%rsi), %r9 108 movq -3*8(%rsi), %r10 109 movq -4*8(%rsi), %r11 [all …]
|
D | copy_user_nocache_64.S | 29 100: movb (%rsi),%al 31 incq %rsi 60 1: movq (%rsi),%r8 61 2: movq 1*8(%rsi),%r9 62 3: movq 2*8(%rsi),%r10 63 4: movq 3*8(%rsi),%r11 68 9: movq 4*8(%rsi),%r8 69 10: movq 5*8(%rsi),%r9 70 11: movq 6*8(%rsi),%r10 71 12: movq 7*8(%rsi),%r11 [all …]
|
D | csum-copy_64.S | 121 movq %rbx, (%rsi) 123 movq %r8, 8(%rsi) 125 movq %r11, 16(%rsi) 127 movq %rdx, 24(%rsi) 130 movq %r10, 32(%rsi) 132 movq %rbp, 40(%rsi) 134 movq %r14, 48(%rsi) 136 movq %r13, 56(%rsi) 141 leaq 64(%rsi), %rsi 162 movq %rbx, (%rsi) [all …]
|
D | copy_user_64.S | 56 100: movb (%rsi),%al 58 incq %rsi 92 movq %rsi,%rcx 142 1: movq (%rsi),%r8 143 2: movq 1*8(%rsi),%r9 144 3: movq 2*8(%rsi),%r10 145 4: movq 3*8(%rsi),%r11 150 9: movq 4*8(%rsi),%r8 151 10: movq 5*8(%rsi),%r9 152 11: movq 6*8(%rsi),%r10 [all …]
|
D | cmpxchg16b_emu.S | 38 cmpq PER_CPU_VAR((%rsi)), %rax 40 cmpq PER_CPU_VAR(8(%rsi)), %rdx 43 movq %rbx, PER_CPU_VAR((%rsi)) 44 movq %rcx, PER_CPU_VAR(8(%rsi))
|
D | rwsem.S | 68 pushq_cfi %rsi; CFI_REL_OFFSET rsi, 0; \ 81 popq_cfi %rsi; CFI_RESTORE rsi; \
|
/arch/x86/crypto/ |
D | salsa20-x86_64-asm_64.S | 12 mov %rsi,%rsi 91 leaq 192(%rsp),%rsi 98 movq %rsi,144(%rsp) 108 movq 64(%rsp),%rsi 110 mov %rsi,%rcx 112 shr $32,%rsi 248 xor %rbp,%rsi 250 lea (%r15,%rsi),%rbp 256 lea (%rsi,%r8),%rbp 272 lea (%rsi,%rdx),%rbp [all …]
|
D | camellia-aesni-avx-asm_64.S | 899 movq %rsi, %rax; 905 %xmm8, %rsi); 927 movq %rsi, %rax; 933 %xmm8, %rsi); 982 %xmm8, %rsi); 1093 %xmm8, %rsi); 1125 vmovdqu %xmm0, 0 * 16(%rsi); 1131 vmovdqu %xmm0, 1 * 16(%rsi); 1135 vmovdqu %xmm0, 2 * 16(%rsi); 1139 vmovdqu %xmm0, 3 * 16(%rsi); [all …]
|
D | camellia-aesni-avx2-asm_64.S | 941 movq %rsi, %rax; 947 %ymm8, %rsi); 973 movq %rsi, %rax; 979 %ymm8, %rsi); 1005 cmpq %rsi, %rdx; 1009 movq %rsi, %rax; 1045 %ymm8, %rsi); 1077 cmpq %rsi, %rdx; 1081 movq %rsi, %rax; 1183 %ymm8, %rsi); [all …]
|
D | blowfish-x86_64-asm_64.S | 37 #define RIO %rsi 60 #define RT1 %rsi 130 movq %rsi, %r10; 166 movq %rsi, %r10; 312 movq %rsi, %r11; 357 movq %rsi, %r11;
|
D | ghash-clmulni-intel_asm.S | 98 movups (%rsi), SHASH 120 movups (%rsi), IN1 125 add $16, %rsi
|
D | cast6-avx-x86_64-asm_64.S | 74 #define RID2 %rsi 353 movq %rsi, %r11; 371 movq %rsi, %r11; 391 movq %rsi, %r11; 415 movq %rsi, %r11; 438 movq %rsi, %r11; 441 load_xts_8way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2, 460 movq %rsi, %r11; 463 load_xts_8way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2,
|
/arch/x86/kernel/ |
D | relocate_kernel_64.S | 61 movq PTR(VA_CONTROL_PAGE)(%rsi), %r11 78 movq PTR(PA_CONTROL_PAGE)(%rsi), %r8 81 movq PTR(PA_TABLE_PAGE)(%rsi), %r9 84 movq PTR(PA_SWAP_PAGE)(%rsi), %r10 242 movq %rcx, %rsi /* For ever source page do a copy */ 243 andq $0xfffffffffffff000, %rsi 246 movq %rsi, %rax 253 movq %rdx, %rsi 258 movq %r10, %rsi 262 lea PAGE_SIZE(%rax), %rsi
|
D | mcount_64.S | 42 movq SS+16(%rsp), %rsi 44 movq 8(%rbp), %rsi 155 movq SS+16(%rsp), %rsi 157 movq 8(%rbp), %rsi 181 movq RIP(%rsp), %rsi 182 subq $MCOUNT_INSN_SIZE, %rsi
|
D | entry_64.S | 218 CFI_REL_OFFSET rsi, RSI+\offset-ARGOFFSET 246 movq_cfi rsi, (RSI-RBP) 259 movq %rsp, %rsi 262 testl $3, CS-RBP(%rsi) 273 CFI_DEF_CFA_REGISTER rsi 276 pushq %rsi 290 movq %rsi, RSI+8(%rsp) 489 movq RAX-ARGOFFSET(%rsp),%rsi /* second arg, syscall return value */ 490 cmpq $-MAX_ERRNO,%rsi /* is it < -MAX_ERRNO? */ 501 movq $AUDIT_ARCH_X86_64, %rsi [all …]
|
/arch/x86/boot/compressed/ |
D | head_64.S | 271 mov %rax, %rsi 273 movl %eax, BP_code32_start(%rsi) 289 movq %rax,%rsi 328 movl BP_kernel_alignment(%rsi), %eax 353 pushq %rsi 354 leaq (_bss-8)(%rip), %rsi 361 popq %rsi 373 movq %rsi, efi64_config+8(%rip) /* EFI System table pointer */ 378 movq %rdx, %rsi 412 pushq %rsi /* Save the real mode argument */ [all …]
|
/arch/x86/um/ |
D | setjmp_64.S | 24 pop %rsi # Return address, and adjust the stack 28 push %rsi # Make the call/return stack happy 34 movq %rsi,56(%rdi) # Return address
|
/arch/x86/purgatory/ |
D | entry64.S | 44 movq rsi(%rip), %rsi 69 rsi: .quad 0x0 label
|
/arch/x86/kernel/acpi/ |
D | wakeup_64.S | 31 movq saved_rsi, %rsi 49 movq %rsi, pt_regs_si(%rax) 71 movq %rsi, saved_rsi 96 movq pt_regs_si(%rax), %rsi
|
/arch/x86/platform/efi/ |
D | efi_stub_64.S | 37 mov 0x8(%rsp), %rsi; \ 38 mov %rsi, %cr0; \ 85 mov %rsi, %rcx
|
/arch/x86/include/asm/ |
D | ftrace.h | 15 movq %rsi, RSI(%rsp) 28 movq RSI(%rsp), %rsi
|
/arch/x86/net/ |
D | bpf_jit.S | 39 mov (SKBDATA,%rsi),%eax 56 movzwl (SKBDATA,%rsi),%eax 71 movzbl (SKBDATA,%rsi),%eax
|
/arch/x86/xen/ |
D | xen-asm.S | 125 push %rsi 137 pop %rsi
|