/arch/x86/lib/ |
D | copy_page_64.S | 33 movq 0x8*0(%rsi), %rax 34 movq 0x8*1(%rsi), %rbx 35 movq 0x8*2(%rsi), %rdx 36 movq 0x8*3(%rsi), %r8 37 movq 0x8*4(%rsi), %r9 38 movq 0x8*5(%rsi), %r10 39 movq 0x8*6(%rsi), %r11 40 movq 0x8*7(%rsi), %r12 42 prefetcht0 5*64(%rsi) 53 leaq 64 (%rsi), %rsi [all …]
|
D | memmove_64.S | 33 cmp %rdi, %rsi 35 mov %rsi, %r8 64 movq 0*8(%rsi), %r11 65 movq 1*8(%rsi), %r10 66 movq 2*8(%rsi), %r9 67 movq 3*8(%rsi), %r8 68 leaq 4*8(%rsi), %rsi 84 movq -8(%rsi, %rdx), %r11 98 movq (%rsi), %r11 100 leaq -8(%rsi, %rdx), %rsi [all …]
|
D | memcpy_64.S | 78 movq 0*8(%rsi), %r8 79 movq 1*8(%rsi), %r9 80 movq 2*8(%rsi), %r10 81 movq 3*8(%rsi), %r11 82 leaq 4*8(%rsi), %rsi 97 addq %rdx, %rsi 107 movq -1*8(%rsi), %r8 108 movq -2*8(%rsi), %r9 109 movq -3*8(%rsi), %r10 110 movq -4*8(%rsi), %r11 [all …]
|
D | csum-copy_64.S | 107 movq %rbx, (%rsi) 109 movq %r8, 8(%rsi) 111 movq %r11, 16(%rsi) 113 movq %rdx, 24(%rsi) 116 movq %r10, 32(%rsi) 118 movq %r15, 40(%rsi) 120 movq %r14, 48(%rsi) 122 movq %r13, 56(%rsi) 125 leaq 64(%rsi), %rsi 146 movq %rbx, (%rsi) [all …]
|
D | copy_user_64.S | 28 100: movb (%rsi),%al 30 incq %rsi 66 1: movq (%rsi),%r8 67 2: movq 1*8(%rsi),%r9 68 3: movq 2*8(%rsi),%r10 69 4: movq 3*8(%rsi),%r11 74 9: movq 4*8(%rsi),%r8 75 10: movq 5*8(%rsi),%r9 76 11: movq 6*8(%rsi),%r10 77 12: movq 7*8(%rsi),%r11 [all …]
|
D | copy_mc_64.S | 37 movb (%rsi), %al 40 incq %rsi 52 movq (%rsi), %r8 55 addq $8, %rsi 68 movb (%rsi), %al 71 incq %rsi
|
D | cmpxchg16b_emu.S | 30 cmpq PER_CPU_VAR((%rsi)), %rax 32 cmpq PER_CPU_VAR(8(%rsi)), %rdx 35 movq %rbx, PER_CPU_VAR((%rsi)) 36 movq %rcx, PER_CPU_VAR(8(%rsi))
|
/arch/x86/kernel/ |
D | sev_verify_cbit.S | 25 movq sme_me_mask(%rip), %rsi 26 testq %rsi, %rsi 30 movq sev_status(%rip), %rsi 31 testq %rsi, %rsi 35 movq %cr4, %rsi 38 movq %rsi, %rdx 71 movq %rsi, %cr4
|
D | head_64.S | 68 pushq %rsi 70 popq %rsi 91 pushq %rsi 93 popq %rsi 135 pushq %rsi 137 popq %rsi 164 pushq %rsi 167 popq %rsi 220 pushq %rsi 222 popq %rsi [all …]
|
D | relocate_kernel_64.S | 63 movq PTR(VA_CONTROL_PAGE)(%rsi), %r11 86 movq PTR(PA_CONTROL_PAGE)(%rsi), %r8 89 movq PTR(PA_TABLE_PAGE)(%rsi), %r9 92 movq PTR(PA_SWAP_PAGE)(%rsi), %r10 277 movq %rcx, %rsi /* For ever source page do a copy */ 278 andq $0xfffffffffffff000, %rsi 281 movq %rsi, %rax 288 movq %rdx, %rsi 293 movq %r10, %rsi 297 lea PAGE_SIZE(%rax), %rsi
|
/arch/x86/crypto/ |
D | sm4-aesni-avx-asm_64.S | 236 vmovdqu RA0, 0*16(%rsi); 239 vmovdqu RA1, 1*16(%rsi); 241 vmovdqu RA2, 2*16(%rsi); 244 vmovdqu RA3, 3*16(%rsi); 399 vmovdqu RA0, (0 * 16)(%rsi); 400 vmovdqu RA1, (1 * 16)(%rsi); 401 vmovdqu RA2, (2 * 16)(%rsi); 402 vmovdqu RA3, (3 * 16)(%rsi); 403 vmovdqu RB0, (4 * 16)(%rsi); 405 vmovdqu RB1, (5 * 16)(%rsi); [all …]
|
D | blake2s-core.S | 71 movd (%rsi,%rax,4),%xmm4 73 movd (%rsi,%rax,4),%xmm5 75 movd (%rsi,%rax,4),%xmm6 77 movd (%rsi,%rax,4),%xmm7 92 movd (%rsi,%rax,4),%xmm5 94 movd (%rsi,%rax,4),%xmm6 96 movd (%rsi,%rax,4),%xmm7 98 movd (%rsi,%rax,4),%xmm4 116 movd (%rsi,%rax,4),%xmm6 118 movd (%rsi,%rax,4),%xmm7 [all …]
|
D | sm4-aesni-avx2-asm_64.S | 379 vmovdqu RA0, (0 * 32)(%rsi); 380 vmovdqu RA1, (1 * 32)(%rsi); 381 vmovdqu RA2, (2 * 32)(%rsi); 382 vmovdqu RA3, (3 * 32)(%rsi); 383 vmovdqu RB0, (4 * 32)(%rsi); 384 vmovdqu RB1, (5 * 32)(%rsi); 385 vmovdqu RB2, (6 * 32)(%rsi); 386 vmovdqu RB3, (7 * 32)(%rsi); 433 vmovdqu RA0, (0 * 32)(%rsi); 434 vmovdqu RA1, (1 * 32)(%rsi); [all …]
|
D | chacha-avx512vl-x86_64.S | 29 # %rsi: up to 2 data blocks output, o 118 vmovdqu %xmm6,0x00(%rsi) 125 vmovdqu %xmm6,0x10(%rsi) 132 vmovdqu %xmm6,0x20(%rsi) 139 vmovdqu %xmm6,0x30(%rsi) 147 vmovdqu %xmm6,0x40(%rsi) 153 vmovdqu %xmm6,0x50(%rsi) 159 vmovdqu %xmm6,0x60(%rsi) 165 vmovdqu %xmm6,0x70(%rsi) 186 vmovdqu8 %xmm1,(%rsi,%r9){%k1} [all …]
|
D | chacha-avx2-x86_64.S | 39 # %rsi: up to 2 data blocks output, o 145 vmovdqu %xmm6,0x00(%rsi) 152 vmovdqu %xmm6,0x10(%rsi) 159 vmovdqu %xmm6,0x20(%rsi) 166 vmovdqu %xmm6,0x30(%rsi) 174 vmovdqu %xmm6,0x40(%rsi) 180 vmovdqu %xmm6,0x50(%rsi) 186 vmovdqu %xmm6,0x60(%rsi) 192 vmovdqu %xmm6,0x70(%rsi) 205 mov %rsi,%r11 [all …]
|
D | chacha-ssse3-x86_64.S | 116 # %rsi: up to 1 data block output, o 141 movdqu %xmm0,0x00(%rsi) 149 movdqu %xmm0,0x10(%rsi) 157 movdqu %xmm0,0x20(%rsi) 165 movdqu %xmm0,0x30(%rsi) 178 mov %rsi,%r11 184 lea (%rdx,%rax),%rsi 192 mov %rsp,%rsi 204 # %rsi: output (8 32-bit words) 216 movdqu %xmm0,0x00(%rsi) [all …]
|
D | blowfish-x86_64-asm_64.S | 22 #define RIO %rsi 45 #define RT1 %rsi 116 movq %rsi, %r10; 153 movq %rsi, %r10; 298 movq %rsi, %r11; 345 movq %rsi, %r11
|
/arch/x86/mm/ |
D | mem_encrypt_boot.S | 43 movq %rsi, %r11 /* Decrypted area */ 48 leaq __enc_copy(%rip), %rsi /* Encryption routine */ 54 movq %r11, %rsi /* Decrypted area */ 113 movq %rsi, %r11 /* Save decrypted area address */ 133 movq %r11, %rsi /* Source - decrypted area */ 138 movq %r8, %rsi /* Source - intermediate copy buffer */
|
/arch/x86/boot/compressed/ |
D | head_64.S | 389 movl BP_kernel_alignment(%rsi), %eax 401 movl BP_init_size(%rsi), %ebx 446 pushq %rsi 448 popq %rsi 463 pushq %rsi 464 movq %rsi, %rdi /* real mode address */ 466 popq %rsi 481 pushq %rsi 498 popq %rsi 514 pushq %rsi [all …]
|
D | idt_handlers_64.S | 24 pushq %rsi 42 movq ORIG_RAX(%rsp), %rsi 59 popq %rsi
|
/arch/x86/um/ |
D | setjmp_64.S | 25 pop %rsi # Return address, and adjust the stack 29 push %rsi # Make the call/return stack happy 35 movq %rsi,56(%rdi) # Return address
|
/arch/x86/purgatory/ |
D | entry64.S | 42 movq rsi(%rip), %rsi 68 rsi: .quad 0x0 label
|
/arch/x86/entry/ |
D | thunk_64.S | 19 pushq %rsi 47 popq %rsi
|
/arch/x86/include/uapi/asm/ |
D | sgx.h | 92 typedef int (*sgx_enclave_user_handler_t)(long rdi, long rsi, long rdx, 163 typedef int (*vdso_sgx_enter_enclave_t)(unsigned long rdi, unsigned long rsi,
|
/arch/x86/kernel/acpi/ |
D | wakeup_64.S | 40 movq saved_rsi, %rsi 57 movq %rsi, pt_regs_si(%rax) 79 movq %rsi, saved_rsi 104 movq pt_regs_si(%rax), %rsi
|