/arch/x86/kernel/ |
D | ftrace_64.S | 68 pushq \added+8*2(%rsp) 70 movq %rsp, %rbp 72 pushq \added+8*3(%rsp) 74 movq %rsp, %rbp 80 subq $(FRAME_SIZE), %rsp 81 movq %rax, RAX(%rsp) 82 movq %rcx, RCX(%rsp) 83 movq %rdx, RDX(%rsp) 84 movq %rsi, RSI(%rsp) 85 movq %rdi, RDI(%rsp) [all …]
|
D | head_64.S | 74 leaq (__end_init_task - SIZEOF_PTREGS)(%rip), %rsp 226 movq initial_stack(%rip), %rsp 307 movq initial_stack(%rip), %rsp 330 movq %rsp, %rdi 331 movq ORIG_RAX(%rsp), %rsi 340 addq $8, %rsp 373 pushq $i # 72(%rsp) Vector number 394 movq 8(%rsp), %rsi /* RSI = vector number */ 395 movq %rdi, 8(%rsp) /* pt_regs->di = RDI */ 411 movq %rsp,%rdi /* RDI = pt_regs; RSI is already trapnr */ [all …]
|
D | sev_verify_cbit.S | 80 xorq %rsp, %rsp 81 subq $0x1000, %rsp
|
/arch/x86/entry/ |
D | entry_64.S | 100 movq %rsp, PER_CPU_VAR(cpu_tss_rw + TSS_sp2) 101 SWITCH_TO_KERNEL_CR3 scratch_reg=%rsp 102 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp 119 movq %rsp, %rsi 132 movq RCX(%rsp), %rcx 133 movq RIP(%rsp), %r11 161 cmpq $__USER_CS, CS(%rsp) /* CS must match SYSRET */ 164 movq R11(%rsp), %r11 165 cmpq %r11, EFLAGS(%rsp) /* R11 == RFLAGS */ 191 cmpq $__USER_DS, SS(%rsp) /* SS must match SYSRET */ [all …]
|
D | entry_64_compat.S | 60 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp 139 testl $X86_EFLAGS_NT|X86_EFLAGS_AC|X86_EFLAGS_TF, EFLAGS(%rsp) 143 movq %rsp, %rdi 213 SWITCH_TO_KERNEL_CR3 scratch_reg=%rsp 216 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp 263 movq %rsp, %rdi 279 movq RBX(%rsp), %rbx /* pt_regs->rbx */ 280 movq RBP(%rsp), %rbp /* pt_regs->rbp */ 281 movq EFLAGS(%rsp), %r11 /* pt_regs->flags (in r11) */ 282 movq RIP(%rsp), %rcx /* pt_regs->ip (in rcx) */ [all …]
|
/arch/x86/platform/efi/ |
D | efi_thunk_64.S | 36 movq %rsp, %rax 37 movq efi_scratch(%rip), %rsp 51 subq $28, %rsp 52 movl %ebx, 0x0(%rsp) /* return address */ 53 movl %esi, 0x4(%rsp) 54 movl %edx, 0x8(%rsp) 55 movl %ecx, 0xc(%rsp) 56 movl %r8d, 0x10(%rsp) 57 movl %r9d, 0x14(%rsp) 64 1: movq 24(%rsp), %rsp
|
D | efi_stub_64.S | 15 movq %rsp, %rbp 16 and $~0xf, %rsp 18 subq $48, %rsp 19 mov %r9, 32(%rsp) 20 mov %rax, 40(%rsp)
|
/arch/x86/crypto/ |
D | chacha-ssse3-x86_64.S | 180 lea 8(%rsp),%r10 181 sub $0x10,%rsp 182 and $~31,%rsp 185 mov %rsp,%rdi 189 pxor 0x00(%rsp),%xmm0 190 movdqa %xmm0,0x00(%rsp) 192 mov %rsp,%rsi 197 lea -8(%r10),%rsp 240 lea 8(%rsp),%r10 241 sub $0x80,%rsp [all …]
|
D | chacha-avx2-x86_64.S | 207 lea 8(%rsp),%r10 208 sub $0x10,%rsp 209 and $~31,%rsp 212 mov %rsp,%rdi 216 vpxor 0x00(%rsp),%xmm7,%xmm7 217 vmovdqa %xmm7,0x00(%rsp) 219 mov %rsp,%rsi 224 lea -8(%r10),%rsp 512 lea 8(%rsp),%r10 513 sub $0x10,%rsp [all …]
|
D | sha1_ni_asm.S | 99 mov %rsp, RSPSAVE 100 sub $FRAME_SIZE, %rsp 101 and $~0xF, %rsp 117 movdqa E0, (0*16)(%rsp) 118 movdqa ABCD, (1*16)(%rsp) 277 sha1nexte (0*16)(%rsp), E0 278 paddd (1*16)(%rsp), ABCD 291 mov RSPSAVE, %rsp
|
D | sha512-avx2-asm.S | 188 add frame_XFER(%rsp),h # h = k + w + h # -- 250 add 1*8+frame_XFER(%rsp), h # h = k + w + h # -- 307 add 2*8+frame_XFER(%rsp), h # h = k + w + h # -- 365 add 3*8+frame_XFER(%rsp), h # h = k + w + h # -- 426 add frame_XFER(%rsp), h # h = k + w + h # -- 464 add 8*1+frame_XFER(%rsp), h # h = k + w + h # -- 502 add 8*2+frame_XFER(%rsp), h # h = k + w + h # -- 540 add 8*3+frame_XFER(%rsp), h # h = k + w + h # -- 574 mov %rsp, %rax 575 sub $frame_size, %rsp [all …]
|
D | poly1305-x86_64-cryptogams.pl | 355 mov 0(%rsp),$ctx 362 mov 8(%rsp),%r15 364 mov 16(%rsp),%r14 366 mov 24(%rsp),%r13 368 mov 32(%rsp),%r12 370 mov 40(%rsp),%rbx 372 lea 48(%rsp),%rsp 439 mov %rsp,%rbp 624 mov %rsp,%rbp 758 mov %rsp,%rbp [all …]
|
D | sha512-avx-asm.S | 102 #define W_t(i) 8*i+frame_W(%rsp) 105 #define WK_2(i) 8*((i%2))+frame_WK(%rsp) 285 mov %rsp, %rax 286 sub $frame_size, %rsp 287 and $~(0x20 - 1), %rsp 288 mov %rax, frame_RSPSAVE(%rsp) 291 mov %rbx, frame_GPRSAVE(%rsp) 292 mov %r12, frame_GPRSAVE +8*1(%rsp) 293 mov %r13, frame_GPRSAVE +8*2(%rsp) 294 mov %r14, frame_GPRSAVE +8*3(%rsp) [all …]
|
D | sha512-ssse3-asm.S | 100 #define W_t(i) 8*i+frame_W(%rsp) 103 #define WK_2(i) 8*((i%2))+frame_WK(%rsp) 287 mov %rsp, %rax 288 sub $frame_size, %rsp 289 and $~(0x20 - 1), %rsp 290 mov %rax, frame_RSPSAVE(%rsp) 293 mov %rbx, frame_GPRSAVE(%rsp) 294 mov %r12, frame_GPRSAVE +8*1(%rsp) 295 mov %r13, frame_GPRSAVE +8*2(%rsp) 296 mov %r14, frame_GPRSAVE +8*3(%rsp) [all …]
|
/arch/x86/um/ |
D | stub_64.S | 9 mov %rbx, %rsp 10 add $0x10, %rsp 13 mov 0x0(%rsp), %rax 22 mov %rsp, 8(%rbx) 25 add %rax, %rsp
|
D | setjmp_64.S | 11 # %rsp (post-return) 28 movq %rsp,8(%rdi) # Post-return %rsp! 47 movq 8(%rdi),%rsp
|
/arch/x86/lib/ |
D | csum-copy_64.S | 40 subq $5*8, %rsp 41 movq %rbx, 0*8(%rsp) 42 movq %r12, 1*8(%rsp) 43 movq %r14, 2*8(%rsp) 44 movq %r13, 3*8(%rsp) 45 movq %r15, 4*8(%rsp) 198 movq 0*8(%rsp), %rbx 199 movq 1*8(%rsp), %r12 200 movq 2*8(%rsp), %r14 201 movq 3*8(%rsp), %r13 [all …]
|
D | copy_page_64.S | 25 subq $2*8, %rsp 26 movq %rbx, (%rsp) 27 movq %r12, 1*8(%rsp) 85 movq (%rsp), %rbx 86 movq 1*8(%rsp), %r12 87 addq $2*8, %rsp
|
/arch/x86/boot/compressed/ |
D | efi_thunk_64.S | 42 subq $32, %rsp 43 movl %esi, 0x0(%rsp) 44 movl %edx, 0x4(%rsp) 45 movl %ecx, 0x8(%rsp) 46 movl %r8d, 0xc(%rsp) 47 movl %r9d, 0x10(%rsp) 49 leaq 0x14(%rsp), %rbx 70 1: addq $32, %rsp
|
D | idt_handlers_64.S | 40 movq %rsp, %rdi 42 movq ORIG_RAX(%rsp), %rsi 63 addq $8, %rsp
|
/arch/x86/xen/ |
D | xen-asm.S | 231 movq %rsp, PER_CPU_VAR(cpu_tss_rw + TSS_sp2) 232 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp 259 addq $8, %rsp /* skip regs->orig_ax */ 289 movq $__USER_DS, 4*8(%rsp) 290 movq $__USER_CS, 1*8(%rsp) 308 movq $__USER32_DS, 4*8(%rsp) 309 movq $__USER32_CS, 1*8(%rsp) 329 movq $__USER32_DS, 4*8(%rsp) 330 movq $__USER32_CS, 1*8(%rsp) 340 lea 16(%rsp), %rsp /* strip %rcx, %r11 */
|
/arch/x86/kernel/acpi/ |
D | wakeup_64.S | 34 movq saved_rsp, %rsp 47 subq $8, %rsp 52 movq %rsp, pt_regs_sp(%rax) 72 movq %rsp, saved_rsp 78 addq $8, %rsp 99 movq pt_regs_sp(%rax), %rsp 120 movq %rsp, %rdi 125 addq $8, %rsp
|
/arch/s390/pci/ |
D | pci_clp.c | 37 static inline void zpci_err_clp(unsigned int rsp, int rc) in zpci_err_clp() argument 40 unsigned int rsp; in zpci_err_clp() member 42 } __packed data = {rsp, rc}; in zpci_err_clp() 135 if (!rc && rrb->response.hdr.rsp == CLP_RC_OK) in clp_query_pci_fngrp() 139 zpci_err_clp(rrb->response.hdr.rsp, rc); in clp_query_pci_fngrp() 202 if (!rc && rrb->response.hdr.rsp == CLP_RC_OK) { in clp_query_pci_fn() 209 zpci_err_clp(rrb->response.hdr.rsp, rc); in clp_query_pci_fn() 246 if (rrb->response.hdr.rsp == CLP_RC_SETPCIFN_BUSY) { in clp_set_pci_fn() 252 } while (rrb->response.hdr.rsp == CLP_RC_SETPCIFN_BUSY); in clp_set_pci_fn() 254 if (!rc && rrb->response.hdr.rsp == CLP_RC_OK) { in clp_set_pci_fn() [all …]
|
/arch/x86/purgatory/ |
D | entry64.S | 30 leaq stack_init(%rip), %rsp 44 movq rsp(%rip), %rsp 66 rsp: .quad 0x0 label
|
/arch/x86/include/asm/ |
D | unwind_hints.h | 18 .macro UNWIND_HINT_REGS base=%rsp offset=0 indirect=0 extra=1 partial=0 19 .if \base == %rsp 51 .macro UNWIND_HINT_IRET_REGS base=%rsp offset=0
|