Lines Matching refs:rsp
92 movq %rsp, PER_CPU_VAR(cpu_tss_rw + TSS_sp2)
93 SWITCH_TO_KERNEL_CR3 scratch_reg=%rsp
94 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
110 movq %rsp, %rdi
130 movq RCX(%rsp), %rcx
131 movq RIP(%rsp), %r11
159 cmpq $__USER_CS, CS(%rsp) /* CS must match SYSRET */
162 movq R11(%rsp), %r11
163 cmpq %r11, EFLAGS(%rsp) /* R11 == RFLAGS */
189 cmpq $__USER_DS, SS(%rsp) /* SS must match SYSRET */
204 movq %rsp, %rdi
205 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
220 popq %rsp
243 movq %rsp, TASK_threadsp(%rdi)
244 movq TASK_threadsp(%rsi), %rsp
290 movq %rsp, %rdi
304 movq $0, RAX(%rsp)
350 movq %rsp, %rdi /* pt_regs pointer into 1st argument*/
353 movq ORIG_RAX(%rsp), %rsi /* get error code into 2nd argument*/
354 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
387 testb $3, CS-ORIG_RAX(%rsp)
390 pushq 5*8(%rsp)
452 testb $3, CS-ORIG_RAX(%rsp)
460 movq %rsp, %rdi /* pt_regs pointer */
506 testb $3, CS-ORIG_RAX(%rsp)
523 movq %rsp, %rdi /* pt_regs pointer */
525 movq %rax, %rsp /* Switch to new stack */
531 movq ORIG_RAX(%rsp), %rsi /* get error code into 2nd argument*/
532 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
534 movq %rsp, %rdi /* pt_regs pointer */
569 movq %rsp, %rdi /* pt_regs pointer into first argument */
570 movq ORIG_RAX(%rsp), %rsi /* get error code into 2nd argument*/
571 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
600 testb $3, CS(%rsp)
615 movq %rsp, %rdi
616 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
646 testb $3, CS(%rsp)
652 addq $8, %rsp /* skip regs->orig_ax */
666 testb $4, (SS-RIP)(%rsp)
709 movq (1*8)(%rsp), %rax /* user RIP */
711 movq (2*8)(%rsp), %rax /* user CS */
713 movq (3*8)(%rsp), %rax /* user RFLAGS */
715 movq (5*8)(%rsp), %rax /* user SS */
717 movq (4*8)(%rsp), %rax /* user RSP */
737 movq %rax, %rsp
814 movq %rdi, %rsp /* we don't return, adjust the stack frame */
838 cmpw %cx, 0x10(%rsp)
841 cmpw %cx, 0x18(%rsp)
844 cmpw %cx, 0x20(%rsp)
847 cmpw %cx, 0x28(%rsp)
850 movq (%rsp), %rcx
851 movq 8(%rsp), %r11
852 addq $0x30, %rsp
857 movq (%rsp), %rcx
858 movq 8(%rsp), %r11
859 addq $0x30, %rsp
1022 testb $3, CS+8(%rsp)
1036 leaq 8(%rsp), %rdi /* arg0 = pt_regs pointer */
1051 cmpq %rcx, RIP+8(%rsp)
1054 cmpq %rax, RIP+8(%rsp)
1056 cmpq $.Lgs_change, RIP+8(%rsp)
1072 leaq 8(%rsp), %rax /* return pt_regs pointer */
1078 movq %rcx, RIP+8(%rsp)
1096 leaq 8(%rsp), %rdi /* arg0 = pt_regs pointer */
1105 testb $3, CS(%rsp)
1165 testb $3, CS-RIP+8(%rsp)
1182 movq %rsp, %rdx
1183 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
1204 movq %rsp, %rdi
1268 cmpq 8(%rsp), %rdx
1271 cmpq 8(%rsp), %rdx
1280 cmpl $1, -8(%rsp)
1295 lea 6*8(%rsp), %rdx
1297 cmpq %rdx, 4*8(%rsp)
1302 cmpq %rdx, 4*8(%rsp)
1308 testb $(X86_EFLAGS_DF >> 8), (3*8 + 1)(%rsp)
1318 subq $8, %rsp
1319 leaq -10*8(%rsp), %rdx
1327 addq $(6*8), %rsp
1337 movq (%rsp), %rdx
1343 subq $(5*8), %rsp
1347 pushq 11*8(%rsp)
1359 pushq %rsp /* RSP (minus 8 because of the previous push) */
1360 addq $8, (%rsp) /* Fix up RSP */
1385 movq $1, 10*8(%rsp) /* Set "NMI executing". */
1392 addq $(10*8), %rsp
1394 pushq -6*8(%rsp)
1396 subq $(5*8), %rsp
1416 movq %rsp, %rdi
1454 addq $6*8, %rsp
1466 movq $0, 5*8(%rsp) /* clear "NMI executing" */
1496 leaq -PTREGS_SIZE(%rax), %rsp