• Home
  • Raw
  • Download

Lines Matching refs:rsp

66 	TRACE_IRQS_FLAGS EFLAGS(%rsp)
95 btl $9, EFLAGS(%rsp) /* interrupts off? */
155 movq %rsp, PER_CPU_VAR(cpu_tss_rw + TSS_sp2)
156 SWITCH_TO_KERNEL_CR3 scratch_reg=%rsp
157 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
174 movq %rsp, %rsi
188 movq RCX(%rsp), %rcx
189 movq RIP(%rsp), %r11
217 cmpq $__USER_CS, CS(%rsp) /* CS must match SYSRET */
220 movq R11(%rsp), %r11
221 cmpq %r11, EFLAGS(%rsp) /* R11 == RFLAGS */
247 cmpq $__USER_DS, SS(%rsp) /* SS must match SYSRET */
262 movq %rsp, %rdi
263 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
278 popq %rsp
300 movq %rsp, TASK_threadsp(%rdi)
301 movq TASK_threadsp(%rsi), %rsp
345 movq %rsp, %rdi
360 movq $0, RAX(%rsp)
420 leaq 8(%rsp), \old_rsp
422 movq %rsp, \old_rsp
448 movq PER_CPU_VAR(hardirq_stack_ptr), %rsp
457 cmpq -8(%rsp), \old_rsp
486 popq %rsp
521 testb $3, CS-ORIG_RAX+8(%rsp)
535 movq %rsp, %rdi
536 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
561 testb $3, CS+8(%rsp)
595 addq $-0x80, (%rsp) /* Adjust vector to [-256, -1] range */
606 addq $-0x80, (%rsp) /* Adjust vector to [-256, -1] range */
617 testb $3, CS(%rsp)
622 mov %rsp,%rdi
630 testb $3, CS(%rsp)
641 movq %rsp, %rdi
642 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
674 btl $9, EFLAGS(%rsp) /* were interrupts off? */
689 testb $3, CS(%rsp)
695 addq $8, %rsp /* skip regs->orig_ax */
709 testb $4, (SS-RIP)(%rsp)
752 movq (1*8)(%rsp), %rax /* user RIP */
754 movq (2*8)(%rsp), %rax /* user CS */
756 movq (3*8)(%rsp), %rax /* user RFLAGS */
758 movq (5*8)(%rsp), %rax /* user SS */
760 movq (4*8)(%rsp), %rax /* user RSP */
780 movq %rax, %rsp
899 testb $3, CS(%rsp)
905 movq %rsp, %rdi /* pt_regs pointer */
908 movq ORIG_RAX(%rsp), %rsi /* get error code */
909 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
995 testb $3, CS-ORIG_RAX(%rsp) /* If coming from userspace, switch stacks */
1004 testb $3, CS-ORIG_RAX(%rsp)
1007 pushq 5*8(%rsp)
1058 TRACE_IRQS_FLAGS (%rsp)
1084 mov %rsp, %rbp
1115 movq %rdi, %rsp /* we don't return, adjust the stack frame */
1144 cmpw %cx, 0x10(%rsp)
1147 cmpw %cx, 0x18(%rsp)
1150 cmpw %cx, 0x20(%rsp)
1153 cmpw %cx, 0x28(%rsp)
1156 movq (%rsp), %rcx
1157 movq 8(%rsp), %r11
1158 addq $0x30, %rsp
1163 movq (%rsp), %rcx
1164 movq 8(%rsp), %r11
1165 addq $0x30, %rsp
1310 testb $3, CS+8(%rsp)
1326 movq %rsp, %rdi /* arg0 = pt_regs pointer */
1328 movq %rax, %rsp /* switch stack */
1346 cmpq %rcx, RIP+8(%rsp)
1349 cmpq %rax, RIP+8(%rsp)
1351 cmpq $.Lgs_change, RIP+8(%rsp)
1366 movq %rcx, RIP+8(%rsp)
1383 mov %rsp, %rdi
1385 mov %rax, %rsp
1393 testb $3, CS(%rsp)
1452 testb $3, CS-RIP+8(%rsp)
1470 movq %rsp, %rdx
1471 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
1491 movq %rsp, %rdi
1555 cmpq 8(%rsp), %rdx
1558 cmpq 8(%rsp), %rdx
1567 cmpl $1, -8(%rsp)
1582 lea 6*8(%rsp), %rdx
1584 cmpq %rdx, 4*8(%rsp)
1589 cmpq %rdx, 4*8(%rsp)
1595 testb $(X86_EFLAGS_DF >> 8), (3*8 + 1)(%rsp)
1605 subq $8, %rsp
1606 leaq -10*8(%rsp), %rdx
1614 addq $(6*8), %rsp
1624 movq (%rsp), %rdx
1630 subq $(5*8), %rsp
1634 pushq 11*8(%rsp)
1646 pushq %rsp /* RSP (minus 8 because of the previous push) */
1647 addq $8, (%rsp) /* Fix up RSP */
1672 movq $1, 10*8(%rsp) /* Set "NMI executing". */
1679 addq $(10*8), %rsp
1681 pushq -6*8(%rsp)
1683 subq $(5*8), %rsp
1704 movq %rsp, %rdi
1725 addq $6*8, %rsp
1737 movq $0, 5*8(%rsp) /* clear "NMI executing" */
1766 leaq -PTREGS_SIZE(%rax), %rsp