Lines Matching refs:rsp
66 TRACE_IRQS_FLAGS EFLAGS(%rsp)
95 btl $9, EFLAGS(%rsp) /* interrupts off? */
171 movq %rsp, RSP_SCRATCH
174 SWITCH_TO_KERNEL_CR3 scratch_reg=%rsp
177 movq CPU_ENTRY_AREA_tss + TSS_sp1 + CPU_ENTRY_AREA, %rsp
219 movq %rsp, PER_CPU_VAR(rsp_scratch)
220 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
237 movq %rsp, %rsi
247 movq RCX(%rsp), %rcx
248 movq RIP(%rsp), %r11
276 cmpq $__USER_CS, CS(%rsp) /* CS must match SYSRET */
279 movq R11(%rsp), %r11
280 cmpq %r11, EFLAGS(%rsp) /* R11 == RFLAGS */
306 cmpq $__USER_DS, SS(%rsp) /* SS must match SYSRET */
322 movq %rsp, %rdi
323 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
335 popq %rsp
357 movq %rsp, TASK_threadsp(%rdi)
358 movq TASK_threadsp(%rsi), %rsp
404 movq %rsp, %rdi
419 movq $0, RAX(%rsp)
467 leaq 8(%rsp), \old_rsp
469 movq %rsp, \old_rsp
495 movq PER_CPU_VAR(irq_stack_ptr), %rsp
504 cmpq -8(%rsp), \old_rsp
533 popq %rsp
568 testb $3, CS-ORIG_RAX+8(%rsp)
582 movq %rsp, %rdi
583 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
606 testb $3, CS+8(%rsp)
640 addq $-0x80, (%rsp) /* Adjust vector to [-256, -1] range */
651 testb $3, CS(%rsp)
656 mov %rsp,%rdi
663 testb $3, CS(%rsp)
674 movq %rsp, %rdi
675 movq PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %rsp
705 btl $9, EFLAGS(%rsp) /* were interrupts off? */
721 testb $3, CS(%rsp)
727 addq $8, %rsp /* skip regs->orig_ax */
741 testb $4, (SS-RIP)(%rsp)
784 movq (1*8)(%rsp), %rax /* user RIP */
786 movq (2*8)(%rsp), %rax /* user CS */
788 movq (3*8)(%rsp), %rax /* user RFLAGS */
790 movq (5*8)(%rsp), %rax /* user SS */
792 movq (4*8)(%rsp), %rax /* user RSP */
812 movq %rax, %rsp
919 testb $3, CS-ORIG_RAX(%rsp) /* If coming from userspace, switch stacks */
939 movq %rsp, %rdi /* pt_regs pointer */
942 movq ORIG_RAX(%rsp), %rsi /* get error code */
943 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
974 movq %rsp, %rdi /* pt_regs pointer */
977 movq ORIG_RAX(%rsp), %rsi /* get error code */
978 movq $-1, ORIG_RAX(%rsp) /* no syscall to restart */
1019 TRACE_IRQS_FLAGS (%rsp)
1045 mov %rsp, %rbp
1076 movq %rdi, %rsp /* we don't return, adjust the stack frame */
1105 cmpw %cx, 0x10(%rsp)
1108 cmpw %cx, 0x18(%rsp)
1111 cmpw %cx, 0x20(%rsp)
1114 cmpw %cx, 0x28(%rsp)
1117 movq (%rsp), %rcx
1118 movq 8(%rsp), %r11
1119 addq $0x30, %rsp
1124 movq (%rsp), %rcx
1125 movq 8(%rsp), %r11
1126 addq $0x30, %rsp
1244 testb $3, CS+8(%rsp)
1258 movq %rsp, %rdi /* arg0 = pt_regs pointer */
1260 movq %rax, %rsp /* switch stack */
1285 cmpq %rcx, RIP+8(%rsp)
1288 cmpq %rax, RIP+8(%rsp)
1290 cmpq $.Lgs_change, RIP+8(%rsp)
1304 movq %rcx, RIP+8(%rsp)
1319 mov %rsp, %rdi
1321 mov %rax, %rsp
1329 testb $3, CS(%rsp)
1388 testb $3, CS-RIP+8(%rsp)
1405 movq %rsp, %rdx
1406 movq PER_CPU_VAR(cpu_current_top_of_stack), %rsp
1424 movq %rsp, %rdi
1488 cmpq 8(%rsp), %rdx
1491 cmpq 8(%rsp), %rdx
1500 cmpl $1, -8(%rsp)
1515 lea 6*8(%rsp), %rdx
1517 cmpq %rdx, 4*8(%rsp)
1522 cmpq %rdx, 4*8(%rsp)
1528 testb $(X86_EFLAGS_DF >> 8), (3*8 + 1)(%rsp)
1538 subq $8, %rsp
1539 leaq -10*8(%rsp), %rdx
1547 addq $(6*8), %rsp
1557 movq (%rsp), %rdx
1563 subq $(5*8), %rsp
1567 pushq 11*8(%rsp)
1579 pushq %rsp /* RSP (minus 8 because of the previous push) */
1580 addq $8, (%rsp) /* Fix up RSP */
1605 movq $1, 10*8(%rsp) /* Set "NMI executing". */
1612 addq $(10*8), %rsp
1614 pushq -6*8(%rsp)
1616 subq $(5*8), %rsp
1637 movq %rsp, %rdi
1655 addq $6*8, %rsp
1667 movq $0, 5*8(%rsp) /* clear "NMI executing" */
1690 leaq -PTREGS_SIZE(%rax), %rsp