Lines Matching refs:esp
71 testl $USER_SEGMENT_RPL_MASK, PT_CS(%esp)
111 andl $0x0000ffff, 4*4(%esp)
114 testl $X86_EFLAGS_VM, 5*4(%esp)
117 testl $USER_SEGMENT_RPL_MASK, 4*4(%esp)
120 orl $CS_FROM_KERNEL, 4*4(%esp)
155 pushl %esp # sp (points at ss)
156 addl $7*4, (%esp) # point sp back at the previous context
157 pushl 7*4(%esp) # flags
158 pushl 7*4(%esp) # cs
159 pushl 7*4(%esp) # ip
160 pushl 7*4(%esp) # orig_eax
161 pushl 7*4(%esp) # gs / function
162 pushl 7*4(%esp) # fs
174 testl $CS_FROM_KERNEL, 1*4(%esp)
184 movl 5*4(%esp), %eax # (modified) regs->sp
186 movl 4*4(%esp), %ecx # flags
189 movl 3*4(%esp), %ecx # cs
193 movl 2*4(%esp), %ecx # ip
196 movl 1*4(%esp), %ecx # eax
200 lea -4*4(%eax), %esp
271 addl $(4 + \pop), %esp /* pop the unused "gs" slot */
274 4: movl $0, (%esp)
276 5: movl $0, (%esp)
278 6: movl $0, (%esp)
315 movl PT_EFLAGS(%esp), %eax # mix EFLAGS, SS and CS
321 movb PT_OLDSS(%esp), %ah
322 movb PT_CS(%esp), %al
338 mov %esp, %edx /* load kernel esp */
339 mov PT_OLDESP(%esp), %eax /* load userspace esp */
353 lss (%esp), %esp /* switch to espfix segment */
383 subl %esp, %ecx /* ecx = (end of entry_stack) - esp */
388 movl %esp, %esi
400 movl PT_EFLAGS(%esp), %ecx # mix EFLAGS and CS
401 movb PT_CS(%esp), %cl
404 movl PT_CS(%esp), %ecx
430 movl %edi, %esp
483 orl $CS_FROM_ENTRY_STACK, PT_CS(%esp)
491 orl $CS_FROM_USER_CR3, PT_CS(%esp)
523 testl $(X86_EFLAGS_VM), PT_EFLAGS(%esp)
535 movl %esp, %esi
550 movl %ebx, %esp
569 testl $CS_FROM_ENTRY_STACK, PT_CS(%esp)
575 andl $(~CS_FROM_ENTRY_STACK), PT_CS(%esp)
578 movl %esp, %esi
601 movl %ebx, %esp
607 testl $CS_FROM_USER_CR3, PT_CS(%esp)
611 andl $(~CS_FROM_USER_CR3), PT_CS(%esp)
647 movl %esp, %eax
648 movl PT_ORIG_EAX(%esp), %edx /* get the vector from stack */
649 movl $-1, PT_ORIG_EAX(%esp) /* no syscall to restart */
696 movl %esp, TASK_threadsp(%eax)
697 movl TASK_threadsp(%edx), %esp
763 movl %esp, %eax
775 movl $0, PT_EAX(%esp)
836 movl TSS_entry2task_stack(%esp), %esp
866 testl $X86_EFLAGS_NT|X86_EFLAGS_AC|X86_EFLAGS_TF, PT_EFLAGS(%esp)
870 movl %esp, %eax
889 movl PT_EFLAGS(%esp), %edi
890 movl PT_EAX(%esp), %esi
895 movl PT_EIP(%esp), %edx /* pt_regs->ip */
896 movl PT_OLDESP(%esp), %ecx /* pt_regs->sp */
897 1: mov PT_FS(%esp), %fs
900 addl $2*4, %esp /* skip pt_regs->cx and pt_regs->dx */
906 movl %eax, %esp
916 btrl $X86_EFLAGS_IF_BIT, (%esp)
929 2: movl $0, PT_FS(%esp)
975 movl %esp, %eax
1037 subl $2*4, %esp
1038 sgdt (%esp)
1039 movl 2(%esp), %ecx /* GDT address */
1047 addl $2*4, %esp
1049 addl %esp, %eax /* the adjusted stack pointer */
1052 lss (%esp), %esp /* switch to the normal stack segment */
1074 movl PT_GS(%esp), %edi # get the function address
1077 movl PT_ORIG_EAX(%esp), %edx # get the error code
1078 movl $-1, PT_ORIG_EAX(%esp) # no syscall to restart
1080 movl %esp, %eax # pt_regs pointer
1085 movl PT_EFLAGS(%esp), %eax # mix EFLAGS and CS
1086 movb PT_CS(%esp), %al
1092 movl PT_CS(%esp), %eax
1104 movl %esp, %eax
1174 movl %esp, %eax # pt_regs pointer
1192 movl %esp, %ebx
1193 movl PER_CPU_VAR(cpu_current_top_of_stack), %esp
1195 movl %ebx, %esp
1199 testl $CS_FROM_ESPFIX, PT_CS(%esp)
1213 pushl %esp
1214 addl $4, (%esp)
1217 pushl 4*4(%esp) # flags
1218 pushl 4*4(%esp) # cs
1219 pushl 4*4(%esp) # ip
1227 xorl $(CS_FROM_ESPFIX | CS_FROM_KERNEL), PT_CS(%esp)
1230 movl %esp, %eax # pt_regs pointer
1245 lss (1+5+6)*4(%esp), %esp # back to espfix stack
1256 leal -TOP_OF_KERNEL_STACK_PADDING-PTREGS_SIZE(%esi), %esp