Lines Matching refs:eax
179 movl %cr3, %eax
180 testl $PTI_SWITCH_MASK, %eax
274 pushl %eax
276 movl 5*4(%esp), %eax # (modified) regs->sp
279 movl %ecx, -4(%eax)
283 movl %ecx, -8(%eax)
286 movl %ecx, -12(%eax)
288 movl 1*4(%esp), %ecx # eax
289 movl %ecx, -16(%eax)
292 lea -16(%eax), %esp
293 popl %eax
297 .macro SAVE_ALL pt_regs_ax=%eax switch_stacks=0 skip_gs=0
351 popl %eax
402 movl PT_EFLAGS(%esp), %eax # mix EFLAGS, SS and CS
410 andl $(X86_EFLAGS_VM | (SEGMENT_TI_MASK << 8) | SEGMENT_RPL_MASK), %eax
411 cmpl $((SEGMENT_LDT << 8) | USER_RPL), %eax
426 mov PT_OLDESP(%esp), %eax /* load userspace esp */
428 sub %eax, %edx /* offset (low word is 0) */
433 pushl %eax /* new kernel esp */
462 SWITCH_TO_KERNEL_CR3 scratch_reg=%eax
578 testl $PTI_SWITCH_MASK, %eax
704 SWITCH_TO_USER_CR3 scratch_reg=%eax
724 movl %esp, TASK_threadsp(%eax)
763 pushl %eax
765 popl %eax
785 movl %esp, %eax
791 1: movl %edi, %eax
815 movl PT_EFLAGS(%esp), %eax # mix EFLAGS and CS
817 andl $(X86_EFLAGS_VM | SEGMENT_RPL_MASK), %eax
822 movl PT_CS(%esp), %eax
823 andl $SEGMENT_RPL_MASK, %eax
825 cmpl $USER_RPL, %eax
831 movl %esp, %eax
895 pushl %eax
897 SWITCH_TO_KERNEL_CR3 scratch_reg=%eax
898 popl %eax
911 pushl %eax /* pt_regs->orig_ax */
943 movl %esp, %eax
960 movl PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %eax
961 subl $(2*4), %eax
966 movl %edi, (%eax)
967 movl %esi, 4(%eax)
982 movl %eax, %esp
985 SWITCH_TO_USER_CR3 scratch_reg=%eax
995 popl %eax
1048 pushl %eax /* pt_regs->orig_ax */
1058 movl %esp, %eax
1071 SWITCH_TO_USER_CR3 scratch_reg=%eax
1114 pushl %eax
1115 SWITCH_TO_USER_CR3 scratch_reg=%eax
1116 popl %eax
1136 shl $16, %eax
1137 addl %esp, %eax /* the adjusted stack pointer */
1139 pushl %eax
1145 movl %ss, %eax
1149 movl $__KERNEL_DS, %eax
1150 movl %eax, %ds
1151 movl %eax, %es
1191 movl %esp, %eax
1209 movl %esp, %eax
1221 movl %esp, %eax; \
1356 movl PT_EIP(%esp), %eax
1357 cmpl $xen_iret_start_crit, %eax
1359 cmpl $xen_iret_end_crit, %eax
1365 1: mov %esp, %eax
1386 pushl %eax
1387 movl $1, %eax
1394 testl %eax, %eax
1395 popl %eax
1405 6: xorl %eax, %eax
1406 movl %eax, 4(%esp)
1408 7: xorl %eax, %eax
1409 movl %eax, 8(%esp)
1411 8: xorl %eax, %eax
1412 movl %eax, 12(%esp)
1414 9: xorl %eax, %eax
1415 movl %eax, 16(%esp)
1463 GET_CR2_INTO(%ecx) # might clobber %eax
1470 movl %esp, %eax # pt_regs pointer
1492 movl %esp, %eax # pt_regs pointer
1518 pushl %eax
1519 movl %ss, %eax
1521 popl %eax
1525 pushl %eax # pt_regs->orig_ax
1529 movl %esp, %eax # pt_regs pointer
1534 subl %eax, %ecx /* ecx = (end of entry_stack) - esp */
1569 pushl %eax
1572 FIXUP_ESPFIX_STACK # %eax == %esp
1589 movl %esp, %eax # pt_regs pointer