Lines Matching refs:eax

178 	movl	%cr3, %eax
179 testl $PTI_SWITCH_MASK, %eax
204 .macro SAVE_ALL pt_regs_ax=%eax switch_stacks=0
271 popl %eax
322 movl PT_EFLAGS(%esp), %eax # mix EFLAGS, SS and CS
330 andl $(X86_EFLAGS_VM | (SEGMENT_TI_MASK << 8) | SEGMENT_RPL_MASK), %eax
331 cmpl $((SEGMENT_LDT << 8) | USER_RPL), %eax
346 mov PT_OLDESP(%esp), %eax /* load userspace esp */
348 sub %eax, %edx /* offset (low word is 0) */
353 pushl %eax /* new kernel esp */
385 SWITCH_TO_KERNEL_CR3 scratch_reg=%eax
508 testl $PTI_SWITCH_MASK, %eax
634 SWITCH_TO_USER_CR3 scratch_reg=%eax
653 movl %esp, TASK_threadsp(%eax)
691 pushl %eax
693 popl %eax
713 movl %esp, %eax
718 1: movl %edi, %eax
742 movl PT_EFLAGS(%esp), %eax # mix EFLAGS and CS
744 andl $(X86_EFLAGS_VM | SEGMENT_RPL_MASK), %eax
749 movl PT_CS(%esp), %eax
750 andl $SEGMENT_RPL_MASK, %eax
752 cmpl $USER_RPL, %eax
758 movl %esp, %eax
835 pushl %eax
837 SWITCH_TO_KERNEL_CR3 scratch_reg=%eax
838 popl %eax
851 pushl %eax /* pt_regs->orig_ax */
883 movl %esp, %eax
898 movl PER_CPU_VAR(cpu_tss_rw + TSS_sp0), %eax
899 subl $(2*4), %eax
904 movl %edi, (%eax)
905 movl %esi, 4(%eax)
920 movl %eax, %esp
923 SWITCH_TO_USER_CR3 scratch_reg=%eax
933 popl %eax
986 pushl %eax /* pt_regs->orig_ax */
996 movl %esp, %eax
1007 SWITCH_TO_USER_CR3 scratch_reg=%eax
1040 pushl %eax
1041 SWITCH_TO_USER_CR3 scratch_reg=%eax
1042 popl %eax
1062 shl $16, %eax
1063 addl %esp, %eax /* the adjusted stack pointer */
1065 pushl %eax
1071 movl %ss, %eax
1075 movl $__KERNEL_DS, %eax
1076 movl %eax, %ds
1077 movl %eax, %es
1111 movl %esp, %eax
1123 movl %esp, %eax; \
1258 movl PT_EIP(%esp), %eax
1259 cmpl $xen_iret_start_crit, %eax
1261 cmpl $xen_iret_end_crit, %eax
1267 1: mov %esp, %eax
1288 pushl %eax
1289 movl $1, %eax
1296 testl %eax, %eax
1297 popl %eax
1307 6: xorl %eax, %eax
1308 movl %eax, 4(%esp)
1310 7: xorl %eax, %eax
1311 movl %eax, 8(%esp)
1313 8: xorl %eax, %eax
1314 movl %eax, 12(%esp)
1316 9: xorl %eax, %eax
1317 movl %eax, 16(%esp)
1356 pushl %eax
1357 movl $(__USER_DS), %eax
1358 movl %eax, %ds
1359 movl %eax, %es
1360 movl $(__KERNEL_PERCPU), %eax
1361 movl %eax, %fs
1379 movl %esp, %eax # pt_regs pointer
1405 pushl %eax
1406 movl %ss, %eax
1408 popl %eax
1412 pushl %eax # pt_regs->orig_ax
1416 movl %esp, %eax # pt_regs pointer
1421 subl %eax, %ecx /* ecx = (end of entry_stack) - esp */
1456 pushl %eax
1459 FIXUP_ESPFIX_STACK # %eax == %esp
1476 movl %esp, %eax # pt_regs pointer