Lines Matching refs:start
50 uintptr_t start, end; in z_x86_check_stack_bounds() local
62 start = (uintptr_t)K_KERNEL_STACK_BUFFER( in z_x86_check_stack_bounds()
64 end = start + CONFIG_ISR_STACK_SIZE; in z_x86_check_stack_bounds()
75 start = arch_current_thread()->stack_info.start - CONFIG_PRIVILEGED_STACK_SIZE; in z_x86_check_stack_bounds()
76 end = arch_current_thread()->stack_info.start; in z_x86_check_stack_bounds()
80 start = arch_current_thread()->stack_info.start; in z_x86_check_stack_bounds()
81 end = Z_STACK_PTR_ALIGN(arch_current_thread()->stack_info.start + in z_x86_check_stack_bounds()
85 return (addr <= start) || (addr + size > end); in z_x86_check_stack_bounds()
101 uintptr_t start, end; in z_x86_check_guard_page() local
104 start = (uintptr_t)thread->stack_info.mapped.addr - CONFIG_MMU_PAGE_SIZE; in z_x86_check_guard_page()
107 if ((addr >= start) && (addr < end)) { in z_x86_check_guard_page()
112 start = (uintptr_t)thread->stack_info.mapped.addr + thread->stack_info.mapped.sz; in z_x86_check_guard_page()
113 end = start + CONFIG_MMU_PAGE_SIZE; in z_x86_check_guard_page()
115 if ((addr >= start) && (addr < end)) { in z_x86_check_guard_page()
486 if ((void *)esf->rip >= exceptions[i].start && in z_x86_page_fault_handler()
492 if ((void *)esf->eip >= exceptions[i].start && in z_x86_page_fault_handler()