/Zephyr-latest/tests/arch/arm/arm_thread_swap/src/ |
D | arm_syscalls.c | 40 zassert_true((_current->arch.mode & CONTROL_nPRIV_Msk) == 0, in z_impl_test_arm_user_syscall() 46 ((__get_PSP() >= _current->arch.priv_stack_start) && in z_impl_test_arm_user_syscall() 47 (__get_PSP() < (_current->arch.priv_stack_start + CONFIG_PRIVILEGED_STACK_SIZE))), in z_impl_test_arm_user_syscall() 51 zassert_true(__get_PSPLIM() == _current->arch.priv_stack_start, in z_impl_test_arm_user_syscall() 77 zassert_true((_current->arch.mode & CONTROL_nPRIV_Msk) != 0, in arm_isr_handler() 82 zassert_true(((__get_PSP() >= _current->stack_info.start) && in arm_isr_handler() 83 (__get_PSP() < (_current->stack_info.start + _current->stack_info.size))), in arm_isr_handler() 93 z_move_thread_to_end_of_prio_q(_current); in arm_isr_handler() 160 zassert_true((_current->arch.mode & CONTROL_nPRIV_Msk) == 0, in ZTEST() 165 zassert_true(((__get_PSP() >= _current->stack_info.start) && in ZTEST() [all …]
|
D | arm_thread_arch.c | 225 zassert_true((_current->arch.mode_exc_return & EXC_RETURN_FTYPE) != 0, in alt_thread_entry() 231 zassert_true((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0, in alt_thread_entry() 233 zassert_true((_current->base.user_options & K_FP_REGS) != 0, in alt_thread_entry() 273 z_move_thread_to_end_of_prio_q(_current); in alt_thread_entry() 392 k_thread_priority_set(_current, K_PRIO_COOP(PRIORITY)); in ZTEST() 399 p_ztest_thread = _current; in ZTEST() 404 zassert_true(_current->arch.basepri == 0, in ZTEST() 422 zassert_true((_current->arch.mode & CONTROL_nPRIV_Msk) == 0, in ZTEST() 424 _current->arch.mode); in ZTEST() 429 zassert_true((_current->arch.mode_exc_return & EXC_RETURN_FTYPE) != 0, in ZTEST() [all …]
|
/Zephyr-latest/arch/riscv/core/ |
D | thread.c | 144 _current->arch.priv_stack_start = in arch_user_mode_enter() 145 (unsigned long)z_priv_stack_find(_current->stack_obj); in arch_user_mode_enter() 147 _current->stack_info.start -= K_THREAD_STACK_RESERVED; in arch_user_mode_enter() 148 _current->stack_info.size += K_THREAD_STACK_RESERVED; in arch_user_mode_enter() 150 _current->arch.priv_stack_start = (unsigned long)_current->stack_obj; in arch_user_mode_enter() 152 top_of_priv_stack = Z_STACK_PTR_ALIGN(_current->arch.priv_stack_start + in arch_user_mode_enter() 158 (void)memset((void *)_current->arch.priv_stack_start, 0xaa, in arch_user_mode_enter() 163 _current->stack_info.start + in arch_user_mode_enter() 164 _current->stack_info.size - in arch_user_mode_enter() 165 _current->stack_info.delta); in arch_user_mode_enter() [all …]
|
D | fpu.c | 39 buf[8] = '0' + _current->arch.exception_depth; in DBG() 40 strcat(buf, _current->name); in DBG() 85 atomic_ptr_set(&_current_cpu->arch.fpu_owner, _current); in z_riscv_fpu_load() 89 z_riscv_fpu_restore(&_current->arch.saved_fp_context); in z_riscv_fpu_load() 90 DBG("restore", _current); in z_riscv_fpu_load() 171 if (thread == _current) { in flush_owned_fpu() 216 if (_current->arch.exception_depth > 0) { in z_riscv_fpu_trap() 236 flush_owned_fpu(_current); in z_riscv_fpu_trap() 259 if (_current->arch.exception_depth == exc_update_level) { in fpu_access_allowed() 261 if (_current_cpu->arch.fpu_owner == _current) { in fpu_access_allowed() [all …]
|
D | fatal.c | 161 if (_current->arch.priv_stack_start != 0 && in bad_stack_pointer() 162 sp >= _current->arch.priv_stack_start && in bad_stack_pointer() 163 sp < _current->arch.priv_stack_start + Z_RISCV_STACK_GUARD_SIZE) { in bad_stack_pointer() 167 if (z_stack_is_user_capable(_current->stack_obj) && in bad_stack_pointer() 168 sp >= _current->stack_info.start - K_THREAD_STACK_RESERVED && in bad_stack_pointer() 169 sp < _current->stack_info.start - K_THREAD_STACK_RESERVED in bad_stack_pointer() 176 if (sp >= _current->stack_info.start - K_KERNEL_STACK_RESERVED && in bad_stack_pointer() 177 sp < _current->stack_info.start - K_KERNEL_STACK_RESERVED in bad_stack_pointer() 194 (esf->sp < _current->stack_info.start || in bad_stack_pointer() 195 esf->sp > _current->stack_info.start + in bad_stack_pointer() [all …]
|
/Zephyr-latest/arch/arm/core/cortex_a_r/ |
D | thread.c | 201 _current->arch.priv_stack_start = in arch_user_mode_enter() 202 (uint32_t)z_priv_stack_find(_current->stack_obj); in arch_user_mode_enter() 211 if ((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0) { in arch_user_mode_enter() 212 _current->stack_info.start -= FP_GUARD_EXTRA_SIZE; in arch_user_mode_enter() 213 _current->stack_info.size += FP_GUARD_EXTRA_SIZE; in arch_user_mode_enter() 216 _current->stack_info.start -= MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter() 217 _current->stack_info.size += MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter() 225 _current->arch.priv_stack_start += in arch_user_mode_enter() 226 ((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0) ? in arch_user_mode_enter() 229 _current->arch.priv_stack_start += MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter() [all …]
|
/Zephyr-latest/kernel/ |
D | sched.c | 106 return !IS_ENABLED(CONFIG_SMP) || (thread != _current); in should_queue_thread() 116 if (thread == _current) { in queue_thread() 170 if (is_halting(_current)) { in next_up() 171 halt_thread(_current, is_aborting(_current) ? in next_up() 219 bool queued = z_is_thread_queued(_current); in next_up() 220 bool active = !z_is_thread_prevented_from_running(_current); in next_up() 227 int32_t cmp = z_sched_prio_cmp(_current, thread); in next_up() 231 thread = _current; in next_up() 235 thread = _current; in next_up() 240 if ((thread != _current) && active && in next_up() [all …]
|
D | spinlock_validate.c | 29 if (arch_is_in_isr() && _current->base.thread_state & _THREAD_DUMMY) { in z_spin_unlock_valid() 33 if (tcpu != (_current_cpu->id | (uintptr_t)_current)) { in z_spin_unlock_valid() 42 l->thread_cpu = _current_cpu->id | (uintptr_t)_current; in z_spin_lock_set_owner()
|
D | thread.c | 85 _current->custom_data = value; in z_impl_k_thread_custom_data_set() 98 return _current->custom_data; in z_impl_k_thread_custom_data_get() 113 return !arch_is_in_isr() && thread_is_preemptible(_current); in z_impl_k_is_preempt_thread() 142 thread = _current; in z_impl_k_thread_name_set() 337 if ((_current->base.thread_state & _THREAD_DUMMY) != 0) { in z_check_stack_sentinel() 341 stack = (uint32_t *)_current->stack_info.start; in z_check_stack_sentinel() 619 if (!_current) { in z_setup_new_thread() 628 k_thread_perms_inherit(_current, new_thread); in z_setup_new_thread() 634 new_thread->resource_pool = _current->resource_pool; in z_setup_new_thread() 729 _current->base.prio))); in z_vrfy_k_thread_create() [all …]
|
D | errno.c | 39 return &_current->userspace_local_data->errno_var; in z_impl_z_errno() 51 return &_current->errno_var; in z_impl_z_errno()
|
D | mutex.c | 117 if (likely((mutex->lock_count == 0U) || (mutex->owner == _current))) { in z_impl_k_mutex_lock() 120 _current->base.prio : in z_impl_k_mutex_lock() 124 mutex->owner = _current; in z_impl_k_mutex_lock() 127 _current, mutex, mutex->lock_count, in z_impl_k_mutex_lock() 147 new_prio = new_prio_for_inheritance(_current->base.prio, in z_impl_k_mutex_lock() 160 LOG_DBG("%p got mutex %p (y/n): %c", _current, mutex, in z_impl_k_mutex_lock() 170 LOG_DBG("%p timeout on mutex %p", _current, mutex); in z_impl_k_mutex_lock() 227 CHECKIF(mutex->owner != _current) { in z_impl_k_mutex_unlock()
|
D | smp.c | 61 if (!_current->base.global_lock_count) { in z_smp_global_lock() 67 _current->base.global_lock_count++; in z_smp_global_lock() 74 if (_current->base.global_lock_count != 0U) { in z_smp_global_unlock() 75 _current->base.global_lock_count--; in z_smp_global_unlock() 77 if (!_current->base.global_lock_count) { in z_smp_global_unlock()
|
/Zephyr-latest/arch/posix/core/ |
D | swap.c | 37 _current->callee_saved.key = key; in arch_swap() 38 _current->callee_saved.retval = -EAGAIN; in arch_swap() 50 _current->callee_saved.thread_status; in arch_swap() 69 irq_unlock(_current->callee_saved.key); in arch_swap() 71 return _current->callee_saved.retval; in arch_swap()
|
/Zephyr-latest/arch/arm/core/cortex_m/ |
D | thread.c | 220 _current->arch.priv_stack_start = (uint32_t)z_priv_stack_find(_current->stack_obj); in arch_user_mode_enter() 229 if ((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0) { in arch_user_mode_enter() 230 _current->stack_info.start -= FP_GUARD_EXTRA_SIZE; in arch_user_mode_enter() 231 _current->stack_info.size += FP_GUARD_EXTRA_SIZE; in arch_user_mode_enter() 234 _current->stack_info.start -= MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter() 235 _current->stack_info.size += MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter() 243 _current->arch.priv_stack_start += in arch_user_mode_enter() 244 ((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0) in arch_user_mode_enter() 248 _current->arch.priv_stack_start += MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter() 252 z_arm_userspace_enter(user_entry, p1, p2, p3, (uint32_t)_current->stack_info.start, in arch_user_mode_enter() [all …]
|
/Zephyr-latest/arch/x86/core/ |
D | userspace.c | 135 stack_end = Z_STACK_PTR_ALIGN(_current->stack_info.start + in arch_user_mode_enter() 136 _current->stack_info.size - in arch_user_mode_enter() 137 _current->stack_info.delta); in arch_user_mode_enter() 166 stack_start = POINTER_TO_UINT(_current->stack_obj); in arch_user_mode_enter() 167 stack_size = K_THREAD_STACK_LEN(_current->stack_info.size); in arch_user_mode_enter() 185 _current->stack_info.start); in arch_user_mode_enter()
|
/Zephyr-latest/tests/kernel/threads/thread_apis/src/ |
D | test_essential_thread.c | 30 z_thread_essential_set(_current); in thread_entry() 32 if (z_is_thread_essential(_current)) { in thread_entry() 38 z_thread_essential_clear(_current); in thread_entry() 39 zassert_false(z_is_thread_essential(_current), in thread_entry() 78 if (z_is_thread_essential(_current)) { in abort_thread_entry()
|
/Zephyr-latest/arch/arm64/core/ |
D | thread.c | 162 z_arm64_thread_mem_domains_init(_current); in arch_user_mode_enter() 165 stack_el0 = Z_STACK_PTR_ALIGN(_current->stack_info.start + in arch_user_mode_enter() 166 _current->stack_info.size - in arch_user_mode_enter() 167 _current->stack_info.delta); in arch_user_mode_enter() 170 stack_el1 = (uintptr_t)(_current->stack_obj + ARCH_THREAD_STACK_RESERVED); in arch_user_mode_enter()
|
D | fpu.c | 39 strcat(buf, _current->name); in DBG() 128 if (thread == _current) { in flush_owned_fpu() 263 flush_owned_fpu(_current); in z_arm64_fpu_trap() 267 atomic_ptr_set(&_current_cpu->arch.fpu_owner, _current); in z_arm64_fpu_trap() 270 z_arm64_fpu_restore(&_current->arch.saved_fp_context); in z_arm64_fpu_trap() 271 DBG("restore", _current); in z_arm64_fpu_trap() 290 if (atomic_ptr_get(&_current_cpu->arch.fpu_owner) == _current) { in fpu_access_update()
|
/Zephyr-latest/tests/kernel/usage/thread_runtime_stats/src/ |
D | test_thread_runtime_stats.c | 75 priority = k_thread_priority_get(_current); in ZTEST() 205 priority = k_thread_priority_get(_current); in ZTEST() 218 k_thread_runtime_stats_get(_current, &stats1); in ZTEST() 234 k_thread_runtime_stats_get(_current, &stats2); in ZTEST() 289 k_thread_runtime_stats_get(_current, &thread_stats1); in ZTEST() 294 k_thread_runtime_stats_get(_current, &thread_stats2); in ZTEST() 306 k_thread_runtime_stats_get(_current, &thread_stats3); in ZTEST() 407 priority = k_thread_priority_get(_current); in ZTEST() 417 status = k_thread_runtime_stats_get(_current, NULL); in ZTEST() 431 main_thread = _current; in ZTEST() [all …]
|
/Zephyr-latest/subsys/profiling/perf/backends/ |
D | perf_x86_64.c | 43 buf[idx++] = (uintptr_t)_current->callee_saved.rip; in arch_perf_current_stack_trace() 44 void **fp = (void **)_current->callee_saved.rbp; in arch_perf_current_stack_trace() 56 while (valid_stack((uintptr_t)fp, _current)) { in arch_perf_current_stack_trace()
|
/Zephyr-latest/arch/arm/include/cortex_a_r/ |
D | kernel_arch_func.h | 43 _current->arch.basepri = key; in arch_swap() 44 _current->arch.swap_return_value = -EAGAIN; in arch_swap() 52 return _current->arch.swap_return_value; in arch_swap()
|
/Zephyr-latest/arch/arm/include/cortex_m/ |
D | kernel_arch_func.h | 85 _current->arch.basepri = key; in arch_swap() 86 _current->arch.swap_return_value = -EAGAIN; in arch_swap() 97 return _current->arch.swap_return_value; in arch_swap()
|
/Zephyr-latest/arch/x86/core/ia32/ |
D | float.c | 210 if (thread == _current) { in z_float_enable() 225 if ((_current->base.user_options & _FP_USER_MASK) == 0) { in z_float_enable() 281 if (thread == _current) { in z_float_disable() 317 k_float_enable(_current, _FP_USER_MASK); in _FpNotAvailableExcHandler()
|
/Zephyr-latest/kernel/include/ |
D | priority_q.h | 132 n = sys_dlist_peek_next_no_check(pq, &_current->base.qnode_dlist); in z_priq_simple_yield() 134 sys_dlist_dequeue(&_current->base.qnode_dlist); in z_priq_simple_yield() 146 if (z_sched_prio_cmp(_current, t) > 0) { in z_priq_simple_yield() 148 &_current->base.qnode_dlist); in z_priq_simple_yield() 154 sys_dlist_append(pq, &_current->base.qnode_dlist); in z_priq_simple_yield() 233 z_priq_rb_remove(pq, _current); in z_priq_rb_yield() 234 z_priq_rb_add(pq, _current); in z_priq_rb_yield() 324 struct prio_info pos = get_prio_info(_current->base.prio); in z_priq_mq_yield() 326 sys_dlist_dequeue(&_current->base.qnode_dlist); in z_priq_mq_yield() 328 &_current->base.qnode_dlist); in z_priq_mq_yield()
|
/Zephyr-latest/arch/arc/core/ |
D | thread.c | 213 *old_thread = _current; in z_arch_get_next_switch_handle() 230 setup_stack_vars(_current); in arch_user_mode_enter() 234 configure_mpu_thread(_current); in arch_user_mode_enter() 237 (uint32_t)_current->stack_info.start, in arch_user_mode_enter() 238 (_current->stack_info.size - in arch_user_mode_enter() 239 _current->stack_info.delta), _current); in arch_user_mode_enter() 339 __ASSERT(!arch_is_in_isr() && (_current->base.cpu_mask == BIT(id)), ""); in arc_vpx_lock() 358 __ASSERT(!arch_is_in_isr() && (_current->base.cpu_mask == BIT(id)), ""); in arc_vpx_unlock()
|