Home
last modified time | relevance | path

Searched refs:_current (Results 1 – 25 of 100) sorted by relevance

1234

/Zephyr-latest/tests/arch/arm/arm_thread_swap/src/
Darm_syscalls.c40 zassert_true((_current->arch.mode & CONTROL_nPRIV_Msk) == 0, in z_impl_test_arm_user_syscall()
46 ((__get_PSP() >= _current->arch.priv_stack_start) && in z_impl_test_arm_user_syscall()
47 (__get_PSP() < (_current->arch.priv_stack_start + CONFIG_PRIVILEGED_STACK_SIZE))), in z_impl_test_arm_user_syscall()
51 zassert_true(__get_PSPLIM() == _current->arch.priv_stack_start, in z_impl_test_arm_user_syscall()
77 zassert_true((_current->arch.mode & CONTROL_nPRIV_Msk) != 0, in arm_isr_handler()
82 zassert_true(((__get_PSP() >= _current->stack_info.start) && in arm_isr_handler()
83 (__get_PSP() < (_current->stack_info.start + _current->stack_info.size))), in arm_isr_handler()
93 z_move_thread_to_end_of_prio_q(_current); in arm_isr_handler()
160 zassert_true((_current->arch.mode & CONTROL_nPRIV_Msk) == 0, in ZTEST()
165 zassert_true(((__get_PSP() >= _current->stack_info.start) && in ZTEST()
[all …]
Darm_thread_arch.c225 zassert_true((_current->arch.mode_exc_return & EXC_RETURN_FTYPE) != 0, in alt_thread_entry()
231 zassert_true((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0, in alt_thread_entry()
233 zassert_true((_current->base.user_options & K_FP_REGS) != 0, in alt_thread_entry()
273 z_move_thread_to_end_of_prio_q(_current); in alt_thread_entry()
392 k_thread_priority_set(_current, K_PRIO_COOP(PRIORITY)); in ZTEST()
399 p_ztest_thread = _current; in ZTEST()
404 zassert_true(_current->arch.basepri == 0, in ZTEST()
422 zassert_true((_current->arch.mode & CONTROL_nPRIV_Msk) == 0, in ZTEST()
424 _current->arch.mode); in ZTEST()
429 zassert_true((_current->arch.mode_exc_return & EXC_RETURN_FTYPE) != 0, in ZTEST()
[all …]
/Zephyr-latest/arch/riscv/core/
Dthread.c144 _current->arch.priv_stack_start = in arch_user_mode_enter()
145 (unsigned long)z_priv_stack_find(_current->stack_obj); in arch_user_mode_enter()
147 _current->stack_info.start -= K_THREAD_STACK_RESERVED; in arch_user_mode_enter()
148 _current->stack_info.size += K_THREAD_STACK_RESERVED; in arch_user_mode_enter()
150 _current->arch.priv_stack_start = (unsigned long)_current->stack_obj; in arch_user_mode_enter()
152 top_of_priv_stack = Z_STACK_PTR_ALIGN(_current->arch.priv_stack_start + in arch_user_mode_enter()
158 (void)memset((void *)_current->arch.priv_stack_start, 0xaa, in arch_user_mode_enter()
163 _current->stack_info.start + in arch_user_mode_enter()
164 _current->stack_info.size - in arch_user_mode_enter()
165 _current->stack_info.delta); in arch_user_mode_enter()
[all …]
Dfpu.c39 buf[8] = '0' + _current->arch.exception_depth; in DBG()
40 strcat(buf, _current->name); in DBG()
85 atomic_ptr_set(&_current_cpu->arch.fpu_owner, _current); in z_riscv_fpu_load()
89 z_riscv_fpu_restore(&_current->arch.saved_fp_context); in z_riscv_fpu_load()
90 DBG("restore", _current); in z_riscv_fpu_load()
171 if (thread == _current) { in flush_owned_fpu()
216 if (_current->arch.exception_depth > 0) { in z_riscv_fpu_trap()
236 flush_owned_fpu(_current); in z_riscv_fpu_trap()
259 if (_current->arch.exception_depth == exc_update_level) { in fpu_access_allowed()
261 if (_current_cpu->arch.fpu_owner == _current) { in fpu_access_allowed()
[all …]
Dfatal.c161 if (_current->arch.priv_stack_start != 0 && in bad_stack_pointer()
162 sp >= _current->arch.priv_stack_start && in bad_stack_pointer()
163 sp < _current->arch.priv_stack_start + Z_RISCV_STACK_GUARD_SIZE) { in bad_stack_pointer()
167 if (z_stack_is_user_capable(_current->stack_obj) && in bad_stack_pointer()
168 sp >= _current->stack_info.start - K_THREAD_STACK_RESERVED && in bad_stack_pointer()
169 sp < _current->stack_info.start - K_THREAD_STACK_RESERVED in bad_stack_pointer()
176 if (sp >= _current->stack_info.start - K_KERNEL_STACK_RESERVED && in bad_stack_pointer()
177 sp < _current->stack_info.start - K_KERNEL_STACK_RESERVED in bad_stack_pointer()
194 (esf->sp < _current->stack_info.start || in bad_stack_pointer()
195 esf->sp > _current->stack_info.start + in bad_stack_pointer()
[all …]
/Zephyr-latest/arch/arm/core/cortex_a_r/
Dthread.c201 _current->arch.priv_stack_start = in arch_user_mode_enter()
202 (uint32_t)z_priv_stack_find(_current->stack_obj); in arch_user_mode_enter()
211 if ((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0) { in arch_user_mode_enter()
212 _current->stack_info.start -= FP_GUARD_EXTRA_SIZE; in arch_user_mode_enter()
213 _current->stack_info.size += FP_GUARD_EXTRA_SIZE; in arch_user_mode_enter()
216 _current->stack_info.start -= MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter()
217 _current->stack_info.size += MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter()
225 _current->arch.priv_stack_start += in arch_user_mode_enter()
226 ((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0) ? in arch_user_mode_enter()
229 _current->arch.priv_stack_start += MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter()
[all …]
/Zephyr-latest/kernel/
Dsched.c106 return !IS_ENABLED(CONFIG_SMP) || (thread != _current); in should_queue_thread()
116 if (thread == _current) { in queue_thread()
170 if (is_halting(_current)) { in next_up()
171 halt_thread(_current, is_aborting(_current) ? in next_up()
219 bool queued = z_is_thread_queued(_current); in next_up()
220 bool active = !z_is_thread_prevented_from_running(_current); in next_up()
227 int32_t cmp = z_sched_prio_cmp(_current, thread); in next_up()
231 thread = _current; in next_up()
235 thread = _current; in next_up()
240 if ((thread != _current) && active && in next_up()
[all …]
Dspinlock_validate.c29 if (arch_is_in_isr() && _current->base.thread_state & _THREAD_DUMMY) { in z_spin_unlock_valid()
33 if (tcpu != (_current_cpu->id | (uintptr_t)_current)) { in z_spin_unlock_valid()
42 l->thread_cpu = _current_cpu->id | (uintptr_t)_current; in z_spin_lock_set_owner()
Dthread.c85 _current->custom_data = value; in z_impl_k_thread_custom_data_set()
98 return _current->custom_data; in z_impl_k_thread_custom_data_get()
113 return !arch_is_in_isr() && thread_is_preemptible(_current); in z_impl_k_is_preempt_thread()
142 thread = _current; in z_impl_k_thread_name_set()
337 if ((_current->base.thread_state & _THREAD_DUMMY) != 0) { in z_check_stack_sentinel()
341 stack = (uint32_t *)_current->stack_info.start; in z_check_stack_sentinel()
619 if (!_current) { in z_setup_new_thread()
628 k_thread_perms_inherit(_current, new_thread); in z_setup_new_thread()
634 new_thread->resource_pool = _current->resource_pool; in z_setup_new_thread()
729 _current->base.prio))); in z_vrfy_k_thread_create()
[all …]
Derrno.c39 return &_current->userspace_local_data->errno_var; in z_impl_z_errno()
51 return &_current->errno_var; in z_impl_z_errno()
Dmutex.c117 if (likely((mutex->lock_count == 0U) || (mutex->owner == _current))) { in z_impl_k_mutex_lock()
120 _current->base.prio : in z_impl_k_mutex_lock()
124 mutex->owner = _current; in z_impl_k_mutex_lock()
127 _current, mutex, mutex->lock_count, in z_impl_k_mutex_lock()
147 new_prio = new_prio_for_inheritance(_current->base.prio, in z_impl_k_mutex_lock()
160 LOG_DBG("%p got mutex %p (y/n): %c", _current, mutex, in z_impl_k_mutex_lock()
170 LOG_DBG("%p timeout on mutex %p", _current, mutex); in z_impl_k_mutex_lock()
227 CHECKIF(mutex->owner != _current) { in z_impl_k_mutex_unlock()
Dsmp.c61 if (!_current->base.global_lock_count) { in z_smp_global_lock()
67 _current->base.global_lock_count++; in z_smp_global_lock()
74 if (_current->base.global_lock_count != 0U) { in z_smp_global_unlock()
75 _current->base.global_lock_count--; in z_smp_global_unlock()
77 if (!_current->base.global_lock_count) { in z_smp_global_unlock()
/Zephyr-latest/arch/posix/core/
Dswap.c37 _current->callee_saved.key = key; in arch_swap()
38 _current->callee_saved.retval = -EAGAIN; in arch_swap()
50 _current->callee_saved.thread_status; in arch_swap()
69 irq_unlock(_current->callee_saved.key); in arch_swap()
71 return _current->callee_saved.retval; in arch_swap()
/Zephyr-latest/arch/arm/core/cortex_m/
Dthread.c220 _current->arch.priv_stack_start = (uint32_t)z_priv_stack_find(_current->stack_obj); in arch_user_mode_enter()
229 if ((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0) { in arch_user_mode_enter()
230 _current->stack_info.start -= FP_GUARD_EXTRA_SIZE; in arch_user_mode_enter()
231 _current->stack_info.size += FP_GUARD_EXTRA_SIZE; in arch_user_mode_enter()
234 _current->stack_info.start -= MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter()
235 _current->stack_info.size += MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter()
243 _current->arch.priv_stack_start += in arch_user_mode_enter()
244 ((_current->arch.mode & Z_ARM_MODE_MPU_GUARD_FLOAT_Msk) != 0) in arch_user_mode_enter()
248 _current->arch.priv_stack_start += MPU_GUARD_ALIGN_AND_SIZE; in arch_user_mode_enter()
252 z_arm_userspace_enter(user_entry, p1, p2, p3, (uint32_t)_current->stack_info.start, in arch_user_mode_enter()
[all …]
/Zephyr-latest/arch/x86/core/
Duserspace.c135 stack_end = Z_STACK_PTR_ALIGN(_current->stack_info.start + in arch_user_mode_enter()
136 _current->stack_info.size - in arch_user_mode_enter()
137 _current->stack_info.delta); in arch_user_mode_enter()
166 stack_start = POINTER_TO_UINT(_current->stack_obj); in arch_user_mode_enter()
167 stack_size = K_THREAD_STACK_LEN(_current->stack_info.size); in arch_user_mode_enter()
185 _current->stack_info.start); in arch_user_mode_enter()
/Zephyr-latest/tests/kernel/threads/thread_apis/src/
Dtest_essential_thread.c30 z_thread_essential_set(_current); in thread_entry()
32 if (z_is_thread_essential(_current)) { in thread_entry()
38 z_thread_essential_clear(_current); in thread_entry()
39 zassert_false(z_is_thread_essential(_current), in thread_entry()
78 if (z_is_thread_essential(_current)) { in abort_thread_entry()
/Zephyr-latest/arch/arm64/core/
Dthread.c162 z_arm64_thread_mem_domains_init(_current); in arch_user_mode_enter()
165 stack_el0 = Z_STACK_PTR_ALIGN(_current->stack_info.start + in arch_user_mode_enter()
166 _current->stack_info.size - in arch_user_mode_enter()
167 _current->stack_info.delta); in arch_user_mode_enter()
170 stack_el1 = (uintptr_t)(_current->stack_obj + ARCH_THREAD_STACK_RESERVED); in arch_user_mode_enter()
Dfpu.c39 strcat(buf, _current->name); in DBG()
128 if (thread == _current) { in flush_owned_fpu()
263 flush_owned_fpu(_current); in z_arm64_fpu_trap()
267 atomic_ptr_set(&_current_cpu->arch.fpu_owner, _current); in z_arm64_fpu_trap()
270 z_arm64_fpu_restore(&_current->arch.saved_fp_context); in z_arm64_fpu_trap()
271 DBG("restore", _current); in z_arm64_fpu_trap()
290 if (atomic_ptr_get(&_current_cpu->arch.fpu_owner) == _current) { in fpu_access_update()
/Zephyr-latest/tests/kernel/usage/thread_runtime_stats/src/
Dtest_thread_runtime_stats.c75 priority = k_thread_priority_get(_current); in ZTEST()
205 priority = k_thread_priority_get(_current); in ZTEST()
218 k_thread_runtime_stats_get(_current, &stats1); in ZTEST()
234 k_thread_runtime_stats_get(_current, &stats2); in ZTEST()
289 k_thread_runtime_stats_get(_current, &thread_stats1); in ZTEST()
294 k_thread_runtime_stats_get(_current, &thread_stats2); in ZTEST()
306 k_thread_runtime_stats_get(_current, &thread_stats3); in ZTEST()
407 priority = k_thread_priority_get(_current); in ZTEST()
417 status = k_thread_runtime_stats_get(_current, NULL); in ZTEST()
431 main_thread = _current; in ZTEST()
[all …]
/Zephyr-latest/subsys/profiling/perf/backends/
Dperf_x86_64.c43 buf[idx++] = (uintptr_t)_current->callee_saved.rip; in arch_perf_current_stack_trace()
44 void **fp = (void **)_current->callee_saved.rbp; in arch_perf_current_stack_trace()
56 while (valid_stack((uintptr_t)fp, _current)) { in arch_perf_current_stack_trace()
/Zephyr-latest/arch/arm/include/cortex_a_r/
Dkernel_arch_func.h43 _current->arch.basepri = key; in arch_swap()
44 _current->arch.swap_return_value = -EAGAIN; in arch_swap()
52 return _current->arch.swap_return_value; in arch_swap()
/Zephyr-latest/arch/arm/include/cortex_m/
Dkernel_arch_func.h85 _current->arch.basepri = key; in arch_swap()
86 _current->arch.swap_return_value = -EAGAIN; in arch_swap()
97 return _current->arch.swap_return_value; in arch_swap()
/Zephyr-latest/arch/x86/core/ia32/
Dfloat.c210 if (thread == _current) { in z_float_enable()
225 if ((_current->base.user_options & _FP_USER_MASK) == 0) { in z_float_enable()
281 if (thread == _current) { in z_float_disable()
317 k_float_enable(_current, _FP_USER_MASK); in _FpNotAvailableExcHandler()
/Zephyr-latest/kernel/include/
Dpriority_q.h132 n = sys_dlist_peek_next_no_check(pq, &_current->base.qnode_dlist); in z_priq_simple_yield()
134 sys_dlist_dequeue(&_current->base.qnode_dlist); in z_priq_simple_yield()
146 if (z_sched_prio_cmp(_current, t) > 0) { in z_priq_simple_yield()
148 &_current->base.qnode_dlist); in z_priq_simple_yield()
154 sys_dlist_append(pq, &_current->base.qnode_dlist); in z_priq_simple_yield()
233 z_priq_rb_remove(pq, _current); in z_priq_rb_yield()
234 z_priq_rb_add(pq, _current); in z_priq_rb_yield()
324 struct prio_info pos = get_prio_info(_current->base.prio); in z_priq_mq_yield()
326 sys_dlist_dequeue(&_current->base.qnode_dlist); in z_priq_mq_yield()
328 &_current->base.qnode_dlist); in z_priq_mq_yield()
/Zephyr-latest/arch/arc/core/
Dthread.c213 *old_thread = _current; in z_arch_get_next_switch_handle()
230 setup_stack_vars(_current); in arch_user_mode_enter()
234 configure_mpu_thread(_current); in arch_user_mode_enter()
237 (uint32_t)_current->stack_info.start, in arch_user_mode_enter()
238 (_current->stack_info.size - in arch_user_mode_enter()
239 _current->stack_info.delta), _current); in arch_user_mode_enter()
339 __ASSERT(!arch_is_in_isr() && (_current->base.cpu_mask == BIT(id)), ""); in arc_vpx_lock()
358 __ASSERT(!arch_is_in_isr() && (_current->base.cpu_mask == BIT(id)), ""); in arc_vpx_unlock()

1234