/Zephyr-latest/kernel/ |
D | nothread.c | 14 return arch_is_in_isr(); in k_is_in_isr() 25 __ASSERT(!arch_is_in_isr(), ""); in z_impl_k_sleep()
|
D | spinlock_validate.c | 27 if (arch_is_in_isr() && arch_current_thread()->base.thread_state & _THREAD_DUMMY) { in z_spin_unlock_valid()
|
D | sched.c | 462 if (arch_is_in_isr()) { in z_thread_halt() 470 if ((thread == arch_current_thread()) && !arch_is_in_isr()) { in z_thread_halt() 491 if (thread == arch_current_thread() && !arch_is_in_isr() && !IS_ENABLED(CONFIG_SMP)) { in z_impl_k_thread_suspend() 748 return arch_irq_unlocked(key) && !arch_is_in_isr(); in resched() 802 __ASSERT(!arch_is_in_isr(), ""); in k_sched_unlock() 1079 __ASSERT(!arch_is_in_isr(), ""); in z_impl_k_yield() 1106 __ASSERT(!arch_is_in_isr(), ""); in z_tick_sleep() 1150 __ASSERT(!arch_is_in_isr(), ""); in z_impl_k_sleep() 1215 if (arch_is_in_isr()) { in z_impl_k_wakeup() 1294 if (thread == arch_current_thread() && arch_is_in_isr()) { in halt_thread() [all …]
|
D | kheap.c | 75 __ASSERT(!arch_is_in_isr() || K_TIMEOUT_EQ(timeout, K_NO_WAIT), ""); in k_heap_aligned_alloc() 147 __ASSERT(!arch_is_in_isr() || K_TIMEOUT_EQ(timeout, K_NO_WAIT), ""); in k_heap_realloc()
|
D | mutex.c | 111 __ASSERT(!arch_is_in_isr(), "mutexes cannot be used inside ISRs"); in z_impl_k_mutex_lock() 215 __ASSERT(!arch_is_in_isr(), "mutexes cannot be used inside ISRs"); in z_impl_k_mutex_unlock()
|
D | smp.c | 246 bool pinned = arch_is_in_isr() || !arch_irq_unlocked(k); in z_smp_cpu_mobile()
|
D | sem.c | 136 __ASSERT(((arch_is_in_isr() == false) || in z_impl_k_sem_take()
|
D | msg_q.c | 126 __ASSERT(!arch_is_in_isr() || K_TIMEOUT_EQ(timeout, K_NO_WAIT), ""); in z_impl_k_msgq_put() 218 __ASSERT(!arch_is_in_isr() || K_TIMEOUT_EQ(timeout, K_NO_WAIT), ""); in z_impl_k_msgq_get()
|
D | thread.c | 78 return arch_is_in_isr(); in k_is_in_isr() 113 return !arch_is_in_isr() && thread_is_preemptible(arch_current_thread()); in z_impl_k_is_preempt_thread() 657 __ASSERT(!arch_is_in_isr(), "Threads may not be created in ISRs"); in z_impl_k_thread_create()
|
/Zephyr-latest/arch/arm64/include/ |
D | exception.h | 29 static ALWAYS_INLINE bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/arm/core/cortex_m/ |
D | thread_abort.c | 31 if (arch_is_in_isr()) { in z_impl_k_thread_abort()
|
/Zephyr-latest/arch/posix/include/ |
D | kernel_arch_func.h | 40 static inline bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/mips/include/ |
D | kernel_arch_func.h | 45 static inline bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/arm/include/cortex_a_r/ |
D | exception.h | 36 static ALWAYS_INLINE bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/sparc/include/ |
D | kernel_arch_func.h | 53 static inline bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/nios2/include/ |
D | kernel_arch_func.h | 49 static inline bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/arc/include/ |
D | kernel_arch_func.h | 59 static inline bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/x86/include/ |
D | kernel_arch_func.h | 19 static inline bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/x86/core/ |
D | efi.c | 131 if (arch_is_in_isr()) { in efi_console_putchar()
|
/Zephyr-latest/arch/xtensa/include/ |
D | kernel_arch_func.h | 141 static inline bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/kernel/include/ |
D | kernel_arch_interface.h | 270 static inline bool arch_is_in_isr(void);
|
D | ksched.h | 145 __ASSERT(!arch_is_in_isr(), ""); in z_sched_lock()
|
/Zephyr-latest/arch/riscv/include/ |
D | kernel_arch_func.h | 84 static inline bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/arm/include/cortex_m/ |
D | exception.h | 94 static ALWAYS_INLINE bool arch_is_in_isr(void) in arch_is_in_isr() function
|
/Zephyr-latest/arch/arc/core/ |
D | thread.c | 339 __ASSERT(!arch_is_in_isr() && (arch_current_thread()->base.cpu_mask == BIT(id)), ""); in arc_vpx_lock() 358 __ASSERT(!arch_is_in_isr() && (arch_current_thread()->base.cpu_mask == BIT(id)), ""); in arc_vpx_unlock()
|