Searched refs:fpu_owner (Results 1 – 7 of 7) sorted by relevance
71 struct k_thread *owner = atomic_ptr_get(&_current_cpu->arch.fpu_owner); in arch_flush_local_fpu()85 atomic_ptr_clear(&_current_cpu->arch.fpu_owner); in arch_flush_local_fpu()105 if (atomic_ptr_get(&_kernel.cpus[i].arch.fpu_owner) != thread) { in flush_owned_fpu()130 while (atomic_ptr_get(&_kernel.cpus[i].arch.fpu_owner) == thread) { in flush_owned_fpu()238 struct k_thread *owner = atomic_ptr_get(&_current_cpu->arch.fpu_owner); in z_arm64_fpu_trap()243 atomic_ptr_clear(&_current_cpu->arch.fpu_owner); in z_arm64_fpu_trap()267 atomic_ptr_set(&_current_cpu->arch.fpu_owner, arch_current_thread()); in z_arm64_fpu_trap()290 if (atomic_ptr_get(&_current_cpu->arch.fpu_owner) == arch_current_thread()) { in fpu_access_update()336 if (thread == atomic_ptr_get(&_current_cpu->arch.fpu_owner)) { in arch_float_disable()
291 arch_float_disable(_current_cpu->arch.fpu_owner); in arch_spin_relax()
85 atomic_ptr_set(&_current_cpu->arch.fpu_owner, arch_current_thread()); in z_riscv_fpu_load()108 struct k_thread *owner = atomic_ptr_get(&_current_cpu->arch.fpu_owner); in arch_flush_local_fpu()127 atomic_ptr_clear(&_current_cpu->arch.fpu_owner); in arch_flush_local_fpu()145 owner = atomic_ptr_get(&_kernel.cpus[i].arch.fpu_owner); in flush_owned_fpu()176 owner = atomic_ptr_get(&_kernel.cpus[i].arch.fpu_owner); in flush_owned_fpu()261 if (_current_cpu->arch.fpu_owner == arch_current_thread()) { in fpu_access_allowed()330 if (thread == _current_cpu->arch.fpu_owner) { in arch_float_disable()
89 arch_float_disable(_current_cpu->arch.fpu_owner); in arch_spin_relax()
13 atomic_ptr_val_t fpu_owner; member
22 atomic_ptr_val_t fpu_owner; member
238 k_float_disable(_current_cpu->arch.fpu_owner); in cpu_hold()