Lines Matching refs:flags

50 			     uint32_t flags)  in flags_set()  argument
52 *flagp = flags; in flags_set()
73 flag_set(&work->flags, K_WORK_FLUSHING_BIT); in init_flusher()
110 flag_clear(&work->flags, K_WORK_FLUSHING_BIT); in finalize_flush_locked()
133 flag_clear(&work->flags, K_WORK_CANCELING_BIT); in finalize_cancel_locked()
163 return flags_get(&work->flags) & K_WORK_MASK; in work_busy_get_locked()
193 if ((flags_get(&work->flags) & K_WORK_QUEUED) != 0U) { in queue_flusher_locked()
211 if (flag_test_and_clear(&work->flags, K_WORK_QUEUED_BIT)) { in queue_remove_locked()
266 bool draining = flag_test(&queue->flags, K_WORK_QUEUE_DRAIN_BIT); in queue_submit_locked()
267 bool plugged = flag_test(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT); in queue_submit_locked()
276 if (!flag_test(&queue->flags, K_WORK_QUEUE_STARTED_BIT)) { in queue_submit_locked()
322 if (flag_test(&work->flags, K_WORK_CANCELING_BIT)) { in submit_to_queue_locked()
325 } else if (!flag_test(&work->flags, K_WORK_QUEUED_BIT)) { in submit_to_queue_locked()
339 if (flag_test(&work->flags, K_WORK_RUNNING_BIT)) { in submit_to_queue_locked()
350 flag_set(&work->flags, K_WORK_QUEUED_BIT); in submit_to_queue_locked()
440 bool need_flush = (flags_get(&work->flags) in work_flush_locked()
459 __ASSERT_NO_MSG(!flag_test(&work->flags, K_WORK_DELAYABLE_BIT)); in k_work_flush()
501 if (!flag_test(&work->flags, K_WORK_CANCELING_BIT)) { in cancel_async_locked()
512 flag_set(&work->flags, K_WORK_CANCELING_BIT); in cancel_async_locked()
537 bool ret = flag_test(&work->flags, K_WORK_CANCELING_BIT); in cancel_sync_locked()
553 __ASSERT_NO_MSG(!flag_test(&work->flags, K_WORK_DELAYABLE_BIT)); in k_work_cancel()
572 __ASSERT_NO_MSG(!flag_test(&work->flags, K_WORK_DELAYABLE_BIT)); in k_work_cancel_sync()
626 flag_set(&queue->flags, K_WORK_QUEUE_BUSY_BIT); in work_queue_main()
628 flag_set(&work->flags, K_WORK_RUNNING_BIT); in work_queue_main()
629 flag_clear(&work->flags, K_WORK_QUEUED_BIT); in work_queue_main()
644 } else if (flag_test_and_clear(&queue->flags, in work_queue_main()
656 } else if (flag_test(&queue->flags, K_WORK_QUEUE_STOP_BIT)) { in work_queue_main()
659 flags_set(&queue->flags, 0); in work_queue_main()
693 flag_clear(&work->flags, K_WORK_RUNNING_BIT); in work_queue_main()
694 if (flag_test(&work->flags, K_WORK_FLUSHING_BIT)) { in work_queue_main()
697 if (flag_test(&work->flags, K_WORK_CANCELING_BIT)) { in work_queue_main()
701 flag_clear(&queue->flags, K_WORK_QUEUE_BUSY_BIT); in work_queue_main()
702 yield = !flag_test(&queue->flags, K_WORK_QUEUE_NO_YIELD_BIT); in work_queue_main()
719 .flags = 0, in k_work_queue_init()
733 __ASSERT_NO_MSG(!flag_test(&queue->flags, K_WORK_QUEUE_STARTED_BIT)); in k_work_queue_start()
734 uint32_t flags = K_WORK_QUEUE_STARTED; in k_work_queue_start() local
743 flags |= K_WORK_QUEUE_NO_YIELD; in k_work_queue_start()
750 flags_set(&queue->flags, flags); in k_work_queue_start()
780 if (((flags_get(&queue->flags) in k_work_queue_drain()
784 flag_set(&queue->flags, K_WORK_QUEUE_DRAIN_BIT); in k_work_queue_drain()
786 flag_set(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT); in k_work_queue_drain()
810 if (flag_test_and_clear(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT)) { in k_work_queue_unplug()
828 if (!flag_test(&queue->flags, K_WORK_QUEUE_STARTED_BIT)) { in k_work_queue_stop()
834 if (!flag_test(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT)) { in k_work_queue_stop()
840 flag_set(&queue->flags, K_WORK_QUEUE_STOP_BIT); in k_work_queue_stop()
846 flag_clear(&queue->flags, K_WORK_QUEUE_STOP_BIT); in k_work_queue_stop()
879 if (flag_test_and_clear(&wp->flags, K_WORK_DELAYED_BIT)) { in work_timeout()
896 .flags = K_WORK_DELAYABLE, in k_work_init_delayable()
906 return flags_get(&dwork->work.flags) & K_WORK_MASK; in work_delayable_busy_get_locked()
954 flag_set(&work->flags, K_WORK_DELAYED_BIT); in schedule_for_queue_locked()
985 if (flag_test_and_clear(&work->flags, K_WORK_DELAYED_BIT)) { in unschedule_locked()