| /Linux-v5.4/include/linux/ | 
| D | blkdev.h | 433 	unsigned long		queue_flags;  member630 #define blk_queue_stopped(q)	test_bit(QUEUE_FLAG_STOPPED, &(q)->queue_flags)
 631 #define blk_queue_dying(q)	test_bit(QUEUE_FLAG_DYING, &(q)->queue_flags)
 632 #define blk_queue_dead(q)	test_bit(QUEUE_FLAG_DEAD, &(q)->queue_flags)
 633 #define blk_queue_init_done(q)	test_bit(QUEUE_FLAG_INIT_DONE, &(q)->queue_flags)
 634 #define blk_queue_nomerges(q)	test_bit(QUEUE_FLAG_NOMERGES, &(q)->queue_flags)
 636 	test_bit(QUEUE_FLAG_NOXMERGES, &(q)->queue_flags)
 637 #define blk_queue_nonrot(q)	test_bit(QUEUE_FLAG_NONROT, &(q)->queue_flags)
 638 #define blk_queue_io_stat(q)	test_bit(QUEUE_FLAG_IO_STAT, &(q)->queue_flags)
 639 #define blk_queue_add_random(q)	test_bit(QUEUE_FLAG_ADD_RANDOM, &(q)->queue_flags)
 [all …]
 
 | 
| /Linux-v5.4/block/ | 
| D | blk-timeout.c | 25 	if (!test_bit(QUEUE_FLAG_FAIL_IO, &q->queue_flags))  in blk_should_fake_timeout()45 	int set = test_bit(QUEUE_FLAG_FAIL_IO, &disk->queue->queue_flags);  in part_timeout_show()
 
 | 
| D | blk-softirq.c | 113 	if (test_bit(QUEUE_FLAG_SAME_COMP, &q->queue_flags) && ccpu != -1) {  in __blk_complete_request()114 		if (!test_bit(QUEUE_FLAG_SAME_FORCE, &q->queue_flags))  in __blk_complete_request()
 
 | 
| D | blk-sysfs.c | 258 	bit = test_bit(QUEUE_FLAG_##flag, &q->queue_flags);		\328 	bool set = test_bit(QUEUE_FLAG_SAME_COMP, &q->queue_flags);  in queue_rq_affinity_show()
 329 	bool force = test_bit(QUEUE_FLAG_SAME_FORCE, &q->queue_flags);  in queue_rq_affinity_show()
 395 	return queue_var_show(test_bit(QUEUE_FLAG_POLL, &q->queue_flags), page);  in queue_poll_show()
 494 	if (test_bit(QUEUE_FLAG_WC, &q->queue_flags))  in queue_wc_show()
 524 	return sprintf(page, "%u\n", test_bit(QUEUE_FLAG_FUA, &q->queue_flags));  in queue_fua_show()
 888 	if (test_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags))  in __blk_release_queue()
 
 | 
| D | blk-core.c | 80 	set_bit(flag, &q->queue_flags);  in blk_queue_flag_set()91 	clear_bit(flag, &q->queue_flags);  in blk_queue_flag_clear()
 105 	return test_and_set_bit(flag, &q->queue_flags);  in blk_queue_flag_test_and_set()
 914 	    !test_bit(QUEUE_FLAG_WC, &q->queue_flags)) {  in generic_make_request_checks()
 922 	if (!test_bit(QUEUE_FLAG_POLL, &q->queue_flags))  in generic_make_request_checks()
 
 | 
| D | blk-mq.c | 609 	    !test_bit(QUEUE_FLAG_SAME_COMP, &q->queue_flags)) {  in __blk_mq_complete_request()615 	if (!test_bit(QUEUE_FLAG_SAME_FORCE, &q->queue_flags))  in __blk_mq_complete_request()
 684 	if (test_bit(QUEUE_FLAG_STATS, &q->queue_flags)) {  in blk_mq_start_request()
 2897 	q->queue_flags |= QUEUE_FLAG_MQ_DEFAULT;  in blk_mq_init_allocated_queue()
 3331 	if (test_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags) ||  in blk_poll_stats_enable()
 3344 	if (!test_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags) ||  in blk_mq_poll_stats_start()
 3492 	    !test_bit(QUEUE_FLAG_POLL, &q->queue_flags))  in blk_poll()
 
 | 
| D | blk-flush.c | 362 	unsigned long fflags = q->queue_flags;	/* may change, cache */  in blk_insert_flush()
 | 
| D | blk-settings.c | 832 	wbt_set_write_cache(q, test_bit(QUEUE_FLAG_WC, &q->queue_flags));  in blk_queue_write_cache()
 | 
| D | blk-wbt.c | 857 	wbt_set_write_cache(q, test_bit(QUEUE_FLAG_WC, &q->queue_flags));  in wbt_init()
 | 
| D | elevator.c | 676 	WARN_ON_ONCE(test_bit(QUEUE_FLAG_REGISTERED, &q->queue_flags));  in elevator_init_mq()
 | 
| D | blk-mq-debugfs.c | 135 	blk_flags_show(m, q->queue_flags, blk_queue_flag_name,  in queue_state_show()
 | 
| D | blk-iocost.c | 623 	if (test_bit(QUEUE_FLAG_REGISTERED, &q->queue_flags))  in q_name()
 | 
| /Linux-v5.4/drivers/target/ | 
| D | target_core_iblock.c | 701 		if (test_bit(QUEUE_FLAG_FUA, &q->queue_flags)) {  in iblock_execute_rw()704 			else if (!test_bit(QUEUE_FLAG_WC, &q->queue_flags))  in iblock_execute_rw()
 857 	return test_bit(QUEUE_FLAG_WC, &q->queue_flags);  in iblock_get_write_cache()
 
 | 
| /Linux-v5.4/kernel/sched/ | 
| D | core.c | 4761 	int queue_flags = DEQUEUE_SAVE | DEQUEUE_MOVE | DEQUEUE_NOCLOCK;  in __sched_setscheduler()  local4965 			queue_flags &= ~DEQUEUE_MOVE;  in __sched_setscheduler()
 4971 		dequeue_task(rq, p, queue_flags);  in __sched_setscheduler()
 4986 			queue_flags |= ENQUEUE_HEAD;  in __sched_setscheduler()
 4988 		enqueue_task(rq, p, queue_flags);  in __sched_setscheduler()
 7036 	int queued, running, queue_flags =  in sched_move_task()  local
 7048 		dequeue_task(rq, tsk, queue_flags);  in sched_move_task()
 7055 		enqueue_task(rq, tsk, queue_flags);  in sched_move_task()
 
 | 
| /Linux-v5.4/drivers/block/xen-blkback/ | 
| D | xenbus.c | 482 	if (q && test_bit(QUEUE_FLAG_WC, &q->queue_flags))  in xen_vbd_create()
 | 
| /Linux-v5.4/drivers/md/ | 
| D | raid5-ppl.c | 1327 	if (test_bit(QUEUE_FLAG_WC, &q->queue_flags))  in ppl_init_child_log()
 | 
| D | dm-table.c | 1637 	return q && (q->queue_flags & flush);  in device_flush_capable()
 | 
| D | raid5-cache.c | 3097 	log->need_cache_flush = test_bit(QUEUE_FLAG_WC, &q->queue_flags) != 0;  in r5l_init_log()
 | 
| /Linux-v5.4/drivers/nvme/host/ | 
| D | core.c | 795 	WARN_ON_ONCE(!test_bit(QUEUE_FLAG_POLL, &q->queue_flags));  in nvme_execute_rq_polled()
 | 
| /Linux-v5.4/fs/btrfs/ | 
| D | disk-io.c | 3625 	if (!test_bit(QUEUE_FLAG_WC, &q->queue_flags))  in write_dev_flush()
 |