/Linux-v4.19/block/ |
D | blk.h | 60 if (test_bit(QUEUE_FLAG_INIT_DONE, &q->queue_flags) && in queue_flag_set_unlocked() 63 __set_bit(flag, &q->queue_flags); in queue_flag_set_unlocked() 69 if (test_bit(QUEUE_FLAG_INIT_DONE, &q->queue_flags) && in queue_flag_clear_unlocked() 72 __clear_bit(flag, &q->queue_flags); in queue_flag_clear_unlocked() 80 if (test_bit(flag, &q->queue_flags)) { in queue_flag_test_and_clear() 81 __clear_bit(flag, &q->queue_flags); in queue_flag_test_and_clear() 93 if (!test_bit(flag, &q->queue_flags)) { in queue_flag_test_and_set() 94 __set_bit(flag, &q->queue_flags); in queue_flag_test_and_set() 104 __set_bit(flag, &q->queue_flags); in queue_flag_set() 110 __clear_bit(flag, &q->queue_flags); in queue_flag_clear()
|
D | blk-sysfs.c | 265 bit = test_bit(QUEUE_FLAG_##flag, &q->queue_flags); \ 332 bool set = test_bit(QUEUE_FLAG_SAME_COMP, &q->queue_flags); in queue_rq_affinity_show() 333 bool force = test_bit(QUEUE_FLAG_SAME_FORCE, &q->queue_flags); in queue_rq_affinity_show() 399 return queue_var_show(test_bit(QUEUE_FLAG_POLL, &q->queue_flags), page); in queue_poll_show() 481 if (test_bit(QUEUE_FLAG_WC, &q->queue_flags)) in queue_wc_show() 511 return sprintf(page, "%u\n", test_bit(QUEUE_FLAG_FUA, &q->queue_flags)); in queue_fua_show() 818 if (test_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags)) in __blk_release_queue() 896 WARN_ONCE(test_bit(QUEUE_FLAG_REGISTERED, &q->queue_flags), in blk_register_queue() 973 if (!test_bit(QUEUE_FLAG_REGISTERED, &q->queue_flags)) in blk_unregister_queue()
|
D | blk-timeout.c | 24 if (!test_bit(QUEUE_FLAG_FAIL_IO, &q->queue_flags)) in blk_should_fake_timeout() 44 int set = test_bit(QUEUE_FLAG_FAIL_IO, &disk->queue->queue_flags); in part_timeout_show()
|
D | blk-softirq.c | 115 if (!test_bit(QUEUE_FLAG_SAME_FORCE, &q->queue_flags)) in __blk_complete_request()
|
D | blk-mq.c | 568 if (!test_bit(QUEUE_FLAG_SAME_COMP, &rq->q->queue_flags)) { in __blk_mq_complete_request() 574 if (!test_bit(QUEUE_FLAG_SAME_FORCE, &rq->q->queue_flags)) in __blk_mq_complete_request() 638 if (test_bit(QUEUE_FLAG_STATS, &q->queue_flags)) { in blk_mq_start_request() 2602 q->queue_flags |= QUEUE_FLAG_MQ_DEFAULT; in blk_mq_init_allocated_queue() 3016 if (test_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags) || in blk_poll_stats_enable() 3029 if (!test_bit(QUEUE_FLAG_POLL_STATS, &q->queue_flags) || in blk_mq_poll_stats_start() 3187 if (!test_bit(QUEUE_FLAG_POLL, &q->queue_flags)) in blk_mq_poll()
|
D | blk-wbt.c | 709 if (!test_bit(QUEUE_FLAG_REGISTERED, &q->queue_flags)) in wbt_enable_default() 816 wbt_set_write_cache(q, test_bit(QUEUE_FLAG_WC, &q->queue_flags)); in wbt_init()
|
D | blk-merge.c | 303 &rq->q->queue_flags); in blk_recalc_rq_segments() 319 if (test_bit(QUEUE_FLAG_NO_SG_MERGE, &q->queue_flags) && in blk_recount_segments()
|
D | blk-flush.c | 442 unsigned long fflags = q->queue_flags; /* may change, cache */ in blk_insert_flush()
|
D | bsg.c | 175 if (!test_bit(QUEUE_FLAG_BIDI, &q->queue_flags)) { in bsg_map_hdr()
|
D | blk-settings.c | 903 wbt_set_write_cache(q, test_bit(QUEUE_FLAG_WC, &q->queue_flags)); in blk_queue_write_cache()
|
D | blk-core.c | 1174 q->queue_flags |= QUEUE_FLAG_DEFAULT; in blk_init_allocated_queue() 2081 if (test_bit(QUEUE_FLAG_SAME_COMP, &q->queue_flags)) in blk_queue_bio() 2289 !test_bit(QUEUE_FLAG_WC, &q->queue_flags)) { in generic_make_request_checks() 2989 if (test_bit(QUEUE_FLAG_STATS, &req->q->queue_flags)) { in blk_start_request()
|
D | blk-mq-debugfs.c | 143 blk_flags_show(m, q->queue_flags, blk_queue_flag_name, in queue_state_show()
|
D | elevator.c | 1090 if (!test_bit(QUEUE_FLAG_REGISTERED, &q->queue_flags)) in __elevator_change()
|
/Linux-v4.19/include/linux/ |
D | blkdev.h | 506 unsigned long queue_flags; member 716 #define blk_queue_tagged(q) test_bit(QUEUE_FLAG_QUEUED, &(q)->queue_flags) 717 #define blk_queue_stopped(q) test_bit(QUEUE_FLAG_STOPPED, &(q)->queue_flags) 718 #define blk_queue_dying(q) test_bit(QUEUE_FLAG_DYING, &(q)->queue_flags) 719 #define blk_queue_dead(q) test_bit(QUEUE_FLAG_DEAD, &(q)->queue_flags) 720 #define blk_queue_bypass(q) test_bit(QUEUE_FLAG_BYPASS, &(q)->queue_flags) 721 #define blk_queue_init_done(q) test_bit(QUEUE_FLAG_INIT_DONE, &(q)->queue_flags) 722 #define blk_queue_nomerges(q) test_bit(QUEUE_FLAG_NOMERGES, &(q)->queue_flags) 724 test_bit(QUEUE_FLAG_NOXMERGES, &(q)->queue_flags) 725 #define blk_queue_nonrot(q) test_bit(QUEUE_FLAG_NONROT, &(q)->queue_flags) [all …]
|
/Linux-v4.19/drivers/target/ |
D | target_core_iblock.c | 704 if (test_bit(QUEUE_FLAG_FUA, &q->queue_flags)) { in iblock_execute_rw() 707 else if (!test_bit(QUEUE_FLAG_WC, &q->queue_flags)) in iblock_execute_rw() 850 return test_bit(QUEUE_FLAG_WC, &q->queue_flags); in iblock_get_write_cache()
|
/Linux-v4.19/kernel/sched/ |
D | core.c | 4126 int queue_flags = DEQUEUE_SAVE | DEQUEUE_MOVE | DEQUEUE_NOCLOCK; in __sched_setscheduler() local 4316 queue_flags &= ~DEQUEUE_MOVE; in __sched_setscheduler() 4322 dequeue_task(rq, p, queue_flags); in __sched_setscheduler() 4335 queue_flags |= ENQUEUE_HEAD; in __sched_setscheduler() 4337 enqueue_task(rq, p, queue_flags); in __sched_setscheduler() 6345 int queued, running, queue_flags = in sched_move_task() local 6357 dequeue_task(rq, tsk, queue_flags); in sched_move_task() 6364 enqueue_task(rq, tsk, queue_flags); in sched_move_task()
|
/Linux-v4.19/drivers/md/ |
D | dm-table.c | 1643 return q && (q->queue_flags & flush); in device_flush_capable() 1726 return q && !test_bit(QUEUE_FLAG_NO_SG_MERGE, &q->queue_flags); in queue_supports_sg_merge()
|
D | raid5-ppl.c | 1332 if (test_bit(QUEUE_FLAG_WC, &q->queue_flags)) in ppl_init_child_log()
|
D | raid5-cache.c | 3095 log->need_cache_flush = test_bit(QUEUE_FLAG_WC, &q->queue_flags) != 0; in r5l_init_log()
|
/Linux-v4.19/drivers/block/xen-blkback/ |
D | xenbus.c | 490 if (q && test_bit(QUEUE_FLAG_WC, &q->queue_flags)) in xen_vbd_create()
|
/Linux-v4.19/fs/btrfs/ |
D | disk-io.c | 3579 if (!test_bit(QUEUE_FLAG_WC, &q->queue_flags)) in write_dev_flush()
|