Lines Matching refs:q

51 static inline void queue_lockdep_assert_held(struct request_queue *q)  in queue_lockdep_assert_held()  argument
53 if (q->queue_lock) in queue_lockdep_assert_held()
54 lockdep_assert_held(q->queue_lock); in queue_lockdep_assert_held()
58 struct request_queue *q) in queue_flag_set_unlocked() argument
60 if (test_bit(QUEUE_FLAG_INIT_DONE, &q->queue_flags) && in queue_flag_set_unlocked()
61 kref_read(&q->kobj.kref)) in queue_flag_set_unlocked()
62 lockdep_assert_held(q->queue_lock); in queue_flag_set_unlocked()
63 __set_bit(flag, &q->queue_flags); in queue_flag_set_unlocked()
67 struct request_queue *q) in queue_flag_clear_unlocked() argument
69 if (test_bit(QUEUE_FLAG_INIT_DONE, &q->queue_flags) && in queue_flag_clear_unlocked()
70 kref_read(&q->kobj.kref)) in queue_flag_clear_unlocked()
71 lockdep_assert_held(q->queue_lock); in queue_flag_clear_unlocked()
72 __clear_bit(flag, &q->queue_flags); in queue_flag_clear_unlocked()
76 struct request_queue *q) in queue_flag_test_and_clear() argument
78 queue_lockdep_assert_held(q); in queue_flag_test_and_clear()
80 if (test_bit(flag, &q->queue_flags)) { in queue_flag_test_and_clear()
81 __clear_bit(flag, &q->queue_flags); in queue_flag_test_and_clear()
89 struct request_queue *q) in queue_flag_test_and_set() argument
91 queue_lockdep_assert_held(q); in queue_flag_test_and_set()
93 if (!test_bit(flag, &q->queue_flags)) { in queue_flag_test_and_set()
94 __set_bit(flag, &q->queue_flags); in queue_flag_test_and_set()
101 static inline void queue_flag_set(unsigned int flag, struct request_queue *q) in queue_flag_set() argument
103 queue_lockdep_assert_held(q); in queue_flag_set()
104 __set_bit(flag, &q->queue_flags); in queue_flag_set()
107 static inline void queue_flag_clear(unsigned int flag, struct request_queue *q) in queue_flag_clear() argument
109 queue_lockdep_assert_held(q); in queue_flag_clear()
110 __clear_bit(flag, &q->queue_flags); in queue_flag_clear()
114 struct request_queue *q, struct blk_mq_ctx *ctx) in blk_get_flush_queue() argument
116 if (q->mq_ops) in blk_get_flush_queue()
117 return blk_mq_map_queue(q, ctx->cpu)->fq; in blk_get_flush_queue()
118 return q->fq; in blk_get_flush_queue()
121 static inline void __blk_get_queue(struct request_queue *q) in __blk_get_queue() argument
123 kobject_get(&q->kobj); in __blk_get_queue()
126 struct blk_flush_queue *blk_alloc_flush_queue(struct request_queue *q,
128 void blk_free_flush_queue(struct blk_flush_queue *q);
130 int blk_init_rl(struct request_list *rl, struct request_queue *q,
132 void blk_exit_rl(struct request_queue *q, struct request_list *rl);
133 void blk_exit_queue(struct request_queue *q);
134 void blk_rq_bio_prep(struct request_queue *q, struct request *rq,
136 void blk_queue_bypass_start(struct request_queue *q);
137 void blk_queue_bypass_end(struct request_queue *q);
138 void __blk_queue_free_tags(struct request_queue *q);
139 void blk_freeze_queue(struct request_queue *q);
141 static inline void blk_queue_enter_live(struct request_queue *q) in blk_queue_enter_live() argument
149 percpu_ref_get(&q->q_usage_counter); in blk_queue_enter_live()
177 bool bio_attempt_front_merge(struct request_queue *q, struct request *req,
179 bool bio_attempt_back_merge(struct request_queue *q, struct request *req,
181 bool bio_attempt_discard_merge(struct request_queue *q, struct request *req,
183 bool blk_attempt_plug_merge(struct request_queue *q, struct bio *bio,
186 unsigned int blk_plug_queued_count(struct request_queue *q);
219 static inline void elv_activate_rq(struct request_queue *q, struct request *rq) in elv_activate_rq() argument
221 struct elevator_queue *e = q->elevator; in elv_activate_rq()
224 e->type->ops.sq.elevator_activate_req_fn(q, rq); in elv_activate_rq()
227 static inline void elv_deactivate_rq(struct request_queue *q, struct request *rq) in elv_deactivate_rq() argument
229 struct elevator_queue *e = q->elevator; in elv_deactivate_rq()
232 e->type->ops.sq.elevator_deactivate_req_fn(q, rq); in elv_deactivate_rq()
236 int elevator_init_mq(struct request_queue *q);
237 int elevator_switch_mq(struct request_queue *q,
240 int elv_register_queue(struct request_queue *q);
241 void elv_unregister_queue(struct request_queue *q);
251 static inline int blk_should_fake_timeout(struct request_queue *q) in blk_should_fake_timeout() argument
257 int ll_back_merge_fn(struct request_queue *q, struct request *req,
259 int ll_front_merge_fn(struct request_queue *q, struct request *req,
261 struct request *attempt_back_merge(struct request_queue *q, struct request *rq);
262 struct request *attempt_front_merge(struct request_queue *q, struct request *rq);
263 int blk_attempt_req_merge(struct request_queue *q, struct request *rq,
270 void blk_queue_congestion_threshold(struct request_queue *q);
280 static inline int queue_congestion_on_threshold(struct request_queue *q) in queue_congestion_on_threshold() argument
282 return q->nr_congestion_on; in queue_congestion_on_threshold()
288 static inline int queue_congestion_off_threshold(struct request_queue *q) in queue_congestion_off_threshold() argument
290 return q->nr_congestion_off; in queue_congestion_off_threshold()
309 static inline void req_set_nomerge(struct request_queue *q, struct request *req) in req_set_nomerge() argument
312 if (req == q->last_merge) in req_set_nomerge()
313 q->last_merge = NULL; in req_set_nomerge()
335 struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q);
336 struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
338 void ioc_clear_queue(struct request_queue *q);
382 extern void blk_throtl_drain(struct request_queue *q);
383 extern int blk_throtl_init(struct request_queue *q);
384 extern void blk_throtl_exit(struct request_queue *q);
385 extern void blk_throtl_register_queue(struct request_queue *q);
387 static inline void blk_throtl_drain(struct request_queue *q) { } in blk_throtl_drain() argument
388 static inline int blk_throtl_init(struct request_queue *q) { return 0; } in blk_throtl_init() argument
389 static inline void blk_throtl_exit(struct request_queue *q) { } in blk_throtl_exit() argument
390 static inline void blk_throtl_register_queue(struct request_queue *q) { } in blk_throtl_register_queue() argument
393 extern ssize_t blk_throtl_sample_time_show(struct request_queue *q, char *page);
394 extern ssize_t blk_throtl_sample_time_store(struct request_queue *q,
405 extern void blk_queue_bounce(struct request_queue *q, struct bio **bio);
411 static inline void blk_queue_bounce(struct request_queue *q, struct bio **bio) in blk_queue_bounce() argument
416 extern void blk_drain_queue(struct request_queue *q);
419 extern int blk_iolatency_init(struct request_queue *q);
421 static inline int blk_iolatency_init(struct request_queue *q) { return 0; } in blk_iolatency_init() argument