Lines Matching refs:request_queue
29 struct request_queue;
119 struct request_queue *q;
373 struct request_queue { struct
594 void blk_queue_flag_set(unsigned int flag, struct request_queue *q);
595 void blk_queue_flag_clear(unsigned int flag, struct request_queue *q);
596 bool blk_queue_flag_test_and_set(unsigned int flag, struct request_queue *q);
636 extern void blk_set_pm_only(struct request_queue *q);
637 extern void blk_clear_pm_only(struct request_queue *q);
650 static inline bool queue_is_mq(struct request_queue *q) in queue_is_mq()
656 static inline enum rpm_status queue_rpm_status(struct request_queue *q) in queue_rpm_status()
661 static inline enum rpm_status queue_rpm_status(struct request_queue *q) in queue_rpm_status()
668 blk_queue_zoned_model(struct request_queue *q) in blk_queue_zoned_model()
675 static inline bool blk_queue_is_zoned(struct request_queue *q) in blk_queue_is_zoned()
686 static inline sector_t blk_queue_zone_sectors(struct request_queue *q) in blk_queue_zone_sectors()
692 static inline unsigned int blk_queue_nr_zones(struct request_queue *q) in blk_queue_nr_zones()
697 static inline unsigned int blk_queue_zone_no(struct request_queue *q, in blk_queue_zone_no()
705 static inline bool blk_queue_zone_is_seq(struct request_queue *q, in blk_queue_zone_is_seq()
715 static inline void blk_queue_max_open_zones(struct request_queue *q, in blk_queue_max_open_zones()
721 static inline unsigned int queue_max_open_zones(const struct request_queue *q) in queue_max_open_zones()
726 static inline void blk_queue_max_active_zones(struct request_queue *q, in blk_queue_max_active_zones()
732 static inline unsigned int queue_max_active_zones(const struct request_queue *q) in queue_max_active_zones()
737 static inline unsigned int blk_queue_nr_zones(struct request_queue *q) in blk_queue_nr_zones()
741 static inline bool blk_queue_zone_is_seq(struct request_queue *q, in blk_queue_zone_is_seq()
746 static inline unsigned int blk_queue_zone_no(struct request_queue *q, in blk_queue_zone_no()
751 static inline unsigned int queue_max_open_zones(const struct request_queue *q) in queue_max_open_zones()
755 static inline unsigned int queue_max_active_zones(const struct request_queue *q) in queue_max_active_zones()
797 static inline unsigned int blk_queue_depth(struct request_queue *q) in blk_queue_depth()
858 extern void blk_rq_init(struct request_queue *q, struct request *rq);
860 extern struct request *blk_get_request(struct request_queue *, unsigned int op,
862 extern int blk_lld_busy(struct request_queue *q);
868 extern blk_status_t blk_insert_cloned_request(struct request_queue *q,
872 extern int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags);
873 extern void blk_queue_exit(struct request_queue *q);
874 extern void blk_sync_queue(struct request_queue *q);
875 extern int blk_rq_map_user(struct request_queue *, struct request *,
879 extern int blk_rq_map_kern(struct request_queue *, struct request *, void *, unsigned int, gfp_t);
880 extern int blk_rq_map_user_iov(struct request_queue *, struct request *,
895 int blk_poll(struct request_queue *q, blk_qc_t cookie, bool spin);
897 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue()
1012 static inline unsigned int blk_queue_get_max_sectors(struct request_queue *q, in blk_queue_get_max_sectors()
1032 static inline unsigned int blk_max_size_offset(struct request_queue *q, in blk_max_size_offset()
1054 struct request_queue *q = rq->q; in blk_rq_get_max_sectors()
1095 extern void blk_cleanup_queue(struct request_queue *);
1096 void blk_queue_bounce_limit(struct request_queue *q, enum blk_bounce limit);
1097 extern void blk_queue_max_hw_sectors(struct request_queue *, unsigned int);
1098 extern void blk_queue_chunk_sectors(struct request_queue *, unsigned int);
1099 extern void blk_queue_max_segments(struct request_queue *, unsigned short);
1100 extern void blk_queue_max_discard_segments(struct request_queue *,
1102 extern void blk_queue_max_segment_size(struct request_queue *, unsigned int);
1103 extern void blk_queue_max_discard_sectors(struct request_queue *q,
1105 extern void blk_queue_max_write_same_sectors(struct request_queue *q,
1107 extern void blk_queue_max_write_zeroes_sectors(struct request_queue *q,
1109 extern void blk_queue_logical_block_size(struct request_queue *, unsigned int);
1110 extern void blk_queue_max_zone_append_sectors(struct request_queue *q,
1112 extern void blk_queue_physical_block_size(struct request_queue *, unsigned int);
1113 void blk_queue_zone_write_granularity(struct request_queue *q,
1115 extern void blk_queue_alignment_offset(struct request_queue *q,
1119 extern void blk_queue_io_min(struct request_queue *q, unsigned int min);
1121 extern void blk_queue_io_opt(struct request_queue *q, unsigned int opt);
1122 extern void blk_set_queue_depth(struct request_queue *q, unsigned int depth);
1129 extern void blk_queue_update_dma_pad(struct request_queue *, unsigned int);
1130 extern void blk_queue_segment_boundary(struct request_queue *, unsigned long);
1131 extern void blk_queue_virt_boundary(struct request_queue *, unsigned long);
1132 extern void blk_queue_dma_alignment(struct request_queue *, int);
1133 extern void blk_queue_update_dma_alignment(struct request_queue *, int);
1134 extern void blk_queue_rq_timeout(struct request_queue *, unsigned int);
1135 extern void blk_queue_write_cache(struct request_queue *q, bool enabled, bool fua);
1136 extern void blk_queue_required_elevator_features(struct request_queue *q,
1138 extern bool blk_queue_can_use_dma_map_merging(struct request_queue *q,
1166 int __blk_rq_map_sg(struct request_queue *q, struct request *rq,
1168 static inline int blk_rq_map_sg(struct request_queue *q, struct request *rq, in blk_rq_map_sg()
1177 bool __must_check blk_get_queue(struct request_queue *);
1178 extern void blk_put_queue(struct request_queue *);
1179 extern void blk_set_queue_dying(struct request_queue *);
1337 static inline unsigned long queue_segment_boundary(const struct request_queue *q) in queue_segment_boundary()
1342 static inline unsigned long queue_virt_boundary(const struct request_queue *q) in queue_virt_boundary()
1347 static inline unsigned int queue_max_sectors(const struct request_queue *q) in queue_max_sectors()
1352 static inline unsigned int queue_max_bytes(struct request_queue *q) in queue_max_bytes()
1357 static inline unsigned int queue_max_hw_sectors(const struct request_queue *q) in queue_max_hw_sectors()
1362 static inline unsigned short queue_max_segments(const struct request_queue *q) in queue_max_segments()
1367 static inline unsigned short queue_max_discard_segments(const struct request_queue *q) in queue_max_discard_segments()
1372 static inline unsigned int queue_max_segment_size(const struct request_queue *q) in queue_max_segment_size()
1377 static inline unsigned int queue_max_zone_append_sectors(const struct request_queue *q) in queue_max_zone_append_sectors()
1385 static inline unsigned queue_logical_block_size(const struct request_queue *q) in queue_logical_block_size()
1400 static inline unsigned int queue_physical_block_size(const struct request_queue *q) in queue_physical_block_size()
1410 static inline unsigned int queue_io_min(const struct request_queue *q) in queue_io_min()
1420 static inline unsigned int queue_io_opt(const struct request_queue *q) in queue_io_opt()
1431 queue_zone_write_granularity(const struct request_queue *q) in queue_zone_write_granularity()
1442 static inline int queue_alignment_offset(const struct request_queue *q) in queue_alignment_offset()
1461 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset()
1471 static inline int queue_discard_alignment(const struct request_queue *q) in queue_discard_alignment()
1520 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment()
1530 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_same()
1540 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_zeroes_sectors()
1550 struct request_queue *q = bdev_get_queue(bdev); in bdev_zoned_model()
1560 struct request_queue *q = bdev_get_queue(bdev); in bdev_is_zoned()
1570 struct request_queue *q = bdev_get_queue(bdev); in bdev_zone_sectors()
1579 struct request_queue *q = bdev_get_queue(bdev); in bdev_max_open_zones()
1588 struct request_queue *q = bdev_get_queue(bdev); in bdev_max_active_zones()
1595 static inline int queue_dma_alignment(const struct request_queue *q) in queue_dma_alignment()
1600 static inline int blk_rq_aligned(struct request_queue *q, unsigned long addr, in blk_rq_aligned()
1664 extern int blk_rq_map_integrity_sg(struct request_queue *, struct bio *,
1666 extern int blk_rq_count_integrity_sg(struct request_queue *, struct bio *);
1685 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity()
1695 static inline void blk_queue_max_integrity_segments(struct request_queue *q, in blk_queue_max_integrity_segments()
1702 queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments()
1751 static inline int blk_rq_count_integrity_sg(struct request_queue *q, in blk_rq_count_integrity_sg()
1756 static inline int blk_rq_map_integrity_sg(struct request_queue *q, in blk_rq_map_integrity_sg()
1771 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity()
1786 static inline void blk_queue_max_integrity_segments(struct request_queue *q, in blk_queue_max_integrity_segments()
1790 static inline unsigned short queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments()
1816 bool blk_ksm_register(struct blk_keyslot_manager *ksm, struct request_queue *q);
1818 void blk_ksm_unregister(struct request_queue *q);
1823 struct request_queue *q) in blk_ksm_register()
1828 static inline void blk_ksm_unregister(struct request_queue *q) { } in blk_ksm_unregister()