Lines Matching refs:bdev_get_queue

858 static inline struct request_queue *bdev_get_queue(struct block_device *bdev)  in bdev_get_queue()  function
1136 return queue_max_zone_append_sectors(bdev_get_queue(bdev)); in bdev_max_zone_append_sectors()
1141 return queue_max_segments(bdev_get_queue(bdev)); in bdev_max_segments()
1156 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size()
1166 return queue_physical_block_size(bdev_get_queue(bdev)); in bdev_physical_block_size()
1176 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min()
1186 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt()
1198 return queue_zone_write_granularity(bdev_get_queue(bdev)); in bdev_zone_write_granularity()
1206 return bdev_get_queue(bdev)->limits.max_discard_sectors; in bdev_max_discard_sectors()
1211 return bdev_get_queue(bdev)->limits.discard_granularity; in bdev_discard_granularity()
1217 return bdev_get_queue(bdev)->limits.max_secure_erase_sectors; in bdev_max_secure_erase_sectors()
1222 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_zeroes_sectors()
1232 return blk_queue_nonrot(bdev_get_queue(bdev)); in bdev_nonrot()
1238 &bdev_get_queue(bdev)->queue_flags); in bdev_synchronous()
1244 &bdev_get_queue(bdev)->queue_flags); in bdev_stable_writes()
1249 return test_bit(QUEUE_FLAG_WC, &bdev_get_queue(bdev)->queue_flags); in bdev_write_cache()
1254 return test_bit(QUEUE_FLAG_FUA, &bdev_get_queue(bdev)->queue_flags); in bdev_fua()
1259 return test_bit(QUEUE_FLAG_NOWAIT, &bdev_get_queue(bdev)->queue_flags); in bdev_nowait()
1264 return blk_queue_zoned_model(bdev_get_queue(bdev)); in bdev_zoned_model()
1269 return blk_queue_is_zoned(bdev_get_queue(bdev)); in bdev_is_zoned()
1291 struct request_queue *q = bdev_get_queue(bdev); in bdev_zone_sectors()
1317 return queue_dma_alignment(bdev_get_queue(bdev)); in bdev_dma_alignment()