Lines Matching refs:bdev

331 int blkdev_report_zones(struct block_device *bdev, sector_t sector,
333 unsigned int bdev_nr_zones(struct block_device *bdev);
334 extern int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op,
340 extern int blkdev_report_zones_ioctl(struct block_device *bdev, fmode_t mode,
342 extern int blkdev_zone_mgmt_ioctl(struct block_device *bdev, fmode_t mode,
347 static inline unsigned int bdev_nr_zones(struct block_device *bdev) in bdev_nr_zones() argument
352 static inline int blkdev_report_zones_ioctl(struct block_device *bdev, in blkdev_report_zones_ioctl() argument
359 static inline int blkdev_zone_mgmt_ioctl(struct block_device *bdev, in blkdev_zone_mgmt_ioctl() argument
703 static inline unsigned int bdev_max_open_zones(struct block_device *bdev) in bdev_max_open_zones() argument
705 return bdev->bd_disk->max_open_zones; in bdev_max_open_zones()
708 static inline unsigned int bdev_max_active_zones(struct block_device *bdev) in bdev_max_active_zones() argument
710 return bdev->bd_disk->max_active_zones; in bdev_max_active_zones()
726 static inline unsigned int bdev_max_open_zones(struct block_device *bdev) in bdev_max_open_zones() argument
731 static inline unsigned int bdev_max_active_zones(struct block_device *bdev) in bdev_max_active_zones() argument
772 static inline int bdev_read_only(struct block_device *bdev) in bdev_read_only() argument
774 return bdev->bd_read_only || get_disk_ro(bdev->bd_disk); in bdev_read_only()
783 static inline sector_t get_start_sect(struct block_device *bdev) in get_start_sect() argument
785 return bdev->bd_start_sect; in get_start_sect()
788 static inline sector_t bdev_nr_sectors(struct block_device *bdev) in bdev_nr_sectors() argument
790 return bdev->bd_nr_sectors; in bdev_nr_sectors()
793 static inline loff_t bdev_nr_bytes(struct block_device *bdev) in bdev_nr_bytes() argument
795 return (loff_t)bdev_nr_sectors(bdev) << SECTOR_SHIFT; in bdev_nr_bytes()
836 bool bdev_check_media_change(struct block_device *bdev);
837 int __invalidate_device(struct block_device *bdev, bool kill_dirty);
841 int bd_link_disk_holder(struct block_device *bdev, struct gendisk *disk);
842 void bd_unlink_disk_holder(struct block_device *bdev, struct gendisk *disk);
845 static inline int bd_link_disk_holder(struct block_device *bdev, in bd_link_disk_holder() argument
850 static inline void bd_unlink_disk_holder(struct block_device *bdev, in bd_unlink_disk_holder() argument
889 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue() argument
891 return bdev->bd_queue; /* this is never NULL */ in bdev_get_queue()
951 extern void disk_stack_limits(struct gendisk *disk, struct block_device *bdev,
1030 int blkdev_issue_flush(struct block_device *bdev);
1053 static inline int blkdev_issue_flush(struct block_device *bdev) in blkdev_issue_flush() argument
1066 int blkdev_issue_discard(struct block_device *bdev, sector_t sector,
1068 int __blkdev_issue_discard(struct block_device *bdev, sector_t sector,
1070 int blkdev_issue_secure_erase(struct block_device *bdev, sector_t sector,
1076 extern int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
1079 extern int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
1103 static inline bool bdev_is_partition(struct block_device *bdev) in bdev_is_partition() argument
1105 return bdev->bd_partno; in bdev_is_partition()
1165 bdev_max_zone_append_sectors(struct block_device *bdev) in bdev_max_zone_append_sectors() argument
1167 return queue_max_zone_append_sectors(bdev_get_queue(bdev)); in bdev_max_zone_append_sectors()
1170 static inline unsigned int bdev_max_segments(struct block_device *bdev) in bdev_max_segments() argument
1172 return queue_max_segments(bdev_get_queue(bdev)); in bdev_max_segments()
1185 static inline unsigned int bdev_logical_block_size(struct block_device *bdev) in bdev_logical_block_size() argument
1187 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size()
1195 static inline unsigned int bdev_physical_block_size(struct block_device *bdev) in bdev_physical_block_size() argument
1197 return queue_physical_block_size(bdev_get_queue(bdev)); in bdev_physical_block_size()
1205 static inline int bdev_io_min(struct block_device *bdev) in bdev_io_min() argument
1207 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min()
1215 static inline int bdev_io_opt(struct block_device *bdev) in bdev_io_opt() argument
1217 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt()
1227 bdev_zone_write_granularity(struct block_device *bdev) in bdev_zone_write_granularity() argument
1229 return queue_zone_write_granularity(bdev_get_queue(bdev)); in bdev_zone_write_granularity()
1232 int bdev_alignment_offset(struct block_device *bdev);
1233 unsigned int bdev_discard_alignment(struct block_device *bdev);
1235 static inline unsigned int bdev_max_discard_sectors(struct block_device *bdev) in bdev_max_discard_sectors() argument
1237 return bdev_get_queue(bdev)->limits.max_discard_sectors; in bdev_max_discard_sectors()
1240 static inline unsigned int bdev_discard_granularity(struct block_device *bdev) in bdev_discard_granularity() argument
1242 return bdev_get_queue(bdev)->limits.discard_granularity; in bdev_discard_granularity()
1246 bdev_max_secure_erase_sectors(struct block_device *bdev) in bdev_max_secure_erase_sectors() argument
1248 return bdev_get_queue(bdev)->limits.max_secure_erase_sectors; in bdev_max_secure_erase_sectors()
1251 static inline unsigned int bdev_write_zeroes_sectors(struct block_device *bdev) in bdev_write_zeroes_sectors() argument
1253 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_zeroes_sectors()
1261 static inline bool bdev_nonrot(struct block_device *bdev) in bdev_nonrot() argument
1263 return blk_queue_nonrot(bdev_get_queue(bdev)); in bdev_nonrot()
1266 static inline bool bdev_stable_writes(struct block_device *bdev) in bdev_stable_writes() argument
1269 &bdev_get_queue(bdev)->queue_flags); in bdev_stable_writes()
1272 static inline bool bdev_write_cache(struct block_device *bdev) in bdev_write_cache() argument
1274 return test_bit(QUEUE_FLAG_WC, &bdev_get_queue(bdev)->queue_flags); in bdev_write_cache()
1277 static inline bool bdev_fua(struct block_device *bdev) in bdev_fua() argument
1279 return test_bit(QUEUE_FLAG_FUA, &bdev_get_queue(bdev)->queue_flags); in bdev_fua()
1282 static inline bool bdev_nowait(struct block_device *bdev) in bdev_nowait() argument
1284 return test_bit(QUEUE_FLAG_NOWAIT, &bdev_get_queue(bdev)->queue_flags); in bdev_nowait()
1287 static inline enum blk_zoned_model bdev_zoned_model(struct block_device *bdev) in bdev_zoned_model() argument
1289 struct request_queue *q = bdev_get_queue(bdev); in bdev_zoned_model()
1297 static inline bool bdev_is_zoned(struct block_device *bdev) in bdev_is_zoned() argument
1299 struct request_queue *q = bdev_get_queue(bdev); in bdev_is_zoned()
1307 static inline bool bdev_op_is_zoned_write(struct block_device *bdev, in bdev_op_is_zoned_write() argument
1310 if (!bdev_is_zoned(bdev)) in bdev_op_is_zoned_write()
1316 static inline sector_t bdev_zone_sectors(struct block_device *bdev) in bdev_zone_sectors() argument
1318 struct request_queue *q = bdev_get_queue(bdev); in bdev_zone_sectors()
1330 static inline unsigned int bdev_dma_alignment(struct block_device *bdev) in bdev_dma_alignment() argument
1332 return queue_dma_alignment(bdev_get_queue(bdev)); in bdev_dma_alignment()
1335 static inline bool bdev_iter_is_aligned(struct block_device *bdev, in bdev_iter_is_aligned() argument
1338 return iov_iter_is_aligned(iter, bdev_dma_alignment(bdev), in bdev_iter_is_aligned()
1339 bdev_logical_block_size(bdev) - 1); in bdev_iter_is_aligned()
1360 static inline unsigned int block_size(struct block_device *bdev) in block_size() argument
1362 return 1 << bdev->bd_inode->i_blkbits; in block_size()
1410 int (*set_read_only)(struct block_device *bdev, bool ro);
1455 unsigned long bdev_start_io_acct(struct block_device *bdev,
1458 void bdev_end_io_acct(struct block_device *bdev, enum req_op op,
1476 int bdev_read_only(struct block_device *bdev);
1477 int set_blocksize(struct block_device *bdev, int size);
1494 int bd_prepare_to_claim(struct block_device *bdev, void *holder);
1495 void bd_abort_claiming(struct block_device *bdev, void *holder);
1496 void blkdev_put(struct block_device *bdev, fmode_t mode);
1500 void blkdev_put_no_open(struct block_device *bdev);
1503 void bdev_add(struct block_device *bdev, dev_t dev);
1505 int truncate_bdev_range(struct block_device *bdev, fmode_t mode, loff_t lstart,
1509 void invalidate_bdev(struct block_device *bdev);
1510 int sync_blockdev(struct block_device *bdev);
1511 int sync_blockdev_range(struct block_device *bdev, loff_t lstart, loff_t lend);
1512 int sync_blockdev_nowait(struct block_device *bdev);
1517 static inline void invalidate_bdev(struct block_device *bdev) in invalidate_bdev() argument
1520 static inline int sync_blockdev(struct block_device *bdev) in sync_blockdev() argument
1524 static inline int sync_blockdev_nowait(struct block_device *bdev) in sync_blockdev_nowait() argument
1539 int fsync_bdev(struct block_device *bdev);
1541 int freeze_bdev(struct block_device *bdev);
1542 int thaw_bdev(struct block_device *bdev);