Lines Matching refs:bdev
336 int blkdev_report_zones(struct block_device *bdev, sector_t sector,
339 extern int blkdev_zone_mgmt(struct block_device *bdev, enum req_opf op,
345 extern int blkdev_report_zones_ioctl(struct block_device *bdev, fmode_t mode,
347 extern int blkdev_zone_mgmt_ioctl(struct block_device *bdev, fmode_t mode,
357 static inline int blkdev_report_zones_ioctl(struct block_device *bdev, in blkdev_report_zones_ioctl() argument
364 static inline int blkdev_zone_mgmt_ioctl(struct block_device *bdev, in blkdev_zone_mgmt_ioctl() argument
897 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue() argument
899 return bdev->bd_disk->queue; /* this is never NULL */ in bdev_get_queue()
1127 extern void disk_stack_limits(struct gendisk *disk, struct block_device *bdev,
1242 int blkdev_issue_flush(struct block_device *bdev);
1270 static inline int blkdev_issue_flush(struct block_device *bdev) in blkdev_issue_flush() argument
1283 extern int blkdev_issue_write_same(struct block_device *bdev, sector_t sector,
1288 extern int blkdev_issue_discard(struct block_device *bdev, sector_t sector,
1290 extern int __blkdev_issue_discard(struct block_device *bdev, sector_t sector,
1297 extern int __blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
1300 extern int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
1324 static inline bool bdev_is_partition(struct block_device *bdev) in bdev_is_partition() argument
1326 return bdev->bd_partno; in bdev_is_partition()
1395 static inline unsigned int bdev_logical_block_size(struct block_device *bdev) in bdev_logical_block_size() argument
1397 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size()
1405 static inline unsigned int bdev_physical_block_size(struct block_device *bdev) in bdev_physical_block_size() argument
1407 return queue_physical_block_size(bdev_get_queue(bdev)); in bdev_physical_block_size()
1415 static inline int bdev_io_min(struct block_device *bdev) in bdev_io_min() argument
1417 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min()
1425 static inline int bdev_io_opt(struct block_device *bdev) in bdev_io_opt() argument
1427 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt()
1437 bdev_zone_write_granularity(struct block_device *bdev) in bdev_zone_write_granularity() argument
1439 return queue_zone_write_granularity(bdev_get_queue(bdev)); in bdev_zone_write_granularity()
1459 static inline int bdev_alignment_offset(struct block_device *bdev) in bdev_alignment_offset() argument
1461 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset()
1465 if (bdev_is_partition(bdev)) in bdev_alignment_offset()
1467 bdev->bd_start_sect); in bdev_alignment_offset()
1518 static inline int bdev_discard_alignment(struct block_device *bdev) in bdev_discard_alignment() argument
1520 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment()
1522 if (bdev_is_partition(bdev)) in bdev_discard_alignment()
1524 bdev->bd_start_sect); in bdev_discard_alignment()
1528 static inline unsigned int bdev_write_same(struct block_device *bdev) in bdev_write_same() argument
1530 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_same()
1538 static inline unsigned int bdev_write_zeroes_sectors(struct block_device *bdev) in bdev_write_zeroes_sectors() argument
1540 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_zeroes_sectors()
1548 static inline enum blk_zoned_model bdev_zoned_model(struct block_device *bdev) in bdev_zoned_model() argument
1550 struct request_queue *q = bdev_get_queue(bdev); in bdev_zoned_model()
1558 static inline bool bdev_is_zoned(struct block_device *bdev) in bdev_is_zoned() argument
1560 struct request_queue *q = bdev_get_queue(bdev); in bdev_is_zoned()
1568 static inline sector_t bdev_zone_sectors(struct block_device *bdev) in bdev_zone_sectors() argument
1570 struct request_queue *q = bdev_get_queue(bdev); in bdev_zone_sectors()
1577 static inline unsigned int bdev_max_open_zones(struct block_device *bdev) in bdev_max_open_zones() argument
1579 struct request_queue *q = bdev_get_queue(bdev); in bdev_max_open_zones()
1586 static inline unsigned int bdev_max_active_zones(struct block_device *bdev) in bdev_max_active_zones() argument
1588 struct request_queue *q = bdev_get_queue(bdev); in bdev_max_active_zones()
1618 static inline unsigned int block_size(struct block_device *bdev) in block_size() argument
1620 return 1 << bdev->bd_inode->i_blkbits; in block_size()
1679 struct blk_integrity *bdev_get_integrity(struct block_device *bdev) in bdev_get_integrity() argument
1681 return blk_get_integrity(bdev->bd_disk); in bdev_get_integrity()
1844 int (*set_read_only)(struct block_device *bdev, bool ro);
1958 int bdev_read_only(struct block_device *bdev);
1959 int set_blocksize(struct block_device *bdev, int size);
1961 const char *bdevname(struct block_device *bdev, char *buffer);
1977 int bd_prepare_to_claim(struct block_device *bdev, void *holder);
1978 void bd_abort_claiming(struct block_device *bdev, void *holder);
1979 void blkdev_put(struct block_device *bdev, fmode_t mode);
1983 void blkdev_put_no_open(struct block_device *bdev);
1986 void bdev_add(struct block_device *bdev, dev_t dev);
1988 int truncate_bdev_range(struct block_device *bdev, fmode_t mode, loff_t lstart,
1992 void invalidate_bdev(struct block_device *bdev);
1993 int sync_blockdev(struct block_device *bdev);
1995 static inline void invalidate_bdev(struct block_device *bdev) in invalidate_bdev() argument
1998 static inline int sync_blockdev(struct block_device *bdev) in sync_blockdev() argument
2003 int fsync_bdev(struct block_device *bdev);
2005 int freeze_bdev(struct block_device *bdev);
2006 int thaw_bdev(struct block_device *bdev);