/Linux-v4.19/block/ |
D | blk-zoned.c | 113 struct request_queue *q = bdev_get_queue(bdev); in blkdev_report_zones() 250 struct request_queue *q = bdev_get_queue(bdev); in blkdev_reset_zones() 315 q = bdev_get_queue(bdev); in blkdev_report_zones_ioctl() 376 q = bdev_get_queue(bdev); in blkdev_reset_zones_ioctl()
|
D | blk-lib.c | 30 struct request_queue *q = bdev_get_queue(bdev); in __blkdev_issue_discard() 146 struct request_queue *q = bdev_get_queue(bdev); in __blkdev_issue_write_same() 229 struct request_queue *q = bdev_get_queue(bdev); in __blkdev_issue_write_zeroes() 283 struct request_queue *q = bdev_get_queue(bdev); in __blkdev_issue_zero_pages()
|
D | ioctl.c | 205 struct request_queue *q = bdev_get_queue(bdev); in blk_ioctl_discard() 560 queue_max_sectors(bdev_get_queue(bdev))); in blkdev_ioctl() 563 return put_ushort(arg, !blk_queue_nonrot(bdev_get_queue(bdev))); in blkdev_ioctl()
|
D | compat_ioctl.c | 378 queue_max_sectors(bdev_get_queue(bdev))); in compat_blkdev_ioctl() 382 !blk_queue_nonrot(bdev_get_queue(bdev))); in compat_blkdev_ioctl()
|
D | blk-flush.c | 536 q = bdev_get_queue(bdev); in blkdev_issue_flush()
|
D | blk-settings.c | 688 struct request_queue *bq = bdev_get_queue(bdev); in bdev_stack_limits()
|
/Linux-v4.19/drivers/md/ |
D | dm-table.c | 299 q = bdev_get_queue(bdev); in device_area_is_invalid() 475 struct request_queue *q = bdev_get_queue(bdev); in dm_set_device_limits() 921 struct request_queue *q = bdev_get_queue(dev->bdev); in device_is_rq_based() 1428 struct request_queue *q = bdev_get_queue(dev->bdev); in device_is_zoned_model() 1458 struct request_queue *q = bdev_get_queue(dev->bdev); in device_matches_zone_sectors() 1641 struct request_queue *q = bdev_get_queue(dev->bdev); in device_flush_capable() 1708 struct request_queue *q = bdev_get_queue(dev->bdev); in device_is_nonrot() 1716 struct request_queue *q = bdev_get_queue(dev->bdev); in device_is_not_random() 1724 struct request_queue *q = bdev_get_queue(dev->bdev); in queue_supports_sg_merge() 1763 struct request_queue *q = bdev_get_queue(dev->bdev); in device_not_write_same_capable() [all …]
|
D | md-linear.c | 71 struct request_queue *q = bdev_get_queue(conf->disks[i].rdev->bdev); in linear_congested() 131 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in linear_conf()
|
D | raid0.c | 44 struct request_queue *q = bdev_get_queue(devlist[i]->bdev); in raid0_congested() 400 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid0_run() 547 trace_block_bio_remap(bdev_get_queue(rdev->bdev), in raid0_handle_discard()
|
D | dm-mpath.c | 522 q = bdev_get_queue(bdev); in multipath_clone_and_map() 811 struct request_queue *q = bdev_get_queue(bdev); in setup_scsi_dh() 890 q = bdev_get_queue(p->path.dev->bdev); in parse_path() 1532 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in activate_or_offline_path() 1949 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in pgpath_busy()
|
D | md-multipath.c | 172 struct request_queue *q = bdev_get_queue(rdev->bdev); in multipath_congested()
|
D | dm-io.c | 306 struct request_queue *q = bdev_get_queue(where->bdev); in do_region()
|
D | raid1.c | 667 nonrot = blk_queue_nonrot(bdev_get_queue(rdev->bdev)); in read_balance() 764 struct request_queue *q = bdev_get_queue(rdev->bdev); in raid1_congested() 1771 if (mddev->queue && blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid1_add_disk() 3087 if (blk_queue_discard(bdev_get_queue(rdev->bdev))) in raid1_run()
|
/Linux-v4.19/include/linux/ |
D | blkdev.h | 1015 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue() function 1484 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size() 1494 return queue_physical_block_size(bdev_get_queue(bdev)); in bdev_physical_block_size() 1504 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min() 1514 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt() 1536 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset() 1580 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment() 1590 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_same() 1600 struct request_queue *q = bdev_get_queue(bdev); in bdev_write_zeroes_sectors() 1610 struct request_queue *q = bdev_get_queue(bdev); in bdev_zoned_model() [all …]
|
/Linux-v4.19/drivers/target/ |
D | target_core_iblock.c | 119 q = bdev_get_queue(bd); in iblock_configure_device() 698 struct request_queue *q = bdev_get_queue(ib_dev->ibd_bd); in iblock_execute_rw() 787 struct request_queue *q = bdev_get_queue(bd); in iblock_get_blocks() 848 struct request_queue *q = bdev_get_queue(bd); in iblock_get_write_cache()
|
/Linux-v4.19/fs/jfs/ |
D | ioctl.c | 133 struct request_queue *q = bdev_get_queue(sb->s_bdev); in jfs_ioctl()
|
D | super.c | 406 struct request_queue *q = bdev_get_queue(sb->s_bdev); in parse_options() 425 struct request_queue *q = bdev_get_queue(sb->s_bdev); in parse_options()
|
/Linux-v4.19/fs/xfs/ |
D | xfs_discard.c | 152 struct request_queue *q = bdev_get_queue(mp->m_ddev_targp->bt_bdev); in xfs_ioc_trim()
|
/Linux-v4.19/drivers/md/bcache/ |
D | request.c | 999 !blk_queue_discard(bdev_get_queue(dc->bdev))) in cached_dev_write() 1105 !blk_queue_discard(bdev_get_queue(dc->bdev))) in detached_dev_do_request() 1228 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_congested() 1239 q = bdev_get_queue(ca->bdev); in cached_dev_congested() 1351 q = bdev_get_queue(ca->bdev); in flash_dev_congested()
|
D | super.c | 879 q = bdev_get_queue(dc->bdev); in cached_dev_status_update() 1232 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_init() 2107 if (blk_queue_discard(bdev_get_queue(bdev))) in register_cache()
|
/Linux-v4.19/drivers/nvme/target/ |
D | io-cmd-bdev.c | 111 blk_poll(bdev_get_queue(req->ns->bdev), cookie); in nvmet_bdev_execute_rw()
|
/Linux-v4.19/drivers/block/xen-blkback/ |
D | xenbus.c | 489 q = bdev_get_queue(bdev); in xen_vbd_create() 548 struct request_queue *q = bdev_get_queue(bdev); in xen_blkbk_discard()
|
/Linux-v4.19/drivers/dax/ |
D | super.c | 102 q = bdev_get_queue(bdev); in __bdev_dax_supported()
|
/Linux-v4.19/fs/fat/ |
D | file.c | 129 struct request_queue *q = bdev_get_queue(sb->s_bdev); in fat_ioctl_fitrim()
|
/Linux-v4.19/mm/ |
D | swapfile.c | 2609 if (!p->bdev || !blk_queue_nonrot(bdev_get_queue(p->bdev))) in SYSCALL_DEFINE1() 3087 struct request_queue *q = bdev_get_queue(si->bdev); in swap_discardable() 3184 if (p->bdev && blk_queue_nonrot(bdev_get_queue(p->bdev))) { in SYSCALL_DEFINE2() 3753 blkcg_schedule_throttle(bdev_get_queue(si->bdev), in mem_cgroup_throttle_swaprate()
|