/Linux-v5.10/drivers/md/bcache/ |
D | writeback.h | 76 unsigned int nr_sectors) in bcache_dev_stripe_dirty() argument 87 if (nr_sectors <= dc->disk.stripe_size) in bcache_dev_stripe_dirty() 90 nr_sectors -= dc->disk.stripe_size; in bcache_dev_stripe_dirty() 145 uint64_t offset, int nr_sectors);
|
D | writeback.c | 521 uint64_t offset, int nr_sectors) in bcache_dev_sectors_dirty_add() argument 535 atomic_long_add(nr_sectors, &c->flash_dev_dirty_sectors); in bcache_dev_sectors_dirty_add() 539 while (nr_sectors) { in bcache_dev_sectors_dirty_add() 540 int s = min_t(unsigned int, abs(nr_sectors), in bcache_dev_sectors_dirty_add() 543 if (nr_sectors < 0) in bcache_dev_sectors_dirty_add() 556 nr_sectors -= s; in bcache_dev_sectors_dirty_add()
|
/Linux-v5.10/drivers/block/ |
D | null_blk_zoned.c | 207 unsigned int nr_sectors = len >> SECTOR_SHIFT; in null_zone_valid_read_len() local 211 sector + nr_sectors <= zone->wp) in null_zone_valid_read_len() 326 unsigned int nr_sectors, bool append) in null_zone_write() argument 336 return null_process_cmd(cmd, REQ_OP_WRITE, sector, nr_sectors); in null_zone_write() 377 if (zone->wp + nr_sectors > zone->start + zone->capacity) { in null_zone_write() 398 ret = null_process_cmd(cmd, REQ_OP_WRITE, sector, nr_sectors); in null_zone_write() 405 zone->wp += nr_sectors; in null_zone_write() 579 sector_t sector, sector_t nr_sectors) in null_process_zoned_cmd() argument 587 sts = null_zone_write(cmd, sector, nr_sectors, false); in null_process_zoned_cmd() 590 sts = null_zone_write(cmd, sector, nr_sectors, true); in null_process_zoned_cmd() [all …]
|
D | null_blk.h | 98 unsigned int nr_sectors); 108 sector_t nr_sectors); 124 enum req_opf op, sector_t sector, sector_t nr_sectors) in null_process_zoned_cmd() argument
|
D | null_blk_main.c | 1253 sector_t nr_sectors) in null_handle_badblocks() argument 1259 if (badblocks_check(bb, sector, nr_sectors, &first_bad, &bad_sectors)) in null_handle_badblocks() 1334 unsigned int nr_sectors) in null_process_cmd() argument 1340 ret = null_handle_badblocks(cmd, sector, nr_sectors); in null_process_cmd() 1352 sector_t nr_sectors, enum req_opf op) in null_handle_cmd() argument 1371 sector, nr_sectors); in null_handle_cmd() 1373 cmd->error = null_process_cmd(cmd, op, sector, nr_sectors); in null_handle_cmd() 1420 sector_t nr_sectors = bio_sectors(bio); in null_submit_bio() local 1428 null_handle_cmd(cmd, sector, nr_sectors, bio_op(bio)); in null_submit_bio() 1462 sector_t nr_sectors = blk_rq_sectors(bd->rq); in null_queue_rq() local [all …]
|
D | floppy.c | 2260 unsigned int nr_sectors = current_count_sectors; in floppy_end_request() local 2265 nr_sectors = blk_rq_cur_sectors(req); in floppy_end_request() 2266 if (blk_update_request(req, error, nr_sectors << 9)) in floppy_end_request() 2322 int nr_sectors; in rw_interrupt() local 2334 nr_sectors = 0; in rw_interrupt() 2347 nr_sectors = (((reply_buffer[R_TRACK] - raw_cmd->cmd[TRACK]) * heads + in rw_interrupt() 2351 if (nr_sectors / ssize > in rw_interrupt() 2354 nr_sectors, current_count_sectors); in rw_interrupt() 2367 nr_sectors -= in_sector_offset; in rw_interrupt() 2368 INFBOUND(nr_sectors, 0); in rw_interrupt() [all …]
|
/Linux-v5.10/block/ |
D | blk-zoned.c | 174 sector_t nr_sectors) in blkdev_allow_reset_all_zones() argument 183 return !sector && nr_sectors == get_capacity(bdev->bd_disk); in blkdev_allow_reset_all_zones() 203 sector_t sector, sector_t nr_sectors, in blkdev_zone_mgmt() argument 209 sector_t end_sector = sector + nr_sectors; in blkdev_zone_mgmt() 230 if ((nr_sectors & (zone_sectors - 1)) && end_sector != capacity) in blkdev_zone_mgmt() 242 blkdev_allow_reset_all_zones(bdev, sector, nr_sectors)) { in blkdev_zone_mgmt() 369 return blkdev_zone_mgmt(bdev, op, zrange.sector, zrange.nr_sectors, in blkdev_zone_mgmt_ioctl()
|
D | blk-core.c | 729 unsigned int nr_sectors = bio_sectors(bio); in bio_check_eod() local 731 if (nr_sectors && maxsector && in bio_check_eod() 732 (nr_sectors > maxsector || in bio_check_eod() 733 bio->bi_iter.bi_sector > maxsector - nr_sectors)) { in bio_check_eod() 778 int nr_sectors = bio_sectors(bio); in blk_check_zone_append() local 794 if (nr_sectors > q->limits.chunk_sectors) in blk_check_zone_append() 798 if (nr_sectors > q->limits.max_zone_append_sectors) in blk_check_zone_append()
|
D | blk-merge.c | 368 unsigned int nr_sectors = 0; in blk_recalc_rq_segments() local 385 bvec_split_segs(rq->q, &bv, &nr_phys_segs, &nr_sectors, in blk_recalc_rq_segments()
|
/Linux-v5.10/drivers/block/xen-blkback/ |
D | common.h | 93 uint64_t nr_sectors; member 147 uint64_t nr_sectors; member 420 dst->u.discard.nr_sectors = src->u.discard.nr_sectors; in blkif_get_x86_32_req() 468 dst->u.discard.nr_sectors = src->u.discard.nr_sectors; in blkif_get_x86_64_req()
|
D | blkback.c | 975 preq.nr_sects = req->u.discard.nr_sectors; in dispatch_discard_io() 991 req->u.discard.nr_sectors, in dispatch_discard_io()
|
/Linux-v5.10/Documentation/block/ |
D | request.rst | 67 unsigned long nr_sectors DBI Total number of sectors in request 69 unsigned long hard_nr_sectors B Used to keep nr_sectors sane
|
/Linux-v5.10/drivers/md/ |
D | dm-log-writes.c | 97 __le64 nr_sectors; member 126 sector_t nr_sectors; member 340 entry.nr_sectors = cpu_to_le64(block->nr_sectors); in log_one_block() 473 lc->next_sector += dev_to_bio_sectors(lc, block->nr_sectors); in log_writes_kthread() 726 block->nr_sectors = bio_to_dev_sectors(lc, bio_sectors(bio)); in log_writes_map() 937 block->nr_sectors = ALIGN(bytes, lc->sectorsize) >> lc->sectorshift; in log_dax()
|
D | md.h | 552 static inline void md_sync_acct(struct block_device *bdev, unsigned long nr_sectors) in md_sync_acct() argument 554 atomic_add(nr_sectors, &bdev->bd_disk->sync_io); in md_sync_acct() 557 static inline void md_sync_acct_bio(struct bio *bio, unsigned long nr_sectors) in md_sync_acct_bio() argument 559 atomic_add(nr_sectors, &bio->bi_disk->sync_io); in md_sync_acct_bio()
|
D | dm-zoned-target.c | 632 unsigned int nr_sectors = bio_sectors(bio); in dmz_map() local 641 bio_op(bio), (unsigned long long)sector, nr_sectors, in dmz_map() 646 if (!nr_sectors && bio_op(bio) != REQ_OP_WRITE) in dmz_map() 650 if ((nr_sectors & DMZ_BLOCK_SECTORS_MASK) || (sector & DMZ_BLOCK_SECTORS_MASK)) in dmz_map() 660 if (!nr_sectors && bio_op(bio) == REQ_OP_WRITE) { in dmz_map() 670 if (chunk_sector + nr_sectors > dmz_zone_nr_sectors(zmd)) in dmz_map()
|
D | raid10.c | 2890 sector_t max_sector, nr_sectors; in raid10_sync_request() local 3401 nr_sectors = 0; in raid10_sync_request() 3420 nr_sectors += len>>9; in raid10_sync_request() 3423 r10_bio->sectors = nr_sectors; in raid10_sync_request() 3428 if (conf->cluster_sync_high < sector_nr + nr_sectors) { in raid10_sync_request() 3449 if (conf->cluster_sync_high < sect_va1 + nr_sectors) { in raid10_sync_request() 3477 r10_bio->sectors = nr_sectors; in raid10_sync_request() 3480 md_sync_acct_bio(bio, nr_sectors); in raid10_sync_request() 3492 return sectors_skipped + nr_sectors; in raid10_sync_request() 4412 int nr_sectors; in reshape_request() local [all …]
|
D | raid1.c | 2620 sector_t max_sector, nr_sectors; in raid1_sync_request() local 2836 nr_sectors = 0; in raid1_sync_request() 2870 nr_sectors += len>>9; in raid1_sync_request() 2875 r1_bio->sectors = nr_sectors; in raid1_sync_request() 2878 conf->cluster_sync_high < sector_nr + nr_sectors) { in raid1_sync_request() 2896 md_sync_acct_bio(bio, nr_sectors); in raid1_sync_request() 2905 md_sync_acct_bio(bio, nr_sectors); in raid1_sync_request() 2910 return nr_sectors; in raid1_sync_request()
|
/Linux-v5.10/include/uapi/linux/ |
D | blkzoned.h | 145 __u64 nr_sectors; member
|
/Linux-v5.10/drivers/block/drbd/ |
D | drbd_actlog.c | 852 sector_t esector, nr_sectors; in __drbd_change_sync() local 868 nr_sectors = get_capacity(device->vdisk); in __drbd_change_sync() 871 if (!expect(sector < nr_sectors)) in __drbd_change_sync() 873 if (!expect(esector < nr_sectors)) in __drbd_change_sync() 874 esector = nr_sectors - 1; in __drbd_change_sync() 876 lbnr = BM_SECT_TO_BIT(nr_sectors-1); in __drbd_change_sync() 883 if (unlikely(esector == (nr_sectors-1))) in __drbd_change_sync()
|
D | drbd_req.c | 884 sector_t esector, nr_sectors; in drbd_may_do_local_read() local 891 nr_sectors = get_capacity(device->vdisk); in drbd_may_do_local_read() 892 D_ASSERT(device, sector < nr_sectors); in drbd_may_do_local_read() 893 D_ASSERT(device, esector < nr_sectors); in drbd_may_do_local_read()
|
D | drbd_receiver.c | 1512 …discard_or_zero_out(struct drbd_device *device, sector_t start, unsigned int nr_sectors, int flags) in drbd_issue_discard_or_zero_out() argument 1533 if (nr_sectors < granularity) in drbd_issue_discard_or_zero_out() 1538 if (nr_sectors < 2*granularity) in drbd_issue_discard_or_zero_out() 1548 nr_sectors -= nr; in drbd_issue_discard_or_zero_out() 1551 while (nr_sectors >= max_discard_sectors) { in drbd_issue_discard_or_zero_out() 1553 nr_sectors -= max_discard_sectors; in drbd_issue_discard_or_zero_out() 1556 if (nr_sectors) { in drbd_issue_discard_or_zero_out() 1561 nr = nr_sectors; in drbd_issue_discard_or_zero_out() 1565 nr_sectors -= nr; in drbd_issue_discard_or_zero_out() 1570 if (nr_sectors) { in drbd_issue_discard_or_zero_out() [all …]
|
/Linux-v5.10/include/xen/interface/io/ |
D | blkif.h | 214 uint64_t nr_sectors; member
|
/Linux-v5.10/fs/fat/ |
D | inode.c | 68 unsigned nr_sectors; member 75 .nr_sectors = 160 * KB_IN_SECTORS, 82 .nr_sectors = 180 * KB_IN_SECTORS, 89 .nr_sectors = 320 * KB_IN_SECTORS, 96 .nr_sectors = 360 * KB_IN_SECTORS, 1566 if (floppy_defaults[i].nr_sectors == bd_sects) { in fat_read_static_bpb() 1590 bpb->fat_sectors = fdefaults->nr_sectors; in fat_read_static_bpb()
|
/Linux-v5.10/fs/btrfs/ |
D | file-item.c | 536 int nr_sectors; in btrfs_csum_one_bio() local 574 nr_sectors = BTRFS_BYTES_TO_BLKS(fs_info, in btrfs_csum_one_bio() 578 for (i = 0; i < nr_sectors; i++) { in btrfs_csum_one_bio()
|
D | compression.c | 711 unsigned int nr_sectors; in btrfs_submit_compressed_read() local 731 nr_sectors = DIV_ROUND_UP(comp_bio->bi_iter.bi_size, in btrfs_submit_compressed_read() 733 sums += csum_size * nr_sectors; in btrfs_submit_compressed_read()
|