Home
last modified time | relevance | path

Searched refs:bio_sectors (Results 1 – 25 of 42) sorted by relevance

12

/Linux-v5.4/drivers/md/bcache/
Drequest.c126 bio_sectors(bio), (uint64_t) bio->bi_iter.bi_sector); in bch_data_invalidate()
128 while (bio_sectors(bio)) { in bch_data_invalidate()
129 unsigned int sectors = min(bio_sectors(bio), in bch_data_invalidate()
207 if (atomic_sub_return(bio_sectors(bio), &op->c->sectors_to_gc) < 0) in bch_data_insert_start()
234 if (!bch_alloc_sectors(op->c, k, bio_sectors(bio), in bch_data_insert_start()
403 bio_sectors(bio) & (c->sb.block_size - 1)) { in check_should_bypass()
459 bch_rescale_priorities(c, bio_sectors(bio)); in check_should_bypass()
462 bch_mark_sectors_bypassed(c, dc, bio_sectors(bio)); in check_should_bypass()
529 unsigned int bio_sectors = bio_sectors(bio); in cache_lookup_fn() local
540 BUG_ON(bio_sectors <= sectors); in cache_lookup_fn()
[all …]
Dwriteback.h79 bio_sectors(bio))) in should_writeback()
/Linux-v5.4/block/
Dblk-merge.c85 if (bio_sectors(bio) <= max_discard_sectors) in blk_bio_discard_split()
113 if (bio_sectors(bio) <= q->limits.max_write_zeroes_sectors) in blk_bio_write_zeroes_split()
129 if (bio_sectors(bio) <= q->limits.max_write_same_sectors) in blk_bio_write_same_split()
579 if (blk_rq_sectors(req) + bio_sectors(bio) > in ll_back_merge_fn()
595 if (blk_rq_sectors(req) + bio_sectors(bio) > in ll_front_merge_fn()
611 if (blk_rq_sectors(req) + bio_sectors(next->bio) > in req_attempt_discard_merge()
892 else if (blk_rq_pos(rq) - bio_sectors(bio) == bio->bi_iter.bi_sector) in blk_try_merge()
Dbio-integrity.c217 if (!bio_sectors(bio)) in bio_integrity_prep()
233 intervals = bio_integrity_intervals(bi, bio_sectors(bio)); in bio_integrity_prep()
394 bip->bip_iter.bi_size = bio_integrity_bytes(bi, bio_sectors(bio)); in bio_integrity_trim()
Dblk-core.c652 if (blk_rq_sectors(req) + bio_sectors(bio) > in bio_attempt_discard_merge()
794 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) in bio_check_ro()
823 unsigned int nr_sectors = bio_sectors(bio); in bio_check_eod()
855 if (bio_sectors(bio) || bio_op(bio) == REQ_OP_ZONE_RESET) { in blk_partition_remap()
873 int nr_sectors = bio_sectors(bio); in generic_make_request_checks()
1160 count = bio_sectors(bio); in submit_bio()
Dbounce.c307 if (!passthrough && sectors < bio_sectors(*bio_orig)) { in __blk_queue_bounce()
/Linux-v5.4/include/trace/events/
Dblock.h242 __entry->nr_sector = bio_sectors(bio);
279 __entry->nr_sector = bio_sectors(bio);
307 __entry->nr_sector = bio_sectors(bio);
374 __entry->nr_sector = bio_sectors(bio);
402 __entry->nr_sector = bio ? bio_sectors(bio) : 0;
578 __entry->nr_sector = bio_sectors(bio);
/Linux-v5.4/drivers/lightnvm/
Dpblk-cache.c32 generic_start_io_acct(q, REQ_OP_WRITE, bio_sectors(bio), in pblk_write_to_cache()
Dpblk-read.c275 generic_start_io_acct(q, REQ_OP_READ, bio_sectors(bio), in pblk_submit_read()
/Linux-v5.4/drivers/md/
Draid0.c469 + bio_sectors(bio)); in is_io_in_chunk_boundary()
473 + bio_sectors(bio)); in is_io_in_chunk_boundary()
600 if (sectors < bio_sectors(bio)) { in raid0_make_request()
Ddm-zoned.h46 #define dmz_bio_blocks(bio) dmz_sect2blk(bio_sectors(bio))
Ddm-linear.c93 if (bio_sectors(bio) || bio_op(bio) == REQ_OP_ZONE_RESET) in linear_map_bio()
Ddm-log-writes.c692 if (!bio_sectors(bio) && !flush_bio) in log_writes_map()
726 block->nr_sectors = bio_to_dev_sectors(lc, bio_sectors(bio)); in log_writes_map()
738 if (flush_bio && !bio_sectors(bio)) { in log_writes_map()
Ddm.c658 generic_start_io_acct(md->queue, bio_op(bio), bio_sectors(bio), in start_io_acct()
663 bio->bi_iter.bi_sector, bio_sectors(bio), in start_io_acct()
678 bio->bi_iter.bi_sector, bio_sectors(bio), in end_io_acct()
1636 ci.sector_count = bio_sectors(bio); in __split_and_process_bio()
1648 struct bio *b = bio_split(bio, bio_sectors(bio) - ci.sector_count, in __split_and_process_bio()
1708 ci.sector_count = bio_sectors(bio); in __process_bio()
1725 sector_count = bio_sectors(*bio); in dm_queue_split()
Ddm-crypt.c819 if (!bio_sectors(bio) || !io->cc->on_disk_tag_size) in dm_crypt_integrity_io_alloc()
826 tag_len = io->cc->on_disk_tag_size * (bio_sectors(bio) >> io->cc->sector_shift); in dm_crypt_integrity_io_alloc()
1623 sector += bio_sectors(clone); in kcryptd_crypt_write_convert()
2758 if (bio_sectors(bio)) in crypt_map()
2785 unsigned tag_len = cc->on_disk_tag_size * (bio_sectors(bio) >> cc->sector_shift); in crypt_map()
2790 if (bio_sectors(bio) > cc->tag_pool_max_sectors) in crypt_map()
Ddm-delay.c298 if (bio_sectors(bio)) in delay_map()
Ddm-flakey.c283 if (bio_sectors(bio) || bio_op(bio) == REQ_OP_ZONE_RESET) in flakey_map_bio()
Draid1.c1193 r1_bio->sectors = bio_sectors(bio); in init_r1bio()
1293 if (max_sectors < bio_sectors(bio)) { in raid1_read_request()
1464 if (max_sectors < bio_sectors(bio)) { in raid1_write_request()
1583 bio->bi_iter.bi_sector, bio_sectors(bio)); in raid1_make_request()
2229 md_sync_acct(conf->mirrors[i].rdev->bdev, bio_sectors(wbio)); in sync_request_write()
/Linux-v5.4/include/linux/
Dbio.h49 #define bio_sectors(bio) bvec_iter_sectors((bio)->bi_iter) macro
375 if (sectors >= bio_sectors(bio)) in bio_next_split()
Dblk-cgroup.h723 bio_issue_init(&bio->bi_issue, bio_sectors(bio)); in blkcg_bio_issue_init()
/Linux-v5.4/drivers/block/rsxx/
Ddev.c101 generic_start_io_acct(card->queue, bio_op(bio), bio_sectors(bio), in disk_stats_start()
/Linux-v5.4/drivers/ide/
Dide-cd.c150 unsigned long bio_sectors; in cdrom_analyze_sense_data() local
191 bio_sectors = max(bio_sectors(failed_command->bio), 4U); in cdrom_analyze_sense_data()
192 sector &= ~(bio_sectors - 1); in cdrom_analyze_sense_data()
/Linux-v5.4/drivers/nvdimm/
Dnd.h408 generic_start_io_acct(disk->queue, bio_op(bio), bio_sectors(bio), in nd_iostat_start()
/Linux-v5.4/fs/ext4/
Dpage-io.c298 (unsigned) bio_sectors(bio), in ext4_end_bio()
/Linux-v5.4/drivers/block/
Dps3disk.c94 __func__, __LINE__, i, bio_sectors(iter.bio), in ps3disk_scatter_gather()

12