/Linux-v5.4/include/linux/ |
D | bvec.h | 20 unsigned int bv_len; member 53 __bvec_iter_bvec((bvec), (iter))->bv_len - (iter).bi_bvec_done) 64 .bv_len = mp_bvec_iter_len((bvec), (iter)), \ 83 .bv_len = bvec_iter_len((bvec), (iter)), \ 99 cur->bv_len - iter->bi_bvec_done); in bvec_iter_advance() 105 if (iter->bi_bvec_done == cur->bv_len) { in bvec_iter_advance() 117 bvec_iter_advance((bio_vec), &(iter), (bvl).bv_len)) 148 bv->bv_len = min_t(unsigned int, PAGE_SIZE - bv->bv_offset, in bvec_advance() 149 bvec->bv_len - iter_all->done); in bvec_advance() 150 iter_all->done += bv->bv_len; in bvec_advance() [all …]
|
D | bio.h | 44 ((bio)->bi_iter.bi_size != bio_iovec(bio).bv_len) 92 return bio_iovec(bio).bv_len; in bio_cur_bytes() 157 bio_advance_iter((bio), &(iter), (bvl).bv_len)) 166 bio_advance_iter((bio), &(iter), (bvl).bv_len)) 172 #define bio_iter_last(bvec, iter) ((iter).bi_size == (bvec).bv_len) 276 bv->bv_len = iter.bi_bvec_done; in bio_get_last_bvec()
|
/Linux-v5.4/lib/ |
D | iov_iter.c | 71 if (!__v.bv_len) \ 613 (from += v.bv_len) - v.bv_len, v.bv_len), in _copy_to_iter() 714 (from += v.bv_len) - v.bv_len, v.bv_len); in _copy_to_iter_mcsafe() 748 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter() 749 v.bv_offset, v.bv_len), in _copy_from_iter() 774 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter_full() 775 v.bv_offset, v.bv_len), in _copy_from_iter_full() 794 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter_nocache() 795 v.bv_offset, v.bv_len), in _copy_from_iter_nocache() 828 memcpy_page_flushcache((to += v.bv_len) - v.bv_len, v.bv_page, in _copy_from_iter_flushcache() [all …]
|
/Linux-v5.4/block/ |
D | bounce.c | 81 memcpy(vto + to->bv_offset, vfrom, to->bv_len); in bounce_copy_vec() 88 memcpy(page_address((to)->bv_page) + (to)->bv_offset, vfrom, (to)->bv_len) 158 bio_advance_iter(from, &from_iter, tovec.bv_len); in copy_to_high_bio_irq() 178 bio_advance_iter(bio_orig, &orig_iter, orig_vec.bv_len); in bounce_end_io() 300 sectors += from.bv_len >> 9; in __blk_queue_bounce() 337 memcpy(vto, vfrom, to->bv_len); in __blk_queue_bounce()
|
D | blk-integrity.c | 40 if (seg_size + iv.bv_len > queue_max_segment_size(q)) in blk_rq_count_integrity_sg() 43 seg_size += iv.bv_len; in blk_rq_count_integrity_sg() 47 seg_size = iv.bv_len; in blk_rq_count_integrity_sg() 82 if (sg->length + iv.bv_len > queue_max_segment_size(q)) in blk_rq_map_integrity_sg() 85 sg->length += iv.bv_len; in blk_rq_map_integrity_sg() 95 sg_set_page(sg, iv.bv_page, iv.bv_len, iv.bv_offset); in blk_rq_map_integrity_sg()
|
D | blk-merge.c | 199 unsigned len = min(bv->bv_len, max_len); in bvec_split_segs() 218 return len > 0 || bv->bv_len > max_len; in bvec_split_segs() 260 sectors + (bv.bv_len >> 9) <= max_sectors && in blk_bio_segment_split() 261 bv.bv_offset + bv.bv_len <= PAGE_SIZE) { in blk_bio_segment_split() 263 sectors += bv.bv_len >> 9; in blk_bio_segment_split() 402 unsigned nbytes = bvec->bv_len; in blk_bvec_map_sg() 436 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); in __blk_bvec_map_sg() 446 int nbytes = bvec->bv_len; in __blk_segment_map_sg_merge() 482 if (bvec.bv_offset + bvec.bv_len <= PAGE_SIZE) in __blk_bios_map_sg()
|
D | bio.c | 531 memset(data, 0, bv.bv_len); in zero_fill_bio_iter() 635 bv->bv_offset + bv->bv_len - 1; in page_is_mergeable() 660 if (bv->bv_len + len > queue_max_segment_size(q)) in bio_try_merge_pc_page() 717 bvec->bv_len = len; in __bio_add_pc_page() 758 bv->bv_len += len; in __bio_try_merge_page() 787 bv->bv_len = len; in __bio_add_page() 842 if (WARN_ON_ONCE(iter->iov_offset > bv->bv_len)) in __bio_iov_bvec_add_pages() 845 len = min_t(size_t, bv->bv_len - iter->iov_offset, iter->count); in __bio_iov_bvec_add_pages() 1010 bytes = min(src_bv.bv_len, dst_bv.bv_len); in bio_copy_data_iter() 1124 bvec->bv_len, in bio_copy_from_iter() [all …]
|
D | blk.h | 81 if (addr1 + vec1->bv_len != addr2) in biovec_phys_mergeable() 85 if ((addr1 | mask) != ((addr2 + vec2->bv_len - 1) | mask)) in biovec_phys_mergeable() 94 ((bprv->bv_offset + bprv->bv_len) & queue_virt_boundary(q)); in __bvec_gap_to_prev()
|
D | t10-pi.c | 154 for (j = 0; j < iv.bv_len; j += tuple_sz) { in t10_pi_type1_prepare() 202 for (j = 0; j < iv.bv_len && intervals; j += tuple_sz) { in t10_pi_type1_complete()
|
D | bio-integrity.c | 139 iv->bv_len = len; in bio_integrity_add_page() 174 iter.data_size = bv.bv_len; in bio_integrity_process()
|
/Linux-v5.4/drivers/md/bcache/ |
D | util.c | 244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map() 251 base += bv->bv_len; in bch_bio_map() 254 size -= bv->bv_len; in bch_bio_map()
|
D | debug.c | 138 bv.bv_len), in bch_data_verify() 145 bio_advance_iter(check, &citer, bv.bv_len); in bch_data_verify()
|
/Linux-v5.4/drivers/s390/block/ |
D | dasd_fba.c | 470 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular() 473 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular() 474 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular() 475 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular() 517 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular() 521 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_build_cp_regular() 595 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_free_cp() 606 memcpy(dst, cda, bv.bv_len); in dasd_fba_free_cp()
|
D | dasd_diag.c | 528 if (bv.bv_len & (blksize - 1)) in dasd_diag_build_cp() 531 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_diag_build_cp() 550 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_diag_build_cp()
|
D | dcssblk.c | 899 if (unlikely((page_addr & 4095) != 0) || (bvec.bv_len & 4095) != 0) in dcssblk_make_request() 904 bvec.bv_len); in dcssblk_make_request() 907 bvec.bv_len); in dcssblk_make_request() 909 bytes_done += bvec.bv_len; in dcssblk_make_request()
|
/Linux-v5.4/drivers/block/zram/ |
D | zram_drv.c | 147 return bvec->bv_len != PAGE_SIZE; in is_partial_io() 182 *index += (*offset + bvec->bv_len) / PAGE_SIZE; in update_position() 183 *offset = (*offset + bvec->bv_len) % PAGE_SIZE; in update_position() 600 if (!bio_add_page(bio, bvec->bv_page, bvec->bv_len, bvec->bv_offset)) { in read_from_bdev_async() 661 bvec.bv_len = PAGE_SIZE; in writeback_store() 716 bio_add_page(&bio, bvec.bv_page, bvec.bv_len, in writeback_store() 1228 bvec.bv_len = PAGE_SIZE; in __zram_bvec_read() 1296 memcpy(dst + bvec->bv_offset, src + offset, bvec->bv_len); in zram_bvec_read() 1448 memcpy(dst + offset, src + bvec->bv_offset, bvec->bv_len); in zram_bvec_write() 1453 vec.bv_len = PAGE_SIZE; in zram_bvec_write() [all …]
|
/Linux-v5.4/drivers/xen/ |
D | biomerge.c | 15 return bfn1 + PFN_DOWN(vec1->bv_offset + vec1->bv_len) == bfn2; in xen_biovec_phys_mergeable()
|
/Linux-v5.4/drivers/md/ |
D | dm-log-writes.c | 385 block->vecs[i].bv_len, 0); in log_one_block() 386 if (ret != block->vecs[i].bv_len) { in log_one_block() 402 block->vecs[i].bv_len, 0); in log_one_block() 403 if (ret != block->vecs[i].bv_len) { in log_one_block() 409 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block() 770 memcpy(dst, src + bv.bv_offset, bv.bv_len); in log_writes_map() 774 block->vecs[i].bv_len = bv.bv_len; in log_writes_map()
|
/Linux-v5.4/drivers/block/rsxx/ |
D | dma.c | 677 unsigned int bv_len; in rsxx_dma_queue_bio() local 695 bv_len = bio->bi_iter.bi_size; in rsxx_dma_queue_bio() 697 while (bv_len > 0) { in rsxx_dma_queue_bio() 709 bv_len -= RSXX_HW_BLK_SIZE; in rsxx_dma_queue_bio() 713 bv_len = bvec.bv_len; in rsxx_dma_queue_bio() 716 while (bv_len > 0) { in rsxx_dma_queue_bio() 720 dma_len = min(bv_len, in rsxx_dma_queue_bio() 735 bv_len -= dma_len; in rsxx_dma_queue_bio()
|
/Linux-v5.4/fs/9p/ |
D | vfs_addr.c | 42 struct bio_vec bvec = {.bv_page = page, .bv_len = PAGE_SIZE}; in v9fs_fid_readpage() 164 bvec.bv_len = len; in v9fs_vfs_writepage_locked()
|
/Linux-v5.4/drivers/nvme/target/ |
D | io-cmd-file.c | 87 bv->bv_len = sg->length; in nvmet_file_init_bvec() 157 len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io() 158 total_len += req->f.bvec[bv_cnt].bv_len; in nvmet_file_execute_io()
|
/Linux-v5.4/drivers/block/ |
D | loop.c | 272 iov_iter_bvec(&i, WRITE, bvec, 1, bvec->bv_len); in lo_write_bvec() 278 if (likely(bw == bvec->bv_len)) in lo_write_bvec() 283 (unsigned long long)*ppos, bvec->bv_len); in lo_write_bvec() 325 bvec.bv_offset, bvec.bv_len, pos >> 9); in lo_write_transfer() 331 b.bv_len = bvec.bv_len; in lo_write_transfer() 350 iov_iter_bvec(&i, READ, &bvec, 1, bvec.bv_len); in lo_read_simple() 357 if (len != bvec.bv_len) { in lo_read_simple() 389 b.bv_len = bvec.bv_len; in lo_read_transfer() 391 iov_iter_bvec(&i, READ, &b, 1, b.bv_len); in lo_read_transfer() 405 if (len != bvec.bv_len) { in lo_read_transfer()
|
D | umem.c | 366 vec.bv_len, in add_bio() 385 desc->transfer_size = cpu_to_le32(vec.bv_len); in add_bio() 401 bio_advance_iter(bio, &card->current_iter, vec.bv_len); in add_bio() 440 bio_advance_iter(bio, &page->iter, vec.bv_len); in process_page() 449 vec.bv_len, in process_page()
|
/Linux-v5.4/fs/crypto/ |
D | bio.c | 36 int ret = fscrypt_decrypt_pagecache_blocks(page, bv->bv_len, in __fscrypt_decrypt_bio()
|
/Linux-v5.4/drivers/nvdimm/ |
D | blk.c | 91 cur_len = min(len, bv.bv_len); in nd_blk_rw_integrity() 183 unsigned int len = bvec.bv_len; in nd_blk_make_request()
|