Lines Matching refs:bv
105 struct bio_vec bv, bvprv, *bvprvp = NULL; in blk_bio_segment_split() local
113 bio_for_each_segment(bv, bio, iter) { in blk_bio_segment_split()
118 if (bvprvp && bvec_gap_to_prev(q, bvprvp, bv.bv_offset)) in blk_bio_segment_split()
121 if (sectors + (bv.bv_len >> 9) > max_sectors) { in blk_bio_segment_split()
135 if (seg_size + bv.bv_len > queue_max_segment_size(q)) in blk_bio_segment_split()
137 if (!BIOVEC_PHYS_MERGEABLE(bvprvp, &bv)) in blk_bio_segment_split()
139 if (!BIOVEC_SEG_BOUNDARY(q, bvprvp, &bv)) in blk_bio_segment_split()
142 seg_size += bv.bv_len; in blk_bio_segment_split()
143 bvprv = bv; in blk_bio_segment_split()
145 sectors += bv.bv_len >> 9; in blk_bio_segment_split()
157 bvprv = bv; in blk_bio_segment_split()
159 seg_size = bv.bv_len; in blk_bio_segment_split()
160 sectors += bv.bv_len >> 9; in blk_bio_segment_split()
235 struct bio_vec bv, bvprv = { NULL }; in __blk_recalc_rq_segments() local
258 bio_for_each_segment(bv, bio, iter) { in __blk_recalc_rq_segments()
267 if (seg_size + bv.bv_len in __blk_recalc_rq_segments()
270 if (!BIOVEC_PHYS_MERGEABLE(&bvprv, &bv)) in __blk_recalc_rq_segments()
272 if (!BIOVEC_SEG_BOUNDARY(q, &bvprv, &bv)) in __blk_recalc_rq_segments()
275 seg_size += bv.bv_len; in __blk_recalc_rq_segments()
276 bvprv = bv; in __blk_recalc_rq_segments()
285 bvprv = bv; in __blk_recalc_rq_segments()
287 seg_size = bv.bv_len; in __blk_recalc_rq_segments()
408 static inline int __blk_bvec_map_sg(struct request_queue *q, struct bio_vec bv, in __blk_bvec_map_sg() argument
412 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); in __blk_bvec_map_sg()