Home
last modified time | relevance | path

Searched refs:nr_segs (Results 1 – 25 of 38) sorted by relevance

12

/Linux-v5.15/include/linux/
Duio.h33 unsigned long nr_segs; member
49 unsigned long nr_segs; member
68 state->nr_segs = iter->nr_segs; in iov_iter_save_state()
113 static inline size_t iov_length(const struct iovec *iov, unsigned long nr_segs) in iov_length() argument
118 for (seg = 0; seg < nr_segs; seg++) in iov_length()
234 unsigned long nr_segs, size_t count);
236 unsigned long nr_segs, size_t count);
238 unsigned long nr_segs, size_t count);
307 unsigned long nr_segs, unsigned long fast_segs,
310 unsigned nr_segs, unsigned fast_segs, struct iovec **iovp,
[all …]
Dblk-mq.h625 unsigned int nr_segs) in blk_rq_bio_prep() argument
627 rq->nr_phys_segments = nr_segs; in blk_rq_bio_prep()
Dbio.h24 static inline unsigned int bio_max_segs(unsigned int nr_segs) in bio_max_segs() argument
26 return min(nr_segs, BIO_MAX_VECS); in bio_max_segs()
/Linux-v5.15/block/
Dblk-merge.c305 void __blk_queue_split(struct bio **bio, unsigned int *nr_segs) in __blk_queue_split() argument
313 split = blk_bio_discard_split(q, *bio, &q->bio_split, nr_segs); in __blk_queue_split()
317 nr_segs); in __blk_queue_split()
321 nr_segs); in __blk_queue_split()
336 *nr_segs = 1; in __blk_queue_split()
339 split = blk_bio_segment_split(q, *bio, &q->bio_split, nr_segs); in __blk_queue_split()
368 unsigned int nr_segs; in blk_queue_split() local
370 __blk_queue_split(bio, &nr_segs); in blk_queue_split()
586 int ll_back_merge_fn(struct request *req, struct bio *bio, unsigned int nr_segs) in ll_back_merge_fn() argument
601 return ll_new_hw_segment(req, bio, nr_segs); in ll_back_merge_fn()
[all …]
Dblk-mq-sched.h13 unsigned int nr_segs, struct request **merged_request);
15 unsigned int nr_segs);
35 unsigned int nr_segs) in blk_mq_sched_bio_merge() argument
40 return __blk_mq_sched_bio_merge(q, bio, nr_segs); in blk_mq_sched_bio_merge()
Dblk-map.c26 if (data->nr_segs > UIO_MAXIOV) in bio_alloc_map_data()
29 bmd = kmalloc(struct_size(bmd, iov, data->nr_segs), gfp_mask); in bio_alloc_map_data()
32 memcpy(bmd->iov, data->iov, sizeof(struct iovec) * data->nr_segs); in bio_alloc_map_data()
486 unsigned int nr_segs = 0; in blk_rq_append_bio() local
489 nr_segs++; in blk_rq_append_bio()
492 blk_rq_bio_prep(rq, bio, nr_segs); in blk_rq_append_bio()
494 if (!ll_back_merge_fn(rq, bio, nr_segs)) in blk_rq_append_bio()
Dblk.h180 unsigned int nr_segs, struct request **same_queue_rq);
182 struct bio *bio, unsigned int nr_segs);
223 void __blk_queue_split(struct bio **bio, unsigned int *nr_segs);
225 unsigned int nr_segs);
Dblk-mq-sched.c367 unsigned int nr_segs) in __blk_mq_sched_bio_merge() argument
376 return e->type->ops.bio_merge(q, bio, nr_segs); in __blk_mq_sched_bio_merge()
392 if (blk_bio_list_merge(q, &ctx->rq_lists[type], bio, nr_segs)) { in __blk_mq_sched_bio_merge()
Dblk-mq.c1953 unsigned int nr_segs) in blk_mq_bio_to_request() argument
1962 blk_rq_bio_prep(rq, bio, nr_segs); in blk_mq_bio_to_request()
2188 unsigned int nr_segs; in blk_mq_submit_bio() local
2194 __blk_queue_split(&bio, &nr_segs); in blk_mq_submit_bio()
2200 blk_attempt_plug_merge(q, bio, nr_segs, &same_queue_rq)) in blk_mq_submit_bio()
2203 if (blk_mq_sched_bio_merge(q, bio, nr_segs)) in blk_mq_submit_bio()
2225 blk_mq_bio_to_request(rq, bio, nr_segs); in blk_mq_submit_bio()
Dkyber-iosched.c567 unsigned int nr_segs) in kyber_bio_merge() argument
578 merged = blk_bio_list_merge(hctx->queue, rq_list, bio, nr_segs); in kyber_bio_merge()
Dmq-deadline.c640 unsigned int nr_segs) in dd_bio_merge() argument
647 ret = blk_mq_sched_try_merge(q, bio, nr_segs, &free); in dd_bio_merge()
/Linux-v5.15/lib/
Diov_iter.c121 i->nr_segs -= iov - i->iov; \
129 i->nr_segs -= bvec - i->bvec; \
137 i->nr_segs -= kvec - i->kvec; \
251 i->nr_segs -= iov - i->iov; in copy_page_to_iter_iovec()
335 i->nr_segs -= iov - i->iov; in copy_page_from_iter_iovec()
464 const struct iovec *iov, unsigned long nr_segs, in iov_iter_init() argument
472 .nr_segs = nr_segs, in iov_iter_init()
993 i->nr_segs -= bi.bi_idx; in iov_iter_bvec_advance()
1007 for (iov = i->iov, end = iov + i->nr_segs; iov < end; iov++) { in iov_iter_iovec_advance()
1013 i->nr_segs -= iov - i->iov; in iov_iter_iovec_advance()
[all …]
/Linux-v5.15/fs/ksmbd/
Dtransport_tcp.c115 unsigned int nr_segs, size_t bytes) in kvec_array_init() argument
126 nr_segs--; in kvec_array_init()
131 memcpy(new, iov, sizeof(*iov) * nr_segs); in kvec_array_init()
134 return nr_segs; in kvec_array_init()
144 static struct kvec *get_conn_iovec(struct tcp_transport *t, unsigned int nr_segs) in get_conn_iovec() argument
148 if (t->iov && nr_segs <= t->nr_iov) in get_conn_iovec()
152 new_iov = kmalloc_array(nr_segs, sizeof(*new_iov), GFP_KERNEL); in get_conn_iovec()
156 t->nr_iov = nr_segs; in get_conn_iovec()
290 unsigned int nr_segs, unsigned int to_read) in ksmbd_tcp_readv() argument
299 iov = get_conn_iovec(t, nr_segs); in ksmbd_tcp_readv()
[all …]
/Linux-v5.15/arch/powerpc/mm/
Ddma-noncoherent.c65 int nr_segs = 1 + ((size - seg_size) + PAGE_SIZE - 1)/PAGE_SIZE; in __dma_sync_page_highmem() local
84 } while (seg_nr < nr_segs); in __dma_sync_page_highmem()
/Linux-v5.15/drivers/scsi/
Dxen-scsifront.c1064 unsigned int sg_grant, nr_segs; in scsifront_read_backend_params() local
1068 nr_segs = min_t(unsigned int, sg_grant, SG_ALL); in scsifront_read_backend_params()
1069 nr_segs = max_t(unsigned int, nr_segs, VSCSIIF_SG_TABLESIZE); in scsifront_read_backend_params()
1070 nr_segs = min_t(unsigned int, nr_segs, in scsifront_read_backend_params()
1075 dev_info(&dev->dev, "using up to %d SG entries\n", nr_segs); in scsifront_read_backend_params()
1076 else if (info->pause && nr_segs < host->sg_tablesize) in scsifront_read_backend_params()
1079 host->sg_tablesize, nr_segs); in scsifront_read_backend_params()
1081 host->sg_tablesize = nr_segs; in scsifront_read_backend_params()
1082 host->max_sectors = (nr_segs - 1) * PAGE_SIZE / 512; in scsifront_read_backend_params()
Dscsi_lib.c984 unsigned short nr_segs = blk_rq_nr_phys_segments(rq); in scsi_alloc_sgtables() local
990 if (WARN_ON_ONCE(!nr_segs)) in scsi_alloc_sgtables()
998 nr_segs++; in scsi_alloc_sgtables()
1003 if (unlikely(sg_alloc_table_chained(&cmd->sdb.table, nr_segs, in scsi_alloc_sgtables()
/Linux-v5.15/drivers/hwtracing/intel_th/
Dmsu.c73 unsigned int nr_segs; member
330 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { in msc_win_total_sz()
425 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { in msc_win_oldest_sg()
664 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { in msc_buffer_clear_hw_header()
994 unsigned int nr_segs) in __msc_buffer_win_alloc() argument
1000 ret = sg_alloc_table(win->sgt, nr_segs, GFP_KERNEL); in __msc_buffer_win_alloc()
1004 for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) { in __msc_buffer_win_alloc()
1014 return nr_segs; in __msc_buffer_win_alloc()
1039 for_each_sg(win->sgt->sgl, sg_ptr, win->nr_segs, i) { in msc_buffer_set_uc()
1059 for_each_sg(win->sgt->sgl, sg_ptr, win->nr_segs, i) { in msc_buffer_set_wb()
[all …]
/Linux-v5.15/fs/fuse/
Ddev.c657 unsigned long nr_segs; member
713 BUG_ON(!cs->nr_segs); in fuse_copy_fill()
719 cs->nr_segs--; in fuse_copy_fill()
721 if (cs->nr_segs >= cs->pipe->max_usage) in fuse_copy_fill()
737 cs->nr_segs++; in fuse_copy_fill()
813 BUG_ON(!cs->nr_segs); in fuse_try_move_page()
817 cs->nr_segs--; in fuse_try_move_page()
899 if (cs->nr_segs >= cs->pipe->max_usage) in fuse_ref_page()
917 cs->nr_segs++; in fuse_ref_page()
1384 if (pipe_occupancy(pipe->head, pipe->tail) + cs.nr_segs > pipe->max_usage) { in fuse_dev_splice_read()
[all …]
/Linux-v5.15/drivers/block/xen-blkback/
Dblkback.c710 invcount = xen_blkbk_unmap_prepare(ring, pages, req->nr_segs, in xen_blkbk_unmap_and_respond()
907 pending_req->nr_segs, in xen_blkbk_map_seg()
923 nseg = pending_req->nr_segs; in xen_blkbk_parse_indirect()
1256 pending_req->nr_segs = nseg; in dispatch_rw_block_io()
1377 pending_req->nr_segs); in dispatch_rw_block_io()
Dcommon.h344 int nr_segs; member
/Linux-v5.15/drivers/nvme/target/
Dio-cmd-file.c102 unsigned long nr_segs, size_t count, int ki_flags) in nvmet_file_submit_bvec() argument
119 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); in nvmet_file_submit_bvec()
/Linux-v5.15/mm/
Dswap.c165 int get_kernel_pages(const struct kvec *kiov, int nr_segs, int write, in get_kernel_pages() argument
170 for (seg = 0; seg < nr_segs; seg++) { in get_kernel_pages()
/Linux-v5.15/sound/core/
Dpcm_native.c3443 if (to->nr_segs > 1024 || to->nr_segs != runtime->channels) in snd_pcm_readv()
3448 bufs = kmalloc_array(to->nr_segs, sizeof(void *), GFP_KERNEL); in snd_pcm_readv()
3451 for (i = 0; i < to->nr_segs; ++i) in snd_pcm_readv()
3479 if (from->nr_segs > 128 || from->nr_segs != runtime->channels || in snd_pcm_writev()
3483 bufs = kmalloc_array(from->nr_segs, sizeof(void *), GFP_KERNEL); in snd_pcm_writev()
3486 for (i = 0; i < from->nr_segs; ++i) in snd_pcm_writev()
/Linux-v5.15/fs/cifs/
Dsmb1ops.c747 struct kvec *iov, unsigned long nr_segs) in cifs_sync_write() argument
751 return CIFSSMBWrite2(xid, parms, written, iov, nr_segs); in cifs_sync_write()
/Linux-v5.15/fs/
Dsplice.c1297 unsigned long, nr_segs, unsigned int, flags) in SYSCALL_DEFINE4() argument
1314 error = import_iovec(type, uiov, nr_segs, in SYSCALL_DEFINE4()

12