Home
last modified time | relevance | path

Searched refs:sg_list (Results 1 – 25 of 142) sorted by relevance

123456

/Linux-v5.10/drivers/virt/
Dfsl_hypervisor.c153 struct fh_sg_list *sg_list = NULL; in ioctl_memcpy() local
246 sg_list = PTR_ALIGN(sg_list_unaligned, sizeof(struct fh_sg_list)); in ioctl_memcpy()
263 sg_list[0].source = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy()
264 sg_list[0].target = param.remote_paddr; in ioctl_memcpy()
266 sg_list[0].source = param.remote_paddr; in ioctl_memcpy()
267 sg_list[0].target = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy()
269 sg_list[0].size = min_t(uint64_t, param.count, PAGE_SIZE - lb_offset); in ioctl_memcpy()
271 remote_paddr = param.remote_paddr + sg_list[0].size; in ioctl_memcpy()
272 count = param.count - sg_list[0].size; in ioctl_memcpy()
277 sg_list[i].source = page_to_phys(pages[i]); in ioctl_memcpy()
[all …]
/Linux-v5.10/drivers/scsi/qla2xxx/
Dqla_bsg.c46 bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free()
51 bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free()
54 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free()
57 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free()
192 bsg_job->reply_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd()
219 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd()
346 dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els()
349 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els()
355 rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els()
358 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els()
[all …]
/Linux-v5.10/drivers/tee/amdtee/
Dcall.c257 cmd->sg_list.count = count; in handle_map_shmem()
262 cmd->sg_list.buf[i].hi_addr = upper_32_bits(paddr); in handle_map_shmem()
263 cmd->sg_list.buf[i].low_addr = lower_32_bits(paddr); in handle_map_shmem()
264 cmd->sg_list.buf[i].size = start[i].size; in handle_map_shmem()
265 cmd->sg_list.size += cmd->sg_list.buf[i].size; in handle_map_shmem()
268 cmd->sg_list.buf[i].hi_addr); in handle_map_shmem()
270 cmd->sg_list.buf[i].low_addr); in handle_map_shmem()
271 pr_debug("buf[%d]:size = 0x%x\n", i, cmd->sg_list.buf[i].size); in handle_map_shmem()
272 pr_debug("list size = 0x%x\n", cmd->sg_list.size); in handle_map_shmem()
/Linux-v5.10/block/
Dbsg-lib.c155 kfree(job->request_payload.sg_list); in bsg_teardown_job()
156 kfree(job->reply_payload.sg_list); in bsg_teardown_job()
210 buf->sg_list = kmalloc(sz, GFP_KERNEL); in bsg_map_buffer()
211 if (!buf->sg_list) in bsg_map_buffer()
213 sg_init_table(buf->sg_list, req->nr_phys_segments); in bsg_map_buffer()
214 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in bsg_map_buffer()
248 kfree(job->request_payload.sg_list); in bsg_prepare_job()
/Linux-v5.10/drivers/infiniband/sw/rdmavt/
Drc.c206 ss->sge = wqe->sg_list[0]; in rvt_restart_sge()
207 ss->sg_list = wqe->sg_list + 1; in rvt_restart_sge()
Dqp.c668 struct rvt_sge *sge = &wqe->sg_list[i]; in rvt_swqe_has_lkey()
1118 sz = struct_size(swq, sg_list, init_attr->cap.max_send_sge); in rvt_create_qp()
1883 wqe->sg_list[i].addr = wr->sg_list[i].addr; in rvt_post_recv()
1884 wqe->sg_list[i].length = wr->sg_list[i].length; in rvt_post_recv()
1885 wqe->sg_list[i].lkey = wr->sg_list[i].lkey; in rvt_post_recv()
1932 wr->sg_list[0].length < sizeof(u64) || in rvt_qp_valid_operation()
1933 wr->sg_list[0].addr & (sizeof(u64) - 1))) in rvt_qp_valid_operation()
2104 u32 length = wr->sg_list[i].length; in rvt_post_one_wr()
2108 ret = rvt_lkey_ok(rkt, pd, &wqe->sg_list[j], last_sge, in rvt_post_one_wr()
2109 &wr->sg_list[i], acc); in rvt_post_one_wr()
[all …]
/Linux-v5.10/include/rdma/
Drdmavt_mr.h77 struct rvt_sge *sg_list; /* next SGE to be used if any */ member
98 ss->sge = *ss->sg_list++; in rvt_put_ss()
126 *sge = *ss->sg_list++; in rvt_update_sge()
/Linux-v5.10/drivers/dma/
Dimx-dma.c159 struct scatterlist *sg_list; member
810 kfree(imxdmac->sg_list); in imxdma_free_chan_resources()
811 imxdmac->sg_list = NULL; in imxdma_free_chan_resources()
885 kfree(imxdmac->sg_list); in imxdma_prep_dma_cyclic()
887 imxdmac->sg_list = kcalloc(periods + 1, in imxdma_prep_dma_cyclic()
889 if (!imxdmac->sg_list) in imxdma_prep_dma_cyclic()
892 sg_init_table(imxdmac->sg_list, periods); in imxdma_prep_dma_cyclic()
895 sg_assign_page(&imxdmac->sg_list[i], NULL); in imxdma_prep_dma_cyclic()
896 imxdmac->sg_list[i].offset = 0; in imxdma_prep_dma_cyclic()
897 imxdmac->sg_list[i].dma_address = dma_addr; in imxdma_prep_dma_cyclic()
[all …]
/Linux-v5.10/drivers/infiniband/hw/qib/
Dqib_ud.c171 ssge.sg_list = swqe->sg_list + 1; in qib_ud_loopback()
172 ssge.sge = *swqe->sg_list; in qib_ud_loopback()
184 *sge = *ssge.sg_list++; in qib_ud_loopback()
317 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_ud_req()
318 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_ud_req()
Dqib_uc.c98 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_uc_req()
99 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_uc_req()
418 qp->r_sge.sg_list = NULL; in qib_uc_rcv()
/Linux-v5.10/drivers/infiniband/hw/qedr/
Dqedr_roce_cm.c110 qp->rqe_wr_id[qp->rq.gsi_cons].sg_list[0].length = in qedr_ll2_complete_rx_packet()
409 send_size += swr->sg_list[i].length; in qedr_gsi_build_header()
534 packet->payload[i].baddr = swr->sg_list[i].addr; in qedr_gsi_build_packet()
535 packet->payload[i].len = swr->sg_list[i].length; in qedr_gsi_build_packet()
642 wr->sg_list[0].addr, in qedr_gsi_post_recv()
643 wr->sg_list[0].length, in qedr_gsi_post_recv()
655 qp->rqe_wr_id[qp->rq.prod].sg_list[0] = wr->sg_list[0]; in qedr_gsi_post_recv()
693 wc[i].byte_len = qp->rqe_wr_id[qp->rq.cons].sg_list[0].length; in qedr_gsi_poll_cq()
/Linux-v5.10/drivers/scsi/aacraid/
Dcommctrl.c476 void *sg_list[HBA_MAX_SG_EMBEDDED]; in aac_send_raw_srb() local
501 memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */ in aac_send_raw_srb()
535 if (user_srbcmd->sg.count > ARRAY_SIZE(sg_list)) { in aac_send_raw_srb()
663 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
724 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
779 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
836 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
877 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb()
917 if (copy_to_user(sg_user[i], sg_list[i], sg_count[i])) { in aac_send_raw_srb()
972 kfree(sg_list[i]); in aac_send_raw_srb()
/Linux-v5.10/drivers/infiniband/core/
Duverbs_std_types_mr.c52 struct ib_sge *sg_list; in UVERBS_HANDLER() local
75 sg_list = uverbs_attr_get_alloced_ptr(attrs, in UVERBS_HANDLER()
77 return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge, in UVERBS_HANDLER()
/Linux-v5.10/drivers/net/ethernet/marvell/octeontx2/nic/
Dotx2_txrx.h72 struct sg_list { struct
93 struct sg_list *sg; argument
Dotx2_txrx.c67 static void otx2_dma_unmap_skb_frags(struct otx2_nic *pfvf, struct sg_list *sg) in otx2_dma_unmap_skb_frags()
88 struct sg_list *sg; in otx2_snd_pkt_handler()
629 struct sg_list *sg = &sq->sg[sqe]; in otx2_dma_map_tso_skb()
661 struct sg_list *sg = &sq->sg[sqe]; in otx2_tso_frag_dma_addr()
676 struct sg_list *list, int *offset) in otx2_sqe_tso_add_sg()
713 struct sg_list list; in otx2_sq_append_tso()
939 struct sg_list *sg; in otx2_cleanup_tx_cqes()
/Linux-v5.10/drivers/infiniband/hw/bnxt_re/
Dib_verbs.c108 struct bnxt_qplib_sge *sg_list, int num) in bnxt_re_build_sgl() argument
113 sg_list[i].addr = ib_sg_list[i].addr; in bnxt_re_build_sgl()
114 sg_list[i].lkey = ib_sg_list[i].lkey; in bnxt_re_build_sgl()
115 sg_list[i].size = ib_sg_list[i].length; in bnxt_re_build_sgl()
116 total += sg_list[i].size; in bnxt_re_build_sgl()
1774 bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge); in bnxt_re_post_srq_recv()
2250 wqe->sg_list[i].addr = wqe->sg_list[i - 1].addr; in bnxt_re_build_qp1_send_v2()
2251 wqe->sg_list[i].lkey = wqe->sg_list[i - 1].lkey; in bnxt_re_build_qp1_send_v2()
2252 wqe->sg_list[i].size = wqe->sg_list[i - 1].size; in bnxt_re_build_qp1_send_v2()
2276 wqe->sg_list[0].addr = sge.addr; in bnxt_re_build_qp1_send_v2()
[all …]
/Linux-v5.10/drivers/infiniband/hw/cxgb4/
Dqp.c423 if ((plen + wr->sg_list[i].length) > max) in build_immd()
425 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; in build_immd()
426 plen += wr->sg_list[i].length; in build_immd()
427 rem = wr->sg_list[i].length; in build_immd()
453 struct fw_ri_isgl *isglp, struct ib_sge *sg_list, in build_isgl() argument
467 if ((plen + sg_list[i].length) < plen) in build_isgl()
469 plen += sg_list[i].length; in build_isgl()
470 *flitp = cpu_to_be64(((u64)sg_list[i].lkey << 32) | in build_isgl()
471 sg_list[i].length); in build_isgl()
474 *flitp = cpu_to_be64(sg_list[i].addr); in build_isgl()
[all …]
/Linux-v5.10/arch/powerpc/include/asm/
Dfsl_hcalls.h358 unsigned int target, phys_addr_t sg_list, unsigned int count) in fh_partition_memcpy() argument
370 r5 = (uint32_t) sg_list; in fh_partition_memcpy()
373 r6 = sg_list >> 32; in fh_partition_memcpy()
/Linux-v5.10/drivers/xen/
Defi.c219 unsigned long count, unsigned long sg_list) in xen_efi_update_capsule() argument
229 efi_data(op).u.update_capsule.sg_list = sg_list; in xen_efi_update_capsule()
/Linux-v5.10/drivers/net/ethernet/ibm/ehea/
Dehea_qmr.h106 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
115 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES-1]; member
132 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
/Linux-v5.10/drivers/infiniband/hw/i40iw/
Di40iw_verbs.c2077 static void i40iw_copy_sg_list(struct i40iw_sge *sg_list, struct ib_sge *sgl, int num_sges) in i40iw_copy_sg_list() argument
2082 sg_list[i].tag_off = sgl[i].addr; in i40iw_copy_sg_list()
2083 sg_list[i].len = sgl[i].length; in i40iw_copy_sg_list()
2084 sg_list[i].stag = sgl[i].lkey; in i40iw_copy_sg_list()
2141 info.op.inline_send.data = (void *)(unsigned long)ib_wr->sg_list[0].addr; in i40iw_post_send()
2142 info.op.inline_send.len = ib_wr->sg_list[0].length; in i40iw_post_send()
2146 info.op.send.sg_list = (struct i40iw_sge *)ib_wr->sg_list; in i40iw_post_send()
2161 info.op.inline_rdma_write.data = (void *)(unsigned long)ib_wr->sg_list[0].addr; in i40iw_post_send()
2162 info.op.inline_rdma_write.len = ib_wr->sg_list[0].length; in i40iw_post_send()
2167 info.op.rdma_write.lo_sg_list = (void *)ib_wr->sg_list; in i40iw_post_send()
[all …]
/Linux-v5.10/include/uapi/rdma/
Drvt-abi.h47 struct rvt_wqe_sge sg_list[]; member
/Linux-v5.10/drivers/scsi/ufs/
Dufs_bsg.c70 sg_copy_to_buffer(job->request_payload.sg_list, in ufs_bsg_alloc_desc_buffer()
148 sg_copy_from_buffer(job->request_payload.sg_list, in ufs_bsg_request()
/Linux-v5.10/include/linux/
Dbsg-lib.h26 struct scatterlist *sg_list; member
/Linux-v5.10/drivers/infiniband/hw/mlx4/
Dsrq.c343 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx4_ib_post_srq_recv()
344 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx4_ib_post_srq_recv()
345 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx4_ib_post_srq_recv()

123456