| /Linux-v5.4/drivers/virt/ |
| D | fsl_hypervisor.c | 153 struct fh_sg_list *sg_list = NULL; in ioctl_memcpy() local 246 sg_list = PTR_ALIGN(sg_list_unaligned, sizeof(struct fh_sg_list)); in ioctl_memcpy() 264 sg_list[0].source = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy() 265 sg_list[0].target = param.remote_paddr; in ioctl_memcpy() 267 sg_list[0].source = param.remote_paddr; in ioctl_memcpy() 268 sg_list[0].target = page_to_phys(pages[0]) + lb_offset; in ioctl_memcpy() 270 sg_list[0].size = min_t(uint64_t, param.count, PAGE_SIZE - lb_offset); in ioctl_memcpy() 272 remote_paddr = param.remote_paddr + sg_list[0].size; in ioctl_memcpy() 273 count = param.count - sg_list[0].size; in ioctl_memcpy() 278 sg_list[i].source = page_to_phys(pages[i]); in ioctl_memcpy() [all …]
|
| /Linux-v5.4/drivers/infiniband/hw/cxgb3/ |
| D | iwch_qp.c | 73 if ((plen + wr->sg_list[i].length) < plen) in build_rdma_send() 76 plen += wr->sg_list[i].length; in build_rdma_send() 77 wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); in build_rdma_send() 78 wqe->send.sgl[i].len = cpu_to_be32(wr->sg_list[i].length); in build_rdma_send() 79 wqe->send.sgl[i].to = cpu_to_be64(wr->sg_list[i].addr); in build_rdma_send() 110 if ((plen + wr->sg_list[i].length) < plen) { in build_rdma_write() 113 plen += wr->sg_list[i].length; in build_rdma_write() 115 cpu_to_be32(wr->sg_list[i].lkey); in build_rdma_write() 117 cpu_to_be32(wr->sg_list[i].length); in build_rdma_write() 119 cpu_to_be64(wr->sg_list[i].addr); in build_rdma_write() [all …]
|
| /Linux-v5.4/drivers/scsi/qla2xxx/ |
| D | qla_bsg.c | 39 bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free() 44 bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free() 47 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_bsg_sp_free() 50 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_bsg_sp_free() 182 bsg_job->reply_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd() 209 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla24xx_proc_fcp_prio_cfg_cmd() 337 dma_map_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els() 340 dma_unmap_sg(&ha->pdev->dev, bsg_job->request_payload.sg_list, in qla2x00_process_els() 346 rsp_sg_cnt = dma_map_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els() 349 dma_unmap_sg(&ha->pdev->dev, bsg_job->reply_payload.sg_list, in qla2x00_process_els() [all …]
|
| /Linux-v5.4/block/ |
| D | bsg-lib.c | 155 kfree(job->request_payload.sg_list); in bsg_teardown_job() 156 kfree(job->reply_payload.sg_list); in bsg_teardown_job() 207 buf->sg_list = kzalloc(sz, GFP_KERNEL); in bsg_map_buffer() 208 if (!buf->sg_list) in bsg_map_buffer() 210 sg_init_table(buf->sg_list, req->nr_phys_segments); in bsg_map_buffer() 211 buf->sg_cnt = blk_rq_map_sg(req->q, req, buf->sg_list); in bsg_map_buffer() 245 kfree(job->request_payload.sg_list); in bsg_prepare_job()
|
| /Linux-v5.4/drivers/infiniband/sw/rdmavt/ |
| D | rc.c | 201 ss->sge = wqe->sg_list[0]; in rvt_restart_sge() 202 ss->sg_list = wqe->sg_list + 1; in rvt_restart_sge()
|
| D | qp.c | 662 struct rvt_sge *sge = &wqe->sg_list[i]; in rvt_swqe_has_lkey() 1088 sz = struct_size(swq, sg_list, init_attr->cap.max_send_sge); in rvt_create_qp() 1851 wqe->sg_list[i].addr = wr->sg_list[i].addr; in rvt_post_recv() 1852 wqe->sg_list[i].length = wr->sg_list[i].length; in rvt_post_recv() 1853 wqe->sg_list[i].lkey = wr->sg_list[i].lkey; in rvt_post_recv() 1900 wr->sg_list[0].length < sizeof(u64) || in rvt_qp_valid_operation() 1901 wr->sg_list[0].addr & (sizeof(u64) - 1))) in rvt_qp_valid_operation() 2072 u32 length = wr->sg_list[i].length; in rvt_post_one_wr() 2076 ret = rvt_lkey_ok(rkt, pd, &wqe->sg_list[j], last_sge, in rvt_post_one_wr() 2077 &wr->sg_list[i], acc); in rvt_post_one_wr() [all …]
|
| /Linux-v5.4/include/rdma/ |
| D | rdmavt_mr.h | 119 struct rvt_sge *sg_list; /* next SGE to be used if any */ member 140 ss->sge = *ss->sg_list++; in rvt_put_ss() 168 *sge = *ss->sg_list++; in rvt_update_sge()
|
| /Linux-v5.4/drivers/dma/ |
| D | imx-dma.c | 159 struct scatterlist *sg_list; member 811 kfree(imxdmac->sg_list); in imxdma_free_chan_resources() 812 imxdmac->sg_list = NULL; in imxdma_free_chan_resources() 886 kfree(imxdmac->sg_list); in imxdma_prep_dma_cyclic() 888 imxdmac->sg_list = kcalloc(periods + 1, in imxdma_prep_dma_cyclic() 890 if (!imxdmac->sg_list) in imxdma_prep_dma_cyclic() 893 sg_init_table(imxdmac->sg_list, periods); in imxdma_prep_dma_cyclic() 896 sg_assign_page(&imxdmac->sg_list[i], NULL); in imxdma_prep_dma_cyclic() 897 imxdmac->sg_list[i].offset = 0; in imxdma_prep_dma_cyclic() 898 imxdmac->sg_list[i].dma_address = dma_addr; in imxdma_prep_dma_cyclic() [all …]
|
| /Linux-v5.4/drivers/infiniband/core/ |
| D | uverbs_std_types_mr.c | 52 struct ib_sge *sg_list; in UVERBS_HANDLER() local 75 sg_list = uverbs_attr_get_alloced_ptr(attrs, in UVERBS_HANDLER() 77 return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge, in UVERBS_HANDLER()
|
| /Linux-v5.4/drivers/infiniband/hw/mlx5/ |
| D | odp.c | 1630 struct ib_sge sg_list[0]; member 1634 struct ib_sge *sg_list, u32 num_sge, in num_pending_prefetch_dec() argument 1647 mlx5_base_mkey(sg_list[i].lkey)); in num_pending_prefetch_dec() 1656 struct ib_sge *sg_list, u32 num_sge) in num_pending_prefetch_inc() argument 1667 mlx5_base_mkey(sg_list[i].lkey)); in num_pending_prefetch_inc() 1668 if (!mmkey || mmkey->key != sg_list[i].lkey) { in num_pending_prefetch_inc() 1694 num_pending_prefetch_dec(dev, sg_list, i, 0); in num_pending_prefetch_inc() 1700 struct ib_sge *sg_list, u32 num_sge) in mlx5_ib_prefetch_sg_list() argument 1707 struct ib_sge *sg = &sg_list[i]; in mlx5_ib_prefetch_sg_list() 1727 mlx5_ib_prefetch_sg_list(w->pd, w->pf_flags, w->sg_list, in mlx5_ib_prefetch_mr_work() [all …]
|
| /Linux-v5.4/drivers/infiniband/hw/qib/ |
| D | qib_ud.c | 171 ssge.sg_list = swqe->sg_list + 1; in qib_ud_loopback() 172 ssge.sge = *swqe->sg_list; in qib_ud_loopback() 184 *sge = *ssge.sg_list++; in qib_ud_loopback() 317 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_ud_req() 318 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_ud_req()
|
| D | qib_uc.c | 98 qp->s_sge.sge = wqe->sg_list[0]; in qib_make_uc_req() 99 qp->s_sge.sg_list = wqe->sg_list + 1; in qib_make_uc_req() 418 qp->r_sge.sg_list = NULL; in qib_uc_rcv()
|
| /Linux-v5.4/drivers/infiniband/hw/qedr/ |
| D | qedr_roce_cm.c | 110 qp->rqe_wr_id[qp->rq.gsi_cons].sg_list[0].length = in qedr_ll2_complete_rx_packet() 409 send_size += swr->sg_list[i].length; in qedr_gsi_build_header() 534 packet->payload[i].baddr = swr->sg_list[i].addr; in qedr_gsi_build_packet() 535 packet->payload[i].len = swr->sg_list[i].length; in qedr_gsi_build_packet() 642 wr->sg_list[0].addr, in qedr_gsi_post_recv() 643 wr->sg_list[0].length, in qedr_gsi_post_recv() 655 qp->rqe_wr_id[qp->rq.prod].sg_list[0] = wr->sg_list[0]; in qedr_gsi_post_recv() 693 wc[i].byte_len = qp->rqe_wr_id[qp->rq.cons].sg_list[0].length; in qedr_gsi_poll_cq()
|
| /Linux-v5.4/drivers/scsi/aacraid/ |
| D | commctrl.c | 478 void *sg_list[HBA_MAX_SG_EMBEDDED]; in aac_send_raw_srb() local 503 memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */ in aac_send_raw_srb() 542 if (user_srbcmd->sg.count > ARRAY_SIZE(sg_list)) { in aac_send_raw_srb() 670 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 731 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 786 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 843 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 882 sg_list[i] = p; // save so we can clean up later in aac_send_raw_srb() 922 if (copy_to_user(sg_user[i], sg_list[i], sg_count[i])) { in aac_send_raw_srb() 977 kfree(sg_list[i]); in aac_send_raw_srb()
|
| /Linux-v5.4/drivers/infiniband/hw/bnxt_re/ |
| D | ib_verbs.c | 108 struct bnxt_qplib_sge *sg_list, int num) in bnxt_re_build_sgl() argument 113 sg_list[i].addr = ib_sg_list[i].addr; in bnxt_re_build_sgl() 114 sg_list[i].lkey = ib_sg_list[i].lkey; in bnxt_re_build_sgl() 115 sg_list[i].size = ib_sg_list[i].length; in bnxt_re_build_sgl() 116 total += sg_list[i].size; in bnxt_re_build_sgl() 1487 bnxt_re_build_sgl(wr->sg_list, wqe.sg_list, wr->num_sge); in bnxt_re_post_srq_recv() 1965 wqe->sg_list[i].addr = wqe->sg_list[i - 1].addr; in bnxt_re_build_qp1_send_v2() 1966 wqe->sg_list[i].lkey = wqe->sg_list[i - 1].lkey; in bnxt_re_build_qp1_send_v2() 1967 wqe->sg_list[i].size = wqe->sg_list[i - 1].size; in bnxt_re_build_qp1_send_v2() 1991 wqe->sg_list[0].addr = sge.addr; in bnxt_re_build_qp1_send_v2() [all …]
|
| /Linux-v5.4/drivers/infiniband/hw/cxgb4/ |
| D | qp.c | 423 if ((plen + wr->sg_list[i].length) > max) in build_immd() 425 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; in build_immd() 426 plen += wr->sg_list[i].length; in build_immd() 427 rem = wr->sg_list[i].length; in build_immd() 453 struct fw_ri_isgl *isglp, struct ib_sge *sg_list, in build_isgl() argument 467 if ((plen + sg_list[i].length) < plen) in build_isgl() 469 plen += sg_list[i].length; in build_isgl() 470 *flitp = cpu_to_be64(((u64)sg_list[i].lkey << 32) | in build_isgl() 471 sg_list[i].length); in build_isgl() 474 *flitp = cpu_to_be64(sg_list[i].addr); in build_isgl() [all …]
|
| /Linux-v5.4/arch/powerpc/include/asm/ |
| D | fsl_hcalls.h | 358 unsigned int target, phys_addr_t sg_list, unsigned int count) in fh_partition_memcpy() argument 370 r5 = (uint32_t) sg_list; in fh_partition_memcpy() 373 r6 = sg_list >> 32; in fh_partition_memcpy()
|
| /Linux-v5.4/drivers/xen/ |
| D | efi.c | 219 unsigned long count, unsigned long sg_list) in xen_efi_update_capsule() argument 229 efi_data(op).u.update_capsule.sg_list = sg_list; in xen_efi_update_capsule()
|
| /Linux-v5.4/drivers/net/ethernet/ibm/ehea/ |
| D | ehea_qmr.h | 106 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member 115 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES-1]; member 132 struct ehea_vsgentry sg_list[EHEA_MAX_WQE_SG_ENTRIES]; member
|
| /Linux-v5.4/include/uapi/rdma/ |
| D | rvt-abi.h | 47 struct rvt_wqe_sge sg_list[]; member
|
| /Linux-v5.4/drivers/infiniband/hw/i40iw/ |
| D | i40iw_verbs.c | 2098 static void i40iw_copy_sg_list(struct i40iw_sge *sg_list, struct ib_sge *sgl, int num_sges) in i40iw_copy_sg_list() argument 2103 sg_list[i].tag_off = sgl[i].addr; in i40iw_copy_sg_list() 2104 sg_list[i].len = sgl[i].length; in i40iw_copy_sg_list() 2105 sg_list[i].stag = sgl[i].lkey; in i40iw_copy_sg_list() 2163 info.op.inline_send.data = (void *)(unsigned long)ib_wr->sg_list[0].addr; in i40iw_post_send() 2164 info.op.inline_send.len = ib_wr->sg_list[0].length; in i40iw_post_send() 2168 info.op.send.sg_list = (struct i40iw_sge *)ib_wr->sg_list; in i40iw_post_send() 2183 info.op.inline_rdma_write.data = (void *)(unsigned long)ib_wr->sg_list[0].addr; in i40iw_post_send() 2184 info.op.inline_rdma_write.len = ib_wr->sg_list[0].length; in i40iw_post_send() 2189 info.op.rdma_write.lo_sg_list = (void *)ib_wr->sg_list; in i40iw_post_send() [all …]
|
| /Linux-v5.4/drivers/scsi/ufs/ |
| D | ufs_bsg.c | 71 sg_copy_to_buffer(job->request_payload.sg_list, in ufs_bsg_alloc_desc_buffer() 147 sg_copy_from_buffer(job->request_payload.sg_list, in ufs_bsg_request()
|
| /Linux-v5.4/include/linux/ |
| D | bsg-lib.h | 26 struct scatterlist *sg_list; member
|
| /Linux-v5.4/drivers/infiniband/hw/mlx4/ |
| D | srq.c | 340 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx4_ib_post_srq_recv() 341 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx4_ib_post_srq_recv() 342 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx4_ib_post_srq_recv()
|
| /Linux-v5.4/drivers/scsi/qla4xxx/ |
| D | ql4_bsg.c | 63 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_flash() 123 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_flash() 187 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_get_acb_state() 258 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_read_nvram() 322 sg_copy_to_buffer(bsg_job->request_payload.sg_list, in qla4xxx_update_nvram() 436 sg_copy_from_buffer(bsg_job->reply_payload.sg_list, in qla4xxx_bsg_get_acb()
|