/Linux-v5.4/net/rds/ |
D | page.c | 68 int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes, in rds_page_remainder_alloc() argument 84 sg_set_page(scat, page, PAGE_SIZE, 0); in rds_page_remainder_alloc() 103 sg_set_page(scat, rem->r_page, bytes, rem->r_offset); in rds_page_remainder_alloc() 104 get_page(sg_page(scat)); in rds_page_remainder_alloc() 147 ret ? NULL : sg_page(scat), ret ? 0 : scat->offset, in rds_page_remainder_alloc() 148 ret ? 0 : scat->length); in rds_page_remainder_alloc()
|
D | ib_send.c | 492 struct scatterlist *scat; in rds_ib_xmit() local 512 scat = &rm->data.op_sg[sg]; in rds_ib_xmit() 513 ret = max_t(int, RDS_CONG_MAP_BYTES, scat->length); in rds_ib_xmit() 621 scat = &ic->i_data_op->op_sg[rm->data.op_dmasg]; in rds_ib_xmit() 642 && scat != &rm->data.op_sg[rm->data.op_count]) { in rds_ib_xmit() 644 sg_dma_len(scat) - rm->data.op_dmaoff); in rds_ib_xmit() 647 send->s_sge[1].addr = sg_dma_address(scat); in rds_ib_xmit() 653 if (rm->data.op_dmaoff == sg_dma_len(scat)) { in rds_ib_xmit() 654 scat++; in rds_ib_xmit() 695 && scat != &rm->data.op_sg[rm->data.op_count]); in rds_ib_xmit() [all …]
|
D | ib_fmr.c | 104 struct scatterlist *scat = sg; in rds_ib_map_fmr() local 122 unsigned int dma_len = sg_dma_len(&scat[i]); in rds_ib_map_fmr() 123 u64 dma_addr = sg_dma_address(&scat[i]); in rds_ib_map_fmr() 162 unsigned int dma_len = sg_dma_len(&scat[i]); in rds_ib_map_fmr() 163 u64 dma_addr = sg_dma_address(&scat[i]); in rds_ib_map_fmr() 181 ibmr->sg = scat; in rds_ib_map_fmr()
|
D | rds.h | 885 int rds_page_remainder_alloc(struct scatterlist *scat, unsigned long bytes,
|
/Linux-v5.4/drivers/infiniband/hw/mlx4/ |
D | srq.c | 305 struct mlx4_wqe_data_seg *scat; in mlx4_ib_post_srq_recv() local 337 scat = (struct mlx4_wqe_data_seg *) (next + 1); in mlx4_ib_post_srq_recv() 340 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx4_ib_post_srq_recv() 341 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx4_ib_post_srq_recv() 342 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx4_ib_post_srq_recv() 346 scat[i].byte_count = 0; in mlx4_ib_post_srq_recv() 347 scat[i].lkey = cpu_to_be32(MLX4_INVALID_LKEY); in mlx4_ib_post_srq_recv() 348 scat[i].addr = 0; in mlx4_ib_post_srq_recv()
|
D | qp.c | 3851 struct mlx4_wqe_data_seg *scat; in _mlx4_ib_post_recv() local 3886 scat = get_recv_wqe(qp, ind); in _mlx4_ib_post_recv() 3894 scat->byte_count = in _mlx4_ib_post_recv() 3897 scat->lkey = cpu_to_be32(wr->sg_list->lkey); in _mlx4_ib_post_recv() 3898 scat->addr = cpu_to_be64(qp->sqp_proxy_rcv[ind].map); in _mlx4_ib_post_recv() 3899 scat++; in _mlx4_ib_post_recv() 3904 __set_data_seg(scat + i, wr->sg_list + i); in _mlx4_ib_post_recv() 3907 scat[i].byte_count = 0; in _mlx4_ib_post_recv() 3908 scat[i].lkey = cpu_to_be32(MLX4_INVALID_LKEY); in _mlx4_ib_post_recv() 3909 scat[i].addr = 0; in _mlx4_ib_post_recv()
|
/Linux-v5.4/drivers/infiniband/hw/mlx5/ |
D | srq.c | 425 struct mlx5_wqe_data_seg *scat; in mlx5_ib_post_srq_recv() local 458 scat = (struct mlx5_wqe_data_seg *)(next + 1); in mlx5_ib_post_srq_recv() 461 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx5_ib_post_srq_recv() 462 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx5_ib_post_srq_recv() 463 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx5_ib_post_srq_recv() 467 scat[i].byte_count = 0; in mlx5_ib_post_srq_recv() 468 scat[i].lkey = cpu_to_be32(MLX5_INVALID_LKEY); in mlx5_ib_post_srq_recv() 469 scat[i].addr = 0; in mlx5_ib_post_srq_recv()
|
D | qp.c | 5361 struct mlx5_wqe_data_seg *scat; in _mlx5_ib_post_recv() local 5397 scat = mlx5_frag_buf_get_wqe(&qp->rq.fbc, ind); in _mlx5_ib_post_recv() 5399 scat++; in _mlx5_ib_post_recv() 5402 set_data_ptr_seg(scat + i, wr->sg_list + i); in _mlx5_ib_post_recv() 5405 scat[i].byte_count = 0; in _mlx5_ib_post_recv() 5406 scat[i].lkey = cpu_to_be32(MLX5_INVALID_LKEY); in _mlx5_ib_post_recv() 5407 scat[i].addr = 0; in _mlx5_ib_post_recv() 5411 sig = (struct mlx5_rwqe_sig *)scat; in _mlx5_ib_post_recv()
|
/Linux-v5.4/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 1644 struct srp_request *req, struct scatterlist *scat, in srp_map_sg_fmr() argument 1654 for_each_sg(scat, sg, count, i) { in srp_map_sg_fmr() 1668 struct srp_request *req, struct scatterlist *scat, in srp_map_sg_fr() argument 1675 state->sg = scat; in srp_map_sg_fr() 1696 struct srp_request *req, struct scatterlist *scat, in srp_map_sg_dma() argument 1703 for_each_sg(scat, sg, count, i) { in srp_map_sg_dma() 1768 struct scatterlist *scat, int count) in srp_check_mapping() argument 1804 struct scatterlist *scat, *sg; in srp_map_data() local 1830 scat = scsi_sglist(scmnd); in srp_map_data() 1836 count = ib_dma_map_sg(ibdev, scat, nents, scmnd->sc_data_direction); in srp_map_data() [all …]
|
/Linux-v5.4/Documentation/m68k/ |
D | kernel-options.rst | 594 :Syntax: atascsi=<can_queue>[,<cmd_per_lun>[,<scat-gat>[,<host-id>[,<tagged>]]]] 631 <scat-gat>:
|
/Linux-v5.4/drivers/infiniband/hw/hns/ |
D | hns_roce_hw_v1.c | 358 struct hns_roce_wqe_data_seg *scat = NULL; in hns_roce_v1_post_recv() local 391 scat = (struct hns_roce_wqe_data_seg *)(ctrl + 1); in hns_roce_v1_post_recv() 394 set_data_seg(scat + i, wr->sg_list + i); in hns_roce_v1_post_recv()
|