/Linux-v4.19/drivers/infiniband/hw/mthca/ |
D | mthca_provider.c | 446 struct mthca_create_srq ucmd; in mthca_create_srq() local 461 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mthca_create_srq() 467 context->db_tab, ucmd.db_index, in mthca_create_srq() 468 ucmd.db_page); in mthca_create_srq() 473 srq->mr.ibmr.lkey = ucmd.lkey; in mthca_create_srq() 474 srq->db_index = ucmd.db_index; in mthca_create_srq() 482 context->db_tab, ucmd.db_index); in mthca_create_srq() 522 struct mthca_create_qp ucmd; in mthca_create_qp() local 543 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mthca_create_qp() 550 ucmd.sq_db_index, ucmd.sq_db_page); in mthca_create_qp() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/mlx5/ |
D | qp.c | 245 int has_rq, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) in set_rq_size() argument 261 if (ucmd) { in set_rq_size() 262 qp->rq.wqe_cnt = ucmd->rq_wqe_count; in set_rq_size() 263 if (ucmd->rq_wqe_shift > BITS_PER_BYTE * sizeof(ucmd->rq_wqe_shift)) in set_rq_size() 265 qp->rq.wqe_shift = ucmd->rq_wqe_shift; in set_rq_size() 440 struct mlx5_ib_create_qp *ucmd, in set_user_buf_size() argument 452 if (ucmd->sq_wqe_count && ((1 << ilog2(ucmd->sq_wqe_count)) != ucmd->sq_wqe_count)) { in set_user_buf_size() 454 ucmd->sq_wqe_count, ucmd->sq_wqe_count); in set_user_buf_size() 458 qp->sq.wqe_cnt = ucmd->sq_wqe_count; in set_user_buf_size() 713 struct mlx5_ib_create_wq *ucmd) in create_user_rq() argument [all …]
|
D | srq.c | 80 struct mlx5_ib_create_srq ucmd = {}; in create_srq_user() local 89 ucmdlen = min(udata->inlen, sizeof(ucmd)); in create_srq_user() 91 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) { in create_srq_user() 96 if (ucmd.reserved0 || ucmd.reserved1) in create_srq_user() 99 if (udata->inlen > sizeof(ucmd) && in create_srq_user() 100 !ib_is_udata_cleared(udata, sizeof(ucmd), in create_srq_user() 101 udata->inlen - sizeof(ucmd))) in create_srq_user() 106 &ucmd, udata->inlen, &uidx); in create_srq_user() 111 srq->wq_sig = !!(ucmd.flags & MLX5_SRQ_FLAG_SIGNATURE); in create_srq_user() 113 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, buf_size, in create_srq_user() [all …]
|
D | cq.c | 781 struct mlx5_ib_create_cq ucmd = {}; in create_cq_user() local 790 ucmdlen = udata->inlen < sizeof(ucmd) ? in create_cq_user() 791 (sizeof(ucmd) - sizeof(ucmd.flags)) : sizeof(ucmd); in create_cq_user() 793 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) in create_cq_user() 796 if (ucmdlen == sizeof(ucmd) && in create_cq_user() 797 (ucmd.flags & ~(MLX5_IB_CREATE_CQ_FLAGS_CQE_128B_PAD))) in create_cq_user() 800 if (ucmd.cqe_size != 64 && ucmd.cqe_size != 128) in create_cq_user() 803 *cqe_size = ucmd.cqe_size; in create_cq_user() 805 cq->buf.umem = ib_umem_get(context, ucmd.buf_addr, in create_cq_user() 806 entries * ucmd.cqe_size, in create_cq_user() [all …]
|
D | mlx5_ib.h | 1311 struct mlx5_ib_create_qp *ucmd, in get_qp_user_index() argument 1318 !cqe_version && (ucmd->uidx == MLX5_IB_DEFAULT_UIDX)) in get_qp_user_index() 1325 return verify_assign_uidx(cqe_version, ucmd->uidx, user_index); in get_qp_user_index() 1329 struct mlx5_ib_create_srq *ucmd, in get_srq_user_index() argument 1336 !cqe_version && (ucmd->uidx == MLX5_IB_DEFAULT_UIDX)) in get_srq_user_index() 1343 return verify_assign_uidx(cqe_version, ucmd->uidx, user_index); in get_srq_user_index()
|
D | main.c | 3181 struct mlx5_ib_create_flow *ucmd) in flow_counters_set_data() argument 3189 if (ucmd && ucmd->ncounters_data != 0) { in flow_counters_set_data() 3190 cntrs_data = ucmd->data; in flow_counters_set_data() 3256 struct mlx5_ib_create_flow *ucmd) in _create_flow_rule() argument 3323 err = flow_counters_set_data(flow_act.counters, ucmd); in _create_flow_rule() 3527 struct mlx5_ib_create_flow *ucmd = NULL, ucmd_hdr; in mlx5_ib_create_flow() local 3554 ucmd = kzalloc(required_ucmd_sz, GFP_KERNEL); in mlx5_ib_create_flow() 3555 if (!ucmd) in mlx5_ib_create_flow() 3558 err = ib_copy_from_udata(ucmd, udata, required_ucmd_sz); in mlx5_ib_create_flow() 3624 dst, underlay_qpn, ucmd); in mlx5_ib_create_flow() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/mlx4/ |
D | qp.c | 437 struct mlx4_ib_create_qp *ucmd) in set_user_sq_size() argument 440 if ((1 << ucmd->log_sq_bb_count) > dev->dev->caps.max_wqes || in set_user_sq_size() 441 ucmd->log_sq_stride > in set_user_sq_size() 443 ucmd->log_sq_stride < MLX4_IB_MIN_SQ_STRIDE) in set_user_sq_size() 446 qp->sq.wqe_cnt = 1 << ucmd->log_sq_bb_count; in set_user_sq_size() 447 qp->sq.wqe_shift = ucmd->log_sq_stride; in set_user_sq_size() 539 struct mlx4_ib_create_qp_rss *ucmd) in set_qp_rss() argument 544 if ((ucmd->rx_hash_function == MLX4_IB_RX_HASH_FUNC_TOEPLITZ) && in set_qp_rss() 546 memcpy(rss_ctx->rss_key, ucmd->rx_hash_key, in set_qp_rss() 553 if (ucmd->rx_hash_fields_mask & ~(MLX4_IB_RX_HASH_SRC_IPV4 | in set_qp_rss() [all …]
|
D | srq.c | 109 struct mlx4_ib_create_srq ucmd; in mlx4_ib_create_srq() local 111 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mlx4_ib_create_srq() 116 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, in mlx4_ib_create_srq() 133 ucmd.db_addr, &srq->db); in mlx4_ib_create_srq()
|
D | cq.c | 207 struct mlx4_ib_create_cq ucmd; in mlx4_ib_create_cq() local 209 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mlx4_ib_create_cq() 215 ucmd.buf_addr, entries); in mlx4_ib_create_cq() 219 err = mlx4_ib_db_map_user(to_mucontext(context), ucmd.db_addr, in mlx4_ib_create_cq() 319 struct mlx4_ib_resize_cq ucmd; in mlx4_alloc_resize_umem() local 325 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) in mlx4_alloc_resize_umem() 333 &cq->resize_umem, ucmd.buf_addr, entries); in mlx4_alloc_resize_umem()
|
/Linux-v4.19/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_srq.c | 110 struct pvrdma_create_srq ucmd; in pvrdma_create_srq() local 151 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_srq() 157 ucmd.buf_addr, in pvrdma_create_srq() 158 ucmd.buf_size, 0, 0); in pvrdma_create_srq()
|
D | pvrdma_cq.c | 118 struct pvrdma_create_cq ucmd; in pvrdma_create_cq() local 139 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_cq() 144 cq->umem = ib_umem_get(context, ucmd.buf_addr, ucmd.buf_size, in pvrdma_create_cq()
|
D | pvrdma_qp.c | 198 struct pvrdma_create_qp ucmd; in pvrdma_create_qp() local 258 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_qp() 266 ucmd.rbuf_addr, in pvrdma_create_qp() 267 ucmd.rbuf_size, 0, 0); in pvrdma_create_qp() 279 ucmd.sbuf_addr, in pvrdma_create_qp() 280 ucmd.sbuf_size, 0, 0); in pvrdma_create_qp()
|
/Linux-v4.19/drivers/infiniband/hw/hns/ |
D | hns_roce_qp.c | 342 struct hns_roce_ib_create_qp *ucmd) in hns_roce_set_user_sq_size() argument 350 if ((u32)(1 << ucmd->log_sq_bb_count) > hr_dev->caps.max_wqes || in hns_roce_set_user_sq_size() 351 ucmd->log_sq_stride > max_sq_stride || in hns_roce_set_user_sq_size() 352 ucmd->log_sq_stride < HNS_ROCE_IB_MIN_SQ_STRIDE) { in hns_roce_set_user_sq_size() 363 hr_qp->sq.wqe_cnt = 1 << ucmd->log_sq_bb_count; in hns_roce_set_user_sq_size() 364 hr_qp->sq.wqe_shift = ucmd->log_sq_stride; in hns_roce_set_user_sq_size() 519 struct hns_roce_ib_create_qp ucmd; in hns_roce_create_qp_common() local 578 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in hns_roce_create_qp_common() 585 &ucmd); in hns_roce_create_qp_common() 592 ucmd.buf_addr, hr_qp->buff_size, 0, in hns_roce_create_qp_common() [all …]
|
D | hns_roce_cq.c | 317 struct hns_roce_ib_create_cq ucmd; in hns_roce_ib_create_cq() local 343 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in hns_roce_ib_create_cq() 351 &hr_cq->umem, ucmd.buf_addr, in hns_roce_ib_create_cq() 361 ucmd.db_addr, &hr_cq->db); in hns_roce_ib_create_cq()
|
/Linux-v4.19/arch/mips/cavium-octeon/ |
D | octeon-platform.c | 34 u32 ucmd; in octeon2_usb_reset() local 41 ucmd = cvmx_read64_uint32(CVMX_UAHCX_EHCI_USBCMD); in octeon2_usb_reset() 42 ucmd &= ~CMD_RUN; in octeon2_usb_reset() 43 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset() 45 ucmd |= CMD_RESET; in octeon2_usb_reset() 46 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset() 47 ucmd = cvmx_read64_uint32(CVMX_UAHCX_OHCI_USBCMD); in octeon2_usb_reset() 48 ucmd |= CMD_RUN; in octeon2_usb_reset() 49 cvmx_write64_uint32(CVMX_UAHCX_OHCI_USBCMD, ucmd); in octeon2_usb_reset()
|
/Linux-v4.19/drivers/infiniband/sw/rxe/ |
D | rxe_srq.c | 146 struct rxe_modify_srq_cmd *ucmd) in rxe_srq_from_attr() argument 157 mi = u64_to_user_ptr(ucmd->mmap_info_addr); in rxe_srq_from_attr()
|
D | rxe_verbs.c | 389 struct rxe_modify_srq_cmd ucmd = {}; in rxe_modify_srq() local 392 if (udata->inlen < sizeof(ucmd)) in rxe_modify_srq() 395 err = ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)); in rxe_modify_srq() 404 err = rxe_srq_from_attr(rxe, srq, attr, mask, &ucmd); in rxe_modify_srq()
|
D | rxe_loc.h | 233 struct rxe_modify_srq_cmd *ucmd);
|
/Linux-v4.19/drivers/infiniband/hw/qib/ |
D | qib_file_ops.c | 2038 const struct qib_cmd __user *ucmd; in qib_write() local 2057 ucmd = (const struct qib_cmd __user *) data; in qib_write() 2059 if (copy_from_user(&cmd.type, &ucmd->type, sizeof(cmd.type))) { in qib_write() 2071 src = &ucmd->cmd.user_info; in qib_write() 2077 src = &ucmd->cmd.recv_ctrl; in qib_write() 2083 src = &ucmd->cmd.ctxt_info; in qib_write() 2090 src = &ucmd->cmd.tid_info; in qib_write() 2096 src = &ucmd->cmd.part_key; in qib_write() 2109 src = &ucmd->cmd.poll_type; in qib_write() 2115 src = &ucmd->cmd.armlaunch_ctrl; in qib_write() [all …]
|
/Linux-v4.19/drivers/block/rsxx/ |
D | cregs.c | 678 struct rsxx_reg_access __user *ucmd, in rsxx_reg_access() argument 684 st = copy_from_user(&cmd, ucmd, sizeof(cmd)); in rsxx_reg_access() 695 st = put_user(cmd.stat, &ucmd->stat); in rsxx_reg_access() 700 st = copy_to_user(ucmd->data, cmd.data, cmd.cnt); in rsxx_reg_access()
|
D | rsxx_priv.h | 427 struct rsxx_reg_access __user *ucmd,
|
/Linux-v4.19/drivers/scsi/sym53c8xx_2/ |
D | sym_glue.c | 142 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); in sym_xpt_done() local 145 if (ucmd->eh_done) in sym_xpt_done() 146 complete(ucmd->eh_done); in sym_xpt_done() 593 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); in sym_eh_handler() local 673 ucmd->eh_done = &eh_done; in sym_eh_handler() 676 ucmd->eh_done = NULL; in sym_eh_handler()
|
/Linux-v4.19/drivers/infiniband/hw/cxgb4/ |
D | cq.c | 1002 struct c4iw_create_cq ucmd; in c4iw_create_cq() local 1020 if (udata->inlen < sizeof(ucmd)) in c4iw_create_cq()
|
/Linux-v4.19/drivers/nvme/host/ |
D | core.c | 1214 struct nvme_passthru_cmd __user *ucmd) in nvme_user_cmd() argument 1224 if (copy_from_user(&cmd, ucmd, sizeof(cmd))) in nvme_user_cmd() 1253 if (put_user(cmd.result, &ucmd->result)) in nvme_user_cmd()
|