Home
last modified time | relevance | path

Searched refs:ucmd (Results 1 – 24 of 24) sorted by relevance

/Linux-v4.19/drivers/infiniband/hw/mthca/
Dmthca_provider.c446 struct mthca_create_srq ucmd; in mthca_create_srq() local
461 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mthca_create_srq()
467 context->db_tab, ucmd.db_index, in mthca_create_srq()
468 ucmd.db_page); in mthca_create_srq()
473 srq->mr.ibmr.lkey = ucmd.lkey; in mthca_create_srq()
474 srq->db_index = ucmd.db_index; in mthca_create_srq()
482 context->db_tab, ucmd.db_index); in mthca_create_srq()
522 struct mthca_create_qp ucmd; in mthca_create_qp() local
543 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mthca_create_qp()
550 ucmd.sq_db_index, ucmd.sq_db_page); in mthca_create_qp()
[all …]
/Linux-v4.19/drivers/infiniband/hw/mlx5/
Dqp.c245 int has_rq, struct mlx5_ib_qp *qp, struct mlx5_ib_create_qp *ucmd) in set_rq_size() argument
261 if (ucmd) { in set_rq_size()
262 qp->rq.wqe_cnt = ucmd->rq_wqe_count; in set_rq_size()
263 if (ucmd->rq_wqe_shift > BITS_PER_BYTE * sizeof(ucmd->rq_wqe_shift)) in set_rq_size()
265 qp->rq.wqe_shift = ucmd->rq_wqe_shift; in set_rq_size()
440 struct mlx5_ib_create_qp *ucmd, in set_user_buf_size() argument
452 if (ucmd->sq_wqe_count && ((1 << ilog2(ucmd->sq_wqe_count)) != ucmd->sq_wqe_count)) { in set_user_buf_size()
454 ucmd->sq_wqe_count, ucmd->sq_wqe_count); in set_user_buf_size()
458 qp->sq.wqe_cnt = ucmd->sq_wqe_count; in set_user_buf_size()
713 struct mlx5_ib_create_wq *ucmd) in create_user_rq() argument
[all …]
Dsrq.c80 struct mlx5_ib_create_srq ucmd = {}; in create_srq_user() local
89 ucmdlen = min(udata->inlen, sizeof(ucmd)); in create_srq_user()
91 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) { in create_srq_user()
96 if (ucmd.reserved0 || ucmd.reserved1) in create_srq_user()
99 if (udata->inlen > sizeof(ucmd) && in create_srq_user()
100 !ib_is_udata_cleared(udata, sizeof(ucmd), in create_srq_user()
101 udata->inlen - sizeof(ucmd))) in create_srq_user()
106 &ucmd, udata->inlen, &uidx); in create_srq_user()
111 srq->wq_sig = !!(ucmd.flags & MLX5_SRQ_FLAG_SIGNATURE); in create_srq_user()
113 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, buf_size, in create_srq_user()
[all …]
Dcq.c781 struct mlx5_ib_create_cq ucmd = {}; in create_cq_user() local
790 ucmdlen = udata->inlen < sizeof(ucmd) ? in create_cq_user()
791 (sizeof(ucmd) - sizeof(ucmd.flags)) : sizeof(ucmd); in create_cq_user()
793 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) in create_cq_user()
796 if (ucmdlen == sizeof(ucmd) && in create_cq_user()
797 (ucmd.flags & ~(MLX5_IB_CREATE_CQ_FLAGS_CQE_128B_PAD))) in create_cq_user()
800 if (ucmd.cqe_size != 64 && ucmd.cqe_size != 128) in create_cq_user()
803 *cqe_size = ucmd.cqe_size; in create_cq_user()
805 cq->buf.umem = ib_umem_get(context, ucmd.buf_addr, in create_cq_user()
806 entries * ucmd.cqe_size, in create_cq_user()
[all …]
Dmlx5_ib.h1311 struct mlx5_ib_create_qp *ucmd, in get_qp_user_index() argument
1318 !cqe_version && (ucmd->uidx == MLX5_IB_DEFAULT_UIDX)) in get_qp_user_index()
1325 return verify_assign_uidx(cqe_version, ucmd->uidx, user_index); in get_qp_user_index()
1329 struct mlx5_ib_create_srq *ucmd, in get_srq_user_index() argument
1336 !cqe_version && (ucmd->uidx == MLX5_IB_DEFAULT_UIDX)) in get_srq_user_index()
1343 return verify_assign_uidx(cqe_version, ucmd->uidx, user_index); in get_srq_user_index()
Dmain.c3181 struct mlx5_ib_create_flow *ucmd) in flow_counters_set_data() argument
3189 if (ucmd && ucmd->ncounters_data != 0) { in flow_counters_set_data()
3190 cntrs_data = ucmd->data; in flow_counters_set_data()
3256 struct mlx5_ib_create_flow *ucmd) in _create_flow_rule() argument
3323 err = flow_counters_set_data(flow_act.counters, ucmd); in _create_flow_rule()
3527 struct mlx5_ib_create_flow *ucmd = NULL, ucmd_hdr; in mlx5_ib_create_flow() local
3554 ucmd = kzalloc(required_ucmd_sz, GFP_KERNEL); in mlx5_ib_create_flow()
3555 if (!ucmd) in mlx5_ib_create_flow()
3558 err = ib_copy_from_udata(ucmd, udata, required_ucmd_sz); in mlx5_ib_create_flow()
3624 dst, underlay_qpn, ucmd); in mlx5_ib_create_flow()
[all …]
/Linux-v4.19/drivers/infiniband/hw/mlx4/
Dqp.c437 struct mlx4_ib_create_qp *ucmd) in set_user_sq_size() argument
440 if ((1 << ucmd->log_sq_bb_count) > dev->dev->caps.max_wqes || in set_user_sq_size()
441 ucmd->log_sq_stride > in set_user_sq_size()
443 ucmd->log_sq_stride < MLX4_IB_MIN_SQ_STRIDE) in set_user_sq_size()
446 qp->sq.wqe_cnt = 1 << ucmd->log_sq_bb_count; in set_user_sq_size()
447 qp->sq.wqe_shift = ucmd->log_sq_stride; in set_user_sq_size()
539 struct mlx4_ib_create_qp_rss *ucmd) in set_qp_rss() argument
544 if ((ucmd->rx_hash_function == MLX4_IB_RX_HASH_FUNC_TOEPLITZ) && in set_qp_rss()
546 memcpy(rss_ctx->rss_key, ucmd->rx_hash_key, in set_qp_rss()
553 if (ucmd->rx_hash_fields_mask & ~(MLX4_IB_RX_HASH_SRC_IPV4 | in set_qp_rss()
[all …]
Dsrq.c109 struct mlx4_ib_create_srq ucmd; in mlx4_ib_create_srq() local
111 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mlx4_ib_create_srq()
116 srq->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, in mlx4_ib_create_srq()
133 ucmd.db_addr, &srq->db); in mlx4_ib_create_srq()
Dcq.c207 struct mlx4_ib_create_cq ucmd; in mlx4_ib_create_cq() local
209 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mlx4_ib_create_cq()
215 ucmd.buf_addr, entries); in mlx4_ib_create_cq()
219 err = mlx4_ib_db_map_user(to_mucontext(context), ucmd.db_addr, in mlx4_ib_create_cq()
319 struct mlx4_ib_resize_cq ucmd; in mlx4_alloc_resize_umem() local
325 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) in mlx4_alloc_resize_umem()
333 &cq->resize_umem, ucmd.buf_addr, entries); in mlx4_alloc_resize_umem()
/Linux-v4.19/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_srq.c110 struct pvrdma_create_srq ucmd; in pvrdma_create_srq() local
151 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_srq()
157 ucmd.buf_addr, in pvrdma_create_srq()
158 ucmd.buf_size, 0, 0); in pvrdma_create_srq()
Dpvrdma_cq.c118 struct pvrdma_create_cq ucmd; in pvrdma_create_cq() local
139 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_cq()
144 cq->umem = ib_umem_get(context, ucmd.buf_addr, ucmd.buf_size, in pvrdma_create_cq()
Dpvrdma_qp.c198 struct pvrdma_create_qp ucmd; in pvrdma_create_qp() local
258 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_qp()
266 ucmd.rbuf_addr, in pvrdma_create_qp()
267 ucmd.rbuf_size, 0, 0); in pvrdma_create_qp()
279 ucmd.sbuf_addr, in pvrdma_create_qp()
280 ucmd.sbuf_size, 0, 0); in pvrdma_create_qp()
/Linux-v4.19/drivers/infiniband/hw/hns/
Dhns_roce_qp.c342 struct hns_roce_ib_create_qp *ucmd) in hns_roce_set_user_sq_size() argument
350 if ((u32)(1 << ucmd->log_sq_bb_count) > hr_dev->caps.max_wqes || in hns_roce_set_user_sq_size()
351 ucmd->log_sq_stride > max_sq_stride || in hns_roce_set_user_sq_size()
352 ucmd->log_sq_stride < HNS_ROCE_IB_MIN_SQ_STRIDE) { in hns_roce_set_user_sq_size()
363 hr_qp->sq.wqe_cnt = 1 << ucmd->log_sq_bb_count; in hns_roce_set_user_sq_size()
364 hr_qp->sq.wqe_shift = ucmd->log_sq_stride; in hns_roce_set_user_sq_size()
519 struct hns_roce_ib_create_qp ucmd; in hns_roce_create_qp_common() local
578 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in hns_roce_create_qp_common()
585 &ucmd); in hns_roce_create_qp_common()
592 ucmd.buf_addr, hr_qp->buff_size, 0, in hns_roce_create_qp_common()
[all …]
Dhns_roce_cq.c317 struct hns_roce_ib_create_cq ucmd; in hns_roce_ib_create_cq() local
343 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in hns_roce_ib_create_cq()
351 &hr_cq->umem, ucmd.buf_addr, in hns_roce_ib_create_cq()
361 ucmd.db_addr, &hr_cq->db); in hns_roce_ib_create_cq()
/Linux-v4.19/arch/mips/cavium-octeon/
Docteon-platform.c34 u32 ucmd; in octeon2_usb_reset() local
41 ucmd = cvmx_read64_uint32(CVMX_UAHCX_EHCI_USBCMD); in octeon2_usb_reset()
42 ucmd &= ~CMD_RUN; in octeon2_usb_reset()
43 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset()
45 ucmd |= CMD_RESET; in octeon2_usb_reset()
46 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset()
47 ucmd = cvmx_read64_uint32(CVMX_UAHCX_OHCI_USBCMD); in octeon2_usb_reset()
48 ucmd |= CMD_RUN; in octeon2_usb_reset()
49 cvmx_write64_uint32(CVMX_UAHCX_OHCI_USBCMD, ucmd); in octeon2_usb_reset()
/Linux-v4.19/drivers/infiniband/sw/rxe/
Drxe_srq.c146 struct rxe_modify_srq_cmd *ucmd) in rxe_srq_from_attr() argument
157 mi = u64_to_user_ptr(ucmd->mmap_info_addr); in rxe_srq_from_attr()
Drxe_verbs.c389 struct rxe_modify_srq_cmd ucmd = {}; in rxe_modify_srq() local
392 if (udata->inlen < sizeof(ucmd)) in rxe_modify_srq()
395 err = ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)); in rxe_modify_srq()
404 err = rxe_srq_from_attr(rxe, srq, attr, mask, &ucmd); in rxe_modify_srq()
Drxe_loc.h233 struct rxe_modify_srq_cmd *ucmd);
/Linux-v4.19/drivers/infiniband/hw/qib/
Dqib_file_ops.c2038 const struct qib_cmd __user *ucmd; in qib_write() local
2057 ucmd = (const struct qib_cmd __user *) data; in qib_write()
2059 if (copy_from_user(&cmd.type, &ucmd->type, sizeof(cmd.type))) { in qib_write()
2071 src = &ucmd->cmd.user_info; in qib_write()
2077 src = &ucmd->cmd.recv_ctrl; in qib_write()
2083 src = &ucmd->cmd.ctxt_info; in qib_write()
2090 src = &ucmd->cmd.tid_info; in qib_write()
2096 src = &ucmd->cmd.part_key; in qib_write()
2109 src = &ucmd->cmd.poll_type; in qib_write()
2115 src = &ucmd->cmd.armlaunch_ctrl; in qib_write()
[all …]
/Linux-v4.19/drivers/block/rsxx/
Dcregs.c678 struct rsxx_reg_access __user *ucmd, in rsxx_reg_access() argument
684 st = copy_from_user(&cmd, ucmd, sizeof(cmd)); in rsxx_reg_access()
695 st = put_user(cmd.stat, &ucmd->stat); in rsxx_reg_access()
700 st = copy_to_user(ucmd->data, cmd.data, cmd.cnt); in rsxx_reg_access()
Drsxx_priv.h427 struct rsxx_reg_access __user *ucmd,
/Linux-v4.19/drivers/scsi/sym53c8xx_2/
Dsym_glue.c142 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); in sym_xpt_done() local
145 if (ucmd->eh_done) in sym_xpt_done()
146 complete(ucmd->eh_done); in sym_xpt_done()
593 struct sym_ucmd *ucmd = SYM_UCMD_PTR(cmd); in sym_eh_handler() local
673 ucmd->eh_done = &eh_done; in sym_eh_handler()
676 ucmd->eh_done = NULL; in sym_eh_handler()
/Linux-v4.19/drivers/infiniband/hw/cxgb4/
Dcq.c1002 struct c4iw_create_cq ucmd; in c4iw_create_cq() local
1020 if (udata->inlen < sizeof(ucmd)) in c4iw_create_cq()
/Linux-v4.19/drivers/nvme/host/
Dcore.c1214 struct nvme_passthru_cmd __user *ucmd) in nvme_user_cmd() argument
1224 if (copy_from_user(&cmd, ucmd, sizeof(cmd))) in nvme_user_cmd()
1253 if (put_user(cmd.result, &ucmd->result)) in nvme_user_cmd()