| /Linux-v5.15/drivers/nvme/target/ |
| D | fabrics-cmd.c | 110 u16 sqsize = le16_to_cpu(c->sqsize); in nvmet_install_queue() local 115 if (!sqsize) { in nvmet_install_queue() 117 req->error_loc = offsetof(struct nvmf_connect_command, sqsize); in nvmet_install_queue() 118 req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(sqsize); in nvmet_install_queue() 129 if (sqsize > mqes) { in nvmet_install_queue() 131 sqsize, mqes, ctrl->cntlid); in nvmet_install_queue() 132 req->error_loc = offsetof(struct nvmf_connect_command, sqsize); in nvmet_install_queue() 133 req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(sqsize); in nvmet_install_queue() 145 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); in nvmet_install_queue() 146 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); in nvmet_install_queue()
|
| D | fc.c | 132 u16 sqsize; member 633 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_prep_fcp_iodlist() 668 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_destroy_fcp_iodlist() 792 u16 qid, u16 sqsize) in nvmet_fc_alloc_target_queue() argument 800 queue = kzalloc(struct_size(queue, fod, sqsize), GFP_KERNEL); in nvmet_fc_alloc_target_queue() 814 queue->sqsize = sqsize; in nvmet_fc_alloc_target_queue() 896 for (i = 0; i < queue->sqsize; fod++, i++) { in nvmet_fc_delete_target_queue() 1676 be16_to_cpu(rqst->assoc_cmd.sqsize))) in nvmet_fc_ls_create_association() 1687 be16_to_cpu(rqst->assoc_cmd.sqsize)); in nvmet_fc_ls_create_association() 1765 be16_to_cpu(rqst->connect_cmd.sqsize))) in nvmet_fc_ls_create_connection() [all …]
|
| D | trace.c | 157 u16 sqsize = get_unaligned_le16(spc + 4); in nvmet_trace_fabrics_connect() local 162 recfmt, qid, sqsize, cattr, kato); in nvmet_trace_fabrics_connect()
|
| D | loop.c | 608 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_loop_create_ctrl()
|
| /Linux-v5.15/include/linux/ |
| D | nvme-fc.h | 260 __be16 sqsize; member 288 __be16 sqsize; member
|
| D | nvme.h | 1355 __le16 sqsize; member
|
| /Linux-v5.15/drivers/nvme/host/ |
| D | trace.c | 269 u16 sqsize = get_unaligned_le16(spc + 4); in nvme_trace_fabrics_connect() local 274 recfmt, qid, sqsize, cattr, kato); in nvme_trace_fabrics_connect()
|
| D | rdma.c | 778 ctrl->ctrl.sqsize + 1); in nvme_rdma_alloc_io_queues() 816 set->queue_depth = nctrl->sqsize + 1; in nvme_rdma_alloc_tagset() 1108 if (ctrl->ctrl.opts->queue_size > ctrl->ctrl.sqsize + 1) { in nvme_rdma_setup_ctrl() 1111 ctrl->ctrl.opts->queue_size, ctrl->ctrl.sqsize + 1); in nvme_rdma_setup_ctrl() 1114 if (ctrl->ctrl.sqsize + 1 > ctrl->ctrl.maxcmd) { in nvme_rdma_setup_ctrl() 1117 ctrl->ctrl.sqsize + 1, ctrl->ctrl.maxcmd); in nvme_rdma_setup_ctrl() 1118 ctrl->ctrl.sqsize = ctrl->ctrl.maxcmd - 1; in nvme_rdma_setup_ctrl() 1908 priv.hsqsize = cpu_to_le16(queue->ctrl->ctrl.sqsize); in nvme_rdma_route_resolved() 2366 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_rdma_create_ctrl()
|
| D | fabrics.c | 373 cmd.connect.sqsize = cpu_to_le16(NVME_AQ_DEPTH - 1); in nvmf_connect_admin_queue() 439 cmd.connect.sqsize = cpu_to_le16(ctrl->sqsize); in nvmf_connect_io_queue()
|
| D | tcp.c | 1631 set->queue_depth = nctrl->sqsize + 1; in nvme_tcp_alloc_tagset() 1719 ctrl->sqsize + 1); in __nvme_tcp_alloc_io_queues() 2018 if (opts->queue_size > ctrl->sqsize + 1) in nvme_tcp_setup_ctrl() 2021 opts->queue_size, ctrl->sqsize + 1); in nvme_tcp_setup_ctrl() 2023 if (ctrl->sqsize + 1 > ctrl->maxcmd) { in nvme_tcp_setup_ctrl() 2026 ctrl->sqsize + 1, ctrl->maxcmd); in nvme_tcp_setup_ctrl() 2027 ctrl->sqsize = ctrl->maxcmd - 1; in nvme_tcp_setup_ctrl() 2516 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_tcp_create_ctrl()
|
| D | fc.c | 1216 assoc_rqst->assoc_cmd.sqsize = cpu_to_be16(qsize - 1); in nvme_fc_connect_admin_queue() 1338 conn_rqst->connect_cmd.sqsize = cpu_to_be16(qsize - 1); in nvme_fc_connect_queue() 2901 ret = nvme_fc_create_hw_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_create_io_queues() 2905 ret = nvme_fc_connect_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_create_io_queues() 2963 ret = nvme_fc_create_hw_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_recreate_io_queues() 2967 ret = nvme_fc_connect_io_queues(ctrl, ctrl->ctrl.sqsize + 1); in nvme_fc_recreate_io_queues() 3131 if (opts->queue_size > ctrl->ctrl.sqsize + 1) { in nvme_fc_create_association() 3136 opts->queue_size, ctrl->ctrl.sqsize + 1); in nvme_fc_create_association() 3137 opts->queue_size = ctrl->ctrl.sqsize + 1; in nvme_fc_create_association() 3498 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_fc_init_ctrl()
|
| D | nvme.h | 285 u16 sqsize; member
|
| D | core.c | 3019 ctrl->sqsize = min_t(u16, NVME_CAP_MQES(ctrl->cap), ctrl->sqsize); in nvme_init_ctrl_finish() 3283 nvme_show_int_function(sqsize);
|
| D | pci.c | 2504 dev->ctrl.sqsize = dev->q_depth - 1; /* 0's based queue depth */ in nvme_pci_enable()
|
| /Linux-v5.15/drivers/infiniband/hw/irdma/ |
| D | hw.c | 900 u32 sqsize = IRDMA_CQP_SW_SQSIZE_2048; in irdma_create_cqp() local 908 cqp->cqp_requests = kcalloc(sqsize, sizeof(*cqp->cqp_requests), GFP_KERNEL); in irdma_create_cqp() 912 cqp->scratch_array = kcalloc(sqsize, sizeof(*cqp->scratch_array), GFP_KERNEL); in irdma_create_cqp() 920 cqp->sq.size = ALIGN(sizeof(struct irdma_cqp_sq_wqe) * sqsize, in irdma_create_cqp() 939 cqp_init_info.sq_size = sqsize; in irdma_create_cqp() 977 for (i = 0; i < sqsize; i++) { in irdma_create_cqp()
|
| /Linux-v5.15/drivers/infiniband/hw/cxgb4/ |
| D | qp.c | 2119 unsigned int sqsize, rqsize = 0; in c4iw_create_qp() local 2149 sqsize = attrs->cap.max_send_wr + 1; in c4iw_create_qp() 2150 if (sqsize < 8) in c4iw_create_qp() 2151 sqsize = 8; in c4iw_create_qp() 2157 qhp->wq.sq.size = sqsize; in c4iw_create_qp() 2159 (sqsize + rhp->rdev.hw_queue.t4_eq_status_entries) * in c4iw_create_qp() 2183 attrs->cap.max_send_wr = sqsize - 1; in c4iw_create_qp()
|
| /Linux-v5.15/drivers/infiniband/sw/rdmavt/ |
| D | qp.c | 1041 size_t sqsize; in rvt_create_qp() local 1067 sqsize = in rvt_create_qp() 1081 swq = vzalloc_node(array_size(sz, sqsize), rdi->dparms.node); in rvt_create_qp() 1155 qp->s_size = sqsize; in rvt_create_qp()
|