Home
last modified time | relevance | path

Searched refs:rq_depth (Results 1 – 25 of 28) sorted by relevance

12

/Linux-v6.1/block/
Dblk-wbt.c238 struct rq_depth *rqd = &rwb->rq_depth; in latency_exceeded()
291 struct rq_depth *rqd = &rwb->rq_depth; in rwb_trace_step()
301 } else if (rwb->rq_depth.max_depth <= 2) { in calc_wb_limits()
302 rwb->wb_normal = rwb->rq_depth.max_depth; in calc_wb_limits()
305 rwb->wb_normal = (rwb->rq_depth.max_depth + 1) / 2; in calc_wb_limits()
306 rwb->wb_background = (rwb->rq_depth.max_depth + 3) / 4; in calc_wb_limits()
312 if (!rq_depth_scale_up(&rwb->rq_depth)) in scale_up()
322 if (!rq_depth_scale_down(&rwb->rq_depth, hard_throttle)) in scale_down()
331 struct rq_depth *rqd = &rwb->rq_depth; in rwb_arm_timer()
356 struct rq_depth *rqd = &rwb->rq_depth; in wb_timer_fn()
[all …]
Dblk-iolatency.c144 struct rq_depth rq_depth; member
283 return rq_wait_inc_below(rqw, iolat->rq_depth.max_depth); in iolat_acquire_inflight()
375 unsigned long old = iolat->rq_depth.max_depth; in scale_change()
387 iolat->rq_depth.max_depth = old; in scale_change()
392 iolat->rq_depth.max_depth = max(old, 1UL); in scale_change()
448 if (iolat->rq_depth.max_depth == 1 && direction < 0) { in check_scale_change()
456 iolat->rq_depth.max_depth = UINT_MAX; in check_scale_change()
511 if (unlikely(issue_as_root && iolat->rq_depth.max_depth != UINT_MAX)) { in iolatency_record_time()
923 if (iolat->rq_depth.max_depth == UINT_MAX) in iolatency_ssd_stat()
931 iolat->rq_depth.max_depth); in iolatency_ssd_stat()
[all …]
Dblk-rq-qos.h51 struct rq_depth { struct
155 bool rq_depth_scale_up(struct rq_depth *rqd);
156 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle);
157 bool rq_depth_calc_max_depth(struct rq_depth *rqd);
Dblk-wbt.h70 struct rq_depth rq_depth; member
Dblk-rq-qos.c110 bool rq_depth_calc_max_depth(struct rq_depth *rqd) in rq_depth_calc_max_depth()
158 bool rq_depth_scale_up(struct rq_depth *rqd) in rq_depth_scale_up()
177 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle) in rq_depth_scale_down()
/Linux-v6.1/net/9p/
Dtrans_rdma.c86 int rq_depth; member
127 int rq_depth; member
159 if (rdma->rq_depth != P9_RDMA_RQ_DEPTH) in p9_rdma_show_options()
160 seq_printf(m, ",rq=%u", rdma->rq_depth); in p9_rdma_show_options()
184 opts->rq_depth = P9_RDMA_RQ_DEPTH; in parse_opts()
221 opts->rq_depth = option; in parse_opts()
234 opts->rq_depth = max(opts->rq_depth, opts->sq_depth); in parse_opts()
574 rdma->rq_depth = opts->rq_depth; in alloc_rdma()
579 sema_init(&rdma->rq_sem, rdma->rq_depth); in alloc_rdma()
690 opts.sq_depth + opts.rq_depth + 1, in rdma_create_trans()
[all …]
/Linux-v6.1/drivers/net/ethernet/fungible/funcore/
Dfun_queue.c175 for (i = 0; i < funq->rq_depth; i++) { in fun_clean_rq()
192 for (i = 0; i < funq->rq_depth; i++) { in fun_fill_rq()
209 funq->rq_tail = funq->rq_depth - 1; in fun_fill_rq()
222 if (++funq->rq_buf_idx == funq->rq_depth) in fun_rq_update_pos()
344 funq->rq_depth; in fun_process_cq()
375 funq->rqes = fun_alloc_ring_mem(funq->fdev->dev, funq->rq_depth, in fun_alloc_rqes()
396 fun_free_ring_mem(dev, funq->rq_depth, sizeof(*funq->rqes), in fun_free_queue()
419 if (req->rq_depth) { in fun_alloc_queue()
458 if (req->rq_depth) { in fun_alloc_queue()
460 funq->rq_depth = req->rq_depth; in fun_alloc_queue()
[all …]
Dfun_queue.h47 u32 rq_depth; member
124 u32 rq_depth; member
Dfun_dev.c233 .rq_depth = areq->rq_depth, in fun_enable_admin_queue()
281 if (areq->rq_depth) { in fun_enable_admin_queue()
583 if (cq_count < 2 || sq_count < 2 + !!fdev->admin_q->rq_depth) in fun_get_dev_limits()
Dfun_dev.h94 u16 rq_depth; member
/Linux-v6.1/net/sunrpc/xprtrdma/
Dsvc_rdma_transport.c373 unsigned int ctxts, rq_depth; in svc_rdma_accept() local
408 rq_depth = newxprt->sc_max_requests + newxprt->sc_max_bc_requests + in svc_rdma_accept()
410 if (rq_depth > dev->attrs.max_qp_wr) { in svc_rdma_accept()
413 rq_depth = dev->attrs.max_qp_wr; in svc_rdma_accept()
415 newxprt->sc_max_requests = rq_depth - 2; in svc_rdma_accept()
421 newxprt->sc_sq_depth = rq_depth + ctxts; in svc_rdma_accept()
439 ib_alloc_cq_any(dev, newxprt, rq_depth, IB_POLL_WORKQUEUE); in svc_rdma_accept()
449 qp_attr.cap.max_recv_wr = rq_depth; in svc_rdma_accept()
/Linux-v6.1/drivers/net/ethernet/fungible/funeth/
Dfuneth.h76 unsigned int rq_depth; member
116 unsigned int rq_depth; member
Dfuneth_ethtool.c566 ring->rx_pending = fp->rq_depth; in fun_get_ringparam()
594 fp->rq_depth == ring->rx_pending) in fun_set_ringparam()
600 .rq_depth = ring->rx_pending, in fun_set_ringparam()
610 fp->rq_depth = ring->rx_pending; in fun_set_ringparam()
611 fp->cq_depth = 2 * fp->rq_depth; in fun_set_ringparam()
Dfuneth_main.c509 qset->rq_depth, qset->rxq_start, qset->state); in fun_alloc_rings()
842 .rq_depth = fp->rq_depth, in funeth_open()
1646 .rq_depth = fp->rq_depth, in fun_change_num_queues()
1784 fp->rq_depth = min_t(unsigned int, RQ_DEPTH, fdev->q_depth); in fun_create_netdev()
2008 .rq_depth = ADMIN_RQ_DEPTH, in funeth_probe()
/Linux-v6.1/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_dev.c271 unsigned int rq_depth) in set_hw_ioctxt() argument
285 hw_ioctxt.rq_depth = ilog2(rq_depth); in set_hw_ioctxt()
438 int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth) in hinic_hwdev_ifup() argument
462 func_to_io->rq_depth = rq_depth; in hinic_hwdev_ifup()
489 err = set_hw_ioctxt(hwdev, sq_depth, rq_depth); in hinic_hwdev_ifup()
Dhinic_dev.h96 u16 rq_depth; member
Dhinic_hw_io.h74 u16 rq_depth; member
Dhinic_hw_dev.h220 u16 rq_depth; member
561 int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth);
Dhinic_ethtool.c557 ring->rx_pending = nic_dev->rq_depth; in hinic_get_ringparam()
600 new_rq_depth == nic_dev->rq_depth) in hinic_set_ringparam()
605 nic_dev->sq_depth, nic_dev->rq_depth, in hinic_set_ringparam()
609 nic_dev->rq_depth = new_rq_depth; in hinic_set_ringparam()
Dhinic_hw_mbox.c1289 ((hw_ctxt)->rq_depth >= HINIC_QUEUE_MIN_DEPTH && \
1290 (hw_ctxt)->rq_depth <= HINIC_QUEUE_MAX_DEPTH && \
1300 if (!hw_ctxt->rq_depth && !hw_ctxt->sq_depth && in hw_ctxt_qps_param_valid()
Dhinic_hw_io.c293 func_to_io->rq_depth, HINIC_RQ_WQE_SIZE); in init_qp()
Dhinic_port.c476 rq_num.rq_depth = ilog2(nic_dev->rq_depth); in hinic_set_max_qnum()
Dhinic_main.c423 nic_dev->rq_depth); in hinic_open()
1202 nic_dev->rq_depth = HINIC_RQ_DEPTH; in nic_dev_init()
Dhinic_port.h317 u32 rq_depth; member
/Linux-v6.1/drivers/infiniband/hw/efa/
Defa_com_cmd.h26 u32 rq_depth; member

12