Lines Matching refs:queue_depth

2329 		for (i = 0; i < set->queue_depth; i++) {  in blk_mq_clear_rq_mapping()
2652 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() argument
2663 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping()
2687 set->queue_depth, flush_rq); in blk_mq_exit_hctx()
2863 set->queue_depth, set->reserved_tags, flags); in __blk_mq_alloc_map_and_request()
2868 set->queue_depth); in __blk_mq_alloc_map_and_request()
3311 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()
3376 depth = set->queue_depth; in blk_mq_alloc_map_and_requests()
3382 set->queue_depth >>= 1; in blk_mq_alloc_map_and_requests()
3383 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) { in blk_mq_alloc_map_and_requests()
3387 } while (set->queue_depth); in blk_mq_alloc_map_and_requests()
3389 if (!set->queue_depth || err) { in blk_mq_alloc_map_and_requests()
3394 if (depth != set->queue_depth) in blk_mq_alloc_map_and_requests()
3396 depth, set->queue_depth); in blk_mq_alloc_map_and_requests()
3481 if (!set->queue_depth) in blk_mq_alloc_tag_set()
3483 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) in blk_mq_alloc_tag_set()
3492 if (set->queue_depth > BLK_MQ_MAX_DEPTH) { in blk_mq_alloc_tag_set()
3495 set->queue_depth = BLK_MQ_MAX_DEPTH; in blk_mq_alloc_tag_set()
3511 set->queue_depth = min(64U, set->queue_depth); in blk_mq_alloc_tag_set()
3571 const struct blk_mq_ops *ops, unsigned int queue_depth, in blk_mq_alloc_sq_tag_set() argument
3578 set->queue_depth = queue_depth; in blk_mq_alloc_sq_tag_set()