Lines Matching refs:nq

624 static void put_tag(struct nullb_queue *nq, unsigned int tag)  in put_tag()  argument
626 clear_bit_unlock(tag, nq->tag_map); in put_tag()
628 if (waitqueue_active(&nq->wait)) in put_tag()
629 wake_up(&nq->wait); in put_tag()
632 static unsigned int get_tag(struct nullb_queue *nq) in get_tag() argument
637 tag = find_first_zero_bit(nq->tag_map, nq->queue_depth); in get_tag()
638 if (tag >= nq->queue_depth) in get_tag()
640 } while (test_and_set_bit_lock(tag, nq->tag_map)); in get_tag()
647 put_tag(cmd->nq, cmd->tag); in free_cmd()
652 static struct nullb_cmd *__alloc_cmd(struct nullb_queue *nq) in __alloc_cmd() argument
657 tag = get_tag(nq); in __alloc_cmd()
659 cmd = &nq->cmds[tag]; in __alloc_cmd()
662 cmd->nq = nq; in __alloc_cmd()
663 if (nq->dev->irqmode == NULL_IRQ_TIMER) { in __alloc_cmd()
674 static struct nullb_cmd *alloc_cmd(struct nullb_queue *nq, int can_wait) in alloc_cmd() argument
679 cmd = __alloc_cmd(nq); in alloc_cmd()
684 prepare_to_wait(&nq->wait, &wait, TASK_UNINTERRUPTIBLE); in alloc_cmd()
685 cmd = __alloc_cmd(nq); in alloc_cmd()
692 finish_wait(&nq->wait, &wait); in alloc_cmd()
698 int queue_mode = cmd->nq->dev->queue_mode; in end_cmd()
722 ktime_t kt = cmd->nq->dev->completion_nsec; in null_cmd_end_timer()
1164 struct nullb *nullb = cmd->nq->dev->nullb; in null_handle_rq()
1191 struct nullb *nullb = cmd->nq->dev->nullb; in null_handle_bio()
1232 struct nullb_device *dev = cmd->nq->dev; in null_handle_throttled()
1255 struct badblocks *bb = &cmd->nq->dev->badblocks; in null_handle_badblocks()
1270 struct nullb_device *dev = cmd->nq->dev; in null_handle_memory_backed()
1286 struct nullb_device *dev = cmd->nq->dev; in nullb_zero_read_cmd_buffer()
1313 switch (cmd->nq->dev->irqmode) { in nullb_complete_cmd()
1315 switch (cmd->nq->dev->queue_mode) { in nullb_complete_cmd()
1341 struct nullb_device *dev = cmd->nq->dev; in null_process_cmd()
1359 struct nullb_device *dev = cmd->nq->dev; in null_handle_cmd()
1430 struct nullb_queue *nq = nullb_to_queue(nullb); in null_submit_bio() local
1433 cmd = alloc_cmd(nq, 1); in null_submit_bio()
1481 struct nullb_queue *nq = hctx->driver_data; in null_queue_rq() local
1487 if (nq->dev->irqmode == NULL_IRQ_TIMER) { in null_queue_rq()
1493 cmd->nq = nq; in null_queue_rq()
1503 nq->requeue_selection++; in null_queue_rq()
1504 if (nq->requeue_selection & 1) in null_queue_rq()
1517 static void cleanup_queue(struct nullb_queue *nq) in cleanup_queue() argument
1519 kfree(nq->tag_map); in cleanup_queue()
1520 kfree(nq->cmds); in cleanup_queue()
1535 struct nullb_queue *nq = hctx->driver_data; in null_exit_hctx() local
1536 struct nullb *nullb = nq->dev->nullb; in null_exit_hctx()
1541 static void null_init_queue(struct nullb *nullb, struct nullb_queue *nq) in null_init_queue() argument
1543 init_waitqueue_head(&nq->wait); in null_init_queue()
1544 nq->queue_depth = nullb->queue_depth; in null_init_queue()
1545 nq->dev = nullb->dev; in null_init_queue()
1552 struct nullb_queue *nq; in null_init_hctx() local
1559 nq = &nullb->queues[hctx_idx]; in null_init_hctx()
1560 hctx->driver_data = nq; in null_init_hctx()
1561 null_init_queue(nullb, nq); in null_init_hctx()
1641 static int setup_commands(struct nullb_queue *nq) in setup_commands() argument
1646 nq->cmds = kcalloc(nq->queue_depth, sizeof(*cmd), GFP_KERNEL); in setup_commands()
1647 if (!nq->cmds) in setup_commands()
1650 tag_size = ALIGN(nq->queue_depth, BITS_PER_LONG) / BITS_PER_LONG; in setup_commands()
1651 nq->tag_map = kcalloc(tag_size, sizeof(unsigned long), GFP_KERNEL); in setup_commands()
1652 if (!nq->tag_map) { in setup_commands()
1653 kfree(nq->cmds); in setup_commands()
1657 for (i = 0; i < nq->queue_depth; i++) { in setup_commands()
1658 cmd = &nq->cmds[i]; in setup_commands()
1679 struct nullb_queue *nq; in init_driver_queues() local
1683 nq = &nullb->queues[i]; in init_driver_queues()
1685 null_init_queue(nullb, nq); in init_driver_queues()
1687 ret = setup_commands(nq); in init_driver_queues()