Lines Matching refs:cmdq

65 	struct cmdq		*cmdq;  member
72 struct cmdq { struct
95 struct cmdq *cmdq = container_of(chan->mbox, struct cmdq, mbox); in cmdq_get_shift_pa() argument
97 return cmdq->shift_pa; in cmdq_get_shift_pa()
101 static int cmdq_thread_suspend(struct cmdq *cmdq, struct cmdq_thread *thread) in cmdq_thread_suspend() argument
113 dev_err(cmdq->mbox.dev, "suspend GCE thread 0x%x failed\n", in cmdq_thread_suspend()
114 (u32)(thread->base - cmdq->base)); in cmdq_thread_suspend()
126 static void cmdq_init(struct cmdq *cmdq) in cmdq_init() argument
130 WARN_ON(clk_bulk_enable(cmdq->gce_num, cmdq->clocks)); in cmdq_init()
131 if (cmdq->control_by_sw) in cmdq_init()
132 writel(0x7, cmdq->base + GCE_GCTL_VALUE); in cmdq_init()
133 writel(CMDQ_THR_ACTIVE_SLOT_CYCLES, cmdq->base + CMDQ_THR_SLOT_CYCLES); in cmdq_init()
135 writel(i, cmdq->base + CMDQ_SYNC_TOKEN_UPDATE); in cmdq_init()
136 clk_bulk_disable(cmdq->gce_num, cmdq->clocks); in cmdq_init()
139 static int cmdq_thread_reset(struct cmdq *cmdq, struct cmdq_thread *thread) in cmdq_thread_reset() argument
147 dev_err(cmdq->mbox.dev, "reset GCE thread 0x%x failed\n", in cmdq_thread_reset()
148 (u32)(thread->base - cmdq->base)); in cmdq_thread_reset()
155 static void cmdq_thread_disable(struct cmdq *cmdq, struct cmdq_thread *thread) in cmdq_thread_disable() argument
157 cmdq_thread_reset(cmdq, thread); in cmdq_thread_disable()
170 struct device *dev = task->cmdq->mbox.dev; in cmdq_task_insert_into_thread()
181 (task->pa_base >> task->cmdq->shift_pa); in cmdq_task_insert_into_thread()
214 struct cmdq *cmdq = task->cmdq; in cmdq_task_handle_error() local
216 dev_err(cmdq->mbox.dev, "task 0x%p error\n", task); in cmdq_task_handle_error()
217 WARN_ON(cmdq_thread_suspend(cmdq, thread) < 0); in cmdq_task_handle_error()
221 writel(next_task->pa_base >> cmdq->shift_pa, in cmdq_task_handle_error()
226 static void cmdq_thread_irq_handler(struct cmdq *cmdq, in cmdq_thread_irq_handler() argument
252 curr_pa = readl(thread->base + CMDQ_THR_CURR_ADDR) << cmdq->shift_pa; in cmdq_thread_irq_handler()
274 cmdq_thread_disable(cmdq, thread); in cmdq_thread_irq_handler()
275 clk_bulk_disable(cmdq->gce_num, cmdq->clocks); in cmdq_thread_irq_handler()
281 struct cmdq *cmdq = dev; in cmdq_irq_handler() local
285 irq_status = readl(cmdq->base + CMDQ_CURR_IRQ_STATUS) & cmdq->irq_mask; in cmdq_irq_handler()
286 if (!(irq_status ^ cmdq->irq_mask)) in cmdq_irq_handler()
289 for_each_clear_bit(bit, &irq_status, cmdq->thread_nr) { in cmdq_irq_handler()
290 struct cmdq_thread *thread = &cmdq->thread[bit]; in cmdq_irq_handler()
293 cmdq_thread_irq_handler(cmdq, thread); in cmdq_irq_handler()
302 struct cmdq *cmdq = dev_get_drvdata(dev); in cmdq_suspend() local
307 cmdq->suspended = true; in cmdq_suspend()
309 for (i = 0; i < cmdq->thread_nr; i++) { in cmdq_suspend()
310 thread = &cmdq->thread[i]; in cmdq_suspend()
320 clk_bulk_unprepare(cmdq->gce_num, cmdq->clocks); in cmdq_suspend()
327 struct cmdq *cmdq = dev_get_drvdata(dev); in cmdq_resume() local
329 WARN_ON(clk_bulk_prepare(cmdq->gce_num, cmdq->clocks)); in cmdq_resume()
330 cmdq->suspended = false; in cmdq_resume()
336 struct cmdq *cmdq = platform_get_drvdata(pdev); in cmdq_remove() local
338 clk_bulk_unprepare(cmdq->gce_num, cmdq->clocks); in cmdq_remove()
346 struct cmdq *cmdq = dev_get_drvdata(chan->mbox->dev); in cmdq_mbox_send_data() local
351 WARN_ON(cmdq->suspended); in cmdq_mbox_send_data()
357 task->cmdq = cmdq; in cmdq_mbox_send_data()
364 WARN_ON(clk_bulk_enable(cmdq->gce_num, cmdq->clocks)); in cmdq_mbox_send_data()
372 WARN_ON(cmdq_thread_reset(cmdq, thread) < 0); in cmdq_mbox_send_data()
374 writel(task->pa_base >> cmdq->shift_pa, in cmdq_mbox_send_data()
376 writel((task->pa_base + pkt->cmd_buf_size) >> cmdq->shift_pa, in cmdq_mbox_send_data()
383 WARN_ON(cmdq_thread_suspend(cmdq, thread) < 0); in cmdq_mbox_send_data()
385 cmdq->shift_pa; in cmdq_mbox_send_data()
387 cmdq->shift_pa; in cmdq_mbox_send_data()
392 writel(task->pa_base >> cmdq->shift_pa, in cmdq_mbox_send_data()
398 writel((task->pa_base + pkt->cmd_buf_size) >> cmdq->shift_pa, in cmdq_mbox_send_data()
415 struct cmdq *cmdq = dev_get_drvdata(chan->mbox->dev); in cmdq_mbox_shutdown() local
423 WARN_ON(cmdq_thread_suspend(cmdq, thread) < 0); in cmdq_mbox_shutdown()
426 cmdq_thread_irq_handler(cmdq, thread); in cmdq_mbox_shutdown()
436 cmdq_thread_disable(cmdq, thread); in cmdq_mbox_shutdown()
437 clk_bulk_disable(cmdq->gce_num, cmdq->clocks); in cmdq_mbox_shutdown()
454 struct cmdq *cmdq = dev_get_drvdata(chan->mbox->dev); in cmdq_mbox_flush() local
463 WARN_ON(cmdq_thread_suspend(cmdq, thread) < 0); in cmdq_mbox_flush()
482 cmdq_thread_disable(cmdq, thread); in cmdq_mbox_flush()
483 clk_bulk_disable(cmdq->gce_num, cmdq->clocks); in cmdq_mbox_flush()
494 dev_err(cmdq->mbox.dev, "Fail to wait GCE thread 0x%x done\n", in cmdq_mbox_flush()
495 (u32)(thread->base - cmdq->base)); in cmdq_mbox_flush()
529 struct cmdq *cmdq; in cmdq_probe() local
537 cmdq = devm_kzalloc(dev, sizeof(*cmdq), GFP_KERNEL); in cmdq_probe()
538 if (!cmdq) in cmdq_probe()
542 cmdq->base = devm_ioremap_resource(dev, res); in cmdq_probe()
543 if (IS_ERR(cmdq->base)) in cmdq_probe()
544 return PTR_ERR(cmdq->base); in cmdq_probe()
546 cmdq->irq = platform_get_irq(pdev, 0); in cmdq_probe()
547 if (cmdq->irq < 0) in cmdq_probe()
548 return cmdq->irq; in cmdq_probe()
556 cmdq->thread_nr = plat_data->thread_nr; in cmdq_probe()
557 cmdq->shift_pa = plat_data->shift; in cmdq_probe()
558 cmdq->control_by_sw = plat_data->control_by_sw; in cmdq_probe()
559 cmdq->gce_num = plat_data->gce_num; in cmdq_probe()
560 cmdq->irq_mask = GENMASK(cmdq->thread_nr - 1, 0); in cmdq_probe()
561 err = devm_request_irq(dev, cmdq->irq, cmdq_irq_handler, IRQF_SHARED, in cmdq_probe()
562 "mtk_cmdq", cmdq); in cmdq_probe()
569 dev, cmdq->base, cmdq->irq); in cmdq_probe()
571 if (cmdq->gce_num > 1) { in cmdq_probe()
576 if (alias_id < cmdq->gce_num) { in cmdq_probe()
578 cmdq->clocks[alias_id].id = clk_id; in cmdq_probe()
579 cmdq->clocks[alias_id].clk = of_clk_get(node, 0); in cmdq_probe()
580 if (IS_ERR(cmdq->clocks[alias_id].clk)) { in cmdq_probe()
582 return PTR_ERR(cmdq->clocks[alias_id].clk); in cmdq_probe()
587 cmdq->clocks[alias_id].id = clk_name; in cmdq_probe()
588 cmdq->clocks[alias_id].clk = devm_clk_get(&pdev->dev, clk_name); in cmdq_probe()
589 if (IS_ERR(cmdq->clocks[alias_id].clk)) { in cmdq_probe()
591 return PTR_ERR(cmdq->clocks[alias_id].clk); in cmdq_probe()
595 cmdq->mbox.dev = dev; in cmdq_probe()
596 cmdq->mbox.chans = devm_kcalloc(dev, cmdq->thread_nr, in cmdq_probe()
597 sizeof(*cmdq->mbox.chans), GFP_KERNEL); in cmdq_probe()
598 if (!cmdq->mbox.chans) in cmdq_probe()
601 cmdq->mbox.num_chans = cmdq->thread_nr; in cmdq_probe()
602 cmdq->mbox.ops = &cmdq_mbox_chan_ops; in cmdq_probe()
603 cmdq->mbox.of_xlate = cmdq_xlate; in cmdq_probe()
606 cmdq->mbox.txdone_irq = false; in cmdq_probe()
607 cmdq->mbox.txdone_poll = false; in cmdq_probe()
609 cmdq->thread = devm_kcalloc(dev, cmdq->thread_nr, in cmdq_probe()
610 sizeof(*cmdq->thread), GFP_KERNEL); in cmdq_probe()
611 if (!cmdq->thread) in cmdq_probe()
614 for (i = 0; i < cmdq->thread_nr; i++) { in cmdq_probe()
615 cmdq->thread[i].base = cmdq->base + CMDQ_THR_BASE + in cmdq_probe()
617 INIT_LIST_HEAD(&cmdq->thread[i].task_busy_list); in cmdq_probe()
618 cmdq->mbox.chans[i].con_priv = (void *)&cmdq->thread[i]; in cmdq_probe()
621 err = devm_mbox_controller_register(dev, &cmdq->mbox); in cmdq_probe()
627 platform_set_drvdata(pdev, cmdq); in cmdq_probe()
629 WARN_ON(clk_bulk_prepare(cmdq->gce_num, cmdq->clocks)); in cmdq_probe()
631 cmdq_init(cmdq); in cmdq_probe()