Home
last modified time | relevance | path

Searched refs:cmdq (Results 1 – 25 of 32) sorted by relevance

12

/Linux-v5.4/drivers/crypto/cavium/nitrox/
Dnitrox_lib.c25 static int nitrox_cmdq_init(struct nitrox_cmdq *cmdq, int align_bytes) in nitrox_cmdq_init() argument
27 struct nitrox_device *ndev = cmdq->ndev; in nitrox_cmdq_init()
29 cmdq->qsize = (ndev->qlen * cmdq->instr_size) + align_bytes; in nitrox_cmdq_init()
30 cmdq->unalign_base = dma_alloc_coherent(DEV(ndev), cmdq->qsize, in nitrox_cmdq_init()
31 &cmdq->unalign_dma, in nitrox_cmdq_init()
33 if (!cmdq->unalign_base) in nitrox_cmdq_init()
36 cmdq->dma = PTR_ALIGN(cmdq->unalign_dma, align_bytes); in nitrox_cmdq_init()
37 cmdq->base = cmdq->unalign_base + (cmdq->dma - cmdq->unalign_dma); in nitrox_cmdq_init()
38 cmdq->write_idx = 0; in nitrox_cmdq_init()
40 spin_lock_init(&cmdq->cmd_qlock); in nitrox_cmdq_init()
[all …]
Dnitrox_reqmgr.c228 struct nitrox_cmdq *cmdq) in backlog_list_add() argument
232 spin_lock_bh(&cmdq->backlog_qlock); in backlog_list_add()
233 list_add_tail(&sr->backlog, &cmdq->backlog_head); in backlog_list_add()
234 atomic_inc(&cmdq->backlog_count); in backlog_list_add()
236 spin_unlock_bh(&cmdq->backlog_qlock); in backlog_list_add()
240 struct nitrox_cmdq *cmdq) in response_list_add() argument
244 spin_lock_bh(&cmdq->resp_qlock); in response_list_add()
245 list_add_tail(&sr->response, &cmdq->response_head); in response_list_add()
246 spin_unlock_bh(&cmdq->resp_qlock); in response_list_add()
250 struct nitrox_cmdq *cmdq) in response_list_del() argument
[all …]
Dnitrox_isr.c31 struct nitrox_cmdq *cmdq = qvec->cmdq; in nps_pkt_slc_isr() local
33 slc_cnts.value = readq(cmdq->compl_cnt_csr_addr); in nps_pkt_slc_isr()
332 qvec->cmdq = &ndev->pkt_inq[qvec->ring]; in nitrox_register_interrupts()
Dnitrox_hal.c124 struct nitrox_cmdq *cmdq = &ndev->pkt_inq[i]; in nitrox_config_pkt_input_rings() local
137 nitrox_write_csr(ndev, offset, cmdq->dma); in nitrox_config_pkt_input_rings()
356 struct nitrox_cmdq *cmdq = ndev->aqmq[ring]; in nitrox_config_aqm_rings() local
379 nitrox_write_csr(ndev, offset, cmdq->dma); in nitrox_config_aqm_rings()
Dnitrox_dev.h107 struct nitrox_cmdq *cmdq; member
Dnitrox_req.h513 struct nitrox_cmdq *cmdq; member
/Linux-v5.4/drivers/mailbox/
Dmtk-cmdq-mailbox.c63 struct cmdq *cmdq; member
70 struct cmdq { struct
81 static int cmdq_thread_suspend(struct cmdq *cmdq, struct cmdq_thread *thread) in cmdq_thread_suspend() argument
93 dev_err(cmdq->mbox.dev, "suspend GCE thread 0x%x failed\n", in cmdq_thread_suspend()
94 (u32)(thread->base - cmdq->base)); in cmdq_thread_suspend()
106 static void cmdq_init(struct cmdq *cmdq) in cmdq_init() argument
110 WARN_ON(clk_enable(cmdq->clock) < 0); in cmdq_init()
111 writel(CMDQ_THR_ACTIVE_SLOT_CYCLES, cmdq->base + CMDQ_THR_SLOT_CYCLES); in cmdq_init()
113 writel(i, cmdq->base + CMDQ_SYNC_TOKEN_UPDATE); in cmdq_init()
114 clk_disable(cmdq->clock); in cmdq_init()
[all …]
DMakefile48 obj-$(CONFIG_MTK_CMDQ_MBOX) += mtk-cmdq-mailbox.o
/Linux-v5.4/drivers/net/ethernet/brocade/bna/
Dbfa_msgq.c31 static void bfa_msgq_cmdq_dbell(struct bfa_msgq_cmdq *cmdq);
32 static void bfa_msgq_cmdq_copy_rsp(struct bfa_msgq_cmdq *cmdq);
43 bfa_fsm_state_decl(cmdq, stopped, struct bfa_msgq_cmdq, enum cmdq_event);
44 bfa_fsm_state_decl(cmdq, init_wait, struct bfa_msgq_cmdq, enum cmdq_event);
45 bfa_fsm_state_decl(cmdq, ready, struct bfa_msgq_cmdq, enum cmdq_event);
46 bfa_fsm_state_decl(cmdq, dbell_wait, struct bfa_msgq_cmdq,
50 cmdq_sm_stopped_entry(struct bfa_msgq_cmdq *cmdq) in cmdq_sm_stopped_entry() argument
54 cmdq->producer_index = 0; in cmdq_sm_stopped_entry()
55 cmdq->consumer_index = 0; in cmdq_sm_stopped_entry()
56 cmdq->flags = 0; in cmdq_sm_stopped_entry()
[all …]
Dbfa_msgq.h104 struct bfa_msgq_cmdq cmdq; member
Dbfi.h420 struct bfi_msgq cmdq; member
/Linux-v5.4/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_cmdq.c78 #define cmdq_to_cmdqs(cmdq) container_of((cmdq) - (cmdq)->cmdq_type, \ argument
79 struct hinic_cmdqs, cmdq[0])
325 static void cmdq_set_db(struct hinic_cmdq *cmdq, in cmdq_set_db() argument
337 writel(db_info, CMDQ_DB_ADDR(cmdq->db_base, prod_idx)); in cmdq_set_db()
340 static int cmdq_sync_cmd_direct_resp(struct hinic_cmdq *cmdq, in cmdq_sync_cmd_direct_resp() argument
348 struct hinic_wq *wq = cmdq->wq; in cmdq_sync_cmd_direct_resp()
353 spin_lock_bh(&cmdq->cmdq_lock); in cmdq_sync_cmd_direct_resp()
358 spin_unlock_bh(&cmdq->cmdq_lock); in cmdq_sync_cmd_direct_resp()
364 wrapped = cmdq->wrapped; in cmdq_sync_cmd_direct_resp()
369 cmdq->wrapped = !cmdq->wrapped; in cmdq_sync_cmd_direct_resp()
[all …]
Dhinic_hw_io.c502 enum hinic_cmdq_type cmdq, type; in hinic_io_init() local
533 for (cmdq = HINIC_CMDQ_SYNC; cmdq < HINIC_MAX_CMDQ_TYPES; cmdq++) { in hinic_io_init()
541 func_to_io->cmdq_db_area[cmdq] = db_area; in hinic_io_init()
555 for (type = HINIC_CMDQ_SYNC; type < cmdq; type++) in hinic_io_init()
574 enum hinic_cmdq_type cmdq; in hinic_io_free() local
578 for (cmdq = HINIC_CMDQ_SYNC; cmdq < HINIC_MAX_CMDQ_TYPES; cmdq++) in hinic_io_free()
579 return_db_area(func_to_io, func_to_io->cmdq_db_area[cmdq]); in hinic_io_free()
Dhinic_hw_cmdq.h157 struct hinic_cmdq cmdq[HINIC_MAX_CMDQ_TYPES]; member
/Linux-v5.4/drivers/infiniband/hw/bnxt_re/
Dqplib_rcfw.c88 struct bnxt_qplib_hwq *cmdq = &rcfw->cmdq; in __send_message() local
119 spin_lock_irqsave(&cmdq->lock, flags); in __send_message()
120 if (req->cmd_size >= HWQ_FREE_SLOTS(cmdq)) { in __send_message()
122 spin_unlock_irqrestore(&cmdq->lock, flags); in __send_message()
136 spin_unlock_irqrestore(&cmdq->lock, flags); in __send_message()
158 cmdq_ptr = (struct bnxt_qplib_cmdqe **)cmdq->pbl_ptr; in __send_message()
162 sw_prod = HWQ_CMP(cmdq->prod, cmdq); in __send_message()
175 cmdq->prod++; in __send_message()
181 cmdq_prod = cmdq->prod; in __send_message()
199 spin_unlock_irqrestore(&cmdq->lock, flags); in __send_message()
[all …]
Dqplib_rcfw.h261 struct bnxt_qplib_hwq cmdq; member
/Linux-v5.4/drivers/iommu/
Darm-smmu-v3.c604 struct arm_smmu_cmdq cmdq; member
917 struct arm_smmu_queue *q = &smmu->cmdq.q; in arm_smmu_cmdq_build_sync_cmd()
946 struct arm_smmu_queue *q = &smmu->cmdq.q; in arm_smmu_cmdq_skip_err()
1005 static void arm_smmu_cmdq_shared_lock(struct arm_smmu_cmdq *cmdq) in arm_smmu_cmdq_shared_lock() argument
1015 if (atomic_fetch_inc_relaxed(&cmdq->lock) >= 0) in arm_smmu_cmdq_shared_lock()
1019 val = atomic_cond_read_relaxed(&cmdq->lock, VAL >= 0); in arm_smmu_cmdq_shared_lock()
1020 } while (atomic_cmpxchg_relaxed(&cmdq->lock, val, val + 1) != val); in arm_smmu_cmdq_shared_lock()
1023 static void arm_smmu_cmdq_shared_unlock(struct arm_smmu_cmdq *cmdq) in arm_smmu_cmdq_shared_unlock() argument
1025 (void)atomic_dec_return_release(&cmdq->lock); in arm_smmu_cmdq_shared_unlock()
1028 static bool arm_smmu_cmdq_shared_tryunlock(struct arm_smmu_cmdq *cmdq) in arm_smmu_cmdq_shared_tryunlock() argument
[all …]
/Linux-v5.4/drivers/atm/
Dfore200e.c558 struct host_cmdq* cmdq = &fore200e->host_cmdq; in fore200e_pca_prom_read() local
559 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_pca_prom_read()
564 FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD); in fore200e_pca_prom_read()
1227 struct host_cmdq* cmdq = &fore200e->host_cmdq; in fore200e_activate_vcin() local
1228 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_activate_vcin()
1235 FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD); in fore200e_activate_vcin()
1660 struct host_cmdq* cmdq = &fore200e->host_cmdq; in fore200e_getstats() local
1661 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_getstats()
1677 FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD); in fore200e_getstats()
1731 struct host_cmdq* cmdq = &fore200e->host_cmdq;
[all …]
/Linux-v5.4/drivers/soc/mediatek/
DMakefile2 obj-$(CONFIG_MTK_CMDQ) += mtk-cmdq-helper.o
/Linux-v5.4/Documentation/devicetree/bindings/iommu/
Darm,smmu-v3.txt27 * "cmdq-sync" - CMD_SYNC complete
73 interrupt-names = "eventq", "priq", "cmdq-sync", "gerror";
/Linux-v5.4/drivers/scsi/aacraid/
Ddpcsup.c199 list_add_tail(&fib->fiblink, &q->cmdq); in aac_command_normal()
310 list_add_tail(&fib->fiblink, &q->cmdq); in aac_intr_normal()
Dcomminit.c264 INIT_LIST_HEAD(&q->cmdq); in aac_queue_init()
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb/
Dsge.c1435 struct cmdQ *cmdq = &sge->cmdQ[0]; in update_tx_info() local
1437 cmdq->processed += pr0; in update_tx_info()
1443 clear_bit(CMDQ_STAT_RUNNING, &cmdq->status); in update_tx_info()
1445 if (cmdq->cleaned + cmdq->in_use != cmdq->processed && in update_tx_info()
1446 !test_and_set_bit(CMDQ_STAT_LAST_PKT_DB, &cmdq->status)) { in update_tx_info()
1447 set_bit(CMDQ_STAT_RUNNING, &cmdq->status); in update_tx_info()
/Linux-v5.4/drivers/net/wireless/intel/iwlwifi/pcie/
Dtrans.c3198 struct iwl_txq *cmdq = trans_pcie->txq[trans_pcie->cmd_queue]; in iwl_trans_pcie_dump_data() local
3214 if (dump_mask & BIT(IWL_FW_ERROR_DUMP_TXCMD) && cmdq) in iwl_trans_pcie_dump_data()
3216 cmdq->n_window * (sizeof(*txcmd) + in iwl_trans_pcie_dump_data()
3266 if (dump_mask & BIT(IWL_FW_ERROR_DUMP_TXCMD) && cmdq) { in iwl_trans_pcie_dump_data()
3271 spin_lock_bh(&cmdq->lock); in iwl_trans_pcie_dump_data()
3272 ptr = cmdq->write_ptr; in iwl_trans_pcie_dump_data()
3273 for (i = 0; i < cmdq->n_window; i++) { in iwl_trans_pcie_dump_data()
3274 u8 idx = iwl_pcie_get_cmd_index(cmdq, ptr); in iwl_trans_pcie_dump_data()
3284 (u8 *)cmdq->tfds + in iwl_trans_pcie_dump_data()
3292 memcpy(txcmd->data, cmdq->entries[idx].cmd, in iwl_trans_pcie_dump_data()
[all …]
/Linux-v5.4/arch/arm64/boot/dts/arm/
Dfvp-base-revc.dts182 interrupt-names = "eventq", "priq", "cmdq-sync", "gerror";

12