Searched refs:dma_mem (Results 1 – 12 of 12) sorted by relevance
27 if (dev && dev->dma_mem) in dev_get_coherent_memory()28 return dev->dma_mem; in dev_get_coherent_memory()45 struct dma_coherent_mem *dma_mem = NULL; in dma_init_coherent_memory() local61 dma_mem = kzalloc(sizeof(struct dma_coherent_mem), GFP_KERNEL); in dma_init_coherent_memory()62 if (!dma_mem) { in dma_init_coherent_memory()66 dma_mem->bitmap = kzalloc(bitmap_size, GFP_KERNEL); in dma_init_coherent_memory()67 if (!dma_mem->bitmap) { in dma_init_coherent_memory()72 dma_mem->virt_base = mem_base; in dma_init_coherent_memory()73 dma_mem->device_base = device_addr; in dma_init_coherent_memory()74 dma_mem->pfn_base = PFN_DOWN(phys_addr); in dma_init_coherent_memory()[all …]
36 struct be_dma_mem dma_mem; member58 return q->dma_mem.va + q->head * q->entry_size; in queue_head_node()63 return q->dma_mem.va + wrb_num * q->entry_size; in queue_get_wrb()68 return q->dma_mem.va + q->tail * q->entry_size; in queue_tail_node()
746 struct be_dma_mem *q_mem = &eq->dma_mem; in beiscsi_cmd_eq_create()788 struct be_dma_mem *q_mem = &cq->dma_mem; in beiscsi_cmd_cq_create()861 struct be_dma_mem *q_mem = &mccq->dma_mem; in beiscsi_cmd_mccq_create()987 struct be_dma_mem *q_mem = &dq->dma_mem; in be_cmd_create_default_pdu_queue()
3000 struct be_dma_mem *mem = &q->dma_mem; in be_fill_queue()3032 mem = &eq->dma_mem; in beiscsi_create_eqs()3070 mem = &eq->dma_mem; in beiscsi_create_eqs()3099 mem = &cq->dma_mem; in beiscsi_create_cqs()3135 mem = &cq->dma_mem; in beiscsi_create_cqs()3160 mem = &dq->dma_mem; in beiscsi_create_def_hdr()3212 mem = &dataq->dma_mem; in beiscsi_create_def_data()3327 struct be_dma_mem *mem = &q->dma_mem; in be_queue_free()3338 struct be_dma_mem *mem = &q->dma_mem; in be_queue_alloc()
316 void *dma_mem; member416 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req()426 return (struct req_msg *)(hba->dma_mem + in stex_ss_alloc_req()1151 msg_h = (struct st_msg_header *)hba->dma_mem; in stex_ss_handshake()1751 hba->dma_mem = dma_alloc_coherent(&pdev->dev, in stex_probe()1753 if (!hba->dma_mem) { in stex_probe()1762 hba->dma_mem = dma_alloc_coherent(&pdev->dev, in stex_probe()1766 if (!hba->dma_mem) { in stex_probe()1783 hba->scratch = (__le32 *)(hba->dma_mem + scratch_offset); in stex_probe()1784 hba->status_buffer = (struct status_msg *)(hba->dma_mem + sts_offset); in stex_probe()[all …]
130 struct i40iw_dma_mem dma_mem; member
425 i40iw_free_dma_mem(iwdev->sc_dev.hw, &iwqp->kqp.dma_mem); in i40iw_free_qp_resources()504 struct i40iw_dma_mem *mem = &iwqp->kqp.dma_mem; in i40iw_setup_kmode_qp()
136 struct be_dma_mem dma_mem; member158 return q->dma_mem.va + q->head * q->entry_size; in queue_head_node()163 return q->dma_mem.va + q->tail * q->entry_size; in queue_tail_node()168 return q->dma_mem.va + index * q->entry_size; in queue_index_node()
1002 struct be_dma_mem *q_mem = &eqo->q.dma_mem; in be_cmd_eq_create()1167 struct be_dma_mem *q_mem = &cq->dma_mem; in be_cmd_cq_create()1245 struct be_dma_mem *q_mem = &mccq->dma_mem; in be_cmd_mccq_ext_create()1310 struct be_dma_mem *q_mem = &mccq->dma_mem; in be_cmd_mccq_org_create()1369 struct be_dma_mem *q_mem = &txq->dma_mem; in be_cmd_txq_create()1417 struct be_dma_mem *q_mem = &rxq->dma_mem; in be_cmd_rxq_create()
152 struct be_dma_mem *mem = &q->dma_mem; in be_queue_free()164 struct be_dma_mem *mem = &q->dma_mem; in be_queue_alloc()1440 entry = txo->q.dma_mem.va; in be_tx_timeout()1450 entry = txo->cq.dma_mem.va; in be_tx_timeout()
2078 u8 *dma_mem; in dpaa2_eth_set_hash() local2110 dma_mem = kzalloc(DPAA2_CLASSIFIER_DMA_SIZE, GFP_KERNEL); in dpaa2_eth_set_hash()2111 if (!dma_mem) in dpaa2_eth_set_hash()2114 err = dpni_prepare_key_cfg(&cls_cfg, dma_mem); in dpaa2_eth_set_hash()2123 dist_cfg.key_cfg_iova = dma_map_single(dev, dma_mem, in dpaa2_eth_set_hash()2143 kfree(dma_mem); in dpaa2_eth_set_hash()
990 struct dma_coherent_mem *dma_mem; /* internal for coherent mem member