/Linux-v4.19/drivers/infiniband/hw/cxgb4/ |
D | device.c | 110 le.poll_sge_ts = cxgb4_read_sge_timestamp(wq->rdev->lldi.ports[0]); in c4iw_log_wr_stats() 138 #define ts2ns(ts) div64_u64((ts) * dev->rdev.lldi.cclk_ps, 1000) in wr_log_show() 388 ret = cxgb4_read_tpte(stagd->devp->rdev.lldi.ports[0], (u32)id<<8, in dump_stag() 391 dev_err(&stagd->devp->rdev.lldi.pdev->dev, in dump_stag() 794 if (rdev->lldi.udb_density != rdev->lldi.ucq_density) { in c4iw_rdev_open() 796 pci_name(rdev->lldi.pdev), rdev->lldi.udb_density, in c4iw_rdev_open() 797 rdev->lldi.ucq_density); in c4iw_rdev_open() 800 if (rdev->lldi.vr->qp.start != rdev->lldi.vr->cq.start || in c4iw_rdev_open() 801 rdev->lldi.vr->qp.size != rdev->lldi.vr->cq.size) { in c4iw_rdev_open() 803 pci_name(rdev->lldi.pdev), rdev->lldi.vr->qp.start, in c4iw_rdev_open() [all …]
|
D | provider.c | 168 if ((addr >= pci_resource_start(rdev->lldi.pdev, 0)) && in c4iw_mmap() 169 (addr < (pci_resource_start(rdev->lldi.pdev, 0) + in c4iw_mmap() 170 pci_resource_len(rdev->lldi.pdev, 0)))) { in c4iw_mmap() 179 } else if ((addr >= pci_resource_start(rdev->lldi.pdev, 2)) && in c4iw_mmap() 180 (addr < (pci_resource_start(rdev->lldi.pdev, 2) + in c4iw_mmap() 181 pci_resource_len(rdev->lldi.pdev, 2)))) { in c4iw_mmap() 189 if (!is_t4(rdev->lldi.adapter_type)) in c4iw_mmap() 284 memcpy(&(gid->raw[0]), dev->rdev.lldi.ports[port-1]->dev_addr, 6); in c4iw_query_gid() 301 memcpy(&props->sys_image_guid, dev->rdev.lldi.ports[0]->dev_addr, 6); in c4iw_query_device() 302 props->hw_ver = CHELSIO_CHIP_RELEASE(dev->rdev.lldi.adapter_type); in c4iw_query_device() [all …]
|
D | cm.c | 212 error = cxgb4_l2t_send(rdev->lldi.ports[0], skb, l2e); in c4iw_l2t_send() 229 error = cxgb4_ofld_send(rdev->lldi.ports[0], skb); in c4iw_ofld_send() 250 ep->emss = ep->com.dev->rdev.lldi.mtus[TCPOPT_MSS_G(opt)] - in set_emss() 398 ep->com.dev->rdev.lldi.ports[0], in _c4iw_free_ep() 402 cxgb4_remove_tid(ep->com.dev->rdev.lldi.tids, 0, ep->hwtid, in _c4iw_free_ep() 556 cxgb4_clip_release(ep->com.dev->rdev.lldi.ports[0], in act_open_req_arp_failure() 560 cxgb4_free_atid(ep->com.dev->rdev.lldi.tids, ep->atid); in act_open_req_arp_failure() 616 (ep->com.dev->rdev.lldi.pf)); in send_flowc() 697 enum chip_type adapter_type = ep->com.dev->rdev.lldi.adapter_type; in send_connect() 702 netdev = ep->com.dev->rdev.lldi.ports[0]; in send_connect() [all …]
|
D | resource.c | 43 rdev->lldi.vr->qp.start, in c4iw_init_qid_table() 44 rdev->lldi.vr->qp.size, in c4iw_init_qid_table() 45 rdev->lldi.vr->qp.size, 0)) in c4iw_init_qid_table() 48 for (i = rdev->lldi.vr->qp.start; in c4iw_init_qid_table() 49 i < rdev->lldi.vr->qp.start + rdev->lldi.vr->qp.size; i++) in c4iw_init_qid_table() 308 pbl_start = rdev->lldi.vr->pbl.start; in c4iw_pblpool_create() 309 pbl_chunk = rdev->lldi.vr->pbl.size; in c4iw_pblpool_create() 350 pci_name(rdev->lldi.pdev)); in c4iw_rqtpool_alloc() 395 if (rdev->lldi.vr->srq.size) in c4iw_rqtpool_create() 398 rqt_start = rdev->lldi.vr->rq.start + skip; in c4iw_rqtpool_create() [all …]
|
D | mem.c | 56 return (is_t4(dev->rdev.lldi.adapter_type) || in mr_exceeds_hw_limits() 57 is_t5(dev->rdev.lldi.adapter_type)) && in mr_exceeds_hw_limits() 92 T5_ULP_MEMIO_FID_V(rdev->lldi.rxq_ids[0])); in _c4iw_write_mem_dma_aligned() 120 if (is_t4(rdev->lldi.adapter_type)) in _c4iw_write_mem_inline() 199 daddr = dma_map_single(&rdev->lldi.pdev->dev, data, len, DMA_TO_DEVICE); in _c4iw_write_mem_dma() 200 if (dma_mapping_error(&rdev->lldi.pdev->dev, daddr)) in _c4iw_write_mem_dma() 225 dma_unmap_single(&rdev->lldi.pdev->dev, save, len, DMA_TO_DEVICE); in _c4iw_write_mem_dma() 239 if (!rdev->lldi.ulptx_memwrite_dsgl || !use_dsgl) { in write_adapter_mem() 254 pci_name(rdev->lldi.pdev)); in write_adapter_mem() 327 (rdev->lldi.vr->stag.start >> 5), in write_tpt_entry() [all …]
|
D | qp.c | 73 dev_warn(&dev->rdev.lldi.pdev->dev, in alloc_ird() 101 dma_free_coherent(&(rdev->lldi.pdev->dev), sq->memsize, sq->queue, in dealloc_host_sq() 115 if (!ocqp_support || !ocqp_supported(&rdev->lldi)) in alloc_oc_sq() 121 rdev->lldi.vr->ocq.start; in alloc_oc_sq() 123 rdev->lldi.vr->ocq.start); in alloc_oc_sq() 130 sq->queue = dma_alloc_coherent(&(rdev->lldi.pdev->dev), sq->memsize, in alloc_host_sq() 161 dma_free_coherent(&rdev->lldi.pdev->dev, in destroy_qp() 183 ret = cxgb4_bar2_sge_qregs(rdev->lldi.ports[0], qid, qtype, in c4iw_bar2_addrs() 192 if (is_t4(rdev->lldi.adapter_type)) in c4iw_bar2_addrs() 263 wq->rq.queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, in create_qp() [all …]
|
D | ev.c | 43 ret = cxgb4_read_tpte(dev->rdev.lldi.ports[0], stag, in print_tpte() 46 dev_err(&dev->rdev.lldi.pdev->dev, in print_tpte() 66 dev_err(&dev->rdev.lldi.pdev->dev, in dump_err_cqe()
|
D | iw_cxgb4.h | 76 #define PBL_OFF(rdev_p, a) ((a) - (rdev_p)->lldi.vr->pbl.start) 77 #define RQT_OFF(rdev_p, a) ((a) - (rdev_p)->lldi.vr->rq.start) 180 struct cxgb4_lld_info lldi; member 205 return (int)(rdev->lldi.vr->stag.size >> 5); in c4iw_num_stags() 274 func, pci_name(rdev->lldi.pdev), hwtid, qpid); in c4iw_wait_for_reply() 281 pci_name(rdev->lldi.pdev), wr_waitp->ret, hwtid, qpid); in c4iw_wait_for_reply() 335 struct cxgb4_lld_info lldi; member 422 return min(dev->rdev.lldi.max_ordird_qp, c4iw_max_read_depth); in cur_max_read_depth()
|
D | cq.c | 63 dma_free_coherent(&(rdev->lldi.pdev->dev), in destroy_cq() 98 cq->queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, cq->memsize, in create_cq() 143 rdev->lldi.ciq_ids[cq->vector])); in create_cq() 161 cq->gts = rdev->lldi.gts_reg; in create_cq() 169 pci_name(rdev->lldi.pdev), cq->cqid); in create_cq() 175 dma_free_coherent(&rdev->lldi.pdev->dev, cq->memsize, cq->queue, in create_cq() 1015 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq()
|
D | restrack.c | 454 ret = cxgb4_read_tpte(dev->rdev.lldi.ports[0], stag, (__be32 *)&tpte); in fill_res_mr_entry() 456 dev_err(&dev->rdev.lldi.pdev->dev, in fill_res_mr_entry()
|
/Linux-v4.19/drivers/scsi/cxgbi/cxgb4i/ |
D | cxgb4i.c | 192 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(csk->cdev); in send_act_open_req() local 212 if (is_t4(lldi->adapter_type)) { in send_act_open_req() 235 } else if (is_t5(lldi->adapter_type)) { in send_act_open_req() 301 CHELSIO_CHIP_VERSION(lldi->adapter_type), csk, in send_act_open_req() 311 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(csk->cdev); in send_act_open_req6() local 331 if (is_t4(lldi->adapter_type)) { in send_act_open_req6() 356 } else if (is_t5(lldi->adapter_type)) { in send_act_open_req6() 412 CHELSIO_CHIP_VERSION(lldi->adapter_type), csk, csk->state, in send_act_open_req6() 769 struct cxgb4_lld_info *lldi = cxgbi_cdev_priv(csk->cdev); in free_atid() local 772 cxgb4_free_atid(lldi->tids, csk->atid); in free_atid() [all …]
|
/Linux-v4.19/drivers/crypto/chelsio/chtls/ |
D | chtls_main.c | 131 for (i = 0; i < cdev->lldi->nports; i++) { in chtls_inline_feature() 157 strlcat(tlsdev->name, cdev->lldi->ports[0]->name, in chtls_register_dev() 200 struct cxgb4_lld_info *lldi; in chtls_uld_add() local 209 lldi = kzalloc(sizeof(*lldi), GFP_KERNEL); in chtls_uld_add() 210 if (!lldi) in chtls_uld_add() 216 *lldi = *info; in chtls_uld_add() 217 cdev->lldi = lldi; in chtls_uld_add() 218 cdev->pdev = lldi->pdev; in chtls_uld_add() 219 cdev->tids = lldi->tids; in chtls_uld_add() 220 cdev->ports = lldi->ports; in chtls_uld_add() [all …]
|
D | chtls_cm.c | 171 cxgb4_ofld_send(cdev->lldi->ports[0], skb); in abort_arp_failure() 647 cdev->lldi->rxq_ids[0]); in chtls_listen_start() 676 cxgb4_remove_server(cdev->lldi->ports[0], stid, in chtls_listen_stop() 677 cdev->lldi->rxq_ids[0], 0); in chtls_listen_stop() 883 tp->advmss = cxgb4_best_aligned_mtu(cdev->lldi->mtus, in chtls_select_mss() 933 struct cxgb4_lld_info *lldi; in chtls_pass_accept_rpl() local 946 lldi = csk->cdev->lldi; in chtls_pass_accept_rpl() 970 if (!is_t5(lldi->adapter_type)) in chtls_pass_accept_rpl() 1053 csk->l2t_entry = cxgb4_l2t_get(cdev->lldi->l2t, n, ndev, 0); in chtls_recv_sock() 1076 step = cdev->lldi->nrxq / cdev->lldi->nchan; in chtls_recv_sock() [all …]
|
D | chtls_hw.c | 120 int chtls_init_kmap(struct chtls_dev *cdev, struct cxgb4_lld_info *lldi) in chtls_init_kmap() argument 125 num_key_ctx = (lldi->vr->key.size / TLS_KEY_CONTEXT_SZ); in chtls_init_kmap() 135 cdev->kmap.start = lldi->vr->key.start; in chtls_init_kmap()
|
D | chtls.h | 126 struct cxgb4_lld_info *lldi; member 463 int chtls_init_kmap(struct chtls_dev *cdev, struct cxgb4_lld_info *lldi);
|
D | chtls_io.c | 188 FLOWC_PARAM(PFNVFN, FW_PFVF_CMD_PFN_V(csk->cdev->lldi->pf)); in send_tx_flowc_wr()
|
/Linux-v4.19/drivers/target/iscsi/cxgbit/ |
D | cxgbit_ddp.c | 300 struct cxgb4_lld_info *lldi = &cdev->lldi; in cxgbit_ddp_init() local 301 struct net_device *ndev = cdev->lldi.ports[0]; in cxgbit_ddp_init() 306 if (!lldi->vr->iscsi.size) { in cxgbit_ddp_init() 311 ppmax = lldi->vr->iscsi.size >> PPOD_SIZE_SHIFT; in cxgbit_ddp_init() 315 tformat.pgsz_order[i] = (lldi->iscsi_pgsz_order >> (i << 3)) in cxgbit_ddp_init() 317 cxgbi_tagmask_check(lldi->iscsi_tagmask, &tformat); in cxgbit_ddp_init() 319 ret = cxgbi_ppm_init(lldi->iscsi_ppm, cdev->lldi.ports[0], in cxgbit_ddp_init() 320 cdev->lldi.pdev, &cdev->lldi, &tformat, in cxgbit_ddp_init() 321 ppmax, lldi->iscsi_llimit, in cxgbit_ddp_init() 322 lldi->vr->iscsi.start, 2); in cxgbit_ddp_init() [all …]
|
D | cxgbit_cm.c | 66 func, pci_name(cdev->lldi.pdev), tid); in cxgbit_wait_for_reply() 72 pci_name(cdev->lldi.pdev), wr_waitp->ret, tid); in cxgbit_wait_for_reply() 156 __func__, cdev->lldi.ports[0]->name, stid, sin6->sin6_port); in cxgbit_create_server6() 161 ret = cxgb4_clip_get(cdev->lldi.ports[0], in cxgbit_create_server6() 173 ret = cxgb4_create_server6(cdev->lldi.ports[0], in cxgbit_create_server6() 176 cdev->lldi.rxq_ids[0]); in cxgbit_create_server6() 187 cxgb4_clip_release(cdev->lldi.ports[0], in cxgbit_create_server6() 207 __func__, cdev->lldi.ports[0]->name, stid, sin->sin_port); in cxgbit_create_server4() 212 ret = cxgb4_create_server(cdev->lldi.ports[0], in cxgbit_create_server4() 215 cdev->lldi.rxq_ids[0]); in cxgbit_create_server4() [all …]
|
D | cxgbit_main.c | 36 struct cxgb4_lld_info *lldi = &cdev->lldi; in cxgbit_set_mdsl() local 41 mdsl = min_t(u32, lldi->iscsi_iolen - ISCSI_PDU_NONPAYLOAD_LEN, in cxgbit_set_mdsl() 49 static void *cxgbit_uld_add(const struct cxgb4_lld_info *lldi) in cxgbit_uld_add() argument 53 if (is_t4(lldi->adapter_type)) in cxgbit_uld_add() 62 cdev->lldi = *lldi; in cxgbit_uld_add() 73 pci_name(lldi->pdev)); in cxgbit_uld_add() 75 if (lldi->fw_vers >= 0x10d2b00) in cxgbit_uld_add() 86 pci_name(lldi->pdev)); in cxgbit_uld_add() 144 pr_info("cdev %s state UP.\n", pci_name(cdev->lldi.pdev)); in cxgbit_uld_state_change() 149 pr_info("cdev %s state RECOVERY.\n", pci_name(cdev->lldi.pdev)); in cxgbit_uld_state_change() [all …]
|
D | cxgbit.h | 136 struct cxgb4_lld_info lldi; member 353 return (struct cxgbi_ppm *)(*cdev->lldi.iscsi_ppm); in cdev2ppm()
|
D | cxgbit_target.c | 167 const struct cxgb4_lld_info *lldi = &csk->com.cdev->lldi; in cxgbit_tx_data_wr() local 173 u32 force = is_t5(lldi->adapter_type) ? TX_FORCE_V(!submode) : in cxgbit_tx_data_wr() 739 struct cxgbi_ppm *ppm = *csk->com.cdev->lldi.iscsi_ppm; in cxgbit_set_params() 770 if (is_t5(cdev->lldi.adapter_type)) in cxgbit_set_params() 1566 struct cxgb4_lld_info *lldi = &csk->com.cdev->lldi; in cxgbit_rx_skb() local 1570 if (is_t5(lldi->adapter_type)) in cxgbit_rx_skb()
|
/Linux-v4.19/drivers/net/ethernet/chelsio/libcxgb/ |
D | libcxgb_cm.c | 77 cxgb_our_interface(struct cxgb4_lld_info *lldi, in cxgb_our_interface() argument 84 for (i = 0; i < lldi->nports; i++) in cxgb_our_interface() 85 if (lldi->ports[i] == egress_dev) in cxgb_our_interface() 91 cxgb_find_route(struct cxgb4_lld_info *lldi, in cxgb_find_route() argument 108 if (!cxgb_our_interface(lldi, get_real_dev, n->dev) && in cxgb_find_route() 120 cxgb_find_route6(struct cxgb4_lld_info *lldi, in cxgb_find_route6() argument 137 (!cxgb_our_interface(lldi, get_real_dev, in cxgb_find_route6()
|
/Linux-v4.19/drivers/crypto/chelsio/ |
D | chcr_algo.c | 710 int qid = u_ctx->lldi.rxq_ids[ctx->rx_qidx]; in create_wreq() 1107 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, in chcr_handle_cipher_resp() 1130 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, in chcr_handle_cipher_resp() 1145 wrparam.qid = u_ctx->lldi.rxq_ids[c_ctx(tfm)->rx_qidx]; in chcr_handle_cipher_resp() 1154 skb->dev = u_ctx->lldi.ports[0]; in chcr_handle_cipher_resp() 1161 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); in chcr_handle_cipher_resp() 1189 chcr_cipher_dma_map(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); in process_cipher() 1241 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, in process_cipher() 1270 chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req); in process_cipher() 1282 if (unlikely(cxgb4_is_crypto_q_full(u_ctx->lldi.ports[0], in chcr_aes_encrypt() [all …]
|
D | chcr_core.h | 65 #define padap(dev) pci_get_drvdata(dev->u_ctx->lldi.pdev) 141 struct cxgb4_lld_info lldi; member
|
D | chcr_core.c | 169 u_ctx->lldi = *lld; in chcr_uld_add()
|