/Linux-v5.10/drivers/infiniband/hw/bnxt_re/ |
D | qplib_res.c | 61 /* PBL */ 62 static void __free_pbl(struct bnxt_qplib_res *res, struct bnxt_qplib_pbl *pbl, in __free_pbl() argument 69 for (i = 0; i < pbl->pg_count; i++) { in __free_pbl() 70 if (pbl->pg_arr[i]) in __free_pbl() 71 dma_free_coherent(&pdev->dev, pbl->pg_size, in __free_pbl() 73 pbl->pg_arr[i] & in __free_pbl() 75 pbl->pg_map_arr[i]); in __free_pbl() 78 "PBL free pg_arr[%d] empty?!\n", i); in __free_pbl() 79 pbl->pg_arr[i] = NULL; in __free_pbl() 82 vfree(pbl->pg_arr); in __free_pbl() [all …]
|
D | qplib_res.h | 150 struct bnxt_qplib_pbl pbl[PBL_LVL_MAX + 1]; member 152 /* ptr for easy access to the PBL entries */ 294 struct bnxt_qplib_pbl *pbl; in bnxt_qplib_base_pg_size() local 296 pbl = &hwq->pbl[PBL_LVL_0]; in bnxt_qplib_base_pg_size() 297 switch (pbl->pg_size) { in bnxt_qplib_base_pg_size()
|
D | qplib_fp.c | 608 struct bnxt_qplib_pbl *pbl; in bnxt_qplib_create_srq() local 636 pbl = &srq->hwq.pbl[PBL_LVL_0]; in bnxt_qplib_create_srq() 642 req.pbl = cpu_to_le64(pbl->pg_map_arr[0]); in bnxt_qplib_create_srq() 816 struct bnxt_qplib_pbl *pbl; in bnxt_qplib_create_qp1() local 845 pbl = &sq->hwq.pbl[PBL_LVL_0]; in bnxt_qplib_create_qp1() 846 req.sq_pbl = cpu_to_le64(pbl->pg_map_arr[0]); in bnxt_qplib_create_qp1() 870 pbl = &rq->hwq.pbl[PBL_LVL_0]; in bnxt_qplib_create_qp1() 871 req.rq_pbl = cpu_to_le64(pbl->pg_map_arr[0]); in bnxt_qplib_create_qp1() 958 struct bnxt_qplib_pbl *pbl; in bnxt_qplib_create_qp() local 999 pbl = &sq->hwq.pbl[PBL_LVL_0]; in bnxt_qplib_create_qp() [all …]
|
/Linux-v5.10/drivers/infiniband/hw/efa/ |
D | efa_verbs.c | 1190 static int pbl_chunk_list_create(struct efa_dev *dev, struct pbl_context *pbl) in pbl_chunk_list_create() argument 1192 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; in pbl_chunk_list_create() 1193 int page_cnt = pbl->phys.indirect.pbl_buf_size_in_pages; in pbl_chunk_list_create() 1194 struct scatterlist *pages_sgl = pbl->phys.indirect.sgl; in pbl_chunk_list_create() 1196 int sg_dma_cnt = pbl->phys.indirect.sg_dma_cnt; in pbl_chunk_list_create() 1290 static void pbl_chunk_list_destroy(struct efa_dev *dev, struct pbl_context *pbl) in pbl_chunk_list_destroy() argument 1292 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; in pbl_chunk_list_destroy() 1304 /* initialize pbl continuous mode: map pbl buffer to a dma address. */ 1306 struct pbl_context *pbl) in pbl_continuous_initialize() argument 1310 dma_addr = dma_map_single(&dev->pdev->dev, pbl->pbl_buf, in pbl_continuous_initialize() [all …]
|
D | efa_com_cmd.c | 239 memcpy(mr_cmd.pbl.inline_pbl_array, in efa_com_register_mr() 240 params->pbl.inline_pbl_array, in efa_com_register_mr() 241 sizeof(mr_cmd.pbl.inline_pbl_array)); in efa_com_register_mr() 243 mr_cmd.pbl.pbl.length = params->pbl.pbl.length; in efa_com_register_mr() 244 mr_cmd.pbl.pbl.address.mem_addr_low = in efa_com_register_mr() 245 params->pbl.pbl.address.mem_addr_low; in efa_com_register_mr() 246 mr_cmd.pbl.pbl.address.mem_addr_high = in efa_com_register_mr() 247 params->pbl.pbl.address.mem_addr_high; in efa_com_register_mr()
|
D | efa_com_cmd.h | 176 struct efa_com_ctrl_buff_info pbl; member 177 } pbl; member 178 /* number of pages in PBL (redundant, could be calculated) */
|
/Linux-v5.10/include/linux/qed/ |
D | qed_chain.h | 84 /* Fastpath portions of the PBL [if exists] */ 89 * in the pbl table. 97 } pbl; member 126 /* Base address of a pre-allocated buffer for pbl */ 310 *p_next_elem = p_chain->pbl.pp_addr_tbl[page_index].virt_addr; in qed_chain_advance_page() 374 p_prod_page_idx = &p_chain->pbl.c.u16.prod_page_idx; in qed_chain_produce() 383 p_prod_page_idx = &p_chain->pbl.c.u32.prod_page_idx; in qed_chain_produce() 447 p_cons_page_idx = &p_chain->pbl.c.u16.cons_page_idx; in qed_chain_consume() 456 p_cons_page_idx = &p_chain->pbl.c.u32.cons_page_idx; in qed_chain_consume() 498 p_chain->pbl.c.u16.prod_page_idx = (u16)reset_val; in qed_chain_reset() [all …]
|
/Linux-v5.10/drivers/infiniband/sw/siw/ |
D | siw_mem.c | 144 siw_dbg_mem(mem, "free mem, pbl: %s\n", mem->is_pbl ? "y" : "n"); in siw_free_mem() 150 kfree(mem->pbl); in siw_free_mem() 320 * Gets physical address backed by PBL element. Address is referenced 323 * current PBL index for later resume at same element. 325 dma_addr_t siw_pbl_get_buffer(struct siw_pbl *pbl, u64 off, int *len, int *idx) in siw_pbl_get_buffer() argument 329 while (i < pbl->num_buf) { in siw_pbl_get_buffer() 330 struct siw_pble *pble = &pbl->pbe[i]; in siw_pbl_get_buffer() 351 struct siw_pbl *pbl; in siw_pbl_alloc() local 356 pbl = kzalloc(struct_size(pbl, pbe, num_buf), GFP_KERNEL); in siw_pbl_alloc() 357 if (!pbl) in siw_pbl_alloc() [all …]
|
D | siw_verbs.c | 1373 struct siw_pbl *pbl = NULL; in siw_alloc_mr() local 1391 pbl = siw_pbl_alloc(max_sge); in siw_alloc_mr() 1392 if (IS_ERR(pbl)) { in siw_alloc_mr() 1393 rv = PTR_ERR(pbl); in siw_alloc_mr() 1394 siw_dbg_pd(pd, "pbl allocation failed: %d\n", rv); in siw_alloc_mr() 1395 pbl = NULL; in siw_alloc_mr() 1403 rv = siw_mr_add_mem(mr, pd, pbl, 0, max_sge * PAGE_SIZE, 0); in siw_alloc_mr() 1417 kfree(pbl); in siw_alloc_mr() 1440 struct siw_pbl *pbl = mem->pbl; in siw_map_mr_sg() local 1445 if (!pbl) { in siw_map_mr_sg() [all …]
|
/Linux-v5.10/drivers/infiniband/hw/cxgb4/ |
D | resource.c | 260 * PBL Memory Manager. Uses Linux generic allocator. 263 #define MIN_PBL_SHIFT 8 /* 256B == min PBL size (32 entries) */ 271 rdev->stats.pbl.cur += roundup(size, 1 << MIN_PBL_SHIFT); in c4iw_pblpool_alloc() 272 if (rdev->stats.pbl.cur > rdev->stats.pbl.max) in c4iw_pblpool_alloc() 273 rdev->stats.pbl.max = rdev->stats.pbl.cur; in c4iw_pblpool_alloc() 276 rdev->stats.pbl.fail++; in c4iw_pblpool_alloc() 294 rdev->stats.pbl.cur -= roundup(size, 1 << MIN_PBL_SHIFT); in c4iw_pblpool_free() 308 pbl_start = rdev->lldi.vr->pbl.start; in c4iw_pblpool_create() 309 pbl_chunk = rdev->lldi.vr->pbl.size; in c4iw_pblpool_create() 315 pr_debug("failed to add PBL chunk (%x/%x)\n", in c4iw_pblpool_create() [all …]
|
/Linux-v5.10/drivers/infiniband/hw/qedr/ |
D | verbs.c | 505 struct qedr_pbl_info *pbl_info, struct qedr_pbl *pbl) in qedr_free_pbl() argument 511 if (!pbl[i].va) in qedr_free_pbl() 514 pbl[i].va, pbl[i].pa); in qedr_free_pbl() 517 kfree(pbl); in qedr_free_pbl() 552 /* Two-Layer PBLs, if we have more than one pbl we need to initialize in qedr_alloc_pbl_tbl() 581 DP_ERR(dev, "prepare pbl table: too many pages %d\n", in qedr_prepare_pbl_tbl() 586 /* calculate required pbl page size */ in qedr_prepare_pbl_tbl() 601 /* One layered PBL */ in qedr_prepare_pbl_tbl() 613 "prepare pbl table: num_pbes=%d, num_pbls=%d, pbl_size=%d\n", in qedr_prepare_pbl_tbl() 620 struct qedr_pbl *pbl, in qedr_populate_pbls() argument [all …]
|
D | qedr.h | 75 struct qed_chain pbl; member 291 struct qed_chain pbl; member 320 struct qed_chain pbl; member 342 qed_chain_get_capacity(p_info->pbl) \ 348 struct qed_chain pbl; member
|
D | main.c | 346 dev->ops->common->chain_free(dev->cdev, &dev->cnq_array[i].pbl); in qedr_free_resources() 408 rc = dev->ops->common->chain_alloc(dev->cdev, &cnq->pbl, in qedr_alloc_resources() 421 i, qed_chain_get_cons_idx(&cnq->pbl)); in qedr_alloc_resources() 429 dev->ops->common->chain_free(dev->cdev, &dev->cnq_array[i].pbl); in qedr_alloc_resources() 470 sw_comp_cons = qed_chain_get_cons_idx(&cnq->pbl); in qedr_irq_handler() 476 cq_handle = (struct regpair *)qed_chain_consume(&cnq->pbl); in qedr_irq_handler() 510 sw_comp_cons = qed_chain_get_cons_idx(&cnq->pbl); in qedr_irq_handler() 825 page_cnt = qed_chain_get_page_cnt(&dev->cnq_array[i].pbl); in qedr_init_hw() 828 p_phys_table = qed_chain_get_pbl_phys(&dev->cnq_array[i].pbl); in qedr_init_hw()
|
/Linux-v5.10/drivers/scsi/bnx2fc/ |
D | bnx2fc_tgt.c | 667 u32 *pbl; in bnx2fc_alloc_session_resc() local 695 /* Allocate and map RQ and RQ PBL */ in bnx2fc_alloc_session_resc() 715 printk(KERN_ERR PFX "unable to allocate RQ PBL %d\n", in bnx2fc_alloc_session_resc() 722 pbl = (u32 *)tgt->rq_pbl; in bnx2fc_alloc_session_resc() 725 *pbl = (u32)page; in bnx2fc_alloc_session_resc() 726 pbl++; in bnx2fc_alloc_session_resc() 727 *pbl = (u32)((u64)page >> 32); in bnx2fc_alloc_session_resc() 728 pbl++; in bnx2fc_alloc_session_resc() 746 /* Allocate and map CONFQ & CONFQ PBL */ in bnx2fc_alloc_session_resc() 769 printk(KERN_ERR PFX "unable to allocate CONFQ PBL %d\n", in bnx2fc_alloc_session_resc() [all …]
|
D | bnx2fc_hwi.c | 1988 u32 *pbl; in bnx2fc_free_hash_table() local 1992 pbl = hba->hash_tbl_pbl; in bnx2fc_free_hash_table() 1993 if (pbl) { in bnx2fc_free_hash_table() 1998 dma_address = le32_to_cpu(*pbl); in bnx2fc_free_hash_table() 1999 ++pbl; in bnx2fc_free_hash_table() 2000 dma_address += ((u64)le32_to_cpu(*pbl)) << 32; in bnx2fc_free_hash_table() 2001 ++pbl; in bnx2fc_free_hash_table() 2029 u32 *pbl; in bnx2fc_allocate_hash_table() local 2066 printk(KERN_ERR PFX "hash table pbl alloc failed\n"); in bnx2fc_allocate_hash_table() 2070 pbl = hba->hash_tbl_pbl; in bnx2fc_allocate_hash_table() [all …]
|
/Linux-v5.10/drivers/scsi/qedi/ |
D | qedi_main.c | 1540 struct scsi_bd *pbl; in qedi_alloc_bdq() local 1570 QEDI_ERR(&qedi->dbg_ctx, "Could not allocate BDQ PBL.\n"); in qedi_alloc_bdq() 1575 * Populate BDQ PBL with physical and virtual address of individual in qedi_alloc_bdq() 1578 pbl = (struct scsi_bd *)qedi->bdq_pbl; in qedi_alloc_bdq() 1580 pbl->address.hi = in qedi_alloc_bdq() 1582 pbl->address.lo = in qedi_alloc_bdq() 1585 "pbl [0x%p] pbl->address hi [0x%llx] lo [0x%llx], idx [%d]\n", in qedi_alloc_bdq() 1586 pbl, pbl->address.hi, pbl->address.lo, i); in qedi_alloc_bdq() 1587 pbl->opaque.iscsi_opaque.reserved_zero[0] = 0; in qedi_alloc_bdq() 1588 pbl->opaque.iscsi_opaque.reserved_zero[1] = 0; in qedi_alloc_bdq() [all …]
|
/Linux-v5.10/drivers/net/ethernet/qlogic/qed/ |
D | qed_chain.c | 111 if (!chain->pbl.pp_addr_tbl) in qed_chain_free_pbl() 115 entry = chain->pbl.pp_addr_tbl + i; in qed_chain_free_pbl() 128 vfree(chain->pbl.pp_addr_tbl); in qed_chain_free_pbl() 129 chain->pbl.pp_addr_tbl = NULL; in qed_chain_free_pbl() 269 chain->pbl.pp_addr_tbl = addr_tbl; in qed_chain_alloc_pbl() 300 /* Fill the PBL table with the physical address of the page */ in qed_chain_alloc_pbl()
|
/Linux-v5.10/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_dma.c | 44 int fix_burst, int pbl, dma_addr_t dma_tx, in sxgbe_dma_channel_init() argument 51 /* set the pbl */ in sxgbe_dma_channel_init() 55 /* program the TX pbl */ in sxgbe_dma_channel_init() 57 reg_val |= (pbl << SXGBE_DMA_TXPBL_LSHIFT); in sxgbe_dma_channel_init() 59 /* program the RX pbl */ in sxgbe_dma_channel_init() 61 reg_val |= (pbl << SXGBE_DMA_RXPBL_LSHIFT); in sxgbe_dma_channel_init()
|
/Linux-v5.10/Documentation/devicetree/bindings/net/ |
D | stm32-dwmac.yaml | 117 snps,pbl = <2>; 133 snps,pbl = <8>; 148 snps,pbl = <8>;
|
D | snps,dwmac.yaml | 307 snps,pbl: 316 value rather than snps,pbl. 323 value rather than snps,pbl. 327 snps,no-pbl-x8: 330 Don\'t multiply the pbl/txpbl/rxpbl values by 8. For core
|
D | samsung-sxgbe.txt | 15 - samsung,pbl: Integer, Programmable Burst Length. 47 samsung,pbl = <0x08>
|
/Linux-v5.10/drivers/scsi/qedf/ |
D | qedf_main.c | 1245 u32 *pbl; in qedf_alloc_sq() local 1249 /* Calculate appropriate queue and PBL sizes */ in qedf_alloc_sq() 1268 QEDF_WARN(&(qedf->dbg_ctx), "Could not allocate send queue PBL.\n"); in qedf_alloc_sq() 1273 /* Create PBL */ in qedf_alloc_sq() 1276 pbl = (u32 *)fcport->sq_pbl; in qedf_alloc_sq() 1279 *pbl = U64_LO(page); in qedf_alloc_sq() 1280 pbl++; in qedf_alloc_sq() 1281 *pbl = U64_HI(page); in qedf_alloc_sq() 1282 pbl++; in qedf_alloc_sq() 2926 struct scsi_bd *pbl; in qedf_alloc_bdq() local [all …]
|
/Linux-v5.10/drivers/net/ethernet/stmicro/stmmac/ |
D | dwmac1000_dma.c | 77 int txpbl = dma_cfg->txpbl ?: dma_cfg->pbl; in dwmac1000_dma_init() 78 int rxpbl = dma_cfg->rxpbl ?: dma_cfg->pbl; in dwmac1000_dma_init() 81 * Set the DMA PBL (Programmable Burst Length) mode. in dwmac1000_dma_init() 84 * post 3.5 mode bit acts as 8*PBL. in dwmac1000_dma_init()
|
/Linux-v5.10/drivers/infiniband/hw/i40iw/ |
D | i40iw_verbs.c | 318 * i40iw_get_pbl - Retrieve pbl from a list given a virtual 321 * @pbl_list: pbl list to search in (QP's or CQ's) 633 i40iw_pr_err("no pbl info\n"); in i40iw_create_qp() 1273 * i40iw_next_pbl_addr - Get next pbl address 1274 * @pbl: pointer to a pble 1278 static inline u64 *i40iw_next_pbl_addr(u64 *pbl, in i40iw_next_pbl_addr() argument 1284 return ++pbl; in i40iw_next_pbl_addr() 1293 * @pbl: ple pointer to save 1 level or 0 level pble 1297 u64 *pbl, in i40iw_copy_user_pgaddrs() argument 1313 *pbl = rdma_block_iter_dma_address(&biter); in i40iw_copy_user_pgaddrs() [all …]
|
/Linux-v5.10/kernel/irq/ |
D | debugfs.c | 36 seq_printf(m, "affinity: %*pbl\n", cpumask_pr_args(msk)); in irq_debug_show_masks() 39 seq_printf(m, "effectiv: %*pbl\n", cpumask_pr_args(msk)); in irq_debug_show_masks() 43 seq_printf(m, "pending: %*pbl\n", cpumask_pr_args(msk)); in irq_debug_show_masks()
|