Home
last modified time | relevance | path

Searched refs:pg_size (Results 1 – 16 of 16) sorted by relevance

/Linux-v5.10/drivers/infiniband/hw/bnxt_re/
Dqplib_res.h123 u32 pg_size; member
293 u8 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size() local
297 switch (pbl->pg_size) { in bnxt_qplib_base_pg_size()
299 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size()
302 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8K; in bnxt_qplib_base_pg_size()
305 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_64K; in bnxt_qplib_base_pg_size()
308 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_2M; in bnxt_qplib_base_pg_size()
311 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8M; in bnxt_qplib_base_pg_size()
314 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_1G; in bnxt_qplib_base_pg_size()
320 return pg_size; in bnxt_qplib_base_pg_size()
Dqplib_res.c71 dma_free_coherent(&pdev->dev, pbl->pg_size, in __free_pbl()
87 pbl->pg_size = 0; in __free_pbl()
131 pbl->pg_size = sginfo->pgsize; in __alloc_pbl()
136 pbl->pg_size, in __alloc_pbl()
185 u32 npages, aux_slots, pg_size, aux_pages = 0, aux_size = 0; in bnxt_qplib_alloc_init_hwq() local
195 pg_size = hwq_attr->sginfo->pgsize; in bnxt_qplib_alloc_init_hwq()
203 aux_pages = (aux_slots * aux_size) / pg_size; in bnxt_qplib_alloc_init_hwq()
204 if ((aux_slots * aux_size) % pg_size) in bnxt_qplib_alloc_init_hwq()
210 npages = (depth * stride) / pg_size + aux_pages; in bnxt_qplib_alloc_init_hwq()
211 if ((depth * stride) % pg_size) in bnxt_qplib_alloc_init_hwq()
[all …]
Dqplib_sp.c663 u32 pg_size; in bnxt_qplib_reg_mr() local
712 pg_size = PAGE_SIZE; in bnxt_qplib_reg_mr()
717 pg_size = buf_pg_size ? buf_pg_size : PAGE_SIZE; in bnxt_qplib_reg_mr()
719 ((ilog2(pg_size) << in bnxt_qplib_reg_mr()
Dib_verbs.c3922 resp.pg_size = PAGE_SIZE; in bnxt_re_alloc_ucontext()
/Linux-v5.10/include/uapi/rdma/
Dbnxt_re-abi.h58 __u32 pg_size; member
/Linux-v5.10/drivers/net/ethernet/chelsio/inline_crypto/chtls/
Dchtls_io.c1106 int pg_size = PAGE_SIZE; in chtls_sendmsg() local
1111 pg_size = page_size(page); in chtls_sendmsg()
1112 if (off < pg_size && in chtls_sendmsg()
1122 if (page && off == pg_size) { in chtls_sendmsg()
1125 pg_size = PAGE_SIZE; in chtls_sendmsg()
1138 pg_size <<= order; in chtls_sendmsg()
1142 pg_size = PAGE_SIZE; in chtls_sendmsg()
1149 if (copy > pg_size - off) in chtls_sendmsg()
1150 copy = pg_size - off; in chtls_sendmsg()
1171 if (off + copy < pg_size) { in chtls_sendmsg()
/Linux-v5.10/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_eqs.c27 #define GET_EQ_NUM_PAGES(eq, pg_size) \ argument
28 (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size))
30 #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size) argument
Dhinic_hw_io.c38 #define HINIC_PAGE_SIZE_HW(pg_size) ((u8)ilog2((u32)((pg_size) >> 12))) argument
/Linux-v5.10/drivers/crypto/caam/
Dctrl.c625 int pg_size; in caam_probe() local
695 pg_size = (comp_params & CTPR_MS_PG_SZ_MASK) >> CTPR_MS_PG_SZ_SHIFT; in caam_probe()
696 if (pg_size == 0) in caam_probe()
/Linux-v5.10/drivers/infiniband/hw/i40iw/
Di40iw_verbs.c1325 static bool i40iw_check_mem_contiguous(u64 *arr, u32 npages, u32 pg_size) in i40iw_check_mem_contiguous() argument
1330 if ((*arr + (pg_size * pg_idx)) != arr[pg_idx]) in i40iw_check_mem_contiguous()
1341 static bool i40iw_check_mr_contiguous(struct i40iw_pble_alloc *palloc, u32 pg_size) in i40iw_check_mr_contiguous() argument
1352 ret = i40iw_check_mem_contiguous(arr, palloc->total_cnt, pg_size); in i40iw_check_mr_contiguous()
1360 if ((*start_addr + (i * pg_size * PBLE_PER_PAGE)) != *arr) in i40iw_check_mr_contiguous()
1362 ret = i40iw_check_mem_contiguous(arr, leaf->cnt, pg_size); in i40iw_check_mr_contiguous()
1428 u32 pg_size; in i40iw_handle_q_mem() local
1434 pg_size = iwmr->page_size; in i40iw_handle_q_mem()
1454 ret = i40iw_check_mem_contiguous(arr, req->sq_pages, pg_size); in i40iw_handle_q_mem()
1456 ret = i40iw_check_mem_contiguous(&arr[req->sq_pages], req->rq_pages, pg_size); in i40iw_handle_q_mem()
[all …]
/Linux-v5.10/drivers/iommu/
Diommu.c726 unsigned long pg_size; in iommu_create_device_direct_mappings() local
734 pg_size = 1UL << __ffs(domain->pgsize_bitmap); in iommu_create_device_direct_mappings()
746 start = ALIGN(entry->start, pg_size); in iommu_create_device_direct_mappings()
747 end = ALIGN(entry->start + entry->length, pg_size); in iommu_create_device_direct_mappings()
753 for (addr = start; addr < end; addr += pg_size) { in iommu_create_device_direct_mappings()
760 ret = iommu_map(domain, addr, addr, pg_size, entry->prot); in iommu_create_device_direct_mappings()
/Linux-v5.10/drivers/net/wireless/realtek/rtw88/
Dfw.c1165 u32 pg_size; in rtw_download_drv_rsvd_page() local
1169 pg_size = rtwdev->chip->page_size; in rtw_download_drv_rsvd_page()
1170 pg_num = size / pg_size + ((size & (pg_size - 1)) ? 1 : 0); in rtw_download_drv_rsvd_page()
/Linux-v5.10/drivers/target/sbp/
Dsbp_target.c1244 int tcode, sg_miter_flags, max_payload, pg_size, speed, node_id, in sbp_rw_data() local
1263 pg_size = CMDBLK_ORB_PG_SIZE(be32_to_cpu(req->orb.misc)); in sbp_rw_data()
1264 if (pg_size) { in sbp_rw_data()
1266 pg_size = 0x100 << pg_size; in sbp_rw_data()
/Linux-v5.10/drivers/iommu/amd/
Diommu.c447 unsigned long pte_mask, pg_size, cnt; in first_pte_l7() local
450 pg_size = PTE_PAGE_SIZE(*pte); in first_pte_l7()
451 cnt = PAGE_SIZE_PTE_COUNT(pg_size); in first_pte_l7()
456 *page_size = pg_size; in first_pte_l7()
/Linux-v5.10/drivers/net/ethernet/chelsio/cxgb3/
Dt3_hw.c2587 unsigned int pg_size) in pm_num_pages() argument
2589 unsigned int n = mem_size / pg_size; in pm_num_pages()
/Linux-v5.10/drivers/net/ethernet/broadcom/bnxt/
Dbnxt.c6815 u8 pg_size = 0; in bnxt_hwrm_set_pg_attr() local
6818 pg_size = 1 << 4; in bnxt_hwrm_set_pg_attr()
6820 pg_size = 2 << 4; in bnxt_hwrm_set_pg_attr()
6822 *pg_attr = pg_size; in bnxt_hwrm_set_pg_attr()