| /Linux-v5.10/drivers/net/ethernet/qlogic/qed/ |
| D | qed_chain.c | 12 u32 page_cnt) in qed_chain_init() argument 34 chain->page_cnt = page_cnt; in qed_chain_init() 35 chain->capacity = chain->usable_per_page * page_cnt; in qed_chain_init() 36 chain->size = chain->elem_per_page * page_cnt; in qed_chain_init() 80 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_next_ptr() 114 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_pbl() 160 u32 page_cnt) in qed_chain_alloc_sanity_check() argument 165 chain_size *= page_cnt; in qed_chain_alloc_sanity_check() 206 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_alloc_next_ptr() 255 u32 page_cnt, i; in qed_chain_alloc_pbl() local [all …]
|
| D | qed_sp_commands.c | 309 u8 page_cnt, i; in qed_sp_pf_start() local 370 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_eq->chain); in qed_sp_pf_start() 371 p_ramrod->event_ring_num_pages = page_cnt; in qed_sp_pf_start()
|
| /Linux-v5.10/include/linux/qed/ |
| D | qed_chain.h | 106 u32 page_cnt; member 262 return chain->page_cnt; in qed_chain_get_page_cnt() 302 if (++(*(u16 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page() 306 if (++(*(u32 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page() 495 u32 reset_val = p_chain->page_cnt - 1; in qed_chain_reset() 554 last_page_idx = p_chain->page_cnt - 1; in qed_chain_get_last_elem() 575 u32 cur_prod, page_mask, page_cnt, page_diff; in qed_chain_set_prod() local 594 page_cnt = qed_chain_get_page_cnt(p_chain); in qed_chain_set_prod() 598 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod() 602 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod() [all …]
|
| /Linux-v5.10/arch/mips/cavium-octeon/executive/ |
| D | cvmx-helper-util.c | 133 union cvmx_ipd_portx_bp_page_cnt page_cnt; in cvmx_helper_setup_red() local 141 page_cnt.u64 = 0; in cvmx_helper_setup_red() 142 page_cnt.s.bp_enb = 0; in cvmx_helper_setup_red() 143 page_cnt.s.page_cnt = 100; in cvmx_helper_setup_red() 148 page_cnt.u64); in cvmx_helper_setup_red()
|
| /Linux-v5.10/drivers/hv/ |
| D | ring_buffer.c | 193 struct page *pages, u32 page_cnt) in hv_ringbuffer_init() argument 204 pages_wraparound = kcalloc(page_cnt * 2 - 1, sizeof(struct page *), in hv_ringbuffer_init() 210 for (i = 0; i < 2 * (page_cnt - 1); i++) in hv_ringbuffer_init() 211 pages_wraparound[i + 1] = &pages[i % (page_cnt - 1) + 1]; in hv_ringbuffer_init() 214 vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, PAGE_KERNEL); in hv_ringbuffer_init() 228 ring_info->ring_size = page_cnt << PAGE_SHIFT; in hv_ringbuffer_init()
|
| D | hv_balloon.c | 142 __u64 page_cnt:24; member 994 pfn_cnt = dm->ha_wrk.ha_page_range.finfo.page_cnt; in hot_add_req() 997 rg_sz = dm->ha_wrk.ha_region_range.finfo.page_cnt; in hot_add_req() 1191 int num_pages = range_array->finfo.page_cnt; in free_balloon_pages() 1247 bl_resp->range_array[i].finfo.page_cnt = alloc_unit; in alloc_balloon_pages()
|
| /Linux-v5.10/drivers/base/firmware_loader/ |
| D | fallback.c | 323 int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count); in firmware_rw() local 328 memcpy(buffer, page_data + page_ofs, page_cnt); in firmware_rw() 330 memcpy(page_data + page_ofs, buffer, page_cnt); in firmware_rw() 333 buffer += page_cnt; in firmware_rw() 334 offset += page_cnt; in firmware_rw() 335 count -= page_cnt; in firmware_rw()
|
| /Linux-v5.10/drivers/infiniband/hw/hns/ |
| D | hns_roce_mr.c | 844 dma_addr_t *pages, int page_cnt) in hns_roce_mtr_map() argument 862 if (r->offset + r->count > page_cnt) { in hns_roce_mtr_map() 866 i, r->offset, r->count, page_cnt); in hns_roce_mtr_map() 948 int page_cnt = 0; in mtr_init_buf_cfg() local 954 page_cnt = DIV_ROUND_UP(buf_size, HNS_HW_PAGE_SIZE); in mtr_init_buf_cfg() 959 cfg->buf_pg_shift = HNS_HW_PAGE_SHIFT + order_base_2(page_cnt); in mtr_init_buf_cfg() 961 cfg->buf_pg_count = page_cnt; in mtr_init_buf_cfg() 978 for (page_cnt = 0, region_cnt = 0; page_cnt < cfg->buf_pg_count && in mtr_init_buf_cfg() 982 r->offset = page_cnt; in mtr_init_buf_cfg() 985 page_cnt += r->count; in mtr_init_buf_cfg() [all …]
|
| D | hns_roce_device.h | 1158 dma_addr_t *pages, int page_cnt);
|
| /Linux-v5.10/arch/mips/include/asm/octeon/ |
| D | cvmx-ipd-defs.h | 324 uint64_t page_cnt:17; member 326 uint64_t page_cnt:17; 999 uint64_t page_cnt:17; member 1001 uint64_t page_cnt:17; 1014 uint64_t page_cnt:17; member 1016 uint64_t page_cnt:17; 1029 uint64_t page_cnt:17; member 1031 uint64_t page_cnt:17; 1387 uint64_t page_cnt:25; member 1389 uint64_t page_cnt:25;
|
| /Linux-v5.10/drivers/staging/rts5208/ |
| D | xd.c | 1506 u8 reg_val, page_cnt; in xd_read_multiple_pages() local 1512 page_cnt = end_page - start_page; in xd_read_multiple_pages() 1538 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_read_multiple_pages() 1543 page_cnt * 512, DMA_512); in xd_read_multiple_pages() 1553 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 512, in xd_read_multiple_pages() 1709 u8 page_cnt, reg_val; in xd_write_multiple_pages() local 1717 page_cnt = end_page - start_page; in xd_write_multiple_pages() 1739 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_write_multiple_pages() 1743 page_cnt * 512, DMA_512); in xd_write_multiple_pages() 1752 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 512, in xd_write_multiple_pages() [all …]
|
| D | ms.c | 3400 u8 start_page, end_page = 0, page_cnt; local 3526 page_cnt = end_page - start_page; 3529 start_page, end_page, page_cnt); 3566 total_sec_cnt -= page_cnt; 3568 ptr += page_cnt * 512;
|
| /Linux-v5.10/drivers/infiniband/hw/i40iw/ |
| D | i40iw_verbs.h | 95 u32 page_cnt; member
|
| D | i40iw_verbs.c | 1389 status = i40iw_get_pble(&iwdev->sc_dev, iwdev->pble_rsrc, palloc, iwmr->page_cnt); in i40iw_setup_pbles() 1565 iwmr->page_cnt = max_num_sg; in i40iw_alloc_mr() 1567 status = i40iw_get_pble(&iwdev->sc_dev, iwdev->pble_rsrc, palloc, iwmr->page_cnt); in i40iw_alloc_mr() 1601 if (unlikely(iwmr->npages == iwmr->page_cnt)) in i40iw_set_page() 1781 iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size); in i40iw_reg_user_mr() 1806 use_pbles = (iwmr->page_cnt != 1); in i40iw_reg_user_mr() 1897 iwmr->page_cnt = 1; in i40iw_reg_phys_mr()
|
| /Linux-v5.10/drivers/infiniband/hw/efa/ |
| D | efa_verbs.c | 1162 static struct scatterlist *efa_vmalloc_buf_to_sg(u64 *buf, int page_cnt) in efa_vmalloc_buf_to_sg() argument 1168 sglist = kmalloc_array(page_cnt, sizeof(*sglist), GFP_KERNEL); in efa_vmalloc_buf_to_sg() 1171 sg_init_table(sglist, page_cnt); in efa_vmalloc_buf_to_sg() 1172 for (i = 0; i < page_cnt; i++) { in efa_vmalloc_buf_to_sg() 1193 int page_cnt = pbl->phys.indirect.pbl_buf_size_in_pages; in pbl_chunk_list_create() local 1204 chunk_list_size = DIV_ROUND_UP(page_cnt, EFA_PTRS_PER_CHUNK); in pbl_chunk_list_create() 1215 page_cnt); in pbl_chunk_list_create() 1226 ((page_cnt % EFA_PTRS_PER_CHUNK) * EFA_CHUNK_PAYLOAD_PTR_SIZE) + in pbl_chunk_list_create()
|
| /Linux-v5.10/fs/ubifs/ |
| D | file.c | 724 int err, page_idx, page_cnt, ret = 0, n = 0; in ubifs_do_bulk_read() local 739 page_cnt = bu->blk_cnt >> UBIFS_BLOCKS_PER_PAGE_SHIFT; in ubifs_do_bulk_read() 740 if (!page_cnt) { in ubifs_do_bulk_read() 783 for (page_idx = 1; page_idx < page_cnt; page_idx++) { in ubifs_do_bulk_read()
|
| /Linux-v5.10/drivers/infiniband/hw/qedr/ |
| D | main.c | 811 u32 page_cnt; in qedr_init_hw() local 825 page_cnt = qed_chain_get_page_cnt(&dev->cnq_array[i].pbl); in qedr_init_hw() 826 cur_pbl->num_pbl_pages = page_cnt; in qedr_init_hw()
|
| D | verbs.c | 841 int chain_entries, int page_cnt, in qedr_init_cq_params() argument 852 params->pbl_num_pages = page_cnt; in qedr_init_cq_params() 922 int page_cnt; in qedr_create_cq() local 968 page_cnt = cq->q.pbl_info.num_pbes; in qedr_create_cq() 980 page_cnt = qed_chain_get_page_cnt(&cq->pbl); in qedr_create_cq() 985 qedr_init_cq_params(cq, ctx, dev, vector, chain_entries, page_cnt, in qedr_create_cq() 1541 u32 page_cnt, page_size; in qedr_create_srq() local 1573 page_cnt = srq->usrq.pbl_info.num_pbes; in qedr_create_srq() 1585 page_cnt = qed_chain_get_page_cnt(pbl); in qedr_create_srq() 1594 in_params.num_pages = page_cnt; in qedr_create_srq()
|
| /Linux-v5.10/tools/lib/bpf/ |
| D | libbpf.h | 558 perf_buffer__new(int map_fd, size_t page_cnt, 591 perf_buffer__new_raw(int map_fd, size_t page_cnt,
|
| D | libbpf.c | 9992 static struct perf_buffer *__perf_buffer__new(int map_fd, size_t page_cnt, 9995 struct perf_buffer *perf_buffer__new(int map_fd, size_t page_cnt, in perf_buffer__new() argument 10012 return __perf_buffer__new(map_fd, page_cnt, &p); in perf_buffer__new() 10016 perf_buffer__new_raw(int map_fd, size_t page_cnt, in perf_buffer__new_raw() argument 10028 return __perf_buffer__new(map_fd, page_cnt, &p); in perf_buffer__new_raw() 10031 static struct perf_buffer *__perf_buffer__new(int map_fd, size_t page_cnt, in __perf_buffer__new() argument 10042 if (page_cnt & (page_cnt - 1)) { in __perf_buffer__new() 10044 page_cnt); in __perf_buffer__new() 10082 pb->mmap_size = pb->page_size * page_cnt; in __perf_buffer__new()
|
| /Linux-v5.10/fs/btrfs/ |
| D | ioctl.c | 1276 u64 page_cnt; in cluster_pages_for_defrag() local 1291 page_cnt = min_t(u64, (u64)num_pages, (u64)file_end - start_index + 1); in cluster_pages_for_defrag() 1294 start, page_cnt << PAGE_SHIFT); in cluster_pages_for_defrag() 1301 for (i = 0; i < page_cnt; i++) { in cluster_pages_for_defrag() 1378 if (i_done != page_cnt) { in cluster_pages_for_defrag() 1383 start, (page_cnt - i_done) << PAGE_SHIFT, true); in cluster_pages_for_defrag() 1401 btrfs_delalloc_release_extents(BTRFS_I(inode), page_cnt << PAGE_SHIFT); in cluster_pages_for_defrag() 1410 start, page_cnt << PAGE_SHIFT, true); in cluster_pages_for_defrag() 1411 btrfs_delalloc_release_extents(BTRFS_I(inode), page_cnt << PAGE_SHIFT); in cluster_pages_for_defrag()
|
| /Linux-v5.10/drivers/net/ethernet/qlogic/qede/ |
| D | qede_main.c | 2080 u32 page_cnt = qed_chain_get_page_cnt(&txq->tx_pbl); in qede_start_txq() local 2101 page_cnt, &ret_params); in qede_start_txq() 2170 u32 page_cnt; in qede_start_queues() local 2186 page_cnt = qed_chain_get_page_cnt(&rxq->rx_comp_ring); in qede_start_queues() 2192 page_cnt, &ret_params); in qede_start_queues()
|
| /Linux-v5.10/drivers/mmc/core/ |
| D | mmc_test.c | 343 unsigned long page_cnt = 0; in mmc_test_alloc_mem() local 380 if (page_cnt < min_page_cnt) in mmc_test_alloc_mem() 390 page_cnt += 1UL << order; in mmc_test_alloc_mem() 392 if (page_cnt < min_page_cnt) in mmc_test_alloc_mem()
|
| /Linux-v5.10/arch/sparc/kernel/ |
| D | traps_64.c | 2167 unsigned long page_cnt = DIV_ROUND_UP(ent->err_size, in sun4v_nonresum_error_user_handled() local 2174 page_cnt); in sun4v_nonresum_error_user_handled() 2176 while (page_cnt-- > 0) { in sun4v_nonresum_error_user_handled()
|