Home
last modified time | relevance | path

Searched refs:page_cnt (Results 1 – 24 of 24) sorted by relevance

/Linux-v5.10/drivers/net/ethernet/qlogic/qed/
Dqed_chain.c12 u32 page_cnt) in qed_chain_init() argument
34 chain->page_cnt = page_cnt; in qed_chain_init()
35 chain->capacity = chain->usable_per_page * page_cnt; in qed_chain_init()
36 chain->size = chain->elem_per_page * page_cnt; in qed_chain_init()
80 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_next_ptr()
114 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_pbl()
160 u32 page_cnt) in qed_chain_alloc_sanity_check() argument
165 chain_size *= page_cnt; in qed_chain_alloc_sanity_check()
206 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_alloc_next_ptr()
255 u32 page_cnt, i; in qed_chain_alloc_pbl() local
[all …]
Dqed_sp_commands.c309 u8 page_cnt, i; in qed_sp_pf_start() local
370 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_eq->chain); in qed_sp_pf_start()
371 p_ramrod->event_ring_num_pages = page_cnt; in qed_sp_pf_start()
/Linux-v5.10/include/linux/qed/
Dqed_chain.h106 u32 page_cnt; member
262 return chain->page_cnt; in qed_chain_get_page_cnt()
302 if (++(*(u16 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page()
306 if (++(*(u32 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page()
495 u32 reset_val = p_chain->page_cnt - 1; in qed_chain_reset()
554 last_page_idx = p_chain->page_cnt - 1; in qed_chain_get_last_elem()
575 u32 cur_prod, page_mask, page_cnt, page_diff; in qed_chain_set_prod() local
594 page_cnt = qed_chain_get_page_cnt(p_chain); in qed_chain_set_prod()
598 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod()
602 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod()
[all …]
/Linux-v5.10/arch/mips/cavium-octeon/executive/
Dcvmx-helper-util.c133 union cvmx_ipd_portx_bp_page_cnt page_cnt; in cvmx_helper_setup_red() local
141 page_cnt.u64 = 0; in cvmx_helper_setup_red()
142 page_cnt.s.bp_enb = 0; in cvmx_helper_setup_red()
143 page_cnt.s.page_cnt = 100; in cvmx_helper_setup_red()
148 page_cnt.u64); in cvmx_helper_setup_red()
/Linux-v5.10/drivers/hv/
Dring_buffer.c193 struct page *pages, u32 page_cnt) in hv_ringbuffer_init() argument
204 pages_wraparound = kcalloc(page_cnt * 2 - 1, sizeof(struct page *), in hv_ringbuffer_init()
210 for (i = 0; i < 2 * (page_cnt - 1); i++) in hv_ringbuffer_init()
211 pages_wraparound[i + 1] = &pages[i % (page_cnt - 1) + 1]; in hv_ringbuffer_init()
214 vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, PAGE_KERNEL); in hv_ringbuffer_init()
228 ring_info->ring_size = page_cnt << PAGE_SHIFT; in hv_ringbuffer_init()
Dhv_balloon.c142 __u64 page_cnt:24; member
994 pfn_cnt = dm->ha_wrk.ha_page_range.finfo.page_cnt; in hot_add_req()
997 rg_sz = dm->ha_wrk.ha_region_range.finfo.page_cnt; in hot_add_req()
1191 int num_pages = range_array->finfo.page_cnt; in free_balloon_pages()
1247 bl_resp->range_array[i].finfo.page_cnt = alloc_unit; in alloc_balloon_pages()
/Linux-v5.10/drivers/base/firmware_loader/
Dfallback.c323 int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count); in firmware_rw() local
328 memcpy(buffer, page_data + page_ofs, page_cnt); in firmware_rw()
330 memcpy(page_data + page_ofs, buffer, page_cnt); in firmware_rw()
333 buffer += page_cnt; in firmware_rw()
334 offset += page_cnt; in firmware_rw()
335 count -= page_cnt; in firmware_rw()
/Linux-v5.10/drivers/infiniband/hw/hns/
Dhns_roce_mr.c844 dma_addr_t *pages, int page_cnt) in hns_roce_mtr_map() argument
862 if (r->offset + r->count > page_cnt) { in hns_roce_mtr_map()
866 i, r->offset, r->count, page_cnt); in hns_roce_mtr_map()
948 int page_cnt = 0; in mtr_init_buf_cfg() local
954 page_cnt = DIV_ROUND_UP(buf_size, HNS_HW_PAGE_SIZE); in mtr_init_buf_cfg()
959 cfg->buf_pg_shift = HNS_HW_PAGE_SHIFT + order_base_2(page_cnt); in mtr_init_buf_cfg()
961 cfg->buf_pg_count = page_cnt; in mtr_init_buf_cfg()
978 for (page_cnt = 0, region_cnt = 0; page_cnt < cfg->buf_pg_count && in mtr_init_buf_cfg()
982 r->offset = page_cnt; in mtr_init_buf_cfg()
985 page_cnt += r->count; in mtr_init_buf_cfg()
[all …]
Dhns_roce_device.h1158 dma_addr_t *pages, int page_cnt);
/Linux-v5.10/arch/mips/include/asm/octeon/
Dcvmx-ipd-defs.h324 uint64_t page_cnt:17; member
326 uint64_t page_cnt:17;
999 uint64_t page_cnt:17; member
1001 uint64_t page_cnt:17;
1014 uint64_t page_cnt:17; member
1016 uint64_t page_cnt:17;
1029 uint64_t page_cnt:17; member
1031 uint64_t page_cnt:17;
1387 uint64_t page_cnt:25; member
1389 uint64_t page_cnt:25;
/Linux-v5.10/drivers/staging/rts5208/
Dxd.c1506 u8 reg_val, page_cnt; in xd_read_multiple_pages() local
1512 page_cnt = end_page - start_page; in xd_read_multiple_pages()
1538 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_read_multiple_pages()
1543 page_cnt * 512, DMA_512); in xd_read_multiple_pages()
1553 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 512, in xd_read_multiple_pages()
1709 u8 page_cnt, reg_val; in xd_write_multiple_pages() local
1717 page_cnt = end_page - start_page; in xd_write_multiple_pages()
1739 rtsx_add_cmd(chip, WRITE_REG_CMD, XD_PAGE_CNT, 0xFF, page_cnt); in xd_write_multiple_pages()
1743 page_cnt * 512, DMA_512); in xd_write_multiple_pages()
1752 retval = rtsx_transfer_data_partial(chip, XD_CARD, buf, page_cnt * 512, in xd_write_multiple_pages()
[all …]
Dms.c3400 u8 start_page, end_page = 0, page_cnt; local
3526 page_cnt = end_page - start_page;
3529 start_page, end_page, page_cnt);
3566 total_sec_cnt -= page_cnt;
3568 ptr += page_cnt * 512;
/Linux-v5.10/drivers/infiniband/hw/i40iw/
Di40iw_verbs.h95 u32 page_cnt; member
Di40iw_verbs.c1389 status = i40iw_get_pble(&iwdev->sc_dev, iwdev->pble_rsrc, palloc, iwmr->page_cnt); in i40iw_setup_pbles()
1565 iwmr->page_cnt = max_num_sg; in i40iw_alloc_mr()
1567 status = i40iw_get_pble(&iwdev->sc_dev, iwdev->pble_rsrc, palloc, iwmr->page_cnt); in i40iw_alloc_mr()
1601 if (unlikely(iwmr->npages == iwmr->page_cnt)) in i40iw_set_page()
1781 iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size); in i40iw_reg_user_mr()
1806 use_pbles = (iwmr->page_cnt != 1); in i40iw_reg_user_mr()
1897 iwmr->page_cnt = 1; in i40iw_reg_phys_mr()
/Linux-v5.10/drivers/infiniband/hw/efa/
Defa_verbs.c1162 static struct scatterlist *efa_vmalloc_buf_to_sg(u64 *buf, int page_cnt) in efa_vmalloc_buf_to_sg() argument
1168 sglist = kmalloc_array(page_cnt, sizeof(*sglist), GFP_KERNEL); in efa_vmalloc_buf_to_sg()
1171 sg_init_table(sglist, page_cnt); in efa_vmalloc_buf_to_sg()
1172 for (i = 0; i < page_cnt; i++) { in efa_vmalloc_buf_to_sg()
1193 int page_cnt = pbl->phys.indirect.pbl_buf_size_in_pages; in pbl_chunk_list_create() local
1204 chunk_list_size = DIV_ROUND_UP(page_cnt, EFA_PTRS_PER_CHUNK); in pbl_chunk_list_create()
1215 page_cnt); in pbl_chunk_list_create()
1226 ((page_cnt % EFA_PTRS_PER_CHUNK) * EFA_CHUNK_PAYLOAD_PTR_SIZE) + in pbl_chunk_list_create()
/Linux-v5.10/fs/ubifs/
Dfile.c724 int err, page_idx, page_cnt, ret = 0, n = 0; in ubifs_do_bulk_read() local
739 page_cnt = bu->blk_cnt >> UBIFS_BLOCKS_PER_PAGE_SHIFT; in ubifs_do_bulk_read()
740 if (!page_cnt) { in ubifs_do_bulk_read()
783 for (page_idx = 1; page_idx < page_cnt; page_idx++) { in ubifs_do_bulk_read()
/Linux-v5.10/drivers/infiniband/hw/qedr/
Dmain.c811 u32 page_cnt; in qedr_init_hw() local
825 page_cnt = qed_chain_get_page_cnt(&dev->cnq_array[i].pbl); in qedr_init_hw()
826 cur_pbl->num_pbl_pages = page_cnt; in qedr_init_hw()
Dverbs.c841 int chain_entries, int page_cnt, in qedr_init_cq_params() argument
852 params->pbl_num_pages = page_cnt; in qedr_init_cq_params()
922 int page_cnt; in qedr_create_cq() local
968 page_cnt = cq->q.pbl_info.num_pbes; in qedr_create_cq()
980 page_cnt = qed_chain_get_page_cnt(&cq->pbl); in qedr_create_cq()
985 qedr_init_cq_params(cq, ctx, dev, vector, chain_entries, page_cnt, in qedr_create_cq()
1541 u32 page_cnt, page_size; in qedr_create_srq() local
1573 page_cnt = srq->usrq.pbl_info.num_pbes; in qedr_create_srq()
1585 page_cnt = qed_chain_get_page_cnt(pbl); in qedr_create_srq()
1594 in_params.num_pages = page_cnt; in qedr_create_srq()
/Linux-v5.10/tools/lib/bpf/
Dlibbpf.h558 perf_buffer__new(int map_fd, size_t page_cnt,
591 perf_buffer__new_raw(int map_fd, size_t page_cnt,
Dlibbpf.c9992 static struct perf_buffer *__perf_buffer__new(int map_fd, size_t page_cnt,
9995 struct perf_buffer *perf_buffer__new(int map_fd, size_t page_cnt, in perf_buffer__new() argument
10012 return __perf_buffer__new(map_fd, page_cnt, &p); in perf_buffer__new()
10016 perf_buffer__new_raw(int map_fd, size_t page_cnt, in perf_buffer__new_raw() argument
10028 return __perf_buffer__new(map_fd, page_cnt, &p); in perf_buffer__new_raw()
10031 static struct perf_buffer *__perf_buffer__new(int map_fd, size_t page_cnt, in __perf_buffer__new() argument
10042 if (page_cnt & (page_cnt - 1)) { in __perf_buffer__new()
10044 page_cnt); in __perf_buffer__new()
10082 pb->mmap_size = pb->page_size * page_cnt; in __perf_buffer__new()
/Linux-v5.10/fs/btrfs/
Dioctl.c1276 u64 page_cnt; in cluster_pages_for_defrag() local
1291 page_cnt = min_t(u64, (u64)num_pages, (u64)file_end - start_index + 1); in cluster_pages_for_defrag()
1294 start, page_cnt << PAGE_SHIFT); in cluster_pages_for_defrag()
1301 for (i = 0; i < page_cnt; i++) { in cluster_pages_for_defrag()
1378 if (i_done != page_cnt) { in cluster_pages_for_defrag()
1383 start, (page_cnt - i_done) << PAGE_SHIFT, true); in cluster_pages_for_defrag()
1401 btrfs_delalloc_release_extents(BTRFS_I(inode), page_cnt << PAGE_SHIFT); in cluster_pages_for_defrag()
1410 start, page_cnt << PAGE_SHIFT, true); in cluster_pages_for_defrag()
1411 btrfs_delalloc_release_extents(BTRFS_I(inode), page_cnt << PAGE_SHIFT); in cluster_pages_for_defrag()
/Linux-v5.10/drivers/net/ethernet/qlogic/qede/
Dqede_main.c2080 u32 page_cnt = qed_chain_get_page_cnt(&txq->tx_pbl); in qede_start_txq() local
2101 page_cnt, &ret_params); in qede_start_txq()
2170 u32 page_cnt; in qede_start_queues() local
2186 page_cnt = qed_chain_get_page_cnt(&rxq->rx_comp_ring); in qede_start_queues()
2192 page_cnt, &ret_params); in qede_start_queues()
/Linux-v5.10/drivers/mmc/core/
Dmmc_test.c343 unsigned long page_cnt = 0; in mmc_test_alloc_mem() local
380 if (page_cnt < min_page_cnt) in mmc_test_alloc_mem()
390 page_cnt += 1UL << order; in mmc_test_alloc_mem()
392 if (page_cnt < min_page_cnt) in mmc_test_alloc_mem()
/Linux-v5.10/arch/sparc/kernel/
Dtraps_64.c2167 unsigned long page_cnt = DIV_ROUND_UP(ent->err_size, in sun4v_nonresum_error_user_handled() local
2174 page_cnt); in sun4v_nonresum_error_user_handled()
2176 while (page_cnt-- > 0) { in sun4v_nonresum_error_user_handled()