Home
last modified time | relevance | path

Searched refs:page_idx (Results 1 – 25 of 26) sorted by relevance

12

/Linux-v5.4/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_wq.c45 ((void *)((wqs)->page_vaddr[(wq)->page_idx]) \
49 ((wqs)->page_paddr[(wq)->page_idx] \
53 ((void *)((wqs)->shadow_page_vaddr[(wq)->page_idx]) \
136 static int wqs_allocate_page(struct hinic_wqs *wqs, int page_idx) in wqs_allocate_page() argument
138 return queue_alloc_page(wqs->hwif, &wqs->page_vaddr[page_idx], in wqs_allocate_page()
139 &wqs->page_paddr[page_idx], in wqs_allocate_page()
140 &wqs->shadow_page_vaddr[page_idx], in wqs_allocate_page()
149 static void wqs_free_page(struct hinic_wqs *wqs, int page_idx) in wqs_free_page() argument
155 wqs->page_vaddr[page_idx], in wqs_free_page()
156 (dma_addr_t)wqs->page_paddr[page_idx]); in wqs_free_page()
[all …]
Dhinic_hw_wq.h18 int page_idx; member
25 int page_idx; member
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx4/
Dicm.h75 int page_idx; member
103 iter->page_idx = 0; in mlx4_icm_first()
113 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next()
121 iter->page_idx = 0; in mlx4_icm_next()
128 return iter->chunk->buf[iter->page_idx].dma_addr; in mlx4_icm_addr()
130 return sg_dma_address(&iter->chunk->sg[iter->page_idx]); in mlx4_icm_addr()
136 return iter->chunk->buf[iter->page_idx].size; in mlx4_icm_size()
138 return sg_dma_len(&iter->chunk->sg[iter->page_idx]); in mlx4_icm_size()
/Linux-v5.4/drivers/infiniband/hw/mthca/
Dmthca_memfree.h77 int page_idx; member
106 iter->page_idx = 0; in mthca_icm_first()
116 if (++iter->page_idx >= iter->chunk->nsg) { in mthca_icm_next()
124 iter->page_idx = 0; in mthca_icm_next()
130 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mthca_icm_addr()
135 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mthca_icm_size()
/Linux-v5.4/mm/
Dreadahead.c120 unsigned page_idx; in read_pages() local
132 for (page_idx = 0; page_idx < nr_pages; page_idx++) { in read_pages()
163 int page_idx; in __do_page_cache_readahead() local
176 for (page_idx = 0; page_idx < nr_to_read; page_idx++) { in __do_page_cache_readahead()
177 pgoff_t page_offset = offset + page_idx; in __do_page_cache_readahead()
201 if (page_idx == nr_to_read - lookahead_size) in __do_page_cache_readahead()
Dpercpu-vm.c13 unsigned int cpu, int page_idx) in pcpu_chunk_page() argument
18 return vmalloc_to_page((void *)pcpu_chunk_addr(chunk, cpu, page_idx)); in pcpu_chunk_page()
Dpercpu.c256 static int __maybe_unused pcpu_page_idx(unsigned int cpu, int page_idx) in pcpu_page_idx() argument
258 return pcpu_unit_map[cpu] * pcpu_unit_pages + page_idx; in pcpu_page_idx()
261 static unsigned long pcpu_unit_page_offset(unsigned int cpu, int page_idx) in pcpu_unit_page_offset() argument
263 return pcpu_unit_offsets[cpu] + (page_idx << PAGE_SHIFT); in pcpu_unit_page_offset()
267 unsigned int cpu, int page_idx) in pcpu_chunk_addr() argument
270 pcpu_unit_page_offset(cpu, page_idx); in pcpu_chunk_addr()
/Linux-v5.4/drivers/infiniband/sw/siw/
Dsiw_mem.h65 unsigned int page_idx = (addr - umem->fp_addr) >> PAGE_SHIFT, in siw_get_upage() local
66 chunk_idx = page_idx >> CHUNK_SHIFT, in siw_get_upage()
67 page_in_chunk = page_idx & ~CHUNK_MASK; in siw_get_upage()
69 if (likely(page_idx < umem->num_pages)) in siw_get_upage()
/Linux-v5.4/drivers/infiniband/hw/hns/
Dhns_roce_hem.h97 int page_idx; member
157 iter->page_idx = 0; in hns_roce_hem_first()
167 if (++iter->page_idx >= iter->chunk->nsg) { in hns_roce_hem_next()
175 iter->page_idx = 0; in hns_roce_hem_next()
181 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in hns_roce_hem_addr()
/Linux-v5.4/drivers/infiniband/hw/mlx5/
Dcmd.c97 u64 page_idx = 0; in mlx5_cmd_alloc_memic() local
117 while (page_idx < num_memic_hw_pages) { in mlx5_cmd_alloc_memic()
119 page_idx = bitmap_find_next_zero_area(dm->memic_alloc_pages, in mlx5_cmd_alloc_memic()
121 page_idx, in mlx5_cmd_alloc_memic()
124 if (page_idx < num_memic_hw_pages) in mlx5_cmd_alloc_memic()
126 page_idx, num_pages); in mlx5_cmd_alloc_memic()
130 if (page_idx >= num_memic_hw_pages) in mlx5_cmd_alloc_memic()
134 hw_start_addr + (page_idx * PAGE_SIZE)); in mlx5_cmd_alloc_memic()
140 page_idx, num_pages); in mlx5_cmd_alloc_memic()
144 page_idx++; in mlx5_cmd_alloc_memic()
Dmain.c2199 u16 page_idx = get_extended_index(vma->vm_pgoff); in dm_mmap() local
2204 if (find_next_zero_bit(mctx->dm_pages, page_idx + npages, page_idx) != in dm_mmap()
2205 page_idx + npages) in dm_mmap()
2211 page_idx; in dm_mmap()
2295 u32 page_idx; in handle_alloc_dm_memic() local
2305 page_idx = (dm->dev_addr - pci_resource_start(dm_db->dev->pdev, 0) - in handle_alloc_dm_memic()
2311 &page_idx, sizeof(page_idx)); in handle_alloc_dm_memic()
2322 bitmap_set(to_mucontext(ctx)->dm_pages, page_idx, in handle_alloc_dm_memic()
2432 u32 page_idx; in mlx5_ib_dealloc_dm() local
2441 page_idx = (dm->dev_addr - pci_resource_start(dev->pdev, 0) - in mlx5_ib_dealloc_dm()
[all …]
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
Drx.c91 u32 page_idx) in mlx5e_xsk_skb_from_cqe_mpwrq_linear() argument
93 struct mlx5e_dma_info *di = &wi->umr.dma_info[page_idx]; in mlx5e_xsk_skb_from_cqe_mpwrq_linear()
141 __set_bit(page_idx, wi->xdp_xmit_bitmap); /* non-atomic */ in mlx5e_xsk_skb_from_cqe_mpwrq_linear()
Drx.h22 u32 page_idx);
/Linux-v5.4/fs/
Dmpage.c392 unsigned page_idx; in mpage_readpages() local
394 for (page_idx = 0; page_idx < nr_pages; page_idx++) { in mpage_readpages()
403 args.nr_pages = nr_pages - page_idx; in mpage_readpages()
/Linux-v5.4/sound/soc/amd/raven/
Dacp3x-pcm-dma.c213 u16 page_idx; in config_acp3x_dma() local
229 for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) { in config_acp3x_dma()
/Linux-v5.4/drivers/mmc/host/
Dusdhi6rol0.c179 int page_idx; /* page index within an SG segment */ member
418 (host->page_idx << PAGE_SHIFT) + data->blksz - blk_head) in usdhi6_sg_unmap()
442 host->page_idx++; in usdhi6_sg_advance()
452 host->page_idx++; in usdhi6_sg_advance()
462 done = (host->page_idx << PAGE_SHIFT) + host->offset; in usdhi6_sg_advance()
488 host->page_idx = 0; in usdhi6_sg_advance()
505 host->pg.page = nth_page(sg_page(host->sg), host->page_idx); in usdhi6_sg_advance()
1006 host->page_idx = 0; in usdhi6_rq_start()
1708 data->flags & MMC_DATA_READ ? 'R' : 'W', host->page_idx, in usdhi6_timeout_work()
/Linux-v5.4/sound/soc/amd/
Dacp-pcm-dma.c301 u16 page_idx; in acp_pte_config() local
307 for (page_idx = 0; page_idx < (num_of_pages); page_idx++) { in acp_pte_config()
309 acp_reg_write((offset + (page_idx * 8)), in acp_pte_config()
318 acp_reg_write((offset + (page_idx * 8) + 4), in acp_pte_config()
/Linux-v5.4/fs/ubifs/
Dfile.c724 int err, page_idx, page_cnt, ret = 0, n = 0; in ubifs_do_bulk_read() local
783 for (page_idx = 1; page_idx < page_cnt; page_idx++) { in ubifs_do_bulk_read()
784 pgoff_t page_offset = offset + page_idx; in ubifs_do_bulk_read()
800 ui->last_page_read = offset + page_idx - 1; in ubifs_do_bulk_read()
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/
Den_rx.c1237 u16 cqe_bcnt, u32 head_offset, u32 page_idx) in mlx5e_skb_from_cqe_mpwrq_nonlinear() argument
1240 struct mlx5e_dma_info *di = &wi->umr.dma_info[page_idx]; in mlx5e_skb_from_cqe_mpwrq_nonlinear()
1283 u16 cqe_bcnt, u32 head_offset, u32 page_idx) in mlx5e_skb_from_cqe_mpwrq_linear() argument
1285 struct mlx5e_dma_info *di = &wi->umr.dma_info[page_idx]; in mlx5e_skb_from_cqe_mpwrq_linear()
1313 __set_bit(page_idx, wi->xdp_xmit_bitmap); /* non-atomic */ in mlx5e_skb_from_cqe_mpwrq_linear()
1335 u32 page_idx = wqe_offset >> PAGE_SHIFT; in mlx5e_handle_rx_cqe_mpwrq() local
1362 rq, wi, cqe_bcnt, head_offset, page_idx); in mlx5e_handle_rx_cqe_mpwrq()
Den.h603 u16 cqe_bcnt, u32 head_offset, u32 page_idx);
956 u16 cqe_bcnt, u32 head_offset, u32 page_idx);
959 u16 cqe_bcnt, u32 head_offset, u32 page_idx);
/Linux-v5.4/drivers/input/touchscreen/
Draydium_i2c_ts.c574 u16 page_idx, const void *data, size_t len) in raydium_i2c_fw_write_page() argument
585 buf[BL_PAGE_STR] = page_idx ? 0xff : 0; in raydium_i2c_fw_write_page()
599 page_idx, i, error); in raydium_i2c_fw_write_page()
/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/
Damdgpu_ttm.c1007 uint64_t page_idx = 1; in amdgpu_ttm_gart_bind() local
1009 r = amdgpu_gart_bind(adev, gtt->offset, page_idx, in amdgpu_ttm_gart_bind()
1019 gtt->offset + (page_idx << PAGE_SHIFT), in amdgpu_ttm_gart_bind()
1020 ttm->num_pages - page_idx, in amdgpu_ttm_gart_bind()
1021 &ttm->pages[page_idx], in amdgpu_ttm_gart_bind()
1022 &(gtt->ttm.dma_address[page_idx]), flags); in amdgpu_ttm_gart_bind()
/Linux-v5.4/drivers/md/
Draid1.c2648 int page_idx = 0; in raid1_sync_request() local
2879 page = resync_fetch_page(rp, page_idx); in raid1_sync_request()
2891 } while (++page_idx < RESYNC_PAGES); in raid1_sync_request()
Draid10.c2915 int page_idx = 0; in raid10_sync_request() local
3431 page = resync_fetch_page(rp, page_idx); in raid10_sync_request()
3440 } while (++page_idx < RESYNC_PAGES); in raid10_sync_request()
/Linux-v5.4/Documentation/vm/
Dhugetlbfs_reserv.rst431 [page_idx, page_idx + 1).

12