/Linux-v5.4/include/drm/ttm/ |
D | ttm_set_memory.h | 40 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() 45 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc() 50 static inline int ttm_set_pages_array_uc(struct page **pages, int addrinarray) in ttm_set_pages_array_uc() 78 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() 87 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc() 96 static inline int ttm_set_pages_array_uc(struct page **pages, int addrinarray) in ttm_set_pages_array_uc() 116 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() 121 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc() 126 static inline int ttm_set_pages_array_uc(struct page **pages, int addrinarray) in ttm_set_pages_array_uc()
|
/Linux-v5.4/net/ceph/ |
D | pagevec.c | 13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector() 26 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector() 41 struct page **pages; in ceph_alloc_page_vector() local 61 int ceph_copy_user_to_page_vector(struct page **pages, in ceph_copy_user_to_page_vector() 87 void ceph_copy_to_page_vector(struct page **pages, in ceph_copy_to_page_vector() 110 void ceph_copy_from_page_vector(struct page **pages, in ceph_copy_from_page_vector() 137 void ceph_zero_page_vector_range(int off, int len, struct page **pages) in ceph_zero_page_vector_range()
|
/Linux-v5.4/mm/ |
D | gup.c | 54 void put_user_pages_dirty_lock(struct page **pages, unsigned long npages, in put_user_pages_dirty_lock() 108 void put_user_pages(struct page **pages, unsigned long npages) in put_user_pages() 781 unsigned int gup_flags, struct page **pages, in __get_user_pages() 1002 struct page **pages, in __get_user_pages_locked() 1151 unsigned int gup_flags, struct page **pages, in get_user_pages_remote() 1318 unsigned long nr_pages, struct page **pages, in __get_user_pages_locked() 1437 struct page **pages, in check_and_migrate_cma_pages() 1524 struct page **pages, in check_and_migrate_cma_pages() 1540 struct page **pages, in __gup_longterm_locked() 1591 struct page **pages, in __gup_longterm_locked() [all …]
|
D | percpu-vm.c | 33 static struct page **pages; in pcpu_get_pages() local 54 struct page **pages, int page_start, int page_end) in pcpu_free_pages() 82 struct page **pages, int page_start, int page_end, in pcpu_alloc_pages() 153 struct page **pages, int page_start, int page_end) in pcpu_unmap_pages() 192 static int __pcpu_map_pages(unsigned long addr, struct page **pages, in __pcpu_map_pages() 214 struct page **pages, int page_start, int page_end) in pcpu_map_pages() 278 struct page **pages; in pcpu_populate_chunk() local 311 struct page **pages; in pcpu_depopulate_chunk() local
|
D | frame_vector.c | 126 struct page **pages; in put_vaddr_frames() local 158 struct page **pages; in frame_vector_to_pages() local 184 struct page **pages; in frame_vector_to_pfns() local
|
D | mprotect.c | 44 unsigned long pages = 0; in change_pte_range() local 170 unsigned long pages = 0; in change_pmd_range() local 233 unsigned long pages = 0; in change_pud_range() local 253 unsigned long pages = 0; in change_p4d_range() local 275 unsigned long pages = 0; in change_protection_range() local 301 unsigned long pages; in change_protection() local
|
/Linux-v5.4/include/xen/ |
D | xen-ops.h | 72 bool no_translate, struct page **pages) in xen_remap_pfn() 100 struct page **pages) in xen_xlate_remap_gfn_array() 106 int nr, struct page **pages) in xen_xlate_unmap_gfn_range() 137 struct page **pages) in xen_remap_domain_gfn_array() 173 struct page **pages) in xen_remap_domain_mfn_array() 198 struct page **pages) in xen_remap_domain_gfn_range()
|
D | mem-reservation.h | 38 struct page **pages, in xenmem_reservation_va_mapping_update() 48 struct page **pages) in xenmem_reservation_va_mapping_reset()
|
/Linux-v5.4/drivers/gpu/drm/i915/gem/selftests/ |
D | huge_gem_object.c | 12 struct sg_table *pages) in huge_free_pages() 30 struct sg_table *pages; in huge_get_pages() local 79 struct sg_table *pages) in huge_put_pages()
|
/Linux-v5.4/drivers/xen/ |
D | xlate_mmu.c | 48 static void xen_for_each_gfn(struct page **pages, unsigned nr_gfn, in xen_for_each_gfn() 71 struct page **pages; member 148 struct page **pages) in xen_xlate_remap_gfn_array() 184 int nr, struct page **pages) in xen_xlate_unmap_gfn_range() 217 struct page **pages; in xen_xlate_map_ballooned_pages() local 268 struct page **pages; member
|
D | privcmd.c | 93 static void free_page_list(struct list_head *pages) in free_page_list() 342 struct page **pages = vma->vm_private_data; in mmap_batch_fn() local 424 struct page **pages; in alloc_empty_pages() local 585 struct page *pages[], unsigned int nr_pages) in lock_pages() 612 static void unlock_pages(struct page *pages[], unsigned int nr_pages) in unlock_pages() 631 struct page **pages = NULL; in privcmd_ioctl_dm_op() local 761 struct page **pages; in privcmd_ioctl_mmap_resource() local 895 struct page **pages = vma->vm_private_data; in privcmd_close() local
|
D | mem-reservation.c | 30 struct page **pages, in __xenmem_reservation_va_mapping_update() 64 struct page **pages) in __xenmem_reservation_va_mapping_reset()
|
/Linux-v5.4/drivers/hwtracing/coresight/ |
D | coresight-tmc-etr.c | 46 void **pages; member 192 enum dma_data_direction dir, void **pages) in tmc_pages_alloc() 285 static int tmc_alloc_data_pages(struct tmc_sg_table *sg_table, void **pages) in tmc_alloc_data_pages() 319 void **pages) in tmc_alloc_sg_table() 557 unsigned long size, void **pages) in tmc_init_etr_sg_table() 592 void **pages) in tmc_etr_alloc_flat_buf() 672 void **pages) in tmc_etr_alloc_sg_buf() 798 void **pages) in tmc_etr_mode_alloc_buf() 827 int node, void **pages) in tmc_alloc_etr_buf() 1204 int nr_pages, void **pages, bool snapshot) in alloc_etr_buf() [all …]
|
/Linux-v5.4/fs/cifs/ |
D | fscache.h | 94 struct list_head *pages, in cifs_readpages_from_fscache() 111 struct list_head *pages) in cifs_fscache_readpages_cancel() 148 struct list_head *pages, in cifs_readpages_from_fscache() 158 struct list_head *pages) in cifs_fscache_readpages_cancel()
|
/Linux-v5.4/fs/ceph/ |
D | cache.h | 69 struct list_head *pages) in ceph_fscache_readpages_cancel() 122 struct page *pages) in ceph_fscache_uncache_page() 134 struct list_head *pages, in ceph_readpages_from_fscache() 165 struct list_head *pages) in ceph_fscache_readpages_cancel()
|
/Linux-v5.4/include/linux/ |
D | balloon_compaction.h | 57 struct list_head pages; /* Pages enqueued & handled to Host */ member 188 static inline void balloon_page_push(struct list_head *pages, struct page *page) in balloon_page_push() 200 static inline struct page *balloon_page_pop(struct list_head *pages) in balloon_page_pop()
|
/Linux-v5.4/fs/squashfs/ |
D | page_actor.h | 12 int pages; member 18 int pages, int length) in squashfs_page_actor_init() 58 int pages; member
|
D | page_actor.c | 42 int pages, int length) in squashfs_page_actor_init() 82 int pages, int length) in squashfs_page_actor_init_special()
|
/Linux-v5.4/drivers/gpu/drm/omapdrm/ |
D | omap_gem_dmabuf.c | 69 struct page **pages; in omap_gem_dmabuf_begin_cpu_access() local 92 struct page **pages; in omap_gem_dmabuf_kmap() local 102 struct page **pages; in omap_gem_dmabuf_kunmap() local
|
/Linux-v5.4/drivers/gpu/drm/i915/gem/ |
D | i915_gem_pages.c | 12 struct sg_table *pages, in __i915_gem_object_set_pages() 156 struct sg_table *pages; in __i915_gem_object_unset_pages() local 185 struct sg_table *pages; in __i915_gem_object_put_pages() local 235 struct page **pages = stack_pages; in i915_gem_object_map() local
|
/Linux-v5.4/drivers/staging/media/ipu3/ |
D | ipu3-dmamap.c | 20 static void imgu_dmamap_free_buffer(struct page **pages, in imgu_dmamap_free_buffer() 36 struct page **pages; in imgu_dmamap_alloc_buffer() local 99 struct page **pages; in imgu_dmamap_alloc() local
|
/Linux-v5.4/drivers/gpu/drm/ |
D | drm_cache.c | 60 static void drm_cache_flush_clflush(struct page *pages[], in drm_cache_flush_clflush() 81 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages()
|
/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_gart.c | 222 int pages) in amdgpu_gart_unbind() 274 int pages, dma_addr_t *dma_addr, uint64_t flags, in amdgpu_gart_map() 311 int pages, struct page **pagelist, dma_addr_t *dma_addr, in amdgpu_gart_bind()
|
/Linux-v5.4/arch/x86/include/asm/ |
D | pgtable_32.h | 94 #define PAGE_TABLE_SIZE(pages) (((pages) / PTRS_PER_PMD) + PTRS_PER_PGD) argument 96 #define PAGE_TABLE_SIZE(pages) ((pages) / PTRS_PER_PGD) argument
|
/Linux-v5.4/drivers/block/xen-blkback/ |
D | blkback.c | 308 struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; in free_persistent_gnts() local 349 struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; in xen_blkbk_unmap_purged_grants() local 696 struct grant_page **pages, in xen_blkbk_unmap_prepare() 756 struct grant_page **pages = req->segments; in xen_blkbk_unmap_and_respond() local 781 struct grant_page *pages[], in xen_blkbk_unmap() 805 struct grant_page *pages[], in xen_blkbk_map() 958 struct grant_page **pages = pending_req->indirect_pages; in xen_blkbk_parse_indirect() local 1236 struct grant_page **pages = pending_req->segments; in dispatch_rw_block_io() local
|