Home
last modified time | relevance | path

Searched refs:pages (Results 1 – 25 of 1155) sorted by relevance

12345678910>>...47

/Linux-v5.4/net/ceph/
Dpagevec.c13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector() argument
19 set_page_dirty_lock(pages[i]); in ceph_put_page_vector()
20 put_page(pages[i]); in ceph_put_page_vector()
22 kvfree(pages); in ceph_put_page_vector()
26 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector() argument
31 __free_pages(pages[i], 0); in ceph_release_page_vector()
32 kfree(pages); in ceph_release_page_vector()
41 struct page **pages; in ceph_alloc_page_vector() local
44 pages = kmalloc_array(num_pages, sizeof(*pages), flags); in ceph_alloc_page_vector()
45 if (!pages) in ceph_alloc_page_vector()
[all …]
/Linux-v5.4/mm/
Dpercpu-vm.c33 static struct page **pages; in pcpu_get_pages() local
34 size_t pages_size = pcpu_nr_units * pcpu_unit_pages * sizeof(pages[0]); in pcpu_get_pages()
38 if (!pages) in pcpu_get_pages()
39 pages = pcpu_mem_zalloc(pages_size, GFP_KERNEL); in pcpu_get_pages()
40 return pages; in pcpu_get_pages()
54 struct page **pages, int page_start, int page_end) in pcpu_free_pages() argument
61 struct page *page = pages[pcpu_page_idx(cpu, i)]; in pcpu_free_pages()
82 struct page **pages, int page_start, int page_end, in pcpu_alloc_pages() argument
92 struct page **pagep = &pages[pcpu_page_idx(cpu, i)]; in pcpu_alloc_pages()
103 __free_page(pages[pcpu_page_idx(cpu, i)]); in pcpu_alloc_pages()
[all …]
Dgup.c54 void put_user_pages_dirty_lock(struct page **pages, unsigned long npages, in put_user_pages_dirty_lock() argument
66 put_user_pages(pages, npages); in put_user_pages_dirty_lock()
71 struct page *page = compound_head(pages[index]); in put_user_pages_dirty_lock()
108 void put_user_pages(struct page **pages, unsigned long npages) in put_user_pages() argument
118 put_user_page(pages[index]); in put_user_pages()
781 unsigned int gup_flags, struct page **pages, in __get_user_pages() argument
793 VM_BUG_ON(!!pages != !!(gup_flags & FOLL_GET)); in __get_user_pages()
814 pages ? &pages[i] : NULL); in __get_user_pages()
826 i = follow_hugetlb_page(mm, vma, pages, vmas, in __get_user_pages()
871 if (pages) { in __get_user_pages()
[all …]
/Linux-v5.4/fs/isofs/
Dcompress.c42 struct page **pages, unsigned poffset, in zisofs_uncompress_block() argument
68 if (!pages[i]) in zisofs_uncompress_block()
70 memset(page_address(pages[i]), 0, PAGE_SIZE); in zisofs_uncompress_block()
71 flush_dcache_page(pages[i]); in zisofs_uncompress_block()
72 SetPageUptodate(pages[i]); in zisofs_uncompress_block()
122 if (pages[curpage]) { in zisofs_uncompress_block()
123 stream.next_out = page_address(pages[curpage]) in zisofs_uncompress_block()
175 if (pages[curpage]) { in zisofs_uncompress_block()
176 flush_dcache_page(pages[curpage]); in zisofs_uncompress_block()
177 SetPageUptodate(pages[curpage]); in zisofs_uncompress_block()
[all …]
/Linux-v5.4/drivers/gpu/drm/i915/gem/selftests/
Dhuge_gem_object.c12 struct sg_table *pages) in huge_free_pages() argument
17 for (sg = pages->sgl; sg && nreal--; sg = __sg_next(sg)) in huge_free_pages()
20 sg_free_table(pages); in huge_free_pages()
21 kfree(pages); in huge_free_pages()
30 struct sg_table *pages; in huge_get_pages() local
33 pages = kmalloc(sizeof(*pages), GFP); in huge_get_pages()
34 if (!pages) in huge_get_pages()
37 if (sg_alloc_table(pages, npages, GFP)) { in huge_get_pages()
38 kfree(pages); in huge_get_pages()
42 sg = pages->sgl; in huge_get_pages()
[all …]
/Linux-v5.4/include/drm/ttm/
Dttm_set_memory.h40 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() argument
42 return set_pages_array_wb(pages, addrinarray); in ttm_set_pages_array_wb()
45 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc() argument
47 return set_pages_array_wc(pages, addrinarray); in ttm_set_pages_array_wc()
50 static inline int ttm_set_pages_array_uc(struct page **pages, int addrinarray) in ttm_set_pages_array_uc() argument
52 return set_pages_array_uc(pages, addrinarray); in ttm_set_pages_array_uc()
78 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() argument
83 unmap_page_from_agp(pages[i]); in ttm_set_pages_array_wb()
87 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc() argument
92 map_page_into_agp(pages[i]); in ttm_set_pages_array_wc()
[all …]
/Linux-v5.4/drivers/gpu/drm/vkms/
Dvkms_gem.c35 WARN_ON(gem->pages); in vkms_gem_free_object()
59 if (obj->pages) { in vkms_gem_fault()
60 get_page(obj->pages[page_offset]); in vkms_gem_fault()
61 vmf->page = obj->pages[page_offset]; in vkms_gem_fault()
152 if (!vkms_obj->pages) { in _get_pages()
153 struct page **pages = drm_gem_get_pages(gem_obj); in _get_pages() local
155 if (IS_ERR(pages)) in _get_pages()
156 return pages; in _get_pages()
158 if (cmpxchg(&vkms_obj->pages, NULL, pages)) in _get_pages()
159 drm_gem_put_pages(gem_obj, pages, false, true); in _get_pages()
[all …]
/Linux-v5.4/drivers/xen/
Dxlate_mmu.c48 static void xen_for_each_gfn(struct page **pages, unsigned nr_gfn, in xen_for_each_gfn() argument
57 page = pages[i / XEN_PFN_PER_PAGE]; in xen_for_each_gfn()
71 struct page **pages; member
99 struct page *page = info->pages[info->index++]; in remap_pte_fn()
148 struct page **pages) in xen_xlate_remap_gfn_array() argument
163 data.pages = pages; in xen_xlate_remap_gfn_array()
184 int nr, struct page **pages) in xen_xlate_unmap_gfn_range() argument
186 xen_for_each_gfn(pages, nr, unmap_gfn, NULL); in xen_xlate_unmap_gfn_range()
217 struct page **pages; in xen_xlate_map_ballooned_pages() local
226 pages = kcalloc(nr_pages, sizeof(pages[0]), GFP_KERNEL); in xen_xlate_map_ballooned_pages()
[all …]
/Linux-v5.4/drivers/gpu/drm/xen/
Dxen_drm_front_gem.c29 struct page **pages; member
48 xen_obj->pages = kvmalloc_array(xen_obj->num_pages, in gem_alloc_pages_array()
50 return !xen_obj->pages ? -ENOMEM : 0; in gem_alloc_pages_array()
55 kvfree(xen_obj->pages); in gem_free_pages_array()
56 xen_obj->pages = NULL; in gem_free_pages_array()
103 xen_obj->pages); in gem_create()
119 xen_obj->pages = drm_gem_get_pages(&xen_obj->base); in gem_create()
120 if (IS_ERR_OR_NULL(xen_obj->pages)) { in gem_create()
121 ret = PTR_ERR(xen_obj->pages); in gem_create()
122 xen_obj->pages = NULL; in gem_create()
[all …]
/Linux-v5.4/drivers/staging/media/ipu3/
Dipu3-dmamap.c20 static void imgu_dmamap_free_buffer(struct page **pages, in imgu_dmamap_free_buffer() argument
26 __free_page(pages[count]); in imgu_dmamap_free_buffer()
27 kvfree(pages); in imgu_dmamap_free_buffer()
36 struct page **pages; in imgu_dmamap_alloc_buffer() local
42 pages = kvmalloc_array(count, sizeof(*pages), GFP_KERNEL); in imgu_dmamap_alloc_buffer()
44 if (!pages) in imgu_dmamap_alloc_buffer()
72 imgu_dmamap_free_buffer(pages, i << PAGE_SHIFT); in imgu_dmamap_alloc_buffer()
77 pages[i++] = page++; in imgu_dmamap_alloc_buffer()
80 return pages; in imgu_dmamap_alloc_buffer()
99 struct page **pages; in imgu_dmamap_alloc() local
[all …]
/Linux-v5.4/fs/ramfs/
Dfile-nommu.c65 struct page *pages; in ramfs_nommu_expand_for_mapping() local
84 pages = alloc_pages(gfp, order); in ramfs_nommu_expand_for_mapping()
85 if (!pages) in ramfs_nommu_expand_for_mapping()
92 split_page(pages, order); in ramfs_nommu_expand_for_mapping()
96 __free_page(pages + loop); in ramfs_nommu_expand_for_mapping()
100 data = page_address(pages); in ramfs_nommu_expand_for_mapping()
105 struct page *page = pages + loop; in ramfs_nommu_expand_for_mapping()
124 __free_page(pages + loop++); in ramfs_nommu_expand_for_mapping()
207 struct page **pages = NULL, **ptr, *page; in ramfs_nommu_get_unmapped_area() local
223 pages = kcalloc(lpages, sizeof(struct page *), GFP_KERNEL); in ramfs_nommu_get_unmapped_area()
[all …]
/Linux-v5.4/drivers/gpu/drm/ttm/
Dttm_page_alloc.c247 static void ttm_pages_put(struct page *pages[], unsigned npages, in ttm_pages_put() argument
253 if (ttm_set_pages_array_wb(pages, npages)) in ttm_pages_put()
259 if (ttm_set_pages_wb(pages[i], pages_nr)) in ttm_pages_put()
262 __free_pages(pages[i], order); in ttm_pages_put()
442 static int ttm_set_pages_caching(struct page **pages, in ttm_set_pages_caching() argument
449 r = ttm_set_pages_array_uc(pages, cpages); in ttm_set_pages_caching()
454 r = ttm_set_pages_array_wc(pages, cpages); in ttm_set_pages_caching()
469 static void ttm_handle_caching_state_failure(struct list_head *pages, in ttm_handle_caching_state_failure() argument
487 static int ttm_alloc_new_pages(struct list_head *pages, gfp_t gfp_flags, in ttm_alloc_new_pages() argument
519 ttm_handle_caching_state_failure(pages, in ttm_alloc_new_pages()
[all …]
/Linux-v5.4/include/xen/
Dxen-ops.h67 unsigned int domid, bool no_translate, struct page **pages);
72 bool no_translate, struct page **pages) in xen_remap_pfn() argument
87 struct page **pages);
89 int nr, struct page **pages);
100 struct page **pages) in xen_xlate_remap_gfn_array() argument
106 int nr, struct page **pages) in xen_xlate_unmap_gfn_range() argument
137 struct page **pages) in xen_remap_domain_gfn_array() argument
141 prot, domid, pages); in xen_remap_domain_gfn_array()
149 false, pages); in xen_remap_domain_gfn_array()
173 struct page **pages) in xen_remap_domain_mfn_array() argument
[all …]
/Linux-v5.4/fs/squashfs/
Dpage_actor.c30 if (actor->next_page == actor->pages) in cache_next_page()
42 int pages, int length) in squashfs_page_actor_init() argument
49 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init()
51 actor->pages = pages; in squashfs_page_actor_init()
71 return actor->pageaddr = actor->next_page == actor->pages ? NULL : in direct_next_page()
82 int pages, int length) in squashfs_page_actor_init_special() argument
89 actor->length = length ? : pages * PAGE_SIZE; in squashfs_page_actor_init_special()
91 actor->pages = pages; in squashfs_page_actor_init_special()
Dfile_direct.c22 int pages, struct page **page, int bytes);
36 int i, n, pages, missing_pages, bytes, res = -ENOMEM; in squashfs_readpage_block() local
44 pages = end_index - start_index + 1; in squashfs_readpage_block()
46 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL); in squashfs_readpage_block()
54 actor = squashfs_page_actor_init_special(page, pages, 0); in squashfs_readpage_block()
59 for (missing_pages = 0, i = 0, n = start_index; i < pages; i++, n++) { in squashfs_readpage_block()
84 res = squashfs_read_cache(target_page, block, bsize, pages, in squashfs_readpage_block()
105 pageaddr = kmap_atomic(page[pages - 1]); in squashfs_readpage_block()
111 for (i = 0; i < pages; i++) { in squashfs_readpage_block()
128 for (i = 0; i < pages; i++) { in squashfs_readpage_block()
[all …]
/Linux-v5.4/drivers/gpu/drm/lima/
Dlima_object.c13 kfree(bo->pages); in lima_bo_destroy()
27 if (bo->pages) in lima_bo_destroy()
28 drm_gem_put_pages(&bo->gem, bo->pages, true, true); in lima_bo_destroy()
81 bo->pages = kcalloc(npages, sizeof(*bo->pages), GFP_KERNEL); in lima_bo_create()
82 if (!bo->pages) { in lima_bo_create()
88 sgt, bo->pages, bo->pages_dma_addr, npages); in lima_bo_create()
95 bo->pages = drm_gem_get_pages(&bo->gem); in lima_bo_create()
96 if (IS_ERR(bo->pages)) { in lima_bo_create()
97 ret = ERR_CAST(bo->pages); in lima_bo_create()
98 bo->pages = NULL; in lima_bo_create()
[all …]
/Linux-v5.4/drivers/staging/android/ion/
Dion_cma_heap.c33 struct page *pages; in ion_cma_allocate() local
42 pages = cma_alloc(cma_heap->cma, nr_pages, align, false); in ion_cma_allocate()
43 if (!pages) in ion_cma_allocate()
46 if (PageHighMem(pages)) { in ion_cma_allocate()
48 struct page *page = pages; in ion_cma_allocate()
59 memset(page_address(pages), 0, size); in ion_cma_allocate()
70 sg_set_page(table->sgl, pages, size, 0); in ion_cma_allocate()
72 buffer->priv_virt = pages; in ion_cma_allocate()
79 cma_release(cma_heap->cma, pages, nr_pages); in ion_cma_allocate()
86 struct page *pages = buffer->priv_virt; in ion_cma_free() local
[all …]
/Linux-v5.4/arch/arm/plat-omap/
Dsram.c65 int pages; in omap_sram_push() local
73 pages = PAGE_ALIGN(size) / PAGE_SIZE; in omap_sram_push()
75 set_memory_rw(base, pages); in omap_sram_push()
79 set_memory_ro(base, pages); in omap_sram_push()
80 set_memory_x(base, pages); in omap_sram_push()
101 int pages; in omap_map_sram() local
125 pages = PAGE_ALIGN(omap_sram_size) / PAGE_SIZE; in omap_map_sram()
127 set_memory_ro(base, pages); in omap_map_sram()
128 set_memory_x(base, pages); in omap_map_sram()
/Linux-v5.4/drivers/gpu/drm/vgem/
Dvgem_drv.c61 kvfree(vgem_obj->pages); in vgem_gem_free_object()
88 if (obj->pages) { in vgem_gem_fault()
89 get_page(obj->pages[page_offset]); in vgem_gem_fault()
90 vmf->page = obj->pages[page_offset]; in vgem_gem_fault()
291 struct page **pages; in vgem_pin_pages() local
293 pages = drm_gem_get_pages(&bo->base); in vgem_pin_pages()
294 if (IS_ERR(pages)) { in vgem_pin_pages()
297 return pages; in vgem_pin_pages()
300 bo->pages = pages; in vgem_pin_pages()
304 return bo->pages; in vgem_pin_pages()
[all …]
/Linux-v5.4/drivers/gpu/drm/i915/gem/
Di915_gem_pages.c12 struct sg_table *pages, in __i915_gem_object_set_pages() argument
25 drm_clflush_sg(pages); in __i915_gem_object_set_pages()
29 obj->mm.get_page.sg_pos = pages->sgl; in __i915_gem_object_set_pages()
32 obj->mm.pages = pages; in __i915_gem_object_set_pages()
156 struct sg_table *pages; in __i915_gem_object_unset_pages() local
158 pages = fetch_and_zero(&obj->mm.pages); in __i915_gem_object_unset_pages()
159 if (IS_ERR_OR_NULL(pages)) in __i915_gem_object_unset_pages()
160 return pages; in __i915_gem_object_unset_pages()
179 return pages; in __i915_gem_object_unset_pages()
185 struct sg_table *pages; in __i915_gem_object_put_pages() local
[all …]
/Linux-v5.4/net/rds/
Dinfo.c65 struct page **pages; member
122 iter->addr = kmap_atomic(*iter->pages); in rds_info_copy()
127 "bytes %lu\n", *iter->pages, iter->addr, in rds_info_copy()
140 iter->pages++; in rds_info_copy()
167 struct page **pages = NULL; in rds_info_getsockopt() local
191 pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); in rds_info_getsockopt()
192 if (!pages) { in rds_info_getsockopt()
196 ret = get_user_pages_fast(start, nr_pages, FOLL_WRITE, pages); in rds_info_getsockopt()
215 iter.pages = pages; in rds_info_getsockopt()
238 for (i = 0; pages && i < nr_pages; i++) in rds_info_getsockopt()
[all …]
/Linux-v5.4/drivers/gpu/drm/udl/
Dudl_gem.c112 if (!obj->pages) in udl_gem_fault()
115 page = obj->pages[page_offset]; in udl_gem_fault()
121 struct page **pages; in udl_gem_get_pages() local
123 if (obj->pages) in udl_gem_get_pages()
126 pages = drm_gem_get_pages(&obj->base); in udl_gem_get_pages()
127 if (IS_ERR(pages)) in udl_gem_get_pages()
128 return PTR_ERR(pages); in udl_gem_get_pages()
130 obj->pages = pages; in udl_gem_get_pages()
138 kvfree(obj->pages); in udl_gem_put_pages()
139 obj->pages = NULL; in udl_gem_put_pages()
[all …]
/Linux-v5.4/Documentation/admin-guide/mm/
Dhugetlbpage.rst30 persistent hugetlb pages in the kernel's huge page pool. It also displays
32 and surplus huge pages in the pool of huge pages of default size.
48 is the size of the pool of huge pages.
50 is the number of huge pages in the pool that are not yet
53 is short for "reserved," and is the number of huge pages for
55 but no allocation has yet been made. Reserved huge pages
57 huge page from the pool of huge pages at fault time.
59 is short for "surplus," and is the number of huge pages in
61 maximum number of surplus huge pages is controlled by
67 pages of all sizes.
[all …]
/Linux-v5.4/include/linux/
Dballoon_compaction.h57 struct list_head pages; /* Pages enqueued & handled to Host */ member
68 struct list_head *pages);
70 struct list_head *pages, size_t n_req_pages);
76 INIT_LIST_HEAD(&balloon->pages); in balloon_devinfo_init()
105 list_add(&page->lru, &balloon->pages); in balloon_page_insert()
149 list_add(&page->lru, &balloon->pages); in balloon_page_insert()
188 static inline void balloon_page_push(struct list_head *pages, struct page *page) in balloon_page_push() argument
190 list_add(&page->lru, pages); in balloon_page_push()
200 static inline struct page *balloon_page_pop(struct list_head *pages) in balloon_page_pop() argument
202 struct page *page = list_first_entry_or_null(pages, struct page, lru); in balloon_page_pop()
/Linux-v5.4/arch/m68k/mm/
Dsun3kmap.c50 unsigned long type, int pages) in do_pmeg_mapin() argument
56 while(pages) { in do_pmeg_mapin()
60 pages--; in do_pmeg_mapin()
69 int pages; in sun3_ioremap() local
88 pages = size / PAGE_SIZE; in sun3_ioremap()
92 while(pages) { in sun3_ioremap()
96 if(seg_pages > pages) in sun3_ioremap()
97 seg_pages = pages; in sun3_ioremap()
101 pages -= seg_pages; in sun3_ioremap()

12345678910>>...47