/Linux-v5.10/include/linux/ |
D | pagevec.h | 26 void __pagevec_release(struct pagevec *pvec); 27 void __pagevec_lru_add(struct pagevec *pvec); 28 unsigned pagevec_lookup_entries(struct pagevec *pvec, 32 void pagevec_remove_exceptionals(struct pagevec *pvec); 33 unsigned pagevec_lookup_range(struct pagevec *pvec, 36 static inline unsigned pagevec_lookup(struct pagevec *pvec, in pagevec_lookup() argument 40 return pagevec_lookup_range(pvec, mapping, start, (pgoff_t)-1); in pagevec_lookup() 43 unsigned pagevec_lookup_range_tag(struct pagevec *pvec, 46 unsigned pagevec_lookup_range_nr_tag(struct pagevec *pvec, 49 static inline unsigned pagevec_lookup_tag(struct pagevec *pvec, in pagevec_lookup_tag() argument [all …]
|
/Linux-v5.10/mm/ |
D | swap.c | 51 struct pagevec pvec; member 206 static void pagevec_lru_move_fn(struct pagevec *pvec, in pagevec_lru_move_fn() argument 215 for (i = 0; i < pagevec_count(pvec); i++) { in pagevec_lru_move_fn() 216 struct page *page = pvec->pages[i]; in pagevec_lru_move_fn() 231 release_pages(pvec->pages, pvec->nr); in pagevec_lru_move_fn() 232 pagevec_reinit(pvec); in pagevec_lru_move_fn() 252 static void pagevec_move_tail(struct pagevec *pvec) in pagevec_move_tail() argument 256 pagevec_lru_move_fn(pvec, pagevec_move_tail_fn, &pgmoved); in pagevec_move_tail() 269 struct pagevec *pvec; in rotate_reclaimable_page() local 274 pvec = this_cpu_ptr(&lru_rotate.pvec); in rotate_reclaimable_page() [all …]
|
D | truncate.c | 60 struct pagevec *pvec, pgoff_t *indices, in truncate_exceptional_pvec_entries() argument 70 for (j = 0; j < pagevec_count(pvec); j++) in truncate_exceptional_pvec_entries() 71 if (xa_is_value(pvec->pages[j])) in truncate_exceptional_pvec_entries() 74 if (j == pagevec_count(pvec)) in truncate_exceptional_pvec_entries() 82 for (i = j; i < pagevec_count(pvec); i++) { in truncate_exceptional_pvec_entries() 83 struct page *page = pvec->pages[i]; in truncate_exceptional_pvec_entries() 87 pvec->pages[j++] = page; in truncate_exceptional_pvec_entries() 104 pvec->nr = j; in truncate_exceptional_pvec_entries() 298 struct pagevec pvec; in truncate_inode_pages_range() local 327 pagevec_init(&pvec); in truncate_inode_pages_range() [all …]
|
D | mlock.c | 251 static bool __putback_lru_fast_prepare(struct page *page, struct pagevec *pvec, in __putback_lru_fast_prepare() argument 258 pagevec_add(pvec, page); in __putback_lru_fast_prepare() 274 static void __putback_lru_fast(struct pagevec *pvec, int pgrescued) in __putback_lru_fast() argument 276 count_vm_events(UNEVICTABLE_PGMUNLOCKED, pagevec_count(pvec)); in __putback_lru_fast() 281 __pagevec_lru_add(pvec); in __putback_lru_fast() 295 static void __munlock_pagevec(struct pagevec *pvec, struct zone *zone) in __munlock_pagevec() argument 298 int nr = pagevec_count(pvec); in __munlock_pagevec() 308 struct page *page = pvec->pages[i]; in __munlock_pagevec() 329 pagevec_add(&pvec_putback, pvec->pages[i]); in __munlock_pagevec() 330 pvec->pages[i] = NULL; in __munlock_pagevec() [all …]
|
D | shmem.c | 844 struct pagevec pvec; in shmem_unlock_mapping() local 848 pagevec_init(&pvec); in shmem_unlock_mapping() 857 pvec.nr = find_get_entries(mapping, index, in shmem_unlock_mapping() 858 PAGEVEC_SIZE, pvec.pages, indices); in shmem_unlock_mapping() 859 if (!pvec.nr) in shmem_unlock_mapping() 861 index = indices[pvec.nr - 1] + 1; in shmem_unlock_mapping() 862 pagevec_remove_exceptionals(&pvec); in shmem_unlock_mapping() 863 check_move_unevictable_pages(&pvec); in shmem_unlock_mapping() 864 pagevec_release(&pvec); in shmem_unlock_mapping() 908 struct pagevec pvec; in shmem_undo_range() local [all …]
|
D | filemap.c | 296 struct pagevec *pvec) in page_cache_delete_batch() argument 298 XA_STATE(xas, &mapping->i_pages, pvec->pages[0]->index); in page_cache_delete_batch() 305 if (i >= pagevec_count(pvec)) in page_cache_delete_batch() 318 if (page != pvec->pages[i]) { in page_cache_delete_batch() 319 VM_BUG_ON_PAGE(page->index > pvec->pages[i]->index, in page_cache_delete_batch() 344 struct pagevec *pvec) in delete_from_page_cache_batch() argument 349 if (!pagevec_count(pvec)) in delete_from_page_cache_batch() 353 for (i = 0; i < pagevec_count(pvec); i++) { in delete_from_page_cache_batch() 354 trace_mm_filemap_delete_from_page_cache(pvec->pages[i]); in delete_from_page_cache_batch() 356 unaccount_page_cache_page(mapping, pvec->pages[i]); in delete_from_page_cache_batch() [all …]
|
D | page-writeback.c | 2175 struct pagevec pvec; in write_cache_pages() local 2183 pagevec_init(&pvec); in write_cache_pages() 2203 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in write_cache_pages() 2209 struct page *page = pvec.pages[i]; in write_cache_pages() 2285 pagevec_release(&pvec); in write_cache_pages()
|
/Linux-v5.10/fs/nilfs2/ |
D | page.c | 243 struct pagevec pvec; in nilfs_copy_dirty_pages() local 248 pagevec_init(&pvec); in nilfs_copy_dirty_pages() 250 if (!pagevec_lookup_tag(&pvec, smap, &index, PAGECACHE_TAG_DIRTY)) in nilfs_copy_dirty_pages() 253 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_copy_dirty_pages() 254 struct page *page = pvec.pages[i], *dpage; in nilfs_copy_dirty_pages() 278 pagevec_release(&pvec); in nilfs_copy_dirty_pages() 297 struct pagevec pvec; in nilfs_copy_back_pages() local 301 pagevec_init(&pvec); in nilfs_copy_back_pages() 303 n = pagevec_lookup(&pvec, smap, &index); in nilfs_copy_back_pages() 307 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_copy_back_pages() [all …]
|
D | segment.c | 688 struct pagevec pvec; in nilfs_lookup_dirty_data_buffers() local 702 pagevec_init(&pvec); in nilfs_lookup_dirty_data_buffers() 705 !pagevec_lookup_range_tag(&pvec, mapping, &index, last, in nilfs_lookup_dirty_data_buffers() 709 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_lookup_dirty_data_buffers() 711 struct page *page = pvec.pages[i]; in nilfs_lookup_dirty_data_buffers() 726 pagevec_release(&pvec); in nilfs_lookup_dirty_data_buffers() 732 pagevec_release(&pvec); in nilfs_lookup_dirty_data_buffers() 742 struct pagevec pvec; in nilfs_lookup_dirty_node_buffers() local 747 pagevec_init(&pvec); in nilfs_lookup_dirty_node_buffers() 749 while (pagevec_lookup_tag(&pvec, mapping, &index, in nilfs_lookup_dirty_node_buffers() [all …]
|
D | btree.c | 2140 struct pagevec pvec; in nilfs_btree_lookup_dirty_buffers() local 2150 pagevec_init(&pvec); in nilfs_btree_lookup_dirty_buffers() 2152 while (pagevec_lookup_tag(&pvec, btcache, &index, in nilfs_btree_lookup_dirty_buffers() 2154 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_btree_lookup_dirty_buffers() 2155 bh = head = page_buffers(pvec.pages[i]); in nilfs_btree_lookup_dirty_buffers() 2162 pagevec_release(&pvec); in nilfs_btree_lookup_dirty_buffers()
|
/Linux-v5.10/drivers/gpu/drm/i915/gem/ |
D | i915_gem_shmem.c | 21 static void check_release_pagevec(struct pagevec *pvec) in check_release_pagevec() argument 23 check_move_unevictable_pages(pvec); in check_release_pagevec() 24 __pagevec_release(pvec); in check_release_pagevec() 194 struct pagevec pvec; in shmem_get_pages() local 196 pagevec_init(&pvec); in shmem_get_pages() 198 if (!pagevec_add(&pvec, page)) in shmem_get_pages() 199 check_release_pagevec(&pvec); in shmem_get_pages() 201 if (pagevec_count(&pvec)) in shmem_get_pages() 202 check_release_pagevec(&pvec); in shmem_get_pages() 303 struct pagevec pvec; in shmem_put_pages() local [all …]
|
D | i915_gem_userptr.c | 401 struct page **pvec, unsigned long num_pages) in __i915_gem_userptr_alloc_pages() argument 414 sg = __sg_alloc_table_from_pages(st, pvec, num_pages, 0, in __i915_gem_userptr_alloc_pages() 449 struct page **pvec; in __i915_gem_userptr_get_pages_worker() local 455 pvec = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in __i915_gem_userptr_get_pages_worker() 456 if (pvec != NULL) { in __i915_gem_userptr_get_pages_worker() 476 pvec + pinned, NULL, &locked); in __i915_gem_userptr_get_pages_worker() 493 pages = __i915_gem_userptr_alloc_pages(obj, pvec, in __i915_gem_userptr_get_pages_worker() 507 unpin_user_pages(pvec, pinned); in __i915_gem_userptr_get_pages_worker() 508 kvfree(pvec); in __i915_gem_userptr_get_pages_worker() 560 struct page **pvec; in i915_gem_userptr_get_pages() local [all …]
|
/Linux-v5.10/fs/iomap/ |
D | seek.c | 80 struct pagevec pvec; in page_cache_seek_hole_data() local 85 pagevec_init(&pvec); in page_cache_seek_hole_data() 90 nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping, &index, in page_cache_seek_hole_data() 96 struct page *page = pvec.pages[i]; in page_cache_seek_hole_data() 102 pagevec_release(&pvec); in page_cache_seek_hole_data() 115 pagevec_release(&pvec); in page_cache_seek_hole_data()
|
/Linux-v5.10/drivers/gpu/drm/ |
D | drm_gem.c | 524 static void drm_gem_check_release_pagevec(struct pagevec *pvec) in drm_gem_check_release_pagevec() argument 526 check_move_unevictable_pages(pvec); in drm_gem_check_release_pagevec() 527 __pagevec_release(pvec); in drm_gem_check_release_pagevec() 560 struct pagevec pvec; in drm_gem_get_pages() local 603 pagevec_init(&pvec); in drm_gem_get_pages() 605 if (!pagevec_add(&pvec, pages[i])) in drm_gem_get_pages() 606 drm_gem_check_release_pagevec(&pvec); in drm_gem_get_pages() 608 if (pagevec_count(&pvec)) in drm_gem_get_pages() 609 drm_gem_check_release_pagevec(&pvec); in drm_gem_get_pages() 628 struct pagevec pvec; in drm_gem_put_pages() local [all …]
|
/Linux-v5.10/fs/gfs2/ |
D | aops.c | 235 struct pagevec *pvec, in gfs2_write_jdata_pagevec() argument 250 struct page *page = pvec->pages[i]; in gfs2_write_jdata_pagevec() 333 struct pagevec pvec; in gfs2_write_cache_jdata() local 343 pagevec_init(&pvec); in gfs2_write_cache_jdata() 369 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in gfs2_write_cache_jdata() 374 ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, &done_index); in gfs2_write_cache_jdata() 379 pagevec_release(&pvec); in gfs2_write_cache_jdata()
|
/Linux-v5.10/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem.c | 659 struct page **pvec = NULL; in etnaviv_gem_userptr_get_pages() local 668 pvec = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); in etnaviv_gem_userptr_get_pages() 669 if (!pvec) in etnaviv_gem_userptr_get_pages() 675 struct page **pages = pvec + pinned; in etnaviv_gem_userptr_get_pages() 680 unpin_user_pages(pvec, pinned); in etnaviv_gem_userptr_get_pages() 681 kvfree(pvec); in etnaviv_gem_userptr_get_pages() 689 etnaviv_obj->pages = pvec; in etnaviv_gem_userptr_get_pages()
|
/Linux-v5.10/fs/ceph/ |
D | addr.c | 831 struct pagevec pvec; in ceph_writepages_start() local 855 pagevec_init(&pvec); in ceph_writepages_start() 913 pvec_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, in ceph_writepages_start() 919 page = pvec.pages[i]; in ceph_writepages_start() 1039 pvec.pages[i] = NULL; in ceph_writepages_start() 1051 if (!pvec.pages[j]) in ceph_writepages_start() 1054 pvec.pages[n] = pvec.pages[j]; in ceph_writepages_start() 1057 pvec.nr = n; in ceph_writepages_start() 1062 pagevec_release(&pvec); in ceph_writepages_start() 1187 dout("pagevec_release on %d pages (%p)\n", (int)pvec.nr, in ceph_writepages_start() [all …]
|
/Linux-v5.10/fs/hugetlbfs/ |
D | inode.c | 111 static void huge_pagevec_release(struct pagevec *pvec) in huge_pagevec_release() argument 115 for (i = 0; i < pagevec_count(pvec); ++i) in huge_pagevec_release() 116 put_page(pvec->pages[i]); in huge_pagevec_release() 118 pagevec_reinit(pvec); in huge_pagevec_release() 467 struct pagevec pvec; in remove_inode_hugepages() local 474 pagevec_init(&pvec); in remove_inode_hugepages() 480 if (!pagevec_lookup_range(&pvec, mapping, &next, end - 1)) in remove_inode_hugepages() 483 for (i = 0; i < pagevec_count(&pvec); ++i) { in remove_inode_hugepages() 484 struct page *page = pvec.pages[i]; in remove_inode_hugepages() 543 huge_pagevec_release(&pvec); in remove_inode_hugepages()
|
/Linux-v5.10/fs/f2fs/ |
D | node.c | 1449 struct pagevec pvec; in last_fsync_dnode() local 1453 pagevec_init(&pvec); in last_fsync_dnode() 1456 while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index, in last_fsync_dnode() 1461 struct page *page = pvec.pages[i]; in last_fsync_dnode() 1465 pagevec_release(&pvec); in last_fsync_dnode() 1496 pagevec_release(&pvec); in last_fsync_dnode() 1665 struct pagevec pvec; in f2fs_fsync_node_pages() local 1679 pagevec_init(&pvec); in f2fs_fsync_node_pages() 1682 while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index, in f2fs_fsync_node_pages() 1687 struct page *page = pvec.pages[i]; in f2fs_fsync_node_pages() [all …]
|
D | checkpoint.c | 372 struct pagevec pvec; in f2fs_sync_meta_pages() local 380 pagevec_init(&pvec); in f2fs_sync_meta_pages() 384 while ((nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, in f2fs_sync_meta_pages() 389 struct page *page = pvec.pages[i]; in f2fs_sync_meta_pages() 394 pagevec_release(&pvec); in f2fs_sync_meta_pages() 424 pagevec_release(&pvec); in f2fs_sync_meta_pages()
|
/Linux-v5.10/arch/x86/kvm/mmu/ |
D | mmu.c | 1735 static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp, in mmu_pages_add() argument 1741 for (i=0; i < pvec->nr; i++) in mmu_pages_add() 1742 if (pvec->page[i].sp == sp) in mmu_pages_add() 1745 pvec->page[pvec->nr].sp = sp; in mmu_pages_add() 1746 pvec->page[pvec->nr].idx = idx; in mmu_pages_add() 1747 pvec->nr++; in mmu_pages_add() 1748 return (pvec->nr == KVM_PAGE_ARRAY_NR); in mmu_pages_add() 1759 struct kvm_mmu_pages *pvec) in __mmu_unsync_walk() argument 1775 if (mmu_pages_add(pvec, child, i)) in __mmu_unsync_walk() 1778 ret = __mmu_unsync_walk(child, pvec); in __mmu_unsync_walk() [all …]
|
/Linux-v5.10/fs/fscache/ |
D | page.c | 1219 struct pagevec pvec; in __fscache_uncache_all_inode_pages() local 1230 pagevec_init(&pvec); in __fscache_uncache_all_inode_pages() 1233 if (!pagevec_lookup(&pvec, mapping, &next)) in __fscache_uncache_all_inode_pages() 1235 for (i = 0; i < pagevec_count(&pvec); i++) { in __fscache_uncache_all_inode_pages() 1236 struct page *page = pvec.pages[i]; in __fscache_uncache_all_inode_pages() 1242 pagevec_release(&pvec); in __fscache_uncache_all_inode_pages()
|
/Linux-v5.10/fs/ext4/ |
D | inode.c | 1552 struct pagevec pvec; in mpage_release_unused_pages() local 1570 pagevec_init(&pvec); in mpage_release_unused_pages() 1572 nr_pages = pagevec_lookup_range(&pvec, mapping, &index, end); in mpage_release_unused_pages() 1576 struct page *page = pvec.pages[i]; in mpage_release_unused_pages() 1588 pagevec_release(&pvec); in mpage_release_unused_pages() 2306 struct pagevec pvec; in mpage_map_and_submit_buffers() local 2321 pagevec_init(&pvec); in mpage_map_and_submit_buffers() 2323 nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping, in mpage_map_and_submit_buffers() 2328 struct page *page = pvec.pages[i]; in mpage_map_and_submit_buffers() 2344 pagevec_release(&pvec); in mpage_map_and_submit_buffers() [all …]
|
/Linux-v5.10/drivers/mtd/lpddr/ |
D | lpddr_cmds.c | 404 unsigned long adr, const struct kvec **pvec, in do_write_buffer() argument 436 vec = *pvec; in do_write_buffer() 466 *pvec = vec; in do_write_buffer()
|
/Linux-v5.10/fs/ |
D | buffer.c | 1610 struct pagevec pvec; in clean_bdev_aliases() local 1618 pagevec_init(&pvec); in clean_bdev_aliases() 1619 while (pagevec_lookup_range(&pvec, bd_mapping, &index, end)) { in clean_bdev_aliases() 1620 count = pagevec_count(&pvec); in clean_bdev_aliases() 1622 struct page *page = pvec.pages[i]; in clean_bdev_aliases() 1651 pagevec_release(&pvec); in clean_bdev_aliases()
|