Searched refs:pagevec_count (Results 1 – 16 of 16) sorted by relevance
215 for (i = 0; i < pagevec_count(pvec); i++) { in pagevec_lru_move_fn()342 if (pagevec_count(pvec)) in activate_page_drain()348 return pagevec_count(&per_cpu(lru_pvecs.activate_page, cpu)) != 0; in need_activate_page_drain()400 for (i = pagevec_count(pvec) - 1; i >= 0; i--) { in __lru_cache_activate_page()629 if (pagevec_count(pvec)) in lru_add_drain_cpu()634 if (data_race(pagevec_count(pvec))) { in lru_add_drain_cpu()644 if (pagevec_count(pvec)) in lru_add_drain_cpu()648 if (pagevec_count(pvec)) in lru_add_drain_cpu()652 if (pagevec_count(pvec)) in lru_add_drain_cpu()837 if (pagevec_count(&per_cpu(lru_pvecs.lru_add, cpu)) || in lru_add_drain_all()[all …]
70 for (j = 0; j < pagevec_count(pvec); j++) in truncate_exceptional_pvec_entries()74 if (j == pagevec_count(pvec)) in truncate_exceptional_pvec_entries()82 for (i = j; i < pagevec_count(pvec); i++) { in truncate_exceptional_pvec_entries()340 for (i = 0; i < pagevec_count(&pvec); i++) { in truncate_inode_pages_range()364 for (i = 0; i < pagevec_count(&locked_pvec); i++) in truncate_inode_pages_range()367 for (i = 0; i < pagevec_count(&locked_pvec); i++) in truncate_inode_pages_range()432 for (i = 0; i < pagevec_count(&pvec); i++) { in truncate_inode_pages_range()545 for (i = 0; i < pagevec_count(&pvec); i++) { in __invalidate_mapping_pages()725 for (i = 0; i < pagevec_count(&pvec); i++) { in invalidate_inode_pages2_range()
276 count_vm_events(UNEVICTABLE_PGMUNLOCKED, pagevec_count(pvec)); in __putback_lru_fast()298 int nr = pagevec_count(pvec); in __munlock_pagevec()362 if (pagevec_count(&pvec_putback)) in __munlock_pagevec()
305 if (i >= pagevec_count(pvec)) in page_cache_delete_batch()349 if (!pagevec_count(pvec)) in delete_from_page_cache_batch()353 for (i = 0; i < pagevec_count(pvec); i++) { in delete_from_page_cache_batch()361 for (i = 0; i < pagevec_count(pvec); i++) in delete_from_page_cache_batch()
925 for (i = 0; i < pagevec_count(&pvec); i++) { in shmem_undo_range()1002 for (i = 0; i < pagevec_count(&pvec); i++) { in shmem_undo_range()
66 static inline unsigned pagevec_count(struct pagevec *pvec) in pagevec_count() function87 if (pagevec_count(pvec)) in pagevec_release()
253 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_copy_dirty_pages()307 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_copy_back_pages()368 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_clear_dirty_pages()547 } while (++i < pagevec_count(&pvec)); in nilfs_find_uncommitted_extent()
709 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_lookup_dirty_data_buffers()751 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_lookup_dirty_node_buffers()
2154 for (i = 0; i < pagevec_count(&pvec); i++) { in nilfs_btree_lookup_dirty_buffers()
201 if (pagevec_count(&pvec)) in shmem_get_pages()326 if (pagevec_count(&pvec)) in shmem_put_pages()
765 if (pagevec_count(&pagevec) > 0) in cachefiles_read_or_alloc_pages()863 if (pagevec_count(&pagevec) > 0) in cachefiles_allocate_pages()
608 if (pagevec_count(&pvec)) in drm_gem_get_pages()656 if (pagevec_count(&pvec)) in drm_gem_put_pages()
115 for (i = 0; i < pagevec_count(pvec); ++i) in huge_pagevec_release()483 for (i = 0; i < pagevec_count(&pvec); ++i) { in remove_inode_hugepages()
1235 for (i = 0; i < pagevec_count(&pvec); i++) { in __fscache_uncache_all_inode_pages()
219 if (!p && pagevec_count(pv)) in pool_alloc()
1620 count = pagevec_count(&pvec); in clean_bdev_aliases()