Home
last modified time | relevance | path

Searched refs:folios (Results 1 – 19 of 19) sorted by relevance

/Linux-v6.1/include/linux/
Dpagevec.h86 struct folio *folios[PAGEVEC_SIZE]; member
92 offsetof(struct folio_batch, folios));
129 fbatch->folios[fbatch->nr++] = folio; in folio_batch_add()
Dmm.h1196 static inline void folios_put(struct folio **folios, unsigned int nr) in folios_put() argument
1198 release_pages((struct page **)folios, nr); in folios_put()
/Linux-v6.1/fs/ramfs/
Dfile-nommu.c235 ret = (unsigned long) folio_address(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area()
236 pfn = folio_pfn(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area()
240 if (pfn + nr_pages != folio_pfn(fbatch.folios[loop])) { in ramfs_nommu_get_unmapped_area()
244 nr_pages += folio_nr_pages(fbatch.folios[loop]); in ramfs_nommu_get_unmapped_area()
/Linux-v6.1/mm/
Dtruncate.c71 if (xa_is_value(fbatch->folios[j])) in truncate_folio_batch_exceptionals()
84 struct folio *folio = fbatch->folios[i]; in truncate_folio_batch_exceptionals()
88 fbatch->folios[j++] = folio; in truncate_folio_batch_exceptionals()
369 truncate_cleanup_folio(fbatch.folios[i]); in truncate_inode_pages_range()
372 folio_unlock(fbatch.folios[i]); in truncate_inode_pages_range()
415 struct folio *folio = fbatch.folios[i]; in truncate_inode_pages_range()
515 struct folio *folio = fbatch.folios[i]; in invalidate_mapping_pagevec()
646 struct folio *folio = fbatch.folios[i]; in invalidate_inode_pages2_range()
Dswap.c239 struct folio *folio = fbatch->folios[i]; in folio_batch_move_lru()
253 folios_put(fbatch->folios, folio_batch_count(fbatch)); in folio_batch_move_lru()
420 struct folio *batch_folio = fbatch->folios[i]; in __lru_cache_activate_folio()
1094 struct folio *folio = fbatch->folios[i]; in folio_batch_remove_exceptionals()
1096 fbatch->folios[j++] = folio; in folio_batch_remove_exceptionals()
Dfilemap.c280 XA_STATE(xas, &mapping->i_pages, fbatch->folios[0]->index); in page_cache_delete_batch()
300 if (folio != fbatch->folios[i]) { in page_cache_delete_batch()
302 fbatch->folios[i]->index, folio); in page_cache_delete_batch()
329 struct folio *folio = fbatch->folios[i]; in delete_from_page_cache_batch()
341 filemap_free_folio(mapping, fbatch->folios[i]); in delete_from_page_cache_batch()
2253 folio = fbatch->folios[nr - 1]; in filemap_get_folios_contig()
2595 folio = fbatch->folios[folio_batch_count(fbatch) - 1]; in filemap_get_pages()
2703 fbatch.folios[0])) in filemap_read()
2704 folio_mark_accessed(fbatch.folios[0]); in filemap_read()
2707 struct folio *folio = fbatch.folios[i]; in filemap_read()
[all …]
Dshmem.c928 folio = fbatch.folios[i]; in shmem_undo_range()
1001 folio = fbatch.folios[i]; in shmem_undo_range()
1232 struct folio *folio = fbatch->folios[i]; in shmem_unuse_swap_entries()
Dvmscan.c7737 struct folio *folio = fbatch->folios[i]; in check_move_unevictable_folios()
/Linux-v6.1/Documentation/filesystems/
Dnetfs_library.rst104 * Handle folios that span multiple pages.
109 don't match folio sizes or folio alignments and that may cross folios.
363 it transferred. The filesystem also should not deal with setting folios
367 Note that the helpers have the folios locked, but not pinned. It is
391 [Optional] This is called after the folios in the request have all been
438 * Once the data is read, the folios that have been fully read/cleared:
446 * Any folios that need writing to the cache will then have DIO writes issued.
450 * Writes to the cache will proceed asynchronously and the folios will have the
Dvfs.rst628 on dirty pages, and ->release_folio on clean folios with the private
785 If the mapping does not support large folios, the folio will
795 In normal operation, folios are read through the ->readahead()
902 release_folio is called on folios with private data to tell the
912 some or all folios in an address_space. This can happen
917 and needs to be certain that all folios are invalidated, then
957 some filesystems have more complex state (unstable folios in NFS
Dlocking.rst298 ->readahead() unlocks the folios that I/O is attempted on like ->read_folio().
/Linux-v6.1/fs/nilfs2/
Dpage.c308 struct folio *folio = fbatch.folios[i], *dfolio; in nilfs_copy_back_pages()
504 folio = fbatch.folios[i]; in nilfs_find_uncommitted_extent()
/Linux-v6.1/fs/btrfs/tests/
Dextent-io-tests.c37 struct folio *folio = fbatch.folios[i]; in process_page_range()
/Linux-v6.1/fs/btrfs/
Dcompression.c240 struct folio *folio = fbatch.folios[i]; in end_compressed_writeback()
Dextent_io.c305 struct folio *folio = fbatch.folios[i]; in __process_pages_contig()
/Linux-v6.1/fs/hugetlbfs/
Dinode.c632 struct folio *folio = fbatch.folios[i]; in remove_inode_hugepages()
/Linux-v6.1/fs/f2fs/
Dcompress.c1922 struct folio *folio = fbatch.folios[i]; in f2fs_invalidate_compress_pages()
/Linux-v6.1/fs/
Dbuffer.c1604 struct folio *folio = fbatch.folios[i]; in clean_bdev_aliases()
/Linux-v6.1/fs/ext4/
Dinode.c1599 struct folio *folio = fbatch.folios[i]; in mpage_release_unused_pages()
2354 struct page *page = &fbatch.folios[i]->page; in mpage_map_and_submit_buffers()