Searched refs:split_huge_page (Results 1 – 10 of 10) sorted by relevance
75 calling split_huge_page(page). This is what the Linux VM does before76 it tries to swapout the hugepage for example. split_huge_page() can fail97 split_huge_page() or split_huge_pmd() has a cost.151 split_huge_page internally has to distribute the refcounts in the head155 additional pins (i.e. from get_user_pages). split_huge_page() fails any157 the sum of mapcount of all sub-pages plus one (split_huge_page caller must160 split_huge_page uses migration entries to stabilize page->_refcount and
457 We cannot just split the page on partial mlock() as split_huge_page() can
192 static inline int split_huge_page(struct page *page) in split_huge_page() function400 static inline int split_huge_page(struct page *page) in split_huge_page() function
355 err = split_huge_page(page); in madvise_cold_or_pageout_pte_range()423 if (split_huge_page(page)) { in madvise_cold_or_pageout_pte_range()633 if (split_huge_page(page)) { in madvise_free_pte_range()
1604 split_huge_page(page); in madvise_free_huge_pmd()2850 if (!split_huge_page(page)) in deferred_split_scan()2907 if (!split_huge_page(page)) in split_huge_pages_set()
477 ret = split_huge_page(page); in follow_page_pte()645 ret = split_huge_page(page); in follow_pmd_mask()
1225 if (split_huge_page(page)) in try_to_merge_one_page()2191 split_huge_page(page); in cmp_and_merge_page()
1134 if (!PageAnon(page) || unlikely(split_huge_page(page))) { in try_to_split_thp_page()
603 ret = split_huge_page(page); in shmem_unused_huge_shrink()892 return split_huge_page(page) >= 0; in shmem_punch_compound()
2343 ret = split_huge_page(page); in migrate_vma_collect_pmd()