Lines Matching refs:PMD_SIZE
88 ALIGN(unused_pmd_start, PMD_SIZE) - unused_pmd_start); in vmemmap_flush_unused_pmd()
111 if (likely(IS_ALIGNED(unused_pmd_start, PMD_SIZE))) in vmemmap_use_sub_pmd()
121 void *page = __va(ALIGN_DOWN(start, PMD_SIZE)); in vmemmap_use_new_sub_pmd()
129 if (!IS_ALIGNED(start, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
136 if (!IS_ALIGNED(end, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
143 void *page = __va(ALIGN_DOWN(start, PMD_SIZE)); in vmemmap_unuse_sub_pmd()
147 return !memchr_inv(page, PAGE_UNUSED, PMD_SIZE); in vmemmap_unuse_sub_pmd()
227 if (IS_ALIGNED(addr, PMD_SIZE) && in modify_pmd_table()
228 IS_ALIGNED(next, PMD_SIZE)) { in modify_pmd_table()
230 vmem_free_pages(pmd_deref(*pmd), get_order(PMD_SIZE)); in modify_pmd_table()
234 vmem_free_pages(pmd_deref(*pmd), get_order(PMD_SIZE)); in modify_pmd_table()
240 if (IS_ALIGNED(addr, PMD_SIZE) && in modify_pmd_table()
241 IS_ALIGNED(next, PMD_SIZE) && in modify_pmd_table()
257 new_page = vmemmap_alloc_block(PMD_SIZE, NUMA_NO_NODE); in modify_pmd_table()
260 if (!IS_ALIGNED(addr, PMD_SIZE) || in modify_pmd_table()
261 !IS_ALIGNED(next, PMD_SIZE)) { in modify_pmd_table()