Searched refs:pmd_large (Results 1 – 25 of 31) sorted by relevance
12
695 static inline int pmd_large(pmd_t pmd) in pmd_large() function704 if (pmd_large(pmd)) in pmd_bad()748 if (pmd_large(pmd)) in pmd_pfn()762 if (pmd_large(pmd)) in pmd_dirty()770 if (pmd_large(pmd)) in pmd_young()814 return pmd_large(pmd) && !(pmd_val(pmd) & _SEGMENT_ENTRY_READ); in pmd_protnone()1300 if (pmd_large(pmd) && !(pmd_val(pmd) & _SEGMENT_ENTRY_DIRTY)) in pmd_mkwrite()1308 if (pmd_large(pmd)) { in pmd_mkclean()1317 if (pmd_large(pmd)) { in pmd_mkdirty()1382 if (pmd_large(pmd)) { in pmd_mkyoung()[all …]
130 VM_WARN_ON(!(pmd_large(pmd) || pmd_huge(pmd))); in pmd_page()
83 #define pmd_large(pmd) 0 macro
246 if (pmd_large(*pmd_k)) in vmalloc_fault()310 if (!low_pfn(pmd_pfn(*pmd)) || !pmd_present(*pmd) || pmd_large(*pmd)) in dump_pagetable()388 if (pmd_large(*pmd)) in vmalloc_fault()462 if (!pmd_present(*pmd) || pmd_large(*pmd)) in dump_pagetable()1143 if (pmd_large(*pmd)) in spurious_kernel_fault()
256 if (pmd_large(*pmd)) { in pti_user_pagetable_walk_pte()345 if (pmd_large(*pmd) || level == PTI_CLONE_PMD) { in pti_clone_pgtable()
152 if (pmd_large(*pmd)) in sme_populate_pgd_large()176 if (pmd_large(*pmd)) in sme_populate_pgd()
418 if (pmd_large(*start) || !pmd_present(*start)) { in walk_pmd_level()432 #define pud_large(a) pmd_large(__pmd(pud_val(a)))
527 if (!pmd_large(*pmd)) { in phys_pmd_init()1044 if (pmd_large(*pmd)) { in remove_pmd_table()1374 if (pmd_large(*pmd)) in kern_addr_valid()1506 } else if (pmd_large(*pmd)) { in vmemmap_populate_hugepages()
101 if (!pmd_large(*pmd)) in kasan_populate_pud()
772 if (pmd_large(*pmd)) { in pmd_clear_huge()
502 if (pmd_large(*pmd)) { in native_pagetable_init()
191 #define pmd_large(pmd) (pmd_val(pmd) & 2) macro
136 #define pmd_large(pmd) pmd_sect(pmd) macro
344 if (pmd_none(*pmd) || pmd_large(*pmd) || !pmd_present(*pmd)) in walk_pmd()349 if (SECTION_SIZE < PMD_SIZE && pmd_large(pmd[1])) { in walk_pmd()
193 if (pmd_large(*pm_dir)) { in vmem_remove_range()278 } else if (pmd_large(*pm_dir)) { in vmemmap_populate()
113 if (pmd_none(*pmd) || pmd_large(*pmd)) in mark_kernel_pmd()
233 return pmd_large(pmd); in pmd_huge()
594 if (pmd_large(*pmd) && !gmap->mm->context.allow_gmap_hpage_1m) in __gmap_link()606 if (pmd_large(*pmd)) { in __gmap_link()925 if (!pmd_large(*pmdp)) in gmap_pmd_op_walk()937 if (pmd_large(*pmdp)) in gmap_pmd_op_end()1048 if (!pmd_large(*pmdp)) { in gmap_protect_range()2466 if (pmd_large(*pmdp)) { in gmap_sync_dirty_log_pmd()
168 if (pmd_large(*pmd)) { in walk_pmd_level()
773 if (pmd_large(*pmdp)) { in set_guest_storage_key()876 if (pmd_large(*pmdp)) { in reset_guest_reference_bit()932 if (pmd_large(*pmdp)) { in get_guest_storage_key()
175 } else if (pmd_large(*pm_dir)) { in kasan_early_vmemmap_populate()
167 if (pmd_large(*pmdp)) { in walk_pmd_level()
239 if (pmd_large(*pmd)) { in relocate_restore_code()
75 WARN_ON(!(pmd_large(pmd))); in set_pmd_at()
1143 if (pmd_large(*pmd)) { in xen_cleanmfnmap_pmd()2040 if (pmd_large(pmd)) in xen_early_virt_to_phys()2226 if (pmd_present(pmd[pmdidx]) && !pmd_large(pmd[pmdidx])) { in xen_find_pt_base()