Home
last modified time | relevance | path

Searched refs:PUD_SIZE (Results 1 – 25 of 50) sorted by relevance

12

/Linux-v4.19/include/asm-generic/
Dpgtable-nop4d-hack.h19 #define PUD_SIZE (1UL << PUD_SHIFT) macro
20 #define PUD_MASK (~(PUD_SIZE-1))
Dpgtable-nopud.h23 #define PUD_SIZE (1UL << PUD_SHIFT) macro
24 #define PUD_MASK (~(PUD_SIZE-1))
D4level-fixup.h9 #define PUD_SIZE PGDIR_SIZE macro
/Linux-v4.19/arch/powerpc/include/asm/nohash/64/
Dpgtable-4k.h36 #define PUD_SIZE (1UL << PUD_SHIFT) macro
37 #define PUD_MASK (~(PUD_SIZE-1))
/Linux-v4.19/arch/powerpc/mm/
Dpgtable-radix.c95 if (map_page_size == PUD_SIZE) { in early_map_kernel_page()
155 if (map_page_size == PUD_SIZE) { in __map_kernel_page()
279 max_mapping_size = PUD_SIZE; in create_physical_mapping()
282 if (IS_ALIGNED(addr, PUD_SIZE) && gap >= PUD_SIZE && in create_physical_mapping()
284 PUD_SIZE <= max_mapping_size) { in create_physical_mapping()
285 mapping_size = PUD_SIZE; in create_physical_mapping()
296 if (split_text_mapping && (mapping_size == PUD_SIZE) && in create_physical_mapping()
842 split_kernel_mapping(addr, end, PUD_SIZE, (pte_t *)pud); in remove_pud_table()
/Linux-v4.19/arch/x86/include/asm/
Dpgtable_64_types.h100 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro
101 #define PUD_MASK (~(PUD_SIZE - 1))
/Linux-v4.19/arch/arm64/mm/
Dhugetlbpage.c79 case PUD_SIZE: in num_contig_ptes()
217 if (sz == PUD_SIZE) { in huge_pte_alloc()
259 if (sz != PUD_SIZE && pud_none(pud)) in huge_pte_offset()
292 } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { in arch_make_huge_pte()
438 case PUD_SIZE: in setup_hugepagesz()
/Linux-v4.19/arch/x86/boot/compressed/
Dkaslr.c225 if (memparse(p, &p) != PUD_SIZE) { in parse_gb_huge_pages()
514 addr = ALIGN(region->start, PUD_SIZE); in process_gb_huge_pages()
520 while (size > PUD_SIZE && max_gb_huge_pages) { in process_gb_huge_pages()
521 size -= PUD_SIZE; in process_gb_huge_pages()
544 size = region->size - (addr - region->start) - i * PUD_SIZE; in process_gb_huge_pages()
546 tmp.start = addr + i * PUD_SIZE; in process_gb_huge_pages()
/Linux-v4.19/arch/x86/mm/
Dkasan_init_64.c84 ((end - addr) == PUD_SIZE) && in kasan_populate_pud()
85 IS_ALIGNED(addr, PUD_SIZE)) { in kasan_populate_pud()
86 p = early_alloc(PUD_SIZE, nid, false); in kasan_populate_pud()
90 memblock_free(__pa(p), PUD_SIZE); in kasan_populate_pud()
Dkaslr.c142 vaddr = round_up(vaddr + 1, PUD_SIZE); in kernel_randomize_memory()
166 paddr_next = (paddr & PUD_MASK) + PUD_SIZE; in init_trampoline_pud()
Dinit.c300 unsigned long start = round_down(mr[i].start, PUD_SIZE); in adjust_range_page_size_mask()
301 unsigned long end = round_up(mr[i].end, PUD_SIZE); in adjust_range_page_size_mask()
372 end_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
385 start_pfn = round_up(pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
386 end_pfn = round_down(limit_pfn, PFN_DOWN(PUD_SIZE)); in split_mem_range()
Dpageattr.c892 if (start & (PUD_SIZE - 1)) { in unmap_pud_range()
893 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in unmap_pud_range()
905 while (end - start >= PUD_SIZE) { in unmap_pud_range()
910 unmap_pmd_range(pud, start, start + PUD_SIZE); in unmap_pud_range()
912 start += PUD_SIZE; in unmap_pud_range()
1055 if (start & (PUD_SIZE - 1)) { in populate_pud()
1057 unsigned long next_page = (start + PUD_SIZE) & PUD_MASK; in populate_pud()
1090 while (boot_cpu_has(X86_FEATURE_GBPAGES) && end - start >= PUD_SIZE) { in populate_pud()
1094 start += PUD_SIZE; in populate_pud()
1095 cpa->pfn += PUD_SIZE >> PAGE_SHIFT; in populate_pud()
[all …]
Dinit_64.c569 paddr_next = (paddr & PUD_MASK) + PUD_SIZE; in phys_pud_init()
1041 if (IS_ALIGNED(addr, PUD_SIZE) && in remove_pud_table()
1042 IS_ALIGNED(next, PUD_SIZE)) { in remove_pud_table()
1045 get_order(PUD_SIZE)); in remove_pud_table()
1057 PUD_SIZE)) { in remove_pud_table()
1059 get_order(PUD_SIZE)); in remove_pud_table()
Dident_map.c30 next = (addr & PUD_MASK) + PUD_SIZE; in ident_pud_init()
Dmem_encrypt_identity.c253 entries += (DIV_ROUND_UP(len, PUD_SIZE) + 1) * sizeof(pmd_t) * PTRS_PER_PMD; in sme_pgtable_calc()
264 tables += DIV_ROUND_UP(entries, PUD_SIZE) * sizeof(pmd_t) * PTRS_PER_PMD; in sme_pgtable_calc()
Dhugetlbpage.c194 } else if (ps == PUD_SIZE && boot_cpu_has(X86_FEATURE_GBPAGES)) { in setup_hugepagesz()
/Linux-v4.19/arch/s390/mm/
Dvmem.c109 !(address & ~PUD_MASK) && (address + PUD_SIZE <= end) && in vmem_add_mem()
112 address += PUD_SIZE; in vmem_add_mem()
180 address += PUD_SIZE; in vmem_remove_range()
185 address += PUD_SIZE; in vmem_remove_range()
Dhugetlbpage.c137 size = PUD_SIZE; in clear_huge_pte_skeys()
199 if (sz == PUD_SIZE) in huge_pte_alloc()
259 } else if (MACHINE_HAS_EDAT2 && size == PUD_SIZE) { in setup_hugepagesz()
/Linux-v4.19/mm/kasan/
Dkasan_init.c146 if (IS_ALIGNED(addr, PUD_SIZE) && end - addr >= PUD_SIZE) { in zero_pud_populate()
399 if (IS_ALIGNED(addr, PUD_SIZE) && in kasan_remove_pud_table()
400 IS_ALIGNED(next, PUD_SIZE)) in kasan_remove_pud_table()
/Linux-v4.19/lib/
Dioremap.c120 ((next - addr) == PUD_SIZE) && in ioremap_pud_range()
121 IS_ALIGNED(phys_addr + addr, PUD_SIZE) && in ioremap_pud_range()
/Linux-v4.19/drivers/dax/
Ddevice.c341 unsigned int fault_size = PUD_SIZE; in __dev_dax_pud_fault()
348 if (dax_region->align > PUD_SIZE) { in __dev_dax_pud_fault()
367 (pud_addr + PUD_SIZE) > vmf->vma->vm_end) in __dev_dax_pud_fault()
371 phys = dax_pgoff_to_phys(dev_dax, pgoff, PUD_SIZE); in __dev_dax_pud_fault()
415 fault_size = PUD_SIZE; in dev_dax_huge_fault()
/Linux-v4.19/arch/arm64/include/asm/
Dpgtable-hwdef.h71 #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) macro
72 #define PUD_MASK (~(PUD_SIZE-1))
/Linux-v4.19/arch/sparc/mm/
Dhugetlbpage.c287 if (sz >= PUD_SIZE) in huge_pte_alloc()
330 if (size >= PUD_SIZE) in set_huge_pte_at()
367 if (size >= PUD_SIZE) in huge_ptep_get_and_clear()
/Linux-v4.19/arch/mips/include/asm/
Dpgtable-64.h59 #define PUD_SIZE (1UL << PUD_SHIFT) macro
60 #define PUD_MASK (~(PUD_SIZE-1))
/Linux-v4.19/arch/x86/xen/
Dxen-head.S83 ELFNOTE(Xen, XEN_ELFNOTE_INIT_P2M, .quad (PUD_SIZE * PTRS_PER_PUD))

12