Home
last modified time | relevance | path

Searched refs:PMD_SIZE (Results 1 – 25 of 103) sorted by relevance

12345

/Linux-v5.10/arch/s390/mm/
Dvmem.c88 ALIGN(unused_pmd_start, PMD_SIZE) - unused_pmd_start); in vmemmap_flush_unused_pmd()
111 if (likely(IS_ALIGNED(unused_pmd_start, PMD_SIZE))) in vmemmap_use_sub_pmd()
121 void *page = __va(ALIGN_DOWN(start, PMD_SIZE)); in vmemmap_use_new_sub_pmd()
129 if (!IS_ALIGNED(start, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
136 if (!IS_ALIGNED(end, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
143 void *page = __va(ALIGN_DOWN(start, PMD_SIZE)); in vmemmap_unuse_sub_pmd()
147 return !memchr_inv(page, PAGE_UNUSED, PMD_SIZE); in vmemmap_unuse_sub_pmd()
227 if (IS_ALIGNED(addr, PMD_SIZE) && in modify_pmd_table()
228 IS_ALIGNED(next, PMD_SIZE)) { in modify_pmd_table()
230 vmem_free_pages(pmd_deref(*pmd), get_order(PMD_SIZE)); in modify_pmd_table()
[all …]
Dkasan_init.c169 IS_ALIGNED(address, PMD_SIZE) && in kasan_early_vmemmap_populate()
170 end - address >= PMD_SIZE) { in kasan_early_vmemmap_populate()
173 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate()
177 if (has_edat && address && end - address >= PMD_SIZE && in kasan_early_vmemmap_populate()
188 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate()
195 address = (address + PMD_SIZE) & PMD_MASK; in kasan_early_vmemmap_populate()
/Linux-v5.10/arch/m68k/mm/
Dkmap.c51 #define IO_SIZE PMD_SIZE
86 virtaddr += PMD_SIZE; in __free_io_area()
87 size -= PMD_SIZE; in __free_io_area()
249 if (!(virtaddr & (PMD_SIZE-1))) in __ioremap()
264 physaddr += PMD_SIZE; in __ioremap()
265 virtaddr += PMD_SIZE; in __ioremap()
266 size -= PMD_SIZE; in __ioremap()
371 virtaddr += PMD_SIZE; in kernel_set_cachemode()
372 size -= PMD_SIZE; in kernel_set_cachemode()
Dmotorola.c309 if (!(virtaddr & (PMD_SIZE-1))) in map_node()
344 physaddr += PMD_SIZE; in map_node()
358 size -= PMD_SIZE; in map_node()
359 virtaddr += PMD_SIZE; in map_node()
/Linux-v5.10/arch/x86/mm/
Dinit.c323 unsigned long start = round_down(mr[i].start, PMD_SIZE); in adjust_range_page_size_mask()
324 unsigned long end = round_up(mr[i].end, PMD_SIZE); in adjust_range_page_size_mask()
390 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range()
392 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
394 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
404 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
406 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
409 if (end_pfn > round_down(limit_pfn, PFN_DOWN(PMD_SIZE))) in split_mem_range()
410 end_pfn = round_down(limit_pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
431 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
[all …]
Dinit_64.c372 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { in __init_extra_mapping()
424 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; in cleanup_highmap()
435 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap()
515 paddr_next = (paddr & PMD_MASK) + PMD_SIZE; in phys_pmd_init()
900 vmem_altmap_free(altmap, PMD_SIZE / PAGE_SIZE); in free_hugepage_table()
902 free_pagetable(page, get_order(PMD_SIZE)); in free_hugepage_table()
1047 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table()
1048 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table()
1063 PMD_SIZE)) { in remove_pmd_table()
1354 all_end = roundup((unsigned long)_brk_end, PMD_SIZE); in mark_rodata_ro()
[all …]
Dmem_encrypt.c134 vaddr += PMD_SIZE; in __sme_early_map_unmap_mem()
135 paddr += PMD_SIZE; in __sme_early_map_unmap_mem()
136 size = (size <= PMD_SIZE) ? 0 : size - PMD_SIZE; in __sme_early_map_unmap_mem()
/Linux-v5.10/arch/sh/include/asm/
Dpgtable-3level.h23 #define PMD_SIZE (1UL << PMD_SHIFT) macro
24 #define PMD_MASK (~(PMD_SIZE-1))
26 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE)
/Linux-v5.10/arch/parisc/kernel/
Dpci-dma.c85 if (end > PMD_SIZE) in map_pte_uncached()
86 end = PMD_SIZE; in map_pte_uncached()
120 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in map_pmd_uncached()
121 orig_vaddr += PMD_SIZE; in map_pmd_uncached()
170 if (end > PMD_SIZE) in unmap_uncached_pte()
171 end = PMD_SIZE; in unmap_uncached_pte()
210 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in unmap_uncached_pmd()
211 orig_vaddr += PMD_SIZE; in unmap_uncached_pmd()
/Linux-v5.10/arch/riscv/mm/
Dinit.c346 if (sz == PMD_SIZE) { in create_pmd_mapping()
410 if ((base & (PMD_SIZE - 1)) || (size & (PMD_SIZE - 1))) in best_map_size()
413 return PMD_SIZE; in best_map_size()
471 (uintptr_t)fixmap_pte, PMD_SIZE, PAGE_TABLE); in setup_vm()
476 load_pa, PMD_SIZE, PAGE_KERNEL_EXEC); in setup_vm()
499 pa = dtb_pa & ~(PMD_SIZE - 1); in setup_vm()
501 pa, PMD_SIZE, PAGE_KERNEL); in setup_vm()
502 create_pmd_mapping(early_dtb_pmd, DTB_EARLY_BASE_VA + PMD_SIZE, in setup_vm()
503 pa + PMD_SIZE, PMD_SIZE, PAGE_KERNEL); in setup_vm()
504 dtb_early_va = (void *)DTB_EARLY_BASE_VA + (dtb_pa & (PMD_SIZE - 1)); in setup_vm()
/Linux-v5.10/arch/x86/include/asm/
Dpgtable_32_types.h12 # define PMD_SIZE (1UL << PMD_SHIFT) macro
13 # define PMD_MASK (~(PMD_SIZE - 1))
Dpgtable_64_types.h98 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro
99 #define PMD_MASK (~(PMD_SIZE - 1))
/Linux-v5.10/arch/arm64/mm/
Dhugetlbpage.c68 case PMD_SIZE: in arch_hugetlb_migration_supported()
116 *pgsize = PMD_SIZE; in find_num_contig()
132 case PMD_SIZE: in num_contig_ptes()
136 *pgsize = PMD_SIZE; in num_contig_ptes()
286 } else if (sz == PMD_SIZE) { in huge_pte_alloc()
331 if (!(sz == PMD_SIZE || sz == CONT_PMD_SIZE) && in huge_pte_offset()
352 } else if (pagesize != PUD_SIZE && pagesize != PMD_SIZE) { in arch_make_huge_pte()
512 case PMD_SIZE: in arch_hugetlb_valid_size()
/Linux-v5.10/include/asm-generic/
Dpgtable-nopmd.h22 #define PMD_SIZE (1UL << PMD_SHIFT) macro
23 #define PMD_MASK (~(PMD_SIZE-1))
/Linux-v5.10/arch/nios2/mm/
Dioremap.c33 if (end > PMD_SIZE) in remap_area_pte()
34 end = PMD_SIZE; in remap_area_pte()
70 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
/Linux-v5.10/arch/powerpc/include/asm/nohash/64/
Dpgtable-4k.h31 #define PMD_SIZE (1UL << PMD_SHIFT) macro
32 #define PMD_MASK (~(PMD_SIZE-1))
/Linux-v5.10/arch/riscv/include/asm/
Dpgtable-64.h18 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro
19 #define PMD_MASK (~(PMD_SIZE - 1))
/Linux-v5.10/arch/x86/kernel/
Dvmlinux.lds.S69 #define ALIGN_ENTRY_TEXT_BEGIN . = ALIGN(PMD_SIZE);
70 #define ALIGN_ENTRY_TEXT_END . = ALIGN(PMD_SIZE);
81 . = ALIGN(PMD_SIZE); \
86 . = ALIGN(PMD_SIZE); \
/Linux-v5.10/arch/m68k/include/asm/
Dpgtable_mm.h41 #define PMD_SIZE (1UL << PMD_SHIFT) macro
42 #define PMD_MASK (~(PMD_SIZE-1))
/Linux-v5.10/arch/x86/boot/compressed/
Dident_map_64.c98 start = round_down(start, PMD_SIZE); in add_identity_map()
99 end = round_up(end, PMD_SIZE); in add_identity_map()
343 end = address + PMD_SIZE; in do_boot_page_fault()
/Linux-v5.10/arch/um/include/asm/
Dpgtable-3level.h27 #define PMD_SIZE (1UL << PMD_SHIFT) macro
28 #define PMD_MASK (~(PMD_SIZE-1))
/Linux-v5.10/arch/arm/include/asm/
Dpgtable-2level.h87 #define PMD_SIZE (1UL << PMD_SHIFT) macro
88 #define PMD_MASK (~(PMD_SIZE-1))
/Linux-v5.10/arch/sparc/mm/
Dhugetlbpage.c293 if (sz >= PMD_SIZE) in huge_pte_alloc()
337 else if (size >= PMD_SIZE) in set_huge_pte_at()
374 else if (size >= PMD_SIZE) in huge_ptep_get_and_clear()
505 addr += PMD_SIZE; in hugetlb_free_pgd_range()
515 end -= PMD_SIZE; in hugetlb_free_pgd_range()
/Linux-v5.10/drivers/dax/
Ddevice.c113 unsigned int fault_size = PMD_SIZE; in __dev_dax_pmd_fault()
118 if (dev_dax->align > PMD_SIZE) { in __dev_dax_pmd_fault()
131 (pmd_addr + PMD_SIZE) > vmf->vma->vm_end) in __dev_dax_pmd_fault()
135 phys = dax_pgoff_to_phys(dev_dax, pgoff, PMD_SIZE); in __dev_dax_pmd_fault()
216 fault_size = PMD_SIZE; in dev_dax_huge_fault()
/Linux-v5.10/arch/arm64/include/asm/
Dpgtable-hwdef.h50 #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) macro
51 #define PMD_MASK (~(PMD_SIZE-1))
91 #define CONT_PMD_SIZE (CONT_PMDS * PMD_SIZE)

12345