Lines Matching full:end
93 static void vmemmap_mark_sub_pmd_used(unsigned long start, unsigned long end) in vmemmap_mark_sub_pmd_used() argument
104 static void vmemmap_use_sub_pmd(unsigned long start, unsigned long end) in vmemmap_use_sub_pmd() argument
111 unused_sub_pmd_start = end; in vmemmap_use_sub_pmd()
117 vmemmap_mark_sub_pmd_used(start, end); in vmemmap_use_sub_pmd()
120 static void vmemmap_use_new_sub_pmd(unsigned long start, unsigned long end) in vmemmap_use_new_sub_pmd() argument
127 vmemmap_mark_sub_pmd_used(start, end); in vmemmap_use_new_sub_pmd()
137 if (!IS_ALIGNED(end, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
138 unused_sub_pmd_start = end; in vmemmap_use_new_sub_pmd()
142 static bool vmemmap_unuse_sub_pmd(unsigned long start, unsigned long end) in vmemmap_unuse_sub_pmd() argument
147 memset((void *)start, PAGE_UNUSED, end - start); in vmemmap_unuse_sub_pmd()
153 unsigned long end, bool add, bool direct) in modify_pte_table() argument
164 for (; addr < end; addr += PAGE_SIZE, pte++) { in modify_pte_table()
210 unsigned long end, bool add, bool direct) in modify_pmd_table() argument
222 for (; addr < end; addr = next, pmd++) { in modify_pmd_table()
223 next = pmd_addr_end(addr, end); in modify_pmd_table()
292 const unsigned long end = start + PUD_SIZE; in try_free_pmd_table() local
297 if (end > VMALLOC_START) in try_free_pmd_table()
300 if (start < KASAN_SHADOW_END && KASAN_SHADOW_START > end) in try_free_pmd_table()
311 static int modify_pud_table(p4d_t *p4d, unsigned long addr, unsigned long end, in modify_pud_table() argument
323 for (; addr < end; addr = next, pud++) { in modify_pud_table()
324 next = pud_addr_end(addr, end); in modify_pud_table()
367 const unsigned long end = start + P4D_SIZE; in try_free_pud_table() local
372 if (end > VMALLOC_START) in try_free_pud_table()
375 if (start < KASAN_SHADOW_END && KASAN_SHADOW_START > end) in try_free_pud_table()
388 static int modify_p4d_table(pgd_t *pgd, unsigned long addr, unsigned long end, in modify_p4d_table() argument
397 for (; addr < end; addr = next, p4d++) { in modify_p4d_table()
398 next = p4d_addr_end(addr, end); in modify_p4d_table()
421 const unsigned long end = start + PGDIR_SIZE; in try_free_p4d_table() local
426 if (end > VMALLOC_START) in try_free_p4d_table()
429 if (start < KASAN_SHADOW_END && KASAN_SHADOW_START > end) in try_free_p4d_table()
442 static int modify_pagetable(unsigned long start, unsigned long end, bool add, in modify_pagetable() argument
450 if (WARN_ON_ONCE(!PAGE_ALIGNED(start | end))) in modify_pagetable()
452 for (addr = start; addr < end; addr = next) { in modify_pagetable()
453 next = pgd_addr_end(addr, end); in modify_pagetable()
474 flush_tlb_kernel_range(start, end); in modify_pagetable()
478 static int add_pagetable(unsigned long start, unsigned long end, bool direct) in add_pagetable() argument
480 return modify_pagetable(start, end, true, direct); in add_pagetable()
483 static int remove_pagetable(unsigned long start, unsigned long end, bool direct) in remove_pagetable() argument
485 return modify_pagetable(start, end, false, direct); in remove_pagetable()
507 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, in vmemmap_populate() argument
514 ret = add_pagetable(start, end, false); in vmemmap_populate()
516 remove_pagetable(start, end, false); in vmemmap_populate()
521 void vmemmap_free(unsigned long start, unsigned long end, in vmemmap_free() argument
525 remove_pagetable(start, end, false); in vmemmap_free()
541 mhp_range.end = VMEM_MAX_PHYS - 1; in arch_get_mappable_range()
551 start + size > range.end + 1 || in vmem_add_mapping()
570 phys_addr_t base, end; in vmem_map_init() local
573 for_each_mem_range(i, &base, &end) in vmem_map_init()
574 vmem_add_range(base, end - base); in vmem_map_init()