/Linux-v4.19/arch/x86/mm/ |
D | numa_32.c | 55 start = round_down(start, PAGES_PER_SECTION); in memory_present() 56 end = round_up(end, PAGES_PER_SECTION); in memory_present() 57 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present() 58 physnode_map[pfn / PAGES_PER_SECTION] = nid; in memory_present()
|
D | numa.c | 563 if (pfn_align && pfn_align < PAGES_PER_SECTION) { in numa_register_memblks() 566 PFN_PHYS(PAGES_PER_SECTION) >> 20); in numa_register_memblks()
|
/Linux-v4.19/mm/ |
D | sparse.c | 225 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) { in memory_present() 376 return ALIGN(sizeof(struct page) * PAGES_PER_SECTION, PMD_SIZE); in section_map_size() 382 return PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION); in section_map_size() 528 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections() 547 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections() 575 unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION); in __kfree_section_memmap() 583 unsigned long end = (unsigned long)(memmap + PAGES_PER_SECTION); in free_map_bootmem() 592 unsigned long memmap_size = sizeof(struct page) * PAGES_PER_SECTION; in __kmalloc_section_memmap() 623 get_order(sizeof(struct page) * PAGES_PER_SECTION)); in __kfree_section_memmap() 633 nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page)) in free_map_bootmem() [all …]
|
D | page_ext.c | 240 table_size = get_entry_size() * PAGES_PER_SECTION; in init_section_page_ext() 273 table_size = get_entry_size() * PAGES_PER_SECTION; in free_page_ext() 313 for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION) { in online_page_ext() 322 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) in online_page_ext() 336 for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION) in offline_page_ext() 392 pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) { in page_ext_init()
|
D | memory_hotplug.c | 180 mapsize = sizeof(struct page) * PAGES_PER_SECTION; in register_page_bootmem_info_section() 208 register_page_bootmem_memmap(section_nr, memmap, PAGES_PER_SECTION); in register_page_bootmem_info_section() 236 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node() 325 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SECTION) { in find_smallest_section_pfn() 353 for (; pfn >= start_pfn; pfn -= PAGES_PER_SECTION) { in find_biggest_section_pfn() 415 for (; pfn < zone_end_pfn; pfn += PAGES_PER_SECTION) { in shrink_zone_span() 483 for (; pfn < pgdat_end_pfn; pfn += PAGES_PER_SECTION) { in shrink_pgdat_span() 508 int nr_pages = PAGES_PER_SECTION; in __remove_zone() 584 BUG_ON(nr_pages % PAGES_PER_SECTION); in __remove_pages() 586 sections_to_remove = nr_pages / PAGES_PER_SECTION; in __remove_pages() [all …]
|
D | sparse-vmemmap.c | 255 map = pfn_to_page(pnum * PAGES_PER_SECTION); in sparse_mem_map_populate() 257 end = (unsigned long)(map + PAGES_PER_SECTION); in sparse_mem_map_populate()
|
D | page_alloc.c | 321 (pfn & (PAGES_PER_SECTION - 1)) == 0) { in update_defer_init() 356 pfn &= (PAGES_PER_SECTION-1); in pfn_to_bitidx() 1624 unsigned long nr_pages_needed = ALIGN(1 << order, PAGES_PER_SECTION); in deferred_grow_zone() 1669 t = ALIGN(spfn + PAGES_PER_SECTION, PAGES_PER_SECTION); in deferred_grow_zone() 6391 pgdat->static_init_pgcnt = min_t(unsigned long, PAGES_PER_SECTION, in pgdat_set_deferred_range()
|
/Linux-v4.19/arch/x86/include/asm/ |
D | mmzone_32.h | 30 #define PAGES_PER_SECTION (MAX_NR_PAGES/MAX_SECTIONS) macro 37 return((int) physnode_map[(pfn) / PAGES_PER_SECTION]); in pfn_to_nid()
|
/Linux-v4.19/drivers/base/ |
D | memory.c | 139 ret &= is_mem_section_removable(pfn, PAGES_PER_SECTION); in show_mem_removable() 211 section_nr, pfn, pfn + PAGES_PER_SECTION); in pages_correctly_probed() 215 section_nr, pfn, pfn + PAGES_PER_SECTION); in pages_correctly_probed() 219 section_nr, pfn, pfn + PAGES_PER_SECTION); in pages_correctly_probed() 222 pfn += PAGES_PER_SECTION; in pages_correctly_probed() 237 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in memory_block_action() 414 unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; in show_valid_zones() 513 unsigned long pages_per_block = PAGES_PER_SECTION * sections_per_block; in memory_probe_store()
|
D | node.c | 411 sect_end_pfn += PAGES_PER_SECTION - 1; in register_mem_sect_under_node() 420 pfn = round_down(pfn + PAGES_PER_SECTION, in register_mem_sect_under_node() 421 PAGES_PER_SECTION) - 1; in register_mem_sect_under_node() 467 sect_end_pfn = sect_start_pfn + PAGES_PER_SECTION - 1; in unregister_mem_sect_under_nodes()
|
/Linux-v4.19/arch/arm64/mm/ |
D | init.c | 560 start = min(start, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap() 579 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) in free_unused_memmap() 580 free_memmap(prev_end, ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
|
/Linux-v4.19/drivers/xen/ |
D | balloon.c | 287 PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL); in additional_memory_resource() 311 PAGES_PER_SECTION * PAGE_SIZE, NULL, NULL); in additional_memory_resource() 354 balloon_hotplug = round_up(credit, PAGES_PER_SECTION); in reserve_additional_memory()
|
/Linux-v4.19/include/linux/ |
D | memory.h | 124 #define CONFIG_MEM_BLOCK_SIZE (PAGES_PER_SECTION<<PAGE_SHIFT)
|
D | mmzone.h | 1079 #define PAGES_PER_SECTION (1UL << PFN_SECTION_SHIFT) macro 1080 #define PAGE_SECTION_MASK (~(PAGES_PER_SECTION-1)) 1098 #define SECTION_ALIGN_UP(pfn) (((pfn) + PAGES_PER_SECTION - 1) & PAGE_SECTION_MASK)
|
/Linux-v4.19/arch/arm/mm/ |
D | init.c | 403 ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap() 429 if (!IS_ALIGNED(prev_end, PAGES_PER_SECTION)) in free_unused_memmap() 431 ALIGN(prev_end, PAGES_PER_SECTION)); in free_unused_memmap()
|
/Linux-v4.19/arch/powerpc/mm/ |
D | init_64.c | 101 for (; start < end; start += (PAGES_PER_SECTION * sizeof(struct page))) in vmemmap_populated()
|
/Linux-v4.19/arch/powerpc/platforms/pseries/ |
D | hotplug-memory.c | 402 rc &= is_mem_section_removable(pfn, PAGES_PER_SECTION); in lmb_is_removable()
|