/Linux-v5.4/arch/x86/mm/ |
D | init.c | 262 unsigned long start_pfn, unsigned long end_pfn, in save_mr() argument 265 if (start_pfn < end_pfn) { in save_mr() 269 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr() 340 unsigned long start_pfn, end_pfn, limit_pfn; in split_mem_range() local 356 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range() 358 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 360 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range() 362 if (end_pfn > limit_pfn) in split_mem_range() 363 end_pfn = limit_pfn; in split_mem_range() 364 if (start_pfn < end_pfn) { in split_mem_range() [all …]
|
D | init_32.c | 259 unsigned long start_pfn, end_pfn; in kernel_physical_mapping_init() local 270 end_pfn = end >> PAGE_SHIFT; in kernel_physical_mapping_init() 299 if (pfn >= end_pfn) in kernel_physical_mapping_init() 307 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; in kernel_physical_mapping_init() 347 for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; in kernel_physical_mapping_init() 435 unsigned long start_pfn, unsigned long end_pfn) in add_highpages_with_active_regions() argument 442 start_pfn, end_pfn); in add_highpages_with_active_regions() 444 start_pfn, end_pfn); in add_highpages_with_active_regions()
|
/Linux-v5.4/include/trace/events/ |
D | page_isolation.h | 14 unsigned long end_pfn, 17 TP_ARGS(start_pfn, end_pfn, fin_pfn), 21 __field(unsigned long, end_pfn) 27 __entry->end_pfn = end_pfn; 32 __entry->start_pfn, __entry->end_pfn, __entry->fin_pfn, 33 __entry->end_pfn <= __entry->fin_pfn ? "success" : "fail")
|
D | compaction.h | 18 unsigned long end_pfn, 22 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken), 26 __field(unsigned long, end_pfn) 33 __entry->end_pfn = end_pfn; 40 __entry->end_pfn, 49 unsigned long end_pfn, 53 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken) 60 unsigned long end_pfn, 64 TP_ARGS(start_pfn, end_pfn, nr_scanned, nr_taken)
|
/Linux-v5.4/mm/ |
D | page_isolation.c | 193 int start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in start_isolate_page_range() argument 202 BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); in start_isolate_page_range() 205 pfn < end_pfn; in start_isolate_page_range() 233 void undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, in undo_isolate_page_range() argument 240 BUG_ON(!IS_ALIGNED(end_pfn, pageblock_nr_pages)); in undo_isolate_page_range() 243 pfn < end_pfn; in undo_isolate_page_range() 259 __test_page_isolated_in_pageblock(unsigned long pfn, unsigned long end_pfn, in __test_page_isolated_in_pageblock() argument 264 while (pfn < end_pfn) { in __test_page_isolated_in_pageblock() 288 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn, in test_pages_isolated() argument 300 for (pfn = start_pfn; pfn < end_pfn; pfn += pageblock_nr_pages) { in test_pages_isolated() [all …]
|
D | page_idle.c | 127 unsigned long pfn, end_pfn; in page_idle_bitmap_read() local 137 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_read() 138 if (end_pfn > max_pfn) in page_idle_bitmap_read() 139 end_pfn = max_pfn; in page_idle_bitmap_read() 141 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_read() 172 unsigned long pfn, end_pfn; in page_idle_bitmap_write() local 182 end_pfn = pfn + count * BITS_PER_BYTE; in page_idle_bitmap_write() 183 if (end_pfn > max_pfn) in page_idle_bitmap_write() 184 end_pfn = max_pfn; in page_idle_bitmap_write() 186 for (; pfn < end_pfn; pfn++) { in page_idle_bitmap_write()
|
D | memory_hotplug.c | 227 unsigned long i, pfn, end_pfn, nr_pages; in register_page_bootmem_info_node() local 238 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node() 241 for (; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in register_page_bootmem_info_node() 331 unsigned long end_pfn) in find_smallest_section_pfn() argument 333 for (; start_pfn < end_pfn; start_pfn += PAGES_PER_SUBSECTION) { in find_smallest_section_pfn() 352 unsigned long end_pfn) in find_biggest_section_pfn() argument 357 pfn = end_pfn - 1; in find_biggest_section_pfn() 375 unsigned long end_pfn) in shrink_zone_span() argument 391 pfn = find_smallest_section_pfn(nid, zone, end_pfn, in shrink_zone_span() 397 } else if (zone_end_pfn == end_pfn) { in shrink_zone_span() [all …]
|
D | compaction.c | 537 unsigned long end_pfn, in isolate_freepages_block() argument 556 for (; blockpfn < end_pfn; blockpfn += stride, cursor += stride) { in isolate_freepages_block() 644 if (unlikely(blockpfn > end_pfn)) in isolate_freepages_block() 645 blockpfn = end_pfn; in isolate_freepages_block() 658 if (strict && blockpfn < end_pfn) in isolate_freepages_block() 683 unsigned long start_pfn, unsigned long end_pfn) in isolate_freepages_range() argument 694 for (; pfn < end_pfn; pfn += isolated, in isolate_freepages_range() 700 block_end_pfn = min(block_end_pfn, end_pfn); in isolate_freepages_range() 710 block_end_pfn = min(block_end_pfn, end_pfn); in isolate_freepages_range() 738 if (pfn < end_pfn) { in isolate_freepages_range() [all …]
|
D | internal.h | 148 unsigned long end_pfn, struct zone *zone); 151 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument 156 return __pageblock_pfn_to_page(start_pfn, end_pfn, zone); in pageblock_pfn_to_page() 219 unsigned long start_pfn, unsigned long end_pfn); 222 unsigned long low_pfn, unsigned long end_pfn); 442 unsigned long *end_pfn); 445 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument
|
D | page_alloc.c | 405 defer_init(int nid, unsigned long pfn, unsigned long end_pfn) in defer_init() argument 413 if (prev_end_pfn != end_pfn) { in defer_init() 414 prev_end_pfn = end_pfn; in defer_init() 419 if (end_pfn < pgdat_end_pfn(NODE_DATA(nid))) in defer_init() 442 static inline bool defer_init(int nid, unsigned long pfn, unsigned long end_pfn) in defer_init() argument 1400 unsigned long end_pfn = PFN_UP(end); in reserve_bootmem_region() local 1402 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region() 1523 unsigned long end_pfn, struct zone *zone) in __pageblock_pfn_to_page() argument 1529 end_pfn--; in __pageblock_pfn_to_page() 1531 if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn)) in __pageblock_pfn_to_page() [all …]
|
D | sparse.c | 157 unsigned long *end_pfn) in mminit_validate_memmodel_limits() argument 168 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 171 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 172 } else if (*end_pfn > max_sparsemem_pfn) { in mminit_validate_memmodel_limits() 175 *start_pfn, *end_pfn, max_sparsemem_pfn); in mminit_validate_memmodel_limits() 177 *end_pfn = max_sparsemem_pfn; in mminit_validate_memmodel_limits() 609 void online_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in online_mem_sections() argument 613 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in online_mem_sections() 628 void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn) in offline_mem_sections() argument 632 for (pfn = start_pfn; pfn < end_pfn; pfn += PAGES_PER_SECTION) { in offline_mem_sections()
|
/Linux-v5.4/arch/mips/loongson64/loongson-3/ |
D | numa.c | 123 u64 node_id, node_psize, start_pfn, end_pfn, mem_start, mem_size; in szmem() local 139 end_pfn = start_pfn + node_psize; in szmem() 144 start_pfn, end_pfn, num_physpages); in szmem() 146 PFN_PHYS(end_pfn - start_pfn), node); in szmem() 151 end_pfn = start_pfn + node_psize; in szmem() 156 start_pfn, end_pfn, num_physpages); in szmem() 158 PFN_PHYS(end_pfn - start_pfn), node); in szmem() 173 unsigned long start_pfn, end_pfn; in node_mem_init() local 179 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); in node_mem_init() 181 node, start_pfn, end_pfn); in node_mem_init() [all …]
|
/Linux-v5.4/arch/sparc/mm/ |
D | init_32.c | 67 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in calc_highpages() local 69 if (end_pfn <= max_low_pfn) in calc_highpages() 75 nr += end_pfn - start_pfn; in calc_highpages() 233 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) in map_high_region() argument 238 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); in map_high_region() 241 for (tmp = start_pfn; tmp < end_pfn; tmp++) in map_high_region() 282 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; in mem_init() local 284 if (end_pfn <= highstart_pfn) in mem_init() 290 map_high_region(start_pfn, end_pfn); in mem_init()
|
/Linux-v5.4/arch/sh/mm/ |
D | numa.c | 28 unsigned long start_pfn, end_pfn; in setup_bootmem_node() local 34 end_pfn = PFN_DOWN(end); in setup_bootmem_node() 41 __add_active_range(nid, start_pfn, end_pfn); in setup_bootmem_node() 52 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_bootmem_node()
|
D | init.c | 201 unsigned long start_pfn, end_pfn; in allocate_pgdat() local 203 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in allocate_pgdat() 215 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in allocate_pgdat() 224 unsigned long start_pfn, end_pfn; in do_init_bootmem() local 226 end_pfn = memblock_region_memory_end_pfn(reg); in do_init_bootmem() 227 __add_active_range(0, start_pfn, end_pfn); in do_init_bootmem()
|
/Linux-v5.4/arch/s390/numa/ |
D | numa.c | 103 unsigned long start_pfn, end_pfn; in numa_setup_memory() local 108 end_pfn = 0; in numa_setup_memory() 112 if (t_end > end_pfn) in numa_setup_memory() 113 end_pfn = t_end; in numa_setup_memory() 115 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in numa_setup_memory()
|
/Linux-v5.4/arch/powerpc/platforms/powernv/ |
D | memtrace.c | 97 u64 start_pfn, end_pfn, nr_pages, pfn; in memtrace_alloc_node() local 105 end_pfn = node_end_pfn(nid); in memtrace_alloc_node() 109 end_pfn = round_down(end_pfn - nr_pages, nr_pages); in memtrace_alloc_node() 112 for (base_pfn = end_pfn; base_pfn > start_pfn; base_pfn -= nr_pages) { in memtrace_alloc_node() 120 end_pfn = base_pfn + nr_pages; in memtrace_alloc_node() 121 for (pfn = base_pfn; pfn < end_pfn; pfn += bytes>> PAGE_SHIFT) { in memtrace_alloc_node()
|
/Linux-v5.4/include/linux/ |
D | node.h | 103 unsigned long end_pfn); 106 unsigned long end_pfn) in link_mem_sections() argument 125 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node() local 131 error = link_mem_sections(nid, start_pfn, end_pfn); in register_one_node()
|
D | page-isolation.h | 46 start_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 54 undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn, 60 int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
|
/Linux-v5.4/arch/x86/xen/ |
D | setup.c | 255 unsigned long end_pfn, unsigned long nr_pages) in xen_set_identity_and_release_chunk() argument 260 WARN_ON(start_pfn > end_pfn); in xen_set_identity_and_release_chunk() 263 end = min(end_pfn, nr_pages); in xen_set_identity_and_release_chunk() 282 set_phys_range_identity(start_pfn, end_pfn); in xen_set_identity_and_release_chunk() 388 unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, in xen_set_identity_and_remap_chunk() argument 393 unsigned long n = end_pfn - start_pfn; in xen_set_identity_and_remap_chunk() 435 for (pfn = start_pfn; pfn <= max_pfn_mapped && pfn < end_pfn; pfn++) in xen_set_identity_and_remap_chunk() 444 unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, in xen_count_remap_pages() argument 450 return remap_pages + min(end_pfn, nr_pages) - start_pfn; in xen_count_remap_pages() 454 unsigned long (*func)(unsigned long start_pfn, unsigned long end_pfn, in xen_foreach_remap_area() argument [all …]
|
/Linux-v5.4/arch/sh/kernel/ |
D | swsusp.c | 22 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; in pfn_is_nosave() local 24 return (pfn >= begin_pfn) && (pfn < end_pfn); in pfn_is_nosave()
|
D | setup.c | 197 unsigned long end_pfn) in __add_active_range() argument 205 end = end_pfn << PAGE_SHIFT; in __add_active_range() 214 start_pfn, end_pfn); in __add_active_range() 238 memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn), in __add_active_range()
|
/Linux-v5.4/arch/arm64/mm/ |
D | numa.c | 218 static void __init setup_node_data(int nid, u64 start_pfn, u64 end_pfn) in setup_node_data() argument 225 if (start_pfn >= end_pfn) in setup_node_data() 246 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; in setup_node_data() 363 unsigned long start_pfn, end_pfn; in numa_register_nodes() local 365 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in numa_register_nodes() 366 setup_node_data(nid, start_pfn, end_pfn); in numa_register_nodes()
|
/Linux-v5.4/arch/alpha/mm/ |
D | numa.c | 220 unsigned long end_pfn = start_pfn + NODE_DATA(nid)->node_present_pages; in paging_init() local 222 if (dma_local_pfn >= end_pfn - start_pfn) in paging_init() 223 zones_size[ZONE_DMA] = end_pfn - start_pfn; in paging_init() 226 zones_size[ZONE_NORMAL] = (end_pfn - start_pfn) - dma_local_pfn; in paging_init()
|
/Linux-v5.4/drivers/hv/ |
D | hv_balloon.c | 436 unsigned long end_pfn; member 446 unsigned long end_pfn; member 583 if ((pfn >= gap->start_pfn) && (pfn < gap->end_pfn)) in has_pfn_is_backed() 605 (pfn < has->end_pfn) && in hv_page_offline_check() 779 (pfn + (1UL << order) > has->end_pfn)) in hv_online_page() 802 if (start_pfn < has->start_pfn || start_pfn >= has->end_pfn) in pfn_covered() 818 gap->end_pfn = start_pfn; in pfn_covered() 828 if ((start_pfn + pfn_cnt) > has->end_pfn) { in pfn_covered() 829 residual = (start_pfn + pfn_cnt - has->end_pfn); in pfn_covered() 837 has->end_pfn += new_inc; in pfn_covered() [all …]
|