Lines Matching refs:end_pfn

415 defer_init(int nid, unsigned long pfn, unsigned long end_pfn)  in defer_init()  argument
423 if (prev_end_pfn != end_pfn) { in defer_init()
424 prev_end_pfn = end_pfn; in defer_init()
429 if (end_pfn < pgdat_end_pfn(NODE_DATA(nid))) in defer_init()
459 static inline bool defer_init(int nid, unsigned long pfn, unsigned long end_pfn) in defer_init() argument
1614 unsigned long end_pfn = PFN_UP(end); in reserve_bootmem_region() local
1616 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region()
1709 unsigned long start_pfn, end_pfn; in __early_pfn_to_nid() local
1715 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn); in __early_pfn_to_nid()
1718 state->last_end = end_pfn; in __early_pfn_to_nid()
1764 unsigned long end_pfn, struct zone *zone) in __pageblock_pfn_to_page() argument
1770 end_pfn--; in __pageblock_pfn_to_page()
1772 if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn)) in __pageblock_pfn_to_page()
1782 end_page = pfn_to_page(end_pfn); in __pageblock_pfn_to_page()
1877 unsigned long end_pfn) in deferred_free_pages() argument
1882 for (; pfn < end_pfn; pfn++) { in deferred_free_pages()
1904 unsigned long end_pfn) in deferred_init_pages() argument
1912 for (; pfn < end_pfn; pfn++) { in deferred_init_pages()
1969 unsigned long *end_pfn) in deferred_init_maxorder() argument
1972 unsigned long spfn = *start_pfn, epfn = *end_pfn; in deferred_init_maxorder()
1977 for_each_free_mem_pfn_range_in_zone_from(j, zone, start_pfn, end_pfn) { in deferred_init_maxorder()
1983 t = min(mo_pfn, *end_pfn); in deferred_init_maxorder()
1986 if (mo_pfn < *end_pfn) { in deferred_init_maxorder()
2012 deferred_init_memmap_chunk(unsigned long start_pfn, unsigned long end_pfn, in deferred_init_memmap_chunk() argument
2025 while (spfn < end_pfn) { in deferred_init_memmap_chunk()
2504 unsigned long start_pfn, unsigned long end_pfn, in move_freepages() argument
2512 for (pfn = start_pfn; pfn <= end_pfn;) { in move_freepages()
2543 unsigned long start_pfn, end_pfn, pfn; in move_freepages_block() local
2550 end_pfn = start_pfn + pageblock_nr_pages - 1; in move_freepages_block()
2555 if (!zone_spans_pfn(zone, end_pfn)) in move_freepages_block()
2558 return move_freepages(zone, start_pfn, end_pfn, migratetype, in move_freepages_block()
6504 unsigned long pfn, end_pfn = start_pfn + size; in memmap_init_range() local
6507 if (highest_memmap_pfn < end_pfn - 1) in memmap_init_range()
6508 highest_memmap_pfn = end_pfn - 1; in memmap_init_range()
6524 end_pfn = altmap->base_pfn + vmem_altmap_offset(altmap); in memmap_init_range()
6528 for (pfn = start_pfn; pfn < end_pfn; ) { in memmap_init_range()
6564 unsigned long pfn, end_pfn = start_pfn + nr_pages; in memmap_init_zone_device() local
6581 nr_pages = end_pfn - start_pfn; in memmap_init_zone_device()
6584 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in memmap_init_zone_device()
6683 unsigned long end_pfn, in memmap_init_zone_range() argument
6691 end_pfn = clamp(end_pfn, zone_start_pfn, zone_end_pfn); in memmap_init_zone_range()
6693 if (start_pfn >= end_pfn) in memmap_init_zone_range()
6696 memmap_init_range(end_pfn - start_pfn, nid, zone_id, start_pfn, in memmap_init_zone_range()
6702 *hole_pfn = end_pfn; in memmap_init_zone_range()
6707 unsigned long start_pfn, end_pfn; in memmap_init() local
6711 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in memmap_init()
6720 memmap_init_zone_range(zone, start_pfn, end_pfn, in memmap_init()
6736 end_pfn = round_up(end_pfn, PAGES_PER_SECTION); in memmap_init()
6737 if (hole_pfn < end_pfn) in memmap_init()
6739 init_unavailable_range(hole_pfn, end_pfn, zone_id, nid); in memmap_init()
7039 unsigned long *start_pfn, unsigned long *end_pfn) in get_pfn_range_for_nid() argument
7045 *end_pfn = 0; in get_pfn_range_for_nid()
7049 *end_pfn = max(*end_pfn, this_end_pfn); in get_pfn_range_for_nid()
7159 unsigned long start_pfn, end_pfn; in __absent_pages_in_range() local
7162 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in __absent_pages_in_range()
7164 end_pfn = clamp(end_pfn, range_start_pfn, range_end_pfn); in __absent_pages_in_range()
7165 nr_absent -= end_pfn - start_pfn; in __absent_pages_in_range()
7178 unsigned long end_pfn) in absent_pages_in_range() argument
7180 return __absent_pages_in_range(MAX_NUMNODES, start_pfn, end_pfn); in absent_pages_in_range()
7212 unsigned long start_pfn, end_pfn; in zone_absent_pages_in_node() local
7218 end_pfn = clamp(memblock_region_memory_end_pfn(r), in zone_absent_pages_in_node()
7223 nr_absent += end_pfn - start_pfn; in zone_absent_pages_in_node()
7227 nr_absent += end_pfn - start_pfn; in zone_absent_pages_in_node()
7574 unsigned long end_pfn = 0; in free_area_init_node() local
7579 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in free_area_init_node()
7587 end_pfn ? ((u64)end_pfn << PAGE_SHIFT) - 1 : 0); in free_area_init_node()
7588 calculate_node_totalpages(pgdat, start_pfn, end_pfn); in free_area_init_node()
7683 unsigned long start_pfn, end_pfn; in early_calculate_totalpages() local
7686 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in early_calculate_totalpages()
7687 unsigned long pages = end_pfn - start_pfn; in early_calculate_totalpages()
7814 unsigned long start_pfn, end_pfn; in find_zone_movable_pfns_for_nodes() local
7832 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in find_zone_movable_pfns_for_nodes()
7836 if (start_pfn >= end_pfn) in find_zone_movable_pfns_for_nodes()
7842 kernel_pages = min(end_pfn, usable_startpfn) in find_zone_movable_pfns_for_nodes()
7851 if (end_pfn <= usable_startpfn) { in find_zone_movable_pfns_for_nodes()
7859 zone_movable_pfn[nid] = end_pfn; in find_zone_movable_pfns_for_nodes()
7870 size_pages = end_pfn - start_pfn; in find_zone_movable_pfns_for_nodes()
7950 unsigned long start_pfn, end_pfn; in free_area_init() local
7972 end_pfn = max(max_zone_pfn[zone], start_pfn); in free_area_init()
7974 arch_zone_highest_possible_pfn[zone] = end_pfn; in free_area_init()
7976 start_pfn = end_pfn; in free_area_init()
8014 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in free_area_init()
8017 ((u64)end_pfn << PAGE_SHIFT) - 1); in free_area_init()
8018 subsection_map_init(start_pfn, end_pfn - start_pfn); in free_area_init()
9159 unsigned long end_pfn = start_pfn + nr_pages; in __alloc_contig_pages() local
9161 return alloc_contig_range(start_pfn, end_pfn, MIGRATE_MOVABLE, in __alloc_contig_pages()
9168 unsigned long i, end_pfn = start_pfn + nr_pages; in pfn_range_valid_contig() local
9171 for (i = start_pfn; i < end_pfn; i++) { in pfn_range_valid_contig()
9320 void __offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn) in __offline_isolated_pages() argument
9328 offline_mem_sections(pfn, end_pfn); in __offline_isolated_pages()
9331 while (pfn < end_pfn) { in __offline_isolated_pages()