Lines Matching refs:start_pfn

566 	unsigned long sp, start_pfn;  in page_outside_zone_boundaries()  local
570 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries()
579 start_pfn, start_pfn + sp); in page_outside_zone_boundaries()
1399 unsigned long start_pfn = PFN_DOWN(start); in reserve_bootmem_region() local
1402 for (; start_pfn < end_pfn; start_pfn++) { in reserve_bootmem_region()
1403 if (pfn_valid(start_pfn)) { in reserve_bootmem_region()
1404 struct page *page = pfn_to_page(start_pfn); in reserve_bootmem_region()
1406 init_reserved_page(start_pfn); in reserve_bootmem_region()
1522 struct page *__pageblock_pfn_to_page(unsigned long start_pfn, in __pageblock_pfn_to_page() argument
1531 if (!pfn_valid(start_pfn) || !pfn_valid(end_pfn)) in __pageblock_pfn_to_page()
1534 start_page = pfn_to_online_page(start_pfn); in __pageblock_pfn_to_page()
1730 deferred_init_maxorder(u64 *i, struct zone *zone, unsigned long *start_pfn, in deferred_init_maxorder() argument
1733 unsigned long mo_pfn = ALIGN(*start_pfn + 1, MAX_ORDER_NR_PAGES); in deferred_init_maxorder()
1734 unsigned long spfn = *start_pfn, epfn = *end_pfn; in deferred_init_maxorder()
1739 for_each_free_mem_pfn_range_in_zone_from(j, zone, start_pfn, end_pfn) { in deferred_init_maxorder()
1742 if (mo_pfn <= *start_pfn) in deferred_init_maxorder()
1746 nr_pages += deferred_init_pages(zone, *start_pfn, t); in deferred_init_maxorder()
1749 *start_pfn = mo_pfn; in deferred_init_maxorder()
2292 unsigned long start_pfn, end_pfn; in move_freepages_block() local
2298 start_pfn = page_to_pfn(page); in move_freepages_block()
2299 start_pfn = start_pfn & ~(pageblock_nr_pages-1); in move_freepages_block()
2300 start_page = pfn_to_page(start_pfn); in move_freepages_block()
2302 end_pfn = start_pfn + pageblock_nr_pages - 1; in move_freepages_block()
2305 if (!zone_spans_pfn(zone, start_pfn)) in move_freepages_block()
5880 unsigned long start_pfn, enum memmap_context context, in memmap_init_zone() argument
5883 unsigned long pfn, end_pfn = start_pfn + size; in memmap_init_zone()
5901 if (start_pfn == altmap->base_pfn) in memmap_init_zone()
5902 start_pfn += altmap->reserve; in memmap_init_zone()
5907 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in memmap_init_zone()
5949 unsigned long start_pfn, in memmap_init_zone_device() argument
5953 unsigned long pfn, end_pfn = start_pfn + size; in memmap_init_zone_device()
5969 start_pfn = altmap->base_pfn + vmem_altmap_offset(altmap); in memmap_init_zone_device()
5970 size = end_pfn - start_pfn; in memmap_init_zone_device()
5973 for (pfn = start_pfn; pfn < end_pfn; pfn++) { in memmap_init_zone_device()
6031 unsigned long zone, unsigned long start_pfn) in memmap_init() argument
6033 memmap_init_zone(size, nid, zone, start_pfn, MEMMAP_EARLY, NULL); in memmap_init()
6240 unsigned long start_pfn, end_pfn; in __early_pfn_to_nid() local
6246 nid = memblock_search_pfn_nid(pfn, &start_pfn, &end_pfn); in __early_pfn_to_nid()
6248 state->last_start = start_pfn; in __early_pfn_to_nid()
6268 unsigned long start_pfn, end_pfn; in free_bootmem_with_active_regions() local
6271 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, &this_nid) { in free_bootmem_with_active_regions()
6272 start_pfn = min(start_pfn, max_low_pfn); in free_bootmem_with_active_regions()
6275 if (start_pfn < end_pfn) in free_bootmem_with_active_regions()
6276 memblock_free_early_nid(PFN_PHYS(start_pfn), in free_bootmem_with_active_regions()
6277 (end_pfn - start_pfn) << PAGE_SHIFT, in free_bootmem_with_active_regions()
6291 unsigned long start_pfn, end_pfn; in sparse_memory_present_with_active_regions() local
6294 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, &this_nid) in sparse_memory_present_with_active_regions()
6295 memory_present(this_nid, start_pfn, end_pfn); in sparse_memory_present_with_active_regions()
6310 unsigned long *start_pfn, unsigned long *end_pfn) in get_pfn_range_for_nid() argument
6315 *start_pfn = -1UL; in get_pfn_range_for_nid()
6319 *start_pfn = min(*start_pfn, this_start_pfn); in get_pfn_range_for_nid()
6323 if (*start_pfn == -1UL) in get_pfn_range_for_nid()
6324 *start_pfn = 0; in get_pfn_range_for_nid()
6431 unsigned long start_pfn, end_pfn; in __absent_pages_in_range() local
6434 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in __absent_pages_in_range()
6435 start_pfn = clamp(start_pfn, range_start_pfn, range_end_pfn); in __absent_pages_in_range()
6437 nr_absent -= end_pfn - start_pfn; in __absent_pages_in_range()
6449 unsigned long __init absent_pages_in_range(unsigned long start_pfn, in absent_pages_in_range() argument
6452 return __absent_pages_in_range(MAX_NUMNODES, start_pfn, end_pfn); in absent_pages_in_range()
6485 unsigned long start_pfn, end_pfn; in zone_absent_pages_in_node() local
6489 start_pfn = clamp(memblock_region_memory_base_pfn(r), in zone_absent_pages_in_node()
6496 nr_absent += end_pfn - start_pfn; in zone_absent_pages_in_node()
6500 nr_absent += end_pfn - start_pfn; in zone_absent_pages_in_node()
6888 unsigned long start_pfn = 0; in free_area_init_node() local
6898 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); in free_area_init_node()
6900 (u64)start_pfn << PAGE_SHIFT, in free_area_init_node()
6903 start_pfn = node_start_pfn; in free_area_init_node()
6905 calculate_node_totalpages(pgdat, start_pfn, end_pfn, in free_area_init_node()
7045 unsigned long start_pfn; in find_min_pfn_for_node() local
7048 for_each_mem_pfn_range(i, nid, &start_pfn, NULL, NULL) in find_min_pfn_for_node()
7049 min_pfn = min(min_pfn, start_pfn); in find_min_pfn_for_node()
7078 unsigned long start_pfn, end_pfn; in early_calculate_totalpages() local
7081 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in early_calculate_totalpages()
7082 unsigned long pages = end_pfn - start_pfn; in early_calculate_totalpages()
7209 unsigned long start_pfn, end_pfn; in find_zone_movable_pfns_for_nodes() local
7227 for_each_mem_pfn_range(i, nid, &start_pfn, &end_pfn, NULL) { in find_zone_movable_pfns_for_nodes()
7230 start_pfn = max(start_pfn, zone_movable_pfn[nid]); in find_zone_movable_pfns_for_nodes()
7231 if (start_pfn >= end_pfn) in find_zone_movable_pfns_for_nodes()
7235 if (start_pfn < usable_startpfn) { in find_zone_movable_pfns_for_nodes()
7238 - start_pfn; in find_zone_movable_pfns_for_nodes()
7257 start_pfn = usable_startpfn; in find_zone_movable_pfns_for_nodes()
7265 size_pages = end_pfn - start_pfn; in find_zone_movable_pfns_for_nodes()
7268 zone_movable_pfn[nid] = start_pfn + size_pages; in find_zone_movable_pfns_for_nodes()
7336 unsigned long start_pfn, end_pfn; in free_area_init_nodes() local
7345 start_pfn = find_min_pfn_with_active_regions(); in free_area_init_nodes()
7351 end_pfn = max(max_zone_pfn[i], start_pfn); in free_area_init_nodes()
7352 arch_zone_lowest_possible_pfn[i] = start_pfn; in free_area_init_nodes()
7355 start_pfn = end_pfn; in free_area_init_nodes()
7393 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { in free_area_init_nodes()
7395 (u64)start_pfn << PAGE_SHIFT, in free_area_init_nodes()
7397 subsection_map_init(start_pfn, end_pfn - start_pfn); in free_area_init_nodes()
8559 __offline_isolated_pages(unsigned long start_pfn, unsigned long end_pfn) in __offline_isolated_pages() argument
8569 for (pfn = start_pfn; pfn < end_pfn; pfn++) in __offline_isolated_pages()
8578 pfn = start_pfn; in __offline_isolated_pages()