Home
last modified time | relevance | path

Searched refs:pgdat (Results 1 – 25 of 40) sorted by relevance

12

/Linux-v5.15/mm/
Dvmscan.c1336 struct pglist_data *pgdat) in demote_page_list() argument
1338 int target_nid = next_demotion_node(pgdat->node_id); in demote_page_list()
1365 struct pglist_data *pgdat, in shrink_page_list() argument
1379 do_demote_pass = can_demote(pgdat->node_id, sc); in shrink_page_list()
1484 test_bit(PGDAT_WRITEBACK, &pgdat->flags)) { in shrink_page_list()
1636 !test_bit(PGDAT_DIRTY, &pgdat->flags))) { in shrink_page_list()
1798 nr_reclaimed += demote_page_list(&demote_pages, pgdat); in shrink_page_list()
2118 static int too_many_isolated(struct pglist_data *pgdat, int file, in too_many_isolated() argument
2130 inactive = node_page_state(pgdat, NR_INACTIVE_FILE); in too_many_isolated()
2131 isolated = node_page_state(pgdat, NR_ISOLATED_FILE); in too_many_isolated()
[all …]
Dvmstat.c246 struct pglist_data *pgdat; in refresh_zone_stat_thresholds() local
252 for_each_online_pgdat(pgdat) { in refresh_zone_stat_thresholds()
254 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold = 0; in refresh_zone_stat_thresholds()
259 struct pglist_data *pgdat = zone->zone_pgdat; in refresh_zone_stat_thresholds() local
271 pgdat_threshold = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold; in refresh_zone_stat_thresholds()
272 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold in refresh_zone_stat_thresholds()
289 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() argument
297 for (i = 0; i < pgdat->nr_zones; i++) { in set_pgdat_percpu_threshold()
298 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold()
347 void __mod_node_page_state(struct pglist_data *pgdat, enum node_stat_item item, in __mod_node_page_state() argument
[all …]
Dcompaction.c383 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument
388 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable()
762 static bool too_many_isolated(pg_data_t *pgdat) in too_many_isolated() argument
766 inactive = node_page_state(pgdat, NR_INACTIVE_FILE) + in too_many_isolated()
767 node_page_state(pgdat, NR_INACTIVE_ANON); in too_many_isolated()
768 active = node_page_state(pgdat, NR_ACTIVE_FILE) + in too_many_isolated()
769 node_page_state(pgdat, NR_ACTIVE_ANON); in too_many_isolated()
770 isolated = node_page_state(pgdat, NR_ISOLATED_FILE) + in too_many_isolated()
771 node_page_state(pgdat, NR_ISOLATED_ANON); in too_many_isolated()
797 pg_data_t *pgdat = cc->zone->zone_pgdat; in isolate_migratepages_block() local
[all …]
Dmmzone.c18 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) in next_online_pgdat() argument
20 int nid = next_online_node(pgdat->node_id); in next_online_pgdat()
32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() local
34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone()
37 pgdat = next_online_pgdat(pgdat); in next_zone()
38 if (pgdat) in next_zone()
39 zone = pgdat->node_zones; in next_zone()
Dmemory_hotplug.c433 static void update_pgdat_span(struct pglist_data *pgdat) in update_pgdat_span() argument
438 for (zone = pgdat->node_zones; in update_pgdat_span()
439 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span()
457 pgdat->node_start_pfn = node_start_pfn; in update_pgdat_span()
458 pgdat->node_spanned_pages = node_end_pfn - node_start_pfn; in update_pgdat_span()
466 struct pglist_data *pgdat = zone->zone_pgdat; in remove_pfn_range_from_zone() local
491 update_pgdat_span(pgdat); in remove_pfn_range_from_zone()
664 static void __meminit resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, in resize_pgdat_range() argument
667 unsigned long old_end_pfn = pgdat_end_pfn(pgdat); in resize_pgdat_range()
669 if (!pgdat->node_spanned_pages || start_pfn < pgdat->node_start_pfn) in resize_pgdat_range()
[all …]
Dworkingset.c187 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, in pack_shadow() argument
193 eviction = (eviction << NODES_SHIFT) | pgdat->node_id; in pack_shadow()
199 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow() argument
214 *pgdat = NODE_DATA(nid); in unpack_shadow()
257 struct pglist_data *pgdat = page_pgdat(page); in workingset_eviction() local
267 lruvec = mem_cgroup_lruvec(target_memcg, pgdat); in workingset_eviction()
272 return pack_shadow(memcgid, pgdat, eviction, PageWorkingset(page)); in workingset_eviction()
291 struct pglist_data *pgdat; in workingset_refault() local
299 unpack_shadow(shadow, &memcgid, &pgdat, &eviction, &workingset); in workingset_refault()
321 eviction_lruvec = mem_cgroup_lruvec(eviction_memcg, pgdat); in workingset_refault()
[all …]
Dshuffle.h11 extern void __shuffle_free_memory(pg_data_t *pgdat);
13 static inline void __meminit shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
17 __shuffle_free_memory(pgdat); in shuffle_free_memory()
40 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
Dpage_alloc.c1582 pg_data_t *pgdat; in init_reserved_page() local
1589 pgdat = NODE_DATA(nid); in init_reserved_page()
1592 struct zone *zone = &pgdat->node_zones[zid]; in init_reserved_page()
2041 pg_data_t *pgdat = data; in deferred_init_memmap() local
2042 const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id); in deferred_init_memmap()
2054 pgdat_resize_lock(pgdat, &flags); in deferred_init_memmap()
2055 first_init_pfn = pgdat->first_deferred_pfn; in deferred_init_memmap()
2057 pgdat_resize_unlock(pgdat, &flags); in deferred_init_memmap()
2063 BUG_ON(pgdat->first_deferred_pfn < pgdat->node_start_pfn); in deferred_init_memmap()
2064 BUG_ON(pgdat->first_deferred_pfn > pgdat_end_pfn(pgdat)); in deferred_init_memmap()
[all …]
Dbootmem_info.c101 void __init register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument
104 int node = pgdat->node_id; in register_page_bootmem_info_node()
108 page = virt_to_page(pgdat); in register_page_bootmem_info_node()
113 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node()
114 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node()
Dpage_owner.c247 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showmixedcount_print() argument
320 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showmixedcount_print()
533 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument
597 pgdat->node_id, zone->name, count); in init_pages_in_zone()
600 static void init_zones_in_node(pg_data_t *pgdat) in init_zones_in_node() argument
603 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node()
609 init_pages_in_zone(pgdat, zone); in init_zones_in_node()
615 pg_data_t *pgdat; in init_early_allocated_pages() local
617 for_each_online_pgdat(pgdat) in init_early_allocated_pages()
618 init_zones_in_node(pgdat); in init_early_allocated_pages()
Dshuffle.c154 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory() argument
158 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
Dsparse.c321 static inline phys_addr_t pgdat_to_phys(struct pglist_data *pgdat) in pgdat_to_phys() argument
324 VM_BUG_ON(pgdat != &contig_page_data); in pgdat_to_phys()
327 return __pa(pgdat); in pgdat_to_phys()
333 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument
349 goal = pgdat_to_phys(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT); in sparse_early_usemaps_alloc_pgdat_section()
367 struct pglist_data *pgdat = NODE_DATA(nid); in check_usemap_section_nr() local
377 pgdat_snr = pfn_to_section_nr(pgdat_to_phys(pgdat) >> PAGE_SHIFT); in check_usemap_section_nr()
405 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument
408 return memblock_alloc_node(size, SMP_CACHE_BYTES, pgdat->node_id); in sparse_early_usemaps_alloc_pgdat_section()
Dmm_init.c31 pg_data_t *pgdat = NODE_DATA(nid); in mminit_verify_zonelist() local
43 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist()
44 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
/Linux-v5.15/include/linux/
Dmemory_hotplug.h172 extern void arch_free_nodedata(pg_data_t *pgdat);
173 extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat);
178 #define arch_free_nodedata(pgdat) generic_free_nodedata(pgdat) argument
195 #define generic_free_nodedata(pgdat) kfree(pgdat) argument
198 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument
200 node_data[nid] = pgdat; in arch_refresh_nodedata()
211 static inline void generic_free_nodedata(pg_data_t *pgdat) in generic_free_nodedata() argument
214 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument
276 void pgdat_resize_lock(struct pglist_data *pgdat, unsigned long *flags) in pgdat_resize_lock() argument
278 spin_lock_irqsave(&pgdat->node_size_lock, *flags); in pgdat_resize_lock()
[all …]
Dmmzone.h315 struct pglist_data *pgdat; member
908 #define pgdat_page_nr(pgdat, pagenr) ((pgdat)->node_mem_map + (pagenr)) argument
910 #define pgdat_page_nr(pgdat, pagenr) pfn_to_page((pgdat)->node_start_pfn + (pagenr)) argument
917 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() argument
919 return pgdat->node_start_pfn + pgdat->node_spanned_pages; in pgdat_end_pfn()
922 static inline bool pgdat_is_empty(pg_data_t *pgdat) in pgdat_is_empty() argument
924 return !pgdat->node_start_pfn && !pgdat->node_spanned_pages; in pgdat_is_empty()
929 void build_all_zonelists(pg_data_t *pgdat);
957 return lruvec->pgdat; in lruvec_pgdat()
1088 extern struct pglist_data *next_online_pgdat(struct pglist_data *pgdat);
[all …]
Dmemcontrol.h52 pg_data_t *pgdat; member
729 struct pglist_data *pgdat) in mem_cgroup_lruvec() argument
735 lruvec = &pgdat->__lruvec; in mem_cgroup_lruvec()
742 mz = memcg->nodeinfo[pgdat->node_id]; in mem_cgroup_lruvec()
750 if (unlikely(lruvec->pgdat != pgdat)) in mem_cgroup_lruvec()
751 lruvec->pgdat = pgdat; in mem_cgroup_lruvec()
763 pg_data_t *pgdat = page_pgdat(page); in mem_cgroup_page_lruvec() local
767 return mem_cgroup_lruvec(memcg, pgdat); in mem_cgroup_page_lruvec()
1109 unsigned long mem_cgroup_soft_limit_reclaim(pg_data_t *pgdat, int order,
1211 struct pglist_data *pgdat) in mem_cgroup_lruvec() argument
[all …]
Dvmstat.h172 static inline void node_page_state_add(long x, struct pglist_data *pgdat, in node_page_state_add() argument
175 atomic_long_add(x, &pgdat->vm_stat[item]); in node_page_state_add()
261 extern unsigned long node_page_state(struct pglist_data *pgdat,
263 extern unsigned long node_page_state_pages(struct pglist_data *pgdat,
311 void set_pgdat_percpu_threshold(pg_data_t *pgdat,
325 static inline void __mod_node_page_state(struct pglist_data *pgdat, in __mod_node_page_state() argument
339 node_page_state_add(delta, pgdat, item); in __mod_node_page_state()
348 static inline void __inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __inc_node_state() argument
350 atomic_long_inc(&pgdat->vm_stat[item]); in __inc_node_state()
360 static inline void __dec_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __dec_node_state() argument
[all …]
Dcompaction.h98 extern void reset_isolation_suitable(pg_data_t *pgdat);
182 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx);
185 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument
223 static inline void wakeup_kcompactd(pg_data_t *pgdat, in wakeup_kcompactd() argument
Dnode.h124 struct pglist_data *pgdat = NODE_DATA(nid); in register_one_node() local
125 unsigned long start_pfn = pgdat->node_start_pfn; in register_one_node()
126 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node()
Dbootmem_info.h20 void __init register_page_bootmem_info_node(struct pglist_data *pgdat);
47 static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument
Dpage_ext.h40 extern void pgdat_page_ext_init(struct pglist_data *pgdat);
70 static inline void pgdat_page_ext_init(struct pglist_data *pgdat) in pgdat_page_ext_init() argument
Dmm_inline.h31 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in update_lru_size() local
34 __mod_zone_page_state(&pgdat->node_zones[zid], in update_lru_size()
/Linux-v5.15/drivers/base/
Dnode.c374 struct pglist_data *pgdat = NODE_DATA(nid); in node_read_meminfo() local
380 sreclaimable = node_page_state_pages(pgdat, NR_SLAB_RECLAIMABLE_B); in node_read_meminfo()
381 sunreclaimable = node_page_state_pages(pgdat, NR_SLAB_UNRECLAIMABLE_B); in node_read_meminfo()
383 swapcached = node_page_state_pages(pgdat, NR_SWAPCACHE); in node_read_meminfo()
402 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON) + in node_read_meminfo()
403 node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo()
404 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON) + in node_read_meminfo()
405 node_page_state(pgdat, NR_INACTIVE_FILE)), in node_read_meminfo()
406 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON)), in node_read_meminfo()
407 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON)), in node_read_meminfo()
[all …]
/Linux-v5.15/lib/
Dshow_mem.c13 pg_data_t *pgdat; in show_mem() local
19 for_each_online_pgdat(pgdat) { in show_mem()
23 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem()
/Linux-v5.15/arch/ia64/include/asm/
Dnodedata.h57 #define LOCAL_DATA_ADDR(pgdat) \ argument
58 ((struct ia64_node_data *)((u64)(pgdat) + \

12