/Linux-v5.4/mm/ |
D | vmscan.c | 273 static void set_memcg_congestion(pg_data_t *pgdat, in set_memcg_congestion() argument 282 mn = mem_cgroup_nodeinfo(memcg, pgdat->node_id); in set_memcg_congestion() 286 static bool memcg_congested(pg_data_t *pgdat, in memcg_congested() argument 291 mn = mem_cgroup_nodeinfo(memcg, pgdat->node_id); in memcg_congested() 315 static inline void set_memcg_congestion(struct pglist_data *pgdat, in set_memcg_congestion() argument 320 static inline bool memcg_congested(struct pglist_data *pgdat, in memcg_congested() argument 1120 struct pglist_data *pgdat, in shrink_page_list() argument 1237 test_bit(PGDAT_WRITEBACK, &pgdat->flags)) { in shrink_page_list() 1371 !test_bit(PGDAT_DIRTY, &pgdat->flags))) { in shrink_page_list() 1804 pg_data_t *pgdat = page_pgdat(page); in isolate_lru_page() local [all …]
|
D | vmstat.c | 248 struct pglist_data *pgdat; in refresh_zone_stat_thresholds() local 254 for_each_online_pgdat(pgdat) { in refresh_zone_stat_thresholds() 256 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold = 0; in refresh_zone_stat_thresholds() 261 struct pglist_data *pgdat = zone->zone_pgdat; in refresh_zone_stat_thresholds() local 273 pgdat_threshold = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold; in refresh_zone_stat_thresholds() 274 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold in refresh_zone_stat_thresholds() 291 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() argument 299 for (i = 0; i < pgdat->nr_zones; i++) { in set_pgdat_percpu_threshold() 300 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold() 336 void __mod_node_page_state(struct pglist_data *pgdat, enum node_stat_item item, in __mod_node_page_state() argument [all …]
|
D | memory_hotplug.c | 225 void __init register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument 228 int node = pgdat->node_id; in register_page_bootmem_info_node() 232 page = virt_to_page(pgdat); in register_page_bootmem_info_node() 237 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node() 238 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node() 439 static void update_pgdat_span(struct pglist_data *pgdat) in update_pgdat_span() argument 444 for (zone = pgdat->node_zones; in update_pgdat_span() 445 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { in update_pgdat_span() 464 pgdat->node_start_pfn = node_start_pfn; in update_pgdat_span() 465 pgdat->node_spanned_pages = node_end_pfn - node_start_pfn; in update_pgdat_span() [all …]
|
D | mmzone.c | 18 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) in next_online_pgdat() argument 20 int nid = next_online_node(pgdat->node_id); in next_online_pgdat() 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() local 34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 37 pgdat = next_online_pgdat(pgdat); in next_zone() 38 if (pgdat) in next_zone() 39 zone = pgdat->node_zones; in next_zone()
|
D | compaction.c | 369 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument 374 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable() 749 static bool too_many_isolated(pg_data_t *pgdat) in too_many_isolated() argument 753 inactive = node_page_state(pgdat, NR_INACTIVE_FILE) + in too_many_isolated() 754 node_page_state(pgdat, NR_INACTIVE_ANON); in too_many_isolated() 755 active = node_page_state(pgdat, NR_ACTIVE_FILE) + in too_many_isolated() 756 node_page_state(pgdat, NR_ACTIVE_ANON); in too_many_isolated() 757 isolated = node_page_state(pgdat, NR_ISOLATED_FILE) + in too_many_isolated() 758 node_page_state(pgdat, NR_ISOLATED_ANON); in too_many_isolated() 785 pg_data_t *pgdat = cc->zone->zone_pgdat; in isolate_migratepages_block() local [all …]
|
D | page_alloc.c | 1368 pg_data_t *pgdat; in init_reserved_page() local 1375 pgdat = NODE_DATA(nid); in init_reserved_page() 1378 struct zone *zone = &pgdat->node_zones[zid]; in init_reserved_page() 1776 pg_data_t *pgdat = data; in deferred_init_memmap() local 1777 const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id); in deferred_init_memmap() 1789 pgdat_resize_lock(pgdat, &flags); in deferred_init_memmap() 1790 first_init_pfn = pgdat->first_deferred_pfn; in deferred_init_memmap() 1792 pgdat_resize_unlock(pgdat, &flags); in deferred_init_memmap() 1798 BUG_ON(pgdat->first_deferred_pfn < pgdat->node_start_pfn); in deferred_init_memmap() 1799 BUG_ON(pgdat->first_deferred_pfn > pgdat_end_pfn(pgdat)); in deferred_init_memmap() [all …]
|
D | workingset.c | 184 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, in pack_shadow() argument 190 eviction = (eviction << NODES_SHIFT) | pgdat->node_id; in pack_shadow() 196 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow() argument 211 *pgdat = NODE_DATA(nid); in unpack_shadow() 225 struct pglist_data *pgdat = page_pgdat(page); in workingset_eviction() local 236 lruvec = mem_cgroup_lruvec(pgdat, memcg); in workingset_eviction() 238 return pack_shadow(memcgid, pgdat, eviction, PageWorkingset(page)); in workingset_eviction() 252 struct pglist_data *pgdat; in workingset_refault() local 261 unpack_shadow(shadow, &memcgid, &pgdat, &eviction, &workingset); in workingset_refault() 283 lruvec = mem_cgroup_lruvec(pgdat, memcg); in workingset_refault()
|
D | shuffle.h | 24 extern void __shuffle_free_memory(pg_data_t *pgdat); 25 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument 29 __shuffle_free_memory(pgdat); in shuffle_free_memory() 47 static inline void shuffle_free_memory(pg_data_t *pgdat) in shuffle_free_memory() argument
|
D | swap.c | 63 pg_data_t *pgdat = page_pgdat(page); in __page_cache_release() local 67 spin_lock_irqsave(&pgdat->lru_lock, flags); in __page_cache_release() 68 lruvec = mem_cgroup_page_lruvec(page, pgdat); in __page_cache_release() 72 spin_unlock_irqrestore(&pgdat->lru_lock, flags); in __page_cache_release() 195 struct pglist_data *pgdat = NULL; in pagevec_lru_move_fn() local 203 if (pagepgdat != pgdat) { in pagevec_lru_move_fn() 204 if (pgdat) in pagevec_lru_move_fn() 205 spin_unlock_irqrestore(&pgdat->lru_lock, flags); in pagevec_lru_move_fn() 206 pgdat = pagepgdat; in pagevec_lru_move_fn() 207 spin_lock_irqsave(&pgdat->lru_lock, flags); in pagevec_lru_move_fn() [all …]
|
D | page_owner.c | 254 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showmixedcount_print() argument 331 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showmixedcount_print() 546 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument 615 pgdat->node_id, zone->name, count); in init_pages_in_zone() 618 static void init_zones_in_node(pg_data_t *pgdat) in init_zones_in_node() argument 621 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node() 627 init_pages_in_zone(pgdat, zone); in init_zones_in_node() 633 pg_data_t *pgdat; in init_early_allocated_pages() local 635 for_each_online_pgdat(pgdat) in init_early_allocated_pages() 636 init_zones_in_node(pgdat); in init_early_allocated_pages()
|
D | page_idle.c | 34 pg_data_t *pgdat; in page_idle_get_page() local 44 pgdat = page_pgdat(page); in page_idle_get_page() 45 spin_lock_irq(&pgdat->lru_lock); in page_idle_get_page() 50 spin_unlock_irq(&pgdat->lru_lock); in page_idle_get_page()
|
D | mm_init.c | 34 pg_data_t *pgdat = NODE_DATA(nid); in mminit_verify_zonelist() local 46 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 47 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
|
D | shuffle.c | 178 void __meminit __shuffle_free_memory(pg_data_t *pgdat) in __shuffle_free_memory() argument 182 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in __shuffle_free_memory()
|
/Linux-v5.4/include/linux/ |
D | memory_hotplug.h | 167 extern void arch_free_nodedata(pg_data_t *pgdat); 168 extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat); 173 #define arch_free_nodedata(pgdat) generic_free_nodedata(pgdat) argument 190 #define generic_free_nodedata(pgdat) kfree(pgdat) argument 193 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument 195 node_data[nid] = pgdat; in arch_refresh_nodedata() 206 static inline void generic_free_nodedata(pg_data_t *pgdat) in generic_free_nodedata() argument 209 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument 216 extern void __init register_page_bootmem_info_node(struct pglist_data *pgdat); 218 static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument [all …]
|
D | vmstat.h | 170 static inline void node_page_state_add(long x, struct pglist_data *pgdat, in node_page_state_add() argument 173 atomic_long_add(x, &pgdat->vm_stat[item]); in node_page_state_add() 235 extern unsigned long node_page_state(struct pglist_data *pgdat, 278 void set_pgdat_percpu_threshold(pg_data_t *pgdat, 292 static inline void __mod_node_page_state(struct pglist_data *pgdat, in __mod_node_page_state() argument 295 node_page_state_add(delta, pgdat, item); in __mod_node_page_state() 304 static inline void __inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __inc_node_state() argument 306 atomic_long_inc(&pgdat->vm_stat[item]); in __inc_node_state() 316 static inline void __dec_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __dec_node_state() argument 318 atomic_long_dec(&pgdat->vm_stat[item]); in __dec_node_state() [all …]
|
D | mmzone.h | 307 struct pglist_data *pgdat; member 787 #define pgdat_page_nr(pgdat, pagenr) ((pgdat)->node_mem_map + (pagenr)) argument 789 #define pgdat_page_nr(pgdat, pagenr) pfn_to_page((pgdat)->node_start_pfn + (pagenr)) argument 796 static inline struct lruvec *node_lruvec(struct pglist_data *pgdat) in node_lruvec() argument 798 return &pgdat->lruvec; in node_lruvec() 801 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() argument 803 return pgdat->node_start_pfn + pgdat->node_spanned_pages; in pgdat_end_pfn() 806 static inline bool pgdat_is_empty(pg_data_t *pgdat) in pgdat_is_empty() argument 808 return !pgdat->node_start_pfn && !pgdat->node_spanned_pages; in pgdat_is_empty() 813 void build_all_zonelists(pg_data_t *pgdat); [all …]
|
D | compaction.h | 98 extern void reset_isolation_suitable(pg_data_t *pgdat); 185 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int classzone_idx); 188 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument 235 static inline void wakeup_kcompactd(pg_data_t *pgdat, int order, int classzone_idx) in wakeup_kcompactd() argument
|
D | memcontrol.h | 60 pg_data_t *pgdat; member 405 static inline struct lruvec *mem_cgroup_lruvec(struct pglist_data *pgdat, in mem_cgroup_lruvec() argument 412 lruvec = node_lruvec(pgdat); in mem_cgroup_lruvec() 416 mz = mem_cgroup_nodeinfo(memcg, pgdat->node_id); in mem_cgroup_lruvec() 424 if (unlikely(lruvec->pgdat != pgdat)) in mem_cgroup_lruvec() 425 lruvec->pgdat = pgdat; in mem_cgroup_lruvec() 722 pg_data_t *pgdat = page_pgdat(page); in __mod_lruvec_page_state() local 727 __mod_node_page_state(pgdat, idx, val); in __mod_lruvec_page_state() 731 lruvec = mem_cgroup_lruvec(pgdat, page->mem_cgroup); in __mod_lruvec_page_state() 745 unsigned long mem_cgroup_soft_limit_reclaim(pg_data_t *pgdat, int order, [all …]
|
D | node.h | 123 struct pglist_data *pgdat = NODE_DATA(nid); in register_one_node() local 124 unsigned long start_pfn = pgdat->node_start_pfn; in register_one_node() 125 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node()
|
D | page_ext.h | 40 extern void pgdat_page_ext_init(struct pglist_data *pgdat); 66 static inline void pgdat_page_ext_init(struct pglist_data *pgdat) in pgdat_page_ext_init() argument
|
/Linux-v5.4/arch/m68k/include/asm/ |
D | page_mm.h | 158 struct pglist_data *pgdat; \ 159 pgdat = __virt_to_node((unsigned long)pfn_to_virt(__pfn)); \ 160 pgdat->node_mem_map + (__pfn - pgdat->node_start_pfn); \ 164 struct pglist_data *pgdat; \ 165 pgdat = &pg_data_map[page_to_nid(__p)]; \ 166 ((__p) - pgdat->node_mem_map) + pgdat->node_start_pfn; \
|
/Linux-v5.4/drivers/base/ |
D | node.c | 366 struct pglist_data *pgdat = NODE_DATA(nid); in node_read_meminfo() local 371 sreclaimable = node_page_state(pgdat, NR_SLAB_RECLAIMABLE); in node_read_meminfo() 372 sunreclaimable = node_page_state(pgdat, NR_SLAB_UNRECLAIMABLE); in node_read_meminfo() 388 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON) + in node_read_meminfo() 389 node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo() 390 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON) + in node_read_meminfo() 391 node_page_state(pgdat, NR_INACTIVE_FILE)), in node_read_meminfo() 392 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON)), in node_read_meminfo() 393 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON)), in node_read_meminfo() 394 nid, K(node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo() [all …]
|
/Linux-v5.4/lib/ |
D | show_mem.c | 13 pg_data_t *pgdat; in show_mem() local 19 for_each_online_pgdat(pgdat) { in show_mem() 23 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem()
|
/Linux-v5.4/arch/ia64/include/asm/ |
D | nodedata.h | 57 #define LOCAL_DATA_ADDR(pgdat) \ argument 58 ((struct ia64_node_data *)((u64)(pgdat) + \
|
/Linux-v5.4/arch/c6x/mm/ |
D | init.c | 35 struct pglist_data *pgdat = NODE_DATA(0); in paging_init() local 53 pgdat->node_zones[ZONE_NORMAL].zone_start_pfn = in paging_init()
|