/Linux-v4.19/mm/ |
D | vmscan.c | 268 static void set_memcg_congestion(pg_data_t *pgdat, in set_memcg_congestion() argument 277 mn = mem_cgroup_nodeinfo(memcg, pgdat->node_id); in set_memcg_congestion() 281 static bool memcg_congested(pg_data_t *pgdat, in memcg_congested() argument 286 mn = mem_cgroup_nodeinfo(memcg, pgdat->node_id); in memcg_congested() 301 static inline void set_memcg_congestion(struct pglist_data *pgdat, in set_memcg_congestion() argument 306 static inline bool memcg_congested(struct pglist_data *pgdat, in memcg_congested() argument 1097 struct pglist_data *pgdat, in shrink_page_list() argument 1221 test_bit(PGDAT_WRITEBACK, &pgdat->flags)) { in shrink_page_list() 1343 !test_bit(PGDAT_DIRTY, &pgdat->flags))) { in shrink_page_list() 1794 static int too_many_isolated(struct pglist_data *pgdat, int file, in too_many_isolated() argument [all …]
|
D | nobootmem.c | 155 void reset_node_managed_pages(pg_data_t *pgdat) in reset_node_managed_pages() argument 159 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages() 165 struct pglist_data *pgdat; in reset_all_zones_managed_pages() local 170 for_each_online_pgdat(pgdat) in reset_all_zones_managed_pages() 171 reset_node_managed_pages(pgdat); in reset_all_zones_managed_pages() 203 void __init free_bootmem_node(pg_data_t *pgdat, unsigned long physaddr, in free_bootmem_node() argument 307 void * __init ___alloc_bootmem_node_nopanic(pg_data_t *pgdat, in ___alloc_bootmem_node_nopanic() argument 316 ptr = __alloc_memory_core_early(pgdat->node_id, size, align, in ___alloc_bootmem_node_nopanic() 334 void * __init __alloc_bootmem_node_nopanic(pg_data_t *pgdat, unsigned long size, in __alloc_bootmem_node_nopanic() argument 338 return kzalloc_node(size, GFP_NOWAIT, pgdat->node_id); in __alloc_bootmem_node_nopanic() [all …]
|
D | vmstat.c | 247 struct pglist_data *pgdat; in refresh_zone_stat_thresholds() local 253 for_each_online_pgdat(pgdat) { in refresh_zone_stat_thresholds() 255 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold = 0; in refresh_zone_stat_thresholds() 260 struct pglist_data *pgdat = zone->zone_pgdat; in refresh_zone_stat_thresholds() local 272 pgdat_threshold = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold; in refresh_zone_stat_thresholds() 273 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold in refresh_zone_stat_thresholds() 290 void set_pgdat_percpu_threshold(pg_data_t *pgdat, in set_pgdat_percpu_threshold() argument 298 for (i = 0; i < pgdat->nr_zones; i++) { in set_pgdat_percpu_threshold() 299 zone = &pgdat->node_zones[i]; in set_pgdat_percpu_threshold() 335 void __mod_node_page_state(struct pglist_data *pgdat, enum node_stat_item item, in __mod_node_page_state() argument [all …]
|
D | memory_hotplug.c | 220 void __init register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument 223 int node = pgdat->node_id; in register_page_bootmem_info_node() 227 page = virt_to_page(pgdat); in register_page_bootmem_info_node() 232 pfn = pgdat->node_start_pfn; in register_page_bootmem_info_node() 233 end_pfn = pgdat_end_pfn(pgdat); in register_page_bootmem_info_node() 439 static void shrink_pgdat_span(struct pglist_data *pgdat, in shrink_pgdat_span() argument 442 unsigned long pgdat_start_pfn = pgdat->node_start_pfn; in shrink_pgdat_span() 443 unsigned long p = pgdat_end_pfn(pgdat); /* pgdat_end_pfn namespace clash */ in shrink_pgdat_span() 447 int nid = pgdat->node_id; in shrink_pgdat_span() 459 pgdat->node_start_pfn = pfn; in shrink_pgdat_span() [all …]
|
D | bootmem.c | 175 unsigned long __init init_bootmem_node(pg_data_t *pgdat, unsigned long freepfn, in init_bootmem_node() argument 178 return init_bootmem_core(pgdat->bdata, freepfn, startpfn, endpfn); in init_bootmem_node() 286 void reset_node_managed_pages(pg_data_t *pgdat) in reset_node_managed_pages() argument 290 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) in reset_node_managed_pages() 296 struct pglist_data *pgdat; in reset_all_zones_managed_pages() local 301 for_each_online_pgdat(pgdat) in reset_all_zones_managed_pages() 302 reset_node_managed_pages(pgdat); in reset_all_zones_managed_pages() 423 void __init free_bootmem_node(pg_data_t *pgdat, unsigned long physaddr, in free_bootmem_node() argument 433 mark_bootmem_node(pgdat->bdata, start, end, 0, 0); in free_bootmem_node() 461 int __init reserve_bootmem_node(pg_data_t *pgdat, unsigned long physaddr, in reserve_bootmem_node() argument [all …]
|
D | mmzone.c | 18 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) in next_online_pgdat() argument 20 int nid = next_online_node(pgdat->node_id); in next_online_pgdat() 32 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() local 34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 37 pgdat = next_online_pgdat(pgdat); in next_zone() 38 if (pgdat) in next_zone() 39 zone = pgdat->node_zones; in next_zone()
|
D | compaction.c | 272 void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument 277 struct zone *zone = &pgdat->node_zones[zoneid]; in reset_isolation_suitable() 1817 pg_data_t *pgdat = NODE_DATA(nid); in compact_node() local 1833 zone = &pgdat->node_zones[zoneid]; in compact_node() 1915 static inline bool kcompactd_work_requested(pg_data_t *pgdat) in kcompactd_work_requested() argument 1917 return pgdat->kcompactd_max_order > 0 || kthread_should_stop(); in kcompactd_work_requested() 1920 static bool kcompactd_node_suitable(pg_data_t *pgdat) in kcompactd_node_suitable() argument 1924 enum zone_type classzone_idx = pgdat->kcompactd_classzone_idx; in kcompactd_node_suitable() 1927 zone = &pgdat->node_zones[zoneid]; in kcompactd_node_suitable() 1932 if (compaction_suitable(zone, pgdat->kcompactd_max_order, 0, in kcompactd_node_suitable() [all …]
|
D | page_alloc.c | 312 static inline bool update_defer_init(pg_data_t *pgdat, in update_defer_init() argument 317 if (zone_end < pgdat_end_pfn(pgdat)) in update_defer_init() 320 if ((*nr_initialised > pgdat->static_init_pgcnt) && in update_defer_init() 322 pgdat->first_deferred_pfn = pfn; in update_defer_init() 334 static inline bool update_defer_init(pg_data_t *pgdat, in update_defer_init() argument 1191 pg_data_t *pgdat; in init_reserved_page() local 1198 pgdat = NODE_DATA(nid); in init_reserved_page() 1201 struct zone *zone = &pgdat->node_zones[zid]; in init_reserved_page() 1532 pg_data_t *pgdat = data; in deferred_init_memmap() local 1533 int nid = pgdat->node_id; in deferred_init_memmap() [all …]
|
D | workingset.c | 173 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction) in pack_shadow() argument 177 eviction = (eviction << NODES_SHIFT) | pgdat->node_id; in pack_shadow() 183 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow() argument 196 *pgdat = NODE_DATA(nid); in unpack_shadow() 211 struct pglist_data *pgdat = page_pgdat(page); in workingset_eviction() local 221 lruvec = mem_cgroup_lruvec(pgdat, memcg); in workingset_eviction() 223 return pack_shadow(memcgid, pgdat, eviction); in workingset_eviction() 243 struct pglist_data *pgdat; in workingset_refault() local 246 unpack_shadow(shadow, &memcgid, &pgdat, &eviction); in workingset_refault() 270 lruvec = mem_cgroup_lruvec(pgdat, memcg); in workingset_refault()
|
D | sparse.c | 289 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument 305 goal = __pa(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT); in sparse_early_usemaps_alloc_pgdat_section() 324 struct pglist_data *pgdat = NODE_DATA(nid); in check_usemap_section_nr() local 334 pgdat_snr = pfn_to_section_nr(__pa(pgdat) >> PAGE_SHIFT); in check_usemap_section_nr() 362 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, in sparse_early_usemaps_alloc_pgdat_section() argument 365 return memblock_virt_alloc_node_nopanic(size, pgdat->node_id); in sparse_early_usemaps_alloc_pgdat_section() 664 int __meminit sparse_add_one_section(struct pglist_data *pgdat, in sparse_add_one_section() argument 678 ret = sparse_index_init(section_nr, pgdat->node_id); in sparse_add_one_section() 682 memmap = kmalloc_section_memmap(section_nr, pgdat->node_id, altmap); in sparse_add_one_section() 691 pgdat_resize_lock(pgdat, &flags); in sparse_add_one_section() [all …]
|
D | page_owner.c | 256 pg_data_t *pgdat, struct zone *zone) in pagetypeinfo_showmixedcount_print() argument 332 seq_printf(m, "Node %d, zone %8s ", pgdat->node_id, zone->name); in pagetypeinfo_showmixedcount_print() 529 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument 597 pgdat->node_id, zone->name, count); in init_pages_in_zone() 600 static void init_zones_in_node(pg_data_t *pgdat) in init_zones_in_node() argument 603 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node() 609 init_pages_in_zone(pgdat, zone); in init_zones_in_node() 615 pg_data_t *pgdat; in init_early_allocated_pages() local 617 for_each_online_pgdat(pgdat) in init_early_allocated_pages() 618 init_zones_in_node(pgdat); in init_early_allocated_pages()
|
D | swap.c | 194 struct pglist_data *pgdat = NULL; in pagevec_lru_move_fn() local 202 if (pagepgdat != pgdat) { in pagevec_lru_move_fn() 203 if (pgdat) in pagevec_lru_move_fn() 204 spin_unlock_irqrestore(&pgdat->lru_lock, flags); in pagevec_lru_move_fn() 205 pgdat = pagepgdat; in pagevec_lru_move_fn() 206 spin_lock_irqsave(&pgdat->lru_lock, flags); in pagevec_lru_move_fn() 209 lruvec = mem_cgroup_page_lruvec(page, pgdat); in pagevec_lru_move_fn() 212 if (pgdat) in pagevec_lru_move_fn() 213 spin_unlock_irqrestore(&pgdat->lru_lock, flags); in pagevec_lru_move_fn() 765 struct pglist_data *pgdat = page_pgdat(page); in release_pages() local [all …]
|
/Linux-v4.19/include/linux/ |
D | memory_hotplug.h | 153 extern void arch_free_nodedata(pg_data_t *pgdat); 154 extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat); 159 #define arch_free_nodedata(pgdat) generic_free_nodedata(pgdat) argument 176 #define generic_free_nodedata(pgdat) kfree(pgdat) argument 179 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument 181 node_data[nid] = pgdat; in arch_refresh_nodedata() 192 static inline void generic_free_nodedata(pg_data_t *pgdat) in generic_free_nodedata() argument 195 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) in arch_refresh_nodedata() argument 202 extern void __init register_page_bootmem_info_node(struct pglist_data *pgdat); 204 static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) in register_page_bootmem_info_node() argument [all …]
|
D | bootmem.h | 59 extern unsigned long init_bootmem_node(pg_data_t *pgdat, 66 extern void reset_node_managed_pages(pg_data_t *pgdat); 69 extern void free_bootmem_node(pg_data_t *pgdat, 89 extern int reserve_bootmem_node(pg_data_t *pgdat, 100 extern void *__alloc_bootmem_node(pg_data_t *pgdat, 104 void *__alloc_bootmem_node_high(pg_data_t *pgdat, 108 extern void *__alloc_bootmem_node_nopanic(pg_data_t *pgdat, 112 void *___alloc_bootmem_node_nopanic(pg_data_t *pgdat, 123 extern void *__alloc_bootmem_low_node(pg_data_t *pgdat, 149 #define alloc_bootmem_node(pgdat, x) \ argument [all …]
|
D | mmzone.h | 244 struct pglist_data *pgdat; member 720 #define pgdat_page_nr(pgdat, pagenr) ((pgdat)->node_mem_map + (pagenr)) argument 722 #define pgdat_page_nr(pgdat, pagenr) pfn_to_page((pgdat)->node_start_pfn + (pagenr)) argument 733 static inline struct lruvec *node_lruvec(struct pglist_data *pgdat) in node_lruvec() argument 735 return &pgdat->lruvec; in node_lruvec() 738 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) in pgdat_end_pfn() argument 740 return pgdat->node_start_pfn + pgdat->node_spanned_pages; in pgdat_end_pfn() 743 static inline bool pgdat_is_empty(pg_data_t *pgdat) in pgdat_is_empty() argument 745 return !pgdat->node_start_pfn && !pgdat->node_spanned_pages; in pgdat_is_empty() 750 void build_all_zonelists(pg_data_t *pgdat); [all …]
|
D | vmstat.h | 170 static inline void node_page_state_add(long x, struct pglist_data *pgdat, in node_page_state_add() argument 173 atomic_long_add(x, &pgdat->vm_stat[item]); in node_page_state_add() 235 extern unsigned long node_page_state(struct pglist_data *pgdat, 283 void set_pgdat_percpu_threshold(pg_data_t *pgdat, 297 static inline void __mod_node_page_state(struct pglist_data *pgdat, in __mod_node_page_state() argument 300 node_page_state_add(delta, pgdat, item); in __mod_node_page_state() 309 static inline void __inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __inc_node_state() argument 311 atomic_long_inc(&pgdat->vm_stat[item]); in __inc_node_state() 321 static inline void __dec_node_state(struct pglist_data *pgdat, enum node_stat_item item) in __dec_node_state() argument 323 atomic_long_dec(&pgdat->vm_stat[item]); in __dec_node_state() [all …]
|
D | compaction.h | 99 extern void reset_isolation_suitable(pg_data_t *pgdat); 179 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int classzone_idx); 182 static inline void reset_isolation_suitable(pg_data_t *pgdat) in reset_isolation_suitable() argument 224 static inline void wakeup_kcompactd(pg_data_t *pgdat, int order, int classzone_idx) in wakeup_kcompactd() argument
|
D | memcontrol.h | 69 pg_data_t *pgdat; member 369 static inline struct lruvec *mem_cgroup_lruvec(struct pglist_data *pgdat, in mem_cgroup_lruvec() argument 376 lruvec = node_lruvec(pgdat); in mem_cgroup_lruvec() 380 mz = mem_cgroup_nodeinfo(memcg, pgdat->node_id); in mem_cgroup_lruvec() 388 if (unlikely(lruvec->pgdat != pgdat)) in mem_cgroup_lruvec() 389 lruvec->pgdat = pgdat; in mem_cgroup_lruvec() 690 pg_data_t *pgdat = page_pgdat(page); in __mod_lruvec_page_state() local 695 __mod_node_page_state(pgdat, idx, val); in __mod_lruvec_page_state() 699 lruvec = mem_cgroup_lruvec(pgdat, page->mem_cgroup); in __mod_lruvec_page_state() 713 unsigned long mem_cgroup_soft_limit_reclaim(pg_data_t *pgdat, int order, [all …]
|
D | node.h | 56 struct pglist_data *pgdat = NODE_DATA(nid); in register_one_node() local 57 unsigned long start_pfn = pgdat->node_start_pfn; in register_one_node() 58 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages; in register_one_node()
|
D | page_ext.h | 39 extern void pgdat_page_ext_init(struct pglist_data *pgdat); 58 static inline void pgdat_page_ext_init(struct pglist_data *pgdat) in pgdat_page_ext_init() argument
|
D | mm_inline.h | 30 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in __update_lru_size() local 32 __mod_node_page_state(pgdat, NR_LRU_BASE + lru, nr_pages); in __update_lru_size() 33 __mod_zone_page_state(&pgdat->node_zones[zid], in __update_lru_size()
|
/Linux-v4.19/drivers/base/ |
D | node.c | 68 struct pglist_data *pgdat = NODE_DATA(nid); in node_read_meminfo() local 87 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON) + in node_read_meminfo() 88 node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo() 89 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON) + in node_read_meminfo() 90 node_page_state(pgdat, NR_INACTIVE_FILE)), in node_read_meminfo() 91 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON)), in node_read_meminfo() 92 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON)), in node_read_meminfo() 93 nid, K(node_page_state(pgdat, NR_ACTIVE_FILE)), in node_read_meminfo() 94 nid, K(node_page_state(pgdat, NR_INACTIVE_FILE)), in node_read_meminfo() 95 nid, K(node_page_state(pgdat, NR_UNEVICTABLE)), in node_read_meminfo() [all …]
|
/Linux-v4.19/arch/m68k/include/asm/ |
D | page_mm.h | 158 struct pglist_data *pgdat; \ 159 pgdat = __virt_to_node((unsigned long)pfn_to_virt(__pfn)); \ 160 pgdat->node_mem_map + (__pfn - pgdat->node_start_pfn); \ 164 struct pglist_data *pgdat; \ 165 pgdat = &pg_data_map[page_to_nid(__p)]; \ 166 ((__p) - pgdat->node_mem_map) + pgdat->node_start_pfn; \
|
/Linux-v4.19/lib/ |
D | show_mem.c | 14 pg_data_t *pgdat; in show_mem() local 20 for_each_online_pgdat(pgdat) { in show_mem() 24 pgdat_resize_lock(pgdat, &flags); in show_mem() 26 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem() 36 pgdat_resize_unlock(pgdat, &flags); in show_mem()
|
/Linux-v4.19/arch/ia64/include/asm/ |
D | nodedata.h | 57 #define LOCAL_DATA_ADDR(pgdat) \ argument 58 ((struct ia64_node_data *)((u64)(pgdat) + \
|