Lines Matching refs:lruvec
791 void __mod_memcg_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, in __mod_memcg_lruvec_state() argument
797 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in __mod_memcg_lruvec_state()
841 void __mod_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, in __mod_lruvec_state() argument
845 __mod_node_page_state(lruvec_pgdat(lruvec), idx, val); in __mod_lruvec_state()
849 __mod_memcg_lruvec_state(lruvec, idx, val); in __mod_lruvec_state()
858 struct lruvec *lruvec; in __mod_lruvec_page_state() local
869 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __mod_lruvec_page_state()
870 __mod_lruvec_state(lruvec, idx, val); in __mod_lruvec_page_state()
879 struct lruvec *lruvec; in __mod_lruvec_kmem_state() local
893 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __mod_lruvec_kmem_state()
894 __mod_lruvec_state(lruvec, idx, val); in __mod_lruvec_kmem_state()
1288 void lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio) in lruvec_memcg_debug() argument
1298 VM_BUG_ON_FOLIO(!mem_cgroup_is_root(lruvec_memcg(lruvec)), folio); in lruvec_memcg_debug()
1300 VM_BUG_ON_FOLIO(lruvec_memcg(lruvec) != memcg, folio); in lruvec_memcg_debug()
1316 struct lruvec *folio_lruvec_lock(struct folio *folio) in folio_lruvec_lock()
1318 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock() local
1320 spin_lock(&lruvec->lru_lock); in folio_lruvec_lock()
1321 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock()
1323 return lruvec; in folio_lruvec_lock()
1339 struct lruvec *folio_lruvec_lock_irq(struct folio *folio) in folio_lruvec_lock_irq()
1341 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irq() local
1343 spin_lock_irq(&lruvec->lru_lock); in folio_lruvec_lock_irq()
1344 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock_irq()
1346 return lruvec; in folio_lruvec_lock_irq()
1363 struct lruvec *folio_lruvec_lock_irqsave(struct folio *folio, in folio_lruvec_lock_irqsave()
1366 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irqsave() local
1368 spin_lock_irqsave(&lruvec->lru_lock, *flags); in folio_lruvec_lock_irqsave()
1369 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock_irqsave()
1371 return lruvec; in folio_lruvec_lock_irqsave()
1384 void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru, in mem_cgroup_update_lru_size() argument
1394 mz = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in mem_cgroup_update_lru_size()
1403 __func__, lruvec, lru, nr_pages, size)) { in mem_cgroup_update_lru_size()
2878 struct lruvec *lruvec; in mod_objcg_mlstate() local
2882 lruvec = mem_cgroup_lruvec(memcg, pgdat); in mod_objcg_mlstate()
2883 mod_memcg_lruvec_state(lruvec, idx, nr); in mod_objcg_mlstate()
3976 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(nid)); in mem_cgroup_node_nr_lru_pages() local
3986 nr += lruvec_page_state(lruvec, NR_LRU_BASE + lru); in mem_cgroup_node_nr_lru_pages()
3988 nr += lruvec_page_state_local(lruvec, NR_LRU_BASE + lru); in mem_cgroup_node_nr_lru_pages()
4166 anon_cost += mz->lruvec.anon_cost; in memcg1_stat_format()
4167 file_cost += mz->lruvec.file_cost; in memcg1_stat_format()
5249 lruvec_init(&pn->lruvec); in alloc_mem_cgroup_per_node_info()
5768 struct lruvec *from_vec, *to_vec; in mem_cgroup_move_account()
6622 static inline unsigned long lruvec_page_state_output(struct lruvec *lruvec, in lruvec_page_state_output() argument
6625 return lruvec_page_state(lruvec, item) * memcg_page_state_unit(item); in lruvec_page_state_output()
6644 struct lruvec *lruvec; in memory_numa_stat_show() local
6646 lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(nid)); in memory_numa_stat_show()
6647 size = lruvec_page_state_output(lruvec, in memory_numa_stat_show()