Lines Matching refs:lruvec

763 void __mod_memcg_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx,  in __mod_memcg_lruvec_state()  argument
769 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in __mod_memcg_lruvec_state()
813 void __mod_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, in __mod_lruvec_state() argument
817 __mod_node_page_state(lruvec_pgdat(lruvec), idx, val); in __mod_lruvec_state()
821 __mod_memcg_lruvec_state(lruvec, idx, val); in __mod_lruvec_state()
830 struct lruvec *lruvec; in __mod_lruvec_page_state() local
841 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __mod_lruvec_page_state()
842 __mod_lruvec_state(lruvec, idx, val); in __mod_lruvec_page_state()
851 struct lruvec *lruvec; in __mod_lruvec_kmem_state() local
865 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __mod_lruvec_kmem_state()
866 __mod_lruvec_state(lruvec, idx, val); in __mod_lruvec_kmem_state()
1265 void lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio) in lruvec_memcg_debug() argument
1275 VM_BUG_ON_FOLIO(lruvec_memcg(lruvec) != root_mem_cgroup, folio); in lruvec_memcg_debug()
1277 VM_BUG_ON_FOLIO(lruvec_memcg(lruvec) != memcg, folio); in lruvec_memcg_debug()
1293 struct lruvec *folio_lruvec_lock(struct folio *folio) in folio_lruvec_lock()
1295 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock() local
1297 spin_lock(&lruvec->lru_lock); in folio_lruvec_lock()
1298 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock()
1300 return lruvec; in folio_lruvec_lock()
1316 struct lruvec *folio_lruvec_lock_irq(struct folio *folio) in folio_lruvec_lock_irq()
1318 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irq() local
1320 spin_lock_irq(&lruvec->lru_lock); in folio_lruvec_lock_irq()
1321 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock_irq()
1323 return lruvec; in folio_lruvec_lock_irq()
1340 struct lruvec *folio_lruvec_lock_irqsave(struct folio *folio, in folio_lruvec_lock_irqsave()
1343 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irqsave() local
1345 spin_lock_irqsave(&lruvec->lru_lock, *flags); in folio_lruvec_lock_irqsave()
1346 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock_irqsave()
1348 return lruvec; in folio_lruvec_lock_irqsave()
1361 void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru, in mem_cgroup_update_lru_size() argument
1371 mz = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in mem_cgroup_update_lru_size()
1380 __func__, lruvec, lru, nr_pages, size)) { in mem_cgroup_update_lru_size()
2868 struct lruvec *lruvec; in mod_objcg_mlstate() local
2872 lruvec = mem_cgroup_lruvec(memcg, pgdat); in mod_objcg_mlstate()
2873 mod_memcg_lruvec_state(lruvec, idx, nr); in mod_objcg_mlstate()
3942 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(nid)); in mem_cgroup_node_nr_lru_pages() local
3952 nr += lruvec_page_state(lruvec, NR_LRU_BASE + lru); in mem_cgroup_node_nr_lru_pages()
3954 nr += lruvec_page_state_local(lruvec, NR_LRU_BASE + lru); in mem_cgroup_node_nr_lru_pages()
4133 anon_cost += mz->lruvec.anon_cost; in memcg_stat_show()
4134 file_cost += mz->lruvec.file_cost; in memcg_stat_show()
5214 lruvec_init(&pn->lruvec); in alloc_mem_cgroup_per_node_info()
5691 struct lruvec *from_vec, *to_vec; in mem_cgroup_move_account()
6516 static inline unsigned long lruvec_page_state_output(struct lruvec *lruvec, in lruvec_page_state_output() argument
6519 return lruvec_page_state(lruvec, item) * memcg_page_state_unit(item); in lruvec_page_state_output()
6538 struct lruvec *lruvec; in memory_numa_stat_show() local
6540 lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(nid)); in memory_numa_stat_show()
6541 size = lruvec_page_state_output(lruvec, in memory_numa_stat_show()