Lines Matching refs:lruvec

315 unsigned long lruvec_lru_size(struct lruvec *lruvec, enum lru_list lru, int zone_idx)  in lruvec_lru_size()  argument
321 struct zone *zone = &lruvec_pgdat(lruvec)->node_zones[zid]; in lruvec_lru_size()
327 size += mem_cgroup_get_zone_lru_size(lruvec, lru, zid); in lruvec_lru_size()
1616 static __always_inline void update_lru_sizes(struct lruvec *lruvec, in update_lru_sizes() argument
1625 update_lru_size(lruvec, lru, zid, -nr_zone_taken[zid]); in update_lru_sizes()
1650 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_pages() argument
1654 struct list_head *src = &lruvec->lists[lru]; in isolate_lru_pages()
1732 update_lru_sizes(lruvec, lru, nr_zone_taken); in isolate_lru_pages()
1771 struct lruvec *lruvec; in isolate_lru_page() local
1774 lruvec = mem_cgroup_page_lruvec(page, pgdat); in isolate_lru_page()
1779 del_page_from_lru_list(page, lruvec, lru); in isolate_lru_page()
1844 static unsigned noinline_for_stack move_pages_to_lru(struct lruvec *lruvec, in move_pages_to_lru() argument
1847 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in move_pages_to_lru()
1863 lruvec = mem_cgroup_page_lruvec(page, pgdat); in move_pages_to_lru()
1869 update_lru_size(lruvec, lru, page_zonenum(page), nr_pages); in move_pages_to_lru()
1870 list_move(&page->lru, &lruvec->lists[lru]); in move_pages_to_lru()
1875 del_page_from_lru_list(page, lruvec, lru); in move_pages_to_lru()
1886 workingset_age_nonresident(lruvec, nr_pages); in move_pages_to_lru()
1916 shrink_inactive_list(unsigned long nr_to_scan, struct lruvec *lruvec, in shrink_inactive_list() argument
1926 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in shrink_inactive_list()
1946 nr_taken = isolate_lru_pages(nr_to_scan, lruvec, &page_list, in shrink_inactive_list()
1953 __count_memcg_events(lruvec_memcg(lruvec), item, nr_scanned); in shrink_inactive_list()
1966 move_pages_to_lru(lruvec, &page_list); in shrink_inactive_list()
1969 lru_note_cost(lruvec, file, stat.nr_pageout); in shrink_inactive_list()
1973 __count_memcg_events(lruvec_memcg(lruvec), item, nr_reclaimed); in shrink_inactive_list()
2010 struct lruvec *lruvec, in shrink_active_list() argument
2024 struct pglist_data *pgdat = lruvec_pgdat(lruvec); in shrink_active_list()
2030 nr_taken = isolate_lru_pages(nr_to_scan, lruvec, &l_hold, in shrink_active_list()
2037 __count_memcg_events(lruvec_memcg(lruvec), PGREFILL, nr_scanned); in shrink_active_list()
2087 nr_activate = move_pages_to_lru(lruvec, &l_active); in shrink_active_list()
2088 nr_deactivate = move_pages_to_lru(lruvec, &l_inactive); in shrink_active_list()
2093 __count_memcg_events(lruvec_memcg(lruvec), PGDEACTIVATE, nr_deactivate); in shrink_active_list()
2161 struct lruvec *lruvec, struct scan_control *sc) in shrink_list() argument
2165 shrink_active_list(nr_to_scan, lruvec, sc, lru); in shrink_list()
2171 return shrink_inactive_list(nr_to_scan, lruvec, sc, lru); in shrink_list()
2202 static bool inactive_is_low(struct lruvec *lruvec, enum lru_list inactive_lru) in inactive_is_low() argument
2209 inactive = lruvec_page_state(lruvec, NR_LRU_BASE + inactive_lru); in inactive_is_low()
2210 active = lruvec_page_state(lruvec, NR_LRU_BASE + active_lru); in inactive_is_low()
2237 static void get_scan_count(struct lruvec *lruvec, struct scan_control *sc, in get_scan_count() argument
2240 struct mem_cgroup *memcg = lruvec_memcg(lruvec); in get_scan_count()
2331 lruvec_size = lruvec_lru_size(lruvec, lru, sc->reclaim_idx); in get_scan_count()
2425 static void shrink_lruvec(struct lruvec *lruvec, struct scan_control *sc) in shrink_lruvec() argument
2436 get_scan_count(lruvec, sc, nr); in shrink_lruvec()
2467 lruvec, sc); in shrink_lruvec()
2534 if (total_swap_pages && inactive_is_low(lruvec, LRU_INACTIVE_ANON)) in shrink_lruvec()
2535 shrink_active_list(SWAP_CLUSTER_MAX, lruvec, in shrink_lruvec()
2617 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, pgdat); in shrink_node_memcgs() local
2654 shrink_lruvec(lruvec, sc); in shrink_node_memcgs()
2671 struct lruvec *target_lruvec; in shrink_node()
2986 struct lruvec *target_lruvec; in snapshot_refaults()
3055 struct lruvec *lruvec; in do_try_to_free_pages() local
3057 lruvec = mem_cgroup_lruvec(sc->target_mem_cgroup, in do_try_to_free_pages()
3059 clear_bit(LRUVEC_CONGESTED, &lruvec->flags); in do_try_to_free_pages()
3284 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, pgdat); in mem_cgroup_shrink_node() local
3309 shrink_lruvec(lruvec, &sc); in mem_cgroup_shrink_node()
3361 struct lruvec *lruvec; in age_active_anon() local
3366 lruvec = mem_cgroup_lruvec(NULL, pgdat); in age_active_anon()
3367 if (!inactive_is_low(lruvec, LRU_INACTIVE_ANON)) in age_active_anon()
3372 lruvec = mem_cgroup_lruvec(memcg, pgdat); in age_active_anon()
3373 shrink_active_list(SWAP_CLUSTER_MAX, lruvec, in age_active_anon()
3442 struct lruvec *lruvec = mem_cgroup_lruvec(NULL, pgdat); in clear_pgdat_congested() local
3444 clear_bit(LRUVEC_CONGESTED, &lruvec->flags); in clear_pgdat_congested()
4265 struct lruvec *lruvec; in check_move_unevictable_pages() local
4288 lruvec = mem_cgroup_page_lruvec(page, pgdat); in check_move_unevictable_pages()
4298 del_page_from_lru_list(page, lruvec, LRU_UNEVICTABLE); in check_move_unevictable_pages()
4299 add_page_to_lru_list(page, lruvec, lru); in check_move_unevictable_pages()