Lines Matching refs:vmpr
75 static struct vmpressure *vmpressure_parent(struct vmpressure *vmpr) in vmpressure_parent() argument
77 struct cgroup_subsys_state *css = vmpressure_to_css(vmpr); in vmpressure_parent()
158 static bool vmpressure_event(struct vmpressure *vmpr, in vmpressure_event() argument
165 mutex_lock(&vmpr->events_lock); in vmpressure_event()
166 list_for_each_entry(ev, &vmpr->events, node) { in vmpressure_event()
176 mutex_unlock(&vmpr->events_lock); in vmpressure_event()
183 struct vmpressure *vmpr = work_to_vmpressure(work); in vmpressure_work_fn() local
190 spin_lock(&vmpr->sr_lock); in vmpressure_work_fn()
199 scanned = vmpr->tree_scanned; in vmpressure_work_fn()
201 spin_unlock(&vmpr->sr_lock); in vmpressure_work_fn()
205 reclaimed = vmpr->tree_reclaimed; in vmpressure_work_fn()
206 vmpr->tree_scanned = 0; in vmpressure_work_fn()
207 vmpr->tree_reclaimed = 0; in vmpressure_work_fn()
208 spin_unlock(&vmpr->sr_lock); in vmpressure_work_fn()
213 if (vmpressure_event(vmpr, level, ancestor, signalled)) in vmpressure_work_fn()
216 } while ((vmpr = vmpressure_parent(vmpr))); in vmpressure_work_fn()
243 struct vmpressure *vmpr = memcg_to_vmpressure(memcg); in vmpressure() local
271 spin_lock(&vmpr->sr_lock); in vmpressure()
272 scanned = vmpr->tree_scanned += scanned; in vmpressure()
273 vmpr->tree_reclaimed += reclaimed; in vmpressure()
274 spin_unlock(&vmpr->sr_lock); in vmpressure()
278 schedule_work(&vmpr->work); in vmpressure()
286 spin_lock(&vmpr->sr_lock); in vmpressure()
287 scanned = vmpr->scanned += scanned; in vmpressure()
288 reclaimed = vmpr->reclaimed += reclaimed; in vmpressure()
290 spin_unlock(&vmpr->sr_lock); in vmpressure()
293 vmpr->scanned = vmpr->reclaimed = 0; in vmpressure()
294 spin_unlock(&vmpr->sr_lock); in vmpressure()
365 struct vmpressure *vmpr = memcg_to_vmpressure(memcg); in vmpressure_register_event() local
403 mutex_lock(&vmpr->events_lock); in vmpressure_register_event()
404 list_add(&ev->node, &vmpr->events); in vmpressure_register_event()
405 mutex_unlock(&vmpr->events_lock); in vmpressure_register_event()
426 struct vmpressure *vmpr = memcg_to_vmpressure(memcg); in vmpressure_unregister_event() local
429 mutex_lock(&vmpr->events_lock); in vmpressure_unregister_event()
430 list_for_each_entry(ev, &vmpr->events, node) { in vmpressure_unregister_event()
437 mutex_unlock(&vmpr->events_lock); in vmpressure_unregister_event()
447 void vmpressure_init(struct vmpressure *vmpr) in vmpressure_init() argument
449 spin_lock_init(&vmpr->sr_lock); in vmpressure_init()
450 mutex_init(&vmpr->events_lock); in vmpressure_init()
451 INIT_LIST_HEAD(&vmpr->events); in vmpressure_init()
452 INIT_WORK(&vmpr->work, vmpressure_work_fn); in vmpressure_init()
462 void vmpressure_cleanup(struct vmpressure *vmpr) in vmpressure_cleanup() argument
468 flush_work(&vmpr->work); in vmpressure_cleanup()