Lines Matching refs:vmpr

75 static struct vmpressure *vmpressure_parent(struct vmpressure *vmpr)  in vmpressure_parent()  argument
77 struct cgroup_subsys_state *css = vmpressure_to_css(vmpr); in vmpressure_parent()
158 static bool vmpressure_event(struct vmpressure *vmpr, in vmpressure_event() argument
165 mutex_lock(&vmpr->events_lock); in vmpressure_event()
166 list_for_each_entry(ev, &vmpr->events, node) { in vmpressure_event()
176 mutex_unlock(&vmpr->events_lock); in vmpressure_event()
183 struct vmpressure *vmpr = work_to_vmpressure(work); in vmpressure_work_fn() local
190 spin_lock(&vmpr->sr_lock); in vmpressure_work_fn()
199 scanned = vmpr->tree_scanned; in vmpressure_work_fn()
201 spin_unlock(&vmpr->sr_lock); in vmpressure_work_fn()
205 reclaimed = vmpr->tree_reclaimed; in vmpressure_work_fn()
206 vmpr->tree_scanned = 0; in vmpressure_work_fn()
207 vmpr->tree_reclaimed = 0; in vmpressure_work_fn()
208 spin_unlock(&vmpr->sr_lock); in vmpressure_work_fn()
213 if (vmpressure_event(vmpr, level, ancestor, signalled)) in vmpressure_work_fn()
216 } while ((vmpr = vmpressure_parent(vmpr))); in vmpressure_work_fn()
243 struct vmpressure *vmpr = memcg_to_vmpressure(memcg); in vmpressure() local
271 spin_lock(&vmpr->sr_lock); in vmpressure()
272 scanned = vmpr->tree_scanned += scanned; in vmpressure()
273 vmpr->tree_reclaimed += reclaimed; in vmpressure()
274 spin_unlock(&vmpr->sr_lock); in vmpressure()
278 schedule_work(&vmpr->work); in vmpressure()
286 spin_lock(&vmpr->sr_lock); in vmpressure()
287 scanned = vmpr->scanned += scanned; in vmpressure()
288 reclaimed = vmpr->reclaimed += reclaimed; in vmpressure()
290 spin_unlock(&vmpr->sr_lock); in vmpressure()
293 vmpr->scanned = vmpr->reclaimed = 0; in vmpressure()
294 spin_unlock(&vmpr->sr_lock); in vmpressure()
365 struct vmpressure *vmpr = memcg_to_vmpressure(memcg); in vmpressure_register_event() local
405 mutex_lock(&vmpr->events_lock); in vmpressure_register_event()
406 list_add(&ev->node, &vmpr->events); in vmpressure_register_event()
407 mutex_unlock(&vmpr->events_lock); in vmpressure_register_event()
428 struct vmpressure *vmpr = memcg_to_vmpressure(memcg); in vmpressure_unregister_event() local
431 mutex_lock(&vmpr->events_lock); in vmpressure_unregister_event()
432 list_for_each_entry(ev, &vmpr->events, node) { in vmpressure_unregister_event()
439 mutex_unlock(&vmpr->events_lock); in vmpressure_unregister_event()
449 void vmpressure_init(struct vmpressure *vmpr) in vmpressure_init() argument
451 spin_lock_init(&vmpr->sr_lock); in vmpressure_init()
452 mutex_init(&vmpr->events_lock); in vmpressure_init()
453 INIT_LIST_HEAD(&vmpr->events); in vmpressure_init()
454 INIT_WORK(&vmpr->work, vmpressure_work_fn); in vmpressure_init()
464 void vmpressure_cleanup(struct vmpressure *vmpr) in vmpressure_cleanup() argument
470 flush_work(&vmpr->work); in vmpressure_cleanup()