Lines Matching refs:vmpr
78 static struct vmpressure *vmpressure_parent(struct vmpressure *vmpr) in vmpressure_parent() argument
80 struct cgroup_subsys_state *css = vmpressure_to_css(vmpr); in vmpressure_parent()
161 static bool vmpressure_event(struct vmpressure *vmpr, in vmpressure_event() argument
168 mutex_lock(&vmpr->events_lock); in vmpressure_event()
169 list_for_each_entry(ev, &vmpr->events, node) { in vmpressure_event()
179 mutex_unlock(&vmpr->events_lock); in vmpressure_event()
186 struct vmpressure *vmpr = work_to_vmpressure(work); in vmpressure_work_fn() local
193 spin_lock(&vmpr->sr_lock); in vmpressure_work_fn()
202 scanned = vmpr->tree_scanned; in vmpressure_work_fn()
204 spin_unlock(&vmpr->sr_lock); in vmpressure_work_fn()
208 reclaimed = vmpr->tree_reclaimed; in vmpressure_work_fn()
209 vmpr->tree_scanned = 0; in vmpressure_work_fn()
210 vmpr->tree_reclaimed = 0; in vmpressure_work_fn()
211 spin_unlock(&vmpr->sr_lock); in vmpressure_work_fn()
216 if (vmpressure_event(vmpr, level, ancestor, signalled)) in vmpressure_work_fn()
219 } while ((vmpr = vmpressure_parent(vmpr))); in vmpressure_work_fn()
246 struct vmpressure *vmpr = memcg_to_vmpressure(memcg); in vmpressure() local
274 spin_lock(&vmpr->sr_lock); in vmpressure()
275 scanned = vmpr->tree_scanned += scanned; in vmpressure()
276 vmpr->tree_reclaimed += reclaimed; in vmpressure()
277 spin_unlock(&vmpr->sr_lock); in vmpressure()
281 schedule_work(&vmpr->work); in vmpressure()
289 spin_lock(&vmpr->sr_lock); in vmpressure()
290 scanned = vmpr->scanned += scanned; in vmpressure()
291 reclaimed = vmpr->reclaimed += reclaimed; in vmpressure()
293 spin_unlock(&vmpr->sr_lock); in vmpressure()
296 vmpr->scanned = vmpr->reclaimed = 0; in vmpressure()
297 spin_unlock(&vmpr->sr_lock); in vmpressure()
365 struct vmpressure *vmpr = memcg_to_vmpressure(memcg); in vmpressure_register_event() local
407 mutex_lock(&vmpr->events_lock); in vmpressure_register_event()
408 list_add(&ev->node, &vmpr->events); in vmpressure_register_event()
409 mutex_unlock(&vmpr->events_lock); in vmpressure_register_event()
429 struct vmpressure *vmpr = memcg_to_vmpressure(memcg); in vmpressure_unregister_event() local
432 mutex_lock(&vmpr->events_lock); in vmpressure_unregister_event()
433 list_for_each_entry(ev, &vmpr->events, node) { in vmpressure_unregister_event()
440 mutex_unlock(&vmpr->events_lock); in vmpressure_unregister_event()
450 void vmpressure_init(struct vmpressure *vmpr) in vmpressure_init() argument
452 spin_lock_init(&vmpr->sr_lock); in vmpressure_init()
453 mutex_init(&vmpr->events_lock); in vmpressure_init()
454 INIT_LIST_HEAD(&vmpr->events); in vmpressure_init()
455 INIT_WORK(&vmpr->work, vmpressure_work_fn); in vmpressure_init()
465 void vmpressure_cleanup(struct vmpressure *vmpr) in vmpressure_cleanup() argument
471 flush_work(&vmpr->work); in vmpressure_cleanup()