Searched refs:gl_lockref (Results 1 – 6 of 6) sorted by relevance
157 GLOCK_BUG_ON(gl, __lockref_is_dead(&gl->gl_lockref)); in gfs2_glock_hold()158 lockref_get(&gl->gl_lockref); in gfs2_glock_hold()226 GLOCK_BUG_ON(gl, gl->gl_lockref.count < 2); in __gfs2_glock_queue_work()227 gl->gl_lockref.count--; in __gfs2_glock_queue_work()232 spin_lock(&gl->gl_lockref.lock); in gfs2_glock_queue_work()234 spin_unlock(&gl->gl_lockref.lock); in gfs2_glock_queue_work()242 lockref_mark_dead(&gl->gl_lockref); in __gfs2_glock_put()245 spin_unlock(&gl->gl_lockref.lock); in __gfs2_glock_put()268 if (lockref_put_or_lock(&gl->gl_lockref)) in gfs2_glock_put()348 __releases(&gl->gl_lockref.lock) in do_promote()[all …]
142 spin_lock(&gl->gl_lockref.lock); in gfs2_glock_is_locked_by_me()152 spin_unlock(&gl->gl_lockref.lock); in gfs2_glock_is_locked_by_me()276 spin_lock(&gl->gl_lockref.lock); in glock_set_object()280 spin_unlock(&gl->gl_lockref.lock); in glock_set_object()303 spin_lock(&gl->gl_lockref.lock); in glock_clear_object()306 spin_unlock(&gl->gl_lockref.lock); in glock_clear_object()
150 spin_lock(&gl->gl_lockref.lock); in rgrp_go_sync()154 spin_unlock(&gl->gl_lockref.lock); in rgrp_go_sync()167 spin_lock(&gl->gl_lockref.lock); in rgrp_go_sync()171 spin_unlock(&gl->gl_lockref.lock); in rgrp_go_sync()205 spin_lock(&gl->gl_lockref.lock); in gfs2_glock2inode()209 spin_unlock(&gl->gl_lockref.lock); in gfs2_glock2inode()217 spin_lock(&gl->gl_lockref.lock); in gfs2_glock2rgrp()219 spin_unlock(&gl->gl_lockref.lock); in gfs2_glock2rgrp()579 gl->gl_lockref.count++; in iopen_go_callback()581 gl->gl_lockref.count--; in iopen_go_callback()
54 spin_lock_init(&gl->gl_lockref.lock); in gfs2_init_glock_once()
352 struct lockref gl_lockref; member
8 1. A spinlock (gl_lockref.lock) which protects the internal state such85 Operation | GLF_LOCK bit lock held | gl_lockref.lock spinlock held