Lines Matching refs:sdp

52 	struct gfs2_sbd *sdp;		/* incore superblock           */  member
150 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in glock_blocked_by_withdraw() local
152 if (likely(!gfs2_withdrawn(sdp))) in glock_blocked_by_withdraw()
156 if (!sdp->sd_jdesc || in glock_blocked_by_withdraw()
157 gl->gl_name.ln_number == sdp->sd_jdesc->jd_no_addr) in glock_blocked_by_withdraw()
164 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_glock_free() local
171 if (atomic_dec_and_test(&sdp->sd_glock_disposal)) in gfs2_glock_free()
172 wake_up(&sdp->sd_glock_wait); in gfs2_glock_free()
265 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in __gfs2_glock_put() local
275 if (!gfs2_withdrawn(sdp)) in __gfs2_glock_put()
280 sdp->sd_lockstruct.ls_ops->lm_put_lock(gl); in __gfs2_glock_put()
340 struct gfs2_sbd *sdp = gh->gh_gl->gl_name.ln_sbd; in gfs2_holder_wake() local
342 wake_up(&sdp->sd_async_glock_wait); in gfs2_holder_wake()
471 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_set_demote() local
475 wake_up(&sdp->sd_async_glock_wait); in gfs2_set_demote()
585 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in do_xmote() local
622 if (cmpxchg(&sdp->sd_log_error, 0, ret)) { in do_xmote()
623 fs_err(sdp, "Error %d syncing glock \n", ret); in do_xmote()
638 (!cmpxchg(&sdp->sd_log_error, 0, -EIO))) { in do_xmote()
665 if (unlikely(sdp->sd_log_error && !gfs2_withdrawn(sdp))) in do_xmote()
666 gfs2_withdraw_delayed(sdp); in do_xmote()
669 test_bit(SDF_WITHDRAW_RECOVERY, &sdp->sd_flags)) { in do_xmote()
675 if (sdp->sd_lockstruct.ls_ops->lm_lock) { in do_xmote()
677 ret = sdp->sd_lockstruct.ls_ops->lm_lock(gl, target, lck_flags); in do_xmote()
680 test_bit(SDF_SKIP_DLM_UNLOCK, &sdp->sd_flags)) { in do_xmote()
684 fs_err(sdp, "lm_lock ret %d\n", ret); in do_xmote()
685 GLOCK_BUG_ON(gl, !gfs2_withdrawn(sdp)); in do_xmote()
852 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in delete_work_func() local
891 inode = gfs2_lookup_by_inum(sdp, no_addr, gl->gl_no_formal_ino, in delete_work_func()
995 int gfs2_glock_get(struct gfs2_sbd *sdp, u64 number, in gfs2_glock_get() argument
999 struct super_block *s = sdp->sd_vfs; in gfs2_glock_get()
1002 .ln_sbd = sdp }; in gfs2_glock_get()
1034 atomic_inc(&sdp->sd_glock_disposal); in gfs2_glock_get()
1047 gl->gl_stats = this_cpu_ptr(sdp->sd_lkstats)->lkstats[glops->go_type]; in gfs2_glock_get()
1082 if (atomic_dec_and_test(&sdp->sd_glock_disposal)) in gfs2_glock_get()
1083 wake_up(&sdp->sd_glock_wait); in gfs2_glock_get()
1196 struct gfs2_sbd *sdp = ghs[0].gh_gl->gl_name.ln_sbd; in gfs2_glock_async_wait() local
1210 if (!wait_event_timeout(sdp->sd_async_glock_wait, in gfs2_glock_async_wait()
1321 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in add_to_queue() local
1367 if (sdp->sd_lockstruct.ls_ops->lm_cancel) in add_to_queue()
1368 sdp->sd_lockstruct.ls_ops->lm_cancel(gl); in add_to_queue()
1374 fs_err(sdp, "original: %pSR\n", (void *)gh2->gh_ip); in add_to_queue()
1375 fs_err(sdp, "pid: %d\n", pid_nr(gh2->gh_owner_pid)); in add_to_queue()
1376 fs_err(sdp, "lock type: %d req lock state : %d\n", in add_to_queue()
1378 fs_err(sdp, "new: %pSR\n", (void *)gh->gh_ip); in add_to_queue()
1379 fs_err(sdp, "pid: %d\n", pid_nr(gh->gh_owner_pid)); in add_to_queue()
1380 fs_err(sdp, "lock type: %d req lock state : %d\n", in add_to_queue()
1444 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_glock_dq() local
1456 if (test_bit(SDF_WITHDRAW_RECOVERY, &sdp->sd_flags) && in gfs2_glock_dq()
1458 gh->gh_gl != sdp->sd_jinode_gl) { in gfs2_glock_dq()
1459 sdp->sd_glock_dqs_held++; in gfs2_glock_dq()
1461 wait_on_bit(&sdp->sd_flags, SDF_WITHDRAW_RECOVERY, in gfs2_glock_dq()
1522 int gfs2_glock_nq_num(struct gfs2_sbd *sdp, u64 number, in gfs2_glock_nq_num() argument
1529 error = gfs2_glock_get(sdp, number, glops, CREATE, &gl); in gfs2_glock_nq_num()
1857 static void glock_hash_walk(glock_examiner examiner, const struct gfs2_sbd *sdp) in glock_hash_walk() argument
1868 if (gl->gl_name.ln_sbd == sdp && in glock_hash_walk()
1915 void gfs2_flush_delete_work(struct gfs2_sbd *sdp) in gfs2_flush_delete_work() argument
1917 glock_hash_walk(flush_delete_work, sdp); in gfs2_flush_delete_work()
1960 void gfs2_glock_thaw(struct gfs2_sbd *sdp) in gfs2_glock_thaw() argument
1962 glock_hash_walk(thaw_glock, sdp); in gfs2_glock_thaw()
1985 void gfs2_gl_hash_clear(struct gfs2_sbd *sdp) in gfs2_gl_hash_clear() argument
1987 set_bit(SDF_SKIP_DLM_UNLOCK, &sdp->sd_flags); in gfs2_gl_hash_clear()
1989 glock_hash_walk(clear_glock, sdp); in gfs2_gl_hash_clear()
1991 wait_event_timeout(sdp->sd_glock_wait, in gfs2_gl_hash_clear()
1992 atomic_read(&sdp->sd_glock_disposal) == 0, in gfs2_gl_hash_clear()
1994 glock_hash_walk(dump_glock_func, sdp); in gfs2_gl_hash_clear()
2149 struct gfs2_sbd *sdp = gl->gl_name.ln_sbd; in gfs2_dump_glock() local
2150 char fs_id_buf[sizeof(sdp->sd_fsname) + 7]; in gfs2_dump_glock()
2159 if (fsid && sdp) /* safety precaution */ in gfs2_dump_glock()
2160 sprintf(fs_id_buf, "fsid=%s: ", sdp->sd_fsname); in gfs2_dump_glock()
2231 struct gfs2_sbd *sdp = seq->private; in gfs2_sbstats_seq_show() local
2244 const struct gfs2_pcpu_lkstats *lkstats = per_cpu_ptr(sdp->sd_lkstats, i); in gfs2_sbstats_seq_show()
2321 if (gl->gl_name.ln_sbd != gi->sdp) in gfs2_glock_iter_next()
2439 gi->sdp = inode->i_private; in __gfs2_glocks_open()
2493 void gfs2_create_debugfs_file(struct gfs2_sbd *sdp) in gfs2_create_debugfs_file() argument
2495 sdp->debugfs_dir = debugfs_create_dir(sdp->sd_table_name, gfs2_root); in gfs2_create_debugfs_file()
2497 debugfs_create_file("glocks", S_IFREG | S_IRUGO, sdp->debugfs_dir, sdp, in gfs2_create_debugfs_file()
2500 debugfs_create_file("glstats", S_IFREG | S_IRUGO, sdp->debugfs_dir, sdp, in gfs2_create_debugfs_file()
2503 debugfs_create_file("sbstats", S_IFREG | S_IRUGO, sdp->debugfs_dir, sdp, in gfs2_create_debugfs_file()
2507 void gfs2_delete_debugfs_file(struct gfs2_sbd *sdp) in gfs2_delete_debugfs_file() argument
2509 debugfs_remove_recursive(sdp->debugfs_dir); in gfs2_delete_debugfs_file()
2510 sdp->debugfs_dir = NULL; in gfs2_delete_debugfs_file()