Lines Matching refs:dfs_cache
59 struct dfs_cache { struct
67 static struct dfs_cache dfs_cache; argument
298 INIT_LIST_HEAD(&dfs_cache.dc_vol_list); in dfs_cache_init()
299 mutex_init(&dfs_cache.dc_lock); in dfs_cache_init()
300 INIT_DELAYED_WORK(&dfs_cache.dc_refresh, refresh_cache_worker); in dfs_cache_init()
301 dfs_cache.dc_ttl = -1; in dfs_cache_init()
302 dfs_cache.dc_nlsc = load_nls_default(); in dfs_cache_init()
468 mutex_lock(&dfs_cache.dc_lock); in add_cache_entry()
469 if (dfs_cache.dc_ttl < 0) { in add_cache_entry()
470 dfs_cache.dc_ttl = ce->ce_ttl; in add_cache_entry()
471 queue_delayed_work(cifsiod_wq, &dfs_cache.dc_refresh, in add_cache_entry()
472 dfs_cache.dc_ttl * HZ); in add_cache_entry()
474 dfs_cache.dc_ttl = min_t(int, dfs_cache.dc_ttl, ce->ce_ttl); in add_cache_entry()
475 mod_delayed_work(cifsiod_wq, &dfs_cache.dc_refresh, in add_cache_entry()
476 dfs_cache.dc_ttl * HZ); in add_cache_entry()
478 mutex_unlock(&dfs_cache.dc_lock); in add_cache_entry()
542 list_for_each_entry_safe(vi, nvi, &dfs_cache.dc_vol_list, vi_list) in free_vol_list()
551 cancel_delayed_work_sync(&dfs_cache.dc_refresh); in dfs_cache_destroy()
552 unload_nls(dfs_cache.dc_nlsc); in dfs_cache_destroy()
554 mutex_destroy(&dfs_cache.dc_lock); in dfs_cache_destroy()
1176 mutex_lock(&dfs_cache.dc_lock); in dfs_cache_add_vol()
1177 list_add_tail(&vi->vi_list, &dfs_cache.dc_vol_list); in dfs_cache_add_vol()
1178 mutex_unlock(&dfs_cache.dc_lock); in dfs_cache_add_vol()
1192 list_for_each_entry(vi, &dfs_cache.dc_vol_list, vi_list) { in find_vol()
1219 mutex_lock(&dfs_cache.dc_lock); in dfs_cache_update_vol()
1233 mutex_unlock(&dfs_cache.dc_lock); in dfs_cache_update_vol()
1251 mutex_lock(&dfs_cache.dc_lock); in dfs_cache_del_vol()
1255 mutex_unlock(&dfs_cache.dc_lock); in dfs_cache_del_vol()
1377 static void do_refresh_tcon(struct dfs_cache *dc, struct dfs_cache_vol_info *vi, in do_refresh_tcon()
1452 struct dfs_cache *dc = container_of(work, struct dfs_cache, in refresh_cache_worker()