Lines Matching refs:mdsc
45 static u64 __get_oldest_flush_tid(struct ceph_mds_client *mdsc);
46 static void __kick_flushing_caps(struct ceph_mds_client *mdsc,
127 void ceph_caps_init(struct ceph_mds_client *mdsc) in ceph_caps_init() argument
129 INIT_LIST_HEAD(&mdsc->caps_list); in ceph_caps_init()
130 spin_lock_init(&mdsc->caps_list_lock); in ceph_caps_init()
133 void ceph_caps_finalize(struct ceph_mds_client *mdsc) in ceph_caps_finalize() argument
137 spin_lock(&mdsc->caps_list_lock); in ceph_caps_finalize()
138 while (!list_empty(&mdsc->caps_list)) { in ceph_caps_finalize()
139 cap = list_first_entry(&mdsc->caps_list, in ceph_caps_finalize()
144 mdsc->caps_total_count = 0; in ceph_caps_finalize()
145 mdsc->caps_avail_count = 0; in ceph_caps_finalize()
146 mdsc->caps_use_count = 0; in ceph_caps_finalize()
147 mdsc->caps_reserve_count = 0; in ceph_caps_finalize()
148 mdsc->caps_min_count = 0; in ceph_caps_finalize()
149 spin_unlock(&mdsc->caps_list_lock); in ceph_caps_finalize()
152 void ceph_adjust_caps_max_min(struct ceph_mds_client *mdsc, in ceph_adjust_caps_max_min() argument
155 spin_lock(&mdsc->caps_list_lock); in ceph_adjust_caps_max_min()
156 mdsc->caps_min_count = fsopt->max_readdir; in ceph_adjust_caps_max_min()
157 if (mdsc->caps_min_count < 1024) in ceph_adjust_caps_max_min()
158 mdsc->caps_min_count = 1024; in ceph_adjust_caps_max_min()
159 mdsc->caps_use_max = fsopt->caps_max; in ceph_adjust_caps_max_min()
160 if (mdsc->caps_use_max > 0 && in ceph_adjust_caps_max_min()
161 mdsc->caps_use_max < mdsc->caps_min_count) in ceph_adjust_caps_max_min()
162 mdsc->caps_use_max = mdsc->caps_min_count; in ceph_adjust_caps_max_min()
163 spin_unlock(&mdsc->caps_list_lock); in ceph_adjust_caps_max_min()
166 static void __ceph_unreserve_caps(struct ceph_mds_client *mdsc, int nr_caps) in __ceph_unreserve_caps() argument
172 BUG_ON(mdsc->caps_reserve_count < nr_caps); in __ceph_unreserve_caps()
173 mdsc->caps_reserve_count -= nr_caps; in __ceph_unreserve_caps()
174 if (mdsc->caps_avail_count >= in __ceph_unreserve_caps()
175 mdsc->caps_reserve_count + mdsc->caps_min_count) { in __ceph_unreserve_caps()
176 mdsc->caps_total_count -= nr_caps; in __ceph_unreserve_caps()
178 cap = list_first_entry(&mdsc->caps_list, in __ceph_unreserve_caps()
184 mdsc->caps_avail_count += nr_caps; in __ceph_unreserve_caps()
189 mdsc->caps_total_count, mdsc->caps_use_count, in __ceph_unreserve_caps()
190 mdsc->caps_reserve_count, mdsc->caps_avail_count); in __ceph_unreserve_caps()
191 BUG_ON(mdsc->caps_total_count != mdsc->caps_use_count + in __ceph_unreserve_caps()
192 mdsc->caps_reserve_count + in __ceph_unreserve_caps()
193 mdsc->caps_avail_count); in __ceph_unreserve_caps()
200 int ceph_reserve_caps(struct ceph_mds_client *mdsc, in ceph_reserve_caps() argument
216 spin_lock(&mdsc->caps_list_lock); in ceph_reserve_caps()
217 if (mdsc->caps_avail_count >= need) in ceph_reserve_caps()
220 have = mdsc->caps_avail_count; in ceph_reserve_caps()
221 mdsc->caps_avail_count -= have; in ceph_reserve_caps()
222 mdsc->caps_reserve_count += have; in ceph_reserve_caps()
223 BUG_ON(mdsc->caps_total_count != mdsc->caps_use_count + in ceph_reserve_caps()
224 mdsc->caps_reserve_count + in ceph_reserve_caps()
225 mdsc->caps_avail_count); in ceph_reserve_caps()
226 spin_unlock(&mdsc->caps_list_lock); in ceph_reserve_caps()
238 for (j = 0; j < mdsc->max_sessions; j++) { in ceph_reserve_caps()
239 s = __ceph_lookup_mds_session(mdsc, j); in ceph_reserve_caps()
242 mutex_unlock(&mdsc->mutex); in ceph_reserve_caps()
246 ceph_trim_caps(mdsc, s, max_caps); in ceph_reserve_caps()
250 mutex_lock(&mdsc->mutex); in ceph_reserve_caps()
254 spin_lock(&mdsc->caps_list_lock); in ceph_reserve_caps()
255 if (mdsc->caps_avail_count) { in ceph_reserve_caps()
257 if (mdsc->caps_avail_count >= need - i) in ceph_reserve_caps()
260 more_have = mdsc->caps_avail_count; in ceph_reserve_caps()
264 mdsc->caps_avail_count -= more_have; in ceph_reserve_caps()
265 mdsc->caps_reserve_count += more_have; in ceph_reserve_caps()
268 spin_unlock(&mdsc->caps_list_lock); in ceph_reserve_caps()
285 spin_lock(&mdsc->caps_list_lock); in ceph_reserve_caps()
286 mdsc->caps_total_count += alloc; in ceph_reserve_caps()
287 mdsc->caps_reserve_count += alloc; in ceph_reserve_caps()
288 list_splice(&newcaps, &mdsc->caps_list); in ceph_reserve_caps()
290 BUG_ON(mdsc->caps_total_count != mdsc->caps_use_count + in ceph_reserve_caps()
291 mdsc->caps_reserve_count + in ceph_reserve_caps()
292 mdsc->caps_avail_count); in ceph_reserve_caps()
295 __ceph_unreserve_caps(mdsc, have + alloc); in ceph_reserve_caps()
297 spin_unlock(&mdsc->caps_list_lock); in ceph_reserve_caps()
300 ctx, mdsc->caps_total_count, mdsc->caps_use_count, in ceph_reserve_caps()
301 mdsc->caps_reserve_count, mdsc->caps_avail_count); in ceph_reserve_caps()
305 void ceph_unreserve_caps(struct ceph_mds_client *mdsc, in ceph_unreserve_caps() argument
313 spin_lock(&mdsc->caps_list_lock); in ceph_unreserve_caps()
314 __ceph_unreserve_caps(mdsc, ctx->count); in ceph_unreserve_caps()
317 if (mdsc->caps_use_max > 0 && in ceph_unreserve_caps()
318 mdsc->caps_use_count > mdsc->caps_use_max) in ceph_unreserve_caps()
320 spin_unlock(&mdsc->caps_list_lock); in ceph_unreserve_caps()
323 ceph_reclaim_caps_nr(mdsc, ctx->used); in ceph_unreserve_caps()
326 struct ceph_cap *ceph_get_cap(struct ceph_mds_client *mdsc, in ceph_get_cap() argument
335 spin_lock(&mdsc->caps_list_lock); in ceph_get_cap()
336 mdsc->caps_use_count++; in ceph_get_cap()
337 mdsc->caps_total_count++; in ceph_get_cap()
338 spin_unlock(&mdsc->caps_list_lock); in ceph_get_cap()
340 spin_lock(&mdsc->caps_list_lock); in ceph_get_cap()
341 if (mdsc->caps_avail_count) { in ceph_get_cap()
342 BUG_ON(list_empty(&mdsc->caps_list)); in ceph_get_cap()
344 mdsc->caps_avail_count--; in ceph_get_cap()
345 mdsc->caps_use_count++; in ceph_get_cap()
346 cap = list_first_entry(&mdsc->caps_list, in ceph_get_cap()
350 BUG_ON(mdsc->caps_total_count != mdsc->caps_use_count + in ceph_get_cap()
351 mdsc->caps_reserve_count + mdsc->caps_avail_count); in ceph_get_cap()
353 spin_unlock(&mdsc->caps_list_lock); in ceph_get_cap()
359 spin_lock(&mdsc->caps_list_lock); in ceph_get_cap()
361 ctx, ctx->count, mdsc->caps_total_count, mdsc->caps_use_count, in ceph_get_cap()
362 mdsc->caps_reserve_count, mdsc->caps_avail_count); in ceph_get_cap()
364 BUG_ON(ctx->count > mdsc->caps_reserve_count); in ceph_get_cap()
365 BUG_ON(list_empty(&mdsc->caps_list)); in ceph_get_cap()
369 mdsc->caps_reserve_count--; in ceph_get_cap()
370 mdsc->caps_use_count++; in ceph_get_cap()
372 cap = list_first_entry(&mdsc->caps_list, struct ceph_cap, caps_item); in ceph_get_cap()
375 BUG_ON(mdsc->caps_total_count != mdsc->caps_use_count + in ceph_get_cap()
376 mdsc->caps_reserve_count + mdsc->caps_avail_count); in ceph_get_cap()
377 spin_unlock(&mdsc->caps_list_lock); in ceph_get_cap()
381 void ceph_put_cap(struct ceph_mds_client *mdsc, struct ceph_cap *cap) in ceph_put_cap() argument
383 spin_lock(&mdsc->caps_list_lock); in ceph_put_cap()
385 cap, mdsc->caps_total_count, mdsc->caps_use_count, in ceph_put_cap()
386 mdsc->caps_reserve_count, mdsc->caps_avail_count); in ceph_put_cap()
387 mdsc->caps_use_count--; in ceph_put_cap()
392 if (mdsc->caps_avail_count >= mdsc->caps_reserve_count + in ceph_put_cap()
393 mdsc->caps_min_count) { in ceph_put_cap()
394 mdsc->caps_total_count--; in ceph_put_cap()
397 mdsc->caps_avail_count++; in ceph_put_cap()
398 list_add(&cap->caps_item, &mdsc->caps_list); in ceph_put_cap()
401 BUG_ON(mdsc->caps_total_count != mdsc->caps_use_count + in ceph_put_cap()
402 mdsc->caps_reserve_count + mdsc->caps_avail_count); in ceph_put_cap()
403 spin_unlock(&mdsc->caps_list_lock); in ceph_put_cap()
410 struct ceph_mds_client *mdsc = fsc->mdsc; in ceph_reservation_status() local
412 spin_lock(&mdsc->caps_list_lock); in ceph_reservation_status()
415 *total = mdsc->caps_total_count; in ceph_reservation_status()
417 *avail = mdsc->caps_avail_count; in ceph_reservation_status()
419 *used = mdsc->caps_use_count; in ceph_reservation_status()
421 *reserved = mdsc->caps_reserve_count; in ceph_reservation_status()
423 *min = mdsc->caps_min_count; in ceph_reservation_status()
425 spin_unlock(&mdsc->caps_list_lock); in ceph_reservation_status()
489 static void __cap_set_timeouts(struct ceph_mds_client *mdsc, in __cap_set_timeouts() argument
492 struct ceph_mount_options *opt = mdsc->fsc->mount_options; in __cap_set_timeouts()
510 static void __cap_delay_requeue(struct ceph_mds_client *mdsc, in __cap_delay_requeue() argument
516 if (!mdsc->stopping) { in __cap_delay_requeue()
517 spin_lock(&mdsc->cap_delay_lock); in __cap_delay_requeue()
524 __cap_set_timeouts(mdsc, ci); in __cap_delay_requeue()
525 list_add_tail(&ci->i_cap_delay_list, &mdsc->cap_delay_list); in __cap_delay_requeue()
527 spin_unlock(&mdsc->cap_delay_lock); in __cap_delay_requeue()
536 static void __cap_delay_requeue_front(struct ceph_mds_client *mdsc, in __cap_delay_requeue_front() argument
540 spin_lock(&mdsc->cap_delay_lock); in __cap_delay_requeue_front()
544 list_add(&ci->i_cap_delay_list, &mdsc->cap_delay_list); in __cap_delay_requeue_front()
545 spin_unlock(&mdsc->cap_delay_lock); in __cap_delay_requeue_front()
553 static void __cap_delay_cancel(struct ceph_mds_client *mdsc, in __cap_delay_cancel() argument
559 spin_lock(&mdsc->cap_delay_lock); in __cap_delay_cancel()
561 spin_unlock(&mdsc->cap_delay_lock); in __cap_delay_cancel()
612 struct ceph_mds_client *mdsc = ceph_inode_to_client(inode)->mdsc; in ceph_add_cap() local
688 struct ceph_snap_realm *realm = ceph_lookup_snap_realm(mdsc, in ceph_add_cap()
707 ceph_put_snap_realm(mdsc, oldrealm); in ceph_add_cap()
728 __cap_delay_requeue(mdsc, ci, true); in ceph_add_cap()
1041 ceph_put_snap_realm(ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc, in drop_inode_snap_realm()
1055 struct ceph_mds_client *mdsc = in __ceph_remove_cap() local
1056 ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc; in __ceph_remove_cap()
1100 ceph_put_cap(mdsc, cap); in __ceph_remove_cap()
1110 __cap_delay_cancel(mdsc, ci); in __ceph_remove_cap()
1269 static int __send_cap(struct ceph_mds_client *mdsc, struct ceph_cap *cap, in __send_cap() argument
1313 __cap_set_timeouts(mdsc, ci); in __send_cap()
1459 struct ceph_mds_client *mdsc = session->s_mdsc; in __ceph_flush_snaps() local
1483 spin_lock(&mdsc->cap_dirty_lock); in __ceph_flush_snaps()
1484 capsnap->cap_flush.tid = ++mdsc->last_cap_flush_tid; in __ceph_flush_snaps()
1486 &mdsc->cap_flush_list); in __ceph_flush_snaps()
1488 oldest_flush_tid = __get_oldest_flush_tid(mdsc); in __ceph_flush_snaps()
1493 spin_unlock(&mdsc->cap_dirty_lock); in __ceph_flush_snaps()
1552 struct ceph_mds_client *mdsc = ceph_inode_to_client(inode)->mdsc; in ceph_flush_snaps() local
1579 mutex_lock(&mdsc->mutex); in ceph_flush_snaps()
1580 session = __ceph_lookup_mds_session(mdsc, mds); in ceph_flush_snaps()
1581 mutex_unlock(&mdsc->mutex); in ceph_flush_snaps()
1591 __kick_flushing_caps(mdsc, session, ci, 0); in ceph_flush_snaps()
1604 spin_lock(&mdsc->snap_flush_lock); in ceph_flush_snaps()
1606 spin_unlock(&mdsc->snap_flush_lock); in ceph_flush_snaps()
1617 struct ceph_mds_client *mdsc = in __ceph_mark_dirty_caps() local
1618 ceph_sb_to_client(ci->vfs_inode.i_sb)->mdsc; in __ceph_mark_dirty_caps()
1639 WARN_ON_ONCE(!rwsem_is_locked(&mdsc->snap_rwsem)); in __ceph_mark_dirty_caps()
1646 spin_lock(&mdsc->cap_dirty_lock); in __ceph_mark_dirty_caps()
1647 list_add(&ci->i_dirty_item, &mdsc->cap_dirty); in __ceph_mark_dirty_caps()
1648 spin_unlock(&mdsc->cap_dirty_lock); in __ceph_mark_dirty_caps()
1660 __cap_delay_requeue(mdsc, ci, true); in __ceph_mark_dirty_caps()
1675 static u64 __get_oldest_flush_tid(struct ceph_mds_client *mdsc) in __get_oldest_flush_tid() argument
1677 if (!list_empty(&mdsc->cap_flush_list)) { in __get_oldest_flush_tid()
1679 list_first_entry(&mdsc->cap_flush_list, in __get_oldest_flush_tid()
1690 static bool __finish_cap_flush(struct ceph_mds_client *mdsc, in __finish_cap_flush() argument
1696 if (mdsc) { in __finish_cap_flush()
1698 if (wake && cf->g_list.prev != &mdsc->cap_flush_list) { in __finish_cap_flush()
1727 struct ceph_mds_client *mdsc = ceph_sb_to_client(inode->i_sb)->mdsc; in __mark_caps_flushing() local
1749 spin_lock(&mdsc->cap_dirty_lock); in __mark_caps_flushing()
1752 cf->tid = ++mdsc->last_cap_flush_tid; in __mark_caps_flushing()
1753 list_add_tail(&cf->g_list, &mdsc->cap_flush_list); in __mark_caps_flushing()
1754 *oldest_flush_tid = __get_oldest_flush_tid(mdsc); in __mark_caps_flushing()
1758 mdsc->num_cap_flushing++; in __mark_caps_flushing()
1760 spin_unlock(&mdsc->cap_dirty_lock); in __mark_caps_flushing()
1821 struct ceph_mds_client *mdsc = fsc->mdsc; in ceph_check_caps() local
1837 if (mdsc->stopping) in ceph_check_caps()
1847 __cap_delay_cancel(mdsc, ci); in ceph_check_caps()
1860 if (!mdsc->stopping && inode->i_nlink > 0) { in ceph_check_caps()
1906 if ((!no_delay || mdsc->stopping) && in ceph_check_caps()
2014 up_read(&mdsc->snap_rwsem); in ceph_check_caps()
2028 __kick_flushing_caps(mdsc, session, ci, 0); in ceph_check_caps()
2037 if (down_read_trylock(&mdsc->snap_rwsem) == 0) { in ceph_check_caps()
2041 down_read(&mdsc->snap_rwsem); in ceph_check_caps()
2055 spin_lock(&mdsc->cap_dirty_lock); in ceph_check_caps()
2056 oldest_flush_tid = __get_oldest_flush_tid(mdsc); in ceph_check_caps()
2057 spin_unlock(&mdsc->cap_dirty_lock); in ceph_check_caps()
2064 delayed += __send_cap(mdsc, cap, CEPH_CAP_OP_UPDATE, 0, in ceph_check_caps()
2072 __cap_delay_requeue(mdsc, ci, false); in ceph_check_caps()
2082 up_read(&mdsc->snap_rwsem); in ceph_check_caps()
2090 struct ceph_mds_client *mdsc = ceph_sb_to_client(inode->i_sb)->mdsc; in try_flush_caps() local
2119 __kick_flushing_caps(mdsc, session, ci, 0); in try_flush_caps()
2130 delayed = __send_cap(mdsc, cap, CEPH_CAP_OP_FLUSH, in try_flush_caps()
2139 __cap_delay_requeue(mdsc, ci, true); in try_flush_caps()
2291 struct ceph_mds_client *mdsc = in ceph_write_inode() local
2292 ceph_sb_to_client(inode->i_sb)->mdsc; in ceph_write_inode()
2296 __cap_delay_requeue_front(mdsc, ci); in ceph_write_inode()
2302 static void __kick_flushing_caps(struct ceph_mds_client *mdsc, in __kick_flushing_caps() argument
2343 ret = __send_cap(mdsc, cap, CEPH_CAP_OP_FLUSH, in __kick_flushing_caps()
2385 void ceph_early_kick_flushing_caps(struct ceph_mds_client *mdsc, in ceph_early_kick_flushing_caps() argument
2394 spin_lock(&mdsc->cap_dirty_lock); in ceph_early_kick_flushing_caps()
2395 oldest_flush_tid = __get_oldest_flush_tid(mdsc); in ceph_early_kick_flushing_caps()
2396 spin_unlock(&mdsc->cap_dirty_lock); in ceph_early_kick_flushing_caps()
2423 __kick_flushing_caps(mdsc, session, ci, in ceph_early_kick_flushing_caps()
2433 void ceph_kick_flushing_caps(struct ceph_mds_client *mdsc, in ceph_kick_flushing_caps() argument
2442 spin_lock(&mdsc->cap_dirty_lock); in ceph_kick_flushing_caps()
2443 oldest_flush_tid = __get_oldest_flush_tid(mdsc); in ceph_kick_flushing_caps()
2444 spin_unlock(&mdsc->cap_dirty_lock); in ceph_kick_flushing_caps()
2456 __kick_flushing_caps(mdsc, session, ci, in ceph_kick_flushing_caps()
2463 static void kick_flushing_inode_caps(struct ceph_mds_client *mdsc, in kick_flushing_inode_caps() argument
2477 spin_lock(&mdsc->cap_dirty_lock); in kick_flushing_inode_caps()
2480 oldest_flush_tid = __get_oldest_flush_tid(mdsc); in kick_flushing_inode_caps()
2481 spin_unlock(&mdsc->cap_dirty_lock); in kick_flushing_inode_caps()
2483 __kick_flushing_caps(mdsc, session, ci, oldest_flush_tid); in kick_flushing_inode_caps()
2544 struct ceph_mds_client *mdsc = ceph_inode_to_client(inode)->mdsc; in try_get_cap_refs() local
2576 up_read(&mdsc->snap_rwsem); in try_get_cap_refs()
2619 if (!down_read_trylock(&mdsc->snap_rwsem)) { in try_get_cap_refs()
2630 down_read(&mdsc->snap_rwsem); in try_get_cap_refs()
2660 if (READ_ONCE(mdsc->fsc->mount_state) == in try_get_cap_refs()
2683 up_read(&mdsc->snap_rwsem); in try_get_cap_refs()
3375 struct ceph_mds_client *mdsc = ceph_sb_to_client(inode->i_sb)->mdsc; in handle_cap_flush_ack() local
3412 spin_lock(&mdsc->cap_dirty_lock); in handle_cap_flush_ack()
3415 if (__finish_cap_flush(mdsc, NULL, cf)) in handle_cap_flush_ack()
3430 mdsc->num_cap_flushing--; in handle_cap_flush_ack()
3447 spin_unlock(&mdsc->cap_dirty_lock); in handle_cap_flush_ack()
3462 wake_up_all(&mdsc->cap_flushing_wq); in handle_cap_flush_ack()
3478 struct ceph_mds_client *mdsc = ceph_sb_to_client(inode->i_sb)->mdsc; in handle_cap_flushsnap_ack() local
3512 spin_lock(&mdsc->cap_dirty_lock); in handle_cap_flushsnap_ack()
3517 if (__finish_cap_flush(mdsc, NULL, &capsnap->cap_flush)) in handle_cap_flushsnap_ack()
3520 spin_unlock(&mdsc->cap_dirty_lock); in handle_cap_flushsnap_ack()
3529 wake_up_all(&mdsc->cap_flushing_wq); in handle_cap_flushsnap_ack()
3579 struct ceph_mds_client *mdsc = ceph_inode_to_client(inode)->mdsc; in handle_cap_export() local
3645 spin_lock(&mdsc->cap_dirty_lock); in handle_cap_export()
3648 spin_unlock(&mdsc->cap_dirty_lock); in handle_cap_export()
3662 spin_lock(&mdsc->cap_dirty_lock); in handle_cap_export()
3665 spin_unlock(&mdsc->cap_dirty_lock); in handle_cap_export()
3676 tsession = ceph_mdsc_open_export_target_session(mdsc, target); in handle_cap_export()
3687 new_cap = ceph_get_cap(mdsc, NULL); in handle_cap_export()
3703 ceph_put_cap(mdsc, new_cap); in handle_cap_export()
3711 static void handle_cap_import(struct ceph_mds_client *mdsc, in handle_cap_import() argument
3748 new_cap = ceph_get_cap(mdsc, NULL); in handle_cap_import()
3754 ceph_put_cap(mdsc, new_cap); in handle_cap_import()
3799 struct ceph_mds_client *mdsc = session->s_mdsc; in ceph_handle_caps() local
3862 struct ceph_osd_client *osdc = &mdsc->fsc->client->osdc; in ceph_handle_caps()
3911 inode = ceph_find_inode(mdsc->fsc->sb, vino); in ceph_handle_caps()
3925 cap = ceph_get_cap(mdsc, NULL); in ceph_handle_caps()
3953 down_write(&mdsc->snap_rwsem); in ceph_handle_caps()
3954 ceph_update_snap_trace(mdsc, snaptrace, in ceph_handle_caps()
3957 downgrade_write(&mdsc->snap_rwsem); in ceph_handle_caps()
3959 down_read(&mdsc->snap_rwsem); in ceph_handle_caps()
3961 handle_cap_import(mdsc, inode, h, peer, session, in ceph_handle_caps()
3966 ceph_put_snap_realm(mdsc, realm); in ceph_handle_caps()
4020 ceph_flush_cap_releases(mdsc, session); in ceph_handle_caps()
4032 void ceph_check_delayed_caps(struct ceph_mds_client *mdsc) in ceph_check_delayed_caps() argument
4040 spin_lock(&mdsc->cap_delay_lock); in ceph_check_delayed_caps()
4041 if (list_empty(&mdsc->cap_delay_list)) in ceph_check_delayed_caps()
4043 ci = list_first_entry(&mdsc->cap_delay_list, in ceph_check_delayed_caps()
4052 spin_unlock(&mdsc->cap_delay_lock); in ceph_check_delayed_caps()
4061 spin_unlock(&mdsc->cap_delay_lock); in ceph_check_delayed_caps()
4067 void ceph_flush_dirty_caps(struct ceph_mds_client *mdsc) in ceph_flush_dirty_caps() argument
4073 spin_lock(&mdsc->cap_dirty_lock); in ceph_flush_dirty_caps()
4074 while (!list_empty(&mdsc->cap_dirty)) { in ceph_flush_dirty_caps()
4075 ci = list_first_entry(&mdsc->cap_dirty, struct ceph_inode_info, in ceph_flush_dirty_caps()
4080 spin_unlock(&mdsc->cap_dirty_lock); in ceph_flush_dirty_caps()
4083 spin_lock(&mdsc->cap_dirty_lock); in ceph_flush_dirty_caps()
4085 spin_unlock(&mdsc->cap_dirty_lock); in ceph_flush_dirty_caps()
4143 struct ceph_mds_client *mdsc = in ceph_drop_caps_for_unlink() local
4144 ceph_inode_to_client(inode)->mdsc; in ceph_drop_caps_for_unlink()
4145 __cap_delay_requeue_front(mdsc, ci); in ceph_drop_caps_for_unlink()