Lines Matching refs:inode
70 static inline struct inode *wb_inode(struct list_head *head) in wb_inode()
72 return list_entry(head, struct inode, i_io_list); in wb_inode()
118 static bool inode_io_list_move_locked(struct inode *inode, in inode_io_list_move_locked() argument
123 assert_spin_locked(&inode->i_lock); in inode_io_list_move_locked()
125 list_move(&inode->i_io_list, head); in inode_io_list_move_locked()
234 / sizeof(struct inode *))
239 void __inode_attach_wb(struct inode *inode, struct page *page) in __inode_attach_wb() argument
241 struct backing_dev_info *bdi = inode_to_bdi(inode); in __inode_attach_wb()
244 if (inode_cgwb_enabled(inode)) { in __inode_attach_wb()
265 if (unlikely(cmpxchg(&inode->i_wb, NULL, wb))) in __inode_attach_wb()
278 static void inode_cgwb_move_to_attached(struct inode *inode, in inode_cgwb_move_to_attached() argument
282 assert_spin_locked(&inode->i_lock); in inode_cgwb_move_to_attached()
284 inode->i_state &= ~I_SYNC_QUEUED; in inode_cgwb_move_to_attached()
286 list_move(&inode->i_io_list, &wb->b_attached); in inode_cgwb_move_to_attached()
288 list_del_init(&inode->i_io_list); in inode_cgwb_move_to_attached()
301 locked_inode_to_wb_and_lock_list(struct inode *inode) in locked_inode_to_wb_and_lock_list() argument
302 __releases(&inode->i_lock) in locked_inode_to_wb_and_lock_list()
306 struct bdi_writeback *wb = inode_to_wb(inode); in locked_inode_to_wb_and_lock_list()
315 spin_unlock(&inode->i_lock); in locked_inode_to_wb_and_lock_list()
319 if (likely(wb == inode->i_wb)) { in locked_inode_to_wb_and_lock_list()
327 spin_lock(&inode->i_lock); in locked_inode_to_wb_and_lock_list()
338 static struct bdi_writeback *inode_to_wb_and_lock_list(struct inode *inode) in inode_to_wb_and_lock_list() argument
341 spin_lock(&inode->i_lock); in inode_to_wb_and_lock_list()
342 return locked_inode_to_wb_and_lock_list(inode); in inode_to_wb_and_lock_list()
357 struct inode *inodes[];
370 static bool inode_do_switch_wbs(struct inode *inode, in inode_do_switch_wbs() argument
374 struct address_space *mapping = inode->i_mapping; in inode_do_switch_wbs()
379 spin_lock(&inode->i_lock); in inode_do_switch_wbs()
386 if (unlikely(inode->i_state & (I_FREEING | I_WILL_FREE))) in inode_do_switch_wbs()
389 trace_inode_switch_wbs(inode, old_wb, new_wb); in inode_do_switch_wbs()
427 if (!list_empty(&inode->i_io_list)) { in inode_do_switch_wbs()
428 inode->i_wb = new_wb; in inode_do_switch_wbs()
430 if (inode->i_state & I_DIRTY_ALL) { in inode_do_switch_wbs()
431 struct inode *pos; in inode_do_switch_wbs()
434 if (time_after_eq(inode->dirtied_when, in inode_do_switch_wbs()
437 inode_io_list_move_locked(inode, new_wb, in inode_do_switch_wbs()
440 inode_cgwb_move_to_attached(inode, new_wb); in inode_do_switch_wbs()
443 inode->i_wb = new_wb; in inode_do_switch_wbs()
447 inode->i_wb_frn_winner = 0; in inode_do_switch_wbs()
448 inode->i_wb_frn_avg_time = 0; in inode_do_switch_wbs()
449 inode->i_wb_frn_history = 0; in inode_do_switch_wbs()
456 smp_store_release(&inode->i_state, inode->i_state & ~I_WB_SWITCH); in inode_do_switch_wbs()
459 spin_unlock(&inode->i_lock); in inode_do_switch_wbs()
472 struct inode **inodep; in inode_switch_wbs_work_fn()
521 static bool inode_prepare_wbs_switch(struct inode *inode, in inode_prepare_wbs_switch() argument
532 if (IS_DAX(inode)) in inode_prepare_wbs_switch()
536 spin_lock(&inode->i_lock); in inode_prepare_wbs_switch()
537 if (!(inode->i_sb->s_flags & SB_ACTIVE) || in inode_prepare_wbs_switch()
538 inode->i_state & (I_WB_SWITCH | I_FREEING | I_WILL_FREE) || in inode_prepare_wbs_switch()
539 inode_to_wb(inode) == new_wb) { in inode_prepare_wbs_switch()
540 spin_unlock(&inode->i_lock); in inode_prepare_wbs_switch()
543 inode->i_state |= I_WB_SWITCH; in inode_prepare_wbs_switch()
544 __iget(inode); in inode_prepare_wbs_switch()
545 spin_unlock(&inode->i_lock); in inode_prepare_wbs_switch()
558 static void inode_switch_wbs(struct inode *inode, int new_wb_id) in inode_switch_wbs() argument
560 struct backing_dev_info *bdi = inode_to_bdi(inode); in inode_switch_wbs()
565 if (inode->i_state & I_WB_SWITCH) in inode_switch_wbs()
592 if (!inode_prepare_wbs_switch(inode, isw->new_wb)) in inode_switch_wbs()
595 isw->inodes[0] = inode; in inode_switch_wbs()
626 struct inode *inode; in cleanup_offline_cgwb() local
648 list_for_each_entry(inode, &wb->b_attached, i_io_list) { in cleanup_offline_cgwb()
649 if (!inode_prepare_wbs_switch(inode, isw->new_wb)) in cleanup_offline_cgwb()
652 isw->inodes[nr++] = inode; in cleanup_offline_cgwb()
692 struct inode *inode) in wbc_attach_and_unlock_inode() argument
694 if (!inode_cgwb_enabled(inode)) { in wbc_attach_and_unlock_inode()
695 spin_unlock(&inode->i_lock); in wbc_attach_and_unlock_inode()
699 wbc->wb = inode_to_wb(inode); in wbc_attach_and_unlock_inode()
700 wbc->inode = inode; in wbc_attach_and_unlock_inode()
703 wbc->wb_lcand_id = inode->i_wb_frn_winner; in wbc_attach_and_unlock_inode()
710 spin_unlock(&inode->i_lock); in wbc_attach_and_unlock_inode()
720 inode_switch_wbs(inode, wbc->wb_id); in wbc_attach_and_unlock_inode()
764 struct inode *inode = wbc->inode; in wbc_detach_inode() local
772 history = inode->i_wb_frn_history; in wbc_detach_inode()
773 avg_time = inode->i_wb_frn_avg_time; in wbc_detach_inode()
821 trace_inode_foreign_history(inode, wbc, history); in wbc_detach_inode()
831 inode_switch_wbs(inode, max_id); in wbc_detach_inode()
838 inode->i_wb_frn_winner = max_id; in wbc_detach_inode()
839 inode->i_wb_frn_avg_time = min(avg_time, (unsigned long)U16_MAX); in wbc_detach_inode()
840 inode->i_wb_frn_history = history; in wbc_detach_inode()
1127 static void inode_cgwb_move_to_attached(struct inode *inode, in inode_cgwb_move_to_attached() argument
1131 assert_spin_locked(&inode->i_lock); in inode_cgwb_move_to_attached()
1133 inode->i_state &= ~I_SYNC_QUEUED; in inode_cgwb_move_to_attached()
1134 list_del_init(&inode->i_io_list); in inode_cgwb_move_to_attached()
1139 locked_inode_to_wb_and_lock_list(struct inode *inode) in locked_inode_to_wb_and_lock_list() argument
1140 __releases(&inode->i_lock) in locked_inode_to_wb_and_lock_list()
1143 struct bdi_writeback *wb = inode_to_wb(inode); in locked_inode_to_wb_and_lock_list()
1145 spin_unlock(&inode->i_lock); in locked_inode_to_wb_and_lock_list()
1150 static struct bdi_writeback *inode_to_wb_and_lock_list(struct inode *inode) in inode_to_wb_and_lock_list() argument
1153 struct bdi_writeback *wb = inode_to_wb(inode); in inode_to_wb_and_lock_list()
1232 void inode_io_list_del(struct inode *inode) in inode_io_list_del() argument
1236 wb = inode_to_wb_and_lock_list(inode); in inode_io_list_del()
1237 spin_lock(&inode->i_lock); in inode_io_list_del()
1239 inode->i_state &= ~I_SYNC_QUEUED; in inode_io_list_del()
1240 list_del_init(&inode->i_io_list); in inode_io_list_del()
1243 spin_unlock(&inode->i_lock); in inode_io_list_del()
1251 void sb_mark_inode_writeback(struct inode *inode) in sb_mark_inode_writeback() argument
1253 struct super_block *sb = inode->i_sb; in sb_mark_inode_writeback()
1256 if (list_empty(&inode->i_wb_list)) { in sb_mark_inode_writeback()
1258 if (list_empty(&inode->i_wb_list)) { in sb_mark_inode_writeback()
1259 list_add_tail(&inode->i_wb_list, &sb->s_inodes_wb); in sb_mark_inode_writeback()
1260 trace_sb_mark_inode_writeback(inode); in sb_mark_inode_writeback()
1269 void sb_clear_inode_writeback(struct inode *inode) in sb_clear_inode_writeback() argument
1271 struct super_block *sb = inode->i_sb; in sb_clear_inode_writeback()
1274 if (!list_empty(&inode->i_wb_list)) { in sb_clear_inode_writeback()
1276 if (!list_empty(&inode->i_wb_list)) { in sb_clear_inode_writeback()
1277 list_del_init(&inode->i_wb_list); in sb_clear_inode_writeback()
1278 trace_sb_clear_inode_writeback(inode); in sb_clear_inode_writeback()
1293 static void redirty_tail_locked(struct inode *inode, struct bdi_writeback *wb) in redirty_tail_locked() argument
1295 assert_spin_locked(&inode->i_lock); in redirty_tail_locked()
1298 struct inode *tail; in redirty_tail_locked()
1301 if (time_before(inode->dirtied_when, tail->dirtied_when)) in redirty_tail_locked()
1302 inode->dirtied_when = jiffies; in redirty_tail_locked()
1304 inode_io_list_move_locked(inode, wb, &wb->b_dirty); in redirty_tail_locked()
1305 inode->i_state &= ~I_SYNC_QUEUED; in redirty_tail_locked()
1308 static void redirty_tail(struct inode *inode, struct bdi_writeback *wb) in redirty_tail() argument
1310 spin_lock(&inode->i_lock); in redirty_tail()
1311 redirty_tail_locked(inode, wb); in redirty_tail()
1312 spin_unlock(&inode->i_lock); in redirty_tail()
1318 static void requeue_io(struct inode *inode, struct bdi_writeback *wb) in requeue_io() argument
1320 inode_io_list_move_locked(inode, wb, &wb->b_more_io); in requeue_io()
1323 static void inode_sync_complete(struct inode *inode) in inode_sync_complete() argument
1325 inode->i_state &= ~I_SYNC; in inode_sync_complete()
1327 inode_add_lru(inode); in inode_sync_complete()
1330 wake_up_bit(&inode->i_state, __I_SYNC); in inode_sync_complete()
1333 static bool inode_dirtied_after(struct inode *inode, unsigned long t) in inode_dirtied_after() argument
1335 bool ret = time_after(inode->dirtied_when, t); in inode_dirtied_after()
1343 ret = ret && time_before_eq(inode->dirtied_when, jiffies); in inode_dirtied_after()
1361 struct inode *inode; in move_expired_inodes() local
1366 inode = wb_inode(delaying_queue->prev); in move_expired_inodes()
1367 if (inode_dirtied_after(inode, dirtied_before)) in move_expired_inodes()
1369 spin_lock(&inode->i_lock); in move_expired_inodes()
1370 list_move(&inode->i_io_list, &tmp); in move_expired_inodes()
1372 inode->i_state |= I_SYNC_QUEUED; in move_expired_inodes()
1373 spin_unlock(&inode->i_lock); in move_expired_inodes()
1374 if (sb_is_blkdev_sb(inode->i_sb)) in move_expired_inodes()
1376 if (sb && sb != inode->i_sb) in move_expired_inodes()
1378 sb = inode->i_sb; in move_expired_inodes()
1396 inode = wb_inode(pos); in move_expired_inodes()
1397 if (inode->i_sb == sb) in move_expired_inodes()
1398 list_move(&inode->i_io_list, dispatch_queue); in move_expired_inodes()
1434 static int write_inode(struct inode *inode, struct writeback_control *wbc) in write_inode() argument
1438 if (inode->i_sb->s_op->write_inode && !is_bad_inode(inode)) { in write_inode()
1439 trace_writeback_write_inode_start(inode, wbc); in write_inode()
1440 ret = inode->i_sb->s_op->write_inode(inode, wbc); in write_inode()
1441 trace_writeback_write_inode(inode, wbc); in write_inode()
1451 static void __inode_wait_for_writeback(struct inode *inode) in __inode_wait_for_writeback() argument
1452 __releases(inode->i_lock) in __inode_wait_for_writeback()
1453 __acquires(inode->i_lock) in __inode_wait_for_writeback()
1455 DEFINE_WAIT_BIT(wq, &inode->i_state, __I_SYNC); in __inode_wait_for_writeback()
1458 wqh = bit_waitqueue(&inode->i_state, __I_SYNC); in __inode_wait_for_writeback()
1459 while (inode->i_state & I_SYNC) { in __inode_wait_for_writeback()
1460 spin_unlock(&inode->i_lock); in __inode_wait_for_writeback()
1463 spin_lock(&inode->i_lock); in __inode_wait_for_writeback()
1470 void inode_wait_for_writeback(struct inode *inode) in inode_wait_for_writeback() argument
1472 spin_lock(&inode->i_lock); in inode_wait_for_writeback()
1473 __inode_wait_for_writeback(inode); in inode_wait_for_writeback()
1474 spin_unlock(&inode->i_lock); in inode_wait_for_writeback()
1482 static void inode_sleep_on_writeback(struct inode *inode) in inode_sleep_on_writeback() argument
1483 __releases(inode->i_lock) in inode_sleep_on_writeback()
1486 wait_queue_head_t *wqh = bit_waitqueue(&inode->i_state, __I_SYNC); in inode_sleep_on_writeback()
1490 sleep = inode->i_state & I_SYNC; in inode_sleep_on_writeback()
1491 spin_unlock(&inode->i_lock); in inode_sleep_on_writeback()
1505 static void requeue_inode(struct inode *inode, struct bdi_writeback *wb, in requeue_inode() argument
1508 if (inode->i_state & I_FREEING) in requeue_inode()
1516 if ((inode->i_state & I_DIRTY) && in requeue_inode()
1518 inode->dirtied_when = jiffies; in requeue_inode()
1525 redirty_tail_locked(inode, wb); in requeue_inode()
1529 if (mapping_tagged(inode->i_mapping, PAGECACHE_TAG_DIRTY)) { in requeue_inode()
1536 requeue_io(inode, wb); in requeue_inode()
1545 redirty_tail_locked(inode, wb); in requeue_inode()
1547 } else if (inode->i_state & I_DIRTY) { in requeue_inode()
1553 redirty_tail_locked(inode, wb); in requeue_inode()
1554 } else if (inode->i_state & I_DIRTY_TIME) { in requeue_inode()
1555 inode->dirtied_when = jiffies; in requeue_inode()
1556 inode_io_list_move_locked(inode, wb, &wb->b_dirty_time); in requeue_inode()
1557 inode->i_state &= ~I_SYNC_QUEUED; in requeue_inode()
1560 inode_cgwb_move_to_attached(inode, wb); in requeue_inode()
1576 __writeback_single_inode(struct inode *inode, struct writeback_control *wbc) in __writeback_single_inode() argument
1578 struct address_space *mapping = inode->i_mapping; in __writeback_single_inode()
1583 WARN_ON(!(inode->i_state & I_SYNC)); in __writeback_single_inode()
1585 trace_writeback_single_inode_start(inode, wbc, nr_to_write); in __writeback_single_inode()
1607 if ((inode->i_state & I_DIRTY_TIME) && in __writeback_single_inode()
1609 time_after(jiffies, inode->dirtied_time_when + in __writeback_single_inode()
1611 trace_writeback_lazytime(inode); in __writeback_single_inode()
1612 mark_inode_dirty_sync(inode); in __writeback_single_inode()
1621 spin_lock(&inode->i_lock); in __writeback_single_inode()
1622 dirty = inode->i_state & I_DIRTY; in __writeback_single_inode()
1623 inode->i_state &= ~dirty; in __writeback_single_inode()
1639 inode->i_state |= I_DIRTY_PAGES; in __writeback_single_inode()
1640 else if (unlikely(inode->i_state & I_PINNING_FSCACHE_WB)) { in __writeback_single_inode()
1641 if (!(inode->i_state & I_DIRTY_PAGES)) { in __writeback_single_inode()
1642 inode->i_state &= ~I_PINNING_FSCACHE_WB; in __writeback_single_inode()
1648 spin_unlock(&inode->i_lock); in __writeback_single_inode()
1652 int err = write_inode(inode, wbc); in __writeback_single_inode()
1657 trace_writeback_single_inode(inode, wbc, nr_to_write); in __writeback_single_inode()
1670 static int writeback_single_inode(struct inode *inode, in writeback_single_inode() argument
1676 spin_lock(&inode->i_lock); in writeback_single_inode()
1677 if (!atomic_read(&inode->i_count)) in writeback_single_inode()
1678 WARN_ON(!(inode->i_state & (I_WILL_FREE|I_FREEING))); in writeback_single_inode()
1680 WARN_ON(inode->i_state & I_WILL_FREE); in writeback_single_inode()
1682 if (inode->i_state & I_SYNC) { in writeback_single_inode()
1691 __inode_wait_for_writeback(inode); in writeback_single_inode()
1693 WARN_ON(inode->i_state & I_SYNC); in writeback_single_inode()
1701 if (!(inode->i_state & I_DIRTY_ALL) && in writeback_single_inode()
1703 !mapping_tagged(inode->i_mapping, PAGECACHE_TAG_WRITEBACK))) in writeback_single_inode()
1705 inode->i_state |= I_SYNC; in writeback_single_inode()
1706 wbc_attach_and_unlock_inode(wbc, inode); in writeback_single_inode()
1708 ret = __writeback_single_inode(inode, wbc); in writeback_single_inode()
1712 wb = inode_to_wb_and_lock_list(inode); in writeback_single_inode()
1713 spin_lock(&inode->i_lock); in writeback_single_inode()
1718 if (!(inode->i_state & I_FREEING)) { in writeback_single_inode()
1724 if (!(inode->i_state & I_DIRTY_ALL)) in writeback_single_inode()
1725 inode_cgwb_move_to_attached(inode, wb); in writeback_single_inode()
1726 else if (!(inode->i_state & I_SYNC_QUEUED)) { in writeback_single_inode()
1727 if ((inode->i_state & I_DIRTY)) in writeback_single_inode()
1728 redirty_tail_locked(inode, wb); in writeback_single_inode()
1729 else if (inode->i_state & I_DIRTY_TIME) { in writeback_single_inode()
1730 inode->dirtied_when = jiffies; in writeback_single_inode()
1731 inode_io_list_move_locked(inode, in writeback_single_inode()
1739 inode_sync_complete(inode); in writeback_single_inode()
1741 spin_unlock(&inode->i_lock); in writeback_single_inode()
1804 struct inode *inode = wb_inode(wb->b_io.prev); in writeback_sb_inodes() local
1808 if (inode->i_sb != sb) { in writeback_sb_inodes()
1815 redirty_tail(inode, wb); in writeback_sb_inodes()
1832 spin_lock(&inode->i_lock); in writeback_sb_inodes()
1833 if (inode->i_state & (I_NEW | I_FREEING | I_WILL_FREE)) { in writeback_sb_inodes()
1834 redirty_tail_locked(inode, wb); in writeback_sb_inodes()
1835 spin_unlock(&inode->i_lock); in writeback_sb_inodes()
1838 if ((inode->i_state & I_SYNC) && wbc.sync_mode != WB_SYNC_ALL) { in writeback_sb_inodes()
1848 requeue_io(inode, wb); in writeback_sb_inodes()
1849 spin_unlock(&inode->i_lock); in writeback_sb_inodes()
1850 trace_writeback_sb_inodes_requeue(inode); in writeback_sb_inodes()
1860 if (inode->i_state & I_SYNC) { in writeback_sb_inodes()
1862 inode_sleep_on_writeback(inode); in writeback_sb_inodes()
1867 inode->i_state |= I_SYNC; in writeback_sb_inodes()
1868 wbc_attach_and_unlock_inode(&wbc, inode); in writeback_sb_inodes()
1878 __writeback_single_inode(inode, &wbc); in writeback_sb_inodes()
1903 tmp_wb = inode_to_wb_and_lock_list(inode); in writeback_sb_inodes()
1904 spin_lock(&inode->i_lock); in writeback_sb_inodes()
1905 if (!(inode->i_state & I_DIRTY_ALL)) in writeback_sb_inodes()
1907 requeue_inode(inode, tmp_wb, &wbc); in writeback_sb_inodes()
1908 inode_sync_complete(inode); in writeback_sb_inodes()
1909 spin_unlock(&inode->i_lock); in writeback_sb_inodes()
1937 struct inode *inode = wb_inode(wb->b_io.prev); in __writeback_inodes_wb() local
1938 struct super_block *sb = inode->i_sb; in __writeback_inodes_wb()
1946 redirty_tail(inode, wb); in __writeback_inodes_wb()
2006 struct inode *inode; in wb_writeback() local
2078 inode = wb_inode(wb->b_more_io.prev); in wb_writeback()
2079 spin_lock(&inode->i_lock); in wb_writeback()
2082 inode_sleep_on_writeback(inode); in wb_writeback()
2376 void __mark_inode_dirty(struct inode *inode, int flags) in __mark_inode_dirty() argument
2378 struct super_block *sb = inode->i_sb; in __mark_inode_dirty()
2382 trace_writeback_mark_inode_dirty(inode, flags); in __mark_inode_dirty()
2390 if (inode->i_state & I_DIRTY_TIME) { in __mark_inode_dirty()
2391 spin_lock(&inode->i_lock); in __mark_inode_dirty()
2392 if (inode->i_state & I_DIRTY_TIME) { in __mark_inode_dirty()
2393 inode->i_state &= ~I_DIRTY_TIME; in __mark_inode_dirty()
2396 spin_unlock(&inode->i_lock); in __mark_inode_dirty()
2406 trace_writeback_dirty_inode_start(inode, flags); in __mark_inode_dirty()
2408 sb->s_op->dirty_inode(inode, in __mark_inode_dirty()
2410 trace_writeback_dirty_inode(inode, flags); in __mark_inode_dirty()
2430 if ((inode->i_state & flags) == flags) in __mark_inode_dirty()
2433 spin_lock(&inode->i_lock); in __mark_inode_dirty()
2434 if ((inode->i_state & flags) != flags) { in __mark_inode_dirty()
2435 const int was_dirty = inode->i_state & I_DIRTY; in __mark_inode_dirty()
2437 inode_attach_wb(inode, NULL); in __mark_inode_dirty()
2439 inode->i_state |= flags; in __mark_inode_dirty()
2448 wb = locked_inode_to_wb_and_lock_list(inode); in __mark_inode_dirty()
2449 spin_lock(&inode->i_lock); in __mark_inode_dirty()
2458 if (inode->i_state & I_SYNC_QUEUED) in __mark_inode_dirty()
2465 if (!S_ISBLK(inode->i_mode)) { in __mark_inode_dirty()
2466 if (inode_unhashed(inode)) in __mark_inode_dirty()
2469 if (inode->i_state & I_FREEING) in __mark_inode_dirty()
2480 inode->dirtied_when = jiffies; in __mark_inode_dirty()
2482 inode->dirtied_time_when = jiffies; in __mark_inode_dirty()
2484 if (inode->i_state & I_DIRTY) in __mark_inode_dirty()
2489 wakeup_bdi = inode_io_list_move_locked(inode, wb, in __mark_inode_dirty()
2493 spin_unlock(&inode->i_lock); in __mark_inode_dirty()
2494 trace_writeback_dirty_inode_enqueue(inode); in __mark_inode_dirty()
2511 spin_unlock(&inode->i_lock); in __mark_inode_dirty()
2557 struct inode *inode = list_first_entry(&sync_list, struct inode, in wait_sb_inodes() local
2559 struct address_space *mapping = inode->i_mapping; in wait_sb_inodes()
2567 list_move_tail(&inode->i_wb_list, &sb->s_inodes_wb); in wait_sb_inodes()
2579 spin_lock(&inode->i_lock); in wait_sb_inodes()
2580 if (inode->i_state & (I_FREEING|I_WILL_FREE|I_NEW)) { in wait_sb_inodes()
2581 spin_unlock(&inode->i_lock); in wait_sb_inodes()
2586 __iget(inode); in wait_sb_inodes()
2587 spin_unlock(&inode->i_lock); in wait_sb_inodes()
2599 iput(inode); in wait_sb_inodes()
2731 int write_inode_now(struct inode *inode, int sync) in write_inode_now() argument
2740 if (!mapping_can_writeback(inode->i_mapping)) in write_inode_now()
2744 return writeback_single_inode(inode, &wbc); in write_inode_now()
2757 int sync_inode_metadata(struct inode *inode, int wait) in sync_inode_metadata() argument
2764 return writeback_single_inode(inode, &wbc); in sync_inode_metadata()