Lines Matching refs:work

177 				  struct wb_writeback_work *work)  in finish_writeback_work()  argument
179 struct wb_completion *done = work->done; in finish_writeback_work()
181 if (work->auto_free) in finish_writeback_work()
182 kfree(work); in finish_writeback_work()
188 struct wb_writeback_work *work) in wb_queue_work() argument
190 trace_writeback_queue(wb, work); in wb_queue_work()
192 if (work->done) in wb_queue_work()
193 atomic_inc(&work->done->cnt); in wb_queue_work()
198 list_add_tail(&work->list, &wb->work_list); in wb_queue_work()
201 finish_writeback_work(wb, work); in wb_queue_work()
331 struct work_struct work; member
334 static void inode_switch_wbs_work_fn(struct work_struct *work) in inode_switch_wbs_work_fn() argument
337 container_of(work, struct inode_switch_wbs_context, work); in inode_switch_wbs_work_fn()
456 INIT_WORK(&isw->work, inode_switch_wbs_work_fn); in inode_switch_wbs_rcu_fn()
457 queue_work(isw_wq, &isw->work); in inode_switch_wbs_rcu_fn()
814 struct wb_writeback_work *work; in bdi_split_work_to_wbs() local
832 work = kmalloc(sizeof(*work), GFP_ATOMIC); in bdi_split_work_to_wbs()
833 if (work) { in bdi_split_work_to_wbs()
834 *work = *base_work; in bdi_split_work_to_wbs()
835 work->nr_pages = nr_pages; in bdi_split_work_to_wbs()
836 work->auto_free = 1; in bdi_split_work_to_wbs()
837 wb_queue_work(wb, work); in bdi_split_work_to_wbs()
842 work = &fallback_work; in bdi_split_work_to_wbs()
843 *work = *base_work; in bdi_split_work_to_wbs()
844 work->nr_pages = nr_pages; in bdi_split_work_to_wbs()
845 work->auto_free = 0; in bdi_split_work_to_wbs()
846 work->done = &fallback_work_done; in bdi_split_work_to_wbs()
848 wb_queue_work(wb, work); in bdi_split_work_to_wbs()
1100 struct wb_writeback_work *work) in move_expired_inodes() argument
1112 older_than_this = work->older_than_this; in move_expired_inodes()
1113 else if (!work->for_sync) { in move_expired_inodes()
1163 static void queue_io(struct bdi_writeback *wb, struct wb_writeback_work *work) in queue_io() argument
1169 moved = move_expired_inodes(&wb->b_dirty, &wb->b_io, 0, work); in queue_io()
1171 EXPIRE_DIRTY_ATIME, work); in queue_io()
1174 trace_writeback_queue_io(wb, work, moved); in queue_io()
1456 struct wb_writeback_work *work) in writeback_chunk_size() argument
1473 if (work->sync_mode == WB_SYNC_ALL || work->tagged_writepages) in writeback_chunk_size()
1478 pages = min(pages, work->nr_pages); in writeback_chunk_size()
1497 struct wb_writeback_work *work) in writeback_sb_inodes() argument
1500 .sync_mode = work->sync_mode, in writeback_sb_inodes()
1501 .tagged_writepages = work->tagged_writepages, in writeback_sb_inodes()
1502 .for_kupdate = work->for_kupdate, in writeback_sb_inodes()
1503 .for_background = work->for_background, in writeback_sb_inodes()
1504 .for_sync = work->for_sync, in writeback_sb_inodes()
1505 .range_cyclic = work->range_cyclic, in writeback_sb_inodes()
1518 if (work->sb) { in writeback_sb_inodes()
1579 write_chunk = writeback_chunk_size(wb, work); in writeback_sb_inodes()
1590 work->nr_pages -= write_chunk - wbc.nr_to_write; in writeback_sb_inodes()
1630 if (work->nr_pages <= 0) in writeback_sb_inodes()
1638 struct wb_writeback_work *work) in __writeback_inodes_wb() argument
1656 wrote += writeback_sb_inodes(sb, wb, work); in __writeback_inodes_wb()
1663 if (work->nr_pages <= 0) in __writeback_inodes_wb()
1674 struct wb_writeback_work work = { in writeback_inodes_wb() local
1685 queue_io(wb, &work); in writeback_inodes_wb()
1686 __writeback_inodes_wb(wb, &work); in writeback_inodes_wb()
1690 return nr_pages - work.nr_pages; in writeback_inodes_wb()
1709 struct wb_writeback_work *work) in wb_writeback() argument
1712 long nr_pages = work->nr_pages; in wb_writeback()
1719 work->older_than_this = &oldest_jif; in wb_writeback()
1727 if (work->nr_pages <= 0) in wb_writeback()
1736 if ((work->for_background || work->for_kupdate) && in wb_writeback()
1744 if (work->for_background && !wb_over_bg_thresh(wb)) in wb_writeback()
1753 if (work->for_kupdate) { in wb_writeback()
1756 } else if (work->for_background) in wb_writeback()
1759 trace_writeback_start(wb, work); in wb_writeback()
1761 queue_io(wb, work); in wb_writeback()
1762 if (work->sb) in wb_writeback()
1763 progress = writeback_sb_inodes(work->sb, wb, work); in wb_writeback()
1765 progress = __writeback_inodes_wb(wb, work); in wb_writeback()
1766 trace_writeback_written(wb, work); in wb_writeback()
1790 trace_writeback_wait(wb, work); in wb_writeback()
1801 return nr_pages - work->nr_pages; in wb_writeback()
1809 struct wb_writeback_work *work = NULL; in get_next_work_item() local
1813 work = list_entry(wb->work_list.next, in get_next_work_item()
1815 list_del_init(&work->list); in get_next_work_item()
1818 return work; in get_next_work_item()
1825 struct wb_writeback_work work = { in wb_check_background_flush() local
1833 return wb_writeback(wb, &work); in wb_check_background_flush()
1859 struct wb_writeback_work work = { in wb_check_old_data_flush() local
1867 return wb_writeback(wb, &work); in wb_check_old_data_flush()
1882 struct wb_writeback_work work = { in wb_check_start_all() local
1889 nr_pages = wb_writeback(wb, &work); in wb_check_start_all()
1902 struct wb_writeback_work *work; in wb_do_writeback() local
1906 while ((work = get_next_work_item(wb)) != NULL) { in wb_do_writeback()
1907 trace_writeback_exec(wb, work); in wb_do_writeback()
1908 wrote += wb_writeback(wb, work); in wb_do_writeback()
1909 finish_writeback_work(wb, work); in wb_do_writeback()
1931 void wb_workfn(struct work_struct *work) in wb_workfn() argument
1933 struct bdi_writeback *wb = container_of(to_delayed_work(work), in wb_workfn()
2325 struct wb_writeback_work work = { in __writeback_inodes_sb_nr() local
2339 bdi_split_work_to_wbs(sb->s_bdi, &work, skip_if_busy); in __writeback_inodes_sb_nr()
2403 struct wb_writeback_work work = { in sync_inodes_sb() local
2423 bdi_split_work_to_wbs(bdi, &work, false); in sync_inodes_sb()