Lines Matching refs:fio
383 struct f2fs_io_info fio = { in __f2fs_commit_inmem_pages() local
412 fio.page = page; in __f2fs_commit_inmem_pages()
413 fio.old_blkaddr = NULL_ADDR; in __f2fs_commit_inmem_pages()
414 fio.encrypted_page = NULL; in __f2fs_commit_inmem_pages()
415 fio.need_lock = LOCK_DONE; in __f2fs_commit_inmem_pages()
416 err = f2fs_do_write_data_page(&fio); in __f2fs_commit_inmem_pages()
428 cur->old_addr = fio.old_blkaddr; in __f2fs_commit_inmem_pages()
3221 static int __get_segment_type_2(struct f2fs_io_info *fio) in __get_segment_type_2() argument
3223 if (fio->type == DATA) in __get_segment_type_2()
3229 static int __get_segment_type_4(struct f2fs_io_info *fio) in __get_segment_type_4() argument
3231 if (fio->type == DATA) { in __get_segment_type_4()
3232 struct inode *inode = fio->page->mapping->host; in __get_segment_type_4()
3239 if (IS_DNODE(fio->page) && is_cold_node(fio->page)) in __get_segment_type_4()
3246 static int __get_segment_type_6(struct f2fs_io_info *fio) in __get_segment_type_6() argument
3248 if (fio->type == DATA) { in __get_segment_type_6()
3249 struct inode *inode = fio->page->mapping->host; in __get_segment_type_6()
3251 if (is_cold_data(fio->page)) { in __get_segment_type_6()
3252 if (fio->sbi->am.atgc_enabled) in __get_segment_type_6()
3266 if (IS_DNODE(fio->page)) in __get_segment_type_6()
3267 return is_cold_node(fio->page) ? CURSEG_WARM_NODE : in __get_segment_type_6()
3273 static int __get_segment_type(struct f2fs_io_info *fio) in __get_segment_type() argument
3277 switch (F2FS_OPTION(fio->sbi).active_logs) { in __get_segment_type()
3279 type = __get_segment_type_2(fio); in __get_segment_type()
3282 type = __get_segment_type_4(fio); in __get_segment_type()
3285 type = __get_segment_type_6(fio); in __get_segment_type()
3288 f2fs_bug_on(fio->sbi, true); in __get_segment_type()
3292 fio->temp = HOT; in __get_segment_type()
3294 fio->temp = WARM; in __get_segment_type()
3296 fio->temp = COLD; in __get_segment_type()
3303 struct f2fs_io_info *fio) in f2fs_allocate_data_block() argument
3379 fio->retry = false; in f2fs_allocate_data_block()
3381 if (fio) { in f2fs_allocate_data_block()
3384 INIT_LIST_HEAD(&fio->list); in f2fs_allocate_data_block()
3385 fio->in_list = true; in f2fs_allocate_data_block()
3386 io = sbi->write_io[fio->type] + fio->temp; in f2fs_allocate_data_block()
3388 list_add_tail(&fio->list, &io->io_list); in f2fs_allocate_data_block()
3397 static void update_device_state(struct f2fs_io_info *fio) in update_device_state() argument
3399 struct f2fs_sb_info *sbi = fio->sbi; in update_device_state()
3405 devidx = f2fs_target_device_index(sbi, fio->new_blkaddr); in update_device_state()
3408 f2fs_set_dirty_device(sbi, fio->ino, devidx, FLUSH_INO); in update_device_state()
3418 static void do_write_page(struct f2fs_summary *sum, struct f2fs_io_info *fio) in do_write_page() argument
3420 int type = __get_segment_type(fio); in do_write_page()
3421 bool keep_order = (f2fs_lfs_mode(fio->sbi) && type == CURSEG_COLD_DATA); in do_write_page()
3424 down_read(&fio->sbi->io_order_lock); in do_write_page()
3426 f2fs_allocate_data_block(fio->sbi, fio->page, fio->old_blkaddr, in do_write_page()
3427 &fio->new_blkaddr, sum, type, fio); in do_write_page()
3428 if (GET_SEGNO(fio->sbi, fio->old_blkaddr) != NULL_SEGNO) in do_write_page()
3429 invalidate_mapping_pages(META_MAPPING(fio->sbi), in do_write_page()
3430 fio->old_blkaddr, fio->old_blkaddr); in do_write_page()
3433 f2fs_submit_page_write(fio); in do_write_page()
3434 if (fio->retry) { in do_write_page()
3435 fio->old_blkaddr = fio->new_blkaddr; in do_write_page()
3439 update_device_state(fio); in do_write_page()
3442 up_read(&fio->sbi->io_order_lock); in do_write_page()
3448 struct f2fs_io_info fio = { in f2fs_do_write_meta_page() local
3462 fio.op_flags &= ~REQ_META; in f2fs_do_write_meta_page()
3466 f2fs_submit_page_write(&fio); in f2fs_do_write_meta_page()
3472 void f2fs_do_write_node_page(unsigned int nid, struct f2fs_io_info *fio) in f2fs_do_write_node_page() argument
3477 do_write_page(&sum, fio); in f2fs_do_write_node_page()
3479 f2fs_update_iostat(fio->sbi, fio->io_type, F2FS_BLKSIZE); in f2fs_do_write_node_page()
3483 struct f2fs_io_info *fio) in f2fs_outplace_write_data() argument
3485 struct f2fs_sb_info *sbi = fio->sbi; in f2fs_outplace_write_data()
3489 set_summary(&sum, dn->nid, dn->ofs_in_node, fio->version); in f2fs_outplace_write_data()
3490 do_write_page(&sum, fio); in f2fs_outplace_write_data()
3491 f2fs_update_data_blkaddr(dn, fio->new_blkaddr); in f2fs_outplace_write_data()
3493 f2fs_update_iostat(sbi, fio->io_type, F2FS_BLKSIZE); in f2fs_outplace_write_data()
3496 int f2fs_inplace_write_data(struct f2fs_io_info *fio) in f2fs_inplace_write_data() argument
3499 struct f2fs_sb_info *sbi = fio->sbi; in f2fs_inplace_write_data()
3502 fio->new_blkaddr = fio->old_blkaddr; in f2fs_inplace_write_data()
3504 __get_segment_type(fio); in f2fs_inplace_write_data()
3506 segno = GET_SEGNO(sbi, fio->new_blkaddr); in f2fs_inplace_write_data()
3515 stat_inc_inplace_blocks(fio->sbi); in f2fs_inplace_write_data()
3517 if (fio->bio && !(SM_I(sbi)->ipu_policy & (1 << F2FS_IPU_NOCACHE))) in f2fs_inplace_write_data()
3518 err = f2fs_merge_page_bio(fio); in f2fs_inplace_write_data()
3520 err = f2fs_submit_page_bio(fio); in f2fs_inplace_write_data()
3522 update_device_state(fio); in f2fs_inplace_write_data()
3523 f2fs_update_iostat(fio->sbi, fio->io_type, F2FS_BLKSIZE); in f2fs_inplace_write_data()