Lines Matching refs:mapping
514 int sync_mapping_buffers(struct address_space *mapping) in sync_mapping_buffers() argument
516 struct address_space *buffer_mapping = mapping->private_data; in sync_mapping_buffers()
518 if (buffer_mapping == NULL || list_empty(&mapping->private_list)) in sync_mapping_buffers()
522 &mapping->private_list); in sync_mapping_buffers()
545 struct address_space *mapping = inode->i_mapping; in mark_buffer_dirty_inode() local
546 struct address_space *buffer_mapping = bh->b_page->mapping; in mark_buffer_dirty_inode()
549 if (!mapping->private_data) { in mark_buffer_dirty_inode()
550 mapping->private_data = buffer_mapping; in mark_buffer_dirty_inode()
552 BUG_ON(mapping->private_data != buffer_mapping); in mark_buffer_dirty_inode()
557 &mapping->private_list); in mark_buffer_dirty_inode()
558 bh->b_assoc_map = mapping; in mark_buffer_dirty_inode()
573 void __set_page_dirty(struct page *page, struct address_space *mapping, in __set_page_dirty() argument
578 xa_lock_irqsave(&mapping->i_pages, flags); in __set_page_dirty()
579 if (page->mapping) { /* Race with truncate? */ in __set_page_dirty()
581 account_page_dirtied(page, mapping); in __set_page_dirty()
582 radix_tree_tag_set(&mapping->i_pages, in __set_page_dirty()
585 xa_unlock_irqrestore(&mapping->i_pages, flags); in __set_page_dirty()
617 struct address_space *mapping = page_mapping(page); in __set_page_dirty_buffers() local
619 if (unlikely(!mapping)) in __set_page_dirty_buffers()
622 spin_lock(&mapping->private_lock); in __set_page_dirty_buffers()
638 spin_unlock(&mapping->private_lock); in __set_page_dirty_buffers()
641 __set_page_dirty(page, mapping, 1); in __set_page_dirty_buffers()
646 __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); in __set_page_dirty_buffers()
675 struct address_space *mapping; in fsync_buffers_list() local
685 mapping = bh->b_assoc_map; in fsync_buffers_list()
692 bh->b_assoc_map = mapping; in fsync_buffers_list()
724 mapping = bh->b_assoc_map; in fsync_buffers_list()
731 &mapping->private_list); in fsync_buffers_list()
732 bh->b_assoc_map = mapping; in fsync_buffers_list()
762 struct address_space *mapping = &inode->i_data; in invalidate_inode_buffers() local
763 struct list_head *list = &mapping->private_list; in invalidate_inode_buffers()
764 struct address_space *buffer_mapping = mapping->private_data; in invalidate_inode_buffers()
785 struct address_space *mapping = &inode->i_data; in remove_inode_buffers() local
786 struct list_head *list = &mapping->private_list; in remove_inode_buffers()
787 struct address_space *buffer_mapping = mapping->private_data; in remove_inode_buffers()
1104 struct address_space *mapping = NULL; in mark_buffer_dirty() local
1108 mapping = page_mapping(page); in mark_buffer_dirty()
1109 if (mapping) in mark_buffer_dirty()
1110 __set_page_dirty(page, mapping, 0); in mark_buffer_dirty()
1113 if (mapping) in mark_buffer_dirty()
1114 __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); in mark_buffer_dirty()
1123 if (bh->b_page && bh->b_page->mapping) in mark_buffer_write_io_error()
1124 mapping_set_error(bh->b_page->mapping, -EIO); in mark_buffer_write_io_error()
1155 struct address_space *buffer_mapping = bh->b_page->mapping; in __bforget()
1525 spin_lock(&page->mapping->private_lock); in create_empty_buffers()
1537 spin_unlock(&page->mapping->private_lock); in create_empty_buffers()
1814 mapping_set_error(page->mapping, err); in __block_write_full_page()
1933 struct inode *inode = page->mapping->host; in __block_write_begin_int()
2064 int block_write_begin(struct address_space *mapping, loff_t pos, unsigned len, in block_write_begin() argument
2071 page = grab_cache_page_write_begin(mapping, index, flags); in block_write_begin()
2121 int block_write_end(struct file *file, struct address_space *mapping, in block_write_end() argument
2125 struct inode *inode = mapping->host; in block_write_end()
2157 int generic_write_end(struct file *file, struct address_space *mapping, in generic_write_end() argument
2161 copied = block_write_end(file, mapping, pos, len, copied, page, fsdata); in generic_write_end()
2162 return __generic_write_end(mapping->host, pos, copied, page); in generic_write_end()
2220 struct inode *inode = page->mapping->host; in block_read_full_page()
2310 struct address_space *mapping = inode->i_mapping; in generic_cont_expand_simple() local
2319 err = pagecache_write_begin(NULL, mapping, size, 0, in generic_cont_expand_simple()
2324 err = pagecache_write_end(NULL, mapping, size, 0, 0, page, fsdata); in generic_cont_expand_simple()
2332 static int cont_expand_zero(struct file *file, struct address_space *mapping, in cont_expand_zero() argument
2335 struct inode *inode = mapping->host; in cont_expand_zero()
2355 err = pagecache_write_begin(file, mapping, curpos, len, 0, in cont_expand_zero()
2360 err = pagecache_write_end(file, mapping, curpos, len, len, in cont_expand_zero()
2367 balance_dirty_pages_ratelimited(mapping); in cont_expand_zero()
2388 err = pagecache_write_begin(file, mapping, curpos, len, 0, in cont_expand_zero()
2393 err = pagecache_write_end(file, mapping, curpos, len, len, in cont_expand_zero()
2408 int cont_write_begin(struct file *file, struct address_space *mapping, in cont_write_begin() argument
2413 struct inode *inode = mapping->host; in cont_write_begin()
2418 err = cont_expand_zero(file, mapping, pos, bytes); in cont_write_begin()
2428 return block_write_begin(mapping, pos, len, flags, pagep, get_block); in cont_write_begin()
2434 struct inode *inode = page->mapping->host; in block_commit_write()
2469 if ((page->mapping != inode->i_mapping) || in block_page_mkwrite()
2518 spin_lock(&page->mapping->private_lock); in attach_nobh_buffers()
2528 spin_unlock(&page->mapping->private_lock); in attach_nobh_buffers()
2536 int nobh_write_begin(struct address_space *mapping, in nobh_write_begin() argument
2541 struct inode *inode = mapping->host; in nobh_write_begin()
2559 page = grab_cache_page_write_begin(mapping, index, flags); in nobh_write_begin()
2677 int nobh_write_end(struct file *file, struct address_space *mapping, in nobh_write_end() argument
2681 struct inode *inode = page->mapping->host; in nobh_write_end()
2689 return generic_write_end(file, mapping, pos, len, in nobh_write_end()
2720 struct inode * const inode = page->mapping->host; in nobh_writepage()
2740 if (page->mapping->a_ops->invalidatepage) in nobh_writepage()
2741 page->mapping->a_ops->invalidatepage(page, offset); in nobh_writepage()
2764 int nobh_truncate_page(struct address_space *mapping, in nobh_truncate_page() argument
2772 struct inode *inode = mapping->host; in nobh_truncate_page()
2787 page = grab_cache_page(mapping, index); in nobh_truncate_page()
2796 return block_truncate_page(mapping, from, get_block); in nobh_truncate_page()
2817 err = mapping->a_ops->readpage(NULL, page); in nobh_truncate_page()
2842 int block_truncate_page(struct address_space *mapping, in block_truncate_page() argument
2850 struct inode *inode = mapping->host; in block_truncate_page()
2865 page = grab_cache_page(mapping, index); in block_truncate_page()
2924 struct inode * const inode = page->mapping->host; in block_write_full_page()
2960 sector_t generic_block_bmap(struct address_space *mapping, sector_t block, in generic_block_bmap() argument
2963 struct inode *inode = mapping->host; in generic_block_bmap()
3251 struct address_space * const mapping = page->mapping; in try_to_free_buffers() local
3259 if (mapping == NULL) { /* can this still happen? */ in try_to_free_buffers()
3264 spin_lock(&mapping->private_lock); in try_to_free_buffers()
3283 spin_unlock(&mapping->private_lock); in try_to_free_buffers()