Lines Matching refs:cached
583 struct extent_state *cached; in __clear_extent_bit() local
615 cached = *cached_state; in __clear_extent_bit()
622 if (cached && extent_state_in_tree(cached) && in __clear_extent_bit()
623 cached->start <= start && cached->end > start) { in __clear_extent_bit()
625 refcount_dec(&cached->refs); in __clear_extent_bit()
626 state = cached; in __clear_extent_bit()
630 free_extent_state(cached); in __clear_extent_bit()
1293 struct extent_state **cached) in clear_extent_bit() argument
1296 cached, GFP_NOFS, NULL); in clear_extent_bit()
1868 unsigned bits, int filled, struct extent_state *cached) in test_range_bit() argument
1875 if (cached && extent_state_in_tree(cached) && cached->start <= start && in test_range_bit()
1876 cached->end > start) in test_range_bit()
1877 node = &cached->rb_node; in test_range_bit()
2479 struct extent_state *cached = NULL; in endio_readpage_release_extent() local
2483 set_extent_uptodate(tree, start, end, &cached, GFP_ATOMIC); in endio_readpage_release_extent()
2484 unlock_extent_cached_atomic(tree, start, end, &cached); in endio_readpage_release_extent()
2924 struct extent_state *cached = NULL; in __do_readpage() local
2932 &cached, GFP_NOFS); in __do_readpage()
2934 cur + iosize - 1, &cached); in __do_readpage()
3017 struct extent_state *cached = NULL; in __do_readpage() local
3025 &cached, GFP_NOFS); in __do_readpage()
3027 cur + iosize - 1, &cached); in __do_readpage()
4315 bool cached; member
4334 if (!cache->cached) in emit_fiemap_extent()
4372 cache->cached = false; in emit_fiemap_extent()
4376 cache->cached = true; in emit_fiemap_extent()
4385 cache->cached = false; in emit_fiemap_extent()
4407 if (!cache->cached) in emit_last_fiemap_cache()
4412 cache->cached = false; in emit_last_fiemap_cache()