Home
last modified time | relevance | path

Searched refs:cached_state (Results 1 – 24 of 24) sorted by relevance

/Linux-v5.10/fs/btrfs/
Dextent-io-tree.h113 struct extent_state *cached_state);
158 struct extent_state **cached_state, gfp_t mask);
169 u64 end, struct extent_state **cached_state) in clear_extent_uptodate() argument
172 cached_state, GFP_NOFS, NULL); in clear_extent_uptodate()
192 struct extent_state **cached_state);
196 struct extent_state **cached_state) in set_extent_delalloc() argument
200 NULL, cached_state, GFP_NOFS); in set_extent_delalloc()
204 u64 end, struct extent_state **cached_state) in set_extent_defrag() argument
208 NULL, cached_state, GFP_NOFS); in set_extent_defrag()
219 u64 end, struct extent_state **cached_state, gfp_t mask) in set_extent_uptodate() argument
[all …]
Dextent_io.c699 struct extent_state **cached_state, in __clear_extent_bit() argument
734 if (cached_state) { in __clear_extent_bit()
735 cached = *cached_state; in __clear_extent_bit()
738 *cached_state = NULL; in __clear_extent_bit()
739 cached_state = NULL; in __clear_extent_bit()
968 u64 *failed_start, struct extent_state **cached_state, in __set_extent_bit() argument
996 if (cached_state && *cached_state) { in __set_extent_bit()
997 state = *cached_state; in __set_extent_bit()
1017 cache_state(prealloc, cached_state); in __set_extent_bit()
1040 cache_state(state, cached_state); in __set_extent_bit()
[all …]
Dfile.c1410 struct extent_state **cached_state) in lock_and_cleanup_extent_if_need() argument
1425 cached_state); in lock_and_cleanup_extent_if_need()
1432 last_pos, cached_state); in lock_and_cleanup_extent_if_need()
1588 struct extent_state *cached_state = NULL; in btrfs_buffered_write() local
1677 &lockend, &cached_state); in btrfs_buffered_write()
1736 &cached_state); in btrfs_buffered_write()
1747 lockstart, lockend, &cached_state); in btrfs_buffered_write()
1749 free_extent_state(cached_state); in btrfs_buffered_write()
2472 struct extent_state **cached_state) in btrfs_punch_hole_lock_range() argument
2481 cached_state); in btrfs_punch_hole_lock_range()
[all …]
Dinode.c2259 struct extent_state **cached_state) in btrfs_find_new_delalloc_bytes() argument
2286 NULL, cached_state, GFP_NOFS); in btrfs_find_new_delalloc_bytes()
2298 struct extent_state **cached_state) in btrfs_set_extent_delalloc() argument
2314 cached_state); in btrfs_set_extent_delalloc()
2320 cached_state); in btrfs_set_extent_delalloc()
2334 struct extent_state *cached_state = NULL; in btrfs_writepage_fixup_worker() local
2398 lock_extent_bits(&inode->io_tree, page_start, page_end, &cached_state); in btrfs_writepage_fixup_worker()
2407 &cached_state); in btrfs_writepage_fixup_worker()
2415 &cached_state); in btrfs_writepage_fixup_worker()
2434 &cached_state); in btrfs_writepage_fixup_worker()
[all …]
Dtransaction.c973 struct extent_state *cached_state = NULL; in btrfs_write_marked_extents() local
979 mark, &cached_state)) { in btrfs_write_marked_extents()
984 mark, &cached_state); in btrfs_write_marked_extents()
1008 free_extent_state(cached_state); in btrfs_write_marked_extents()
1009 cached_state = NULL; in btrfs_write_marked_extents()
1029 struct extent_state *cached_state = NULL; in __btrfs_wait_marked_extents() local
1034 EXTENT_NEED_WAIT, &cached_state)) { in __btrfs_wait_marked_extents()
1044 EXTENT_NEED_WAIT, 0, 0, &cached_state); in __btrfs_wait_marked_extents()
1051 free_extent_state(cached_state); in __btrfs_wait_marked_extents()
1052 cached_state = NULL; in __btrfs_wait_marked_extents()
Dordered-data.c917 struct extent_state **cached_state) in btrfs_lock_and_flush_ordered_range() argument
923 if (cached_state) in btrfs_lock_and_flush_ordered_range()
924 cachedp = cached_state; in btrfs_lock_and_flush_ordered_range()
936 if (!cached_state) in btrfs_lock_and_flush_ordered_range()
Dordered-data.h195 struct extent_state **cached_state);
Ddev-replace.c627 struct extent_state *cached_state = NULL; in btrfs_set_target_alloc_state() local
637 CHUNK_ALLOCATED, &cached_state)) { in btrfs_set_target_alloc_state()
645 free_extent_state(cached_state); in btrfs_set_target_alloc_state()
Dfree-space-cache.c1158 struct extent_state **cached_state) in cleanup_write_cache_enospc() argument
1162 i_size_read(inode) - 1, cached_state); in cleanup_write_cache_enospc()
1257 struct extent_state *cached_state = NULL; in __btrfs_write_out_cache() local
1293 &cached_state); in __btrfs_write_out_cache()
1335 &cached_state); in __btrfs_write_out_cache()
1349 i_size_read(inode) - 1, &cached_state); in __btrfs_write_out_cache()
1371 cleanup_write_cache_enospc(inode, io_ctl, &cached_state); in __btrfs_write_out_cache()
Ddisk-io.c241 struct extent_state *cached_state = NULL; in verify_parent_transid() local
257 &cached_state); in verify_parent_transid()
281 &cached_state); in verify_parent_transid()
4522 struct extent_state *cached_state = NULL; in btrfs_destroy_pinned_extent() local
4532 EXTENT_DIRTY, &cached_state); in btrfs_destroy_pinned_extent()
4538 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_destroy_pinned_extent()
4539 free_extent_state(cached_state); in btrfs_destroy_pinned_extent()
Dioctl.c1282 struct extent_state *cached_state = NULL; in cluster_pages_for_defrag() local
1313 &cached_state); in cluster_pages_for_defrag()
1317 &cached_state); in cluster_pages_for_defrag()
1373 page_start, page_end - 1, &cached_state); in cluster_pages_for_defrag()
1376 EXTENT_DEFRAG, 0, 0, &cached_state); in cluster_pages_for_defrag()
1388 &cached_state); in cluster_pages_for_defrag()
1391 page_start, page_end - 1, &cached_state); in cluster_pages_for_defrag()
Dextent-tree.c2895 struct extent_state *cached_state = NULL; in btrfs_finish_extent_commit() local
2899 EXTENT_DIRTY, &cached_state); in btrfs_finish_extent_commit()
2912 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_finish_extent_commit()
2915 free_extent_state(cached_state); in btrfs_finish_extent_commit()
Dtree-log.c3906 struct extent_state *cached_state = NULL; in log_csums() local
3924 lock_end, &cached_state); in log_csums()
3941 &cached_state); in log_csums()
Dctree.h3008 struct extent_state **cached_state);
/Linux-v5.10/drivers/gpu/drm/msm/dsi/pll/
Ddsi_pll_28nm_8960.c81 struct pll_28nm_cached_state cached_state; member
340 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_save_state() local
343 cached_state->postdiv3 = in dsi_pll_28nm_save_state()
345 cached_state->postdiv2 = in dsi_pll_28nm_save_state()
347 cached_state->postdiv1 = in dsi_pll_28nm_save_state()
350 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_28nm_save_state()
356 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_restore_state() local
361 cached_state->vco_rate, 0); in dsi_pll_28nm_restore_state()
369 cached_state->postdiv3); in dsi_pll_28nm_restore_state()
371 cached_state->postdiv2); in dsi_pll_28nm_restore_state()
[all …]
Ddsi_pll_28nm.c86 struct pll_28nm_cached_state cached_state; member
442 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_save_state() local
445 cached_state->postdiv3 = in dsi_pll_28nm_save_state()
447 cached_state->postdiv1 = in dsi_pll_28nm_save_state()
449 cached_state->byte_mux = pll_read(base + REG_DSI_28nm_PHY_PLL_VREG_CFG); in dsi_pll_28nm_save_state()
450 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_28nm_save_state()
456 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_restore_state() local
461 cached_state->vco_rate, 0); in dsi_pll_28nm_restore_state()
469 cached_state->postdiv3); in dsi_pll_28nm_restore_state()
471 cached_state->postdiv1); in dsi_pll_28nm_restore_state()
[all …]
Ddsi_pll_14nm.c144 struct pll_14nm_cached_state cached_state; member
798 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_pll_14nm_save_state() local
804 cached_state->n1postdiv = data & 0xf; in dsi_pll_14nm_save_state()
805 cached_state->n2postdiv = (data >> 4) & 0xf; in dsi_pll_14nm_save_state()
808 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_pll_14nm_save_state()
810 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_14nm_save_state()
816 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_pll_14nm_restore_state() local
822 cached_state->vco_rate, 0); in dsi_pll_14nm_restore_state()
829 data = cached_state->n1postdiv | (cached_state->n2postdiv << 4); in dsi_pll_14nm_restore_state()
832 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_pll_14nm_restore_state()
Ddsi_pll_7nm.c118 struct pll_7nm_cached_state cached_state; member
562 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_pll_7nm_save_state()
585 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_pll_7nm_restore_state()
Ddsi_pll_10nm.c118 struct pll_10nm_cached_state cached_state; member
536 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_pll_10nm_save_state()
559 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_pll_10nm_restore_state()
/Linux-v5.10/drivers/mux/
Dcore.c117 mux->cached_state = MUX_CACHE_UNKNOWN; in mux_chip_alloc()
131 mux->cached_state = ret < 0 ? MUX_CACHE_UNKNOWN : state; in mux_control_set()
155 if (mux->idle_state == mux->cached_state) in mux_chip_register()
303 if (mux->cached_state == state) in __mux_control_select()
399 mux->idle_state != mux->cached_state) in mux_control_deselect()
/Linux-v5.10/drivers/iio/multiplexer/
Diio-mux.c29 int cached_state; member
47 mux->cached_state = -1; in iio_mux_select()
51 if (mux->cached_state == chan->channel) in iio_mux_select()
70 mux->cached_state = -1; in iio_mux_select()
75 mux->cached_state = chan->channel; in iio_mux_select()
393 mux->cached_state = -1; in mux_probe()
/Linux-v5.10/include/linux/mux/
Ddriver.h46 int cached_state; member
/Linux-v5.10/drivers/gpu/drm/amd/display/amdgpu_dm/
Damdgpu_dm.h324 struct drm_atomic_state *cached_state; member
Damdgpu_dm.c1751 WARN_ON(adev->dm.cached_state); in dm_suspend()
1752 adev->dm.cached_state = drm_atomic_helper_suspend(adev_to_drm(adev)); in dm_suspend()
2018 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) in dm_resume()
2026 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) { in dm_resume()
2035 for_each_new_plane_in_state(dm->cached_state, plane, new_plane_state, i) { in dm_resume()
2044 drm_atomic_helper_resume(ddev, dm->cached_state); in dm_resume()
2046 dm->cached_state = NULL; in dm_resume()