Home
last modified time | relevance | path

Searched refs:cached_state (Results 1 – 21 of 21) sorted by relevance

/Linux-v5.4/drivers/gpu/drm/msm/dsi/pll/
Ddsi_pll_28nm_8960.c81 struct pll_28nm_cached_state cached_state; member
340 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_save_state() local
343 cached_state->postdiv3 = in dsi_pll_28nm_save_state()
345 cached_state->postdiv2 = in dsi_pll_28nm_save_state()
347 cached_state->postdiv1 = in dsi_pll_28nm_save_state()
350 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_28nm_save_state()
356 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_restore_state() local
361 cached_state->vco_rate, 0); in dsi_pll_28nm_restore_state()
369 cached_state->postdiv3); in dsi_pll_28nm_restore_state()
371 cached_state->postdiv2); in dsi_pll_28nm_restore_state()
[all …]
Ddsi_pll_28nm.c86 struct pll_28nm_cached_state cached_state; member
442 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_save_state() local
445 cached_state->postdiv3 = in dsi_pll_28nm_save_state()
447 cached_state->postdiv1 = in dsi_pll_28nm_save_state()
449 cached_state->byte_mux = pll_read(base + REG_DSI_28nm_PHY_PLL_VREG_CFG); in dsi_pll_28nm_save_state()
450 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_28nm_save_state()
456 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_restore_state() local
461 cached_state->vco_rate, 0); in dsi_pll_28nm_restore_state()
469 cached_state->postdiv3); in dsi_pll_28nm_restore_state()
471 cached_state->postdiv1); in dsi_pll_28nm_restore_state()
[all …]
Ddsi_pll_14nm.c144 struct pll_14nm_cached_state cached_state; member
798 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_pll_14nm_save_state() local
804 cached_state->n1postdiv = data & 0xf; in dsi_pll_14nm_save_state()
805 cached_state->n2postdiv = (data >> 4) & 0xf; in dsi_pll_14nm_save_state()
808 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_pll_14nm_save_state()
810 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_14nm_save_state()
816 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_pll_14nm_restore_state() local
822 cached_state->vco_rate, 0); in dsi_pll_14nm_restore_state()
829 data = cached_state->n1postdiv | (cached_state->n2postdiv << 4); in dsi_pll_14nm_restore_state()
832 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_pll_14nm_restore_state()
Ddsi_pll_10nm.c118 struct pll_10nm_cached_state cached_state; member
530 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_pll_10nm_save_state()
553 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_pll_10nm_restore_state()
/Linux-v5.4/fs/btrfs/
Dextent_io.h289 struct extent_state *cached_state);
334 struct extent_state **cached_state, gfp_t mask);
345 u64 end, struct extent_state **cached_state) in clear_extent_uptodate() argument
348 cached_state, GFP_NOFS, NULL); in clear_extent_uptodate()
368 struct extent_state **cached_state);
372 struct extent_state **cached_state) in set_extent_delalloc() argument
376 NULL, cached_state, GFP_NOFS); in set_extent_delalloc()
380 u64 end, struct extent_state **cached_state) in set_extent_defrag() argument
384 NULL, cached_state, GFP_NOFS); in set_extent_defrag()
395 u64 end, struct extent_state **cached_state, gfp_t mask) in set_extent_uptodate() argument
[all …]
Dextent_io.c670 struct extent_state **cached_state, in __clear_extent_bit() argument
705 if (cached_state) { in __clear_extent_bit()
706 cached = *cached_state; in __clear_extent_bit()
709 *cached_state = NULL; in __clear_extent_bit()
710 cached_state = NULL; in __clear_extent_bit()
939 u64 *failed_start, struct extent_state **cached_state, in __set_extent_bit() argument
967 if (cached_state && *cached_state) { in __set_extent_bit()
968 state = *cached_state; in __set_extent_bit()
988 cache_state(prealloc, cached_state); in __set_extent_bit()
1011 cache_state(state, cached_state); in __set_extent_bit()
[all …]
Dfile.c469 struct extent_state **cached_state) in btrfs_find_new_delalloc_bytes() argument
497 NULL, cached_state, GFP_NOFS); in btrfs_find_new_delalloc_bytes()
1485 struct extent_state **cached_state) in lock_and_cleanup_extent_if_need() argument
1502 cached_state); in lock_and_cleanup_extent_if_need()
1509 last_pos, cached_state); in lock_and_cleanup_extent_if_need()
1613 struct extent_state *cached_state = NULL; in btrfs_buffered_write() local
1702 &lockend, &cached_state); in btrfs_buffered_write()
1760 pos, copied, &cached_state); in btrfs_buffered_write()
1771 lockstart, lockend, &cached_state); in btrfs_buffered_write()
1773 free_extent_state(cached_state); in btrfs_buffered_write()
[all …]
Dinode.c2146 struct extent_state **cached_state) in btrfs_set_extent_delalloc() argument
2150 extra_bits, cached_state); in btrfs_set_extent_delalloc()
2163 struct extent_state *cached_state = NULL; in btrfs_writepage_fixup_worker() local
2185 &cached_state); in btrfs_writepage_fixup_worker()
2195 page_end, &cached_state); in btrfs_writepage_fixup_worker()
2212 &cached_state); in btrfs_writepage_fixup_worker()
2225 &cached_state); in btrfs_writepage_fixup_worker()
3025 struct extent_state *cached_state = NULL; in btrfs_finish_ordered_io() local
3091 &cached_state); in btrfs_finish_ordered_io()
3095 EXTENT_DEFRAG, 0, cached_state); in btrfs_finish_ordered_io()
[all …]
Dtransaction.c905 struct extent_state *cached_state = NULL; in btrfs_write_marked_extents() local
911 mark, &cached_state)) { in btrfs_write_marked_extents()
916 mark, &cached_state); in btrfs_write_marked_extents()
940 free_extent_state(cached_state); in btrfs_write_marked_extents()
941 cached_state = NULL; in btrfs_write_marked_extents()
961 struct extent_state *cached_state = NULL; in __btrfs_wait_marked_extents() local
966 EXTENT_NEED_WAIT, &cached_state)) { in __btrfs_wait_marked_extents()
976 EXTENT_NEED_WAIT, 0, 0, &cached_state); in __btrfs_wait_marked_extents()
983 free_extent_state(cached_state); in __btrfs_wait_marked_extents()
984 cached_state = NULL; in __btrfs_wait_marked_extents()
Dordered-data.c986 struct extent_state **cached_state) in btrfs_lock_and_flush_ordered_range() argument
992 if (cached_state) in btrfs_lock_and_flush_ordered_range()
993 cachedp = cached_state; in btrfs_lock_and_flush_ordered_range()
1005 if (!cached_state) in btrfs_lock_and_flush_ordered_range()
Dordered-data.h194 struct extent_state **cached_state);
Dfree-space-cache.c1135 struct extent_state **cached_state) in cleanup_write_cache_enospc() argument
1139 i_size_read(inode) - 1, cached_state); in cleanup_write_cache_enospc()
1238 struct extent_state *cached_state = NULL; in __btrfs_write_out_cache() local
1274 &cached_state); in __btrfs_write_out_cache()
1315 i_size_read(inode), &cached_state); in __btrfs_write_out_cache()
1328 i_size_read(inode) - 1, &cached_state); in __btrfs_write_out_cache()
1362 cleanup_write_cache_enospc(inode, io_ctl, &cached_state); in __btrfs_write_out_cache()
Ddisk-io.c305 struct extent_state *cached_state = NULL; in verify_parent_transid() local
321 &cached_state); in verify_parent_transid()
345 &cached_state); in verify_parent_transid()
4407 struct extent_state *cached_state = NULL; in btrfs_destroy_pinned_extent() local
4417 EXTENT_DIRTY, &cached_state); in btrfs_destroy_pinned_extent()
4423 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_destroy_pinned_extent()
4424 free_extent_state(cached_state); in btrfs_destroy_pinned_extent()
Dioctl.c1242 struct extent_state *cached_state = NULL; in cluster_pages_for_defrag() local
1274 &cached_state); in cluster_pages_for_defrag()
1278 &cached_state); in cluster_pages_for_defrag()
1334 page_start, page_end - 1, &cached_state); in cluster_pages_for_defrag()
1337 EXTENT_DEFRAG, 0, 0, &cached_state); in cluster_pages_for_defrag()
1350 &cached_state); in cluster_pages_for_defrag()
1353 page_start, page_end - 1, &cached_state); in cluster_pages_for_defrag()
Dextent-tree.c2896 struct extent_state *cached_state = NULL; in btrfs_finish_extent_commit() local
2900 EXTENT_DIRTY, &cached_state); in btrfs_finish_extent_commit()
2910 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_finish_extent_commit()
2913 free_extent_state(cached_state); in btrfs_finish_extent_commit()
Dctree.h2846 struct extent_state **cached_state);
/Linux-v5.4/drivers/mux/
Dcore.c117 mux->cached_state = MUX_CACHE_UNKNOWN; in mux_chip_alloc()
131 mux->cached_state = ret < 0 ? MUX_CACHE_UNKNOWN : state; in mux_control_set()
155 if (mux->idle_state == mux->cached_state) in mux_chip_register()
303 if (mux->cached_state == state) in __mux_control_select()
399 mux->idle_state != mux->cached_state) in mux_control_deselect()
/Linux-v5.4/drivers/iio/multiplexer/
Diio-mux.c29 int cached_state; member
47 mux->cached_state = -1; in iio_mux_select()
51 if (mux->cached_state == chan->channel) in iio_mux_select()
70 mux->cached_state = -1; in iio_mux_select()
75 mux->cached_state = chan->channel; in iio_mux_select()
395 mux->cached_state = -1; in mux_probe()
/Linux-v5.4/include/linux/mux/
Ddriver.h46 int cached_state; member
/Linux-v5.4/drivers/gpu/drm/amd/display/amdgpu_dm/
Damdgpu_dm.h229 struct drm_atomic_state *cached_state; member
Damdgpu_dm.c1045 WARN_ON(adev->dm.cached_state); in dm_suspend()
1046 adev->dm.cached_state = drm_atomic_helper_suspend(adev->ddev); in dm_suspend()
1228 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) in dm_resume()
1236 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) { in dm_resume()
1245 for_each_new_plane_in_state(dm->cached_state, plane, new_plane_state, i) { in dm_resume()
1254 drm_atomic_helper_resume(ddev, dm->cached_state); in dm_resume()
1256 dm->cached_state = NULL; in dm_resume()