/Linux-v5.4/drivers/gpu/drm/msm/dsi/pll/ |
D | dsi_pll_28nm_8960.c | 81 struct pll_28nm_cached_state cached_state; member 340 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_save_state() local 343 cached_state->postdiv3 = in dsi_pll_28nm_save_state() 345 cached_state->postdiv2 = in dsi_pll_28nm_save_state() 347 cached_state->postdiv1 = in dsi_pll_28nm_save_state() 350 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_28nm_save_state() 356 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_restore_state() local 361 cached_state->vco_rate, 0); in dsi_pll_28nm_restore_state() 369 cached_state->postdiv3); in dsi_pll_28nm_restore_state() 371 cached_state->postdiv2); in dsi_pll_28nm_restore_state() [all …]
|
D | dsi_pll_28nm.c | 86 struct pll_28nm_cached_state cached_state; member 442 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_save_state() local 445 cached_state->postdiv3 = in dsi_pll_28nm_save_state() 447 cached_state->postdiv1 = in dsi_pll_28nm_save_state() 449 cached_state->byte_mux = pll_read(base + REG_DSI_28nm_PHY_PLL_VREG_CFG); in dsi_pll_28nm_save_state() 450 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_28nm_save_state() 456 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_restore_state() local 461 cached_state->vco_rate, 0); in dsi_pll_28nm_restore_state() 469 cached_state->postdiv3); in dsi_pll_28nm_restore_state() 471 cached_state->postdiv1); in dsi_pll_28nm_restore_state() [all …]
|
D | dsi_pll_14nm.c | 144 struct pll_14nm_cached_state cached_state; member 798 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_pll_14nm_save_state() local 804 cached_state->n1postdiv = data & 0xf; in dsi_pll_14nm_save_state() 805 cached_state->n2postdiv = (data >> 4) & 0xf; in dsi_pll_14nm_save_state() 808 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_pll_14nm_save_state() 810 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_14nm_save_state() 816 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_pll_14nm_restore_state() local 822 cached_state->vco_rate, 0); in dsi_pll_14nm_restore_state() 829 data = cached_state->n1postdiv | (cached_state->n2postdiv << 4); in dsi_pll_14nm_restore_state() 832 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_pll_14nm_restore_state()
|
D | dsi_pll_10nm.c | 118 struct pll_10nm_cached_state cached_state; member 530 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_pll_10nm_save_state() 553 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_pll_10nm_restore_state()
|
/Linux-v5.4/fs/btrfs/ |
D | extent_io.h | 289 struct extent_state *cached_state); 334 struct extent_state **cached_state, gfp_t mask); 345 u64 end, struct extent_state **cached_state) in clear_extent_uptodate() argument 348 cached_state, GFP_NOFS, NULL); in clear_extent_uptodate() 368 struct extent_state **cached_state); 372 struct extent_state **cached_state) in set_extent_delalloc() argument 376 NULL, cached_state, GFP_NOFS); in set_extent_delalloc() 380 u64 end, struct extent_state **cached_state) in set_extent_defrag() argument 384 NULL, cached_state, GFP_NOFS); in set_extent_defrag() 395 u64 end, struct extent_state **cached_state, gfp_t mask) in set_extent_uptodate() argument [all …]
|
D | extent_io.c | 670 struct extent_state **cached_state, in __clear_extent_bit() argument 705 if (cached_state) { in __clear_extent_bit() 706 cached = *cached_state; in __clear_extent_bit() 709 *cached_state = NULL; in __clear_extent_bit() 710 cached_state = NULL; in __clear_extent_bit() 939 u64 *failed_start, struct extent_state **cached_state, in __set_extent_bit() argument 967 if (cached_state && *cached_state) { in __set_extent_bit() 968 state = *cached_state; in __set_extent_bit() 988 cache_state(prealloc, cached_state); in __set_extent_bit() 1011 cache_state(state, cached_state); in __set_extent_bit() [all …]
|
D | file.c | 469 struct extent_state **cached_state) in btrfs_find_new_delalloc_bytes() argument 497 NULL, cached_state, GFP_NOFS); in btrfs_find_new_delalloc_bytes() 1485 struct extent_state **cached_state) in lock_and_cleanup_extent_if_need() argument 1502 cached_state); in lock_and_cleanup_extent_if_need() 1509 last_pos, cached_state); in lock_and_cleanup_extent_if_need() 1613 struct extent_state *cached_state = NULL; in btrfs_buffered_write() local 1702 &lockend, &cached_state); in btrfs_buffered_write() 1760 pos, copied, &cached_state); in btrfs_buffered_write() 1771 lockstart, lockend, &cached_state); in btrfs_buffered_write() 1773 free_extent_state(cached_state); in btrfs_buffered_write() [all …]
|
D | inode.c | 2146 struct extent_state **cached_state) in btrfs_set_extent_delalloc() argument 2150 extra_bits, cached_state); in btrfs_set_extent_delalloc() 2163 struct extent_state *cached_state = NULL; in btrfs_writepage_fixup_worker() local 2185 &cached_state); in btrfs_writepage_fixup_worker() 2195 page_end, &cached_state); in btrfs_writepage_fixup_worker() 2212 &cached_state); in btrfs_writepage_fixup_worker() 2225 &cached_state); in btrfs_writepage_fixup_worker() 3025 struct extent_state *cached_state = NULL; in btrfs_finish_ordered_io() local 3091 &cached_state); in btrfs_finish_ordered_io() 3095 EXTENT_DEFRAG, 0, cached_state); in btrfs_finish_ordered_io() [all …]
|
D | transaction.c | 905 struct extent_state *cached_state = NULL; in btrfs_write_marked_extents() local 911 mark, &cached_state)) { in btrfs_write_marked_extents() 916 mark, &cached_state); in btrfs_write_marked_extents() 940 free_extent_state(cached_state); in btrfs_write_marked_extents() 941 cached_state = NULL; in btrfs_write_marked_extents() 961 struct extent_state *cached_state = NULL; in __btrfs_wait_marked_extents() local 966 EXTENT_NEED_WAIT, &cached_state)) { in __btrfs_wait_marked_extents() 976 EXTENT_NEED_WAIT, 0, 0, &cached_state); in __btrfs_wait_marked_extents() 983 free_extent_state(cached_state); in __btrfs_wait_marked_extents() 984 cached_state = NULL; in __btrfs_wait_marked_extents()
|
D | ordered-data.c | 986 struct extent_state **cached_state) in btrfs_lock_and_flush_ordered_range() argument 992 if (cached_state) in btrfs_lock_and_flush_ordered_range() 993 cachedp = cached_state; in btrfs_lock_and_flush_ordered_range() 1005 if (!cached_state) in btrfs_lock_and_flush_ordered_range()
|
D | ordered-data.h | 194 struct extent_state **cached_state);
|
D | free-space-cache.c | 1135 struct extent_state **cached_state) in cleanup_write_cache_enospc() argument 1139 i_size_read(inode) - 1, cached_state); in cleanup_write_cache_enospc() 1238 struct extent_state *cached_state = NULL; in __btrfs_write_out_cache() local 1274 &cached_state); in __btrfs_write_out_cache() 1315 i_size_read(inode), &cached_state); in __btrfs_write_out_cache() 1328 i_size_read(inode) - 1, &cached_state); in __btrfs_write_out_cache() 1362 cleanup_write_cache_enospc(inode, io_ctl, &cached_state); in __btrfs_write_out_cache()
|
D | disk-io.c | 305 struct extent_state *cached_state = NULL; in verify_parent_transid() local 321 &cached_state); in verify_parent_transid() 345 &cached_state); in verify_parent_transid() 4407 struct extent_state *cached_state = NULL; in btrfs_destroy_pinned_extent() local 4417 EXTENT_DIRTY, &cached_state); in btrfs_destroy_pinned_extent() 4423 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_destroy_pinned_extent() 4424 free_extent_state(cached_state); in btrfs_destroy_pinned_extent()
|
D | ioctl.c | 1242 struct extent_state *cached_state = NULL; in cluster_pages_for_defrag() local 1274 &cached_state); in cluster_pages_for_defrag() 1278 &cached_state); in cluster_pages_for_defrag() 1334 page_start, page_end - 1, &cached_state); in cluster_pages_for_defrag() 1337 EXTENT_DEFRAG, 0, 0, &cached_state); in cluster_pages_for_defrag() 1350 &cached_state); in cluster_pages_for_defrag() 1353 page_start, page_end - 1, &cached_state); in cluster_pages_for_defrag()
|
D | extent-tree.c | 2896 struct extent_state *cached_state = NULL; in btrfs_finish_extent_commit() local 2900 EXTENT_DIRTY, &cached_state); in btrfs_finish_extent_commit() 2910 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_finish_extent_commit() 2913 free_extent_state(cached_state); in btrfs_finish_extent_commit()
|
D | ctree.h | 2846 struct extent_state **cached_state);
|
/Linux-v5.4/drivers/mux/ |
D | core.c | 117 mux->cached_state = MUX_CACHE_UNKNOWN; in mux_chip_alloc() 131 mux->cached_state = ret < 0 ? MUX_CACHE_UNKNOWN : state; in mux_control_set() 155 if (mux->idle_state == mux->cached_state) in mux_chip_register() 303 if (mux->cached_state == state) in __mux_control_select() 399 mux->idle_state != mux->cached_state) in mux_control_deselect()
|
/Linux-v5.4/drivers/iio/multiplexer/ |
D | iio-mux.c | 29 int cached_state; member 47 mux->cached_state = -1; in iio_mux_select() 51 if (mux->cached_state == chan->channel) in iio_mux_select() 70 mux->cached_state = -1; in iio_mux_select() 75 mux->cached_state = chan->channel; in iio_mux_select() 395 mux->cached_state = -1; in mux_probe()
|
/Linux-v5.4/include/linux/mux/ |
D | driver.h | 46 int cached_state; member
|
/Linux-v5.4/drivers/gpu/drm/amd/display/amdgpu_dm/ |
D | amdgpu_dm.h | 229 struct drm_atomic_state *cached_state; member
|
D | amdgpu_dm.c | 1045 WARN_ON(adev->dm.cached_state); in dm_suspend() 1046 adev->dm.cached_state = drm_atomic_helper_suspend(adev->ddev); in dm_suspend() 1228 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) in dm_resume() 1236 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) { in dm_resume() 1245 for_each_new_plane_in_state(dm->cached_state, plane, new_plane_state, i) { in dm_resume() 1254 drm_atomic_helper_resume(ddev, dm->cached_state); in dm_resume() 1256 dm->cached_state = NULL; in dm_resume()
|