| /Linux-v5.10/fs/btrfs/ |
| D | extent-io-tree.h | 113 struct extent_state *cached_state); 158 struct extent_state **cached_state, gfp_t mask); 169 u64 end, struct extent_state **cached_state) in clear_extent_uptodate() argument 172 cached_state, GFP_NOFS, NULL); in clear_extent_uptodate() 192 struct extent_state **cached_state); 196 struct extent_state **cached_state) in set_extent_delalloc() argument 200 NULL, cached_state, GFP_NOFS); in set_extent_delalloc() 204 u64 end, struct extent_state **cached_state) in set_extent_defrag() argument 208 NULL, cached_state, GFP_NOFS); in set_extent_defrag() 219 u64 end, struct extent_state **cached_state, gfp_t mask) in set_extent_uptodate() argument [all …]
|
| D | extent_io.c | 699 struct extent_state **cached_state, in __clear_extent_bit() argument 734 if (cached_state) { in __clear_extent_bit() 735 cached = *cached_state; in __clear_extent_bit() 738 *cached_state = NULL; in __clear_extent_bit() 739 cached_state = NULL; in __clear_extent_bit() 968 u64 *failed_start, struct extent_state **cached_state, in __set_extent_bit() argument 996 if (cached_state && *cached_state) { in __set_extent_bit() 997 state = *cached_state; in __set_extent_bit() 1017 cache_state(prealloc, cached_state); in __set_extent_bit() 1040 cache_state(state, cached_state); in __set_extent_bit() [all …]
|
| D | file.c | 1410 struct extent_state **cached_state) in lock_and_cleanup_extent_if_need() argument 1425 cached_state); in lock_and_cleanup_extent_if_need() 1432 last_pos, cached_state); in lock_and_cleanup_extent_if_need() 1588 struct extent_state *cached_state = NULL; in btrfs_buffered_write() local 1677 &lockend, &cached_state); in btrfs_buffered_write() 1736 &cached_state); in btrfs_buffered_write() 1747 lockstart, lockend, &cached_state); in btrfs_buffered_write() 1749 free_extent_state(cached_state); in btrfs_buffered_write() 2472 struct extent_state **cached_state) in btrfs_punch_hole_lock_range() argument 2481 cached_state); in btrfs_punch_hole_lock_range() [all …]
|
| D | inode.c | 2259 struct extent_state **cached_state) in btrfs_find_new_delalloc_bytes() argument 2286 NULL, cached_state, GFP_NOFS); in btrfs_find_new_delalloc_bytes() 2298 struct extent_state **cached_state) in btrfs_set_extent_delalloc() argument 2314 cached_state); in btrfs_set_extent_delalloc() 2320 cached_state); in btrfs_set_extent_delalloc() 2334 struct extent_state *cached_state = NULL; in btrfs_writepage_fixup_worker() local 2398 lock_extent_bits(&inode->io_tree, page_start, page_end, &cached_state); in btrfs_writepage_fixup_worker() 2407 &cached_state); in btrfs_writepage_fixup_worker() 2415 &cached_state); in btrfs_writepage_fixup_worker() 2434 &cached_state); in btrfs_writepage_fixup_worker() [all …]
|
| D | transaction.c | 973 struct extent_state *cached_state = NULL; in btrfs_write_marked_extents() local 979 mark, &cached_state)) { in btrfs_write_marked_extents() 984 mark, &cached_state); in btrfs_write_marked_extents() 1008 free_extent_state(cached_state); in btrfs_write_marked_extents() 1009 cached_state = NULL; in btrfs_write_marked_extents() 1029 struct extent_state *cached_state = NULL; in __btrfs_wait_marked_extents() local 1034 EXTENT_NEED_WAIT, &cached_state)) { in __btrfs_wait_marked_extents() 1044 EXTENT_NEED_WAIT, 0, 0, &cached_state); in __btrfs_wait_marked_extents() 1051 free_extent_state(cached_state); in __btrfs_wait_marked_extents() 1052 cached_state = NULL; in __btrfs_wait_marked_extents()
|
| D | ordered-data.c | 917 struct extent_state **cached_state) in btrfs_lock_and_flush_ordered_range() argument 923 if (cached_state) in btrfs_lock_and_flush_ordered_range() 924 cachedp = cached_state; in btrfs_lock_and_flush_ordered_range() 936 if (!cached_state) in btrfs_lock_and_flush_ordered_range()
|
| D | ordered-data.h | 195 struct extent_state **cached_state);
|
| D | dev-replace.c | 627 struct extent_state *cached_state = NULL; in btrfs_set_target_alloc_state() local 637 CHUNK_ALLOCATED, &cached_state)) { in btrfs_set_target_alloc_state() 645 free_extent_state(cached_state); in btrfs_set_target_alloc_state()
|
| D | free-space-cache.c | 1158 struct extent_state **cached_state) in cleanup_write_cache_enospc() argument 1162 i_size_read(inode) - 1, cached_state); in cleanup_write_cache_enospc() 1257 struct extent_state *cached_state = NULL; in __btrfs_write_out_cache() local 1293 &cached_state); in __btrfs_write_out_cache() 1335 &cached_state); in __btrfs_write_out_cache() 1349 i_size_read(inode) - 1, &cached_state); in __btrfs_write_out_cache() 1371 cleanup_write_cache_enospc(inode, io_ctl, &cached_state); in __btrfs_write_out_cache()
|
| D | disk-io.c | 241 struct extent_state *cached_state = NULL; in verify_parent_transid() local 257 &cached_state); in verify_parent_transid() 281 &cached_state); in verify_parent_transid() 4522 struct extent_state *cached_state = NULL; in btrfs_destroy_pinned_extent() local 4532 EXTENT_DIRTY, &cached_state); in btrfs_destroy_pinned_extent() 4538 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_destroy_pinned_extent() 4539 free_extent_state(cached_state); in btrfs_destroy_pinned_extent()
|
| D | ioctl.c | 1282 struct extent_state *cached_state = NULL; in cluster_pages_for_defrag() local 1313 &cached_state); in cluster_pages_for_defrag() 1317 &cached_state); in cluster_pages_for_defrag() 1373 page_start, page_end - 1, &cached_state); in cluster_pages_for_defrag() 1376 EXTENT_DEFRAG, 0, 0, &cached_state); in cluster_pages_for_defrag() 1388 &cached_state); in cluster_pages_for_defrag() 1391 page_start, page_end - 1, &cached_state); in cluster_pages_for_defrag()
|
| D | extent-tree.c | 2895 struct extent_state *cached_state = NULL; in btrfs_finish_extent_commit() local 2899 EXTENT_DIRTY, &cached_state); in btrfs_finish_extent_commit() 2912 clear_extent_dirty(unpin, start, end, &cached_state); in btrfs_finish_extent_commit() 2915 free_extent_state(cached_state); in btrfs_finish_extent_commit()
|
| D | tree-log.c | 3906 struct extent_state *cached_state = NULL; in log_csums() local 3924 lock_end, &cached_state); in log_csums() 3941 &cached_state); in log_csums()
|
| D | ctree.h | 3008 struct extent_state **cached_state);
|
| /Linux-v5.10/drivers/gpu/drm/msm/dsi/pll/ |
| D | dsi_pll_28nm_8960.c | 81 struct pll_28nm_cached_state cached_state; member 340 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_save_state() local 343 cached_state->postdiv3 = in dsi_pll_28nm_save_state() 345 cached_state->postdiv2 = in dsi_pll_28nm_save_state() 347 cached_state->postdiv1 = in dsi_pll_28nm_save_state() 350 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_28nm_save_state() 356 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_restore_state() local 361 cached_state->vco_rate, 0); in dsi_pll_28nm_restore_state() 369 cached_state->postdiv3); in dsi_pll_28nm_restore_state() 371 cached_state->postdiv2); in dsi_pll_28nm_restore_state() [all …]
|
| D | dsi_pll_28nm.c | 86 struct pll_28nm_cached_state cached_state; member 442 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_save_state() local 445 cached_state->postdiv3 = in dsi_pll_28nm_save_state() 447 cached_state->postdiv1 = in dsi_pll_28nm_save_state() 449 cached_state->byte_mux = pll_read(base + REG_DSI_28nm_PHY_PLL_VREG_CFG); in dsi_pll_28nm_save_state() 450 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_28nm_save_state() 456 struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state; in dsi_pll_28nm_restore_state() local 461 cached_state->vco_rate, 0); in dsi_pll_28nm_restore_state() 469 cached_state->postdiv3); in dsi_pll_28nm_restore_state() 471 cached_state->postdiv1); in dsi_pll_28nm_restore_state() [all …]
|
| D | dsi_pll_14nm.c | 144 struct pll_14nm_cached_state cached_state; member 798 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_pll_14nm_save_state() local 804 cached_state->n1postdiv = data & 0xf; in dsi_pll_14nm_save_state() 805 cached_state->n2postdiv = (data >> 4) & 0xf; in dsi_pll_14nm_save_state() 808 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_pll_14nm_save_state() 810 cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw); in dsi_pll_14nm_save_state() 816 struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state; in dsi_pll_14nm_restore_state() local 822 cached_state->vco_rate, 0); in dsi_pll_14nm_restore_state() 829 data = cached_state->n1postdiv | (cached_state->n2postdiv << 4); in dsi_pll_14nm_restore_state() 832 cached_state->n1postdiv, cached_state->n2postdiv); in dsi_pll_14nm_restore_state()
|
| D | dsi_pll_7nm.c | 118 struct pll_7nm_cached_state cached_state; member 562 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_pll_7nm_save_state() 585 struct pll_7nm_cached_state *cached = &pll_7nm->cached_state; in dsi_pll_7nm_restore_state()
|
| D | dsi_pll_10nm.c | 118 struct pll_10nm_cached_state cached_state; member 536 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_pll_10nm_save_state() 559 struct pll_10nm_cached_state *cached = &pll_10nm->cached_state; in dsi_pll_10nm_restore_state()
|
| /Linux-v5.10/drivers/mux/ |
| D | core.c | 117 mux->cached_state = MUX_CACHE_UNKNOWN; in mux_chip_alloc() 131 mux->cached_state = ret < 0 ? MUX_CACHE_UNKNOWN : state; in mux_control_set() 155 if (mux->idle_state == mux->cached_state) in mux_chip_register() 303 if (mux->cached_state == state) in __mux_control_select() 399 mux->idle_state != mux->cached_state) in mux_control_deselect()
|
| /Linux-v5.10/drivers/iio/multiplexer/ |
| D | iio-mux.c | 29 int cached_state; member 47 mux->cached_state = -1; in iio_mux_select() 51 if (mux->cached_state == chan->channel) in iio_mux_select() 70 mux->cached_state = -1; in iio_mux_select() 75 mux->cached_state = chan->channel; in iio_mux_select() 393 mux->cached_state = -1; in mux_probe()
|
| /Linux-v5.10/include/linux/mux/ |
| D | driver.h | 46 int cached_state; member
|
| /Linux-v5.10/drivers/gpu/drm/amd/display/amdgpu_dm/ |
| D | amdgpu_dm.h | 324 struct drm_atomic_state *cached_state; member
|
| D | amdgpu_dm.c | 1751 WARN_ON(adev->dm.cached_state); in dm_suspend() 1752 adev->dm.cached_state = drm_atomic_helper_suspend(adev_to_drm(adev)); in dm_suspend() 2018 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) in dm_resume() 2026 for_each_new_crtc_in_state(dm->cached_state, crtc, new_crtc_state, i) { in dm_resume() 2035 for_each_new_plane_in_state(dm->cached_state, plane, new_plane_state, i) { in dm_resume() 2044 drm_atomic_helper_resume(ddev, dm->cached_state); in dm_resume() 2046 dm->cached_state = NULL; in dm_resume()
|