Lines Matching refs:delayed_refs

810 	struct btrfs_delayed_ref_root *delayed_refs;  in btrfs_lookup_extent_info()  local
891 delayed_refs = &trans->transaction->delayed_refs; in btrfs_lookup_extent_info()
892 spin_lock(&delayed_refs->lock); in btrfs_lookup_extent_info()
893 head = btrfs_find_delayed_ref_head(delayed_refs, bytenr); in btrfs_lookup_extent_info()
897 spin_unlock(&delayed_refs->lock); in btrfs_lookup_extent_info()
920 spin_unlock(&delayed_refs->lock); in btrfs_lookup_extent_info()
2396 static void unselect_delayed_ref_head(struct btrfs_delayed_ref_root *delayed_refs, in unselect_delayed_ref_head() argument
2399 spin_lock(&delayed_refs->lock); in unselect_delayed_ref_head()
2401 delayed_refs->num_heads_ready++; in unselect_delayed_ref_head()
2402 spin_unlock(&delayed_refs->lock); in unselect_delayed_ref_head()
2430 struct btrfs_delayed_ref_root *delayed_refs; in cleanup_ref_head() local
2433 delayed_refs = &trans->transaction->delayed_refs; in cleanup_ref_head()
2437 unselect_delayed_ref_head(delayed_refs, head); in cleanup_ref_head()
2449 spin_lock(&delayed_refs->lock); in cleanup_ref_head()
2453 spin_unlock(&delayed_refs->lock); in cleanup_ref_head()
2456 delayed_refs->num_heads--; in cleanup_ref_head()
2457 rb_erase(&head->href_node, &delayed_refs->href_root); in cleanup_ref_head()
2460 spin_unlock(&delayed_refs->lock); in cleanup_ref_head()
2461 atomic_dec(&delayed_refs->num_entries); in cleanup_ref_head()
2482 spin_lock(&delayed_refs->lock); in cleanup_ref_head()
2483 delayed_refs->pending_csums -= head->num_bytes; in cleanup_ref_head()
2484 spin_unlock(&delayed_refs->lock); in cleanup_ref_head()
2513 struct btrfs_delayed_ref_root *delayed_refs; in __btrfs_run_delayed_refs() local
2523 delayed_refs = &trans->transaction->delayed_refs; in __btrfs_run_delayed_refs()
2529 spin_lock(&delayed_refs->lock); in __btrfs_run_delayed_refs()
2532 spin_unlock(&delayed_refs->lock); in __btrfs_run_delayed_refs()
2539 spin_unlock(&delayed_refs->lock); in __btrfs_run_delayed_refs()
2566 btrfs_merge_delayed_refs(trans, delayed_refs, locked_ref); in __btrfs_run_delayed_refs()
2573 unselect_delayed_ref_head(delayed_refs, locked_ref); in __btrfs_run_delayed_refs()
2619 atomic_dec(&delayed_refs->num_entries); in __btrfs_run_delayed_refs()
2637 unselect_delayed_ref_head(delayed_refs, locked_ref); in __btrfs_run_delayed_refs()
2662 spin_lock(&delayed_refs->lock); in __btrfs_run_delayed_refs()
2665 spin_unlock(&delayed_refs->lock); in __btrfs_run_delayed_refs()
2752 u64 num_heads = trans->transaction->delayed_refs.num_heads_ready; in btrfs_check_space_for_delayed_refs()
2753 u64 csum_bytes = trans->transaction->delayed_refs.pending_csums; in btrfs_check_space_for_delayed_refs()
2789 atomic_read(&trans->transaction->delayed_refs.num_entries); in btrfs_should_throttle_delayed_refs()
2910 struct btrfs_delayed_ref_root *delayed_refs; in btrfs_run_delayed_refs() local
2923 delayed_refs = &trans->transaction->delayed_refs; in btrfs_run_delayed_refs()
2925 count = atomic_read(&delayed_refs->num_entries) * 2; in btrfs_run_delayed_refs()
2929 delayed_refs->run_delayed_start = find_middle(&delayed_refs->root); in btrfs_run_delayed_refs()
2942 spin_lock(&delayed_refs->lock); in btrfs_run_delayed_refs()
2943 node = rb_first(&delayed_refs->href_root); in btrfs_run_delayed_refs()
2945 spin_unlock(&delayed_refs->lock); in btrfs_run_delayed_refs()
2951 spin_unlock(&delayed_refs->lock); in btrfs_run_delayed_refs()
2998 struct btrfs_delayed_ref_root *delayed_refs; in check_delayed_ref() local
3011 delayed_refs = &cur_trans->delayed_refs; in check_delayed_ref()
3012 spin_lock(&delayed_refs->lock); in check_delayed_ref()
3013 head = btrfs_find_delayed_ref_head(delayed_refs, bytenr); in check_delayed_ref()
3015 spin_unlock(&delayed_refs->lock); in check_delayed_ref()
3022 spin_unlock(&delayed_refs->lock); in check_delayed_ref()
3036 spin_unlock(&delayed_refs->lock); in check_delayed_ref()
6918 struct btrfs_delayed_ref_root *delayed_refs; in check_ref_cleanup() local
6921 delayed_refs = &trans->transaction->delayed_refs; in check_ref_cleanup()
6922 spin_lock(&delayed_refs->lock); in check_ref_cleanup()
6923 head = btrfs_find_delayed_ref_head(delayed_refs, bytenr); in check_ref_cleanup()
6949 rb_erase(&head->href_node, &delayed_refs->href_root); in check_ref_cleanup()
6951 atomic_dec(&delayed_refs->num_entries); in check_ref_cleanup()
6957 delayed_refs->num_heads--; in check_ref_cleanup()
6959 delayed_refs->num_heads_ready--; in check_ref_cleanup()
6962 spin_unlock(&delayed_refs->lock); in check_ref_cleanup()
6975 spin_unlock(&delayed_refs->lock); in check_ref_cleanup()