Lines Matching refs:em
53 struct extent_map *em; in alloc_extent_map() local
54 em = kmem_cache_zalloc(extent_map_cache, GFP_NOFS); in alloc_extent_map()
55 if (!em) in alloc_extent_map()
57 RB_CLEAR_NODE(&em->rb_node); in alloc_extent_map()
58 em->compress_type = BTRFS_COMPRESS_NONE; in alloc_extent_map()
59 refcount_set(&em->refs, 1); in alloc_extent_map()
60 INIT_LIST_HEAD(&em->list); in alloc_extent_map()
61 return em; in alloc_extent_map()
71 void free_extent_map(struct extent_map *em) in free_extent_map() argument
73 if (!em) in free_extent_map()
75 if (refcount_dec_and_test(&em->refs)) { in free_extent_map()
76 WARN_ON(extent_map_in_tree(em)); in free_extent_map()
77 WARN_ON(!list_empty(&em->list)); in free_extent_map()
78 if (test_bit(EXTENT_FLAG_FS_MAPPING, &em->flags)) in free_extent_map()
79 kfree(em->map_lookup); in free_extent_map()
80 kmem_cache_free(extent_map_cache, em); in free_extent_map()
92 static int tree_insert(struct rb_root_cached *root, struct extent_map *em) in tree_insert() argument
98 u64 end = range_end(em->start, em->len); in tree_insert()
105 if (em->start < entry->start) { in tree_insert()
107 } else if (em->start >= extent_map_end(entry)) { in tree_insert()
116 while (parent && em->start >= extent_map_end(entry)) { in tree_insert()
121 if (end > entry->start && em->start < extent_map_end(entry)) in tree_insert()
126 while (parent && em->start < entry->start) { in tree_insert()
131 if (end > entry->start && em->start < extent_map_end(entry)) in tree_insert()
134 rb_link_node(&em->rb_node, orig_parent, p); in tree_insert()
135 rb_insert_color_cached(&em->rb_node, root, leftmost); in tree_insert()
239 static void try_merge_map(struct extent_map_tree *tree, struct extent_map *em) in try_merge_map() argument
252 if (refcount_read(&em->refs) > 2) in try_merge_map()
255 if (em->start != 0) { in try_merge_map()
256 rb = rb_prev(&em->rb_node); in try_merge_map()
259 if (rb && mergable_maps(merge, em)) { in try_merge_map()
260 em->start = merge->start; in try_merge_map()
261 em->orig_start = merge->orig_start; in try_merge_map()
262 em->len += merge->len; in try_merge_map()
263 em->block_len += merge->block_len; in try_merge_map()
264 em->block_start = merge->block_start; in try_merge_map()
265 em->mod_len = (em->mod_len + em->mod_start) - merge->mod_start; in try_merge_map()
266 em->mod_start = merge->mod_start; in try_merge_map()
267 em->generation = max(em->generation, merge->generation); in try_merge_map()
268 set_bit(EXTENT_FLAG_MERGED, &em->flags); in try_merge_map()
276 rb = rb_next(&em->rb_node); in try_merge_map()
279 if (rb && mergable_maps(em, merge)) { in try_merge_map()
280 em->len += merge->len; in try_merge_map()
281 em->block_len += merge->block_len; in try_merge_map()
284 em->mod_len = (merge->mod_start + merge->mod_len) - em->mod_start; in try_merge_map()
285 em->generation = max(em->generation, merge->generation); in try_merge_map()
286 set_bit(EXTENT_FLAG_MERGED, &em->flags); in try_merge_map()
306 struct extent_map *em; in unpin_extent_cache() local
310 em = lookup_extent_mapping(tree, start, len); in unpin_extent_cache()
312 WARN_ON(!em || em->start != start); in unpin_extent_cache()
314 if (!em) in unpin_extent_cache()
317 em->generation = gen; in unpin_extent_cache()
318 clear_bit(EXTENT_FLAG_PINNED, &em->flags); in unpin_extent_cache()
319 em->mod_start = em->start; in unpin_extent_cache()
320 em->mod_len = em->len; in unpin_extent_cache()
322 if (test_bit(EXTENT_FLAG_FILLING, &em->flags)) { in unpin_extent_cache()
324 clear_bit(EXTENT_FLAG_FILLING, &em->flags); in unpin_extent_cache()
327 try_merge_map(tree, em); in unpin_extent_cache()
330 em->mod_start = em->start; in unpin_extent_cache()
331 em->mod_len = em->len; in unpin_extent_cache()
334 free_extent_map(em); in unpin_extent_cache()
341 void clear_em_logging(struct extent_map_tree *tree, struct extent_map *em) in clear_em_logging() argument
345 clear_bit(EXTENT_FLAG_LOGGING, &em->flags); in clear_em_logging()
346 if (extent_map_in_tree(em)) in clear_em_logging()
347 try_merge_map(tree, em); in clear_em_logging()
351 struct extent_map *em, in setup_extent_mapping() argument
354 refcount_inc(&em->refs); in setup_extent_mapping()
355 em->mod_start = em->start; in setup_extent_mapping()
356 em->mod_len = em->len; in setup_extent_mapping()
359 list_move(&em->list, &tree->modified_extents); in setup_extent_mapping()
361 try_merge_map(tree, em); in setup_extent_mapping()
364 static void extent_map_device_set_bits(struct extent_map *em, unsigned bits) in extent_map_device_set_bits() argument
366 struct map_lookup *map = em->map_lookup; in extent_map_device_set_bits()
367 u64 stripe_size = em->orig_block_len; in extent_map_device_set_bits()
379 static void extent_map_device_clear_bits(struct extent_map *em, unsigned bits) in extent_map_device_clear_bits() argument
381 struct map_lookup *map = em->map_lookup; in extent_map_device_clear_bits()
382 u64 stripe_size = em->orig_block_len; in extent_map_device_clear_bits()
409 struct extent_map *em, int modified) in add_extent_mapping() argument
415 ret = tree_insert(&tree->map, em); in add_extent_mapping()
419 setup_extent_mapping(tree, em, modified); in add_extent_mapping()
420 if (test_bit(EXTENT_FLAG_FS_MAPPING, &em->flags)) { in add_extent_mapping()
421 extent_map_device_set_bits(em, CHUNK_ALLOCATED); in add_extent_mapping()
422 extent_map_device_clear_bits(em, CHUNK_TRIMMED); in add_extent_mapping()
432 struct extent_map *em; in __lookup_extent_mapping() local
445 em = rb_entry(rb_node, struct extent_map, rb_node); in __lookup_extent_mapping()
447 if (strict && !(end > em->start && start < extent_map_end(em))) in __lookup_extent_mapping()
450 refcount_inc(&em->refs); in __lookup_extent_mapping()
451 return em; in __lookup_extent_mapping()
496 void remove_extent_mapping(struct extent_map_tree *tree, struct extent_map *em) in remove_extent_mapping() argument
500 WARN_ON(test_bit(EXTENT_FLAG_PINNED, &em->flags)); in remove_extent_mapping()
501 rb_erase_cached(&em->rb_node, &tree->map); in remove_extent_mapping()
502 if (!test_bit(EXTENT_FLAG_LOGGING, &em->flags)) in remove_extent_mapping()
503 list_del_init(&em->list); in remove_extent_mapping()
504 if (test_bit(EXTENT_FLAG_FS_MAPPING, &em->flags)) in remove_extent_mapping()
505 extent_map_device_clear_bits(em, CHUNK_ALLOCATED); in remove_extent_mapping()
506 RB_CLEAR_NODE(&em->rb_node); in remove_extent_mapping()
526 static struct extent_map *next_extent_map(struct extent_map *em) in next_extent_map() argument
530 next = rb_next(&em->rb_node); in next_extent_map()
536 static struct extent_map *prev_extent_map(struct extent_map *em) in prev_extent_map() argument
540 prev = rb_prev(&em->rb_node); in prev_extent_map()
554 struct extent_map *em, in merge_extent_mapping() argument
563 BUG_ON(map_start < em->start || map_start >= extent_map_end(em)); in merge_extent_mapping()
573 start = prev ? extent_map_end(prev) : em->start; in merge_extent_mapping()
574 start = max_t(u64, start, em->start); in merge_extent_mapping()
575 end = next ? next->start : extent_map_end(em); in merge_extent_mapping()
576 end = min_t(u64, end, extent_map_end(em)); in merge_extent_mapping()
577 start_diff = start - em->start; in merge_extent_mapping()
578 em->start = start; in merge_extent_mapping()
579 em->len = end - start; in merge_extent_mapping()
580 if (em->block_start < EXTENT_MAP_LAST_BYTE && in merge_extent_mapping()
581 !test_bit(EXTENT_FLAG_COMPRESSED, &em->flags)) { in merge_extent_mapping()
582 em->block_start += start_diff; in merge_extent_mapping()
583 em->block_len = em->len; in merge_extent_mapping()
585 return add_extent_mapping(em_tree, em, 0); in merge_extent_mapping()
614 struct extent_map *em = *em_in; in btrfs_add_extent_mapping() local
616 ret = add_extent_mapping(em_tree, em, 0); in btrfs_add_extent_mapping()
628 trace_btrfs_handle_em_exist(fs_info, existing, em, start, len); in btrfs_add_extent_mapping()
636 free_extent_map(em); in btrfs_add_extent_mapping()
640 u64 orig_start = em->start; in btrfs_add_extent_mapping()
641 u64 orig_len = em->len; in btrfs_add_extent_mapping()
648 em, start); in btrfs_add_extent_mapping()
650 free_extent_map(em); in btrfs_add_extent_mapping()
674 struct extent_map *em; in drop_all_extent_maps_fast() local
678 em = rb_entry(node, struct extent_map, rb_node); in drop_all_extent_maps_fast()
679 clear_bit(EXTENT_FLAG_PINNED, &em->flags); in drop_all_extent_maps_fast()
680 clear_bit(EXTENT_FLAG_LOGGING, &em->flags); in drop_all_extent_maps_fast()
681 remove_extent_mapping(tree, em); in drop_all_extent_maps_fast()
682 free_extent_map(em); in drop_all_extent_maps_fast()
707 struct extent_map *em; in btrfs_drop_extent_map_range() local
735 em = lookup_extent_mapping(em_tree, start, len); in btrfs_drop_extent_map_range()
737 while (em) { in btrfs_drop_extent_map_range()
739 const u64 em_end = extent_map_end(em); in btrfs_drop_extent_map_range()
747 next_em = next_extent_map(em); in btrfs_drop_extent_map_range()
756 if (skip_pinned && test_bit(EXTENT_FLAG_PINNED, &em->flags)) { in btrfs_drop_extent_map_range()
763 clear_bit(EXTENT_FLAG_PINNED, &em->flags); in btrfs_drop_extent_map_range()
765 modified = !list_empty(&em->list); in btrfs_drop_extent_map_range()
771 if (em->start >= start && em_end <= end) in btrfs_drop_extent_map_range()
774 flags = em->flags; in btrfs_drop_extent_map_range()
775 gen = em->generation; in btrfs_drop_extent_map_range()
776 compressed = test_bit(EXTENT_FLAG_COMPRESSED, &em->flags); in btrfs_drop_extent_map_range()
778 if (em->start < start) { in btrfs_drop_extent_map_range()
785 split->start = em->start; in btrfs_drop_extent_map_range()
786 split->len = start - em->start; in btrfs_drop_extent_map_range()
788 if (em->block_start < EXTENT_MAP_LAST_BYTE) { in btrfs_drop_extent_map_range()
789 split->orig_start = em->orig_start; in btrfs_drop_extent_map_range()
790 split->block_start = em->block_start; in btrfs_drop_extent_map_range()
793 split->block_len = em->block_len; in btrfs_drop_extent_map_range()
797 em->orig_block_len); in btrfs_drop_extent_map_range()
798 split->ram_bytes = em->ram_bytes; in btrfs_drop_extent_map_range()
802 split->block_start = em->block_start; in btrfs_drop_extent_map_range()
809 split->compress_type = em->compress_type; in btrfs_drop_extent_map_range()
810 replace_extent_mapping(em_tree, em, split, modified); in btrfs_drop_extent_map_range()
824 split->block_start = em->block_start; in btrfs_drop_extent_map_range()
826 split->compress_type = em->compress_type; in btrfs_drop_extent_map_range()
829 if (em->block_start < EXTENT_MAP_LAST_BYTE) { in btrfs_drop_extent_map_range()
830 split->orig_block_len = max(em->block_len, in btrfs_drop_extent_map_range()
831 em->orig_block_len); in btrfs_drop_extent_map_range()
833 split->ram_bytes = em->ram_bytes; in btrfs_drop_extent_map_range()
835 split->block_len = em->block_len; in btrfs_drop_extent_map_range()
836 split->orig_start = em->orig_start; in btrfs_drop_extent_map_range()
838 const u64 diff = start + len - em->start; in btrfs_drop_extent_map_range()
842 split->orig_start = em->orig_start; in btrfs_drop_extent_map_range()
851 if (extent_map_in_tree(em)) { in btrfs_drop_extent_map_range()
852 replace_extent_mapping(em_tree, em, split, in btrfs_drop_extent_map_range()
868 if (extent_map_in_tree(em)) { in btrfs_drop_extent_map_range()
889 if ((em->start < start || em_end > end) && modified) { in btrfs_drop_extent_map_range()
893 remove_extent_mapping(em_tree, em); in btrfs_drop_extent_map_range()
900 free_extent_map(em); in btrfs_drop_extent_map_range()
903 free_extent_map(em); in btrfs_drop_extent_map_range()
905 em = next_em; in btrfs_drop_extent_map_range()