Lines Matching refs:vmap_area

777 static DEFINE_PER_CPU(struct vmap_area *, ne_fit_preload_node);
780 va_size(struct vmap_area *va) in va_size()
788 struct vmap_area *va; in get_subtree_max_size()
790 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size()
795 struct vmap_area, rb_node, unsigned long, subtree_max_size, va_size)
810 static struct vmap_area *find_vmap_area_exceed_addr(unsigned long addr) in find_vmap_area_exceed_addr()
812 struct vmap_area *va = NULL; in find_vmap_area_exceed_addr()
818 struct vmap_area *tmp; in find_vmap_area_exceed_addr()
820 tmp = rb_entry(n, struct vmap_area, rb_node); in find_vmap_area_exceed_addr()
834 static struct vmap_area *__find_vmap_area(unsigned long addr, struct rb_root *root) in __find_vmap_area()
841 struct vmap_area *va; in __find_vmap_area()
843 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area()
864 find_va_links(struct vmap_area *va, in find_va_links()
868 struct vmap_area *tmp_va; in find_va_links()
887 tmp_va = rb_entry(*link, struct vmap_area, rb_node); in find_va_links()
924 list = &rb_entry(parent, struct vmap_area, rb_node)->list; in get_va_next_sibling()
929 __link_va(struct vmap_area *va, struct rb_root *root, in __link_va()
938 head = &rb_entry(parent, struct vmap_area, rb_node)->list; in __link_va()
969 link_va(struct vmap_area *va, struct rb_root *root, in link_va()
977 link_va_augment(struct vmap_area *va, struct rb_root *root, in link_va_augment()
985 __unlink_va(struct vmap_area *va, struct rb_root *root, bool augment) in __unlink_va()
1001 unlink_va(struct vmap_area *va, struct rb_root *root) in unlink_va()
1007 unlink_va_augment(struct vmap_area *va, struct rb_root *root) in unlink_va_augment()
1017 compute_subtree_max_size(struct vmap_area *va) in compute_subtree_max_size()
1027 struct vmap_area *va; in augment_tree_propagate_check()
1067 augment_tree_propagate_from(struct vmap_area *va) in augment_tree_propagate_from()
1082 insert_vmap_area(struct vmap_area *va, in insert_vmap_area()
1094 insert_vmap_area_augment(struct vmap_area *va, in insert_vmap_area_augment()
1123 static __always_inline struct vmap_area *
1124 __merge_or_add_vmap_area(struct vmap_area *va, in __merge_or_add_vmap_area()
1127 struct vmap_area *sibling; in __merge_or_add_vmap_area()
1156 sibling = list_entry(next, struct vmap_area, list); in __merge_or_add_vmap_area()
1177 sibling = list_entry(next->prev, struct vmap_area, list); in __merge_or_add_vmap_area()
1207 static __always_inline struct vmap_area *
1208 merge_or_add_vmap_area(struct vmap_area *va, in merge_or_add_vmap_area()
1214 static __always_inline struct vmap_area *
1215 merge_or_add_vmap_area_augment(struct vmap_area *va, in merge_or_add_vmap_area_augment()
1226 is_within_this_va(struct vmap_area *va, unsigned long size, in is_within_this_va()
1251 static __always_inline struct vmap_area *
1255 struct vmap_area *va; in find_vmap_lowest_match()
1266 va = rb_entry(node, struct vmap_area, rb_node); in find_vmap_lowest_match()
1292 va = rb_entry(node, struct vmap_area, rb_node); in find_vmap_lowest_match()
1318 static struct vmap_area *
1322 struct vmap_area *va; in find_vmap_lowest_linear_match()
1338 struct vmap_area *va_1, *va_2; in find_vmap_lowest_match_check()
1363 classify_va_fit_type(struct vmap_area *va, in classify_va_fit_type()
1390 struct vmap_area *va, unsigned long nva_start_addr, in adjust_va_to_fit_type()
1393 struct vmap_area *lva = NULL; in adjust_va_to_fit_type()
1499 struct vmap_area *va; in __alloc_vmap_area()
1542 static void free_vmap_area(struct vmap_area *va) in free_vmap_area()
1562 struct vmap_area *va = NULL; in preload_this_cpu_lock()
1586 static struct vmap_area *alloc_vmap_area(unsigned long size, in alloc_vmap_area()
1591 struct vmap_area *va; in alloc_vmap_area()
1729 struct vmap_area *va, *n_va; in __purge_vmap_area_lazy()
1743 struct vmap_area, list)->va_start); in __purge_vmap_area_lazy()
1747 struct vmap_area, list)->va_end); in __purge_vmap_area_lazy()
1812 static void free_vmap_area_noflush(struct vmap_area *va) in free_vmap_area_noflush()
1839 static void free_unmap_vmap_area(struct vmap_area *va) in free_unmap_vmap_area()
1849 struct vmap_area *find_vmap_area(unsigned long addr) in find_vmap_area()
1851 struct vmap_area *va; in find_vmap_area()
1897 struct vmap_area *va;
1950 struct vmap_area *va; in new_vmap_block()
2209 struct vmap_area *va; in vm_unmap_ram()
2259 struct vmap_area *va; in vm_map_ram()
2368 struct vmap_area *busy, *free; in vmap_init_free_space()
2407 struct vmap_area *va; in vmalloc_init()
2414 vmap_area_cachep = KMEM_CACHE(vmap_area, SLAB_PANIC); in vmalloc_init()
2448 struct vmap_area *va, unsigned long flags, const void *caller) in setup_vmalloc_vm_locked()
2457 static void setup_vmalloc_vm(struct vm_struct *vm, struct vmap_area *va, in setup_vmalloc_vm()
2481 struct vmap_area *va; in __get_vm_area_node()
2571 struct vmap_area *va; in find_vm_area()
2592 struct vmap_area *va; in remove_vm_area()
3525 struct vmap_area *va; in vread()
3684 static struct vmap_area *node_to_va(struct rb_node *n) in node_to_va()
3686 return rb_entry_safe(n, struct vmap_area, rb_node); in node_to_va()
3698 static struct vmap_area *
3701 struct vmap_area *va, *tmp; in pvm_find_va_enclose_addr()
3708 tmp = rb_entry(n, struct vmap_area, rb_node); in pvm_find_va_enclose_addr()
3734 pvm_determine_end_from_reverse(struct vmap_area **va, unsigned long align) in pvm_determine_end_from_reverse()
3781 struct vmap_area **vas, *va; in pcpu_get_vm_areas()
4096 struct vmap_area *va; in show_purge_info()
4109 struct vmap_area *va; in s_show()
4112 va = list_entry(p, struct vmap_area, list); in s_show()