Lines Matching refs:vmap_area

371 static DEFINE_PER_CPU(struct vmap_area *, ne_fit_preload_node);
374 va_size(struct vmap_area *va) in va_size()
382 struct vmap_area *va; in get_subtree_max_size()
384 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size()
392 compute_subtree_max_size(struct vmap_area *va) in compute_subtree_max_size()
400 struct vmap_area, rb_node, unsigned long, subtree_max_size, va_size)
413 static struct vmap_area *__find_vmap_area(unsigned long addr) in __find_vmap_area()
418 struct vmap_area *va; in __find_vmap_area()
420 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area()
437 find_va_links(struct vmap_area *va, in find_va_links()
441 struct vmap_area *tmp_va; in find_va_links()
460 tmp_va = rb_entry(*link, struct vmap_area, rb_node); in find_va_links()
495 list = &rb_entry(parent, struct vmap_area, rb_node)->list; in get_va_next_sibling()
500 link_va(struct vmap_area *va, struct rb_root *root, in link_va()
508 head = &rb_entry(parent, struct vmap_area, rb_node)->list; in link_va()
539 unlink_va(struct vmap_area *va, struct rb_root *root) in unlink_va()
558 struct vmap_area *va; in augment_tree_propagate_check()
566 va = rb_entry(n, struct vmap_area, rb_node); in augment_tree_propagate_check()
571 va = rb_entry(node, struct vmap_area, rb_node); in augment_tree_propagate_check()
586 va = rb_entry(n, struct vmap_area, rb_node); in augment_tree_propagate_check()
624 augment_tree_propagate_from(struct vmap_area *va) in augment_tree_propagate_from()
630 va = rb_entry(node, struct vmap_area, rb_node); in augment_tree_propagate_from()
652 insert_vmap_area(struct vmap_area *va, in insert_vmap_area()
663 insert_vmap_area_augment(struct vmap_area *va, in insert_vmap_area_augment()
686 merge_or_add_vmap_area(struct vmap_area *va, in merge_or_add_vmap_area()
689 struct vmap_area *sibling; in merge_or_add_vmap_area()
716 sibling = list_entry(next, struct vmap_area, list); in merge_or_add_vmap_area()
740 sibling = list_entry(next->prev, struct vmap_area, list); in merge_or_add_vmap_area()
764 is_within_this_va(struct vmap_area *va, unsigned long size, in is_within_this_va()
787 static __always_inline struct vmap_area *
791 struct vmap_area *va; in find_vmap_lowest_match()
802 va = rb_entry(node, struct vmap_area, rb_node); in find_vmap_lowest_match()
827 va = rb_entry(node, struct vmap_area, rb_node); in find_vmap_lowest_match()
846 static struct vmap_area *
850 struct vmap_area *va; in find_vmap_lowest_linear_match()
865 struct vmap_area *va_1, *va_2; in find_vmap_lowest_match_check()
890 classify_va_fit_type(struct vmap_area *va, in classify_va_fit_type()
916 adjust_va_to_fit_type(struct vmap_area *va, in adjust_va_to_fit_type()
920 struct vmap_area *lva = NULL; in adjust_va_to_fit_type()
1011 struct vmap_area *va; in __alloc_vmap_area()
1049 static struct vmap_area *alloc_vmap_area(unsigned long size, in alloc_vmap_area()
1054 struct vmap_area *va, *pva; in alloc_vmap_area()
1164 static void __free_vmap_area(struct vmap_area *va) in __free_vmap_area()
1181 static void free_vmap_area(struct vmap_area *va) in free_vmap_area()
1191 static void unmap_vmap_area(struct vmap_area *va) in unmap_vmap_area()
1249 struct vmap_area *va; in __purge_vmap_area_lazy()
1250 struct vmap_area *n_va; in __purge_vmap_area_lazy()
1327 static void free_vmap_area_noflush(struct vmap_area *va) in free_vmap_area_noflush()
1348 static void free_unmap_vmap_area(struct vmap_area *va) in free_unmap_vmap_area()
1358 static struct vmap_area *find_vmap_area(unsigned long addr) in find_vmap_area()
1360 struct vmap_area *va; in find_vmap_area()
1406 struct vmap_area *va;
1460 struct vmap_area *va; in new_vmap_block()
1740 struct vmap_area *va; in vm_unmap_ram()
1789 struct vmap_area *va; in vm_map_ram()
1863 struct vmap_area *busy, *free; in vmap_init_free_space()
1902 struct vmap_area *va; in vmalloc_init()
1909 vmap_area_cachep = KMEM_CACHE(vmap_area, SLAB_PANIC); in vmalloc_init()
2017 static void setup_vmalloc_vm(struct vm_struct *vm, struct vmap_area *va, in setup_vmalloc_vm()
2044 struct vmap_area *va; in __get_vm_area_node()
2127 struct vmap_area *va; in find_vm_area()
2148 struct vmap_area *va; in remove_vm_area()
2847 struct vmap_area *va; in vread()
2926 struct vmap_area *va; in vwrite()
3122 static struct vmap_area *node_to_va(struct rb_node *n) in node_to_va()
3124 return rb_entry_safe(n, struct vmap_area, rb_node); in node_to_va()
3136 static struct vmap_area *
3139 struct vmap_area *va, *tmp; in pvm_find_va_enclose_addr()
3146 tmp = rb_entry(n, struct vmap_area, rb_node); in pvm_find_va_enclose_addr()
3171 pvm_determine_end_from_reverse(struct vmap_area **va, unsigned long align) in pvm_determine_end_from_reverse()
3218 struct vmap_area **vas, *va; in pcpu_get_vm_areas()
3464 struct vmap_area *va; in show_purge_info()
3479 struct vmap_area *va; in s_show()
3482 va = list_entry(p, struct vmap_area, list); in s_show()