Lines Matching refs:vmap_area

773 static DEFINE_PER_CPU(struct vmap_area *, ne_fit_preload_node);
776 va_size(struct vmap_area *va) in va_size()
784 struct vmap_area *va; in get_subtree_max_size()
786 va = rb_entry_safe(node, struct vmap_area, rb_node); in get_subtree_max_size()
791 struct vmap_area, rb_node, unsigned long, subtree_max_size, va_size)
806 static struct vmap_area *find_vmap_area_exceed_addr(unsigned long addr) in find_vmap_area_exceed_addr()
808 struct vmap_area *va = NULL; in find_vmap_area_exceed_addr()
814 struct vmap_area *tmp; in find_vmap_area_exceed_addr()
816 tmp = rb_entry(n, struct vmap_area, rb_node); in find_vmap_area_exceed_addr()
830 static struct vmap_area *__find_vmap_area(unsigned long addr, struct rb_root *root) in __find_vmap_area()
837 struct vmap_area *va; in __find_vmap_area()
839 va = rb_entry(n, struct vmap_area, rb_node); in __find_vmap_area()
860 find_va_links(struct vmap_area *va, in find_va_links()
864 struct vmap_area *tmp_va; in find_va_links()
883 tmp_va = rb_entry(*link, struct vmap_area, rb_node); in find_va_links()
920 list = &rb_entry(parent, struct vmap_area, rb_node)->list; in get_va_next_sibling()
925 __link_va(struct vmap_area *va, struct rb_root *root, in __link_va()
934 head = &rb_entry(parent, struct vmap_area, rb_node)->list; in __link_va()
965 link_va(struct vmap_area *va, struct rb_root *root, in link_va()
973 link_va_augment(struct vmap_area *va, struct rb_root *root, in link_va_augment()
981 __unlink_va(struct vmap_area *va, struct rb_root *root, bool augment) in __unlink_va()
997 unlink_va(struct vmap_area *va, struct rb_root *root) in unlink_va()
1003 unlink_va_augment(struct vmap_area *va, struct rb_root *root) in unlink_va_augment()
1013 compute_subtree_max_size(struct vmap_area *va) in compute_subtree_max_size()
1023 struct vmap_area *va; in augment_tree_propagate_check()
1063 augment_tree_propagate_from(struct vmap_area *va) in augment_tree_propagate_from()
1078 insert_vmap_area(struct vmap_area *va, in insert_vmap_area()
1090 insert_vmap_area_augment(struct vmap_area *va, in insert_vmap_area_augment()
1119 static __always_inline struct vmap_area *
1120 __merge_or_add_vmap_area(struct vmap_area *va, in __merge_or_add_vmap_area()
1123 struct vmap_area *sibling; in __merge_or_add_vmap_area()
1152 sibling = list_entry(next, struct vmap_area, list); in __merge_or_add_vmap_area()
1173 sibling = list_entry(next->prev, struct vmap_area, list); in __merge_or_add_vmap_area()
1203 static __always_inline struct vmap_area *
1204 merge_or_add_vmap_area(struct vmap_area *va, in merge_or_add_vmap_area()
1210 static __always_inline struct vmap_area *
1211 merge_or_add_vmap_area_augment(struct vmap_area *va, in merge_or_add_vmap_area_augment()
1222 is_within_this_va(struct vmap_area *va, unsigned long size, in is_within_this_va()
1247 static __always_inline struct vmap_area *
1251 struct vmap_area *va; in find_vmap_lowest_match()
1262 va = rb_entry(node, struct vmap_area, rb_node); in find_vmap_lowest_match()
1288 va = rb_entry(node, struct vmap_area, rb_node); in find_vmap_lowest_match()
1314 static struct vmap_area *
1318 struct vmap_area *va; in find_vmap_lowest_linear_match()
1334 struct vmap_area *va_1, *va_2; in find_vmap_lowest_match_check()
1359 classify_va_fit_type(struct vmap_area *va, in classify_va_fit_type()
1386 struct vmap_area *va, unsigned long nva_start_addr, in adjust_va_to_fit_type()
1389 struct vmap_area *lva = NULL; in adjust_va_to_fit_type()
1495 struct vmap_area *va; in __alloc_vmap_area()
1538 static void free_vmap_area(struct vmap_area *va) in free_vmap_area()
1558 struct vmap_area *va = NULL; in preload_this_cpu_lock()
1582 static struct vmap_area *alloc_vmap_area(unsigned long size, in alloc_vmap_area()
1588 struct vmap_area *va; in alloc_vmap_area()
1729 struct vmap_area *va, *n_va; in __purge_vmap_area_lazy()
1743 struct vmap_area, list)->va_start); in __purge_vmap_area_lazy()
1747 struct vmap_area, list)->va_end); in __purge_vmap_area_lazy()
1817 static void free_vmap_area_noflush(struct vmap_area *va) in free_vmap_area_noflush()
1847 static void free_unmap_vmap_area(struct vmap_area *va) in free_unmap_vmap_area()
1857 struct vmap_area *find_vmap_area(unsigned long addr) in find_vmap_area()
1859 struct vmap_area *va; in find_vmap_area()
1868 static struct vmap_area *find_unlink_vmap_area(unsigned long addr) in find_unlink_vmap_area()
1870 struct vmap_area *va; in find_unlink_vmap_area()
1935 struct vmap_area *va;
2026 struct vmap_area *va; in new_vmap_block()
2342 struct vmap_area *va; in vm_unmap_ram()
2394 struct vmap_area *va; in vm_map_ram()
2504 struct vmap_area *busy, *free; in vmap_init_free_space()
2542 struct vmap_area *va, unsigned long flags, const void *caller) in setup_vmalloc_vm_locked()
2551 static void setup_vmalloc_vm(struct vm_struct *vm, struct vmap_area *va, in setup_vmalloc_vm()
2575 struct vmap_area *va; in __get_vm_area_node()
2665 struct vmap_area *va; in find_vm_area()
2686 struct vmap_area *va; in remove_vm_area()
3744 struct vmap_area *va; in vread_iter()
3931 static struct vmap_area *node_to_va(struct rb_node *n) in node_to_va()
3933 return rb_entry_safe(n, struct vmap_area, rb_node); in node_to_va()
3945 static struct vmap_area *
3948 struct vmap_area *va, *tmp; in pvm_find_va_enclose_addr()
3955 tmp = rb_entry(n, struct vmap_area, rb_node); in pvm_find_va_enclose_addr()
3981 pvm_determine_end_from_reverse(struct vmap_area **va, unsigned long align) in pvm_determine_end_from_reverse()
4028 struct vmap_area **vas, *va; in pcpu_get_vm_areas()
4284 struct vmap_area *va; in vmalloc_dump_obj()
4361 struct vmap_area *va; in show_purge_info()
4374 struct vmap_area *va; in s_show()
4377 va = list_entry(p, struct vmap_area, list); in s_show()
4456 struct vmap_area *va; in vmalloc_init()
4463 vmap_area_cachep = KMEM_CACHE(vmap_area, SLAB_PANIC); in vmalloc_init()