Lines Matching refs:va_end
353 else if (addr >= va->va_end) in __find_vmap_area()
373 if (va->va_start < tmp_va->va_end) in __insert_vmap_area()
375 else if (va->va_end > tmp_va->va_start) in __insert_vmap_area()
456 addr = ALIGN(first->va_end, align); in alloc_vmap_area()
473 if (tmp->va_end >= addr) { in alloc_vmap_area()
490 addr = ALIGN(first->va_end, align); in alloc_vmap_area()
505 va->va_end = addr + size; in alloc_vmap_area()
513 BUG_ON(va->va_end > vend); in alloc_vmap_area()
558 if (va->va_end < cached_vstart) { in __free_vmap_area()
582 if (va->va_end > VMALLOC_START && va->va_end <= VMALLOC_END) in __free_vmap_area()
583 vmap_area_pcpu_hole = max(vmap_area_pcpu_hole, va->va_end); in __free_vmap_area()
603 vunmap_page_range(va->va_start, va->va_end); in unmap_vmap_area()
668 if (va->va_end > end) in __purge_vmap_area_lazy()
669 end = va->va_end; in __purge_vmap_area_lazy()
680 int nr = (va->va_end - va->va_start) >> PAGE_SHIFT; in __purge_vmap_area_lazy()
722 nr_lazy = atomic_add_return((va->va_end - va->va_start) >> PAGE_SHIFT, in free_vmap_area_noflush()
737 flush_cache_vunmap(va->va_start, va->va_end); in free_unmap_vmap_area()
740 flush_tlb_kernel_range(va->va_start, va->va_end); in free_unmap_vmap_area()
1140 (va->va_end - va->va_start)); in vm_unmap_ram()
1264 va->va_end = va->va_start + tmp->size; in vmalloc_init()
1355 vm->size = va->va_end - va->va_start; in setup_vmalloc_vm()
2385 if (end < va->va_end) in pvm_find_next_prev()
2387 else if (end > va->va_end) in pvm_find_next_prev()
2396 if (va->va_end > end) { in pvm_find_next_prev()
2434 while (*pprev && (*pprev)->va_end > addr) { in pvm_determine_end()
2532 BUG_ON(next && next->va_end <= base + end); in pcpu_get_vm_areas()
2533 BUG_ON(prev && prev->va_end > base + end); in pcpu_get_vm_areas()
2564 if (prev && prev->va_end > base + start) { in pcpu_get_vm_areas()
2589 va->va_end = va->va_start + sizes[area]; in pcpu_get_vm_areas()
2689 (void *)va->va_start, (void *)va->va_end, in s_show()
2690 va->va_end - va->va_start, in s_show()