Lines Matching refs:va_start

351 		if (addr < va->va_start)  in __find_vmap_area()
373 if (va->va_start < tmp_va->va_end) in __insert_vmap_area()
375 else if (va->va_end > tmp_va->va_start) in __insert_vmap_area()
475 if (tmp->va_start <= addr) in alloc_vmap_area()
487 while (addr + size > first->va_start && addr + size <= vend) { in alloc_vmap_area()
488 if (addr + cached_hole_size < first->va_start) in alloc_vmap_area()
489 cached_hole_size = first->va_start - addr; in alloc_vmap_area()
504 va->va_start = addr; in alloc_vmap_area()
511 BUG_ON(!IS_ALIGNED(va->va_start, align)); in alloc_vmap_area()
512 BUG_ON(va->va_start < vstart); in alloc_vmap_area()
563 if (va->va_start <= cache->va_start) { in __free_vmap_area()
603 vunmap_page_range(va->va_start, va->va_end); in unmap_vmap_area()
666 if (va->va_start < start) in __purge_vmap_area_lazy()
667 start = va->va_start; in __purge_vmap_area_lazy()
680 int nr = (va->va_end - va->va_start) >> PAGE_SHIFT; in __purge_vmap_area_lazy()
722 nr_lazy = atomic_add_return((va->va_end - va->va_start) >> PAGE_SHIFT, in free_vmap_area_noflush()
737 flush_cache_vunmap(va->va_start, va->va_end); in free_unmap_vmap_area()
740 flush_tlb_kernel_range(va->va_start, va->va_end); in free_unmap_vmap_area()
828 static void *vmap_block_vaddr(unsigned long va_start, unsigned long pages_off) in vmap_block_vaddr() argument
832 addr = va_start + (pages_off << PAGE_SHIFT); in vmap_block_vaddr()
833 BUG_ON(addr_to_vb_idx(addr) != addr_to_vb_idx(va_start)); in vmap_block_vaddr()
876 vaddr = vmap_block_vaddr(va->va_start, 0); in new_vmap_block()
887 vb_idx = addr_to_vb_idx(va->va_start); in new_vmap_block()
908 vb_idx = addr_to_vb_idx(vb->va->va_start); in free_vmap_block()
992 vaddr = vmap_block_vaddr(vb->va->va_start, pages_off); in vb_alloc()
1090 unsigned long va_start = vb->va->va_start; in vm_unmap_aliases() local
1093 s = va_start + (vb->dirty_min << PAGE_SHIFT); in vm_unmap_aliases()
1094 e = va_start + (vb->dirty_max << PAGE_SHIFT); in vm_unmap_aliases()
1139 debug_check_no_locks_freed((void *)va->va_start, in vm_unmap_ram()
1140 (va->va_end - va->va_start)); in vm_unmap_ram()
1178 addr = va->va_start; in vm_map_ram()
1263 va->va_start = (unsigned long)tmp->addr; in vmalloc_init()
1264 va->va_end = va->va_start + tmp->size; in vmalloc_init()
1354 vm->addr = (void *)va->va_start; in setup_vmalloc_vm()
1355 vm->size = va->va_end - va->va_start; in setup_vmalloc_vm()
2430 addr = min((*pnext)->va_start & ~(align - 1), vmalloc_end); in pvm_determine_end()
2553 if (next && next->va_start < base + end) { in pcpu_get_vm_areas()
2588 va->va_start = base + offsets[area]; in pcpu_get_vm_areas()
2589 va->va_end = va->va_start + sizes[area]; in pcpu_get_vm_areas()
2689 (void *)va->va_start, (void *)va->va_end, in s_show()
2690 va->va_end - va->va_start, in s_show()