Lines Matching refs:vm_start

103 			return vma->vm_end - vma->vm_start;  in kobjsize()
433 BUG_ON(last->vm_end <= last->vm_start); in validate_nommu_regions()
440 BUG_ON(region->vm_end <= region->vm_start); in validate_nommu_regions()
442 BUG_ON(region->vm_start < last->vm_top); in validate_nommu_regions()
468 if (region->vm_start < pregion->vm_start) in add_nommu_region()
470 else if (region->vm_start > pregion->vm_start) in add_nommu_region()
521 if (region->vm_top > region->vm_start) in __put_nommu_region()
531 free_page_series(region->vm_start, region->vm_top); in __put_nommu_region()
584 VMA_ITERATOR(vmi, vma->vm_mm, vma->vm_start); in delete_vma_from_mm()
586 vma_iter_config(&vmi, vma->vm_start, vma->vm_end); in delete_vma_from_mm()
680 if (vma->vm_start != addr) in find_vma_exact()
969 region->vm_start = (unsigned long) base; in do_mmap_private()
970 region->vm_end = region->vm_start + len; in do_mmap_private()
971 region->vm_top = region->vm_start + (total << PAGE_SHIFT); in do_mmap_private()
973 vma->vm_start = region->vm_start; in do_mmap_private()
974 vma->vm_end = region->vm_start + len; in do_mmap_private()
998 free_page_series(region->vm_start, region->vm_top); in do_mmap_private()
999 region->vm_start = vma->vm_start = 0; in do_mmap_private()
1101 rpglen = pregion->vm_end - pregion->vm_start; in do_mmap()
1120 start = pregion->vm_start; in do_mmap()
1122 vma->vm_start = start; in do_mmap()
1131 vma->vm_start = 0; in do_mmap()
1166 vma->vm_start = region->vm_start = addr; in do_mmap()
1189 memset((void *)region->vm_start, 0, in do_mmap()
1190 region->vm_end - region->vm_start); in do_mmap()
1193 result = vma->vm_start; in do_mmap()
1199 vma_iter_config(&vmi, vma->vm_start, vma->vm_end); in do_mmap()
1211 flush_icache_user_range(region->vm_start, region->vm_end); in do_mmap()
1337 npages = (addr - vma->vm_start) >> PAGE_SHIFT; in split_vma()
1342 region->vm_start = new->vm_start = addr; in split_vma()
1346 vma_iter_config(vmi, new->vm_start, new->vm_end); in split_vma()
1359 vma->vm_region->vm_start = vma->vm_start = addr; in split_vma()
1394 if (from > vma->vm_start) { in vmi_shrink_vma()
1399 if (vma_iter_clear_gfp(vmi, vma->vm_start, to, GFP_KERNEL)) in vmi_shrink_vma()
1401 vma->vm_start = to; in vmi_shrink_vma()
1410 if (from > region->vm_start) { in vmi_shrink_vma()
1414 region->vm_start = to; in vmi_shrink_vma()
1457 if (start > vma->vm_start) in do_munmap()
1466 if (start == vma->vm_start && end == vma->vm_end) in do_munmap()
1468 if (start < vma->vm_start || end > vma->vm_end) in do_munmap()
1474 if (start != vma->vm_start && end != vma->vm_end) { in do_munmap()
1571 if (vma->vm_end != vma->vm_start + old_len) in do_mremap()
1577 if (new_len > vma->vm_region->vm_end - vma->vm_region->vm_start) in do_mremap()
1581 vma->vm_end = vma->vm_start + new_len; in do_mremap()
1582 return vma->vm_start; in do_mremap()
1617 unsigned long vm_len = vma->vm_end - vma->vm_start; in vm_iomap_memory()
1620 return io_remap_pfn_range(vma, vma->vm_start, pfn, vm_len, vma->vm_page_prot); in vm_iomap_memory()
1627 unsigned int size = vma->vm_end - vma->vm_start; in remap_vmalloc_range()
1632 vma->vm_start = (unsigned long)(addr + (pgoff << PAGE_SHIFT)); in remap_vmalloc_range()
1633 vma->vm_end = vma->vm_start + size; in remap_vmalloc_range()
1774 r_size = region->vm_top - region->vm_start; in nommu_shrink_inode_mappings()