Lines Matching refs:vm_end
298 (vma_mt->vm_end - 1 != mas.last)) { in validate_mm_mt()
305 vma_mt->vm_start, vma_mt->vm_end); in validate_mm_mt()
308 if (vma_mt->vm_end != mas.last + 1) { in validate_mm_mt()
310 mm, vma_mt->vm_start, vma_mt->vm_end, in validate_mm_mt()
314 VM_BUG_ON_MM(vma_mt->vm_end != mas.last + 1, mm); in validate_mm_mt()
317 mm, vma_mt, vma_mt->vm_start, vma_mt->vm_end); in validate_mm_mt()
401 unsigned long vm_end = min(end, vma->vm_end); in count_vma_pages_range() local
403 nr_pages += PHYS_PFN(vm_end - vm_start); in count_vma_pages_range()
433 mas_set_range(mas, vma->vm_start, vma->vm_end - 1); in vma_mas_store()
448 trace_vma_mas_szero(mas->tree, vma->vm_start, vma->vm_end - 1); in vma_mas_remove()
450 mas->last = vma->vm_end - 1; in vma_mas_remove()
523 if (next && (vma != next) && (end == next->vm_end)) { in vma_expand()
539 VM_BUG_ON(vma->vm_start < start || vma->vm_end > end); in vma_expand()
549 uprobe_munmap(vma, vma->vm_start, vma->vm_end); in vma_expand()
564 vma->vm_end = end; in vma_expand()
591 uprobe_munmap(next, next->vm_start, next->vm_end); in vma_expand()
621 struct vm_area_struct *next = find_vma(mm, vma->vm_end); in __vma_adjust()
634 if (end >= next->vm_end) { in __vma_adjust()
646 VM_WARN_ON(end != next->vm_end); in __vma_adjust()
661 remove_next = 1 + (end > next->vm_end); in __vma_adjust()
663 next_next = find_vma(mm, next->vm_end); in __vma_adjust()
666 end != next_next->vm_end); in __vma_adjust()
688 } else if (end < vma->vm_end) { in __vma_adjust()
694 adjust_next = -(vma->vm_end - end); in __vma_adjust()
722 uprobe_munmap(vma, vma->vm_start, vma->vm_end); in __vma_adjust()
725 uprobe_munmap(next, next->vm_start, next->vm_end); in __vma_adjust()
760 (!insert || (insert->vm_end != start))) { in __vma_adjust()
768 if (end != vma->vm_end) { in __vma_adjust()
769 if (vma->vm_end > end) { in __vma_adjust()
771 vma_mas_szero(&mas, end, vma->vm_end); in __vma_adjust()
774 insert->vm_end < vma->vm_end); in __vma_adjust()
779 vma->vm_end = end; in __vma_adjust()
832 uprobe_munmap(next, next->vm_start, next->vm_end); in __vma_adjust()
840 BUG_ON(vma->vm_end < next->vm_end); in __vma_adjust()
1020 next = find_vma(mm, prev ? prev->vm_end : 0); in vma_merge()
1022 if (next && next->vm_end == end) /* cases 6, 7, 8 */ in vma_merge()
1023 next = find_vma(mm, next->vm_end); in vma_merge()
1027 VM_WARN_ON(mid && end > mid->vm_end); in vma_merge()
1031 if (prev && prev->vm_end == addr && in vma_merge()
1051 next->vm_end, prev->vm_pgoff, NULL, in vma_merge()
1059 if (prev && addr < prev->vm_end) /* case 4 */ in vma_merge()
1063 err = __vma_adjust(mid, addr, next->vm_end, in vma_merge()
1092 return a->vm_end == b->vm_start && in anon_vma_compatible()
1142 MA_STATE(mas, &vma->vm_mm->mm_mt, vma->vm_end, vma->vm_end); in find_mergeable_anon_vma()
1887 vma->vm_end - size; in acct_stack_growth()
1930 next = find_vma_intersection(mm, vma->vm_end, gap_addr); in expand_upwards()
1954 if (address > vma->vm_end) { in expand_upwards()
1958 grow = (address - vma->vm_end) >> PAGE_SHIFT; in expand_upwards()
1978 vma->vm_end = address; in expand_upwards()
2014 if (address - prev->vm_end < stack_guard_gap) in expand_downwards()
2038 size = vma->vm_end - address; in expand_downwards()
2110 populate_vma_page_range(prev, addr, prev->vm_end, NULL); in find_extend_vma()
2185 free_pgtables(&tlb, mt, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS, in unmap_region()
2212 new->vm_end = addr; in __split_vma()
2233 err = vma_adjust(vma, addr, vma->vm_end, vma->vm_pgoff + in __split_vma()
2243 new->vm_start = new->vm_end; in __split_vma()
2275 mas_set_range(mas_detach, vma->vm_start, vma->vm_end - 1); in munmap_sidetree()
2330 if (end < vma->vm_end && mm->map_count >= sysctl_max_map_count) in do_mas_align_munmap()
2355 if (next->vm_end > end) { in do_mas_align_munmap()
2569 merge_end = next->vm_end; in mmap_region()
2575 if (prev && prev->vm_end == addr && !vma_policy(prev) && in mmap_region()
2608 vma->vm_end = end; in mmap_region()
2640 merge = vma_merge(mm, prev, vma->vm_start, vma->vm_end, vma->vm_flags, in mmap_region()
2753 unmap_region(mm, mas.tree, vma, prev, next, vma->vm_start, vma->vm_end); in mmap_region()
2840 if (start + size > vma->vm_end) { in SYSCALL_DEFINE5()
2841 VMA_ITERATOR(vmi, mm, vma->vm_end); in SYSCALL_DEFINE5()
2846 if (next->vm_start != prev->vm_end) in SYSCALL_DEFINE5()
2855 if (start + size <= next->vm_end) in SYSCALL_DEFINE5()
2948 if (vma && vma->vm_end == addr && !vma_policy(vma) && in do_brk_flags()
2960 vma->vm_end = addr + len; in do_brk_flags()
2979 vma->vm_end = addr + len; in do_brk_flags()
3128 if (find_vma_intersection(mm, vma->vm_start, vma->vm_end)) in insert_vm_struct()
3196 vma_start < new_vma->vm_end)) { in copy_vma()
3218 new_vma->vm_end = addr + len; in copy_vma()
3391 vma->vm_end = addr + len; in __install_special_mapping()