/Linux-v5.4/mm/ |
D | vmacache.c | 39 current->vmacache.vmas[VMACACHE_HASH(addr)] = newvma; in vmacache_update() 73 struct vm_area_struct *vma = current->vmacache.vmas[idx]; in vmacache_find() 106 struct vm_area_struct *vma = current->vmacache.vmas[idx]; in vmacache_find_exact()
|
D | gup.c | 782 struct vm_area_struct **vmas, int *nonblocking) in __get_user_pages() argument 826 i = follow_hugetlb_page(mm, vma, pages, vmas, in __get_user_pages() 878 if (vmas) { in __get_user_pages() 879 vmas[i] = vma; in __get_user_pages() 1003 struct vm_area_struct **vmas, in __get_user_pages_locked() argument 1012 BUG_ON(vmas); in __get_user_pages_locked() 1024 vmas, locked); in __get_user_pages_locked() 1152 struct vm_area_struct **vmas, int *locked) in get_user_pages_remote() argument 1163 return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas, in get_user_pages_remote() 1319 struct vm_area_struct **vmas, int *locked, in __get_user_pages_locked() argument [all …]
|
D | hugetlb.c | 4339 struct page **pages, struct vm_area_struct **vmas, in follow_hugetlb_page() argument 4468 if (vmas) in follow_hugetlb_page() 4469 vmas[i] = vma; in follow_hugetlb_page()
|
D | nommu.c | 656 if (curr->vmacache.vmas[i] == vma) { in delete_vma_from_mm()
|
/Linux-v5.4/include/linux/ |
D | vmacache.h | 10 memset(tsk->vmacache.vmas, 0, sizeof(tsk->vmacache.vmas)); in vmacache_flush()
|
D | mm_types_task.h | 36 struct vm_area_struct *vmas[VMACACHE_SIZE]; member
|
D | mm.h | 1523 struct vm_area_struct **vmas, int *locked); 1526 struct vm_area_struct **vmas);
|
/Linux-v5.4/drivers/gpu/drm/msm/ |
D | msm_gem.c | 343 list_add_tail(&vma->list, &msm_obj->vmas); in add_vma() 356 list_for_each_entry(vma, &msm_obj->vmas, list) { in lookup_vma() 382 list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) { in put_iova() 831 if (!list_empty(&msm_obj->vmas)) { in msm_gem_describe() 835 list_for_each_entry(vma, &msm_obj->vmas, list) in msm_gem_describe() 1007 INIT_LIST_HEAD(&msm_obj->vmas); in msm_gem_new_impl()
|
D | msm_gem.h | 74 struct list_head vmas; /* list of msm_gem_vma */ member
|
/Linux-v5.4/drivers/video/fbdev/vermilion/ |
D | vermilion.h | 210 atomic_t vmas; member
|
/Linux-v5.4/kernel/debug/ |
D | debug_core.c | 293 if (!current->vmacache.vmas[i]) in kgdb_flush_swbreak_addr() 295 flush_cache_range(current->vmacache.vmas[i], in kgdb_flush_swbreak_addr()
|
/Linux-v5.4/fs/ |
D | io_uring.c | 3373 struct vm_area_struct **vmas = NULL; in io_sqe_buffer_register() local 3425 kfree(vmas); in io_sqe_buffer_register() 3429 vmas = kvmalloc_array(nr_pages, in io_sqe_buffer_register() 3432 if (!pages || !vmas) { in io_sqe_buffer_register() 3454 pages, vmas); in io_sqe_buffer_register() 3458 struct vm_area_struct *vma = vmas[j]; in io_sqe_buffer_register() 3503 kvfree(vmas); in io_sqe_buffer_register() 3507 kvfree(vmas); in io_sqe_buffer_register()
|
/Linux-v5.4/Documentation/ |
D | robust-futexes.txt | 58 FUTEX_RECOVER. At do_exit() time, all vmas are searched to see whether 69 microsecond on Linux, but with thousands (or tens of thousands) of vmas
|
/Linux-v5.4/drivers/vfio/ |
D | vfio_iommu_type1.c | 343 struct vm_area_struct *vmas[1]; in vaddr_get_pfn() local 353 vmas); in vaddr_get_pfn() 356 vmas, NULL); in vaddr_get_pfn() 364 if (ret > 0 && vma_is_fsdax(vmas[0])) { in vaddr_get_pfn()
|
/Linux-v5.4/Documentation/admin-guide/mm/ |
D | userfaultfd.rst | 35 operations never involve heavyweight structures like vmas (in fact the 40 Terabytes. Too many vmas would be needed for that.
|