Home
last modified time | relevance | path

Searched refs:vmas (Results 1 – 16 of 16) sorted by relevance

/Linux-v5.15/mm/damon/
Dvaddr-test.h17 static void __link_vmas(struct vm_area_struct *vmas, ssize_t nr_vmas) in __link_vmas() argument
26 vmas[i].vm_next = &vmas[i + 1]; in __link_vmas()
28 vmas[i].vm_rb.rb_left = NULL; in __link_vmas()
29 vmas[i].vm_rb.rb_right = &vmas[i + 1].vm_rb; in __link_vmas()
35 gap = vmas[j].vm_start - vmas[j - 1].vm_end; in __link_vmas()
39 vmas[i].rb_subtree_gap = largest_gap; in __link_vmas()
41 vmas[i].vm_next = NULL; in __link_vmas()
42 vmas[i].vm_rb.rb_right = NULL; in __link_vmas()
43 vmas[i].rb_subtree_gap = 0; in __link_vmas()
77 struct vm_area_struct vmas[] = { in damon_test_three_regions_in_vmas() local
[all …]
/Linux-v5.15/mm/
Dvmacache.c38 current->vmacache.vmas[VMACACHE_HASH(addr)] = newvma; in vmacache_update()
72 struct vm_area_struct *vma = current->vmacache.vmas[idx]; in vmacache_find()
105 struct vm_area_struct *vma = current->vmacache.vmas[idx]; in vmacache_find_exact()
Dgup.c1082 struct vm_area_struct **vmas, int *locked) in __get_user_pages() argument
1130 i = follow_hugetlb_page(mm, vma, pages, vmas, in __get_user_pages()
1190 if (vmas) { in __get_user_pages()
1191 vmas[i] = vma; in __get_user_pages()
1311 struct vm_area_struct **vmas, in __get_user_pages_locked() argument
1320 BUG_ON(vmas); in __get_user_pages_locked()
1344 vmas, locked); in __get_user_pages_locked()
1617 struct vm_area_struct **vmas, int *locked, in __get_user_pages_locked() argument
1647 if (vmas) in __get_user_pages_locked()
1648 vmas[i] = vma; in __get_user_pages_locked()
[all …]
Dhugetlb.c5406 struct vm_area_struct **vmas) in record_subpages_vmas() argument
5413 if (vmas) in record_subpages_vmas()
5414 vmas[nr] = vma; in record_subpages_vmas()
5419 struct page **pages, struct vm_area_struct **vmas, in follow_hugetlb_page() argument
5537 if (!pages && !vmas && !pfn_offset && in follow_hugetlb_page()
5551 if (pages || vmas) in follow_hugetlb_page()
5555 vmas ? vmas + i : NULL); in follow_hugetlb_page()
Dnommu.c627 if (curr->vmacache.vmas[i] == vma) { in delete_vma_from_mm()
/Linux-v5.15/include/linux/
Dvmacache.h10 memset(tsk->vmacache.vmas, 0, sizeof(tsk->vmacache.vmas)); in vmacache_flush()
Dmm_types_task.h36 struct vm_area_struct *vmas[VMACACHE_SIZE]; member
Dmm.h1819 struct vm_area_struct **vmas, int *locked);
1823 struct vm_area_struct **vmas, int *locked);
1826 struct vm_area_struct **vmas);
1829 struct vm_area_struct **vmas);
Dhugetlb.h247 struct vm_area_struct **vmas, unsigned long *position, in follow_hugetlb_page() argument
/Linux-v5.15/drivers/gpu/drm/msm/
Dmsm_gem.c311 list_add_tail(&vma->list, &msm_obj->vmas); in add_vma()
324 list_for_each_entry(vma, &msm_obj->vmas, list) { in lookup_vma()
355 list_for_each_entry(vma, &msm_obj->vmas, list) { in put_iova_spaces()
373 list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) { in put_iova_vmas()
928 if (!list_empty(&msm_obj->vmas)) { in msm_gem_describe()
932 list_for_each_entry(vma, &msm_obj->vmas, list) { in msm_gem_describe()
1135 INIT_LIST_HEAD(&msm_obj->vmas); in msm_gem_new_impl()
Dmsm_gem.h95 struct list_head vmas; /* list of msm_gem_vma */ member
/Linux-v5.15/drivers/video/fbdev/vermilion/
Dvermilion.h210 atomic_t vmas; member
/Linux-v5.15/kernel/debug/
Ddebug_core.c289 if (!current->vmacache.vmas[i]) in kgdb_flush_swbreak_addr()
291 flush_cache_range(current->vmacache.vmas[i], in kgdb_flush_swbreak_addr()
/Linux-v5.15/Documentation/locking/
Drobust-futexes.rst58 FUTEX_RECOVER. At do_exit() time, all vmas are searched to see whether
69 microsecond on Linux, but with thousands (or tens of thousands) of vmas
/Linux-v5.15/Documentation/admin-guide/mm/
Duserfaultfd.rst35 operations never involve heavyweight structures like vmas (in fact the
40 Terabytes. Too many vmas would be needed for that.
/Linux-v5.15/fs/
Dio_uring.c8937 struct vm_area_struct **vmas = NULL; in io_sqe_buffer_register() local
8960 vmas = kvmalloc_array(nr_pages, sizeof(struct vm_area_struct *), in io_sqe_buffer_register()
8962 if (!vmas) in io_sqe_buffer_register()
8972 pages, vmas); in io_sqe_buffer_register()
8976 struct vm_area_struct *vma = vmas[i]; in io_sqe_buffer_register()
9028 kvfree(vmas); in io_sqe_buffer_register()