Lines Matching refs:resv_map

100 static struct resv_map *vma_resv_map(struct vm_area_struct *vma);
272 struct resv_map *resv_map = vma_resv_map(vma); in hugetlb_vma_lock_read() local
274 down_read(&resv_map->rw_sema); in hugetlb_vma_lock_read()
285 struct resv_map *resv_map = vma_resv_map(vma); in hugetlb_vma_unlock_read() local
287 up_read(&resv_map->rw_sema); in hugetlb_vma_unlock_read()
298 struct resv_map *resv_map = vma_resv_map(vma); in hugetlb_vma_lock_write() local
300 down_write(&resv_map->rw_sema); in hugetlb_vma_lock_write()
311 struct resv_map *resv_map = vma_resv_map(vma); in hugetlb_vma_unlock_write() local
313 up_write(&resv_map->rw_sema); in hugetlb_vma_unlock_write()
325 struct resv_map *resv_map = vma_resv_map(vma); in hugetlb_vma_trylock_write() local
327 return down_write_trylock(&resv_map->rw_sema); in hugetlb_vma_trylock_write()
340 struct resv_map *resv_map = vma_resv_map(vma); in hugetlb_vma_assert_locked() local
342 lockdep_assert_held(&resv_map->rw_sema); in hugetlb_vma_assert_locked()
376 struct resv_map *resv_map = vma_resv_map(vma); in __hugetlb_vma_unlock_write_free() local
379 up_write(&resv_map->rw_sema); in __hugetlb_vma_unlock_write_free()
437 get_file_region_entry_from_cache(struct resv_map *resv, long from, long to) in get_file_region_entry_from_cache()
467 struct resv_map *resv, in record_hugetlb_cgroup_uncharge_info()
519 static void coalesce_file_region(struct resv_map *resv, struct file_region *rg) in coalesce_file_region()
547 hugetlb_resv_map_add(struct resv_map *map, struct list_head *rg, long from, in hugetlb_resv_map_add()
572 static long add_reservation_in_range(struct resv_map *resv, long f, long t, in add_reservation_in_range()
634 static int allocate_file_region_entries(struct resv_map *resv, in allocate_file_region_entries()
705 static long region_add(struct resv_map *resv, long f, long t, in region_add()
772 static long region_chg(struct resv_map *resv, long f, long t, in region_chg()
808 static void region_abort(struct resv_map *resv, long f, long t, in region_abort()
831 static long region_del(struct resv_map *resv, long f, long t) in region_del()
958 static long region_count(struct resv_map *resv, long f, long t) in region_count()
1071 resv_map_set_hugetlb_cgroup_uncharge_info(struct resv_map *resv_map, in resv_map_set_hugetlb_cgroup_uncharge_info() argument
1077 resv_map->reservation_counter = NULL; in resv_map_set_hugetlb_cgroup_uncharge_info()
1078 resv_map->pages_per_hpage = 0; in resv_map_set_hugetlb_cgroup_uncharge_info()
1079 resv_map->css = NULL; in resv_map_set_hugetlb_cgroup_uncharge_info()
1081 resv_map->reservation_counter = in resv_map_set_hugetlb_cgroup_uncharge_info()
1083 resv_map->pages_per_hpage = pages_per_huge_page(h); in resv_map_set_hugetlb_cgroup_uncharge_info()
1084 resv_map->css = &h_cg->css; in resv_map_set_hugetlb_cgroup_uncharge_info()
1089 struct resv_map *resv_map_alloc(void) in resv_map_alloc()
1091 struct resv_map *resv_map = kmalloc(sizeof(*resv_map), GFP_KERNEL); in resv_map_alloc() local
1094 if (!resv_map || !rg) { in resv_map_alloc()
1095 kfree(resv_map); in resv_map_alloc()
1100 kref_init(&resv_map->refs); in resv_map_alloc()
1101 spin_lock_init(&resv_map->lock); in resv_map_alloc()
1102 INIT_LIST_HEAD(&resv_map->regions); in resv_map_alloc()
1103 init_rwsem(&resv_map->rw_sema); in resv_map_alloc()
1105 resv_map->adds_in_progress = 0; in resv_map_alloc()
1112 resv_map_set_hugetlb_cgroup_uncharge_info(resv_map, NULL, NULL); in resv_map_alloc()
1114 INIT_LIST_HEAD(&resv_map->region_cache); in resv_map_alloc()
1115 list_add(&rg->link, &resv_map->region_cache); in resv_map_alloc()
1116 resv_map->region_cache_count = 1; in resv_map_alloc()
1118 return resv_map; in resv_map_alloc()
1123 struct resv_map *resv_map = container_of(ref, struct resv_map, refs); in resv_map_release() local
1124 struct list_head *head = &resv_map->region_cache; in resv_map_release()
1128 region_del(resv_map, 0, LONG_MAX); in resv_map_release()
1136 VM_BUG_ON(resv_map->adds_in_progress); in resv_map_release()
1138 kfree(resv_map); in resv_map_release()
1141 static inline struct resv_map *inode_resv_map(struct inode *inode) in inode_resv_map()
1151 return (struct resv_map *)(&inode->i_data)->private_data; in inode_resv_map()
1154 static struct resv_map *vma_resv_map(struct vm_area_struct *vma) in vma_resv_map()
1164 return (struct resv_map *)(get_vma_private_data(vma) & in vma_resv_map()
1169 static void set_vma_resv_map(struct vm_area_struct *vma, struct resv_map *map) in set_vma_resv_map()
1237 struct resv_map *reservations = vma_resv_map(vma); in clear_vma_resv_huge_pages()
2711 struct resv_map *resv; in __vma_reservation_common()
4833 struct resv_map *resv = vma_resv_map(vma); in hugetlb_vm_op_open()
4872 struct resv_map *resv; in hugetlb_vm_op_close()
6711 struct resv_map *resv_map; in hugetlb_reserve_pages() local
6747 resv_map = inode_resv_map(inode); in hugetlb_reserve_pages()
6749 chg = region_chg(resv_map, from, to, &regions_needed); in hugetlb_reserve_pages()
6752 resv_map = resv_map_alloc(); in hugetlb_reserve_pages()
6753 if (!resv_map) in hugetlb_reserve_pages()
6758 set_vma_resv_map(vma, resv_map); in hugetlb_reserve_pages()
6773 resv_map_set_hugetlb_cgroup_uncharge_info(resv_map, h_cg, h); in hugetlb_reserve_pages()
6804 add = region_add(resv_map, from, to, regions_needed, h, h_cg); in hugetlb_reserve_pages()
6855 region_abort(resv_map, from, to, regions_needed); in hugetlb_reserve_pages()
6857 kref_put(&resv_map->refs, resv_map_release); in hugetlb_reserve_pages()
6867 struct resv_map *resv_map = inode_resv_map(inode); in hugetlb_unreserve_pages() local
6876 if (resv_map) { in hugetlb_unreserve_pages()
6877 chg = region_del(resv_map, start, end); in hugetlb_unreserve_pages()