Lines Matching refs:vm_area_struct

247 struct vm_area_struct *vm_area_alloc(struct mm_struct *);
248 struct vm_area_struct *vm_area_dup(struct vm_area_struct *);
249 void vm_area_free(struct vm_area_struct *);
251 void __vm_area_free(struct vm_area_struct *vma);
510 struct vm_area_struct *vma; /* Target VMA */
560 void (*open)(struct vm_area_struct * area);
565 void (*close)(struct vm_area_struct * area);
567 int (*may_split)(struct vm_area_struct *area, unsigned long addr);
568 int (*mremap)(struct vm_area_struct *area);
574 int (*mprotect)(struct vm_area_struct *vma, unsigned long start,
580 unsigned long (*pagesize)(struct vm_area_struct * area);
593 int (*access)(struct vm_area_struct *vma, unsigned long addr,
599 const char *(*name)(struct vm_area_struct *vma);
609 int (*set_policy)(struct vm_area_struct *vma, struct mempolicy *new);
621 struct mempolicy *(*get_policy)(struct vm_area_struct *vma,
629 struct page *(*find_special_page)(struct vm_area_struct *vma,
634 static inline void vma_numab_state_init(struct vm_area_struct *vma) in vma_numab_state_init()
638 static inline void vma_numab_state_free(struct vm_area_struct *vma) in vma_numab_state_free()
643 static inline void vma_numab_state_init(struct vm_area_struct *vma) {} in vma_numab_state_init()
644 static inline void vma_numab_state_free(struct vm_area_struct *vma) {} in vma_numab_state_free()
653 static inline bool vma_start_read(struct vm_area_struct *vma) in vma_start_read()
686 static inline void vma_end_read(struct vm_area_struct *vma) in vma_end_read()
694 static bool __is_vma_write_locked(struct vm_area_struct *vma, int *mm_lock_seq) in __is_vma_write_locked()
711 static inline void vma_start_write(struct vm_area_struct *vma) in vma_start_write()
729 static inline void vma_assert_write_locked(struct vm_area_struct *vma) in vma_assert_write_locked()
736 static inline void vma_assert_locked(struct vm_area_struct *vma) in vma_assert_locked()
742 static inline void vma_mark_detached(struct vm_area_struct *vma, bool detached) in vma_mark_detached()
766 struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm,
771 static inline bool vma_start_read(struct vm_area_struct *vma) in vma_start_read()
773 static inline void vma_end_read(struct vm_area_struct *vma) {} in vma_end_read()
774 static inline void vma_start_write(struct vm_area_struct *vma) {} in vma_start_write()
775 static inline void vma_assert_write_locked(struct vm_area_struct *vma) in vma_assert_write_locked()
777 static inline void vma_mark_detached(struct vm_area_struct *vma, in vma_mark_detached()
780 static inline struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm, in lock_vma_under_rcu()
804 static inline void vma_init(struct vm_area_struct *vma, struct mm_struct *mm) in vma_init()
815 static inline void vm_flags_init(struct vm_area_struct *vma, in vm_flags_init()
826 static inline void vm_flags_reset(struct vm_area_struct *vma, in vm_flags_reset()
833 static inline void vm_flags_reset_once(struct vm_area_struct *vma, in vm_flags_reset_once()
840 static inline void vm_flags_set(struct vm_area_struct *vma, in vm_flags_set()
847 static inline void vm_flags_clear(struct vm_area_struct *vma, in vm_flags_clear()
858 static inline void __vm_flags_mod(struct vm_area_struct *vma, in __vm_flags_mod()
868 static inline void vm_flags_mod(struct vm_area_struct *vma, in vm_flags_mod()
875 static inline void vma_set_anonymous(struct vm_area_struct *vma) in vma_set_anonymous()
880 static inline bool vma_is_anonymous(struct vm_area_struct *vma) in vma_is_anonymous()
889 static inline bool vma_is_initial_heap(const struct vm_area_struct *vma) in vma_is_initial_heap()
899 static inline bool vma_is_initial_stack(const struct vm_area_struct *vma) in vma_is_initial_stack()
910 static inline bool vma_is_temporary_stack(struct vm_area_struct *vma) in vma_is_temporary_stack()
924 static inline bool vma_is_foreign(struct vm_area_struct *vma) in vma_is_foreign()
935 static inline bool vma_is_accessible(struct vm_area_struct *vma) in vma_is_accessible()
941 struct vm_area_struct *vma_find(struct vma_iterator *vmi, unsigned long max) in vma_find()
946 static inline struct vm_area_struct *vma_next(struct vma_iterator *vmi) in vma_next()
956 struct vm_area_struct *vma_iter_next_range(struct vma_iterator *vmi) in vma_iter_next_range()
962 static inline struct vm_area_struct *vma_prev(struct vma_iterator *vmi) in vma_prev()
968 struct vm_area_struct *vma_iter_prev_range(struct vma_iterator *vmi) in vma_iter_prev_range()
995 struct vm_area_struct *vma) in vma_iter_bulk_store()
1028 bool vma_is_shmem(struct vm_area_struct *vma);
1029 bool vma_is_anon_shmem(struct vm_area_struct *vma);
1031 static inline bool vma_is_shmem(struct vm_area_struct *vma) { return false; } in vma_is_shmem()
1032 static inline bool vma_is_anon_shmem(struct vm_area_struct *vma) { return false; } in vma_is_anon_shmem()
1035 int vma_is_stack_for_current(struct vm_area_struct *vma);
1328 static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) in maybe_mkwrite()
1724 static inline void vma_set_access_pid_bit(struct vm_area_struct *vma) in vma_set_access_pid_bit()
1783 static inline void vma_set_access_pid_bit(struct vm_area_struct *vma) in vma_set_access_pid_bit()
1947 static inline bool page_needs_cow_for_dma(struct vm_area_struct *vma, in page_needs_cow_for_dma()
2326 struct folio *vm_normal_folio(struct vm_area_struct *vma, unsigned long addr,
2328 struct page *vm_normal_page(struct vm_area_struct *vma, unsigned long addr,
2330 struct page *vm_normal_page_pmd(struct vm_area_struct *vma, unsigned long addr,
2333 void zap_vma_ptes(struct vm_area_struct *vma, unsigned long address,
2335 void zap_page_range_single(struct vm_area_struct *vma, unsigned long address,
2337 static inline void zap_vma_pages(struct vm_area_struct *vma) in zap_vma_pages()
2343 struct vm_area_struct *start_vma, unsigned long start,
2351 copy_page_range(struct vm_area_struct *dst_vma, struct vm_area_struct *src_vma);
2354 int follow_pfn(struct vm_area_struct *vma, unsigned long address,
2356 int follow_phys(struct vm_area_struct *vma, unsigned long address,
2358 int generic_access_phys(struct vm_area_struct *vma, unsigned long addr,
2367 struct vm_area_struct *lock_mm_and_find_vma(struct mm_struct *mm,
2371 extern vm_fault_t handle_mm_fault(struct vm_area_struct *vma,
2382 static inline vm_fault_t handle_mm_fault(struct vm_area_struct *vma, in handle_mm_fault()
2409 static inline struct vm_area_struct *vma_lookup(struct mm_struct *mm,
2431 struct vm_area_struct **vmap) in get_user_page_vma_remote()
2434 struct vm_area_struct *vma; in get_user_page_vma_remote()
2480 extern unsigned long move_page_tables(struct vm_area_struct *vma,
2481 unsigned long old_addr, struct vm_area_struct *new_vma,
2505 bool vma_needs_dirty_tracking(struct vm_area_struct *vma);
2506 int vma_wants_writenotify(struct vm_area_struct *vma, pgprot_t vm_page_prot);
2507 static inline bool vma_wants_manual_pte_write_upgrade(struct vm_area_struct *vma) in vma_wants_manual_pte_write_upgrade()
2520 bool can_change_pte_writable(struct vm_area_struct *vma, unsigned long addr,
2523 struct vm_area_struct *vma, unsigned long start,
2526 struct vm_area_struct *vma, struct vm_area_struct **pprev,
3184 void vma_interval_tree_insert(struct vm_area_struct *node,
3186 void vma_interval_tree_insert_after(struct vm_area_struct *node,
3187 struct vm_area_struct *prev,
3189 void vma_interval_tree_remove(struct vm_area_struct *node,
3191 struct vm_area_struct *vma_interval_tree_iter_first(struct rb_root_cached *root,
3193 struct vm_area_struct *vma_interval_tree_iter_next(struct vm_area_struct *node,
3219 extern int vma_expand(struct vma_iterator *vmi, struct vm_area_struct *vma,
3221 struct vm_area_struct *next);
3222 extern int vma_shrink(struct vma_iterator *vmi, struct vm_area_struct *vma,
3224 extern struct vm_area_struct *vma_merge(struct vma_iterator *vmi,
3225 struct mm_struct *, struct vm_area_struct *prev, unsigned long addr,
3229 extern struct anon_vma *find_mergeable_anon_vma(struct vm_area_struct *);
3230 extern int __split_vma(struct vma_iterator *vmi, struct vm_area_struct *,
3232 extern int split_vma(struct vma_iterator *vmi, struct vm_area_struct *,
3234 extern int insert_vm_struct(struct mm_struct *, struct vm_area_struct *);
3235 extern void unlink_file_vma(struct vm_area_struct *);
3236 extern struct vm_area_struct *copy_vma(struct vm_area_struct **,
3266 extern bool vma_is_special_mapping(const struct vm_area_struct *vma,
3268 extern struct vm_area_struct *_install_special_mapping(struct mm_struct *mm,
3297 extern int do_vma_munmap(struct vma_iterator *vmi, struct vm_area_struct *vma,
3345 int expand_stack_locked(struct vm_area_struct *vma, unsigned long address);
3346 struct vm_area_struct *expand_stack(struct mm_struct * mm, unsigned long addr);
3349 int expand_downwards(struct vm_area_struct *vma, unsigned long address);
3352 extern struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr);
3353 extern struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr,
3354 struct vm_area_struct **pprev);
3360 struct vm_area_struct *find_vma_intersection(struct mm_struct *mm,
3371 struct vm_area_struct *vma_lookup(struct mm_struct *mm, unsigned long addr) in vma_lookup()
3376 static inline unsigned long stack_guard_start_gap(struct vm_area_struct *vma) in stack_guard_start_gap()
3388 static inline unsigned long vm_start_gap(struct vm_area_struct *vma) in vm_start_gap()
3399 static inline unsigned long vm_end_gap(struct vm_area_struct *vma) in vm_end_gap()
3411 static inline unsigned long vma_pages(struct vm_area_struct *vma) in vma_pages()
3417 static inline struct vm_area_struct *find_exact_vma(struct mm_struct *mm, in find_exact_vma()
3420 struct vm_area_struct *vma = vma_lookup(mm, vm_start); in find_exact_vma()
3428 static inline bool range_in_vma(struct vm_area_struct *vma, in range_in_vma()
3436 void vma_set_page_prot(struct vm_area_struct *vma);
3442 static inline void vma_set_page_prot(struct vm_area_struct *vma) in vma_set_page_prot()
3448 void vma_set_file(struct vm_area_struct *vma, struct file *file);
3451 unsigned long change_prot_numa(struct vm_area_struct *vma,
3455 struct vm_area_struct *find_extend_vma_locked(struct mm_struct *,
3457 int remap_pfn_range(struct vm_area_struct *, unsigned long addr,
3459 int remap_pfn_range_notrack(struct vm_area_struct *vma, unsigned long addr,
3461 int vm_insert_page(struct vm_area_struct *, unsigned long addr, struct page *);
3462 int vm_insert_pages(struct vm_area_struct *vma, unsigned long addr,
3464 int vm_map_pages(struct vm_area_struct *vma, struct page **pages,
3466 int vm_map_pages_zero(struct vm_area_struct *vma, struct page **pages,
3468 vm_fault_t vmf_insert_pfn(struct vm_area_struct *vma, unsigned long addr,
3470 vm_fault_t vmf_insert_pfn_prot(struct vm_area_struct *vma, unsigned long addr,
3472 vm_fault_t vmf_insert_mixed(struct vm_area_struct *vma, unsigned long addr,
3474 vm_fault_t vmf_insert_mixed_mkwrite(struct vm_area_struct *vma,
3476 int vm_iomap_memory(struct vm_area_struct *vma, phys_addr_t start, unsigned long len);
3478 static inline vm_fault_t vmf_insert_page(struct vm_area_struct *vma, in vmf_insert_page()
3492 static inline int io_remap_pfn_range(struct vm_area_struct *vma, in io_remap_pfn_range()
3527 struct page *follow_page(struct vm_area_struct *vma, unsigned long address,
3545 static inline bool gup_can_follow_protnone(struct vm_area_struct *vma, in gup_can_follow_protnone()
3716 extern struct vm_area_struct *get_gate_vma(struct mm_struct *mm);
3720 static inline struct vm_area_struct *get_gate_vma(struct mm_struct *mm) in get_gate_vma()
3747 const char * arch_vma_name(struct vm_area_struct *vma);
3876 struct vm_area_struct *vma, struct list_head *to_kill,
3946 struct vm_area_struct *vma);
3961 static inline bool vma_is_special_huge(const struct vm_area_struct *vma) in vma_is_special_huge()
4010 static inline int seal_check_future_write(int seals, struct vm_area_struct *vma) in seal_check_future_write()