Lines Matching refs:vm_area_struct

126 void reset_vma_resv_huge_pages(struct vm_area_struct *vma);
135 int copy_hugetlb_page_range(struct mm_struct *, struct mm_struct *, struct vm_area_struct *);
136 long follow_hugetlb_page(struct mm_struct *, struct vm_area_struct *,
137 struct page **, struct vm_area_struct **,
140 void unmap_hugepage_range(struct vm_area_struct *,
143 struct vm_area_struct *vma,
146 void __unmap_hugepage_range(struct mmu_gather *tlb, struct vm_area_struct *vma,
153 vm_fault_t hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
157 struct vm_area_struct *dst_vma,
164 struct vm_area_struct *vma,
177 pte_t *huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma,
187 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma,
191 int huge_pmd_unshare(struct mm_struct *mm, struct vm_area_struct *vma,
193 void adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma,
197 struct page *follow_huge_pd(struct vm_area_struct *vma,
209 unsigned long hugetlb_change_protection(struct vm_area_struct *vma,
213 void hugetlb_unshare_all_pmds(struct vm_area_struct *vma);
217 static inline void reset_vma_resv_huge_pages(struct vm_area_struct *vma) in reset_vma_resv_huge_pages()
233 struct vm_area_struct *vma, in huge_pmd_unshare()
240 struct vm_area_struct *vma, in adjust_range_if_pmd_sharing_possible()
246 struct vm_area_struct *vma, struct page **pages, in follow_hugetlb_page()
247 struct vm_area_struct **vmas, unsigned long *position, in follow_hugetlb_page()
262 struct mm_struct *src, struct vm_area_struct *vma) in copy_hugetlb_page_range()
281 static inline struct page *follow_huge_pd(struct vm_area_struct *vma, in follow_huge_pd()
338 struct vm_area_struct *dst_vma, in hugetlb_mcopy_atomic_pte()
375 struct vm_area_struct *vma, unsigned long address, in hugetlb_change_protection()
382 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range_final()
389 struct vm_area_struct *vma, unsigned long start, in __unmap_hugepage_range()
396 struct vm_area_struct *vma, unsigned long address, in hugetlb_fault()
403 static inline void hugetlb_unshare_all_pmds(struct vm_area_struct *vma) { } in hugetlb_unshare_all_pmds()
628 struct page *alloc_huge_page(struct vm_area_struct *vma,
632 struct page *alloc_huge_page_vma(struct hstate *h, struct vm_area_struct *vma,
636 void restore_reserve_on_error(struct hstate *h, struct vm_area_struct *vma,
683 static inline struct hstate *hstate_vma(struct vm_area_struct *vma) in hstate_vma()
693 extern unsigned long vma_kernel_pagesize(struct vm_area_struct *vma);
695 extern unsigned long vma_mmu_pagesize(struct vm_area_struct *vma);
886 static inline pte_t huge_ptep_modify_prot_start(struct vm_area_struct *vma, in huge_ptep_modify_prot_start()
895 static inline void huge_ptep_modify_prot_commit(struct vm_area_struct *vma, in huge_ptep_modify_prot_commit()
917 static inline struct page *alloc_huge_page(struct vm_area_struct *vma,
932 struct vm_area_struct *vma,
953 static inline struct hstate *hstate_vma(struct vm_area_struct *vma)
973 static inline unsigned long vma_kernel_pagesize(struct vm_area_struct *vma)
978 static inline unsigned long vma_mmu_pagesize(struct vm_area_struct *vma)
1096 bool want_pmd_share(struct vm_area_struct *vma, unsigned long addr);