Lines Matching refs:vm_mm

77 	if (!vma->vm_mm)		/* vdso */  in hugepage_vma_check()
86 test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags)) in hugepage_vma_check()
123 !enforce_sysfs, vma->vm_mm, vm_flags); in hugepage_vma_check()
657 if (mem_cgroup_charge(folio, vma->vm_mm, gfp)) { in __do_huge_pmd_anonymous_page()
665 pgtable = pte_alloc_one(vma->vm_mm); in __do_huge_pmd_anonymous_page()
679 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page()
685 ret = check_stable_address_space(vma->vm_mm); in __do_huge_pmd_anonymous_page()
693 pte_free(vma->vm_mm, pgtable); in __do_huge_pmd_anonymous_page()
703 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); in __do_huge_pmd_anonymous_page()
704 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); in __do_huge_pmd_anonymous_page()
706 add_mm_counter(vma->vm_mm, MM_ANONPAGES, HPAGE_PMD_NR); in __do_huge_pmd_anonymous_page()
707 mm_inc_nr_ptes(vma->vm_mm); in __do_huge_pmd_anonymous_page()
710 count_memcg_event_mm(vma->vm_mm, THP_FAULT_ALLOC); in __do_huge_pmd_anonymous_page()
718 pte_free(vma->vm_mm, pgtable); in __do_huge_pmd_anonymous_page()
788 !mm_forbids_zeropage(vma->vm_mm) && in do_huge_pmd_anonymous_page()
793 pgtable = pte_alloc_one(vma->vm_mm); in do_huge_pmd_anonymous_page()
796 zero_page = mm_get_huge_zero_page(vma->vm_mm); in do_huge_pmd_anonymous_page()
798 pte_free(vma->vm_mm, pgtable); in do_huge_pmd_anonymous_page()
802 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_anonymous_page()
805 ret = check_stable_address_space(vma->vm_mm); in do_huge_pmd_anonymous_page()
808 pte_free(vma->vm_mm, pgtable); in do_huge_pmd_anonymous_page()
811 pte_free(vma->vm_mm, pgtable); in do_huge_pmd_anonymous_page()
815 set_huge_zero_page(pgtable, vma->vm_mm, vma, in do_huge_pmd_anonymous_page()
822 pte_free(vma->vm_mm, pgtable); in do_huge_pmd_anonymous_page()
839 struct mm_struct *mm = vma->vm_mm; in insert_pfn_pmd()
914 pgtable = pte_alloc_one(vma->vm_mm); in vmf_insert_pfn_pmd()
937 struct mm_struct *mm = vma->vm_mm; in insert_pfn_pud()
1026 struct mm_struct *mm = vma->vm_mm; in follow_devmap_pmd()
1181 struct mm_struct *mm = vma->vm_mm; in follow_devmap_pud()
1265 vmf->ptl = pud_lock(vmf->vma->vm_mm, vmf->pud); in huge_pud_set_accessed()
1279 vmf->ptl = pmd_lock(vmf->vma->vm_mm, vmf->pmd); in huge_pmd_set_accessed()
1298 vmf->ptl = pmd_lockptr(vma->vm_mm, vmf->pmd); in do_huge_pmd_wp_page()
1448 struct mm_struct *mm = vma->vm_mm; in follow_trans_huge_pmd()
1500 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_numa_page()
1549 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_numa_page()
1570 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, pmd); in do_huge_pmd_numa_page()
1753 struct mm_struct *mm = vma->vm_mm; in move_huge_pmd()
1808 struct mm_struct *mm = vma->vm_mm; in change_huge_pmd()
1944 ptl = pmd_lock(vma->vm_mm, pmd); in __pmd_trans_huge_lock()
1962 ptl = pud_lock(vma->vm_mm, pud); in __pud_trans_huge_lock()
2010 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma->vm_mm, in __split_huge_pud()
2014 ptl = pud_lock(vma->vm_mm, pud); in __split_huge_pud()
2028 struct mm_struct *mm = vma->vm_mm; in __split_huge_zero_page_pmd()
2069 struct mm_struct *mm = vma->vm_mm; in __split_huge_pmd_locked()
2272 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma->vm_mm, in __split_huge_pmd()
2276 ptl = pmd_lock(vma->vm_mm, pmd); in __split_huge_pmd()
2304 pmd_t *pmd = mm_find_pmd(vma->vm_mm, address); in split_huge_pmd_address()
2340 struct vm_area_struct *next = find_vma(vma->vm_mm, vma->vm_end); in vma_adjust_trans_huge()
3206 struct mm_struct *mm = vma->vm_mm; in set_pmd_migration_entry()
3254 struct mm_struct *mm = vma->vm_mm; in remove_migration_pmd()