Lines Matching refs:haddr

652 	unsigned long haddr = vmf->address & HPAGE_PMD_MASK;  in __do_huge_pmd_anonymous_page()  local
701 folio_add_new_anon_rmap(folio, vma, haddr); in __do_huge_pmd_anonymous_page()
704 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); in __do_huge_pmd_anonymous_page()
761 struct vm_area_struct *vma, unsigned long haddr, pmd_t *pmd, in set_huge_zero_page() argument
770 set_pmd_at(mm, haddr, pmd, entry); in set_huge_zero_page()
779 unsigned long haddr = vmf->address & HPAGE_PMD_MASK; in do_huge_pmd_anonymous_page() local
781 if (!transhuge_vma_suitable(vma, haddr)) in do_huge_pmd_anonymous_page()
816 haddr, vmf->pmd, zero_page); in do_huge_pmd_anonymous_page()
827 folio = vma_alloc_folio(gfp, HPAGE_PMD_ORDER, vma, haddr, true); in do_huge_pmd_anonymous_page()
1295 unsigned long haddr = vmf->address & HPAGE_PMD_MASK; in do_huge_pmd_wp_page() local
1361 if (pmdp_set_access_flags(vma, haddr, vmf->pmd, entry, 1)) in do_huge_pmd_wp_page()
1494 unsigned long haddr = vmf->address & HPAGE_PMD_MASK; in do_huge_pmd_numa_page() local
1517 page = vm_normal_page_pmd(vma, haddr, pmd); in do_huge_pmd_numa_page()
1532 target_nid = numa_migrate_prep(page, vma, haddr, page_nid, in do_huge_pmd_numa_page()
1570 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, pmd); in do_huge_pmd_numa_page()
1992 unsigned long haddr) in __split_huge_pud_locked() argument
1994 VM_BUG_ON(haddr & ~HPAGE_PUD_MASK); in __split_huge_pud_locked()
1995 VM_BUG_ON_VMA(vma->vm_start > haddr, vma); in __split_huge_pud_locked()
1996 VM_BUG_ON_VMA(vma->vm_end < haddr + HPAGE_PUD_SIZE, vma); in __split_huge_pud_locked()
2001 pudp_huge_clear_flush(vma, haddr, pud); in __split_huge_pud_locked()
2026 unsigned long haddr, pmd_t *pmd) in __split_huge_zero_page_pmd() argument
2043 old_pmd = pmdp_huge_clear_flush(vma, haddr, pmd); in __split_huge_zero_page_pmd()
2048 pte = pte_offset_map(&_pmd, haddr); in __split_huge_zero_page_pmd()
2050 for (i = 0, addr = haddr; i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE) { in __split_huge_zero_page_pmd()
2067 unsigned long haddr, bool freeze) in __split_huge_pmd_locked() argument
2079 VM_BUG_ON(haddr & ~HPAGE_PMD_MASK); in __split_huge_pmd_locked()
2080 VM_BUG_ON_VMA(vma->vm_start > haddr, vma); in __split_huge_pmd_locked()
2081 VM_BUG_ON_VMA(vma->vm_end < haddr + HPAGE_PMD_SIZE, vma); in __split_huge_pmd_locked()
2088 old_pmd = pmdp_huge_clear_flush(vma, haddr, pmd); in __split_huge_pmd_locked()
2125 return __split_huge_zero_page_pmd(vma, haddr, pmd); in __split_huge_pmd_locked()
2148 old_pmd = pmdp_invalidate(vma, haddr, pmd); in __split_huge_pmd_locked()
2205 pte = pte_offset_map(&_pmd, haddr); in __split_huge_pmd_locked()
2207 for (i = 0, addr = haddr; i < HPAGE_PMD_NR; i++, addr += PAGE_SIZE) { in __split_huge_pmd_locked()
3256 unsigned long haddr = address & HPAGE_PMD_MASK; in remove_migration_pmd() local
3284 page_add_anon_rmap(new, vma, haddr, rmap_flags); in remove_migration_pmd()
3289 set_pmd_at(mm, haddr, pvmw->pmd, pmde); in remove_migration_pmd()