Lines Matching refs:dst_pmd

70 int mfill_atomic_install_pte(pmd_t *dst_pmd,  in mfill_atomic_install_pte()  argument
94 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_install_pte()
140 static int mfill_atomic_pte_copy(pmd_t *dst_pmd, in mfill_atomic_pte_copy() argument
205 ret = mfill_atomic_install_pte(dst_pmd, dst_vma, dst_addr, in mfill_atomic_pte_copy()
216 static int mfill_atomic_pte_zeropage(pmd_t *dst_pmd, in mfill_atomic_pte_zeropage() argument
227 dst_pte = pte_offset_map_lock(dst_vma->vm_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_pte_zeropage()
248 static int mfill_atomic_pte_continue(pmd_t *dst_pmd, in mfill_atomic_pte_continue() argument
276 ret = mfill_atomic_install_pte(dst_pmd, dst_vma, dst_addr, in mfill_atomic_pte_continue()
292 static int mfill_atomic_pte_poison(pmd_t *dst_pmd, in mfill_atomic_pte_poison() argument
304 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_pte_poison()
512 static __always_inline ssize_t mfill_atomic_pte(pmd_t *dst_pmd, in mfill_atomic_pte() argument
522 return mfill_atomic_pte_continue(dst_pmd, dst_vma, in mfill_atomic_pte()
525 return mfill_atomic_pte_poison(dst_pmd, dst_vma, in mfill_atomic_pte()
541 err = mfill_atomic_pte_copy(dst_pmd, dst_vma, in mfill_atomic_pte()
545 err = mfill_atomic_pte_zeropage(dst_pmd, in mfill_atomic_pte()
548 err = shmem_mfill_atomic_pte(dst_pmd, dst_vma, in mfill_atomic_pte()
565 pmd_t *dst_pmd; in mfill_atomic() local
649 dst_pmd = mm_alloc_pmd(dst_mm, dst_addr); in mfill_atomic()
650 if (unlikely(!dst_pmd)) { in mfill_atomic()
655 dst_pmdval = pmdp_get_lockless(dst_pmd); in mfill_atomic()
665 unlikely(__pte_alloc(dst_mm, dst_pmd))) { in mfill_atomic()
670 if (unlikely(pmd_trans_huge(*dst_pmd))) { in mfill_atomic()
675 BUG_ON(pmd_none(*dst_pmd)); in mfill_atomic()
676 BUG_ON(pmd_trans_huge(*dst_pmd)); in mfill_atomic()
678 err = mfill_atomic_pte(dst_pmd, dst_vma, dst_addr, in mfill_atomic()