Home
last modified time | relevance | path

Searched refs:vm_mm (Results 1 – 25 of 156) sorted by relevance

1234567

/Linux-v5.4/arch/mips/mm/
Dtlb-r3k.c74 struct mm_struct *mm = vma->vm_mm; in local_flush_tlb_range()
155 if (cpu_context(cpu, vma->vm_mm) != 0) { in local_flush_tlb_page()
160 printk("[tlbpage<%lu,0x%08lx>]", cpu_context(cpu, vma->vm_mm), page); in local_flush_tlb_page()
162 newpid = cpu_context(cpu, vma->vm_mm) & asid_mask; in local_flush_tlb_page()
191 if (current->active_mm != vma->vm_mm) in __update_tlb()
197 if ((pid != (cpu_context(cpu, vma->vm_mm) & asid_mask)) || (cpu_context(cpu, vma->vm_mm) == 0)) { in __update_tlb()
199 (cpu_context(cpu, vma->vm_mm)), pid); in __update_tlb()
Dtlb-r4k.c110 struct mm_struct *mm = vma->vm_mm; in local_flush_tlb_range()
216 if (cpu_context(cpu, vma->vm_mm) != 0) { in local_flush_tlb_page()
228 write_c0_memorymapid(cpu_asid(cpu, vma->vm_mm)); in local_flush_tlb_page()
230 write_c0_entryhi(page | cpu_asid(cpu, vma->vm_mm)); in local_flush_tlb_page()
306 if (current->active_mm != vma->vm_mm) in __update_tlb()
319 pgdp = pgd_offset(vma->vm_mm, address); in __update_tlb()
/Linux-v5.4/arch/powerpc/mm/book3s64/
Dradix_hugetlbpage.c18 radix__flush_tlb_page_psize(vma->vm_mm, vmaddr, psize); in radix__flush_hugetlb_page()
27 radix__local_flush_tlb_page_psize(vma->vm_mm, vmaddr, psize); in radix__local_flush_hugetlb_page()
37 radix__flush_tlb_range_psize(vma->vm_mm, start, end, psize); in radix__flush_hugetlb_tlb_range()
99 struct mm_struct *mm = vma->vm_mm; in radix__huge_ptep_modify_prot_commit()
109 set_huge_pte_at(vma->vm_mm, addr, ptep, pte); in radix__huge_ptep_modify_prot_commit()
Dpgtable.c41 assert_spin_locked(pmd_lockptr(vma->vm_mm, pmdp)); in pmdp_set_access_flags()
58 return __pmdp_test_and_clear_young(vma->vm_mm, address, pmdp); in pmdp_test_and_clear_young()
110 old_pmd = pmd_hugepage_update(vma->vm_mm, address, pmdp, _PAGE_PRESENT, _PAGE_INVALID); in pmdp_invalidate()
119 serialize_against_pte_lookup(vma->vm_mm); in pmdp_invalidate()
437 pte_val = pte_update(vma->vm_mm, addr, ptep, _PAGE_PRESENT, _PAGE_INVALID, 0); in ptep_modify_prot_start()
449 set_pte_at(vma->vm_mm, addr, ptep, pte); in ptep_modify_prot_commit()
/Linux-v5.4/arch/sh/mm/
Dtlbflush_32.c19 if (vma->vm_mm && cpu_context(cpu, vma->vm_mm) != NO_CONTEXT) { in local_flush_tlb_page()
24 asid = cpu_asid(cpu, vma->vm_mm); in local_flush_tlb_page()
28 if (vma->vm_mm != current->mm) { in local_flush_tlb_page()
42 struct mm_struct *mm = vma->vm_mm; in local_flush_tlb_range()
/Linux-v5.4/mm/
Dhuge_memory.c584 if (mem_cgroup_try_charge_delay(page, vma->vm_mm, gfp, &memcg, true)) { in __do_huge_pmd_anonymous_page()
590 pgtable = pte_alloc_one(vma->vm_mm); in __do_huge_pmd_anonymous_page()
604 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page()
610 ret = check_stable_address_space(vma->vm_mm); in __do_huge_pmd_anonymous_page()
621 pte_free(vma->vm_mm, pgtable); in __do_huge_pmd_anonymous_page()
632 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); in __do_huge_pmd_anonymous_page()
633 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); in __do_huge_pmd_anonymous_page()
634 add_mm_counter(vma->vm_mm, MM_ANONPAGES, HPAGE_PMD_NR); in __do_huge_pmd_anonymous_page()
635 mm_inc_nr_ptes(vma->vm_mm); in __do_huge_pmd_anonymous_page()
646 pte_free(vma->vm_mm, pgtable); in __do_huge_pmd_anonymous_page()
[all …]
Dpgtable-generic.c61 set_pte_at(vma->vm_mm, address, ptep, entry); in ptep_set_access_flags()
84 struct mm_struct *mm = (vma)->vm_mm; in ptep_clear_flush()
103 set_pmd_at(vma->vm_mm, address, pmdp, entry); in pmdp_set_access_flags()
131 pmd = pmdp_huge_get_and_clear(vma->vm_mm, address, pmdp); in pmdp_huge_clear_flush()
144 pud = pudp_huge_get_and_clear(vma->vm_mm, address, pudp); in pudp_huge_clear_flush()
206 pmd = pmdp_huge_get_and_clear(vma->vm_mm, address, pmdp); in pmdp_collapse_flush()
Dmemory.c484 pgd_t *pgd = pgd_offset(vma->vm_mm, addr); in print_bad_pte()
1232 pgd = pgd_offset(vma->vm_mm, addr); in unmap_page_range()
1310 mmu_notifier_range_init(&range, MMU_NOTIFY_UNMAP, 0, vma, vma->vm_mm, in unmap_vmas()
1333 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma, vma->vm_mm, in zap_page_range()
1335 tlb_gather_mmu(&tlb, vma->vm_mm, start, range.end); in zap_page_range()
1336 update_hiwater_rss(vma->vm_mm); in zap_page_range()
1360 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma, vma->vm_mm, in zap_page_range_single()
1362 tlb_gather_mmu(&tlb, vma->vm_mm, address, range.end); in zap_page_range_single()
1363 update_hiwater_rss(vma->vm_mm); in zap_page_range_single()
1425 struct mm_struct *mm = vma->vm_mm; in insert_page()
[all …]
Dmremap.c120 struct mm_struct *mm = vma->vm_mm; in move_ptes()
156 flush_tlb_batched_pending(vma->vm_mm); in move_ptes()
200 struct mm_struct *mm = vma->vm_mm; in move_normal_pmd()
218 old_ptl = pmd_lock(vma->vm_mm, old_pmd); in move_normal_pmd()
252 mmu_notifier_range_init(&range, MMU_NOTIFY_UNMAP, 0, vma, vma->vm_mm, in move_page_tables()
263 old_pmd = get_old_pmd(vma->vm_mm, old_addr); in move_page_tables()
266 new_pmd = alloc_new_pmd(vma->vm_mm, vma, new_addr); in move_page_tables()
304 if (pte_alloc(new_vma->vm_mm, new_pmd)) in move_page_tables()
324 struct mm_struct *mm = vma->vm_mm; in move_vma()
Dmprotect.c61 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in change_pte_range()
65 atomic_read(&vma->vm_mm->mm_users) == 1) in change_pte_range()
68 flush_tlb_batched_pending(vma->vm_mm); in change_pte_range()
138 set_pte_at(vma->vm_mm, addr, pte, newpte); in change_pte_range()
152 set_pte_at(vma->vm_mm, addr, pte, newpte); in change_pte_range()
189 vma, vma->vm_mm, addr, end); in change_pmd_range()
271 struct mm_struct *mm = vma->vm_mm; in change_protection_range()
342 struct mm_struct *mm = vma->vm_mm; in mprotect_fixup()
/Linux-v5.4/arch/arm/mm/
Dfault-armv.c58 set_pte_at(vma->vm_mm, address, ptep, entry); in do_adjust_pte()
99 pgd = pgd_offset(vma->vm_mm, address); in adjust_pte()
116 ptl = pte_lockptr(vma->vm_mm, pmd); in adjust_pte()
132 struct mm_struct *mm = vma->vm_mm; in make_coherent()
152 if (mpnt->vm_mm != mm || mpnt == vma) in make_coherent()
/Linux-v5.4/arch/arc/mm/
Dtlb.c338 local_flush_tlb_mm(vma->vm_mm); in local_flush_tlb_range()
351 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) { in local_flush_tlb_range()
353 tlb_entry_erase(start | hw_pid(vma->vm_mm, cpu)); in local_flush_tlb_range()
408 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) { in local_flush_tlb_page()
409 tlb_entry_erase((page & PAGE_MASK) | hw_pid(vma->vm_mm, cpu)); in local_flush_tlb_page()
472 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_tlb_page, &ta, 1); in flush_tlb_page()
484 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_tlb_range, &ta, 1); in flush_tlb_range()
497 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_pmd_tlb_range, &ta, 1); in flush_pmd_tlb_range()
547 if (current->active_mm != vma->vm_mm) in create_tlb()
552 tlb_paranoid_check(asid_mm(vma->vm_mm, smp_processor_id()), vaddr); in create_tlb()
[all …]
/Linux-v5.4/include/linux/
Dkhugepaged.h57 if (!test_bit(MMF_VM_HUGEPAGE, &vma->vm_mm->flags)) in khugepaged_enter()
61 !test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags)) in khugepaged_enter()
62 if (__khugepaged_enter(vma->vm_mm)) in khugepaged_enter()
Dmmu_notifier.h428 __young |= mmu_notifier_clear_flush_young(___vma->vm_mm, \
441 __young |= mmu_notifier_clear_flush_young(___vma->vm_mm, \
454 __young |= mmu_notifier_clear_young(___vma->vm_mm, ___address, \
465 __young |= mmu_notifier_clear_young(___vma->vm_mm, ___address, \
473 struct mm_struct *___mm = (__vma)->vm_mm; \
486 struct mm_struct *___mm = (__vma)->vm_mm; \
499 struct mm_struct *___mm = (__vma)->vm_mm; \
/Linux-v5.4/arch/arm/kernel/
Dsmp_tlb.c202 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_tlb_page, in flush_tlb_page()
206 broadcast_tlb_mm_a15_erratum(vma->vm_mm); in flush_tlb_page()
228 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_tlb_range, in flush_tlb_range()
232 broadcast_tlb_mm_a15_erratum(vma->vm_mm); in flush_tlb_range()
/Linux-v5.4/arch/powerpc/mm/book3s32/
Dtlb.c134 flush_range(mp->vm_mm, mp->vm_start, mp->vm_end); in flush_tlb_mm()
147 mm = (vmaddr < TASK_SIZE)? vma->vm_mm: &init_mm; in flush_tlb_page()
162 flush_range(vma->vm_mm, start, end); in flush_tlb_range()
/Linux-v5.4/arch/m68k/include/asm/
Dtlbflush.h87 if (vma->vm_mm == current->active_mm) { in flush_tlb_page()
98 if (vma->vm_mm == current->active_mm) in flush_tlb_range()
178 sun3_put_context(vma->vm_mm->context); in flush_tlb_page()
195 struct mm_struct *mm = vma->vm_mm; in flush_tlb_range()
/Linux-v5.4/arch/s390/include/asm/
Dhugetlb.h62 huge_ptep_get_and_clear(vma->vm_mm, address, ptep); in huge_ptep_clear_flush()
71 huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); in huge_ptep_set_access_flags()
72 set_huge_pte_at(vma->vm_mm, addr, ptep, pte); in huge_ptep_set_access_flags()
/Linux-v5.4/arch/riscv/mm/
Dtlbflush.c28 __sbi_tlb_flush_range(mm_cpumask(vma->vm_mm), addr, PAGE_SIZE); in flush_tlb_page()
34 __sbi_tlb_flush_range(mm_cpumask(vma->vm_mm), start, end - start); in flush_tlb_range()
/Linux-v5.4/arch/parisc/include/asm/
Dtlbflush.h20 __flush_tlb_range((vma)->vm_mm->context, start, end)
67 purge_tlb_entries(vma->vm_mm, addr); in flush_tlb_page()
/Linux-v5.4/arch/hexagon/mm/
Dvm_tlb.c28 struct mm_struct *mm = vma->vm_mm; in flush_tlb_range()
68 struct mm_struct *mm = vma->vm_mm; in flush_tlb_page()
/Linux-v5.4/arch/arm64/include/asm/
Dtlbflush.h160 unsigned long addr = __TLBI_VADDR(uaddr, ASID(vma->vm_mm)); in flush_tlb_page_nosync()
184 unsigned long asid = ASID(vma->vm_mm); in __flush_tlb_range()
191 flush_tlb_mm(vma->vm_mm); in __flush_tlb_range()
/Linux-v5.4/arch/nds32/mm/
Dtlb.c27 ncid = (ocid & ~TLB_MISC_mskCID) | vma->vm_mm->context.id; in local_flush_tlb_range()
44 ncid = (ocid & ~TLB_MISC_mskCID) | vma->vm_mm->context.id; in local_flush_tlb_page()
/Linux-v5.4/arch/hexagon/kernel/
Dvdso.c85 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) in arch_vma_name()
/Linux-v5.4/arch/csky/kernel/
Dvdso.c79 if (vma->vm_mm == NULL) in arch_vma_name()
82 if (vma->vm_start == (long)vma->vm_mm->context.vdso) in arch_vma_name()

1234567