/Linux-v4.19/arch/mips/mm/ |
D | tlb-r3k.c | 86 struct mm_struct *mm = vma->vm_mm; in local_flush_tlb_range() 167 if (cpu_context(cpu, vma->vm_mm) != 0) { in local_flush_tlb_page() 172 printk("[tlbpage<%lu,0x%08lx>]", cpu_context(cpu, vma->vm_mm), page); in local_flush_tlb_page() 174 newpid = cpu_context(cpu, vma->vm_mm) & asid_mask; in local_flush_tlb_page() 203 if (current->active_mm != vma->vm_mm) in __update_tlb() 209 if ((pid != (cpu_context(cpu, vma->vm_mm) & asid_mask)) || (cpu_context(cpu, vma->vm_mm) == 0)) { in __update_tlb() 211 (cpu_context(cpu, vma->vm_mm)), pid); in __update_tlb()
|
D | tlb-r8k.c | 64 struct mm_struct *mm = vma->vm_mm; in local_flush_tlb_range() 155 if (!cpu_context(cpu, vma->vm_mm)) in local_flush_tlb_page() 158 newpid = cpu_asid(cpu, vma->vm_mm); in local_flush_tlb_page() 194 if (current->active_mm != vma->vm_mm) in __update_tlb() 203 pgdp = pgd_offset(vma->vm_mm, address); in __update_tlb()
|
/Linux-v4.19/arch/sh/mm/ |
D | tlbflush_32.c | 19 if (vma->vm_mm && cpu_context(cpu, vma->vm_mm) != NO_CONTEXT) { in local_flush_tlb_page() 24 asid = cpu_asid(cpu, vma->vm_mm); in local_flush_tlb_page() 28 if (vma->vm_mm != current->mm) { in local_flush_tlb_page() 42 struct mm_struct *mm = vma->vm_mm; in local_flush_tlb_range()
|
D | cache-sh4.c | 223 if (cpu_context(smp_processor_id(), vma->vm_mm) == NO_CONTEXT) in sh4_flush_cache_page() 226 pgd = pgd_offset(vma->vm_mm, address); in sh4_flush_cache_page() 235 if ((vma->vm_mm == current->active_mm)) in sh4_flush_cache_page() 286 if (cpu_context(smp_processor_id(), vma->vm_mm) == NO_CONTEXT) in sh4_flush_cache_range()
|
/Linux-v4.19/mm/ |
D | huge_memory.c | 555 if (mem_cgroup_try_charge_delay(page, vma->vm_mm, gfp, &memcg, true)) { in __do_huge_pmd_anonymous_page() 561 pgtable = pte_alloc_one(vma->vm_mm, haddr); in __do_huge_pmd_anonymous_page() 575 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page() 581 ret = check_stable_address_space(vma->vm_mm); in __do_huge_pmd_anonymous_page() 592 pte_free(vma->vm_mm, pgtable); in __do_huge_pmd_anonymous_page() 603 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); in __do_huge_pmd_anonymous_page() 604 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); in __do_huge_pmd_anonymous_page() 605 add_mm_counter(vma->vm_mm, MM_ANONPAGES, HPAGE_PMD_NR); in __do_huge_pmd_anonymous_page() 606 mm_inc_nr_ptes(vma->vm_mm); in __do_huge_pmd_anonymous_page() 616 pte_free(vma->vm_mm, pgtable); in __do_huge_pmd_anonymous_page() [all …]
|
D | pgtable-generic.c | 60 set_pte_at(vma->vm_mm, address, ptep, entry); in ptep_set_access_flags() 83 struct mm_struct *mm = (vma)->vm_mm; in ptep_clear_flush() 102 set_pmd_at(vma->vm_mm, address, pmdp, entry); in pmdp_set_access_flags() 130 pmd = pmdp_huge_get_and_clear(vma->vm_mm, address, pmdp); in pmdp_huge_clear_flush() 143 pud = pudp_huge_get_and_clear(vma->vm_mm, address, pudp); in pudp_huge_clear_flush() 205 pmd = pmdp_huge_get_and_clear(vma->vm_mm, address, pmdp); in pmdp_collapse_flush()
|
D | memory.c | 729 pgd_t *pgd = pgd_offset(vma->vm_mm, addr); in print_bad_pte() 1506 pgd = pgd_offset(vma->vm_mm, addr); in unmap_page_range() 1582 struct mm_struct *mm = vma->vm_mm; in unmap_vmas() 1601 struct mm_struct *mm = vma->vm_mm; in zap_page_range() 1627 struct mm_struct *mm = vma->vm_mm; in zap_page_range_single() 1695 struct mm_struct *mm = vma->vm_mm; in insert_page() 1762 BUG_ON(down_read_trylock(&vma->vm_mm->mmap_sem)); in vm_insert_page() 1773 struct mm_struct *mm = vma->vm_mm; in insert_pfn() 2085 struct mm_struct *mm = vma->vm_mm; in remap_pfn_range() 2489 struct mm_struct *mm = vma->vm_mm; in wp_page_copy() [all …]
|
D | mremap.c | 120 struct mm_struct *mm = vma->vm_mm; in move_ptes() 156 flush_tlb_batched_pending(vma->vm_mm); in move_ptes() 209 mmu_notifier_invalidate_range_start(vma->vm_mm, mmun_start, mmun_end); in move_page_tables() 218 old_pmd = get_old_pmd(vma->vm_mm, old_addr); in move_page_tables() 221 new_pmd = alloc_new_pmd(vma->vm_mm, vma, new_addr); in move_page_tables() 241 if (pte_alloc(new_vma->vm_mm, new_pmd, new_addr)) in move_page_tables() 250 mmu_notifier_invalidate_range_end(vma->vm_mm, mmun_start, mmun_end); in move_page_tables() 261 struct mm_struct *mm = vma->vm_mm; in move_vma()
|
D | mprotect.c | 42 struct mm_struct *mm = vma->vm_mm; in change_pte_range() 62 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in change_pte_range() 66 atomic_read(&vma->vm_mm->mm_users) == 1) in change_pte_range() 69 flush_tlb_batched_pending(vma->vm_mm); in change_pte_range() 170 struct mm_struct *mm = vma->vm_mm; in change_pmd_range() 269 struct mm_struct *mm = vma->vm_mm; in change_protection_range() 349 struct mm_struct *mm = vma->vm_mm; in mprotect_fixup()
|
/Linux-v4.19/arch/arm/mm/ |
D | fault-armv.c | 61 set_pte_at(vma->vm_mm, address, ptep, entry); in do_adjust_pte() 102 pgd = pgd_offset(vma->vm_mm, address); in adjust_pte() 119 ptl = pte_lockptr(vma->vm_mm, pmd); in adjust_pte() 135 struct mm_struct *mm = vma->vm_mm; in make_coherent() 155 if (mpnt->vm_mm != mm || mpnt == vma) in make_coherent()
|
/Linux-v4.19/include/linux/ |
D | khugepaged.h | 49 if (!test_bit(MMF_VM_HUGEPAGE, &vma->vm_mm->flags)) in khugepaged_enter() 53 !test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags)) in khugepaged_enter() 54 if (__khugepaged_enter(vma->vm_mm)) in khugepaged_enter()
|
D | mmu_notifier.h | 336 __young |= mmu_notifier_clear_flush_young(___vma->vm_mm, \ 349 __young |= mmu_notifier_clear_flush_young(___vma->vm_mm, \ 362 __young |= mmu_notifier_clear_young(___vma->vm_mm, ___address, \ 373 __young |= mmu_notifier_clear_young(___vma->vm_mm, ___address, \ 381 struct mm_struct *___mm = (__vma)->vm_mm; \ 394 struct mm_struct *___mm = (__vma)->vm_mm; \ 407 struct mm_struct *___mm = (__vma)->vm_mm; \
|
/Linux-v4.19/arch/arc/mm/ |
D | tlb.c | 341 local_flush_tlb_mm(vma->vm_mm); in local_flush_tlb_range() 354 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) { in local_flush_tlb_range() 356 tlb_entry_erase(start | hw_pid(vma->vm_mm, cpu)); in local_flush_tlb_range() 411 if (asid_mm(vma->vm_mm, cpu) != MM_CTXT_NO_ASID) { in local_flush_tlb_page() 412 tlb_entry_erase((page & PAGE_MASK) | hw_pid(vma->vm_mm, cpu)); in local_flush_tlb_page() 475 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_tlb_page, &ta, 1); in flush_tlb_page() 487 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_tlb_range, &ta, 1); in flush_tlb_range() 500 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_pmd_tlb_range, &ta, 1); in flush_pmd_tlb_range() 550 if (current->active_mm != vma->vm_mm) in create_tlb() 555 tlb_paranoid_check(asid_mm(vma->vm_mm, smp_processor_id()), vaddr); in create_tlb() [all …]
|
/Linux-v4.19/arch/arm/kernel/ |
D | smp_tlb.c | 205 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_tlb_page, in flush_tlb_page() 209 broadcast_tlb_mm_a15_erratum(vma->vm_mm); in flush_tlb_page() 231 on_each_cpu_mask(mm_cpumask(vma->vm_mm), ipi_flush_tlb_range, in flush_tlb_range() 235 broadcast_tlb_mm_a15_erratum(vma->vm_mm); in flush_tlb_range()
|
/Linux-v4.19/arch/powerpc/mm/ |
D | hugetlbpage-radix.c | 17 radix__flush_tlb_page_psize(vma->vm_mm, vmaddr, psize); in radix__flush_hugetlb_page() 26 radix__local_flush_tlb_page_psize(vma->vm_mm, vmaddr, psize); in radix__local_flush_hugetlb_page() 36 radix__flush_tlb_range_psize(vma->vm_mm, start, end, psize); in radix__flush_hugetlb_tlb_range()
|
D | tlb_hash32.c | 139 flush_range(mp->vm_mm, mp->vm_start, mp->vm_end); in flush_tlb_mm() 152 mm = (vmaddr < TASK_SIZE)? vma->vm_mm: &init_mm; in flush_tlb_page() 167 flush_range(vma->vm_mm, start, end); in flush_tlb_range()
|
/Linux-v4.19/arch/m68k/include/asm/ |
D | tlbflush.h | 87 if (vma->vm_mm == current->active_mm) { in flush_tlb_page() 98 if (vma->vm_mm == current->active_mm) in flush_tlb_range() 178 sun3_put_context(vma->vm_mm->context); in flush_tlb_page() 195 struct mm_struct *mm = vma->vm_mm; in flush_tlb_range()
|
/Linux-v4.19/arch/s390/include/asm/ |
D | hugetlb.h | 57 huge_ptep_get_and_clear(vma->vm_mm, address, ptep); in huge_ptep_clear_flush() 66 huge_ptep_get_and_clear(vma->vm_mm, addr, ptep); in huge_ptep_set_access_flags() 67 set_huge_pte_at(vma->vm_mm, addr, ptep, pte); in huge_ptep_set_access_flags()
|
/Linux-v4.19/arch/hexagon/mm/ |
D | vm_tlb.c | 41 struct mm_struct *mm = vma->vm_mm; in flush_tlb_range() 81 struct mm_struct *mm = vma->vm_mm; in flush_tlb_page()
|
/Linux-v4.19/arch/arm64/include/asm/ |
D | tlbflush.h | 140 unsigned long addr = __TLBI_VADDR(uaddr, ASID(vma->vm_mm)); in flush_tlb_page() 158 unsigned long asid = ASID(vma->vm_mm); in __flush_tlb_range() 162 flush_tlb_mm(vma->vm_mm); in __flush_tlb_range()
|
/Linux-v4.19/arch/nds32/mm/ |
D | tlb.c | 27 ncid = (ocid & ~TLB_MISC_mskCID) | vma->vm_mm->context.id; in local_flush_tlb_range() 44 ncid = (ocid & ~TLB_MISC_mskCID) | vma->vm_mm->context.id; in local_flush_tlb_page()
|
/Linux-v4.19/arch/ia64/include/asm/ |
D | tlbflush.h | 81 if (vma->vm_mm == current->active_mm) in flush_tlb_page() 84 vma->vm_mm->context = 0; in flush_tlb_page()
|
/Linux-v4.19/arch/parisc/include/asm/ |
D | tlbflush.h | 35 __flush_tlb_range((vma)->vm_mm->context, start, end) 84 sid = vma->vm_mm->context; in flush_tlb_page()
|
/Linux-v4.19/arch/hexagon/kernel/ |
D | vdso.c | 98 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) in arch_vma_name()
|
/Linux-v4.19/arch/sh/kernel/vsyscall/ |
D | vsyscall.c | 92 if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso) in arch_vma_name()
|