| /Linux-v5.4/arch/powerpc/include/asm/ |
| D | mman.h | 31 static inline pgprot_t arch_vm_get_page_prot(unsigned long vm_flags) in arch_vm_get_page_prot() argument 34 return (vm_flags & VM_SAO) ? in arch_vm_get_page_prot() 35 __pgprot(_PAGE_SAO | vmflag_to_pte_pkey_bits(vm_flags)) : in arch_vm_get_page_prot() 36 __pgprot(0 | vmflag_to_pte_pkey_bits(vm_flags)); in arch_vm_get_page_prot() 38 return (vm_flags & VM_SAO) ? __pgprot(_PAGE_SAO) : __pgprot(0); in arch_vm_get_page_prot() 41 #define arch_vm_get_page_prot(vm_flags) arch_vm_get_page_prot(vm_flags) argument
|
| D | pkeys.h | 33 static inline u64 vmflag_to_pte_pkey_bits(u64 vm_flags) in vmflag_to_pte_pkey_bits() argument 38 return (((vm_flags & VM_PKEY_BIT0) ? H_PTE_PKEY_BIT4 : 0x0UL) | in vmflag_to_pte_pkey_bits() 39 ((vm_flags & VM_PKEY_BIT1) ? H_PTE_PKEY_BIT3 : 0x0UL) | in vmflag_to_pte_pkey_bits() 40 ((vm_flags & VM_PKEY_BIT2) ? H_PTE_PKEY_BIT2 : 0x0UL) | in vmflag_to_pte_pkey_bits() 41 ((vm_flags & VM_PKEY_BIT3) ? H_PTE_PKEY_BIT1 : 0x0UL) | in vmflag_to_pte_pkey_bits() 42 ((vm_flags & VM_PKEY_BIT4) ? H_PTE_PKEY_BIT0 : 0x0UL)); in vmflag_to_pte_pkey_bits() 49 return (vma->vm_flags & ARCH_VM_PKEY_FLAGS) >> VM_PKEY_SHIFT; in vma_pkey()
|
| /Linux-v5.4/include/trace/events/ |
| D | fs_dax.h | 18 __field(unsigned long, vm_flags) 31 __entry->vm_flags = vmf->vma->vm_flags; 43 __entry->vm_flags & VM_SHARED ? "shared" : "private", 70 __field(unsigned long, vm_flags) 79 __entry->vm_flags = vmf->vma->vm_flags; 89 __entry->vm_flags & VM_SHARED ? "shared" : "private", 111 __field(unsigned long, vm_flags) 122 __entry->vm_flags = vmf->vma->vm_flags; 134 __entry->vm_flags & VM_SHARED ? "shared" : "private", 158 __field(unsigned long, vm_flags) [all …]
|
| /Linux-v5.4/mm/ |
| D | mmap.c | 112 pgprot_t vm_get_page_prot(unsigned long vm_flags) in vm_get_page_prot() argument 114 pgprot_t ret = __pgprot(pgprot_val(protection_map[vm_flags & in vm_get_page_prot() 116 pgprot_val(arch_vm_get_page_prot(vm_flags))); in vm_get_page_prot() 122 static pgprot_t vm_pgprot_modify(pgprot_t oldprot, unsigned long vm_flags) in vm_pgprot_modify() argument 124 return pgprot_modify(oldprot, vm_get_page_prot(vm_flags)); in vm_pgprot_modify() 130 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot() local 133 vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags); in vma_set_page_prot() 135 vm_flags &= ~VM_SHARED; in vma_set_page_prot() 136 vm_page_prot = vm_pgprot_modify(vm_page_prot, vm_flags); in vma_set_page_prot() 148 if (vma->vm_flags & VM_DENYWRITE) in __remove_shared_vm_struct() [all …]
|
| D | mremap.c | 326 unsigned long vm_flags = vma->vm_flags; in move_vma() local 350 MADV_UNMERGEABLE, &vm_flags); in move_vma() 387 if (vm_flags & VM_ACCOUNT) { in move_vma() 388 vma->vm_flags &= ~VM_ACCOUNT; in move_vma() 405 vm_stat_account(mm, vma->vm_flags, new_len >> PAGE_SHIFT); in move_vma() 408 if (unlikely(vma->vm_flags & VM_PFNMAP)) in move_vma() 420 vma->vm_flags |= VM_ACCOUNT; in move_vma() 422 vma->vm_next->vm_flags |= VM_ACCOUNT; in move_vma() 425 if (vm_flags & VM_LOCKED) { in move_vma() 451 if (!old_len && !(vma->vm_flags & (VM_SHARED | VM_MAYSHARE))) { in vma_to_resize() [all …]
|
| D | nommu.c | 127 if (!(vma->vm_flags & (VM_IO | VM_PFNMAP))) in follow_pfn() 169 vma->vm_flags |= VM_USERMAP; in vmalloc_user() 559 if (region->vm_flags & VM_MAPPED_COPY) in __put_nommu_region() 946 unsigned long vm_flags; in determine_vm_flags() local 948 vm_flags = calc_vm_prot_bits(prot, 0) | calc_vm_flag_bits(flags); in determine_vm_flags() 953 vm_flags |= VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC; in determine_vm_flags() 955 vm_flags |= VM_MAYSHARE; in determine_vm_flags() 960 vm_flags |= VM_MAYSHARE | (capabilities & NOMMU_VMFLAGS); in determine_vm_flags() 962 vm_flags |= VM_SHARED; in determine_vm_flags() 970 vm_flags &= ~VM_MAYSHARE; in determine_vm_flags() [all …]
|
| D | mprotect.c | 64 if (prot_numa && !(vma->vm_flags & VM_SHARED) && in change_pte_range() 88 if (is_cow_mapping(vma->vm_flags) && in change_pte_range() 120 !(vma->vm_flags & VM_SOFTDIRTY))) { in change_pte_range() 343 unsigned long oldflags = vma->vm_flags; in mprotect_fixup() 361 (vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)) && in mprotect_fixup() 400 VM_WARN_ON((vma->vm_flags ^ newflags) & ~VM_SOFTDIRTY); in mprotect_fixup() 423 vma->vm_flags = newflags; in mprotect_fixup() 502 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_mprotect_pkey() 510 if (!(vma->vm_flags & VM_GROWSUP)) in do_mprotect_pkey() 525 if (rier && (vma->vm_flags & VM_MAYEXEC)) in do_mprotect_pkey() [all …]
|
| D | mlock.c | 448 vma->vm_flags &= VM_LOCKED_CLEAR_MASK; in munlock_vma_pages_range() 527 vm_flags_t old_flags = vma->vm_flags; in mlock_fixup() 529 if (newflags == vma->vm_flags || (vma->vm_flags & VM_SPECIAL) || in mlock_fixup() 574 vma->vm_flags = newflags; in mlock_fixup() 606 vm_flags_t newflags = vma->vm_flags & VM_LOCKED_CLEAR_MASK; in apply_vma_lock_flags() 657 if (vma->vm_flags & VM_LOCKED) { in count_mm_mlocked_page_nr() 725 vm_flags_t vm_flags = VM_LOCKED; in SYSCALL_DEFINE3() local 731 vm_flags |= VM_LOCKONFAULT; in SYSCALL_DEFINE3() 733 return do_mlock(start, len, vm_flags); in SYSCALL_DEFINE3() 788 newflags = vma->vm_flags & VM_LOCKED_CLEAR_MASK; in apply_mlockall_flags()
|
| D | gup.c | 272 if ((flags & FOLL_MLOCK) && (vma->vm_flags & VM_LOCKED)) { in follow_page_pte() 326 if (pmd_huge(pmdval) && vma->vm_flags & VM_HUGETLB) { in follow_pmd_mask() 436 if (pud_huge(*pud) && vma->vm_flags & VM_HUGETLB) { in follow_pud_mask() 671 if ((ret & VM_FAULT_WRITE) && !(vma->vm_flags & VM_WRITE)) in faultin_page() 678 vm_flags_t vm_flags = vma->vm_flags; in check_vma_flags() local 682 if (vm_flags & (VM_IO | VM_PFNMAP)) in check_vma_flags() 689 if (!(vm_flags & VM_WRITE)) { in check_vma_flags() 701 if (!is_cow_mapping(vm_flags)) in check_vma_flags() 704 } else if (!(vm_flags & VM_READ)) { in check_vma_flags() 711 if (!(vm_flags & VM_MAYREAD)) in check_vma_flags() [all …]
|
| /Linux-v5.4/arch/x86/include/uapi/asm/ |
| D | mman.h | 16 #define arch_vm_get_page_prot(vm_flags) __pgprot( \ argument 17 ((vm_flags) & VM_PKEY_BIT0 ? _PAGE_PKEY_BIT0 : 0) | \ 18 ((vm_flags) & VM_PKEY_BIT1 ? _PAGE_PKEY_BIT1 : 0) | \ 19 ((vm_flags) & VM_PKEY_BIT2 ? _PAGE_PKEY_BIT2 : 0) | \ 20 ((vm_flags) & VM_PKEY_BIT3 ? _PAGE_PKEY_BIT3 : 0))
|
| /Linux-v5.4/arch/sparc/include/asm/ |
| D | mman.h | 49 #define arch_vm_get_page_prot(vm_flags) sparc_vm_get_page_prot(vm_flags) argument 50 static inline pgprot_t sparc_vm_get_page_prot(unsigned long vm_flags) in sparc_vm_get_page_prot() argument 52 return (vm_flags & VM_SPARC_ADI) ? __pgprot(_PAGE_MCD_4V) : __pgprot(0); in sparc_vm_get_page_prot() 72 if (vma->vm_flags & (VM_PFNMAP | VM_MIXEDMAP)) in sparc_validate_prot() 83 if (vma->vm_flags & VM_MERGEABLE) in sparc_validate_prot()
|
| /Linux-v5.4/include/linux/ |
| D | khugepaged.h | 17 unsigned long vm_flags); 55 unsigned long vm_flags) in khugepaged_enter() argument 59 (khugepaged_req_madv() && (vm_flags & VM_HUGEPAGE))) && in khugepaged_enter() 60 !(vm_flags & VM_NOHUGEPAGE) && in khugepaged_enter() 75 unsigned long vm_flags) in khugepaged_enter() argument 80 unsigned long vm_flags) in khugepaged_enter_vma_merge() argument
|
| /Linux-v5.4/arch/nds32/mm/ |
| D | cacheflush.c | 34 cpu_cache_wbinval_page(kaddr, vma->vm_flags & VM_EXEC); in flush_icache_page() 69 (vma->vm_flags & VM_EXEC)) { in update_mmu_cache() 73 cpu_cache_wbinval_page(kaddr, vma->vm_flags & VM_EXEC); in update_mmu_cache() 140 if (vma->vm_flags & VM_EXEC) in flush_cache_range() 147 cpu_cache_wbinval_page(start, vma->vm_flags & VM_EXEC); in flush_cache_range() 161 cpu_cache_wbinval_page(vto, vma->vm_flags & VM_EXEC); in flush_cache_page() 275 if (vma->vm_flags & VM_EXEC) { in copy_to_user_page() 311 if (vma->vm_flags & VM_EXEC) in flush_anon_page()
|
| /Linux-v5.4/arch/hexagon/mm/ |
| D | vm_fault.c | 66 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 78 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 82 if (!(vma->vm_flags & VM_READ)) in do_page_fault() 86 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
| /Linux-v5.4/arch/nios2/mm/ |
| D | cacheflush.c | 87 if (!(mpnt->vm_flags & VM_MAYSHARE)) in flush_aliases() 136 if (vma == NULL || (vma->vm_flags & VM_EXEC)) in flush_cache_range() 156 if (vma->vm_flags & VM_EXEC) in flush_cache_page() 228 if (vma->vm_flags & VM_EXEC) in update_mmu_cache() 259 if (vma->vm_flags & VM_EXEC) in copy_from_user_page() 270 if (vma->vm_flags & VM_EXEC) in copy_to_user_page()
|
| D | fault.c | 98 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 115 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 119 if (!(vma->vm_flags & VM_READ)) in do_page_fault() 123 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
| /Linux-v5.4/arch/m68k/mm/ |
| D | fault.c | 94 if (vma->vm_flags & VM_IO) in do_page_fault() 98 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 121 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 128 if (!(vma->vm_flags & (VM_READ | VM_EXEC | VM_WRITE))) in do_page_fault()
|
| /Linux-v5.4/drivers/gpu/drm/udl/ |
| D | udl_gem.c | 67 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in update_vm_cache_attr() 70 pgprot_writecombine(vm_get_page_prot(vma->vm_flags)); in update_vm_cache_attr() 73 pgprot_noncached(vm_get_page_prot(vma->vm_flags)); in update_vm_cache_attr() 95 vma->vm_flags &= ~VM_PFNMAP; in udl_drm_gem_mmap() 96 vma->vm_flags |= VM_MIXEDMAP; in udl_drm_gem_mmap()
|
| /Linux-v5.4/arch/alpha/mm/ |
| D | fault.c | 126 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 136 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 140 if (!(vma->vm_flags & (VM_READ | VM_WRITE))) in do_page_fault() 143 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
| /Linux-v5.4/fs/ |
| D | userfaultfd.c | 647 vma->vm_flags &= ~(VM_UFFD_WP | VM_UFFD_MISSING); in userfaultfd_event_wait_completion() 679 vma->vm_flags &= ~(VM_UFFD_WP | VM_UFFD_MISSING); in dup_userfaultfd() 761 vma->vm_flags &= ~(VM_UFFD_WP | VM_UFFD_MISSING); in mremap_userfaultfd_prep() 904 !!(vma->vm_flags & (VM_UFFD_MISSING | VM_UFFD_WP))); in userfaultfd_release() 909 new_flags = vma->vm_flags & ~(VM_UFFD_MISSING | VM_UFFD_WP); in userfaultfd_release() 921 vma->vm_flags = new_flags; in userfaultfd_release() 1310 unsigned long vm_flags, new_flags; in userfaultfd_register() local 1328 vm_flags = 0; in userfaultfd_register() 1330 vm_flags |= VM_UFFD_MISSING; in userfaultfd_register() 1332 vm_flags |= VM_UFFD_WP; in userfaultfd_register() [all …]
|
| /Linux-v5.4/arch/mips/mm/ |
| D | fault.c | 106 if (!(vma->vm_flags & VM_GROWSDOWN)) in __do_page_fault() 118 if (!(vma->vm_flags & VM_WRITE)) in __do_page_fault() 123 if (address == regs->cp0_epc && !(vma->vm_flags & VM_EXEC)) { in __do_page_fault() 133 if (!(vma->vm_flags & VM_READ) && in __do_page_fault() 145 if (!(vma->vm_flags & (VM_READ | VM_WRITE | VM_EXEC))) in __do_page_fault()
|
| /Linux-v5.4/arch/riscv/mm/ |
| D | fault.c | 78 if (unlikely(!(vma->vm_flags & VM_GROWSDOWN))) in do_page_fault() 92 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 96 if (!(vma->vm_flags & VM_READ)) in do_page_fault() 100 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
| /Linux-v5.4/arch/ia64/mm/ |
| D | fault.c | 131 if (((isr >> IA64_ISR_R_BIT) & 1UL) && (!(vma->vm_flags & (VM_READ | VM_WRITE)))) in ia64_do_page_fault() 134 if ((vma->vm_flags & mask) != mask) in ia64_do_page_fault() 186 if (!(prev_vma && (prev_vma->vm_flags & VM_GROWSUP) && (address == prev_vma->vm_end))) { in ia64_do_page_fault() 189 if (!(vma->vm_flags & VM_GROWSDOWN)) in ia64_do_page_fault()
|
| /Linux-v5.4/arch/sparc/mm/ |
| D | fault_32.c | 209 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_sparc_fault() 220 if (!(vma->vm_flags & VM_WRITE)) in do_sparc_fault() 224 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in do_sparc_fault() 394 if (!(vma->vm_flags & VM_GROWSDOWN)) in force_user_fault() 401 if (!(vma->vm_flags & VM_WRITE)) in force_user_fault() 405 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in force_user_fault()
|
| /Linux-v5.4/drivers/sbus/char/ |
| D | flash.c | 47 if ((vma->vm_flags & VM_READ) && in flash_mmap() 48 (vma->vm_flags & VM_WRITE)) { in flash_mmap() 52 if (vma->vm_flags & VM_READ) { in flash_mmap() 55 } else if (vma->vm_flags & VM_WRITE) { in flash_mmap()
|