Home
last modified time | relevance | path

Searched refs:vm_userfaultfd_ctx (Results 1 – 11 of 11) sorted by relevance

/Linux-v5.10/include/linux/
Duserfaultfd_k.h50 struct vm_userfaultfd_ctx vm_ctx) in is_mergeable_vm_userfaultfd_ctx()
52 return vma->vm_userfaultfd_ctx.ctx == vm_ctx.ctx; in is_mergeable_vm_userfaultfd_ctx()
86 struct vm_userfaultfd_ctx *);
87 extern void mremap_userfaultfd_complete(struct vm_userfaultfd_ctx *,
111 struct vm_userfaultfd_ctx vm_ctx) in is_mergeable_vm_userfaultfd_ctx()
155 struct vm_userfaultfd_ctx *ctx) in mremap_userfaultfd_prep()
159 static inline void mremap_userfaultfd_complete(struct vm_userfaultfd_ctx *ctx, in mremap_userfaultfd_complete()
Dmm_types.h290 #define NULL_VM_UFFD_CTX ((struct vm_userfaultfd_ctx) { NULL, })
291 struct vm_userfaultfd_ctx { struct
295 #define NULL_VM_UFFD_CTX ((struct vm_userfaultfd_ctx) {}) argument
296 struct vm_userfaultfd_ctx {}; struct
373 struct vm_userfaultfd_ctx vm_userfaultfd_ctx; member
Dmm.h2518 struct mempolicy *, struct vm_userfaultfd_ctx);
/Linux-v5.10/fs/
Duserfaultfd.c397 ctx = vmf->vma->vm_userfaultfd_ctx.ctx; in handle_userfault()
605 if (vma->vm_userfaultfd_ctx.ctx == release_new_ctx) { in userfaultfd_event_wait_completion()
606 vma->vm_userfaultfd_ctx = NULL_VM_UFFD_CTX; in userfaultfd_event_wait_completion()
636 octx = vma->vm_userfaultfd_ctx.ctx; in dup_userfaultfd()
638 vma->vm_userfaultfd_ctx = NULL_VM_UFFD_CTX; in dup_userfaultfd()
676 vma->vm_userfaultfd_ctx.ctx = ctx; in dup_userfaultfd()
705 struct vm_userfaultfd_ctx *vm_ctx) in mremap_userfaultfd_prep()
709 ctx = vma->vm_userfaultfd_ctx.ctx; in mremap_userfaultfd_prep()
720 vma->vm_userfaultfd_ctx = NULL_VM_UFFD_CTX; in mremap_userfaultfd_prep()
725 void mremap_userfaultfd_complete(struct vm_userfaultfd_ctx *vm_ctx, in mremap_userfaultfd_complete()
[all …]
/Linux-v5.10/mm/
Dmmap.c1031 struct vm_userfaultfd_ctx vm_userfaultfd_ctx) in is_mergeable_vma() argument
1047 if (!is_mergeable_vm_userfaultfd_ctx(vma, vm_userfaultfd_ctx)) in is_mergeable_vma()
1081 struct vm_userfaultfd_ctx vm_userfaultfd_ctx) in can_vma_merge_before() argument
1083 if (is_mergeable_vma(vma, file, vm_flags, vm_userfaultfd_ctx) && in can_vma_merge_before()
1102 struct vm_userfaultfd_ctx vm_userfaultfd_ctx) in can_vma_merge_after() argument
1104 if (is_mergeable_vma(vma, file, vm_flags, vm_userfaultfd_ctx) && in can_vma_merge_after()
1162 struct vm_userfaultfd_ctx vm_userfaultfd_ctx) in vma_merge() argument
1192 vm_userfaultfd_ctx)) { in vma_merge()
1201 vm_userfaultfd_ctx) && in vma_merge()
1224 vm_userfaultfd_ctx)) { in vma_merge()
[all …]
Dmremap.c336 struct vm_userfaultfd_ctx *uf, struct list_head *uf_unmap) in move_vma()
546 unsigned long flags, struct vm_userfaultfd_ctx *uf, in mremap_to()
668 struct vm_userfaultfd_ctx uf = NULL_VM_UFFD_CTX; in SYSCALL_DEFINE5()
Duserfaultfd.c45 if (!dst_vma->vm_userfaultfd_ctx.ctx) in find_dst_vma()
Dmprotect.c457 vma->vm_userfaultfd_ctx); in mprotect_fixup()
Dmlock.c543 vma->vm_userfaultfd_ctx); in mlock_fixup()
Dmadvise.c139 vma->vm_userfaultfd_ctx); in madvise_behavior()
Dmempolicy.c832 new_pol, vma->vm_userfaultfd_ctx); in mbind_range()