/Linux-v4.19/mm/ |
D | memory.c | 2380 static vm_fault_t do_page_mkwrite(struct vm_fault *vmf) in do_page_mkwrite() argument 2383 struct page *page = vmf->page; in do_page_mkwrite() 2384 unsigned int old_flags = vmf->flags; in do_page_mkwrite() 2386 vmf->flags = FAULT_FLAG_WRITE|FAULT_FLAG_MKWRITE; in do_page_mkwrite() 2388 ret = vmf->vma->vm_ops->page_mkwrite(vmf); in do_page_mkwrite() 2390 vmf->flags = old_flags; in do_page_mkwrite() 2448 static inline void wp_page_reuse(struct vm_fault *vmf) in wp_page_reuse() argument 2449 __releases(vmf->ptl) in wp_page_reuse() 2451 struct vm_area_struct *vma = vmf->vma; in wp_page_reuse() 2452 struct page *page = vmf->page; in wp_page_reuse() [all …]
|
D | huge_memory.c | 544 static vm_fault_t __do_huge_pmd_anonymous_page(struct vm_fault *vmf, in __do_huge_pmd_anonymous_page() argument 547 struct vm_area_struct *vma = vmf->vma; in __do_huge_pmd_anonymous_page() 550 unsigned long haddr = vmf->address & HPAGE_PMD_MASK; in __do_huge_pmd_anonymous_page() 567 clear_huge_page(page, vmf->address, HPAGE_PMD_NR); in __do_huge_pmd_anonymous_page() 575 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page() 576 if (unlikely(!pmd_none(*vmf->pmd))) { in __do_huge_pmd_anonymous_page() 589 spin_unlock(vmf->ptl); in __do_huge_pmd_anonymous_page() 593 ret2 = handle_userfault(vmf, VM_UFFD_MISSING); in __do_huge_pmd_anonymous_page() 603 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); in __do_huge_pmd_anonymous_page() 604 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, entry); in __do_huge_pmd_anonymous_page() [all …]
|
D | swap_state.c | 567 struct vm_fault *vmf) in swap_cluster_readahead() argument 577 struct vm_area_struct *vma = vmf->vma; in swap_cluster_readahead() 578 unsigned long addr = vmf->address; in swap_cluster_readahead() 664 static void swap_ra_info(struct vm_fault *vmf, in swap_ra_info() argument 667 struct vm_area_struct *vma = vmf->vma; in swap_ra_info() 685 faddr = vmf->address; in swap_ra_info() 686 orig_pte = pte = pte_offset_map(vmf->pmd, faddr); in swap_ra_info() 733 struct vm_fault *vmf) in swap_vma_readahead() argument 736 struct vm_area_struct *vma = vmf->vma; in swap_vma_readahead() 744 swap_ra_info(vmf, &ra_info); in swap_vma_readahead() [all …]
|
/Linux-v4.19/include/trace/events/ |
D | fs_dax.h | 11 TP_PROTO(struct inode *inode, struct vm_fault *vmf, 13 TP_ARGS(inode, vmf, max_pgoff, result), 29 __entry->vm_start = vmf->vma->vm_start; 30 __entry->vm_end = vmf->vma->vm_end; 31 __entry->vm_flags = vmf->vma->vm_flags; 32 __entry->address = vmf->address; 33 __entry->flags = vmf->flags; 34 __entry->pgoff = vmf->pgoff; 56 TP_PROTO(struct inode *inode, struct vm_fault *vmf, \ 58 TP_ARGS(inode, vmf, max_pgoff, result)) [all …]
|
/Linux-v4.19/fs/ |
D | dax.c | 803 struct vm_fault *vmf, in dax_insert_mapping_entry() argument 809 pgoff_t index = vmf->pgoff; in dax_insert_mapping_entry() 821 unmap_mapping_pages(mapping, vmf->pgoff, 1, false); in dax_insert_mapping_entry() 828 dax_associate_entry(new_entry, mapping, vmf->vma, vmf->address); in dax_insert_mapping_entry() 1129 struct vm_fault *vmf) in dax_load_hole() argument 1132 unsigned long vaddr = vmf->address; in dax_load_hole() 1136 dax_insert_mapping_entry(mapping, vmf, entry, pfn, RADIX_DAX_ZERO_PAGE, in dax_load_hole() 1138 ret = vmf_insert_mixed(vmf->vma, vaddr, pfn); in dax_load_hole() 1139 trace_dax_load_hole(inode, vmf, ret); in dax_load_hole() 1340 static vm_fault_t dax_iomap_pte_fault(struct vm_fault *vmf, pfn_t *pfnp, in dax_iomap_pte_fault() argument [all …]
|
/Linux-v4.19/drivers/dax/ |
D | device.c | 252 struct vm_fault *vmf, pfn_t *pfn) in __dev_dax_pte_fault() argument 259 if (check_vma(dev_dax, vmf->vma, __func__)) in __dev_dax_pte_fault() 272 phys = dax_pgoff_to_phys(dev_dax, vmf->pgoff, PAGE_SIZE); in __dev_dax_pte_fault() 274 dev_dbg(dev, "pgoff_to_phys(%#lx) failed\n", vmf->pgoff); in __dev_dax_pte_fault() 280 return vmf_insert_mixed(vmf->vma, vmf->address, *pfn); in __dev_dax_pte_fault() 284 struct vm_fault *vmf, pfn_t *pfn) in __dev_dax_pmd_fault() argument 286 unsigned long pmd_addr = vmf->address & PMD_MASK; in __dev_dax_pmd_fault() 293 if (check_vma(dev_dax, vmf->vma, __func__)) in __dev_dax_pmd_fault() 315 if (pmd_addr < vmf->vma->vm_start || in __dev_dax_pmd_fault() 316 (pmd_addr + PMD_SIZE) > vmf->vma->vm_end) in __dev_dax_pmd_fault() [all …]
|
/Linux-v4.19/fs/ocfs2/ |
D | mmap.c | 47 static vm_fault_t ocfs2_fault(struct vm_fault *vmf) in ocfs2_fault() argument 49 struct vm_area_struct *vma = vmf->vma; in ocfs2_fault() 54 ret = filemap_fault(vmf); in ocfs2_fault() 58 vma, vmf->page, vmf->pgoff); in ocfs2_fault() 129 static vm_fault_t ocfs2_page_mkwrite(struct vm_fault *vmf) in ocfs2_page_mkwrite() argument 131 struct page *page = vmf->page; in ocfs2_page_mkwrite() 132 struct inode *inode = file_inode(vmf->vma->vm_file); in ocfs2_page_mkwrite() 160 ret = __ocfs2_page_mkwrite(vmf->vma->vm_file, di_bh, page); in ocfs2_page_mkwrite()
|
/Linux-v4.19/drivers/video/fbdev/core/ |
D | fb_defio.c | 40 static vm_fault_t fb_deferred_io_fault(struct vm_fault *vmf) in fb_deferred_io_fault() argument 44 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_fault() 46 offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_fault() 56 if (vmf->vma->vm_file) in fb_deferred_io_fault() 57 page->mapping = vmf->vma->vm_file->f_mapping; in fb_deferred_io_fault() 62 page->index = vmf->pgoff; in fb_deferred_io_fault() 64 vmf->page = page; in fb_deferred_io_fault() 93 static vm_fault_t fb_deferred_io_mkwrite(struct vm_fault *vmf) in fb_deferred_io_mkwrite() argument 95 struct page *page = vmf->page; in fb_deferred_io_mkwrite() 96 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_mkwrite() [all …]
|
/Linux-v4.19/drivers/gpu/drm/ |
D | drm_vm.c | 103 static vm_fault_t drm_vm_fault(struct vm_fault *vmf) in drm_vm_fault() argument 105 struct vm_area_struct *vma = vmf->vma; in drm_vm_fault() 132 resource_size_t offset = vmf->address - vma->vm_start; in drm_vm_fault() 162 vmf->page = page; in drm_vm_fault() 176 static vm_fault_t drm_vm_fault(struct vm_fault *vmf) in drm_vm_fault() argument 192 static vm_fault_t drm_vm_shm_fault(struct vm_fault *vmf) in drm_vm_shm_fault() argument 194 struct vm_area_struct *vma = vmf->vma; in drm_vm_shm_fault() 203 offset = vmf->address - vma->vm_start; in drm_vm_shm_fault() 209 vmf->page = page; in drm_vm_shm_fault() 294 static vm_fault_t drm_vm_dma_fault(struct vm_fault *vmf) in drm_vm_dma_fault() argument [all …]
|
/Linux-v4.19/drivers/gpu/drm/ttm/ |
D | ttm_bo_vm.c | 48 struct vm_fault *vmf) in ttm_bo_vm_fault_idle() argument 66 if (vmf->flags & FAULT_FLAG_ALLOW_RETRY) { in ttm_bo_vm_fault_idle() 68 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT) in ttm_bo_vm_fault_idle() 72 up_read(&vmf->vma->vm_mm->mmap_sem); in ttm_bo_vm_fault_idle() 109 static vm_fault_t ttm_bo_vm_fault(struct vm_fault *vmf) in ttm_bo_vm_fault() argument 111 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_fault() 123 unsigned long address = vmf->address; in ttm_bo_vm_fault() 139 if (vmf->flags & FAULT_FLAG_ALLOW_RETRY) { in ttm_bo_vm_fault() 140 if (!(vmf->flags & FAULT_FLAG_RETRY_NOWAIT)) { in ttm_bo_vm_fault() 142 up_read(&vmf->vma->vm_mm->mmap_sem); in ttm_bo_vm_fault() [all …]
|
/Linux-v4.19/arch/x86/entry/vdso/ |
D | vma.c | 43 struct vm_area_struct *vma, struct vm_fault *vmf) in vdso_fault() argument 47 if (!image || (vmf->pgoff << PAGE_SHIFT) >= image->size) in vdso_fault() 50 vmf->page = virt_to_page(image->data + (vmf->pgoff << PAGE_SHIFT)); in vdso_fault() 51 get_page(vmf->page); in vdso_fault() 88 struct vm_area_struct *vma, struct vm_fault *vmf) in vvar_fault() argument 97 sym_offset = (long)(vmf->pgoff << PAGE_SHIFT) + in vvar_fault() 111 ret = vm_insert_pfn(vma, vmf->address, in vvar_fault() 119 vmf->address, in vvar_fault() 127 ret = vm_insert_pfn(vma, vmf->address, in vvar_fault()
|
/Linux-v4.19/drivers/xen/ |
D | privcmd-buf.c | 126 static vm_fault_t privcmd_buf_vma_fault(struct vm_fault *vmf) in privcmd_buf_vma_fault() argument 129 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end, in privcmd_buf_vma_fault() 130 vmf->pgoff, (void *)vmf->address); in privcmd_buf_vma_fault()
|
/Linux-v4.19/fs/ext2/ |
D | file.c | 91 static vm_fault_t ext2_dax_fault(struct vm_fault *vmf) in ext2_dax_fault() argument 93 struct inode *inode = file_inode(vmf->vma->vm_file); in ext2_dax_fault() 97 if (vmf->flags & FAULT_FLAG_WRITE) { in ext2_dax_fault() 99 file_update_time(vmf->vma->vm_file); in ext2_dax_fault() 103 ret = dax_iomap_fault(vmf, PE_SIZE_PTE, NULL, NULL, &ext2_iomap_ops); in ext2_dax_fault() 106 if (vmf->flags & FAULT_FLAG_WRITE) in ext2_dax_fault()
|
/Linux-v4.19/fs/xfs/ |
D | xfs_file.c | 1056 struct vm_fault *vmf, in __xfs_filemap_fault() argument 1060 struct inode *inode = file_inode(vmf->vma->vm_file); in __xfs_filemap_fault() 1068 file_update_time(vmf->vma->vm_file); in __xfs_filemap_fault() 1075 ret = dax_iomap_fault(vmf, pe_size, &pfn, NULL, &xfs_iomap_ops); in __xfs_filemap_fault() 1077 ret = dax_finish_sync_fault(vmf, pe_size, pfn); in __xfs_filemap_fault() 1080 ret = iomap_page_mkwrite(vmf, &xfs_iomap_ops); in __xfs_filemap_fault() 1082 ret = filemap_fault(vmf); in __xfs_filemap_fault() 1093 struct vm_fault *vmf) in xfs_filemap_fault() argument 1096 return __xfs_filemap_fault(vmf, PE_SIZE_PTE, in xfs_filemap_fault() 1097 IS_DAX(file_inode(vmf->vma->vm_file)) && in xfs_filemap_fault() [all …]
|
/Linux-v4.19/drivers/gpu/drm/vkms/ |
D | vkms_gem.c | 46 int vkms_gem_fault(struct vm_fault *vmf) in vkms_gem_fault() argument 48 struct vm_area_struct *vma = vmf->vma; in vkms_gem_fault() 50 unsigned long vaddr = vmf->address; in vkms_gem_fault() 65 vmf->page = obj->pages[page_offset]; in vkms_gem_fault() 77 vmf->page = page; in vkms_gem_fault()
|
/Linux-v4.19/drivers/misc/ocxl/ |
D | sysfs.c | 67 static vm_fault_t global_mmio_fault(struct vm_fault *vmf) in global_mmio_fault() argument 69 struct vm_area_struct *vma = vmf->vma; in global_mmio_fault() 73 if (vmf->pgoff >= (afu->config.global_mmio_size >> PAGE_SHIFT)) in global_mmio_fault() 76 offset = vmf->pgoff; in global_mmio_fault() 78 return vmf_insert_pfn(vma, vmf->address, offset); in global_mmio_fault()
|
D | context.c | 129 static vm_fault_t ocxl_mmap_fault(struct vm_fault *vmf) in ocxl_mmap_fault() argument 131 struct vm_area_struct *vma = vmf->vma; in ocxl_mmap_fault() 136 offset = vmf->pgoff << PAGE_SHIFT; in ocxl_mmap_fault() 138 ctx->pasid, vmf->address, offset); in ocxl_mmap_fault() 141 ret = map_pp_mmio(vma, vmf->address, offset, ctx); in ocxl_mmap_fault() 143 ret = map_afu_irq(vma, vmf->address, offset, ctx); in ocxl_mmap_fault()
|
/Linux-v4.19/sound/usb/usx2y/ |
D | usX2Yhwdep.c | 34 static vm_fault_t snd_us428ctls_vm_fault(struct vm_fault *vmf) in snd_us428ctls_vm_fault() argument 41 vmf->vma->vm_start, in snd_us428ctls_vm_fault() 42 vmf->pgoff); in snd_us428ctls_vm_fault() 44 offset = vmf->pgoff << PAGE_SHIFT; in snd_us428ctls_vm_fault() 45 vaddr = (char *)((struct usX2Ydev *)vmf->vma->vm_private_data)->us428ctls_sharedmem + offset; in snd_us428ctls_vm_fault() 48 vmf->page = page; in snd_us428ctls_vm_fault()
|
/Linux-v4.19/fs/nilfs2/ |
D | file.c | 45 static vm_fault_t nilfs_page_mkwrite(struct vm_fault *vmf) in nilfs_page_mkwrite() argument 47 struct vm_area_struct *vma = vmf->vma; in nilfs_page_mkwrite() 48 struct page *page = vmf->page; in nilfs_page_mkwrite() 99 ret = block_page_mkwrite(vma, vmf, nilfs_get_block); in nilfs_page_mkwrite()
|
/Linux-v4.19/drivers/gpu/drm/gma500/ |
D | gem.c | 137 vm_fault_t psb_gem_fault(struct vm_fault *vmf) in psb_gem_fault() argument 139 struct vm_area_struct *vma = vmf->vma; in psb_gem_fault() 173 page_offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT; in psb_gem_fault() 180 ret = vmf_insert_pfn(vma, vmf->address, pfn); in psb_gem_fault()
|
/Linux-v4.19/fs/ext4/ |
D | file.c | 280 static vm_fault_t ext4_dax_huge_fault(struct vm_fault *vmf, in ext4_dax_huge_fault() argument 287 struct inode *inode = file_inode(vmf->vma->vm_file); in ext4_dax_huge_fault() 301 bool write = (vmf->flags & FAULT_FLAG_WRITE) && in ext4_dax_huge_fault() 302 (vmf->vma->vm_flags & VM_SHARED); in ext4_dax_huge_fault() 307 file_update_time(vmf->vma->vm_file); in ext4_dax_huge_fault() 320 result = dax_iomap_fault(vmf, pe_size, &pfn, &error, &ext4_iomap_ops); in ext4_dax_huge_fault() 329 result = dax_finish_sync_fault(vmf, pe_size, pfn); in ext4_dax_huge_fault() 339 static vm_fault_t ext4_dax_fault(struct vm_fault *vmf) in ext4_dax_fault() argument 341 return ext4_dax_huge_fault(vmf, PE_SIZE_PTE); in ext4_dax_fault()
|
/Linux-v4.19/include/linux/ |
D | huge_mm.h | 10 extern vm_fault_t do_huge_pmd_anonymous_page(struct vm_fault *vmf); 14 extern void huge_pmd_set_accessed(struct vm_fault *vmf, pmd_t orig_pmd); 20 extern void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud); 22 static inline void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud) in huge_pud_set_accessed() argument 27 extern vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf, pmd_t orig_pmd); 220 extern vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf, pmd_t orig_pmd); 325 static inline vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf, in do_huge_pmd_numa_page() argument
|
/Linux-v4.19/drivers/misc/cxl/ |
D | context.c | 130 static vm_fault_t cxl_mmap_fault(struct vm_fault *vmf) in cxl_mmap_fault() argument 132 struct vm_area_struct *vma = vmf->vma; in cxl_mmap_fault() 137 offset = vmf->pgoff << PAGE_SHIFT; in cxl_mmap_fault() 140 __func__, ctx->pe, vmf->address, offset); in cxl_mmap_fault() 165 vmf->page = ctx->ff_page; in cxl_mmap_fault() 172 ret = vmf_insert_pfn(vma, vmf->address, (area + offset) >> PAGE_SHIFT); in cxl_mmap_fault()
|
/Linux-v4.19/arch/s390/kernel/ |
D | vdso.c | 51 struct vm_area_struct *vma, struct vm_fault *vmf) in vdso_fault() argument 65 if (vmf->pgoff >= vdso_pages) in vdso_fault() 68 vmf->page = vdso_pagelist[vmf->pgoff]; in vdso_fault() 69 get_page(vmf->page); in vdso_fault()
|
/Linux-v4.19/drivers/char/ |
D | mspec.c | 195 mspec_fault(struct vm_fault *vmf) in mspec_fault() argument 199 pgoff_t index = vmf->pgoff; in mspec_fault() 200 struct vma_data *vdata = vmf->vma->vm_private_data; in mspec_fault() 226 return vmf_insert_pfn(vmf->vma, vmf->address, pfn); in mspec_fault()
|