/Linux-v6.6/arch/riscv/mm/ |
D | pageattr.c | 171 ret = walk_page_range(&init_mm, start, end, &pageattr_ops, &masks); in set_direct_map_invalid_noflush() 188 ret = walk_page_range(&init_mm, start, end, &pageattr_ops, &masks); in set_direct_map_default_noflush()
|
/Linux-v6.6/include/linux/ |
D | pagewalk.h | 117 int walk_page_range(struct mm_struct *mm, unsigned long start,
|
/Linux-v6.6/mm/ |
D | mincore.c | 202 err = walk_page_range(vma->vm_mm, addr, end, &mincore_walk_ops, vec); in do_mincore()
|
D | madvise.c | 290 walk_page_range(vma->vm_mm, start, end, &swapin_walk_ops, vma); in madvise_willneed() 554 walk_page_range(vma->vm_mm, addr, end, &cold_walk_ops, &walk_private); in madvise_cold_page_range() 592 walk_page_range(vma->vm_mm, addr, end, &cold_walk_ops, &walk_private); in madvise_pageout_page_range() 794 walk_page_range(vma->vm_mm, range.start, range.end, in madvise_free_single_vma()
|
D | hmm.c | 603 ret = walk_page_range(mm, hmm_vma_walk.last, range->end, in hmm_range_fault()
|
D | pagewalk.c | 470 int walk_page_range(struct mm_struct *mm, unsigned long start, in walk_page_range() function
|
D | migrate_device.c | 307 walk_page_range(migrate->vma->vm_mm, migrate->start, migrate->end, in migrate_vma_collect()
|
D | mlock.c | 394 walk_page_range(vma->vm_mm, start, end, &mlock_walk_ops, NULL); in mlock_vma_pages_range()
|
D | mprotect.c | 602 error = walk_page_range(current->mm, start, end, in mprotect_fixup()
|
D | memory-failure.c | 859 ret = walk_page_range(p->mm, 0, TASK_SIZE, &hwpoison_walk_ops, in kill_accessing_process()
|
D | mempolicy.c | 760 err = walk_page_range(mm, start, end, ops, &qp); in queue_pages_range()
|
D | memcontrol.c | 6053 walk_page_range(mm, 0, ULONG_MAX, &precharge_walk_ops, NULL); in mem_cgroup_count_precharge() 6355 walk_page_range(mc.mm, 0, ULONG_MAX, &charge_walk_ops, NULL); in mem_cgroup_move_charge()
|
D | vmscan.c | 4310 err = walk_page_range(mm, walk->next_addr, ULONG_MAX, &mm_walk_ops, walk); in walk_mm()
|
/Linux-v6.6/mm/damon/ |
D | vaddr.c | 399 walk_page_range(mm, addr, addr + 1, &damon_mkold_ops, NULL); in damon_va_mkold() 549 walk_page_range(mm, addr, addr + 1, &damon_young_ops, &arg); in damon_va_young()
|
/Linux-v6.6/Documentation/mm/ |
D | multigen_lru.rst | 121 and calls ``walk_page_range()`` with each ``mm_struct`` on this list 167 ``walk_page_range()`` with each ``mm_struct`` on this list to scan
|
D | unevictable-lru.rst | 302 mlock_pte_range() via walk_page_range() via mlock_vma_pages_range(). 377 mlock_pte_range() via walk_page_range() via mlock_vma_pages_range() - the same
|
/Linux-v6.6/arch/s390/mm/ |
D | gmap.c | 2589 walk_page_range(mm, 0, TASK_SIZE, &zap_zero_walk_ops, NULL); in s390_enable_sie() 2678 walk_page_range(mm, 0, TASK_SIZE, &enable_skey_walk_ops, NULL); in s390_enable_skey() 2704 walk_page_range(mm, 0, TASK_SIZE, &reset_cmma_walk_ops, NULL); in s390_reset_cmma() 2779 r = walk_page_range(mm, state.next, end, &gather_pages_ops, &state); in __s390_uv_destroy_range()
|
/Linux-v6.6/fs/proc/ |
D | task_mmu.c | 802 walk_page_range(vma->vm_mm, start, vma->vm_end, ops, mss); in smap_gather_stats() 1305 walk_page_range(mm, 0, -1, &clear_refs_walk_ops, &cp); in clear_refs_write() 1719 ret = walk_page_range(mm, start_vaddr, end, &pagemap_ops, &pm); in pagemap_read()
|