/Linux-v6.1/arch/riscv/mm/ |
D | pageattr.c | 170 ret = walk_page_range(&init_mm, start, end, &pageattr_ops, &masks); in set_direct_map_invalid_noflush() 187 ret = walk_page_range(&init_mm, start, end, &pageattr_ops, &masks); in set_direct_map_default_noflush()
|
/Linux-v6.1/include/linux/ |
D | pagewalk.h | 95 int walk_page_range(struct mm_struct *mm, unsigned long start,
|
/Linux-v6.1/mm/ |
D | madvise.c | 287 walk_page_range(vma->vm_mm, start, end, &swapin_walk_ops, vma); in madvise_willneed() 514 walk_page_range(vma->vm_mm, addr, end, &cold_walk_ops, &walk_private); in madvise_cold_page_range() 552 walk_page_range(vma->vm_mm, addr, end, &cold_walk_ops, &walk_private); in madvise_pageout_page_range() 762 walk_page_range(vma->vm_mm, range.start, range.end, in madvise_free_single_vma()
|
D | mincore.c | 202 err = walk_page_range(vma->vm_mm, addr, end, &mincore_walk_ops, vec); in do_mincore()
|
D | hmm.c | 588 ret = walk_page_range(mm, hmm_vma_walk.last, range->end, in hmm_range_fault()
|
D | pagewalk.c | 427 int walk_page_range(struct mm_struct *mm, unsigned long start, in walk_page_range() function
|
D | migrate_device.c | 313 walk_page_range(migrate->vma->vm_mm, migrate->start, migrate->end, in migrate_vma_collect()
|
D | mlock.c | 386 walk_page_range(vma->vm_mm, start, end, &mlock_walk_ops, NULL); in mlock_vma_pages_range()
|
D | mprotect.c | 575 error = walk_page_range(current->mm, start, end, in mprotect_fixup()
|
D | memory-failure.c | 753 ret = walk_page_range(p->mm, 0, TASK_SIZE, &hwp_walk_ops, in kill_accessing_process()
|
D | mempolicy.c | 741 err = walk_page_range(mm, start, end, &queue_pages_walk_ops, &qp); in queue_pages_range()
|
D | memcontrol.c | 5953 walk_page_range(mm, 0, ULONG_MAX, &precharge_walk_ops, NULL); in mem_cgroup_count_precharge() 6251 walk_page_range(mc.mm, 0, ULONG_MAX, &charge_walk_ops, NULL); in mem_cgroup_move_charge()
|
D | vmscan.c | 4183 err = walk_page_range(mm, walk->next_addr, ULONG_MAX, &mm_walk_ops, walk); in walk_mm()
|
/Linux-v6.1/mm/damon/ |
D | vaddr.c | 392 walk_page_range(mm, addr, addr + 1, &damon_mkold_ops, NULL); in damon_va_mkold() 535 walk_page_range(mm, addr, addr + 1, &damon_young_ops, &arg); in damon_va_young()
|
/Linux-v6.1/arch/s390/mm/ |
D | gmap.c | 2577 walk_page_range(mm, 0, TASK_SIZE, &zap_zero_walk_ops, NULL); in s390_enable_sie() 2672 walk_page_range(mm, 0, TASK_SIZE, &enable_skey_walk_ops, NULL); in s390_enable_skey() 2697 walk_page_range(mm, 0, TASK_SIZE, &reset_cmma_walk_ops, NULL); in s390_reset_cmma() 2771 r = walk_page_range(mm, state.next, end, &gather_pages_ops, &state); in __s390_uv_destroy_range()
|
/Linux-v6.1/fs/proc/ |
D | task_mmu.c | 805 walk_page_range(vma->vm_mm, start, vma->vm_end, ops, mss); in smap_gather_stats() 1306 walk_page_range(mm, 0, -1, &clear_refs_walk_ops, &cp); in clear_refs_write() 1713 ret = walk_page_range(mm, start_vaddr, end, &pagemap_ops, &pm); in pagemap_read()
|
/Linux-v6.1/Documentation/mm/ |
D | multigen_lru.rst | 120 and calls ``walk_page_range()`` with each ``mm_struct`` on this list
|
D | unevictable-lru.rst | 299 mlock_pte_range() via walk_page_range() via mlock_vma_pages_range(). 374 mlock_pte_range() via walk_page_range() via mlock_vma_pages_range() - the same
|