/Linux-v5.4/mm/ |
D | page_ext.c | 121 base = NODE_DATA(page_to_nid(page))->node_page_ext; in lookup_page_ext() 130 index = pfn - round_down(node_start_pfn(page_to_nid(page)), in lookup_page_ext()
|
D | list_lru.c | 127 int nid = page_to_nid(virt_to_page(item)); in list_lru_add() 151 int nid = page_to_nid(virt_to_page(item)); in list_lru_del()
|
D | sparse.c | 47 int page_to_nid(const struct page *page) in page_to_nid() function 51 EXPORT_SYMBOL(page_to_nid);
|
D | hugetlb.c | 867 int nid = page_to_nid(page); in enqueue_huge_page() 1182 h->nr_huge_pages_node[page_to_nid(page)]--; in update_and_free_page() 1265 int nid = page_to_nid(page); in free_huge_page() 1473 prep_new_huge_page(h, page, page_to_nid(page)); in alloc_fresh_huge_page() 1569 int nid = page_to_nid(head); in dissolve_free_huge_page() 1654 h->surplus_huge_pages_node[page_to_nid(page)]++; in alloc_surplus_huge_page() 2223 prep_new_huge_page(h, page, page_to_nid(page)); in gather_bootmem_prealloc() 2331 h->free_huge_pages_node[page_to_nid(page)]--; in try_to_free_low() 5119 int old_nid = page_to_nid(oldpage); in move_hugetlb_state() 5120 int new_nid = page_to_nid(newpage); in move_hugetlb_state()
|
D | slub.c | 1173 struct kmem_cache_node *n = get_node(s, page_to_nid(page)); in free_debug_processing() 1691 inc_slabs_node(s, page_to_nid(page), page->objects); in allocate_slab() 1751 dec_slabs_node(s, page_to_nid(page), page->objects); in discard_slab() 2042 struct kmem_cache_node *n = get_node(s, page_to_nid(page)); in deactivate_slab() 2196 n2 = get_node(s, page_to_nid(page)); in unfreeze_partials() 2372 if (node != NUMA_NO_NODE && page_to_nid(page) != node) in node_match() 2871 n = get_node(s, page_to_nid(page)); in __slab_free() 3376 if (page_to_nid(page) != node) { in early_kmem_cache_node_alloc() 4561 node_set(page_to_nid(virt_to_page(track)), l->nodes); in add_location() 4592 node_set(page_to_nid(virt_to_page(track)), l->nodes); in add_location() [all …]
|
D | huge_memory.c | 503 struct pglist_data *pgdat = NODE_DATA(page_to_nid(page)); in get_deferred_split_queue() 513 struct pglist_data *pgdat = NODE_DATA(page_to_nid(page)); in get_deferred_split_queue() 1218 vmf->address, page_to_nid(page)); in do_huge_pmd_wp_page_fallback() 1566 page_nid = page_to_nid(page); in do_huge_pmd_numa_page() 2696 struct pglist_data *pgdata = NODE_DATA(page_to_nid(head)); in split_huge_page_to_list() 2880 memcg_set_shrinker_bit(memcg, page_to_nid(page), in deferred_split_huge_page()
|
D | memremap.c | 119 nid = page_to_nid(first_page); in memunmap_pages()
|
D | mempolicy.c | 425 int nid = page_to_nid(page); in queue_pages_required() 862 err = page_to_nid(p); in lookup_node() 2066 if (page && page_to_nid(page) == nid) { in alloc_page_interleave() 2389 int curnid = page_to_nid(page); in mpol_misplaced()
|
D | memory-failure.c | 260 drop_slab_node(page_to_nid(p)); in shake_page() 1615 int nid = page_to_nid(p); in new_page()
|
D | mprotect.c | 108 if (target_node == page_to_nid(page)) in change_pte_range()
|
D | memory_hotplug.c | 1175 if (!node_online(page_to_nid(page))) in is_pageblock_removable_nolock() 1289 int nid = page_to_nid(page); in new_node_page()
|
D | slab.c | 558 page_node = page_to_nid(page); in cache_free_pfmemalloc() 785 int page_node = page_to_nid(virt_to_page(objp)); in cache_free_alien() 2603 page_node = page_to_nid(page); in cache_grow_begin() 2658 n = get_node(cachep, page_to_nid(page)); in cache_grow_end() 3149 nid = page_to_nid(page); in fallback_alloc()
|
D | slob.c | 326 if (node != NUMA_NO_NODE && page_to_nid(sp) != node) in slob_alloc()
|
D | khugepaged.c | 1184 node = page_to_nid(page); in khugepaged_scan_pmd() 1845 node = page_to_nid(page); in khugepaged_scan_file()
|
D | migrate.c | 1550 if (page_to_nid(page) == node) in add_page_for_migration() 1705 err = page ? page_to_nid(page) : -ENOENT; in do_pages_stat_array()
|
D | kmemleak.c | 1454 if (page_to_nid(page) != i) in kmemleak_scan()
|
D | vmscan.c | 2164 nid = page_to_nid(page); in reclaim_pages() 2168 if (nid == page_to_nid(page)) { in reclaim_pages()
|
/Linux-v5.4/include/asm-generic/ |
D | memory_model.h | 46 struct pglist_data *__pgdat = NODE_DATA(page_to_nid(__pg)); \
|
/Linux-v5.4/arch/m68k/include/asm/ |
D | page_mm.h | 165 pgdat = &pg_data_map[page_to_nid(__p)]; \
|
/Linux-v5.4/include/linux/ |
D | mm.h | 1096 extern int page_to_nid(const struct page *page); 1098 static inline int page_to_nid(const struct page *page) in page_to_nid() function 1173 return page_to_nid(page); /* XXX */ in page_cpupid_xchg_last() 1178 return page_to_nid(page); /* XXX */ in page_cpupid_last() 1244 return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)]; in page_zone() 1249 return NODE_DATA(page_to_nid(page)); in page_pgdat()
|
D | mmzone.h | 1379 page_to_nid(pfn_to_page(__pfn_to_nid_pfn)); \
|
/Linux-v5.4/drivers/net/ethernet/cavium/liquidio/ |
D | octeon_network.h | 342 unlikely(page_to_nid(pg_info->page) != numa_node_id())) { in recv_buffer_recycle()
|
/Linux-v5.4/fs/proc/ |
D | task_mmu.c | 1691 md->node[page_to_nid(page)] += nr_pages; in gather_stats() 1710 nid = page_to_nid(page); in can_gather_numa_stats() 1735 nid = page_to_nid(page); in can_gather_numa_stats_pmd()
|
/Linux-v5.4/Documentation/vm/ |
D | memory-model.rst | 97 :c:func:`page_to_nid` is generic as it uses the node number encoded in
|
/Linux-v5.4/drivers/net/ethernet/hisilicon/hns/ |
D | hns_enet.c | 435 if (unlikely(page_to_nid(desc_cb->priv) != numa_node_id())) in hns_nic_reuse_page() 583 if (likely(page_to_nid(desc_cb->priv) == numa_node_id())) in hns_nic_poll_rx_skb()
|