/Linux-v4.19/include/trace/events/ |
D | fs_dax.h | 65 struct page *zero_page, 67 TP_ARGS(inode, vmf, zero_page, radix_entry), 72 __field(struct page *, zero_page) 81 __entry->zero_page = zero_page; 91 __entry->zero_page, 99 struct page *zero_page, void *radix_entry), \ 100 TP_ARGS(inode, vmf, zero_page, radix_entry))
|
/Linux-v4.19/arch/nds32/mm/ |
D | init.c | 144 void *zero_page; in paging_init() local 156 zero_page = __va(memblock_alloc(PAGE_SIZE, PAGE_SIZE)); in paging_init() 157 memset(zero_page, 0, PAGE_SIZE); in paging_init() 160 empty_zero_page = virt_to_page(zero_page); in paging_init()
|
/Linux-v4.19/arch/arm64/kernel/ |
D | hibernate-asm.S | 36 .macro break_before_make_ttbr_switch zero_page, page_table, tmp 37 phys_to_ttbr \tmp, \zero_page
|
D | hibernate.c | 480 void *zero_page; in swsusp_arch_resume() local 506 zero_page = (void *)get_safe_page(GFP_ATOMIC); in swsusp_arch_resume() 507 if (!zero_page) { in swsusp_arch_resume() 554 resume_hdr.__hyp_stub_vectors, virt_to_phys(zero_page)); in swsusp_arch_resume()
|
/Linux-v4.19/mm/ |
D | huge_memory.c | 67 struct page *zero_page; in get_huge_zero_page() local 72 zero_page = alloc_pages((GFP_TRANSHUGE | __GFP_ZERO) & ~__GFP_MOVABLE, in get_huge_zero_page() 74 if (!zero_page) { in get_huge_zero_page() 80 if (cmpxchg(&huge_zero_page, NULL, zero_page)) { in get_huge_zero_page() 82 __free_pages(zero_page, compound_order(zero_page)); in get_huge_zero_page() 132 struct page *zero_page = xchg(&huge_zero_page, NULL); in shrink_huge_zero_page_scan() local 133 BUG_ON(zero_page == NULL); in shrink_huge_zero_page_scan() 134 __free_pages(zero_page, compound_order(zero_page)); in shrink_huge_zero_page_scan() 652 struct page *zero_page) in set_huge_zero_page() argument 657 entry = mk_pmd(zero_page, vma->vm_page_prot); in set_huge_zero_page() [all …]
|
/Linux-v4.19/arch/unicore32/mm/ |
D | mmu.c | 424 void *zero_page; in paging_init() local 435 zero_page = early_alloc(PAGE_SIZE); in paging_init() 439 empty_zero_page = virt_to_page(zero_page); in paging_init()
|
/Linux-v4.19/fs/ |
D | dax.c | 1514 struct page *zero_page; in dax_pmd_load_hole() local 1520 zero_page = mm_get_huge_zero_page(vmf->vma->vm_mm); in dax_pmd_load_hole() 1522 if (unlikely(!zero_page)) in dax_pmd_load_hole() 1525 pfn = page_to_pfn_t(zero_page); in dax_pmd_load_hole() 1535 pmd_entry = mk_pmd(zero_page, vmf->vma->vm_page_prot); in dax_pmd_load_hole() 1539 trace_dax_pmd_load_hole(inode, vmf, zero_page, ret); in dax_pmd_load_hole() 1543 trace_dax_pmd_load_hole_fallback(inode, vmf, zero_page, ret); in dax_pmd_load_hole()
|
/Linux-v4.19/include/target/ |
D | target_core_fabric.h | 188 u32 length, bool zero_page, bool chainable);
|
/Linux-v4.19/net/ceph/ |
D | messenger.c | 188 static struct page *zero_page; /* used in certain error cases */ variable 267 BUG_ON(zero_page == NULL); in _ceph_msgr_exit() 268 put_page(zero_page); in _ceph_msgr_exit() 269 zero_page = NULL; in _ceph_msgr_exit() 279 BUG_ON(zero_page != NULL); in ceph_msgr_init() 280 zero_page = ZERO_PAGE(0); in ceph_msgr_init() 281 get_page(zero_page); in ceph_msgr_init() 1656 ret = ceph_tcp_sendpage(con->sock, zero_page, 0, size, true); in write_partial_skip()
|
/Linux-v4.19/arch/arm/mm/ |
D | mmu.c | 1624 void *zero_page; in paging_init() local 1638 zero_page = early_alloc(PAGE_SIZE); in paging_init() 1642 empty_zero_page = virt_to_page(zero_page); in paging_init()
|
/Linux-v4.19/drivers/nvdimm/ |
D | btt.c | 518 void *zero_page = page_address(ZERO_PAGE(0)); in arena_clear_freelist_error() local 528 ret = arena_write_bytes(arena, nsoff, zero_page, in arena_clear_freelist_error()
|
/Linux-v4.19/drivers/target/ |
D | target_core_transport.c | 2489 bool zero_page, bool chainable) in target_alloc_sgl() argument 2491 gfp_t gfp = GFP_KERNEL | (zero_page ? __GFP_ZERO : 0); in target_alloc_sgl()
|
/Linux-v4.19/virt/kvm/ |
D | kvm_main.c | 2022 const void *zero_page = (const void *) __va(page_to_phys(ZERO_PAGE(0))); in kvm_clear_guest_page() local 2024 return kvm_write_guest_page(kvm, gfn, zero_page, offset, len); in kvm_clear_guest_page()
|