Searched refs:alloc_page (Results 1 – 25 of 223) sorted by relevance
123456789
52 lock_op_page = alloc_page(GFP_NOIO); in ceph_cls_lock()111 unlock_op_page = alloc_page(GFP_NOIO); in ceph_cls_unlock()164 break_op_page = alloc_page(GFP_NOIO); in ceph_cls_break_lock()213 cookie_op_page = alloc_page(GFP_NOIO); in ceph_cls_set_cookie()352 get_info_op_page = alloc_page(GFP_NOIO); in ceph_cls_lock_info()356 reply_page = alloc_page(GFP_NOIO); in ceph_cls_lock_info()
80 page = alloc_page(gfp); in rds_page_remainder_alloc()122 page = alloc_page(gfp); in rds_page_remainder_alloc()
33 um_vdso = alloc_page(GFP_KERNEL); in init_vdso()
236 src_page = alloc_page(GFP_KERNEL);239 dst_page = alloc_page(GFP_KERNEL);
24 vdso_page = alloc_page(GFP_KERNEL); in vdso_init()
124 out_page = alloc_page(GFP_NOFS); in zlib_compress_pages()203 out_page = alloc_page(GFP_NOFS); in zlib_compress_pages()239 out_page = alloc_page(GFP_NOFS); in zlib_compress_pages()
412 out_page = alloc_page(GFP_NOFS); in zstd_compress_pages()457 out_page = alloc_page(GFP_NOFS); in zstd_compress_pages()514 out_page = alloc_page(GFP_NOFS); in zstd_compress_pages()
155 cur_page = alloc_page(GFP_NOFS); in copy_compressed_data_to_page()181 cur_page = alloc_page(GFP_NOFS); in copy_compressed_data_to_page()
49 selinux_state.status_page = alloc_page(GFP_KERNEL|__GFP_ZERO); in selinux_kernel_status_page()
32 p = alloc_page(GFP_KERNEL | __GFP_ZERO); in kvm_arch_ptp_init()
214 io->header_page = alloc_page(gfp_mask); in ppl_io_pool_alloc()804 page1 = alloc_page(GFP_KERNEL); in ppl_recover_entry()805 page2 = alloc_page(GFP_KERNEL); in ppl_recover_entry()977 page = alloc_page(GFP_KERNEL); in ppl_recover()1051 page = alloc_page(GFP_NOIO | __GFP_ZERO); in ppl_write_empty_header()1088 page = alloc_page(GFP_KERNEL); in ppl_load_distributed()1092 page2 = alloc_page(GFP_KERNEL); in ppl_load_distributed()
49 rp->pages[i] = alloc_page(gfp_flags); in resync_alloc_pages()
158 p = alloc_page(GFP_KERNEL); in do_mount_root()187 struct page *page = alloc_page(GFP_KERNEL); in mount_root_generic()
167 pd->p = alloc_page(GFP_DMA32); in psb_mmu_alloc_pd()170 pd->dummy_pt = alloc_page(GFP_DMA32); in psb_mmu_alloc_pd()173 pd->dummy_page = alloc_page(GFP_DMA32); in psb_mmu_alloc_pd()274 pt->p = alloc_page(GFP_DMA32); in psb_mmu_alloc_pt()
154 struct page *page = alloc_page(GFP_KERNEL); in gve_prefill_rx_pages()459 bool alloc_page = false; in gve_rx_copy_to_pool() local471 alloc_page = !recycle; in gve_rx_copy_to_pool()474 if (alloc_page) { in gve_rx_copy_to_pool()484 page = alloc_page(GFP_ATOMIC); in gve_rx_copy_to_pool()
54 page = alloc_page(GFP | __GFP_HIGHMEM); in huge_get_pages()
112 mock->pages[i] = alloc_page(GFP_KERNEL); in mock_dmabuf()
275 cp = alloc_page(GFP_KERNEL); in init_vdso_image()296 dp = alloc_page(GFP_KERNEL); in init_vdso_image()
373 if (trans_pcie->alloc_page) { in iwl_pcie_rx_alloc_page()376 if (trans_pcie->alloc_page) { in iwl_pcie_rx_alloc_page()378 page = trans_pcie->alloc_page; in iwl_pcie_rx_alloc_page()381 trans_pcie->alloc_page = NULL; in iwl_pcie_rx_alloc_page()408 if (!trans_pcie->alloc_page) { in iwl_pcie_rx_alloc_page()410 trans_pcie->alloc_page = page; in iwl_pcie_rx_alloc_page()1242 if (trans_pcie->alloc_page) in iwl_pcie_rx_free()1243 __free_pages(trans_pcie->alloc_page, trans_pcie->rx_page_order); in iwl_pcie_rx_free()
148 vma_priv->pages[i] = alloc_page(GFP_KERNEL | __GFP_ZERO); in privcmd_buf_mmap()
126 struct page *page = alloc_page(balloon_mapping_gfp_mask() | in balloon_page_alloc()
48 page = alloc_page(GFP_KERNEL | __GFP_ZERO); in swap_cgroup_prepare()
27 page = alloc_page(GFP_KERNEL); in dump_rsc()
238 sg_pages[i] = alloc_page(GFP_KERNEL); in efi_capsule_update()
243 page = alloc_page(GFP_KERNEL); in check_writebuffer_bugs()