Lines Matching refs:gfp

79 		gfp_t gfp)  in __dma_direct_alloc_pages()  argument
87 gfp |= dma_direct_optimal_gfp_mask(dev, dev->coherent_dma_mask, in __dma_direct_alloc_pages()
89 page = dma_alloc_contiguous(dev, size, gfp); in __dma_direct_alloc_pages()
96 page = alloc_pages_node(node, gfp, get_order(size)); in __dma_direct_alloc_pages()
103 !(gfp & (GFP_DMA32 | GFP_DMA))) { in __dma_direct_alloc_pages()
104 gfp |= GFP_DMA32; in __dma_direct_alloc_pages()
108 if (IS_ENABLED(CONFIG_ZONE_DMA) && !(gfp & GFP_DMA)) { in __dma_direct_alloc_pages()
109 gfp = (gfp & ~GFP_DMA32) | GFP_DMA; in __dma_direct_alloc_pages()
118 dma_addr_t *dma_handle, gfp_t gfp) in dma_direct_alloc_from_pool() argument
124 gfp |= dma_direct_optimal_gfp_mask(dev, dev->coherent_dma_mask, in dma_direct_alloc_from_pool()
126 page = dma_alloc_from_pool(dev, size, &ret, gfp, dma_coherent_ok); in dma_direct_alloc_from_pool()
134 dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) in dma_direct_alloc() argument
142 gfp |= __GFP_NOWARN; in dma_direct_alloc()
146 page = __dma_direct_alloc_pages(dev, size, gfp & ~__GFP_ZERO); in dma_direct_alloc()
160 return arch_dma_alloc(dev, size, dma_handle, gfp, attrs); in dma_direct_alloc()
167 !gfpflags_allow_blocking(gfp) && in dma_direct_alloc()
170 return dma_direct_alloc_from_pool(dev, size, dma_handle, gfp); in dma_direct_alloc()
173 page = __dma_direct_alloc_pages(dev, size, gfp & ~__GFP_ZERO); in dma_direct_alloc()
280 dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp) in dma_direct_alloc_pages() argument
286 force_dma_unencrypted(dev) && !gfpflags_allow_blocking(gfp)) in dma_direct_alloc_pages()
287 return dma_direct_alloc_from_pool(dev, size, dma_handle, gfp); in dma_direct_alloc_pages()
289 page = __dma_direct_alloc_pages(dev, size, gfp); in dma_direct_alloc_pages()