Searched refs:gfpflags_allow_blocking (Results 1 – 25 of 34) sorted by relevance
12
339 might_sleep_if(gfpflags_allow_blocking(gfp)); in __i915_sw_fence_await_sw_fence()356 if (!gfpflags_allow_blocking(gfp)) in __i915_sw_fence_await_sw_fence()475 might_sleep_if(gfpflags_allow_blocking(gfp)); in i915_sw_fence_await_dma_fence()487 if (!gfpflags_allow_blocking(gfp)) in i915_sw_fence_await_dma_fence()574 might_sleep_if(gfpflags_allow_blocking(gfp)); in i915_sw_fence_await_reservation()
739 if (!gfpflags_allow_blocking(gfp)) { in request_alloc_slow()794 might_sleep_if(gfpflags_allow_blocking(gfp)); in __i915_request_create()
28 static inline bool gfpflags_allow_blocking(const gfp_t gfp_flags) in gfpflags_allow_blocking() function
167 !gfpflags_allow_blocking(gfp) && in dma_direct_alloc()286 force_dma_unencrypted(dev) && !gfpflags_allow_blocking(gfp)) in dma_direct_alloc_pages()
311 if (!gfpflags_allow_blocking(gfp)) in dma_alloc_contiguous()
116 !gfpflags_allow_blocking(gfp_mask)); in cn_netlink_send_mult()
441 if (gfpflags_allow_blocking(flags)) in __kasan_kmalloc()486 if (gfpflags_allow_blocking(flags)) in kasan_kmalloc_large()
243 if (!gfpflags_allow_blocking(gfp_mask) && !in_interrupt()) { in radix_tree_node_alloc()366 WARN_ON_ONCE(!gfpflags_allow_blocking(gfp_mask)); in radix_tree_preload()378 if (gfpflags_allow_blocking(gfp_mask)) in radix_tree_maybe_preload()
500 (alloc_pages_nr && gfpflags_allow_blocking(alloc_pages_gfp))) && in test_lockup_init()
335 if (gfpflags_allow_blocking(gfp)) { in __xas_nomem()
308 might_sleep_if(gfpflags_allow_blocking(gfp_flags)); in get_task_io_context()
332 static inline bool gfpflags_allow_blocking(const gfp_t gfp_flags) in gfpflags_allow_blocking() function
1634 might_sleep_if(gfpflags_allow_blocking(pri)); in skb_unclone()1663 might_sleep_if(gfpflags_allow_blocking(pri)); in skb_header_unclone()1709 might_sleep_if(gfpflags_allow_blocking(pri)); in skb_share_check()1745 might_sleep_if(gfpflags_allow_blocking(pri)); in skb_unshare()
1032 if (IS_ENABLED(CONFIG_DMA_REMAP) && gfpflags_allow_blocking(gfp) && in iommu_dma_alloc()1039 !gfpflags_allow_blocking(gfp) && !coherent) in iommu_dma_alloc()1061 if (!gfpflags_allow_blocking(gfp)) { in iommu_dma_alloc_noncoherent()
322 might_sleep_if(gfpflags_allow_blocking(mem_flags)); in dma_pool_alloc()
516 might_sleep_if(gfpflags_allow_blocking(flags)); in slab_pre_alloc_hook()
2585 if (gfpflags_allow_blocking(local_flags)) in cache_grow_begin()2627 if (gfpflags_allow_blocking(local_flags)) in cache_grow_begin()2635 if (gfpflags_allow_blocking(local_flags)) in cache_grow_begin()2986 might_sleep_if(gfpflags_allow_blocking(flags)); in cache_alloc_debugcheck_before()
583 might_sleep_if(gfpflags_allow_blocking(gfp)); in wb_get_create()
1248 if (gfpflags_allow_blocking(gfp_mask)) { in alloc_vmap_area()2503 if (gfpflags_allow_blocking(gfp_mask)) in __vmalloc_area_node()
1098 bool can_sleep = gfpflags_allow_blocking(gfp); in z3fold_alloc()
722 if (!prealloc && gfpflags_allow_blocking(mask)) { in __clear_extent_bit()839 if (gfpflags_allow_blocking(mask)) in __clear_extent_bit()984 if (!prealloc && gfpflags_allow_blocking(mask)) { in __set_extent_bit()1169 if (gfpflags_allow_blocking(mask)) in __set_extent_bit()4444 if (gfpflags_allow_blocking(mask) && in try_release_extent_mapping()
649 if (!gfpflags_allow_blocking(gfp)) { in rxrpc_wait_for_channel()
709 allowblock = gfpflags_allow_blocking(gfp); in __dma_alloc()1433 if (coherent_flag == COHERENT || !gfpflags_allow_blocking(gfp)) in __arm_iommu_alloc_attrs()
1618 bool preload = gfpflags_allow_blocking(gfp); in sctp_assoc_set_id()
1839 if (gfpflags_allow_blocking(gfp_mask) && (stime > 0)) { in audit_log_start()