Lines Matching refs:gfp_mask

2798 noinline bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order)  in should_fail_alloc_page()  argument
2800 return __should_fail_alloc_page(gfp_mask, order); in should_fail_alloc_page()
2922 unsigned int alloc_flags, gfp_t gfp_mask) in zone_watermark_fast() argument
3001 alloc_flags_nofragment(struct zone *zone, gfp_t gfp_mask) in alloc_flags_nofragment() argument
3009 alloc_flags = (__force int) (gfp_mask & __GFP_KSWAPD_RECLAIM); in alloc_flags_nofragment()
3033 static inline unsigned int gfp_to_alloc_flags_cma(gfp_t gfp_mask, in gfp_to_alloc_flags_cma() argument
3037 if (gfp_migratetype(gfp_mask) == MIGRATE_MOVABLE) in gfp_to_alloc_flags_cma()
3048 get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags, in get_page_from_freelist() argument
3071 !__cpuset_zone_allowed(zone, gfp_mask)) in get_page_from_freelist()
3121 gfp_mask)) { in get_page_from_freelist()
3148 ret = node_reclaim(zone->zone_pgdat, gfp_mask, order); in get_page_from_freelist()
3168 gfp_mask, alloc_flags, ac->migratetype); in get_page_from_freelist()
3170 prep_new_page(page, order, gfp_mask, alloc_flags); in get_page_from_freelist()
3208 static void warn_alloc_show_mem(gfp_t gfp_mask, nodemask_t *nodemask) in warn_alloc_show_mem() argument
3217 if (!(gfp_mask & __GFP_NOMEMALLOC)) in warn_alloc_show_mem()
3221 if (!in_task() || !(gfp_mask & __GFP_DIRECT_RECLAIM)) in warn_alloc_show_mem()
3224 __show_mem(filter, nodemask, gfp_zone(gfp_mask)); in warn_alloc_show_mem()
3227 void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...) in warn_alloc() argument
3233 if ((gfp_mask & __GFP_NOWARN) || in warn_alloc()
3235 ((gfp_mask & __GFP_DMA) && !has_managed_dma())) in warn_alloc()
3242 current->comm, &vaf, gfp_mask, &gfp_mask, in warn_alloc()
3249 warn_alloc_show_mem(gfp_mask, nodemask); in warn_alloc()
3253 __alloc_pages_cpuset_fallback(gfp_t gfp_mask, unsigned int order, in __alloc_pages_cpuset_fallback() argument
3259 page = get_page_from_freelist(gfp_mask, order, in __alloc_pages_cpuset_fallback()
3266 page = get_page_from_freelist(gfp_mask, order, in __alloc_pages_cpuset_fallback()
3273 __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order, in __alloc_pages_may_oom() argument
3280 .gfp_mask = gfp_mask, in __alloc_pages_may_oom()
3304 page = get_page_from_freelist((gfp_mask | __GFP_HARDWALL) & in __alloc_pages_may_oom()
3324 if (gfp_mask & (__GFP_RETRY_MAYFAIL | __GFP_THISNODE)) in __alloc_pages_may_oom()
3343 WARN_ON_ONCE_GFP(gfp_mask & __GFP_NOFAIL, gfp_mask)) { in __alloc_pages_may_oom()
3350 if (gfp_mask & __GFP_NOFAIL) in __alloc_pages_may_oom()
3351 page = __alloc_pages_cpuset_fallback(gfp_mask, order, in __alloc_pages_may_oom()
3368 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, in __alloc_pages_direct_compact() argument
3383 *compact_result = try_to_compact_pages(gfp_mask, order, alloc_flags, ac, in __alloc_pages_direct_compact()
3400 prep_new_page(page, order, gfp_mask, alloc_flags); in __alloc_pages_direct_compact()
3404 page = get_page_from_freelist(gfp_mask, order, alloc_flags, ac); in __alloc_pages_direct_compact()
3493 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, in __alloc_pages_direct_compact() argument
3533 static bool __need_reclaim(gfp_t gfp_mask) in __need_reclaim() argument
3536 if (!(gfp_mask & __GFP_DIRECT_RECLAIM)) in __need_reclaim()
3543 if (gfp_mask & __GFP_NOLOCKDEP) in __need_reclaim()
3559 void fs_reclaim_acquire(gfp_t gfp_mask) in fs_reclaim_acquire() argument
3561 gfp_mask = current_gfp_context(gfp_mask); in fs_reclaim_acquire()
3563 if (__need_reclaim(gfp_mask)) { in fs_reclaim_acquire()
3564 if (gfp_mask & __GFP_FS) in fs_reclaim_acquire()
3576 void fs_reclaim_release(gfp_t gfp_mask) in fs_reclaim_release() argument
3578 gfp_mask = current_gfp_context(gfp_mask); in fs_reclaim_release()
3580 if (__need_reclaim(gfp_mask)) { in fs_reclaim_release()
3581 if (gfp_mask & __GFP_FS) in fs_reclaim_release()
3614 __perform_reclaim(gfp_t gfp_mask, unsigned int order, in __perform_reclaim() argument
3624 fs_reclaim_acquire(gfp_mask); in __perform_reclaim()
3627 progress = try_to_free_pages(ac->zonelist, order, gfp_mask, in __perform_reclaim()
3631 fs_reclaim_release(gfp_mask); in __perform_reclaim()
3640 __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, in __alloc_pages_direct_reclaim() argument
3649 *did_some_progress = __perform_reclaim(gfp_mask, order, ac); in __alloc_pages_direct_reclaim()
3654 page = get_page_from_freelist(gfp_mask, order, alloc_flags, ac); in __alloc_pages_direct_reclaim()
3673 static void wake_all_kswapds(unsigned int order, gfp_t gfp_mask, in wake_all_kswapds() argument
3686 wakeup_kswapd(zone, gfp_mask, order, highest_zoneidx); in wake_all_kswapds()
3693 gfp_to_alloc_flags(gfp_t gfp_mask, unsigned int order) in gfp_to_alloc_flags() argument
3712 (gfp_mask & (__GFP_HIGH | __GFP_KSWAPD_RECLAIM)); in gfp_to_alloc_flags()
3714 if (!(gfp_mask & __GFP_DIRECT_RECLAIM)) { in gfp_to_alloc_flags()
3719 if (!(gfp_mask & __GFP_NOMEMALLOC)) { in gfp_to_alloc_flags()
3736 alloc_flags = gfp_to_alloc_flags_cma(gfp_mask, alloc_flags); in gfp_to_alloc_flags()
3760 static inline int __gfp_pfmemalloc_flags(gfp_t gfp_mask) in __gfp_pfmemalloc_flags() argument
3762 if (unlikely(gfp_mask & __GFP_NOMEMALLOC)) in __gfp_pfmemalloc_flags()
3764 if (gfp_mask & __GFP_MEMALLOC) in __gfp_pfmemalloc_flags()
3778 bool gfp_pfmemalloc_allowed(gfp_t gfp_mask) in gfp_pfmemalloc_allowed() argument
3780 return !!__gfp_pfmemalloc_flags(gfp_mask); in gfp_pfmemalloc_allowed()
3794 should_reclaim_retry(gfp_t gfp_mask, unsigned order, in should_reclaim_retry() argument
3899 __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, in __alloc_pages_slowpath() argument
3902 bool can_direct_reclaim = gfp_mask & __GFP_DIRECT_RECLAIM; in __alloc_pages_slowpath()
3927 alloc_flags = gfp_to_alloc_flags(gfp_mask, order); in __alloc_pages_slowpath()
3945 if (cpusets_insane_config() && (gfp_mask & __GFP_HARDWALL)) { in __alloc_pages_slowpath()
3954 wake_all_kswapds(order, gfp_mask, ac); in __alloc_pages_slowpath()
3960 page = get_page_from_freelist(gfp_mask, order, alloc_flags, ac); in __alloc_pages_slowpath()
3976 && !gfp_pfmemalloc_allowed(gfp_mask)) { in __alloc_pages_slowpath()
3977 page = __alloc_pages_direct_compact(gfp_mask, order, in __alloc_pages_slowpath()
3988 if (costly_order && (gfp_mask & __GFP_NORETRY)) { in __alloc_pages_slowpath()
4022 wake_all_kswapds(order, gfp_mask, ac); in __alloc_pages_slowpath()
4024 reserve_flags = __gfp_pfmemalloc_flags(gfp_mask); in __alloc_pages_slowpath()
4026 alloc_flags = gfp_to_alloc_flags_cma(gfp_mask, reserve_flags) | in __alloc_pages_slowpath()
4041 page = get_page_from_freelist(gfp_mask, order, alloc_flags, ac); in __alloc_pages_slowpath()
4054 page = __alloc_pages_direct_reclaim(gfp_mask, order, alloc_flags, ac, in __alloc_pages_slowpath()
4060 page = __alloc_pages_direct_compact(gfp_mask, order, alloc_flags, ac, in __alloc_pages_slowpath()
4066 if (gfp_mask & __GFP_NORETRY) in __alloc_pages_slowpath()
4073 if (costly_order && !(gfp_mask & __GFP_RETRY_MAYFAIL)) in __alloc_pages_slowpath()
4076 if (should_reclaim_retry(gfp_mask, order, ac, alloc_flags, in __alloc_pages_slowpath()
4102 page = __alloc_pages_may_oom(gfp_mask, order, ac, &did_some_progress); in __alloc_pages_slowpath()
4109 (gfp_mask & __GFP_NOMEMALLOC))) in __alloc_pages_slowpath()
4131 if (gfp_mask & __GFP_NOFAIL) { in __alloc_pages_slowpath()
4136 if (WARN_ON_ONCE_GFP(!can_direct_reclaim, gfp_mask)) in __alloc_pages_slowpath()
4144 WARN_ON_ONCE_GFP(current->flags & PF_MEMALLOC, gfp_mask); in __alloc_pages_slowpath()
4152 WARN_ON_ONCE_GFP(costly_order, gfp_mask); in __alloc_pages_slowpath()
4161 page = __alloc_pages_cpuset_fallback(gfp_mask, order, ALLOC_MIN_RESERVE, ac); in __alloc_pages_slowpath()
4169 warn_alloc(gfp_mask, ac->nodemask, in __alloc_pages_slowpath()
4175 static inline bool prepare_alloc_pages(gfp_t gfp_mask, unsigned int order, in prepare_alloc_pages() argument
4180 ac->highest_zoneidx = gfp_zone(gfp_mask); in prepare_alloc_pages()
4181 ac->zonelist = node_zonelist(preferred_nid, gfp_mask); in prepare_alloc_pages()
4183 ac->migratetype = gfp_migratetype(gfp_mask); in prepare_alloc_pages()
4197 might_alloc(gfp_mask); in prepare_alloc_pages()
4199 if (should_fail_alloc_page(gfp_mask, order)) in prepare_alloc_pages()
4202 *alloc_flags = gfp_to_alloc_flags_cma(gfp_mask, *alloc_flags); in prepare_alloc_pages()
4205 ac->spread_dirty_pages = (gfp_mask & __GFP_WRITE); in prepare_alloc_pages()
4473 unsigned long __get_free_pages(gfp_t gfp_mask, unsigned int order) in __get_free_pages() argument
4477 page = alloc_pages(gfp_mask & ~__GFP_HIGHMEM, order); in __get_free_pages()
4484 unsigned long get_zeroed_page(gfp_t gfp_mask) in get_zeroed_page() argument
4486 return __get_free_page(gfp_mask | __GFP_ZERO); in get_zeroed_page()
4545 gfp_t gfp_mask) in __page_frag_cache_refill() argument
4548 gfp_t gfp = gfp_mask; in __page_frag_cache_refill()
4551 gfp_mask |= __GFP_COMP | __GFP_NOWARN | __GFP_NORETRY | in __page_frag_cache_refill()
4553 page = alloc_pages_node(NUMA_NO_NODE, gfp_mask, in __page_frag_cache_refill()
4575 unsigned int fragsz, gfp_t gfp_mask, in page_frag_alloc_align() argument
4584 page = __page_frag_cache_refill(nc, gfp_mask); in page_frag_alloc_align()
4694 void *alloc_pages_exact(size_t size, gfp_t gfp_mask) in alloc_pages_exact() argument
4699 if (WARN_ON_ONCE(gfp_mask & (__GFP_COMP | __GFP_HIGHMEM))) in alloc_pages_exact()
4700 gfp_mask &= ~(__GFP_COMP | __GFP_HIGHMEM); in alloc_pages_exact()
4702 addr = __get_free_pages(gfp_mask, order); in alloc_pages_exact()
4719 void * __meminit alloc_pages_exact_nid(int nid, size_t size, gfp_t gfp_mask) in alloc_pages_exact_nid() argument
4724 if (WARN_ON_ONCE(gfp_mask & (__GFP_COMP | __GFP_HIGHMEM))) in alloc_pages_exact_nid()
4725 gfp_mask &= ~(__GFP_COMP | __GFP_HIGHMEM); in alloc_pages_exact_nid()
4727 p = alloc_pages_node(nid, gfp_mask, order); in alloc_pages_exact_nid()
6040 .gfp_mask = GFP_USER | __GFP_MOVABLE | __GFP_RETRY_MAYFAIL, in __alloc_contig_migrate_range()
6080 if (!(cc->gfp_mask & __GFP_NOWARN) && ret == -EBUSY) in __alloc_contig_migrate_range()
6110 unsigned migratetype, gfp_t gfp_mask) in alloc_contig_range() argument
6123 .gfp_mask = current_gfp_context(gfp_mask), in alloc_contig_range()
6149 ret = start_isolate_page_range(start, end, migratetype, 0, gfp_mask); in alloc_contig_range()
6236 unsigned long nr_pages, gfp_t gfp_mask) in __alloc_contig_pages() argument
6241 gfp_mask); in __alloc_contig_pages()
6296 struct page *alloc_contig_pages(unsigned long nr_pages, gfp_t gfp_mask, in alloc_contig_pages() argument
6304 zonelist = node_zonelist(nid, gfp_mask); in alloc_contig_pages()
6306 gfp_zone(gfp_mask), nodemask) { in alloc_contig_pages()
6321 gfp_mask); in alloc_contig_pages()