/Linux-v4.19/include/linux/ |
D | gfp.h | 467 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid, 471 __alloc_pages(gfp_t gfp_mask, unsigned int order, int preferred_nid) in __alloc_pages() argument 473 return __alloc_pages_nodemask(gfp_mask, order, preferred_nid, NULL); in __alloc_pages() 481 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument 484 VM_WARN_ON((gfp_mask & __GFP_THISNODE) && !node_online(nid)); in __alloc_pages_node() 486 return __alloc_pages(gfp_mask, order, nid); in __alloc_pages_node() 494 static inline struct page *alloc_pages_node(int nid, gfp_t gfp_mask, in alloc_pages_node() argument 500 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node() 504 extern struct page *alloc_pages_current(gfp_t gfp_mask, unsigned order); 507 alloc_pages(gfp_t gfp_mask, unsigned int order) in alloc_pages() argument [all …]
|
D | cpuset.h | 65 extern bool __cpuset_node_allowed(int node, gfp_t gfp_mask); 67 static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) in cpuset_node_allowed() argument 70 return __cpuset_node_allowed(node, gfp_mask); in cpuset_node_allowed() 74 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() argument 76 return __cpuset_node_allowed(zone_to_nid(z), gfp_mask); in __cpuset_zone_allowed() 79 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() argument 82 return __cpuset_zone_allowed(z, gfp_mask); in cpuset_zone_allowed() 202 static inline bool cpuset_node_allowed(int node, gfp_t gfp_mask) in cpuset_node_allowed() argument 207 static inline bool __cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in __cpuset_zone_allowed() argument 212 static inline bool cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() argument
|
D | mempool.h | 13 typedef void * (mempool_alloc_t)(gfp_t gfp_mask, void *pool_data); 36 gfp_t gfp_mask, int node_id); 44 gfp_t gfp_mask, int nid); 48 extern void *mempool_alloc(mempool_t *pool, gfp_t gfp_mask) __malloc; 56 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data); 77 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data); 96 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data);
|
D | pagemap.h | 102 return mapping->gfp_mask; in mapping_gfp_mask() 107 gfp_t gfp_mask) in mapping_gfp_constraint() argument 109 return mapping_gfp_mask(mapping) & gfp_mask; in mapping_gfp_constraint() 118 m->gfp_mask = mask; in mapping_set_gfp_mask() 320 pgoff_t offset, gfp_t gfp_mask) in find_or_create_page() argument 324 gfp_mask); in find_or_create_page() 394 pgoff_t index, gfp_t gfp_mask); 617 pgoff_t index, gfp_t gfp_mask); 619 pgoff_t index, gfp_t gfp_mask); 622 int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask); [all …]
|
D | page_owner.h | 13 unsigned int order, gfp_t gfp_mask); 28 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument 31 __set_page_owner(page, order, gfp_mask); in set_page_owner() 59 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument
|
D | swap.h | 346 gfp_t gfp_mask, nodemask_t *mask); 350 gfp_t gfp_mask, 353 gfp_t gfp_mask, bool noswap, 508 static inline int add_swap_count_continuation(swp_entry_t swp, gfp_t gfp_mask) in add_swap_count_continuation() argument 531 gfp_t gfp_mask, struct vm_fault *vmf) in swap_cluster_readahead() argument 536 static inline struct page *swapin_readahead(swp_entry_t swp, gfp_t gfp_mask, in swapin_readahead() argument 560 gfp_t gfp_mask) in add_to_swap_cache() argument 641 gfp_t gfp_mask); 644 int node, gfp_t gfp_mask) in mem_cgroup_throttle_swaprate() argument
|
/Linux-v4.19/block/ |
D | blk-lib.c | 27 sector_t nr_sects, gfp_t gfp_mask, int flags, in __blkdev_issue_discard() argument 66 bio = next_bio(bio, 0, gfp_mask); in __blkdev_issue_discard() 109 sector_t nr_sects, gfp_t gfp_mask, unsigned long flags) in blkdev_issue_discard() argument 116 ret = __blkdev_issue_discard(bdev, sector, nr_sects, gfp_mask, flags, in blkdev_issue_discard() 143 sector_t nr_sects, gfp_t gfp_mask, struct page *page, in __blkdev_issue_write_same() argument 168 bio = next_bio(bio, 1, gfp_mask); in __blkdev_issue_write_same() 204 sector_t nr_sects, gfp_t gfp_mask, in blkdev_issue_write_same() argument 212 ret = __blkdev_issue_write_same(bdev, sector, nr_sects, gfp_mask, page, in blkdev_issue_write_same() 224 sector_t sector, sector_t nr_sects, gfp_t gfp_mask, in __blkdev_issue_write_zeroes() argument 244 bio = next_bio(bio, 0, gfp_mask); in __blkdev_issue_write_zeroes() [all …]
|
D | blk-map.c | 60 gfp_t gfp_mask, bool copy) in __blk_rq_map_user_iov() argument 67 bio = bio_copy_user_iov(q, map_data, iter, gfp_mask); in __blk_rq_map_user_iov() 69 bio = bio_map_user_iov(q, iter, gfp_mask); in __blk_rq_map_user_iov() 116 const struct iov_iter *iter, gfp_t gfp_mask) in blk_rq_map_user_iov() argument 136 ret =__blk_rq_map_user_iov(rq, map_data, &i, gfp_mask, copy); in blk_rq_map_user_iov() 157 unsigned long len, gfp_t gfp_mask) in blk_rq_map_user() argument 166 return blk_rq_map_user_iov(q, rq, map_data, &i, gfp_mask); in blk_rq_map_user() 216 unsigned int len, gfp_t gfp_mask) in blk_rq_map_kern() argument 231 bio = bio_copy_kern(q, kbuf, len, gfp_mask, reading); in blk_rq_map_kern() 233 bio = bio_map_kern(q, kbuf, len, gfp_mask); in blk_rq_map_kern()
|
D | bio.c | 181 struct bio_vec *bvec_alloc(gfp_t gfp_mask, int nr, unsigned long *idx, in bvec_alloc() argument 218 bvl = mempool_alloc(pool, gfp_mask); in bvec_alloc() 221 gfp_t __gfp_mask = gfp_mask & ~(__GFP_DIRECT_RECLAIM | __GFP_IO); in bvec_alloc() 235 if (unlikely(!bvl && (gfp_mask & __GFP_DIRECT_RECLAIM))) { in bvec_alloc() 438 struct bio *bio_alloc_bioset(gfp_t gfp_mask, unsigned int nr_iovecs, in bio_alloc_bioset() argument 441 gfp_t saved_gfp = gfp_mask; in bio_alloc_bioset() 454 gfp_mask); in bio_alloc_bioset() 487 gfp_mask &= ~__GFP_DIRECT_RECLAIM; in bio_alloc_bioset() 489 p = mempool_alloc(&bs->bio_pool, gfp_mask); in bio_alloc_bioset() 490 if (!p && gfp_mask != saved_gfp) { in bio_alloc_bioset() [all …]
|
/Linux-v4.19/mm/ |
D | mempool.c | 182 gfp_t gfp_mask, int node_id) in mempool_init_node() argument 192 gfp_mask, node_id); in mempool_init_node() 202 element = pool->alloc(gfp_mask, pool->pool_data); in mempool_init_node() 259 gfp_t gfp_mask, int node_id) in mempool_create_node() argument 263 pool = kzalloc_node(sizeof(*pool), gfp_mask, node_id); in mempool_create_node() 268 gfp_mask, node_id)) { in mempool_create_node() 367 void *mempool_alloc(mempool_t *pool, gfp_t gfp_mask) in mempool_alloc() argument 374 VM_WARN_ON_ONCE(gfp_mask & __GFP_ZERO); in mempool_alloc() 375 might_sleep_if(gfp_mask & __GFP_DIRECT_RECLAIM); in mempool_alloc() 377 gfp_mask |= __GFP_NOMEMALLOC; /* don't allocate emergency reserves */ in mempool_alloc() [all …]
|
D | page_owner.c | 25 gfp_t gfp_mask; member 173 depot_stack_handle_t handle, unsigned int order, gfp_t gfp_mask) in __set_page_owner_handle() argument 180 page_owner->gfp_mask = gfp_mask; in __set_page_owner_handle() 187 gfp_t gfp_mask) in __set_page_owner() argument 195 handle = save_stack(gfp_mask); in __set_page_owner() 196 __set_page_owner_handle(page_ext, handle, order, gfp_mask); in __set_page_owner() 238 new_page_owner->gfp_mask = old_page_owner->gfp_mask; in __copy_page_owner() 317 page_owner->gfp_mask); in pagetypeinfo_showmixedcount_print() 360 page_owner->order, page_owner->gfp_mask, in print_page_owner() 361 &page_owner->gfp_mask); in print_page_owner() [all …]
|
D | page_alloc.c | 3045 static bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) in should_fail_alloc_page() argument 3049 if (gfp_mask & __GFP_NOFAIL) in should_fail_alloc_page() 3051 if (fail_page_alloc.ignore_gfp_highmem && (gfp_mask & __GFP_HIGHMEM)) in should_fail_alloc_page() 3054 (gfp_mask & __GFP_DIRECT_RECLAIM)) in should_fail_alloc_page() 3095 static inline bool should_fail_alloc_page(gfp_t gfp_mask, unsigned int order) in should_fail_alloc_page() argument 3250 get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags, in get_page_from_freelist() argument 3268 !__cpuset_zone_allowed(zone, gfp_mask)) in get_page_from_freelist() 3323 ret = node_reclaim(zone->zone_pgdat, gfp_mask, order); in get_page_from_freelist() 3343 gfp_mask, alloc_flags, ac->migratetype); in get_page_from_freelist() 3345 prep_new_page(page, order, gfp_mask, alloc_flags); in get_page_from_freelist() [all …]
|
D | vmscan.c | 112 gfp_t gfp_mask; member 579 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid, in shrink_slab_memcg() argument 599 .gfp_mask = gfp_mask, in shrink_slab_memcg() 649 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid, in shrink_slab_memcg() argument 676 static unsigned long shrink_slab(gfp_t gfp_mask, int nid, in shrink_slab() argument 684 return shrink_slab_memcg(gfp_mask, nid, memcg, priority); in shrink_slab() 691 .gfp_mask = gfp_mask, in shrink_slab() 1147 may_enter_fs = (sc->gfp_mask & __GFP_FS) || in shrink_page_list() 1148 (PageSwapCache(page) && (sc->gfp_mask & __GFP_IO)); in shrink_page_list() 1274 if (!(sc->gfp_mask & __GFP_IO)) in shrink_page_list() [all …]
|
D | swap_state.c | 160 int add_to_swap_cache(struct page *page, swp_entry_t entry, gfp_t gfp_mask) in add_to_swap_cache() argument 164 error = radix_tree_maybe_preload_order(gfp_mask, compound_order(page)); in add_to_swap_cache() 377 struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, in __read_swap_cache_async() argument 411 new_page = alloc_page_vma(gfp_mask, vma, addr); in __read_swap_cache_async() 419 err = radix_tree_maybe_preload(gfp_mask & GFP_KERNEL); in __read_swap_cache_async() 475 struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, in read_swap_cache_async() argument 479 struct page *retpage = __read_swap_cache_async(entry, gfp_mask, in read_swap_cache_async() 566 struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask, in swap_cluster_readahead() argument 598 gfp_mask, vma, addr, &page_allocated); in swap_cluster_readahead() 614 return read_swap_cache_async(entry, gfp_mask, vma, addr, do_poll); in swap_cluster_readahead() [all …]
|
/Linux-v4.19/fs/btrfs/ |
D | ulist.h | 48 struct ulist *ulist_alloc(gfp_t gfp_mask); 50 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask); 52 u64 *old_aux, gfp_t gfp_mask); 57 void **old_aux, gfp_t gfp_mask) in ulist_add_merge_ptr() argument 61 int ret = ulist_add_merge(ulist, val, (uintptr_t)aux, &old64, gfp_mask); in ulist_add_merge_ptr() 65 return ulist_add_merge(ulist, val, (u64)aux, (u64 *)old_aux, gfp_mask); in ulist_add_merge_ptr()
|
D | ulist.c | 92 struct ulist *ulist_alloc(gfp_t gfp_mask) in ulist_alloc() argument 94 struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); in ulist_alloc() 186 int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) in ulist_add() argument 188 return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); in ulist_add() 192 u64 *old_aux, gfp_t gfp_mask) in ulist_add_merge() argument 203 node = kmalloc(sizeof(*node), gfp_mask); in ulist_add_merge()
|
/Linux-v4.19/fs/nfs/blocklayout/ |
D | dev.c | 231 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask); 236 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_simple() argument 242 dev = bl_resolve_deviceid(server, v, gfp_mask); in bl_parse_simple() 353 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_scsi() argument 402 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_slice() argument 407 ret = bl_parse_deviceid(server, d, volumes, v->slice.volume, gfp_mask); in bl_parse_slice() 418 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_concat() argument 431 volumes, v->concat.volumes[i], gfp_mask); in bl_parse_concat() 447 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_stripe() argument 460 volumes, v->stripe.volumes[i], gfp_mask); in bl_parse_stripe() [all …]
|
/Linux-v4.19/drivers/staging/android/ion/ |
D | ion_page_pool.c | 16 return alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages() 97 int ion_page_pool_shrink(struct ion_page_pool *pool, gfp_t gfp_mask, in ion_page_pool_shrink() argument 106 high = !!(gfp_mask & __GFP_HIGHMEM); in ion_page_pool_shrink() 131 struct ion_page_pool *ion_page_pool_create(gfp_t gfp_mask, unsigned int order) in ion_page_pool_create() argument 141 pool->gfp_mask = gfp_mask | __GFP_COMP; in ion_page_pool_create()
|
/Linux-v4.19/fs/ntfs/ |
D | malloc.h | 42 static inline void *__ntfs_malloc(unsigned long size, gfp_t gfp_mask) in __ntfs_malloc() argument 47 return kmalloc(PAGE_SIZE, gfp_mask & ~__GFP_HIGHMEM); in __ntfs_malloc() 51 return __vmalloc(size, gfp_mask, PAGE_KERNEL); in __ntfs_malloc()
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx4/ |
D | icm.c | 99 gfp_t gfp_mask, int node) in mlx4_alloc_icm_pages() argument 103 page = alloc_pages_node(node, gfp_mask, order); in mlx4_alloc_icm_pages() 105 page = alloc_pages(gfp_mask, order); in mlx4_alloc_icm_pages() 115 int order, gfp_t gfp_mask) in mlx4_alloc_icm_coherent() argument 118 &sg_dma_address(mem), gfp_mask); in mlx4_alloc_icm_coherent() 134 gfp_t gfp_mask, int coherent) in mlx4_alloc_icm() argument 143 BUG_ON(coherent && (gfp_mask & __GFP_HIGHMEM)); in mlx4_alloc_icm() 146 gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN), in mlx4_alloc_icm() 150 gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN)); in mlx4_alloc_icm() 163 gfp_mask & ~(__GFP_HIGHMEM | in mlx4_alloc_icm() [all …]
|
/Linux-v4.19/net/sunrpc/auth_gss/ |
D | gss_krb5_mech.c | 379 context_derive_keys_des3(struct krb5_ctx *ctx, gfp_t gfp_mask) in context_derive_keys_des3() argument 406 err = krb5_derive_key(ctx->gk5e, &keyin, &keyout, &c, gfp_mask); in context_derive_keys_des3() 499 context_derive_keys_new(struct krb5_ctx *ctx, gfp_t gfp_mask) in context_derive_keys_new() argument 515 err = krb5_derive_key(ctx->gk5e, &keyin, &keyout, &c, gfp_mask); in context_derive_keys_new() 530 err = krb5_derive_key(ctx->gk5e, &keyin, &keyout, &c, gfp_mask); in context_derive_keys_new() 545 err = krb5_derive_key(ctx->gk5e, &keyin, &keyout, &c, gfp_mask); in context_derive_keys_new() 555 err = krb5_derive_key(ctx->gk5e, &keyin, &keyout, &c, gfp_mask); in context_derive_keys_new() 565 err = krb5_derive_key(ctx->gk5e, &keyin, &keyout, &c, gfp_mask); in context_derive_keys_new() 575 err = krb5_derive_key(ctx->gk5e, &keyin, &keyout, &c, gfp_mask); in context_derive_keys_new() 611 gfp_t gfp_mask) in gss_import_v2_context() argument [all …]
|
/Linux-v4.19/drivers/connector/ |
D | connector.c | 75 gfp_t gfp_mask) in cn_netlink_send_mult() argument 109 skb = nlmsg_new(size, gfp_mask); in cn_netlink_send_mult() 127 gfp_mask); in cn_netlink_send_mult() 129 !gfpflags_allow_blocking(gfp_mask)); in cn_netlink_send_mult() 135 gfp_t gfp_mask) in cn_netlink_send() argument 137 return cn_netlink_send_mult(msg, msg->len, portid, __group, gfp_mask); in cn_netlink_send()
|
/Linux-v4.19/lib/ |
D | scatterlist.c | 151 static struct scatterlist *sg_kmalloc(unsigned int nents, gfp_t gfp_mask) in sg_kmalloc() argument 163 void *ptr = (void *) __get_free_page(gfp_mask); in sg_kmalloc() 164 kmemleak_alloc(ptr, PAGE_SIZE, 1, gfp_mask); in sg_kmalloc() 168 gfp_mask); in sg_kmalloc() 265 gfp_t gfp_mask, sg_alloc_fn *alloc_fn) in __sg_alloc_table() argument 296 sg = alloc_fn(alloc_size, gfp_mask); in __sg_alloc_table() 347 int sg_alloc_table(struct sg_table *table, unsigned int nents, gfp_t gfp_mask) in sg_alloc_table() argument 352 NULL, gfp_mask, sg_kmalloc); in sg_alloc_table() 384 gfp_t gfp_mask) in __sg_alloc_table_from_pages() argument 405 ret = sg_alloc_table(sgt, chunks, gfp_mask); in __sg_alloc_table_from_pages() [all …]
|
D | radix-tree.c | 149 return root->gfp_mask & (__GFP_BITS_MASK & ~GFP_ZONEMASK); in root_gfp_mask() 172 root->gfp_mask |= (__force gfp_t)(1 << (tag + ROOT_TAG_SHIFT)); in root_tag_set() 177 root->gfp_mask &= (__force gfp_t)~(1 << (tag + ROOT_TAG_SHIFT)); in root_tag_clear() 182 root->gfp_mask &= (1 << ROOT_TAG_SHIFT) - 1; in root_tag_clear_all() 187 return (__force int)root->gfp_mask & (1 << (tag + ROOT_TAG_SHIFT)); in root_tag_get() 192 return (__force unsigned)root->gfp_mask >> ROOT_TAG_SHIFT; in root_tags_get() 197 return !!(root->gfp_mask & ROOT_IS_IDR); in is_idr() 318 root->gfp_mask >> ROOT_TAG_SHIFT); in radix_tree_dump() 368 root->gfp_mask >> ROOT_TAG_SHIFT); in ida_dump() 378 radix_tree_node_alloc(gfp_t gfp_mask, struct radix_tree_node *parent, in radix_tree_node_alloc() argument [all …]
|
/Linux-v4.19/kernel/power/ |
D | snapshot.c | 158 static void *get_image_page(gfp_t gfp_mask, int safe_needed) in get_image_page() argument 162 res = (void *)get_zeroed_page(gfp_mask); in get_image_page() 168 res = (void *)get_zeroed_page(gfp_mask); in get_image_page() 177 static void *__get_safe_page(gfp_t gfp_mask) in __get_safe_page() argument 186 return get_image_page(gfp_mask, PG_SAFE); in __get_safe_page() 189 unsigned long get_safe_page(gfp_t gfp_mask) in get_safe_page() argument 191 return (unsigned long)__get_safe_page(gfp_mask); in get_safe_page() 194 static struct page *alloc_image_page(gfp_t gfp_mask) in alloc_image_page() argument 198 page = alloc_page(gfp_mask); in alloc_image_page() 264 gfp_t gfp_mask; /* mask for allocating pages */ member [all …]
|