| /Linux-v6.6/mm/ | 
| D | page_alloc.c | 1541 							unsigned int alloc_flags)  in prep_new_page()  argument1554 	if (alloc_flags & ALLOC_NO_WATERMARKS)  in prep_new_page()
 1768 		unsigned int alloc_flags, int start_type, bool whole_block)  in steal_suitable_fallback()  argument
 1794 	if (boost_watermark(zone) && (alloc_flags & ALLOC_KSWAPD))  in steal_suitable_fallback()
 2005 						unsigned int alloc_flags)  in __rmqueue_fallback()  argument
 2019 	if (order < pageblock_order && alloc_flags & ALLOC_NOFRAGMENT)  in __rmqueue_fallback()
 2071 	steal_suitable_fallback(zone, page, alloc_flags, start_migratetype,  in __rmqueue_fallback()
 2087 						unsigned int alloc_flags)  in __rmqueue()  argument
 2097 		if (alloc_flags & ALLOC_CMA &&  in __rmqueue()
 2108 		if (alloc_flags & ALLOC_CMA)  in __rmqueue()
 [all …]
 
 | 
| D | compaction.c | 2351 		int alloc_flags)  in compaction_zonelist_suitable()  argument2409 					cc->alloc_flags & ALLOC_WMARK_MASK);  in compact_zone()
 2411 				      cc->highest_zoneidx, cc->alloc_flags))  in compact_zone()
 2620 		unsigned int alloc_flags, int highest_zoneidx,  in compact_zone_order()  argument
 2631 		.alloc_flags = alloc_flags,  in compact_zone_order()
 2684 		unsigned int alloc_flags, const struct alloc_context *ac,  in try_to_compact_pages()  argument
 2713 				alloc_flags, ac->highest_zoneidx, capture);  in try_to_compact_pages()
 
 | 
| D | internal.h | 496 	const unsigned int alloc_flags;	/* alloc flags of a direct compactor */  member
 | 
| /Linux-v6.6/drivers/base/regmap/ | 
| D | regcache-maple.c | 77 			map->alloc_flags);  in regcache_maple_write()95 	ret = mas_store_gfp(&mas, entry, map->alloc_flags);  in regcache_maple_write()
 137 					map->alloc_flags);  in regcache_maple_drop()
 151 					map->alloc_flags);  in regcache_maple_drop()
 165 			ret = mas_store_gfp(&mas, lower, map->alloc_flags);  in regcache_maple_drop()
 173 			ret = mas_store_gfp(&mas, upper, map->alloc_flags);  in regcache_maple_drop()
 207 		buf = kmalloc(val_bytes * (max - min), map->alloc_flags);  in regcache_maple_sync_block()
 323 	entry = kcalloc(last - first + 1, sizeof(unsigned long), map->alloc_flags);  in regcache_maple_insert_block()
 334 	ret = mas_store_gfp(&mas, entry, map->alloc_flags);  in regcache_maple_insert_block()
 
 | 
| D | regcache-rbtree.c | 280 		       map->alloc_flags);  in regcache_rbtree_insert_to_block()289 				   map->alloc_flags);  in regcache_rbtree_insert_to_block()
 323 	rbnode = kzalloc(sizeof(*rbnode), map->alloc_flags);  in regcache_rbtree_node_alloc()
 349 				      map->alloc_flags);  in regcache_rbtree_node_alloc()
 355 					map->alloc_flags);  in regcache_rbtree_node_alloc()
 
 | 
| D | internal.h | 65 	gfp_t alloc_flags;  member
 | 
| D | regmap.c | 755 		map->alloc_flags = GFP_ATOMIC;  in __regmap_init()757 		map->alloc_flags = GFP_KERNEL;  in __regmap_init()
 2346 		wval = kmemdup(val, val_count * val_bytes, map->alloc_flags);  in regmap_bulk_write()
 
 | 
| /Linux-v6.6/lib/ | 
| D | stackdepot.c | 360 					gfp_t alloc_flags, bool can_alloc)  in __stack_depot_save()  argument407 		alloc_flags &= ~GFP_ZONEMASK;  in __stack_depot_save()
 408 		alloc_flags &= (GFP_ATOMIC | GFP_KERNEL);  in __stack_depot_save()
 409 		alloc_flags |= __GFP_NOWARN;  in __stack_depot_save()
 410 		page = alloc_pages(alloc_flags, DEPOT_POOL_ORDER);  in __stack_depot_save()
 454 				      gfp_t alloc_flags)  in stack_depot_save()  argument
 456 	return __stack_depot_save(entries, nr_entries, alloc_flags, true);  in stack_depot_save()
 
 | 
| /Linux-v6.6/fs/xfs/libxfs/ | 
| D | xfs_alloc.c | 1540 	uint32_t		alloc_flags)  in xfs_alloc_ag_vextent_near()  argument1560 	alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH;  in xfs_alloc_ag_vextent_near()
 1625 					acur.busy_gen, alloc_flags);  in xfs_alloc_ag_vextent_near()
 1629 			alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH;  in xfs_alloc_ag_vextent_near()
 1655 	uint32_t		alloc_flags)  in xfs_alloc_ag_vextent_size()  argument
 1670 	alloc_flags |= XFS_ALLOC_FLAG_TRYFLUSH;  in xfs_alloc_ag_vextent_size()
 1740 					busy_gen, alloc_flags);  in xfs_alloc_ag_vextent_size()
 1744 			alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH;  in xfs_alloc_ag_vextent_size()
 1835 					busy_gen, alloc_flags);  in xfs_alloc_ag_vextent_size()
 1839 			alloc_flags &= ~XFS_ALLOC_FLAG_TRYFLUSH;  in xfs_alloc_ag_vextent_size()
 [all …]
 
 | 
| D | xfs_alloc.h | 199 int xfs_alloc_fix_freelist(struct xfs_alloc_arg *args, uint32_t alloc_flags);
 | 
| /Linux-v6.6/drivers/gpu/drm/amd/amdgpu/ | 
| D | amdgpu_amdkfd_gpuvm.c | 287 	u32 alloc_flags = bo->kfd_bo->alloc_flags;  in amdgpu_amdkfd_release_notify()  local290 	amdgpu_amdkfd_unreserve_mem_limit(adev, size, alloc_flags,  in amdgpu_amdkfd_release_notify()
 317 	if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR)  in create_dmamap_sg_bo()
 469 	if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE)  in get_pte_flags()
 471 	if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_EXECUTABLE)  in get_pte_flags()
 514 		mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?  in kfd_mem_dmamap_userptr()
 616 	mmio = (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP);  in kfd_mem_dmamap_sg_bo()
 622 	dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?  in kfd_mem_dmamap_sg_bo()
 681 		mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?  in kfd_mem_dmaunmap_userptr()
 743 	dir = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE ?  in kfd_mem_dmaunmap_sg_bo()
 [all …]
 
 | 
| D | amdgpu_amdkfd.h | 79 	uint32_t alloc_flags;  member
 | 
| /Linux-v6.6/include/linux/ | 
| D | compaction.h | 88 		unsigned int order, unsigned int alloc_flags,99 					int alloc_flags);
 
 | 
| D | mmzone.h | 1420 			 int highest_zoneidx, unsigned int alloc_flags,1424 		unsigned int alloc_flags);
 
 | 
| /Linux-v6.6/drivers/gpu/drm/amd/amdkfd/ | 
| D | kfd_chardev.c | 1932 			bo_bucket->alloc_flags = (uint32_t)kgd_mem->alloc_flags;  in criu_checkpoint_bos()1935 			if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_USERPTR) {  in criu_checkpoint_bos()
 1943 			if (bo_bucket->alloc_flags  in criu_checkpoint_bos()
 1946 						bo_bucket->alloc_flags &  in criu_checkpoint_bos()
 1955 			if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL)  in criu_checkpoint_bos()
 1958 			else if (bo_bucket->alloc_flags &  in criu_checkpoint_bos()
 1976 					bo_bucket->alloc_flags,  in criu_checkpoint_bos()
 2000 		if (bo_buckets[bo_index].alloc_flags  in criu_checkpoint_bos()
 2295 	if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_DOORBELL) {  in criu_restore_memory_of_gpu()
 2303 	} else if (bo_bucket->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_MMIO_REMAP) {  in criu_restore_memory_of_gpu()
 [all …]
 
 | 
| /Linux-v6.6/drivers/md/ | 
| D | dm-zoned-reclaim.c | 286 	int alloc_flags = DMZ_ALLOC_SEQ;  in dmz_reclaim_rnd_data()  local292 			       alloc_flags | DMZ_ALLOC_RECLAIM);  in dmz_reclaim_rnd_data()
 293 	if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) {  in dmz_reclaim_rnd_data()
 294 		alloc_flags = DMZ_ALLOC_RND;  in dmz_reclaim_rnd_data()
 
 | 
| D | dm-zoned-metadata.c | 2053 	int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND;  in dmz_get_chunk_mapping()  local2068 		dzone = dmz_alloc_zone(zmd, 0, alloc_flags);  in dmz_get_chunk_mapping()
 2165 	int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND;  in dmz_get_chunk_buffer()  local
 2174 	bzone = dmz_alloc_zone(zmd, 0, alloc_flags);  in dmz_get_chunk_buffer()
 
 | 
| /Linux-v6.6/fs/xfs/ | 
| D | xfs_extent_busy.c | 609 	uint32_t		alloc_flags)  in xfs_extent_busy_flush()  argument620 		if (alloc_flags & XFS_ALLOC_FLAG_TRYFLUSH)  in xfs_extent_busy_flush()
 626 		if (alloc_flags & XFS_ALLOC_FLAG_FREEING)  in xfs_extent_busy_flush()
 
 | 
| D | xfs_extent_busy.h | 74 		unsigned busy_gen, uint32_t alloc_flags);
 | 
| /Linux-v6.6/fs/btrfs/ | 
| D | block-group.c | 2844 	u64 alloc_flags;  in btrfs_inc_block_group_ro()  local2892 		alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags);  in btrfs_inc_block_group_ro()
 2893 		if (alloc_flags != cache->flags) {  in btrfs_inc_block_group_ro()
 2894 			ret = btrfs_chunk_alloc(trans, alloc_flags,  in btrfs_inc_block_group_ro()
 2922 	alloc_flags = btrfs_get_alloc_profile(fs_info, cache->space_info->flags);  in btrfs_inc_block_group_ro()
 2923 	ret = btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE);  in btrfs_inc_block_group_ro()
 2939 		alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags);  in btrfs_inc_block_group_ro()
 2941 		check_system_chunk(trans, alloc_flags);  in btrfs_inc_block_group_ro()
 3783 	u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type);  in btrfs_force_chunk_alloc()  local
 3785 	return btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE);  in btrfs_force_chunk_alloc()
 
 | 
| /Linux-v6.6/drivers/iommu/ | 
| D | dma-iommu.c | 806 			gfp_t alloc_flags = gfp;  in __iommu_dma_alloc_pages()  local810 				alloc_flags |= __GFP_NORETRY;  in __iommu_dma_alloc_pages()
 811 			page = alloc_pages_node(nid, alloc_flags, order);  in __iommu_dma_alloc_pages()
 
 | 
| /Linux-v6.6/drivers/gpu/drm/i915/gem/ | 
| D | i915_gem_object.h | 50 			  unsigned alloc_flags);
 | 
| /Linux-v6.6/include/uapi/linux/ | 
| D | kfd_ioctl.h | 633 	__u32 alloc_flags;  member
 | 
| /Linux-v6.6/drivers/net/ethernet/mellanox/mlx5/core/ | 
| D | cmd.c | 118 	gfp_t alloc_flags = cbk ? GFP_ATOMIC : GFP_KERNEL;  in cmd_alloc_ent()  local121 	ent = kzalloc(sizeof(*ent), alloc_flags);  in cmd_alloc_ent()
 
 | 
| /Linux-v6.6/arch/s390/kvm/ | 
| D | kvm-s390.c | 3292 	gfp_t alloc_flags = GFP_KERNEL_ACCOUNT;  in kvm_arch_init_vm()  local3315 		alloc_flags |= GFP_DMA;  in kvm_arch_init_vm()
 3318 	kvm->arch.sca = (struct bsca_block *) get_zeroed_page(alloc_flags);  in kvm_arch_init_vm()
 
 |