/Linux-v5.4/mm/ |
D | mempool.c | 25 static void poison_error(mempool_t *pool, void *element, size_t size, in poison_error() argument 28 const int nr = pool->curr_nr; in poison_error() 34 pr_err("Mempool %p size %zu\n", pool, size); in poison_error() 42 static void __check_element(mempool_t *pool, void *element, size_t size) in __check_element() argument 51 poison_error(pool, element, size, i); in __check_element() 58 static void check_element(mempool_t *pool, void *element) in check_element() argument 61 if (pool->free == mempool_free_slab || pool->free == mempool_kfree) in check_element() 62 __check_element(pool, element, ksize(element)); in check_element() 65 if (pool->free == mempool_free_pages) { in check_element() 66 int order = (int)(long)pool->pool_data; in check_element() [all …]
|
D | zbud.c | 129 static int zbud_zpool_evict(struct zbud_pool *pool, unsigned long handle) in zbud_zpool_evict() argument 131 if (pool->zpool && pool->zpool_ops && pool->zpool_ops->evict) in zbud_zpool_evict() 132 return pool->zpool_ops->evict(pool->zpool, handle); in zbud_zpool_evict() 145 struct zbud_pool *pool; in zbud_zpool_create() local 147 pool = zbud_create_pool(gfp, zpool_ops ? &zbud_zpool_ops : NULL); in zbud_zpool_create() 148 if (pool) { in zbud_zpool_create() 149 pool->zpool = zpool; in zbud_zpool_create() 150 pool->zpool_ops = zpool_ops; in zbud_zpool_create() 152 return pool; in zbud_zpool_create() 155 static void zbud_zpool_destroy(void *pool) in zbud_zpool_destroy() argument [all …]
|
D | dmapool.c | 71 struct dma_pool *pool; in show_pools() local 81 list_for_each_entry(pool, &dev->dma_pools, pools) { in show_pools() 85 spin_lock_irq(&pool->lock); in show_pools() 86 list_for_each_entry(page, &pool->page_list, page_list) { in show_pools() 90 spin_unlock_irq(&pool->lock); in show_pools() 94 pool->name, blocks, in show_pools() 95 pages * (pool->allocation / pool->size), in show_pools() 96 pool->size, pages); in show_pools() 205 static void pool_initialise_page(struct dma_pool *pool, struct dma_page *page) in pool_initialise_page() argument 208 unsigned int next_boundary = pool->boundary; in pool_initialise_page() [all …]
|
D | z3fold.c | 75 int (*evict)(struct z3fold_pool *pool, unsigned long handle); 92 unsigned long pool; /* back link + flags */ member 119 struct z3fold_pool *pool; member 196 static inline struct z3fold_buddy_slots *alloc_slots(struct z3fold_pool *pool, in alloc_slots() argument 201 slots = kmem_cache_alloc(pool->c_handle, in alloc_slots() 206 slots->pool = (unsigned long)pool; in alloc_slots() 214 return (struct z3fold_pool *)(s->pool & ~HANDLE_FLAG_MASK); in slots_to_pool() 243 struct z3fold_pool *pool = slots_to_pool(slots); in free_handle() local 245 kmem_cache_free(pool->c_handle, slots); in free_handle() 278 static int z3fold_register_migration(struct z3fold_pool *pool) in z3fold_register_migration() argument [all …]
|
D | zswap.c | 156 struct zswap_pool *pool; member 203 static int zswap_writeback_entry(struct zpool *pool, unsigned long handle); 204 static int zswap_pool_get(struct zswap_pool *pool); 205 static void zswap_pool_put(struct zswap_pool *pool); 219 struct zswap_pool *pool; in zswap_update_total_size() local 224 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_update_total_size() 225 total += zpool_get_total_size(pool->zpool); in zswap_update_total_size() 328 zpool_free(entry->pool->zpool, entry->handle); in zswap_free_entry() 329 zswap_pool_put(entry->pool); in zswap_free_entry() 400 struct zswap_pool *pool = hlist_entry(node, struct zswap_pool, node); in zswap_cpu_comp_prepare() local [all …]
|
/Linux-v5.4/drivers/staging/android/ion/ |
D | ion_page_pool.c | 15 static inline struct page *ion_page_pool_alloc_pages(struct ion_page_pool *pool) in ion_page_pool_alloc_pages() argument 19 return alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages() 22 static void ion_page_pool_free_pages(struct ion_page_pool *pool, in ion_page_pool_free_pages() argument 25 __free_pages(page, pool->order); in ion_page_pool_free_pages() 28 static void ion_page_pool_add(struct ion_page_pool *pool, struct page *page) in ion_page_pool_add() argument 30 mutex_lock(&pool->mutex); in ion_page_pool_add() 32 list_add_tail(&page->lru, &pool->high_items); in ion_page_pool_add() 33 pool->high_count++; in ion_page_pool_add() 35 list_add_tail(&page->lru, &pool->low_items); in ion_page_pool_add() 36 pool->low_count++; in ion_page_pool_add() [all …]
|
/Linux-v5.4/net/core/ |
D | page_pool.c | 21 static int page_pool_init(struct page_pool *pool, in page_pool_init() argument 26 memcpy(&pool->p, params, sizeof(pool->p)); in page_pool_init() 29 if (pool->p.flags & ~(PP_FLAG_ALL)) in page_pool_init() 32 if (pool->p.pool_size) in page_pool_init() 33 ring_qsize = pool->p.pool_size; in page_pool_init() 43 if ((pool->p.dma_dir != DMA_FROM_DEVICE) && in page_pool_init() 44 (pool->p.dma_dir != DMA_BIDIRECTIONAL)) in page_pool_init() 47 if (ptr_ring_init(&pool->ring, ring_qsize, GFP_KERNEL) < 0) in page_pool_init() 50 atomic_set(&pool->pages_state_release_cnt, 0); in page_pool_init() 53 refcount_set(&pool->user_cnt, 1); in page_pool_init() [all …]
|
/Linux-v5.4/drivers/infiniband/sw/rxe/ |
D | rxe_pool.c | 108 static inline const char *pool_name(struct rxe_pool *pool) in pool_name() argument 110 return rxe_type_info[pool->type].name; in pool_name() 113 static inline struct kmem_cache *pool_cache(struct rxe_pool *pool) in pool_cache() argument 115 return rxe_type_info[pool->type].cache; in pool_cache() 169 static int rxe_pool_init_index(struct rxe_pool *pool, u32 max, u32 min) in rxe_pool_init_index() argument 174 if ((max - min + 1) < pool->max_elem) { in rxe_pool_init_index() 180 pool->max_index = max; in rxe_pool_init_index() 181 pool->min_index = min; in rxe_pool_init_index() 184 pool->table = kmalloc(size, GFP_KERNEL); in rxe_pool_init_index() 185 if (!pool->table) { in rxe_pool_init_index() [all …]
|
/Linux-v5.4/drivers/infiniband/core/ |
D | fmr_pool.c | 95 void (*flush_function)(struct ib_fmr_pool *pool, 115 static inline struct ib_pool_fmr *ib_fmr_cache_lookup(struct ib_fmr_pool *pool, in ib_fmr_cache_lookup() argument 123 if (!pool->cache_bucket) in ib_fmr_cache_lookup() 126 bucket = pool->cache_bucket + ib_fmr_hash(*page_list); in ib_fmr_cache_lookup() 138 static void ib_fmr_batch_release(struct ib_fmr_pool *pool) in ib_fmr_batch_release() argument 145 spin_lock_irq(&pool->pool_lock); in ib_fmr_batch_release() 147 list_for_each_entry(fmr, &pool->dirty_list, list) { in ib_fmr_batch_release() 153 list_splice_init(&pool->dirty_list, &unmap_list); in ib_fmr_batch_release() 154 pool->dirty_len = 0; in ib_fmr_batch_release() 156 spin_unlock_irq(&pool->pool_lock); in ib_fmr_batch_release() [all …]
|
/Linux-v5.4/sound/core/seq/ |
D | seq_memory.c | 22 static inline int snd_seq_pool_available(struct snd_seq_pool *pool) in snd_seq_pool_available() argument 24 return pool->total_elements - atomic_read(&pool->counter); in snd_seq_pool_available() 27 static inline int snd_seq_output_ok(struct snd_seq_pool *pool) in snd_seq_output_ok() argument 29 return snd_seq_pool_available(pool) >= pool->room; in snd_seq_output_ok() 163 static inline void free_cell(struct snd_seq_pool *pool, in free_cell() argument 166 cell->next = pool->free; in free_cell() 167 pool->free = cell; in free_cell() 168 atomic_dec(&pool->counter); in free_cell() 174 struct snd_seq_pool *pool; in snd_seq_cell_free() local 178 pool = cell->pool; in snd_seq_cell_free() [all …]
|
/Linux-v5.4/drivers/md/ |
D | dm-thin.c | 229 struct pool { struct 286 static void metadata_operation_failed(struct pool *pool, const char *op, int r); argument 288 static enum pool_mode get_pool_mode(struct pool *pool) in get_pool_mode() argument 290 return pool->pf.mode; in get_pool_mode() 293 static void notify_of_pool_mode_change(struct pool *pool) in notify_of_pool_mode_change() argument 303 enum pool_mode mode = get_pool_mode(pool); in notify_of_pool_mode_change() 306 if (!pool->pf.error_if_no_space) in notify_of_pool_mode_change() 312 dm_table_event(pool->ti->table); in notify_of_pool_mode_change() 314 dm_device_name(pool->pool_md), in notify_of_pool_mode_change() 323 struct pool *pool; member [all …]
|
/Linux-v5.4/net/ceph/ |
D | msgpool.c | 14 struct ceph_msgpool *pool = arg; in msgpool_alloc() local 17 msg = ceph_msg_new2(pool->type, pool->front_len, pool->max_data_items, in msgpool_alloc() 20 dout("msgpool_alloc %s failed\n", pool->name); in msgpool_alloc() 22 dout("msgpool_alloc %s %p\n", pool->name, msg); in msgpool_alloc() 23 msg->pool = pool; in msgpool_alloc() 30 struct ceph_msgpool *pool = arg; in msgpool_free() local 33 dout("msgpool_release %s %p\n", pool->name, msg); in msgpool_free() 34 msg->pool = NULL; in msgpool_free() 38 int ceph_msgpool_init(struct ceph_msgpool *pool, int type, in ceph_msgpool_init() argument 43 pool->type = type; in ceph_msgpool_init() [all …]
|
/Linux-v5.4/drivers/gpu/drm/ttm/ |
D | ttm_page_alloc_dma.c | 151 struct dma_pool *pool; member 266 static int ttm_set_pages_caching(struct dma_pool *pool, in ttm_set_pages_caching() argument 271 if (pool->type & IS_UC) { in ttm_set_pages_caching() 275 pool->dev_name, cpages); in ttm_set_pages_caching() 277 if (pool->type & IS_WC) { in ttm_set_pages_caching() 281 pool->dev_name, cpages); in ttm_set_pages_caching() 286 static void __ttm_dma_free_page(struct dma_pool *pool, struct dma_page *d_page) in __ttm_dma_free_page() argument 291 if (pool->type & IS_HUGE) in __ttm_dma_free_page() 294 dma_free_attrs(pool->dev, pool->size, (void *)d_page->vaddr, dma, attrs); in __ttm_dma_free_page() 299 static struct dma_page *__ttm_dma_alloc_page(struct dma_pool *pool) in __ttm_dma_alloc_page() argument [all …]
|
/Linux-v5.4/drivers/gpu/drm/amd/display/dc/dce80/ |
D | dce80_resource.c | 730 static void destruct(struct dce110_resource_pool *pool) in destruct() argument 734 for (i = 0; i < pool->base.pipe_count; i++) { in destruct() 735 if (pool->base.opps[i] != NULL) in destruct() 736 dce110_opp_destroy(&pool->base.opps[i]); in destruct() 738 if (pool->base.transforms[i] != NULL) in destruct() 739 dce80_transform_destroy(&pool->base.transforms[i]); in destruct() 741 if (pool->base.ipps[i] != NULL) in destruct() 742 dce_ipp_destroy(&pool->base.ipps[i]); in destruct() 744 if (pool->base.mis[i] != NULL) { in destruct() 745 kfree(TO_DCE_MEM_INPUT(pool->base.mis[i])); in destruct() [all …]
|
/Linux-v5.4/net/rds/ |
D | ib_rdma.c | 191 struct rds_ib_mr *rds_ib_reuse_mr(struct rds_ib_mr_pool *pool) in rds_ib_reuse_mr() argument 197 spin_lock_irqsave(&pool->clean_lock, flags); in rds_ib_reuse_mr() 198 ret = llist_del_first(&pool->clean_list); in rds_ib_reuse_mr() 199 spin_unlock_irqrestore(&pool->clean_lock, flags); in rds_ib_reuse_mr() 202 if (pool->pool_type == RDS_IB_MR_8K_POOL) in rds_ib_reuse_mr() 265 struct rds_ib_mr_pool *pool = ibmr->pool; in rds_ib_teardown_mr() local 267 atomic_sub(pinned, &pool->free_pinned); in rds_ib_teardown_mr() 271 static inline unsigned int rds_ib_flush_goal(struct rds_ib_mr_pool *pool, int free_all) in rds_ib_flush_goal() argument 275 item_count = atomic_read(&pool->item_count); in rds_ib_flush_goal() 332 int rds_ib_flush_mr_pool(struct rds_ib_mr_pool *pool, in rds_ib_flush_mr_pool() argument [all …]
|
/Linux-v5.4/include/net/ |
D | page_pool.h | 112 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp); 114 static inline struct page *page_pool_dev_alloc_pages(struct page_pool *pool) in page_pool_dev_alloc_pages() argument 118 return page_pool_alloc_pages(pool, gfp); in page_pool_dev_alloc_pages() 125 inline enum dma_data_direction page_pool_get_dma_dir(struct page_pool *pool) in page_pool_get_dma_dir() argument 127 return pool->p.dma_dir; in page_pool_get_dma_dir() 132 void __page_pool_free(struct page_pool *pool); 133 static inline void page_pool_free(struct page_pool *pool) in page_pool_free() argument 139 __page_pool_free(pool); in page_pool_free() 144 static inline void page_pool_destroy(struct page_pool *pool) in page_pool_destroy() argument 146 if (!pool) in page_pool_destroy() [all …]
|
/Linux-v5.4/include/linux/ |
D | genalloc.h | 52 void *data, struct gen_pool *pool, 97 extern phys_addr_t gen_pool_virt_to_phys(struct gen_pool *pool, unsigned long); 101 static inline int gen_pool_add_virt(struct gen_pool *pool, unsigned long addr, in gen_pool_add_virt() argument 104 return gen_pool_add_owner(pool, addr, phys, size, nid, NULL); in gen_pool_add_virt() 119 static inline int gen_pool_add(struct gen_pool *pool, unsigned long addr, in gen_pool_add() argument 122 return gen_pool_add_virt(pool, addr, -1, size, nid); in gen_pool_add() 125 unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size, 128 static inline unsigned long gen_pool_alloc_owner(struct gen_pool *pool, in gen_pool_alloc_owner() argument 131 return gen_pool_alloc_algo_owner(pool, size, pool->algo, pool->data, in gen_pool_alloc_owner() 135 static inline unsigned long gen_pool_alloc_algo(struct gen_pool *pool, in gen_pool_alloc_algo() argument [all …]
|
/Linux-v5.4/drivers/gpu/drm/i915/gt/ |
D | intel_engine_pool.c | 13 static struct intel_engine_cs *to_engine(struct intel_engine_pool *pool) in to_engine() argument 15 return container_of(pool, struct intel_engine_cs, pool); in to_engine() 19 bucket_for_size(struct intel_engine_pool *pool, size_t sz) in bucket_for_size() argument 29 if (n >= ARRAY_SIZE(pool->cache_list)) in bucket_for_size() 30 n = ARRAY_SIZE(pool->cache_list) - 1; in bucket_for_size() 32 return &pool->cache_list[n]; in bucket_for_size() 68 struct intel_engine_pool *pool = node->pool; in pool_retire() local 69 struct list_head *list = bucket_for_size(pool, node->obj->base.size); in pool_retire() 72 GEM_BUG_ON(!intel_engine_pm_is_awake(to_engine(pool))); in pool_retire() 79 spin_lock_irqsave(&pool->lock, flags); in pool_retire() [all …]
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_icm_pool.c | 12 struct mlx5dr_icm_pool *pool; member 61 struct mlx5dr_icm_pool *pool; member 98 dr_icm_pool_mr_create(struct mlx5dr_icm_pool *pool, in dr_icm_pool_mr_create() argument 102 struct mlx5_core_dev *mdev = pool->dmn->mdev; in dr_icm_pool_mr_create() 111 icm_mr->pool = pool; in dr_icm_pool_mr_create() 117 icm_mr->dm.length = mlx5dr_icm_pool_chunk_size_to_byte(pool->max_log_chunk_sz, in dr_icm_pool_mr_create() 118 pool->icm_type) * 2; in dr_icm_pool_mr_create() 123 mlx5dr_err(pool->dmn, "Failed to allocate SW ICM memory, err (%d)\n", err); in dr_icm_pool_mr_create() 128 err = dr_icm_create_dm_mkey(mdev, pool->dmn->pdn, in dr_icm_pool_mr_create() 134 mlx5dr_err(pool->dmn, "Failed to create SW ICM MKEY, err (%d)\n", err); in dr_icm_pool_mr_create() [all …]
|
/Linux-v5.4/lib/ |
D | genalloc.c | 153 struct gen_pool *pool; in gen_pool_create() local 155 pool = kmalloc_node(sizeof(struct gen_pool), GFP_KERNEL, nid); in gen_pool_create() 156 if (pool != NULL) { in gen_pool_create() 157 spin_lock_init(&pool->lock); in gen_pool_create() 158 INIT_LIST_HEAD(&pool->chunks); in gen_pool_create() 159 pool->min_alloc_order = min_alloc_order; in gen_pool_create() 160 pool->algo = gen_pool_first_fit; in gen_pool_create() 161 pool->data = NULL; in gen_pool_create() 162 pool->name = NULL; in gen_pool_create() 164 return pool; in gen_pool_create() [all …]
|
/Linux-v5.4/drivers/gpu/drm/amd/display/dc/dce100/ |
D | dce100_resource.c | 682 static void destruct(struct dce110_resource_pool *pool) in destruct() argument 686 for (i = 0; i < pool->base.pipe_count; i++) { in destruct() 687 if (pool->base.opps[i] != NULL) in destruct() 688 dce110_opp_destroy(&pool->base.opps[i]); in destruct() 690 if (pool->base.transforms[i] != NULL) in destruct() 691 dce100_transform_destroy(&pool->base.transforms[i]); in destruct() 693 if (pool->base.ipps[i] != NULL) in destruct() 694 dce_ipp_destroy(&pool->base.ipps[i]); in destruct() 696 if (pool->base.mis[i] != NULL) { in destruct() 697 kfree(TO_DCE_MEM_INPUT(pool->base.mis[i])); in destruct() [all …]
|
/Linux-v5.4/drivers/media/platform/vsp1/ |
D | vsp1_dl.c | 110 struct vsp1_dl_body_pool *pool; member 227 struct vsp1_dl_body_pool *pool; member 251 struct vsp1_dl_body_pool *pool; in vsp1_dl_body_pool_create() local 255 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in vsp1_dl_body_pool_create() 256 if (!pool) in vsp1_dl_body_pool_create() 259 pool->vsp1 = vsp1; in vsp1_dl_body_pool_create() 268 pool->size = dlb_size * num_bodies; in vsp1_dl_body_pool_create() 270 pool->bodies = kcalloc(num_bodies, sizeof(*pool->bodies), GFP_KERNEL); in vsp1_dl_body_pool_create() 271 if (!pool->bodies) { in vsp1_dl_body_pool_create() 272 kfree(pool); in vsp1_dl_body_pool_create() [all …]
|
/Linux-v5.4/drivers/staging/media/ipu3/ |
D | ipu3-css-pool.c | 25 void imgu_css_pool_cleanup(struct imgu_device *imgu, struct imgu_css_pool *pool) in imgu_css_pool_cleanup() argument 30 imgu_dmamap_free(imgu, &pool->entry[i].param); in imgu_css_pool_cleanup() 33 int imgu_css_pool_init(struct imgu_device *imgu, struct imgu_css_pool *pool, in imgu_css_pool_init() argument 39 pool->entry[i].valid = false; in imgu_css_pool_init() 41 pool->entry[i].param.vaddr = NULL; in imgu_css_pool_init() 45 if (!imgu_dmamap_alloc(imgu, &pool->entry[i].param, size)) in imgu_css_pool_init() 49 pool->last = IPU3_CSS_POOL_SIZE; in imgu_css_pool_init() 54 imgu_css_pool_cleanup(imgu, pool); in imgu_css_pool_init() 61 void imgu_css_pool_get(struct imgu_css_pool *pool) in imgu_css_pool_get() argument 64 u32 n = (pool->last + 1) % IPU3_CSS_POOL_SIZE; in imgu_css_pool_get() [all …]
|
/Linux-v5.4/drivers/staging/octeon/ |
D | ethernet-mem.c | 24 static int cvm_oct_fill_hw_skbuff(int pool, int size, int elements) in cvm_oct_fill_hw_skbuff() argument 35 cvmx_fpa_free(skb->data, pool, size / 128); in cvm_oct_fill_hw_skbuff() 47 static void cvm_oct_free_hw_skbuff(int pool, int size, int elements) in cvm_oct_free_hw_skbuff() argument 52 memory = cvmx_fpa_alloc(pool); in cvm_oct_free_hw_skbuff() 63 pool, elements); in cvm_oct_free_hw_skbuff() 66 pool, elements); in cvm_oct_free_hw_skbuff() 77 static int cvm_oct_fill_hw_memory(int pool, int size, int elements) in cvm_oct_fill_hw_memory() argument 97 elements * size, pool); in cvm_oct_fill_hw_memory() 102 cvmx_fpa_free(fpa, pool, 0); in cvm_oct_fill_hw_memory() 114 static void cvm_oct_free_hw_memory(int pool, int size, int elements) in cvm_oct_free_hw_memory() argument [all …]
|
/Linux-v5.4/kernel/ |
D | workqueue.c | 200 struct worker_pool *pool; /* I: the associated pool */ member 378 #define for_each_cpu_worker_pool(pool, cpu) \ argument 379 for ((pool) = &per_cpu(cpu_worker_pools, cpu)[0]; \ 380 (pool) < &per_cpu(cpu_worker_pools, cpu)[NR_STD_WORKER_POOLS]; \ 381 (pool)++) 395 #define for_each_pool(pool, pi) \ argument 396 idr_for_each_entry(&worker_pool_idr, pool, pi) \ 410 #define for_each_pool_worker(worker, pool) \ argument 411 list_for_each_entry((worker), &(pool)->workers, node) \ 536 static int worker_pool_assign_id(struct worker_pool *pool) in worker_pool_assign_id() argument [all …]
|