Lines Matching refs:pool
153 struct gen_pool *pool; in gen_pool_create() local
155 pool = kmalloc_node(sizeof(struct gen_pool), GFP_KERNEL, nid); in gen_pool_create()
156 if (pool != NULL) { in gen_pool_create()
157 spin_lock_init(&pool->lock); in gen_pool_create()
158 INIT_LIST_HEAD(&pool->chunks); in gen_pool_create()
159 pool->min_alloc_order = min_alloc_order; in gen_pool_create()
160 pool->algo = gen_pool_first_fit; in gen_pool_create()
161 pool->data = NULL; in gen_pool_create()
162 pool->name = NULL; in gen_pool_create()
164 return pool; in gen_pool_create()
182 int gen_pool_add_owner(struct gen_pool *pool, unsigned long virt, phys_addr_t phys, in gen_pool_add_owner() argument
186 int nbits = size >> pool->min_alloc_order; in gen_pool_add_owner()
200 spin_lock(&pool->lock); in gen_pool_add_owner()
201 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_owner()
202 spin_unlock(&pool->lock); in gen_pool_add_owner()
215 phys_addr_t gen_pool_virt_to_phys(struct gen_pool *pool, unsigned long addr) in gen_pool_virt_to_phys() argument
221 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_virt_to_phys()
240 void gen_pool_destroy(struct gen_pool *pool) in gen_pool_destroy() argument
244 int order = pool->min_alloc_order; in gen_pool_destroy()
247 list_for_each_safe(_chunk, _next_chunk, &pool->chunks) { in gen_pool_destroy()
257 kfree_const(pool->name); in gen_pool_destroy()
258 kfree(pool); in gen_pool_destroy()
275 unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size, in gen_pool_alloc_algo_owner() argument
280 int order = pool->min_alloc_order; in gen_pool_alloc_algo_owner()
295 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_alloc_algo_owner()
303 nbits, data, pool, chunk->start_addr); in gen_pool_alloc_algo_owner()
339 void *gen_pool_dma_alloc(struct gen_pool *pool, size_t size, dma_addr_t *dma) in gen_pool_dma_alloc() argument
341 return gen_pool_dma_alloc_algo(pool, size, dma, pool->algo, pool->data); in gen_pool_dma_alloc()
360 void *gen_pool_dma_alloc_algo(struct gen_pool *pool, size_t size, in gen_pool_dma_alloc_algo() argument
365 if (!pool) in gen_pool_dma_alloc_algo()
368 vaddr = gen_pool_alloc_algo(pool, size, algo, data); in gen_pool_dma_alloc_algo()
373 *dma = gen_pool_virt_to_phys(pool, vaddr); in gen_pool_dma_alloc_algo()
393 void *gen_pool_dma_alloc_align(struct gen_pool *pool, size_t size, in gen_pool_dma_alloc_align() argument
398 return gen_pool_dma_alloc_algo(pool, size, dma, in gen_pool_dma_alloc_align()
417 void *gen_pool_dma_zalloc(struct gen_pool *pool, size_t size, dma_addr_t *dma) in gen_pool_dma_zalloc() argument
419 return gen_pool_dma_zalloc_algo(pool, size, dma, pool->algo, pool->data); in gen_pool_dma_zalloc()
438 void *gen_pool_dma_zalloc_algo(struct gen_pool *pool, size_t size, in gen_pool_dma_zalloc_algo() argument
441 void *vaddr = gen_pool_dma_alloc_algo(pool, size, dma, algo, data); in gen_pool_dma_zalloc_algo()
464 void *gen_pool_dma_zalloc_align(struct gen_pool *pool, size_t size, in gen_pool_dma_zalloc_align() argument
469 return gen_pool_dma_zalloc_algo(pool, size, dma, in gen_pool_dma_zalloc_align()
485 void gen_pool_free_owner(struct gen_pool *pool, unsigned long addr, size_t size, in gen_pool_free_owner() argument
489 int order = pool->min_alloc_order; in gen_pool_free_owner()
501 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_free_owner()
529 void gen_pool_for_each_chunk(struct gen_pool *pool, in gen_pool_for_each_chunk() argument
530 void (*func)(struct gen_pool *pool, struct gen_pool_chunk *chunk, void *data), in gen_pool_for_each_chunk() argument
536 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) in gen_pool_for_each_chunk()
537 func(pool, chunk, data); in gen_pool_for_each_chunk()
551 bool gen_pool_has_addr(struct gen_pool *pool, unsigned long start, in gen_pool_has_addr() argument
559 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) { in gen_pool_has_addr()
578 size_t gen_pool_avail(struct gen_pool *pool) in gen_pool_avail() argument
584 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_avail()
597 size_t gen_pool_size(struct gen_pool *pool) in gen_pool_size() argument
603 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_size()
620 void gen_pool_set_algo(struct gen_pool *pool, genpool_algo_t algo, void *data) in gen_pool_set_algo() argument
624 pool->algo = algo; in gen_pool_set_algo()
625 if (!pool->algo) in gen_pool_set_algo()
626 pool->algo = gen_pool_first_fit; in gen_pool_set_algo()
628 pool->data = data; in gen_pool_set_algo()
646 struct gen_pool *pool, unsigned long start_addr) in gen_pool_first_fit() argument
664 struct gen_pool *pool, unsigned long start_addr) in gen_pool_first_fit_align() argument
671 order = pool->min_alloc_order; in gen_pool_first_fit_align()
691 struct gen_pool *pool, unsigned long start_addr) in gen_pool_fixed_alloc() argument
699 order = pool->min_alloc_order; in gen_pool_fixed_alloc()
725 unsigned int nr, void *data, struct gen_pool *pool, in gen_pool_first_fit_order_align() argument
749 struct gen_pool *pool, unsigned long start_addr) in gen_pool_best_fit() argument
825 struct gen_pool **ptr, *pool; in devm_gen_pool_create() local
842 pool = gen_pool_create(min_alloc_order, nid); in devm_gen_pool_create()
843 if (!pool) in devm_gen_pool_create()
846 *ptr = pool; in devm_gen_pool_create()
847 pool->name = pool_name; in devm_gen_pool_create()
850 return pool; in devm_gen_pool_create()
878 struct gen_pool *pool = NULL; in of_gen_pool_get() local
896 pool = gen_pool_get(&pdev->dev, name); in of_gen_pool_get()
899 return pool; in of_gen_pool_get()