Lines Matching refs:pool
153 struct gen_pool *pool; in gen_pool_create() local
155 pool = kmalloc_node(sizeof(struct gen_pool), GFP_KERNEL, nid); in gen_pool_create()
156 if (pool != NULL) { in gen_pool_create()
157 spin_lock_init(&pool->lock); in gen_pool_create()
158 INIT_LIST_HEAD(&pool->chunks); in gen_pool_create()
159 pool->min_alloc_order = min_alloc_order; in gen_pool_create()
160 pool->algo = gen_pool_first_fit; in gen_pool_create()
161 pool->data = NULL; in gen_pool_create()
162 pool->name = NULL; in gen_pool_create()
164 return pool; in gen_pool_create()
182 int gen_pool_add_owner(struct gen_pool *pool, unsigned long virt, phys_addr_t phys, in gen_pool_add_owner() argument
186 int nbits = size >> pool->min_alloc_order; in gen_pool_add_owner()
200 spin_lock(&pool->lock); in gen_pool_add_owner()
201 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_owner()
202 spin_unlock(&pool->lock); in gen_pool_add_owner()
215 phys_addr_t gen_pool_virt_to_phys(struct gen_pool *pool, unsigned long addr) in gen_pool_virt_to_phys() argument
221 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_virt_to_phys()
240 void gen_pool_destroy(struct gen_pool *pool) in gen_pool_destroy() argument
244 int order = pool->min_alloc_order; in gen_pool_destroy()
247 list_for_each_safe(_chunk, _next_chunk, &pool->chunks) { in gen_pool_destroy()
257 kfree_const(pool->name); in gen_pool_destroy()
258 kfree(pool); in gen_pool_destroy()
275 unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size, in gen_pool_alloc_algo_owner() argument
280 int order = pool->min_alloc_order; in gen_pool_alloc_algo_owner()
295 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_alloc_algo_owner()
303 nbits, data, pool, chunk->start_addr); in gen_pool_alloc_algo_owner()
339 void *gen_pool_dma_alloc(struct gen_pool *pool, size_t size, dma_addr_t *dma) in gen_pool_dma_alloc() argument
341 return gen_pool_dma_alloc_algo(pool, size, dma, pool->algo, pool->data); in gen_pool_dma_alloc()
360 void *gen_pool_dma_alloc_algo(struct gen_pool *pool, size_t size, in gen_pool_dma_alloc_algo() argument
365 if (!pool) in gen_pool_dma_alloc_algo()
368 vaddr = gen_pool_alloc_algo(pool, size, algo, data); in gen_pool_dma_alloc_algo()
373 *dma = gen_pool_virt_to_phys(pool, vaddr); in gen_pool_dma_alloc_algo()
393 void *gen_pool_dma_alloc_align(struct gen_pool *pool, size_t size, in gen_pool_dma_alloc_align() argument
398 return gen_pool_dma_alloc_algo(pool, size, dma, in gen_pool_dma_alloc_align()
417 void *gen_pool_dma_zalloc(struct gen_pool *pool, size_t size, dma_addr_t *dma) in gen_pool_dma_zalloc() argument
419 return gen_pool_dma_zalloc_algo(pool, size, dma, pool->algo, pool->data); in gen_pool_dma_zalloc()
438 void *gen_pool_dma_zalloc_algo(struct gen_pool *pool, size_t size, in gen_pool_dma_zalloc_algo() argument
441 void *vaddr = gen_pool_dma_alloc_algo(pool, size, dma, algo, data); in gen_pool_dma_zalloc_algo()
464 void *gen_pool_dma_zalloc_align(struct gen_pool *pool, size_t size, in gen_pool_dma_zalloc_align() argument
469 return gen_pool_dma_zalloc_algo(pool, size, dma, in gen_pool_dma_zalloc_align()
485 void gen_pool_free_owner(struct gen_pool *pool, unsigned long addr, size_t size, in gen_pool_free_owner() argument
489 int order = pool->min_alloc_order; in gen_pool_free_owner()
501 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_free_owner()
529 void gen_pool_for_each_chunk(struct gen_pool *pool, in gen_pool_for_each_chunk() argument
530 void (*func)(struct gen_pool *pool, struct gen_pool_chunk *chunk, void *data), in gen_pool_for_each_chunk() argument
536 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) in gen_pool_for_each_chunk()
537 func(pool, chunk, data); in gen_pool_for_each_chunk()
551 bool addr_in_gen_pool(struct gen_pool *pool, unsigned long start, in addr_in_gen_pool() argument
559 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) { in addr_in_gen_pool()
577 size_t gen_pool_avail(struct gen_pool *pool) in gen_pool_avail() argument
583 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_avail()
596 size_t gen_pool_size(struct gen_pool *pool) in gen_pool_size() argument
602 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_size()
619 void gen_pool_set_algo(struct gen_pool *pool, genpool_algo_t algo, void *data) in gen_pool_set_algo() argument
623 pool->algo = algo; in gen_pool_set_algo()
624 if (!pool->algo) in gen_pool_set_algo()
625 pool->algo = gen_pool_first_fit; in gen_pool_set_algo()
627 pool->data = data; in gen_pool_set_algo()
645 struct gen_pool *pool, unsigned long start_addr) in gen_pool_first_fit() argument
663 struct gen_pool *pool, unsigned long start_addr) in gen_pool_first_fit_align() argument
670 order = pool->min_alloc_order; in gen_pool_first_fit_align()
690 struct gen_pool *pool, unsigned long start_addr) in gen_pool_fixed_alloc() argument
698 order = pool->min_alloc_order; in gen_pool_fixed_alloc()
724 unsigned int nr, void *data, struct gen_pool *pool, in gen_pool_first_fit_order_align() argument
748 struct gen_pool *pool, unsigned long start_addr) in gen_pool_best_fit() argument
824 struct gen_pool **ptr, *pool; in devm_gen_pool_create() local
841 pool = gen_pool_create(min_alloc_order, nid); in devm_gen_pool_create()
842 if (!pool) in devm_gen_pool_create()
845 *ptr = pool; in devm_gen_pool_create()
846 pool->name = pool_name; in devm_gen_pool_create()
849 return pool; in devm_gen_pool_create()
877 struct gen_pool *pool = NULL; in of_gen_pool_get() local
895 pool = gen_pool_get(&pdev->dev, name); in of_gen_pool_get()
898 return pool; in of_gen_pool_get()