Lines Matching refs:pool
154 struct gen_pool *pool; in gen_pool_create() local
156 pool = kmalloc_node(sizeof(struct gen_pool), GFP_KERNEL, nid); in gen_pool_create()
157 if (pool != NULL) { in gen_pool_create()
158 spin_lock_init(&pool->lock); in gen_pool_create()
159 INIT_LIST_HEAD(&pool->chunks); in gen_pool_create()
160 pool->min_alloc_order = min_alloc_order; in gen_pool_create()
161 pool->algo = gen_pool_first_fit; in gen_pool_create()
162 pool->data = NULL; in gen_pool_create()
163 pool->name = NULL; in gen_pool_create()
165 return pool; in gen_pool_create()
183 int gen_pool_add_owner(struct gen_pool *pool, unsigned long virt, phys_addr_t phys, in gen_pool_add_owner() argument
187 unsigned long nbits = size >> pool->min_alloc_order; in gen_pool_add_owner()
201 spin_lock(&pool->lock); in gen_pool_add_owner()
202 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_owner()
203 spin_unlock(&pool->lock); in gen_pool_add_owner()
216 phys_addr_t gen_pool_virt_to_phys(struct gen_pool *pool, unsigned long addr) in gen_pool_virt_to_phys() argument
222 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_virt_to_phys()
241 void gen_pool_destroy(struct gen_pool *pool) in gen_pool_destroy() argument
245 int order = pool->min_alloc_order; in gen_pool_destroy()
248 list_for_each_safe(_chunk, _next_chunk, &pool->chunks) { in gen_pool_destroy()
258 kfree_const(pool->name); in gen_pool_destroy()
259 kfree(pool); in gen_pool_destroy()
276 unsigned long gen_pool_alloc_algo_owner(struct gen_pool *pool, size_t size, in gen_pool_alloc_algo_owner() argument
281 int order = pool->min_alloc_order; in gen_pool_alloc_algo_owner()
296 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_alloc_algo_owner()
304 nbits, data, pool, chunk->start_addr); in gen_pool_alloc_algo_owner()
340 void *gen_pool_dma_alloc(struct gen_pool *pool, size_t size, dma_addr_t *dma) in gen_pool_dma_alloc() argument
342 return gen_pool_dma_alloc_algo(pool, size, dma, pool->algo, pool->data); in gen_pool_dma_alloc()
361 void *gen_pool_dma_alloc_algo(struct gen_pool *pool, size_t size, in gen_pool_dma_alloc_algo() argument
366 if (!pool) in gen_pool_dma_alloc_algo()
369 vaddr = gen_pool_alloc_algo(pool, size, algo, data); in gen_pool_dma_alloc_algo()
374 *dma = gen_pool_virt_to_phys(pool, vaddr); in gen_pool_dma_alloc_algo()
394 void *gen_pool_dma_alloc_align(struct gen_pool *pool, size_t size, in gen_pool_dma_alloc_align() argument
399 return gen_pool_dma_alloc_algo(pool, size, dma, in gen_pool_dma_alloc_align()
418 void *gen_pool_dma_zalloc(struct gen_pool *pool, size_t size, dma_addr_t *dma) in gen_pool_dma_zalloc() argument
420 return gen_pool_dma_zalloc_algo(pool, size, dma, pool->algo, pool->data); in gen_pool_dma_zalloc()
439 void *gen_pool_dma_zalloc_algo(struct gen_pool *pool, size_t size, in gen_pool_dma_zalloc_algo() argument
442 void *vaddr = gen_pool_dma_alloc_algo(pool, size, dma, algo, data); in gen_pool_dma_zalloc_algo()
465 void *gen_pool_dma_zalloc_align(struct gen_pool *pool, size_t size, in gen_pool_dma_zalloc_align() argument
470 return gen_pool_dma_zalloc_algo(pool, size, dma, in gen_pool_dma_zalloc_align()
486 void gen_pool_free_owner(struct gen_pool *pool, unsigned long addr, size_t size, in gen_pool_free_owner() argument
490 int order = pool->min_alloc_order; in gen_pool_free_owner()
502 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_free_owner()
530 void gen_pool_for_each_chunk(struct gen_pool *pool, in gen_pool_for_each_chunk() argument
531 void (*func)(struct gen_pool *pool, struct gen_pool_chunk *chunk, void *data), in gen_pool_for_each_chunk() argument
537 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) in gen_pool_for_each_chunk()
538 func(pool, chunk, data); in gen_pool_for_each_chunk()
552 bool gen_pool_has_addr(struct gen_pool *pool, unsigned long start, in gen_pool_has_addr() argument
560 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) { in gen_pool_has_addr()
579 size_t gen_pool_avail(struct gen_pool *pool) in gen_pool_avail() argument
585 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_avail()
598 size_t gen_pool_size(struct gen_pool *pool) in gen_pool_size() argument
604 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_size()
621 void gen_pool_set_algo(struct gen_pool *pool, genpool_algo_t algo, void *data) in gen_pool_set_algo() argument
625 pool->algo = algo; in gen_pool_set_algo()
626 if (!pool->algo) in gen_pool_set_algo()
627 pool->algo = gen_pool_first_fit; in gen_pool_set_algo()
629 pool->data = data; in gen_pool_set_algo()
647 struct gen_pool *pool, unsigned long start_addr) in gen_pool_first_fit() argument
665 struct gen_pool *pool, unsigned long start_addr) in gen_pool_first_fit_align() argument
672 order = pool->min_alloc_order; in gen_pool_first_fit_align()
692 struct gen_pool *pool, unsigned long start_addr) in gen_pool_fixed_alloc() argument
700 order = pool->min_alloc_order; in gen_pool_fixed_alloc()
726 unsigned int nr, void *data, struct gen_pool *pool, in gen_pool_first_fit_order_align() argument
750 struct gen_pool *pool, unsigned long start_addr) in gen_pool_best_fit() argument
826 struct gen_pool **ptr, *pool; in devm_gen_pool_create() local
843 pool = gen_pool_create(min_alloc_order, nid); in devm_gen_pool_create()
844 if (!pool) in devm_gen_pool_create()
847 *ptr = pool; in devm_gen_pool_create()
848 pool->name = pool_name; in devm_gen_pool_create()
851 return pool; in devm_gen_pool_create()
879 struct gen_pool *pool = NULL; in of_gen_pool_get() local
897 pool = gen_pool_get(&pdev->dev, name); in of_gen_pool_get()
900 return pool; in of_gen_pool_get()