Home
last modified time | relevance | path

Searched refs:pool (Results 1 – 8 of 8) sorted by relevance

/mm/
Dmempool.c25 static void poison_error(mempool_t *pool, void *element, size_t size, in poison_error() argument
28 const int nr = pool->curr_nr; in poison_error()
34 pr_err("Mempool %p size %zu\n", pool, size); in poison_error()
42 static void __check_element(mempool_t *pool, void *element, size_t size) in __check_element() argument
51 poison_error(pool, element, size, i); in __check_element()
58 static void check_element(mempool_t *pool, void *element) in check_element() argument
61 if (pool->free == mempool_free_slab || pool->free == mempool_kfree) { in check_element()
62 __check_element(pool, element, ksize(element)); in check_element()
63 } else if (pool->free == mempool_free_pages) { in check_element()
65 int order = (int)(long)pool->pool_data; in check_element()
[all …]
Dzbud.c78 int (*evict)(struct zbud_pool *pool, unsigned long handle);
222 struct zbud_pool *pool; in zbud_create_pool() local
225 pool = kzalloc(sizeof(struct zbud_pool), gfp); in zbud_create_pool()
226 if (!pool) in zbud_create_pool()
228 spin_lock_init(&pool->lock); in zbud_create_pool()
230 INIT_LIST_HEAD(&pool->unbuddied[i]); in zbud_create_pool()
231 INIT_LIST_HEAD(&pool->buddied); in zbud_create_pool()
232 INIT_LIST_HEAD(&pool->lru); in zbud_create_pool()
233 pool->pages_nr = 0; in zbud_create_pool()
234 pool->ops = ops; in zbud_create_pool()
[all …]
Ddmapool.c71 struct dma_pool *pool; in pools_show() local
81 list_for_each_entry(pool, &dev->dma_pools, pools) { in pools_show()
85 spin_lock_irq(&pool->lock); in pools_show()
86 list_for_each_entry(page, &pool->page_list, page_list) { in pools_show()
90 spin_unlock_irq(&pool->lock); in pools_show()
94 pool->name, blocks, in pools_show()
95 pages * (pool->allocation / pool->size), in pools_show()
96 pool->size, pages); in pools_show()
203 static void pool_initialise_page(struct dma_pool *pool, struct dma_page *page) in pool_initialise_page() argument
206 unsigned int next_boundary = pool->boundary; in pool_initialise_page()
[all …]
Dz3fold.c76 int (*evict)(struct z3fold_pool *pool, unsigned long handle);
93 unsigned long pool; /* back link */ member
121 struct z3fold_pool *pool; member
212 static inline struct z3fold_buddy_slots *alloc_slots(struct z3fold_pool *pool, in alloc_slots() argument
217 slots = kmem_cache_zalloc(pool->c_handle, in alloc_slots()
223 slots->pool = (unsigned long)pool; in alloc_slots()
232 return (struct z3fold_pool *)(s->pool & ~HANDLE_FLAG_MASK); in slots_to_pool()
310 if (test_bit(HANDLES_NOFREE, &slots->pool)) { in free_handle()
328 struct z3fold_pool *pool = slots_to_pool(slots); in free_handle() local
332 kmem_cache_free(pool->c_handle, slots); in free_handle()
[all …]
Dzswap.c176 struct zswap_pool *pool; member
223 static int zswap_writeback_entry(struct zpool *pool, unsigned long handle);
224 static int zswap_pool_get(struct zswap_pool *pool);
225 static void zswap_pool_put(struct zswap_pool *pool);
246 struct zswap_pool *pool; in zswap_update_total_size() local
251 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_update_total_size()
252 total += zpool_get_total_size(pool->zpool); in zswap_update_total_size()
355 zpool_free(entry->pool->zpool, entry->handle); in zswap_free_entry()
356 zswap_pool_put(entry->pool); in zswap_free_entry()
447 struct zswap_pool *pool = hlist_entry(node, struct zswap_pool, node); in zswap_cpu_comp_prepare() local
[all …]
Dzsmalloc.c302 static int zs_register_migration(struct zs_pool *pool);
303 static void zs_unregister_migration(struct zs_pool *pool);
307 static void kick_deferred_free(struct zs_pool *pool);
308 static void init_deferred_free(struct zs_pool *pool);
309 static void SetZsPageMovable(struct zs_pool *pool, struct zspage *zspage);
313 static int zs_register_migration(struct zs_pool *pool) { return 0; } in zs_register_migration() argument
314 static void zs_unregister_migration(struct zs_pool *pool) {} in zs_unregister_migration() argument
318 static void kick_deferred_free(struct zs_pool *pool) {} in kick_deferred_free() argument
319 static void init_deferred_free(struct zs_pool *pool) {} in init_deferred_free() argument
320 static void SetZsPageMovable(struct zs_pool *pool, struct zspage *zspage) {} in SetZsPageMovable() argument
[all …]
Dzpool.c23 void *pool; member
184 zpool->pool = driver->create(name, gfp, ops, zpool); in zpool_create_pool()
189 if (!zpool->pool) { in zpool_create_pool()
223 zpool->driver->destroy(zpool->pool); in zpool_destroy_pool()
278 return zpool->driver->malloc(zpool->pool, size, gfp, handle); in zpool_malloc()
297 zpool->driver->free(zpool->pool, handle); in zpool_free()
321 zpool->driver->shrink(zpool->pool, pages, reclaimed) : -EINVAL; in zpool_shrink()
349 return zpool->driver->map(zpool->pool, handle, mapmode); in zpool_map_handle()
364 zpool->driver->unmap(zpool->pool, handle); in zpool_unmap_handle()
377 return zpool->driver->total_size(zpool->pool); in zpool_get_total_size()
DKconfig538 compress them into a dynamically allocated RAM-based memory pool.