/mm/ |
D | mempool.c | 25 static void poison_error(mempool_t *pool, void *element, size_t size, in poison_error() argument 28 const int nr = pool->curr_nr; in poison_error() 34 pr_err("Mempool %p size %zu\n", pool, size); in poison_error() 42 static void __check_element(mempool_t *pool, void *element, size_t size) in __check_element() argument 51 poison_error(pool, element, size, i); in __check_element() 58 static void check_element(mempool_t *pool, void *element) in check_element() argument 61 if (pool->free == mempool_free_slab || pool->free == mempool_kfree) { in check_element() 62 __check_element(pool, element, ksize(element)); in check_element() 63 } else if (pool->free == mempool_free_pages) { in check_element() 65 int order = (int)(long)pool->pool_data; in check_element() [all …]
|
D | zbud.c | 78 int (*evict)(struct zbud_pool *pool, unsigned long handle); 222 struct zbud_pool *pool; in zbud_create_pool() local 225 pool = kzalloc(sizeof(struct zbud_pool), gfp); in zbud_create_pool() 226 if (!pool) in zbud_create_pool() 228 spin_lock_init(&pool->lock); in zbud_create_pool() 230 INIT_LIST_HEAD(&pool->unbuddied[i]); in zbud_create_pool() 231 INIT_LIST_HEAD(&pool->buddied); in zbud_create_pool() 232 INIT_LIST_HEAD(&pool->lru); in zbud_create_pool() 233 pool->pages_nr = 0; in zbud_create_pool() 234 pool->ops = ops; in zbud_create_pool() [all …]
|
D | dmapool.c | 71 struct dma_pool *pool; in pools_show() local 81 list_for_each_entry(pool, &dev->dma_pools, pools) { in pools_show() 85 spin_lock_irq(&pool->lock); in pools_show() 86 list_for_each_entry(page, &pool->page_list, page_list) { in pools_show() 90 spin_unlock_irq(&pool->lock); in pools_show() 94 pool->name, blocks, in pools_show() 95 pages * (pool->allocation / pool->size), in pools_show() 96 pool->size, pages); in pools_show() 203 static void pool_initialise_page(struct dma_pool *pool, struct dma_page *page) in pool_initialise_page() argument 206 unsigned int next_boundary = pool->boundary; in pool_initialise_page() [all …]
|
D | z3fold.c | 76 int (*evict)(struct z3fold_pool *pool, unsigned long handle); 93 unsigned long pool; /* back link */ member 121 struct z3fold_pool *pool; member 212 static inline struct z3fold_buddy_slots *alloc_slots(struct z3fold_pool *pool, in alloc_slots() argument 217 slots = kmem_cache_zalloc(pool->c_handle, in alloc_slots() 223 slots->pool = (unsigned long)pool; in alloc_slots() 232 return (struct z3fold_pool *)(s->pool & ~HANDLE_FLAG_MASK); in slots_to_pool() 310 if (test_bit(HANDLES_NOFREE, &slots->pool)) { in free_handle() 328 struct z3fold_pool *pool = slots_to_pool(slots); in free_handle() local 332 kmem_cache_free(pool->c_handle, slots); in free_handle() [all …]
|
D | zswap.c | 176 struct zswap_pool *pool; member 223 static int zswap_writeback_entry(struct zpool *pool, unsigned long handle); 224 static int zswap_pool_get(struct zswap_pool *pool); 225 static void zswap_pool_put(struct zswap_pool *pool); 246 struct zswap_pool *pool; in zswap_update_total_size() local 251 list_for_each_entry_rcu(pool, &zswap_pools, list) in zswap_update_total_size() 252 total += zpool_get_total_size(pool->zpool); in zswap_update_total_size() 355 zpool_free(entry->pool->zpool, entry->handle); in zswap_free_entry() 356 zswap_pool_put(entry->pool); in zswap_free_entry() 447 struct zswap_pool *pool = hlist_entry(node, struct zswap_pool, node); in zswap_cpu_comp_prepare() local [all …]
|
D | zsmalloc.c | 302 static int zs_register_migration(struct zs_pool *pool); 303 static void zs_unregister_migration(struct zs_pool *pool); 307 static void kick_deferred_free(struct zs_pool *pool); 308 static void init_deferred_free(struct zs_pool *pool); 309 static void SetZsPageMovable(struct zs_pool *pool, struct zspage *zspage); 313 static int zs_register_migration(struct zs_pool *pool) { return 0; } in zs_register_migration() argument 314 static void zs_unregister_migration(struct zs_pool *pool) {} in zs_unregister_migration() argument 318 static void kick_deferred_free(struct zs_pool *pool) {} in kick_deferred_free() argument 319 static void init_deferred_free(struct zs_pool *pool) {} in init_deferred_free() argument 320 static void SetZsPageMovable(struct zs_pool *pool, struct zspage *zspage) {} in SetZsPageMovable() argument [all …]
|
D | zpool.c | 23 void *pool; member 184 zpool->pool = driver->create(name, gfp, ops, zpool); in zpool_create_pool() 189 if (!zpool->pool) { in zpool_create_pool() 223 zpool->driver->destroy(zpool->pool); in zpool_destroy_pool() 278 return zpool->driver->malloc(zpool->pool, size, gfp, handle); in zpool_malloc() 297 zpool->driver->free(zpool->pool, handle); in zpool_free() 321 zpool->driver->shrink(zpool->pool, pages, reclaimed) : -EINVAL; in zpool_shrink() 349 return zpool->driver->map(zpool->pool, handle, mapmode); in zpool_map_handle() 364 zpool->driver->unmap(zpool->pool, handle); in zpool_unmap_handle() 377 return zpool->driver->total_size(zpool->pool); in zpool_get_total_size()
|
D | Kconfig | 538 compress them into a dynamically allocated RAM-based memory pool.
|