Home
last modified time | relevance | path

Searched refs:pool (Results 1 – 25 of 156) sorted by relevance

1234567

/drivers/staging/android/ion/
Dion_page_pool.c27 static void *ion_page_pool_alloc_pages(struct ion_page_pool *pool) in ion_page_pool_alloc_pages() argument
29 struct page *page = alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages()
33 if (!pool->cached) in ion_page_pool_alloc_pages()
34 ion_pages_sync_for_device(NULL, page, PAGE_SIZE << pool->order, in ion_page_pool_alloc_pages()
39 static void ion_page_pool_free_pages(struct ion_page_pool *pool, in ion_page_pool_free_pages() argument
42 __free_pages(page, pool->order); in ion_page_pool_free_pages()
45 static int ion_page_pool_add(struct ion_page_pool *pool, struct page *page) in ion_page_pool_add() argument
47 mutex_lock(&pool->mutex); in ion_page_pool_add()
49 list_add_tail(&page->lru, &pool->high_items); in ion_page_pool_add()
50 pool->high_count++; in ion_page_pool_add()
[all …]
Dion_system_heap.c68 struct ion_page_pool *pool; in alloc_buffer_page() local
72 pool = heap->uncached_pools[order_to_index(order)]; in alloc_buffer_page()
74 pool = heap->cached_pools[order_to_index(order)]; in alloc_buffer_page()
76 page = ion_page_pool_alloc(pool); in alloc_buffer_page()
87 struct ion_page_pool *pool; in free_buffer_page() local
98 pool = heap->uncached_pools[order_to_index(order)]; in free_buffer_page()
100 pool = heap->cached_pools[order_to_index(order)]; in free_buffer_page()
102 ion_page_pool_free(pool, page); in free_buffer_page()
271 struct ion_page_pool *pool; in ion_system_heap_debug_show() local
274 pool = sys_heap->uncached_pools[i]; in ion_system_heap_debug_show()
[all …]
/drivers/md/
Ddm-thin.c223 struct pool { struct
278 static enum pool_mode get_pool_mode(struct pool *pool); argument
279 static void metadata_operation_failed(struct pool *pool, const char *op, int r);
286 struct pool *pool; member
306 struct pool *pool; member
327 static bool block_size_is_power_of_two(struct pool *pool) in block_size_is_power_of_two() argument
329 return pool->sectors_per_block_shift >= 0; in block_size_is_power_of_two()
332 static sector_t block_to_sectors(struct pool *pool, dm_block_t b) in block_to_sectors() argument
334 return block_size_is_power_of_two(pool) ? in block_to_sectors()
335 (b << pool->sectors_per_block_shift) : in block_to_sectors()
[all …]
/drivers/infiniband/sw/rxe/
Drxe_pool.c105 static inline char *pool_name(struct rxe_pool *pool) in pool_name() argument
107 return rxe_type_info[pool->type].name; in pool_name()
110 static inline struct kmem_cache *pool_cache(struct rxe_pool *pool) in pool_cache() argument
112 return rxe_type_info[pool->type].cache; in pool_cache()
119 return elem->pool->type; in rxe_type()
166 static int rxe_pool_init_index(struct rxe_pool *pool, u32 max, u32 min) in rxe_pool_init_index() argument
171 if ((max - min + 1) < pool->max_elem) { in rxe_pool_init_index()
177 pool->max_index = max; in rxe_pool_init_index()
178 pool->min_index = min; in rxe_pool_init_index()
181 pool->table = kmalloc(size, GFP_KERNEL); in rxe_pool_init_index()
[all …]
/drivers/infiniband/core/
Dfmr_pool.c95 void (*flush_function)(struct ib_fmr_pool *pool,
114 static inline struct ib_pool_fmr *ib_fmr_cache_lookup(struct ib_fmr_pool *pool, in ib_fmr_cache_lookup() argument
122 if (!pool->cache_bucket) in ib_fmr_cache_lookup()
125 bucket = pool->cache_bucket + ib_fmr_hash(*page_list); in ib_fmr_cache_lookup()
137 static void ib_fmr_batch_release(struct ib_fmr_pool *pool) in ib_fmr_batch_release() argument
144 spin_lock_irq(&pool->pool_lock); in ib_fmr_batch_release()
146 list_for_each_entry(fmr, &pool->dirty_list, list) { in ib_fmr_batch_release()
159 list_splice_init(&pool->dirty_list, &unmap_list); in ib_fmr_batch_release()
160 pool->dirty_len = 0; in ib_fmr_batch_release()
162 spin_unlock_irq(&pool->pool_lock); in ib_fmr_batch_release()
[all …]
/drivers/gpu/drm/ttm/
Dttm_page_alloc_dma.c157 struct dma_pool *pool; member
306 static int ttm_set_pages_caching(struct dma_pool *pool, in ttm_set_pages_caching() argument
311 if (pool->type & IS_UC) { in ttm_set_pages_caching()
315 pool->dev_name, cpages); in ttm_set_pages_caching()
317 if (pool->type & IS_WC) { in ttm_set_pages_caching()
321 pool->dev_name, cpages); in ttm_set_pages_caching()
326 static void __ttm_dma_free_page(struct dma_pool *pool, struct dma_page *d_page) in __ttm_dma_free_page() argument
329 dma_free_coherent(pool->dev, pool->size, d_page->vaddr, dma); in __ttm_dma_free_page()
334 static struct dma_page *__ttm_dma_alloc_page(struct dma_pool *pool) in __ttm_dma_alloc_page() argument
342 d_page->vaddr = dma_alloc_coherent(pool->dev, pool->size, in __ttm_dma_alloc_page()
[all …]
Dttm_page_alloc.c285 static void ttm_pool_update_free_locked(struct ttm_page_pool *pool, in ttm_pool_update_free_locked() argument
288 pool->npages -= freed_pages; in ttm_pool_update_free_locked()
289 pool->nfrees += freed_pages; in ttm_pool_update_free_locked()
302 static int ttm_page_pool_free(struct ttm_page_pool *pool, unsigned nr_free, in ttm_page_pool_free() argument
326 spin_lock_irqsave(&pool->lock, irq_flags); in ttm_page_pool_free()
328 list_for_each_entry_reverse(p, &pool->list, lru) { in ttm_page_pool_free()
336 __list_del(p->lru.prev, &pool->list); in ttm_page_pool_free()
338 ttm_pool_update_free_locked(pool, freed_pages); in ttm_page_pool_free()
343 spin_unlock_irqrestore(&pool->lock, irq_flags); in ttm_page_pool_free()
371 __list_del(&p->lru, &pool->list); in ttm_page_pool_free()
[all …]
/drivers/staging/lustre/lustre/lov/
Dlov_pool.c51 static void lov_pool_getref(struct pool_desc *pool) in lov_pool_getref() argument
53 CDEBUG(D_INFO, "pool %p\n", pool); in lov_pool_getref()
54 atomic_inc(&pool->pool_refcount); in lov_pool_getref()
57 void lov_pool_putref(struct pool_desc *pool) in lov_pool_putref() argument
59 CDEBUG(D_INFO, "pool %p\n", pool); in lov_pool_putref()
60 if (atomic_dec_and_test(&pool->pool_refcount)) { in lov_pool_putref()
61 LASSERT(hlist_unhashed(&pool->pool_hash)); in lov_pool_putref()
62 LASSERT(list_empty(&pool->pool_list)); in lov_pool_putref()
63 LASSERT(!pool->pool_debugfs_entry); in lov_pool_putref()
64 lov_ost_pool_free(&pool->pool_obds); in lov_pool_putref()
[all …]
/drivers/dma/
Dcoh901318_lli.c19 #define DEBUGFS_POOL_COUNTER_RESET(pool) (pool->debugfs_pool_counter = 0) argument
20 #define DEBUGFS_POOL_COUNTER_ADD(pool, add) (pool->debugfs_pool_counter += add) argument
22 #define DEBUGFS_POOL_COUNTER_RESET(pool) argument
23 #define DEBUGFS_POOL_COUNTER_ADD(pool, add) argument
35 int coh901318_pool_create(struct coh901318_pool *pool, in coh901318_pool_create() argument
39 spin_lock_init(&pool->lock); in coh901318_pool_create()
40 pool->dev = dev; in coh901318_pool_create()
41 pool->dmapool = dma_pool_create("lli_pool", dev, size, align, 0); in coh901318_pool_create()
43 DEBUGFS_POOL_COUNTER_RESET(pool); in coh901318_pool_create()
47 int coh901318_pool_destroy(struct coh901318_pool *pool) in coh901318_pool_destroy() argument
[all …]
/drivers/staging/octeon/
Dethernet-mem.c30 static int cvm_oct_fill_hw_skbuff(int pool, int size, int elements) in cvm_oct_fill_hw_skbuff() argument
41 cvmx_fpa_free(skb->data, pool, size / 128); in cvm_oct_fill_hw_skbuff()
53 static void cvm_oct_free_hw_skbuff(int pool, int size, int elements) in cvm_oct_free_hw_skbuff() argument
58 memory = cvmx_fpa_alloc(pool); in cvm_oct_free_hw_skbuff()
69 pool, elements); in cvm_oct_free_hw_skbuff()
72 pool, elements); in cvm_oct_free_hw_skbuff()
83 static int cvm_oct_fill_hw_memory(int pool, int size, int elements) in cvm_oct_fill_hw_memory() argument
103 elements * size, pool); in cvm_oct_fill_hw_memory()
108 cvmx_fpa_free(fpa, pool, 0); in cvm_oct_fill_hw_memory()
120 static void cvm_oct_free_hw_memory(int pool, int size, int elements) in cvm_oct_free_hw_memory() argument
[all …]
/drivers/net/ethernet/marvell/
Dmvneta_bm.h33 #define MVNETA_BM_XBAR_POOL_REG(pool) \ argument
34 (((pool) < 2) ? MVNETA_BM_XBAR_01_REG : MVNETA_BM_XBAR_23_REG)
35 #define MVNETA_BM_TARGET_ID_OFFS(pool) (((pool) & 1) ? 16 : 0) argument
36 #define MVNETA_BM_TARGET_ID_MASK(pool) \ argument
37 (0xf << MVNETA_BM_TARGET_ID_OFFS(pool))
38 #define MVNETA_BM_TARGET_ID_VAL(pool, id) \ argument
39 ((id) << MVNETA_BM_TARGET_ID_OFFS(pool))
40 #define MVNETA_BM_XBAR_ATTR_OFFS(pool) (((pool) & 1) ? 20 : 4) argument
41 #define MVNETA_BM_XBAR_ATTR_MASK(pool) \ argument
42 (0xff << MVNETA_BM_XBAR_ATTR_OFFS(pool))
[all …]
/drivers/net/ethernet/ti/
Ddavinci_cpdma.c104 struct cpdma_desc_pool *pool; member
154 static void cpdma_desc_pool_destroy(struct cpdma_desc_pool *pool) in cpdma_desc_pool_destroy() argument
156 if (!pool) in cpdma_desc_pool_destroy()
159 WARN(gen_pool_size(pool->gen_pool) != gen_pool_avail(pool->gen_pool), in cpdma_desc_pool_destroy()
161 gen_pool_size(pool->gen_pool), in cpdma_desc_pool_destroy()
162 gen_pool_avail(pool->gen_pool)); in cpdma_desc_pool_destroy()
163 if (pool->cpumap) in cpdma_desc_pool_destroy()
164 dma_free_coherent(pool->dev, pool->mem_size, pool->cpumap, in cpdma_desc_pool_destroy()
165 pool->phys); in cpdma_desc_pool_destroy()
167 iounmap(pool->iomap); in cpdma_desc_pool_destroy()
[all …]
/drivers/gpu/drm/i915/
Di915_gem_batch_pool.c48 struct i915_gem_batch_pool *pool) in i915_gem_batch_pool_init() argument
52 pool->engine = engine; in i915_gem_batch_pool_init()
54 for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++) in i915_gem_batch_pool_init()
55 INIT_LIST_HEAD(&pool->cache_list[n]); in i915_gem_batch_pool_init()
64 void i915_gem_batch_pool_fini(struct i915_gem_batch_pool *pool) in i915_gem_batch_pool_fini() argument
68 lockdep_assert_held(&pool->engine->i915->drm.struct_mutex); in i915_gem_batch_pool_fini()
70 for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++) { in i915_gem_batch_pool_fini()
74 &pool->cache_list[n], in i915_gem_batch_pool_fini()
78 INIT_LIST_HEAD(&pool->cache_list[n]); in i915_gem_batch_pool_fini()
96 i915_gem_batch_pool_get(struct i915_gem_batch_pool *pool, in i915_gem_batch_pool_get() argument
[all …]
/drivers/net/ethernet/ibm/
Dibmveth.c156 static void ibmveth_init_buffer_pool(struct ibmveth_buff_pool *pool, in ibmveth_init_buffer_pool() argument
160 pool->size = pool_size; in ibmveth_init_buffer_pool()
161 pool->index = pool_index; in ibmveth_init_buffer_pool()
162 pool->buff_size = buff_size; in ibmveth_init_buffer_pool()
163 pool->threshold = pool_size * 7 / 8; in ibmveth_init_buffer_pool()
164 pool->active = pool_active; in ibmveth_init_buffer_pool()
168 static int ibmveth_alloc_buffer_pool(struct ibmveth_buff_pool *pool) in ibmveth_alloc_buffer_pool() argument
172 pool->free_map = kmalloc(sizeof(u16) * pool->size, GFP_KERNEL); in ibmveth_alloc_buffer_pool()
174 if (!pool->free_map) in ibmveth_alloc_buffer_pool()
177 pool->dma_addr = kcalloc(pool->size, sizeof(dma_addr_t), GFP_KERNEL); in ibmveth_alloc_buffer_pool()
[all …]
/drivers/mtd/ubi/
Dfastmap-wl.c57 struct ubi_fm_pool *pool) in return_unused_pool_pebs() argument
62 for (i = pool->used; i < pool->size; i++) { in return_unused_pool_pebs()
63 e = ubi->lookuptbl[pool->pebs[i]]; in return_unused_pool_pebs()
122 struct ubi_fm_pool *pool = &ubi->fm_pool; in ubi_refill_pools() local
129 return_unused_pool_pebs(ubi, pool); in ubi_refill_pools()
132 pool->size = 0; in ubi_refill_pools()
136 if (pool->size < pool->max_size) { in ubi_refill_pools()
144 pool->pebs[pool->size] = e->pnum; in ubi_refill_pools()
145 pool->size++; in ubi_refill_pools()
169 pool->used = 0; in ubi_refill_pools()
[all …]
/drivers/scsi/
Dscsi.c137 struct scsi_host_cmd_pool *pool = shost->cmd_pool; in scsi_host_free_command() local
141 kmem_cache_free(pool->sense_slab, cmd->sense_buffer); in scsi_host_free_command()
142 kmem_cache_free(pool->cmd_slab, cmd); in scsi_host_free_command()
156 struct scsi_host_cmd_pool *pool = shost->cmd_pool; in scsi_host_alloc_command() local
159 cmd = kmem_cache_zalloc(pool->cmd_slab, gfp_mask | pool->gfp_mask); in scsi_host_alloc_command()
163 cmd->sense_buffer = kmem_cache_alloc(pool->sense_slab, in scsi_host_alloc_command()
164 gfp_mask | pool->gfp_mask); in scsi_host_alloc_command()
177 kmem_cache_free(pool->sense_slab, cmd->sense_buffer); in scsi_host_alloc_command()
179 kmem_cache_free(pool->cmd_slab, cmd); in scsi_host_alloc_command()
305 scsi_free_host_cmd_pool(struct scsi_host_cmd_pool *pool) in scsi_free_host_cmd_pool() argument
[all …]
/drivers/xen/
Dtmem.c167 static void tmem_cleancache_put_page(int pool, struct cleancache_filekey key, in tmem_cleancache_put_page() argument
173 if (pool < 0) in tmem_cleancache_put_page()
178 (void)xen_tmem_put_page((u32)pool, oid, ind, page); in tmem_cleancache_put_page()
181 static int tmem_cleancache_get_page(int pool, struct cleancache_filekey key, in tmem_cleancache_get_page() argument
189 if (pool < 0) in tmem_cleancache_get_page()
193 ret = xen_tmem_get_page((u32)pool, oid, ind, page); in tmem_cleancache_get_page()
200 static void tmem_cleancache_flush_page(int pool, struct cleancache_filekey key, in tmem_cleancache_flush_page() argument
206 if (pool < 0) in tmem_cleancache_flush_page()
210 (void)xen_tmem_flush_page((u32)pool, oid, ind); in tmem_cleancache_flush_page()
213 static void tmem_cleancache_flush_inode(int pool, struct cleancache_filekey key) in tmem_cleancache_flush_inode() argument
[all …]
/drivers/scsi/lpfc/
Dlpfc_mem.c83 struct lpfc_dma_pool *pool = &phba->lpfc_mbuf_safety_pool; in lpfc_mem_alloc() local
115 pool->elements = kmalloc(sizeof(struct lpfc_dmabuf) * in lpfc_mem_alloc()
117 if (!pool->elements) in lpfc_mem_alloc()
120 pool->max_count = 0; in lpfc_mem_alloc()
121 pool->current_count = 0; in lpfc_mem_alloc()
123 pool->elements[i].virt = pci_pool_alloc(phba->lpfc_mbuf_pool, in lpfc_mem_alloc()
124 GFP_KERNEL, &pool->elements[i].phys); in lpfc_mem_alloc()
125 if (!pool->elements[i].virt) in lpfc_mem_alloc()
127 pool->max_count++; in lpfc_mem_alloc()
128 pool->current_count++; in lpfc_mem_alloc()
[all …]
/drivers/atm/
Dambassador.c687 static int rx_give (amb_dev * dev, rx_in * rx, unsigned char pool) { in rx_give() argument
688 amb_rxq * rxq = &dev->rxq[pool]; in rx_give()
691 PRINTD (DBG_FLOW|DBG_RX, "rx_give %p[%hu]", dev, pool); in rx_give()
702 wr_mem (dev, offsetof(amb_mem, mb.adapter.rx_address[pool]), virt_to_bus (rxq->in.ptr)); in rx_give()
712 static int rx_take (amb_dev * dev, unsigned char pool) { in rx_take() argument
713 amb_rxq * rxq = &dev->rxq[pool]; in rx_take()
716 PRINTD (DBG_FLOW|DBG_RX, "rx_take %p[%hu]", dev, pool); in rx_take()
745 static void drain_rx_pool (amb_dev * dev, unsigned char pool) { in drain_rx_pool() argument
746 amb_rxq * rxq = &dev->rxq[pool]; in drain_rx_pool()
748 PRINTD (DBG_FLOW|DBG_POOL, "drain_rx_pool %p %hu", dev, pool); in drain_rx_pool()
[all …]
/drivers/s390/scsi/
Dzfcp_aux.c204 adapter->pool.erp_req = in zfcp_allocate_low_mem_buffers()
206 if (!adapter->pool.erp_req) in zfcp_allocate_low_mem_buffers()
209 adapter->pool.gid_pn_req = in zfcp_allocate_low_mem_buffers()
211 if (!adapter->pool.gid_pn_req) in zfcp_allocate_low_mem_buffers()
214 adapter->pool.scsi_req = in zfcp_allocate_low_mem_buffers()
216 if (!adapter->pool.scsi_req) in zfcp_allocate_low_mem_buffers()
219 adapter->pool.scsi_abort = in zfcp_allocate_low_mem_buffers()
221 if (!adapter->pool.scsi_abort) in zfcp_allocate_low_mem_buffers()
224 adapter->pool.status_read_req = in zfcp_allocate_low_mem_buffers()
227 if (!adapter->pool.status_read_req) in zfcp_allocate_low_mem_buffers()
[all …]
/drivers/net/ethernet/chelsio/libcxgb/
Dlibcxgb_ppm.c122 struct cxgbi_ppm_pool *pool; in ppm_get_cpu_entries() local
127 pool = per_cpu_ptr(ppm->pool, cpu); in ppm_get_cpu_entries()
128 spin_lock_bh(&pool->lock); in ppm_get_cpu_entries()
131 i = ppm_find_unused_entries(pool->bmap, ppm->pool_index_max, in ppm_get_cpu_entries()
132 pool->next, count, 0); in ppm_get_cpu_entries()
134 pool->next = 0; in ppm_get_cpu_entries()
135 spin_unlock_bh(&pool->lock); in ppm_get_cpu_entries()
139 pool->next = i + count; in ppm_get_cpu_entries()
140 if (pool->next >= ppm->pool_index_max) in ppm_get_cpu_entries()
141 pool->next = 0; in ppm_get_cpu_entries()
[all …]
/drivers/net/ethernet/mellanox/mlxsw/
Dspectrum_buffers.c47 u8 pool, in mlxsw_sp_sb_pr_get() argument
50 return &mlxsw_sp->sb.prs[dir][pool]; in mlxsw_sp_sb_pr_get()
61 u8 local_port, u8 pool, in mlxsw_sp_sb_pm_get() argument
64 return &mlxsw_sp->sb.ports[local_port].pms[dir][pool]; in mlxsw_sp_sb_pm_get()
67 static int mlxsw_sp_sb_pr_write(struct mlxsw_sp *mlxsw_sp, u8 pool, in mlxsw_sp_sb_pr_write() argument
75 mlxsw_reg_sbpr_pack(sbpr_pl, pool, dir, mode, size); in mlxsw_sp_sb_pr_write()
80 pr = mlxsw_sp_sb_pr_get(mlxsw_sp, pool, dir); in mlxsw_sp_sb_pr_write()
88 u32 min_buff, u32 max_buff, u8 pool) in mlxsw_sp_sb_cm_write() argument
94 min_buff, max_buff, pool); in mlxsw_sp_sb_cm_write()
104 cm->pool = pool; in mlxsw_sp_sb_cm_write()
[all …]
/drivers/soc/ti/
Dknav_qmss_queue.c676 static void kdesc_fill_pool(struct knav_pool *pool) in kdesc_fill_pool() argument
681 region = pool->region; in kdesc_fill_pool()
682 pool->desc_size = region->desc_size; in kdesc_fill_pool()
683 for (i = 0; i < pool->num_desc; i++) { in kdesc_fill_pool()
684 int index = pool->region_offset + i; in kdesc_fill_pool()
688 dma_size = ALIGN(pool->desc_size, SMP_CACHE_BYTES); in kdesc_fill_pool()
689 dma_sync_single_for_device(pool->dev, dma_addr, dma_size, in kdesc_fill_pool()
691 knav_queue_push(pool->queue, dma_addr, dma_size, 0); in kdesc_fill_pool()
696 static void kdesc_empty_pool(struct knav_pool *pool) in kdesc_empty_pool() argument
703 if (!pool->queue) in kdesc_empty_pool()
[all …]
/drivers/scsi/megaraid/
Dmegaraid_mm.c518 mm_dmapool_t *pool; in mraid_mm_attach_buf() local
535 pool = &adp->dma_pool_list[i]; in mraid_mm_attach_buf()
537 if (xferlen > pool->buf_size) in mraid_mm_attach_buf()
543 spin_lock_irqsave(&pool->lock, flags); in mraid_mm_attach_buf()
545 if (!pool->in_use) { in mraid_mm_attach_buf()
547 pool->in_use = 1; in mraid_mm_attach_buf()
549 kioc->buf_vaddr = pool->vaddr; in mraid_mm_attach_buf()
550 kioc->buf_paddr = pool->paddr; in mraid_mm_attach_buf()
552 spin_unlock_irqrestore(&pool->lock, flags); in mraid_mm_attach_buf()
556 spin_unlock_irqrestore(&pool->lock, flags); in mraid_mm_attach_buf()
[all …]
/drivers/tee/
Dtee_shm_pool.c164 struct tee_shm_pool *pool; in tee_shm_pool_alloc() local
169 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in tee_shm_pool_alloc()
170 if (!pool) in tee_shm_pool_alloc()
173 pool->private_mgr = priv_mgr; in tee_shm_pool_alloc()
174 pool->dma_buf_mgr = dmabuf_mgr; in tee_shm_pool_alloc()
176 return pool; in tee_shm_pool_alloc()
187 void tee_shm_pool_free(struct tee_shm_pool *pool) in tee_shm_pool_free() argument
189 if (pool->private_mgr) in tee_shm_pool_free()
190 tee_shm_pool_mgr_destroy(pool->private_mgr); in tee_shm_pool_free()
191 if (pool->dma_buf_mgr) in tee_shm_pool_free()
[all …]

1234567