/drivers/infiniband/core/ |
D | fmr_pool.c | 86 int pool_size; member 259 pool->pool_size = 0; in ib_create_fmr_pool() 291 for (i = 0; i < params->pool_size; ++i) { in ib_create_fmr_pool() 310 ++pool->pool_size; in ib_create_fmr_pool() 358 if (i < pool->pool_size) in ib_destroy_fmr_pool() 360 pool->pool_size - i); in ib_destroy_fmr_pool()
|
/drivers/mtd/ubi/ |
D | fastmap.c | 411 __be32 *pebs, int pool_size, unsigned long long *max_sqnum, in scan_pool() argument 432 dbg_bld("scanning fastmap pool: size = %i", pool_size); in scan_pool() 438 for (i = 0; i < pool_size; i++) { in scan_pool() 569 int ret, i, j, pool_size, wl_pool_size; in ubi_attach_fastmap() local 615 pool_size = be16_to_cpu(fmpl->size); in ubi_attach_fastmap() 620 if (pool_size > UBI_FM_MAX_POOL_SIZE || pool_size < 0) { in ubi_attach_fastmap() 621 ubi_err(ubi, "bad pool size: %i", pool_size); in ubi_attach_fastmap() 765 ret = scan_pool(ubi, ai, fmpl->pebs, pool_size, &max_sqnum, &free); in ubi_attach_fastmap()
|
/drivers/dma/ppc4xx/ |
D | adma.h | 70 size_t pool_size; member
|
D | adma.c | 1797 for (; i < (ppc440spe_chan->device->pool_size / db_sz); i++) { in ppc440spe_adma_alloc_chan_resources() 4030 u32 id, pool_size; in ppc440spe_adma_probe() local 4038 pool_size = PAGE_SIZE << 1; in ppc440spe_adma_probe() 4060 pool_size = (id == PPC440SPE_DMA0_ID) ? in ppc440spe_adma_probe() 4062 pool_size <<= 2; in ppc440spe_adma_probe() 4090 adev->pool_size = pool_size; in ppc440spe_adma_probe() 4093 adev->pool_size, &adev->dma_desc_pool, in ppc440spe_adma_probe() 4098 adev->pool_size); in ppc440spe_adma_probe() 4232 dma_free_coherent(adev->dev, adev->pool_size, in ppc440spe_adma_probe() 4288 dma_free_coherent(adev->dev, adev->pool_size, in ppc440spe_adma_remove()
|
/drivers/md/ |
D | dm.c | 2563 unsigned int pool_size = 0; in dm_alloc_md_mempools() local 2573 pool_size = dm_get_reserved_bio_based_ios(); in dm_alloc_md_mempools() 2578 pool_size = dm_get_reserved_rq_based_ios(); in dm_alloc_md_mempools() 2579 pools->rq_pool = mempool_create_slab_pool(pool_size, _rq_cache); in dm_alloc_md_mempools() 2584 if (!pool_size) in dm_alloc_md_mempools() 2585 pool_size = dm_get_reserved_rq_based_ios(); in dm_alloc_md_mempools() 2594 pools->io_pool = mempool_create_slab_pool(pool_size, cachep); in dm_alloc_md_mempools() 2599 pools->bs = bioset_create_nobvec(pool_size, front_pad); in dm_alloc_md_mempools() 2603 if (integrity && bioset_integrity_create(pools->bs, pool_size)) in dm_alloc_md_mempools()
|
D | raid5.h | 540 int pool_size; /* number of disks in stripeheads in pool */ member
|
D | raid5.c | 485 int num = sh->raid_conf->pool_size; in shrink_buffers() 500 int num = sh->raid_conf->pool_size; in grow_buffers() 2028 sh = alloc_stripe(conf->slab_cache, gfp, conf->pool_size); in grow_one_stripe() 2069 conf->pool_size = devs; in grow_stripes() 2181 if (newsize <= conf->pool_size) in resize_stripes() 2232 for(i=0; i<conf->pool_size; i++) { in resize_stripes() 2284 conf->pool_size = newsize; in resize_stripes()
|
/drivers/char/ |
D | random.c | 646 const int pool_size = r->poolinfo->poolfracbits; in credit_entropy_bits() local 684 unsigned int anfrac = min(pnfrac, pool_size/2); in credit_entropy_bits() 686 ((pool_size - entropy_count)*anfrac*3) >> s; in credit_entropy_bits() 690 } while (unlikely(entropy_count < pool_size-2 && pnfrac)); in credit_entropy_bits() 698 } else if (entropy_count > pool_size) in credit_entropy_bits() 699 entropy_count = pool_size; in credit_entropy_bits()
|
/drivers/net/ethernet/neterion/vxge/ |
D | vxge-config.c | 1119 blockpool->pool_size--; in __vxge_hw_blockpool_destroy() 1137 u32 pool_size, in __vxge_hw_blockpool_create() argument 1155 blockpool->pool_size = 0; in __vxge_hw_blockpool_create() 1162 for (i = 0; i < pool_size + pool_max; i++) { in __vxge_hw_blockpool_create() 1173 for (i = 0; i < pool_size; i++) { in __vxge_hw_blockpool_create() 1214 blockpool->pool_size++; in __vxge_hw_blockpool_create() 2302 blockpool->pool_size++; in vxge_hw_blockpool_block_add() 2338 if ((blockpool->pool_size + blockpool->req_out) < in __vxge_hw_blockpool_blocks_add() 2404 blockpool->pool_size--; in __vxge_hw_blockpool_malloc() 2424 if (blockpool->pool_size < blockpool->pool_max) in __vxge_hw_blockpool_blocks_remove() [all …]
|
D | vxge-config.h | 555 u32 pool_size; member
|
/drivers/dma/ |
D | mv_xor.h | 125 size_t pool_size; member
|
D | iop-adma.c | 437 int num_descs_in_pool = plat_data->pool_size/IOP_ADMA_SLOT_SIZE; in iop_adma_alloc_chan_resources() 1262 dma_free_coherent(&dev->dev, plat_data->pool_size, in iop_adma_remove() 1303 plat_data->pool_size, in iop_adma_probe() 1438 dma_free_coherent(&adev->pdev->dev, plat_data->pool_size, in iop_adma_probe()
|
/drivers/net/ethernet/ibm/ |
D | ibmveth.h | 114 static int pool_size[] = { 512, 1024 * 2, 1024 * 16, 1024 * 32, 1024 * 64 }; variable
|
D | ibmveth.c | 157 u32 pool_index, u32 pool_size, in ibmveth_init_buffer_pool() argument 160 pool->size = pool_size; in ibmveth_init_buffer_pool() 163 pool->threshold = pool_size * 7 / 8; in ibmveth_init_buffer_pool() 1636 pool_count[i], pool_size[i], in ibmveth_probe()
|
/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_buffers.c | 611 u32 pool_size = MLXSW_SP_BYTES_TO_CELLS(size); in mlxsw_sp_sb_pool_set() local 615 return mlxsw_sp_sb_pr_write(mlxsw_sp, pool, dir, mode, pool_size); in mlxsw_sp_sb_pool_set()
|
/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 323 fmr_param.pool_size = target->mr_pool_size; in srp_alloc_fmr_pool() 324 fmr_param.dirty_watermark = fmr_param.pool_size / 4; in srp_alloc_fmr_pool() 362 struct ib_pd *pd, int pool_size, in srp_create_fr_pool() argument 370 if (pool_size <= 0) in srp_create_fr_pool() 374 pool_size * sizeof(struct srp_fr_desc), GFP_KERNEL); in srp_create_fr_pool() 377 pool->size = pool_size; in srp_create_fr_pool()
|
/drivers/scsi/libfc/ |
D | fc_exch.c | 2369 size_t pool_size; in fc_exch_mgr_alloc() local 2417 pool_size = sizeof(*pool) + pool_exch_range * sizeof(struct fc_exch *); in fc_exch_mgr_alloc() 2418 mp->pool = __alloc_percpu(pool_size, __alignof__(struct fc_exch_pool)); in fc_exch_mgr_alloc()
|
/drivers/atm/ |
D | fore200e.h | 659 u32 pool_size; /* number of rbds */ member
|
D | fore200e.c | 2372 int pool_size, int supply_blksize) in fore200e_param_bs_queue() argument 2378 fore200e->bus->write(pool_size, &bs_spec->pool_size); in fore200e_param_bs_queue()
|
/drivers/net/wireless/ath/ath10k/ |
D | wmi.c | 4477 u32 pool_size = 0; in ath10k_wmi_alloc_chunk() local 4485 pool_size = num_units * round_up(unit_len, 4); in ath10k_wmi_alloc_chunk() 4486 if (!pool_size) in ath10k_wmi_alloc_chunk() 4489 vaddr = kzalloc(pool_size, GFP_KERNEL | __GFP_NOWARN); in ath10k_wmi_alloc_chunk() 4497 paddr = dma_map_single(ar->dev, vaddr, pool_size, DMA_TO_DEVICE); in ath10k_wmi_alloc_chunk() 4505 ar->wmi.mem_chunks[idx].len = pool_size; in ath10k_wmi_alloc_chunk() 4534 u32 pool_size; in ath10k_wmi_is_host_mem_allocated() local 4561 pool_size = num_units * round_up(unit_size, 4); in ath10k_wmi_is_host_mem_allocated() 4562 if (ar->wmi.mem_chunks[j].len == pool_size) { in ath10k_wmi_is_host_mem_allocated()
|
/drivers/infiniband/ulp/iser/ |
D | iser_verbs.c | 189 params.pool_size = cmds_max * 2; in iser_alloc_fmr_pool()
|
/drivers/staging/lustre/lnet/klnds/o2iblnd/ |
D | o2iblnd.c | 1368 .pool_size = fps->fps_pool_size, in kiblnd_alloc_fmr_pool()
|