/drivers/md/ |
D | dm-thin.c | 232 struct pool { struct 290 static void metadata_operation_failed(struct pool *pool, const char *op, int r); argument 292 static enum pool_mode get_pool_mode(struct pool *pool) in get_pool_mode() argument 294 return pool->pf.mode; in get_pool_mode() 297 static void notify_of_pool_mode_change(struct pool *pool) in notify_of_pool_mode_change() argument 307 enum pool_mode mode = get_pool_mode(pool); in notify_of_pool_mode_change() 310 if (!pool->pf.error_if_no_space) in notify_of_pool_mode_change() 316 dm_table_event(pool->ti->table); in notify_of_pool_mode_change() 318 dm_device_name(pool->pool_md), in notify_of_pool_mode_change() 327 struct pool *pool; member [all …]
|
/drivers/net/ethernet/ti/ |
D | k3-cppi-desc-pool.c | 27 void k3_cppi_desc_pool_destroy(struct k3_cppi_desc_pool *pool) in k3_cppi_desc_pool_destroy() argument 29 if (!pool) in k3_cppi_desc_pool_destroy() 32 WARN(gen_pool_size(pool->gen_pool) != gen_pool_avail(pool->gen_pool), in k3_cppi_desc_pool_destroy() 34 gen_pool_size(pool->gen_pool), in k3_cppi_desc_pool_destroy() 35 gen_pool_avail(pool->gen_pool)); in k3_cppi_desc_pool_destroy() 36 if (pool->cpumem) in k3_cppi_desc_pool_destroy() 37 dma_free_coherent(pool->dev, pool->mem_size, pool->cpumem, in k3_cppi_desc_pool_destroy() 38 pool->dma_addr); in k3_cppi_desc_pool_destroy() 40 gen_pool_destroy(pool->gen_pool); /* frees pool->name */ in k3_cppi_desc_pool_destroy() 49 struct k3_cppi_desc_pool *pool; in k3_cppi_desc_pool_create_name() local [all …]
|
/drivers/dma-buf/heaps/ |
D | page_pool.c | 53 struct page *dmabuf_page_pool_alloc_pages(struct dmabuf_page_pool *pool) in dmabuf_page_pool_alloc_pages() argument 57 return alloc_pages(pool->gfp_mask, pool->order); in dmabuf_page_pool_alloc_pages() 60 static inline void dmabuf_page_pool_free_pages(struct dmabuf_page_pool *pool, in dmabuf_page_pool_free_pages() argument 63 __free_pages(page, pool->order); in dmabuf_page_pool_free_pages() 66 static void dmabuf_page_pool_add(struct dmabuf_page_pool *pool, struct page *page) in dmabuf_page_pool_add() argument 75 spin_lock(&pool->lock); in dmabuf_page_pool_add() 76 list_add_tail(&page->lru, &pool->items[index]); in dmabuf_page_pool_add() 77 pool->count[index]++; in dmabuf_page_pool_add() 78 spin_unlock(&pool->lock); in dmabuf_page_pool_add() 80 1 << pool->order); in dmabuf_page_pool_add() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_icm_pool.c | 82 u32 offset = mlx5dr_icm_pool_dm_type_to_entry_size(chunk->buddy_mem->pool->icm_type); in mlx5dr_icm_pool_get_chunk_mr_addr() 94 u32 size = mlx5dr_icm_pool_dm_type_to_entry_size(chunk->buddy_mem->pool->icm_type); in mlx5dr_icm_pool_get_chunk_icm_addr() 102 chunk->buddy_mem->pool->icm_type); in mlx5dr_icm_pool_get_chunk_byte_size() 111 dr_icm_pool_mr_create(struct mlx5dr_icm_pool *pool) in dr_icm_pool_mr_create() argument 113 struct mlx5_core_dev *mdev = pool->dmn->mdev; in dr_icm_pool_mr_create() 123 icm_mr->dmn = pool->dmn; in dr_icm_pool_mr_create() 125 icm_mr->dm.length = mlx5dr_icm_pool_chunk_size_to_byte(pool->max_log_chunk_sz, in dr_icm_pool_mr_create() 126 pool->icm_type); in dr_icm_pool_mr_create() 128 switch (pool->icm_type) { in dr_icm_pool_mr_create() 144 WARN_ON(pool->icm_type); in dr_icm_pool_mr_create() [all …]
|
D | dr_arg.c | 31 static int dr_arg_pool_alloc_objs(struct dr_arg_pool *pool) in dr_arg_pool_alloc_objs() argument 43 pool->dmn->info.caps.log_header_modify_argument_granularity; in dr_arg_pool_alloc_objs() 46 max_t(u32, pool->dmn->info.caps.log_header_modify_argument_granularity, in dr_arg_pool_alloc_objs() 49 min_t(u32, pool->dmn->info.caps.log_header_modify_argument_max_alloc, in dr_arg_pool_alloc_objs() 52 if (pool->log_chunk_size > object_range) { in dr_arg_pool_alloc_objs() 53 mlx5dr_err(pool->dmn, "Required chunk size (%d) is not supported\n", in dr_arg_pool_alloc_objs() 54 pool->log_chunk_size); in dr_arg_pool_alloc_objs() 58 num_of_objects = (1 << (object_range - pool->log_chunk_size)); in dr_arg_pool_alloc_objs() 60 ret = mlx5dr_cmd_create_modify_header_arg(pool->dmn->mdev, in dr_arg_pool_alloc_objs() 62 pool->dmn->pdn, in dr_arg_pool_alloc_objs() [all …]
|
/drivers/gpu/drm/amd/display/dc/dce80/ |
D | dce80_resource.c | 804 static void dce80_resource_destruct(struct dce110_resource_pool *pool) in dce80_resource_destruct() argument 808 for (i = 0; i < pool->base.pipe_count; i++) { in dce80_resource_destruct() 809 if (pool->base.opps[i] != NULL) in dce80_resource_destruct() 810 dce110_opp_destroy(&pool->base.opps[i]); in dce80_resource_destruct() 812 if (pool->base.transforms[i] != NULL) in dce80_resource_destruct() 813 dce80_transform_destroy(&pool->base.transforms[i]); in dce80_resource_destruct() 815 if (pool->base.ipps[i] != NULL) in dce80_resource_destruct() 816 dce_ipp_destroy(&pool->base.ipps[i]); in dce80_resource_destruct() 818 if (pool->base.mis[i] != NULL) { in dce80_resource_destruct() 819 kfree(TO_DCE_MEM_INPUT(pool->base.mis[i])); in dce80_resource_destruct() [all …]
|
/drivers/gpu/drm/amd/display/dc/dce60/ |
D | dce60_resource.c | 797 static void dce60_resource_destruct(struct dce110_resource_pool *pool) in dce60_resource_destruct() argument 801 for (i = 0; i < pool->base.pipe_count; i++) { in dce60_resource_destruct() 802 if (pool->base.opps[i] != NULL) in dce60_resource_destruct() 803 dce110_opp_destroy(&pool->base.opps[i]); in dce60_resource_destruct() 805 if (pool->base.transforms[i] != NULL) in dce60_resource_destruct() 806 dce60_transform_destroy(&pool->base.transforms[i]); in dce60_resource_destruct() 808 if (pool->base.ipps[i] != NULL) in dce60_resource_destruct() 809 dce_ipp_destroy(&pool->base.ipps[i]); in dce60_resource_destruct() 811 if (pool->base.mis[i] != NULL) { in dce60_resource_destruct() 812 kfree(TO_DCE_MEM_INPUT(pool->base.mis[i])); in dce60_resource_destruct() [all …]
|
/drivers/staging/media/atomisp/pci/runtime/rmgr/src/ |
D | rmgr_vbuf.c | 128 int ia_css_rmgr_init_vbuf(struct ia_css_rmgr_vbuf_pool *pool) in ia_css_rmgr_init_vbuf() argument 134 assert(pool); in ia_css_rmgr_init_vbuf() 135 if (!pool) in ia_css_rmgr_init_vbuf() 138 if (pool->recycle && pool->size) { in ia_css_rmgr_init_vbuf() 142 pool->size; in ia_css_rmgr_init_vbuf() 143 pool->handles = kvmalloc(bytes_needed, GFP_KERNEL); in ia_css_rmgr_init_vbuf() 144 if (pool->handles) in ia_css_rmgr_init_vbuf() 145 memset(pool->handles, 0, bytes_needed); in ia_css_rmgr_init_vbuf() 150 pool->size = 0; in ia_css_rmgr_init_vbuf() 151 pool->handles = NULL; in ia_css_rmgr_init_vbuf() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | irq_affinity.c | 8 static void cpu_put(struct mlx5_irq_pool *pool, int cpu) in cpu_put() argument 10 pool->irqs_per_cpu[cpu]--; in cpu_put() 13 static void cpu_get(struct mlx5_irq_pool *pool, int cpu) in cpu_get() argument 15 pool->irqs_per_cpu[cpu]++; in cpu_get() 19 static int cpu_get_least_loaded(struct mlx5_irq_pool *pool, in cpu_get_least_loaded() argument 27 if (!pool->irqs_per_cpu[cpu]) { in cpu_get_least_loaded() 33 if (pool->irqs_per_cpu[cpu] < pool->irqs_per_cpu[best_cpu]) in cpu_get_least_loaded() 38 mlx5_core_err(pool->dev, "NO online CPUs in req_mask (%*pbl)\n", in cpu_get_least_loaded() 42 pool->irqs_per_cpu[best_cpu]++; in cpu_get_least_loaded() 48 irq_pool_request_irq(struct mlx5_irq_pool *pool, struct irq_affinity_desc *af_desc) in irq_pool_request_irq() argument [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
D | crypto.c | 19 #define MLX5_CRYPTO_DEK_POOL_CALC_FREED(pool) MLX5_CRYPTO_DEK_CALC_FREED(pool) argument 288 mlx5_crypto_dek_bulk_create(struct mlx5_crypto_dek_pool *pool) in mlx5_crypto_dek_bulk_create() argument 290 struct mlx5_crypto_dek_priv *dek_priv = pool->mdev->mlx5e_res.dek_priv; in mlx5_crypto_dek_bulk_create() 291 struct mlx5_core_dev *mdev = pool->mdev; in mlx5_crypto_dek_bulk_create() 313 err = mlx5_crypto_create_dek_bulk(mdev, pool->key_purpose, in mlx5_crypto_dek_bulk_create() 334 mlx5_crypto_dek_pool_add_bulk(struct mlx5_crypto_dek_pool *pool) in mlx5_crypto_dek_pool_add_bulk() argument 338 bulk = mlx5_crypto_dek_bulk_create(pool); in mlx5_crypto_dek_pool_add_bulk() 342 pool->avail_deks += bulk->num_deks; in mlx5_crypto_dek_pool_add_bulk() 343 pool->num_deks += bulk->num_deks; in mlx5_crypto_dek_pool_add_bulk() 344 list_add(&bulk->entry, &pool->partial_list); in mlx5_crypto_dek_pool_add_bulk() [all …]
|
/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_cnt.c | 54 struct mlxsw_sp_counter_pool *pool = mlxsw_sp->counter_pool; in mlxsw_sp_counter_sub_pools_init() local 62 for (i = 0; i < pool->sub_pools_count; i++) { in mlxsw_sp_counter_sub_pools_init() 63 sub_pool = &pool->sub_pools[i]; in mlxsw_sp_counter_sub_pools_init() 89 sub_pool = &pool->sub_pools[i]; in mlxsw_sp_counter_sub_pools_init() 99 struct mlxsw_sp_counter_pool *pool = mlxsw_sp->counter_pool; in mlxsw_sp_counter_sub_pools_fini() local 104 for (i = 0; i < pool->sub_pools_count; i++) { in mlxsw_sp_counter_sub_pools_fini() 105 sub_pool = &pool->sub_pools[i]; in mlxsw_sp_counter_sub_pools_fini() 115 const struct mlxsw_sp_counter_pool *pool = priv; in mlxsw_sp_counter_pool_occ_get() local 117 return atomic_read(&pool->active_entries_count); in mlxsw_sp_counter_pool_occ_get() 124 struct mlxsw_sp_counter_pool *pool; in mlxsw_sp_counter_pool_init() local [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn302/ |
D | dcn302_resource.c | 705 static bool dcn302_dwbc_create(struct dc_context *ctx, struct resource_pool *pool) in dcn302_dwbc_create() argument 708 uint32_t pipe_count = pool->res_cap->num_dwb; in dcn302_dwbc_create() 720 pool->dwbc[i] = &dwbc30->base; in dcn302_dwbc_create() 740 static bool dcn302_mmhubbub_create(struct dc_context *ctx, struct resource_pool *pool) in dcn302_mmhubbub_create() argument 743 uint32_t pipe_count = pool->res_cap->num_dwb; in dcn302_mmhubbub_create() 755 pool->mcif_wb[i] = &mcif_wb30->base; in dcn302_mmhubbub_create() 951 static bool init_soc_bounding_box(struct dc *dc, struct resource_pool *pool) in init_soc_bounding_box() argument 963 loaded_ip->max_num_otg = pool->pipe_count; in init_soc_bounding_box() 964 loaded_ip->max_num_dpp = pool->pipe_count; in init_soc_bounding_box() 985 static void dcn302_resource_destruct(struct resource_pool *pool) in dcn302_resource_destruct() argument [all …]
|
/drivers/gpu/drm/ttm/tests/ |
D | ttm_pool_test.c | 79 struct ttm_pool *pool; in ttm_pool_pre_populated() local 87 pool = kunit_kzalloc(test, sizeof(*pool), GFP_KERNEL); in ttm_pool_pre_populated() 88 KUNIT_ASSERT_NOT_NULL(test, pool); in ttm_pool_pre_populated() 90 ttm_pool_init(pool, devs->dev, NUMA_NO_NODE, true, false); in ttm_pool_pre_populated() 92 err = ttm_pool_alloc(pool, tt, &simple_ctx); in ttm_pool_pre_populated() 95 ttm_pool_free(pool, tt); in ttm_pool_pre_populated() 98 return pool; in ttm_pool_pre_populated() 141 struct ttm_pool *pool; in ttm_pool_alloc_basic() local 151 pool = kunit_kzalloc(test, sizeof(*pool), GFP_KERNEL); in ttm_pool_alloc_basic() 152 KUNIT_ASSERT_NOT_NULL(test, pool); in ttm_pool_alloc_basic() [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn303/ |
D | dcn303_resource.c | 648 static bool dcn303_dwbc_create(struct dc_context *ctx, struct resource_pool *pool) in dcn303_dwbc_create() argument 651 uint32_t pipe_count = pool->res_cap->num_dwb; in dcn303_dwbc_create() 663 pool->dwbc[i] = &dwbc30->base; in dcn303_dwbc_create() 683 static bool dcn303_mmhubbub_create(struct dc_context *ctx, struct resource_pool *pool) in dcn303_mmhubbub_create() argument 686 uint32_t pipe_count = pool->res_cap->num_dwb; in dcn303_mmhubbub_create() 698 pool->mcif_wb[i] = &mcif_wb30->base; in dcn303_mmhubbub_create() 878 static bool init_soc_bounding_box(struct dc *dc, struct resource_pool *pool) in init_soc_bounding_box() argument 890 loaded_ip->max_num_otg = pool->pipe_count; in init_soc_bounding_box() 891 loaded_ip->max_num_dpp = pool->pipe_count; in init_soc_bounding_box() 911 static void dcn303_resource_destruct(struct resource_pool *pool) in dcn303_resource_destruct() argument [all …]
|
/drivers/media/platform/renesas/vsp1/ |
D | vsp1_dl.c | 110 struct vsp1_dl_body_pool *pool; member 227 struct vsp1_dl_body_pool *pool; member 251 struct vsp1_dl_body_pool *pool; in vsp1_dl_body_pool_create() local 255 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in vsp1_dl_body_pool_create() 256 if (!pool) in vsp1_dl_body_pool_create() 259 pool->vsp1 = vsp1; in vsp1_dl_body_pool_create() 268 pool->size = dlb_size * num_bodies; in vsp1_dl_body_pool_create() 270 pool->bodies = kcalloc(num_bodies, sizeof(*pool->bodies), GFP_KERNEL); in vsp1_dl_body_pool_create() 271 if (!pool->bodies) { in vsp1_dl_body_pool_create() 272 kfree(pool); in vsp1_dl_body_pool_create() [all …]
|
/drivers/gpu/drm/i915/gt/ |
D | intel_gt_buffer_pool.c | 14 bucket_for_size(struct intel_gt_buffer_pool *pool, size_t sz) in bucket_for_size() argument 24 if (n >= ARRAY_SIZE(pool->cache_list)) in bucket_for_size() 25 n = ARRAY_SIZE(pool->cache_list) - 1; in bucket_for_size() 27 return &pool->cache_list[n]; in bucket_for_size() 37 static bool pool_free_older_than(struct intel_gt_buffer_pool *pool, long keep) in pool_free_older_than() argument 44 for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++) { in pool_free_older_than() 45 struct list_head *list = &pool->cache_list[n]; in pool_free_older_than() 50 if (spin_trylock_irq(&pool->lock)) { in pool_free_older_than() 73 spin_unlock_irq(&pool->lock); in pool_free_older_than() 89 struct intel_gt_buffer_pool *pool = in pool_free_work() local [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn301/ |
D | dcn301_resource.c | 92 #define TO_DCN301_RES_POOL(pool)\ argument 93 container_of(pool, struct dcn301_resource_pool, base) 1033 static void dcn301_destruct(struct dcn301_resource_pool *pool) in dcn301_destruct() argument 1037 for (i = 0; i < pool->base.stream_enc_count; i++) { in dcn301_destruct() 1038 if (pool->base.stream_enc[i] != NULL) { in dcn301_destruct() 1039 if (pool->base.stream_enc[i]->vpg != NULL) { in dcn301_destruct() 1040 kfree(DCN30_VPG_FROM_VPG(pool->base.stream_enc[i]->vpg)); in dcn301_destruct() 1041 pool->base.stream_enc[i]->vpg = NULL; in dcn301_destruct() 1043 if (pool->base.stream_enc[i]->afmt != NULL) { in dcn301_destruct() 1044 kfree(DCN30_AFMT_FROM_AFMT(pool->base.stream_enc[i]->afmt)); in dcn301_destruct() [all …]
|
/drivers/gpu/drm/amd/display/dc/dce100/ |
D | dce100_resource.c | 756 static void dce100_resource_destruct(struct dce110_resource_pool *pool) in dce100_resource_destruct() argument 760 for (i = 0; i < pool->base.pipe_count; i++) { in dce100_resource_destruct() 761 if (pool->base.opps[i] != NULL) in dce100_resource_destruct() 762 dce110_opp_destroy(&pool->base.opps[i]); in dce100_resource_destruct() 764 if (pool->base.transforms[i] != NULL) in dce100_resource_destruct() 765 dce100_transform_destroy(&pool->base.transforms[i]); in dce100_resource_destruct() 767 if (pool->base.ipps[i] != NULL) in dce100_resource_destruct() 768 dce_ipp_destroy(&pool->base.ipps[i]); in dce100_resource_destruct() 770 if (pool->base.mis[i] != NULL) { in dce100_resource_destruct() 771 kfree(TO_DCE_MEM_INPUT(pool->base.mis[i])); in dce100_resource_destruct() [all …]
|
/drivers/infiniband/sw/rxe/ |
D | rxe_pool.c | 92 void rxe_pool_init(struct rxe_dev *rxe, struct rxe_pool *pool, in rxe_pool_init() argument 97 memset(pool, 0, sizeof(*pool)); in rxe_pool_init() 99 pool->rxe = rxe; in rxe_pool_init() 100 pool->name = info->name; in rxe_pool_init() 101 pool->type = type; in rxe_pool_init() 102 pool->max_elem = info->max_elem; in rxe_pool_init() 103 pool->elem_size = ALIGN(info->size, RXE_POOL_ALIGN); in rxe_pool_init() 104 pool->elem_offset = info->elem_offset; in rxe_pool_init() 105 pool->cleanup = info->cleanup; in rxe_pool_init() 107 atomic_set(&pool->num_elem, 0); in rxe_pool_init() [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn316/ |
D | dcn316_resource.c | 1339 static void dcn316_resource_destruct(struct dcn316_resource_pool *pool) in dcn316_resource_destruct() argument 1343 for (i = 0; i < pool->base.stream_enc_count; i++) { in dcn316_resource_destruct() 1344 if (pool->base.stream_enc[i] != NULL) { in dcn316_resource_destruct() 1345 if (pool->base.stream_enc[i]->vpg != NULL) { in dcn316_resource_destruct() 1346 kfree(DCN30_VPG_FROM_VPG(pool->base.stream_enc[i]->vpg)); in dcn316_resource_destruct() 1347 pool->base.stream_enc[i]->vpg = NULL; in dcn316_resource_destruct() 1349 if (pool->base.stream_enc[i]->afmt != NULL) { in dcn316_resource_destruct() 1350 kfree(DCN30_AFMT_FROM_AFMT(pool->base.stream_enc[i]->afmt)); in dcn316_resource_destruct() 1351 pool->base.stream_enc[i]->afmt = NULL; in dcn316_resource_destruct() 1353 kfree(DCN10STRENC_FROM_STRENC(pool->base.stream_enc[i])); in dcn316_resource_destruct() [all …]
|
/drivers/staging/media/ipu3/ |
D | ipu3-css-pool.c | 25 void imgu_css_pool_cleanup(struct imgu_device *imgu, struct imgu_css_pool *pool) in imgu_css_pool_cleanup() argument 30 imgu_dmamap_free(imgu, &pool->entry[i].param); in imgu_css_pool_cleanup() 33 int imgu_css_pool_init(struct imgu_device *imgu, struct imgu_css_pool *pool, in imgu_css_pool_init() argument 39 pool->entry[i].valid = false; in imgu_css_pool_init() 41 pool->entry[i].param.vaddr = NULL; in imgu_css_pool_init() 45 if (!imgu_dmamap_alloc(imgu, &pool->entry[i].param, size)) in imgu_css_pool_init() 49 pool->last = IPU3_CSS_POOL_SIZE; in imgu_css_pool_init() 54 imgu_css_pool_cleanup(imgu, pool); in imgu_css_pool_init() 61 void imgu_css_pool_get(struct imgu_css_pool *pool) in imgu_css_pool_get() argument 64 u32 n = (pool->last + 1) % IPU3_CSS_POOL_SIZE; in imgu_css_pool_get() [all …]
|
/drivers/staging/octeon/ |
D | ethernet-mem.c | 24 static int cvm_oct_fill_hw_skbuff(int pool, int size, int elements) in cvm_oct_fill_hw_skbuff() argument 35 cvmx_fpa_free(skb->data, pool, size / 128); in cvm_oct_fill_hw_skbuff() 47 static void cvm_oct_free_hw_skbuff(int pool, int size, int elements) in cvm_oct_free_hw_skbuff() argument 52 memory = cvmx_fpa_alloc(pool); in cvm_oct_free_hw_skbuff() 63 pool, elements); in cvm_oct_free_hw_skbuff() 66 pool, elements); in cvm_oct_free_hw_skbuff() 77 static int cvm_oct_fill_hw_memory(int pool, int size, int elements) in cvm_oct_fill_hw_memory() argument 97 elements * size, pool); in cvm_oct_fill_hw_memory() 102 cvmx_fpa_free(fpa, pool, 0); in cvm_oct_fill_hw_memory() 114 static void cvm_oct_free_hw_memory(int pool, int size, int elements) in cvm_oct_free_hw_memory() argument [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn21/ |
D | dcn21_resource.c | 679 static void dcn21_resource_destruct(struct dcn21_resource_pool *pool) in dcn21_resource_destruct() argument 683 for (i = 0; i < pool->base.stream_enc_count; i++) { in dcn21_resource_destruct() 684 if (pool->base.stream_enc[i] != NULL) { in dcn21_resource_destruct() 685 kfree(DCN10STRENC_FROM_STRENC(pool->base.stream_enc[i])); in dcn21_resource_destruct() 686 pool->base.stream_enc[i] = NULL; in dcn21_resource_destruct() 690 for (i = 0; i < pool->base.res_cap->num_dsc; i++) { in dcn21_resource_destruct() 691 if (pool->base.dscs[i] != NULL) in dcn21_resource_destruct() 692 dcn20_dsc_destroy(&pool->base.dscs[i]); in dcn21_resource_destruct() 695 if (pool->base.mpc != NULL) { in dcn21_resource_destruct() 696 kfree(TO_DCN20_MPC(pool->base.mpc)); in dcn21_resource_destruct() [all …]
|
/drivers/gpu/drm/ttm/ |
D | ttm_pool.c | 79 static struct page *ttm_pool_alloc_page(struct ttm_pool *pool, gfp_t gfp_flags, in ttm_pool_alloc_page() argument 95 if (!pool->use_dma_alloc) { in ttm_pool_alloc_page() 96 p = alloc_pages_node(pool->nid, gfp_flags, order); in ttm_pool_alloc_page() 109 vaddr = dma_alloc_attrs(pool->dev, (1ULL << order) * PAGE_SIZE, in ttm_pool_alloc_page() 132 static void ttm_pool_free_page(struct ttm_pool *pool, enum ttm_caching caching, in ttm_pool_free_page() argument 147 if (!pool || !pool->use_dma_alloc) { in ttm_pool_free_page() 157 dma_free_attrs(pool->dev, (1UL << order) * PAGE_SIZE, vaddr, dma->addr, in ttm_pool_free_page() 185 static int ttm_pool_map(struct ttm_pool *pool, unsigned int order, in ttm_pool_map() argument 191 if (pool->use_dma_alloc) { in ttm_pool_map() 198 addr = dma_map_page(pool->dev, p, 0, size, DMA_BIDIRECTIONAL); in ttm_pool_map() [all …]
|
/drivers/gpu/drm/amd/display/dc/dcn10/ |
D | dcn10_resource.c | 907 static void dcn10_resource_destruct(struct dcn10_resource_pool *pool) in dcn10_resource_destruct() argument 911 for (i = 0; i < pool->base.stream_enc_count; i++) { in dcn10_resource_destruct() 912 if (pool->base.stream_enc[i] != NULL) { in dcn10_resource_destruct() 913 kfree(DCN10STRENC_FROM_STRENC(pool->base.stream_enc[i])); in dcn10_resource_destruct() 914 pool->base.stream_enc[i] = NULL; in dcn10_resource_destruct() 918 if (pool->base.mpc != NULL) { in dcn10_resource_destruct() 919 kfree(TO_DCN10_MPC(pool->base.mpc)); in dcn10_resource_destruct() 920 pool->base.mpc = NULL; in dcn10_resource_destruct() 923 kfree(pool->base.hubbub); in dcn10_resource_destruct() 924 pool->base.hubbub = NULL; in dcn10_resource_destruct() [all …]
|