Home
last modified time | relevance | path

Searched refs:p_size (Results 1 – 13 of 13) sorted by relevance

/drivers/crypto/qat/qat_common/
Dqat_asym_algs.c154 unsigned int p_size; member
197 dma_free_coherent(dev, req->ctx.dh->p_size, in qat_dh_cb()
201 req->ctx.dh->p_size, DMA_TO_DEVICE); in qat_dh_cb()
204 areq->dst_len = req->ctx.dh->p_size; in qat_dh_cb()
209 dma_free_coherent(dev, req->ctx.dh->p_size, req->dst_align, in qat_dh_cb()
212 dma_unmap_single(dev, req->out.dh.r, req->ctx.dh->p_size, in qat_dh_cb()
272 if (req->dst_len < ctx->p_size) { in qat_dh_compute_value()
273 req->dst_len = ctx->p_size; in qat_dh_compute_value()
280 msg->pke_hdr.cd_pars.func_id = qat_dh_fn_id(ctx->p_size, in qat_dh_compute_value()
322 if (sg_is_last(req->src) && req->src_len == ctx->p_size) { in qat_dh_compute_value()
[all …]
/drivers/net/ethernet/qlogic/qed/
Dqed_cxt.c178 struct ilt_cfg_pair p_size; member
448 u32 page_sz = p_mgr->clients[ILT_CLI_CDUC].p_size.val; in qed_cxt_set_proto_cid_count()
501 u32 ilt_size = ILT_PAGE_IN_BYTES(p_cli->p_size.val); in qed_ilt_cli_blk_fill()
547 cxts_per_p = ILT_PAGE_IN_BYTES(p_cli->p_size.val) / in qed_ilt_get_dynamic_line_cnt()
863 ilt_page_size = ILT_PAGE_IN_BYTES(p_cli->p_size.val); in qed_cxt_cfg_ilt_compute_excess()
923 psz = ILT_PAGE_IN_BYTES(p_src->p_size.val); in qed_cxt_src_t2_alloc()
1221 clients[ILT_CLI_CDUC].p_size.reg = ILT_CFG_REG(CDUC, P_SIZE); in qed_cxt_mngr_alloc()
1225 clients[ILT_CLI_QM].p_size.reg = ILT_CFG_REG(QM, P_SIZE); in qed_cxt_mngr_alloc()
1229 clients[ILT_CLI_TM].p_size.reg = ILT_CFG_REG(TM, P_SIZE); in qed_cxt_mngr_alloc()
1233 clients[ILT_CLI_SRC].p_size.reg = ILT_CFG_REG(SRC, P_SIZE); in qed_cxt_mngr_alloc()
[all …]
Dqed_debug.c5276 u32 needed_buf_size_in_dwords, *p_size = &needed_buf_size_in_dwords; in qed_dbg_protection_override_dump() local
5284 p_size); in qed_dbg_protection_override_dump()
5330 u32 needed_buf_size_in_dwords, *p_size = &needed_buf_size_in_dwords; in qed_dbg_fw_asserts_dump() local
5338 p_size); in qed_dbg_fw_asserts_dump()
/drivers/gpu/drm/ttm/
Dttm_bo_manager.c107 unsigned long p_size) in ttm_bo_man_init() argument
115 drm_mm_init(&rman->mm, 0, p_size); in ttm_bo_man_init()
Dttm_bo.c1431 unsigned long p_size) in ttm_bo_init_mm() argument
1452 ret = (*man->func->init)(man, p_size); in ttm_bo_init_mm()
1458 man->size = p_size; in ttm_bo_init_mm()
/drivers/gpu/drm/vmwgfx/
Dvmwgfx_gmrid_manager.c117 unsigned long p_size) in vmw_gmrid_man_init() argument
131 switch (p_size) { in vmw_gmrid_man_init()
/drivers/gpu/drm/amd/amdgpu/
Damdgpu_gtt_mgr.c43 unsigned long p_size) in amdgpu_gtt_mgr_init() argument
57 atomic64_set(&mgr->available, p_size); in amdgpu_gtt_mgr_init()
Damdgpu_vram_mgr.c44 unsigned long p_size) in amdgpu_vram_mgr_init() argument
52 drm_mm_init(&mgr->mm, 0, p_size); in amdgpu_vram_mgr_init()
/drivers/block/drbd/
Ddrbd_nl.c1081 sector_t p_size = device->p_size; /* partner's disk size. */ in drbd_new_dev_size() local
1090 p_size = m_size; in drbd_new_dev_size()
1093 if (p_size && m_size) { in drbd_new_dev_size()
1094 size = min_t(sector_t, p_size, m_size); in drbd_new_dev_size()
1100 if (p_size && p_size < size) in drbd_new_dev_size()
1101 size = p_size; in drbd_new_dev_size()
1105 if (p_size) in drbd_new_dev_size()
1106 size = p_size; in drbd_new_dev_size()
Ddrbd_receiver.c3979 sector_t p_size, p_usize, p_csize, my_usize; in receive_sizes() local
3990 p_size = be64_to_cpu(p->d_size); in receive_sizes()
3996 device->p_size = p_size; in receive_sizes()
4004 p_size, drbd_get_max_capacity(device->ldev)); in receive_sizes()
4084 sector_t new_size = p_csize ?: p_usize ?: p_size; in receive_sizes()
4090 } else if (cur_size != 0 && p_size == 0) { in receive_sizes()
Ddrbd_int.h851 sector_t p_size; /* partner's disk size */ member
Ddrbd_main.c2060 device->p_size = in drbd_device_cleanup()
/drivers/gpu/drm/virtio/
Dvirtgpu_ttm.c184 unsigned long p_size) in ttm_bo_man_init() argument