Searched refs:p_size (Results 1 – 10 of 10) sorted by relevance
/drivers/crypto/intel/qat/qat_common/ |
D | qat_asym_algs.c | 110 unsigned int p_size; member 167 dma_unmap_single(dev, req->in.dh.in.b, req->ctx.dh->p_size, in qat_dh_cb() 172 areq->dst_len = req->ctx.dh->p_size; in qat_dh_cb() 173 dma_unmap_single(dev, req->out.dh.r, req->ctx.dh->p_size, in qat_dh_cb() 234 if (req->dst_len < ctx->p_size) { in qat_dh_compute_value() 235 req->dst_len = ctx->p_size; in qat_dh_compute_value() 239 if (req->src_len > ctx->p_size) in qat_dh_compute_value() 246 msg->pke_hdr.cd_pars.func_id = qat_dh_fn_id(ctx->p_size, in qat_dh_compute_value() 288 if (sg_is_last(req->src) && req->src_len == ctx->p_size) { in qat_dh_compute_value() 292 int shift = ctx->p_size - req->src_len; in qat_dh_compute_value() [all …]
|
/drivers/net/ethernet/qlogic/qed/ |
D | qed_cxt.c | 286 return ILT_PAGE_IN_BYTES(p_cli->p_size.val); in qed_cxt_get_ilt_page_size() 319 u32 page_sz = p_mgr->clients[ILT_CLI_CDUC].p_size.val; in qed_cxt_set_proto_cid_count() 372 u32 ilt_size = ILT_PAGE_IN_BYTES(p_cli->p_size.val); in qed_ilt_cli_blk_fill() 418 cxts_per_p = ILT_PAGE_IN_BYTES(p_cli->p_size.val) / in qed_ilt_get_dynamic_line_cnt() 753 ilt_page_size = ILT_PAGE_IN_BYTES(p_cli->p_size.val); in qed_cxt_cfg_ilt_compute_excess() 843 psz = ILT_PAGE_IN_BYTES(p_src->p_size.val); in qed_cxt_src_t2_alloc() 1125 clients[ILT_CLI_CDUC].p_size.reg = ILT_CFG_REG(CDUC, P_SIZE); in qed_cxt_mngr_alloc() 1129 clients[ILT_CLI_QM].p_size.reg = ILT_CFG_REG(QM, P_SIZE); in qed_cxt_mngr_alloc() 1133 clients[ILT_CLI_TM].p_size.reg = ILT_CFG_REG(TM, P_SIZE); in qed_cxt_mngr_alloc() 1137 clients[ILT_CLI_SRC].p_size.reg = ILT_CFG_REG(SRC, P_SIZE); in qed_cxt_mngr_alloc() [all …]
|
D | qed_cxt.h | 285 struct ilt_cfg_pair p_size; member
|
D | qed_debug.c | 5019 (clients[ILT_CLI_CDUC].p_size.val + PXP_ILT_PAGE_SIZE_NUM_BITS_MIN); in qed_ilt_dump() 5021 (clients[ILT_CLI_CDUT].p_size.val + PXP_ILT_PAGE_SIZE_NUM_BITS_MIN); in qed_ilt_dump() 5604 u32 needed_buf_size_in_dwords, *p_size = &needed_buf_size_in_dwords; in qed_dbg_protection_override_dump() local 5612 p_size); in qed_dbg_protection_override_dump() 5658 u32 needed_buf_size_in_dwords, *p_size = &needed_buf_size_in_dwords; in qed_dbg_fw_asserts_dump() local 5666 p_size); in qed_dbg_fw_asserts_dump()
|
/drivers/gpu/drm/ttm/ |
D | ttm_range_manager.c | 180 unsigned long p_size) in ttm_range_man_init_nocheck() argument 194 ttm_resource_manager_init(man, bdev, p_size); in ttm_range_man_init_nocheck() 196 drm_mm_init(&rman->mm, 0, p_size); in ttm_range_man_init_nocheck()
|
/drivers/block/drbd/ |
D | drbd_nl.c | 1097 sector_t p_size = device->p_size; /* partner's disk size. */ in drbd_new_dev_size() local 1106 p_size = m_size; in drbd_new_dev_size() 1109 if (p_size && m_size) { in drbd_new_dev_size() 1110 size = min_t(sector_t, p_size, m_size); in drbd_new_dev_size() 1116 if (p_size && p_size < size) in drbd_new_dev_size() 1117 size = p_size; in drbd_new_dev_size() 1121 if (p_size) in drbd_new_dev_size() 1122 size = p_size; in drbd_new_dev_size()
|
D | drbd_receiver.c | 4068 sector_t p_size, p_usize, p_csize, my_usize; in receive_sizes() local 4079 p_size = be64_to_cpu(p->d_size); in receive_sizes() 4085 device->p_size = p_size; in receive_sizes() 4093 p_size, drbd_get_max_capacity(device->ldev)); in receive_sizes() 4172 sector_t new_size = p_csize ?: p_usize ?: p_size; in receive_sizes() 4178 } else if (cur_size != 0 && p_size == 0) { in receive_sizes()
|
D | drbd_int.h | 772 sector_t p_size; /* partner's disk size */ member
|
D | drbd_main.c | 2032 device->p_size = in drbd_device_cleanup()
|
/drivers/crypto/hisilicon/hpre/ |
D | hpre_crypto.c | 658 if (params->p_size > HPRE_DH_MAX_P_SZ) in hpre_dh_set_params() 661 if (hpre_is_dh_params_length_valid(params->p_size << in hpre_dh_set_params() 665 sz = ctx->key_sz = params->p_size; in hpre_dh_set_params()
|