Home
last modified time | relevance | path

Searched refs:sq_size (Results 1 – 25 of 26) sorted by relevance

12

/drivers/scsi/bnx2i/
Dbnx2i_init.c61 unsigned int sq_size; variable
62 module_param(sq_size, int, 0664);
63 MODULE_PARM_DESC(sq_size, "Configure SQ size");
474 if (sq_size && !is_power_of_2(sq_size)) in bnx2i_mod_init()
475 sq_size = roundup_pow_of_two(sq_size); in bnx2i_mod_init()
Dbnx2i_sysfs.c140 static DEVICE_ATTR(sq_size, S_IRUGO | S_IWUSR,
Dbnx2i_iscsi.c844 if (sq_size && sq_size <= BNX2I_5770X_SQ_WQES_MAX) in bnx2i_alloc_hba()
845 hba->max_sqes = sq_size; in bnx2i_alloc_hba()
849 if (sq_size && sq_size <= BNX2I_570X_SQ_WQES_MAX) in bnx2i_alloc_hba()
850 hba->max_sqes = sq_size; in bnx2i_alloc_hba()
Dbnx2i.h795 extern unsigned int sq_size;
/drivers/infiniband/hw/erdma/
Derdma_qp.c217 data = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, qp->attrs.sq_size, in fill_inline_data()
242 qp->attrs.sq_size, SQEBB_SHIFT); in fill_inline_data()
271 qp->attrs.sq_size, SQEBB_SHIFT); in fill_sgl()
289 u32 idx = *pi & (qp->attrs.sq_size - 1); in erdma_push_one_sqe()
304 entry = get_queue_entry(qp->kern_qp.sq_buf, idx, qp->attrs.sq_size, in erdma_push_one_sqe()
366 qp->attrs.sq_size, SQEBB_SHIFT); in erdma_push_one_sqe()
412 qp->attrs.sq_size, SQEBB_SHIFT), in erdma_push_one_sqe()
448 qp->attrs.sq_size, SQEBB_SHIFT); in erdma_push_one_sqe()
515 if ((u16)(sq_pi - qp->kern_qp.sq_ci) >= qp->attrs.sq_size) { in erdma_post_send()
Derdma_verbs.c52 ilog2(qp->attrs.sq_size)) | in create_qp_cmd()
80 (qp->attrs.sq_size << SQEBB_SHIFT); in create_qp_cmd()
487 WARPPED_BUFSIZE(qp->attrs.sq_size << SQEBB_SHIFT), in free_kernel_qp()
514 kqp->swr_tbl = vmalloc_array(qp->attrs.sq_size, sizeof(u64)); in init_kernel_qp()
519 size = (qp->attrs.sq_size << SQEBB_SHIFT) + ERDMA_EXTRA_BUFFER_SIZE; in init_kernel_qp()
531 kqp->sq_db_info = kqp->sq_buf + (qp->attrs.sq_size << SQEBB_SHIFT); in init_kernel_qp()
873 if (len < (ALIGN(qp->attrs.sq_size * SQEBB_SIZE, ERDMA_HW_PAGE_SIZE) + in init_user_qp()
878 qp->attrs.sq_size << SQEBB_SHIFT, 0, va, in init_user_qp()
883 rq_offset = ALIGN(qp->attrs.sq_size << SQEBB_SHIFT, ERDMA_HW_PAGE_SIZE); in init_user_qp()
955 qp->attrs.sq_size = roundup_pow_of_two(attrs->cap.max_send_wr * in erdma_create_qp()
[all …]
Derdma_cq.c147 depth = qp->attrs.sq_size; in erdma_poll_one_cqe()
149 qp->attrs.sq_size, SQEBB_SHIFT); in erdma_poll_one_cqe()
Derdma_verbs.h204 u32 sq_size; member
/drivers/infiniband/hw/irdma/
Dpuda.c649 u32 sq_size, rq_size; in irdma_puda_qp_create() local
652 sq_size = rsrc->sq_size * IRDMA_QP_WQE_MIN_SIZE; in irdma_puda_qp_create()
654 rsrc->qpmem.size = ALIGN((sq_size + rq_size + (IRDMA_SHADOW_AREA_SIZE << 3) + IRDMA_QP_CTX_SIZE), in irdma_puda_qp_create()
664 qp->hw_sq_size = irdma_get_encoded_wqe_size(rsrc->sq_size, IRDMA_QUEUE_TYPE_SQ_RQ); in irdma_puda_qp_create()
671 qp->rq_pa = qp->sq_pa + sq_size; in irdma_puda_qp_create()
674 ukqp->rq_base = &ukqp->sq_base[rsrc->sq_size]; in irdma_puda_qp_create()
684 ukqp->sq_size = rsrc->sq_size; in irdma_puda_qp_create()
687 IRDMA_RING_INIT(ukqp->sq_ring, ukqp->sq_size); in irdma_puda_qp_create()
688 IRDMA_RING_INIT(ukqp->initial_ring, ukqp->sq_size); in irdma_puda_qp_create()
1012 sqwridsize = info->sq_size * sizeof(struct irdma_sq_uk_wr_trk_info); in irdma_puda_create_rsrc()
[all …]
Duser.h327 u32 sq_size; member
373 u32 sq_size; member
408 int irdma_get_sqdepth(struct irdma_uk_attrs *uk_attrs, u32 sq_size, u8 shift,
Dpuda.h90 u32 sq_size; member
113 u32 sq_size; member
Duk.c1266 int irdma_get_sqdepth(struct irdma_uk_attrs *uk_attrs, u32 sq_size, u8 shift, in irdma_get_sqdepth() argument
1271 *sqdepth = irdma_qp_round_up((sq_size << shift) + IRDMA_SQ_RSVD); in irdma_get_sqdepth()
1379 status = irdma_get_sqdepth(ukinfo->uk_attrs, ukinfo->sq_size, in irdma_uk_calc_depth_shift_sq()
1440 qp->sq_size = info->sq_size; in irdma_uk_qp_init()
1442 sq_ring_size = qp->sq_size << info->sq_shift; in irdma_uk_qp_init()
Dhw.c967 cqp_init_info.sq_size = sqsize; in irdma_create_cqp()
1484 info.sq_size = min(iwdev->rf->max_qp / 2, (u32)32768); in irdma_initialize_ilq()
1485 info.rq_size = info.sq_size; in irdma_initialize_ilq()
1487 info.tx_buf_cnt = 2 * info.sq_size; in irdma_initialize_ilq()
1514 info.sq_size = min(iwdev->rf->max_qp / 2, (u32)32768); in irdma_initialize_ieq()
1515 info.rq_size = info.sq_size; in irdma_initialize_ieq()
Dctrl.c3093 if (info->sq_size > IRDMA_CQP_SW_SQSIZE_2048 || in irdma_sc_cqp_init()
3094 info->sq_size < IRDMA_CQP_SW_SQSIZE_4 || in irdma_sc_cqp_init()
3095 ((info->sq_size & (info->sq_size - 1)))) in irdma_sc_cqp_init()
3098 hw_sq_size = irdma_get_encoded_wqe_size(info->sq_size, in irdma_sc_cqp_init()
3101 cqp->sq_size = info->sq_size; in irdma_sc_cqp_init()
3123 IRDMA_RING_INIT(cqp->sq_ring, cqp->sq_size); in irdma_sc_cqp_init()
3135 cqp->sq_size, cqp->hw_sq_size, cqp->sq_base, in irdma_sc_cqp_init()
3154 cqp->sdbuf.size = ALIGN(IRDMA_UPDATE_SD_BUFF_SIZE * cqp->sq_size, in irdma_sc_cqp_create()
Dtype.h235 u32 sq_size; member
371 u32 sq_size; member
Dverbs.c627 ukinfo->sq_size = init_attr->cap.max_send_wr; in irdma_setup_umode_qp()
646 ukinfo->sq_size = ukinfo->sq_depth >> ukinfo->sq_shift; in irdma_setup_umode_qp()
720 ukinfo->sq_size = ukinfo->sq_depth >> ukinfo->sq_shift; in irdma_setup_kmode_qp()
903 init_info.qp_uk_init_info.sq_size = init_attr->cap.max_send_wr; in irdma_create_qp()
1031 uresp.actual_sq_size = init_info.qp_uk_init_info.sq_size; in irdma_create_qp()
/drivers/infiniband/sw/siw/
Dsiw.h261 u32 sq_size; member
630 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_empty()
637 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in sq_get_next()
Dsiw_verbs.c425 qp->attrs.sq_size = num_sqe; in siw_create_qp()
519 qp_attr->cap.max_send_wr = qp->attrs.sq_size; in siw_query_qp()
818 u32 idx = qp->sq_put % qp->attrs.sq_size; in siw_post_send()
Dsiw_qp.c1236 while (qp->attrs.sq_size) { in siw_sq_flush()
1237 sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_flush()
/drivers/infiniband/ulp/srpt/
Dib_srpt.c1794 int sq_size = sport->port_attrib.srp_sq_size; in srpt_create_ch_ib() local
1805 ch->cq = ib_cq_pool_get(sdev->device, ch->rq_size + sq_size, -1, in srpt_create_ch_ib()
1810 ch->rq_size + sq_size, ret); in srpt_create_ch_ib()
1813 ch->cq_size = ch->rq_size + sq_size; in srpt_create_ch_ib()
1828 qp_init->cap.max_send_wr = min(sq_size / 2, attrs->max_qp_wr); in srpt_create_ch_ib()
1829 qp_init->cap.max_rdma_ctxs = sq_size / 2; in srpt_create_ch_ib()
1852 bool retry = sq_size > MIN_SRPT_SQ_SIZE; in srpt_create_ch_ib()
1856 sq_size, ret); in srpt_create_ch_ib()
1858 sq_size = max(sq_size / 2, MIN_SRPT_SQ_SIZE); in srpt_create_ch_ib()
1862 sq_size, ret); in srpt_create_ch_ib()
/drivers/net/ethernet/mellanox/mlx5/core/en/
Dparams.c489 size_t sq_size = 1 << params->log_sq_size; in mlx5e_validate_params() local
493 if (stop_room >= sq_size) { in mlx5e_validate_params()
495 stop_room, sq_size); in mlx5e_validate_params()
/drivers/dma/
Dhisi_dma.c585 size_t sq_size = sizeof(struct hisi_dma_sqe) * hdma_dev->chan_depth; in hisi_dma_alloc_qps_mem() local
593 chan->sq = dmam_alloc_coherent(dev, sq_size, &chan->sq_dma, in hisi_dma_alloc_qps_mem()
/drivers/infiniband/hw/bnxt_re/
Droce_hsi.h475 __le32 sq_size; member
663 __le32 sq_size; member
816 __le32 sq_size; member
1739 __le32 sq_size; member
Dqplib_fp.c861 req.sq_size = cpu_to_le32(bnxt_qplib_set_sq_size(sq, qp->wqe_mode)); in bnxt_qplib_create_qp1()
1034 req.sq_size = cpu_to_le32(bnxt_qplib_set_sq_size(sq, qp->wqe_mode)); in bnxt_qplib_create_qp()
1354 req.sq_size = cpu_to_le32(qp->sq.hwq.max_elements); in bnxt_qplib_modify_qp()
/drivers/net/ethernet/mellanox/mlx4/
Dresource_tracker.c2728 int sq_size; in qp_get_mtt_size() local
2735 sq_size = 1 << (log_sq_size + log_sq_sride + 4); in qp_get_mtt_size()
2737 total_mem = sq_size + rq_size; in qp_get_mtt_size()

12