/drivers/scsi/bnx2i/ |
D | bnx2i_init.c | 61 unsigned int sq_size; variable 62 module_param(sq_size, int, 0664); 63 MODULE_PARM_DESC(sq_size, "Configure SQ size"); 474 if (sq_size && !is_power_of_2(sq_size)) in bnx2i_mod_init() 475 sq_size = roundup_pow_of_two(sq_size); in bnx2i_mod_init()
|
D | bnx2i_sysfs.c | 140 static DEVICE_ATTR(sq_size, S_IRUGO | S_IWUSR,
|
D | bnx2i_iscsi.c | 844 if (sq_size && sq_size <= BNX2I_5770X_SQ_WQES_MAX) in bnx2i_alloc_hba() 845 hba->max_sqes = sq_size; in bnx2i_alloc_hba() 849 if (sq_size && sq_size <= BNX2I_570X_SQ_WQES_MAX) in bnx2i_alloc_hba() 850 hba->max_sqes = sq_size; in bnx2i_alloc_hba()
|
D | bnx2i.h | 795 extern unsigned int sq_size;
|
/drivers/infiniband/hw/erdma/ |
D | erdma_qp.c | 217 data = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, qp->attrs.sq_size, in fill_inline_data() 242 qp->attrs.sq_size, SQEBB_SHIFT); in fill_inline_data() 271 qp->attrs.sq_size, SQEBB_SHIFT); in fill_sgl() 289 u32 idx = *pi & (qp->attrs.sq_size - 1); in erdma_push_one_sqe() 304 entry = get_queue_entry(qp->kern_qp.sq_buf, idx, qp->attrs.sq_size, in erdma_push_one_sqe() 366 qp->attrs.sq_size, SQEBB_SHIFT); in erdma_push_one_sqe() 412 qp->attrs.sq_size, SQEBB_SHIFT), in erdma_push_one_sqe() 448 qp->attrs.sq_size, SQEBB_SHIFT); in erdma_push_one_sqe() 515 if ((u16)(sq_pi - qp->kern_qp.sq_ci) >= qp->attrs.sq_size) { in erdma_post_send()
|
D | erdma_verbs.c | 52 ilog2(qp->attrs.sq_size)) | in create_qp_cmd() 80 (qp->attrs.sq_size << SQEBB_SHIFT); in create_qp_cmd() 487 WARPPED_BUFSIZE(qp->attrs.sq_size << SQEBB_SHIFT), in free_kernel_qp() 514 kqp->swr_tbl = vmalloc_array(qp->attrs.sq_size, sizeof(u64)); in init_kernel_qp() 519 size = (qp->attrs.sq_size << SQEBB_SHIFT) + ERDMA_EXTRA_BUFFER_SIZE; in init_kernel_qp() 531 kqp->sq_db_info = kqp->sq_buf + (qp->attrs.sq_size << SQEBB_SHIFT); in init_kernel_qp() 873 if (len < (ALIGN(qp->attrs.sq_size * SQEBB_SIZE, ERDMA_HW_PAGE_SIZE) + in init_user_qp() 878 qp->attrs.sq_size << SQEBB_SHIFT, 0, va, in init_user_qp() 883 rq_offset = ALIGN(qp->attrs.sq_size << SQEBB_SHIFT, ERDMA_HW_PAGE_SIZE); in init_user_qp() 955 qp->attrs.sq_size = roundup_pow_of_two(attrs->cap.max_send_wr * in erdma_create_qp() [all …]
|
D | erdma_cq.c | 147 depth = qp->attrs.sq_size; in erdma_poll_one_cqe() 149 qp->attrs.sq_size, SQEBB_SHIFT); in erdma_poll_one_cqe()
|
D | erdma_verbs.h | 204 u32 sq_size; member
|
/drivers/infiniband/hw/irdma/ |
D | puda.c | 649 u32 sq_size, rq_size; in irdma_puda_qp_create() local 652 sq_size = rsrc->sq_size * IRDMA_QP_WQE_MIN_SIZE; in irdma_puda_qp_create() 654 rsrc->qpmem.size = ALIGN((sq_size + rq_size + (IRDMA_SHADOW_AREA_SIZE << 3) + IRDMA_QP_CTX_SIZE), in irdma_puda_qp_create() 664 qp->hw_sq_size = irdma_get_encoded_wqe_size(rsrc->sq_size, IRDMA_QUEUE_TYPE_SQ_RQ); in irdma_puda_qp_create() 671 qp->rq_pa = qp->sq_pa + sq_size; in irdma_puda_qp_create() 674 ukqp->rq_base = &ukqp->sq_base[rsrc->sq_size]; in irdma_puda_qp_create() 684 ukqp->sq_size = rsrc->sq_size; in irdma_puda_qp_create() 687 IRDMA_RING_INIT(ukqp->sq_ring, ukqp->sq_size); in irdma_puda_qp_create() 688 IRDMA_RING_INIT(ukqp->initial_ring, ukqp->sq_size); in irdma_puda_qp_create() 1012 sqwridsize = info->sq_size * sizeof(struct irdma_sq_uk_wr_trk_info); in irdma_puda_create_rsrc() [all …]
|
D | user.h | 327 u32 sq_size; member 373 u32 sq_size; member 408 int irdma_get_sqdepth(struct irdma_uk_attrs *uk_attrs, u32 sq_size, u8 shift,
|
D | puda.h | 90 u32 sq_size; member 113 u32 sq_size; member
|
D | uk.c | 1266 int irdma_get_sqdepth(struct irdma_uk_attrs *uk_attrs, u32 sq_size, u8 shift, in irdma_get_sqdepth() argument 1271 *sqdepth = irdma_qp_round_up((sq_size << shift) + IRDMA_SQ_RSVD); in irdma_get_sqdepth() 1379 status = irdma_get_sqdepth(ukinfo->uk_attrs, ukinfo->sq_size, in irdma_uk_calc_depth_shift_sq() 1440 qp->sq_size = info->sq_size; in irdma_uk_qp_init() 1442 sq_ring_size = qp->sq_size << info->sq_shift; in irdma_uk_qp_init()
|
D | hw.c | 967 cqp_init_info.sq_size = sqsize; in irdma_create_cqp() 1484 info.sq_size = min(iwdev->rf->max_qp / 2, (u32)32768); in irdma_initialize_ilq() 1485 info.rq_size = info.sq_size; in irdma_initialize_ilq() 1487 info.tx_buf_cnt = 2 * info.sq_size; in irdma_initialize_ilq() 1514 info.sq_size = min(iwdev->rf->max_qp / 2, (u32)32768); in irdma_initialize_ieq() 1515 info.rq_size = info.sq_size; in irdma_initialize_ieq()
|
D | ctrl.c | 3093 if (info->sq_size > IRDMA_CQP_SW_SQSIZE_2048 || in irdma_sc_cqp_init() 3094 info->sq_size < IRDMA_CQP_SW_SQSIZE_4 || in irdma_sc_cqp_init() 3095 ((info->sq_size & (info->sq_size - 1)))) in irdma_sc_cqp_init() 3098 hw_sq_size = irdma_get_encoded_wqe_size(info->sq_size, in irdma_sc_cqp_init() 3101 cqp->sq_size = info->sq_size; in irdma_sc_cqp_init() 3123 IRDMA_RING_INIT(cqp->sq_ring, cqp->sq_size); in irdma_sc_cqp_init() 3135 cqp->sq_size, cqp->hw_sq_size, cqp->sq_base, in irdma_sc_cqp_init() 3154 cqp->sdbuf.size = ALIGN(IRDMA_UPDATE_SD_BUFF_SIZE * cqp->sq_size, in irdma_sc_cqp_create()
|
D | type.h | 235 u32 sq_size; member 371 u32 sq_size; member
|
D | verbs.c | 627 ukinfo->sq_size = init_attr->cap.max_send_wr; in irdma_setup_umode_qp() 646 ukinfo->sq_size = ukinfo->sq_depth >> ukinfo->sq_shift; in irdma_setup_umode_qp() 720 ukinfo->sq_size = ukinfo->sq_depth >> ukinfo->sq_shift; in irdma_setup_kmode_qp() 903 init_info.qp_uk_init_info.sq_size = init_attr->cap.max_send_wr; in irdma_create_qp() 1031 uresp.actual_sq_size = init_info.qp_uk_init_info.sq_size; in irdma_create_qp()
|
/drivers/infiniband/sw/siw/ |
D | siw.h | 261 u32 sq_size; member 630 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_empty() 637 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in sq_get_next()
|
D | siw_verbs.c | 425 qp->attrs.sq_size = num_sqe; in siw_create_qp() 519 qp_attr->cap.max_send_wr = qp->attrs.sq_size; in siw_query_qp() 818 u32 idx = qp->sq_put % qp->attrs.sq_size; in siw_post_send()
|
D | siw_qp.c | 1236 while (qp->attrs.sq_size) { in siw_sq_flush() 1237 sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_flush()
|
/drivers/infiniband/ulp/srpt/ |
D | ib_srpt.c | 1794 int sq_size = sport->port_attrib.srp_sq_size; in srpt_create_ch_ib() local 1805 ch->cq = ib_cq_pool_get(sdev->device, ch->rq_size + sq_size, -1, in srpt_create_ch_ib() 1810 ch->rq_size + sq_size, ret); in srpt_create_ch_ib() 1813 ch->cq_size = ch->rq_size + sq_size; in srpt_create_ch_ib() 1828 qp_init->cap.max_send_wr = min(sq_size / 2, attrs->max_qp_wr); in srpt_create_ch_ib() 1829 qp_init->cap.max_rdma_ctxs = sq_size / 2; in srpt_create_ch_ib() 1852 bool retry = sq_size > MIN_SRPT_SQ_SIZE; in srpt_create_ch_ib() 1856 sq_size, ret); in srpt_create_ch_ib() 1858 sq_size = max(sq_size / 2, MIN_SRPT_SQ_SIZE); in srpt_create_ch_ib() 1862 sq_size, ret); in srpt_create_ch_ib()
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | params.c | 489 size_t sq_size = 1 << params->log_sq_size; in mlx5e_validate_params() local 493 if (stop_room >= sq_size) { in mlx5e_validate_params() 495 stop_room, sq_size); in mlx5e_validate_params()
|
/drivers/dma/ |
D | hisi_dma.c | 585 size_t sq_size = sizeof(struct hisi_dma_sqe) * hdma_dev->chan_depth; in hisi_dma_alloc_qps_mem() local 593 chan->sq = dmam_alloc_coherent(dev, sq_size, &chan->sq_dma, in hisi_dma_alloc_qps_mem()
|
/drivers/infiniband/hw/bnxt_re/ |
D | roce_hsi.h | 475 __le32 sq_size; member 663 __le32 sq_size; member 816 __le32 sq_size; member 1739 __le32 sq_size; member
|
D | qplib_fp.c | 861 req.sq_size = cpu_to_le32(bnxt_qplib_set_sq_size(sq, qp->wqe_mode)); in bnxt_qplib_create_qp1() 1034 req.sq_size = cpu_to_le32(bnxt_qplib_set_sq_size(sq, qp->wqe_mode)); in bnxt_qplib_create_qp() 1354 req.sq_size = cpu_to_le32(qp->sq.hwq.max_elements); in bnxt_qplib_modify_qp()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | resource_tracker.c | 2728 int sq_size; in qp_get_mtt_size() local 2735 sq_size = 1 << (log_sq_size + log_sq_sride + 4); in qp_get_mtt_size() 2737 total_mem = sq_size + rq_size; in qp_get_mtt_size()
|