Home
last modified time | relevance | path

Searched refs:rq_size (Results 1 – 25 of 31) sorted by relevance

12

/drivers/infiniband/hw/i40iw/
Di40iw_puda.c151 ((rsrc->rx_wqe_idx + 1) % rsrc->rq_size); in i40iw_puda_replenish_rq()
557 u32 sq_size, rq_size, t_size; in i40iw_puda_qp_create() local
561 rq_size = rsrc->rq_size * I40IW_QP_WQE_MIN_SIZE; in i40iw_puda_qp_create()
562 t_size = (sq_size + rq_size + (I40IW_SHADOW_AREA_SIZE << 3) + in i40iw_puda_qp_create()
576 qp->hw_rq_size = i40iw_get_encoded_wqe_size(rsrc->rq_size, false); in i40iw_puda_qp_create()
586 ukqp->shadow_area = ukqp->rq_base[rsrc->rq_size].elem; in i40iw_puda_qp_create()
587 qp->shadow_area_pa = qp->rq_pa + rq_size; in i40iw_puda_qp_create()
597 ukqp->rq_size = rsrc->rq_size; in i40iw_puda_qp_create()
601 I40IW_RING_INIT(ukqp->rq_ring, ukqp->rq_size); in i40iw_puda_qp_create()
895 rqwridsize = info->rq_size * 8; in i40iw_puda_create_rsrc()
[all …]
Di40iw_user.h350 u32 rq_size; member
388 u32 rq_size; member
429 enum i40iw_status_code i40iw_get_rqdepth(u32 rq_size, u8 shift, u32 *rqdepth);
Di40iw_puda.h106 u32 rq_size; member
130 u32 rq_size; member
Di40iw_uk.c940 enum i40iw_status_code i40iw_get_rqdepth(u32 rq_size, u8 shift, u32 *rqdepth) in i40iw_get_rqdepth() argument
942 *rqdepth = roundup_pow_of_two((rq_size << shift) + I40IW_RQ_RSVD); in i40iw_get_rqdepth()
1031 qp->rq_size = info->rq_size; in i40iw_qp_uk_init()
1033 I40IW_RING_INIT(qp->rq_ring, qp->rq_size); in i40iw_qp_uk_init()
Di40iw_verbs.c473 status = i40iw_get_rqdepth(ukinfo->rq_size, I40IW_MAX_RQ_WQE_SHIFT, &rqdepth); in i40iw_setup_kmode_qp()
506 ukinfo->rq_size = rqdepth >> I40IW_MAX_RQ_WQE_SHIFT; in i40iw_setup_kmode_qp()
534 int rq_size; in i40iw_create_qp() local
563 rq_size = init_attr->cap.max_recv_wr; in i40iw_create_qp()
567 init_info.qp_uk_init_info.rq_size = rq_size; in i40iw_create_qp()
729 uresp.actual_rq_size = rq_size; in i40iw_create_qp()
768 attr->cap.max_recv_wr = qp->qp_uk.rq_size; in i40iw_query_qp()
Di40iw_main.c941 info.rq_size = 8192; in i40iw_initialize_ilq()
970 info.rq_size = 8192; in i40iw_initialize_ieq()
/drivers/net/ethernet/qlogic/netxen/
Dnetxen_nic_ctx.c269 size_t rq_size, rsp_size; in nx_fw_cmd_create_rx_ctx() local
279 rq_size = in nx_fw_cmd_create_rx_ctx()
285 rq_size, &hostrq_phys_addr); in nx_fw_cmd_create_rx_ctx()
349 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_rx_ctx()
392 pci_free_consistent(adapter->pdev, rq_size, prq, hostrq_phys_addr); in nx_fw_cmd_create_rx_ctx()
422 size_t rq_size, rsp_size; in nx_fw_cmd_create_tx_ctx() local
431 rq_size = SIZEOF_HOSTRQ_TX(nx_hostrq_tx_ctx_t); in nx_fw_cmd_create_tx_ctx()
433 rq_size, &rq_phys_addr); in nx_fw_cmd_create_tx_ctx()
474 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_tx_ctx()
497 pci_free_consistent(adapter->pdev, rq_size, rq_addr, rq_phys_addr); in nx_fw_cmd_create_tx_ctx()
/drivers/scsi/
Dstex.c343 u16 rq_size; member
357 u16 rq_size; member
411 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req()
422 hba->req_head * hba->rq_size + sizeof(struct st_msg_header)); in stex_ss_alloc_req()
548 addr = hba->dma_handle + hba->req_head * hba->rq_size; in stex_ss_send_cmd()
1061 h->req_sz = cpu_to_le16(hba->rq_size); in stex_common_handshake()
1073 status_phys = hba->dma_handle + (hba->rq_count+1) * hba->rq_size; in stex_common_handshake()
1152 h->req_sz = cpu_to_le16(hba->rq_size); in stex_ss_handshake()
1543 .rq_size = 1048,
1556 .rq_size = 1048,
[all …]
/drivers/scsi/bnx2i/
Dbnx2i_init.c65 unsigned int rq_size = BNX2I_RQ_WQES_DEFAULT; variable
66 module_param(rq_size, int, 0664);
67 MODULE_PARM_DESC(rq_size, "Configure RQ size");
Dbnx2i.h796 extern unsigned int rq_size;
Dbnx2i_iscsi.c857 hba->max_rqes = rq_size; in bnx2i_alloc_hba()
858 hba->max_cqes = hba->max_sqes + rq_size; in bnx2i_alloc_hba()
/drivers/infiniband/hw/efa/
Defa_verbs.c502 qp->rq_cpu_addr, qp->rq_size, in efa_destroy_qp()
504 dma_unmap_single(&dev->pdev->dev, qp->rq_dma_addr, qp->rq_size, in efa_destroy_qp()
542 if (qp->rq_size) { in qp_mmap_entries_setup()
555 qp->rq_size, EFA_MMAP_DMA_PAGE); in qp_mmap_entries_setup()
559 resp->rq_mmap_size = qp->rq_size; in qp_mmap_entries_setup()
714 qp->rq_size = PAGE_ALIGN(create_qp_params.rq_ring_size_in_bytes); in efa_create_qp()
715 if (qp->rq_size) { in efa_create_qp()
717 qp->rq_size, DMA_TO_DEVICE); in efa_create_qp()
725 qp->rq_cpu_addr, qp->rq_size, &qp->rq_dma_addr); in efa_create_qp()
773 if (qp->rq_size) { in efa_create_qp()
[all …]
Defa.h102 size_t rq_size; member
/drivers/net/ethernet/qlogic/qlcnic/
Dqlcnic_ctx.c251 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_rx_ctx() local
261 rq_size = SIZEOF_HOSTRQ_RX(struct qlcnic_hostrq_rx_ctx, nrds_rings, in qlcnic_82xx_fw_cmd_create_rx_ctx()
266 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_rx_ctx()
344 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_rx_ctx()
388 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); in qlcnic_82xx_fw_cmd_create_rx_ctx()
427 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_tx_ctx() local
436 rq_size = SIZEOF_HOSTRQ_TX(struct qlcnic_hostrq_tx_ctx); in qlcnic_82xx_fw_cmd_create_tx_ctx()
437 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_tx_ctx()
490 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_tx_ctx()
519 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_addr, rq_phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx()
/drivers/infiniband/ulp/srpt/
Dib_srpt.c1786 WARN_ON(ch->rq_size < 1); in srpt_create_ch_ib()
1794 ch->cq = ib_alloc_cq_any(sdev->device, ch, ch->rq_size + sq_size, in srpt_create_ch_ib()
1799 ch->rq_size + sq_size, ret); in srpt_create_ch_ib()
1827 qp_init->cap.max_recv_wr = ch->rq_size; in srpt_create_ch_ib()
1868 for (i = 0; i < ch->rq_size; i++) in srpt_create_ch_ib()
2122 ch->sport->sdev, ch->rq_size, in srpt_release_channel_work()
2128 sdev, ch->rq_size, in srpt_release_channel_work()
2249 ch->rq_size = min(MAX_SRPT_RQ_SIZE, sdev->device->attrs.max_qp_wr); in srpt_cm_req_recv()
2261 srpt_alloc_ioctx_ring(ch->sport->sdev, ch->rq_size, in srpt_cm_req_recv()
2270 for (i = 0; i < ch->rq_size; i++) in srpt_cm_req_recv()
[all …]
Dib_srpt.h309 int rq_size; member
/drivers/gpu/drm/i915/gem/
Di915_gem_execbuffer.c257 unsigned int rq_size; member
909 cache->rq_size = 0; in reloc_cache_init()
933 GEM_BUG_ON(cache->rq_size >= cache->rq->batch->obj->base.size / sizeof(u32)); in reloc_gpu_flush()
934 cache->rq_cmd[cache->rq_size] = MI_BATCH_BUFFER_END; in reloc_gpu_flush()
936 __i915_gem_object_flush_map(cache->rq->batch->obj, 0, cache->rq_size); in reloc_gpu_flush()
1206 cache->rq_size = 0; in __reloc_gpu_alloc()
1231 if (cache->rq_size > PAGE_SIZE/sizeof(u32) - (len + 1)) in reloc_gpu()
1249 cmd = cache->rq_cmd + cache->rq_size; in reloc_gpu()
1250 cache->rq_size += len; in reloc_gpu()
/drivers/infiniband/hw/cxgb3/
Dcxio_wr.h342 u32 rq_size; member
365 __be32 rq_size; member
Diwch_qp.c828 init_attr.rq_size = 1 << qhp->wq.rq_size_log2; in rdma_init()
860 init_attr.rq_addr, init_attr.rq_size, in rdma_init()
Dcxio_hal.c826 wqe->rq_size = cpu_to_be32(attr->rq_size); in cxio_rdma_init()
/drivers/infiniband/sw/siw/
Dsiw_verbs.c427 qp->attrs.rq_size = 0; in siw_create_qp()
441 qp->attrs.rq_size = num_rqe; in siw_create_qp()
538 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; in siw_query_qp()
1045 u32 idx = qp->rq_put % qp->attrs.rq_size; in siw_post_receive()
Dsiw_qp.c1283 while (qp->attrs.rq_size) { in siw_rq_flush()
1285 &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rq_flush()
Dsiw.h274 u32 rq_size; member
/drivers/lightnvm/
Dpblk-core.c273 int rq_size; in pblk_alloc_rqd() local
279 rq_size = pblk_w_rq_size; in pblk_alloc_rqd()
283 rq_size = pblk_g_rq_size; in pblk_alloc_rqd()
287 rq_size = pblk_g_rq_size; in pblk_alloc_rqd()
291 memset(rqd, 0, rq_size); in pblk_alloc_rqd()
/drivers/infiniband/hw/bnxt_re/
Droce_hsi.h1059 __le32 rq_size; member
1191 __le32 rq_size; member
1635 __le32 rq_size; member
2216 __le32 rq_size; member

12