/drivers/infiniband/hw/i40iw/ |
D | i40iw_puda.c | 151 ((rsrc->rx_wqe_idx + 1) % rsrc->rq_size); in i40iw_puda_replenish_rq() 557 u32 sq_size, rq_size, t_size; in i40iw_puda_qp_create() local 561 rq_size = rsrc->rq_size * I40IW_QP_WQE_MIN_SIZE; in i40iw_puda_qp_create() 562 t_size = (sq_size + rq_size + (I40IW_SHADOW_AREA_SIZE << 3) + in i40iw_puda_qp_create() 576 qp->hw_rq_size = i40iw_get_encoded_wqe_size(rsrc->rq_size, false); in i40iw_puda_qp_create() 586 ukqp->shadow_area = ukqp->rq_base[rsrc->rq_size].elem; in i40iw_puda_qp_create() 587 qp->shadow_area_pa = qp->rq_pa + rq_size; in i40iw_puda_qp_create() 597 ukqp->rq_size = rsrc->rq_size; in i40iw_puda_qp_create() 601 I40IW_RING_INIT(ukqp->rq_ring, ukqp->rq_size); in i40iw_puda_qp_create() 895 rqwridsize = info->rq_size * 8; in i40iw_puda_create_rsrc() [all …]
|
D | i40iw_user.h | 350 u32 rq_size; member 388 u32 rq_size; member 429 enum i40iw_status_code i40iw_get_rqdepth(u32 rq_size, u8 shift, u32 *rqdepth);
|
D | i40iw_puda.h | 106 u32 rq_size; member 130 u32 rq_size; member
|
D | i40iw_uk.c | 940 enum i40iw_status_code i40iw_get_rqdepth(u32 rq_size, u8 shift, u32 *rqdepth) in i40iw_get_rqdepth() argument 942 *rqdepth = roundup_pow_of_two((rq_size << shift) + I40IW_RQ_RSVD); in i40iw_get_rqdepth() 1031 qp->rq_size = info->rq_size; in i40iw_qp_uk_init() 1033 I40IW_RING_INIT(qp->rq_ring, qp->rq_size); in i40iw_qp_uk_init()
|
D | i40iw_verbs.c | 473 status = i40iw_get_rqdepth(ukinfo->rq_size, I40IW_MAX_RQ_WQE_SHIFT, &rqdepth); in i40iw_setup_kmode_qp() 506 ukinfo->rq_size = rqdepth >> I40IW_MAX_RQ_WQE_SHIFT; in i40iw_setup_kmode_qp() 534 int rq_size; in i40iw_create_qp() local 563 rq_size = init_attr->cap.max_recv_wr; in i40iw_create_qp() 567 init_info.qp_uk_init_info.rq_size = rq_size; in i40iw_create_qp() 729 uresp.actual_rq_size = rq_size; in i40iw_create_qp() 768 attr->cap.max_recv_wr = qp->qp_uk.rq_size; in i40iw_query_qp()
|
D | i40iw_main.c | 941 info.rq_size = 8192; in i40iw_initialize_ilq() 970 info.rq_size = 8192; in i40iw_initialize_ieq()
|
/drivers/net/ethernet/qlogic/netxen/ |
D | netxen_nic_ctx.c | 269 size_t rq_size, rsp_size; in nx_fw_cmd_create_rx_ctx() local 279 rq_size = in nx_fw_cmd_create_rx_ctx() 285 rq_size, &hostrq_phys_addr); in nx_fw_cmd_create_rx_ctx() 349 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_rx_ctx() 392 pci_free_consistent(adapter->pdev, rq_size, prq, hostrq_phys_addr); in nx_fw_cmd_create_rx_ctx() 422 size_t rq_size, rsp_size; in nx_fw_cmd_create_tx_ctx() local 431 rq_size = SIZEOF_HOSTRQ_TX(nx_hostrq_tx_ctx_t); in nx_fw_cmd_create_tx_ctx() 433 rq_size, &rq_phys_addr); in nx_fw_cmd_create_tx_ctx() 474 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_tx_ctx() 497 pci_free_consistent(adapter->pdev, rq_size, rq_addr, rq_phys_addr); in nx_fw_cmd_create_tx_ctx()
|
/drivers/scsi/ |
D | stex.c | 343 u16 rq_size; member 357 u16 rq_size; member 411 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req() 422 hba->req_head * hba->rq_size + sizeof(struct st_msg_header)); in stex_ss_alloc_req() 548 addr = hba->dma_handle + hba->req_head * hba->rq_size; in stex_ss_send_cmd() 1061 h->req_sz = cpu_to_le16(hba->rq_size); in stex_common_handshake() 1073 status_phys = hba->dma_handle + (hba->rq_count+1) * hba->rq_size; in stex_common_handshake() 1152 h->req_sz = cpu_to_le16(hba->rq_size); in stex_ss_handshake() 1543 .rq_size = 1048, 1556 .rq_size = 1048, [all …]
|
/drivers/scsi/bnx2i/ |
D | bnx2i_init.c | 65 unsigned int rq_size = BNX2I_RQ_WQES_DEFAULT; variable 66 module_param(rq_size, int, 0664); 67 MODULE_PARM_DESC(rq_size, "Configure RQ size");
|
D | bnx2i.h | 796 extern unsigned int rq_size;
|
D | bnx2i_iscsi.c | 857 hba->max_rqes = rq_size; in bnx2i_alloc_hba() 858 hba->max_cqes = hba->max_sqes + rq_size; in bnx2i_alloc_hba()
|
/drivers/infiniband/hw/efa/ |
D | efa_verbs.c | 502 qp->rq_cpu_addr, qp->rq_size, in efa_destroy_qp() 504 dma_unmap_single(&dev->pdev->dev, qp->rq_dma_addr, qp->rq_size, in efa_destroy_qp() 542 if (qp->rq_size) { in qp_mmap_entries_setup() 555 qp->rq_size, EFA_MMAP_DMA_PAGE); in qp_mmap_entries_setup() 559 resp->rq_mmap_size = qp->rq_size; in qp_mmap_entries_setup() 714 qp->rq_size = PAGE_ALIGN(create_qp_params.rq_ring_size_in_bytes); in efa_create_qp() 715 if (qp->rq_size) { in efa_create_qp() 717 qp->rq_size, DMA_TO_DEVICE); in efa_create_qp() 725 qp->rq_cpu_addr, qp->rq_size, &qp->rq_dma_addr); in efa_create_qp() 773 if (qp->rq_size) { in efa_create_qp() [all …]
|
D | efa.h | 102 size_t rq_size; member
|
/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_ctx.c | 251 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_rx_ctx() local 261 rq_size = SIZEOF_HOSTRQ_RX(struct qlcnic_hostrq_rx_ctx, nrds_rings, in qlcnic_82xx_fw_cmd_create_rx_ctx() 266 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_rx_ctx() 344 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_rx_ctx() 388 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); in qlcnic_82xx_fw_cmd_create_rx_ctx() 427 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_tx_ctx() local 436 rq_size = SIZEOF_HOSTRQ_TX(struct qlcnic_hostrq_tx_ctx); in qlcnic_82xx_fw_cmd_create_tx_ctx() 437 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_tx_ctx() 490 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_tx_ctx() 519 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_addr, rq_phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx()
|
/drivers/infiniband/ulp/srpt/ |
D | ib_srpt.c | 1786 WARN_ON(ch->rq_size < 1); in srpt_create_ch_ib() 1794 ch->cq = ib_alloc_cq_any(sdev->device, ch, ch->rq_size + sq_size, in srpt_create_ch_ib() 1799 ch->rq_size + sq_size, ret); in srpt_create_ch_ib() 1827 qp_init->cap.max_recv_wr = ch->rq_size; in srpt_create_ch_ib() 1868 for (i = 0; i < ch->rq_size; i++) in srpt_create_ch_ib() 2122 ch->sport->sdev, ch->rq_size, in srpt_release_channel_work() 2128 sdev, ch->rq_size, in srpt_release_channel_work() 2249 ch->rq_size = min(MAX_SRPT_RQ_SIZE, sdev->device->attrs.max_qp_wr); in srpt_cm_req_recv() 2261 srpt_alloc_ioctx_ring(ch->sport->sdev, ch->rq_size, in srpt_cm_req_recv() 2270 for (i = 0; i < ch->rq_size; i++) in srpt_cm_req_recv() [all …]
|
D | ib_srpt.h | 309 int rq_size; member
|
/drivers/gpu/drm/i915/gem/ |
D | i915_gem_execbuffer.c | 257 unsigned int rq_size; member 909 cache->rq_size = 0; in reloc_cache_init() 933 GEM_BUG_ON(cache->rq_size >= cache->rq->batch->obj->base.size / sizeof(u32)); in reloc_gpu_flush() 934 cache->rq_cmd[cache->rq_size] = MI_BATCH_BUFFER_END; in reloc_gpu_flush() 936 __i915_gem_object_flush_map(cache->rq->batch->obj, 0, cache->rq_size); in reloc_gpu_flush() 1206 cache->rq_size = 0; in __reloc_gpu_alloc() 1231 if (cache->rq_size > PAGE_SIZE/sizeof(u32) - (len + 1)) in reloc_gpu() 1249 cmd = cache->rq_cmd + cache->rq_size; in reloc_gpu() 1250 cache->rq_size += len; in reloc_gpu()
|
/drivers/infiniband/hw/cxgb3/ |
D | cxio_wr.h | 342 u32 rq_size; member 365 __be32 rq_size; member
|
D | iwch_qp.c | 828 init_attr.rq_size = 1 << qhp->wq.rq_size_log2; in rdma_init() 860 init_attr.rq_addr, init_attr.rq_size, in rdma_init()
|
D | cxio_hal.c | 826 wqe->rq_size = cpu_to_be32(attr->rq_size); in cxio_rdma_init()
|
/drivers/infiniband/sw/siw/ |
D | siw_verbs.c | 427 qp->attrs.rq_size = 0; in siw_create_qp() 441 qp->attrs.rq_size = num_rqe; in siw_create_qp() 538 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; in siw_query_qp() 1045 u32 idx = qp->rq_put % qp->attrs.rq_size; in siw_post_receive()
|
D | siw_qp.c | 1283 while (qp->attrs.rq_size) { in siw_rq_flush() 1285 &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rq_flush()
|
D | siw.h | 274 u32 rq_size; member
|
/drivers/lightnvm/ |
D | pblk-core.c | 273 int rq_size; in pblk_alloc_rqd() local 279 rq_size = pblk_w_rq_size; in pblk_alloc_rqd() 283 rq_size = pblk_g_rq_size; in pblk_alloc_rqd() 287 rq_size = pblk_g_rq_size; in pblk_alloc_rqd() 291 memset(rqd, 0, rq_size); in pblk_alloc_rqd()
|
/drivers/infiniband/hw/bnxt_re/ |
D | roce_hsi.h | 1059 __le32 rq_size; member 1191 __le32 rq_size; member 1635 __le32 rq_size; member 2216 __le32 rq_size; member
|