/drivers/scsi/ |
D | stex.c | 347 u16 rq_size; member 361 u16 rq_size; member 412 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req() 423 hba->req_head * hba->rq_size + sizeof(struct st_msg_header)); in stex_ss_alloc_req() 549 addr = hba->dma_handle + hba->req_head * hba->rq_size; in stex_ss_send_cmd() 1065 h->req_sz = cpu_to_le16(hba->rq_size); in stex_common_handshake() 1077 status_phys = hba->dma_handle + (hba->rq_count+1) * hba->rq_size; in stex_common_handshake() 1156 h->req_sz = cpu_to_le16(hba->rq_size); in stex_ss_handshake() 1547 .rq_size = 1048, 1560 .rq_size = 1048, [all …]
|
/drivers/net/ethernet/qlogic/netxen/ |
D | netxen_nic_ctx.c | 270 size_t rq_size, rsp_size; in nx_fw_cmd_create_rx_ctx() local 280 rq_size = in nx_fw_cmd_create_rx_ctx() 285 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in nx_fw_cmd_create_rx_ctx() 350 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_rx_ctx() 394 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); in nx_fw_cmd_create_rx_ctx() 424 size_t rq_size, rsp_size; in nx_fw_cmd_create_tx_ctx() local 433 rq_size = SIZEOF_HOSTRQ_TX(nx_hostrq_tx_ctx_t); in nx_fw_cmd_create_tx_ctx() 434 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in nx_fw_cmd_create_tx_ctx() 476 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_tx_ctx() 500 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_addr, rq_phys_addr); in nx_fw_cmd_create_tx_ctx()
|
/drivers/gpu/drm/i915/gem/selftests/ |
D | i915_gem_execbuffer.c | 55 i -= eb->reloc_cache.rq_size; in __igt_gpu_reloc() 56 memset32(eb->reloc_cache.rq_cmd + eb->reloc_cache.rq_size, in __igt_gpu_reloc() 58 eb->reloc_cache.rq_size += i; in __igt_gpu_reloc()
|
/drivers/infiniband/hw/irdma/ |
D | puda.c | 129 rsrc->rx_wqe_idx = ((rsrc->rx_wqe_idx + 1) % rsrc->rq_size); in irdma_puda_replenish_rq() 657 u32 sq_size, rq_size; in irdma_puda_qp_create() local 661 rq_size = rsrc->rq_size * IRDMA_QP_WQE_MIN_SIZE; in irdma_puda_qp_create() 662 rsrc->qpmem.size = ALIGN((sq_size + rq_size + (IRDMA_SHADOW_AREA_SIZE << 3) + IRDMA_QP_CTX_SIZE), in irdma_puda_qp_create() 673 qp->hw_rq_size = irdma_get_encoded_wqe_size(rsrc->rq_size, IRDMA_QUEUE_TYPE_SQ_RQ); in irdma_puda_qp_create() 683 ukqp->shadow_area = ukqp->rq_base[rsrc->rq_size].elem; in irdma_puda_qp_create() 685 qp->shadow_area_pa = qp->rq_pa + rq_size; in irdma_puda_qp_create() 693 ukqp->rq_size = rsrc->rq_size; in irdma_puda_qp_create() 697 IRDMA_RING_INIT(ukqp->rq_ring, ukqp->rq_size); in irdma_puda_qp_create() 1023 rqwridsize = info->rq_size * 8; in irdma_puda_create_rsrc() [all …]
|
D | puda.h | 91 u32 rq_size; member 114 u32 rq_size; member
|
D | user.h | 353 u32 rq_size; member 402 u32 rq_size; member 436 u32 rq_size, u8 shift, u32 *wqdepth);
|
D | uk.c | 1389 u32 rq_size, u8 shift, u32 *rqdepth) in irdma_get_rqdepth() argument 1391 *rqdepth = irdma_qp_round_up((rq_size << shift) + IRDMA_RQ_RSVD); in irdma_get_rqdepth() 1493 qp->rq_size = info->rq_size; in irdma_uk_qp_init() 1497 IRDMA_RING_INIT(qp->rq_ring, qp->rq_size); in irdma_uk_qp_init()
|
D | verbs.c | 591 status = irdma_get_rqdepth(uk_attrs, ukinfo->rq_size, rqshift, in irdma_setup_kmode_qp() 633 ukinfo->rq_size = rqdepth >> rqshift; in irdma_setup_kmode_qp() 797 int rq_size; in irdma_create_qp() local 810 rq_size = init_attr->cap.max_recv_wr; in irdma_create_qp() 815 init_info.qp_uk_init_info.rq_size = rq_size; in irdma_create_qp() 952 iwqp->max_recv_wr = rq_size; in irdma_create_qp() 974 uresp.actual_rq_size = rq_size; in irdma_create_qp()
|
D | hw.c | 1464 info.rq_size = info.sq_size; in irdma_initialize_ilq() 1494 info.rq_size = info.sq_size; in irdma_initialize_ieq()
|
/drivers/scsi/bnx2i/ |
D | bnx2i_init.c | 65 unsigned int rq_size = BNX2I_RQ_WQES_DEFAULT; variable 66 module_param(rq_size, int, 0664); 67 MODULE_PARM_DESC(rq_size, "Configure RQ size");
|
D | bnx2i.h | 796 extern unsigned int rq_size;
|
D | bnx2i_iscsi.c | 855 hba->max_rqes = rq_size; in bnx2i_alloc_hba() 856 hba->max_cqes = hba->max_sqes + rq_size; in bnx2i_alloc_hba()
|
/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_ctx.c | 250 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_rx_ctx() local 260 rq_size = SIZEOF_HOSTRQ_RX(struct qlcnic_hostrq_rx_ctx, nrds_rings, in qlcnic_82xx_fw_cmd_create_rx_ctx() 265 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_rx_ctx() 343 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_rx_ctx() 387 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); in qlcnic_82xx_fw_cmd_create_rx_ctx() 426 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_tx_ctx() local 435 rq_size = SIZEOF_HOSTRQ_TX(struct qlcnic_hostrq_tx_ctx); in qlcnic_82xx_fw_cmd_create_tx_ctx() 436 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_tx_ctx() 489 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_tx_ctx() 518 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_addr, rq_phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx()
|
/drivers/infiniband/hw/efa/ |
D | efa_verbs.c | 447 qp->rq_cpu_addr, qp->rq_size, in efa_destroy_qp() 450 qp->rq_size, DMA_TO_DEVICE); in efa_destroy_qp() 515 if (qp->rq_size) { in qp_mmap_entries_setup() 531 address, qp->rq_size, in qp_mmap_entries_setup() 537 resp->rq_mmap_size = qp->rq_size; in qp_mmap_entries_setup() 689 qp->rq_size = PAGE_ALIGN(create_qp_params.rq_ring_size_in_bytes); in efa_create_qp() 690 if (qp->rq_size) { in efa_create_qp() 692 qp->rq_size, DMA_TO_DEVICE); in efa_create_qp() 700 qp->rq_cpu_addr, qp->rq_size, &qp->rq_dma_addr); in efa_create_qp() 748 if (qp->rq_size) in efa_create_qp() [all …]
|
D | efa.h | 95 size_t rq_size; member
|
/drivers/infiniband/ulp/srpt/ |
D | ib_srpt.c | 1791 WARN_ON(ch->rq_size < 1); in srpt_create_ch_ib() 1799 ch->cq = ib_cq_pool_get(sdev->device, ch->rq_size + sq_size, -1, in srpt_create_ch_ib() 1804 ch->rq_size + sq_size, ret); in srpt_create_ch_ib() 1807 ch->cq_size = ch->rq_size + sq_size; in srpt_create_ch_ib() 1831 qp_init->cap.max_recv_wr = ch->rq_size; in srpt_create_ch_ib() 1869 for (i = 0; i < ch->rq_size; i++) in srpt_create_ch_ib() 2114 ch->sport->sdev, ch->rq_size, in srpt_release_channel_work() 2120 sdev, ch->rq_size, in srpt_release_channel_work() 2237 ch->rq_size = min(MAX_SRPT_RQ_SIZE, sdev->device->attrs.max_qp_wr); in srpt_cm_req_recv() 2249 srpt_alloc_ioctx_ring(ch->sport->sdev, ch->rq_size, in srpt_cm_req_recv() [all …]
|
D | ib_srpt.h | 309 int rq_size; member
|
/drivers/infiniband/sw/siw/ |
D | siw_verbs.c | 407 qp->attrs.rq_size = 0; in siw_create_qp() 421 qp->attrs.rq_size = num_rqe; in siw_create_qp() 519 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; in siw_query_qp() 1002 if (qp->srq || qp->attrs.rq_size == 0) { in siw_post_receive() 1062 u32 idx = qp->rq_put % qp->attrs.rq_size; in siw_post_receive()
|
D | siw_qp.c | 1296 while (qp->attrs.rq_size) { in siw_rq_flush() 1298 &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rq_flush()
|
D | siw.h | 261 u32 rq_size; member
|
D | siw_qp_rx.c | 349 rqe = &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rqe_get()
|
/drivers/net/ethernet/microsoft/mana/ |
D | mana_en.c | 1412 u32 cq_size, rq_size; in mana_create_rxq() local 1429 err = mana_alloc_rx_wqe(apc, rxq, &rq_size, &cq_size); in mana_create_rxq() 1433 rq_size = PAGE_ALIGN(rq_size); in mana_create_rxq() 1440 spec.queue_size = rq_size; in mana_create_rxq()
|
/drivers/infiniband/hw/bnxt_re/ |
D | roce_hsi.h | 1166 __le32 rq_size; member 1298 __le32 rq_size; member 1742 __le32 rq_size; member 2323 __le32 rq_size; member
|
D | qplib_fp.c | 900 req.rq_size = cpu_to_le32(rq->max_wqe); in bnxt_qplib_create_qp1() 1057 req.rq_size = cpu_to_le32(rq->max_wqe); in bnxt_qplib_create_qp() 1346 req.rq_size = cpu_to_le32(qp->rq.hwq.max_elements); in bnxt_qplib_modify_qp()
|
/drivers/net/ |
D | virtio_net.c | 2903 unsigned int rq_size = virtqueue_get_vring_size(vq); in mergeable_min_buf_len() local 2906 unsigned int min_buf_len = DIV_ROUND_UP(buf_len, rq_size); in mergeable_min_buf_len()
|