Home
last modified time | relevance | path

Searched refs:cqe_size (Results 1 – 17 of 17) sorted by relevance

/drivers/infiniband/hw/mlx5/
Dcq.c728 int nent, int cqe_size) in alloc_cq_buf() argument
732 err = mlx5_buf_alloc(dev->mdev, nent * cqe_size, &buf->buf); in alloc_cq_buf()
736 buf->cqe_size = cqe_size; in alloc_cq_buf()
745 int *cqe_size, int *index, int *inlen) in create_cq_user() argument
768 if (ucmd.cqe_size != 64 && ucmd.cqe_size != 128) in create_cq_user()
771 *cqe_size = ucmd.cqe_size; in create_cq_user()
774 entries * ucmd.cqe_size, in create_cq_user()
789 ucmd.buf_addr, entries * ucmd.cqe_size, npages, page_shift, ncont); in create_cq_user()
831 cqe = get_cqe_from_buf(buf, i, buf->cqe_size); in init_cq_buf()
832 cqe64 = buf->cqe_size == 64 ? cqe : cqe + 64; in init_cq_buf()
[all …]
Dmlx5_ib.h395 int cqe_size; member
451 int cqe_size; member
/drivers/infiniband/hw/mlx4/
Dcq.c104 err = mlx4_buf_alloc(dev->dev, nent * dev->dev->caps.cqe_size, in mlx4_ib_alloc_cq_buf()
110 buf->entry_size = dev->dev->caps.cqe_size; in mlx4_ib_alloc_cq_buf()
142 int cqe_size = dev->dev->caps.cqe_size; in mlx4_ib_get_cq_umem() local
144 *umem = ib_umem_get(context, buf_addr, cqe * cqe_size, in mlx4_ib_get_cq_umem()
354 int cqe_size = cq->buf.entry_size; in mlx4_ib_cq_resize_copy_cqes() local
355 int cqe_inc = cqe_size == 64 ? 1 : 0; in mlx4_ib_cq_resize_copy_cqes()
364 memcpy(new_cqe, get_cqe(cq, i & cq->ibcq.cqe), cqe_size); in mlx4_ib_cq_resize_copy_cqes()
Dmain.c1052 resp.cqe_size = dev->dev->caps.cqe_size; in mlx4_ib_alloc_ucontext()
/drivers/net/ethernet/mellanox/mlx4/
Dfw.h204 u16 cqe_size; /* For use only when CQE stride feature enabled */ member
Den_cq.c65 cq->buf_size = cq->size * mdev->dev->caps.cqe_size; in mlx4_en_create_cq()
Den_tx.c432 cqe = mlx4_en_get_cqe(buf, index, priv->cqe_size) + factor; in mlx4_en_process_tx_cq()
485 cqe = mlx4_en_get_cqe(buf, index, priv->cqe_size) + factor; in mlx4_en_process_tx_cq()
Den_rx.c833 cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor; in mlx4_en_process_rx_cq()
1093 cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor; in mlx4_en_process_rx_cq()
Dmlx4_en.h565 int cqe_size; member
Dfw.c1918 dev->caps.cqe_size = 64; in mlx4_INIT_HCA()
1921 dev->caps.cqe_size = 32; in mlx4_INIT_HCA()
1928 dev->caps.cqe_size = cache_line_size(); in mlx4_INIT_HCA()
2131 param->cqe_size = 1 << ((byte_field & in mlx4_QUERY_HCA()
Den_netdev.c1601 cqe = mlx4_en_get_cqe(cq->buf, j, priv->cqe_size) + in mlx4_en_start_port()
3122 priv->cqe_factor = (mdev->dev->caps.cqe_size == 64) ? 1 : 0; in mlx4_en_init_netdev()
3123 priv->cqe_size = mdev->dev->caps.cqe_size; in mlx4_en_init_netdev()
Dmain.c980 dev->caps.cqe_size = 64; in mlx4_slave_cap()
983 dev->caps.cqe_size = 32; in mlx4_slave_cap()
992 dev->caps.cqe_size = hca_param.cqe_size; in mlx4_slave_cap()
/drivers/infiniband/hw/ocrdma/
Docrdma_hw.c1790 u32 hw_pages, cqe_size, page_size, cqe_count; in ocrdma_mbx_create_cq() local
1803 cqe_size = OCRDMA_DPP_CQE_SIZE; in ocrdma_mbx_create_cq()
1808 cqe_size = sizeof(struct ocrdma_cqe); in ocrdma_mbx_create_cq()
1812 cq->len = roundup(max_hw_cqe * cqe_size, OCRDMA_MIN_Q_PAGE_SIZE); in ocrdma_mbx_create_cq()
1833 cqe_count = cq->len / cqe_size; in ocrdma_mbx_create_cq()
1862 cmd->cmd.pdid_cqecnt = (cq->len / cqe_size); in ocrdma_mbx_create_cq()
1864 cmd->cmd.pdid_cqecnt = (cq->len / cqe_size) - 1; in ocrdma_mbx_create_cq()
/drivers/scsi/bnx2i/
Dbnx2i.h661 u32 cqe_size; member
Dbnx2i_hwi.c171 if (cq_index > ep->qp.cqe_size * 2) in bnx2i_arm_cq_event_coalescing()
172 cq_index -= ep->qp.cqe_size * 2; in bnx2i_arm_cq_event_coalescing()
1140 ep->qp.cqe_size = hba->max_cqes; in bnx2i_alloc_qp_resc()
2084 if (qp->cqe_exp_seq_sn == (qp->cqe_size * 2 + 1)) in bnx2i_process_new_cqes()
/drivers/scsi/lpfc/
Dlpfc_sli4.h410 uint32_t cqe_size; member
Dlpfc_init.c9500 sli4_params->cqe_size = bf_get(cqe_size, &mqe->un.sli4_params); in lpfc_pc_sli4_params_get()