Home
last modified time | relevance | path

Searched refs:wqe_shift (Results 1 – 8 of 8) sorted by relevance

/drivers/infiniband/hw/mthca/
Dmthca_srq.c75 return srq->queue.direct.buf + (n << srq->wqe_shift); in get_wqe()
77 return srq->queue.page_list[(n << srq->wqe_shift) >> PAGE_SHIFT].buf + in get_wqe()
78 ((n << srq->wqe_shift) & (PAGE_SIZE - 1)); in get_wqe()
102 context->wqe_base_ds = cpu_to_be64(1 << (srq->wqe_shift - 4)); in mthca_tavor_init_srq_context()
131 context->logstride_usrpage = cpu_to_be32((srq->wqe_shift - 4) << 29); in mthca_arbel_init_srq_context()
142 mthca_buf_free(dev, srq->max << srq->wqe_shift, &srq->queue, in mthca_free_srq_buf()
162 err = mthca_buf_alloc(dev, srq->max << srq->wqe_shift, in mthca_alloc_srq_buf()
182 next->nda_op = htonl(((i + 1) << srq->wqe_shift) | 1); in mthca_alloc_srq_buf()
189 (void *) scatter < wqe + (1 << srq->wqe_shift); in mthca_alloc_srq_buf()
227 srq->wqe_shift = ilog2(ds); in mthca_alloc_srq()
[all …]
Dmthca_qp.c210 return qp->queue.direct.buf + (n << qp->rq.wqe_shift); in get_recv_wqe()
212 return qp->queue.page_list[(n << qp->rq.wqe_shift) >> PAGE_SHIFT].buf + in get_recv_wqe()
213 ((n << qp->rq.wqe_shift) & (PAGE_SIZE - 1)); in get_recv_wqe()
220 (n << qp->sq.wqe_shift); in get_send_wqe()
223 (n << qp->sq.wqe_shift)) >> in get_send_wqe()
225 ((qp->send_wqe_offset + (n << qp->sq.wqe_shift)) & in get_send_wqe()
600 qp_context->rq_size_stride |= qp->rq.wqe_shift - 4; in __mthca_modify_qp()
604 qp_context->sq_size_stride |= qp->sq.wqe_shift - 4; in __mthca_modify_qp()
950 1 << qp->sq.wqe_shift)); in mthca_adjust_qp_caps()
957 (min(dev->limits.max_desc_sz, 1 << qp->rq.wqe_shift) - in mthca_adjust_qp_caps()
[all …]
Dmthca_provider.h229 int wqe_shift; member
255 int wqe_shift; member
Dmthca_cq.c541 >> wq->wqe_shift); in mthca_poll_one()
548 wqe_index = wqe >> srq->wqe_shift; in mthca_poll_one()
555 wqe_index = wqe >> wq->wqe_shift; in mthca_poll_one()
/drivers/infiniband/hw/mlx4/
Dqp.c180 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); in get_recv_wqe()
185 return get_wqe(qp, qp->sq.offset + (n << qp->sq.wqe_shift)); in get_send_wqe()
208 s = roundup(size, 1U << qp->sq.wqe_shift); in stamp_send_wqe()
210 ind = (i >> qp->sq.wqe_shift) + n; in stamp_send_wqe()
214 wqe = buf + (i & ((1 << qp->sq.wqe_shift) - 1)); in stamp_send_wqe()
269 post_nop_wqe(qp, ind, s << qp->sq.wqe_shift); in pad_wraparound()
387 qp->rq.wqe_shift = ilog2(qp->rq.max_gs * sizeof (struct mlx4_wqe_data_seg)); in set_rq_size()
469 qp->sq.wqe_shift = ilog2(64); in set_kernel_sq_size()
471 qp->sq.wqe_shift = ilog2(roundup_pow_of_two(s)); in set_kernel_sq_size()
474 qp->sq_max_wqes_per_wr = DIV_ROUND_UP(s, 1U << qp->sq.wqe_shift); in set_kernel_sq_size()
[all …]
Dsrq.c43 return mlx4_buf_offset(&srq->buf, n << srq->msrq.wqe_shift); in get_wqe()
104 srq->msrq.wqe_shift = ilog2(desc_size); in mlx4_ib_create_srq()
284 mlx4_buf_free(dev->dev, msrq->msrq.max << msrq->msrq.wqe_shift, in mlx4_ib_destroy_srq()
Dmlx4_ib.h142 int wqe_shift; member
/drivers/net/ethernet/mellanox/mlx4/
Dsrq.c195 srq_context->logstride = srq->wqe_shift - 4; in mlx4_srq_alloc()