/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
D | hns_roce_qp.c | 363 hr_qp->rq.wqe_cnt = 0; in set_rq_size() 365 hr_qp->rq_inl_buf.wqe_cnt = 0; in set_rq_size() 392 hr_qp->rq.wqe_cnt = cnt; in set_rq_size() 394 hr_qp->rq_inl_buf.wqe_cnt = cnt; in set_rq_size() 396 hr_qp->rq_inl_buf.wqe_cnt = 0; in set_rq_size() 489 hr_qp->sq.wqe_cnt = cnt; in set_user_sq_size() 505 buf_size = to_hr_hem_entries_size(hr_qp->sq.wqe_cnt, in set_wqe_buf_attr() 527 buf_size = to_hr_hem_entries_size(hr_qp->rq.wqe_cnt, in set_wqe_buf_attr() 569 hr_qp->sq.wqe_cnt = cnt; in set_kernel_sq_size() 604 u32 wqe_cnt = hr_qp->rq_inl_buf.wqe_cnt; in alloc_rq_inline_buf() local [all …]
|
D | hns_roce_srq.c | 194 buf_attr.region[0].size = to_hr_hem_entries_size(srq->wqe_cnt, in alloc_srq_buf() 226 buf_attr.region[0].size = to_hr_hem_entries_size(srq->wqe_cnt, in alloc_srq_idx() 242 idx_que->bitmap = bitmap_zalloc(srq->wqe_cnt, GFP_KERNEL); in alloc_srq_idx() 270 srq->tail = srq->wqe_cnt - 1; in alloc_srq_wrid() 271 srq->wrid = kvmalloc_array(srq->wqe_cnt, sizeof(u64), GFP_KERNEL); in alloc_srq_wrid() 304 srq->wqe_cnt = roundup_pow_of_two(init_attr->attr.max_wr + 1); in hns_roce_create_srq()
|
D | hns_roce_hw_v2.c | 674 wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1); in hns_roce_v2_post_send() 687 ~(((qp->sq.head + nreq) >> ilog2(qp->sq.wqe_cnt)) & 0x1); in hns_roce_v2_post_send() 759 wqe_idx = (hr_qp->rq.head + nreq) & (hr_qp->rq.wqe_cnt - 1); in hns_roce_v2_post_recv() 784 if (hr_qp->rq_inl_buf.wqe_cnt) { in hns_roce_v2_post_recv() 882 ind = srq->head & (srq->wqe_cnt - 1); in hns_roce_v2_post_srq_recv() 897 wqe_idx = find_empty_entry(&srq->idx_que, srq->wqe_cnt); in hns_roce_v2_post_srq_recv() 923 ind = (ind + 1) & (srq->wqe_cnt - 1); in hns_roce_v2_post_srq_recv() 3169 wr_cnt = wr_num & ((*cur_qp)->rq.wqe_cnt - 1); in hns_roce_handle_recv_inl_wqe() 3204 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in sw_comp() 3362 (wq->wqe_cnt - 1); in hns_roce_v2_poll_one() [all …]
|
D | hns_roce_hw_v1.c | 99 wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1); in hns_roce_v1_post_send() 322 (qp->sq.head & ((qp->sq.wqe_cnt << 1) - 1))); in hns_roce_v1_post_send() 370 wqe_idx = (hr_qp->rq.head + nreq) & (hr_qp->rq.wqe_cnt - 1); in hns_roce_v1_post_recv() 2211 ((*cur_qp)->sq.wqe_cnt-1)); in hns_roce_v1_poll_one() 2247 (wq->wqe_cnt - 1); in hns_roce_v1_poll_one() 2249 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in hns_roce_v1_poll_one() 2284 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in hns_roce_v1_poll_one() 2528 ilog2((unsigned int)hr_qp->sq.wqe_cnt)); in hns_roce_v1_m_sqp() 2532 ilog2((unsigned int)hr_qp->rq.wqe_cnt)); in hns_roce_v1_m_sqp() 2735 ilog2((unsigned int)hr_qp->sq.wqe_cnt)); in hns_roce_v1_m_qp() [all …]
|
D | hns_roce_device.h | 402 u32 wqe_cnt; /* WQE num */ member 490 u32 wqe_cnt; member 612 u32 wqe_cnt; member
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
D | qp.c | 183 wq->offset, wq->wqe_cnt, in mlx5_ib_read_user_wqe_sq() 209 wq->wqe_cnt, wq->wqe_shift, in mlx5_ib_read_user_wqe_sq() 245 wq->offset, wq->wqe_cnt, in mlx5_ib_read_user_wqe_rq() 363 qp->rq.wqe_cnt = 0; in set_rq_size() 371 qp->rq.wqe_cnt = ucmd->rq_wqe_count; in set_rq_size() 383 qp->rq.max_post = qp->rq.wqe_cnt; in set_rq_size() 392 qp->rq.wqe_cnt = wq_size / wqe_size; in set_rq_size() 405 qp->rq.max_post = qp->rq.wqe_cnt; in set_rq_size() 534 qp->sq.wqe_cnt = wq_size / MLX5_SEND_WQE_BB; in calc_sq_size() 535 if (qp->sq.wqe_cnt > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz))) { in calc_sq_size() [all …]
|
D | srq.h | 21 u32 wqe_cnt; member
|
D | cq.c | 197 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in handle_responder() 341 idx = tail & (qp->sq.wqe_cnt - 1); in handle_atomics() 407 idx &= (wq->wqe_cnt - 1); in sw_comp() 498 idx = wqe_ctr & (wq->wqe_cnt - 1); in mlx5_poll_one() 526 idx = wqe_ctr & (wq->wqe_cnt - 1); in mlx5_poll_one() 539 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx5_poll_one()
|
D | wr.c | 43 idx = (sq->cur_post + (wqe_sz >> 2)) & (sq->wqe_cnt - 1); in handle_post_send_edge() 934 idx = (idx + 1) & (qp->sq.wqe_cnt - 1); in dump_wqe() 951 *idx = qp->sq.cur_post & (qp->sq.wqe_cnt - 1); in __begin_wqe() 1004 (qp->sq.wqe_cnt - 1)) : in finish_wqe() 1484 ind = qp->rq.head & (qp->rq.wqe_cnt - 1); in mlx5_ib_post_recv() 1519 ind = (ind + 1) & (qp->rq.wqe_cnt - 1); in mlx5_ib_post_recv()
|
D | mem.c | 216 idx = qp->sq.cur_post & (qp->sq.wqe_cnt - 1); in post_send_nop()
|
D | mlx5_ib.h | 276 int wqe_cnt; member
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
D | qp.c | 194 buf = get_send_wqe(qp, n & (qp->sq.wqe_cnt - 1)); in stamp_send_wqe() 314 qp->rq.wqe_cnt = qp->rq.max_gs = 0; in set_rq_size() 325 qp->rq.wqe_cnt = roundup_pow_of_two(max(1U, cap->max_recv_wr)); in set_rq_size() 333 cap->max_recv_wr = qp->rq.max_post = qp->rq.wqe_cnt; in set_rq_size() 337 min(dev->dev->caps.max_wqes - MLX4_IB_SQ_MAX_SPARE, qp->rq.wqe_cnt); in set_rq_size() 381 qp->sq.wqe_cnt = roundup_pow_of_two(cap->max_send_wr + in set_kernel_sq_size() 390 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_kernel_sq_size() 391 (qp->sq.wqe_cnt << qp->sq.wqe_shift); in set_kernel_sq_size() 394 qp->sq.offset = qp->rq.wqe_cnt << qp->rq.wqe_shift; in set_kernel_sq_size() 396 qp->rq.offset = qp->sq.wqe_cnt << qp->sq.wqe_shift; in set_kernel_sq_size() [all …]
|
D | cq.c | 621 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_qp_sw_comp() 737 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_poll_one() 751 tail = wq->tail & (wq->wqe_cnt - 1); in mlx4_ib_poll_one()
|
D | mlx4_ib.h | 165 int wqe_cnt; member
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_qp.c | 140 qp->rq.wqe_cnt = roundup_pow_of_two(max(1U, req_cap->max_recv_wr)); in pvrdma_set_rq_size() 144 req_cap->max_recv_wr = qp->rq.wqe_cnt; in pvrdma_set_rq_size() 150 qp->npages_recv = (qp->rq.wqe_cnt * qp->rq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_rq_size() 165 qp->sq.wqe_cnt = roundup_pow_of_two(max(1U, req_cap->max_send_wr)); in pvrdma_set_sq_size() 169 req_cap->max_send_wr = qp->sq.wqe_cnt; in pvrdma_set_sq_size() 177 (qp->sq.wqe_cnt * qp->sq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_sq_size() 704 qp->sq.ring, qp->sq.wqe_cnt, &tail))) { in pvrdma_post_send() 865 qp->sq.wqe_cnt); in pvrdma_post_send() 930 qp->rq.ring, qp->rq.wqe_cnt, &tail))) { in pvrdma_post_recv() 956 qp->rq.wqe_cnt); in pvrdma_post_recv()
|
D | pvrdma.h | 154 int wqe_cnt; member 169 int wqe_cnt; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_send.c | 58 (dr_cq->qp->sq.wqe_cnt - 1); in dr_parse_cqe() 64 (dr_cq->qp->sq.wqe_cnt - 1); in dr_parse_cqe() 124 dr_qp->rq.wqe_cnt = 4; in dr_create_rc_qp() 127 dr_qp->sq.wqe_cnt = roundup_pow_of_two(attr->max_send_wr); in dr_create_rc_qp() 130 MLX5_SET(qpc, temp_qpc, log_rq_size, ilog2(dr_qp->rq.wqe_cnt)); in dr_create_rc_qp() 131 MLX5_SET(qpc, temp_qpc, log_sq_size, ilog2(dr_qp->sq.wqe_cnt)); in dr_create_rc_qp() 139 dr_qp->sq.wqe_head = kcalloc(dr_qp->sq.wqe_cnt, in dr_create_rc_qp() 169 MLX5_SET(qpc, qpc, log_rq_size, ilog2(dr_qp->rq.wqe_cnt)); in dr_create_rc_qp() 171 MLX5_SET(qpc, qpc, log_sq_size, ilog2(dr_qp->sq.wqe_cnt)); in dr_create_rc_qp() 236 idx = dr_qp->sq.pc & (dr_qp->sq.wqe_cnt - 1); in dr_rdma_segments()
|
D | dr_types.h | 1029 unsigned int wqe_cnt; member 1035 unsigned int wqe_cnt; member
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/ |
D | mthca_srq.c | 56 __be16 wqe_cnt; member 68 __be16 wqe_cnt; member
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/bnxt_re/ |
D | qplib_fp.h | 291 u64 wqe_cnt; member
|
D | ib_verbs.c | 2553 qp->qplib_qp.wqe_cnt == BNXT_RE_UD_QP_HW_STALL) { in bnxt_ud_qp_hw_stall_workaround() 2560 qp->qplib_qp.wqe_cnt = 0; in bnxt_ud_qp_hw_stall_workaround()
|
D | qplib_fp.c | 1926 qp->wqe_cnt++; in bnxt_qplib_post_send()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/ |
D | mlx4.h | 376 __be16 wqe_cnt; member
|
/kernel/linux/linux-5.10/include/linux/mlx5/ |
D | mlx5_ifc.h | 3104 u8 wqe_cnt[0x10]; member 3262 u8 wqe_cnt[0x10]; member
|