/drivers/infiniband/hw/qib/ |
D | qib_srq.c | 61 if ((unsigned) wr->num_sge > srq->rq.max_sge) { in qib_post_srq_receive() 81 wqe->num_sge = wr->num_sge; in qib_post_srq_receive() 82 for (i = 0; i < wr->num_sge; i++) in qib_post_srq_receive() 287 p->num_sge = wqe->num_sge; in qib_modify_srq() 288 for (i = 0; i < wqe->num_sge; i++) in qib_modify_srq()
|
D | qib_ruc.c | 94 for (i = j = 0; i < wqe->num_sge; i++) { in qib_init_sge() 104 ss->num_sge = j; in qib_init_sge() 115 ss->num_sge = 0; in qib_init_sge() 429 sqp->s_sge.num_sge = wqe->wr.num_sge; in qib_ruc_loopback() 466 qp->r_sge.num_sge = 1; in qib_ruc_loopback() 480 sqp->s_sge.num_sge = 1; in qib_ruc_loopback() 483 qp->r_sge.num_sge = wqe->wr.num_sge; in qib_ruc_loopback() 505 qp->r_sge.num_sge = 0; in qib_ruc_loopback() 529 if (--sqp->s_sge.num_sge) in qib_ruc_loopback() 778 for (i = 0; i < wqe->wr.num_sge; i++) { in qib_send_complete()
|
D | qib_uc.c | 101 qp->s_sge.num_sge = wqe->wr.num_sge; in qib_make_uc_req() 282 qp->r_sge.num_sge = 0; in qib_uc_rcv() 445 qp->r_sge.num_sge = 1; in qib_uc_rcv() 447 qp->r_sge.num_sge = 0; in qib_uc_rcv() 527 qp->r_sge.num_sge = 0; in qib_uc_rcv()
|
D | qib_verbs.c | 187 if (--ss->num_sge) in qib_copy_sge() 228 if (--ss->num_sge) in qib_skip_sge() 254 u8 num_sge = ss->num_sge; in qib_count_sge() local 275 if (--num_sge) in qib_count_sge() 313 if (--ss->num_sge) in qib_copy_from_sge() 356 if (wr->num_sge > qp->s_max_sge) in qib_post_one_send() 381 (wr->num_sge == 0 || in qib_post_one_send() 402 if (wr->num_sge) { in qib_post_one_send() 405 for (i = 0; i < wr->num_sge; i++) { in qib_post_one_send() 418 wqe->wr.num_sge = j; in qib_post_one_send() [all …]
|
D | qib_ud.c | 167 ssge.num_sge = swqe->wr.num_sge; in qib_ud_loopback() 182 if (--ssge.num_sge) in qib_ud_loopback() 315 qp->s_sge.num_sge = wqe->wr.num_sge; in qib_make_ud_req()
|
D | qib_rc.c | 51 ss->num_sge = wqe->wr.num_sge; in restart_sge() 138 qp->s_ack_rdma_sge.num_sge = 1; in qib_make_rc_ack() 489 qp->s_sge.num_sge = wqe->wr.num_sge; in qib_make_rc_req() 1012 for (i = 0; i < wqe->wr.num_sge; i++) { in qib_rc_send_complete() 1068 for (i = 0; i < wqe->wr.num_sge; i++) { in do_rc_completion() 1580 WARN_ON(qp->s_rdma_read_sge.num_sge); in qib_rc_rcv_resp() 2074 qp->r_sge.num_sge = 1; in qib_rc_rcv() 2076 qp->r_sge.num_sge = 0; in qib_rc_rcv() 2210 qp->r_sge.num_sge = 0; in qib_rc_rcv()
|
D | qib_verbs.h | 354 u8 num_sge; member 392 u8 num_sge; member 1035 while (ss->num_sge) { in qib_put_ss() 1037 if (--ss->num_sge) in qib_put_ss()
|
/drivers/infiniband/hw/ipath/ |
D | ipath_srq.c | 61 if ((unsigned) wr->num_sge > srq->rq.max_sge) { in ipath_post_srq_receive() 81 wqe->num_sge = wr->num_sge; in ipath_post_srq_receive() 82 for (i = 0; i < wr->num_sge; i++) in ipath_post_srq_receive() 291 p->num_sge = wqe->num_sge; in ipath_modify_srq() 292 for (i = 0; i < wqe->num_sge; i++) in ipath_modify_srq()
|
D | ipath_ruc.c | 130 for (i = j = 0; i < wqe->num_sge; i++) { in ipath_init_sge() 140 ss->num_sge = j; in ipath_init_sge() 330 sqp->s_sge.num_sge = wqe->wr.num_sge; in ipath_ruc_loopback() 372 qp->r_sge.num_sge = wqe->wr.num_sge; in ipath_ruc_loopback() 413 if (--sqp->s_sge.num_sge) in ipath_ruc_loopback()
|
D | ipath_verbs.c | 188 if (--ss->num_sge) in ipath_copy_sge() 227 if (--ss->num_sge) in ipath_skip_sge() 253 u8 num_sge = ss->num_sge; in ipath_count_sge() local 274 if (--num_sge) in ipath_count_sge() 313 if (--ss->num_sge) in ipath_copy_from_sge() 360 if (wr->num_sge > qp->s_max_sge) in ipath_post_one_send() 382 (wr->num_sge == 0 || in ipath_post_one_send() 400 if (wr->num_sge) { in ipath_post_one_send() 403 for (i = 0, j = 0; i < wr->num_sge; i++) { in ipath_post_one_send() 416 wqe->wr.num_sge = j; in ipath_post_one_send() [all …]
|
D | ipath_keys.c | 227 ss->num_sge = 1; in ipath_rkey_ok() 264 ss->num_sge = 1; in ipath_rkey_ok()
|
D | ipath_ud.c | 198 if (--swqe->wr.num_sge) in ipath_ud_loopback() 325 qp->s_sge.num_sge = wqe->wr.num_sge; in ipath_make_ud_req()
|
/drivers/infiniband/hw/ehca/ |
D | ehca_reqs.c | 76 if (unlikely((recv_wr->num_sge < 0) || in ehca_write_rwqe() 77 (recv_wr->num_sge > ipz_rqueue->act_nr_of_sg))) { in ehca_write_rwqe() 80 recv_wr->num_sge, ipz_rqueue->act_nr_of_sg); in ehca_write_rwqe() 88 wqe_p->nr_of_data_seg = recv_wr->num_sge; in ehca_write_rwqe() 90 for (cnt_ds = 0; cnt_ds < recv_wr->num_sge; cnt_ds++) { in ehca_write_rwqe() 122 send_wr->num_sge, send_wr->send_flags, in trace_send_wr_ud() 137 for (j = 0; j < send_wr->num_sge; j++) { in trace_send_wr_ud() 166 if (unlikely((send_wr->num_sge < 0) || in ehca_write_swqe() 167 (send_wr->num_sge > qp->ipz_squeue.act_nr_of_sg))) { in ehca_write_swqe() 170 send_wr->num_sge, qp->ipz_squeue.act_nr_of_sg); in ehca_write_swqe() [all …]
|
/drivers/infiniband/hw/amso1100/ |
D | c2_qp.c | 836 sizeof(struct c2_data_addr) * ib_wr->num_sge; in c2_post_send() 837 if (ib_wr->num_sge > qp->send_sgl_depth) { in c2_post_send() 846 ib_wr->num_sge, in c2_post_send() 854 (sizeof(struct c2_data_addr) * ib_wr->num_sge); in c2_post_send() 855 if (ib_wr->num_sge > qp->rdma_write_sgl_depth) { in c2_post_send() 869 ib_wr->num_sge, in c2_post_send() 879 if (ib_wr->num_sge > 1) { in c2_post_send() 965 if (ib_wr->num_sge > qp->recv_sgl_depth) { in c2_post_receive() 978 BUG_ON(ib_wr->num_sge >= 256); in c2_post_receive() 981 ib_wr->num_sge, &tot_len, &actual_sge_count); in c2_post_receive()
|
/drivers/infiniband/hw/cxgb3/ |
D | iwch_qp.c | 66 if (wr->num_sge > T3_MAX_SGE) in build_rdma_send() 72 for (i = 0; i < wr->num_sge; i++) { in build_rdma_send() 81 wqe->send.num_sgle = cpu_to_be32(wr->num_sge); in build_rdma_send() 82 *flit_cnt = 4 + ((wr->num_sge) << 1); in build_rdma_send() 92 if (wr->num_sge > T3_MAX_SGE) in build_rdma_write() 109 for (i = 0; i < wr->num_sge; i++) { in build_rdma_write() 121 wqe->write.num_sgle = cpu_to_be32(wr->num_sge); in build_rdma_write() 122 *flit_cnt = 5 + ((wr->num_sge) << 1); in build_rdma_write() 131 if (wr->num_sge > 1) in build_rdma_read() 254 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, in build_rdma_recv() [all …]
|
/drivers/infiniband/ulp/isert/ |
D | ib_isert.h | 43 int num_sge; member 53 int num_sge; member
|
D | ib_isert.c | 615 rx_wr->num_sge = 1; in isert_post_recv() 649 send_wr.num_sge = tx_desc->num_sge; in isert_post_send() 677 tx_desc->num_sge = 1; in isert_create_send_desc() 720 send_wr->num_sge = isert_cmd->tx_desc.num_sge; in isert_init_send_wr() 741 rx_wr.num_sge = 1; in isert_rdma_post_recvl() 784 tx_desc->num_sge = 2; in isert_put_login_tx() 1171 ib_dma_unmap_sg(ib_dev, wr->sge, wr->num_sge, DMA_TO_DEVICE); in isert_unmap_cmd() 1276 ib_dma_unmap_sg(ib_dev, wr->sge, wr->num_sge, DMA_TO_DEVICE); in isert_completion_rdma_read() 1564 isert_cmd->tx_desc.num_sge = 2; in isert_put_response() 1669 send_wr->num_sge = sg_nents; in isert_build_rdma_wr() [all …]
|
/drivers/infiniband/hw/cxgb4/ |
D | qp.c | 335 for (i = 0; i < wr->num_sge; i++) { in build_immd() 367 int num_sge, u32 *plenp) in build_isgl() argument 374 for (i = 0; i < num_sge; i++) { in build_isgl() 389 isglp->nsge = cpu_to_be16(num_sge); in build_isgl() 403 if (wr->num_sge > T4_MAX_SEND_SGE) in build_rdma_send() 430 if (wr->num_sge) { in build_rdma_send() 442 wr->sg_list, wr->num_sge, &plen); in build_rdma_send() 446 wr->num_sge * sizeof(struct fw_ri_sge); in build_rdma_send() 468 if (wr->num_sge > T4_MAX_SEND_SGE) in build_rdma_write() 473 if (wr->num_sge) { in build_rdma_write() [all …]
|
/drivers/infiniband/ulp/iser/ |
D | iser_initiator.c | 147 iser_task->desc.num_sge = 2; in iser_prepare_write_cmd() 165 tx_desc->num_sge = 1; in iser_create_send_desc() 360 tx_desc->num_sge = 2; in iser_send_data_out() 424 mdesc->num_sge = 2; in iser_send_control()
|
D | iser_verbs.c | 757 rx_wr.num_sge = 1; in iser_post_recvl() 780 rx_wr->num_sge = 1; in iser_post_recvm() 815 send_wr.num_sge = tx_desc->num_sge; in iser_post_send()
|
/drivers/infiniband/hw/mthca/ |
D | mthca_srq.c | 514 if (unlikely(wr->num_sge > srq->max_gs)) { in mthca_tavor_post_srq_recv() 521 for (i = 0; i < wr->num_sge; ++i) { in mthca_tavor_post_srq_recv() 607 if (unlikely(wr->num_sge > srq->max_gs)) { in mthca_arbel_post_srq_recv() 613 for (i = 0; i < wr->num_sge; ++i) { in mthca_arbel_post_srq_recv()
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.c | 1657 struct ocrdma_sge *sge, int num_sge, in ocrdma_build_sges() argument 1662 for (i = 0; i < num_sge; i++) { in ocrdma_build_sges() 1669 if (num_sge == 0) in ocrdma_build_sges() 1692 ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); in ocrdma_build_inline_sges() 1693 if (wr->num_sge) in ocrdma_build_inline_sges() 1694 wqe_size += (wr->num_sge * sizeof(struct ocrdma_sge)); in ocrdma_build_inline_sges() 1744 u32 wqe_size = ((wr->num_sge + 1) * sizeof(struct ocrdma_sge)) + in ocrdma_build_read() 1747 ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); in ocrdma_build_read() 1781 wr->num_sge > qp->sq.max_sges) { in ocrdma_post_send() 1869 if (wr->num_sge) in ocrdma_build_rqe() [all …]
|
/drivers/infiniband/ulp/ipoib/ |
D | ipoib_verbs.c | 220 priv->rx_wr.num_sge = IPOIB_UD_RX_SG; in ipoib_transport_dev_init() 223 priv->rx_wr.num_sge = 1; in ipoib_transport_dev_init()
|
/drivers/scsi/lpfc/ |
D | lpfc_scsi.c | 2414 int i = 0, num_sge = 0, status; in lpfc_bg_setup_sgl() local 2480 num_sge++; in lpfc_bg_setup_sgl() 2500 num_sge++; in lpfc_bg_setup_sgl() 2504 return num_sge; in lpfc_bg_setup_sgl() 2568 int num_sge = 0; in lpfc_bg_setup_sgl_prot() local 2601 if (num_sge >= (phba->cfg_total_seg_cnt - 2)) in lpfc_bg_setup_sgl_prot() 2602 return num_sge + 3; in lpfc_bg_setup_sgl_prot() 2649 num_sge++; in lpfc_bg_setup_sgl_prot() 2680 num_sge++; in lpfc_bg_setup_sgl_prot() 2687 if (num_sge >= phba->cfg_total_seg_cnt) in lpfc_bg_setup_sgl_prot() [all …]
|
/drivers/infiniband/core/ |
D | uverbs_cmd.c | 2083 if (user_wr->num_sge + sg_ind > cmd.sge_count) { in ib_uverbs_post_send() 2089 user_wr->num_sge * sizeof (struct ib_sge), in ib_uverbs_post_send() 2104 next->num_sge = user_wr->num_sge; in ib_uverbs_post_send() 2151 if (next->num_sge) { in ib_uverbs_post_send() 2158 next->num_sge * sizeof (struct ib_sge))) { in ib_uverbs_post_send() 2162 sg_ind += next->num_sge; in ib_uverbs_post_send() 2229 if (user_wr->num_sge + sg_ind > sge_count) { in ib_uverbs_unmarshall_recv() 2235 user_wr->num_sge * sizeof (struct ib_sge), in ib_uverbs_unmarshall_recv() 2250 next->num_sge = user_wr->num_sge; in ib_uverbs_unmarshall_recv() 2252 if (next->num_sge) { in ib_uverbs_unmarshall_recv() [all …]
|