/drivers/infiniband/ulp/ipoib/ |
D | ipoib_verbs.c | 41 struct ib_qp_attr *qp_attr = NULL; in ipoib_mcast_attach() local 54 qp_attr = kmalloc(sizeof *qp_attr, GFP_KERNEL); in ipoib_mcast_attach() 55 if (!qp_attr) in ipoib_mcast_attach() 59 qp_attr->qkey = priv->qkey; in ipoib_mcast_attach() 60 ret = ib_modify_qp(priv->qp, qp_attr, IB_QP_QKEY); in ipoib_mcast_attach() 73 kfree(qp_attr); in ipoib_mcast_attach() 81 struct ib_qp_attr qp_attr; in ipoib_init_qp() local 87 qp_attr.qp_state = IB_QPS_INIT; in ipoib_init_qp() 88 qp_attr.qkey = 0; in ipoib_init_qp() 89 qp_attr.port_num = priv->port; in ipoib_init_qp() [all …]
|
D | ipoib_cm.c | 276 struct ib_qp_attr qp_attr; in ipoib_cm_modify_rx_qp() local 279 qp_attr.qp_state = IB_QPS_INIT; in ipoib_cm_modify_rx_qp() 280 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); in ipoib_cm_modify_rx_qp() 285 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask); in ipoib_cm_modify_rx_qp() 290 qp_attr.qp_state = IB_QPS_RTR; in ipoib_cm_modify_rx_qp() 291 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); in ipoib_cm_modify_rx_qp() 296 qp_attr.rq_psn = psn; in ipoib_cm_modify_rx_qp() 297 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask); in ipoib_cm_modify_rx_qp() 311 qp_attr.qp_state = IB_QPS_RTS; in ipoib_cm_modify_rx_qp() 312 ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); in ipoib_cm_modify_rx_qp() [all …]
|
D | ipoib_ib.c | 818 struct ib_qp_attr qp_attr; in ipoib_ib_dev_stop() local 832 qp_attr.qp_state = IB_QPS_ERR; in ipoib_ib_dev_stop() 833 if (ib_modify_qp(priv->qp, &qp_attr, IB_QP_STATE)) in ipoib_ib_dev_stop() 881 qp_attr.qp_state = IB_QPS_RESET; in ipoib_ib_dev_stop() 882 if (ib_modify_qp(priv->qp, &qp_attr, IB_QP_STATE)) in ipoib_ib_dev_stop()
|
/drivers/infiniband/hw/ehca/ |
D | ehca_qp.c | 1888 struct ib_qp_attr *qp_attr, in ehca_query_qp() argument 1926 qp_attr->cur_qp_state = ehca2ib_qp_state(qpcb->qp_state); in ehca_query_qp() 1927 qp_attr->qp_state = qp_attr->cur_qp_state; in ehca_query_qp() 1929 if (qp_attr->cur_qp_state == -EINVAL) { in ehca_query_qp() 1937 if (qp_attr->qp_state == IB_QPS_SQD) in ehca_query_qp() 1938 qp_attr->sq_draining = 1; in ehca_query_qp() 1940 qp_attr->qkey = qpcb->qkey; in ehca_query_qp() 1941 qp_attr->path_mtu = qpcb->path_mtu; in ehca_query_qp() 1942 qp_attr->path_mig_state = qpcb->path_migration_state - 1; in ehca_query_qp() 1943 qp_attr->rq_psn = qpcb->receive_psn; in ehca_query_qp() [all …]
|
D | ehca_iverbs.h | 152 int ehca_query_qp(struct ib_qp *qp, struct ib_qp_attr *qp_attr,
|
/drivers/infiniband/core/ |
D | iwcm.c | 225 struct ib_qp_attr qp_attr; in iwcm_modify_qp_err() local 230 qp_attr.qp_state = IB_QPS_ERR; in iwcm_modify_qp_err() 231 return ib_modify_qp(qp, &qp_attr, IB_QP_STATE); in iwcm_modify_qp_err() 240 struct ib_qp_attr qp_attr; in iwcm_modify_qp_sqd() local 243 qp_attr.qp_state = IB_QPS_SQD; in iwcm_modify_qp_sqd() 244 return ib_modify_qp(qp, &qp_attr, IB_QP_STATE); in iwcm_modify_qp_sqd() 946 struct ib_qp_attr *qp_attr, in iwcm_init_qp_init_attr() argument 959 qp_attr->qp_access_flags = IB_ACCESS_REMOTE_WRITE| in iwcm_init_qp_init_attr() 972 struct ib_qp_attr *qp_attr, in iwcm_init_qp_rts_attr() argument 996 struct ib_qp_attr *qp_attr, in iw_cm_init_qp_attr() argument [all …]
|
D | cma.c | 452 struct ib_qp_attr qp_attr; in cma_init_ud_qp() local 455 qp_attr.qp_state = IB_QPS_INIT; in cma_init_ud_qp() 456 ret = rdma_init_qp_attr(&id_priv->id, &qp_attr, &qp_attr_mask); in cma_init_ud_qp() 460 ret = ib_modify_qp(qp, &qp_attr, qp_attr_mask); in cma_init_ud_qp() 464 qp_attr.qp_state = IB_QPS_RTR; in cma_init_ud_qp() 465 ret = ib_modify_qp(qp, &qp_attr, IB_QP_STATE); in cma_init_ud_qp() 469 qp_attr.qp_state = IB_QPS_RTS; in cma_init_ud_qp() 470 qp_attr.sq_psn = 0; in cma_init_ud_qp() 471 ret = ib_modify_qp(qp, &qp_attr, IB_QP_STATE | IB_QP_SQ_PSN); in cma_init_ud_qp() 478 struct ib_qp_attr qp_attr; in cma_init_conn_qp() local [all …]
|
D | cm.c | 3452 struct ib_qp_attr *qp_attr, in cm_init_qp_init_attr() argument 3471 qp_attr->qp_access_flags = IB_ACCESS_REMOTE_WRITE; in cm_init_qp_init_attr() 3473 qp_attr->qp_access_flags |= IB_ACCESS_REMOTE_READ | in cm_init_qp_init_attr() 3475 qp_attr->pkey_index = cm_id_priv->av.pkey_index; in cm_init_qp_init_attr() 3476 qp_attr->port_num = cm_id_priv->av.port->port_num; in cm_init_qp_init_attr() 3488 struct ib_qp_attr *qp_attr, in cm_init_qp_rtr_attr() argument 3505 qp_attr->ah_attr = cm_id_priv->av.ah_attr; in cm_init_qp_rtr_attr() 3506 qp_attr->path_mtu = cm_id_priv->path_mtu; in cm_init_qp_rtr_attr() 3507 qp_attr->dest_qp_num = be32_to_cpu(cm_id_priv->remote_qpn); in cm_init_qp_rtr_attr() 3508 qp_attr->rq_psn = be32_to_cpu(cm_id_priv->rq_psn); in cm_init_qp_rtr_attr() [all …]
|
D | ucm.c | 600 struct ib_qp_attr qp_attr; in ib_ucm_init_qp_attr() local 614 memset(&qp_attr, 0, sizeof qp_attr); in ib_ucm_init_qp_attr() 615 qp_attr.qp_state = cmd.qp_state; in ib_ucm_init_qp_attr() 616 result = ib_cm_init_qp_attr(ctx->cm_id, &qp_attr, &resp.qp_attr_mask); in ib_ucm_init_qp_attr() 620 ib_copy_qp_attr_to_user(&resp, &qp_attr); in ib_ucm_init_qp_attr()
|
D | ucma.c | 845 struct ib_qp_attr qp_attr; in ucma_init_qp_attr() local 859 memset(&qp_attr, 0, sizeof qp_attr); in ucma_init_qp_attr() 860 qp_attr.qp_state = cmd.qp_state; in ucma_init_qp_attr() 861 ret = rdma_init_qp_attr(ctx->cm_id, &qp_attr, &resp.qp_attr_mask); in ucma_init_qp_attr() 865 ib_copy_qp_attr_to_user(&resp, &qp_attr); in ucma_init_qp_attr()
|
D | verbs.c | 806 struct ib_qp_attr *qp_attr, in ib_modify_qp() argument 809 return qp->device->modify_qp(qp->real_qp, qp_attr, qp_attr_mask, NULL); in ib_modify_qp() 814 struct ib_qp_attr *qp_attr, in ib_query_qp() argument 819 qp->device->query_qp(qp->real_qp, qp_attr, qp_attr_mask, qp_init_attr) : in ib_query_qp()
|
/drivers/infiniband/hw/mthca/ |
D | mthca_qp.c | 424 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, in mthca_query_qp() argument 438 qp_attr->qp_state = IB_QPS_RESET; in mthca_query_qp() 459 qp_attr->qp_state = qp->state; in mthca_query_qp() 460 qp_attr->path_mtu = context->mtu_msgmax >> 5; in mthca_query_qp() 461 qp_attr->path_mig_state = in mthca_query_qp() 463 qp_attr->qkey = be32_to_cpu(context->qkey); in mthca_query_qp() 464 qp_attr->rq_psn = be32_to_cpu(context->rnr_nextrecvpsn) & 0xffffff; in mthca_query_qp() 465 qp_attr->sq_psn = be32_to_cpu(context->next_send_psn) & 0xffffff; in mthca_query_qp() 466 qp_attr->dest_qp_num = be32_to_cpu(context->remote_qpn) & 0xffffff; in mthca_query_qp() 467 qp_attr->qp_access_flags = in mthca_query_qp() [all …]
|
D | mthca_dev.h | 529 int mthca_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask,
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.c | 1210 struct ib_qp_attr *qp_attr, in ocrdma_query_qp() argument 1225 qp_attr->qp_state = get_ibqp_state(IB_QPS_INIT); in ocrdma_query_qp() 1226 qp_attr->cur_qp_state = get_ibqp_state(IB_QPS_INIT); in ocrdma_query_qp() 1227 qp_attr->path_mtu = in ocrdma_query_qp() 1231 qp_attr->path_mig_state = IB_MIG_MIGRATED; in ocrdma_query_qp() 1232 qp_attr->rq_psn = params.hop_lmt_rq_psn & OCRDMA_QP_PARAMS_RQ_PSN_MASK; in ocrdma_query_qp() 1233 qp_attr->sq_psn = params.tclass_sq_psn & OCRDMA_QP_PARAMS_SQ_PSN_MASK; in ocrdma_query_qp() 1234 qp_attr->dest_qp_num = in ocrdma_query_qp() 1237 qp_attr->qp_access_flags = ocrdma_to_ib_qp_acc_flags(qp->cap_flags); in ocrdma_query_qp() 1238 qp_attr->cap.max_send_wr = qp->sq.max_cnt - 1; in ocrdma_query_qp() [all …]
|
D | ocrdma_verbs.h | 72 struct ib_qp_attr *qp_attr,
|
/drivers/infiniband/hw/mlx4/ |
D | qp.c | 2674 int mlx4_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, in mlx4_ib_query_qp() argument 2686 qp_attr->qp_state = IB_QPS_RESET; in mlx4_ib_query_qp() 2699 qp_attr->qp_state = qp->state; in mlx4_ib_query_qp() 2700 qp_attr->path_mtu = context.mtu_msgmax >> 5; in mlx4_ib_query_qp() 2701 qp_attr->path_mig_state = in mlx4_ib_query_qp() 2703 qp_attr->qkey = be32_to_cpu(context.qkey); in mlx4_ib_query_qp() 2704 qp_attr->rq_psn = be32_to_cpu(context.rnr_nextrecvpsn) & 0xffffff; in mlx4_ib_query_qp() 2705 qp_attr->sq_psn = be32_to_cpu(context.next_send_psn) & 0xffffff; in mlx4_ib_query_qp() 2706 qp_attr->dest_qp_num = be32_to_cpu(context.remote_qpn) & 0xffffff; in mlx4_ib_query_qp() 2707 qp_attr->qp_access_flags = in mlx4_ib_query_qp() [all …]
|
D | mlx4_ib.h | 636 int mlx4_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask,
|
/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 1460 static uint32_t srp_compute_rq_tmo(struct ib_qp_attr *qp_attr, int attr_mask) in srp_compute_rq_tmo() argument 1479 T_tr_ns = 4096 * (1ULL << qp_attr->timeout); in srp_compute_rq_tmo() 1480 max_compl_time_ms = qp_attr->retry_cnt * 4 * T_tr_ns; in srp_compute_rq_tmo() 1491 struct ib_qp_attr *qp_attr = NULL; in srp_cm_rep_handler() local 1521 qp_attr = kmalloc(sizeof *qp_attr, GFP_KERNEL); in srp_cm_rep_handler() 1522 if (!qp_attr) in srp_cm_rep_handler() 1525 qp_attr->qp_state = IB_QPS_RTR; in srp_cm_rep_handler() 1526 ret = ib_cm_init_qp_attr(cm_id, qp_attr, &attr_mask); in srp_cm_rep_handler() 1530 ret = ib_modify_qp(target->qp, qp_attr, attr_mask); in srp_cm_rep_handler() 1541 qp_attr->qp_state = IB_QPS_RTS; in srp_cm_rep_handler() [all …]
|
/drivers/infiniband/ulp/srpt/ |
D | ib_srpt.c | 987 struct ib_qp_attr qp_attr; in srpt_ch_qp_rtr() local 991 qp_attr.qp_state = IB_QPS_RTR; in srpt_ch_qp_rtr() 992 ret = ib_cm_init_qp_attr(ch->cm_id, &qp_attr, &attr_mask); in srpt_ch_qp_rtr() 996 qp_attr.max_dest_rd_atomic = 4; in srpt_ch_qp_rtr() 998 ret = ib_modify_qp(qp, &qp_attr, attr_mask); in srpt_ch_qp_rtr() 1017 struct ib_qp_attr qp_attr; in srpt_ch_qp_rts() local 1021 qp_attr.qp_state = IB_QPS_RTS; in srpt_ch_qp_rts() 1022 ret = ib_cm_init_qp_attr(ch->cm_id, &qp_attr, &attr_mask); in srpt_ch_qp_rts() 1026 qp_attr.max_rd_atomic = 4; in srpt_ch_qp_rts() 1028 ret = ib_modify_qp(qp, &qp_attr, attr_mask); in srpt_ch_qp_rts() [all …]
|
/drivers/net/ethernet/ibm/ehea/ |
D | ehea_qmr.c | 511 struct ehea_qp_init_attr *qp_attr = &qp->init_attr; in ehea_destroy_qp_res() local 522 if (qp_attr->rq_count > 1) in ehea_destroy_qp_res() 524 if (qp_attr->rq_count > 2) in ehea_destroy_qp_res()
|