/drivers/infiniband/hw/qib/ |
D | qib_qp.c | 565 int attr_mask, struct ib_udata *udata) in qib_modify_qp() argument 579 cur_state = attr_mask & IB_QP_CUR_STATE ? in qib_modify_qp() 581 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; in qib_modify_qp() 584 attr_mask)) in qib_modify_qp() 587 if (attr_mask & IB_QP_AV) { in qib_modify_qp() 594 if (attr_mask & IB_QP_ALT_PATH) { in qib_modify_qp() 603 if (attr_mask & IB_QP_PKEY_INDEX) in qib_modify_qp() 607 if (attr_mask & IB_QP_MIN_RNR_TIMER) in qib_modify_qp() 611 if (attr_mask & IB_QP_PORT) in qib_modify_qp() 618 if (attr_mask & IB_QP_DEST_QPN) in qib_modify_qp() [all …]
|
D | qib_srq.c | 212 enum ib_srq_attr_mask attr_mask, in qib_modify_srq() argument 219 if (attr_mask & IB_SRQ_MAX_WR) { in qib_modify_srq() 226 ((attr_mask & IB_SRQ_LIMIT) ? in qib_modify_srq() 299 if (attr_mask & IB_SRQ_LIMIT) in qib_modify_srq() 333 } else if (attr_mask & IB_SRQ_LIMIT) { in qib_modify_srq()
|
D | qib_verbs.h | 902 int attr_mask, struct ib_udata *udata); 905 int attr_mask, struct ib_qp_init_attr *init_attr); 968 enum ib_srq_attr_mask attr_mask,
|
/drivers/infiniband/hw/ipath/ |
D | ipath_qp.c | 451 int attr_mask, struct ib_udata *udata) in ipath_modify_qp() argument 461 cur_state = attr_mask & IB_QP_CUR_STATE ? in ipath_modify_qp() 463 new_state = attr_mask & IB_QP_STATE ? attr->qp_state : cur_state; in ipath_modify_qp() 466 attr_mask)) in ipath_modify_qp() 469 if (attr_mask & IB_QP_AV) { in ipath_modify_qp() 479 if (attr_mask & IB_QP_PKEY_INDEX) in ipath_modify_qp() 483 if (attr_mask & IB_QP_MIN_RNR_TIMER) in ipath_modify_qp() 487 if (attr_mask & IB_QP_PORT) in ipath_modify_qp() 496 if ((attr_mask & IB_QP_PATH_MTU) && in ipath_modify_qp() 501 if (attr_mask & IB_QP_PATH_MIG_STATE) in ipath_modify_qp() [all …]
|
D | ipath_srq.c | 216 enum ib_srq_attr_mask attr_mask, in ipath_modify_srq() argument 223 if (attr_mask & IB_SRQ_MAX_WR) { in ipath_modify_srq() 230 ((attr_mask & IB_SRQ_LIMIT) ? in ipath_modify_srq() 303 if (attr_mask & IB_SRQ_LIMIT) in ipath_modify_srq() 333 } else if (attr_mask & IB_SRQ_LIMIT) { in ipath_modify_srq()
|
D | ipath_verbs.h | 745 int attr_mask, struct ib_udata *udata); 748 int attr_mask, struct ib_qp_init_attr *init_attr); 799 enum ib_srq_attr_mask attr_mask,
|
/drivers/infiniband/ulp/ipoib/ |
D | ipoib_verbs.c | 82 int attr_mask; in ipoib_init_qp() local 91 attr_mask = in ipoib_init_qp() 96 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); in ipoib_init_qp() 104 attr_mask &= ~IB_QP_PORT; in ipoib_init_qp() 105 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); in ipoib_init_qp() 113 attr_mask |= IB_QP_SQ_PSN; in ipoib_init_qp() 114 attr_mask &= ~IB_QP_PKEY_INDEX; in ipoib_init_qp() 115 ret = ib_modify_qp(priv->qp, &qp_attr, attr_mask); in ipoib_init_qp()
|
/drivers/infiniband/hw/mthca/ |
D | mthca_qp.c | 299 int attr_mask) in store_attrs() argument 301 if (attr_mask & IB_QP_PKEY_INDEX) in store_attrs() 303 if (attr_mask & IB_QP_QKEY) in store_attrs() 305 if (attr_mask & IB_QP_SQ_PSN) in store_attrs() 328 int attr_mask) in get_hw_access_flags() argument 334 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) in get_hw_access_flags() 339 if (attr_mask & IB_QP_ACCESS_FLAGS) in get_hw_access_flags() 544 const struct ib_qp_attr *attr, int attr_mask, in __mthca_modify_qp() argument 567 if (!(attr_mask & IB_QP_PATH_MIG_STATE)) in __mthca_modify_qp() 588 else if (attr_mask & IB_QP_PATH_MTU) { in __mthca_modify_qp() [all …]
|
D | mthca_srq.c | 365 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) in mthca_modify_srq() argument 372 if (attr_mask & IB_SRQ_MAX_WR) in mthca_modify_srq() 375 if (attr_mask & IB_SRQ_LIMIT) { in mthca_modify_srq()
|
D | mthca_dev.h | 516 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata); 531 int mthca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask,
|
/drivers/infiniband/hw/ehca/ |
D | ehca_qp.c | 1239 int attr_mask, int smi_reset2init) in internal_modify_qp() argument 1293 (attr_mask & IB_QP_STATE) && in internal_modify_qp() 1315 if ((attr_mask & IB_QP_CUR_STATE) && in internal_modify_qp() 1327 my_qp, ibqp->qp_num, qp_cur_state, attr->qp_state, attr_mask); in internal_modify_qp() 1329 qp_new_state = attr_mask & IB_QP_STATE ? attr->qp_state : qp_cur_state; in internal_modify_qp() 1332 attr_mask)) { in internal_modify_qp() 1337 qp_cur_state, my_qp, ibqp->qp_num, attr_mask); in internal_modify_qp() 1428 !(attr_mask & IB_QP_MAX_DEST_RD_ATOMIC)) { in internal_modify_qp() 1434 if (attr_mask & IB_QP_PKEY_INDEX) { in internal_modify_qp() 1445 if (attr_mask & IB_QP_PORT) { in internal_modify_qp() [all …]
|
D | ehca_iverbs.h | 149 int ehca_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, int attr_mask, 170 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata);
|
/drivers/infiniband/hw/mlx4/ |
D | qp.c | 1102 int attr_mask) in to_mlx4_access_flags() argument 1108 if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC) in to_mlx4_access_flags() 1113 if (attr_mask & IB_QP_ACCESS_FLAGS) in to_mlx4_access_flags() 1132 int attr_mask) in store_sqp_attrs() argument 1134 if (attr_mask & IB_QP_PKEY_INDEX) in store_sqp_attrs() 1136 if (attr_mask & IB_QP_QKEY) in store_sqp_attrs() 1138 if (attr_mask & IB_QP_SQ_PSN) in store_sqp_attrs() 1228 const struct ib_qp_attr *attr, int attr_mask, in __mlx4_ib_modify_qp() argument 1247 if (!(attr_mask & IB_QP_PATH_MIG_STATE)) in __mlx4_ib_modify_qp() 1274 } else if (attr_mask & IB_QP_PATH_MTU) { in __mlx4_ib_modify_qp() [all …]
|
D | srq.c | 228 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) in mlx4_ib_modify_srq() argument 235 if (attr_mask & IB_SRQ_MAX_WR) in mlx4_ib_modify_srq() 238 if (attr_mask & IB_SRQ_LIMIT) { in mlx4_ib_modify_srq()
|
D | mlx4_ib.h | 623 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata); 635 int attr_mask, struct ib_udata *udata);
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.h | 68 int attr_mask); 70 int attr_mask, struct ib_udata *udata);
|
D | ocrdma_hw.c | 2176 struct ib_qp_attr *attrs, int attr_mask, in ocrdma_set_qp_params() argument 2183 if (attr_mask & IB_QP_PKEY_INDEX) { in ocrdma_set_qp_params() 2188 if (attr_mask & IB_QP_QKEY) { in ocrdma_set_qp_params() 2193 if (attr_mask & IB_QP_AV) in ocrdma_set_qp_params() 2204 if ((attr_mask & IB_QP_EN_SQD_ASYNC_NOTIFY) && in ocrdma_set_qp_params() 2210 if (attr_mask & IB_QP_DEST_QPN) { in ocrdma_set_qp_params() 2215 if (attr_mask & IB_QP_PATH_MTU) { in ocrdma_set_qp_params() 2227 if (attr_mask & IB_QP_TIMEOUT) { in ocrdma_set_qp_params() 2232 if (attr_mask & IB_QP_RETRY_CNT) { in ocrdma_set_qp_params() 2238 if (attr_mask & IB_QP_MIN_RNR_TIMER) { in ocrdma_set_qp_params() [all …]
|
D | ocrdma_verbs.c | 1121 int attr_mask) in _ocrdma_modify_qp() argument 1130 if (attr_mask & IB_QP_STATE) in _ocrdma_modify_qp() 1137 status = ocrdma_mbx_modify_qp(dev, qp, attr, attr_mask, old_qps); in _ocrdma_modify_qp() 1142 int attr_mask, struct ib_udata *udata) in ocrdma_modify_qp() argument 1158 if (attr_mask & IB_QP_STATE) in ocrdma_modify_qp() 1164 if (!ib_modify_qp_is_ok(old_qps, new_qps, ibqp->qp_type, attr_mask)) { in ocrdma_modify_qp() 1167 __func__, dev->id, attr_mask, qp->id, ibqp->qp_type, in ocrdma_modify_qp() 1172 status = _ocrdma_modify_qp(ibqp, attr, attr_mask); in ocrdma_modify_qp() 1211 int attr_mask, struct ib_qp_init_attr *qp_init_attr) in ocrdma_query_qp() argument 1431 int attr_mask = IB_QP_STATE; in ocrdma_destroy_qp() local [all …]
|
D | ocrdma_hw.h | 110 struct ib_qp_attr *attrs, int attr_mask,
|
D | ocrdma_main.c | 486 int attr_mask = IB_QP_STATE; in ocrdma_close() local 496 _ocrdma_modify_qp(&qp->ibqp, &attrs, attr_mask); in ocrdma_close()
|
/drivers/infiniband/hw/cxgb4/ |
D | qp.c | 1724 int attr_mask, struct ib_udata *udata) in c4iw_ib_modify_qp() argument 1734 if ((attr_mask & IB_QP_STATE) && (attr->qp_state == IB_QPS_RTR)) in c4iw_ib_modify_qp() 1735 attr_mask &= ~IB_QP_STATE; in c4iw_ib_modify_qp() 1738 if (!attr_mask) in c4iw_ib_modify_qp() 1753 mask |= (attr_mask & IB_QP_STATE) ? C4IW_QP_ATTR_NEXT_STATE : 0; in c4iw_ib_modify_qp() 1754 mask |= (attr_mask & IB_QP_ACCESS_FLAGS) ? in c4iw_ib_modify_qp() 1765 mask |= (attr_mask & IB_QP_SQ_PSN) ? C4IW_QP_ATTR_SQ_DB : 0; in c4iw_ib_modify_qp() 1766 mask |= (attr_mask & IB_QP_RQ_PSN) ? C4IW_QP_ATTR_RQ_DB : 0; in c4iw_ib_modify_qp() 1778 int attr_mask, struct ib_qp_init_attr *init_attr) in c4iw_ib_query_qp() argument
|
/drivers/infiniband/ulp/srp/ |
D | ib_srp.c | 1460 static uint32_t srp_compute_rq_tmo(struct ib_qp_attr *qp_attr, int attr_mask) in srp_compute_rq_tmo() argument 1470 WARN_ON_ONCE((attr_mask & (IB_QP_TIMEOUT | IB_QP_RETRY_CNT)) != in srp_compute_rq_tmo() 1492 int attr_mask = 0; in srp_cm_rep_handler() local 1526 ret = ib_cm_init_qp_attr(cm_id, qp_attr, &attr_mask); in srp_cm_rep_handler() 1530 ret = ib_modify_qp(target->qp, qp_attr, attr_mask); in srp_cm_rep_handler() 1542 ret = ib_cm_init_qp_attr(cm_id, qp_attr, &attr_mask); in srp_cm_rep_handler() 1546 target->rq_tmo_jiffies = srp_compute_rq_tmo(qp_attr, attr_mask); in srp_cm_rep_handler() 1548 ret = ib_modify_qp(target->qp, qp_attr, attr_mask); in srp_cm_rep_handler()
|
/drivers/infiniband/hw/cxgb3/ |
D | iwch_provider.c | 1051 int attr_mask, struct ib_udata *udata) in iwch_ib_modify_qp() argument 1061 if ((attr_mask & IB_QP_STATE) && (attr->qp_state == IB_QPS_RTR)) in iwch_ib_modify_qp() 1062 attr_mask &= ~IB_QP_STATE; in iwch_ib_modify_qp() 1065 if (!attr_mask) in iwch_ib_modify_qp() 1080 mask |= (attr_mask & IB_QP_STATE) ? IWCH_QP_ATTR_NEXT_STATE : 0; in iwch_ib_modify_qp() 1081 mask |= (attr_mask & IB_QP_ACCESS_FLAGS) ? in iwch_ib_modify_qp()
|
/drivers/infiniband/hw/amso1100/ |
D | c2_qp.c | 135 struct ib_qp_attr *attr, int attr_mask) in c2_qp_modify() argument 163 if (attr_mask & IB_QP_STATE) { in c2_qp_modify() 185 } else if (attr_mask & IB_QP_CUR_STATE) { in c2_qp_modify()
|
D | c2_provider.c | 564 int attr_mask, struct ib_udata *udata) in c2_modify_qp() argument 570 attr_mask); in c2_modify_qp()
|