/drivers/infiniband/hw/hfi1/ |
D | ud.c | 72 u8 sc5 = ibp->sl_to_sc[rdma_ah_get_sl(ah_attr)]; in ud_loopback() local 77 if (unlikely(ingress_pkey_check(ppd, pkey, sc5, in ud_loopback() 267 u8 sc5; in hfi1_make_ud_req_9B() local 299 sc5 = ibp->sl_to_sc[rdma_ah_get_sl(ah_attr)]; in hfi1_make_ud_req_9B() 305 lrh0 |= (sc5 & 0xf) << 12; in hfi1_make_ud_req_9B() 306 priv->s_sc = sc5; in hfi1_make_ud_req_9B() 344 u8 l4, sc5; in hfi1_make_ud_req_16B() local 396 sc5 = ibp->sl_to_sc[rdma_ah_get_sl(ah_attr)]; in hfi1_make_ud_req_16B() 400 priv->s_sc = sc5; in hfi1_make_ud_req_16B() 593 u8 sc5, const struct ib_grh *old_grh) in return_cnp_16B() argument [all …]
|
D | trace_ibhdrs.h | 78 void hfi1_trace_parse_9b_hdr(struct ib_header *hdr, bool sc5, 117 bool sc5), 118 TP_ARGS(dd, packet, sc5), 189 hfi1_trace_parse_9b_hdr(packet->hdr, sc5, 272 struct hfi1_packet *packet, bool sc5), 273 TP_ARGS(dd, packet, sc5)); 277 struct hfi1_opa_header *opah, bool sc5), 278 TP_ARGS(dd, opah, sc5), 354 hfi1_trace_parse_9b_hdr(&opah->ibh, sc5, 435 struct hfi1_opa_header *opah, bool sc5), [all …]
|
D | verbs.c | 808 u8 sc5 = priv->s_sc; in hfi1_verbs_send_dma() local 825 u32 vl = sc_to_vlt(dd_from_ibdev(qp->ibqp.device), sc5); in hfi1_verbs_send_dma() 833 pbc |= (ib_is_sc5(sc5) << PBC_DC_INFO_SHIFT); in hfi1_verbs_send_dma() 861 &ps->s_txreq->phdr.hdr, ib_is_sc5(sc5)); in hfi1_verbs_send_dma() 948 u8 sc5; in hfi1_verbs_send_pio() local 980 sc5 = priv->s_sc; in hfi1_verbs_send_pio() 984 u8 vl = sc_to_vlt(dd_from_ibdev(qp->ibqp.device), sc5); in hfi1_verbs_send_pio() 990 pbc |= (ib_is_sc5(sc5) << PBC_DC_INFO_SHIFT); in hfi1_verbs_send_pio() 1058 &ps->s_txreq->phdr.hdr, ib_is_sc5(sc5)); in hfi1_verbs_send_pio() 1117 u8 sc5, int8_t s_pkey_index) in egress_pkey_check() argument [all …]
|
D | ipoib_tx.c | 337 lrh0 |= (txp->flow.sc5 & 0xf) << 12; in hfi1_ipoib_build_ib_tx_headers() 376 ib_is_sc5(txp->flow.sc5) << in hfi1_ipoib_build_ib_tx_headers() 379 sc_to_vlt(priv->dd, txp->flow.sc5), in hfi1_ipoib_build_ib_tx_headers() 411 txp->txq->flow.sc5 = txp->flow.sc5; in hfi1_ipoib_send_dma_common() 415 txp->flow.sc5); in hfi1_ipoib_send_dma_common() 508 ib_is_sc5(txp->flow.sc5)); in hfi1_ipoib_send_dma_single() 566 ib_is_sc5(txp->flow.sc5)); in hfi1_ipoib_send_dma_list() 609 txp.flow.sc5 = txp.ibp->sl_to_sc[rdma_ah_get_sl(txp.ah_attr)]; in hfi1_ipoib_send() 753 txq->flow.sc5 = 0xff; in hfi1_ipoib_txreq_init()
|
D | qp.h | 80 struct sdma_engine *qp_to_sdma_engine(struct rvt_qp *qp, u8 sc5); 81 struct send_context *qp_to_send_context(struct rvt_qp *qp, u8 sc5);
|
D | hfi.h | 1700 u16 pkey, u32 slid, u32 dlid, u8 sc5, 1704 u8 sc5, const struct ib_grh *old_grh); 1707 u8 sc5, const struct ib_grh *old_grh); 1711 u8 sc5, int8_t s_pkey_index); 1727 static inline u8 sc_to_vlt(struct hfi1_devdata *dd, u8 sc5) in sc_to_vlt() argument 1732 if (sc5 >= OPA_MAX_SCS) in sc_to_vlt() 1737 rval = *(((u8 *)dd->sc2vl) + sc5); in sc_to_vlt() 1813 u8 sc5, u8 idx, u32 slid, bool force) in ingress_pkey_check() argument 1819 if ((sc5 == 0xf) && ((pkey & PKEY_LOW_15_MASK) != PKEY_LOW_15_MASK)) in ingress_pkey_check() 1846 u8 sc5, u16 slid) in rcv_pkey_check() argument [all …]
|
D | ruc.c | 32 u8 sc5 = ibp->sl_to_sc[rdma_ah_get_sl(&qp->remote_ah_attr)]; in hfi1_ruc_check_hdr() local 63 sc5, slid))) { in hfi1_ruc_check_hdr() 100 sc5, slid))) { in hfi1_ruc_check_hdr()
|
D | ipoib.h | 42 u8 sc5; member
|
D | rc.c | 1221 u8 sc5, bool is_fecn, in hfi1_make_rc_ack_9B() argument 1247 *pbc_flags |= ((!!(sc5 & 0x10)) << PBC_DC_INFO_SHIFT); in hfi1_make_rc_ack_9B() 1252 lrh0 |= (sc5 & IB_SC_MASK) << IB_SC_SHIFT | in hfi1_make_rc_ack_9B() 1274 u8 sc5, bool is_fecn, in hfi1_make_rc_ack_16B() argument 1316 16B), len, pkey, becn, 0, l4, sc5); in hfi1_make_rc_ack_16B() 1327 u8 sc5, bool is_fecn, 1351 u8 sc5 = ibp->sl_to_sc[rdma_ah_get_sl(&qp->remote_ah_attr)]; in hfi1_send_rc_ack() local 1379 hfi1_make_rc_ack_tbl[priv->hdr_type](packet, &opa_hdr, sc5, is_fecn, in hfi1_send_rc_ack() 1384 sc_to_vlt(ppd->dd, sc5), plen); in hfi1_send_rc_ack() 1397 &opa_hdr, ib_is_sc5(sc5)); in hfi1_send_rc_ack() [all …]
|
D | qp.c | 539 struct sdma_engine *qp_to_sdma_engine(struct rvt_qp *qp, u8 sc5) in qp_to_sdma_engine() argument 552 sde = sdma_select_engine_sc(dd, qp->ibqp.qp_num >> dd->qos_shift, sc5); in qp_to_sdma_engine() 564 struct send_context *qp_to_send_context(struct rvt_qp *qp, u8 sc5) in qp_to_send_context() argument 577 sc5); in qp_to_send_context()
|
D | pio.h | 233 u32 selector, u8 sc5);
|
D | trace.c | 155 void hfi1_trace_parse_9b_hdr(struct ib_header *hdr, bool sc5, in hfi1_trace_parse_9b_hdr() argument 162 *sc = ib_get_sc(hdr) | (sc5 << 4); in hfi1_trace_parse_9b_hdr()
|
D | driver.c | 329 u8 svc_type, sl, sc5; in rcv_hdrerr() local 331 sc5 = hfi1_9B_get_sc5(rhdr, packet->rhf); in rcv_hdrerr() 332 sl = ibp->sc_to_sl[sc5]; in rcv_hdrerr()
|
D | sdma.h | 1021 u8 sc5);
|
D | pio.c | 1801 u32 selector, u8 sc5) in pio_select_send_context_sc() argument 1803 u8 vl = sc_to_vlt(dd, sc5); in pio_select_send_context_sc()
|
D | sdma.c | 785 u8 sc5) in sdma_select_engine_sc() argument 787 u8 vl = sc_to_vlt(dd, sc5); in sdma_select_engine_sc()
|
D | chip.c | 184 sc5, sc5val, \ argument 193 ((u64)(sc5val) << SEND_SC2VLT##num##_SC##sc5##_SHIFT) | \
|