/drivers/infiniband/sw/rxe/ |
D | rxe_hdr.h | 86 struct rxe_bth *bth = arg; in __bth_opcode() local 88 return bth->opcode; in __bth_opcode() 93 struct rxe_bth *bth = arg; in __bth_set_opcode() local 95 bth->opcode = opcode; in __bth_set_opcode() 100 struct rxe_bth *bth = arg; in __bth_se() local 102 return 0 != (BTH_SE_MASK & bth->flags); in __bth_se() 107 struct rxe_bth *bth = arg; in __bth_set_se() local 110 bth->flags |= BTH_SE_MASK; in __bth_set_se() 112 bth->flags &= ~BTH_SE_MASK; in __bth_set_se() 117 struct rxe_bth *bth = arg; in __bth_mig() local [all …]
|
D | rxe_icrc.c | 78 struct rxe_bth *bth; in rxe_icrc_hdr() local 118 bth = (struct rxe_bth *)&pshdr[bth_offset]; in rxe_icrc_hdr() 121 bth->qpn |= cpu_to_be32(~BTH_QPN_MASK); in rxe_icrc_hdr()
|
/drivers/infiniband/core/ |
D | ud_header.c | 222 { STRUCT_FIELD(bth, opcode), 226 { STRUCT_FIELD(bth, solicited_event), 230 { STRUCT_FIELD(bth, mig_req), 234 { STRUCT_FIELD(bth, pad_count), 238 { STRUCT_FIELD(bth, transport_header_version), 242 { STRUCT_FIELD(bth, pkey), 250 { STRUCT_FIELD(bth, destination_qpn), 254 { STRUCT_FIELD(bth, ack_req), 262 { STRUCT_FIELD(bth, psn), 388 header->bth.opcode = IB_OPCODE_UD_SEND_ONLY_WITH_IMMEDIATE; in ib_ud_header_init() [all …]
|
/drivers/infiniband/hw/qib/ |
D | qib_ud.c | 368 ohdr->bth[0] = cpu_to_be32(bth0); in qib_make_ud_req() 372 ohdr->bth[1] = rdma_ah_get_dlid(ah_attr) >= in qib_make_ud_req() 377 ohdr->bth[2] = cpu_to_be32(wqe->psn & QIB_PSN_MASK); in qib_make_ud_req() 454 pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; in qib_ud_rcv() 471 pkey1 = be32_to_cpu(ohdr->bth[0]); in qib_ud_rcv() 508 opcode = be32_to_cpu(ohdr->bth[0]) >> 24; in qib_ud_rcv() 567 qib_lookup_pkey(ibp, be32_to_cpu(ohdr->bth[0])) : 0; in qib_ud_rcv()
|
D | qib_uc.c | 260 opcode = be32_to_cpu(ohdr->bth[0]); in qib_uc_rcv() 264 psn = be32_to_cpu(ohdr->bth[2]); in qib_uc_rcv() 378 pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; in qib_uc_rcv() 462 pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; in qib_uc_rcv() 489 pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; in qib_uc_rcv()
|
D | qib_ruc.c | 240 ohdr->bth[0] = cpu_to_be32(bth0); in qib_make_ruc_header() 241 ohdr->bth[1] = cpu_to_be32(qp->remote_qpn); in qib_make_ruc_header() 242 ohdr->bth[2] = cpu_to_be32(bth2); in qib_make_ruc_header()
|
D | qib_driver.c | 342 opcode = be32_to_cpu(ohdr->bth[0]); in qib_rcv_hdrerr() 344 psn = be32_to_cpu(ohdr->bth[2]); in qib_rcv_hdrerr() 347 qp_num = be32_to_cpu(ohdr->bth[1]) & RVT_QPN_MASK; in qib_rcv_hdrerr() 378 be32_to_cpu(ohdr->bth[0])); in qib_rcv_hdrerr()
|
D | qib_rc.c | 653 ohdr->bth[0] = cpu_to_be32(bth0); in qib_send_rc_ack() 654 ohdr->bth[1] = cpu_to_be32(qp->remote_qpn); in qib_send_rc_ack() 655 ohdr->bth[2] = cpu_to_be32(qp->r_ack_psn & QIB_PSN_MASK); in qib_send_rc_ack() 898 opcode = be32_to_cpu(ohdr->bth[0]) >> 24; in qib_rc_send_complete() 906 psn = be32_to_cpu(ohdr->bth[2]); in qib_rc_send_complete() 1410 pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; in qib_rc_rcv_resp() 1435 pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; in qib_rc_rcv_resp() 1739 opcode = be32_to_cpu(ohdr->bth[0]); in qib_rc_rcv() 1743 psn = be32_to_cpu(ohdr->bth[2]); in qib_rc_rcv() 1859 pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; in qib_rc_rcv()
|
D | qib_user_sdma.c | 459 hdr->bth[0] = cpu_to_be32(be32_to_cpu(hdr->bth[0])&0xFFCFFFFF); in qib_user_sdma_page_to_frags() 516 seqnum.val = be32_to_cpu(hdr->bth[2]); in qib_user_sdma_page_to_frags() 521 hdr->bth[2] = cpu_to_be32(seqnum.val); in qib_user_sdma_page_to_frags()
|
D | qib_common.h | 701 __be32 bth[3]; member
|
/drivers/infiniband/hw/hfi1/ |
D | ud.c | 248 ohdr->bth[0] = cpu_to_be32(bth0); in hfi1_make_bth_deth() 249 ohdr->bth[1] = cpu_to_be32(rvt_get_swqe_remote_qpn(wqe)); in hfi1_make_bth_deth() 250 ohdr->bth[2] = cpu_to_be32(mask_psn(wqe->psn)); in hfi1_make_bth_deth() 630 ohdr->bth[0] = cpu_to_be32(bth0); in return_cnp_16B() 632 ohdr->bth[1] = cpu_to_be32(remote_qpn); in return_cnp_16B() 633 ohdr->bth[2] = 0; /* PSN 0 */ in return_cnp_16B() 688 ohdr->bth[0] = cpu_to_be32(bth0); in return_cnp() 690 ohdr->bth[1] = cpu_to_be32(remote_qpn | (1 << IB_BECN_SHIFT)); in return_cnp() 691 ohdr->bth[2] = 0; /* PSN 0 */ in return_cnp()
|
D | ruc.c | 150 hdr.ibh.u.oth.bth[2]) / 4) 211 ohdr->bth[0] = cpu_to_be32(bth0); in hfi1_make_ruc_bth() 212 ohdr->bth[1] = cpu_to_be32(bth1); in hfi1_make_ruc_bth() 213 ohdr->bth[2] = cpu_to_be32(bth2); in hfi1_make_ruc_bth()
|
D | user_sdma.c | 424 opcode = (be32_to_cpu(req->hdr.bth[0]) >> 24) & 0xff; in hfi1_user_sdma_process_request() 447 pkey = (u16)be32_to_cpu(req->hdr.bth[0]); in hfi1_user_sdma_process_request() 996 hdr->bth[2] = cpu_to_be32( in set_txreq_header() 997 set_pkt_bth_psn(hdr->bth[2], in set_txreq_header() 1003 hdr->bth[2] |= cpu_to_be32(1UL << 31); in set_txreq_header() 1087 val32 = (be32_to_cpu(hdr->bth[2]) + req->seqnum) & in set_txreq_header_ahg()
|
D | trace_tx.h | 495 __be32 *bth = (__be32 *)hdr->bth; 506 __entry->bth0 = be32_to_cpu(bth[0]); 507 __entry->bth1 = be32_to_cpu(bth[1]); 508 __entry->bth2 = be32_to_cpu(bth[2]);
|
D | driver.c | 484 u32 lqpn = be32_to_cpu(ohdr->bth[1]) & RVT_QPN_MASK; in hfi1_process_ecn_slowpath() 614 bth1 = be32_to_cpu(packet->ohdr->bth[1]); in __prescan_rxq() 629 packet->ohdr->bth[1] = cpu_to_be32(bth1); in __prescan_rxq() 1874 opcode = (be32_to_cpu(packet.ohdr->bth[0]) >> 24); in seqfile_dump_rcd() 1875 qpn = be32_to_cpu(packet.ohdr->bth[1]) & RVT_QPN_MASK; in seqfile_dump_rcd() 1876 psn = mask_psn(be32_to_cpu(packet.ohdr->bth[2])); in seqfile_dump_rcd()
|
D | tid_rdma.c | 2061 bth0 = be32_to_cpu(ohdr->bth[0]); in tid_rdma_rcv_error() 2245 bth0 = be32_to_cpu(ohdr->bth[0]); in hfi1_rc_rcv_tid_rdma_read_req() 2250 psn = mask_psn(be32_to_cpu(ohdr->bth[2])); in hfi1_rc_rcv_tid_rdma_read_req() 2464 kpsn = mask_psn(be32_to_cpu(ohdr->bth[2])); in hfi1_rc_rcv_tid_rdma_read_resp() 2466 opcode = (be32_to_cpu(ohdr->bth[0]) >> 24) & 0xff; in hfi1_rc_rcv_tid_rdma_read_resp() 2884 psn = mask_psn(be32_to_cpu(ohdr->bth[2])); in hfi1_handle_kdeth_eflags() 2885 opcode = (be32_to_cpu(ohdr->bth[0]) >> 24) & 0xff; in hfi1_handle_kdeth_eflags() 3674 bth0 = be32_to_cpu(ohdr->bth[0]); in hfi1_rc_rcv_tid_rdma_write_req() 3679 psn = mask_psn(be32_to_cpu(ohdr->bth[2])); in hfi1_rc_rcv_tid_rdma_write_req() 4053 psn = mask_psn(be32_to_cpu(ohdr->bth[2])); in hfi1_rc_rcv_tid_rdma_write_resp() [all …]
|
D | ipoib_tx.c | 328 ohdr->bth[0] = cpu_to_be32(bth0); in hfi1_ipoib_build_ib_tx_headers() 329 ohdr->bth[1] = cpu_to_be32(txp->dqpn); in hfi1_ipoib_build_ib_tx_headers() 330 ohdr->bth[2] = cpu_to_be32(mask_psn((u32)txp->txq->tx_ring.sent_txreqs)); in hfi1_ipoib_build_ib_tx_headers()
|
D | verbs.h | 442 return ohdr->bth[1] & cpu_to_be32(OPA_BTH_MIG_REQ); in opa_bth_is_migration()
|
D | rc.c | 1191 ohdr->bth[0] = cpu_to_be32(bth0); in hfi1_make_bth_aeth() 1192 ohdr->bth[1] = cpu_to_be32(bth1 | qp->remote_qpn); in hfi1_make_bth_aeth() 1193 ohdr->bth[2] = cpu_to_be32(mask_psn(qp->r_ack_psn)); in hfi1_make_bth_aeth() 1686 ohdr->bth[2] = cpu_to_be32(psn); in hfi1_rc_verbs_aborted() 2792 opfn_trigger_conn_request(qp, be32_to_cpu(ohdr->bth[1])); in hfi1_rc_rcv()
|
/drivers/infiniband/hw/qedr/ |
D | qedr_roce_cm.c | 447 udh->bth.solicited_event = !!(swr->send_flags & IB_SEND_SOLICITED); in qedr_gsi_build_header() 448 udh->bth.pkey = QEDR_ROCE_PKEY_DEFAULT; in qedr_gsi_build_header() 449 udh->bth.destination_qpn = htonl(ud_wr(swr)->remote_qpn); in qedr_gsi_build_header() 450 udh->bth.psn = htonl((qp->sq_psn++) & ((1 << 24) - 1)); in qedr_gsi_build_header() 451 udh->bth.opcode = IB_OPCODE_UD_SEND_ONLY; in qedr_gsi_build_header()
|
/drivers/infiniband/hw/mlx4/ |
D | qp.c | 2874 sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED); in build_sriov_qp0_header() 2878 sqp->ud_header.bth.pkey = cpu_to_be16(pkey); in build_sriov_qp0_header() 2880 sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->remote_qpn); in build_sriov_qp0_header() 2882 sqp->ud_header.bth.destination_qpn = in build_sriov_qp0_header() 2885 sqp->ud_header.bth.psn = cpu_to_be32((sqp->send_psn++) & ((1 << 24) - 1)); in build_sriov_qp0_header() 2896 sqp->ud_header.bth.opcode = IB_OPCODE_UD_SEND_ONLY; in build_sriov_qp0_header() 3124 sqp->ud_header.bth.opcode = IB_OPCODE_UD_SEND_ONLY; in build_mlx_header() 3128 sqp->ud_header.bth.opcode = IB_OPCODE_UD_SEND_ONLY_WITH_IMMEDIATE; in build_mlx_header() 3170 sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED); in build_mlx_header() 3180 sqp->ud_header.bth.pkey = cpu_to_be16(pkey); in build_mlx_header() [all …]
|
/drivers/infiniband/hw/mthca/ |
D | mthca_qp.c | 1527 sqp->ud_header.bth.opcode = IB_OPCODE_UD_SEND_ONLY; in build_mlx_header() 1531 sqp->ud_header.bth.opcode = IB_OPCODE_UD_SEND_ONLY_WITH_IMMEDIATE; in build_mlx_header() 1542 sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED); in build_mlx_header() 1549 sqp->ud_header.bth.pkey = cpu_to_be16(pkey); in build_mlx_header() 1550 sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->remote_qpn); in build_mlx_header() 1551 sqp->ud_header.bth.psn = cpu_to_be32((sqp->send_psn++) & ((1 << 24) - 1)); in build_mlx_header()
|
/drivers/net/ethernet/freescale/enetc/ |
D | enetc_hw.h | 749 __le32 bth; member 884 u32 bth; member
|
D | enetc_qos.c | 99 gcl_data->bth = cpu_to_le32(upper_32_bits(admin_conf->base_time)); in enetc_setup_taprio() 786 sgcl_data->bth = upper_32_bits(start); in enetc_streamgate_hw_set() 792 sgcl_data->bth = hi; in enetc_streamgate_hw_set()
|
/drivers/infiniband/hw/bnxt_re/ |
D | ib_verbs.c | 2224 qp->qp1_hdr.bth.opcode = IB_OPCODE_UD_SEND_ONLY_WITH_IMMEDIATE; in bnxt_re_build_qp1_send_v2() 2227 qp->qp1_hdr.bth.opcode = IB_OPCODE_UD_SEND_ONLY; in bnxt_re_build_qp1_send_v2() 2230 qp->qp1_hdr.bth.solicited_event = 1; in bnxt_re_build_qp1_send_v2() 2232 qp->qp1_hdr.bth.pad_count = (4 - payload_size) & 3; in bnxt_re_build_qp1_send_v2() 2235 qp->qp1_hdr.bth.pkey = cpu_to_be16(0xFFFF); in bnxt_re_build_qp1_send_v2() 2236 qp->qp1_hdr.bth.destination_qpn = IB_QP1; in bnxt_re_build_qp1_send_v2() 2237 qp->qp1_hdr.bth.ack_req = 0; in bnxt_re_build_qp1_send_v2() 2240 qp->qp1_hdr.bth.psn = cpu_to_be32(qp->send_psn); in bnxt_re_build_qp1_send_v2()
|