/drivers/infiniband/core/ |
D | ud_header.c | 48 { STRUCT_FIELD(lrh, virtual_lane), 52 { STRUCT_FIELD(lrh, link_version), 56 { STRUCT_FIELD(lrh, service_level), 64 { STRUCT_FIELD(lrh, link_next_header), 68 { STRUCT_FIELD(lrh, destination_lid), 76 { STRUCT_FIELD(lrh, packet_length), 80 { STRUCT_FIELD(lrh, source_lid), 339 header->lrh.link_version = 0; in ib_ud_header_init() 340 header->lrh.link_next_header = in ib_ud_header_init() 349 header->lrh.packet_length = cpu_to_be16(packet_length); in ib_ud_header_init() [all …]
|
/drivers/infiniband/hw/qib/ |
D | qib_ruc.c | 113 (be16_to_cpu(hdr->lrh[0]) >> 4) & 0xF, in qib_ruc_check_hdr() 115 hdr->lrh[3], hdr->lrh[1]); in qib_ruc_check_hdr() 119 if ((be16_to_cpu(hdr->lrh[3]) != in qib_ruc_check_hdr() 152 (be16_to_cpu(hdr->lrh[0]) >> 4) & 0xF, in qib_ruc_check_hdr() 154 hdr->lrh[3], hdr->lrh[1]); in qib_ruc_check_hdr() 158 if (be16_to_cpu(hdr->lrh[3]) != in qib_ruc_check_hdr() 228 priv->s_hdr->lrh[0] = cpu_to_be16(lrh0); in qib_make_ruc_header() 229 priv->s_hdr->lrh[1] = in qib_make_ruc_header() 231 priv->s_hdr->lrh[2] = in qib_make_ruc_header() 233 priv->s_hdr->lrh[3] = in qib_make_ruc_header()
|
D | qib_ud.c | 350 priv->s_hdr->lrh[0] = cpu_to_be16(lrh0); in qib_make_ud_req() 351 priv->s_hdr->lrh[1] = in qib_make_ud_req() 353 priv->s_hdr->lrh[2] = in qib_make_ud_req() 359 priv->s_hdr->lrh[3] = cpu_to_be16(lid); in qib_make_ud_req() 361 priv->s_hdr->lrh[3] = IB_LID_PERMISSIVE; in qib_make_ud_req() 465 if (unlikely(hdr->lrh[1] == IB_LID_PERMISSIVE || in qib_ud_rcv() 466 hdr->lrh[3] == IB_LID_PERMISSIVE)) in qib_ud_rcv() 476 (be16_to_cpu(hdr->lrh[0]) >> 4) & in qib_ud_rcv() 479 hdr->lrh[3], hdr->lrh[1]); in qib_ud_rcv() 489 (be16_to_cpu(hdr->lrh[0]) >> 12) == 15))) in qib_ud_rcv() [all …]
|
D | qib_driver.c | 308 u16 lid = be16_to_cpu(hdr->lrh[1]); in qib_rcv_hdrerr() 309 int lnh = be16_to_cpu(hdr->lrh[0]) & 3; in qib_rcv_hdrerr() 489 u16 lrh_len = be16_to_cpu(hdr->lrh[2]) << 2; in qib_kreceive()
|
D | qib_user_sdma.c | 449 hdr->lrh[2] = cpu_to_be16(le16_to_cpu(pbc16[0])); in qib_user_sdma_page_to_frags() 467 be16_to_cpu(hdr->lrh[2]) - in qib_user_sdma_page_to_frags() 495 hdr->lrh[2] = cpu_to_be16(le16_to_cpu(pbc16[0])); in qib_user_sdma_page_to_frags() 511 be16_to_cpu(hdr->lrh[2]) - in qib_user_sdma_page_to_frags()
|
D | qib_rc.c | 648 hdr.lrh[0] = cpu_to_be16(lrh0); in qib_send_rc_ack() 649 hdr.lrh[1] = cpu_to_be16(rdma_ah_get_dlid(&qp->remote_ah_attr)); in qib_send_rc_ack() 650 hdr.lrh[2] = cpu_to_be16(hwords + SIZE_OF_CRC); in qib_send_rc_ack() 651 hdr.lrh[3] = cpu_to_be16(ppd->lid | in qib_send_rc_ack() 893 if ((be16_to_cpu(hdr->lrh[0]) & 3) == QIB_LRH_BTH) in qib_rc_send_complete()
|
D | qib_common.h | 700 __be16 lrh[4]; member
|
D | qib_verbs.c | 290 lid = be16_to_cpu(hdr->lrh[1]); in qib_ib_rcv() 298 lnh = be16_to_cpu(hdr->lrh[0]) & 3; in qib_ib_rcv() 798 be16_to_cpu(hdr->lrh[0]) >> 12); in qib_verbs_send_dma() 922 be16_to_cpu(ibhdr->lrh[0]) >> 12); in qib_verbs_send_pio()
|
D | qib_iba7322.c | 1403 .lrh[0] = cpu_to_be16(0xF000 | QIB_LRH_BTH), in flush_fifo() 1404 .lrh[1] = IB_LID_PERMISSIVE, in flush_fifo() 1405 .lrh[2] = cpu_to_be16(hdrwords + SIZE_OF_CRC), in flush_fifo() 1406 .lrh[3] = IB_LID_PERMISSIVE, in flush_fifo()
|
/drivers/infiniband/hw/hfi1/ |
D | hfi.h | 415 return (u8)(hdr->lrh[2] & OPA_16B_L4_MASK); in hfi1_16B_get_l4() 420 return (u8)((hdr->lrh[1] & OPA_16B_SC_MASK) >> OPA_16B_SC_SHIFT); in hfi1_16B_get_sc() 425 return (u32)((hdr->lrh[1] & OPA_16B_LID_MASK) | in hfi1_16B_get_dlid() 426 (((hdr->lrh[2] & OPA_16B_DLID_MASK) >> in hfi1_16B_get_dlid() 432 return (u32)((hdr->lrh[0] & OPA_16B_LID_MASK) | in hfi1_16B_get_slid() 433 (((hdr->lrh[2] & OPA_16B_SLID_MASK) >> in hfi1_16B_get_slid() 439 return (u8)((hdr->lrh[0] & OPA_16B_BECN_MASK) >> OPA_16B_BECN_SHIFT); in hfi1_16B_get_becn() 444 return (u8)((hdr->lrh[1] & OPA_16B_FECN_MASK) >> OPA_16B_FECN_SHIFT); in hfi1_16B_get_fecn() 449 return (u8)((hdr->lrh[1] & OPA_16B_L2_MASK) >> OPA_16B_L2_SHIFT); in hfi1_16B_get_l2() 454 return (u16)((hdr->lrh[2] & OPA_16B_PKEY_MASK) >> OPA_16B_PKEY_SHIFT); in hfi1_16B_get_pkey() [all …]
|
D | trace_ibhdrs.h | 108 #define lrh_name(lrh) { HFI1_##lrh, #lrh } argument 109 #define show_lnh(lrh) \ argument 110 __print_symbolic(lrh, \
|
D | verbs.h | 68 #define LRH_16B_BYTES (sizeof_field(struct hfi1_16b_header, lrh)) 70 #define LRH_9B_BYTES (sizeof_field(struct ib_header, lrh)) 80 u32 lrh[4]; member
|
D | user_sdma.c | 437 sc = (((be16_to_cpu(req->hdr.lrh[0]) >> 12) & 0xF) | in hfi1_user_sdma_process_request() 448 slid = be16_to_cpu(req->hdr.lrh[3]); in hfi1_user_sdma_process_request() 459 if ((be16_to_cpu(req->hdr.lrh[0]) & 0x3) == HFI1_LRH_GRH) { in hfi1_user_sdma_process_request() 531 dlid = be16_to_cpu(req->hdr.lrh[1]); in hfi1_user_sdma_process_request() 611 len = ((be16_to_cpu(req->hdr.lrh[2]) << 2) - in compute_data_length() 965 hdr->lrh[2] = cpu_to_be16(lrhlen >> 2); in set_txreq_header() 981 req->hdr.lrh[2] = hdr->lrh[2]; in set_txreq_header()
|
D | trace_tx.h | 494 __be32 *lrh = (__be32 *)hdr->lrh; 504 __entry->lrh0 = be32_to_cpu(lrh[0]); 505 __entry->lrh1 = be32_to_cpu(lrh[1]);
|
D | trace.c | 152 return be16_to_cpu(hdr->lrh[2]); in ib_get_len()
|
D | verbs.c | 345 lnh = be16_to_cpu(hdr->lrh[0]) & 3; in hfi1_kdeth_eager_rcv() 399 lnh = be16_to_cpu(hdr->lrh[0]) & 3; in hfi1_kdeth_expected_rcv()
|
D | driver.c | 1870 lnh = be16_to_cpu(hdr->lrh[0]) & 3; in seqfile_dump_rcd()
|
D | tid_rdma.c | 2856 int lnh = be16_to_cpu(hdr->lrh[0]) & 3; in hfi1_handle_kdeth_eflags() 2857 u16 lid = be16_to_cpu(hdr->lrh[1]); in hfi1_handle_kdeth_eflags()
|
/drivers/infiniband/hw/mthca/ |
D | mthca_av.c | 275 header->lrh.service_level = be32_to_cpu(ah->av->sl_tclass_flowlabel) >> 28; in mthca_read_ah() 276 header->lrh.destination_lid = ah->av->dlid; in mthca_read_ah() 277 header->lrh.source_lid = cpu_to_be16(ah->av->g_slid & 0x7f); in mthca_read_ah()
|
D | mthca_qp.c | 1519 (sqp->ud_header.lrh.destination_lid == in build_mlx_header() 1521 (sqp->ud_header.lrh.service_level << 8)); in build_mlx_header() 1522 mlx->rlid = sqp->ud_header.lrh.destination_lid; in build_mlx_header() 1539 sqp->ud_header.lrh.virtual_lane = !qp->ibqp.qp_num ? 15 : 0; in build_mlx_header() 1540 if (sqp->ud_header.lrh.destination_lid == IB_LID_PERMISSIVE) in build_mlx_header() 1541 sqp->ud_header.lrh.source_lid = IB_LID_PERMISSIVE; in build_mlx_header()
|
/drivers/infiniband/hw/mlx4/ |
D | qp.c | 2859 sqp->ud_header.lrh.service_level = in build_sriov_qp0_header() 2861 sqp->ud_header.lrh.destination_lid = in build_sriov_qp0_header() 2863 sqp->ud_header.lrh.source_lid = in build_sriov_qp0_header() 2871 mlx->rlid = sqp->ud_header.lrh.destination_lid; in build_sriov_qp0_header() 2873 sqp->ud_header.lrh.virtual_lane = 0; in build_sriov_qp0_header() 3047 sqp->ud_header.lrh.service_level = in build_mlx_header() 3049 sqp->ud_header.lrh.destination_lid = ah->av.ib.dlid; in build_mlx_header() 3050 sqp->ud_header.lrh.source_lid = cpu_to_be16(ah->av.ib.g_slid & 0x7f); in build_mlx_header() 3112 (sqp->ud_header.lrh.destination_lid == in build_mlx_header() 3116 (sqp->ud_header.lrh.service_level << 8)); in build_mlx_header() [all …]
|