/drivers/infiniband/sw/rdmavt/ |
D | mr.c | 180 mr->lkey = (r << (32 - dev->dparms.lkey_table_size)) | in rvt_alloc_lkey() 183 if (mr->lkey == 0) { in rvt_alloc_lkey() 184 mr->lkey |= 1 << 8; in rvt_alloc_lkey() 208 u32 lkey = mr->lkey; in rvt_free_lkey() local 215 if (!lkey) { in rvt_free_lkey() 225 r = lkey >> (32 - dev->dparms.lkey_table_size); in rvt_free_lkey() 259 mr->ibmr.lkey = mr->mr.lkey; in __rvt_alloc_mr() 260 mr->ibmr.rkey = mr->mr.lkey; in __rvt_alloc_mr() 414 rvt_qp_mr_clean(qp, mr->lkey); in rvt_dereg_clean_qp_cb() 447 if (mr->lkey) { in rvt_check_refs() [all …]
|
D | trace_mr.h | 31 __field(u32, lkey) 42 __entry->lkey = mr->lkey; 52 __entry->lkey, 91 __field(u32, lkey) 106 __entry->lkey = sge->mr->lkey; 122 __entry->lkey,
|
D | qp.c | 621 static bool rvt_swqe_has_lkey(struct rvt_swqe *wqe, u32 lkey) in rvt_swqe_has_lkey() argument 628 if (rvt_mr_has_lkey(sge->mr, lkey)) in rvt_swqe_has_lkey() 639 static bool rvt_qp_sends_has_lkey(struct rvt_qp *qp, u32 lkey) in rvt_qp_sends_has_lkey() argument 646 if (rvt_swqe_has_lkey(wqe, lkey)) in rvt_qp_sends_has_lkey() 653 if (rvt_mr_has_lkey(qp->s_rdma_mr, lkey)) in rvt_qp_sends_has_lkey() 663 static bool rvt_qp_acks_has_lkey(struct rvt_qp *qp, u32 lkey) in rvt_qp_acks_has_lkey() argument 671 if (rvt_mr_has_lkey(e->rdma_sge.mr, lkey)) in rvt_qp_acks_has_lkey() 688 void rvt_qp_mr_clean(struct rvt_qp *qp, u32 lkey) in rvt_qp_mr_clean() argument 703 if (rvt_ss_has_lkey(&qp->r_sge, lkey) || in rvt_qp_mr_clean() 704 rvt_qp_sends_has_lkey(qp, lkey) || in rvt_qp_mr_clean() [all …]
|
/drivers/infiniband/hw/mthca/ |
D | mthca_wqe.h | 66 __be32 lkey; member 85 __be32 lkey; member 103 __be32 lkey; member 120 dseg->lkey = cpu_to_be32(sg->lkey); in mthca_set_data_seg() 127 dseg->lkey = cpu_to_be32(MTHCA_INVAL_LKEY); in mthca_set_data_seg_inval()
|
D | mthca_mr.c | 57 __be32 lkey; member 444 mr->ibmr.rkey = mr->ibmr.lkey = hw_index_to_key(dev, key); in mthca_mr_alloc() 472 memset(&mpt_entry->lkey, 0, in mthca_mr_alloc() 473 sizeof *mpt_entry - offsetof(struct mthca_mpt_entry, lkey)); in mthca_mr_alloc() 481 mthca_dbg(dev, "Dumping MPT entry %08x:\n", mr->ibmr.lkey); in mthca_mr_alloc() 545 static void mthca_free_region(struct mthca_dev *dev, u32 lkey) in mthca_free_region() argument 548 key_to_hw_index(dev, lkey)); in mthca_free_region() 550 mthca_free(&dev->mr_table.mpt_alloc, key_to_hw_index(dev, lkey)); in mthca_free_region() 558 key_to_hw_index(dev, mr->ibmr.lkey) & in mthca_free_mr() 563 mthca_free_region(dev, mr->ibmr.lkey); in mthca_free_mr()
|
D | mthca_provider.c | 423 srq->mr.ibmr.lkey = ucmd.lkey; in mthca_create_srq() 503 qp->mr.ibmr.lkey = ucmd.lkey; in mthca_create_qp() 622 cq->buf.mr.ibmr.lkey = ucmd.lkey; in mthca_create_cq() 709 u32 lkey; in mthca_resize_cq() local 727 lkey = cq->resize_buf->buf.mr.ibmr.lkey; in mthca_resize_cq() 733 lkey = ucmd.lkey; in mthca_resize_cq() 736 ret = mthca_RESIZE_CQ(dev, cq->cqn, lkey, ilog2(entries)); in mthca_resize_cq()
|
D | mthca_srq.c | 53 __be32 lkey; member 62 __be32 lkey; member 110 context->lkey = cpu_to_be32(srq->mr.ibmr.lkey); in mthca_tavor_init_srq_context() 137 context->lkey = cpu_to_be32(srq->mr.ibmr.lkey); in mthca_arbel_init_srq_context() 198 scatter->lkey = cpu_to_be32(MTHCA_INVAL_LKEY); in mthca_alloc_srq_buf()
|
/drivers/infiniband/sw/rxe/ |
D | rxe_mr.c | 48 u32 lkey = mr->pelem.index << 8 | rxe_get_next_key(-1); in rxe_mr_init() local 49 u32 rkey = (access & IB_ACCESS_REMOTE) ? lkey : 0; in rxe_mr_init() 56 mr->lkey = mr->ibmr.lkey = lkey; in rxe_mr_init() 396 mr = lookup_mr(pd, access, sge->lkey, RXE_LOOKUP_LOCAL); in copy_data() 421 mr = lookup_mr(pd, access, sge->lkey, in copy_data() 514 if (unlikely((type == RXE_LOOKUP_LOCAL && mr->lkey != key) || in lookup_mr() 538 if (mr->rkey ? (key != mr->rkey) : (key != mr->lkey)) { in rxe_invalidate_mr() 540 __func__, key, (mr->rkey ? mr->rkey : mr->lkey)); in rxe_invalidate_mr() 577 __func__, mr->lkey); in rxe_reg_fast_mr() 589 if (unlikely((mr->lkey & ~0xff) != (key & ~0xff))) { in rxe_reg_fast_mr() [all …]
|
/drivers/infiniband/ulp/iser/ |
D | iser_memory.c | 109 reg->sge.lkey = device->pd->local_dma_lkey; in iser_reg_dma() 123 " length=0x%x\n", reg->sge.lkey, reg->rkey, in iser_reg_dma() 279 sig_reg->sge.lkey = mr->lkey; in iser_reg_sig_mr() 285 sig_reg->sge.lkey, sig_reg->rkey, sig_reg->sge.addr, in iser_reg_sig_mr() 327 reg->sge.lkey = mr->lkey; in iser_fast_reg_mr() 333 reg->sge.lkey, reg->rkey, reg->sge.addr, reg->sge.length); in iser_fast_reg_mr()
|
D | iser_initiator.c | 155 tx_dsg->lkey = mem_reg->sge.lkey; in iser_prepare_write_cmd() 280 rx_sg->lkey = device->pd->local_dma_lkey; in iser_alloc_rx_descriptors() 476 tx_dsg->lkey = mem_reg->sge.lkey; in iser_send_data_out() 538 tx_dsg->lkey = device->pd->local_dma_lkey; in iser_send_control()
|
/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_mr.c | 94 mr->ibmr.lkey = resp->lkey; in pvrdma_get_dma_mr() 182 mr->ibmr.lkey = resp->lkey; in pvrdma_reg_user_mr() 254 mr->ibmr.lkey = resp->lkey; in pvrdma_alloc_mr()
|
/drivers/infiniband/core/ |
D | uverbs_std_types_mr.c | 146 ret = uverbs_copy_to(attrs, UVERBS_ATTR_REG_DM_MR_RESP_LKEY, &mr->lkey, in UVERBS_HANDLER() 147 sizeof(mr->lkey)); in UVERBS_HANDLER() 163 ret = uverbs_copy_to(attrs, UVERBS_ATTR_QUERY_MR_RESP_LKEY, &mr->lkey, in UVERBS_HANDLER() 164 sizeof(mr->lkey)); in UVERBS_HANDLER() 260 &mr->lkey, sizeof(mr->lkey)); in UVERBS_HANDLER()
|
D | rw.c | 79 reg->inv_wr.ex.invalidate_rkey = reg->mr->lkey; in rdma_rw_inv_key() 168 reg->wr.wr.ex.invalidate_rkey = reg->mr->lkey; in rdma_rw_init_mr_wrs() 229 sge->lkey = qp->pd->local_dma_lkey; in rdma_rw_init_map_wrs() 258 ctx->single.sge.lkey = qp->pd->local_dma_lkey; in rdma_rw_init_single_wr() 444 ctx->reg->reg_wr.key = ctx->reg->mr->lkey; in rdma_rw_ctx_signature_init() 488 ib_update_fast_reg_key(reg->mr, ib_inc_rkey(reg->mr->lkey)); in rdma_rw_update_lkey() 489 reg->reg_wr.key = reg->mr->lkey; in rdma_rw_update_lkey() 490 reg->sge.lkey = reg->mr->lkey; in rdma_rw_update_lkey()
|
/drivers/infiniband/sw/siw/ |
D | siw_mem.c | 121 mr->base_mr.lkey = mr->base_mr.rkey = mem->stag; in siw_mr_add_mem() 234 new = siw_mem_id2obj(sdev, sge->lkey >> 8); in siw_check_sge() 236 siw_dbg_pd(pd, "STag unknown: 0x%08x\n", sge->lkey); in siw_check_sge() 243 if (unlikely((*mem)->stag != sge->lkey)) { in siw_check_sge() 244 siw_dbg_mem((*mem), "STag mismatch: 0x%08x\n", sge->lkey); in siw_check_sge()
|
D | siw_qp.c | 279 wqe->sqe.sge[0].lkey = 0; in siw_qp_mpa_rts() 486 rreq->source_stag = htonl(wqe->sqe.sge[0].lkey); in siw_send_terminate() 875 rreq->sge[0].lkey = sqe->sge[0].lkey; in siw_read_to_orq() 876 rreq->sge[1].lkey = sqe->sge[1].lkey; in siw_read_to_orq() 912 wqe->sqe.sge[0].lkey = 0; in siw_activate_tx_from_sq() 1008 wqe->sqe.sge[0].lkey = irqe->sge[0].lkey; in siw_activate_tx()
|
D | siw_verbs.h | 30 siw_sge->lkey = sge->lkey; in siw_copy_sgl()
|
D | siw_qp_rx.c | 176 srx->ddp_stag = wqe->sqe.sge[0].lkey; in siw_rresp_check_ntoh() 368 wqe->rqe.sge[i].lkey = rqe->sge[i].lkey; in siw_rqe_get() 668 lkey = be32_to_cpu(srx->hdr.rreq.source_stag), in siw_init_rresp() local 705 resp->sge[0].lkey = lkey; in siw_init_rresp() 761 wqe->sqe.sge[0].lkey = orqe->sge[0].lkey; in siw_orqe_start_rx() 1286 rv = siw_invalidate_stag(qp->pd, wqe->sqe.sge[0].lkey); in siw_rdmap_complete()
|
/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_send.c | 16 u32 lkey; member 256 wq_dseg->lkey = cpu_to_be32(data_seg->lkey); in dr_rdma_segments() 353 send_info->read.lkey = send_ring->mr->mkey.key; in dr_fill_data_segs() 391 send_info->write.lkey = send_ring->mr->mkey.key; in dr_postsend_icm_data() 454 send_info.write.lkey = 0; in mlx5dr_send_postsend_ste() 512 send_info.write.lkey = 0; in mlx5dr_send_postsend_htbl() 569 send_info.write.lkey = 0; in mlx5dr_send_postsend_formatted_htbl() 593 send_info.write.lkey = 0; in mlx5dr_send_postsend_action() 1039 send_info.write.lkey = 0; in mlx5dr_send_ring_force_drain()
|
/drivers/infiniband/hw/mlx4/ |
D | srq.c | 160 scatter->lkey = cpu_to_be32(MLX4_INVALID_LKEY); in mlx4_ib_create_srq() 348 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx4_ib_post_srq_recv() 354 scat[i].lkey = cpu_to_be32(MLX4_INVALID_LKEY); in mlx4_ib_post_srq_recv()
|
/drivers/infiniband/hw/bnxt_re/ |
D | ib_verbs.c | 114 sg_list[i].lkey = ib_sg_list[i].lkey; in bnxt_re_build_sgl() 400 wqe->bind.parent_l_key = ib_mr->lkey; in bnxt_re_create_fence_wqe() 455 if (mr->ib_mr.lkey) in bnxt_re_destroy_fence_mr() 507 mr->ib_mr.lkey = mr->qplib_mr.lkey; in bnxt_re_create_fence_mr() 2248 wqe->sg_list[i].lkey = wqe->sg_list[i - 1].lkey; in bnxt_re_build_qp1_send_v2() 2274 wqe->sg_list[0].lkey = sge.lkey; in bnxt_re_build_qp1_send_v2() 2313 ref.lkey = wqe->sg_list[0].lkey; in bnxt_re_build_qp1_shadow_qp_recv() 2320 wqe->sg_list[0].lkey = sge.lkey; in bnxt_re_build_qp1_shadow_qp_recv() 2325 sqp_entry->sge.lkey = ref.lkey; in bnxt_re_build_qp1_shadow_qp_recv() 3204 s_sge[0].lkey = 0xFFFFFFFF; in bnxt_re_process_raw_qp_pkt_rx() [all …]
|
D | qplib_sp.c | 573 if (mrw->lkey == 0xFFFFFFFF) { in bnxt_qplib_free_mrw() 587 req.key = cpu_to_le32(mrw->lkey); in bnxt_qplib_free_mrw() 632 mrw->lkey = le32_to_cpu(resp.xid); in bnxt_qplib_alloc_mrw() 647 req.lkey = cpu_to_le32(mrw->lkey); in bnxt_qplib_dereg_mrw() 722 req.key = cpu_to_le32(mr->lkey); in bnxt_qplib_reg_mr()
|
/drivers/infiniband/hw/mlx5/ |
D | odp.c | 137 pklm->key = cpu_to_be32(mtt->ibmr.lkey); in populate_klm() 431 mr->ibmr.lkey = mr->mmkey.key; in implicit_get_child_mr() 505 imr->ibmr.lkey = imr->mmkey.key; in mlx5_ib_alloc_implicit_mr() 1020 key = be32_to_cpu(dseg->lkey); in pagefault_data_segments() 1704 u32 lkey) in get_prefetchable_mr() argument 1711 mmkey = xa_load(&dev->odp_mkeys, mlx5_base_mkey(lkey)); in get_prefetchable_mr() 1712 if (!mmkey || mmkey->key != lkey || mmkey->type != MLX5_MKEY_MR) in get_prefetchable_mr() 1771 get_prefetchable_mr(pd, advice, sg_list[i].lkey); in init_prefetch_work() 1793 mr = get_prefetchable_mr(pd, advice, sg_list[i].lkey); in mlx5_ib_prefetch_sg_list()
|
/drivers/net/ethernet/ibm/ehea/ |
D | ehea_main.c | 489 rwqe->sg_list[0].l_key = pr->recv_mr.lkey; in ehea_refill_rq_def() 1602 u32 lkey) in write_swqe2_immediate() argument 1626 sg1entry->l_key = lkey; in write_swqe2_immediate() 1639 struct ehea_swqe *swqe, u32 lkey) in write_swqe2_data() argument 1650 write_swqe2_immediate(skb, swqe, lkey); in write_swqe2_data() 1659 sg1entry->l_key = lkey; in write_swqe2_data() 1672 sgentry->l_key = lkey; in write_swqe2_data() 1987 struct ehea_swqe *swqe, u32 lkey) in ehea_xmit2() argument 1993 write_swqe2_data(skb, dev, swqe, lkey); in ehea_xmit2() 2016 u32 lkey; in ehea_start_xmit() local [all …]
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.c | 710 mr->ibmr.lkey = mr->hwmr.lkey; in ocrdma_alloc_lkey() 712 mr->ibmr.rkey = mr->hwmr.lkey; in ocrdma_alloc_lkey() 888 mr->ibmr.lkey = mr->hwmr.lkey; in ocrdma_reg_user_mr() 890 mr->ibmr.rkey = mr->hwmr.lkey; in ocrdma_reg_user_mr() 906 (void) ocrdma_mbx_dealloc_lkey(dev, mr->hwmr.fr_mr, mr->hwmr.lkey); in ocrdma_dereg_mr() 1900 sge[i].lrkey = sg_list[i].lkey; in ocrdma_build_sges() 2049 hdr->lkey = wr->key; in ocrdma_build_reg() 2141 hdr->lkey = wr->ex.invalidate_rkey; in ocrdma_post_send() 2161 hdr->lkey = wr->ex.invalidate_rkey; in ocrdma_post_send() 2938 mr->ibmr.rkey = mr->hwmr.lkey; in ocrdma_alloc_mr() [all …]
|
/drivers/infiniband/ulp/ipoib/ |
D | ipoib_verbs.c | 222 priv->tx_sge[i].lkey = priv->pd->local_dma_lkey; in ipoib_transport_dev_init() 228 priv->rx_sge[0].lkey = priv->pd->local_dma_lkey; in ipoib_transport_dev_init()
|