Home
last modified time | relevance | path

Searched refs:rdi (Results 1 – 25 of 52) sorted by relevance

123

/drivers/infiniband/sw/rdmavt/
Dvt.c50 struct rvt_dev_info *rdi; in rvt_alloc_device() local
52 rdi = container_of(_ib_alloc_device(size), struct rvt_dev_info, ibdev); in rvt_alloc_device()
53 if (!rdi) in rvt_alloc_device()
54 return rdi; in rvt_alloc_device()
56 rdi->ports = kcalloc(nports, sizeof(*rdi->ports), GFP_KERNEL); in rvt_alloc_device()
57 if (!rdi->ports) in rvt_alloc_device()
58 ib_dealloc_device(&rdi->ibdev); in rvt_alloc_device()
60 return rdi; in rvt_alloc_device()
70 void rvt_dealloc_device(struct rvt_dev_info *rdi) in rvt_dealloc_device() argument
72 kfree(rdi->ports); in rvt_dealloc_device()
[all …]
Dmmap.c16 void rvt_mmap_init(struct rvt_dev_info *rdi) in rvt_mmap_init() argument
18 INIT_LIST_HEAD(&rdi->pending_mmaps); in rvt_mmap_init()
19 spin_lock_init(&rdi->pending_lock); in rvt_mmap_init()
20 rdi->mmap_offset = PAGE_SIZE; in rvt_mmap_init()
21 spin_lock_init(&rdi->mmap_offset_lock); in rvt_mmap_init()
32 struct rvt_dev_info *rdi = ib_to_rvt(ip->context->device); in rvt_release_mmap_info() local
34 spin_lock_irq(&rdi->pending_lock); in rvt_release_mmap_info()
36 spin_unlock_irq(&rdi->pending_lock); in rvt_release_mmap_info()
70 struct rvt_dev_info *rdi = ib_to_rvt(context->device); in rvt_mmap() local
81 spin_lock_irq(&rdi->pending_lock); in rvt_mmap()
[all …]
Dqp.c22 static void rvt_reset_qp(struct rvt_dev_info *rdi, struct rvt_qp *qp,
103 void rvt_wss_exit(struct rvt_dev_info *rdi) in rvt_wss_exit() argument
105 struct rvt_wss *wss = rdi->wss; in rvt_wss_exit()
113 kfree(rdi->wss); in rvt_wss_exit()
114 rdi->wss = NULL; in rvt_wss_exit()
122 int rvt_wss_init(struct rvt_dev_info *rdi) in rvt_wss_init() argument
124 unsigned int sge_copy_mode = rdi->dparms.sge_copy_mode; in rvt_wss_init()
125 unsigned int wss_threshold = rdi->dparms.wss_threshold; in rvt_wss_init()
126 unsigned int wss_clean_period = rdi->dparms.wss_clean_period; in rvt_wss_init()
132 int node = rdi->dparms.node; in rvt_wss_init()
[all …]
Dmad.c60 int rvt_create_mad_agents(struct rvt_dev_info *rdi) in rvt_create_mad_agents() argument
67 for (p = 0; p < rdi->dparms.nports; p++) { in rvt_create_mad_agents()
68 rvp = rdi->ports[p]; in rvt_create_mad_agents()
69 agent = ib_register_mad_agent(&rdi->ibdev, p + 1, in rvt_create_mad_agents()
80 if (rdi->driver_f.notify_create_mad_agent) in rvt_create_mad_agents()
81 rdi->driver_f.notify_create_mad_agent(rdi, p); in rvt_create_mad_agents()
87 for (p = 0; p < rdi->dparms.nports; p++) { in rvt_create_mad_agents()
88 rvp = rdi->ports[p]; in rvt_create_mad_agents()
93 if (rdi->driver_f.notify_free_mad_agent) in rvt_create_mad_agents()
94 rdi->driver_f.notify_free_mad_agent(rdi, p); in rvt_create_mad_agents()
[all …]
Dvt.h21 #define rvt_pr_info(rdi, fmt, ...) \ argument
22 __rvt_pr_info(rdi->driver_f.get_pci_dev(rdi), \
23 rvt_get_ibdev_name(rdi), \
27 #define rvt_pr_warn(rdi, fmt, ...) \ argument
28 __rvt_pr_warn(rdi->driver_f.get_pci_dev(rdi), \
29 rvt_get_ibdev_name(rdi), \
33 #define rvt_pr_err(rdi, fmt, ...) \ argument
34 __rvt_pr_err(rdi->driver_f.get_pci_dev(rdi), \
35 rvt_get_ibdev_name(rdi), \
39 #define rvt_pr_err_ratelimited(rdi, fmt, ...) \ argument
[all …]
Dcq.c62 struct rvt_dev_info *rdi = cq->rdi; in rvt_cq_enter() local
65 rvt_pr_err_ratelimited(rdi, "CQ is full!\n"); in rvt_cq_enter()
162 struct rvt_dev_info *rdi = ib_to_rvt(ibdev); in rvt_create_cq() local
174 if (entries < 1 || entries > rdi->dparms.props.max_cqe) in rvt_create_cq()
180 comp_vector = comp_vector % rdi->ibdev.num_comp_vectors; in rvt_create_cq()
198 k_wc = vzalloc_node(sz, rdi->dparms.node); in rvt_create_cq()
208 cq->ip = rvt_create_mmap_info(rdi, sz, udata, u_wc); in rvt_create_cq()
220 spin_lock_irq(&rdi->n_cqs_lock); in rvt_create_cq()
221 if (rdi->n_cqs_allocated == rdi->dparms.props.max_cq) { in rvt_create_cq()
222 spin_unlock_irq(&rdi->n_cqs_lock); in rvt_create_cq()
[all …]
Dmcast.c20 void rvt_driver_mcast_init(struct rvt_dev_info *rdi) in rvt_driver_mcast_init() argument
26 spin_lock_init(&rdi->n_mcast_grps_lock); in rvt_driver_mcast_init()
151 static int rvt_mcast_add(struct rvt_dev_info *rdi, struct rvt_ibport *ibp, in rvt_mcast_add() argument
192 rdi->dparms.props.max_mcast_qp_attach) { in rvt_mcast_add()
204 spin_lock(&rdi->n_mcast_grps_lock); in rvt_mcast_add()
205 if (rdi->n_mcast_grps_allocated == rdi->dparms.props.max_mcast_grp) { in rvt_mcast_add()
206 spin_unlock(&rdi->n_mcast_grps_lock); in rvt_mcast_add()
211 rdi->n_mcast_grps_allocated++; in rvt_mcast_add()
212 spin_unlock(&rdi->n_mcast_grps_lock); in rvt_mcast_add()
241 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); in rvt_attach_mcast() local
[all …]
Dmr.c22 int rvt_driver_mr_init(struct rvt_dev_info *rdi) in rvt_driver_mr_init() argument
24 unsigned int lkey_table_size = rdi->dparms.lkey_table_size; in rvt_driver_mr_init()
36 spin_lock_init(&rdi->lkey_table.lock); in rvt_driver_mr_init()
40 rvt_pr_warn(rdi, "lkey bits %u too large, reduced to %u\n", in rvt_driver_mr_init()
42 rdi->dparms.lkey_table_size = RVT_MAX_LKEY_TABLE_BITS; in rvt_driver_mr_init()
43 lkey_table_size = rdi->dparms.lkey_table_size; in rvt_driver_mr_init()
45 rdi->lkey_table.max = 1 << lkey_table_size; in rvt_driver_mr_init()
46 rdi->lkey_table.shift = 32 - lkey_table_size; in rvt_driver_mr_init()
47 lk_tab_size = rdi->lkey_table.max * sizeof(*rdi->lkey_table.table); in rvt_driver_mr_init()
48 rdi->lkey_table.table = (struct rvt_mregion __rcu **) in rvt_driver_mr_init()
[all …]
Dtrace.h6 #define RDI_DEV_ENTRY(rdi) __string(dev, rvt_get_ibdev_name(rdi)) argument
7 #define RDI_DEV_ASSIGN(rdi) __assign_str(dev, rvt_get_ibdev_name(rdi)) argument
Dtrace_rvt.h18 TP_PROTO(struct rvt_dev_info *rdi,
20 TP_ARGS(rdi, msg),
22 RDI_DEV_ENTRY(rdi)
26 RDI_DEV_ASSIGN(rdi);
Dqp.h11 int rvt_driver_qp_init(struct rvt_dev_info *rdi);
12 void rvt_qp_exit(struct rvt_dev_info *rdi);
26 int rvt_wss_init(struct rvt_dev_info *rdi);
27 void rvt_wss_exit(struct rvt_dev_info *rdi);
Dmmap.h11 void rvt_mmap_init(struct rvt_dev_info *rdi);
14 struct rvt_mmap_info *rvt_create_mmap_info(struct rvt_dev_info *rdi, u32 size,
16 void rvt_update_mmap_info(struct rvt_dev_info *rdi, struct rvt_mmap_info *ip,
Dtrace_cq.h39 TP_STRUCT__entry(RDI_DEV_ENTRY(cq->rdi)
46 TP_fast_assign(RDI_DEV_ASSIGN(cq->rdi);
73 RDI_DEV_ENTRY(cq->rdi)
84 RDI_DEV_ASSIGN(cq->rdi);
Dah.c26 struct rvt_dev_info *rdi = ib_to_rvt(ibdev); in rvt_check_ah() local
42 if (rdi->driver_f.check_ah) in rvt_check_ah()
43 return rdi->driver_f.check_ah(ibdev, ah_attr); in rvt_check_ah()
Drc.c123 struct rvt_dev_info *rdi = ib_to_rvt(qp->ibqp.device); in rvt_get_credit() local
137 rdi->driver_f.schedule_send(qp); in rvt_get_credit()
147 rdi->driver_f.schedule_send(qp); in rvt_get_credit()
Dmcast.h11 void rvt_driver_mcast_init(struct rvt_dev_info *rdi);
14 int rvt_mcast_tree_empty(struct rvt_dev_info *rdi);
/drivers/infiniband/hw/qib/
Dqib_verbs.c277 struct rvt_dev_info *rdi = &dd->verbs_dev.rdi; in qib_ib_rcv() local
343 qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); in qib_ib_rcv()
370 spin_lock_irqsave(&dev->rdi.pending_lock, flags); in mem_timer()
379 spin_unlock_irqrestore(&dev->rdi.pending_lock, flags); in mem_timer()
567 spin_lock(&dev->rdi.pending_lock); in __get_txreq()
573 spin_unlock(&dev->rdi.pending_lock); in __get_txreq()
584 spin_unlock(&dev->rdi.pending_lock); in __get_txreq()
597 spin_lock_irqsave(&dev->rdi.pending_lock, flags); in get_txreq()
603 spin_unlock_irqrestore(&dev->rdi.pending_lock, flags); in get_txreq()
607 spin_unlock_irqrestore(&dev->rdi.pending_lock, flags); in get_txreq()
[all …]
Dqib_qp.c127 int qib_alloc_qpn(struct rvt_dev_info *rdi, struct rvt_qpn_table *qpt, in qib_alloc_qpn() argument
133 struct qib_ibdev *verbs_dev = container_of(rdi, struct qib_ibdev, rdi); in qib_alloc_qpn()
213 unsigned qib_free_all_qps(struct rvt_dev_info *rdi) in qib_free_all_qps() argument
215 struct qib_ibdev *verbs_dev = container_of(rdi, struct qib_ibdev, rdi); in qib_free_all_qps()
245 spin_lock(&dev->rdi.pending_lock); in qib_notify_error_qp()
250 spin_unlock(&dev->rdi.pending_lock); in qib_notify_error_qp()
291 int qib_get_pmtu_from_attr(struct rvt_dev_info *rdi, struct rvt_qp *qp, in qib_get_pmtu_from_attr() argument
295 struct qib_ibdev *verbs_dev = container_of(rdi, struct qib_ibdev, rdi); in qib_get_pmtu_from_attr()
314 u32 qib_mtu_from_qp(struct rvt_dev_info *rdi, struct rvt_qp *qp, u32 pmtu) in qib_mtu_from_qp() argument
319 void *qib_qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp) in qib_qp_priv_alloc() argument
[all …]
Dqib_verbs.h170 struct rvt_dev_info rdi; member
208 struct rvt_dev_info *rdi; in to_idev() local
210 rdi = container_of(ibdev, struct rvt_dev_info, ibdev); in to_idev()
211 return container_of(rdi, struct qib_ibdev, rdi); in to_idev()
242 void qib_cap_mask_chg(struct rvt_dev_info *rdi, u32 port_num);
249 void qib_notify_create_mad_agent(struct rvt_dev_info *rdi, int port_idx);
250 void qib_notify_free_mad_agent(struct rvt_dev_info *rdi, int port_idx);
271 unsigned qib_free_all_qps(struct rvt_dev_info *rdi);
272 void *qib_qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp);
273 void qib_qp_priv_free(struct rvt_dev_info *rdi, struct rvt_qp *qp);
[all …]
/drivers/media/platform/qcom/camss/
Dcamss-csid-gen2.c38 #define CSID_CSI2_RDIN_IRQ_STATUS(rdi) ((IS_LITE ? 0x30 : 0x40) \ argument
39 + 0x10 * (rdi))
40 #define CSID_CSI2_RDIN_IRQ_MASK(rdi) ((IS_LITE ? 0x34 : 0x44) \ argument
41 + 0x10 * (rdi))
42 #define CSID_CSI2_RDIN_IRQ_CLEAR(rdi) ((IS_LITE ? 0x38 : 0x48) \ argument
43 + 0x10 * (rdi))
44 #define CSID_CSI2_RDIN_IRQ_SET(rdi) ((IS_LITE ? 0x3C : 0x4C) \ argument
45 + 0x10 * (rdi))
76 #define CSID_RDI_CFG0(rdi) ((IS_LITE ? 0x200 : 0x300) \ argument
77 + 0x100 * (rdi))
[all …]
/drivers/infiniband/hw/hfi1/
Dverbs.c334 struct rvt_dev_info *rdi = &ppd->dd->verbs_dev.rdi; in hfi1_kdeth_eager_rcv() local
360 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); in hfi1_kdeth_eager_rcv()
388 struct rvt_dev_info *rdi = &ppd->dd->verbs_dev.rdi; in hfi1_kdeth_expected_rcv() local
414 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); in hfi1_kdeth_expected_rcv()
460 struct rvt_dev_info *rdi = &ppd->dd->verbs_dev.rdi; in hfi1_handle_packet() local
506 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); in hfi1_handle_packet()
1290 struct rvt_dev_info *rdi = &dd->verbs_dev.rdi; in hfi1_fill_device_attr() local
1293 memset(&rdi->dparms.props, 0, sizeof(rdi->dparms.props)); in hfi1_fill_device_attr()
1295 rdi->dparms.props.fw_ver = ((u64)(dc8051_ver_maj(ver)) << 32) | in hfi1_fill_device_attr()
1299 rdi->dparms.props.device_cap_flags = IB_DEVICE_BAD_PKEY_CNTR | in hfi1_fill_device_attr()
[all …]
Dqp.h93 void *qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp);
94 void qp_priv_free(struct rvt_dev_info *rdi, struct rvt_qp *qp);
95 unsigned free_all_qps(struct rvt_dev_info *rdi);
97 int get_pmtu_from_attr(struct rvt_dev_info *rdi, struct rvt_qp *qp,
103 u32 mtu_from_qp(struct rvt_dev_info *rdi, struct rvt_qp *qp, u32 pmtu);
Dqp.c639 rvt_max_atomic(&to_idev(qp->ibqp.device)->rdi), in qp_iter_print()
668 void *qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp) in qp_priv_alloc() argument
672 priv = kzalloc_node(sizeof(*priv), GFP_KERNEL, rdi->dparms.node); in qp_priv_alloc()
679 rdi->dparms.node); in qp_priv_alloc()
698 void qp_priv_free(struct rvt_dev_info *rdi, struct rvt_qp *qp) in qp_priv_free() argument
702 hfi1_qp_priv_tid_free(rdi, qp); in qp_priv_free()
707 unsigned free_all_qps(struct rvt_dev_info *rdi) in free_all_qps() argument
709 struct hfi1_ibdev *verbs_dev = container_of(rdi, in free_all_qps()
711 rdi); in free_all_qps()
799 u32 mtu_from_qp(struct rvt_dev_info *rdi, struct rvt_qp *qp, u32 pmtu) in mtu_from_qp() argument
[all …]
Dsysfs.c410 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in hw_rev_show()
420 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in board_id_show()
434 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in boardversion_show()
446 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in nctxts_show()
465 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in nfreectxts_show()
477 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in serial_show()
490 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in chip_reset_store()
519 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in tempsense_show()
655 struct ib_device *dev = &dd->verbs_dev.rdi.ibdev; in hfi1_verbs_register_sysfs()
Dverbs.h236 struct rvt_dev_info rdi; /* Must be first */ member
267 struct rvt_dev_info *rdi; in to_idev() local
269 rdi = container_of(ibdev, struct rvt_dev_info, ibdev); in to_idev()
270 return container_of(rdi, struct hfi1_ibdev, rdi); in to_idev()
286 void hfi1_cap_mask_chg(struct rvt_dev_info *rdi, u32 port_num);

123