/drivers/infiniband/sw/rdmavt/ |
D | vt.c | 50 struct rvt_dev_info *rdi; in rvt_alloc_device() local 52 rdi = container_of(_ib_alloc_device(size), struct rvt_dev_info, ibdev); in rvt_alloc_device() 53 if (!rdi) in rvt_alloc_device() 54 return rdi; in rvt_alloc_device() 56 rdi->ports = kcalloc(nports, sizeof(*rdi->ports), GFP_KERNEL); in rvt_alloc_device() 57 if (!rdi->ports) in rvt_alloc_device() 58 ib_dealloc_device(&rdi->ibdev); in rvt_alloc_device() 60 return rdi; in rvt_alloc_device() 70 void rvt_dealloc_device(struct rvt_dev_info *rdi) in rvt_dealloc_device() argument 72 kfree(rdi->ports); in rvt_dealloc_device() [all …]
|
D | mmap.c | 16 void rvt_mmap_init(struct rvt_dev_info *rdi) in rvt_mmap_init() argument 18 INIT_LIST_HEAD(&rdi->pending_mmaps); in rvt_mmap_init() 19 spin_lock_init(&rdi->pending_lock); in rvt_mmap_init() 20 rdi->mmap_offset = PAGE_SIZE; in rvt_mmap_init() 21 spin_lock_init(&rdi->mmap_offset_lock); in rvt_mmap_init() 32 struct rvt_dev_info *rdi = ib_to_rvt(ip->context->device); in rvt_release_mmap_info() local 34 spin_lock_irq(&rdi->pending_lock); in rvt_release_mmap_info() 36 spin_unlock_irq(&rdi->pending_lock); in rvt_release_mmap_info() 70 struct rvt_dev_info *rdi = ib_to_rvt(context->device); in rvt_mmap() local 81 spin_lock_irq(&rdi->pending_lock); in rvt_mmap() [all …]
|
D | qp.c | 22 static void rvt_reset_qp(struct rvt_dev_info *rdi, struct rvt_qp *qp, 103 void rvt_wss_exit(struct rvt_dev_info *rdi) in rvt_wss_exit() argument 105 struct rvt_wss *wss = rdi->wss; in rvt_wss_exit() 113 kfree(rdi->wss); in rvt_wss_exit() 114 rdi->wss = NULL; in rvt_wss_exit() 122 int rvt_wss_init(struct rvt_dev_info *rdi) in rvt_wss_init() argument 124 unsigned int sge_copy_mode = rdi->dparms.sge_copy_mode; in rvt_wss_init() 125 unsigned int wss_threshold = rdi->dparms.wss_threshold; in rvt_wss_init() 126 unsigned int wss_clean_period = rdi->dparms.wss_clean_period; in rvt_wss_init() 132 int node = rdi->dparms.node; in rvt_wss_init() [all …]
|
D | mad.c | 60 int rvt_create_mad_agents(struct rvt_dev_info *rdi) in rvt_create_mad_agents() argument 67 for (p = 0; p < rdi->dparms.nports; p++) { in rvt_create_mad_agents() 68 rvp = rdi->ports[p]; in rvt_create_mad_agents() 69 agent = ib_register_mad_agent(&rdi->ibdev, p + 1, in rvt_create_mad_agents() 80 if (rdi->driver_f.notify_create_mad_agent) in rvt_create_mad_agents() 81 rdi->driver_f.notify_create_mad_agent(rdi, p); in rvt_create_mad_agents() 87 for (p = 0; p < rdi->dparms.nports; p++) { in rvt_create_mad_agents() 88 rvp = rdi->ports[p]; in rvt_create_mad_agents() 93 if (rdi->driver_f.notify_free_mad_agent) in rvt_create_mad_agents() 94 rdi->driver_f.notify_free_mad_agent(rdi, p); in rvt_create_mad_agents() [all …]
|
D | vt.h | 21 #define rvt_pr_info(rdi, fmt, ...) \ argument 22 __rvt_pr_info(rdi->driver_f.get_pci_dev(rdi), \ 23 rvt_get_ibdev_name(rdi), \ 27 #define rvt_pr_warn(rdi, fmt, ...) \ argument 28 __rvt_pr_warn(rdi->driver_f.get_pci_dev(rdi), \ 29 rvt_get_ibdev_name(rdi), \ 33 #define rvt_pr_err(rdi, fmt, ...) \ argument 34 __rvt_pr_err(rdi->driver_f.get_pci_dev(rdi), \ 35 rvt_get_ibdev_name(rdi), \ 39 #define rvt_pr_err_ratelimited(rdi, fmt, ...) \ argument [all …]
|
D | cq.c | 62 struct rvt_dev_info *rdi = cq->rdi; in rvt_cq_enter() local 65 rvt_pr_err_ratelimited(rdi, "CQ is full!\n"); in rvt_cq_enter() 162 struct rvt_dev_info *rdi = ib_to_rvt(ibdev); in rvt_create_cq() local 174 if (entries < 1 || entries > rdi->dparms.props.max_cqe) in rvt_create_cq() 180 comp_vector = comp_vector % rdi->ibdev.num_comp_vectors; in rvt_create_cq() 198 k_wc = vzalloc_node(sz, rdi->dparms.node); in rvt_create_cq() 208 cq->ip = rvt_create_mmap_info(rdi, sz, udata, u_wc); in rvt_create_cq() 220 spin_lock_irq(&rdi->n_cqs_lock); in rvt_create_cq() 221 if (rdi->n_cqs_allocated == rdi->dparms.props.max_cq) { in rvt_create_cq() 222 spin_unlock_irq(&rdi->n_cqs_lock); in rvt_create_cq() [all …]
|
D | mcast.c | 20 void rvt_driver_mcast_init(struct rvt_dev_info *rdi) in rvt_driver_mcast_init() argument 26 spin_lock_init(&rdi->n_mcast_grps_lock); in rvt_driver_mcast_init() 151 static int rvt_mcast_add(struct rvt_dev_info *rdi, struct rvt_ibport *ibp, in rvt_mcast_add() argument 192 rdi->dparms.props.max_mcast_qp_attach) { in rvt_mcast_add() 204 spin_lock(&rdi->n_mcast_grps_lock); in rvt_mcast_add() 205 if (rdi->n_mcast_grps_allocated == rdi->dparms.props.max_mcast_grp) { in rvt_mcast_add() 206 spin_unlock(&rdi->n_mcast_grps_lock); in rvt_mcast_add() 211 rdi->n_mcast_grps_allocated++; in rvt_mcast_add() 212 spin_unlock(&rdi->n_mcast_grps_lock); in rvt_mcast_add() 241 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); in rvt_attach_mcast() local [all …]
|
D | mr.c | 22 int rvt_driver_mr_init(struct rvt_dev_info *rdi) in rvt_driver_mr_init() argument 24 unsigned int lkey_table_size = rdi->dparms.lkey_table_size; in rvt_driver_mr_init() 36 spin_lock_init(&rdi->lkey_table.lock); in rvt_driver_mr_init() 40 rvt_pr_warn(rdi, "lkey bits %u too large, reduced to %u\n", in rvt_driver_mr_init() 42 rdi->dparms.lkey_table_size = RVT_MAX_LKEY_TABLE_BITS; in rvt_driver_mr_init() 43 lkey_table_size = rdi->dparms.lkey_table_size; in rvt_driver_mr_init() 45 rdi->lkey_table.max = 1 << lkey_table_size; in rvt_driver_mr_init() 46 rdi->lkey_table.shift = 32 - lkey_table_size; in rvt_driver_mr_init() 47 lk_tab_size = rdi->lkey_table.max * sizeof(*rdi->lkey_table.table); in rvt_driver_mr_init() 48 rdi->lkey_table.table = (struct rvt_mregion __rcu **) in rvt_driver_mr_init() [all …]
|
D | trace.h | 6 #define RDI_DEV_ENTRY(rdi) __string(dev, rvt_get_ibdev_name(rdi)) argument 7 #define RDI_DEV_ASSIGN(rdi) __assign_str(dev, rvt_get_ibdev_name(rdi)) argument
|
D | trace_rvt.h | 18 TP_PROTO(struct rvt_dev_info *rdi, 20 TP_ARGS(rdi, msg), 22 RDI_DEV_ENTRY(rdi) 26 RDI_DEV_ASSIGN(rdi);
|
D | qp.h | 11 int rvt_driver_qp_init(struct rvt_dev_info *rdi); 12 void rvt_qp_exit(struct rvt_dev_info *rdi); 26 int rvt_wss_init(struct rvt_dev_info *rdi); 27 void rvt_wss_exit(struct rvt_dev_info *rdi);
|
D | mmap.h | 11 void rvt_mmap_init(struct rvt_dev_info *rdi); 14 struct rvt_mmap_info *rvt_create_mmap_info(struct rvt_dev_info *rdi, u32 size, 16 void rvt_update_mmap_info(struct rvt_dev_info *rdi, struct rvt_mmap_info *ip,
|
D | trace_cq.h | 39 TP_STRUCT__entry(RDI_DEV_ENTRY(cq->rdi) 46 TP_fast_assign(RDI_DEV_ASSIGN(cq->rdi); 73 RDI_DEV_ENTRY(cq->rdi) 84 RDI_DEV_ASSIGN(cq->rdi);
|
D | ah.c | 26 struct rvt_dev_info *rdi = ib_to_rvt(ibdev); in rvt_check_ah() local 42 if (rdi->driver_f.check_ah) in rvt_check_ah() 43 return rdi->driver_f.check_ah(ibdev, ah_attr); in rvt_check_ah()
|
D | rc.c | 123 struct rvt_dev_info *rdi = ib_to_rvt(qp->ibqp.device); in rvt_get_credit() local 137 rdi->driver_f.schedule_send(qp); in rvt_get_credit() 147 rdi->driver_f.schedule_send(qp); in rvt_get_credit()
|
D | mcast.h | 11 void rvt_driver_mcast_init(struct rvt_dev_info *rdi); 14 int rvt_mcast_tree_empty(struct rvt_dev_info *rdi);
|
/drivers/infiniband/hw/qib/ |
D | qib_verbs.c | 277 struct rvt_dev_info *rdi = &dd->verbs_dev.rdi; in qib_ib_rcv() local 343 qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); in qib_ib_rcv() 370 spin_lock_irqsave(&dev->rdi.pending_lock, flags); in mem_timer() 379 spin_unlock_irqrestore(&dev->rdi.pending_lock, flags); in mem_timer() 567 spin_lock(&dev->rdi.pending_lock); in __get_txreq() 573 spin_unlock(&dev->rdi.pending_lock); in __get_txreq() 584 spin_unlock(&dev->rdi.pending_lock); in __get_txreq() 597 spin_lock_irqsave(&dev->rdi.pending_lock, flags); in get_txreq() 603 spin_unlock_irqrestore(&dev->rdi.pending_lock, flags); in get_txreq() 607 spin_unlock_irqrestore(&dev->rdi.pending_lock, flags); in get_txreq() [all …]
|
D | qib_qp.c | 127 int qib_alloc_qpn(struct rvt_dev_info *rdi, struct rvt_qpn_table *qpt, in qib_alloc_qpn() argument 133 struct qib_ibdev *verbs_dev = container_of(rdi, struct qib_ibdev, rdi); in qib_alloc_qpn() 213 unsigned qib_free_all_qps(struct rvt_dev_info *rdi) in qib_free_all_qps() argument 215 struct qib_ibdev *verbs_dev = container_of(rdi, struct qib_ibdev, rdi); in qib_free_all_qps() 245 spin_lock(&dev->rdi.pending_lock); in qib_notify_error_qp() 250 spin_unlock(&dev->rdi.pending_lock); in qib_notify_error_qp() 291 int qib_get_pmtu_from_attr(struct rvt_dev_info *rdi, struct rvt_qp *qp, in qib_get_pmtu_from_attr() argument 295 struct qib_ibdev *verbs_dev = container_of(rdi, struct qib_ibdev, rdi); in qib_get_pmtu_from_attr() 314 u32 qib_mtu_from_qp(struct rvt_dev_info *rdi, struct rvt_qp *qp, u32 pmtu) in qib_mtu_from_qp() argument 319 void *qib_qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp) in qib_qp_priv_alloc() argument [all …]
|
D | qib_verbs.h | 170 struct rvt_dev_info rdi; member 208 struct rvt_dev_info *rdi; in to_idev() local 210 rdi = container_of(ibdev, struct rvt_dev_info, ibdev); in to_idev() 211 return container_of(rdi, struct qib_ibdev, rdi); in to_idev() 242 void qib_cap_mask_chg(struct rvt_dev_info *rdi, u32 port_num); 249 void qib_notify_create_mad_agent(struct rvt_dev_info *rdi, int port_idx); 250 void qib_notify_free_mad_agent(struct rvt_dev_info *rdi, int port_idx); 271 unsigned qib_free_all_qps(struct rvt_dev_info *rdi); 272 void *qib_qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp); 273 void qib_qp_priv_free(struct rvt_dev_info *rdi, struct rvt_qp *qp); [all …]
|
/drivers/media/platform/qcom/camss/ |
D | camss-csid-gen2.c | 38 #define CSID_CSI2_RDIN_IRQ_STATUS(rdi) ((IS_LITE ? 0x30 : 0x40) \ argument 39 + 0x10 * (rdi)) 40 #define CSID_CSI2_RDIN_IRQ_MASK(rdi) ((IS_LITE ? 0x34 : 0x44) \ argument 41 + 0x10 * (rdi)) 42 #define CSID_CSI2_RDIN_IRQ_CLEAR(rdi) ((IS_LITE ? 0x38 : 0x48) \ argument 43 + 0x10 * (rdi)) 44 #define CSID_CSI2_RDIN_IRQ_SET(rdi) ((IS_LITE ? 0x3C : 0x4C) \ argument 45 + 0x10 * (rdi)) 76 #define CSID_RDI_CFG0(rdi) ((IS_LITE ? 0x200 : 0x300) \ argument 77 + 0x100 * (rdi)) [all …]
|
/drivers/infiniband/hw/hfi1/ |
D | verbs.c | 334 struct rvt_dev_info *rdi = &ppd->dd->verbs_dev.rdi; in hfi1_kdeth_eager_rcv() local 360 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); in hfi1_kdeth_eager_rcv() 388 struct rvt_dev_info *rdi = &ppd->dd->verbs_dev.rdi; in hfi1_kdeth_expected_rcv() local 414 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); in hfi1_kdeth_expected_rcv() 460 struct rvt_dev_info *rdi = &ppd->dd->verbs_dev.rdi; in hfi1_handle_packet() local 506 packet->qp = rvt_lookup_qpn(rdi, &ibp->rvp, qp_num); in hfi1_handle_packet() 1290 struct rvt_dev_info *rdi = &dd->verbs_dev.rdi; in hfi1_fill_device_attr() local 1293 memset(&rdi->dparms.props, 0, sizeof(rdi->dparms.props)); in hfi1_fill_device_attr() 1295 rdi->dparms.props.fw_ver = ((u64)(dc8051_ver_maj(ver)) << 32) | in hfi1_fill_device_attr() 1299 rdi->dparms.props.device_cap_flags = IB_DEVICE_BAD_PKEY_CNTR | in hfi1_fill_device_attr() [all …]
|
D | qp.h | 93 void *qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp); 94 void qp_priv_free(struct rvt_dev_info *rdi, struct rvt_qp *qp); 95 unsigned free_all_qps(struct rvt_dev_info *rdi); 97 int get_pmtu_from_attr(struct rvt_dev_info *rdi, struct rvt_qp *qp, 103 u32 mtu_from_qp(struct rvt_dev_info *rdi, struct rvt_qp *qp, u32 pmtu);
|
D | qp.c | 639 rvt_max_atomic(&to_idev(qp->ibqp.device)->rdi), in qp_iter_print() 668 void *qp_priv_alloc(struct rvt_dev_info *rdi, struct rvt_qp *qp) in qp_priv_alloc() argument 672 priv = kzalloc_node(sizeof(*priv), GFP_KERNEL, rdi->dparms.node); in qp_priv_alloc() 679 rdi->dparms.node); in qp_priv_alloc() 698 void qp_priv_free(struct rvt_dev_info *rdi, struct rvt_qp *qp) in qp_priv_free() argument 702 hfi1_qp_priv_tid_free(rdi, qp); in qp_priv_free() 707 unsigned free_all_qps(struct rvt_dev_info *rdi) in free_all_qps() argument 709 struct hfi1_ibdev *verbs_dev = container_of(rdi, in free_all_qps() 711 rdi); in free_all_qps() 799 u32 mtu_from_qp(struct rvt_dev_info *rdi, struct rvt_qp *qp, u32 pmtu) in mtu_from_qp() argument [all …]
|
D | sysfs.c | 410 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in hw_rev_show() 420 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in board_id_show() 434 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in boardversion_show() 446 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in nctxts_show() 465 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in nfreectxts_show() 477 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in serial_show() 490 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in chip_reset_store() 519 rdma_device_to_drv_device(device, struct hfi1_ibdev, rdi.ibdev); in tempsense_show() 655 struct ib_device *dev = &dd->verbs_dev.rdi.ibdev; in hfi1_verbs_register_sysfs()
|
D | verbs.h | 236 struct rvt_dev_info rdi; /* Must be first */ member 267 struct rvt_dev_info *rdi; in to_idev() local 269 rdi = container_of(ibdev, struct rvt_dev_info, ibdev); in to_idev() 270 return container_of(rdi, struct hfi1_ibdev, rdi); in to_idev() 286 void hfi1_cap_mask_chg(struct rvt_dev_info *rdi, u32 port_num);
|