/drivers/atm/ |
D | nicstar.c | 114 #define scq_virt_to_bus(scq, p) \ argument 115 (scq->dma + ((unsigned long)(p) - (unsigned long)(scq)->org)) 125 static void free_scq(ns_dev *card, scq_info * scq, struct atm_vcc *vcc); 132 static int push_scqe(ns_dev * card, vc_map * vc, scq_info * scq, ns_scqe * tbd, 135 static void drain_scq(ns_dev * card, scq_info * scq, int pos); 244 free_scq(card, card->scd2vc[j]->scq, card->scd2vc[j]->tx_vcc); in nicstar_remove_one() 860 scq_info *scq; in get_scq() local 866 scq = kmalloc(sizeof(*scq), GFP_KERNEL); in get_scq() 867 if (!scq) in get_scq() 869 scq->org = dma_alloc_coherent(&card->pcidev->dev, in get_scq() [all …]
|
D | idt77252.c | 639 struct scq_info *scq; in alloc_scq() local 641 scq = kzalloc(sizeof(struct scq_info), GFP_KERNEL); in alloc_scq() 642 if (!scq) in alloc_scq() 644 scq->base = dma_zalloc_coherent(&card->pcidev->dev, SCQ_SIZE, in alloc_scq() 645 &scq->paddr, GFP_KERNEL); in alloc_scq() 646 if (scq->base == NULL) { in alloc_scq() 647 kfree(scq); in alloc_scq() 651 scq->next = scq->base; in alloc_scq() 652 scq->last = scq->base + (SCQ_ENTRIES - 1); in alloc_scq() 653 atomic_set(&scq->used, 0); in alloc_scq() [all …]
|
D | nicstar.h | 702 scq_info *scq; /* To keep track of the SCQ */ member
|
D | idt77252.h | 214 struct scq_info *scq; /* To keep track of the SCQ */ member
|
/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_qp.c | 62 static void pvrdma_lock_cqs(struct pvrdma_cq *scq, struct pvrdma_cq *rcq, in pvrdma_lock_cqs() argument 65 __acquires(scq->cq_lock) __acquires(rcq->cq_lock) in pvrdma_lock_cqs() 67 if (scq == rcq) { in pvrdma_lock_cqs() 68 spin_lock_irqsave(&scq->cq_lock, *scq_flags); in pvrdma_lock_cqs() 70 } else if (scq->cq_handle < rcq->cq_handle) { in pvrdma_lock_cqs() 71 spin_lock_irqsave(&scq->cq_lock, *scq_flags); in pvrdma_lock_cqs() 76 spin_lock_irqsave_nested(&scq->cq_lock, *scq_flags, in pvrdma_lock_cqs() 81 static void pvrdma_unlock_cqs(struct pvrdma_cq *scq, struct pvrdma_cq *rcq, in pvrdma_unlock_cqs() argument 84 __releases(scq->cq_lock) __releases(rcq->cq_lock) in pvrdma_unlock_cqs() 86 if (scq == rcq) { in pvrdma_unlock_cqs() [all …]
|
/drivers/infiniband/hw/bnxt_re/ |
D | qplib_fp.c | 66 struct bnxt_qplib_cq *scq, *rcq; in __bnxt_qplib_add_flush_qp() local 68 scq = qp->scq; in __bnxt_qplib_add_flush_qp() 72 dev_dbg(&scq->hwq.pdev->dev, in __bnxt_qplib_add_flush_qp() 76 list_add_tail(&qp->sq_flush, &scq->sqf_head); in __bnxt_qplib_add_flush_qp() 92 __acquires(&qp->scq->hwq.lock) __acquires(&qp->rcq->hwq.lock) in bnxt_qplib_acquire_cq_locks() 94 spin_lock_irqsave(&qp->scq->hwq.lock, *flags); in bnxt_qplib_acquire_cq_locks() 95 if (qp->scq == qp->rcq) in bnxt_qplib_acquire_cq_locks() 103 __releases(&qp->scq->hwq.lock) __releases(&qp->rcq->hwq.lock) in bnxt_qplib_release_cq_locks() 105 if (qp->scq == qp->rcq) in bnxt_qplib_release_cq_locks() 109 spin_unlock_irqrestore(&qp->scq->hwq.lock, *flags); in bnxt_qplib_release_cq_locks() [all …]
|
D | qplib_fp.h | 272 struct bnxt_qplib_cq *scq; member
|
D | ib_verbs.c | 991 qp->qplib_qp.scq = qp1_qp->scq; in bnxt_re_create_shadow_qp() 1072 qp->qplib_qp.scq = &cq->qplib_cq; in bnxt_re_create_qp()
|
/drivers/infiniband/sw/rxe/ |
D | rxe_qp.c | 340 struct rxe_cq *scq = to_rcq(init->send_cq); in rxe_qp_from_init() local 346 rxe_add_ref(scq); in rxe_qp_from_init() 352 qp->scq = scq; in rxe_qp_from_init() 375 rxe_drop_ref(scq); in rxe_qp_from_init() 843 if (qp->scq) in rxe_qp_do_cleanup() 844 rxe_drop_ref(qp->scq); in rxe_qp_do_cleanup()
|
D | rxe_verbs.h | 246 struct rxe_cq *scq; member
|
D | rxe_comp.c | 450 rxe_cq_post(qp->scq, &cqe, 0); in do_complete()
|
/drivers/infiniband/hw/cxgb4/ |
D | ev.c | 138 cqid = qhp->attr.scq; in c4iw_ev_dispatch()
|
D | qp.c | 196 struct t4_cq *rcq, struct t4_cq *scq, in create_qp() argument 318 FW_RI_RES_WR_IQID_V(scq->cqid)); in create_qp() 1540 wqe->u.init.scqid = cpu_to_be32(qhp->attr.scq); in rdma_init() 1912 qhp->attr.scq = ((struct c4iw_cq *) attrs->send_cq)->cq.cqid; in c4iw_create_qp()
|
D | iw_cxgb4.h | 448 u32 scq; member
|
/drivers/infiniband/hw/cxgb3/ |
D | iwch_provider.h | 132 u32 scq; member
|
D | iwch_qp.c | 775 schp = get_chp(qhp->rhp, qhp->attr.scq); in flush_qp() 822 init_attr.scqid = qhp->attr.scq; in rdma_init()
|
D | iwch_provider.c | 897 qhp->attr.scq = ((struct iwch_cq *) attrs->send_cq)->cq.cqid; in iwch_create_qp()
|
/drivers/infiniband/core/ |
D | uverbs_cmd.c | 1367 struct ib_cq *scq = NULL, *rcq = NULL; in create_qp() local 1459 scq = uobj_get_obj_read(cq, cmd->send_cq_handle, in create_qp() 1462 rcq = rcq ?: scq; in create_qp() 1464 if (!pd || (!scq && has_sq)) { in create_qp() 1474 attr.send_cq = scq; in create_qp() 1591 if (scq) in create_qp() 1592 uobj_put_obj_read(scq); in create_qp() 1593 if (rcq && rcq != scq) in create_qp() 1611 if (scq) in create_qp() 1612 uobj_put_obj_read(scq); in create_qp() [all …]
|
D | verbs.c | 1450 struct ib_cq *scq, *rcq; in ib_destroy_qp() local 1465 scq = qp->send_cq; in ib_destroy_qp() 1480 if (scq) in ib_destroy_qp() 1481 atomic_dec(&scq->usecnt); in ib_destroy_qp()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | resource_tracker.c | 112 struct res_cq *scq; member 2963 struct res_cq *scq; in mlx4_RST2INIT_QP_wrapper() local 3001 err = get_res(dev, slave, scqn, RES_CQ, &scq); in mlx4_RST2INIT_QP_wrapper() 3005 scq = rcq; in mlx4_RST2INIT_QP_wrapper() 3022 atomic_inc(&scq->ref_count); in mlx4_RST2INIT_QP_wrapper() 3023 qp->scq = scq; in mlx4_RST2INIT_QP_wrapper() 4001 atomic_dec(&qp->scq->ref_count); in mlx4_2RST_QP_wrapper() 4714 atomic_dec(&qp->scq->ref_count); in rem_slave_qps()
|