Home
last modified time | relevance | path

Searched refs:cq (Results 1 – 11 of 11) sorted by relevance

/include/linux/mlx5/
Dcq.h49 void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe);
57 void (*comp)(struct mlx5_core_cq *cq, struct mlx5_eqe *eqe);
142 static inline void mlx5_cq_set_ci(struct mlx5_core_cq *cq) in mlx5_cq_set_ci() argument
144 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx5_cq_set_ci()
152 static inline void mlx5_cq_arm(struct mlx5_core_cq *cq, u32 cmd, in mlx5_cq_arm() argument
160 sn = cq->arm_sn & 3; in mlx5_cq_arm()
163 *cq->arm_db = cpu_to_be32(sn << 28 | cmd | ci); in mlx5_cq_arm()
171 doorbell[1] = cpu_to_be32(cq->cqn); in mlx5_cq_arm()
176 static inline void mlx5_cq_hold(struct mlx5_core_cq *cq) in mlx5_cq_hold() argument
178 refcount_inc(&cq->refcount); in mlx5_cq_hold()
[all …]
/include/trace/events/
Drdma_core.h51 struct ib_cq *cq
54 TP_ARGS(cq),
61 cq->timestamp = ktime_get();
62 cq->interrupt = true;
64 __entry->cq_id = cq->res.id;
72 struct ib_cq *cq
75 TP_ARGS(cq),
82 cq->timestamp = ktime_get();
83 cq->interrupt = false;
85 __entry->cq_id = cq->res.id;
[all …]
/include/linux/mlx4/
Dcq.h147 static inline void mlx4_cq_arm(struct mlx4_cq *cq, u32 cmd, in mlx4_cq_arm() argument
155 sn = cq->arm_sn & 3; in mlx4_cq_arm()
156 ci = cq->cons_index & 0xffffff; in mlx4_cq_arm()
158 *cq->arm_db = cpu_to_be32(sn << 28 | cmd | ci); in mlx4_cq_arm()
166 doorbell[0] = cpu_to_be32(sn << 28 | cmd | cq->cqn); in mlx4_cq_arm()
172 static inline void mlx4_cq_set_ci(struct mlx4_cq *cq) in mlx4_cq_set_ci() argument
174 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx4_cq_set_ci()
182 int mlx4_cq_modify(struct mlx4_dev *dev, struct mlx4_cq *cq,
184 int mlx4_cq_resize(struct mlx4_dev *dev, struct mlx4_cq *cq,
Ddevice.h832 int cq; member
1147 struct mlx4_uar *uar, u64 db_rec, struct mlx4_cq *cq,
1150 void mlx4_cq_free(struct mlx4_dev *dev, struct mlx4_cq *cq);
/include/rdma/
Drdmavt_cq.h65 bool rvt_cq_enter(struct rvt_cq *cq, struct ib_wc *entry, bool solicited);
Drdmavt_qp.h786 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.recv_cq); in rvt_recv_cq() local
788 if (unlikely(!rvt_cq_enter(cq, wc, solicited))) in rvt_recv_cq()
806 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.send_cq); in rvt_send_cq() local
808 if (unlikely(!rvt_cq_enter(cq, wc, solicited))) in rvt_send_cq()
923 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_tail() local
926 RDMA_READ_UAPI_ATOMIC(cq->queue->tail) : in ib_cq_tail()
939 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_head() local
942 RDMA_READ_UAPI_ATOMIC(cq->queue->head) : in ib_cq_head()
Dib_verbs.h757 struct ib_cq *cq; member
1077 struct ib_cq *cq; member
1377 void (*done)(struct ib_cq *cq, struct ib_wc *wc);
1582 typedef void (*ib_comp_handler)(struct ib_cq *cq, void *cq_context);
1634 struct ib_cq *cq; member
1685 struct ib_cq *cq; member
1705 struct ib_cq *cq; member
2341 int (*poll_cq)(struct ib_cq *cq, int num_entries, struct ib_wc *wc);
2342 int (*peek_cq)(struct ib_cq *cq, int wc_cnt);
2343 int (*req_notify_cq)(struct ib_cq *cq, enum ib_cq_notify_flags flags);
[all …]
/include/net/mana/
Dgdma.h92 } cq; member
327 } cq; member
352 } cq; member
438 int mana_gd_poll_cq(struct gdma_queue *cq, struct gdma_comp *comp, int num_cqe);
440 void mana_gd_ring_cq(struct gdma_queue *cq, u8 arm_bit);
Dhw_channel.h182 struct hwc_cq *cq; member
/include/net/
Dxsk_buff_pool.h65 struct xsk_queue *cq; member
/include/linux/
Dio_uring_types.h126 struct io_uring sq, cq; member