/drivers/infiniband/sw/rxe/ |
D | rxe_cq.c | 39 int cqe, int comp_vector, struct ib_udata *udata) in rxe_cq_chk_attr() 76 int rxe_cq_from_init(struct rxe_dev *rxe, struct rxe_cq *cq, int cqe, in rxe_cq_from_init() 107 int rxe_cq_resize_queue(struct rxe_cq *cq, int cqe, struct ib_udata *udata) in rxe_cq_resize_queue() 121 int rxe_cq_post(struct rxe_cq *cq, struct rxe_cqe *cqe, int solicited) in rxe_cq_post()
|
D | rxe_comp.c | 386 struct rxe_cqe *cqe) in make_send_cqe() 417 struct rxe_cqe cqe; in do_complete() local
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_rx.c | 77 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(&cq->wq, ci); in mlx5e_cqes_update_owner() local 85 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(&cq->wq, ci); in mlx5e_cqes_update_owner() local 483 static void mlx5e_lro_update_hdr(struct sk_buff *skb, struct mlx5_cqe64 *cqe, in mlx5e_lro_update_hdr() 539 static inline void mlx5e_skb_set_hash(struct mlx5_cqe64 *cqe, in mlx5e_skb_set_hash() 557 struct mlx5_cqe64 *cqe, in mlx5e_handle_csum() 592 static inline void mlx5e_build_rx_skb(struct mlx5_cqe64 *cqe, in mlx5e_build_rx_skb() 632 struct mlx5_cqe64 *cqe, in mlx5e_complete_rx_cqe() 743 struct sk_buff *skb_from_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe, in skb_from_cqe() 788 void mlx5e_handle_rx_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe) in mlx5e_handle_rx_cqe() 813 void mlx5e_handle_rx_cqe_rep(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe) in mlx5e_handle_rx_cqe_rep() [all …]
|
D | en_txrx.c | 39 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci); in mlx5e_get_cqe() local 56 struct mlx5_cqe64 *cqe; in mlx5e_poll_ico_cq() local 125 struct mlx5_cqe64 *cqe; in mlx5e_poll_xdp_tx_cq() local
|
/drivers/infiniband/hw/mlx4/ |
D | cq.c | 80 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local 132 static void mlx4_ib_free_cq_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq_buf *buf, int cqe) in mlx4_ib_free_cq_buf() 139 u64 buf_addr, int cqe) in mlx4_ib_get_cq_umem() 352 struct mlx4_cqe *cqe, *new_cqe; in mlx4_ib_cq_resize_copy_cqes() local 497 static void dump_cqe(void *cqe) in dump_cqe() 507 static void mlx4_ib_handle_error_cqe(struct mlx4_err_cqe *cqe, in mlx4_ib_handle_error_cqe() 583 unsigned tail, struct mlx4_cqe *cqe, int is_eth) in use_tunnel_data() 661 struct mlx4_cqe *cqe; in mlx4_ib_poll_one() local 916 struct mlx4_cqe *cqe, *dest; in __mlx4_ib_cq_clean() local
|
/drivers/infiniband/hw/cxgb4/ |
D | cq.c | 181 struct t4_cqe cqe; in insert_recv_cqe() local 214 struct t4_cqe cqe; in insert_sq_cqe() local 413 static int cqe_completes_wr(struct t4_cqe *cqe, struct t4_wq *wq) in cqe_completes_wr() 431 struct t4_cqe *cqe; in c4iw_count_rcqes() local 464 static int poll_cq(struct t4_wq *wq, struct t4_cq *cq, struct t4_cqe *cqe, in poll_cq() 1017 int c4iw_resize_cq(struct ib_cq *cq, int cqe, struct ib_udata *udata) in c4iw_resize_cq()
|
/drivers/scsi/bnx2i/ |
D | bnx2i_hwi.c | 1355 struct cqe *cqe) in bnx2i_process_scsi_cmd_resp() 1452 struct cqe *cqe) in bnx2i_process_login_resp() 1520 struct cqe *cqe) in bnx2i_process_text_resp() 1581 struct cqe *cqe) in bnx2i_process_tmf_resp() 1620 struct cqe *cqe) in bnx2i_process_logout_resp() 1666 struct cqe *cqe) in bnx2i_process_nopin_local_cmpl() 1707 struct cqe *cqe) in bnx2i_process_nopin_mesg() 1759 struct cqe *cqe) in bnx2i_process_async_mesg() 1809 struct cqe *cqe) in bnx2i_process_reject_mesg() 1846 struct cqe *cqe) in bnx2i_process_cmd_cleanup_resp() [all …]
|
/drivers/infiniband/hw/mthca/ |
D | mthca_cq.c | 174 static inline struct mthca_cqe *cqe_sw(struct mthca_cqe *cqe) in cqe_sw() 184 static inline void set_cqe_hw(struct mthca_cqe *cqe) in set_cqe_hw() 191 __be32 *cqe = cqe_ptr; in dump_cqe() local 269 static inline int is_recv_cqe(struct mthca_cqe *cqe) in is_recv_cqe() 281 struct mthca_cqe *cqe; in mthca_cq_clean() local 371 void mthca_free_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int cqe) in mthca_free_cq_buf() 379 struct mthca_err_cqe *cqe, in handle_error_cqe() 490 struct mthca_cqe *cqe; in mthca_poll_one() local
|
/drivers/infiniband/hw/cxgb3/ |
D | cxio_hal.c | 75 struct t3_cqe *cqe; in cxio_hal_cq_op() local 352 struct t3_cqe cqe; in insert_recv_cqe() local 389 struct t3_cqe cqe; in insert_sq_cqe() local 430 struct t3_cqe *cqe, *swcqe; in cxio_flush_hw_cq() local 446 static int cqe_completes_wr(struct t3_cqe *cqe, struct t3_wq *wq) in cqe_completes_wr() 466 struct t3_cqe *cqe; in cxio_count_scqes() local 484 struct t3_cqe *cqe; in cxio_count_rcqes() local 1147 int cxio_poll_cq(struct t3_wq *wq, struct t3_cq *cq, struct t3_cqe *cqe, in cxio_poll_cq()
|
D | iwch_cq.c | 48 struct t3_cqe cqe, *rd_cqe; in iwch_poll_cq_one() local
|
/drivers/infiniband/hw/mlx5/ |
D | cq.c | 84 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local 120 static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_good_req() 167 static void handle_responder(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_responder() 270 static void dump_cqe(struct mlx5_ib_dev *dev, struct mlx5_err_cqe *cqe) in dump_cqe() 283 struct mlx5_err_cqe *cqe, in mlx5_handle_error_cqe() 407 static void get_sig_err_item(struct mlx5_sig_err_cqe *cqe, in get_sig_err_item() 531 void *cqe; in mlx5_poll_one() local 827 void *cqe; in init_cq_buf() local 1041 void *cqe, *dest; in __mlx5_ib_cq_clean() local
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.c | 1144 struct ocrdma_cqe *cqe = NULL; in ocrdma_flush_cq() local 1707 struct ocrdma_cqe *cqe; in ocrdma_discard_cqes() local 2560 struct ocrdma_cqe *cqe) in ocrdma_set_cqe_status_flushed() 2591 static bool ocrdma_update_err_cqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, in ocrdma_update_err_cqe() 2613 static int ocrdma_update_err_rcqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, in ocrdma_update_err_rcqe() 2623 static int ocrdma_update_err_scqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe, in ocrdma_update_err_scqe() 2634 struct ocrdma_cqe *cqe, struct ib_wc *ibwc, in ocrdma_poll_err_scqe() 2678 struct ocrdma_cqe *cqe, in ocrdma_poll_success_scqe() 2703 static bool ocrdma_poll_scqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe, in ocrdma_poll_scqe() 2720 struct ocrdma_cqe *cqe) in ocrdma_update_ud_rcqe() [all …]
|
D | ocrdma_hw.c | 122 struct ocrdma_mcqe *cqe = (struct ocrdma_mcqe *) in ocrdma_get_mcqe() local 673 struct ocrdma_ae_mcqe *cqe) in ocrdma_dispatch_ibevent() 799 struct ocrdma_ae_mcqe *cqe) in ocrdma_process_grp5_aync() 825 struct ocrdma_ae_mcqe *cqe) in ocrdma_process_link_state() 843 struct ocrdma_ae_mcqe *cqe = ae_cqe; in ocrdma_process_acqe() local 862 static void ocrdma_process_mcqe(struct ocrdma_dev *dev, struct ocrdma_mcqe *cqe) in ocrdma_process_mcqe() 880 struct ocrdma_mcqe *cqe; in ocrdma_mq_cq_handler() local
|
/drivers/nvme/host/ |
D | fabrics.c | 164 struct nvme_completion cqe; in nvmf_reg_read32() local 210 struct nvme_completion cqe; in nvmf_reg_read64() local 371 struct nvme_completion cqe; in nvmf_connect_admin_queue() local 444 struct nvme_completion cqe; in nvmf_connect_io_queue() local
|
D | core.c | 347 struct nvme_completion *cqe, void *buffer, unsigned bufflen, in __nvme_submit_sync_cmd() 388 struct nvme_completion cqe; in __nvme_submit_user_cmd() local 605 struct nvme_completion cqe; in nvme_get_features() local 624 struct nvme_completion cqe; in nvme_set_features() local 1892 struct nvme_completion *cqe) in nvme_complete_async_event()
|
/drivers/infiniband/ulp/iser/ |
D | iscsi_iser.h | 261 struct ib_cqe cqe; member 293 struct ib_cqe cqe; member 313 struct ib_cqe cqe; member 703 iser_rx(struct ib_cqe *cqe) in iser_rx() 709 iser_tx(struct ib_cqe *cqe) in iser_tx() 715 iser_login(struct ib_cqe *cqe) in iser_login()
|
D | iser_memory.c | 373 struct ib_cqe *cqe) in iser_inv_rkey() 391 struct ib_cqe *cqe = &iser_task->iser_conn->ib_conn.reg_cqe; in iser_reg_sig_mr() local 443 struct ib_cqe *cqe = &iser_task->iser_conn->ib_conn.reg_cqe; in iser_fast_reg_mr() local
|
/drivers/infiniband/hw/hns/ |
D | hns_roce_cq.c | 213 struct ib_umem **umem, u64 buf_addr, int cqe) in hns_roce_ib_get_cq_umem() 274 struct hns_roce_cq_buf *buf, int cqe) in hns_roce_ib_free_cq_buf()
|
/drivers/net/ethernet/ibm/ehea/ |
D | ehea_main.c | 543 static inline int ehea_check_cqe(struct ehea_cqe *cqe, int *rq_num) in ehea_check_cqe() 555 struct sk_buff *skb, struct ehea_cqe *cqe, in ehea_fill_skb() 576 struct ehea_cqe *cqe) in get_skb_by_index() 629 struct ehea_cqe *cqe, int *processed_rq2, in ehea_treat_poll_error() 670 struct ehea_cqe *cqe; in ehea_proc_rwqes() local 817 struct ehea_cqe *cqe; in ehea_proc_cqes() local 894 struct ehea_cqe *cqe; in ehea_poll() local
|
/drivers/net/ethernet/qlogic/qed/ |
D | qed_spq.c | 371 struct eth_slow_path_rx_cqe *cqe, in qed_cqe_completion() 385 struct eth_slow_path_rx_cqe *cqe) in qed_eth_cqe_completion()
|
/drivers/infiniband/core/ |
D | rw.c | 496 u8 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr) in rdma_rw_ctx_wrs() 569 struct ib_cqe *cqe, struct ib_send_wr *chain_wr) in rdma_rw_ctx_post()
|
/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_cmn.c | 354 struct eth_end_agg_rx_cqe *cqe) in bnx2x_update_sge_prod() 409 const struct eth_fast_path_rx_cqe *cqe, in bnx2x_get_rxhash() 430 struct eth_fast_path_rx_cqe *cqe) in bnx2x_tpa_start() 592 struct eth_end_agg_rx_cqe *cqe, in bnx2x_fill_frag_skb() 758 struct eth_end_agg_rx_cqe *cqe, in bnx2x_tpa_stop() 859 void bnx2x_csum_validate(struct sk_buff *skb, union eth_rx_cqe *cqe, in bnx2x_csum_validate() 888 union eth_rx_cqe *cqe; in bnx2x_rx_int() local
|
/drivers/net/ethernet/qlogic/qede/ |
D | qede_main.c | 1015 struct eth_fast_path_rx_tpa_start_cqe *cqe) in qede_set_gro_params() 1071 struct eth_fast_path_rx_tpa_start_cqe *cqe) in qede_tpa_start() 1224 struct eth_fast_path_rx_tpa_cont_cqe *cqe) in qede_tpa_cont() 1239 struct eth_fast_path_rx_tpa_end_cqe *cqe) in qede_tpa_end() 1369 static bool qede_pkt_is_ip_fragmented(struct eth_fast_path_rx_reg_cqe *cqe, in qede_pkt_is_ip_fragmented() 1408 union eth_rx_cqe *cqe; in qede_rx_int() local
|
/drivers/infiniband/ulp/isert/ |
D | ib_isert.h | 91 static inline struct iser_rx_desc *cqe_to_rx_desc(struct ib_cqe *cqe) in cqe_to_rx_desc() 107 static inline struct iser_tx_desc *cqe_to_tx_desc(struct ib_cqe *cqe) in cqe_to_tx_desc()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | en_clock.c | 50 u64 mlx4_en_get_cqe_ts(struct mlx4_cqe *cqe) in mlx4_en_get_cqe_ts()
|