/net/sunrpc/xprtrdma/ |
D | frwr_ops.c | 360 static void frwr_wc_fastreg(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_fastreg() argument 362 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_fastreg() 367 trace_xprtrdma_wc_fastreg(wc, frwr); in frwr_wc_fastreg() 370 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_fastreg() 430 static void __frwr_release_mr(struct ib_wc *wc, struct rpcrdma_mr *mr) in __frwr_release_mr() argument 432 if (wc->status != IB_WC_SUCCESS) in __frwr_release_mr() 444 static void frwr_wc_localinv(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_localinv() argument 446 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_localinv() 452 trace_xprtrdma_wc_li(wc, frwr); in frwr_wc_localinv() 453 __frwr_release_mr(wc, mr); in frwr_wc_localinv() [all …]
|
D | verbs.c | 150 void rpcrdma_flush_disconnect(struct rpcrdma_xprt *r_xprt, struct ib_wc *wc) in rpcrdma_flush_disconnect() argument 152 if (wc->status != IB_WC_SUCCESS) in rpcrdma_flush_disconnect() 162 static void rpcrdma_wc_send(struct ib_cq *cq, struct ib_wc *wc) in rpcrdma_wc_send() argument 164 struct ib_cqe *cqe = wc->wr_cqe; in rpcrdma_wc_send() 170 trace_xprtrdma_wc_send(sc, wc); in rpcrdma_wc_send() 172 rpcrdma_flush_disconnect(r_xprt, wc); in rpcrdma_wc_send() 181 static void rpcrdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc) in rpcrdma_wc_receive() argument 183 struct ib_cqe *cqe = wc->wr_cqe; in rpcrdma_wc_receive() 189 trace_xprtrdma_wc_receive(wc); in rpcrdma_wc_receive() 191 if (wc->status != IB_WC_SUCCESS) in rpcrdma_wc_receive() [all …]
|
D | svc_rdma_rw.c | 17 static void svc_rdma_write_done(struct ib_cq *cq, struct ib_wc *wc); 18 static void svc_rdma_wc_read_done(struct ib_cq *cq, struct ib_wc *wc); 238 static void svc_rdma_write_done(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_write_done() argument 240 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_write_done() 247 trace_svcrdma_wc_write(wc, &cc->cc_cid); in svc_rdma_write_done() 252 if (unlikely(wc->status != IB_WC_SUCCESS)) in svc_rdma_write_done() 296 static void svc_rdma_wc_read_done(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_wc_read_done() argument 298 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_read_done() 305 trace_svcrdma_wc_read(wc, &cc->cc_cid); in svc_rdma_wc_read_done() 310 if (unlikely(wc->status != IB_WC_SUCCESS)) { in svc_rdma_wc_read_done()
|
D | svc_rdma_recvfrom.c | 111 static void svc_rdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc); 316 static void svc_rdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_wc_receive() argument 319 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_receive() 325 trace_svcrdma_wc_receive(wc, &ctxt->rc_cid); in svc_rdma_wc_receive() 326 if (wc->status != IB_WC_SUCCESS) in svc_rdma_wc_receive() 333 ctxt->rc_byte_len = wc->byte_len; in svc_rdma_wc_receive() 336 wc->byte_len, DMA_FROM_DEVICE); in svc_rdma_wc_receive()
|
D | svc_rdma_sendto.c | 116 static void svc_rdma_wc_send(struct ib_cq *cq, struct ib_wc *wc); 276 static void svc_rdma_wc_send(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_wc_send() argument 279 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_send() 283 trace_svcrdma_wc_send(wc, &ctxt->sc_cid); in svc_rdma_wc_send() 290 if (unlikely(wc->status != IB_WC_SUCCESS)) { in svc_rdma_wc_send()
|
D | xprt_rdma.h | 450 void rpcrdma_flush_disconnect(struct rpcrdma_xprt *r_xprt, struct ib_wc *wc);
|
/net/smc/ |
D | smc_wr.c | 81 static inline void smc_wr_tx_process_cqe(struct ib_wc *wc) in smc_wr_tx_process_cqe() argument 87 link = wc->qp->qp_context; in smc_wr_tx_process_cqe() 89 if (wc->opcode == IB_WC_REG_MR) { in smc_wr_tx_process_cqe() 90 if (wc->status) in smc_wr_tx_process_cqe() 98 pnd_snd_idx = smc_wr_tx_find_pending_index(link, wc->wr_id); in smc_wr_tx_process_cqe() 101 link->wr_tx_pends[pnd_snd_idx].wc_status = wc->status; in smc_wr_tx_process_cqe() 112 if (wc->status) { in smc_wr_tx_process_cqe() 117 pnd_snd.handler(&pnd_snd.priv, link, wc->status); in smc_wr_tx_process_cqe() 124 struct ib_wc wc[SMC_WR_MAX_POLL_CQE]; in smc_wr_tx_tasklet_fn() local 131 memset(&wc, 0, sizeof(wc)); in smc_wr_tx_tasklet_fn() [all …]
|
D | smc_cdc.c | 422 static void smc_cdc_rx_handler(struct ib_wc *wc, void *buf) in smc_cdc_rx_handler() argument 424 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_cdc_rx_handler() 430 if (wc->byte_len < offsetof(struct smc_cdc_msg, reserved)) in smc_cdc_rx_handler()
|
D | smc_llc.c | 1730 static void smc_llc_rx_handler(struct ib_wc *wc, void *buf) in smc_llc_rx_handler() argument 1732 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_llc_rx_handler() 1735 if (wc->byte_len < sizeof(*llc)) in smc_llc_rx_handler()
|
/net/rds/ |
D | ib_cm.c | 297 struct ib_wc *wc; in poll_scq() local 301 wc = wcs + i; in poll_scq() 303 (unsigned long long)wc->wr_id, wc->status, in poll_scq() 304 wc->byte_len, be32_to_cpu(wc->ex.imm_data)); in poll_scq() 306 if (wc->wr_id <= ic->i_send_ring.w_nr || in poll_scq() 307 wc->wr_id == RDS_IB_ACK_WR_ID) in poll_scq() 308 rds_ib_send_cqe_handler(ic, wc); in poll_scq() 310 rds_ib_mr_cqe_handler(ic, wc); in poll_scq() 342 struct ib_wc *wc; in poll_rcq() local 346 wc = wcs + i; in poll_rcq() [all …]
|
D | ib_send.c | 243 void rds_ib_send_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc) in rds_ib_send_cqe_handler() argument 255 (unsigned long long)wc->wr_id, wc->status, in rds_ib_send_cqe_handler() 256 ib_wc_status_msg(wc->status), wc->byte_len, in rds_ib_send_cqe_handler() 257 be32_to_cpu(wc->ex.imm_data)); in rds_ib_send_cqe_handler() 260 if (wc->wr_id == RDS_IB_ACK_WR_ID) { in rds_ib_send_cqe_handler() 269 completed = rds_ib_ring_completed(&ic->i_send_ring, wc->wr_id, oldest); in rds_ib_send_cqe_handler() 276 rm = rds_ib_send_unmap_op(ic, send, wc->status); in rds_ib_send_cqe_handler() 304 if (wc->status != IB_WC_SUCCESS && rds_conn_up(conn)) { in rds_ib_send_cqe_handler() 307 conn->c_tos, wc->status, in rds_ib_send_cqe_handler() 308 ib_wc_status_msg(wc->status), wc->vendor_err); in rds_ib_send_cqe_handler()
|
D | ib_recv.c | 984 struct ib_wc *wc, in rds_ib_recv_cqe_handler() argument 991 (unsigned long long)wc->wr_id, wc->status, in rds_ib_recv_cqe_handler() 992 ib_wc_status_msg(wc->status), wc->byte_len, in rds_ib_recv_cqe_handler() 993 be32_to_cpu(wc->ex.imm_data)); in rds_ib_recv_cqe_handler() 1004 if (wc->status == IB_WC_SUCCESS) { in rds_ib_recv_cqe_handler() 1005 rds_ib_process_recv(conn, recv, wc->byte_len, state); in rds_ib_recv_cqe_handler() 1011 conn->c_tos, wc->status, in rds_ib_recv_cqe_handler() 1012 ib_wc_status_msg(wc->status), in rds_ib_recv_cqe_handler() 1013 wc->vendor_err); in rds_ib_recv_cqe_handler()
|
D | ib_frmr.c | 322 void rds_ib_mr_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc) in rds_ib_mr_cqe_handler() argument 324 struct rds_ib_mr *ibmr = (void *)(unsigned long)wc->wr_id; in rds_ib_mr_cqe_handler() 327 if (wc->status != IB_WC_SUCCESS) { in rds_ib_mr_cqe_handler() 334 wc->status, in rds_ib_mr_cqe_handler() 335 ib_wc_status_msg(wc->status), in rds_ib_mr_cqe_handler() 336 wc->vendor_err); in rds_ib_mr_cqe_handler()
|
D | ib.h | 392 void rds_ib_mr_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc); 403 void rds_ib_recv_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc, 430 void rds_ib_send_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc);
|
/net/9p/ |
D | trans_rdma.c | 291 recv_done(struct ib_cq *cq, struct ib_wc *wc) in recv_done() argument 296 container_of(wc->wr_cqe, struct p9_rdma_context, cqe); in recv_done() 305 if (wc->status != IB_WC_SUCCESS) in recv_done() 308 c->rc.size = wc->byte_len; in recv_done() 335 req, err, wc->status); in recv_done() 342 send_done(struct ib_cq *cq, struct ib_wc *wc) in send_done() argument 347 container_of(wc->wr_cqe, struct p9_rdma_context, cqe); in send_done()
|
/net/netfilter/ |
D | nfnetlink_osf.c | 81 if (f->wss.wc >= OSF_WSS_MAX) in nf_osf_match_one() 95 check_WSS = f->wss.wc; in nf_osf_match_one()
|