/net/smc/ |
D | smc_wr.c | 64 static inline void smc_wr_tx_process_cqe(struct ib_wc *wc) in smc_wr_tx_process_cqe() argument 71 link = wc->qp->qp_context; in smc_wr_tx_process_cqe() 73 if (wc->opcode == IB_WC_REG_MR) { in smc_wr_tx_process_cqe() 74 if (wc->status) in smc_wr_tx_process_cqe() 82 pnd_snd_idx = smc_wr_tx_find_pending_index(link, wc->wr_id); in smc_wr_tx_process_cqe() 85 link->wr_tx_pends[pnd_snd_idx].wc_status = wc->status; in smc_wr_tx_process_cqe() 94 if (wc->status) { in smc_wr_tx_process_cqe() 107 pnd_snd.handler(&pnd_snd.priv, link, wc->status); in smc_wr_tx_process_cqe() 114 struct ib_wc wc[SMC_WR_MAX_POLL_CQE]; in smc_wr_tx_tasklet_fn() local 121 memset(&wc, 0, sizeof(wc)); in smc_wr_tx_tasklet_fn() [all …]
|
D | smc_cdc.c | 351 static void smc_cdc_rx_handler(struct ib_wc *wc, void *buf) in smc_cdc_rx_handler() argument 353 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_cdc_rx_handler() 359 if (wc->byte_len < offsetof(struct smc_cdc_msg, reserved)) in smc_cdc_rx_handler()
|
D | smc_llc.c | 556 static void smc_llc_rx_handler(struct ib_wc *wc, void *buf) in smc_llc_rx_handler() argument 558 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_llc_rx_handler() 561 if (wc->byte_len < sizeof(*llc)) in smc_llc_rx_handler()
|
/net/sunrpc/xprtrdma/ |
D | frwr_ops.c | 401 static void frwr_wc_fastreg(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_fastreg() argument 403 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_fastreg() 408 trace_xprtrdma_wc_fastreg(wc, frwr); in frwr_wc_fastreg() 469 static void __frwr_release_mr(struct ib_wc *wc, struct rpcrdma_mr *mr) in __frwr_release_mr() argument 471 if (wc->status != IB_WC_SUCCESS) in __frwr_release_mr() 483 static void frwr_wc_localinv(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_localinv() argument 485 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_localinv() 491 trace_xprtrdma_wc_li(wc, frwr); in frwr_wc_localinv() 492 __frwr_release_mr(wc, mr); in frwr_wc_localinv() 502 static void frwr_wc_localinv_wake(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_localinv_wake() argument [all …]
|
D | svc_rdma_rw.c | 19 static void svc_rdma_write_done(struct ib_cq *cq, struct ib_wc *wc); 20 static void svc_rdma_wc_read_done(struct ib_cq *cq, struct ib_wc *wc); 202 static void svc_rdma_write_done(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_write_done() argument 204 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_write_done() 211 trace_svcrdma_wc_write(wc); in svc_rdma_write_done() 216 if (unlikely(wc->status != IB_WC_SUCCESS)) in svc_rdma_write_done() 260 static void svc_rdma_wc_read_done(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_wc_read_done() argument 262 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_read_done() 269 trace_svcrdma_wc_read(wc); in svc_rdma_wc_read_done() 274 if (unlikely(wc->status != IB_WC_SUCCESS)) { in svc_rdma_wc_read_done()
|
D | svc_rdma_recvfrom.c | 111 static void svc_rdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc); 308 static void svc_rdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_wc_receive() argument 311 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_receive() 314 trace_svcrdma_wc_receive(wc); in svc_rdma_wc_receive() 319 if (wc->status != IB_WC_SUCCESS) in svc_rdma_wc_receive() 326 ctxt->rc_byte_len = wc->byte_len; in svc_rdma_wc_receive() 329 wc->byte_len, DMA_FROM_DEVICE); in svc_rdma_wc_receive()
|
D | verbs.c | 132 rpcrdma_wc_send(struct ib_cq *cq, struct ib_wc *wc) in rpcrdma_wc_send() argument 134 struct ib_cqe *cqe = wc->wr_cqe; in rpcrdma_wc_send() 139 trace_xprtrdma_wc_send(sc, wc); in rpcrdma_wc_send() 150 rpcrdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc) in rpcrdma_wc_receive() argument 152 struct ib_cqe *cqe = wc->wr_cqe; in rpcrdma_wc_receive() 158 trace_xprtrdma_wc_receive(wc); in rpcrdma_wc_receive() 160 if (wc->status != IB_WC_SUCCESS) in rpcrdma_wc_receive() 164 rpcrdma_set_xdrlen(&rep->rr_hdrbuf, wc->byte_len); in rpcrdma_wc_receive() 165 rep->rr_wc_flags = wc->wc_flags; in rpcrdma_wc_receive() 166 rep->rr_inv_rkey = wc->ex.invalidate_rkey; in rpcrdma_wc_receive() [all …]
|
D | svc_rdma_sendto.c | 117 static void svc_rdma_wc_send(struct ib_cq *cq, struct ib_wc *wc); 258 static void svc_rdma_wc_send(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_wc_send() argument 261 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_send() 264 trace_svcrdma_wc_send(wc); in svc_rdma_wc_send() 272 if (unlikely(wc->status != IB_WC_SUCCESS)) { in svc_rdma_wc_send()
|
/net/rds/ |
D | ib_cm.c | 296 struct ib_wc *wc; in poll_scq() local 300 wc = wcs + i; in poll_scq() 302 (unsigned long long)wc->wr_id, wc->status, in poll_scq() 303 wc->byte_len, be32_to_cpu(wc->ex.imm_data)); in poll_scq() 305 if (wc->wr_id <= ic->i_send_ring.w_nr || in poll_scq() 306 wc->wr_id == RDS_IB_ACK_WR_ID) in poll_scq() 307 rds_ib_send_cqe_handler(ic, wc); in poll_scq() 309 rds_ib_mr_cqe_handler(ic, wc); in poll_scq() 341 struct ib_wc *wc; in poll_rcq() local 345 wc = wcs + i; in poll_rcq() [all …]
|
D | ib_send.c | 241 void rds_ib_send_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc) in rds_ib_send_cqe_handler() argument 253 (unsigned long long)wc->wr_id, wc->status, in rds_ib_send_cqe_handler() 254 ib_wc_status_msg(wc->status), wc->byte_len, in rds_ib_send_cqe_handler() 255 be32_to_cpu(wc->ex.imm_data)); in rds_ib_send_cqe_handler() 258 if (wc->wr_id == RDS_IB_ACK_WR_ID) { in rds_ib_send_cqe_handler() 267 completed = rds_ib_ring_completed(&ic->i_send_ring, wc->wr_id, oldest); in rds_ib_send_cqe_handler() 274 rm = rds_ib_send_unmap_op(ic, send, wc->status); in rds_ib_send_cqe_handler() 302 if (wc->status != IB_WC_SUCCESS && rds_conn_up(conn)) { in rds_ib_send_cqe_handler() 305 conn->c_tos, wc->status, in rds_ib_send_cqe_handler() 306 ib_wc_status_msg(wc->status)); in rds_ib_send_cqe_handler()
|
D | ib_frmr.c | 322 void rds_ib_mr_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc) in rds_ib_mr_cqe_handler() argument 324 struct rds_ib_mr *ibmr = (void *)(unsigned long)wc->wr_id; in rds_ib_mr_cqe_handler() 327 if (wc->status != IB_WC_SUCCESS) { in rds_ib_mr_cqe_handler() 334 wc->status, in rds_ib_mr_cqe_handler() 335 ib_wc_status_msg(wc->status), in rds_ib_mr_cqe_handler() 336 wc->vendor_err); in rds_ib_mr_cqe_handler()
|
D | ib_recv.c | 972 struct ib_wc *wc, in rds_ib_recv_cqe_handler() argument 979 (unsigned long long)wc->wr_id, wc->status, in rds_ib_recv_cqe_handler() 980 ib_wc_status_msg(wc->status), wc->byte_len, in rds_ib_recv_cqe_handler() 981 be32_to_cpu(wc->ex.imm_data)); in rds_ib_recv_cqe_handler() 992 if (wc->status == IB_WC_SUCCESS) { in rds_ib_recv_cqe_handler() 993 rds_ib_process_recv(conn, recv, wc->byte_len, state); in rds_ib_recv_cqe_handler() 999 conn->c_tos, wc->status, in rds_ib_recv_cqe_handler() 1000 ib_wc_status_msg(wc->status)); in rds_ib_recv_cqe_handler()
|
D | ib.h | 395 void rds_ib_mr_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc); 406 void rds_ib_recv_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc, 433 void rds_ib_send_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc);
|
/net/9p/ |
D | trans_rdma.c | 290 recv_done(struct ib_cq *cq, struct ib_wc *wc) in recv_done() argument 295 container_of(wc->wr_cqe, struct p9_rdma_context, cqe); in recv_done() 304 if (wc->status != IB_WC_SUCCESS) in recv_done() 307 c->rc.size = wc->byte_len; in recv_done() 334 req, err, wc->status); in recv_done() 341 send_done(struct ib_cq *cq, struct ib_wc *wc) in send_done() argument 346 container_of(wc->wr_cqe, struct p9_rdma_context, cqe); in send_done()
|
/net/netfilter/ |
D | nfnetlink_osf.c | 81 if (f->wss.wc >= OSF_WSS_MAX) in nf_osf_match_one() 95 check_WSS = f->wss.wc; in nf_osf_match_one()
|