Home
last modified time | relevance | path

Searched refs:wc (Results 1 – 16 of 16) sorted by relevance

/net/sunrpc/xprtrdma/
Dfrwr_ops.c360 static void frwr_wc_fastreg(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_fastreg() argument
362 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_fastreg()
367 trace_xprtrdma_wc_fastreg(wc, frwr); in frwr_wc_fastreg()
370 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_fastreg()
430 static void __frwr_release_mr(struct ib_wc *wc, struct rpcrdma_mr *mr) in __frwr_release_mr() argument
432 if (wc->status != IB_WC_SUCCESS) in __frwr_release_mr()
444 static void frwr_wc_localinv(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_localinv() argument
446 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_localinv()
452 trace_xprtrdma_wc_li(wc, frwr); in frwr_wc_localinv()
453 __frwr_release_mr(wc, mr); in frwr_wc_localinv()
[all …]
Dverbs.c150 void rpcrdma_flush_disconnect(struct rpcrdma_xprt *r_xprt, struct ib_wc *wc) in rpcrdma_flush_disconnect() argument
152 if (wc->status != IB_WC_SUCCESS) in rpcrdma_flush_disconnect()
162 static void rpcrdma_wc_send(struct ib_cq *cq, struct ib_wc *wc) in rpcrdma_wc_send() argument
164 struct ib_cqe *cqe = wc->wr_cqe; in rpcrdma_wc_send()
170 trace_xprtrdma_wc_send(sc, wc); in rpcrdma_wc_send()
172 rpcrdma_flush_disconnect(r_xprt, wc); in rpcrdma_wc_send()
181 static void rpcrdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc) in rpcrdma_wc_receive() argument
183 struct ib_cqe *cqe = wc->wr_cqe; in rpcrdma_wc_receive()
189 trace_xprtrdma_wc_receive(wc); in rpcrdma_wc_receive()
191 if (wc->status != IB_WC_SUCCESS) in rpcrdma_wc_receive()
[all …]
Dsvc_rdma_rw.c17 static void svc_rdma_write_done(struct ib_cq *cq, struct ib_wc *wc);
18 static void svc_rdma_wc_read_done(struct ib_cq *cq, struct ib_wc *wc);
238 static void svc_rdma_write_done(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_write_done() argument
240 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_write_done()
247 trace_svcrdma_wc_write(wc, &cc->cc_cid); in svc_rdma_write_done()
252 if (unlikely(wc->status != IB_WC_SUCCESS)) in svc_rdma_write_done()
296 static void svc_rdma_wc_read_done(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_wc_read_done() argument
298 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_read_done()
305 trace_svcrdma_wc_read(wc, &cc->cc_cid); in svc_rdma_wc_read_done()
310 if (unlikely(wc->status != IB_WC_SUCCESS)) { in svc_rdma_wc_read_done()
Dsvc_rdma_recvfrom.c111 static void svc_rdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc);
316 static void svc_rdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_wc_receive() argument
319 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_receive()
325 trace_svcrdma_wc_receive(wc, &ctxt->rc_cid); in svc_rdma_wc_receive()
326 if (wc->status != IB_WC_SUCCESS) in svc_rdma_wc_receive()
333 ctxt->rc_byte_len = wc->byte_len; in svc_rdma_wc_receive()
336 wc->byte_len, DMA_FROM_DEVICE); in svc_rdma_wc_receive()
Dsvc_rdma_sendto.c116 static void svc_rdma_wc_send(struct ib_cq *cq, struct ib_wc *wc);
276 static void svc_rdma_wc_send(struct ib_cq *cq, struct ib_wc *wc) in svc_rdma_wc_send() argument
279 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_send()
283 trace_svcrdma_wc_send(wc, &ctxt->sc_cid); in svc_rdma_wc_send()
290 if (unlikely(wc->status != IB_WC_SUCCESS)) { in svc_rdma_wc_send()
Dxprt_rdma.h450 void rpcrdma_flush_disconnect(struct rpcrdma_xprt *r_xprt, struct ib_wc *wc);
/net/smc/
Dsmc_wr.c81 static inline void smc_wr_tx_process_cqe(struct ib_wc *wc) in smc_wr_tx_process_cqe() argument
87 link = wc->qp->qp_context; in smc_wr_tx_process_cqe()
89 if (wc->opcode == IB_WC_REG_MR) { in smc_wr_tx_process_cqe()
90 if (wc->status) in smc_wr_tx_process_cqe()
98 pnd_snd_idx = smc_wr_tx_find_pending_index(link, wc->wr_id); in smc_wr_tx_process_cqe()
101 link->wr_tx_pends[pnd_snd_idx].wc_status = wc->status; in smc_wr_tx_process_cqe()
112 if (wc->status) { in smc_wr_tx_process_cqe()
117 pnd_snd.handler(&pnd_snd.priv, link, wc->status); in smc_wr_tx_process_cqe()
124 struct ib_wc wc[SMC_WR_MAX_POLL_CQE]; in smc_wr_tx_tasklet_fn() local
131 memset(&wc, 0, sizeof(wc)); in smc_wr_tx_tasklet_fn()
[all …]
Dsmc_cdc.c422 static void smc_cdc_rx_handler(struct ib_wc *wc, void *buf) in smc_cdc_rx_handler() argument
424 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_cdc_rx_handler()
430 if (wc->byte_len < offsetof(struct smc_cdc_msg, reserved)) in smc_cdc_rx_handler()
Dsmc_llc.c1730 static void smc_llc_rx_handler(struct ib_wc *wc, void *buf) in smc_llc_rx_handler() argument
1732 struct smc_link *link = (struct smc_link *)wc->qp->qp_context; in smc_llc_rx_handler()
1735 if (wc->byte_len < sizeof(*llc)) in smc_llc_rx_handler()
/net/rds/
Dib_cm.c297 struct ib_wc *wc; in poll_scq() local
301 wc = wcs + i; in poll_scq()
303 (unsigned long long)wc->wr_id, wc->status, in poll_scq()
304 wc->byte_len, be32_to_cpu(wc->ex.imm_data)); in poll_scq()
306 if (wc->wr_id <= ic->i_send_ring.w_nr || in poll_scq()
307 wc->wr_id == RDS_IB_ACK_WR_ID) in poll_scq()
308 rds_ib_send_cqe_handler(ic, wc); in poll_scq()
310 rds_ib_mr_cqe_handler(ic, wc); in poll_scq()
342 struct ib_wc *wc; in poll_rcq() local
346 wc = wcs + i; in poll_rcq()
[all …]
Dib_send.c243 void rds_ib_send_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc) in rds_ib_send_cqe_handler() argument
255 (unsigned long long)wc->wr_id, wc->status, in rds_ib_send_cqe_handler()
256 ib_wc_status_msg(wc->status), wc->byte_len, in rds_ib_send_cqe_handler()
257 be32_to_cpu(wc->ex.imm_data)); in rds_ib_send_cqe_handler()
260 if (wc->wr_id == RDS_IB_ACK_WR_ID) { in rds_ib_send_cqe_handler()
269 completed = rds_ib_ring_completed(&ic->i_send_ring, wc->wr_id, oldest); in rds_ib_send_cqe_handler()
276 rm = rds_ib_send_unmap_op(ic, send, wc->status); in rds_ib_send_cqe_handler()
304 if (wc->status != IB_WC_SUCCESS && rds_conn_up(conn)) { in rds_ib_send_cqe_handler()
307 conn->c_tos, wc->status, in rds_ib_send_cqe_handler()
308 ib_wc_status_msg(wc->status), wc->vendor_err); in rds_ib_send_cqe_handler()
Dib_recv.c984 struct ib_wc *wc, in rds_ib_recv_cqe_handler() argument
991 (unsigned long long)wc->wr_id, wc->status, in rds_ib_recv_cqe_handler()
992 ib_wc_status_msg(wc->status), wc->byte_len, in rds_ib_recv_cqe_handler()
993 be32_to_cpu(wc->ex.imm_data)); in rds_ib_recv_cqe_handler()
1004 if (wc->status == IB_WC_SUCCESS) { in rds_ib_recv_cqe_handler()
1005 rds_ib_process_recv(conn, recv, wc->byte_len, state); in rds_ib_recv_cqe_handler()
1011 conn->c_tos, wc->status, in rds_ib_recv_cqe_handler()
1012 ib_wc_status_msg(wc->status), in rds_ib_recv_cqe_handler()
1013 wc->vendor_err); in rds_ib_recv_cqe_handler()
Dib_frmr.c322 void rds_ib_mr_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc) in rds_ib_mr_cqe_handler() argument
324 struct rds_ib_mr *ibmr = (void *)(unsigned long)wc->wr_id; in rds_ib_mr_cqe_handler()
327 if (wc->status != IB_WC_SUCCESS) { in rds_ib_mr_cqe_handler()
334 wc->status, in rds_ib_mr_cqe_handler()
335 ib_wc_status_msg(wc->status), in rds_ib_mr_cqe_handler()
336 wc->vendor_err); in rds_ib_mr_cqe_handler()
Dib.h392 void rds_ib_mr_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc);
403 void rds_ib_recv_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc,
430 void rds_ib_send_cqe_handler(struct rds_ib_connection *ic, struct ib_wc *wc);
/net/9p/
Dtrans_rdma.c291 recv_done(struct ib_cq *cq, struct ib_wc *wc) in recv_done() argument
296 container_of(wc->wr_cqe, struct p9_rdma_context, cqe); in recv_done()
305 if (wc->status != IB_WC_SUCCESS) in recv_done()
308 c->rc.size = wc->byte_len; in recv_done()
335 req, err, wc->status); in recv_done()
342 send_done(struct ib_cq *cq, struct ib_wc *wc) in send_done() argument
347 container_of(wc->wr_cqe, struct p9_rdma_context, cqe); in send_done()
/net/netfilter/
Dnfnetlink_osf.c81 if (f->wss.wc >= OSF_WSS_MAX) in nf_osf_match_one()
95 check_WSS = f->wss.wc; in nf_osf_match_one()