/drivers/md/ |
D | dm-writecache.c | 104 #define WC_MODE_PMEM(wc) ((wc)->pmem_mode) argument 105 #define WC_MODE_FUA(wc) ((wc)->writeback_fua) argument 107 #define WC_MODE_PMEM(wc) false argument 108 #define WC_MODE_FUA(wc) false argument 110 #define WC_MODE_SORT_FREELIST(wc) (!WC_MODE_PMEM(wc)) argument 228 struct dm_writecache *wc; member 237 struct dm_writecache *wc; member 246 static void wc_lock(struct dm_writecache *wc) in wc_lock() argument 248 mutex_lock(&wc->lock); in wc_lock() 251 static void wc_unlock(struct dm_writecache *wc) in wc_unlock() argument [all …]
|
/drivers/infiniband/hw/mlx4/ |
D | cq.c | 511 struct ib_wc *wc) in mlx4_ib_handle_error_cqe() argument 525 wc->status = IB_WC_LOC_LEN_ERR; in mlx4_ib_handle_error_cqe() 528 wc->status = IB_WC_LOC_QP_OP_ERR; in mlx4_ib_handle_error_cqe() 531 wc->status = IB_WC_LOC_PROT_ERR; in mlx4_ib_handle_error_cqe() 534 wc->status = IB_WC_WR_FLUSH_ERR; in mlx4_ib_handle_error_cqe() 537 wc->status = IB_WC_MW_BIND_ERR; in mlx4_ib_handle_error_cqe() 540 wc->status = IB_WC_BAD_RESP_ERR; in mlx4_ib_handle_error_cqe() 543 wc->status = IB_WC_LOC_ACCESS_ERR; in mlx4_ib_handle_error_cqe() 546 wc->status = IB_WC_REM_INV_REQ_ERR; in mlx4_ib_handle_error_cqe() 549 wc->status = IB_WC_REM_ACCESS_ERR; in mlx4_ib_handle_error_cqe() [all …]
|
D | mad.c | 512 enum ib_qp_type dest_qpt, struct ib_wc *wc, in mlx4_ib_send_to_slave() argument 550 ret = ib_get_cached_pkey(&dev->ib_dev, port, wc->pkey_index, &cached_pkey); in mlx4_ib_send_to_slave() 553 is_proxy_qp0(dev, wc->src_qp, slave) ? "SMI" : "GSI", in mlx4_ib_send_to_slave() 554 wc->pkey_index, ret); in mlx4_ib_send_to_slave() 561 is_proxy_qp0(dev, wc->src_qp, slave) ? "SMI" : "GSI", in mlx4_ib_send_to_slave() 619 tun_mad->hdr.flags_src_qp = cpu_to_be32(wc->src_qp & 0xFFFFFF); in mlx4_ib_send_to_slave() 620 tun_mad->hdr.g_ml_path = (grh && (wc->wc_flags & IB_WC_GRH)) ? 0x80 : 0; in mlx4_ib_send_to_slave() 627 if (vlan != wc->vlan_id) in mlx4_ib_send_to_slave() 638 vlan = wc->vlan_id; in mlx4_ib_send_to_slave() 642 memcpy((char *)&tun_mad->hdr.mac_31_0, &(wc->smac[0]), 4); in mlx4_ib_send_to_slave() [all …]
|
/drivers/infiniband/hw/mlx5/ |
D | cq.c | 117 static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_good_req() argument 120 wc->wc_flags = 0; in handle_good_req() 123 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 126 wc->opcode = IB_WC_RDMA_WRITE; in handle_good_req() 129 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 133 wc->opcode = IB_WC_SEND; in handle_good_req() 136 wc->opcode = IB_WC_RDMA_READ; in handle_good_req() 137 wc->byte_len = be32_to_cpu(cqe->byte_cnt); in handle_good_req() 140 wc->opcode = IB_WC_COMP_SWAP; in handle_good_req() 141 wc->byte_len = 8; in handle_good_req() [all …]
|
D | gsi.c | 37 struct ib_wc wc; member 61 WARN_ON_ONCE(mlx5_ib_generate_wc(gsi_cq, &wr->wc)); in generate_completions() 68 static void handle_single_completion(struct ib_cq *cq, struct ib_wc *wc) in handle_single_completion() argument 72 container_of(wc->wr_cqe, struct mlx5_ib_gsi_wr, cqe); in handle_single_completion() 79 wr_id = wr->wc.wr_id; in handle_single_completion() 80 wr->wc = *wc; in handle_single_completion() 81 wr->wc.wr_id = wr_id; in handle_single_completion() 82 wr->wc.qp = &mqp->ibqp; in handle_single_completion() 360 struct ib_ud_wr *wr, struct ib_wc *wc) in mlx5_ib_add_outstanding_wr() argument 375 if (!wc) { in mlx5_ib_add_outstanding_wr() [all …]
|
/drivers/net/ethernet/brocade/bna/ |
D | bfa_cs.h | 73 bfa_wc_up(struct bfa_wc *wc) in bfa_wc_up() argument 75 wc->wc_count++; in bfa_wc_up() 79 bfa_wc_down(struct bfa_wc *wc) in bfa_wc_down() argument 81 wc->wc_count--; in bfa_wc_down() 82 if (wc->wc_count == 0) in bfa_wc_down() 83 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down() 88 bfa_wc_init(struct bfa_wc *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument 90 wc->wc_resume = wc_resume; in bfa_wc_init() 91 wc->wc_cbarg = wc_cbarg; in bfa_wc_init() 92 wc->wc_count = 0; in bfa_wc_init() [all …]
|
/drivers/infiniband/sw/siw/ |
D | siw_cq.c | 48 int siw_reap_cqe(struct siw_cq *cq, struct ib_wc *wc) in siw_reap_cqe() argument 57 memset(wc, 0, sizeof(*wc)); in siw_reap_cqe() 58 wc->wr_id = cqe->id; in siw_reap_cqe() 59 wc->byte_len = cqe->bytes; in siw_reap_cqe() 68 wc->ex.invalidate_rkey = cqe->inval_stag; in siw_reap_cqe() 69 wc->wc_flags = IB_WC_WITH_INVALIDATE; in siw_reap_cqe() 71 wc->qp = cqe->base_qp; in siw_reap_cqe() 72 wc->opcode = map_wc_opcode[cqe->opcode]; in siw_reap_cqe() 73 wc->status = map_cqe_status[cqe->status].ib; in siw_reap_cqe() 95 wc->opcode = map_wc_opcode[opcode]; in siw_reap_cqe() [all …]
|
/drivers/infiniband/hw/qib/ |
D | qib_ud.c | 62 struct ib_wc wc; in qib_ud_loopback() local 125 memset(&wc, 0, sizeof(wc)); in qib_ud_loopback() 126 wc.byte_len = length + sizeof(struct ib_grh); in qib_ud_loopback() 129 wc.wc_flags = IB_WC_WITH_IMM; in qib_ud_loopback() 130 wc.ex.imm_data = swqe->wr.ex.imm_data; in qib_ud_loopback() 155 if (unlikely(wc.byte_len > qp->r_len)) { in qib_ud_loopback() 168 wc.wc_flags |= IB_WC_GRH; in qib_ud_loopback() 201 wc.wr_id = qp->r_wr_id; in qib_ud_loopback() 202 wc.status = IB_WC_SUCCESS; in qib_ud_loopback() 203 wc.opcode = IB_WC_RECV; in qib_ud_loopback() [all …]
|
D | qib_uc.c | 246 struct ib_wc wc; in qib_uc_rcv() local 368 wc.ex.imm_data = ohdr->u.imm_data; in qib_uc_rcv() 370 wc.wc_flags = IB_WC_WITH_IMM; in qib_uc_rcv() 374 wc.ex.imm_data = 0; in qib_uc_rcv() 375 wc.wc_flags = 0; in qib_uc_rcv() 385 wc.byte_len = tlen + qp->r_rcv_len; in qib_uc_rcv() 386 if (unlikely(wc.byte_len > qp->r_len)) in qib_uc_rcv() 388 wc.opcode = IB_WC_RECV; in qib_uc_rcv() 392 wc.wr_id = qp->r_wr_id; in qib_uc_rcv() 393 wc.status = IB_WC_SUCCESS; in qib_uc_rcv() [all …]
|
/drivers/infiniband/sw/rdmavt/ |
D | trace_cq.h | 70 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), 71 TP_ARGS(cq, wc, idx), 85 __entry->wr_id = wc->wr_id; 86 __entry->status = wc->status; 87 __entry->opcode = wc->opcode; 88 __entry->length = wc->byte_len; 89 __entry->qpn = wc->qp->qp_num; 91 __entry->flags = wc->wc_flags; 92 __entry->imm = be32_to_cpu(wc->ex.imm_data); 110 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), [all …]
|
/drivers/infiniband/hw/hfi1/ |
D | ud.c | 41 struct ib_wc wc; in ud_loopback() local 107 memset(&wc, 0, sizeof(wc)); in ud_loopback() 108 wc.byte_len = length + sizeof(struct ib_grh); in ud_loopback() 111 wc.wc_flags = IB_WC_WITH_IMM; in ud_loopback() 112 wc.ex.imm_data = swqe->wr.ex.imm_data; in ud_loopback() 137 if (unlikely(wc.byte_len > qp->r_len)) { in ud_loopback() 174 wc.wc_flags |= IB_WC_GRH; in ud_loopback() 193 wc.wr_id = qp->r_wr_id; in ud_loopback() 194 wc.status = IB_WC_SUCCESS; in ud_loopback() 195 wc.opcode = IB_WC_RECV; in ud_loopback() [all …]
|
D | uc.c | 270 struct ib_wc wc; in hfi1_uc_rcv() local 389 wc.ex.imm_data = ohdr->u.imm_data; in hfi1_uc_rcv() 390 wc.wc_flags = IB_WC_WITH_IMM; in hfi1_uc_rcv() 394 wc.ex.imm_data = 0; in hfi1_uc_rcv() 395 wc.wc_flags = 0; in hfi1_uc_rcv() 403 wc.byte_len = tlen + qp->r_rcv_len; in hfi1_uc_rcv() 404 if (unlikely(wc.byte_len > qp->r_len)) in hfi1_uc_rcv() 406 wc.opcode = IB_WC_RECV; in hfi1_uc_rcv() 410 wc.wr_id = qp->r_wr_id; in hfi1_uc_rcv() 411 wc.status = IB_WC_SUCCESS; in hfi1_uc_rcv() [all …]
|
/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_cq.c | 323 struct ib_wc *wc) in pvrdma_poll_one() argument 356 wc->opcode = pvrdma_wc_opcode_to_ib(cqe->opcode); in pvrdma_poll_one() 357 wc->status = pvrdma_wc_status_to_ib(cqe->status); in pvrdma_poll_one() 358 wc->wr_id = cqe->wr_id; in pvrdma_poll_one() 359 wc->qp = &(*cur_qp)->ibqp; in pvrdma_poll_one() 360 wc->byte_len = cqe->byte_len; in pvrdma_poll_one() 361 wc->ex.imm_data = cqe->imm_data; in pvrdma_poll_one() 362 wc->src_qp = cqe->src_qp; in pvrdma_poll_one() 363 wc->wc_flags = pvrdma_wc_flags_to_ib(cqe->wc_flags); in pvrdma_poll_one() 364 wc->pkey_index = cqe->pkey_index; in pvrdma_poll_one() [all …]
|
/drivers/scsi/bfa/ |
D | bfa_cs.h | 253 bfa_wc_up(struct bfa_wc_s *wc) in bfa_wc_up() argument 255 wc->wc_count++; in bfa_wc_up() 259 bfa_wc_down(struct bfa_wc_s *wc) in bfa_wc_down() argument 261 wc->wc_count--; in bfa_wc_down() 262 if (wc->wc_count == 0) in bfa_wc_down() 263 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down() 270 bfa_wc_init(struct bfa_wc_s *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument 272 wc->wc_resume = wc_resume; in bfa_wc_init() 273 wc->wc_cbarg = wc_cbarg; in bfa_wc_init() 274 wc->wc_count = 0; in bfa_wc_init() [all …]
|
/drivers/infiniband/ulp/iser/ |
D | iser_initiator.c | 562 void iser_login_rsp(struct ib_cq *cq, struct ib_wc *wc) in iser_login_rsp() argument 564 struct ib_conn *ib_conn = wc->qp->qp_context; in iser_login_rsp() 566 struct iser_login_desc *desc = iser_login(wc->wr_cqe); in iser_login_rsp() 571 if (unlikely(wc->status != IB_WC_SUCCESS)) { in iser_login_rsp() 572 iser_err_comp(wc, "login_rsp"); in iser_login_rsp() 582 length = wc->byte_len - ISER_HEADERS_LEN; in iser_login_rsp() 615 struct ib_wc *wc, in iser_check_remote_inv() argument 618 if (wc->wc_flags & IB_WC_WITH_INVALIDATE) { in iser_check_remote_inv() 620 u32 rkey = wc->ex.invalidate_rkey; in iser_check_remote_inv() 657 void iser_task_rsp(struct ib_cq *cq, struct ib_wc *wc) in iser_task_rsp() argument [all …]
|
/drivers/infiniband/hw/cxgb4/ |
D | cq.c | 755 struct ib_wc *wc, struct c4iw_srq *srq) in __c4iw_poll_cq_one() argument 769 wc->wr_id = cookie; in __c4iw_poll_cq_one() 770 wc->qp = qhp ? &qhp->ibqp : NULL; in __c4iw_poll_cq_one() 771 wc->vendor_err = CQE_STATUS(&cqe); in __c4iw_poll_cq_one() 772 wc->wc_flags = 0; in __c4iw_poll_cq_one() 790 wc->byte_len = CQE_LEN(&cqe); in __c4iw_poll_cq_one() 792 wc->byte_len = 0; in __c4iw_poll_cq_one() 796 wc->opcode = IB_WC_RECV; in __c4iw_poll_cq_one() 800 wc->opcode = IB_WC_RECV; in __c4iw_poll_cq_one() 801 wc->ex.invalidate_rkey = CQE_WRID_STAG(&cqe); in __c4iw_poll_cq_one() [all …]
|
/drivers/infiniband/ulp/ipoib/ |
D | ipoib_ib.c | 173 static void ipoib_ib_handle_rx_wc(struct net_device *dev, struct ib_wc *wc) in ipoib_ib_handle_rx_wc() argument 176 unsigned int wr_id = wc->wr_id & ~IPOIB_OP_RECV; in ipoib_ib_handle_rx_wc() 183 wr_id, wc->status); in ipoib_ib_handle_rx_wc() 193 if (unlikely(wc->status != IB_WC_SUCCESS)) { in ipoib_ib_handle_rx_wc() 194 if (wc->status != IB_WC_WR_FLUSH_ERR) in ipoib_ib_handle_rx_wc() 197 wc->status, wr_id, wc->vendor_err); in ipoib_ib_handle_rx_wc() 217 wc->byte_len, wc->slid); in ipoib_ib_handle_rx_wc() 221 skb_put(skb, wc->byte_len); in ipoib_ib_handle_rx_wc() 226 if (!(wc->wc_flags & IB_WC_GRH) || dgid->raw[0] != 0xff) in ipoib_ib_handle_rx_wc() 239 if (wc->slid == priv->local_lid && wc->src_qp == priv->qp->qp_num) { in ipoib_ib_handle_rx_wc() [all …]
|
/drivers/infiniband/core/ |
D | cq.c | 84 static int __poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc) in __poll_cq() argument 88 rc = ib_poll_cq(cq, num_entries, wc); in __poll_cq() 108 struct ib_wc *wc = &wcs[i]; in __ib_process_cq() local 110 if (wc->wr_cqe) in __ib_process_cq() 111 wc->wr_cqe->done(cq, wc); in __ib_process_cq() 113 WARN_ON_ONCE(wc->status == IB_WC_SUCCESS); in __ib_process_cq() 158 completed = __ib_process_cq(cq, budget, cq->wc, IB_POLL_BATCH); in ib_poll_handler() 184 completed = __ib_process_cq(cq, IB_POLL_BUDGET_WORKQUEUE, cq->wc, in ib_cq_poll_work() 234 cq->wc = kmalloc_array(IB_POLL_BATCH, sizeof(*cq->wc), GFP_KERNEL); in __ib_alloc_cq() 235 if (!cq->wc) in __ib_alloc_cq() [all …]
|
D | mad.c | 108 struct ib_wc *wc); 109 static void ib_mad_send_done(struct ib_cq *cq, struct ib_wc *wc); 546 u16 pkey_index, u32 port_num, struct ib_wc *wc) in build_smp_wc() argument 548 memset(wc, 0, sizeof *wc); in build_smp_wc() 549 wc->wr_cqe = cqe; in build_smp_wc() 550 wc->status = IB_WC_SUCCESS; in build_smp_wc() 551 wc->opcode = IB_WC_RECV; in build_smp_wc() 552 wc->pkey_index = pkey_index; in build_smp_wc() 553 wc->byte_len = sizeof(struct ib_mad) + sizeof(struct ib_grh); in build_smp_wc() 554 wc->src_qp = IB_QP0; in build_smp_wc() [all …]
|
D | mad_rmpp.c | 137 msg = ib_create_send_mad(&rmpp_recv->agent->agent, recv_wc->wc->src_qp, in ack_recv() 138 recv_wc->wc->pkey_index, 1, hdr_len, in ack_recv() 158 ah = ib_create_ah_from_wc(agent->qp->pd, recv_wc->wc, in alloc_response_msg() 164 msg = ib_create_send_mad(agent, recv_wc->wc->src_qp, in alloc_response_msg() 165 recv_wc->wc->pkey_index, 1, in alloc_response_msg() 287 mad_recv_wc->wc, in create_rmpp_recv() 310 rmpp_recv->src_qp = mad_recv_wc->wc->src_qp; in create_rmpp_recv() 311 rmpp_recv->slid = mad_recv_wc->wc->slid; in create_rmpp_recv() 331 rmpp_recv->src_qp == mad_recv_wc->wc->src_qp && in find_rmpp_recv() 332 rmpp_recv->slid == mad_recv_wc->wc->slid && in find_rmpp_recv() [all …]
|
/drivers/infiniband/hw/qedr/ |
D | qedr_roce_cm.c | 671 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in qedr_gsi_poll_cq() argument 683 memset(&wc[i], 0, sizeof(*wc)); in qedr_gsi_poll_cq() 685 wc[i].qp = &qp->ibqp; in qedr_gsi_poll_cq() 686 wc[i].wr_id = qp->rqe_wr_id[qp->rq.cons].wr_id; in qedr_gsi_poll_cq() 687 wc[i].opcode = IB_WC_RECV; in qedr_gsi_poll_cq() 688 wc[i].pkey_index = 0; in qedr_gsi_poll_cq() 689 wc[i].status = (qp->rqe_wr_id[qp->rq.cons].rc) ? in qedr_gsi_poll_cq() 692 wc[i].byte_len = qp->rqe_wr_id[qp->rq.cons].sg_list[0].length; in qedr_gsi_poll_cq() 693 wc[i].wc_flags |= IB_WC_GRH | IB_WC_IP_CSUM_OK; in qedr_gsi_poll_cq() 694 ether_addr_copy(wc[i].smac, qp->rqe_wr_id[qp->rq.cons].smac); in qedr_gsi_poll_cq() [all …]
|
/drivers/net/wireless/intersil/hostap/ |
D | hostap_cs.c | 107 u8 *buf, int wc) in hfa384x_outsw_debug() argument 116 prism2_io_debug_add(dev, PRISM2_IO_DEBUG_CMD_OUTSW, a, wc); in hfa384x_outsw_debug() 117 outsw(dev->base_addr + a, buf, wc); in hfa384x_outsw_debug() 122 u8 *buf, int wc) in hfa384x_insw_debug() argument 131 prism2_io_debug_add(dev, PRISM2_IO_DEBUG_CMD_INSW, a, wc); in hfa384x_insw_debug() 132 insw(dev->base_addr + a, buf, wc); in hfa384x_insw_debug() 140 #define HFA384X_OUTSW(a, buf, wc) hfa384x_outsw_debug(dev, (a), (buf), (wc)) argument 141 #define HFA384X_INSW(a, buf, wc) hfa384x_insw_debug(dev, (a), (buf), (wc)) argument 149 #define HFA384X_INSW(a, buf, wc) insw(dev->base_addr + (a), buf, wc) argument 150 #define HFA384X_OUTSW(a, buf, wc) outsw(dev->base_addr + (a), buf, wc) argument
|
D | hostap_plx.c | 173 u8 *buf, int wc) in hfa384x_outsw_debug() argument 183 prism2_io_debug_add(dev, PRISM2_IO_DEBUG_CMD_OUTSW, a, wc); in hfa384x_outsw_debug() 184 outsw(dev->base_addr + a, buf, wc); in hfa384x_outsw_debug() 189 u8 *buf, int wc) in hfa384x_insw_debug() argument 199 prism2_io_debug_add(dev, PRISM2_IO_DEBUG_CMD_INSW, a, wc); in hfa384x_insw_debug() 200 insw(dev->base_addr + a, buf, wc); in hfa384x_insw_debug() 208 #define HFA384X_OUTSW(a, buf, wc) hfa384x_outsw_debug(dev, (a), (buf), (wc)) argument 209 #define HFA384X_INSW(a, buf, wc) hfa384x_insw_debug(dev, (a), (buf), (wc)) argument 217 #define HFA384X_INSW(a, buf, wc) insw(dev->base_addr + (a), buf, wc) argument 218 #define HFA384X_OUTSW(a, buf, wc) outsw(dev->base_addr + (a), buf, wc) argument
|
/drivers/infiniband/ulp/isert/ |
D | ib_isert.c | 57 static void isert_recv_done(struct ib_cq *cq, struct ib_wc *wc); 58 static void isert_send_done(struct ib_cq *cq, struct ib_wc *wc); 59 static void isert_login_recv_done(struct ib_cq *cq, struct ib_wc *wc); 60 static void isert_login_send_done(struct ib_cq *cq, struct ib_wc *wc); 1307 isert_print_wc(struct ib_wc *wc, const char *type) in isert_print_wc() argument 1309 if (wc->status != IB_WC_WR_FLUSH_ERR) in isert_print_wc() 1311 ib_wc_status_msg(wc->status), wc->status, in isert_print_wc() 1312 wc->vendor_err); in isert_print_wc() 1315 ib_wc_status_msg(wc->status), wc->status); in isert_print_wc() 1319 isert_recv_done(struct ib_cq *cq, struct ib_wc *wc) in isert_recv_done() argument [all …]
|
/drivers/infiniband/hw/bnxt_re/ |
D | ib_verbs.c | 3003 static void bnxt_re_process_req_wc(struct ib_wc *wc, struct bnxt_qplib_cqe *cqe) in bnxt_re_process_req_wc() argument 3007 wc->opcode = IB_WC_SEND; in bnxt_re_process_req_wc() 3010 wc->opcode = IB_WC_SEND; in bnxt_re_process_req_wc() 3011 wc->wc_flags |= IB_WC_WITH_IMM; in bnxt_re_process_req_wc() 3014 wc->opcode = IB_WC_SEND; in bnxt_re_process_req_wc() 3015 wc->wc_flags |= IB_WC_WITH_INVALIDATE; in bnxt_re_process_req_wc() 3018 wc->opcode = IB_WC_RDMA_WRITE; in bnxt_re_process_req_wc() 3021 wc->opcode = IB_WC_RDMA_WRITE; in bnxt_re_process_req_wc() 3022 wc->wc_flags |= IB_WC_WITH_IMM; in bnxt_re_process_req_wc() 3025 wc->opcode = IB_WC_RDMA_READ; in bnxt_re_process_req_wc() [all …]
|