| /kernel/linux/linux-4.19/drivers/md/ |
| D | dm-writecache.c | 98 #define WC_MODE_PMEM(wc) ((wc)->pmem_mode) argument 99 #define WC_MODE_FUA(wc) ((wc)->writeback_fua) argument 101 #define WC_MODE_PMEM(wc) false argument 102 #define WC_MODE_FUA(wc) false argument 104 #define WC_MODE_SORT_FREELIST(wc) (!WC_MODE_PMEM(wc)) argument 190 struct dm_writecache *wc; member 201 struct dm_writecache *wc; member 210 static void wc_lock(struct dm_writecache *wc) in wc_lock() argument 212 mutex_lock(&wc->lock); in wc_lock() 215 static void wc_unlock(struct dm_writecache *wc) in wc_unlock() argument [all …]
|
| /kernel/linux/linux-5.10/drivers/md/ |
| D | dm-writecache.c | 101 #define WC_MODE_PMEM(wc) ((wc)->pmem_mode) argument 102 #define WC_MODE_FUA(wc) ((wc)->writeback_fua) argument 104 #define WC_MODE_PMEM(wc) false argument 105 #define WC_MODE_FUA(wc) false argument 107 #define WC_MODE_SORT_FREELIST(wc) (!WC_MODE_PMEM(wc)) argument 206 struct dm_writecache *wc; member 215 struct dm_writecache *wc; member 224 static void wc_lock(struct dm_writecache *wc) in wc_lock() argument 226 mutex_lock(&wc->lock); in wc_lock() 229 static void wc_unlock(struct dm_writecache *wc) in wc_unlock() argument [all …]
|
| /kernel/linux/linux-5.10/include/math-emu/ |
| D | op-common.h | 27 #define _FP_DECL(wc, X) \ argument 29 _FP_FRAC_DECL_##wc(X) 36 #define _FP_UNPACK_CANONICAL(fs, wc, X) \ argument 42 _FP_FRAC_SLL_##wc(X, _FP_WORKBITS); \ 48 if (_FP_FRAC_ZEROP_##wc(X)) \ 54 _FP_FRAC_CLZ_##wc(_shift, X); \ 56 _FP_FRAC_SLL_##wc(X, (_shift+_FP_WORKBITS)); \ 69 if (_FP_FRAC_ZEROP_##wc(X)) \ 89 #define _FP_PACK_CANONICAL(fs, wc, X) \ argument 97 _FP_ROUND(wc, X); \ [all …]
|
| D | soft-fp.h | 135 #define _FP_ROUND_NEAREST(wc, X) \ argument 137 if ((_FP_FRAC_LOW_##wc(X) & 15) != _FP_WORK_ROUND) \ 138 _FP_FRAC_ADDI_##wc(X, _FP_WORK_ROUND); \ 141 #define _FP_ROUND_ZERO(wc, X) (void)0 argument 143 #define _FP_ROUND_PINF(wc, X) \ argument 145 if (!X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \ 146 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \ 149 #define _FP_ROUND_MINF(wc, X) \ argument 151 if (X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \ 152 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \ [all …]
|
| /kernel/linux/linux-4.19/include/math-emu/ |
| D | op-common.h | 27 #define _FP_DECL(wc, X) \ argument 29 _FP_FRAC_DECL_##wc(X) 36 #define _FP_UNPACK_CANONICAL(fs, wc, X) \ argument 42 _FP_FRAC_SLL_##wc(X, _FP_WORKBITS); \ 48 if (_FP_FRAC_ZEROP_##wc(X)) \ 54 _FP_FRAC_CLZ_##wc(_shift, X); \ 56 _FP_FRAC_SLL_##wc(X, (_shift+_FP_WORKBITS)); \ 69 if (_FP_FRAC_ZEROP_##wc(X)) \ 89 #define _FP_PACK_CANONICAL(fs, wc, X) \ argument 97 _FP_ROUND(wc, X); \ [all …]
|
| D | soft-fp.h | 135 #define _FP_ROUND_NEAREST(wc, X) \ argument 137 if ((_FP_FRAC_LOW_##wc(X) & 15) != _FP_WORK_ROUND) \ 138 _FP_FRAC_ADDI_##wc(X, _FP_WORK_ROUND); \ 141 #define _FP_ROUND_ZERO(wc, X) (void)0 argument 143 #define _FP_ROUND_PINF(wc, X) \ argument 145 if (!X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \ 146 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \ 149 #define _FP_ROUND_MINF(wc, X) \ argument 151 if (X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \ 152 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \ [all …]
|
| /kernel/linux/linux-4.19/drivers/infiniband/hw/cxgb3/ |
| D | iwch_cq.c | 36 struct iwch_qp *qhp, struct ib_wc *wc) in __iwch_poll_cq_one() argument 59 wc->wr_id = cookie; in __iwch_poll_cq_one() 60 wc->qp = qhp ? &qhp->ibqp : NULL; in __iwch_poll_cq_one() 61 wc->vendor_err = CQE_STATUS(cqe); in __iwch_poll_cq_one() 62 wc->wc_flags = 0; in __iwch_poll_cq_one() 72 wc->byte_len = CQE_LEN(cqe); in __iwch_poll_cq_one() 74 wc->byte_len = 0; in __iwch_poll_cq_one() 75 wc->opcode = IB_WC_RECV; in __iwch_poll_cq_one() 78 wc->ex.invalidate_rkey = CQE_WRID_STAG(cqe); in __iwch_poll_cq_one() 79 wc->wc_flags |= IB_WC_WITH_INVALIDATE; in __iwch_poll_cq_one() [all …]
|
| /kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
| D | cq.c | 511 struct ib_wc *wc) in mlx4_ib_handle_error_cqe() argument 525 wc->status = IB_WC_LOC_LEN_ERR; in mlx4_ib_handle_error_cqe() 528 wc->status = IB_WC_LOC_QP_OP_ERR; in mlx4_ib_handle_error_cqe() 531 wc->status = IB_WC_LOC_PROT_ERR; in mlx4_ib_handle_error_cqe() 534 wc->status = IB_WC_WR_FLUSH_ERR; in mlx4_ib_handle_error_cqe() 537 wc->status = IB_WC_MW_BIND_ERR; in mlx4_ib_handle_error_cqe() 540 wc->status = IB_WC_BAD_RESP_ERR; in mlx4_ib_handle_error_cqe() 543 wc->status = IB_WC_LOC_ACCESS_ERR; in mlx4_ib_handle_error_cqe() 546 wc->status = IB_WC_REM_INV_REQ_ERR; in mlx4_ib_handle_error_cqe() 549 wc->status = IB_WC_REM_ACCESS_ERR; in mlx4_ib_handle_error_cqe() [all …]
|
| /kernel/linux/linux-4.19/drivers/infiniband/hw/mlx4/ |
| D | cq.c | 514 struct ib_wc *wc) in mlx4_ib_handle_error_cqe() argument 528 wc->status = IB_WC_LOC_LEN_ERR; in mlx4_ib_handle_error_cqe() 531 wc->status = IB_WC_LOC_QP_OP_ERR; in mlx4_ib_handle_error_cqe() 534 wc->status = IB_WC_LOC_PROT_ERR; in mlx4_ib_handle_error_cqe() 537 wc->status = IB_WC_WR_FLUSH_ERR; in mlx4_ib_handle_error_cqe() 540 wc->status = IB_WC_MW_BIND_ERR; in mlx4_ib_handle_error_cqe() 543 wc->status = IB_WC_BAD_RESP_ERR; in mlx4_ib_handle_error_cqe() 546 wc->status = IB_WC_LOC_ACCESS_ERR; in mlx4_ib_handle_error_cqe() 549 wc->status = IB_WC_REM_INV_REQ_ERR; in mlx4_ib_handle_error_cqe() 552 wc->status = IB_WC_REM_ACCESS_ERR; in mlx4_ib_handle_error_cqe() [all …]
|
| /kernel/linux/linux-4.19/drivers/infiniband/hw/mlx5/ |
| D | cq.c | 115 static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_good_req() argument 118 wc->wc_flags = 0; in handle_good_req() 121 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 124 wc->opcode = IB_WC_RDMA_WRITE; in handle_good_req() 127 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 131 wc->opcode = IB_WC_SEND; in handle_good_req() 134 wc->opcode = IB_WC_RDMA_READ; in handle_good_req() 135 wc->byte_len = be32_to_cpu(cqe->byte_cnt); in handle_good_req() 138 wc->opcode = IB_WC_COMP_SWAP; in handle_good_req() 139 wc->byte_len = 8; in handle_good_req() [all …]
|
| /kernel/linux/linux-4.19/fs/ocfs2/ |
| D | aops.c | 816 static void ocfs2_unlock_pages(struct ocfs2_write_ctxt *wc) in ocfs2_unlock_pages() argument 825 if (wc->w_target_locked) { in ocfs2_unlock_pages() 826 BUG_ON(!wc->w_target_page); in ocfs2_unlock_pages() 827 for (i = 0; i < wc->w_num_pages; i++) { in ocfs2_unlock_pages() 828 if (wc->w_target_page == wc->w_pages[i]) { in ocfs2_unlock_pages() 829 wc->w_pages[i] = NULL; in ocfs2_unlock_pages() 833 mark_page_accessed(wc->w_target_page); in ocfs2_unlock_pages() 834 put_page(wc->w_target_page); in ocfs2_unlock_pages() 836 ocfs2_unlock_and_free_pages(wc->w_pages, wc->w_num_pages); in ocfs2_unlock_pages() 855 struct ocfs2_write_ctxt *wc) in ocfs2_free_write_ctxt() argument [all …]
|
| /kernel/linux/linux-5.10/fs/ocfs2/ |
| D | aops.c | 794 static void ocfs2_unlock_pages(struct ocfs2_write_ctxt *wc) in ocfs2_unlock_pages() argument 803 if (wc->w_target_locked) { in ocfs2_unlock_pages() 804 BUG_ON(!wc->w_target_page); in ocfs2_unlock_pages() 805 for (i = 0; i < wc->w_num_pages; i++) { in ocfs2_unlock_pages() 806 if (wc->w_target_page == wc->w_pages[i]) { in ocfs2_unlock_pages() 807 wc->w_pages[i] = NULL; in ocfs2_unlock_pages() 811 mark_page_accessed(wc->w_target_page); in ocfs2_unlock_pages() 812 put_page(wc->w_target_page); in ocfs2_unlock_pages() 814 ocfs2_unlock_and_free_pages(wc->w_pages, wc->w_num_pages); in ocfs2_unlock_pages() 833 struct ocfs2_write_ctxt *wc) in ocfs2_free_write_ctxt() argument [all …]
|
| /kernel/linux/linux-4.19/drivers/infiniband/sw/rdmavt/ |
| D | cq.c | 66 struct rvt_cq_wc *wc; in rvt_cq_enter() local 77 wc = cq->queue; in rvt_cq_enter() 78 head = wc->head; in rvt_cq_enter() 86 if (unlikely(next == wc->tail)) { in rvt_cq_enter() 100 wc->uqueue[head].wr_id = entry->wr_id; in rvt_cq_enter() 101 wc->uqueue[head].status = entry->status; in rvt_cq_enter() 102 wc->uqueue[head].opcode = entry->opcode; in rvt_cq_enter() 103 wc->uqueue[head].vendor_err = entry->vendor_err; in rvt_cq_enter() 104 wc->uqueue[head].byte_len = entry->byte_len; in rvt_cq_enter() 105 wc->uqueue[head].ex.imm_data = entry->ex.imm_data; in rvt_cq_enter() [all …]
|
| D | trace_cq.h | 112 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), 113 TP_ARGS(cq, wc, idx), 125 __entry->wr_id = wc->wr_id; 126 __entry->status = wc->status; 127 __entry->opcode = wc->opcode; 128 __entry->length = wc->byte_len; 129 __entry->qpn = wc->qp->qp_num; 146 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), 147 TP_ARGS(cq, wc, idx)); 151 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), [all …]
|
| /kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
| D | cq.c | 117 static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_good_req() argument 120 wc->wc_flags = 0; in handle_good_req() 123 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 126 wc->opcode = IB_WC_RDMA_WRITE; in handle_good_req() 129 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req() 133 wc->opcode = IB_WC_SEND; in handle_good_req() 136 wc->opcode = IB_WC_RDMA_READ; in handle_good_req() 137 wc->byte_len = be32_to_cpu(cqe->byte_cnt); in handle_good_req() 140 wc->opcode = IB_WC_COMP_SWAP; in handle_good_req() 141 wc->byte_len = 8; in handle_good_req() [all …]
|
| /kernel/linux/linux-5.10/drivers/net/ethernet/brocade/bna/ |
| D | bfa_cs.h | 73 bfa_wc_up(struct bfa_wc *wc) in bfa_wc_up() argument 75 wc->wc_count++; in bfa_wc_up() 79 bfa_wc_down(struct bfa_wc *wc) in bfa_wc_down() argument 81 wc->wc_count--; in bfa_wc_down() 82 if (wc->wc_count == 0) in bfa_wc_down() 83 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down() 88 bfa_wc_init(struct bfa_wc *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument 90 wc->wc_resume = wc_resume; in bfa_wc_init() 91 wc->wc_cbarg = wc_cbarg; in bfa_wc_init() 92 wc->wc_count = 0; in bfa_wc_init() [all …]
|
| /kernel/linux/linux-4.19/drivers/net/ethernet/brocade/bna/ |
| D | bfa_cs.h | 81 bfa_wc_up(struct bfa_wc *wc) in bfa_wc_up() argument 83 wc->wc_count++; in bfa_wc_up() 87 bfa_wc_down(struct bfa_wc *wc) in bfa_wc_down() argument 89 wc->wc_count--; in bfa_wc_down() 90 if (wc->wc_count == 0) in bfa_wc_down() 91 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down() 96 bfa_wc_init(struct bfa_wc *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument 98 wc->wc_resume = wc_resume; in bfa_wc_init() 99 wc->wc_cbarg = wc_cbarg; in bfa_wc_init() 100 wc->wc_count = 0; in bfa_wc_init() [all …]
|
| /kernel/linux/linux-4.19/Documentation/x86/ |
| D | pat.txt | 41 ioremap_wc | -- | -- | WC | 48 set_memory_wc | WC | -- | -- | 56 pci sysfs resource_wc | -- | -- | WC | 62 pci proc | -- | -- | WC | 65 /dev/mem | -- | WB/WC/UC- | WB/WC/UC- | 71 /dev/mem | -- | WB/WC/UC- | WB/WC/UC- | 104 In addition, step 2 internally tracks the region as UC or WC in memtype 119 is made, should already have been ioremapped with WC attributes or PAT entries, 137 WC 000 WB _PAGE_CACHE_MODE_WB WC | WC 138 WC 001 WC _PAGE_CACHE_MODE_WC WC* | WC [all …]
|
| /kernel/linux/linux-5.10/drivers/infiniband/sw/siw/ |
| D | siw_cq.c | 48 int siw_reap_cqe(struct siw_cq *cq, struct ib_wc *wc) in siw_reap_cqe() argument 57 memset(wc, 0, sizeof(*wc)); in siw_reap_cqe() 58 wc->wr_id = cqe->id; in siw_reap_cqe() 59 wc->status = map_cqe_status[cqe->status].ib; in siw_reap_cqe() 60 wc->opcode = map_wc_opcode[cqe->opcode]; in siw_reap_cqe() 61 wc->byte_len = cqe->bytes; in siw_reap_cqe() 70 wc->ex.invalidate_rkey = cqe->inval_stag; in siw_reap_cqe() 71 wc->wc_flags = IB_WC_WITH_INVALIDATE; in siw_reap_cqe() 73 wc->qp = cqe->base_qp; in siw_reap_cqe() 98 struct ib_wc wc; in siw_cq_flush() local [all …]
|
| /kernel/linux/linux-4.19/drivers/infiniband/hw/qib/ |
| D | qib_ud.c | 61 struct ib_wc wc; in qib_ud_loopback() local 124 memset(&wc, 0, sizeof(wc)); in qib_ud_loopback() 125 wc.byte_len = length + sizeof(struct ib_grh); in qib_ud_loopback() 128 wc.wc_flags = IB_WC_WITH_IMM; in qib_ud_loopback() 129 wc.ex.imm_data = swqe->wr.ex.imm_data; in qib_ud_loopback() 154 if (unlikely(wc.byte_len > qp->r_len)) { in qib_ud_loopback() 167 wc.wc_flags |= IB_WC_GRH; in qib_ud_loopback() 205 wc.wr_id = qp->r_wr_id; in qib_ud_loopback() 206 wc.status = IB_WC_SUCCESS; in qib_ud_loopback() 207 wc.opcode = IB_WC_RECV; in qib_ud_loopback() [all …]
|
| /kernel/linux/linux-5.10/drivers/infiniband/hw/qib/ |
| D | qib_ud.c | 62 struct ib_wc wc; in qib_ud_loopback() local 125 memset(&wc, 0, sizeof(wc)); in qib_ud_loopback() 126 wc.byte_len = length + sizeof(struct ib_grh); in qib_ud_loopback() 129 wc.wc_flags = IB_WC_WITH_IMM; in qib_ud_loopback() 130 wc.ex.imm_data = swqe->wr.ex.imm_data; in qib_ud_loopback() 155 if (unlikely(wc.byte_len > qp->r_len)) { in qib_ud_loopback() 168 wc.wc_flags |= IB_WC_GRH; in qib_ud_loopback() 201 wc.wr_id = qp->r_wr_id; in qib_ud_loopback() 202 wc.status = IB_WC_SUCCESS; in qib_ud_loopback() 203 wc.opcode = IB_WC_RECV; in qib_ud_loopback() [all …]
|
| /kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/ |
| D | trace_cq.h | 112 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), 113 TP_ARGS(cq, wc, idx), 127 __entry->wr_id = wc->wr_id; 128 __entry->status = wc->status; 129 __entry->opcode = wc->opcode; 130 __entry->length = wc->byte_len; 131 __entry->qpn = wc->qp->qp_num; 133 __entry->flags = wc->wc_flags; 134 __entry->imm = be32_to_cpu(wc->ex.imm_data); 152 TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), [all …]
|
| /kernel/linux/linux-5.10/Documentation/x86/ |
| D | pat.rst | 22 WC Write-combined 47 | ioremap_wc | -- | -- | WC | 54 | set_memory_wc, | WC | -- | -- | 62 | pci sysfs resource_wc | -- | -- | WC | 68 | pci proc | -- | -- | WC | 71 | /dev/mem | -- | WB/WC/UC- | WB/WC/UC- | 77 | /dev/mem | -- | WB/WC/UC- | WB/WC/UC- | 112 In addition, step 2 internally tracks the region as UC or WC in memtype 127 is made, should already have been ioremapped with WC attributes or PAT entries, 145 WC 000 WB _PAGE_CACHE_MODE_WB WC | WC [all …]
|
| /kernel/linux/linux-5.10/net/sunrpc/xprtrdma/ |
| D | frwr_ops.c | 355 * frwr_wc_fastreg - Invoked by RDMA provider for a flushed FastReg WC 357 * @wc: WCE for a completed FastReg WR 360 static void frwr_wc_fastreg(struct ib_cq *cq, struct ib_wc *wc) in frwr_wc_fastreg() argument 362 struct ib_cqe *cqe = wc->wr_cqe; in frwr_wc_fastreg() 367 trace_xprtrdma_wc_fastreg(wc, frwr); in frwr_wc_fastreg() 370 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_fastreg() 430 static void __frwr_release_mr(struct ib_wc *wc, struct rpcrdma_mr *mr) in __frwr_release_mr() argument 432 if (wc->status != IB_WC_SUCCESS) in __frwr_release_mr() 439 * frwr_wc_localinv - Invoked by RDMA provider for a LOCAL_INV WC 441 * @wc: WCE for a completed LocalInv WR [all …]
|
| /kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/ |
| D | pvrdma_cq.c | 320 struct ib_wc *wc) in pvrdma_poll_one() argument 353 wc->opcode = pvrdma_wc_opcode_to_ib(cqe->opcode); in pvrdma_poll_one() 354 wc->status = pvrdma_wc_status_to_ib(cqe->status); in pvrdma_poll_one() 355 wc->wr_id = cqe->wr_id; in pvrdma_poll_one() 356 wc->qp = &(*cur_qp)->ibqp; in pvrdma_poll_one() 357 wc->byte_len = cqe->byte_len; in pvrdma_poll_one() 358 wc->ex.imm_data = cqe->imm_data; in pvrdma_poll_one() 359 wc->src_qp = cqe->src_qp; in pvrdma_poll_one() 360 wc->wc_flags = pvrdma_wc_flags_to_ib(cqe->wc_flags); in pvrdma_poll_one() 361 wc->pkey_index = cqe->pkey_index; in pvrdma_poll_one() [all …]
|