/drivers/media/platform/rockchip/rkisp1/ |
D | rkisp1-capture.c | 63 void (*config)(struct rkisp1_capture *cap); 64 void (*stop)(struct rkisp1_capture *cap); 65 void (*enable)(struct rkisp1_capture *cap); 66 void (*disable)(struct rkisp1_capture *cap); 67 void (*set_data_path)(struct rkisp1_capture *cap); 68 bool (*is_stopped)(struct rkisp1_capture *cap); 345 int rkisp1_cap_enum_mbus_codes(struct rkisp1_capture *cap, in rkisp1_cap_enum_mbus_codes() argument 348 const struct rkisp1_capture_fmt_cfg *fmts = cap->config->fmts; in rkisp1_cap_enum_mbus_codes() 356 for (i = 0; i < cap->config->fmt_size; i++) { in rkisp1_cap_enum_mbus_codes() 373 static void rkisp1_mi_config_ctrl(struct rkisp1_capture *cap) in rkisp1_mi_config_ctrl() argument [all …]
|
/drivers/staging/greybus/ |
D | authentication.c | 47 struct gb_cap *cap = container_of(kref, struct gb_cap, kref); in cap_kref_release() local 49 kfree(cap); in cap_kref_release() 57 static void put_cap(struct gb_cap *cap) in put_cap() argument 59 kref_put(&cap->kref, cap_kref_release); in put_cap() 65 struct gb_cap *cap; in get_cap() local 69 list_for_each_entry(cap, &cap_list, node) { in get_cap() 70 if (&cap->cdev == cdev) { in get_cap() 71 kref_get(&cap->kref); in get_cap() 76 cap = NULL; in get_cap() 81 return cap; in get_cap() [all …]
|
/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_reg.h | 409 #define SXGBE_HW_FEAT_GMII(cap) ((cap & 0x00000002) >> 1) argument 410 #define SXGBE_HW_FEAT_VLAN_HASH_FILTER(cap) ((cap & 0x00000010) >> 4) argument 411 #define SXGBE_HW_FEAT_SMA(cap) ((cap & 0x00000020) >> 5) argument 412 #define SXGBE_HW_FEAT_PMT_TEMOTE_WOP(cap) ((cap & 0x00000040) >> 6) argument 413 #define SXGBE_HW_FEAT_PMT_MAGIC_PKT(cap) ((cap & 0x00000080) >> 7) argument 414 #define SXGBE_HW_FEAT_RMON(cap) ((cap & 0x00000100) >> 8) argument 415 #define SXGBE_HW_FEAT_ARP_OFFLOAD(cap) ((cap & 0x00000200) >> 9) argument 416 #define SXGBE_HW_FEAT_IEEE1500_2008(cap) ((cap & 0x00001000) >> 12) argument 417 #define SXGBE_HW_FEAT_EEE(cap) ((cap & 0x00002000) >> 13) argument 418 #define SXGBE_HW_FEAT_TX_CSUM_OFFLOAD(cap) ((cap & 0x00004000) >> 14) argument [all …]
|
/drivers/vfio/pci/ |
D | vfio_pci_zdev.c | 24 struct vfio_device_info_cap_zpci_base cap = { in zpci_base_cap() local 36 return vfio_info_add_capability(caps, &cap.header, sizeof(cap)); in zpci_base_cap() 44 struct vfio_device_info_cap_zpci_group cap = { in zpci_group_cap() local 56 return vfio_info_add_capability(caps, &cap.header, sizeof(cap)); in zpci_group_cap() 64 struct vfio_device_info_cap_zpci_util *cap; in zpci_util_cap() local 65 int cap_size = sizeof(*cap) + CLP_UTIL_STR_LEN; in zpci_util_cap() 68 cap = kmalloc(cap_size, GFP_KERNEL); in zpci_util_cap() 69 if (!cap) in zpci_util_cap() 72 cap->header.id = VFIO_DEVICE_INFO_CAP_ZPCI_UTIL; in zpci_util_cap() 73 cap->header.version = 1; in zpci_util_cap() [all …]
|
/drivers/iommu/intel/ |
D | cap_audit.h | 67 #define DO_CHECK_FEATURE_MISMATCH(a, b, cap, feature, MASK) \ argument 69 if (cap##_##feature(a) != cap##_##feature(b)) { \ 70 intel_iommu_##cap##_sanity &= ~(MASK); \ 75 #define CHECK_FEATURE_MISMATCH(a, b, cap, feature, MASK) \ argument 76 DO_CHECK_FEATURE_MISMATCH((a)->cap, (b)->cap, cap, feature, MASK) 78 #define CHECK_FEATURE_MISMATCH_HOTPLUG(b, cap, feature, MASK) \ argument 80 if (cap##_##feature(intel_iommu_##cap##_sanity)) \ 81 DO_CHECK_FEATURE_MISMATCH(intel_iommu_##cap##_sanity, \ 82 (b)->cap, cap, feature, MASK); \ 85 #define MINIMAL_FEATURE_IOMMU(iommu, cap, MASK) \ argument [all …]
|
D | cap_audit.c | 22 CHECK_FEATURE_MISMATCH(a, b, cap, pi_support, CAP_PI_MASK); in check_irq_capabilities() 29 MINIMAL_FEATURE_IOMMU(b, cap, CAP_MAMV_MASK); in check_dmar_capabilities() 30 MINIMAL_FEATURE_IOMMU(b, cap, CAP_NFR_MASK); in check_dmar_capabilities() 31 MINIMAL_FEATURE_IOMMU(b, cap, CAP_SLLPS_MASK); in check_dmar_capabilities() 32 MINIMAL_FEATURE_IOMMU(b, cap, CAP_FRO_MASK); in check_dmar_capabilities() 33 MINIMAL_FEATURE_IOMMU(b, cap, CAP_MGAW_MASK); in check_dmar_capabilities() 34 MINIMAL_FEATURE_IOMMU(b, cap, CAP_SAGAW_MASK); in check_dmar_capabilities() 35 MINIMAL_FEATURE_IOMMU(b, cap, CAP_NDOMS_MASK); in check_dmar_capabilities() 40 CHECK_FEATURE_MISMATCH(a, b, cap, 5lp_support, CAP_FL5LP_MASK); in check_dmar_capabilities() 41 CHECK_FEATURE_MISMATCH(a, b, cap, fl1gp_support, CAP_FL1GP_MASK); in check_dmar_capabilities() [all …]
|
/drivers/pci/pcie/ |
D | dpc.c | 45 u16 *cap; in pci_save_dpc_state() local 54 cap = (u16 *)&save_state->cap.data[0]; in pci_save_dpc_state() 55 pci_read_config_word(dev, dev->dpc_cap + PCI_EXP_DPC_CTL, cap); in pci_save_dpc_state() 61 u16 *cap; in pci_restore_dpc_state() local 70 cap = (u16 *)&save_state->cap.data[0]; in pci_restore_dpc_state() 71 pci_write_config_word(dev, dev->dpc_cap + PCI_EXP_DPC_CTL, *cap); in pci_restore_dpc_state() 129 u16 cap = pdev->dpc_cap, status; in dpc_wait_rp_inactive() local 131 pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status); in dpc_wait_rp_inactive() 135 pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status); in dpc_wait_rp_inactive() 147 u16 cap; in dpc_reset_link() local [all …]
|
D | ptm.c | 53 u16 *cap; in pci_save_ptm_state() local 66 cap = (u16 *)&save_state->cap.data[0]; in pci_save_ptm_state() 67 pci_read_config_word(dev, ptm + PCI_PTM_CTRL, cap); in pci_save_ptm_state() 74 u16 *cap; in pci_restore_ptm_state() local 84 cap = (u16 *)&save_state->cap.data[0]; in pci_restore_ptm_state() 85 pci_write_config_word(dev, ptm + PCI_PTM_CTRL, *cap); in pci_restore_ptm_state() 91 u32 cap, ctrl; in pci_ptm_init() local 126 pci_read_config_dword(dev, pos + PCI_PTM_CAP, &cap); in pci_ptm_init() 127 local_clock = (cap & PCI_PTM_GRANULARITY_MASK) >> 8; in pci_ptm_init() 142 if (cap & PCI_PTM_CAP_ROOT) { in pci_ptm_init() [all …]
|
/drivers/media/platform/qcom/venus/ |
D | hfi_parser.c | 14 typedef void (*func)(struct hfi_plat_caps *cap, const void *data, 19 struct hfi_plat_caps *caps = core->caps, *cap; in init_codecs() local 26 cap = &caps[core->codecs_count++]; in init_codecs() 27 cap->codec = BIT(bit); in init_codecs() 28 cap->domain = VIDC_SESSION_TYPE_DEC; in init_codecs() 29 cap->valid = false; in init_codecs() 33 cap = &caps[core->codecs_count++]; in init_codecs() 34 cap->codec = BIT(bit); in init_codecs() 35 cap->domain = VIDC_SESSION_TYPE_ENC; in init_codecs() 36 cap->valid = false; in init_codecs() [all …]
|
/drivers/infiniband/core/ |
D | uverbs_std_types_qp.c | 65 struct ib_uverbs_qp_cap *cap, bool req) in set_caps() argument 68 attr->cap.max_send_wr = cap->max_send_wr; in set_caps() 69 attr->cap.max_recv_wr = cap->max_recv_wr; in set_caps() 70 attr->cap.max_send_sge = cap->max_send_sge; in set_caps() 71 attr->cap.max_recv_sge = cap->max_recv_sge; in set_caps() 72 attr->cap.max_inline_data = cap->max_inline_data; in set_caps() 74 cap->max_send_wr = attr->cap.max_send_wr; in set_caps() 75 cap->max_recv_wr = attr->cap.max_recv_wr; in set_caps() 76 cap->max_send_sge = attr->cap.max_send_sge; in set_caps() 77 cap->max_recv_sge = attr->cap.max_recv_sge; in set_caps() [all …]
|
/drivers/thunderbolt/ |
D | debugfs.c | 255 struct tb_port *port, unsigned int cap, in cap_show_by_dw() argument 264 ret = tb_port_read(port, &data, TB_CFG_PORT, cap + offset + i, 1); in cap_show_by_dw() 266 ret = tb_sw_read(sw, &data, TB_CFG_SWITCH, cap + offset + i, 1); in cap_show_by_dw() 268 seq_printf(s, "0x%04x <not accessible>\n", cap + offset + i); in cap_show_by_dw() 272 seq_printf(s, "0x%04x %4d 0x%02x 0x%02x 0x%08x\n", cap + offset + i, in cap_show_by_dw() 278 struct tb_port *port, unsigned int cap, u8 cap_id, in cap_show() argument 288 ret = tb_port_read(port, data, TB_CFG_PORT, cap + offset, in cap_show() 291 ret = tb_sw_read(sw, data, TB_CFG_SWITCH, cap + offset, dwords); in cap_show() 293 cap_show_by_dw(s, sw, port, cap, offset, cap_id, vsec_id, length); in cap_show() 299 cap + offset + i, offset + i, in cap_show() [all …]
|
D | lc.c | 53 int cap, ret; in tb_lc_set_port_configured() local 58 cap = find_port_lc_cap(port); in tb_lc_set_port_configured() 59 if (cap < 0) in tb_lc_set_port_configured() 60 return cap; in tb_lc_set_port_configured() 62 ret = tb_sw_read(sw, &ctrl, TB_CFG_SWITCH, cap + TB_LC_SX_CTRL, 1); in tb_lc_set_port_configured() 82 return tb_sw_write(sw, &ctrl, TB_CFG_SWITCH, cap + TB_LC_SX_CTRL, 1); in tb_lc_set_port_configured() 111 int cap, ret; in tb_lc_set_xdomain_configured() local 116 cap = find_port_lc_cap(port); in tb_lc_set_xdomain_configured() 117 if (cap < 0) in tb_lc_set_xdomain_configured() 118 return cap; in tb_lc_set_xdomain_configured() [all …]
|
D | cap.c | 87 static int __tb_port_find_cap(struct tb_port *port, enum tb_port_cap cap) in __tb_port_find_cap() argument 103 if (header.basic.cap == cap) in __tb_port_find_cap() 119 int tb_port_find_cap(struct tb_port *port, enum tb_port_cap cap) in tb_port_find_cap() argument 127 ret = __tb_port_find_cap(port, cap); in tb_port_find_cap() 157 switch (header.basic.cap) { in tb_switch_next_cap() 171 header.basic.cap, offset); in tb_switch_next_cap() 188 int tb_switch_find_cap(struct tb_switch *sw, enum tb_switch_cap cap) in tb_switch_find_cap() argument 204 if (header.basic.cap == cap) in tb_switch_find_cap() 237 if (header.extended_short.cap == TB_SWITCH_CAP_VSE && in tb_switch_find_vse_cap()
|
D | tmu.c | 46 sw->tmu.cap + TMU_RTR_CS_0, 1); in tb_switch_tmu_ucap_supported() 59 sw->tmu.cap + TMU_RTR_CS_3, 1); in tb_switch_tmu_rate_read() 73 sw->tmu.cap + TMU_RTR_CS_3, 1); in tb_switch_tmu_rate_write() 81 sw->tmu.cap + TMU_RTR_CS_3, 1); in tb_switch_tmu_rate_write() 137 sw->tmu.cap + TMU_RTR_CS_0, 1); in tb_switch_tmu_set_time_disruption() 147 sw->tmu.cap + TMU_RTR_CS_0, 1); in tb_switch_tmu_set_time_disruption() 168 sw->tmu.cap = ret; in tb_switch_tmu_init() 171 int cap; in tb_switch_tmu_init() local 173 cap = tb_port_find_cap(port, TB_PORT_CAP_TIME1); in tb_switch_tmu_init() 174 if (cap > 0) in tb_switch_tmu_init() [all …]
|
/drivers/char/tpm/ |
D | tpm-sysfs.c | 85 cap_t cap; in pcrs_show() local 94 if (tpm1_getcap(chip, TPM_CAP_PROP_PCR, &cap, in pcrs_show() 96 sizeof(cap.num_pcrs))) { in pcrs_show() 101 num_pcrs = be32_to_cpu(cap.num_pcrs); in pcrs_show() 122 cap_t cap; in enabled_show() local 127 if (tpm1_getcap(chip, TPM_CAP_FLAG_PERM, &cap, in enabled_show() 129 sizeof(cap.perm_flags))) in enabled_show() 132 rc = sprintf(buf, "%d\n", !cap.perm_flags.disable); in enabled_show() 144 cap_t cap; in active_show() local 149 if (tpm1_getcap(chip, TPM_CAP_FLAG_PERM, &cap, in active_show() [all …]
|
/drivers/infiniband/hw/hfi1/ |
D | common.h | 45 #define HFI1_CAP_KSET(cap) ({ hfi1_cap_mask |= HFI1_CAP_##cap; hfi1_cap_mask; }) argument 46 #define HFI1_CAP_KCLEAR(cap) \ argument 48 hfi1_cap_mask &= ~HFI1_CAP_##cap; \ 51 #define HFI1_CAP_USET(cap) \ argument 53 hfi1_cap_mask |= (HFI1_CAP_##cap << HFI1_CAP_USER_SHIFT); \ 56 #define HFI1_CAP_UCLEAR(cap) \ argument 58 hfi1_cap_mask &= ~(HFI1_CAP_##cap << HFI1_CAP_USER_SHIFT); \ 61 #define HFI1_CAP_SET(cap) \ argument 63 hfi1_cap_mask |= (HFI1_CAP_##cap | (HFI1_CAP_##cap << \ 67 #define HFI1_CAP_CLEAR(cap) \ argument [all …]
|
/drivers/platform/surface/aggregator/ |
D | ssh_parser.h | 27 size_t cap; member 39 static inline void sshp_buf_init(struct sshp_buf *buf, u8 *ptr, size_t cap) in sshp_buf_init() argument 43 buf->cap = cap; in sshp_buf_init() 57 static inline int sshp_buf_alloc(struct sshp_buf *buf, size_t cap, gfp_t flags) in sshp_buf_alloc() argument 61 ptr = kzalloc(cap, flags); in sshp_buf_alloc() 65 sshp_buf_init(buf, ptr, cap); in sshp_buf_alloc() 82 buf->cap = 0; in sshp_buf_free() 116 n = kfifo_out(fifo, buf->ptr + buf->len, buf->cap - buf->len); in sshp_buf_read_from_fifo()
|
/drivers/net/ethernet/netronome/nfp/bpf/ |
D | main.c | 29 return nn->cap & NFP_NET_CFG_CTRL_BPF && in nfp_net_ebpf_capable() 209 struct nfp_bpf_cap_tlv_adjust_head __iomem *cap = value; in nfp_bpf_parse_cap_adjust_head() local 212 if (length < sizeof(*cap)) { in nfp_bpf_parse_cap_adjust_head() 217 bpf->adjust_head.flags = readl(&cap->flags); in nfp_bpf_parse_cap_adjust_head() 218 bpf->adjust_head.off_min = readl(&cap->off_min); in nfp_bpf_parse_cap_adjust_head() 219 bpf->adjust_head.off_max = readl(&cap->off_max); in nfp_bpf_parse_cap_adjust_head() 220 bpf->adjust_head.guaranteed_sub = readl(&cap->guaranteed_sub); in nfp_bpf_parse_cap_adjust_head() 221 bpf->adjust_head.guaranteed_add = readl(&cap->guaranteed_add); in nfp_bpf_parse_cap_adjust_head() 240 struct nfp_bpf_cap_tlv_func __iomem *cap = value; in nfp_bpf_parse_cap_func() local 242 if (length < sizeof(*cap)) { in nfp_bpf_parse_cap_func() [all …]
|
/drivers/media/platform/exynos4-is/ |
D | common.c | 37 void __fimc_vidioc_querycap(struct device *dev, struct v4l2_capability *cap) in __fimc_vidioc_querycap() argument 39 strscpy(cap->driver, dev->driver->name, sizeof(cap->driver)); in __fimc_vidioc_querycap() 40 strscpy(cap->card, dev->driver->name, sizeof(cap->card)); in __fimc_vidioc_querycap() 41 snprintf(cap->bus_info, sizeof(cap->bus_info), in __fimc_vidioc_querycap()
|
/drivers/pci/ |
D | pci.c | 410 u8 pos, int cap, int *ttl) in __pci_find_next_cap_ttl() argument 426 if (id == cap) in __pci_find_next_cap_ttl() 434 u8 pos, int cap) in __pci_find_next_cap() argument 438 return __pci_find_next_cap_ttl(bus, devfn, pos, cap, &ttl); in __pci_find_next_cap() 441 u8 pci_find_next_capability(struct pci_dev *dev, u8 pos, int cap) in pci_find_next_capability() argument 444 pos + PCI_CAP_LIST_NEXT, cap); in pci_find_next_capability() 487 u8 pci_find_capability(struct pci_dev *dev, int cap) in pci_find_capability() argument 493 pos = __pci_find_next_cap(dev->bus, dev->devfn, pos, cap); in pci_find_capability() 512 u8 pci_bus_find_capability(struct pci_bus *bus, unsigned int devfn, int cap) in pci_bus_find_capability() argument 520 pos = __pci_find_next_cap(bus, devfn, pos, cap); in pci_bus_find_capability() [all …]
|
/drivers/usb/typec/ucsi/ |
D | ucsi.c | 259 override = !!(con->ucsi->cap.features & UCSI_CAP_ALT_MODE_OVERRIDE); in ucsi_register_altmode() 352 max_altmodes = con->ucsi->cap.num_alt_modes; in ucsi_register_altmodes_nvidia() 423 if (!(con->ucsi->cap.features & UCSI_CAP_ALT_MODE_DETAILS)) in ucsi_register_altmodes() 433 max_altmodes = con->ucsi->cap.num_alt_modes; in ucsi_register_altmodes() 1053 struct typec_capability *cap = &con->typec_cap; in ucsi_register_port() local 1054 enum typec_accessory *accessory = cap->accessory; in ucsi_register_port() 1065 cap->fwnode = ucsi_find_fwnode(con); in ucsi_register_port() 1066 con->usb_role_sw = fwnode_usb_role_switch_get(cap->fwnode); in ucsi_register_port() 1080 ret = ucsi_send_command(ucsi, command, &con->cap, sizeof(con->cap)); in ucsi_register_port() 1084 if (con->cap.op_mode & UCSI_CONCAP_OPMODE_DRP) in ucsi_register_port() [all …]
|
/drivers/cxl/core/ |
D | regs.c | 37 int cap, cap_count; in cxl_probe_component_regs() local 59 for (cap = 1; cap <= cap_count; cap++) { in cxl_probe_component_regs() 66 hdr = readl(base + cap * 0x4); in cxl_probe_component_regs() 106 int cap, cap_count; in cxl_probe_device_regs() local 118 for (cap = 1; cap <= cap_count; cap++) { in cxl_probe_device_regs() 123 readl(base + cap * 0x10)); in cxl_probe_device_regs() 124 offset = readl(base + cap * 0x10 + 0x4); in cxl_probe_device_regs() 125 length = readl(base + cap * 0x10 + 0x8); in cxl_probe_device_regs()
|
/drivers/net/arcnet/ |
D | capmode.c | 77 memcpy(pktbuf, pkthdrbuf, ARC_HDR_SIZE + sizeof(pkt->soft.cap.proto)); in rx() 78 memcpy(pktbuf + ARC_HDR_SIZE + sizeof(pkt->soft.cap.proto) + sizeof(int), in rx() 79 pkthdrbuf + ARC_HDR_SIZE + sizeof(pkt->soft.cap.proto), in rx() 80 sizeof(struct archdr) - ARC_HDR_SIZE - sizeof(pkt->soft.cap.proto)); in rx() 107 *((int *)&pkt->soft.cap.cookie[0])); in build_header() 148 *((int *)&pkt->soft.cap.cookie[0])); in prepare_tx() 171 lp->hw.copy_to_card(dev, bufnum, ofs, &pkt->soft.cap.proto, in prepare_tx() 172 sizeof(pkt->soft.cap.proto)); in prepare_tx() 178 ((unsigned char *)&pkt->soft.cap.mes), length - 1); in prepare_tx() 212 ackpkt->soft.cap.proto = 0; /* using protocol 0 for acknowledge */ in ack_tx() [all …]
|
/drivers/gpu/drm/msm/disp/dpu1/ |
D | dpu_vbif.c | 34 if (!vbif || !vbif->cap || !vbif->ops.get_halt_ctrl) { in _dpu_vbif_wait_for_xin_halt() 39 timeout = ktime_add_us(ktime_get(), vbif->cap->xin_halt_timeout); in _dpu_vbif_wait_for_xin_halt() 77 if (!vbif || !(vbif->cap->features & BIT(DPU_VBIF_QOS_OTLIM))) in _dpu_vbif_apply_dynamic_ot_limit() 88 tbl = params->rd ? &vbif->cap->dynamic_ot_rd_tbl : in _dpu_vbif_apply_dynamic_ot_limit() 89 &vbif->cap->dynamic_ot_wr_tbl; in _dpu_vbif_apply_dynamic_ot_limit() 116 if (!vbif || !vbif->cap) { in _dpu_vbif_get_ot_limit() 121 if (vbif->cap->default_ot_wr_limit && !params->rd) in _dpu_vbif_get_ot_limit() 122 ot_lim = vbif->cap->default_ot_wr_limit; in _dpu_vbif_get_ot_limit() 123 else if (vbif->cap->default_ot_rd_limit && params->rd) in _dpu_vbif_get_ot_limit() 124 ot_lim = vbif->cap->default_ot_rd_limit; in _dpu_vbif_get_ot_limit() [all …]
|
/drivers/infiniband/sw/rxe/ |
D | rxe_qp.c | 18 static int rxe_qp_chk_cap(struct rxe_dev *rxe, struct ib_qp_cap *cap, in rxe_qp_chk_cap() argument 21 if (cap->max_send_wr > rxe->attr.max_qp_wr) { in rxe_qp_chk_cap() 23 cap->max_send_wr, rxe->attr.max_qp_wr); in rxe_qp_chk_cap() 27 if (cap->max_send_sge > rxe->attr.max_send_sge) { in rxe_qp_chk_cap() 29 cap->max_send_sge, rxe->attr.max_send_sge); in rxe_qp_chk_cap() 34 if (cap->max_recv_wr > rxe->attr.max_qp_wr) { in rxe_qp_chk_cap() 36 cap->max_recv_wr, rxe->attr.max_qp_wr); in rxe_qp_chk_cap() 40 if (cap->max_recv_sge > rxe->attr.max_recv_sge) { in rxe_qp_chk_cap() 42 cap->max_recv_sge, rxe->attr.max_recv_sge); in rxe_qp_chk_cap() 47 if (cap->max_inline_data > rxe->max_inline_data) { in rxe_qp_chk_cap() [all …]
|