Home
last modified time | relevance | path

Searched refs:cap (Results 1 – 25 of 631) sorted by relevance

12345678910>>...26

/drivers/media/platform/rockchip/rkisp1/
Drkisp1-capture.c63 void (*config)(struct rkisp1_capture *cap);
64 void (*stop)(struct rkisp1_capture *cap);
65 void (*enable)(struct rkisp1_capture *cap);
66 void (*disable)(struct rkisp1_capture *cap);
67 void (*set_data_path)(struct rkisp1_capture *cap);
68 bool (*is_stopped)(struct rkisp1_capture *cap);
345 int rkisp1_cap_enum_mbus_codes(struct rkisp1_capture *cap, in rkisp1_cap_enum_mbus_codes() argument
348 const struct rkisp1_capture_fmt_cfg *fmts = cap->config->fmts; in rkisp1_cap_enum_mbus_codes()
356 for (i = 0; i < cap->config->fmt_size; i++) { in rkisp1_cap_enum_mbus_codes()
373 static void rkisp1_mi_config_ctrl(struct rkisp1_capture *cap) in rkisp1_mi_config_ctrl() argument
[all …]
/drivers/staging/greybus/
Dauthentication.c47 struct gb_cap *cap = container_of(kref, struct gb_cap, kref); in cap_kref_release() local
49 kfree(cap); in cap_kref_release()
57 static void put_cap(struct gb_cap *cap) in put_cap() argument
59 kref_put(&cap->kref, cap_kref_release); in put_cap()
65 struct gb_cap *cap; in get_cap() local
69 list_for_each_entry(cap, &cap_list, node) { in get_cap()
70 if (&cap->cdev == cdev) { in get_cap()
71 kref_get(&cap->kref); in get_cap()
76 cap = NULL; in get_cap()
81 return cap; in get_cap()
[all …]
/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_reg.h409 #define SXGBE_HW_FEAT_GMII(cap) ((cap & 0x00000002) >> 1) argument
410 #define SXGBE_HW_FEAT_VLAN_HASH_FILTER(cap) ((cap & 0x00000010) >> 4) argument
411 #define SXGBE_HW_FEAT_SMA(cap) ((cap & 0x00000020) >> 5) argument
412 #define SXGBE_HW_FEAT_PMT_TEMOTE_WOP(cap) ((cap & 0x00000040) >> 6) argument
413 #define SXGBE_HW_FEAT_PMT_MAGIC_PKT(cap) ((cap & 0x00000080) >> 7) argument
414 #define SXGBE_HW_FEAT_RMON(cap) ((cap & 0x00000100) >> 8) argument
415 #define SXGBE_HW_FEAT_ARP_OFFLOAD(cap) ((cap & 0x00000200) >> 9) argument
416 #define SXGBE_HW_FEAT_IEEE1500_2008(cap) ((cap & 0x00001000) >> 12) argument
417 #define SXGBE_HW_FEAT_EEE(cap) ((cap & 0x00002000) >> 13) argument
418 #define SXGBE_HW_FEAT_TX_CSUM_OFFLOAD(cap) ((cap & 0x00004000) >> 14) argument
[all …]
/drivers/vfio/pci/
Dvfio_pci_zdev.c24 struct vfio_device_info_cap_zpci_base cap = { in zpci_base_cap() local
36 return vfio_info_add_capability(caps, &cap.header, sizeof(cap)); in zpci_base_cap()
44 struct vfio_device_info_cap_zpci_group cap = { in zpci_group_cap() local
56 return vfio_info_add_capability(caps, &cap.header, sizeof(cap)); in zpci_group_cap()
64 struct vfio_device_info_cap_zpci_util *cap; in zpci_util_cap() local
65 int cap_size = sizeof(*cap) + CLP_UTIL_STR_LEN; in zpci_util_cap()
68 cap = kmalloc(cap_size, GFP_KERNEL); in zpci_util_cap()
69 if (!cap) in zpci_util_cap()
72 cap->header.id = VFIO_DEVICE_INFO_CAP_ZPCI_UTIL; in zpci_util_cap()
73 cap->header.version = 1; in zpci_util_cap()
[all …]
/drivers/iommu/intel/
Dcap_audit.h67 #define DO_CHECK_FEATURE_MISMATCH(a, b, cap, feature, MASK) \ argument
69 if (cap##_##feature(a) != cap##_##feature(b)) { \
70 intel_iommu_##cap##_sanity &= ~(MASK); \
75 #define CHECK_FEATURE_MISMATCH(a, b, cap, feature, MASK) \ argument
76 DO_CHECK_FEATURE_MISMATCH((a)->cap, (b)->cap, cap, feature, MASK)
78 #define CHECK_FEATURE_MISMATCH_HOTPLUG(b, cap, feature, MASK) \ argument
80 if (cap##_##feature(intel_iommu_##cap##_sanity)) \
81 DO_CHECK_FEATURE_MISMATCH(intel_iommu_##cap##_sanity, \
82 (b)->cap, cap, feature, MASK); \
85 #define MINIMAL_FEATURE_IOMMU(iommu, cap, MASK) \ argument
[all …]
Dcap_audit.c22 CHECK_FEATURE_MISMATCH(a, b, cap, pi_support, CAP_PI_MASK); in check_irq_capabilities()
29 MINIMAL_FEATURE_IOMMU(b, cap, CAP_MAMV_MASK); in check_dmar_capabilities()
30 MINIMAL_FEATURE_IOMMU(b, cap, CAP_NFR_MASK); in check_dmar_capabilities()
31 MINIMAL_FEATURE_IOMMU(b, cap, CAP_SLLPS_MASK); in check_dmar_capabilities()
32 MINIMAL_FEATURE_IOMMU(b, cap, CAP_FRO_MASK); in check_dmar_capabilities()
33 MINIMAL_FEATURE_IOMMU(b, cap, CAP_MGAW_MASK); in check_dmar_capabilities()
34 MINIMAL_FEATURE_IOMMU(b, cap, CAP_SAGAW_MASK); in check_dmar_capabilities()
35 MINIMAL_FEATURE_IOMMU(b, cap, CAP_NDOMS_MASK); in check_dmar_capabilities()
40 CHECK_FEATURE_MISMATCH(a, b, cap, 5lp_support, CAP_FL5LP_MASK); in check_dmar_capabilities()
41 CHECK_FEATURE_MISMATCH(a, b, cap, fl1gp_support, CAP_FL1GP_MASK); in check_dmar_capabilities()
[all …]
/drivers/pci/pcie/
Ddpc.c45 u16 *cap; in pci_save_dpc_state() local
54 cap = (u16 *)&save_state->cap.data[0]; in pci_save_dpc_state()
55 pci_read_config_word(dev, dev->dpc_cap + PCI_EXP_DPC_CTL, cap); in pci_save_dpc_state()
61 u16 *cap; in pci_restore_dpc_state() local
70 cap = (u16 *)&save_state->cap.data[0]; in pci_restore_dpc_state()
71 pci_write_config_word(dev, dev->dpc_cap + PCI_EXP_DPC_CTL, *cap); in pci_restore_dpc_state()
129 u16 cap = pdev->dpc_cap, status; in dpc_wait_rp_inactive() local
131 pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status); in dpc_wait_rp_inactive()
135 pci_read_config_word(pdev, cap + PCI_EXP_DPC_STATUS, &status); in dpc_wait_rp_inactive()
147 u16 cap; in dpc_reset_link() local
[all …]
Dptm.c53 u16 *cap; in pci_save_ptm_state() local
66 cap = (u16 *)&save_state->cap.data[0]; in pci_save_ptm_state()
67 pci_read_config_word(dev, ptm + PCI_PTM_CTRL, cap); in pci_save_ptm_state()
74 u16 *cap; in pci_restore_ptm_state() local
84 cap = (u16 *)&save_state->cap.data[0]; in pci_restore_ptm_state()
85 pci_write_config_word(dev, ptm + PCI_PTM_CTRL, *cap); in pci_restore_ptm_state()
91 u32 cap, ctrl; in pci_ptm_init() local
126 pci_read_config_dword(dev, pos + PCI_PTM_CAP, &cap); in pci_ptm_init()
127 local_clock = (cap & PCI_PTM_GRANULARITY_MASK) >> 8; in pci_ptm_init()
142 if (cap & PCI_PTM_CAP_ROOT) { in pci_ptm_init()
[all …]
/drivers/media/platform/qcom/venus/
Dhfi_parser.c14 typedef void (*func)(struct hfi_plat_caps *cap, const void *data,
19 struct hfi_plat_caps *caps = core->caps, *cap; in init_codecs() local
26 cap = &caps[core->codecs_count++]; in init_codecs()
27 cap->codec = BIT(bit); in init_codecs()
28 cap->domain = VIDC_SESSION_TYPE_DEC; in init_codecs()
29 cap->valid = false; in init_codecs()
33 cap = &caps[core->codecs_count++]; in init_codecs()
34 cap->codec = BIT(bit); in init_codecs()
35 cap->domain = VIDC_SESSION_TYPE_ENC; in init_codecs()
36 cap->valid = false; in init_codecs()
[all …]
/drivers/infiniband/core/
Duverbs_std_types_qp.c65 struct ib_uverbs_qp_cap *cap, bool req) in set_caps() argument
68 attr->cap.max_send_wr = cap->max_send_wr; in set_caps()
69 attr->cap.max_recv_wr = cap->max_recv_wr; in set_caps()
70 attr->cap.max_send_sge = cap->max_send_sge; in set_caps()
71 attr->cap.max_recv_sge = cap->max_recv_sge; in set_caps()
72 attr->cap.max_inline_data = cap->max_inline_data; in set_caps()
74 cap->max_send_wr = attr->cap.max_send_wr; in set_caps()
75 cap->max_recv_wr = attr->cap.max_recv_wr; in set_caps()
76 cap->max_send_sge = attr->cap.max_send_sge; in set_caps()
77 cap->max_recv_sge = attr->cap.max_recv_sge; in set_caps()
[all …]
/drivers/thunderbolt/
Ddebugfs.c255 struct tb_port *port, unsigned int cap, in cap_show_by_dw() argument
264 ret = tb_port_read(port, &data, TB_CFG_PORT, cap + offset + i, 1); in cap_show_by_dw()
266 ret = tb_sw_read(sw, &data, TB_CFG_SWITCH, cap + offset + i, 1); in cap_show_by_dw()
268 seq_printf(s, "0x%04x <not accessible>\n", cap + offset + i); in cap_show_by_dw()
272 seq_printf(s, "0x%04x %4d 0x%02x 0x%02x 0x%08x\n", cap + offset + i, in cap_show_by_dw()
278 struct tb_port *port, unsigned int cap, u8 cap_id, in cap_show() argument
288 ret = tb_port_read(port, data, TB_CFG_PORT, cap + offset, in cap_show()
291 ret = tb_sw_read(sw, data, TB_CFG_SWITCH, cap + offset, dwords); in cap_show()
293 cap_show_by_dw(s, sw, port, cap, offset, cap_id, vsec_id, length); in cap_show()
299 cap + offset + i, offset + i, in cap_show()
[all …]
Dlc.c53 int cap, ret; in tb_lc_set_port_configured() local
58 cap = find_port_lc_cap(port); in tb_lc_set_port_configured()
59 if (cap < 0) in tb_lc_set_port_configured()
60 return cap; in tb_lc_set_port_configured()
62 ret = tb_sw_read(sw, &ctrl, TB_CFG_SWITCH, cap + TB_LC_SX_CTRL, 1); in tb_lc_set_port_configured()
82 return tb_sw_write(sw, &ctrl, TB_CFG_SWITCH, cap + TB_LC_SX_CTRL, 1); in tb_lc_set_port_configured()
111 int cap, ret; in tb_lc_set_xdomain_configured() local
116 cap = find_port_lc_cap(port); in tb_lc_set_xdomain_configured()
117 if (cap < 0) in tb_lc_set_xdomain_configured()
118 return cap; in tb_lc_set_xdomain_configured()
[all …]
Dcap.c87 static int __tb_port_find_cap(struct tb_port *port, enum tb_port_cap cap) in __tb_port_find_cap() argument
103 if (header.basic.cap == cap) in __tb_port_find_cap()
119 int tb_port_find_cap(struct tb_port *port, enum tb_port_cap cap) in tb_port_find_cap() argument
127 ret = __tb_port_find_cap(port, cap); in tb_port_find_cap()
157 switch (header.basic.cap) { in tb_switch_next_cap()
171 header.basic.cap, offset); in tb_switch_next_cap()
188 int tb_switch_find_cap(struct tb_switch *sw, enum tb_switch_cap cap) in tb_switch_find_cap() argument
204 if (header.basic.cap == cap) in tb_switch_find_cap()
237 if (header.extended_short.cap == TB_SWITCH_CAP_VSE && in tb_switch_find_vse_cap()
Dtmu.c46 sw->tmu.cap + TMU_RTR_CS_0, 1); in tb_switch_tmu_ucap_supported()
59 sw->tmu.cap + TMU_RTR_CS_3, 1); in tb_switch_tmu_rate_read()
73 sw->tmu.cap + TMU_RTR_CS_3, 1); in tb_switch_tmu_rate_write()
81 sw->tmu.cap + TMU_RTR_CS_3, 1); in tb_switch_tmu_rate_write()
137 sw->tmu.cap + TMU_RTR_CS_0, 1); in tb_switch_tmu_set_time_disruption()
147 sw->tmu.cap + TMU_RTR_CS_0, 1); in tb_switch_tmu_set_time_disruption()
168 sw->tmu.cap = ret; in tb_switch_tmu_init()
171 int cap; in tb_switch_tmu_init() local
173 cap = tb_port_find_cap(port, TB_PORT_CAP_TIME1); in tb_switch_tmu_init()
174 if (cap > 0) in tb_switch_tmu_init()
[all …]
/drivers/char/tpm/
Dtpm-sysfs.c85 cap_t cap; in pcrs_show() local
94 if (tpm1_getcap(chip, TPM_CAP_PROP_PCR, &cap, in pcrs_show()
96 sizeof(cap.num_pcrs))) { in pcrs_show()
101 num_pcrs = be32_to_cpu(cap.num_pcrs); in pcrs_show()
122 cap_t cap; in enabled_show() local
127 if (tpm1_getcap(chip, TPM_CAP_FLAG_PERM, &cap, in enabled_show()
129 sizeof(cap.perm_flags))) in enabled_show()
132 rc = sprintf(buf, "%d\n", !cap.perm_flags.disable); in enabled_show()
144 cap_t cap; in active_show() local
149 if (tpm1_getcap(chip, TPM_CAP_FLAG_PERM, &cap, in active_show()
[all …]
/drivers/infiniband/hw/hfi1/
Dcommon.h45 #define HFI1_CAP_KSET(cap) ({ hfi1_cap_mask |= HFI1_CAP_##cap; hfi1_cap_mask; }) argument
46 #define HFI1_CAP_KCLEAR(cap) \ argument
48 hfi1_cap_mask &= ~HFI1_CAP_##cap; \
51 #define HFI1_CAP_USET(cap) \ argument
53 hfi1_cap_mask |= (HFI1_CAP_##cap << HFI1_CAP_USER_SHIFT); \
56 #define HFI1_CAP_UCLEAR(cap) \ argument
58 hfi1_cap_mask &= ~(HFI1_CAP_##cap << HFI1_CAP_USER_SHIFT); \
61 #define HFI1_CAP_SET(cap) \ argument
63 hfi1_cap_mask |= (HFI1_CAP_##cap | (HFI1_CAP_##cap << \
67 #define HFI1_CAP_CLEAR(cap) \ argument
[all …]
/drivers/platform/surface/aggregator/
Dssh_parser.h27 size_t cap; member
39 static inline void sshp_buf_init(struct sshp_buf *buf, u8 *ptr, size_t cap) in sshp_buf_init() argument
43 buf->cap = cap; in sshp_buf_init()
57 static inline int sshp_buf_alloc(struct sshp_buf *buf, size_t cap, gfp_t flags) in sshp_buf_alloc() argument
61 ptr = kzalloc(cap, flags); in sshp_buf_alloc()
65 sshp_buf_init(buf, ptr, cap); in sshp_buf_alloc()
82 buf->cap = 0; in sshp_buf_free()
116 n = kfifo_out(fifo, buf->ptr + buf->len, buf->cap - buf->len); in sshp_buf_read_from_fifo()
/drivers/net/ethernet/netronome/nfp/bpf/
Dmain.c29 return nn->cap & NFP_NET_CFG_CTRL_BPF && in nfp_net_ebpf_capable()
209 struct nfp_bpf_cap_tlv_adjust_head __iomem *cap = value; in nfp_bpf_parse_cap_adjust_head() local
212 if (length < sizeof(*cap)) { in nfp_bpf_parse_cap_adjust_head()
217 bpf->adjust_head.flags = readl(&cap->flags); in nfp_bpf_parse_cap_adjust_head()
218 bpf->adjust_head.off_min = readl(&cap->off_min); in nfp_bpf_parse_cap_adjust_head()
219 bpf->adjust_head.off_max = readl(&cap->off_max); in nfp_bpf_parse_cap_adjust_head()
220 bpf->adjust_head.guaranteed_sub = readl(&cap->guaranteed_sub); in nfp_bpf_parse_cap_adjust_head()
221 bpf->adjust_head.guaranteed_add = readl(&cap->guaranteed_add); in nfp_bpf_parse_cap_adjust_head()
240 struct nfp_bpf_cap_tlv_func __iomem *cap = value; in nfp_bpf_parse_cap_func() local
242 if (length < sizeof(*cap)) { in nfp_bpf_parse_cap_func()
[all …]
/drivers/media/platform/exynos4-is/
Dcommon.c37 void __fimc_vidioc_querycap(struct device *dev, struct v4l2_capability *cap) in __fimc_vidioc_querycap() argument
39 strscpy(cap->driver, dev->driver->name, sizeof(cap->driver)); in __fimc_vidioc_querycap()
40 strscpy(cap->card, dev->driver->name, sizeof(cap->card)); in __fimc_vidioc_querycap()
41 snprintf(cap->bus_info, sizeof(cap->bus_info), in __fimc_vidioc_querycap()
/drivers/pci/
Dpci.c410 u8 pos, int cap, int *ttl) in __pci_find_next_cap_ttl() argument
426 if (id == cap) in __pci_find_next_cap_ttl()
434 u8 pos, int cap) in __pci_find_next_cap() argument
438 return __pci_find_next_cap_ttl(bus, devfn, pos, cap, &ttl); in __pci_find_next_cap()
441 u8 pci_find_next_capability(struct pci_dev *dev, u8 pos, int cap) in pci_find_next_capability() argument
444 pos + PCI_CAP_LIST_NEXT, cap); in pci_find_next_capability()
487 u8 pci_find_capability(struct pci_dev *dev, int cap) in pci_find_capability() argument
493 pos = __pci_find_next_cap(dev->bus, dev->devfn, pos, cap); in pci_find_capability()
512 u8 pci_bus_find_capability(struct pci_bus *bus, unsigned int devfn, int cap) in pci_bus_find_capability() argument
520 pos = __pci_find_next_cap(bus, devfn, pos, cap); in pci_bus_find_capability()
[all …]
/drivers/usb/typec/ucsi/
Ducsi.c259 override = !!(con->ucsi->cap.features & UCSI_CAP_ALT_MODE_OVERRIDE); in ucsi_register_altmode()
352 max_altmodes = con->ucsi->cap.num_alt_modes; in ucsi_register_altmodes_nvidia()
423 if (!(con->ucsi->cap.features & UCSI_CAP_ALT_MODE_DETAILS)) in ucsi_register_altmodes()
433 max_altmodes = con->ucsi->cap.num_alt_modes; in ucsi_register_altmodes()
1053 struct typec_capability *cap = &con->typec_cap; in ucsi_register_port() local
1054 enum typec_accessory *accessory = cap->accessory; in ucsi_register_port()
1065 cap->fwnode = ucsi_find_fwnode(con); in ucsi_register_port()
1066 con->usb_role_sw = fwnode_usb_role_switch_get(cap->fwnode); in ucsi_register_port()
1080 ret = ucsi_send_command(ucsi, command, &con->cap, sizeof(con->cap)); in ucsi_register_port()
1084 if (con->cap.op_mode & UCSI_CONCAP_OPMODE_DRP) in ucsi_register_port()
[all …]
/drivers/cxl/core/
Dregs.c37 int cap, cap_count; in cxl_probe_component_regs() local
59 for (cap = 1; cap <= cap_count; cap++) { in cxl_probe_component_regs()
66 hdr = readl(base + cap * 0x4); in cxl_probe_component_regs()
106 int cap, cap_count; in cxl_probe_device_regs() local
118 for (cap = 1; cap <= cap_count; cap++) { in cxl_probe_device_regs()
123 readl(base + cap * 0x10)); in cxl_probe_device_regs()
124 offset = readl(base + cap * 0x10 + 0x4); in cxl_probe_device_regs()
125 length = readl(base + cap * 0x10 + 0x8); in cxl_probe_device_regs()
/drivers/net/arcnet/
Dcapmode.c77 memcpy(pktbuf, pkthdrbuf, ARC_HDR_SIZE + sizeof(pkt->soft.cap.proto)); in rx()
78 memcpy(pktbuf + ARC_HDR_SIZE + sizeof(pkt->soft.cap.proto) + sizeof(int), in rx()
79 pkthdrbuf + ARC_HDR_SIZE + sizeof(pkt->soft.cap.proto), in rx()
80 sizeof(struct archdr) - ARC_HDR_SIZE - sizeof(pkt->soft.cap.proto)); in rx()
107 *((int *)&pkt->soft.cap.cookie[0])); in build_header()
148 *((int *)&pkt->soft.cap.cookie[0])); in prepare_tx()
171 lp->hw.copy_to_card(dev, bufnum, ofs, &pkt->soft.cap.proto, in prepare_tx()
172 sizeof(pkt->soft.cap.proto)); in prepare_tx()
178 ((unsigned char *)&pkt->soft.cap.mes), length - 1); in prepare_tx()
212 ackpkt->soft.cap.proto = 0; /* using protocol 0 for acknowledge */ in ack_tx()
[all …]
/drivers/gpu/drm/msm/disp/dpu1/
Ddpu_vbif.c34 if (!vbif || !vbif->cap || !vbif->ops.get_halt_ctrl) { in _dpu_vbif_wait_for_xin_halt()
39 timeout = ktime_add_us(ktime_get(), vbif->cap->xin_halt_timeout); in _dpu_vbif_wait_for_xin_halt()
77 if (!vbif || !(vbif->cap->features & BIT(DPU_VBIF_QOS_OTLIM))) in _dpu_vbif_apply_dynamic_ot_limit()
88 tbl = params->rd ? &vbif->cap->dynamic_ot_rd_tbl : in _dpu_vbif_apply_dynamic_ot_limit()
89 &vbif->cap->dynamic_ot_wr_tbl; in _dpu_vbif_apply_dynamic_ot_limit()
116 if (!vbif || !vbif->cap) { in _dpu_vbif_get_ot_limit()
121 if (vbif->cap->default_ot_wr_limit && !params->rd) in _dpu_vbif_get_ot_limit()
122 ot_lim = vbif->cap->default_ot_wr_limit; in _dpu_vbif_get_ot_limit()
123 else if (vbif->cap->default_ot_rd_limit && params->rd) in _dpu_vbif_get_ot_limit()
124 ot_lim = vbif->cap->default_ot_rd_limit; in _dpu_vbif_get_ot_limit()
[all …]
/drivers/infiniband/sw/rxe/
Drxe_qp.c18 static int rxe_qp_chk_cap(struct rxe_dev *rxe, struct ib_qp_cap *cap, in rxe_qp_chk_cap() argument
21 if (cap->max_send_wr > rxe->attr.max_qp_wr) { in rxe_qp_chk_cap()
23 cap->max_send_wr, rxe->attr.max_qp_wr); in rxe_qp_chk_cap()
27 if (cap->max_send_sge > rxe->attr.max_send_sge) { in rxe_qp_chk_cap()
29 cap->max_send_sge, rxe->attr.max_send_sge); in rxe_qp_chk_cap()
34 if (cap->max_recv_wr > rxe->attr.max_qp_wr) { in rxe_qp_chk_cap()
36 cap->max_recv_wr, rxe->attr.max_qp_wr); in rxe_qp_chk_cap()
40 if (cap->max_recv_sge > rxe->attr.max_recv_sge) { in rxe_qp_chk_cap()
42 cap->max_recv_sge, rxe->attr.max_recv_sge); in rxe_qp_chk_cap()
47 if (cap->max_inline_data > rxe->max_inline_data) { in rxe_qp_chk_cap()
[all …]

12345678910>>...26