Home
last modified time | relevance | path

Searched refs:event (Results 1 – 25 of 1163) sorted by relevance

12345678910>>...47

/drivers/gpu/drm/nouveau/nvkm/core/
Devent.c26 nvkm_event_put(struct nvkm_event *event, u32 types, int index) in nvkm_event_put() argument
28 assert_spin_locked(&event->refs_lock); in nvkm_event_put()
31 if (--event->refs[index * event->types_nr + type] == 0) { in nvkm_event_put()
32 if (event->func->fini) in nvkm_event_put()
33 event->func->fini(event, 1 << type, index); in nvkm_event_put()
39 nvkm_event_get(struct nvkm_event *event, u32 types, int index) in nvkm_event_get() argument
41 assert_spin_locked(&event->refs_lock); in nvkm_event_get()
44 if (++event->refs[index * event->types_nr + type] == 1) { in nvkm_event_get()
45 if (event->func->init) in nvkm_event_get()
46 event->func->init(event, 1 << type, index); in nvkm_event_get()
[all …]
Dnotify.c31 nvkm_event_put(notify->event, notify->types, notify->index); in nvkm_notify_put_locked()
37 struct nvkm_event *event = notify->event; in nvkm_notify_put() local
39 if (likely(event) && in nvkm_notify_put()
41 spin_lock_irqsave(&event->refs_lock, flags); in nvkm_notify_put()
43 spin_unlock_irqrestore(&event->refs_lock, flags); in nvkm_notify_put()
53 nvkm_event_get(notify->event, notify->types, notify->index); in nvkm_notify_get_locked()
59 struct nvkm_event *event = notify->event; in nvkm_notify_get() local
61 if (likely(event) && in nvkm_notify_get()
63 spin_lock_irqsave(&event->refs_lock, flags); in nvkm_notify_get()
65 spin_unlock_irqrestore(&event->refs_lock, flags); in nvkm_notify_get()
[all …]
/drivers/md/
Ddm-uevent.c39 static void dm_uevent_free(struct dm_uevent *event) in dm_uevent_free() argument
41 kmem_cache_free(_dm_event_cache, event); in dm_uevent_free()
46 struct dm_uevent *event; in dm_uevent_alloc() local
48 event = kmem_cache_zalloc(_dm_event_cache, GFP_ATOMIC); in dm_uevent_alloc()
49 if (!event) in dm_uevent_alloc()
52 INIT_LIST_HEAD(&event->elist); in dm_uevent_alloc()
53 event->md = md; in dm_uevent_alloc()
55 return event; in dm_uevent_alloc()
65 struct dm_uevent *event; in dm_build_path_uevent() local
67 event = dm_uevent_alloc(md); in dm_build_path_uevent()
[all …]
/drivers/firmware/
Darm_sdei.c76 struct sdei_event *event; member
81 #define CROSSCALL_INIT(arg, event) \ argument
83 arg.event = event; \
89 struct sdei_event *event) in sdei_do_local_call() argument
93 CROSSCALL_INIT(arg, event); in sdei_do_local_call()
100 struct sdei_event *event) in sdei_do_cross_call() argument
104 CROSSCALL_INIT(arg, event); in sdei_do_cross_call()
190 static int sdei_api_event_get_info(u32 event, u32 info, u64 *result) in sdei_api_event_get_info() argument
192 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_GET_INFO, event, info, 0, in sdei_api_event_get_info()
202 struct sdei_event *event; in sdei_event_create() local
[all …]
/drivers/perf/hisilicon/
Dhisi_uncore_pmu.c66 static bool hisi_validate_event_group(struct perf_event *event) in hisi_validate_event_group() argument
68 struct perf_event *sibling, *leader = event->group_leader; in hisi_validate_event_group()
69 struct hisi_pmu *hisi_pmu = to_hisi_pmu(event->pmu); in hisi_validate_event_group()
78 if (leader->pmu != event->pmu) in hisi_validate_event_group()
82 if (leader != event) in hisi_validate_event_group()
86 for_each_sibling_event(sibling, event->group_leader) { in hisi_validate_event_group()
89 if (sibling->pmu != event->pmu) in hisi_validate_event_group()
105 int hisi_uncore_pmu_get_event_idx(struct perf_event *event) in hisi_uncore_pmu_get_event_idx() argument
107 struct hisi_pmu *hisi_pmu = to_hisi_pmu(event->pmu); in hisi_uncore_pmu_get_event_idx()
132 int hisi_uncore_pmu_event_init(struct perf_event *event) in hisi_uncore_pmu_event_init() argument
[all …]
/drivers/perf/
Darm_pmu.c108 static inline u64 arm_pmu_event_max_period(struct perf_event *event) in arm_pmu_event_max_period() argument
110 if (event->hw.flags & ARMPMU_EVT_64BIT) in arm_pmu_event_max_period()
170 armpmu_map_event(struct perf_event *event, in armpmu_map_event() argument
178 u64 config = event->attr.config; in armpmu_map_event()
179 int type = event->attr.type; in armpmu_map_event()
181 if (type == event->pmu->type) in armpmu_map_event()
196 int armpmu_event_set_period(struct perf_event *event) in armpmu_event_set_period() argument
198 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period()
199 struct hw_perf_event *hwc = &event->hw; in armpmu_event_set_period()
205 max_period = arm_pmu_event_max_period(event); in armpmu_event_set_period()
[all …]
Dqcom_l3_pmu.c136 static inline u32 get_event_type(struct perf_event *event) in get_event_type() argument
138 return (event->attr.config) & L3_EVTYPE_MASK; in get_event_type()
141 static inline bool event_uses_long_counter(struct perf_event *event) in event_uses_long_counter() argument
143 return !!(event->attr.config & BIT_ULL(L3_EVENT_LC_BIT)); in event_uses_long_counter()
146 static inline int event_num_counters(struct perf_event *event) in event_num_counters() argument
148 return event_uses_long_counter(event) ? 2 : 1; in event_num_counters()
175 void (*start)(struct perf_event *event);
177 void (*stop)(struct perf_event *event, int flags);
179 void (*update)(struct perf_event *event);
193 static void qcom_l3_cache__64bit_counter_start(struct perf_event *event) in qcom_l3_cache__64bit_counter_start() argument
[all …]
Darm_dsu_pmu.c177 DSU_FORMAT_ATTR(event, "config:0-31"),
249 static inline u64 dsu_pmu_read_counter(struct perf_event *event) in dsu_pmu_read_counter() argument
253 struct dsu_pmu *dsu_pmu = to_dsu_pmu(event->pmu); in dsu_pmu_read_counter()
254 int idx = event->hw.idx; in dsu_pmu_read_counter()
261 dev_err(event->pmu->dev, in dsu_pmu_read_counter()
276 static void dsu_pmu_write_counter(struct perf_event *event, u64 val) in dsu_pmu_write_counter() argument
279 struct dsu_pmu *dsu_pmu = to_dsu_pmu(event->pmu); in dsu_pmu_write_counter()
280 int idx = event->hw.idx; in dsu_pmu_write_counter()
287 dev_err(event->pmu->dev, in dsu_pmu_write_counter()
301 struct perf_event *event) in dsu_pmu_get_event_idx() argument
[all …]
Dthunderx2_pmu.c111 void (*init_cntr_base)(struct perf_event *event,
113 void (*stop_event)(struct perf_event *event);
114 void (*start_event)(struct perf_event *event, int flags);
137 TX2_PMU_FORMAT_ATTR(event, event, "config:0-4");
138 TX2_PMU_FORMAT_ATTR(event_ccpi2, event, "config:0-9");
322 static void init_cntr_base_l3c(struct perf_event *event, in init_cntr_base_l3c() argument
325 struct hw_perf_event *hwc = &event->hw; in init_cntr_base_l3c()
328 tx2_pmu = pmu_to_tx2_pmu(event->pmu); in init_cntr_base_l3c()
333 + L3C_COUNTER_CTL + (8 * GET_COUNTERID(event, cmask)); in init_cntr_base_l3c()
335 + L3C_COUNTER_DATA + (8 * GET_COUNTERID(event, cmask)); in init_cntr_base_l3c()
[all …]
Dqcom_l2_pmu.c75 #define L2_EVT_CODE(event) (((event) & L2_EVT_CODE_MASK) >> L2_EVT_CODE_SHIFT) argument
76 #define L2_EVT_GROUP(event) (((event) & L2_EVT_GRP_MASK) >> L2_EVT_GRP_SHIFT) argument
297 static void l2_cache_event_update(struct perf_event *event) in l2_cache_event_update() argument
299 struct hw_perf_event *hwc = &event->hw; in l2_cache_event_update()
316 local64_add(delta, &event->count); in l2_cache_event_update()
340 struct perf_event *event) in l2_cache_get_event_idx() argument
342 struct hw_perf_event *hwc = &event->hw; in l2_cache_get_event_idx()
375 struct perf_event *event) in l2_cache_clear_event_idx() argument
377 struct hw_perf_event *hwc = &event->hw; in l2_cache_clear_event_idx()
397 struct perf_event *event = cluster->events[idx]; in l2_cache_handle_irq() local
[all …]
Dfsl_imx8_ddr_perf.c220 PMU_FORMAT_ATTR(event, "config:0-7");
244 static bool ddr_perf_is_filtered(struct perf_event *event) in ddr_perf_is_filtered() argument
246 return event->attr.config == 0x41 || event->attr.config == 0x42; in ddr_perf_is_filtered()
249 static u32 ddr_perf_filter_val(struct perf_event *event) in ddr_perf_filter_val() argument
251 return event->attr.config1; in ddr_perf_filter_val()
264 static bool ddr_perf_is_enhanced_filtered(struct perf_event *event) in ddr_perf_is_enhanced_filtered() argument
267 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_is_enhanced_filtered()
271 ddr_perf_is_filtered(event); in ddr_perf_is_enhanced_filtered()
274 static u32 ddr_perf_alloc_counter(struct ddr_pmu *pmu, int event) in ddr_perf_alloc_counter() argument
283 if (event == EVENT_CYCLES_ID) { in ddr_perf_alloc_counter()
[all …]
Darm-ccn.c154 struct perf_event *event; member
235 static CCN_FORMAT_ATTR(event, "config:16-23");
268 u32 event; member
287 .type = CCN_TYPE_MN, .event = CCN_EVENT_WATCHPOINT, \
293 .event = CCN_EVENT_WATCHPOINT, .num_ports = CCN_NUM_XP_PORTS, \
298 .event = CCN_EVENT_WATCHPOINT, .num_ports = CCN_NUM_XP_PORTS, \
302 .type = CCN_TYPE_HNF, .event = _event, }
305 .type = CCN_TYPE_XP, .event = _event, \
314 .type = CCN_TYPE_RNI_3P, .event = _event, }
317 .type = CCN_TYPE_SBAS, .event = _event, }
[all …]
Darm-cci.c193 #define CCI400_PMU_EVENT_SOURCE(event) \ argument
194 ((event >> CCI400_PMU_EVENT_SOURCE_SHIFT) & \
196 #define CCI400_PMU_EVENT_CODE(event) \ argument
197 ((event >> CCI400_PMU_EVENT_CODE_SHIFT) & CCI400_PMU_EVENT_CODE_MASK)
217 CCI_FORMAT_EXT_ATTR_ENTRY(event, "config:0-4"),
432 #define CCI5xx_PMU_EVENT_SOURCE(event) \ argument
433 ((event >> CCI5xx_PMU_EVENT_SOURCE_SHIFT) & CCI5xx_PMU_EVENT_SOURCE_MASK)
434 #define CCI5xx_PMU_EVENT_CODE(event) \ argument
435 ((event >> CCI5xx_PMU_EVENT_CODE_SHIFT) & CCI5xx_PMU_EVENT_CODE_MASK)
453 CCI_FORMAT_EXT_ATTR_ENTRY(event, "config:0-4"),
[all …]
Darm-cmn.c131 #define CMN_EVENT_TYPE(event) FIELD_GET(CMN_CONFIG_TYPE, (event)->attr.config) argument
132 #define CMN_EVENT_EVENTID(event) FIELD_GET(CMN_CONFIG_EVENTID, (event)->attr.config) argument
133 #define CMN_EVENT_OCCUPID(event) FIELD_GET(CMN_CONFIG_OCCUPID, (event)->attr.config) argument
134 #define CMN_EVENT_BYNODEID(event) FIELD_GET(CMN_CONFIG_BYNODEID, (event)->attr.config) argument
135 #define CMN_EVENT_NODEID(event) FIELD_GET(CMN_CONFIG_NODEID, (event)->attr.config) argument
145 #define CMN_EVENT_WP_COMBINE(event) FIELD_GET(CMN_CONFIG_WP_COMBINE, (event)->attr.config) argument
146 #define CMN_EVENT_WP_DEV_SEL(event) FIELD_GET(CMN_CONFIG_WP_DEV_SEL, (event)->attr.config) argument
147 #define CMN_EVENT_WP_CHN_SEL(event) FIELD_GET(CMN_CONFIG_WP_CHN_SEL, (event)->attr.config) argument
148 #define CMN_EVENT_WP_GRP(event) FIELD_GET(CMN_CONFIG_WP_GRP, (event)->attr.config) argument
149 #define CMN_EVENT_WP_EXCLUSIVE(event) FIELD_GET(CMN_CONFIG_WP_EXCLUSIVE, (event)->attr.config) argument
[all …]
/drivers/gpu/drm/amd/amdgpu/
Damdgpu_pmu.c46 static int amdgpu_perf_event_init(struct perf_event *event) in amdgpu_perf_event_init() argument
48 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_event_init()
51 if (event->attr.type != event->pmu->type) in amdgpu_perf_event_init()
55 hwc->config = event->attr.config; in amdgpu_perf_event_init()
61 static void amdgpu_perf_start(struct perf_event *event, int flags) in amdgpu_perf_start() argument
63 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_start()
64 struct amdgpu_pmu_entry *pe = container_of(event->pmu, in amdgpu_perf_start()
85 perf_event_update_userpage(event); in amdgpu_perf_start()
90 static void amdgpu_perf_read(struct perf_event *event) in amdgpu_perf_read() argument
92 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_read()
[all …]
/drivers/oprofile/
Doprofile_perf.c19 unsigned long event; member
36 static void op_overflow_handler(struct perf_event *event, in op_overflow_handler() argument
43 if (per_cpu(perf_events, cpu)[id] == event) in op_overflow_handler()
69 attr->config = counter_config[i].event; in op_perf_setup()
75 static int op_create_counter(int cpu, int event) in op_create_counter() argument
79 if (!counter_config[event].enabled || per_cpu(perf_events, cpu)[event]) in op_create_counter()
82 pevent = perf_event_create_kernel_counter(&counter_config[event].attr, in op_create_counter()
92 event, cpu); in op_create_counter()
96 per_cpu(perf_events, cpu)[event] = pevent; in op_create_counter()
101 static void op_destroy_counter(int cpu, int event) in op_destroy_counter() argument
[all …]
Dnmi_timer_int.c29 static void nmi_timer_callback(struct perf_event *event, in nmi_timer_callback() argument
33 event->hw.interrupts = 0; /* don't throttle interrupts */ in nmi_timer_callback()
39 struct perf_event *event = per_cpu(nmi_timer_events, cpu); in nmi_timer_start_cpu() local
41 if (!event) { in nmi_timer_start_cpu()
42 event = perf_event_create_kernel_counter(&nmi_timer_attr, cpu, NULL, in nmi_timer_start_cpu()
44 if (IS_ERR(event)) in nmi_timer_start_cpu()
45 return PTR_ERR(event); in nmi_timer_start_cpu()
46 per_cpu(nmi_timer_events, cpu) = event; in nmi_timer_start_cpu()
49 if (event && ctr_running) in nmi_timer_start_cpu()
50 perf_event_enable(event); in nmi_timer_start_cpu()
[all …]
/drivers/net/wireless/broadcom/brcm80211/brcmfmac/
Dfweh.c85 struct brcmf_fweh_queue_item *event) in brcmf_fweh_queue_event() argument
90 list_add_tail(&event->q, &fweh->event_q); in brcmf_fweh_queue_event()
194 struct brcmf_fweh_queue_item *event = NULL; in brcmf_fweh_dequeue_event() local
199 event = list_first_entry(&fweh->event_q, in brcmf_fweh_dequeue_event()
201 list_del(&event->q); in brcmf_fweh_dequeue_event()
205 return event; in brcmf_fweh_dequeue_event()
218 struct brcmf_fweh_queue_item *event; in brcmf_fweh_event_worker() local
226 while ((event = brcmf_fweh_dequeue_event(fweh))) { in brcmf_fweh_event_worker()
228 brcmf_fweh_event_name(event->code), event->code, in brcmf_fweh_event_worker()
229 event->emsg.ifidx, event->emsg.bsscfgidx, in brcmf_fweh_event_worker()
[all …]
/drivers/net/ethernet/huawei/hinic/
Dhinic_devlink.c331 struct hinic_fault_event *event) in chip_fault_show() argument
338 fault_level = (event->event.chip.err_level < FAULT_LEVEL_MAX) ? in chip_fault_show()
339 event->event.chip.err_level : FAULT_LEVEL_MAX; in chip_fault_show()
342 (u32)event->event.chip.func_id); in chip_fault_show()
347 err = devlink_fmsg_u8_pair_put(fmsg, "module_id", event->event.chip.node_id); in chip_fault_show()
351 err = devlink_fmsg_u32_pair_put(fmsg, "err_type", (u32)event->event.chip.err_type); in chip_fault_show()
360 event->event.chip.err_csr_addr); in chip_fault_show()
365 event->event.chip.err_csr_value); in chip_fault_show()
373 struct hinic_fault_event *event) in fault_report_show() argument
381 fault_type = (event->type < FAULT_TYPE_MAX) ? event->type : FAULT_TYPE_MAX; in fault_report_show()
[all …]
/drivers/input/
Dinput-compat.c15 struct input_event *event) in input_event_from_user() argument
24 event->input_event_sec = compat_event.sec; in input_event_from_user()
25 event->input_event_usec = compat_event.usec; in input_event_from_user()
26 event->type = compat_event.type; in input_event_from_user()
27 event->code = compat_event.code; in input_event_from_user()
28 event->value = compat_event.value; in input_event_from_user()
31 if (copy_from_user(event, buffer, sizeof(struct input_event))) in input_event_from_user()
39 const struct input_event *event) in input_event_to_user() argument
44 compat_event.sec = event->input_event_sec; in input_event_to_user()
45 compat_event.usec = event->input_event_usec; in input_event_to_user()
[all …]
/drivers/char/tpm/eventlog/
Dtpm1.c77 struct tcpa_event *event; in tpm1_bios_measurements_start() local
83 event = addr; in tpm1_bios_measurements_start()
90 do_endian_conversion(event->event_size); in tpm1_bios_measurements_start()
92 do_endian_conversion(event->event_type); in tpm1_bios_measurements_start()
111 struct tcpa_event *event = v; in tpm1_bios_measurements_next() local
119 converted_event_size = do_endian_conversion(event->event_size); in tpm1_bios_measurements_next()
127 event = v; in tpm1_bios_measurements_next()
129 converted_event_size = do_endian_conversion(event->event_size); in tpm1_bios_measurements_next()
130 converted_event_type = do_endian_conversion(event->event_type); in tpm1_bios_measurements_next()
143 static int get_event_name(char *dest, struct tcpa_event *event, in get_event_name() argument
[all …]
Dtpm2.c36 static size_t calc_tpm2_event_size(struct tcg_pcr_event2_head *event, in calc_tpm2_event_size() argument
39 return __calc_tpm2_event_size(event, event_header, false); in calc_tpm2_event_size()
49 struct tcg_pcr_event2_head *event; in tpm2_bios_measurements_start() local
54 size = struct_size(event_header, event, event_header->event_size); in tpm2_bios_measurements_start()
67 event = addr; in tpm2_bios_measurements_start()
68 size = calc_tpm2_event_size(event, event_header); in tpm2_bios_measurements_start()
74 event = addr; in tpm2_bios_measurements_start()
75 size = calc_tpm2_event_size(event, event_header); in tpm2_bios_measurements_start()
89 struct tcg_pcr_event2_head *event; in tpm2_bios_measurements_next() local
100 event_size = struct_size(event_header, event, in tpm2_bios_measurements_next()
[all …]
/drivers/acpi/
Devged.c58 struct acpi_ged_event *event = data; in acpi_ged_irq_handler() local
61 acpi_ret = acpi_execute_simple_method(event->handle, NULL, event->gsi); in acpi_ged_irq_handler()
63 dev_err_once(event->dev, "IRQ method execution failed\n"); in acpi_ged_irq_handler()
71 struct acpi_ged_event *event; in acpi_ged_request_interrupt() local
118 event = devm_kzalloc(dev, sizeof(*event), GFP_KERNEL); in acpi_ged_request_interrupt()
119 if (!event) in acpi_ged_request_interrupt()
122 event->gsi = gsi; in acpi_ged_request_interrupt()
123 event->dev = dev; in acpi_ged_request_interrupt()
124 event->irq = irq; in acpi_ged_request_interrupt()
125 event->handle = evt_handle; in acpi_ged_request_interrupt()
[all …]
/drivers/fpga/
Ddfl-fme-perf.c170 int (*event_init)(struct fme_perf_priv *priv, u32 event, u32 portid);
171 void (*event_destroy)(struct fme_perf_priv *priv, u32 event,
173 u64 (*read_counter)(struct fme_perf_priv *priv, u32 event, u32 portid);
217 PMU_FORMAT_ATTR(event, "config:0-11");
288 static int basic_event_init(struct fme_perf_priv *priv, u32 event, u32 portid) in basic_event_init() argument
290 if (event <= BASIC_EVNT_MAX && is_portid_root(portid)) in basic_event_init()
297 u32 event, u32 portid) in basic_read_event_counter() argument
304 static int cache_event_init(struct fme_perf_priv *priv, u32 event, u32 portid) in cache_event_init() argument
307 event <= CACHE_EVNT_MAX && is_portid_root(portid)) in cache_event_init()
314 u32 event, u32 portid) in cache_read_event_counter() argument
[all …]
/drivers/misc/ibmasm/
Devent.c46 struct ibmasm_event *event; in ibmasm_receive_event() local
53 event = &buffer->events[buffer->next_index]; in ibmasm_receive_event()
54 memcpy_fromio(event->data, data, data_size); in ibmasm_receive_event()
55 event->data_size = data_size; in ibmasm_receive_event()
56 event->serial_number = buffer->next_serial_number; in ibmasm_receive_event()
80 struct ibmasm_event *event; in ibmasm_get_next_event() local
96 event = &buffer->events[index]; in ibmasm_get_next_event()
97 while (event->serial_number < reader->next_serial_number) { in ibmasm_get_next_event()
99 event = &buffer->events[index]; in ibmasm_get_next_event()
101 memcpy(reader->data, event->data, event->data_size); in ibmasm_get_next_event()
[all …]

12345678910>>...47