/drivers/gpu/drm/nouveau/nvkm/core/ |
D | event.c | 26 nvkm_event_put(struct nvkm_event *event, u32 types, int index) in nvkm_event_put() argument 28 assert_spin_locked(&event->refs_lock); in nvkm_event_put() 31 if (--event->refs[index * event->types_nr + type] == 0) { in nvkm_event_put() 32 if (event->func->fini) in nvkm_event_put() 33 event->func->fini(event, 1 << type, index); in nvkm_event_put() 39 nvkm_event_get(struct nvkm_event *event, u32 types, int index) in nvkm_event_get() argument 41 assert_spin_locked(&event->refs_lock); in nvkm_event_get() 44 if (++event->refs[index * event->types_nr + type] == 1) { in nvkm_event_get() 45 if (event->func->init) in nvkm_event_get() 46 event->func->init(event, 1 << type, index); in nvkm_event_get() [all …]
|
D | notify.c | 31 nvkm_event_put(notify->event, notify->types, notify->index); in nvkm_notify_put_locked() 37 struct nvkm_event *event = notify->event; in nvkm_notify_put() local 39 if (likely(event) && in nvkm_notify_put() 41 spin_lock_irqsave(&event->refs_lock, flags); in nvkm_notify_put() 43 spin_unlock_irqrestore(&event->refs_lock, flags); in nvkm_notify_put() 53 nvkm_event_get(notify->event, notify->types, notify->index); in nvkm_notify_get_locked() 59 struct nvkm_event *event = notify->event; in nvkm_notify_get() local 61 if (likely(event) && in nvkm_notify_get() 63 spin_lock_irqsave(&event->refs_lock, flags); in nvkm_notify_get() 65 spin_unlock_irqrestore(&event->refs_lock, flags); in nvkm_notify_get() [all …]
|
/drivers/md/ |
D | dm-uevent.c | 39 static void dm_uevent_free(struct dm_uevent *event) in dm_uevent_free() argument 41 kmem_cache_free(_dm_event_cache, event); in dm_uevent_free() 46 struct dm_uevent *event; in dm_uevent_alloc() local 48 event = kmem_cache_zalloc(_dm_event_cache, GFP_ATOMIC); in dm_uevent_alloc() 49 if (!event) in dm_uevent_alloc() 52 INIT_LIST_HEAD(&event->elist); in dm_uevent_alloc() 53 event->md = md; in dm_uevent_alloc() 55 return event; in dm_uevent_alloc() 65 struct dm_uevent *event; in dm_build_path_uevent() local 67 event = dm_uevent_alloc(md); in dm_build_path_uevent() [all …]
|
/drivers/firmware/ |
D | arm_sdei.c | 76 struct sdei_event *event; member 81 #define CROSSCALL_INIT(arg, event) \ argument 83 arg.event = event; \ 89 struct sdei_event *event) in sdei_do_local_call() argument 93 CROSSCALL_INIT(arg, event); in sdei_do_local_call() 100 struct sdei_event *event) in sdei_do_cross_call() argument 104 CROSSCALL_INIT(arg, event); in sdei_do_cross_call() 190 static int sdei_api_event_get_info(u32 event, u32 info, u64 *result) in sdei_api_event_get_info() argument 192 return invoke_sdei_fn(SDEI_1_0_FN_SDEI_EVENT_GET_INFO, event, info, 0, in sdei_api_event_get_info() 202 struct sdei_event *event; in sdei_event_create() local [all …]
|
/drivers/perf/hisilicon/ |
D | hisi_uncore_pmu.c | 66 static bool hisi_validate_event_group(struct perf_event *event) in hisi_validate_event_group() argument 68 struct perf_event *sibling, *leader = event->group_leader; in hisi_validate_event_group() 69 struct hisi_pmu *hisi_pmu = to_hisi_pmu(event->pmu); in hisi_validate_event_group() 78 if (leader->pmu != event->pmu) in hisi_validate_event_group() 82 if (leader != event) in hisi_validate_event_group() 86 for_each_sibling_event(sibling, event->group_leader) { in hisi_validate_event_group() 89 if (sibling->pmu != event->pmu) in hisi_validate_event_group() 105 int hisi_uncore_pmu_get_event_idx(struct perf_event *event) in hisi_uncore_pmu_get_event_idx() argument 107 struct hisi_pmu *hisi_pmu = to_hisi_pmu(event->pmu); in hisi_uncore_pmu_get_event_idx() 132 int hisi_uncore_pmu_event_init(struct perf_event *event) in hisi_uncore_pmu_event_init() argument [all …]
|
/drivers/perf/ |
D | arm_pmu.c | 108 static inline u64 arm_pmu_event_max_period(struct perf_event *event) in arm_pmu_event_max_period() argument 110 if (event->hw.flags & ARMPMU_EVT_64BIT) in arm_pmu_event_max_period() 170 armpmu_map_event(struct perf_event *event, in armpmu_map_event() argument 178 u64 config = event->attr.config; in armpmu_map_event() 179 int type = event->attr.type; in armpmu_map_event() 181 if (type == event->pmu->type) in armpmu_map_event() 196 int armpmu_event_set_period(struct perf_event *event) in armpmu_event_set_period() argument 198 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period() 199 struct hw_perf_event *hwc = &event->hw; in armpmu_event_set_period() 205 max_period = arm_pmu_event_max_period(event); in armpmu_event_set_period() [all …]
|
D | qcom_l3_pmu.c | 136 static inline u32 get_event_type(struct perf_event *event) in get_event_type() argument 138 return (event->attr.config) & L3_EVTYPE_MASK; in get_event_type() 141 static inline bool event_uses_long_counter(struct perf_event *event) in event_uses_long_counter() argument 143 return !!(event->attr.config & BIT_ULL(L3_EVENT_LC_BIT)); in event_uses_long_counter() 146 static inline int event_num_counters(struct perf_event *event) in event_num_counters() argument 148 return event_uses_long_counter(event) ? 2 : 1; in event_num_counters() 175 void (*start)(struct perf_event *event); 177 void (*stop)(struct perf_event *event, int flags); 179 void (*update)(struct perf_event *event); 193 static void qcom_l3_cache__64bit_counter_start(struct perf_event *event) in qcom_l3_cache__64bit_counter_start() argument [all …]
|
D | arm_dsu_pmu.c | 177 DSU_FORMAT_ATTR(event, "config:0-31"), 249 static inline u64 dsu_pmu_read_counter(struct perf_event *event) in dsu_pmu_read_counter() argument 253 struct dsu_pmu *dsu_pmu = to_dsu_pmu(event->pmu); in dsu_pmu_read_counter() 254 int idx = event->hw.idx; in dsu_pmu_read_counter() 261 dev_err(event->pmu->dev, in dsu_pmu_read_counter() 276 static void dsu_pmu_write_counter(struct perf_event *event, u64 val) in dsu_pmu_write_counter() argument 279 struct dsu_pmu *dsu_pmu = to_dsu_pmu(event->pmu); in dsu_pmu_write_counter() 280 int idx = event->hw.idx; in dsu_pmu_write_counter() 287 dev_err(event->pmu->dev, in dsu_pmu_write_counter() 301 struct perf_event *event) in dsu_pmu_get_event_idx() argument [all …]
|
D | thunderx2_pmu.c | 111 void (*init_cntr_base)(struct perf_event *event, 113 void (*stop_event)(struct perf_event *event); 114 void (*start_event)(struct perf_event *event, int flags); 137 TX2_PMU_FORMAT_ATTR(event, event, "config:0-4"); 138 TX2_PMU_FORMAT_ATTR(event_ccpi2, event, "config:0-9"); 322 static void init_cntr_base_l3c(struct perf_event *event, in init_cntr_base_l3c() argument 325 struct hw_perf_event *hwc = &event->hw; in init_cntr_base_l3c() 328 tx2_pmu = pmu_to_tx2_pmu(event->pmu); in init_cntr_base_l3c() 333 + L3C_COUNTER_CTL + (8 * GET_COUNTERID(event, cmask)); in init_cntr_base_l3c() 335 + L3C_COUNTER_DATA + (8 * GET_COUNTERID(event, cmask)); in init_cntr_base_l3c() [all …]
|
D | qcom_l2_pmu.c | 75 #define L2_EVT_CODE(event) (((event) & L2_EVT_CODE_MASK) >> L2_EVT_CODE_SHIFT) argument 76 #define L2_EVT_GROUP(event) (((event) & L2_EVT_GRP_MASK) >> L2_EVT_GRP_SHIFT) argument 297 static void l2_cache_event_update(struct perf_event *event) in l2_cache_event_update() argument 299 struct hw_perf_event *hwc = &event->hw; in l2_cache_event_update() 316 local64_add(delta, &event->count); in l2_cache_event_update() 340 struct perf_event *event) in l2_cache_get_event_idx() argument 342 struct hw_perf_event *hwc = &event->hw; in l2_cache_get_event_idx() 375 struct perf_event *event) in l2_cache_clear_event_idx() argument 377 struct hw_perf_event *hwc = &event->hw; in l2_cache_clear_event_idx() 397 struct perf_event *event = cluster->events[idx]; in l2_cache_handle_irq() local [all …]
|
D | fsl_imx8_ddr_perf.c | 220 PMU_FORMAT_ATTR(event, "config:0-7"); 244 static bool ddr_perf_is_filtered(struct perf_event *event) in ddr_perf_is_filtered() argument 246 return event->attr.config == 0x41 || event->attr.config == 0x42; in ddr_perf_is_filtered() 249 static u32 ddr_perf_filter_val(struct perf_event *event) in ddr_perf_filter_val() argument 251 return event->attr.config1; in ddr_perf_filter_val() 264 static bool ddr_perf_is_enhanced_filtered(struct perf_event *event) in ddr_perf_is_enhanced_filtered() argument 267 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_is_enhanced_filtered() 271 ddr_perf_is_filtered(event); in ddr_perf_is_enhanced_filtered() 274 static u32 ddr_perf_alloc_counter(struct ddr_pmu *pmu, int event) in ddr_perf_alloc_counter() argument 283 if (event == EVENT_CYCLES_ID) { in ddr_perf_alloc_counter() [all …]
|
D | arm-ccn.c | 154 struct perf_event *event; member 235 static CCN_FORMAT_ATTR(event, "config:16-23"); 268 u32 event; member 287 .type = CCN_TYPE_MN, .event = CCN_EVENT_WATCHPOINT, \ 293 .event = CCN_EVENT_WATCHPOINT, .num_ports = CCN_NUM_XP_PORTS, \ 298 .event = CCN_EVENT_WATCHPOINT, .num_ports = CCN_NUM_XP_PORTS, \ 302 .type = CCN_TYPE_HNF, .event = _event, } 305 .type = CCN_TYPE_XP, .event = _event, \ 314 .type = CCN_TYPE_RNI_3P, .event = _event, } 317 .type = CCN_TYPE_SBAS, .event = _event, } [all …]
|
D | arm-cci.c | 193 #define CCI400_PMU_EVENT_SOURCE(event) \ argument 194 ((event >> CCI400_PMU_EVENT_SOURCE_SHIFT) & \ 196 #define CCI400_PMU_EVENT_CODE(event) \ argument 197 ((event >> CCI400_PMU_EVENT_CODE_SHIFT) & CCI400_PMU_EVENT_CODE_MASK) 217 CCI_FORMAT_EXT_ATTR_ENTRY(event, "config:0-4"), 432 #define CCI5xx_PMU_EVENT_SOURCE(event) \ argument 433 ((event >> CCI5xx_PMU_EVENT_SOURCE_SHIFT) & CCI5xx_PMU_EVENT_SOURCE_MASK) 434 #define CCI5xx_PMU_EVENT_CODE(event) \ argument 435 ((event >> CCI5xx_PMU_EVENT_CODE_SHIFT) & CCI5xx_PMU_EVENT_CODE_MASK) 453 CCI_FORMAT_EXT_ATTR_ENTRY(event, "config:0-4"), [all …]
|
D | arm-cmn.c | 131 #define CMN_EVENT_TYPE(event) FIELD_GET(CMN_CONFIG_TYPE, (event)->attr.config) argument 132 #define CMN_EVENT_EVENTID(event) FIELD_GET(CMN_CONFIG_EVENTID, (event)->attr.config) argument 133 #define CMN_EVENT_OCCUPID(event) FIELD_GET(CMN_CONFIG_OCCUPID, (event)->attr.config) argument 134 #define CMN_EVENT_BYNODEID(event) FIELD_GET(CMN_CONFIG_BYNODEID, (event)->attr.config) argument 135 #define CMN_EVENT_NODEID(event) FIELD_GET(CMN_CONFIG_NODEID, (event)->attr.config) argument 145 #define CMN_EVENT_WP_COMBINE(event) FIELD_GET(CMN_CONFIG_WP_COMBINE, (event)->attr.config) argument 146 #define CMN_EVENT_WP_DEV_SEL(event) FIELD_GET(CMN_CONFIG_WP_DEV_SEL, (event)->attr.config) argument 147 #define CMN_EVENT_WP_CHN_SEL(event) FIELD_GET(CMN_CONFIG_WP_CHN_SEL, (event)->attr.config) argument 148 #define CMN_EVENT_WP_GRP(event) FIELD_GET(CMN_CONFIG_WP_GRP, (event)->attr.config) argument 149 #define CMN_EVENT_WP_EXCLUSIVE(event) FIELD_GET(CMN_CONFIG_WP_EXCLUSIVE, (event)->attr.config) argument [all …]
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_pmu.c | 46 static int amdgpu_perf_event_init(struct perf_event *event) in amdgpu_perf_event_init() argument 48 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_event_init() 51 if (event->attr.type != event->pmu->type) in amdgpu_perf_event_init() 55 hwc->config = event->attr.config; in amdgpu_perf_event_init() 61 static void amdgpu_perf_start(struct perf_event *event, int flags) in amdgpu_perf_start() argument 63 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_start() 64 struct amdgpu_pmu_entry *pe = container_of(event->pmu, in amdgpu_perf_start() 85 perf_event_update_userpage(event); in amdgpu_perf_start() 90 static void amdgpu_perf_read(struct perf_event *event) in amdgpu_perf_read() argument 92 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_read() [all …]
|
/drivers/oprofile/ |
D | oprofile_perf.c | 19 unsigned long event; member 36 static void op_overflow_handler(struct perf_event *event, in op_overflow_handler() argument 43 if (per_cpu(perf_events, cpu)[id] == event) in op_overflow_handler() 69 attr->config = counter_config[i].event; in op_perf_setup() 75 static int op_create_counter(int cpu, int event) in op_create_counter() argument 79 if (!counter_config[event].enabled || per_cpu(perf_events, cpu)[event]) in op_create_counter() 82 pevent = perf_event_create_kernel_counter(&counter_config[event].attr, in op_create_counter() 92 event, cpu); in op_create_counter() 96 per_cpu(perf_events, cpu)[event] = pevent; in op_create_counter() 101 static void op_destroy_counter(int cpu, int event) in op_destroy_counter() argument [all …]
|
D | nmi_timer_int.c | 29 static void nmi_timer_callback(struct perf_event *event, in nmi_timer_callback() argument 33 event->hw.interrupts = 0; /* don't throttle interrupts */ in nmi_timer_callback() 39 struct perf_event *event = per_cpu(nmi_timer_events, cpu); in nmi_timer_start_cpu() local 41 if (!event) { in nmi_timer_start_cpu() 42 event = perf_event_create_kernel_counter(&nmi_timer_attr, cpu, NULL, in nmi_timer_start_cpu() 44 if (IS_ERR(event)) in nmi_timer_start_cpu() 45 return PTR_ERR(event); in nmi_timer_start_cpu() 46 per_cpu(nmi_timer_events, cpu) = event; in nmi_timer_start_cpu() 49 if (event && ctr_running) in nmi_timer_start_cpu() 50 perf_event_enable(event); in nmi_timer_start_cpu() [all …]
|
/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
D | fweh.c | 85 struct brcmf_fweh_queue_item *event) in brcmf_fweh_queue_event() argument 90 list_add_tail(&event->q, &fweh->event_q); in brcmf_fweh_queue_event() 194 struct brcmf_fweh_queue_item *event = NULL; in brcmf_fweh_dequeue_event() local 199 event = list_first_entry(&fweh->event_q, in brcmf_fweh_dequeue_event() 201 list_del(&event->q); in brcmf_fweh_dequeue_event() 205 return event; in brcmf_fweh_dequeue_event() 218 struct brcmf_fweh_queue_item *event; in brcmf_fweh_event_worker() local 226 while ((event = brcmf_fweh_dequeue_event(fweh))) { in brcmf_fweh_event_worker() 228 brcmf_fweh_event_name(event->code), event->code, in brcmf_fweh_event_worker() 229 event->emsg.ifidx, event->emsg.bsscfgidx, in brcmf_fweh_event_worker() [all …]
|
/drivers/net/ethernet/huawei/hinic/ |
D | hinic_devlink.c | 331 struct hinic_fault_event *event) in chip_fault_show() argument 338 fault_level = (event->event.chip.err_level < FAULT_LEVEL_MAX) ? in chip_fault_show() 339 event->event.chip.err_level : FAULT_LEVEL_MAX; in chip_fault_show() 342 (u32)event->event.chip.func_id); in chip_fault_show() 347 err = devlink_fmsg_u8_pair_put(fmsg, "module_id", event->event.chip.node_id); in chip_fault_show() 351 err = devlink_fmsg_u32_pair_put(fmsg, "err_type", (u32)event->event.chip.err_type); in chip_fault_show() 360 event->event.chip.err_csr_addr); in chip_fault_show() 365 event->event.chip.err_csr_value); in chip_fault_show() 373 struct hinic_fault_event *event) in fault_report_show() argument 381 fault_type = (event->type < FAULT_TYPE_MAX) ? event->type : FAULT_TYPE_MAX; in fault_report_show() [all …]
|
/drivers/input/ |
D | input-compat.c | 15 struct input_event *event) in input_event_from_user() argument 24 event->input_event_sec = compat_event.sec; in input_event_from_user() 25 event->input_event_usec = compat_event.usec; in input_event_from_user() 26 event->type = compat_event.type; in input_event_from_user() 27 event->code = compat_event.code; in input_event_from_user() 28 event->value = compat_event.value; in input_event_from_user() 31 if (copy_from_user(event, buffer, sizeof(struct input_event))) in input_event_from_user() 39 const struct input_event *event) in input_event_to_user() argument 44 compat_event.sec = event->input_event_sec; in input_event_to_user() 45 compat_event.usec = event->input_event_usec; in input_event_to_user() [all …]
|
/drivers/char/tpm/eventlog/ |
D | tpm1.c | 77 struct tcpa_event *event; in tpm1_bios_measurements_start() local 83 event = addr; in tpm1_bios_measurements_start() 90 do_endian_conversion(event->event_size); in tpm1_bios_measurements_start() 92 do_endian_conversion(event->event_type); in tpm1_bios_measurements_start() 111 struct tcpa_event *event = v; in tpm1_bios_measurements_next() local 119 converted_event_size = do_endian_conversion(event->event_size); in tpm1_bios_measurements_next() 127 event = v; in tpm1_bios_measurements_next() 129 converted_event_size = do_endian_conversion(event->event_size); in tpm1_bios_measurements_next() 130 converted_event_type = do_endian_conversion(event->event_type); in tpm1_bios_measurements_next() 143 static int get_event_name(char *dest, struct tcpa_event *event, in get_event_name() argument [all …]
|
D | tpm2.c | 36 static size_t calc_tpm2_event_size(struct tcg_pcr_event2_head *event, in calc_tpm2_event_size() argument 39 return __calc_tpm2_event_size(event, event_header, false); in calc_tpm2_event_size() 49 struct tcg_pcr_event2_head *event; in tpm2_bios_measurements_start() local 54 size = struct_size(event_header, event, event_header->event_size); in tpm2_bios_measurements_start() 67 event = addr; in tpm2_bios_measurements_start() 68 size = calc_tpm2_event_size(event, event_header); in tpm2_bios_measurements_start() 74 event = addr; in tpm2_bios_measurements_start() 75 size = calc_tpm2_event_size(event, event_header); in tpm2_bios_measurements_start() 89 struct tcg_pcr_event2_head *event; in tpm2_bios_measurements_next() local 100 event_size = struct_size(event_header, event, in tpm2_bios_measurements_next() [all …]
|
/drivers/acpi/ |
D | evged.c | 58 struct acpi_ged_event *event = data; in acpi_ged_irq_handler() local 61 acpi_ret = acpi_execute_simple_method(event->handle, NULL, event->gsi); in acpi_ged_irq_handler() 63 dev_err_once(event->dev, "IRQ method execution failed\n"); in acpi_ged_irq_handler() 71 struct acpi_ged_event *event; in acpi_ged_request_interrupt() local 118 event = devm_kzalloc(dev, sizeof(*event), GFP_KERNEL); in acpi_ged_request_interrupt() 119 if (!event) in acpi_ged_request_interrupt() 122 event->gsi = gsi; in acpi_ged_request_interrupt() 123 event->dev = dev; in acpi_ged_request_interrupt() 124 event->irq = irq; in acpi_ged_request_interrupt() 125 event->handle = evt_handle; in acpi_ged_request_interrupt() [all …]
|
/drivers/fpga/ |
D | dfl-fme-perf.c | 170 int (*event_init)(struct fme_perf_priv *priv, u32 event, u32 portid); 171 void (*event_destroy)(struct fme_perf_priv *priv, u32 event, 173 u64 (*read_counter)(struct fme_perf_priv *priv, u32 event, u32 portid); 217 PMU_FORMAT_ATTR(event, "config:0-11"); 288 static int basic_event_init(struct fme_perf_priv *priv, u32 event, u32 portid) in basic_event_init() argument 290 if (event <= BASIC_EVNT_MAX && is_portid_root(portid)) in basic_event_init() 297 u32 event, u32 portid) in basic_read_event_counter() argument 304 static int cache_event_init(struct fme_perf_priv *priv, u32 event, u32 portid) in cache_event_init() argument 307 event <= CACHE_EVNT_MAX && is_portid_root(portid)) in cache_event_init() 314 u32 event, u32 portid) in cache_read_event_counter() argument [all …]
|
/drivers/misc/ibmasm/ |
D | event.c | 46 struct ibmasm_event *event; in ibmasm_receive_event() local 53 event = &buffer->events[buffer->next_index]; in ibmasm_receive_event() 54 memcpy_fromio(event->data, data, data_size); in ibmasm_receive_event() 55 event->data_size = data_size; in ibmasm_receive_event() 56 event->serial_number = buffer->next_serial_number; in ibmasm_receive_event() 80 struct ibmasm_event *event; in ibmasm_get_next_event() local 96 event = &buffer->events[index]; in ibmasm_get_next_event() 97 while (event->serial_number < reader->next_serial_number) { in ibmasm_get_next_event() 99 event = &buffer->events[index]; in ibmasm_get_next_event() 101 memcpy(reader->data, event->data, event->data_size); in ibmasm_get_next_event() [all …]
|