Home
last modified time | relevance | path

Searched refs:hwc (Results 1 – 23 of 23) sorted by relevance

/drivers/iio/buffer/
Dindustrialio-hw-consumer.c53 struct iio_hw_consumer *hwc, struct iio_dev *indio_dev) in iio_hw_consumer_get_buffer() argument
58 list_for_each_entry(buf, &hwc->buffers, head) { in iio_hw_consumer_get_buffer()
72 list_add_tail(&buf->head, &hwc->buffers); in iio_hw_consumer_get_buffer()
86 struct iio_hw_consumer *hwc; in iio_hw_consumer_alloc() local
90 hwc = kzalloc(sizeof(*hwc), GFP_KERNEL); in iio_hw_consumer_alloc()
91 if (!hwc) in iio_hw_consumer_alloc()
94 INIT_LIST_HEAD(&hwc->buffers); in iio_hw_consumer_alloc()
96 hwc->channels = iio_channel_get_all(dev); in iio_hw_consumer_alloc()
97 if (IS_ERR(hwc->channels)) { in iio_hw_consumer_alloc()
98 ret = PTR_ERR(hwc->channels); in iio_hw_consumer_alloc()
[all …]
/drivers/perf/hisilicon/
Dhisi_uncore_pmu.c128 struct hw_perf_event *hwc = &event->hw; in hisi_uncore_pmu_event_init() local
167 hwc->idx = -1; in hisi_uncore_pmu_event_init()
168 hwc->config_base = event->attr.config; in hisi_uncore_pmu_event_init()
183 struct hw_perf_event *hwc = &event->hw; in hisi_uncore_pmu_enable_event() local
185 hisi_pmu->ops->write_evtype(hisi_pmu, hwc->idx, in hisi_uncore_pmu_enable_event()
188 hisi_pmu->ops->enable_counter_int(hisi_pmu, hwc); in hisi_uncore_pmu_enable_event()
189 hisi_pmu->ops->enable_counter(hisi_pmu, hwc); in hisi_uncore_pmu_enable_event()
198 struct hw_perf_event *hwc = &event->hw; in hisi_uncore_pmu_disable_event() local
200 hisi_pmu->ops->disable_counter(hisi_pmu, hwc); in hisi_uncore_pmu_disable_event()
201 hisi_pmu->ops->disable_counter_int(hisi_pmu, hwc); in hisi_uncore_pmu_disable_event()
[all …]
Dhisi_uncore_ddrc_pmu.c47 #define GET_DDRC_EVENTID(hwc) (hwc->config_base & 0x7) argument
65 struct hw_perf_event *hwc) in hisi_ddrc_pmu_read_counter() argument
68 u32 idx = GET_DDRC_EVENTID(hwc); in hisi_ddrc_pmu_read_counter()
79 struct hw_perf_event *hwc, u64 val) in hisi_ddrc_pmu_write_counter() argument
81 u32 idx = GET_DDRC_EVENTID(hwc); in hisi_ddrc_pmu_write_counter()
122 struct hw_perf_event *hwc) in hisi_ddrc_pmu_enable_counter() argument
128 val |= (1 << GET_DDRC_EVENTID(hwc)); in hisi_ddrc_pmu_enable_counter()
133 struct hw_perf_event *hwc) in hisi_ddrc_pmu_disable_counter() argument
139 val &= ~(1 << GET_DDRC_EVENTID(hwc)); in hisi_ddrc_pmu_disable_counter()
147 struct hw_perf_event *hwc = &event->hw; in hisi_ddrc_pmu_get_event_idx() local
[all …]
Dhisi_uncore_l3c_pmu.c50 struct hw_perf_event *hwc) in hisi_l3c_pmu_read_counter() argument
52 u32 idx = hwc->idx; in hisi_l3c_pmu_read_counter()
64 struct hw_perf_event *hwc, u64 val) in hisi_l3c_pmu_write_counter() argument
66 u32 idx = hwc->idx; in hisi_l3c_pmu_write_counter()
127 struct hw_perf_event *hwc) in hisi_l3c_pmu_enable_counter() argument
133 val |= (1 << hwc->idx); in hisi_l3c_pmu_enable_counter()
138 struct hw_perf_event *hwc) in hisi_l3c_pmu_disable_counter() argument
144 val &= ~(1 << hwc->idx); in hisi_l3c_pmu_disable_counter()
149 struct hw_perf_event *hwc) in hisi_l3c_pmu_enable_counter_int() argument
155 val &= ~(1 << hwc->idx); in hisi_l3c_pmu_enable_counter_int()
[all …]
Dhisi_uncore_hha_pmu.c51 struct hw_perf_event *hwc) in hisi_hha_pmu_read_counter() argument
53 u32 idx = hwc->idx; in hisi_hha_pmu_read_counter()
65 struct hw_perf_event *hwc, u64 val) in hisi_hha_pmu_write_counter() argument
67 u32 idx = hwc->idx; in hisi_hha_pmu_write_counter()
128 struct hw_perf_event *hwc) in hisi_hha_pmu_enable_counter() argument
134 val |= (1 << hwc->idx); in hisi_hha_pmu_enable_counter()
139 struct hw_perf_event *hwc) in hisi_hha_pmu_disable_counter() argument
145 val &= ~(1 << hwc->idx); in hisi_hha_pmu_disable_counter()
150 struct hw_perf_event *hwc) in hisi_hha_pmu_enable_counter_int() argument
156 val &= ~(1 << hwc->idx); in hisi_hha_pmu_enable_counter_int()
[all …]
/drivers/gpu/drm/amd/amdgpu/
Damdgpu_pmu.c48 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_event_init() local
55 hwc->conf = event->attr.config; in amdgpu_perf_event_init()
63 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_start() local
68 if (WARN_ON_ONCE(!(hwc->state & PERF_HES_STOPPED))) in amdgpu_perf_start()
71 WARN_ON_ONCE(!(hwc->state & PERF_HES_UPTODATE)); in amdgpu_perf_start()
72 hwc->state = 0; in amdgpu_perf_start()
77 pe->adev->df_funcs->pmc_start(pe->adev, hwc->conf, 1); in amdgpu_perf_start()
79 pe->adev->df_funcs->pmc_start(pe->adev, hwc->conf, 0); in amdgpu_perf_start()
92 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_read() local
100 prev = local64_read(&hwc->prev_count); in amdgpu_perf_read()
[all …]
/drivers/perf/
Darm_pmu.c123 struct hw_perf_event *hwc = &event->hw; in armpmu_event_set_period() local
124 s64 left = local64_read(&hwc->period_left); in armpmu_event_set_period()
125 s64 period = hwc->sample_period; in armpmu_event_set_period()
132 local64_set(&hwc->period_left, left); in armpmu_event_set_period()
133 hwc->last_period = period; in armpmu_event_set_period()
139 local64_set(&hwc->period_left, left); in armpmu_event_set_period()
140 hwc->last_period = period; in armpmu_event_set_period()
153 local64_set(&hwc->prev_count, (u64)-left); in armpmu_event_set_period()
165 struct hw_perf_event *hwc = &event->hw; in armpmu_event_update() local
170 prev_raw_count = local64_read(&hwc->prev_count); in armpmu_event_update()
[all …]
Dthunderx2_pmu.c247 struct hw_perf_event *hwc = &event->hw; in init_cntr_base_l3c() local
250 hwc->config_base = (unsigned long)tx2_pmu->base in init_cntr_base_l3c()
252 hwc->event_base = (unsigned long)tx2_pmu->base in init_cntr_base_l3c()
259 struct hw_perf_event *hwc = &event->hw; in init_cntr_base_dmc() local
261 hwc->config_base = (unsigned long)tx2_pmu->base in init_cntr_base_dmc()
264 hwc->event_base = (unsigned long)tx2_pmu->base in init_cntr_base_dmc()
271 struct hw_perf_event *hwc = &event->hw; in uncore_start_event_l3c() local
275 reg_writel(val, hwc->config_base); in uncore_start_event_l3c()
276 local64_set(&hwc->prev_count, 0); in uncore_start_event_l3c()
277 reg_writel(0, hwc->event_base); in uncore_start_event_l3c()
[all …]
Dqcom_l2_pmu.c341 struct hw_perf_event *hwc = &event->hw; in l2_cache_event_update() local
343 u32 idx = hwc->idx; in l2_cache_event_update()
346 prev = local64_read(&hwc->prev_count); in l2_cache_event_update()
348 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); in l2_cache_event_update()
362 struct hw_perf_event *hwc) in l2_cache_cluster_set_period() argument
364 u32 idx = hwc->idx; in l2_cache_cluster_set_period()
377 local64_set(&hwc->prev_count, new); in l2_cache_cluster_set_period()
384 struct hw_perf_event *hwc = &event->hw; in l2_cache_get_event_idx() local
389 if (hwc->config_base == L2CYCLE_CTR_RAW_CODE) { in l2_cache_get_event_idx()
406 group = L2_EVT_GROUP(hwc->config_base); in l2_cache_get_event_idx()
[all …]
Darm_smmuv3_pmu.c204 struct hw_perf_event *hwc = &event->hw; in smmu_pmu_event_update() local
207 u32 idx = hwc->idx; in smmu_pmu_event_update()
210 prev = local64_read(&hwc->prev_count); in smmu_pmu_event_update()
212 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); in smmu_pmu_event_update()
222 struct hw_perf_event *hwc) in smmu_pmu_set_period() argument
224 u32 idx = hwc->idx; in smmu_pmu_set_period()
247 local64_set(&hwc->prev_count, new); in smmu_pmu_set_period()
343 struct hw_perf_event *hwc = &event->hw; in smmu_pmu_event_init() local
353 if (hwc->sample_period) { in smmu_pmu_event_init()
391 hwc->idx = -1; in smmu_pmu_event_init()
[all …]
Dfsl_imx8_ddr_perf.c259 struct hw_perf_event *hwc = &event->hw; in ddr_perf_event_init() local
298 hwc->idx = -1; in ddr_perf_event_init()
307 struct hw_perf_event *hwc = &event->hw; in ddr_perf_event_update() local
309 int counter = hwc->idx; in ddr_perf_event_update()
312 prev_raw_count = local64_read(&hwc->prev_count); in ddr_perf_event_update()
314 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in ddr_perf_event_update()
347 struct hw_perf_event *hwc = &event->hw; in ddr_perf_event_start() local
348 int counter = hwc->idx; in ddr_perf_event_start()
350 local64_set(&hwc->prev_count, 0); in ddr_perf_event_start()
354 hwc->state = 0; in ddr_perf_event_start()
[all …]
Darm-cci.c985 struct hw_perf_event *hwc = &event->hw; in pmu_event_update() local
989 prev_raw_count = local64_read(&hwc->prev_count); in pmu_event_update()
991 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in pmu_event_update()
1008 struct hw_perf_event *hwc = &event->hw; in pmu_event_set_period() local
1016 local64_set(&hwc->prev_count, val); in pmu_event_set_period()
1024 hwc->state |= PERF_HES_ARCH; in pmu_event_set_period()
1137 struct hw_perf_event *hwc = &event->hw; in cci_pmu_start() local
1138 int idx = hwc->idx; in cci_pmu_start()
1146 WARN_ON_ONCE(!(hwc->state & PERF_HES_UPTODATE)); in cci_pmu_start()
1148 hwc->state = 0; in cci_pmu_start()
[all …]
Dqcom_l3_pmu.c481 struct hw_perf_event *hwc = &event->hw; in qcom_l3_cache__event_init() local
492 if (hwc->sample_period) in qcom_l3_cache__event_init()
506 hwc->idx = -1; in qcom_l3_cache__event_init()
526 struct hw_perf_event *hwc = &event->hw; in qcom_l3_cache__event_start() local
529 hwc->state = 0; in qcom_l3_cache__event_start()
535 struct hw_perf_event *hwc = &event->hw; in qcom_l3_cache__event_stop() local
538 if (hwc->state & PERF_HES_STOPPED) in qcom_l3_cache__event_stop()
544 hwc->state |= PERF_HES_STOPPED | PERF_HES_UPTODATE; in qcom_l3_cache__event_stop()
550 struct hw_perf_event *hwc = &event->hw; in qcom_l3_cache__event_add() local
562 hwc->idx = idx; in qcom_l3_cache__event_add()
[all …]
Darm_dsu_pmu.c351 struct hw_perf_event *hwc = &event->hw; in dsu_pmu_event_update() local
356 prev_count = local64_read(&hwc->prev_count); in dsu_pmu_event_update()
358 } while (local64_cmpxchg(&hwc->prev_count, prev_count, new_count) != in dsu_pmu_event_update()
360 delta = (new_count - prev_count) & DSU_PMU_COUNTER_MASK(hwc->idx); in dsu_pmu_event_update()
444 struct hw_perf_event *hwc = &event->hw; in dsu_pmu_add() local
455 hwc->idx = idx; in dsu_pmu_add()
457 hwc->state = PERF_HES_STOPPED | PERF_HES_UPTODATE; in dsu_pmu_add()
470 struct hw_perf_event *hwc = &event->hw; in dsu_pmu_del() local
471 int idx = hwc->idx; in dsu_pmu_del()
Darm_spe_pmu.c716 struct hw_perf_event *hwc = &event->hw; in arm_spe_pmu_start() local
719 hwc->state = 0; in arm_spe_pmu_start()
721 if (hwc->state) in arm_spe_pmu_start()
737 reg = local64_read(&hwc->period_left); in arm_spe_pmu_start()
749 struct hw_perf_event *hwc = &event->hw; in arm_spe_pmu_stop() local
753 if (hwc->state & PERF_HES_STOPPED) in arm_spe_pmu_stop()
780 local64_set(&hwc->period_left, read_sysreg_s(SYS_PMSICR_EL1)); in arm_spe_pmu_stop()
781 hwc->state |= PERF_HES_UPTODATE; in arm_spe_pmu_stop()
784 hwc->state |= PERF_HES_STOPPED; in arm_spe_pmu_stop()
791 struct hw_perf_event *hwc = &event->hw; in arm_spe_pmu_add() local
[all …]
/drivers/clk/
Dclk-qoriq.c788 struct mux_hwclock *hwc = to_mux_hwclock(hw); in mux_set_parent() local
791 if (idx >= hwc->num_parents) in mux_set_parent()
794 clksel = hwc->parent_to_clksel[idx]; in mux_set_parent()
795 cg_out(hwc->cg, (clksel << CLKSEL_SHIFT) & CLKSEL_MASK, hwc->reg); in mux_set_parent()
802 struct mux_hwclock *hwc = to_mux_hwclock(hw); in mux_get_parent() local
806 clksel = (cg_in(hwc->cg, hwc->reg) & CLKSEL_MASK) >> CLKSEL_SHIFT; in mux_get_parent()
808 ret = hwc->clksel_to_parent[clksel]; in mux_get_parent()
810 pr_err("%s: mux at %p has bad clksel\n", __func__, hwc->reg); in mux_get_parent()
831 struct mux_hwclock *hwc, in get_pll_div() argument
836 if (!(hwc->info->clksel[idx].flags & CLKSEL_VALID)) in get_pll_div()
[all …]
/drivers/iio/adc/
Dstm32-dfsdm-adc.c79 struct iio_hw_consumer *hwc; member
1026 if (adc->hwc) { in __stm32_dfsdm_postenable()
1027 ret = iio_hw_consumer_enable(adc->hwc); in __stm32_dfsdm_postenable()
1055 if (adc->hwc) in __stm32_dfsdm_postenable()
1056 iio_hw_consumer_disable(adc->hwc); in __stm32_dfsdm_postenable()
1094 if (adc->hwc) in __stm32_dfsdm_predisable()
1095 iio_hw_consumer_disable(adc->hwc); in __stm32_dfsdm_predisable()
1275 ret = iio_hw_consumer_enable(adc->hwc); in stm32_dfsdm_read_raw()
1284 iio_hw_consumer_disable(adc->hwc); in stm32_dfsdm_read_raw()
1490 adc->hwc = devm_iio_hw_consumer_alloc(&indio_dev->dev); in stm32_dfsdm_adc_init()
[all …]
/drivers/video/fbdev/
Dau1200fb.c879 lcd->hwc.cursorctrl = 0; in au1200_setpanel()
880 lcd->hwc.cursorpos = 0; in au1200_setpanel()
881 lcd->hwc.cursorcolor0 = 0; in au1200_setpanel()
882 lcd->hwc.cursorcolor1 = 0; in au1200_setpanel()
883 lcd->hwc.cursorcolor2 = 0; in au1200_setpanel()
884 lcd->hwc.cursorcolor3 = 0; in au1200_setpanel()
928 D(lcd->hwc.cursorctrl); in au1200_setpanel()
929 D(lcd->hwc.cursorpos); in au1200_setpanel()
930 D(lcd->hwc.cursorcolor0); in au1200_setpanel()
931 D(lcd->hwc.cursorcolor1); in au1200_setpanel()
[all …]
Dpxa168fb.h309 #define CFG_CSB_256x32(hwc) ((hwc) << 15) /* HWC */ argument
Dau1200fb.h59 } hwc; member
/drivers/hwtracing/coresight/
Dcoresight-etm-perf.c392 struct hw_perf_event *hwc = &event->hw; in etm_event_add() local
396 if (hwc->state & PERF_HES_STOPPED) in etm_event_add()
399 hwc->state = PERF_HES_STOPPED; in etm_event_add()
/drivers/gpu/drm/i915/
Di915_pmu.c576 struct hw_perf_event *hwc = &event->hw; in i915_pmu_event_read() local
580 prev = local64_read(&hwc->prev_count); in i915_pmu_event_read()
583 if (local64_cmpxchg(&hwc->prev_count, prev, new) != prev) in i915_pmu_event_read()
/drivers/video/fbdev/mmp/hw/
Dmmp_ctrl.h534 #define CFG_CSB_256x32(hwc) ((hwc)<<15) /* HWC */ argument