/drivers/iio/buffer/ |
D | industrialio-hw-consumer.c | 53 struct iio_hw_consumer *hwc, struct iio_dev *indio_dev) in iio_hw_consumer_get_buffer() argument 58 list_for_each_entry(buf, &hwc->buffers, head) { in iio_hw_consumer_get_buffer() 72 list_add_tail(&buf->head, &hwc->buffers); in iio_hw_consumer_get_buffer() 86 struct iio_hw_consumer *hwc; in iio_hw_consumer_alloc() local 90 hwc = kzalloc(sizeof(*hwc), GFP_KERNEL); in iio_hw_consumer_alloc() 91 if (!hwc) in iio_hw_consumer_alloc() 94 INIT_LIST_HEAD(&hwc->buffers); in iio_hw_consumer_alloc() 96 hwc->channels = iio_channel_get_all(dev); in iio_hw_consumer_alloc() 97 if (IS_ERR(hwc->channels)) { in iio_hw_consumer_alloc() 98 ret = PTR_ERR(hwc->channels); in iio_hw_consumer_alloc() [all …]
|
/drivers/perf/hisilicon/ |
D | hisi_uncore_pmu.c | 128 struct hw_perf_event *hwc = &event->hw; in hisi_uncore_pmu_event_init() local 167 hwc->idx = -1; in hisi_uncore_pmu_event_init() 168 hwc->config_base = event->attr.config; in hisi_uncore_pmu_event_init() 183 struct hw_perf_event *hwc = &event->hw; in hisi_uncore_pmu_enable_event() local 185 hisi_pmu->ops->write_evtype(hisi_pmu, hwc->idx, in hisi_uncore_pmu_enable_event() 188 hisi_pmu->ops->enable_counter_int(hisi_pmu, hwc); in hisi_uncore_pmu_enable_event() 189 hisi_pmu->ops->enable_counter(hisi_pmu, hwc); in hisi_uncore_pmu_enable_event() 198 struct hw_perf_event *hwc = &event->hw; in hisi_uncore_pmu_disable_event() local 200 hisi_pmu->ops->disable_counter(hisi_pmu, hwc); in hisi_uncore_pmu_disable_event() 201 hisi_pmu->ops->disable_counter_int(hisi_pmu, hwc); in hisi_uncore_pmu_disable_event() [all …]
|
D | hisi_uncore_ddrc_pmu.c | 47 #define GET_DDRC_EVENTID(hwc) (hwc->config_base & 0x7) argument 65 struct hw_perf_event *hwc) in hisi_ddrc_pmu_read_counter() argument 68 u32 idx = GET_DDRC_EVENTID(hwc); in hisi_ddrc_pmu_read_counter() 79 struct hw_perf_event *hwc, u64 val) in hisi_ddrc_pmu_write_counter() argument 81 u32 idx = GET_DDRC_EVENTID(hwc); in hisi_ddrc_pmu_write_counter() 122 struct hw_perf_event *hwc) in hisi_ddrc_pmu_enable_counter() argument 128 val |= (1 << GET_DDRC_EVENTID(hwc)); in hisi_ddrc_pmu_enable_counter() 133 struct hw_perf_event *hwc) in hisi_ddrc_pmu_disable_counter() argument 139 val &= ~(1 << GET_DDRC_EVENTID(hwc)); in hisi_ddrc_pmu_disable_counter() 147 struct hw_perf_event *hwc = &event->hw; in hisi_ddrc_pmu_get_event_idx() local [all …]
|
D | hisi_uncore_l3c_pmu.c | 50 struct hw_perf_event *hwc) in hisi_l3c_pmu_read_counter() argument 52 u32 idx = hwc->idx; in hisi_l3c_pmu_read_counter() 64 struct hw_perf_event *hwc, u64 val) in hisi_l3c_pmu_write_counter() argument 66 u32 idx = hwc->idx; in hisi_l3c_pmu_write_counter() 127 struct hw_perf_event *hwc) in hisi_l3c_pmu_enable_counter() argument 133 val |= (1 << hwc->idx); in hisi_l3c_pmu_enable_counter() 138 struct hw_perf_event *hwc) in hisi_l3c_pmu_disable_counter() argument 144 val &= ~(1 << hwc->idx); in hisi_l3c_pmu_disable_counter() 149 struct hw_perf_event *hwc) in hisi_l3c_pmu_enable_counter_int() argument 155 val &= ~(1 << hwc->idx); in hisi_l3c_pmu_enable_counter_int() [all …]
|
D | hisi_uncore_hha_pmu.c | 51 struct hw_perf_event *hwc) in hisi_hha_pmu_read_counter() argument 53 u32 idx = hwc->idx; in hisi_hha_pmu_read_counter() 65 struct hw_perf_event *hwc, u64 val) in hisi_hha_pmu_write_counter() argument 67 u32 idx = hwc->idx; in hisi_hha_pmu_write_counter() 128 struct hw_perf_event *hwc) in hisi_hha_pmu_enable_counter() argument 134 val |= (1 << hwc->idx); in hisi_hha_pmu_enable_counter() 139 struct hw_perf_event *hwc) in hisi_hha_pmu_disable_counter() argument 145 val &= ~(1 << hwc->idx); in hisi_hha_pmu_disable_counter() 150 struct hw_perf_event *hwc) in hisi_hha_pmu_enable_counter_int() argument 156 val &= ~(1 << hwc->idx); in hisi_hha_pmu_enable_counter_int() [all …]
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_pmu.c | 48 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_event_init() local 55 hwc->conf = event->attr.config; in amdgpu_perf_event_init() 63 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_start() local 68 if (WARN_ON_ONCE(!(hwc->state & PERF_HES_STOPPED))) in amdgpu_perf_start() 71 WARN_ON_ONCE(!(hwc->state & PERF_HES_UPTODATE)); in amdgpu_perf_start() 72 hwc->state = 0; in amdgpu_perf_start() 77 pe->adev->df_funcs->pmc_start(pe->adev, hwc->conf, 1); in amdgpu_perf_start() 79 pe->adev->df_funcs->pmc_start(pe->adev, hwc->conf, 0); in amdgpu_perf_start() 92 struct hw_perf_event *hwc = &event->hw; in amdgpu_perf_read() local 100 prev = local64_read(&hwc->prev_count); in amdgpu_perf_read() [all …]
|
/drivers/perf/ |
D | arm_pmu.c | 123 struct hw_perf_event *hwc = &event->hw; in armpmu_event_set_period() local 124 s64 left = local64_read(&hwc->period_left); in armpmu_event_set_period() 125 s64 period = hwc->sample_period; in armpmu_event_set_period() 132 local64_set(&hwc->period_left, left); in armpmu_event_set_period() 133 hwc->last_period = period; in armpmu_event_set_period() 139 local64_set(&hwc->period_left, left); in armpmu_event_set_period() 140 hwc->last_period = period; in armpmu_event_set_period() 153 local64_set(&hwc->prev_count, (u64)-left); in armpmu_event_set_period() 165 struct hw_perf_event *hwc = &event->hw; in armpmu_event_update() local 170 prev_raw_count = local64_read(&hwc->prev_count); in armpmu_event_update() [all …]
|
D | thunderx2_pmu.c | 247 struct hw_perf_event *hwc = &event->hw; in init_cntr_base_l3c() local 250 hwc->config_base = (unsigned long)tx2_pmu->base in init_cntr_base_l3c() 252 hwc->event_base = (unsigned long)tx2_pmu->base in init_cntr_base_l3c() 259 struct hw_perf_event *hwc = &event->hw; in init_cntr_base_dmc() local 261 hwc->config_base = (unsigned long)tx2_pmu->base in init_cntr_base_dmc() 264 hwc->event_base = (unsigned long)tx2_pmu->base in init_cntr_base_dmc() 271 struct hw_perf_event *hwc = &event->hw; in uncore_start_event_l3c() local 275 reg_writel(val, hwc->config_base); in uncore_start_event_l3c() 276 local64_set(&hwc->prev_count, 0); in uncore_start_event_l3c() 277 reg_writel(0, hwc->event_base); in uncore_start_event_l3c() [all …]
|
D | qcom_l2_pmu.c | 341 struct hw_perf_event *hwc = &event->hw; in l2_cache_event_update() local 343 u32 idx = hwc->idx; in l2_cache_event_update() 346 prev = local64_read(&hwc->prev_count); in l2_cache_event_update() 348 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); in l2_cache_event_update() 362 struct hw_perf_event *hwc) in l2_cache_cluster_set_period() argument 364 u32 idx = hwc->idx; in l2_cache_cluster_set_period() 377 local64_set(&hwc->prev_count, new); in l2_cache_cluster_set_period() 384 struct hw_perf_event *hwc = &event->hw; in l2_cache_get_event_idx() local 389 if (hwc->config_base == L2CYCLE_CTR_RAW_CODE) { in l2_cache_get_event_idx() 406 group = L2_EVT_GROUP(hwc->config_base); in l2_cache_get_event_idx() [all …]
|
D | arm_smmuv3_pmu.c | 204 struct hw_perf_event *hwc = &event->hw; in smmu_pmu_event_update() local 207 u32 idx = hwc->idx; in smmu_pmu_event_update() 210 prev = local64_read(&hwc->prev_count); in smmu_pmu_event_update() 212 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); in smmu_pmu_event_update() 222 struct hw_perf_event *hwc) in smmu_pmu_set_period() argument 224 u32 idx = hwc->idx; in smmu_pmu_set_period() 247 local64_set(&hwc->prev_count, new); in smmu_pmu_set_period() 343 struct hw_perf_event *hwc = &event->hw; in smmu_pmu_event_init() local 353 if (hwc->sample_period) { in smmu_pmu_event_init() 391 hwc->idx = -1; in smmu_pmu_event_init() [all …]
|
D | fsl_imx8_ddr_perf.c | 259 struct hw_perf_event *hwc = &event->hw; in ddr_perf_event_init() local 298 hwc->idx = -1; in ddr_perf_event_init() 307 struct hw_perf_event *hwc = &event->hw; in ddr_perf_event_update() local 309 int counter = hwc->idx; in ddr_perf_event_update() 312 prev_raw_count = local64_read(&hwc->prev_count); in ddr_perf_event_update() 314 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in ddr_perf_event_update() 347 struct hw_perf_event *hwc = &event->hw; in ddr_perf_event_start() local 348 int counter = hwc->idx; in ddr_perf_event_start() 350 local64_set(&hwc->prev_count, 0); in ddr_perf_event_start() 354 hwc->state = 0; in ddr_perf_event_start() [all …]
|
D | arm-cci.c | 985 struct hw_perf_event *hwc = &event->hw; in pmu_event_update() local 989 prev_raw_count = local64_read(&hwc->prev_count); in pmu_event_update() 991 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in pmu_event_update() 1008 struct hw_perf_event *hwc = &event->hw; in pmu_event_set_period() local 1016 local64_set(&hwc->prev_count, val); in pmu_event_set_period() 1024 hwc->state |= PERF_HES_ARCH; in pmu_event_set_period() 1137 struct hw_perf_event *hwc = &event->hw; in cci_pmu_start() local 1138 int idx = hwc->idx; in cci_pmu_start() 1146 WARN_ON_ONCE(!(hwc->state & PERF_HES_UPTODATE)); in cci_pmu_start() 1148 hwc->state = 0; in cci_pmu_start() [all …]
|
D | qcom_l3_pmu.c | 481 struct hw_perf_event *hwc = &event->hw; in qcom_l3_cache__event_init() local 492 if (hwc->sample_period) in qcom_l3_cache__event_init() 506 hwc->idx = -1; in qcom_l3_cache__event_init() 526 struct hw_perf_event *hwc = &event->hw; in qcom_l3_cache__event_start() local 529 hwc->state = 0; in qcom_l3_cache__event_start() 535 struct hw_perf_event *hwc = &event->hw; in qcom_l3_cache__event_stop() local 538 if (hwc->state & PERF_HES_STOPPED) in qcom_l3_cache__event_stop() 544 hwc->state |= PERF_HES_STOPPED | PERF_HES_UPTODATE; in qcom_l3_cache__event_stop() 550 struct hw_perf_event *hwc = &event->hw; in qcom_l3_cache__event_add() local 562 hwc->idx = idx; in qcom_l3_cache__event_add() [all …]
|
D | arm_dsu_pmu.c | 351 struct hw_perf_event *hwc = &event->hw; in dsu_pmu_event_update() local 356 prev_count = local64_read(&hwc->prev_count); in dsu_pmu_event_update() 358 } while (local64_cmpxchg(&hwc->prev_count, prev_count, new_count) != in dsu_pmu_event_update() 360 delta = (new_count - prev_count) & DSU_PMU_COUNTER_MASK(hwc->idx); in dsu_pmu_event_update() 444 struct hw_perf_event *hwc = &event->hw; in dsu_pmu_add() local 455 hwc->idx = idx; in dsu_pmu_add() 457 hwc->state = PERF_HES_STOPPED | PERF_HES_UPTODATE; in dsu_pmu_add() 470 struct hw_perf_event *hwc = &event->hw; in dsu_pmu_del() local 471 int idx = hwc->idx; in dsu_pmu_del()
|
D | arm_spe_pmu.c | 716 struct hw_perf_event *hwc = &event->hw; in arm_spe_pmu_start() local 719 hwc->state = 0; in arm_spe_pmu_start() 721 if (hwc->state) in arm_spe_pmu_start() 737 reg = local64_read(&hwc->period_left); in arm_spe_pmu_start() 749 struct hw_perf_event *hwc = &event->hw; in arm_spe_pmu_stop() local 753 if (hwc->state & PERF_HES_STOPPED) in arm_spe_pmu_stop() 780 local64_set(&hwc->period_left, read_sysreg_s(SYS_PMSICR_EL1)); in arm_spe_pmu_stop() 781 hwc->state |= PERF_HES_UPTODATE; in arm_spe_pmu_stop() 784 hwc->state |= PERF_HES_STOPPED; in arm_spe_pmu_stop() 791 struct hw_perf_event *hwc = &event->hw; in arm_spe_pmu_add() local [all …]
|
/drivers/clk/ |
D | clk-qoriq.c | 788 struct mux_hwclock *hwc = to_mux_hwclock(hw); in mux_set_parent() local 791 if (idx >= hwc->num_parents) in mux_set_parent() 794 clksel = hwc->parent_to_clksel[idx]; in mux_set_parent() 795 cg_out(hwc->cg, (clksel << CLKSEL_SHIFT) & CLKSEL_MASK, hwc->reg); in mux_set_parent() 802 struct mux_hwclock *hwc = to_mux_hwclock(hw); in mux_get_parent() local 806 clksel = (cg_in(hwc->cg, hwc->reg) & CLKSEL_MASK) >> CLKSEL_SHIFT; in mux_get_parent() 808 ret = hwc->clksel_to_parent[clksel]; in mux_get_parent() 810 pr_err("%s: mux at %p has bad clksel\n", __func__, hwc->reg); in mux_get_parent() 831 struct mux_hwclock *hwc, in get_pll_div() argument 836 if (!(hwc->info->clksel[idx].flags & CLKSEL_VALID)) in get_pll_div() [all …]
|
/drivers/iio/adc/ |
D | stm32-dfsdm-adc.c | 79 struct iio_hw_consumer *hwc; member 1026 if (adc->hwc) { in __stm32_dfsdm_postenable() 1027 ret = iio_hw_consumer_enable(adc->hwc); in __stm32_dfsdm_postenable() 1055 if (adc->hwc) in __stm32_dfsdm_postenable() 1056 iio_hw_consumer_disable(adc->hwc); in __stm32_dfsdm_postenable() 1094 if (adc->hwc) in __stm32_dfsdm_predisable() 1095 iio_hw_consumer_disable(adc->hwc); in __stm32_dfsdm_predisable() 1275 ret = iio_hw_consumer_enable(adc->hwc); in stm32_dfsdm_read_raw() 1284 iio_hw_consumer_disable(adc->hwc); in stm32_dfsdm_read_raw() 1490 adc->hwc = devm_iio_hw_consumer_alloc(&indio_dev->dev); in stm32_dfsdm_adc_init() [all …]
|
/drivers/video/fbdev/ |
D | au1200fb.c | 879 lcd->hwc.cursorctrl = 0; in au1200_setpanel() 880 lcd->hwc.cursorpos = 0; in au1200_setpanel() 881 lcd->hwc.cursorcolor0 = 0; in au1200_setpanel() 882 lcd->hwc.cursorcolor1 = 0; in au1200_setpanel() 883 lcd->hwc.cursorcolor2 = 0; in au1200_setpanel() 884 lcd->hwc.cursorcolor3 = 0; in au1200_setpanel() 928 D(lcd->hwc.cursorctrl); in au1200_setpanel() 929 D(lcd->hwc.cursorpos); in au1200_setpanel() 930 D(lcd->hwc.cursorcolor0); in au1200_setpanel() 931 D(lcd->hwc.cursorcolor1); in au1200_setpanel() [all …]
|
D | pxa168fb.h | 309 #define CFG_CSB_256x32(hwc) ((hwc) << 15) /* HWC */ argument
|
D | au1200fb.h | 59 } hwc; member
|
/drivers/hwtracing/coresight/ |
D | coresight-etm-perf.c | 392 struct hw_perf_event *hwc = &event->hw; in etm_event_add() local 396 if (hwc->state & PERF_HES_STOPPED) in etm_event_add() 399 hwc->state = PERF_HES_STOPPED; in etm_event_add()
|
/drivers/gpu/drm/i915/ |
D | i915_pmu.c | 576 struct hw_perf_event *hwc = &event->hw; in i915_pmu_event_read() local 580 prev = local64_read(&hwc->prev_count); in i915_pmu_event_read() 583 if (local64_cmpxchg(&hwc->prev_count, prev, new) != prev) in i915_pmu_event_read()
|
/drivers/video/fbdev/mmp/hw/ |
D | mmp_ctrl.h | 534 #define CFG_CSB_256x32(hwc) ((hwc)<<15) /* HWC */ argument
|