Home
last modified time | relevance | path

Searched refs:smu (Results 1 – 25 of 41) sorted by relevance

12

/drivers/gpu/drm/amd/pm/swsmu/
Dsmu_internal.h30 #define smu_ppt_funcs(intf, ret, smu, args...) \ argument
31 ((smu)->ppt_funcs ? ((smu)->ppt_funcs->intf ? (smu)->ppt_funcs->intf(smu, ##args) : ret) : -EINVAL)
33 #define smu_init_microcode(smu) smu_ppt_funcs(init_microcode, 0, smu) argument
34 #define smu_fini_microcode(smu) smu_ppt_funcs(fini_microcode, 0, smu) argument
35 #define smu_init_smc_tables(smu) smu_ppt_funcs(init_smc_tables, 0, smu) argument
36 #define smu_fini_smc_tables(smu) smu_ppt_funcs(fini_smc_tables, 0, smu) argument
37 #define smu_init_power(smu) smu_ppt_funcs(init_power, 0, smu) argument
38 #define smu_fini_power(smu) smu_ppt_funcs(fini_power, 0, smu) argument
39 #define smu_setup_pptable(smu) smu_ppt_funcs(setup_pptable, 0, smu) argument
40 #define smu_powergate_sdma(smu, gate) smu_ppt_funcs(powergate_sdma, 0, smu, gate) argument
[all …]
Damdgpu_smu.c59 static int smu_force_smuclk_levels(struct smu_context *smu,
62 static int smu_handle_task(struct smu_context *smu,
65 static int smu_reset(struct smu_context *smu);
70 static int smu_set_gfx_cgpg(struct smu_context *smu, bool enabled);
76 struct smu_context *smu = handle; in smu_sys_get_pp_feature_mask() local
78 if (!smu->pm_enabled || !smu->adev->pm.dpm_enabled) in smu_sys_get_pp_feature_mask()
81 return smu_get_pp_feature_mask(smu, buf); in smu_sys_get_pp_feature_mask()
87 struct smu_context *smu = handle; in smu_sys_set_pp_feature_mask() local
89 if (!smu->pm_enabled || !smu->adev->pm.dpm_enabled) in smu_sys_set_pp_feature_mask()
92 return smu_set_pp_feature_mask(smu, new_mask); in smu_sys_set_pp_feature_mask()
[all …]
Dsmu_cmn.c48 #define smu_cmn_call_asic_func(intf, smu, args...) \ argument
49 ((smu)->ppt_funcs ? ((smu)->ppt_funcs->intf ? \
50 (smu)->ppt_funcs->intf(smu, ##args) : \
54 static const char *smu_get_message_name(struct smu_context *smu, in smu_get_message_name() argument
63 static void smu_cmn_read_arg(struct smu_context *smu, in smu_cmn_read_arg() argument
66 struct amdgpu_device *adev = smu->adev; in smu_cmn_read_arg()
68 *arg = RREG32(smu->param_reg); in smu_cmn_read_arg()
107 static u32 __smu_cmn_poll_stat(struct smu_context *smu) in __smu_cmn_poll_stat() argument
109 struct amdgpu_device *adev = smu->adev; in __smu_cmn_poll_stat()
114 reg = RREG32(smu->resp_reg); in __smu_cmn_poll_stat()
[all …]
Dsmu_cmn.h33 int smu_cmn_send_msg_without_waiting(struct smu_context *smu,
36 int smu_cmn_send_smc_msg_with_param(struct smu_context *smu,
41 int smu_cmn_send_smc_msg(struct smu_context *smu,
45 int smu_cmn_send_debug_smc_msg(struct smu_context *smu,
48 int smu_cmn_wait_for_response(struct smu_context *smu);
50 int smu_cmn_to_asic_specific_index(struct smu_context *smu,
54 int smu_cmn_feature_is_supported(struct smu_context *smu,
57 int smu_cmn_feature_is_enabled(struct smu_context *smu,
60 bool smu_cmn_clk_dpm_is_enabled(struct smu_context *smu,
63 int smu_cmn_get_enabled_mask(struct smu_context *smu,
[all …]
/drivers/gpu/drm/amd/pm/swsmu/smu13/
Dsmu_v13_0.c88 int smu_v13_0_init_microcode(struct smu_context *smu) in smu_v13_0_init_microcode() argument
90 struct amdgpu_device *adev = smu->adev; in smu_v13_0_init_microcode()
144 void smu_v13_0_fini_microcode(struct smu_context *smu) in smu_v13_0_fini_microcode() argument
146 struct amdgpu_device *adev = smu->adev; in smu_v13_0_fini_microcode()
153 int smu_v13_0_load_microcode(struct smu_context *smu) in smu_v13_0_load_microcode() argument
156 struct amdgpu_device *adev = smu->adev; in smu_v13_0_load_microcode()
195 int smu_v13_0_init_pptable_microcode(struct smu_context *smu) in smu_v13_0_init_pptable_microcode() argument
197 struct amdgpu_device *adev = smu->adev; in smu_v13_0_init_pptable_microcode()
223 pptable_id = smu->smu_table.boot_values.pp_table_id; in smu_v13_0_init_pptable_microcode()
230 ret = smu_v13_0_get_pptable_from_firmware(smu, &table, &size, pptable_id); in smu_v13_0_init_pptable_microcode()
[all …]
Dsmu_v13_0_5_ppt.c124 static int smu_v13_0_5_init_smc_tables(struct smu_context *smu) in smu_v13_0_5_init_smc_tables() argument
126 struct smu_table_context *smu_table = &smu->smu_table; in smu_v13_0_5_init_smc_tables()
166 static int smu_v13_0_5_fini_smc_tables(struct smu_context *smu) in smu_v13_0_5_fini_smc_tables() argument
168 struct smu_table_context *smu_table = &smu->smu_table; in smu_v13_0_5_fini_smc_tables()
185 static int smu_v13_0_5_system_features_control(struct smu_context *smu, bool en) in smu_v13_0_5_system_features_control() argument
187 struct amdgpu_device *adev = smu->adev; in smu_v13_0_5_system_features_control()
191 ret = smu_cmn_send_smc_msg(smu, SMU_MSG_PrepareMp1ForUnload, NULL); in smu_v13_0_5_system_features_control()
196 static int smu_v13_0_5_dpm_set_vcn_enable(struct smu_context *smu, bool enable) in smu_v13_0_5_dpm_set_vcn_enable() argument
202 ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_PowerUpVcn, in smu_v13_0_5_dpm_set_vcn_enable()
205 ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_PowerDownVcn, in smu_v13_0_5_dpm_set_vcn_enable()
[all …]
Dyellow_carp_ppt.c143 static int yellow_carp_init_smc_tables(struct smu_context *smu) in yellow_carp_init_smc_tables() argument
145 struct smu_table_context *smu_table = &smu->smu_table; in yellow_carp_init_smc_tables()
185 static int yellow_carp_fini_smc_tables(struct smu_context *smu) in yellow_carp_fini_smc_tables() argument
187 struct smu_table_context *smu_table = &smu->smu_table; in yellow_carp_fini_smc_tables()
204 static int yellow_carp_system_features_control(struct smu_context *smu, bool en) in yellow_carp_system_features_control() argument
206 struct amdgpu_device *adev = smu->adev; in yellow_carp_system_features_control()
210 ret = smu_cmn_send_smc_msg(smu, SMU_MSG_PrepareMp1ForUnload, NULL); in yellow_carp_system_features_control()
215 static int yellow_carp_dpm_set_vcn_enable(struct smu_context *smu, bool enable) in yellow_carp_dpm_set_vcn_enable() argument
221 ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_PowerUpVcn, in yellow_carp_dpm_set_vcn_enable()
224 ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_PowerDownVcn, in yellow_carp_dpm_set_vcn_enable()
[all …]
Daldebaran_ppt.c228 static int aldebaran_tables_init(struct smu_context *smu) in aldebaran_tables_init() argument
230 struct smu_table_context *smu_table = &smu->smu_table; in aldebaran_tables_init()
270 static int aldebaran_allocate_dpm_context(struct smu_context *smu) in aldebaran_allocate_dpm_context() argument
272 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in aldebaran_allocate_dpm_context()
283 static int aldebaran_init_smc_tables(struct smu_context *smu) in aldebaran_init_smc_tables() argument
287 ret = aldebaran_tables_init(smu); in aldebaran_init_smc_tables()
291 ret = aldebaran_allocate_dpm_context(smu); in aldebaran_init_smc_tables()
295 return smu_v13_0_init_smc_tables(smu); in aldebaran_init_smc_tables()
298 static int aldebaran_get_allowed_feature_mask(struct smu_context *smu, in aldebaran_get_allowed_feature_mask() argument
310 static int aldebaran_set_default_dpm_table(struct smu_context *smu) in aldebaran_set_default_dpm_table() argument
[all …]
Dsmu_v13_0_4_ppt.c146 static int smu_v13_0_4_init_smc_tables(struct smu_context *smu) in smu_v13_0_4_init_smc_tables() argument
148 struct smu_table_context *smu_table = &smu->smu_table; in smu_v13_0_4_init_smc_tables()
188 static int smu_v13_0_4_fini_smc_tables(struct smu_context *smu) in smu_v13_0_4_fini_smc_tables() argument
190 struct smu_table_context *smu_table = &smu->smu_table; in smu_v13_0_4_fini_smc_tables()
207 static bool smu_v13_0_4_is_dpm_running(struct smu_context *smu) in smu_v13_0_4_is_dpm_running() argument
212 ret = smu_cmn_get_enabled_mask(smu, &feature_enabled); in smu_v13_0_4_is_dpm_running()
220 static int smu_v13_0_4_system_features_control(struct smu_context *smu, bool en) in smu_v13_0_4_system_features_control() argument
222 struct amdgpu_device *adev = smu->adev; in smu_v13_0_4_system_features_control()
226 ret = smu_cmn_send_smc_msg(smu, SMU_MSG_PrepareMp1ForUnload, NULL); in smu_v13_0_4_system_features_control()
231 static ssize_t smu_v13_0_4_get_gpu_metrics(struct smu_context *smu, in smu_v13_0_4_get_gpu_metrics() argument
[all …]
Dsmu_v13_0_0_ppt.c268 smu_v13_0_0_get_allowed_feature_mask(struct smu_context *smu, in smu_v13_0_0_get_allowed_feature_mask() argument
271 struct amdgpu_device *adev = smu->adev; in smu_v13_0_0_get_allowed_feature_mask()
292 smu_cmn_get_smc_version(smu, NULL, &smu_version); in smu_v13_0_0_get_allowed_feature_mask()
317 static int smu_v13_0_0_check_powerplay_table(struct smu_context *smu) in smu_v13_0_0_check_powerplay_table() argument
319 struct smu_table_context *table_context = &smu->smu_table; in smu_v13_0_0_check_powerplay_table()
322 struct smu_baco_context *smu_baco = &smu->smu_baco; in smu_v13_0_0_check_powerplay_table()
325 smu->dc_controlled_by_gpio = true; in smu_v13_0_0_check_powerplay_table()
341 smu->od_settings = &powerplay_table->overdrive_table; in smu_v13_0_0_check_powerplay_table()
346 static int smu_v13_0_0_store_powerplay_table(struct smu_context *smu) in smu_v13_0_0_store_powerplay_table() argument
348 struct smu_table_context *table_context = &smu->smu_table; in smu_v13_0_0_store_powerplay_table()
[all …]
Dsmu_v13_0_7_ppt.c248 smu_v13_0_7_get_allowed_feature_mask(struct smu_context *smu, in smu_v13_0_7_get_allowed_feature_mask() argument
251 struct amdgpu_device *adev = smu->adev; in smu_v13_0_7_get_allowed_feature_mask()
317 static int smu_v13_0_7_check_powerplay_table(struct smu_context *smu) in smu_v13_0_7_check_powerplay_table() argument
319 struct smu_table_context *table_context = &smu->smu_table; in smu_v13_0_7_check_powerplay_table()
322 struct smu_baco_context *smu_baco = &smu->smu_baco; in smu_v13_0_7_check_powerplay_table()
327 smu->dc_controlled_by_gpio = true; in smu_v13_0_7_check_powerplay_table()
344 smu->od_settings = &powerplay_table->overdrive_table; in smu_v13_0_7_check_powerplay_table()
349 static int smu_v13_0_7_store_powerplay_table(struct smu_context *smu) in smu_v13_0_7_store_powerplay_table() argument
351 struct smu_table_context *table_context = &smu->smu_table; in smu_v13_0_7_store_powerplay_table()
354 struct amdgpu_device *adev = smu->adev; in smu_v13_0_7_store_powerplay_table()
[all …]
/drivers/gpu/drm/amd/pm/swsmu/inc/
Dsmu_v13_0.h130 int smu_v13_0_init_microcode(struct smu_context *smu);
132 void smu_v13_0_fini_microcode(struct smu_context *smu);
134 int smu_v13_0_load_microcode(struct smu_context *smu);
136 int smu_v13_0_init_smc_tables(struct smu_context *smu);
138 int smu_v13_0_fini_smc_tables(struct smu_context *smu);
140 int smu_v13_0_init_power(struct smu_context *smu);
142 int smu_v13_0_fini_power(struct smu_context *smu);
144 int smu_v13_0_check_fw_status(struct smu_context *smu);
146 int smu_v13_0_setup_pptable(struct smu_context *smu);
148 int smu_v13_0_get_vbios_bootup_values(struct smu_context *smu);
[all …]
Dsmu_v11_0.h152 int smu_v11_0_init_microcode(struct smu_context *smu);
154 void smu_v11_0_fini_microcode(struct smu_context *smu);
156 int smu_v11_0_load_microcode(struct smu_context *smu);
158 int smu_v11_0_init_smc_tables(struct smu_context *smu);
160 int smu_v11_0_fini_smc_tables(struct smu_context *smu);
162 int smu_v11_0_init_power(struct smu_context *smu);
164 int smu_v11_0_fini_power(struct smu_context *smu);
166 int smu_v11_0_check_fw_status(struct smu_context *smu);
168 int smu_v11_0_setup_pptable(struct smu_context *smu);
170 int smu_v11_0_get_vbios_bootup_values(struct smu_context *smu);
[all …]
Damdgpu_smu.h591 int (*run_btc)(struct smu_context *smu);
598 int (*get_allowed_feature_mask)(struct smu_context *smu, uint32_t *feature_mask, uint32_t num);
605 enum amd_pm_state_type (*get_current_power_state)(struct smu_context *smu);
611 int (*set_default_dpm_table)(struct smu_context *smu);
613 int (*set_power_state)(struct smu_context *smu);
619 int (*populate_umd_state_clk)(struct smu_context *smu);
628 int (*print_clk_levels)(struct smu_context *smu, enum smu_clk_type clk_type, char *buf);
641 …int (*emit_clk_levels)(struct smu_context *smu, enum smu_clk_type clk_type, char *buf, int *offset…
649 int (*force_clk_levels)(struct smu_context *smu, enum smu_clk_type clk_type, uint32_t mask);
657 int (*od_edit_dpm_table)(struct smu_context *smu,
[all …]
Dsmu_v12_0.h36 int smu_v12_0_check_fw_status(struct smu_context *smu);
38 int smu_v12_0_check_fw_version(struct smu_context *smu);
40 int smu_v12_0_powergate_sdma(struct smu_context *smu, bool gate);
42 int smu_v12_0_powergate_vcn(struct smu_context *smu, bool gate);
44 int smu_v12_0_powergate_jpeg(struct smu_context *smu, bool gate);
46 int smu_v12_0_set_gfx_cgpg(struct smu_context *smu, bool enable);
48 uint32_t smu_v12_0_get_gfxoff_status(struct smu_context *smu);
50 int smu_v12_0_gfx_off_control(struct smu_context *smu, bool enable);
52 int smu_v12_0_fini_smc_tables(struct smu_context *smu);
54 int smu_v12_0_set_default_dpm_tables(struct smu_context *smu);
[all …]
/drivers/gpu/drm/amd/pm/swsmu/smu12/
Dsmu_v12_0.c58 int smu_v12_0_check_fw_status(struct smu_context *smu) in smu_v12_0_check_fw_status() argument
60 struct amdgpu_device *adev = smu->adev; in smu_v12_0_check_fw_status()
73 int smu_v12_0_check_fw_version(struct smu_context *smu) in smu_v12_0_check_fw_version() argument
75 struct amdgpu_device *adev = smu->adev; in smu_v12_0_check_fw_version()
80 ret = smu_cmn_get_smc_version(smu, &if_version, &smu_version); in smu_v12_0_check_fw_version()
88 if (smu->is_apu) in smu_v12_0_check_fw_version()
99 if (if_version != smu->smc_driver_if_version) { in smu_v12_0_check_fw_version()
100 dev_info(smu->adev->dev, "smu driver if version = 0x%08x, smu fw if version = 0x%08x, " in smu_v12_0_check_fw_version()
102 smu->smc_driver_if_version, if_version, in smu_v12_0_check_fw_version()
104 dev_warn(smu->adev->dev, "SMU driver if version not matched\n"); in smu_v12_0_check_fw_version()
[all …]
Drenoir_ppt.c156 static int renoir_init_smc_tables(struct smu_context *smu) in renoir_init_smc_tables() argument
158 struct smu_table_context *smu_table = &smu->smu_table; in renoir_init_smc_tables()
202 static int renoir_get_dpm_clk_limited(struct smu_context *smu, enum smu_clk_type clk_type, in renoir_get_dpm_clk_limited() argument
205 DpmClocks_t *clk_table = smu->smu_table.clocks_table; in renoir_get_dpm_clk_limited()
250 static int renoir_get_profiling_clk_mask(struct smu_context *smu, in renoir_get_profiling_clk_mask() argument
280 static int renoir_get_dpm_ultimate_freq(struct smu_context *smu, in renoir_get_dpm_ultimate_freq() argument
289 if (!smu_cmn_clk_dpm_is_enabled(smu, clk_type)) { in renoir_get_dpm_ultimate_freq()
293 clock_limit = smu->smu_table.boot_values.uclk; in renoir_get_dpm_ultimate_freq()
297 clock_limit = smu->smu_table.boot_values.gfxclk; in renoir_get_dpm_ultimate_freq()
300 clock_limit = smu->smu_table.boot_values.socclk; in renoir_get_dpm_ultimate_freq()
[all …]
/drivers/gpu/drm/amd/pm/swsmu/smu11/
Dsmu_v11_0.c82 int smu_v11_0_init_microcode(struct smu_context *smu) in smu_v11_0_init_microcode() argument
84 struct amdgpu_device *adev = smu->adev; in smu_v11_0_init_microcode()
160 void smu_v11_0_fini_microcode(struct smu_context *smu) in smu_v11_0_fini_microcode() argument
162 struct amdgpu_device *adev = smu->adev; in smu_v11_0_fini_microcode()
169 int smu_v11_0_load_microcode(struct smu_context *smu) in smu_v11_0_load_microcode() argument
171 struct amdgpu_device *adev = smu->adev; in smu_v11_0_load_microcode()
209 int smu_v11_0_check_fw_status(struct smu_context *smu) in smu_v11_0_check_fw_status() argument
211 struct amdgpu_device *adev = smu->adev; in smu_v11_0_check_fw_status()
224 int smu_v11_0_check_fw_version(struct smu_context *smu) in smu_v11_0_check_fw_version() argument
226 struct amdgpu_device *adev = smu->adev; in smu_v11_0_check_fw_version()
[all …]
Dsienna_cichlid_ppt.c77 if (smu->adev->ip_versions[MP1_HWIP][0] == IP_VERSION(11, 0, 13))\
78 (*member) = (smu->smu_table.driver_pptable + offsetof(PPTable_beige_goby_t, field));\
80 (*member) = (smu->smu_table.driver_pptable + offsetof(PPTable_t, field));\
92 static int get_table_size(struct smu_context *smu) in get_table_size() argument
94 if (smu->adev->ip_versions[MP1_HWIP][0] == IP_VERSION(11, 0, 13)) in get_table_size()
275 sienna_cichlid_get_allowed_feature_mask(struct smu_context *smu, in sienna_cichlid_get_allowed_feature_mask() argument
278 struct amdgpu_device *adev = smu->adev; in sienna_cichlid_get_allowed_feature_mask()
339 if (smu->adev->pg_flags & AMD_PG_SUPPORT_ATHUB) in sienna_cichlid_get_allowed_feature_mask()
342 if (smu->adev->pg_flags & AMD_PG_SUPPORT_MMHUB) in sienna_cichlid_get_allowed_feature_mask()
345 if (smu->adev->pg_flags & AMD_PG_SUPPORT_VCN || in sienna_cichlid_get_allowed_feature_mask()
[all …]
Dvangogh_ppt.c222 static int vangogh_tables_init(struct smu_context *smu) in vangogh_tables_init() argument
224 struct smu_table_context *smu_table = &smu->smu_table; in vangogh_tables_init()
230 ret = smu_cmn_get_smc_version(smu, &if_version, &smu_version); in vangogh_tables_init()
285 static int vangogh_get_legacy_smu_metrics_data(struct smu_context *smu, in vangogh_get_legacy_smu_metrics_data() argument
289 struct smu_table_context *smu_table = &smu->smu_table; in vangogh_get_legacy_smu_metrics_data()
293 ret = smu_cmn_get_metrics_table(smu, in vangogh_get_legacy_smu_metrics_data()
344 smu->cpu_core_num * sizeof(uint16_t)); in vangogh_get_legacy_smu_metrics_data()
354 static int vangogh_get_smu_metrics_data(struct smu_context *smu, in vangogh_get_smu_metrics_data() argument
358 struct smu_table_context *smu_table = &smu->smu_table; in vangogh_get_smu_metrics_data()
362 ret = smu_cmn_get_metrics_table(smu, in vangogh_get_smu_metrics_data()
[all …]
Darcturus_ppt.c250 static int arcturus_tables_init(struct smu_context *smu) in arcturus_tables_init() argument
252 struct smu_table_context *smu_table = &smu->smu_table; in arcturus_tables_init()
286 static int arcturus_allocate_dpm_context(struct smu_context *smu) in arcturus_allocate_dpm_context() argument
288 struct smu_dpm_context *smu_dpm = &smu->smu_dpm; in arcturus_allocate_dpm_context()
299 static int arcturus_init_smc_tables(struct smu_context *smu) in arcturus_init_smc_tables() argument
303 ret = arcturus_tables_init(smu); in arcturus_init_smc_tables()
307 ret = arcturus_allocate_dpm_context(smu); in arcturus_init_smc_tables()
311 return smu_v11_0_init_smc_tables(smu); in arcturus_init_smc_tables()
315 arcturus_get_allowed_feature_mask(struct smu_context *smu, in arcturus_get_allowed_feature_mask() argument
327 static int arcturus_set_default_dpm_table(struct smu_context *smu) in arcturus_set_default_dpm_table() argument
[all …]
Dnavi10_ppt.c262 static bool is_asic_secure(struct smu_context *smu) in is_asic_secure() argument
264 struct amdgpu_device *adev = smu->adev; in is_asic_secure()
278 navi10_get_allowed_feature_mask(struct smu_context *smu, in navi10_get_allowed_feature_mask() argument
281 struct amdgpu_device *adev = smu->adev; in navi10_get_allowed_feature_mask()
327 if (smu->adev->pg_flags & AMD_PG_SUPPORT_MMHUB) in navi10_get_allowed_feature_mask()
330 if (smu->adev->pg_flags & AMD_PG_SUPPORT_ATHUB) in navi10_get_allowed_feature_mask()
333 if (smu->adev->pg_flags & AMD_PG_SUPPORT_VCN) in navi10_get_allowed_feature_mask()
336 if (smu->adev->pg_flags & AMD_PG_SUPPORT_JPEG) in navi10_get_allowed_feature_mask()
339 if (smu->dc_controlled_by_gpio) in navi10_get_allowed_feature_mask()
346 if (!(is_asic_secure(smu) && in navi10_get_allowed_feature_mask()
[all …]
Dcyan_skillfish_ppt.c87 static int cyan_skillfish_tables_init(struct smu_context *smu) in cyan_skillfish_tables_init() argument
89 struct smu_table_context *smu_table = &smu->smu_table; in cyan_skillfish_tables_init()
117 static int cyan_skillfish_init_smc_tables(struct smu_context *smu) in cyan_skillfish_init_smc_tables() argument
121 ret = cyan_skillfish_tables_init(smu); in cyan_skillfish_init_smc_tables()
125 return smu_v11_0_init_smc_tables(smu); in cyan_skillfish_init_smc_tables()
129 cyan_skillfish_get_smu_metrics_data(struct smu_context *smu, in cyan_skillfish_get_smu_metrics_data() argument
133 struct smu_table_context *smu_table = &smu->smu_table; in cyan_skillfish_get_smu_metrics_data()
137 ret = smu_cmn_get_metrics_table(smu, NULL, false); in cyan_skillfish_get_smu_metrics_data()
186 static int cyan_skillfish_read_sensor(struct smu_context *smu, in cyan_skillfish_read_sensor() argument
198 ret = cyan_skillfish_get_smu_metrics_data(smu, in cyan_skillfish_read_sensor()
[all …]
/drivers/macintosh/
Dsmu.c98 static struct smu_device *smu; variable
114 if (list_empty(&smu->cmd_list)) in smu_start_cmd()
118 cmd = list_entry(smu->cmd_list.next, struct smu_cmd, link); in smu_start_cmd()
119 smu->cmd_cur = cmd; in smu_start_cmd()
127 smu->cmd_buf->cmd = cmd->cmd; in smu_start_cmd()
128 smu->cmd_buf->length = cmd->data_len; in smu_start_cmd()
129 memcpy(smu->cmd_buf->data, cmd->data_buf, cmd->data_len); in smu_start_cmd()
132 faddr = (unsigned long)smu->cmd_buf; in smu_start_cmd()
133 fend = faddr + smu->cmd_buf->length + 2; in smu_start_cmd()
146 if (smu->broken_nap) in smu_start_cmd()
[all …]
/drivers/mtd/nand/raw/atmel/
Dpmecc.c175 s16 *smu; member
374 user->smu = user->lmu + (req->ecc.strength + 1); in atmel_pmecc_create_user()
375 user->mu = (s32 *)PTR_ALIGN(user->smu + in atmel_pmecc_create_user()
502 s16 *smu = user->smu; in atmel_pmecc_get_sigma() local
516 memset(smu, 0, sizeof(s16) * num); in atmel_pmecc_get_sigma()
517 smu[0] = 1; in atmel_pmecc_get_sigma()
530 memset(&smu[num], 0, sizeof(s16) * num); in atmel_pmecc_get_sigma()
531 smu[num] = 1; in atmel_pmecc_get_sigma()
542 memset(&smu[(strength + 1) * num], 0, sizeof(s16) * num); in atmel_pmecc_get_sigma()
559 smu[(strength + 1) * num + j] = in atmel_pmecc_get_sigma()
[all …]

12