1 /*
2 * sysfs.c - ACPI sysfs interface to userspace.
3 */
4
5 #include <linux/init.h>
6 #include <linux/kernel.h>
7 #include <linux/moduleparam.h>
8 #include <linux/acpi.h>
9
10 #include "internal.h"
11
12 #define _COMPONENT ACPI_SYSTEM_COMPONENT
13 ACPI_MODULE_NAME("sysfs");
14
15 #ifdef CONFIG_ACPI_DEBUG
16 /*
17 * ACPI debug sysfs I/F, including:
18 * /sys/modules/acpi/parameters/debug_layer
19 * /sys/modules/acpi/parameters/debug_level
20 * /sys/modules/acpi/parameters/trace_method_name
21 * /sys/modules/acpi/parameters/trace_state
22 * /sys/modules/acpi/parameters/trace_debug_layer
23 * /sys/modules/acpi/parameters/trace_debug_level
24 */
25
26 struct acpi_dlayer {
27 const char *name;
28 unsigned long value;
29 };
30 struct acpi_dlevel {
31 const char *name;
32 unsigned long value;
33 };
34 #define ACPI_DEBUG_INIT(v) { .name = #v, .value = v }
35
36 static const struct acpi_dlayer acpi_debug_layers[] = {
37 ACPI_DEBUG_INIT(ACPI_UTILITIES),
38 ACPI_DEBUG_INIT(ACPI_HARDWARE),
39 ACPI_DEBUG_INIT(ACPI_EVENTS),
40 ACPI_DEBUG_INIT(ACPI_TABLES),
41 ACPI_DEBUG_INIT(ACPI_NAMESPACE),
42 ACPI_DEBUG_INIT(ACPI_PARSER),
43 ACPI_DEBUG_INIT(ACPI_DISPATCHER),
44 ACPI_DEBUG_INIT(ACPI_EXECUTER),
45 ACPI_DEBUG_INIT(ACPI_RESOURCES),
46 ACPI_DEBUG_INIT(ACPI_CA_DEBUGGER),
47 ACPI_DEBUG_INIT(ACPI_OS_SERVICES),
48 ACPI_DEBUG_INIT(ACPI_CA_DISASSEMBLER),
49 ACPI_DEBUG_INIT(ACPI_COMPILER),
50 ACPI_DEBUG_INIT(ACPI_TOOLS),
51
52 ACPI_DEBUG_INIT(ACPI_BUS_COMPONENT),
53 ACPI_DEBUG_INIT(ACPI_AC_COMPONENT),
54 ACPI_DEBUG_INIT(ACPI_BATTERY_COMPONENT),
55 ACPI_DEBUG_INIT(ACPI_BUTTON_COMPONENT),
56 ACPI_DEBUG_INIT(ACPI_SBS_COMPONENT),
57 ACPI_DEBUG_INIT(ACPI_FAN_COMPONENT),
58 ACPI_DEBUG_INIT(ACPI_PCI_COMPONENT),
59 ACPI_DEBUG_INIT(ACPI_POWER_COMPONENT),
60 ACPI_DEBUG_INIT(ACPI_CONTAINER_COMPONENT),
61 ACPI_DEBUG_INIT(ACPI_SYSTEM_COMPONENT),
62 ACPI_DEBUG_INIT(ACPI_THERMAL_COMPONENT),
63 ACPI_DEBUG_INIT(ACPI_MEMORY_DEVICE_COMPONENT),
64 ACPI_DEBUG_INIT(ACPI_VIDEO_COMPONENT),
65 ACPI_DEBUG_INIT(ACPI_PROCESSOR_COMPONENT),
66 };
67
68 static const struct acpi_dlevel acpi_debug_levels[] = {
69 ACPI_DEBUG_INIT(ACPI_LV_INIT),
70 ACPI_DEBUG_INIT(ACPI_LV_DEBUG_OBJECT),
71 ACPI_DEBUG_INIT(ACPI_LV_INFO),
72 ACPI_DEBUG_INIT(ACPI_LV_REPAIR),
73 ACPI_DEBUG_INIT(ACPI_LV_TRACE_POINT),
74
75 ACPI_DEBUG_INIT(ACPI_LV_INIT_NAMES),
76 ACPI_DEBUG_INIT(ACPI_LV_PARSE),
77 ACPI_DEBUG_INIT(ACPI_LV_LOAD),
78 ACPI_DEBUG_INIT(ACPI_LV_DISPATCH),
79 ACPI_DEBUG_INIT(ACPI_LV_EXEC),
80 ACPI_DEBUG_INIT(ACPI_LV_NAMES),
81 ACPI_DEBUG_INIT(ACPI_LV_OPREGION),
82 ACPI_DEBUG_INIT(ACPI_LV_BFIELD),
83 ACPI_DEBUG_INIT(ACPI_LV_TABLES),
84 ACPI_DEBUG_INIT(ACPI_LV_VALUES),
85 ACPI_DEBUG_INIT(ACPI_LV_OBJECTS),
86 ACPI_DEBUG_INIT(ACPI_LV_RESOURCES),
87 ACPI_DEBUG_INIT(ACPI_LV_USER_REQUESTS),
88 ACPI_DEBUG_INIT(ACPI_LV_PACKAGE),
89
90 ACPI_DEBUG_INIT(ACPI_LV_ALLOCATIONS),
91 ACPI_DEBUG_INIT(ACPI_LV_FUNCTIONS),
92 ACPI_DEBUG_INIT(ACPI_LV_OPTIMIZATIONS),
93
94 ACPI_DEBUG_INIT(ACPI_LV_MUTEX),
95 ACPI_DEBUG_INIT(ACPI_LV_THREADS),
96 ACPI_DEBUG_INIT(ACPI_LV_IO),
97 ACPI_DEBUG_INIT(ACPI_LV_INTERRUPTS),
98
99 ACPI_DEBUG_INIT(ACPI_LV_AML_DISASSEMBLE),
100 ACPI_DEBUG_INIT(ACPI_LV_VERBOSE_INFO),
101 ACPI_DEBUG_INIT(ACPI_LV_FULL_TABLES),
102 ACPI_DEBUG_INIT(ACPI_LV_EVENTS),
103 };
104
param_get_debug_layer(char * buffer,const struct kernel_param * kp)105 static int param_get_debug_layer(char *buffer, const struct kernel_param *kp)
106 {
107 int result = 0;
108 int i;
109
110 result = sprintf(buffer, "%-25s\tHex SET\n", "Description");
111
112 for (i = 0; i < ARRAY_SIZE(acpi_debug_layers); i++) {
113 result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n",
114 acpi_debug_layers[i].name,
115 acpi_debug_layers[i].value,
116 (acpi_dbg_layer & acpi_debug_layers[i].value)
117 ? '*' : ' ');
118 }
119 result +=
120 sprintf(buffer + result, "%-25s\t0x%08X [%c]\n", "ACPI_ALL_DRIVERS",
121 ACPI_ALL_DRIVERS,
122 (acpi_dbg_layer & ACPI_ALL_DRIVERS) ==
123 ACPI_ALL_DRIVERS ? '*' : (acpi_dbg_layer & ACPI_ALL_DRIVERS)
124 == 0 ? ' ' : '-');
125 result +=
126 sprintf(buffer + result,
127 "--\ndebug_layer = 0x%08X ( * = enabled)\n",
128 acpi_dbg_layer);
129
130 return result;
131 }
132
param_get_debug_level(char * buffer,const struct kernel_param * kp)133 static int param_get_debug_level(char *buffer, const struct kernel_param *kp)
134 {
135 int result = 0;
136 int i;
137
138 result = sprintf(buffer, "%-25s\tHex SET\n", "Description");
139
140 for (i = 0; i < ARRAY_SIZE(acpi_debug_levels); i++) {
141 result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n",
142 acpi_debug_levels[i].name,
143 acpi_debug_levels[i].value,
144 (acpi_dbg_level & acpi_debug_levels[i].value)
145 ? '*' : ' ');
146 }
147 result +=
148 sprintf(buffer + result, "--\ndebug_level = 0x%08X (* = enabled)\n",
149 acpi_dbg_level);
150
151 return result;
152 }
153
154 static const struct kernel_param_ops param_ops_debug_layer = {
155 .set = param_set_uint,
156 .get = param_get_debug_layer,
157 };
158
159 static const struct kernel_param_ops param_ops_debug_level = {
160 .set = param_set_uint,
161 .get = param_get_debug_level,
162 };
163
164 module_param_cb(debug_layer, ¶m_ops_debug_layer, &acpi_dbg_layer, 0644);
165 module_param_cb(debug_level, ¶m_ops_debug_level, &acpi_dbg_level, 0644);
166
167 static char trace_method_name[1024];
168
param_set_trace_method_name(const char * val,const struct kernel_param * kp)169 int param_set_trace_method_name(const char *val, const struct kernel_param *kp)
170 {
171 u32 saved_flags = 0;
172 bool is_abs_path = true;
173
174 if (*val != '\\')
175 is_abs_path = false;
176
177 if ((is_abs_path && strlen(val) > 1023) ||
178 (!is_abs_path && strlen(val) > 1022)) {
179 pr_err("%s: string parameter too long\n", kp->name);
180 return -ENOSPC;
181 }
182
183 /*
184 * It's not safe to update acpi_gbl_trace_method_name without
185 * having the tracer stopped, so we save the original tracer
186 * state and disable it.
187 */
188 saved_flags = acpi_gbl_trace_flags;
189 (void)acpi_debug_trace(NULL,
190 acpi_gbl_trace_dbg_level,
191 acpi_gbl_trace_dbg_layer,
192 0);
193
194 /* This is a hack. We can't kmalloc in early boot. */
195 if (is_abs_path)
196 strcpy(trace_method_name, val);
197 else {
198 trace_method_name[0] = '\\';
199 strcpy(trace_method_name+1, val);
200 }
201
202 /* Restore the original tracer state */
203 (void)acpi_debug_trace(trace_method_name,
204 acpi_gbl_trace_dbg_level,
205 acpi_gbl_trace_dbg_layer,
206 saved_flags);
207
208 return 0;
209 }
210
param_get_trace_method_name(char * buffer,const struct kernel_param * kp)211 static int param_get_trace_method_name(char *buffer, const struct kernel_param *kp)
212 {
213 return scnprintf(buffer, PAGE_SIZE, "%s", acpi_gbl_trace_method_name);
214 }
215
216 static const struct kernel_param_ops param_ops_trace_method = {
217 .set = param_set_trace_method_name,
218 .get = param_get_trace_method_name,
219 };
220
221 static const struct kernel_param_ops param_ops_trace_attrib = {
222 .set = param_set_uint,
223 .get = param_get_uint,
224 };
225
226 module_param_cb(trace_method_name, ¶m_ops_trace_method, &trace_method_name, 0644);
227 module_param_cb(trace_debug_layer, ¶m_ops_trace_attrib, &acpi_gbl_trace_dbg_layer, 0644);
228 module_param_cb(trace_debug_level, ¶m_ops_trace_attrib, &acpi_gbl_trace_dbg_level, 0644);
229
param_set_trace_state(const char * val,const struct kernel_param * kp)230 static int param_set_trace_state(const char *val,
231 const struct kernel_param *kp)
232 {
233 acpi_status status;
234 const char *method = trace_method_name;
235 u32 flags = 0;
236
237 /* So "xxx-once" comparison should go prior than "xxx" comparison */
238 #define acpi_compare_param(val, key) \
239 strncmp((val), (key), sizeof(key) - 1)
240
241 if (!acpi_compare_param(val, "enable")) {
242 method = NULL;
243 flags = ACPI_TRACE_ENABLED;
244 } else if (!acpi_compare_param(val, "disable"))
245 method = NULL;
246 else if (!acpi_compare_param(val, "method-once"))
247 flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT;
248 else if (!acpi_compare_param(val, "method"))
249 flags = ACPI_TRACE_ENABLED;
250 else if (!acpi_compare_param(val, "opcode-once"))
251 flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT | ACPI_TRACE_OPCODE;
252 else if (!acpi_compare_param(val, "opcode"))
253 flags = ACPI_TRACE_ENABLED | ACPI_TRACE_OPCODE;
254 else
255 return -EINVAL;
256
257 status = acpi_debug_trace(method,
258 acpi_gbl_trace_dbg_level,
259 acpi_gbl_trace_dbg_layer,
260 flags);
261 if (ACPI_FAILURE(status))
262 return -EBUSY;
263
264 return 0;
265 }
266
param_get_trace_state(char * buffer,const struct kernel_param * kp)267 static int param_get_trace_state(char *buffer, const struct kernel_param *kp)
268 {
269 if (!(acpi_gbl_trace_flags & ACPI_TRACE_ENABLED))
270 return sprintf(buffer, "disable");
271 else {
272 if (acpi_gbl_trace_method_name) {
273 if (acpi_gbl_trace_flags & ACPI_TRACE_ONESHOT)
274 return sprintf(buffer, "method-once");
275 else
276 return sprintf(buffer, "method");
277 } else
278 return sprintf(buffer, "enable");
279 }
280 return 0;
281 }
282
283 module_param_call(trace_state, param_set_trace_state, param_get_trace_state,
284 NULL, 0644);
285 #endif /* CONFIG_ACPI_DEBUG */
286
287
288 /* /sys/modules/acpi/parameters/aml_debug_output */
289
290 module_param_named(aml_debug_output, acpi_gbl_enable_aml_debug_object,
291 byte, 0644);
292 MODULE_PARM_DESC(aml_debug_output,
293 "To enable/disable the ACPI Debug Object output.");
294
295 /* /sys/module/acpi/parameters/acpica_version */
param_get_acpica_version(char * buffer,const struct kernel_param * kp)296 static int param_get_acpica_version(char *buffer,
297 const struct kernel_param *kp)
298 {
299 int result;
300
301 result = sprintf(buffer, "%x", ACPI_CA_VERSION);
302
303 return result;
304 }
305
306 module_param_call(acpica_version, NULL, param_get_acpica_version, NULL, 0444);
307
308 /*
309 * ACPI table sysfs I/F:
310 * /sys/firmware/acpi/tables/
311 * /sys/firmware/acpi/tables/dynamic/
312 */
313
314 static LIST_HEAD(acpi_table_attr_list);
315 static struct kobject *tables_kobj;
316 static struct kobject *dynamic_tables_kobj;
317 static struct kobject *hotplug_kobj;
318
319 #define ACPI_MAX_TABLE_INSTANCES 999
320 #define ACPI_INST_SIZE 4 /* including trailing 0 */
321
322 struct acpi_table_attr {
323 struct bin_attribute attr;
324 char name[ACPI_NAME_SIZE];
325 int instance;
326 char filename[ACPI_NAME_SIZE+ACPI_INST_SIZE];
327 struct list_head node;
328 };
329
acpi_table_show(struct file * filp,struct kobject * kobj,struct bin_attribute * bin_attr,char * buf,loff_t offset,size_t count)330 static ssize_t acpi_table_show(struct file *filp, struct kobject *kobj,
331 struct bin_attribute *bin_attr, char *buf,
332 loff_t offset, size_t count)
333 {
334 struct acpi_table_attr *table_attr =
335 container_of(bin_attr, struct acpi_table_attr, attr);
336 struct acpi_table_header *table_header = NULL;
337 acpi_status status;
338
339 status = acpi_get_table(table_attr->name, table_attr->instance,
340 &table_header);
341 if (ACPI_FAILURE(status))
342 return -ENODEV;
343
344 return memory_read_from_buffer(buf, count, &offset,
345 table_header, table_header->length);
346 }
347
acpi_table_attr_init(struct kobject * tables_obj,struct acpi_table_attr * table_attr,struct acpi_table_header * table_header)348 static int acpi_table_attr_init(struct kobject *tables_obj,
349 struct acpi_table_attr *table_attr,
350 struct acpi_table_header *table_header)
351 {
352 struct acpi_table_header *header = NULL;
353 struct acpi_table_attr *attr = NULL;
354 char instance_str[ACPI_INST_SIZE];
355
356 sysfs_attr_init(&table_attr->attr.attr);
357 ACPI_MOVE_NAME(table_attr->name, table_header->signature);
358
359 list_for_each_entry(attr, &acpi_table_attr_list, node) {
360 if (ACPI_COMPARE_NAME(table_attr->name, attr->name))
361 if (table_attr->instance < attr->instance)
362 table_attr->instance = attr->instance;
363 }
364 table_attr->instance++;
365 if (table_attr->instance > ACPI_MAX_TABLE_INSTANCES) {
366 pr_warn("%4.4s: too many table instances\n",
367 table_attr->name);
368 return -ERANGE;
369 }
370
371 ACPI_MOVE_NAME(table_attr->filename, table_header->signature);
372 table_attr->filename[ACPI_NAME_SIZE] = '\0';
373 if (table_attr->instance > 1 || (table_attr->instance == 1 &&
374 !acpi_get_table
375 (table_header->signature, 2, &header))) {
376 snprintf(instance_str, sizeof(instance_str), "%u",
377 table_attr->instance);
378 strcat(table_attr->filename, instance_str);
379 }
380
381 table_attr->attr.size = table_header->length;
382 table_attr->attr.read = acpi_table_show;
383 table_attr->attr.attr.name = table_attr->filename;
384 table_attr->attr.attr.mode = 0400;
385
386 return sysfs_create_bin_file(tables_obj, &table_attr->attr);
387 }
388
acpi_sysfs_table_handler(u32 event,void * table,void * context)389 acpi_status acpi_sysfs_table_handler(u32 event, void *table, void *context)
390 {
391 struct acpi_table_attr *table_attr;
392
393 switch (event) {
394 case ACPI_TABLE_EVENT_INSTALL:
395 table_attr =
396 kzalloc(sizeof(struct acpi_table_attr), GFP_KERNEL);
397 if (!table_attr)
398 return AE_NO_MEMORY;
399
400 if (acpi_table_attr_init(dynamic_tables_kobj,
401 table_attr, table)) {
402 kfree(table_attr);
403 return AE_ERROR;
404 }
405 list_add_tail(&table_attr->node, &acpi_table_attr_list);
406 break;
407 case ACPI_TABLE_EVENT_LOAD:
408 case ACPI_TABLE_EVENT_UNLOAD:
409 case ACPI_TABLE_EVENT_UNINSTALL:
410 /*
411 * we do not need to do anything right now
412 * because the table is not deleted from the
413 * global table list when unloading it.
414 */
415 break;
416 default:
417 return AE_BAD_PARAMETER;
418 }
419 return AE_OK;
420 }
421
acpi_tables_sysfs_init(void)422 static int acpi_tables_sysfs_init(void)
423 {
424 struct acpi_table_attr *table_attr;
425 struct acpi_table_header *table_header = NULL;
426 int table_index;
427 acpi_status status;
428 int ret;
429
430 tables_kobj = kobject_create_and_add("tables", acpi_kobj);
431 if (!tables_kobj)
432 goto err;
433
434 dynamic_tables_kobj = kobject_create_and_add("dynamic", tables_kobj);
435 if (!dynamic_tables_kobj)
436 goto err_dynamic_tables;
437
438 for (table_index = 0;; table_index++) {
439 status = acpi_get_table_by_index(table_index, &table_header);
440
441 if (status == AE_BAD_PARAMETER)
442 break;
443
444 if (ACPI_FAILURE(status))
445 continue;
446
447 table_attr = kzalloc(sizeof(*table_attr), GFP_KERNEL);
448 if (!table_attr)
449 return -ENOMEM;
450
451 ret = acpi_table_attr_init(tables_kobj,
452 table_attr, table_header);
453 if (ret) {
454 kfree(table_attr);
455 return ret;
456 }
457 list_add_tail(&table_attr->node, &acpi_table_attr_list);
458 }
459
460 kobject_uevent(tables_kobj, KOBJ_ADD);
461 kobject_uevent(dynamic_tables_kobj, KOBJ_ADD);
462
463 return 0;
464 err_dynamic_tables:
465 kobject_put(tables_kobj);
466 err:
467 return -ENOMEM;
468 }
469
470 /*
471 * Detailed ACPI IRQ counters:
472 * /sys/firmware/acpi/interrupts/
473 */
474
475 u32 acpi_irq_handled;
476 u32 acpi_irq_not_handled;
477
478 #define COUNT_GPE 0
479 #define COUNT_SCI 1 /* acpi_irq_handled */
480 #define COUNT_SCI_NOT 2 /* acpi_irq_not_handled */
481 #define COUNT_ERROR 3 /* other */
482 #define NUM_COUNTERS_EXTRA 4
483
484 struct event_counter {
485 u32 count;
486 u32 flags;
487 };
488
489 static struct event_counter *all_counters;
490 static u32 num_gpes;
491 static u32 num_counters;
492 static struct attribute **all_attrs;
493 static u32 acpi_gpe_count;
494
495 static struct attribute_group interrupt_stats_attr_group = {
496 .name = "interrupts",
497 };
498
499 static struct kobj_attribute *counter_attrs;
500
delete_gpe_attr_array(void)501 static void delete_gpe_attr_array(void)
502 {
503 struct event_counter *tmp = all_counters;
504
505 all_counters = NULL;
506 kfree(tmp);
507
508 if (counter_attrs) {
509 int i;
510
511 for (i = 0; i < num_gpes; i++)
512 kfree(counter_attrs[i].attr.name);
513
514 kfree(counter_attrs);
515 }
516 kfree(all_attrs);
517
518 return;
519 }
520
gpe_count(u32 gpe_number)521 static void gpe_count(u32 gpe_number)
522 {
523 acpi_gpe_count++;
524
525 if (!all_counters)
526 return;
527
528 if (gpe_number < num_gpes)
529 all_counters[gpe_number].count++;
530 else
531 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS +
532 COUNT_ERROR].count++;
533
534 return;
535 }
536
fixed_event_count(u32 event_number)537 static void fixed_event_count(u32 event_number)
538 {
539 if (!all_counters)
540 return;
541
542 if (event_number < ACPI_NUM_FIXED_EVENTS)
543 all_counters[num_gpes + event_number].count++;
544 else
545 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS +
546 COUNT_ERROR].count++;
547
548 return;
549 }
550
acpi_global_event_handler(u32 event_type,acpi_handle device,u32 event_number,void * context)551 static void acpi_global_event_handler(u32 event_type, acpi_handle device,
552 u32 event_number, void *context)
553 {
554 if (event_type == ACPI_EVENT_TYPE_GPE)
555 gpe_count(event_number);
556
557 if (event_type == ACPI_EVENT_TYPE_FIXED)
558 fixed_event_count(event_number);
559 }
560
get_status(u32 index,acpi_event_status * status,acpi_handle * handle)561 static int get_status(u32 index, acpi_event_status *status,
562 acpi_handle *handle)
563 {
564 int result;
565
566 if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS)
567 return -EINVAL;
568
569 if (index < num_gpes) {
570 result = acpi_get_gpe_device(index, handle);
571 if (result) {
572 ACPI_EXCEPTION((AE_INFO, AE_NOT_FOUND,
573 "Invalid GPE 0x%x", index));
574 return result;
575 }
576 result = acpi_get_gpe_status(*handle, index, status);
577 } else if (index < (num_gpes + ACPI_NUM_FIXED_EVENTS))
578 result = acpi_get_event_status(index - num_gpes, status);
579
580 return result;
581 }
582
counter_show(struct kobject * kobj,struct kobj_attribute * attr,char * buf)583 static ssize_t counter_show(struct kobject *kobj,
584 struct kobj_attribute *attr, char *buf)
585 {
586 int index = attr - counter_attrs;
587 int size;
588 acpi_handle handle;
589 acpi_event_status status;
590 int result = 0;
591
592 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI].count =
593 acpi_irq_handled;
594 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT].count =
595 acpi_irq_not_handled;
596 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE].count =
597 acpi_gpe_count;
598 size = sprintf(buf, "%8u", all_counters[index].count);
599
600 /* "gpe_all" or "sci" */
601 if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS)
602 goto end;
603
604 result = get_status(index, &status, &handle);
605 if (result)
606 goto end;
607
608 if (status & ACPI_EVENT_FLAG_ENABLE_SET)
609 size += sprintf(buf + size, " EN");
610 else
611 size += sprintf(buf + size, " ");
612 if (status & ACPI_EVENT_FLAG_STATUS_SET)
613 size += sprintf(buf + size, " STS");
614 else
615 size += sprintf(buf + size, " ");
616
617 if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER))
618 size += sprintf(buf + size, " invalid ");
619 else if (status & ACPI_EVENT_FLAG_ENABLED)
620 size += sprintf(buf + size, " enabled ");
621 else if (status & ACPI_EVENT_FLAG_WAKE_ENABLED)
622 size += sprintf(buf + size, " wake_enabled");
623 else
624 size += sprintf(buf + size, " disabled ");
625 if (status & ACPI_EVENT_FLAG_MASKED)
626 size += sprintf(buf + size, " masked ");
627 else
628 size += sprintf(buf + size, " unmasked");
629
630 end:
631 size += sprintf(buf + size, "\n");
632 return result ? result : size;
633 }
634
635 /*
636 * counter_set() sets the specified counter.
637 * setting the total "sci" file to any value clears all counters.
638 * enable/disable/clear a gpe/fixed event in user space.
639 */
counter_set(struct kobject * kobj,struct kobj_attribute * attr,const char * buf,size_t size)640 static ssize_t counter_set(struct kobject *kobj,
641 struct kobj_attribute *attr, const char *buf,
642 size_t size)
643 {
644 int index = attr - counter_attrs;
645 acpi_event_status status;
646 acpi_handle handle;
647 int result = 0;
648 unsigned long tmp;
649
650 if (index == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI) {
651 int i;
652 for (i = 0; i < num_counters; ++i)
653 all_counters[i].count = 0;
654 acpi_gpe_count = 0;
655 acpi_irq_handled = 0;
656 acpi_irq_not_handled = 0;
657 goto end;
658 }
659
660 /* show the event status for both GPEs and Fixed Events */
661 result = get_status(index, &status, &handle);
662 if (result)
663 goto end;
664
665 if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER)) {
666 printk(KERN_WARNING PREFIX
667 "Can not change Invalid GPE/Fixed Event status\n");
668 return -EINVAL;
669 }
670
671 if (index < num_gpes) {
672 if (!strcmp(buf, "disable\n") &&
673 (status & ACPI_EVENT_FLAG_ENABLED))
674 result = acpi_disable_gpe(handle, index);
675 else if (!strcmp(buf, "enable\n") &&
676 !(status & ACPI_EVENT_FLAG_ENABLED))
677 result = acpi_enable_gpe(handle, index);
678 else if (!strcmp(buf, "clear\n") &&
679 (status & ACPI_EVENT_FLAG_STATUS_SET))
680 result = acpi_clear_gpe(handle, index);
681 else if (!strcmp(buf, "mask\n"))
682 result = acpi_mask_gpe(handle, index, TRUE);
683 else if (!strcmp(buf, "unmask\n"))
684 result = acpi_mask_gpe(handle, index, FALSE);
685 else if (!kstrtoul(buf, 0, &tmp))
686 all_counters[index].count = tmp;
687 else
688 result = -EINVAL;
689 } else if (index < num_gpes + ACPI_NUM_FIXED_EVENTS) {
690 int event = index - num_gpes;
691 if (!strcmp(buf, "disable\n") &&
692 (status & ACPI_EVENT_FLAG_ENABLE_SET))
693 result = acpi_disable_event(event, ACPI_NOT_ISR);
694 else if (!strcmp(buf, "enable\n") &&
695 !(status & ACPI_EVENT_FLAG_ENABLE_SET))
696 result = acpi_enable_event(event, ACPI_NOT_ISR);
697 else if (!strcmp(buf, "clear\n") &&
698 (status & ACPI_EVENT_FLAG_STATUS_SET))
699 result = acpi_clear_event(event);
700 else if (!kstrtoul(buf, 0, &tmp))
701 all_counters[index].count = tmp;
702 else
703 result = -EINVAL;
704 } else
705 all_counters[index].count = strtoul(buf, NULL, 0);
706
707 if (ACPI_FAILURE(result))
708 result = -EINVAL;
709 end:
710 return result ? result : size;
711 }
712
713 /*
714 * A Quirk Mechanism for GPE Flooding Prevention:
715 *
716 * Quirks may be needed to prevent GPE flooding on a specific GPE. The
717 * flooding typically cannot be detected and automatically prevented by
718 * ACPI_GPE_DISPATCH_NONE check because there is a _Lxx/_Exx prepared in
719 * the AML tables. This normally indicates a feature gap in Linux, thus
720 * instead of providing endless quirk tables, we provide a boot parameter
721 * for those who want this quirk. For example, if the users want to prevent
722 * the GPE flooding for GPE 00, they need to specify the following boot
723 * parameter:
724 * acpi_mask_gpe=0x00
725 * The masking status can be modified by the following runtime controlling
726 * interface:
727 * echo unmask > /sys/firmware/acpi/interrupts/gpe00
728 */
729
730 /*
731 * Currently, the GPE flooding prevention only supports to mask the GPEs
732 * numbered from 00 to 7f.
733 */
734 #define ACPI_MASKABLE_GPE_MAX 0x80
735
736 static u64 __initdata acpi_masked_gpes;
737
acpi_gpe_set_masked_gpes(char * val)738 static int __init acpi_gpe_set_masked_gpes(char *val)
739 {
740 u8 gpe;
741
742 if (kstrtou8(val, 0, &gpe) || gpe > ACPI_MASKABLE_GPE_MAX)
743 return -EINVAL;
744 acpi_masked_gpes |= ((u64)1<<gpe);
745
746 return 1;
747 }
748 __setup("acpi_mask_gpe=", acpi_gpe_set_masked_gpes);
749
acpi_gpe_apply_masked_gpes(void)750 void __init acpi_gpe_apply_masked_gpes(void)
751 {
752 acpi_handle handle;
753 acpi_status status;
754 u8 gpe;
755
756 for (gpe = 0;
757 gpe < min_t(u8, ACPI_MASKABLE_GPE_MAX, acpi_current_gpe_count);
758 gpe++) {
759 if (acpi_masked_gpes & ((u64)1<<gpe)) {
760 status = acpi_get_gpe_device(gpe, &handle);
761 if (ACPI_SUCCESS(status)) {
762 pr_info("Masking GPE 0x%x.\n", gpe);
763 (void)acpi_mask_gpe(handle, gpe, TRUE);
764 }
765 }
766 }
767 }
768
acpi_irq_stats_init(void)769 void acpi_irq_stats_init(void)
770 {
771 acpi_status status;
772 int i;
773
774 if (all_counters)
775 return;
776
777 num_gpes = acpi_current_gpe_count;
778 num_counters = num_gpes + ACPI_NUM_FIXED_EVENTS + NUM_COUNTERS_EXTRA;
779
780 all_attrs = kzalloc(sizeof(struct attribute *) * (num_counters + 1),
781 GFP_KERNEL);
782 if (all_attrs == NULL)
783 return;
784
785 all_counters = kzalloc(sizeof(struct event_counter) * (num_counters),
786 GFP_KERNEL);
787 if (all_counters == NULL)
788 goto fail;
789
790 status = acpi_install_global_event_handler(acpi_global_event_handler, NULL);
791 if (ACPI_FAILURE(status))
792 goto fail;
793
794 counter_attrs = kzalloc(sizeof(struct kobj_attribute) * (num_counters),
795 GFP_KERNEL);
796 if (counter_attrs == NULL)
797 goto fail;
798
799 for (i = 0; i < num_counters; ++i) {
800 char buffer[12];
801 char *name;
802
803 if (i < num_gpes)
804 sprintf(buffer, "gpe%02X", i);
805 else if (i == num_gpes + ACPI_EVENT_PMTIMER)
806 sprintf(buffer, "ff_pmtimer");
807 else if (i == num_gpes + ACPI_EVENT_GLOBAL)
808 sprintf(buffer, "ff_gbl_lock");
809 else if (i == num_gpes + ACPI_EVENT_POWER_BUTTON)
810 sprintf(buffer, "ff_pwr_btn");
811 else if (i == num_gpes + ACPI_EVENT_SLEEP_BUTTON)
812 sprintf(buffer, "ff_slp_btn");
813 else if (i == num_gpes + ACPI_EVENT_RTC)
814 sprintf(buffer, "ff_rt_clk");
815 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE)
816 sprintf(buffer, "gpe_all");
817 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI)
818 sprintf(buffer, "sci");
819 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT)
820 sprintf(buffer, "sci_not");
821 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_ERROR)
822 sprintf(buffer, "error");
823 else
824 sprintf(buffer, "bug%02X", i);
825
826 name = kstrdup(buffer, GFP_KERNEL);
827 if (name == NULL)
828 goto fail;
829
830 sysfs_attr_init(&counter_attrs[i].attr);
831 counter_attrs[i].attr.name = name;
832 counter_attrs[i].attr.mode = 0644;
833 counter_attrs[i].show = counter_show;
834 counter_attrs[i].store = counter_set;
835
836 all_attrs[i] = &counter_attrs[i].attr;
837 }
838
839 interrupt_stats_attr_group.attrs = all_attrs;
840 if (!sysfs_create_group(acpi_kobj, &interrupt_stats_attr_group))
841 return;
842
843 fail:
844 delete_gpe_attr_array();
845 return;
846 }
847
interrupt_stats_exit(void)848 static void __exit interrupt_stats_exit(void)
849 {
850 sysfs_remove_group(acpi_kobj, &interrupt_stats_attr_group);
851
852 delete_gpe_attr_array();
853
854 return;
855 }
856
857 static ssize_t
acpi_show_profile(struct device * dev,struct device_attribute * attr,char * buf)858 acpi_show_profile(struct device *dev, struct device_attribute *attr,
859 char *buf)
860 {
861 return sprintf(buf, "%d\n", acpi_gbl_FADT.preferred_profile);
862 }
863
864 static const struct device_attribute pm_profile_attr =
865 __ATTR(pm_profile, S_IRUGO, acpi_show_profile, NULL);
866
hotplug_enabled_show(struct kobject * kobj,struct kobj_attribute * attr,char * buf)867 static ssize_t hotplug_enabled_show(struct kobject *kobj,
868 struct kobj_attribute *attr, char *buf)
869 {
870 struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj);
871
872 return sprintf(buf, "%d\n", hotplug->enabled);
873 }
874
hotplug_enabled_store(struct kobject * kobj,struct kobj_attribute * attr,const char * buf,size_t size)875 static ssize_t hotplug_enabled_store(struct kobject *kobj,
876 struct kobj_attribute *attr,
877 const char *buf, size_t size)
878 {
879 struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj);
880 unsigned int val;
881
882 if (kstrtouint(buf, 10, &val) || val > 1)
883 return -EINVAL;
884
885 acpi_scan_hotplug_enabled(hotplug, val);
886 return size;
887 }
888
889 static struct kobj_attribute hotplug_enabled_attr =
890 __ATTR(enabled, S_IRUGO | S_IWUSR, hotplug_enabled_show,
891 hotplug_enabled_store);
892
893 static struct attribute *hotplug_profile_attrs[] = {
894 &hotplug_enabled_attr.attr,
895 NULL
896 };
897
898 static struct kobj_type acpi_hotplug_profile_ktype = {
899 .sysfs_ops = &kobj_sysfs_ops,
900 .default_attrs = hotplug_profile_attrs,
901 };
902
acpi_sysfs_add_hotplug_profile(struct acpi_hotplug_profile * hotplug,const char * name)903 void acpi_sysfs_add_hotplug_profile(struct acpi_hotplug_profile *hotplug,
904 const char *name)
905 {
906 int error;
907
908 if (!hotplug_kobj)
909 goto err_out;
910
911 error = kobject_init_and_add(&hotplug->kobj,
912 &acpi_hotplug_profile_ktype, hotplug_kobj, "%s", name);
913 if (error)
914 goto err_out;
915
916 kobject_uevent(&hotplug->kobj, KOBJ_ADD);
917 return;
918
919 err_out:
920 pr_err(PREFIX "Unable to add hotplug profile '%s'\n", name);
921 }
922
force_remove_show(struct kobject * kobj,struct kobj_attribute * attr,char * buf)923 static ssize_t force_remove_show(struct kobject *kobj,
924 struct kobj_attribute *attr, char *buf)
925 {
926 return sprintf(buf, "%d\n", !!acpi_force_hot_remove);
927 }
928
force_remove_store(struct kobject * kobj,struct kobj_attribute * attr,const char * buf,size_t size)929 static ssize_t force_remove_store(struct kobject *kobj,
930 struct kobj_attribute *attr,
931 const char *buf, size_t size)
932 {
933 bool val;
934 int ret;
935
936 ret = strtobool(buf, &val);
937 if (ret < 0)
938 return ret;
939
940 lock_device_hotplug();
941 acpi_force_hot_remove = val;
942 unlock_device_hotplug();
943 return size;
944 }
945
946 static const struct kobj_attribute force_remove_attr =
947 __ATTR(force_remove, S_IRUGO | S_IWUSR, force_remove_show,
948 force_remove_store);
949
acpi_sysfs_init(void)950 int __init acpi_sysfs_init(void)
951 {
952 int result;
953
954 result = acpi_tables_sysfs_init();
955 if (result)
956 return result;
957
958 hotplug_kobj = kobject_create_and_add("hotplug", acpi_kobj);
959 if (!hotplug_kobj)
960 return -ENOMEM;
961
962 result = sysfs_create_file(hotplug_kobj, &force_remove_attr.attr);
963 if (result)
964 return result;
965
966 result = sysfs_create_file(acpi_kobj, &pm_profile_attr.attr);
967 return result;
968 }
969