/kernel/ |
D | sysctl.c | 194 static int bpf_stats_handler(struct ctl_table *table, int write, in bpf_stats_handler() argument 208 if (write && !capable(CAP_SYS_ADMIN)) in bpf_stats_handler() 213 ret = proc_dointvec_minmax(&tmp, write, buffer, lenp, ppos); in bpf_stats_handler() 214 if (write && !ret && val != saved_val) { in bpf_stats_handler() 229 static int bpf_unpriv_handler(struct ctl_table *table, int write, in bpf_unpriv_handler() argument 236 if (write && !capable(CAP_SYS_ADMIN)) in bpf_unpriv_handler() 240 ret = proc_dointvec_minmax(&tmp, write, buffer, lenp, ppos); in bpf_unpriv_handler() 241 if (write && !ret) { in bpf_unpriv_handler() 259 static int _proc_do_string(char *data, int maxlen, int write, in _proc_do_string() argument 270 if (write) { in _proc_do_string() [all …]
|
D | watchdog.c | 668 static int proc_watchdog_common(int which, struct ctl_table *table, int write, in proc_watchdog_common() argument 675 if (!write) { in proc_watchdog_common() 681 err = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in proc_watchdog_common() 684 err = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in proc_watchdog_common() 695 int proc_watchdog(struct ctl_table *table, int write, in proc_watchdog() argument 699 table, write, buffer, lenp, ppos); in proc_watchdog() 705 int proc_nmi_watchdog(struct ctl_table *table, int write, in proc_nmi_watchdog() argument 708 if (!nmi_watchdog_available && write) in proc_nmi_watchdog() 711 table, write, buffer, lenp, ppos); in proc_nmi_watchdog() 717 int proc_soft_watchdog(struct ctl_table *table, int write, in proc_soft_watchdog() argument [all …]
|
D | stackleak.c | 22 int stack_erasing_sysctl(struct ctl_table *table, int write, in stack_erasing_sysctl() argument 31 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in stack_erasing_sysctl() 33 if (ret || !write || state == prev_state) in stack_erasing_sysctl()
|
D | utsname_sysctl.c | 32 static int proc_do_uts_string(struct ctl_table *table, int write, in proc_do_uts_string() argument 51 r = proc_dostring(&uts_table, write, buffer, lenp, ppos); in proc_do_uts_string() 53 if (write) { in proc_do_uts_string()
|
D | delayacct.c | 49 int sysctl_delayacct(struct ctl_table *table, int write, void *buffer, in sysctl_delayacct() argument 56 if (write && !capable(CAP_SYS_ADMIN)) in sysctl_delayacct() 61 err = proc_dointvec_minmax(&t, write, buffer, lenp, ppos); in sysctl_delayacct() 64 if (write) in sysctl_delayacct()
|
D | hung_task.c | 234 int proc_dohung_task_timeout_secs(struct ctl_table *table, int write, in proc_dohung_task_timeout_secs() argument 239 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in proc_dohung_task_timeout_secs() 241 if (ret || !write) in proc_dohung_task_timeout_secs()
|
D | umh.c | 488 static int proc_cap_handler(struct ctl_table *table, int write, in proc_cap_handler() argument 496 if (write && (!capable(CAP_SETPCAP) || in proc_cap_handler() 522 err = proc_doulongvec_minmax(&t, write, buffer, lenp, ppos); in proc_cap_handler() 536 if (write) { in proc_cap_handler()
|
D | pid_namespace.c | 276 static int pid_ns_ctl_handler(struct ctl_table *table, int write, in pid_ns_ctl_handler() argument 283 if (write && !checkpoint_restore_ns_capable(pid_ns->user_ns)) in pid_ns_ctl_handler() 295 ret = proc_dointvec_minmax(&tmp, write, buffer, lenp, ppos); in pid_ns_ctl_handler() 296 if (!ret && write) in pid_ns_ctl_handler()
|
D | latencytop.c | 272 int sysctl_latencytop(struct ctl_table *table, int write, void *buffer, in sysctl_latencytop() argument 277 err = proc_dointvec(table, write, buffer, lenp, ppos); in sysctl_latencytop()
|
/kernel/trace/ |
D | trace_stack.c | 371 .write = stack_max_size_write, 509 .write = ftrace_filter_write, 517 stack_trace_sysctl(struct ctl_table *table, int write, void *buffer, in stack_trace_sysctl() argument 526 ret = proc_dointvec(table, write, buffer, lenp, ppos); in stack_trace_sysctl() 528 if (ret || !write || (was_enabled == !!stack_tracer_enabled)) in stack_trace_sysctl()
|
D | ring_buffer.c | 296 local_t write; /* index for next write */ member 694 long commit, write; in verify_event() local 704 write = local_read(&page->write); in verify_event() 706 addr < (unsigned long)&page->page->data[write]) in verify_event() 1433 old_write = local_add_return(RB_WRITE_INTCNT, &next_page->write); in rb_tail_page_update() 1463 (void)local_cmpxchg(&next_page->write, old_write, val); in rb_tail_page_update() 1856 return local_read(&bpage->write) & RB_WRITE_MASK; in rb_page_write() 2582 local_sub(length, &tail_page->write); in rb_reset_tail() 2617 local_sub(length, &tail_page->write); in rb_reset_tail() 2635 local_sub(length, &tail_page->write); in rb_reset_tail() [all …]
|
D | trace.c | 291 export->write(export, entry, size); in trace_process_export() 393 if (WARN_ON_ONCE(!export->write)) in register_ftrace_export() 2878 int tracepoint_printk_sysctl(struct ctl_table *table, int write, in tracepoint_printk_sysctl() argument 2888 ret = proc_dointvec(table, write, buffer, lenp, ppos); in tracepoint_printk_sysctl() 5151 .write = tracing_write_stub, 5263 .write = tracing_cpumask_write, 5516 .write = tracing_trace_options_write, 6007 .write = tracing_saved_cmdlines_size_write, 7645 .write = tracing_thresh_write, 7653 .write = tracing_max_lat_write, [all …]
|
D | Kconfig | 99 bool "Register read/write tracing" 102 Create tracepoints for MMIO read/write operations. These trace events 103 can be used for logging all MMIO read/write operations. 774 it took to write to the tracepoint and the next iteration that 778 to keep the time the same. The initial string is simply a write of 780 write which is not added to the rest of the calculations. 933 a thread per cpu. Each thread will write various size events 935 to each of the threads, where the IPI handler will also write 975 as it will write garbage to IO memory starting at a given address.
|
D | trace_recursion_record.c | 219 .write = recursed_function_write,
|
/kernel/events/ |
D | callchain.c | 236 int perf_event_max_stack_handler(struct ctl_table *table, int write, in perf_event_max_stack_handler() argument 244 ret = proc_dointvec_minmax(&new_table, write, buffer, lenp, ppos); in perf_event_max_stack_handler() 245 if (ret || !write) in perf_event_max_stack_handler()
|
/kernel/sched/ |
D | pelt.c | 482 int sched_pelt_multiplier(struct ctl_table *table, int write, void *buffer, in sched_pelt_multiplier() argument 492 ret = proc_dointvec(table, write, buffer, lenp, ppos); in sched_pelt_multiplier() 495 if (!write) in sched_pelt_multiplier()
|
D | debug.c | 166 .write = sched_feat_write, 214 .write = sched_scaling_write, 276 .write = sched_dynamic_write,
|
/kernel/bpf/ |
D | cgroup.c | 1314 struct ctl_table *table, int write, in __cgroup_bpf_run_filter_sysctl() argument 1321 .write = write, in __cgroup_bpf_run_filter_sysctl() 1340 if (write && *buf && *pcount) { in __cgroup_bpf_run_filter_sysctl() 1788 if (!ctx->write) { in BPF_CALL_3() 1808 if (!ctx->write || !ctx->new_val || !ctx->new_len || !buf || !buf_len) in BPF_CALL_3() 1863 case bpf_ctx_range(struct bpf_sysctl, write): in sysctl_is_valid_access() 1889 case offsetof(struct bpf_sysctl, write): in sysctl_convert_ctx_access() 1892 bpf_target_off(struct bpf_sysctl_kern, write, in sysctl_convert_ctx_access() 1894 write), in sysctl_convert_ctx_access()
|
/kernel/printk/ |
D | printk.c | 180 int devkmsg_sysctl_set_loglvl(struct ctl_table *table, int write, in devkmsg_sysctl_set_loglvl() argument 187 if (write) { in devkmsg_sysctl_set_loglvl() 195 err = proc_dostring(table, write, buffer, lenp, ppos); in devkmsg_sysctl_set_loglvl() 199 if (write) { in devkmsg_sysctl_set_loglvl() 1953 if (!con->write) in call_console_drivers() 1959 con->write(con, ext_text, ext_len); in call_console_drivers() 1962 con->write(con, dropped_text, dropped_len); in call_console_drivers() 1963 con->write(con, text, len); in call_console_drivers() 2368 early_console->write(early_console, buf, n); in early_printk() 3233 init_section_contains(con->write, 0) || in printk_late_init()
|
/kernel/bpf/preload/iterators/ |
D | iterators.c | 33 if (write(to_kernel, &obj, sizeof(obj)) != sizeof(obj)) in send_link_to_kernel()
|
/kernel/locking/ |
D | lock_events.c | 114 .write = lockevent_write,
|
D | locktorture.c | 739 struct lock_stress_stats *statp, bool write) in __torture_print_stats() argument 747 n_stress = write ? cxt.nrealwriters_stress : cxt.nrealreaders_stress; in __torture_print_stats() 760 write ? "Writes" : "Reads ", in __torture_print_stats()
|
/kernel/cgroup/ |
D | cgroup.c | 1524 if (cft->write_u64 || cft->write_s64 || cft->write) { in cgroup_file_mode() 3958 if (cft->write) in cgroup_file_write() 3959 return cft->write(of, buf, nbytes, off); in cgroup_file_write() 4035 .write = cgroup_file_write, 4044 .write = cgroup_file_write, 5070 .write = cgroup_type_write, 5080 .write = cgroup_procs_write, 5089 .write = cgroup_threads_write, 5099 .write = cgroup_subtree_control_write, 5110 .write = cgroup_max_descendants_write, [all …]
|
D | cgroup-v1.c | 638 .write = cgroup1_procs_write, 657 .write = cgroup1_tasks_write, 668 .write = cgroup_release_agent_write,
|
/kernel/time/ |
D | test_udelay.c | 132 .write = udelay_test_write,
|