/kernel/trace/ |
D | ring_buffer_benchmark.c | 33 static unsigned long read; variable 99 read++; in read_event() 148 read++; in read_page() 161 read++; in read_page() 185 read = 0; in ring_buffer_consumer() 328 trace_printk("Read: %ld (by %s)\n", read, in ring_buffer_producer() 331 trace_printk("Total: %lld\n", entries + overruns + read); in ring_buffer_producer()
|
D | ring_buffer.c | 302 unsigned read; /* index for next read */ member 462 unsigned long read; member 1504 cpu_buffer->read = 0; in rb_remove_pages() 1864 cpu_buffer->reader_page->read); in rb_reader_event() 3090 if (reader->read != rb_page_commit(reader)) in rb_per_cpu_empty() 3268 (local_read(&cpu_buffer->overrun) + cpu_buffer->read); in rb_num_of_entries() 3422 return cpu_buffer->read; in ring_buffer_read_events_cpu() 3478 iter->head = cpu_buffer->reader_page->read; in rb_iter_reset() 3481 iter->cache_read = cpu_buffer->read; in rb_iter_reset() 3534 head_page->read == commit && in ring_buffer_iter_empty() [all …]
|
D | trace.c | 936 size_t read = 0; in trace_get_user() local 946 read++; in trace_get_user() 959 read++; in trace_get_user() 965 *ppos += read; in trace_get_user() 966 ret = read; in trace_get_user() 984 read++; in trace_get_user() 1000 *ppos += read; in trace_get_user() 1001 ret = read; in trace_get_user() 3451 .read = seq_read, 3459 .read = seq_read, [all …]
|
D | trace_events.c | 826 ssize_t read, ret; in ftrace_event_write() local 838 read = trace_get_user(&parser, ubuf, cnt, ppos); in ftrace_event_write() 840 if (read >= 0 && trace_parser_loaded((&parser))) { in ftrace_event_write() 853 ret = read; in ftrace_event_write() 1591 ssize_t read = 0; in ftrace_event_pid_write() local 1619 read += ret; in ftrace_event_pid_write() 1762 ret = read; in ftrace_event_pid_write() 1763 *ppos += read; in ftrace_event_pid_write() 1796 .read = seq_read, 1803 .read = seq_read, [all …]
|
D | trace_stack.c | 284 .read = stack_max_size_read, 405 .read = seq_read, 419 .read = seq_read,
|
D | ftrace.c | 1037 .read = ftrace_profile_read, 4042 ssize_t ret, read; in ftrace_regex_write() local 4059 read = trace_get_user(parser, ubuf, cnt, ppos); in ftrace_regex_write() 4061 if (read >= 0 && trace_parser_loaded(parser) && in ftrace_regex_write() 4070 ret = read; in ftrace_regex_write() 4451 .read = seq_read, 4458 .read = seq_read, 4465 .read = seq_read, 4473 .read = seq_read, 4708 ssize_t read, ret = 0; in ftrace_graph_write() local [all …]
|
/kernel/time/ |
D | sched_clock.c | 168 sched_clock_register(u64 (*read)(void), int bits, unsigned long rate) in sched_clock_register() 194 new_epoch = read(); in sched_clock_register() 197 cd.actual_read_sched_clock = read; in sched_clock_register() 199 rd.read_sched_clock = read; in sched_clock_register() 236 pr_debug("Registered %pF as sched_clock source\n", read); in sched_clock_register()
|
D | timecounter.c | 26 tc->cycle_last = cc->read(cc); in timecounter_init() 50 cycle_now = tc->cc->read(tc->cc); in timecounter_read_delta()
|
D | posix-clock.c | 59 if (clk->ops.read) in posix_clock_read() 60 err = clk->ops.read(clk, fp->f_flags, buf, count); in posix_clock_read() 195 .read = posix_clock_read,
|
D | timekeeping_debug.c | 52 .read = seq_read,
|
D | jiffies.c | 70 .read = jiffies_read,
|
D | test_udelay.c | 140 .read = seq_read,
|
/kernel/rcu/ |
D | tree_trace.c | 98 .read = seq_read, 179 .read = seq_read, 208 .read = seq_read, 255 .read = seq_read, 308 .read = seq_read, 350 .read = seq_read, 396 .read = seq_read, 419 .read = seq_read,
|
D | tiny_plugin.h | 102 .read = seq_read,
|
/kernel/ |
D | profile.c | 455 .read = seq_read, 477 ssize_t read; in read_profile() local 486 read = 0; in read_profile() 491 buf++; p++; count--; read++; in read_profile() 496 read += count; in read_profile() 497 *ppos += read; in read_profile() 498 return read; in read_profile() 529 .read = read_profile,
|
D | exec_domain.c | 37 .read = seq_read,
|
D | configs.c | 68 .read = ikconfig_read_current,
|
D | dma.c | 144 .read = seq_read,
|
D | latencytop.c | 281 .read = seq_read,
|
/kernel/irq/ |
D | proc.c | 165 .read = seq_read, 173 .read = seq_read, 180 .read = seq_read, 235 .read = seq_read, 256 .read = seq_read, 279 .read = seq_read,
|
/kernel/locking/ |
D | lockdep.c | 268 if (hlock->read) in lock_release_holdtime() 1771 struct lockdep_map *next_instance, int read) in check_deadlock() argument 1790 if ((read == 2) && prev->read) in check_deadlock() 1872 if (next->read == 2 || prev->read == 2) in check_prev_add() 1961 if (hlock->read != 2 && hlock->check) { in check_prevs_add() 2132 int ret = check_deadlock(curr, hlock, lock, hlock->read); in validate_chain() 2142 hlock->read = 2; in validate_chain() 2465 int read = new_bit & 1; in mark_lock_irq() local 2489 if ((!read || !dir || STRICT_READ_CHECKS) && in mark_lock_irq() 2496 if (!read) { in mark_lock_irq() [all …]
|
D | lockdep_proc.c | 110 .read = seq_read, 178 .read = seq_read, 360 .read = seq_read, 672 .read = seq_read,
|
/kernel/power/ |
D | qos.c | 142 .read = pm_qos_power_read, 257 .read = seq_read,
|
/kernel/sched/ |
D | stats.c | 154 .read = seq_read,
|
/kernel/gcov/ |
D | fs.c | 406 .read = seq_read, 563 .read = reset_read,
|