/kernel/trace/ |
D | ring_buffer_benchmark.c | 33 static unsigned long read; variable 99 read++; in read_event() 148 read++; in read_page() 161 read++; in read_page() 185 read = 0; in ring_buffer_consumer() 328 trace_printk("Read: %ld (by %s)\n", read, in ring_buffer_producer() 331 trace_printk("Total: %lld\n", entries + overruns + read); in ring_buffer_producer()
|
D | ring_buffer.c | 302 unsigned read; /* index for next read */ member 460 unsigned long read; member 1493 cpu_buffer->read = 0; in rb_remove_pages() 1851 cpu_buffer->reader_page->read); in rb_reader_event() 3048 return reader->read == rb_page_commit(reader) && in rb_per_cpu_empty() 3051 head->read == rb_page_commit(commit))); in rb_per_cpu_empty() 3191 (local_read(&cpu_buffer->overrun) + cpu_buffer->read); in rb_num_of_entries() 3345 return cpu_buffer->read; in ring_buffer_read_events_cpu() 3401 iter->head = cpu_buffer->reader_page->read; in rb_iter_reset() 3404 iter->cache_read = cpu_buffer->read; in rb_iter_reset() [all …]
|
D | trace.c | 488 ssize_t read = 0; in trace_pid_write() local 535 read += ret; in trace_pid_write() 565 read = ret; in trace_pid_write() 571 return read; in trace_pid_write() 1170 size_t read = 0; in trace_get_user() local 1180 read++; in trace_get_user() 1193 read++; in trace_get_user() 1199 *ppos += read; in trace_get_user() 1200 ret = read; in trace_get_user() 1218 read++; in trace_get_user() [all …]
|
D | trace_stack.c | 283 .read = stack_max_size_read, 404 .read = seq_read, 418 .read = seq_read,
|
D | trace_events.c | 859 ssize_t read, ret; in ftrace_event_write() local 871 read = trace_get_user(&parser, ubuf, cnt, ppos); in ftrace_event_write() 873 if (read >= 0 && trace_parser_loaded((&parser))) { in ftrace_event_write() 886 ret = read; in ftrace_event_write() 1692 .read = seq_read, 1699 .read = seq_read, 1707 .read = seq_read, 1715 .read = event_enable_read, 1722 .read = seq_read, 1728 .read = event_id_read, [all …]
|
D | ftrace.c | 1018 .read = ftrace_profile_read, 4098 ssize_t ret, read; in ftrace_regex_write() local 4115 read = trace_get_user(parser, ubuf, cnt, ppos); in ftrace_regex_write() 4117 if (read >= 0 && trace_parser_loaded(parser) && in ftrace_regex_write() 4126 ret = read; in ftrace_regex_write() 4504 .read = seq_read, 4511 .read = seq_read, 4518 .read = seq_read, 4526 .read = seq_read, 4764 ssize_t read, ret = 0; in ftrace_graph_write() local [all …]
|
D | trace_hwlat.c | 507 .read = hwlat_read, 513 .read = hwlat_read,
|
/kernel/ |
D | profile.c | 444 .read = seq_read, 466 ssize_t read; in read_profile() local 475 read = 0; in read_profile() 480 buf++; p++; count--; read++; in read_profile() 485 read += count; in read_profile() 486 *ppos += read; in read_profile() 487 return read; in read_profile() 518 .read = read_profile,
|
D | exec_domain.c | 37 .read = seq_read,
|
D | configs.c | 68 .read = ikconfig_read_current,
|
D | dma.c | 144 .read = seq_read,
|
/kernel/time/ |
D | sched_clock.c | 168 sched_clock_register(u64 (*read)(void), int bits, unsigned long rate) in sched_clock_register() 194 new_epoch = read(); in sched_clock_register() 197 cd.actual_read_sched_clock = read; in sched_clock_register() 199 rd.read_sched_clock = read; in sched_clock_register() 236 pr_debug("Registered %pF as sched_clock source\n", read); in sched_clock_register()
|
D | timecounter.c | 26 tc->cycle_last = cc->read(cc); in timecounter_init() 50 cycle_now = tc->cc->read(tc->cc); in timecounter_read_delta()
|
D | posix-clock.c | 61 if (clk->ops.read) in posix_clock_read() 62 err = clk->ops.read(clk, fp->f_flags, buf, count); in posix_clock_read() 197 .read = posix_clock_read,
|
D | timekeeping_debug.c | 52 .read = seq_read,
|
D | jiffies.c | 70 .read = jiffies_read,
|
D | test_udelay.c | 140 .read = seq_read,
|
/kernel/rcu/ |
D | tree_trace.c | 97 .read = seq_read, 178 .read = seq_read, 213 .read = seq_read, 260 .read = seq_read, 313 .read = seq_read, 355 .read = seq_read, 401 .read = seq_read, 424 .read = seq_read,
|
D | tiny_plugin.h | 107 .read = seq_read,
|
/kernel/irq/ |
D | proc.c | 165 .read = seq_read, 173 .read = seq_read, 180 .read = seq_read, 235 .read = seq_read, 256 .read = seq_read, 279 .read = seq_read,
|
/kernel/locking/ |
D | lockdep.c | 267 if (hlock->read) in lock_release_holdtime() 1742 struct lockdep_map *next_instance, int read) in check_deadlock() argument 1761 if ((read == 2) && prev->read) in check_deadlock() 1843 if (next->read == 2 || prev->read == 2) in check_prev_add() 1937 if (hlock->read != 2 && hlock->check) { in check_prevs_add() 2249 int ret = check_deadlock(curr, hlock, lock, hlock->read); in validate_chain() 2259 hlock->read = 2; in validate_chain() 2585 int read = new_bit & 1; in mark_lock_irq() local 2609 if ((!read || !dir || STRICT_READ_CHECKS) && in mark_lock_irq() 2616 if (!read) { in mark_lock_irq() [all …]
|
D | lockdep_proc.c | 110 .read = seq_read, 180 .read = seq_read, 360 .read = seq_read, 672 .read = seq_read,
|
D | qspinlock_stat.h | 201 .read = qstat_read,
|
/kernel/power/ |
D | qos.c | 142 .read = pm_qos_power_read, 257 .read = seq_read,
|
/kernel/sched/ |
D | stats.c | 154 .read = seq_read,
|