/kernel/trace/ |
D | ring_buffer_benchmark.c | 35 static unsigned long read; variable 101 read++; in read_event() 150 read++; in read_page() 163 read++; in read_page() 187 read = 0; in ring_buffer_consumer() 330 trace_printk("Read: %ld (by %s)\n", read, in ring_buffer_producer() 333 trace_printk("Total: %lld\n", entries + overruns + read); in ring_buffer_producer()
|
D | ring_buffer.c | 322 unsigned read; /* index for next read */ member 472 unsigned long read; member 535 size_t read; in ring_buffer_nr_dirty_pages() local 538 read = local_read(&buffer->buffers[cpu]->pages_read); in ring_buffer_nr_dirty_pages() 541 if (cnt < read) { in ring_buffer_nr_dirty_pages() 542 WARN_ON_ONCE(read > cnt + 1); in ring_buffer_nr_dirty_pages() 546 return cnt - read; in ring_buffer_nr_dirty_pages() 1560 cpu_buffer->read = 0; in rb_remove_pages() 1914 cpu_buffer->reader_page->read); in rb_reader_event() 3181 return reader->read == rb_page_commit(reader) && in rb_per_cpu_empty() [all …]
|
D | trace.c | 504 ssize_t read = 0; in trace_pid_write() local 554 read += ret; in trace_pid_write() 582 read = ret; in trace_pid_write() 588 return read; in trace_pid_write() 1409 size_t read = 0; in trace_get_user() local 1419 read++; in trace_get_user() 1432 read++; in trace_get_user() 1440 *ppos += read; in trace_get_user() 1441 ret = read; in trace_get_user() 1457 read++; in trace_get_user() [all …]
|
D | trace_stack.c | 371 .read = stack_max_size_read, 490 .read = seq_read, 509 .read = seq_read,
|
D | trace_events.c | 863 ssize_t read, ret; in ftrace_event_write() local 875 read = trace_get_user(&parser, ubuf, cnt, ppos); in ftrace_event_write() 877 if (read >= 0 && trace_parser_loaded((&parser))) { in ftrace_event_write() 888 ret = read; in ftrace_event_write() 1682 .read = seq_read, 1689 .read = seq_read, 1697 .read = seq_read, 1705 .read = event_enable_read, 1712 .read = seq_read, 1718 .read = event_id_read, [all …]
|
D | trace_hwlat.c | 508 .read = hwlat_read, 514 .read = hwlat_read,
|
D | ftrace.c | 946 .read = ftrace_profile_read, 4681 ssize_t ret, read; in ftrace_regex_write() local 4698 read = trace_get_user(parser, ubuf, cnt, ppos); in ftrace_regex_write() 4700 if (read >= 0 && trace_parser_loaded(parser) && in ftrace_regex_write() 4709 ret = read; in ftrace_regex_write() 5073 .read = seq_read, 5080 .read = seq_read, 5087 .read = seq_read, 5095 .read = seq_read, 5453 ssize_t read, ret = 0; in ftrace_graph_write() local [all …]
|
D | trace_dynevent.c | 200 .read = seq_read,
|
/kernel/ |
D | profile.c | 447 .read = seq_read, 469 ssize_t read; in read_profile() local 478 read = 0; in read_profile() 483 buf++; p++; count--; read++; in read_profile() 488 read += count; in read_profile() 489 *ppos += read; in read_profile() 490 return read; in read_profile() 521 .read = read_profile,
|
D | kheaders.c | 46 .read = &ikheaders_read,
|
D | configs.c | 52 .read = ikconfig_read_current,
|
D | latencytop.c | 260 .read = seq_read,
|
D | ksysfs.c | 205 .read = ¬es_read,
|
/kernel/time/ |
D | sched_clock.c | 168 sched_clock_register(u64 (*read)(void), int bits, unsigned long rate) in sched_clock_register() 194 new_epoch = read(); in sched_clock_register() 197 cd.actual_read_sched_clock = read; in sched_clock_register() 199 rd.read_sched_clock = read; in sched_clock_register() 236 pr_debug("Registered %pS as sched_clock source\n", read); in sched_clock_register()
|
D | timecounter.c | 13 tc->cycle_last = cc->read(cc); in timecounter_init() 37 cycle_now = tc->cc->read(tc->cc); in timecounter_read_delta()
|
D | posix-clock.c | 48 if (clk->ops.read) in posix_clock_read() 49 err = clk->ops.read(clk, fp->f_flags, buf, count); in posix_clock_read() 152 .read = posix_clock_read,
|
D | clocksource.c | 210 csnow = cs->read(cs); in clocksource_watchdog() 211 wdnow = watchdog->read(watchdog); in clocksource_watchdog() 539 suspend_start = suspend_clocksource->read(suspend_clocksource); in clocksource_start_suspend_timing() 571 now = suspend_clocksource->read(suspend_clocksource); in clocksource_stop_suspend_timing()
|
D | jiffies.c | 54 .read = jiffies_read,
|
D | test_udelay.c | 132 .read = seq_read,
|
/kernel/locking/ |
D | lockdep.c | 288 if (hlock->read) in lock_release_holdtime() 2401 if ((next->read == 2) && prev->read) in check_deadlock() 2490 if (next->read == 2 || prev->read == 2) in check_prev_add() 2579 if (hlock->read != 2 && hlock->check) { in check_prevs_add() 2964 hlock->read = 2; in validate_chain() 3291 int read = new_bit & LOCK_USAGE_READ_MASK; in mark_lock_irq() local 3315 if ((!read || STRICT_READ_CHECKS) && in mark_lock_irq() 3322 if (!read) { in mark_lock_irq() 3351 if (hlock->read) in mark_held_locks() 3551 if (hlock->read) { in mark_usage() [all …]
|
D | lock_events.c | 113 .read = lockevent_read,
|
/kernel/irq/ |
D | proc.c | 181 .read = seq_read, 189 .read = seq_read, 251 .read = seq_read,
|
/kernel/bpf/ |
D | sysfs_btf.c | 26 .read = btf_vmlinux_read,
|
/kernel/gcov/ |
D | fs.c | 404 .read = seq_read, 556 .read = reset_read,
|
/kernel/sched/ |
D | psi.c | 1256 .read = seq_read, 1265 .read = seq_read, 1274 .read = seq_read,
|