/kernel/time/ |
D | tick-sched.c | 186 static void tick_sched_do_timer(struct tick_sched *ts, ktime_t now) in tick_sched_do_timer() argument 219 if (ts->last_tick_jiffies != jiffies) { in tick_sched_do_timer() 220 ts->stalled_jiffies = 0; in tick_sched_do_timer() 221 ts->last_tick_jiffies = READ_ONCE(jiffies); in tick_sched_do_timer() 223 if (++ts->stalled_jiffies == MAX_STALLED_JIFFIES) { in tick_sched_do_timer() 225 ts->stalled_jiffies = 0; in tick_sched_do_timer() 226 ts->last_tick_jiffies = READ_ONCE(jiffies); in tick_sched_do_timer() 230 if (ts->inidle) in tick_sched_do_timer() 231 ts->got_idle_tick = 1; in tick_sched_do_timer() 234 static void tick_sched_handle(struct tick_sched *ts, struct pt_regs *regs) in tick_sched_handle() argument [all …]
|
D | time.c | 144 struct timespec64 ts; in SYSCALL_DEFINE2() local 146 ktime_get_real_ts64(&ts); in SYSCALL_DEFINE2() 147 if (put_user(ts.tv_sec, &tv->tv_sec) || in SYSCALL_DEFINE2() 148 put_user(ts.tv_nsec / 1000, &tv->tv_usec)) in SYSCALL_DEFINE2() 228 struct timespec64 ts; in COMPAT_SYSCALL_DEFINE2() local 230 ktime_get_real_ts64(&ts); in COMPAT_SYSCALL_DEFINE2() 231 if (put_user(ts.tv_sec, &tv->tv_sec) || in COMPAT_SYSCALL_DEFINE2() 232 put_user(ts.tv_nsec / 1000, &tv->tv_usec)) in COMPAT_SYSCALL_DEFINE2() 454 struct timespec64 ts = ns_to_timespec64(nsec); in ns_to_kernel_old_timeval() local 457 tv.tv_sec = ts.tv_sec; in ns_to_kernel_old_timeval() [all …]
|
D | timekeeping.c | 131 struct timespec64 ts; in tk_xtime() local 133 ts.tv_sec = tk->xtime_sec; in tk_xtime() 134 ts.tv_nsec = (long)(tk->tkr_mono.xtime_nsec >> tk->tkr_mono.shift); in tk_xtime() 135 return ts; in tk_xtime() 138 static void tk_set_xtime(struct timekeeper *tk, const struct timespec64 *ts) in tk_set_xtime() argument 140 tk->xtime_sec = ts->tv_sec; in tk_set_xtime() 141 tk->tkr_mono.xtime_nsec = (u64)ts->tv_nsec << tk->tkr_mono.shift; in tk_set_xtime() 144 static void tk_xtime_add(struct timekeeper *tk, const struct timespec64 *ts) in tk_xtime_add() argument 146 tk->xtime_sec += ts->tv_sec; in tk_xtime_add() 147 tk->tkr_mono.xtime_nsec += (u64)ts->tv_nsec << tk->tkr_mono.shift; in tk_xtime_add() [all …]
|
D | posix-clock.c | 250 static int pc_clock_gettime(clockid_t id, struct timespec64 *ts) in pc_clock_gettime() argument 260 err = cd.clk->ops.clock_gettime(cd.clk, ts); in pc_clock_gettime() 269 static int pc_clock_getres(clockid_t id, struct timespec64 *ts) in pc_clock_getres() argument 279 err = cd.clk->ops.clock_getres(cd.clk, ts); in pc_clock_getres() 288 static int pc_clock_settime(clockid_t id, const struct timespec64 *ts) in pc_clock_settime() argument 303 err = cd.clk->ops.clock_settime(cd.clk, ts); in pc_clock_settime()
|
D | test_udelay.c | 81 struct timespec64 ts; in udelay_test_show() local 83 ktime_get_ts64(&ts); in udelay_test_show() 85 loops_per_jiffy, (s64)ts.tv_sec, ts.tv_nsec); in udelay_test_show()
|
D | ntp.c | 762 int __do_adjtimex(struct __kernel_timex *txc, const struct timespec64 *ts, in __do_adjtimex() argument 822 txc->time.tv_sec = ts->tv_sec; in __do_adjtimex() 823 txc->time.tv_usec = ts->tv_nsec; in __do_adjtimex() 825 txc->time.tv_usec = ts->tv_nsec / NSEC_PER_USEC; in __do_adjtimex() 828 if (unlikely(ts->tv_sec >= ntp_next_leap_sec)) { in __do_adjtimex() 840 (ts->tv_sec == ntp_next_leap_sec)) { in __do_adjtimex() 861 static inline struct pps_normtime pps_normalize_ts(struct timespec64 ts) in pps_normalize_ts() argument 864 .sec = ts.tv_sec, in pps_normalize_ts() 865 .nsec = ts.tv_nsec in pps_normalize_ts()
|
D | posix-timers.c | 1193 struct timespec64 ts; in SYSCALL_DEFINE2() local 1198 if (get_old_timespec32(&ts, tp)) in SYSCALL_DEFINE2() 1201 return kc->clock_set(which_clock, &ts); in SYSCALL_DEFINE2() 1208 struct timespec64 ts; in SYSCALL_DEFINE2() local 1214 err = kc->clock_get_timespec(which_clock, &ts); in SYSCALL_DEFINE2() 1216 if (!err && put_old_timespec32(&ts, tp)) in SYSCALL_DEFINE2() 1244 struct timespec64 ts; in SYSCALL_DEFINE2() local 1250 err = kc->clock_getres(which_clock, &ts); in SYSCALL_DEFINE2() 1251 if (!err && tp && put_old_timespec32(&ts, tp)) in SYSCALL_DEFINE2()
|
D | timer_list.c | 146 (unsigned long long)(ts->x)) in print_cpu() 149 (unsigned long long)(ktime_to_ns(ts->x))) in print_cpu() 151 struct tick_sched *ts = tick_get_tick_sched(cpu); in print_cpu() local
|
D | namespace.c | 329 static void show_offset(struct seq_file *m, int clockid, struct timespec64 *ts) in show_offset() argument 344 seq_printf(m, "%-10s %10lld %9ld\n", clock, ts->tv_sec, ts->tv_nsec); in show_offset()
|
D | ntp_internal.h | 12 const struct timespec64 *ts,
|
D | hrtimer.c | 2016 int nanosleep_copyout(struct restart_block *restart, struct timespec64 *ts) in nanosleep_copyout() argument 2021 if (put_old_timespec32(ts, restart->nanosleep.compat_rmtp)) in nanosleep_copyout() 2026 if (put_timespec64(ts, restart->nanosleep.rmtp)) in nanosleep_copyout()
|
/kernel/trace/ |
D | trace_stat.c | 34 struct tracer_stat *ts; member 52 if (session->ts->stat_release) in __reset_stat_session() 53 session->ts->stat_release(snode->stat); in __reset_stat_session() 125 struct tracer_stat *ts = session->ts; in stat_seq_init() local 134 if (!ts->stat_cmp) in stat_seq_init() 135 ts->stat_cmp = dummy_cmp; in stat_seq_init() 137 stat = ts->stat_start(ts); in stat_seq_init() 141 ret = insert_stat(root, stat, ts->stat_cmp); in stat_seq_init() 149 stat = ts->stat_next(stat, i); in stat_seq_init() 155 ret = insert_stat(root, stat, ts->stat_cmp); in stat_seq_init() [all …]
|
D | ring_buffer.c | 280 u64 ts; in rb_event_time_stamp() local 282 ts = event->array[0]; in rb_event_time_stamp() 283 ts <<= TS_SHIFT; in rb_event_time_stamp() 284 ts += event->time_delta; in rb_event_time_stamp() 286 return ts; in rb_event_time_stamp() 389 u64 ts; member 778 u64 ts; in ring_buffer_event_time_stamp() local 782 ts = rb_event_time_stamp(event); in ring_buffer_event_time_stamp() 783 return rb_fix_abs_ts(ts, cpu_buffer->tail_page->page->time_stamp); in ring_buffer_event_time_stamp() 800 if (!rb_time_read(&cpu_buffer->write_stamp, &ts)) in ring_buffer_event_time_stamp() [all …]
|
D | trace_output.c | 551 unsigned long long abs_ts = iter->ts - iter->array_buffer->time_start; in lat_print_timestamp() 552 unsigned long long rel_ts = next_ts - iter->ts; in lat_print_timestamp() 568 ns2usecs(iter->ts), in lat_print_timestamp() 575 iter->ts, abs_ts, rel_ts); in lat_print_timestamp() 591 unsigned long long ts) in trace_print_time() argument 597 t = ns2usecs(ts); in trace_print_time() 602 trace_seq_printf(s, " %12llu", ts); in trace_print_time() 630 trace_print_time(s, iter, iter->ts); in trace_print_context() 646 next_ts = iter->ts; in trace_print_lat_context() 1540 iter->ts - FUNC_REPEATS_GET_DELTA_TS(field)); in trace_func_repeats_print()
|
D | ring_buffer_benchmark.c | 16 u64 ts; member 89 u64 ts; in read_event() local 91 event = ring_buffer_consume(buffer, cpu, &ts, NULL); in read_event()
|
D | trace_mmiotrace.c | 171 unsigned long long t = ns2usecs(iter->ts); in mmio_print_rw() 216 unsigned long long t = ns2usecs(iter->ts); in mmio_print_map() 250 unsigned long long t = ns2usecs(iter->ts); in mmio_print_mark()
|
D | trace.c | 771 u64 ts; in buffer_ftrace_now() local 777 ts = ring_buffer_time_stamp(buf->buffer); in buffer_ftrace_now() 778 ring_buffer_normalize_time_stamp(buf->buffer, cpu, &ts); in buffer_ftrace_now() 780 return ts; in buffer_ftrace_now() 3589 peek_next_entry(struct trace_iterator *iter, int cpu, u64 *ts, in peek_next_entry() argument 3596 event = ring_buffer_iter_peek(buf_iter, ts); in peek_next_entry() 3601 event = ring_buffer_peek(iter->array_buffer->buffer, cpu, ts, in peek_next_entry() 3621 u64 next_ts = 0, ts; in __find_next_entry() local 3645 ent = peek_next_entry(iter, cpu, &ts, &lost_events); in __find_next_entry() 3650 if (ent && (!next || ts < next_ts)) { in __find_next_entry() [all …]
|
D | blktrace.c | 1252 unsigned long long ts = iter->ts; in blk_log_action_classic() local 1253 unsigned long nsec_rem = do_div(ts, NSEC_PER_SEC); in blk_log_action_classic() 1254 unsigned secs = (unsigned long)ts; in blk_log_action_classic() 1532 .time = iter->ts, in blk_trace_synthesize_old_trace()
|
D | trace_functions_graph.c | 504 usecs = iter->ts - iter->array_buffer->time_start; in print_graph_rel_time() 525 print_graph_abs_time(iter->ts, s); in print_graph_irq() 737 print_graph_abs_time(iter->ts, s); in print_graph_prologue()
|
/kernel/futex/ |
D | syscalls.c | 153 futex_init_timeout(u32 cmd, u32 op, struct timespec64 *ts, ktime_t *t) in futex_init_timeout() argument 155 if (!timespec64_valid(ts)) in futex_init_timeout() 158 *t = timespec64_to_ktime(*ts); in futex_init_timeout() 172 struct timespec64 ts; in SYSCALL_DEFINE6() local 177 if (get_timespec64(&ts, utime)) in SYSCALL_DEFINE6() 179 ret = futex_init_timeout(cmd, op, &ts, &t); in SYSCALL_DEFINE6() 254 struct timespec64 ts; in SYSCALL_DEFINE5() local 276 if (get_timespec64(&ts, timeout)) in SYSCALL_DEFINE5() 283 ret = futex_init_timeout(FUTEX_WAIT_BITSET, flag_init, &ts, &time); in SYSCALL_DEFINE5() 367 struct timespec64 ts; in SYSCALL_DEFINE6() local [all …]
|
/kernel/irq/ |
D | timings.c | 473 static inline void irq_timings_store(int irq, struct irqt_stat *irqs, u64 ts) in irq_timings_store() argument 482 irqs->last_ts = ts; in irq_timings_store() 489 interval = ts - old_ts; in irq_timings_store() 541 u64 ts, next_evt = U64_MAX; in irq_timings_next_event() local 569 irq = irq_timing_decode(irqts->values[i], &ts); in irq_timings_next_event() 572 irq_timings_store(irq, this_cpu_ptr(s), ts); in irq_timings_next_event() 583 ts = __irq_timings_next_event(irqs, i, now); in irq_timings_next_event() 584 if (ts <= now) in irq_timings_next_event() 587 if (ts < next_evt) in irq_timings_next_event() 588 next_evt = ts; in irq_timings_next_event() [all …]
|
D | internals.h | 364 static __always_inline void irq_timings_push(u64 ts, int irq) in irq_timings_push() argument 369 irq_timing_encode(ts, irq); in irq_timings_push()
|
/kernel/rcu/ |
D | tree_stall.h | 300 struct task_struct *ts[8]; in rcu_print_task_stall() local 313 ts[i++] = t; in rcu_print_task_stall() 314 if (i >= ARRAY_SIZE(ts)) in rcu_print_task_stall() 319 t = ts[--i]; in rcu_print_task_stall()
|
D | rcutorture.c | 271 u64 ts = trace_clock_local(); in rcu_trace_clock_local() local 273 (void)do_div(ts, NSEC_PER_USEC); in rcu_trace_clock_local() 274 return ts; in rcu_trace_clock_local() 411 unsigned long long ts; in rcu_read_delay() local 420 ts = rcu_trace_clock_local(); in rcu_read_delay() 426 do_trace_rcu_torture_read(cur_ops->name, NULL, ts, in rcu_read_delay() 1922 unsigned long long ts; in rcu_torture_one_read() local 1934 ts = rcu_trace_clock_local(); in rcu_torture_one_read() 1955 ts, started, completed); in rcu_torture_one_read()
|
/kernel/ |
D | signal.c | 3591 const struct timespec64 *ts) in do_sigtimedwait() argument 3599 if (ts) { in do_sigtimedwait() 3600 if (!timespec64_valid(ts)) in do_sigtimedwait() 3602 timeout = timespec64_to_ktime(*ts); in do_sigtimedwait() 3655 struct timespec64 ts; in SYSCALL_DEFINE4() local 3667 if (get_timespec64(&ts, uts)) in SYSCALL_DEFINE4() 3671 ret = do_sigtimedwait(&these, &info, uts ? &ts : NULL); in SYSCALL_DEFINE4() 3688 struct timespec64 ts; in SYSCALL_DEFINE4() local 3699 if (get_old_timespec32(&ts, uts)) in SYSCALL_DEFINE4() 3703 ret = do_sigtimedwait(&these, &info, uts ? &ts : NULL); in SYSCALL_DEFINE4()
|