/kernel/time/ |
D | tick-sched.c | 154 static void tick_sched_do_timer(struct tick_sched *ts, ktime_t now) in tick_sched_do_timer() argument 187 if (ts->last_tick_jiffies != jiffies) { in tick_sched_do_timer() 188 ts->stalled_jiffies = 0; in tick_sched_do_timer() 189 ts->last_tick_jiffies = READ_ONCE(jiffies); in tick_sched_do_timer() 191 if (++ts->stalled_jiffies == MAX_STALLED_JIFFIES) { in tick_sched_do_timer() 193 ts->stalled_jiffies = 0; in tick_sched_do_timer() 194 ts->last_tick_jiffies = READ_ONCE(jiffies); in tick_sched_do_timer() 198 if (ts->inidle) in tick_sched_do_timer() 199 ts->got_idle_tick = 1; in tick_sched_do_timer() 202 static void tick_sched_handle(struct tick_sched *ts, struct pt_regs *regs) in tick_sched_handle() argument [all …]
|
D | time.c | 144 struct timespec64 ts; in SYSCALL_DEFINE2() local 146 ktime_get_real_ts64(&ts); in SYSCALL_DEFINE2() 147 if (put_user(ts.tv_sec, &tv->tv_sec) || in SYSCALL_DEFINE2() 148 put_user(ts.tv_nsec / 1000, &tv->tv_usec)) in SYSCALL_DEFINE2() 228 struct timespec64 ts; in COMPAT_SYSCALL_DEFINE2() local 230 ktime_get_real_ts64(&ts); in COMPAT_SYSCALL_DEFINE2() 231 if (put_user(ts.tv_sec, &tv->tv_sec) || in COMPAT_SYSCALL_DEFINE2() 232 put_user(ts.tv_nsec / 1000, &tv->tv_usec)) in COMPAT_SYSCALL_DEFINE2() 454 struct timespec64 ts = ns_to_timespec64(nsec); in ns_to_kernel_old_timeval() local 457 tv.tv_sec = ts.tv_sec; in ns_to_kernel_old_timeval() [all …]
|
D | timekeeping.c | 131 struct timespec64 ts; in tk_xtime() local 133 ts.tv_sec = tk->xtime_sec; in tk_xtime() 134 ts.tv_nsec = (long)(tk->tkr_mono.xtime_nsec >> tk->tkr_mono.shift); in tk_xtime() 135 return ts; in tk_xtime() 138 static void tk_set_xtime(struct timekeeper *tk, const struct timespec64 *ts) in tk_set_xtime() argument 140 tk->xtime_sec = ts->tv_sec; in tk_set_xtime() 141 tk->tkr_mono.xtime_nsec = (u64)ts->tv_nsec << tk->tkr_mono.shift; in tk_set_xtime() 144 static void tk_xtime_add(struct timekeeper *tk, const struct timespec64 *ts) in tk_xtime_add() argument 146 tk->xtime_sec += ts->tv_sec; in tk_xtime_add() 147 tk->tkr_mono.xtime_nsec += (u64)ts->tv_nsec << tk->tkr_mono.shift; in tk_xtime_add() [all …]
|
D | posix-clock.c | 250 static int pc_clock_gettime(clockid_t id, struct timespec64 *ts) in pc_clock_gettime() argument 260 err = cd.clk->ops.clock_gettime(cd.clk, ts); in pc_clock_gettime() 269 static int pc_clock_getres(clockid_t id, struct timespec64 *ts) in pc_clock_getres() argument 279 err = cd.clk->ops.clock_getres(cd.clk, ts); in pc_clock_getres() 288 static int pc_clock_settime(clockid_t id, const struct timespec64 *ts) in pc_clock_settime() argument 303 err = cd.clk->ops.clock_settime(cd.clk, ts); in pc_clock_settime()
|
D | test_udelay.c | 82 struct timespec64 ts; in udelay_test_show() local 84 ktime_get_ts64(&ts); in udelay_test_show() 86 loops_per_jiffy, (s64)ts.tv_sec, ts.tv_nsec); in udelay_test_show()
|
D | ntp.c | 714 int __do_adjtimex(struct __kernel_timex *txc, const struct timespec64 *ts, in __do_adjtimex() argument 774 txc->time.tv_sec = ts->tv_sec; in __do_adjtimex() 775 txc->time.tv_usec = ts->tv_nsec; in __do_adjtimex() 777 txc->time.tv_usec = ts->tv_nsec / NSEC_PER_USEC; in __do_adjtimex() 780 if (unlikely(ts->tv_sec >= ntp_next_leap_sec)) { in __do_adjtimex() 792 (ts->tv_sec == ntp_next_leap_sec)) { in __do_adjtimex() 813 static inline struct pps_normtime pps_normalize_ts(struct timespec64 ts) in pps_normalize_ts() argument 816 .sec = ts.tv_sec, in pps_normalize_ts() 817 .nsec = ts.tv_nsec in pps_normalize_ts()
|
D | posix-timers.c | 1193 struct timespec64 ts; in SYSCALL_DEFINE2() local 1198 if (get_old_timespec32(&ts, tp)) in SYSCALL_DEFINE2() 1201 return kc->clock_set(which_clock, &ts); in SYSCALL_DEFINE2() 1208 struct timespec64 ts; in SYSCALL_DEFINE2() local 1214 err = kc->clock_get_timespec(which_clock, &ts); in SYSCALL_DEFINE2() 1216 if (!err && put_old_timespec32(&ts, tp)) in SYSCALL_DEFINE2() 1244 struct timespec64 ts; in SYSCALL_DEFINE2() local 1250 err = kc->clock_getres(which_clock, &ts); in SYSCALL_DEFINE2() 1251 if (!err && tp && put_old_timespec32(&ts, tp)) in SYSCALL_DEFINE2()
|
D | timer_list.c | 161 (unsigned long long)(ts->x)) in print_cpu() 164 (unsigned long long)(ktime_to_ns(ts->x))) in print_cpu() 166 struct tick_sched *ts = tick_get_tick_sched(cpu); in print_cpu() local
|
D | namespace.c | 334 static void show_offset(struct seq_file *m, int clockid, struct timespec64 *ts) in show_offset() argument 349 seq_printf(m, "%-10s %10lld %9ld\n", clock, ts->tv_sec, ts->tv_nsec); in show_offset()
|
D | ntp_internal.h | 12 const struct timespec64 *ts,
|
D | hrtimer.c | 1914 int nanosleep_copyout(struct restart_block *restart, struct timespec64 *ts) in nanosleep_copyout() argument 1919 if (put_old_timespec32(ts, restart->nanosleep.compat_rmtp)) in nanosleep_copyout() 1924 if (put_timespec64(ts, restart->nanosleep.rmtp)) in nanosleep_copyout()
|
/kernel/trace/ |
D | trace_stat.c | 34 struct tracer_stat *ts; member 52 if (session->ts->stat_release) in __reset_stat_session() 53 session->ts->stat_release(snode->stat); in __reset_stat_session() 125 struct tracer_stat *ts = session->ts; in stat_seq_init() local 134 if (!ts->stat_cmp) in stat_seq_init() 135 ts->stat_cmp = dummy_cmp; in stat_seq_init() 137 stat = ts->stat_start(ts); in stat_seq_init() 141 ret = insert_stat(root, stat, ts->stat_cmp); in stat_seq_init() 149 stat = ts->stat_next(stat, i); in stat_seq_init() 155 ret = insert_stat(root, stat, ts->stat_cmp); in stat_seq_init() [all …]
|
D | ring_buffer.c | 301 u64 ts; in ring_buffer_event_time_stamp() local 303 ts = event->array[0]; in ring_buffer_event_time_stamp() 304 ts <<= TS_SHIFT; in ring_buffer_event_time_stamp() 305 ts += event->time_delta; in ring_buffer_event_time_stamp() 307 return ts; in ring_buffer_event_time_stamp() 428 u64 ts; member 1073 u64 ts; in rb_time_stamp() local 1077 ts = trace_clock_local(); in rb_time_stamp() 1079 ts = buffer->clock(); in rb_time_stamp() 1082 return ts << DEBUG_SHIFT; in rb_time_stamp() [all …]
|
D | ring_buffer_benchmark.c | 16 u64 ts; member 89 u64 ts; in read_event() local 91 event = ring_buffer_consume(buffer, cpu, &ts, NULL); in read_event()
|
D | trace_mmiotrace.c | 173 unsigned long long t = ns2usecs(iter->ts); in mmio_print_rw() 218 unsigned long long t = ns2usecs(iter->ts); in mmio_print_map() 252 unsigned long long t = ns2usecs(iter->ts); in mmio_print_mark()
|
D | trace_output.c | 541 unsigned long long abs_ts = iter->ts - iter->array_buffer->time_start; in lat_print_timestamp() 542 unsigned long long rel_ts = next_ts - iter->ts; in lat_print_timestamp() 558 ns2usecs(iter->ts), in lat_print_timestamp() 565 iter->ts, abs_ts, rel_ts); in lat_print_timestamp() 608 t = ns2usecs(iter->ts); in trace_print_context() 613 trace_seq_printf(s, " %12llu: ", iter->ts); in trace_print_context() 628 next_ts = iter->ts; in trace_print_lat_context()
|
D | trace.c | 771 u64 ts; in buffer_ftrace_now() local 777 ts = ring_buffer_time_stamp(buf->buffer, cpu); in buffer_ftrace_now() 778 ring_buffer_normalize_time_stamp(buf->buffer, cpu, &ts); in buffer_ftrace_now() 780 return ts; in buffer_ftrace_now() 3488 peek_next_entry(struct trace_iterator *iter, int cpu, u64 *ts, in peek_next_entry() argument 3495 event = ring_buffer_iter_peek(buf_iter, ts); in peek_next_entry() 3500 event = ring_buffer_peek(iter->array_buffer->buffer, cpu, ts, in peek_next_entry() 3520 u64 next_ts = 0, ts; in __find_next_entry() local 3544 ent = peek_next_entry(iter, cpu, &ts, &lost_events); in __find_next_entry() 3549 if (ent && (!next || ts < next_ts)) { in __find_next_entry() [all …]
|
D | blktrace.c | 1306 unsigned long long ts = iter->ts; in blk_log_action_classic() local 1307 unsigned long nsec_rem = do_div(ts, NSEC_PER_SEC); in blk_log_action_classic() 1308 unsigned secs = (unsigned long)ts; in blk_log_action_classic() 1586 .time = iter->ts, in blk_trace_synthesize_old_trace()
|
D | trace_functions_graph.c | 506 usecs = iter->ts - iter->array_buffer->time_start; in print_graph_rel_time() 527 print_graph_abs_time(iter->ts, s); in print_graph_irq() 739 print_graph_abs_time(iter->ts, s); in print_graph_prologue()
|
/kernel/irq/ |
D | timings.c | 473 static inline void irq_timings_store(int irq, struct irqt_stat *irqs, u64 ts) in irq_timings_store() argument 482 irqs->last_ts = ts; in irq_timings_store() 489 interval = ts - old_ts; in irq_timings_store() 541 u64 ts, next_evt = U64_MAX; in irq_timings_next_event() local 569 irq = irq_timing_decode(irqts->values[i], &ts); in irq_timings_next_event() 572 irq_timings_store(irq, this_cpu_ptr(s), ts); in irq_timings_next_event() 583 ts = __irq_timings_next_event(irqs, i, now); in irq_timings_next_event() 584 if (ts <= now) in irq_timings_next_event() 587 if (ts < next_evt) in irq_timings_next_event() 588 next_evt = ts; in irq_timings_next_event() [all …]
|
D | internals.h | 364 static __always_inline void irq_timings_push(u64 ts, int irq) in irq_timings_push() argument 369 irq_timing_encode(ts, irq); in irq_timings_push()
|
/kernel/rcu/ |
D | tree_stall.h | 263 struct task_struct *ts[8]; in rcu_print_task_stall() local 276 ts[i++] = t; in rcu_print_task_stall() 277 if (i >= ARRAY_SIZE(ts)) in rcu_print_task_stall() 282 t = ts[--i]; in rcu_print_task_stall()
|
D | rcutorture.c | 232 u64 ts = trace_clock_local(); in rcu_trace_clock_local() local 234 (void)do_div(ts, NSEC_PER_USEC); in rcu_trace_clock_local() 235 return ts; in rcu_trace_clock_local() 347 unsigned long long ts; in rcu_read_delay() local 356 ts = rcu_trace_clock_local(); in rcu_read_delay() 362 do_trace_rcu_torture_read(cur_ops->name, NULL, ts, in rcu_read_delay() 1386 unsigned long long ts; in rcu_torture_one_read() local 1392 ts = rcu_trace_clock_local(); in rcu_torture_one_read() 1416 ts, started, completed); in rcu_torture_one_read()
|
/kernel/futex/ |
D | core.c | 3799 struct timespec64 ts; in SYSCALL_DEFINE6() local 3809 if (get_timespec64(&ts, utime)) in SYSCALL_DEFINE6() 3811 if (!timespec64_valid(&ts)) in SYSCALL_DEFINE6() 3814 t = timespec64_to_ktime(ts); in SYSCALL_DEFINE6() 3995 struct timespec64 ts; in SYSCALL_DEFINE6() local 4003 if (get_old_timespec32(&ts, utime)) in SYSCALL_DEFINE6() 4005 if (!timespec64_valid(&ts)) in SYSCALL_DEFINE6() 4008 t = timespec64_to_ktime(ts); in SYSCALL_DEFINE6()
|
/kernel/ |
D | signal.c | 3486 const struct timespec64 *ts) in do_sigtimedwait() argument 3493 if (ts) { in do_sigtimedwait() 3494 if (!timespec64_valid(ts)) in do_sigtimedwait() 3496 timeout = timespec64_to_ktime(*ts); in do_sigtimedwait() 3549 struct timespec64 ts; in SYSCALL_DEFINE4() local 3561 if (get_timespec64(&ts, uts)) in SYSCALL_DEFINE4() 3565 ret = do_sigtimedwait(&these, &info, uts ? &ts : NULL); in SYSCALL_DEFINE4() 3582 struct timespec64 ts; in SYSCALL_DEFINE4() local 3593 if (get_old_timespec32(&ts, uts)) in SYSCALL_DEFINE4() 3597 ret = do_sigtimedwait(&these, &info, uts ? &ts : NULL); in SYSCALL_DEFINE4()
|