/kernel/ |
D | acct.c | 336 int exp, rnd; in encode_comp_t() local 338 exp = rnd = 0; in encode_comp_t() 342 exp++; in encode_comp_t() 350 exp++; in encode_comp_t() 353 if (exp > (((comp_t) ~0U) >> MANTSIZE)) in encode_comp_t() 358 exp <<= MANTSIZE; /* Shift the exponent into place */ in encode_comp_t() 359 exp += value; /* and add on the mantissa. */ in encode_comp_t() 360 return exp; in encode_comp_t() 380 int exp, rnd; in encode_comp2_t() local 382 exp = (value > (MAXFRACT2>>1)); in encode_comp2_t() [all …]
|
/kernel/rcu/ |
D | refscale.c | 664 int exp, r; in main_func() local 688 for (exp = 0; exp < nruns && !torture_must_stop(); exp++) { in main_func() 698 exp_idx = exp; in main_func() 716 result_avg[exp] = div_u64(1000 * process_durations(nreaders), nreaders * loops); in main_func() 723 for (exp = 0; exp < nruns; exp++) { in main_func() 727 avg = div_u64_rem(result_avg[exp], 1000, &rem); in main_func() 728 sprintf(buf1, "%d\t%llu.%03u\n", exp + 1, avg, rem); in main_func()
|
D | tree_plugin.h | 263 if (blkd_state & RCU_EXP_BLKD && rdp->cpu_no_qs.b.exp) in rcu_preempt_ctxt_queue() 266 WARN_ON_ONCE(rdp->cpu_no_qs.b.exp); in rcu_preempt_ctxt_queue() 357 if (rdp->cpu_no_qs.b.exp) in rcu_note_context_switch() 484 if (!special.s && !rdp->cpu_no_qs.b.exp) { in rcu_preempt_deferred_qs_irqrestore() 505 if (rdp->cpu_no_qs.b.exp) in rcu_preempt_deferred_qs_irqrestore() 588 return (__this_cpu_read(rcu_data.cpu_no_qs.b.exp) || in rcu_preempt_need_deferred_qs() 856 if (__this_cpu_read(rcu_data.cpu_no_qs.b.exp)) in rcu_qs() 946 if (READ_ONCE(rdp->cpu_no_qs.b.exp)) in rcu_preempt_deferred_qs()
|
D | tree_exp.h | 259 WRITE_ONCE(rdp->cpu_no_qs.b.exp, false); in rcu_report_exp_rdp() 643 "D."[!!data_race(rdp->cpu_no_qs.b.exp)]); in synchronize_rcu_expedited_wait() 753 WRITE_ONCE(rdp->cpu_no_qs.b.exp, true); in rcu_exp_handler() 775 WRITE_ONCE(rdp->cpu_no_qs.b.exp, true); in rcu_exp_handler() 822 __this_cpu_write(rcu_data.cpu_no_qs.b.exp, true); in rcu_exp_need_qs() 837 __this_cpu_read(rcu_data.cpu_no_qs.b.exp)) in rcu_exp_handler() 861 READ_ONCE(rdp->cpu_no_qs.b.exp)) { in sync_sched_exp_online_cleanup()
|
D | tree.h | 156 u8 exp; member
|
D | tasks.h | 707 ktime_t exp; in rcu_tasks_wait_gp() local 717 exp = jiffies_to_nsecs(fract); in rcu_tasks_wait_gp() 719 schedule_hrtimeout_range(&exp, jiffies_to_nsecs(HZ / 2), HRTIMER_MODE_REL_HARD); in rcu_tasks_wait_gp()
|
/kernel/time/ |
D | alarmtimer.c | 828 ktime_t exp = restart->nanosleep.expires; in alarm_timer_nsleep_restart() local 833 return alarmtimer_do_nsleep(&alarm, exp, type); in alarm_timer_nsleep_restart() 850 ktime_t exp; in alarm_timer_nsleep() local 864 exp = timespec64_to_ktime(*tsreq); in alarm_timer_nsleep() 869 exp = ktime_add_safe(now, exp); in alarm_timer_nsleep() 871 exp = timens_ktime_to_host(which_clock, exp); in alarm_timer_nsleep() 874 ret = alarmtimer_do_nsleep(&alarm, exp, type); in alarm_timer_nsleep() 883 restart->nanosleep.expires = exp; in alarm_timer_nsleep()
|
D | ntp.c | 512 ktime_t exp = ktime_set(ktime_get_real_seconds(), 0); in sched_sync_hw_clock() local 515 exp = ktime_add_ns(exp, 2ULL * NSEC_PER_SEC - offset_nsec); in sched_sync_hw_clock() 517 exp = ktime_add_ns(exp, SYNC_PERIOD_NS - offset_nsec); in sched_sync_hw_clock() 519 hrtimer_start(&sync_hrtimer, exp, HRTIMER_MODE_ABS); in sched_sync_hw_clock()
|
D | posix-cpu-timers.c | 704 u64 exp = bump_cpu_timer(timer, val); in posix_cpu_timer_set() local 706 if (val < exp) { in posix_cpu_timer_set() 707 old_expires = exp - val; in posix_cpu_timer_set()
|
D | timer.c | 2132 ktime_t exp = ktime_add_us(ktime_get(), min); in usleep_range_state() local 2138 if (!schedule_hrtimeout_range(&exp, delta, HRTIMER_MODE_ABS)) in usleep_range_state()
|
/kernel/kcsan/ |
D | core.c | 1225 int __tsan_atomic##bits##_compare_exchange_##strength(u##bits *ptr, u##bits *exp, \ 1227 int __tsan_atomic##bits##_compare_exchange_##strength(u##bits *ptr, u##bits *exp, \ 1236 return __atomic_compare_exchange_n(ptr, exp, val, weak, mo, fail_mo); \ 1241 u##bits __tsan_atomic##bits##_compare_exchange_val(u##bits *ptr, u##bits exp, u##bits val, \ 1243 u##bits __tsan_atomic##bits##_compare_exchange_val(u##bits *ptr, u##bits exp, u##bits val, \ 1252 __atomic_compare_exchange_n(ptr, &exp, val, 0, mo, fail_mo); \ 1253 return exp; \
|
/kernel/sched/ |
D | loadavg.c | 156 calc_load_n(unsigned long load, unsigned long exp, in calc_load_n() argument 159 return calc_load(load, fixed_power_int(exp, FSHIFT, n), active); in calc_load_n()
|
D | core.c | 10645 #define _POW10(exp) ((unsigned int)1e##exp) argument 10646 #define POW10(exp) _POW10(exp) argument
|