Searched refs:elapsed (Results 1 – 5 of 5) sorted by relevance
53 static inline int virt_timer_forward(u64 elapsed) in virt_timer_forward() argument59 elapsed = atomic64_add_return(elapsed, &virt_timer_elapsed); in virt_timer_forward()60 return elapsed >= atomic64_read(&virt_timer_current); in virt_timer_forward()236 unsigned long elapsed; in virt_timer_expire() local241 elapsed = atomic64_read(&virt_timer_elapsed); in virt_timer_expire()243 if (timer->expires < elapsed) in virt_timer_expire()247 timer->expires -= elapsed; in virt_timer_expire()254 atomic64_sub(elapsed, &virt_timer_elapsed); in virt_timer_expire()
109 s64 elapsed; in __kpit_elapsed() local126 elapsed = ps->period - ktime_to_ns(remaining); in __kpit_elapsed()128 return elapsed; in __kpit_elapsed()
1489 u64 offset, ns, elapsed; in kvm_write_tsc() local1499 elapsed = ns - kvm->arch.last_tsc_nsec; in kvm_write_tsc()1525 do_div(elapsed, 1000); in kvm_write_tsc()1526 usdiff -= elapsed; in kvm_write_tsc()1552 u64 delta = nsec_to_cycles(vcpu, elapsed); in kvm_write_tsc()
796 cycles_t elapsed; in record_send_stats() local799 elapsed = time2 - time1; in record_send_stats()800 stat->s_time += elapsed; in record_send_stats()804 bcp->period_time += elapsed; in record_send_stats()805 if ((elapsed > congested_cycles) && in record_send_stats()
134 Calculate elapsed time for every exit/enter cycle. A per-vcpu