/art/test/149-suspend-all-stress/ |
D | suspend_all.cc | 34 const uint64_t start_time = NanoTime(); in Java_Main_suspendAndResume() 37 while (NanoTime() - start_time < MsToNs(10 * 1000)) { in Java_Main_suspendAndResume()
|
/art/libartbase/base/ |
D | time_utils_test.cc | 53 auto start = NanoTime(); in TEST() 55 EXPECT_GT(NanoTime() - start, MsToNs(1000)); in TEST()
|
D | leb128_test.cc | 280 uint64_t last_time = NanoTime(); in TEST() 285 uint64_t cur_time = NanoTime(); in TEST() 291 last_time = NanoTime(); in TEST() 296 uint64_t cur_time = NanoTime(); in TEST()
|
D | time_utils.h | 67 uint64_t NanoTime();
|
D | time_utils.cc | 183 uint64_t NanoTime() { in NanoTime() function 218 return NanoTime(); in ProcessCpuNanoTime()
|
/art/runtime/ |
D | barrier.cc | 79 uint64_t abs_timeout = NanoTime() + MsToNs(timeout_ms); in Increment() 84 uint64_t now = NanoTime(); in Increment()
|
D | thread_pool.cc | 240 start_time_ = NanoTime(); in StartWorkers() 268 const uint64_t wait_start = kMeasureWaitTime ? NanoTime() : 0; in GetTask() 271 const uint64_t wait_end = NanoTime(); in GetTask()
|
D | thread_list.cc | 523 const uint64_t suspend_start_time = NanoTime(); in FlipThreadRoots() 531 suspend_all_historam_.AdjustAndAddValue(NanoTime() - suspend_start_time); in FlipThreadRoots() 534 collector->RegisterPause(NanoTime() - suspend_start_time); in FlipThreadRoots() 616 const uint64_t start_time = NanoTime(); in SuspendAll() 640 const uint64_t end_time = NanoTime(); in SuspendAll() 722 const uint64_t start_time = NanoTime(); in SuspendAllInternal() 734 const uint64_t wait_time = NanoTime() - start_time; in SuspendAllInternal() 880 const uint64_t start_time = NanoTime(); in SuspendThreadByPeer() 958 const uint64_t total_delay = NanoTime() - start_time; in SuspendThreadByPeer() 1000 const uint64_t start_time = NanoTime(); in SuspendThreadByThreadId() [all …]
|
/art/runtime/jit/ |
D | profile_saver.cc | 139 const uint64_t end_time = NanoTime() + MsToNs(force_early_first_save in Run() 143 const uint64_t current_time = NanoTime(); in Run() 170 uint64_t sleep_start = NanoTime(); in Run() 182 sleep_time = NanoTime() - sleep_start; in Run() 197 sleep_time = NanoTime() - sleep_start; in Run() 205 total_ms_of_sleep_ += NsToMs(NanoTime() - sleep_start); in Run() 213 uint64_t start_work = NanoTime(); in Run() 232 total_ns_of_work_ += NanoTime() - start_work; in Run() 293 last_time_ns_saver_woke_up_ = NanoTime(); in WakeUpSaver() 305 if ((NanoTime() - last_time_ns_saver_woke_up_) > MsToNs(options_.GetMinSavePeriodMs())) { in NotifyJitActivityInternal() [all …]
|
/art/runtime/gc/collector/ |
D | garbage_collector.cc | 159 uint64_t start_time = NanoTime(); in Run() 180 uint64_t end_time = NanoTime(); in Run() 266 : start_time_(NanoTime()), collector_(collector), with_reporting_(with_reporting) { in ScopedPause() 278 collector_->RegisterPause(NanoTime() - start_time_); in ~ScopedPause()
|
/art/runtime/gc/ |
D | task_processor_test.cc | 33 : HeapTask(NanoTime() + MsToNs(10)), task_processor_(task_processor), counter_(counter), in RecursiveTask() 120 const uint64_t current_time = NanoTime(); in TEST_F()
|
D | task_processor.cc | 59 const uint64_t current_time = NanoTime(); in GetTask()
|
D | heap.cc | 372 last_time_homogeneous_space_compaction_by_oom_(NanoTime()), in Heap() 383 (NanoTime() / kGcCountRateHistogramWindowDuration) * kGcCountRateHistogramWindowDuration), in Heap() 955 wait_start = NanoTime(); in IncrementDisableThreadFlip() 964 uint64_t wait_time = NanoTime() - wait_start; in IncrementDisableThreadFlip() 1000 uint64_t wait_start = NanoTime(); in ThreadFlipBegin() 1010 uint64_t wait_time = NanoTime() - wait_start; in ThreadFlipBegin() 1279 (NanoTime() / kGcCountRateHistogramWindowDuration) * kGcCountRateHistogramWindowDuration; in ResetGcPerformanceInfo() 1501 uint64_t start_time = NanoTime(); in Trim() 1504 << PrettyDuration(NanoTime() - start_time); in Trim() 1560 const uint64_t start_ns = NanoTime(); in TrimSpaces() [all …]
|
D | reference_processor.cc | 279 : HeapTask(NanoTime()), cleared_references_(cleared_references) { in ClearedReferenceTask()
|
/art/compiler/jit/ |
D | jit_logger.cc | 240 header.time_stamp_ = art::NanoTime(); // CLOCK_MONOTONIC clock is required. in WriteJitDumpHeader() 277 jit_code.time_stamp_ = art::NanoTime(); // CLOCK_MONOTONIC clock is required. in WriteJitDumpLog()
|
/art/compiler/utils/ |
D | dedupe_set-inl.h | 193 hash_start = NanoTime(); 197 uint64_t hash_end = NanoTime();
|
/art/tools/dexanalyze/ |
D | dexanalyze_strings.cc | 503 uint64_t start = NanoTime(); in Benchmark() 513 timings->time_equal_comparisons_ += NanoTime() - start; in Benchmark() 515 start = NanoTime(); in Benchmark() 526 timings->time_non_equal_comparisons_ += NanoTime() - start; in Benchmark()
|
/art/runtime/gc/space/ |
D | dlmalloc_space.cc | 118 start_time = NanoTime(); in Create() 145 LOG(INFO) << "DlMallocSpace::Create exiting (" << PrettyDuration(NanoTime() - start_time) in Create()
|
D | rosalloc_space.cc | 144 start_time = NanoTime(); in Create() 173 LOG(INFO) << "RosAllocSpace::Create exiting (" << PrettyDuration(NanoTime() - start_time) in Create()
|
D | region_space.cc | 503 uint64_t start_time = NanoTime(); in ClearFromSpace() 507 madvise_time_ += NanoTime() - start_time; in ClearFromSpace()
|
/art/runtime/base/ |
D | timing_logger.h | 106 time_ = NanoTime(); in Timing()
|
D | mutex.cc | 175 start_nano_time_(kLogLockContentions ? NanoTime() : 0) { in ScopedContentionRecorder() 186 uint64_t end_nano_time = NanoTime(); in ~ScopedContentionRecorder()
|
/art/dexlayout/ |
D | compact_dex_writer.cc | 89 uint64_t start_time = NanoTime(); in WriteDebugInfoOffsetTable() 98 uint64_t end_time = NanoTime(); in WriteDebugInfoOffsetTable()
|
/art/compiler/optimizing/ |
D | scheduler.h | 457 seed_ = static_cast<uint32_t>(NanoTime()); in RandomSchedulingNodeSelector()
|
/art/runtime/native/ |
D | dalvik_system_ZygoteHooks.cc | 413 std::srand(static_cast<uint32_t>(NanoTime())); in ZygoteHooks_nativePostForkChild()
|