/art/runtime/gc/ |
D | task_processor_test.cc | 32 RecursiveTask(TaskProcessor* task_processor, Atomic<size_t>* counter, size_t max_recursion) in RecursiveTask() argument 33 : HeapTask(NanoTime() + MsToNs(10)), task_processor_(task_processor), counter_(counter), in RecursiveTask() 71 Atomic<size_t> counter(0); in TEST_F() local 72 task_processor.AddTask(self, new RecursiveTask(&task_processor, &counter, kRecursion)); in TEST_F() 79 while (counter.LoadSequentiallyConsistent() != kRecursion) { in TEST_F() 90 counter.StoreSequentiallyConsistent(0); in TEST_F() 95 task_processor.AddTask(self, new RecursiveTask(&task_processor, &counter, kRecursion)); in TEST_F() 100 ASSERT_EQ(counter.LoadSequentiallyConsistent(), kRecursion); in TEST_F() 105 TestOrderTask(uint64_t expected_time, size_t expected_counter, size_t* counter) in TestOrderTask() argument 106 : HeapTask(expected_time), expected_counter_(expected_counter), counter_(counter) { in TestOrderTask() [all …]
|
/art/profman/ |
D | boot_image_profile.cc | 57 size_t counter = 0; in GenerateBootImageProfile() local 61 ++counter; in GenerateBootImageProfile() 70 if (counter >= options.compiled_method_threshold) { in GenerateBootImageProfile() 111 size_t counter = 0; in GenerateBootImageProfile() local 116 ++counter; in GenerateBootImageProfile() 119 if (counter == 0) { in GenerateBootImageProfile() 122 if (counter >= options.image_class_theshold) { in GenerateBootImageProfile() 125 } else if (is_clean && counter >= options.image_class_clean_theshold) { in GenerateBootImageProfile()
|
/art/test/135-MirandaDispatch/src/ |
D | Main.java | 20 static int counter = 0; field in Main 31 ++counter; in m() 46 if (counter != loopIterations * loopIterations) { in main() 47 System.out.println("Expected " + loopIterations * loopIterations + " got " + counter); in main()
|
/art/runtime/base/ |
D | debug_stack.h | 57 explicit DebugStackReferenceImpl(DebugStackRefCounterImpl<false>* counter ATTRIBUTE_UNUSED) {} in DebugStackReferenceImpl() 91 explicit DebugStackReferenceImpl(DebugStackRefCounterImpl<kIsDebug>* counter) in DebugStackReferenceImpl() argument 92 : counter_(counter), ref_count_(counter->IncrementRefCount()) { in DebugStackReferenceImpl()
|
D | variant_map.h | 80 static size_t counter = 0; in AllocateCounter() local 81 counter++; in AllocateCounter() 83 return counter; in AllocateCounter()
|
/art/test/566-polymorphic-inlining/src/ |
D | Main.java | 81 assertEquals(20001, counter); in main() 132 counter++; in increment() 134 public static int counter = 0; field in Main
|
/art/test/441-checker-inliner/src/ |
D | Main.java | 217 private static int counter = 42; field in Main 220 return ++counter; in incCounter()
|
/art/test/114-ParallelGC/src/ |
D | Main.java | 33 private final static AtomicInteger counter = new AtomicInteger(); field in Main 112 int number = counter.incrementAndGet(); in work()
|
/art/test/906-iterate-heap/ |
D | iterate_heap.cc | 83 : counter(_counter), in Java_art_Test906_iterateThroughHeapCount() 91 counter++; in Java_art_Test906_iterateThroughHeapCount() 92 if (counter == stop_after) { in Java_art_Test906_iterateThroughHeapCount() 98 jint counter; in Java_art_Test906_iterateThroughHeapCount() member in art::Test906IterateHeap::Java_art_Test906_iterateThroughHeapCount::CountIterationConfig 105 if (config.counter > config.stop_after) { in Java_art_Test906_iterateThroughHeapCount() 109 return config.counter; in Java_art_Test906_iterateThroughHeapCount()
|
/art/test/547-regression-trycatch-critic-edge/smali/ |
D | TestCase.smali | 29 move v2, p0 # v2 = outer loop counter 36 move v3, p1 # v3 = inner loop counter
|
/art/runtime/ |
D | oat.cc | 453 ssize_t counter = static_cast<ssize_t>(index); in GetStoreKeyValuePairByIndex() local 455 while (ptr < end && counter >= 0) { in GetStoreKeyValuePairByIndex() 462 if (counter == 0) { in GetStoreKeyValuePairByIndex() 467 counter--; in GetStoreKeyValuePairByIndex()
|
D | debugger.cc | 3214 size_t* counter = GetReferenceCounterForEvent(req.InstrumentationEvent()); in RequestDeoptimizationLocked() local 3215 CHECK(counter != nullptr) << StringPrintf("No counter for instrumentation event 0x%x", in RequestDeoptimizationLocked() 3217 if (*counter == 0) { in RequestDeoptimizationLocked() 3222 *counter = *counter + 1; in RequestDeoptimizationLocked() 3227 size_t* counter = GetReferenceCounterForEvent(req.InstrumentationEvent()); in RequestDeoptimizationLocked() local 3228 CHECK(counter != nullptr) << StringPrintf("No counter for instrumentation event 0x%x", in RequestDeoptimizationLocked() 3230 *counter = *counter - 1; in RequestDeoptimizationLocked() 3231 if (*counter == 0) { in RequestDeoptimizationLocked()
|
/art/runtime/gc/accounting/ |
D | space_bitmap_test.cc | 134 explicit SimpleCounter(size_t* counter) : count_(counter) {} in SimpleCounter() argument
|
/art/test/989-method-trace-throw/src/art/ |
D | Test989.java | 431 static int counter = 0; field in Test989 433 return new TestObject(counter++); in mkTestObject()
|
/art/runtime/jit/ |
D | profile_saver.cc | 348 const uint16_t counter = method.GetCounter(); in SampleClassesAndExecutedMethods() local 353 counter >= hot_method_sample_threshold) { in SampleClassesAndExecutedMethods() 355 } else if (counter != 0) { in SampleClassesAndExecutedMethods()
|
/art/compiler/optimizing/ |
D | intrinsics_x86_64.cc | 1682 CpuRegister counter = locations->GetTemp(0).AsRegister<CpuRegister>(); in GenerateStringIndexOf() local 1689 DCHECK_EQ(counter.AsRegister(), RCX); in GenerateStringIndexOf() 1739 __ movl(counter, string_length); in GenerateStringIndexOf() 1750 __ xorl(counter, counter); in GenerateStringIndexOf() 1752 __ cmov(kGreater, counter, start_index, /* is64bit */ false); // 32-bit copy is enough. in GenerateStringIndexOf() 1758 __ leaq(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_1, value_offset)); in GenerateStringIndexOf() 1762 __ leaq(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_2, value_offset)); in GenerateStringIndexOf() 1765 __ leaq(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_2, value_offset)); in GenerateStringIndexOf() 1768 __ negq(counter); // Needs to be 64-bit negation, as the address computation is 64-bit. in GenerateStringIndexOf() 1769 __ leaq(counter, Address(string_length, counter, ScaleFactor::TIMES_1, 0)); in GenerateStringIndexOf() [all …]
|
D | intrinsics_x86.cc | 1515 Register counter = locations->GetTemp(0).AsRegister<Register>(); in GenerateStringIndexOf() local 1524 DCHECK_EQ(counter, ECX); in GenerateStringIndexOf() 1575 __ movl(counter, string_length); in GenerateStringIndexOf() 1587 __ xorl(counter, counter); in GenerateStringIndexOf() 1589 __ cmovl(kGreater, counter, start_index); in GenerateStringIndexOf() 1596 __ leal(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_1, value_offset)); in GenerateStringIndexOf() 1601 __ leal(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_2, value_offset)); in GenerateStringIndexOf() 1607 __ leal(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_2, value_offset)); in GenerateStringIndexOf() 1609 __ negl(counter); in GenerateStringIndexOf() 1610 __ leal(counter, Address(string_length, counter, ScaleFactor::TIMES_1, 0)); in GenerateStringIndexOf() [all …]
|
/art/runtime/arch/arm64/ |
D | quick_entrypoints_arm64.S | 877 .macro LOADREG counter size register return 879 add \counter, \counter, 12
|