/art/runtime/gc/ |
D | task_processor_test.cc | 32 RecursiveTask(TaskProcessor* task_processor, Atomic<size_t>* counter, size_t max_recursion) in RecursiveTask() argument 33 : HeapTask(NanoTime() + MsToNs(10)), task_processor_(task_processor), counter_(counter), in RecursiveTask() 71 Atomic<size_t> counter(0); in TEST_F() local 72 task_processor.AddTask(self, new RecursiveTask(&task_processor, &counter, kRecursion)); in TEST_F() 79 while (counter.LoadSequentiallyConsistent() != kRecursion) { in TEST_F() 90 counter.StoreSequentiallyConsistent(0); in TEST_F() 95 task_processor.AddTask(self, new RecursiveTask(&task_processor, &counter, kRecursion)); in TEST_F() 100 ASSERT_EQ(counter.LoadSequentiallyConsistent(), kRecursion); in TEST_F() 105 TestOrderTask(uint64_t expected_time, size_t expected_counter, size_t* counter) in TestOrderTask() argument 106 : HeapTask(expected_time), expected_counter_(expected_counter), counter_(counter) { in TestOrderTask() [all …]
|
D | heap.cc | 1929 InstanceCounter counter(classes, use_is_assignable_from, counts); in CountInstances() local 1930 VisitObjects(InstanceCounter::Callback, &counter); in CountInstances()
|
/art/test/135-MirandaDispatch/src/ |
D | Main.java | 20 static int counter = 0; field in Main 31 ++counter; in m() 46 if (counter != loopIterations * loopIterations) { in main() 47 System.out.println("Expected " + loopIterations * loopIterations + " got " + counter); in main()
|
/art/runtime/base/ |
D | debug_stack.h | 57 explicit DebugStackReferenceImpl(DebugStackRefCounterImpl<false>* counter ATTRIBUTE_UNUSED) {} in DebugStackReferenceImpl() 91 explicit DebugStackReferenceImpl(DebugStackRefCounterImpl<kIsDebug>* counter) in DebugStackReferenceImpl() argument 92 : counter_(counter), ref_count_(counter->IncrementRefCount()) { in DebugStackReferenceImpl()
|
D | variant_map.h | 79 static size_t counter = 0; in AllocateCounter() local 80 counter++; in AllocateCounter() 82 return counter; in AllocateCounter()
|
/art/test/566-polymorphic-inlining/src/ |
D | Main.java | 78 assertEquals(20001, counter); in main() 127 counter++; in increment() 129 public static int counter = 0; field in Main
|
/art/test/441-checker-inliner/src/ |
D | Main.java | 217 private static int counter = 42; field in Main 220 return ++counter; in incCounter()
|
/art/test/114-ParallelGC/src/ |
D | Main.java | 33 private final static AtomicInteger counter = new AtomicInteger(); field in Main 112 int number = counter.incrementAndGet(); in work()
|
/art/test/547-regression-trycatch-critical-edge/smali/ |
D | TestCase.smali | 29 move v2, p0 # v2 = outer loop counter 36 move v3, p1 # v3 = inner loop counter
|
/art/runtime/gc/accounting/ |
D | space_bitmap_test.cc | 92 explicit SimpleCounter(size_t* counter) : count_(counter) {} in SimpleCounter() argument
|
/art/runtime/ |
D | oat.cc | 434 ssize_t counter = static_cast<ssize_t>(index); in GetStoreKeyValuePairByIndex() local 436 while (ptr < end && counter >= 0) { in GetStoreKeyValuePairByIndex() 443 if (counter == 0) { in GetStoreKeyValuePairByIndex() 448 counter--; in GetStoreKeyValuePairByIndex()
|
D | debugger.cc | 3163 size_t* counter = GetReferenceCounterForEvent(req.InstrumentationEvent()); in RequestDeoptimizationLocked() local 3164 CHECK(counter != nullptr) << StringPrintf("No counter for instrumentation event 0x%x", in RequestDeoptimizationLocked() 3166 if (*counter == 0) { in RequestDeoptimizationLocked() 3171 *counter = *counter + 1; in RequestDeoptimizationLocked() 3176 size_t* counter = GetReferenceCounterForEvent(req.InstrumentationEvent()); in RequestDeoptimizationLocked() local 3177 CHECK(counter != nullptr) << StringPrintf("No counter for instrumentation event 0x%x", in RequestDeoptimizationLocked() 3179 *counter = *counter - 1; in RequestDeoptimizationLocked() 3180 if (*counter == 0) { in RequestDeoptimizationLocked()
|
/art/compiler/optimizing/ |
D | intrinsics_x86.cc | 1410 Register counter = locations->GetTemp(0).AsRegister<Register>(); in GenerateStringIndexOf() local 1417 DCHECK_EQ(counter, ECX); in GenerateStringIndexOf() 1458 __ movl(counter, string_length); in GenerateStringIndexOf() 1470 __ xorl(counter, counter); in GenerateStringIndexOf() 1472 __ cmovl(kGreater, counter, start_index); in GenerateStringIndexOf() 1475 __ leal(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_2, value_offset)); in GenerateStringIndexOf() 1479 __ negl(counter); in GenerateStringIndexOf() 1480 __ leal(counter, Address(string_length, counter, ScaleFactor::TIMES_1, 0)); in GenerateStringIndexOf() 1492 __ subl(string_length, counter); in GenerateStringIndexOf()
|
D | intrinsics_x86_64.cc | 1512 CpuRegister counter = locations->GetTemp(0).AsRegister<CpuRegister>(); in GenerateStringIndexOf() local 1519 DCHECK_EQ(counter.AsRegister(), RCX); in GenerateStringIndexOf() 1560 __ movl(counter, string_length); in GenerateStringIndexOf() 1572 __ xorl(counter, counter); in GenerateStringIndexOf() 1574 __ cmov(kGreater, counter, start_index, /* is64bit */ false); // 32-bit copy is enough. in GenerateStringIndexOf() 1577 __ leaq(string_obj, Address(string_obj, counter, ScaleFactor::TIMES_2, value_offset)); in GenerateStringIndexOf() 1580 __ negq(counter); // Needs to be 64-bit negation, as the address computation is 64-bit. in GenerateStringIndexOf() 1581 __ leaq(counter, Address(string_length, counter, ScaleFactor::TIMES_1, 0)); in GenerateStringIndexOf() 1593 __ subl(string_length, counter); in GenerateStringIndexOf()
|
/art/runtime/arch/arm64/ |
D | quick_entrypoints_arm64.S | 738 .macro LOADREG counter size register return 740 add \counter, \counter, 12
|