/art/runtime/ |
D | subtype_check.h | 475 SubtypeCheckBitsAndStatus old_value = ReadField(klass); in WriteSubtypeCheckBits() local 478 SubtypeCheckBits old_bits = old_value.subtype_check_info_; in WriteSubtypeCheckBits() 485 full_old = old_value; in WriteSubtypeCheckBits() 486 old_status = old_value.status_; in WriteSubtypeCheckBits() 488 new_value = old_value; in WriteSubtypeCheckBits() 503 old_value.int32_alias_, in WriteSubtypeCheckBits() 519 SubtypeCheckBitsAndStatus old_value = ReadField(klass); in WriteStatusImpl() local 520 old_status = old_value.status_; in WriteStatusImpl() 527 new_value = old_value; in WriteStatusImpl() 532 old_value.int32_alias_, in WriteStatusImpl() [all …]
|
D | subtype_check_test.cc | 90 int32_t old_value, in CasField32() 96 if (old_value == GetField32Volatile(offset)) { in CasField32()
|
D | runtime.cc | 597 auto old_value = gAborting.fetch_add(1); // set before taking any locks in Abort() local 600 if (old_value == 0) { in Abort()
|
/art/runtime/base/ |
D | quasi_atomic.h | 142 static bool Cas64(int64_t old_value, int64_t new_value, volatile int64_t* addr) { in Cas64() argument 144 return __sync_bool_compare_and_swap(addr, old_value, new_value); in Cas64() 146 return SwapMutexCas64(old_value, new_value, addr); in Cas64() 167 static bool SwapMutexCas64(int64_t old_value, int64_t new_value, volatile int64_t* addr);
|
D | quasi_atomic.cc | 58 bool QuasiAtomic::SwapMutexCas64(int64_t old_value, int64_t new_value, volatile int64_t* addr) { in SwapMutexCas64() argument 60 if (*addr == old_value) { in SwapMutexCas64()
|
/art/runtime/mirror/ |
D | object-inl.h | 570 int64_t old_value, in CasFieldWeakSequentiallyConsistent64() argument 574 Runtime::Current()->RecordWriteField64(this, field_offset, old_value, true); in CasFieldWeakSequentiallyConsistent64() 579 return atomic_addr->CompareAndSetWeakSequentiallyConsistent(old_value, new_value); in CasFieldWeakSequentiallyConsistent64() 584 int64_t old_value, in CasFieldStrongSequentiallyConsistent64() argument 588 Runtime::Current()->RecordWriteField64(this, field_offset, old_value, true); in CasFieldStrongSequentiallyConsistent64() 593 return atomic_addr->CompareAndSetStrongSequentiallyConsistent(old_value, new_value); in CasFieldStrongSequentiallyConsistent64() 682 ObjPtr<Object> old_value, in CasFieldObjectWithoutWriteBarrier() argument 687 VerifyCAS<kVerifyFlags>(new_value, old_value); in CasFieldObjectWithoutWriteBarrier() 689 Runtime::Current()->RecordWriteFieldReference(this, field_offset, old_value, true); in CasFieldObjectWithoutWriteBarrier() 691 uint32_t old_ref(PtrCompression<kPoisonHeapReferences, Object>::Compress(old_value)); in CasFieldObjectWithoutWriteBarrier() [all …]
|
D | object-readbarrier-inl.h | 42 int32_t old_value, in CasField32() argument 50 Runtime::Current()->RecordWriteField32(this, field_offset, old_value, true); in CasField32() 58 return atomic_addr->CompareAndSet(old_value, new_value, mode, memory_order); in CasField32()
|
D | var_handle.cc | 503 T old_value = atom->exchange(new_value_, MO); in Access() local 504 StoreResult(old_value, result_); in Access() 531 T old_value = atom->load(std::memory_order_relaxed); in Apply() local 534 new_value = old_value + addend; in Apply() 535 } while (!atom->compare_exchange_weak(old_value, new_value, MO, std::memory_order_relaxed)); in Apply() 536 return old_value; in Apply() 547 T old_value = AtomicGetAndAddOperator<T, kIsFloatingPoint, MO>::Apply(addr, addend_); in Access() local 548 StoreResult(old_value, result_); in Access() 598 T old_value = atom->fetch_or(value_, MO); in Access() local 599 StoreResult(old_value, result_); in Access() [all …]
|
D | object.h | 316 ObjPtr<Object> old_value, 325 ObjPtr<Object> old_value, 335 ObjPtr<Object> old_value, 496 int32_t old_value, 538 int64_t old_value, 546 int64_t old_value, 746 ALWAYS_INLINE void VerifyCAS(ObjPtr<mirror::Object> new_value, ObjPtr<mirror::Object> old_value) { in VerifyCAS() argument 748 VerifyRead<kVerifyFlags>(old_value); in VerifyCAS()
|
D | class-inl.h | 1105 T old_value = GetFieldPtrWithSize<T, kVerifyFlags>(member_offset, pointer_size); in FixupNativePointer() local 1106 T new_value = visitor(old_value, address); in FixupNativePointer() 1107 if (old_value != new_value) { in FixupNativePointer()
|
/art/openjdkjvmti/ |
D | ti_search.cc | 143 ScopedLocalRef<jobject> old_value(self->GetJniEnv(), in Update() local 148 DCHECK(old_value.get() != nullptr); in Update() 150 str_value = self->DecodeJObject(old_value.get())->AsString()->ToModifiedUtf8(); in Update() 151 self->GetJniEnv()->DeleteLocalRef(old_value.release()); in Update()
|
/art/libartbase/base/ |
D | stl_util.h | 108 void ReplaceElement(Container& container, const T& old_value, const T& new_value) { in ReplaceElement() argument 109 auto it = std::find(container.begin(), container.end(), old_value); in ReplaceElement()
|
D | leb128_test.cc | 259 uint32_t old_value = uleb128_tests[i].decoded; in TEST() local 262 if (new_value <= old_value) { in TEST() 264 uint8_t* old_end = EncodeUnsignedLeb128(encoded_data, old_value); in TEST()
|
D | leb128.h | 275 uint32_t old_value = DecodeUnsignedLeb128(&old_end); in UpdateUnsignedLeb128() local 276 DCHECK_LE(UnsignedLeb128Size(value), UnsignedLeb128Size(old_value)); in UpdateUnsignedLeb128()
|
/art/runtime/gc/accounting/ |
D | card_table-inl.h | 33 static inline bool byte_cas(uint8_t old_value, uint8_t new_value, uint8_t* address) { in byte_cas() argument 36 return byte_atomic->CompareAndSetWeakRelaxed(old_value, new_value); in byte_cas() 48 const uintptr_t old_word = cur_word | (static_cast<uintptr_t>(old_value) << shift_in_bits); in byte_cas()
|
/art/runtime/gc/space/ |
D | image_space.cc | 558 T* old_value = root->template Read<kWithoutReadBarrier>(); in PatchGcRoot() local 559 DCHECK(kMayBeNull || old_value != nullptr); in PatchGcRoot() 560 if (!kMayBeNull || old_value != nullptr) { in PatchGcRoot() 561 *root = GcRoot<T>(heap_visitor_(old_value)); in PatchGcRoot() 569 T* old_value = reinterpret_cast64<T*>(*raw_entry); in PatchNativePointer() local 570 DCHECK(kMayBeNull || old_value != nullptr); in PatchNativePointer() 571 if (!kMayBeNull || old_value != nullptr) { in PatchNativePointer() 572 T* new_value = native_visitor_(old_value); in PatchNativePointer() 577 T* old_value = reinterpret_cast32<T*>(*raw_entry); in PatchNativePointer() local 578 DCHECK(kMayBeNull || old_value != nullptr); in PatchNativePointer() [all …]
|
/art/compiler/optimizing/ |
D | intrinsics_arm64.cc | 1024 Register old_value = WRegisterFrom(locations->GetTemp(0)); // The old value from main path. in EmitNativeCode() local 1029 arm64_codegen->GenerateUnsafeCasOldValueMovWithBakerReadBarrier(marked, old_value); in EmitNativeCode() 1056 __ Ccmp(tmp, old_value, ZFlag, ne); in EmitNativeCode() 1088 Register old_value; // Value in memory. in GenCas() local 1106 old_value = WRegisterFrom(locations->GetTemp(0)); in GenCas() 1108 old_value = temps.AcquireSameSizeAs(value); in GenCas() 1120 __ Ldaxr(old_value, MemOperand(tmp_ptr)); in GenCas() 1122 assembler->MaybeUnpoisonHeapReference(old_value); in GenCas() 1124 __ Cmp(old_value, expected); in GenCas() 1129 __ Stlxr(old_value.W(), value, MemOperand(tmp_ptr)); // Reuse `old_value` for STLXR result. in GenCas() [all …]
|
D | code_generator_arm64.h | 676 vixl::aarch64::Register old_value);
|
D | code_generator_arm_vixl.h | 627 void GenerateUnsafeCasOldValueAddWithBakerReadBarrier(vixl::aarch32::Register old_value,
|
D | code_generator_arm64.cc | 5849 vixl::aarch64::Register old_value) { in GenerateUnsafeCasOldValueMovWithBakerReadBarrier() argument 5861 __ mov(marked, old_value); in GenerateUnsafeCasOldValueMovWithBakerReadBarrier()
|
D | code_generator_arm_vixl.cc | 8395 vixl::aarch32::Register old_value, in GenerateUnsafeCasOldValueAddWithBakerReadBarrier() argument 8402 uint32_t custom_data = EncodeBakerReadBarrierUnsafeCasData(old_value.GetCode()); in GenerateUnsafeCasOldValueAddWithBakerReadBarrier() 8413 __ add(EncodingSize(Wide), old_value, adjusted_old_value, Operand(expected)); // Preserves flags. in GenerateUnsafeCasOldValueAddWithBakerReadBarrier()
|