Lines Matching refs:to_ref
167 mirror::Object* to_ref = Mark(self, from_ref); in MarkHeapReference() local
168 if (from_ref != to_ref) { in MarkHeapReference()
174 } while (!field->CasWeakRelaxed(from_ref, to_ref)); in MarkHeapReference()
479 mirror::Object* to_ref = concurrent_copying_->Mark(self, ref); in VisitRoots() local
480 if (to_ref != ref) { in VisitRoots()
481 *root = to_ref; in VisitRoots()
496 mirror::Object* to_ref = concurrent_copying_->Mark(self, ref); in VisitRoots() local
497 if (to_ref != ref) { in VisitRoots()
498 root->Assign(to_ref); in VisitRoots()
1747 void ConcurrentCopying::PushOntoMarkStack(Thread* const self, mirror::Object* to_ref) { in PushOntoMarkStack() argument
1749 << " " << to_ref << " " << mirror::Object::PrettyTypeOf(to_ref); in PushOntoMarkStack()
1759 gc_mark_stack_->PushBack(to_ref); in PushOntoMarkStack()
1779 new_tl_mark_stack->PushBack(to_ref); in PushOntoMarkStack()
1786 tl_mark_stack->PushBack(to_ref); in PushOntoMarkStack()
1795 gc_mark_stack_->PushBack(to_ref); in PushOntoMarkStack()
1799 << "ref=" << to_ref in PushOntoMarkStack()
1809 gc_mark_stack_->PushBack(to_ref); in PushOntoMarkStack()
2046 mirror::Object* to_ref = gc_mark_stack_->PopBack(); in ProcessMarkStackOnce() local
2047 ProcessMarkStackRef(to_ref); in ProcessMarkStackOnce()
2089 mirror::Object* to_ref = gc_mark_stack_->PopBack(); in ProcessMarkStackOnce() local
2090 ProcessMarkStackRef(to_ref); in ProcessMarkStackOnce()
2116 mirror::Object* to_ref = p->AsMirrorPtr(); in ProcessThreadLocalMarkStacks() local
2117 processor(to_ref); in ProcessThreadLocalMarkStacks()
2135 inline void ConcurrentCopying::ProcessMarkStackRef(mirror::Object* to_ref) { in ProcessMarkStackRef() argument
2136 DCHECK(!region_space_->IsInFromSpace(to_ref)); in ProcessMarkStackRef()
2137 space::RegionSpace::RegionType rtype = region_space_->GetRegionType(to_ref); in ProcessMarkStackRef()
2139 DCHECK(to_ref->GetReadBarrierState() == ReadBarrier::GrayState()) in ProcessMarkStackRef()
2140 << " to_ref=" << to_ref in ProcessMarkStackRef()
2141 << " rb_state=" << to_ref->GetReadBarrierState() in ProcessMarkStackRef()
2142 << " is_marked=" << IsMarked(to_ref) in ProcessMarkStackRef()
2143 << " type=" << to_ref->PrettyTypeOf() in ProcessMarkStackRef()
2145 << " space=" << heap_->DumpSpaceNameFromAddress(to_ref) in ProcessMarkStackRef()
2153 DCHECK(!region_space_->IsInNewlyAllocatedRegion(to_ref)) << to_ref; in ProcessMarkStackRef()
2158 if (!kUseBakerReadBarrier || !region_space_bitmap_->Set(to_ref)) { in ProcessMarkStackRef()
2162 CHECK(region_space_->IsLargeObject(to_ref)); in ProcessMarkStackRef()
2163 region_space_->ZeroLiveBytesForLargeObject(to_ref); in ProcessMarkStackRef()
2180 region_space_bitmap_->Set(to_ref); in ProcessMarkStackRef()
2185 DCHECK(!region_space_->HasAddress(to_ref)) << to_ref; in ProcessMarkStackRef()
2186 DCHECK(!immune_spaces_.ContainsObject(to_ref)); in ProcessMarkStackRef()
2191 const bool is_los = !mark_bitmap->HasAddress(to_ref); in ProcessMarkStackRef()
2193 if (!IsAligned<kPageSize>(to_ref)) { in ProcessMarkStackRef()
2201 to_ref, in ProcessMarkStackRef()
2205 << "ref=" << to_ref in ProcessMarkStackRef()
2209 DCHECK(los_bitmap->HasAddress(to_ref)); in ProcessMarkStackRef()
2212 perform_scan = !los_bitmap->Set(to_ref); in ProcessMarkStackRef()
2216 perform_scan = !mark_bitmap->Set(to_ref); in ProcessMarkStackRef()
2224 Scan<true>(to_ref); in ProcessMarkStackRef()
2226 Scan<false>(to_ref); in ProcessMarkStackRef()
2230 DCHECK(to_ref->GetReadBarrierState() == ReadBarrier::GrayState()) in ProcessMarkStackRef()
2231 << " to_ref=" << to_ref in ProcessMarkStackRef()
2232 << " rb_state=" << to_ref->GetReadBarrierState() in ProcessMarkStackRef()
2233 << " is_marked=" << IsMarked(to_ref) in ProcessMarkStackRef()
2234 << " type=" << to_ref->PrettyTypeOf() in ProcessMarkStackRef()
2236 << " space=" << heap_->DumpSpaceNameFromAddress(to_ref) in ProcessMarkStackRef()
2243 if (UNLIKELY((to_ref->GetClass<kVerifyNone, kWithoutReadBarrier>()->IsTypeOfReferenceClass() && in ProcessMarkStackRef()
2244 (referent = to_ref->AsReference()->GetReferent<kWithoutReadBarrier>()) != nullptr && in ProcessMarkStackRef()
2248 DCHECK(to_ref->AsReference()->GetPendingNext() != nullptr) in ProcessMarkStackRef()
2249 << "Left unenqueued ref gray " << to_ref; in ProcessMarkStackRef()
2256 bool success = to_ref->AtomicSetReadBarrierState<std::memory_order_release>( in ProcessMarkStackRef()
2269 DCHECK(region_space_bitmap_->Test(to_ref)); in ProcessMarkStackRef()
2270 size_t obj_size = to_ref->SizeOf<kDefaultVerifyFlags>(); in ProcessMarkStackRef()
2272 region_space_->AddLiveBytes(to_ref, alloc_size); in ProcessMarkStackRef()
2275 CHECK(to_ref != nullptr); in ProcessMarkStackRef()
2277 CHECK(!region_space->IsInFromSpace(to_ref)) << "Scanning object " << to_ref << " in from space"; in ProcessMarkStackRef()
2278 AssertToSpaceInvariant(nullptr, MemberOffset(0), to_ref); in ProcessMarkStackRef()
2280 to_ref->VisitReferences</*kVisitNativeRoots=*/true, kDefaultVerifyFlags, kWithoutReadBarrier>( in ProcessMarkStackRef()
3072 inline void ConcurrentCopying::Scan(mirror::Object* to_ref) { in Scan() argument
3081 DCHECK(!region_space_->IsInFromSpace(to_ref)); in Scan()
3085 to_ref->VisitReferences</*kVisitNativeRoots=*/true, kDefaultVerifyFlags, kWithoutReadBarrier>( in Scan()
3099 mirror::Object* to_ref = Mark</*kGrayImmuneObject=*/false, kNoUnEvac, /*kFromGCThread=*/true>( in Process() local
3104 if (to_ref == ref) { in Process()
3109 mirror::Object* new_ref = to_ref; in Process()
3132 mirror::Object* to_ref = Mark(self, ref); in VisitRoots() local
3133 if (to_ref == ref) { in VisitRoots()
3138 mirror::Object* new_ref = to_ref; in VisitRoots()
3153 mirror::Object* to_ref = Mark<kGrayImmuneObject>(self, ref); in MarkRoot() local
3154 if (to_ref != ref) { in MarkRoot()
3157 auto new_ref = mirror::CompressedReference<mirror::Object>::FromMirrorPtr(to_ref); in MarkRoot()
3339 mirror::Object* to_ref = region_space_->AllocNonvirtual</*kForEvac=*/ true>( in Copy() local
3342 if (LIKELY(to_ref != nullptr)) { in Copy()
3346 to_ref = AllocateInSkippedBlock(self, region_space_alloc_size); in Copy()
3347 if (to_ref != nullptr) { in Copy()
3351 region_space_->RecordAlloc(to_ref); in Copy()
3366 to_ref = heap_->non_moving_space_->Alloc(self, obj_size, in Copy()
3368 if (UNLIKELY(to_ref == nullptr)) { in Copy()
3377 DCHECK(to_ref != nullptr); in Copy()
3380 to_ref->SetClass(klass); in Copy()
3388 memcpy(reinterpret_cast<uint8_t*>(to_ref) + kObjectHeaderSize, in Copy()
3402 FillWithDummyObject(self, to_ref, bytes_allocated); in Copy()
3404 DCHECK(region_space_->IsInToSpace(to_ref)); in Copy()
3407 region_space_->FreeLarge</*kForEvac=*/ true>(to_ref, bytes_allocated); in Copy()
3415 reinterpret_cast<uint8_t*>(to_ref))); in Copy()
3418 DCHECK(heap_->non_moving_space_->HasAddress(to_ref)); in Copy()
3421 heap_->non_moving_space_->Free(self, to_ref); in Copy()
3425 mirror::Object* lost_fwd_ptr = to_ref; in Copy()
3426 to_ref = reinterpret_cast<mirror::Object*>(old_lock_word.ForwardingAddress()); in Copy()
3427 CHECK(to_ref != nullptr); in Copy()
3428 CHECK_NE(to_ref, lost_fwd_ptr); in Copy()
3429 CHECK(region_space_->IsInToSpace(to_ref) || heap_->non_moving_space_->HasAddress(to_ref)) in Copy()
3430 << "to_ref=" << to_ref << " " << heap_->DumpSpaces(); in Copy()
3431 CHECK_NE(to_ref->GetLockWord(false).GetState(), LockWord::kForwardingAddress); in Copy()
3432 return to_ref; in Copy()
3436 to_ref->SetLockWord(old_lock_word, false); in Copy()
3439 to_ref->SetReadBarrierState(ReadBarrier::GrayState()); in Copy()
3446 LockWord new_lock_word = LockWord::FromForwardingAddress(reinterpret_cast<size_t>(to_ref)); in Copy()
3465 DCHECK(region_space_->IsInToSpace(to_ref)); in Copy()
3467 DCHECK(heap_->non_moving_space_->HasAddress(to_ref)); in Copy()
3471 CHECK(!heap_->non_moving_space_->GetLiveBitmap()->AtomicTestAndSet(to_ref)); in Copy()
3475 CHECK(!heap_->non_moving_space_->GetMarkBitmap()->AtomicTestAndSet(to_ref)); in Copy()
3479 DCHECK(to_ref->GetReadBarrierState() == ReadBarrier::GrayState()); in Copy()
3481 DCHECK(GetFwdPtr(from_ref) == to_ref); in Copy()
3482 CHECK_NE(to_ref->GetLockWord(false).GetState(), LockWord::kForwardingAddress); in Copy()
3483 PushOntoMarkStack(self, to_ref); in Copy()
3484 return to_ref; in Copy()
3499 mirror::Object* to_ref; in IsMarked() local
3501 to_ref = GetFwdPtr(from_ref); in IsMarked()
3502 DCHECK(to_ref == nullptr || region_space_->IsInToSpace(to_ref) || in IsMarked()
3503 heap_->non_moving_space_->HasAddress(to_ref)) in IsMarked()
3504 << "from_ref=" << from_ref << " to_ref=" << to_ref; in IsMarked()
3507 to_ref = from_ref; in IsMarked()
3509 to_ref = nullptr; in IsMarked()
3518 to_ref = from_ref; in IsMarked()
3523 to_ref = from_ref; in IsMarked()
3525 to_ref = nullptr; in IsMarked()
3529 return to_ref; in IsMarked()
3694 mirror::Object* to_ref = IsMarked(from_ref); in IsNullOrMarkedHeapReference() local
3695 if (to_ref == nullptr) { in IsNullOrMarkedHeapReference()
3698 if (from_ref != to_ref) { in IsNullOrMarkedHeapReference()
3705 } while (!field->CasWeakRelaxed(from_ref, to_ref)); in IsNullOrMarkedHeapReference()
3708 field->Assign</* kIsVolatile= */ true>(to_ref); in IsNullOrMarkedHeapReference()