Lines Matching refs:heap_
20 : heap_(heap), in IncrementalMarking()
43 heap_->mark_compact_collector()->RecordSlot(HeapObject::RawField(obj, 0), in RecordWriteSlow()
80 Code* host = heap_->isolate() in RecordCodeTargetPatch()
94 heap_->mark_compact_collector()->RecordCodeEntrySlot( in RecordWriteOfCodeEntrySlow()
118 heap_->mark_compact_collector()->RecordRelocSlot(rinfo, in RecordWriteIntoCodeSlow()
373 DeactivateIncrementalWriteBarrierForSpace(heap_->old_pointer_space()); in DeactivateIncrementalWriteBarrier()
374 DeactivateIncrementalWriteBarrierForSpace(heap_->old_data_space()); in DeactivateIncrementalWriteBarrier()
375 DeactivateIncrementalWriteBarrierForSpace(heap_->cell_space()); in DeactivateIncrementalWriteBarrier()
376 DeactivateIncrementalWriteBarrierForSpace(heap_->property_cell_space()); in DeactivateIncrementalWriteBarrier()
377 DeactivateIncrementalWriteBarrierForSpace(heap_->map_space()); in DeactivateIncrementalWriteBarrier()
378 DeactivateIncrementalWriteBarrierForSpace(heap_->code_space()); in DeactivateIncrementalWriteBarrier()
379 DeactivateIncrementalWriteBarrierForSpace(heap_->new_space()); in DeactivateIncrementalWriteBarrier()
381 LargePage* lop = heap_->lo_space()->first_page(); in DeactivateIncrementalWriteBarrier()
408 ActivateIncrementalWriteBarrier(heap_->old_pointer_space()); in ActivateIncrementalWriteBarrier()
409 ActivateIncrementalWriteBarrier(heap_->old_data_space()); in ActivateIncrementalWriteBarrier()
410 ActivateIncrementalWriteBarrier(heap_->cell_space()); in ActivateIncrementalWriteBarrier()
411 ActivateIncrementalWriteBarrier(heap_->property_cell_space()); in ActivateIncrementalWriteBarrier()
412 ActivateIncrementalWriteBarrier(heap_->map_space()); in ActivateIncrementalWriteBarrier()
413 ActivateIncrementalWriteBarrier(heap_->code_space()); in ActivateIncrementalWriteBarrier()
414 ActivateIncrementalWriteBarrier(heap_->new_space()); in ActivateIncrementalWriteBarrier()
416 LargePage* lop = heap_->lo_space()->first_page(); in ActivateIncrementalWriteBarrier()
425 return WorthActivating() && heap_->NextGCIsLikelyToBeFull(); in ShouldActivate()
441 heap_->gc_state() == Heap::NOT_IN_GC && in WorthActivating()
442 !heap_->isolate()->serializer_enabled() && in WorthActivating()
443 heap_->isolate()->IsInitialized() && in WorthActivating()
444 heap_->PromotedSpaceSizeOfObjects() > kActivationThreshold; in WorthActivating()
517 DCHECK(heap_->gc_state() == Heap::NOT_IN_GC); in Start()
518 DCHECK(!heap_->isolate()->serializer_enabled()); in Start()
519 DCHECK(heap_->isolate()->IsInitialized()); in Start()
523 if (!heap_->mark_compact_collector()->sweeping_in_progress()) { in Start()
532 heap_->new_space()->LowerInlineAllocationLimit(kAllocatedThreshold); in Start()
542 heap_->mark_compact_collector()->StartCompaction( in StartMarking()
551 PatchIncrementalMarkingRecordWriteStubs(heap_, mode); in StartMarking()
566 heap_->mark_compact_collector()->VerifyMarkbitsAreClean(); in StartMarking()
570 heap_->CompletelyClearInstanceofCache(); in StartMarking()
571 heap_->isolate()->compilation_cache()->MarkCompactPrologue(); in StartMarking()
576 MarkObjectGreyDoNotEnqueue(heap_->polymorphic_code_cache()); in StartMarking()
581 heap_->IterateStrongRoots(&visitor, VISIT_ONLY_STRONG); in StartMarking()
583 heap_->mark_compact_collector()->MarkWeakObjectToCodeTable(); in StartMarking()
594 NewSpacePageIterator it(heap_->new_space()->FromSpaceStart(), in PrepareForScavenge()
595 heap_->new_space()->FromSpaceEnd()); in PrepareForScavenge()
611 Map* filler_map = heap_->one_pointer_filler_map(); in UpdateMarkingDequeAfterScavenge()
617 if (heap_->InNewSpace(obj)) { in UpdateMarkingDequeAfterScavenge()
672 Map* filler_map = heap_->one_pointer_filler_map(); in ProcessMarkingDeque()
694 Map* filler_map = heap_->one_pointer_filler_map(); in ProcessMarkingDeque()
724 heap_->tracer()->AddMarkingTime(delta); in Hurry()
733 PolymorphicCodeCache* poly_cache = heap_->polymorphic_code_cache(); in Hurry()
739 Object* context = heap_->native_contexts_list(); in Hurry()
762 heap_->new_space()->LowerInlineAllocationLimit(0); in Abort()
766 PatchIncrementalMarkingRecordWriteStubs(heap_, in Abort()
771 LargeObjectIterator it(heap_->lo_space()); in Abort()
780 heap_->isolate()->stack_guard()->ClearGC(); in Abort()
790 heap_->new_space()->LowerInlineAllocationLimit(0); in Finalize()
793 PatchIncrementalMarkingRecordWriteStubs(heap_, in Finalize()
797 heap_->isolate()->stack_guard()->ClearGC(); in Finalize()
813 heap_->isolate()->stack_guard()->RequestGC(); in MarkingComplete()
854 (heap_->PromotedTotalSize() > in SpeedUp()
865 heap_->PromotedTotalSize() - in SpeedUp()
868 intptr_t scavenge_slack = heap_->MaxSemiSpaceSize(); in SpeedUp()
897 if (heap_->gc_state() != Heap::NOT_IN_GC || !FLAG_incremental_marking || in Step()
915 heap_->isolate()->counters()->gc_incremental_marking()); in Step()
936 if (heap_->mark_compact_collector()->sweeping_in_progress() && in Step()
937 heap_->mark_compact_collector()->IsSweepingCompleted()) { in Step()
938 heap_->mark_compact_collector()->EnsureSweepingCompleted(); in Step()
940 if (!heap_->mark_compact_collector()->sweeping_in_progress()) { in Step()
960 heap_->tracer()->AddIncrementalMarkingStep(duration, bytes_processed); in Step()
970 heap_->PromotedTotalSize(); in ResetStepCounters()
979 return heap_->MaxOldGenerationSize() - heap_->PromotedSpaceSizeOfObjects(); in SpaceLeftInOldSpace()