/external/v8/src/heap/ |
D | incremental-marking.cc | 54 : heap_(heap), in IncrementalMarking() 109 : heap_(incremental_marking->heap()) {} in IncrementalMarkingRootMarkingVisitor() 126 heap_->incremental_marking()->WhiteToGreyAndPush(HeapObject::cast(obj)); in MarkObjectByPointer() 129 Heap* heap_; member in v8::internal::IncrementalMarkingRootMarkingVisitor 140 return FLAG_incremental_marking && heap_->gc_state() == Heap::NOT_IN_GC && in CanBeActivated() 141 heap_->deserialization_complete() && in CanBeActivated() 142 !heap_->isolate()->serializer_enabled(); in CanBeActivated() 146 return heap_->OldGenerationSizeOfObjects() <= kV8ActivationThreshold && in IsBelowActivationThresholds() 147 heap_->GlobalSizeOfObjects() <= kGlobalActivationThreshold; in IsBelowActivationThresholds() 172 DCHECK(heap_->gc_state() == Heap::NOT_IN_GC); in Start() [all …]
|
D | stress-scavenge-observer.cc | 18 heap_(heap), in StressScavengeObserver() 24 heap_->isolate()->PrintWithTimestamp( in StressScavengeObserver() 31 if (has_requested_gc_ || heap_->new_space()->Capacity() == 0) { in Step() 36 heap_->new_space()->Size() * 100.0 / heap_->new_space()->Capacity(); in Step() 39 heap_->isolate()->PrintWithTimestamp( in Step() 52 heap_->isolate()->PrintWithTimestamp("[Scavenge] GC requested\n"); in Step() 56 heap_->isolate()->stack_guard()->RequestGC(); in Step() 66 heap_->new_space()->Size() * 100.0 / heap_->new_space()->Capacity(); in RequestedGCDone() 70 heap_->isolate()->PrintWithTimestamp( in RequestedGCDone() 73 heap_->isolate()->PrintWithTimestamp("[Scavenge] %d%% is the new limit\n", in RequestedGCDone() [all …]
|
D | marking-barrier.cc | 23 : heap_(heap), in MarkingBarrier() 24 collector_(heap_->mark_compact_collector()), in MarkingBarrier() 25 incremental_marking_(heap_->incremental_marking()), in MarkingBarrier() 30 : heap_(local_heap->heap()), in MarkingBarrier() 31 collector_(heap_->mark_compact_collector()), in MarkingBarrier() 158 DeactivateSpace(heap_->old_space()); in Deactivate() 159 DeactivateSpace(heap_->map_space()); in Deactivate() 160 DeactivateSpace(heap_->code_space()); in Deactivate() 161 DeactivateSpace(heap_->new_space()); in Deactivate() 162 for (LargePage* p : *heap_->new_lo_space()) { in Deactivate() [all …]
|
D | gc-tracer.cc | 30 return heap_->isolate()->counters()->worker_thread_runtime_call_stats(); in worker_thread_runtime_call_stats() 53 start_time_ = tracer_->heap_->MonotonicallyIncreasingTimeInMs(); in Scope() 55 runtime_stats_ = tracer_->heap_->isolate()->counters()->runtime_call_stats(); in Scope() 61 scope_, tracer_->heap_->MonotonicallyIncreasingTimeInMs() - start_time_); in ~Scope() 69 start_time_ = tracer_->heap_->MonotonicallyIncreasingTimeInMs(); in BackgroundScope() 77 tracer_->heap_->MonotonicallyIncreasingTimeInMs() - start_time_; in ~BackgroundScope() 151 : heap_(heap), in GCTracer() 178 current_.end_time = heap_->MonotonicallyIncreasingTimeInMs(); in GCTracer() 186 current_.end_time = heap_->MonotonicallyIncreasingTimeInMs(); in ResetForTesting() 220 heap_->isolate()->counters()->young_generation_handling()->AddSample( in NotifyYoungGenerationHandling() [all …]
|
D | scavenger.cc | 181 TRACE_GC(outer_->heap_->tracer(), in Run() 186 outer_->heap_->tracer(), in Run() 212 PrintIsolate(outer_->heap_->isolate(), in ProcessItems() 237 : isolate_(heap->isolate()), heap_(heap) {} in ScavengerCollector() 257 TRACE_GC(heap_->tracer(), in CollectGarbage() 259 heap_->array_buffer_sweeper()->EnsureFinished(); in CollectGarbage() 271 Sweeper* sweeper = heap_->mark_compact_collector()->sweeper(); in CollectGarbage() 280 heap_->mark_compact_collector()->EnsureSweepingCompleted(); in CollectGarbage() 298 new Scavenger(this, heap_, is_logging, &empty_chunks, &copied_list, in CollectGarbage() 304 heap_, [&memory_chunks](MemoryChunk* chunk) { in CollectGarbage() [all …]
|
D | finalization-registry-cleanup-task.cc | 19 : CancelableTask(heap->isolate()), heap_(heap) {} in FinalizationRegistryCleanupTask() 32 Isolate* isolate = heap_->isolate(); in SlowAssertNoActiveJavaScript() 39 Isolate* isolate = heap_->isolate(); in RunInternal() 50 if (!heap_->DequeueDirtyJSFinalizationRegistry().ToHandle( in RunInternal() 80 heap_->EnqueueDirtyJSFinalizationRegistry(*finalization_registry, nop); in RunInternal() 84 heap_->set_is_finalization_registry_cleanup_task_posted(false); in RunInternal() 85 heap_->PostFinalizationRegistryCleanupTaskIfNeeded(); in RunInternal()
|
D | collection-barrier.cc | 31 : CancelableTask(heap->isolate()), heap_(heap) {} in BackgroundCollectionInterruptTask() 37 void RunInternal() override { heap_->CheckCollectionRequested(); } in RunInternal() 39 Heap* heap_; member in v8::internal::BackgroundCollectionInterruptTask 71 heap_->isolate()->counters()->time_to_collection()->AddTimedSample(delta); in StopTimeToCollectionTimer() 80 Isolate* isolate = heap_->isolate(); in ActivateStackGuardAndPostTask() 86 std::make_unique<BackgroundCollectionInterruptTask>(heap_)); in ActivateStackGuardAndPostTask() 90 TRACE_BACKGROUND_GC(heap_->tracer(), in BlockUntilCollected()
|
D | local-heap.cc | 29 : heap_(heap), in LocalHeap() 40 heap_->safepoint()->AddLocalHeap(this, [this] { in LocalHeap() 43 if (heap_->incremental_marking()->IsMarking()) { in LocalHeap() 45 heap_->incremental_marking()->IsCompacting()); in LocalHeap() 61 heap_->safepoint()->RemoveLocalHeap(this, [this] { in ~LocalHeap() 77 heap_->isolate()->NewPersistentHandles().release()); in EnsurePersistentHandles() 144 if (state_ == ThreadState::Running) heap_->safepoint()->EnterFromThread(this); in EnterSafepoint() 165 heap_->RequestCollectionBackground(this); in PerformCollection() 184 heap_->FatalProcessOutOfMemory("LocalHeap: allocation failed"); in PerformCollectionAndAllocateAgain()
|
D | heap-inl.h | 651 AlwaysAllocateScope::AlwaysAllocateScope(Heap* heap) : heap_(heap) { in AlwaysAllocateScope() 652 heap_->always_allocate_scope_count_++; in AlwaysAllocateScope() 656 heap_->always_allocate_scope_count_--; in ~AlwaysAllocateScope() 663 : heap_(heap) { in CodeSpaceMemoryModificationScope() 664 if (heap_->write_protect_code_memory()) { in CodeSpaceMemoryModificationScope() 665 heap_->increment_code_space_memory_modification_scope_depth(); in CodeSpaceMemoryModificationScope() 666 heap_->code_space()->SetReadAndWritable(); in CodeSpaceMemoryModificationScope() 667 LargePage* page = heap_->code_lo_space()->first_page(); in CodeSpaceMemoryModificationScope() 670 CHECK(heap_->memory_allocator()->IsMemoryChunkExecutable(page)); in CodeSpaceMemoryModificationScope() 678 if (heap_->write_protect_code_memory()) { in ~CodeSpaceMemoryModificationScope() [all …]
|
D | mark-compact.cc | 79 explicit MarkingVerifier(Heap* heap) : heap_(heap) {} in MarkingVerifier() 113 Heap* heap_; member in v8::internal::__anon5b4405900111::MarkingVerifier 117 heap_->IterateRoots(this, base::EnumSet<SkipRoot>{SkipRoot::kWeak}); in VerifyRoots() 187 VerifyMarking(heap_->new_space()); in Run() 188 VerifyMarking(heap_->new_lo_space()); in Run() 189 VerifyMarking(heap_->old_space()); in Run() 190 VerifyMarking(heap_->code_space()); in Run() 191 VerifyMarking(heap_->map_space()); in Run() 192 VerifyMarking(heap_->lo_space()); in Run() 193 VerifyMarking(heap_->code_lo_space()); in Run() [all …]
|
D | array-buffer-sweeper.cc | 75 heap_->isolate()->cancelable_task_manager()->TryAbort(job_->id_); in EnsureFinished() 124 heap_->DecrementExternalBackingStoreBytes( in DecrementExternalMemoryCounters() 126 heap_->update_external_memory(-static_cast<int64_t>(freed_bytes)); in DecrementExternalMemoryCounters() 148 if (!heap_->IsTearingDown() && !heap_->ShouldReduceMemory() && in RequestSweep() 152 auto task = MakeCancelableTask(heap_->isolate(), [this] { in RequestSweep() 154 heap_->tracer(), in RequestSweep() 235 heap_->IncrementExternalBackingStoreBytes( in IncrementExternalMemoryCounters() 237 reinterpret_cast<v8::Isolate*>(heap_->isolate()) in IncrementExternalMemoryCounters()
|
D | base-space.h | 25 DCHECK_NOT_NULL(heap_); in heap() 26 return heap_; in heap() 63 : heap_(heap), id_(id), committed_(0), max_committed_(0) {} in BaseSpace() 68 Heap* heap_;
|
D | spaces-inl.h | 93 : heap_(heap), in OldGenerationMemoryChunkIterator() 104 if (old_iterator_ != heap_->old_space()->end()) return *(old_iterator_++); in next() 109 if (map_iterator_ != heap_->map_space()->end()) return *(map_iterator_++); in next() 114 if (code_iterator_ != heap_->code_space()->end()) in next() 120 if (lo_iterator_ != heap_->lo_space()->end()) return *(lo_iterator_++); in next() 125 if (code_lo_iterator_ != heap_->code_lo_space()->end()) in next() 148 return Heap::PrecedeWithFiller(ReadOnlyRoots(heap_), in AllocateRawAligned()
|
D | local-allocator.h | 25 : heap_(heap), in EvacuationAllocator() 34 heap_->old_space()->MergeLocalSpace(compaction_spaces_.Get(OLD_SPACE)); in Finalize() 35 heap_->code_space()->MergeLocalSpace(compaction_spaces_.Get(CODE_SPACE)); in Finalize() 58 Heap* const heap_; variable
|
D | stress-marking-observer.cc | 13 : AllocationObserver(64), heap_(heap) {} in StressMarkingObserver() 17 heap_->StartIncrementalMarkingIfAllocationLimitIsReached(Heap::kNoGCFlags, in Step() 19 heap_->incremental_marking()->EnsureBlackAllocated(soon_object, size); in Step()
|
D | object-stats.h | 97 explicit ObjectStats(Heap* heap) : heap_(heap) { ClearObjectStats(true); } in ObjectStats() 135 Heap* heap() { return heap_; } in heap() 155 Heap* heap_; variable 181 : heap_(heap), live_(live), dead_(dead) { in ObjectStatsCollector() 182 DCHECK_NOT_NULL(heap_); in ObjectStatsCollector() 192 Heap* const heap_;
|
D | concurrent-marking.cc | 386 : heap_(heap), in ConcurrentMarking() 401 TRACE_BACKGROUND_GC(heap_->tracer(), in Run() 409 task_id, &local_marking_worklists, weak_objects_, heap_, in Run() 411 heap_->local_embedder_heap_tracer()->InUse(), is_forced_gc, in Run() 418 Isolate* isolate = heap_->isolate(); in Run() 451 Address new_space_top = heap_->new_space()->original_top_acquire(); in Run() 452 Address new_space_limit = heap_->new_space()->original_limit_relaxed(); in Run() 453 Address new_large_object = heap_->new_lo_space()->pending_object(); in Run() 514 heap_->isolate()->PrintWithTimestamp( in Run() 534 DCHECK(!heap_->IsTearingDown()); in ScheduleJob() [all …]
|
D | basic-memory-chunk.cc | 45 chunk->heap_ = heap; in Initialize() 67 reinterpret_cast<base::AtomicWord*>(&heap_))) != nullptr || in SynchronizedHeapLoad() 79 offsetof(BasicMemoryChunk, heap_)); 84 STATIC_ASSERT(offsetof(BasicMemoryChunk, heap_) ==
|
D | object-stats.cc | 407 Isolate* isolate() { return heap_->isolate(); } in isolate() 461 Heap* heap_; member in v8::internal::ObjectStatsCollectorImpl 471 : heap_(heap), in ObjectStatsCollectorImpl() 489 if (obj == ReadOnlyRoots(heap_).empty_property_array()) return false; in ShouldRecordObject() 606 if (properties != ReadOnlyRoots(heap_).empty_property_array()) { in RecordVirtualJSObjectDetails() 631 if (elements != ReadOnlyRoots(heap_).empty_fixed_array()) { in RecordVirtualJSObjectDetails() 721 GetFeedbackSlotType(vector.Get(slot), it.kind(), heap_->isolate()), in RecordVirtualFeedbackVectorDetails() 806 Object list = heap_->allocation_sites_list(); in CollectGlobalStatistics() 814 RecordSimpleVirtualObjectStats(HeapObject(), heap_->serialized_objects(), in CollectGlobalStatistics() 816 RecordSimpleVirtualObjectStats(HeapObject(), heap_->number_string_cache(), in CollectGlobalStatistics() [all …]
|
D | sweeper.cc | 20 : heap_(heap), in Sweeper() 42 sweeper_->heap_->mark_compact_collector()->EnsureSweepingCompleted(); in PauseOrCompleteScope() 154 should_reduce_memory_ = heap_->ShouldReduceMemory(); in StartSweeping() 156 heap_->mark_compact_collector()->non_atomic_marking_state(); in StartSweeping() 180 !heap_->delay_sweeper_tasks_for_testing_) { in StartSweeperTasks() 185 heap_->isolate(), this, &pending_sweeper_tasks_semaphore_, in StartSweeperTasks() 219 if (heap_->isolate()->cancelable_task_manager()->TryAbort(task_ids_[i]) != in AbortAndWaitForTasks() 527 v8::Isolate* isolate = reinterpret_cast<v8::Isolate*>(heap_->isolate()); in ScheduleIncrementalSweepingTask() 531 std::make_unique<IncrementalSweeperTask>(heap_->isolate(), this)); in ScheduleIncrementalSweepingTask() 564 heap_->paged_space(space)->IncreaseAllocatedBytes( in PrepareToBeSweptPage() [all …]
|
/external/libchrome/mojo/core/ports/ |
D | message_queue.cc | 33 for (const auto& message : heap_) in ~MessageQueue() 41 return !heap_.empty() && heap_[0]->sequence_num() == next_sequence_num_; in HasNextMessage() 46 if (!HasNextMessage() || (filter && !filter->Match(*heap_[0]))) { in GetNextMessage() 51 std::pop_heap(heap_.begin(), heap_.end()); in GetNextMessage() 52 *message = std::move(heap_.back()); in GetNextMessage() 54 heap_.pop_back(); in GetNextMessage() 64 heap_.emplace_back(std::move(message)); in AcceptMessage() 65 std::push_heap(heap_.begin(), heap_.end()); in AcceptMessage() 70 *has_next_message = (heap_[0]->sequence_num() == next_sequence_num_); in AcceptMessage() 76 *messages = std::move(heap_); in TakeAllMessages()
|
/external/v8/src/utils/ |
D | identity-map.cc | 28 heap_->UnregisterStrongRoots(strong_roots_entry_); in Clear() 52 Address not_mapped = ReadOnlyRoots(heap_).not_mapped_symbol().ptr(); in ScanKeysFor() 66 DCHECK_EQ(gc_counter_, heap_->gc_count()); in InsertKey() 73 Address not_mapped = ReadOnlyRoots(heap_).not_mapped_symbol().ptr(); in InsertKey() 95 Address not_mapped = ReadOnlyRoots(heap_).not_mapped_symbol().ptr(); in DeleteIndex() 136 if (index < 0 && gc_counter_ != heap_->gc_count()) { in Lookup() 151 if (gc_counter_ != heap_->gc_count()) Rehash(); in LookupOrInsert() 161 CHECK_NE(address, ReadOnlyRoots(heap_).not_mapped_symbol().ptr()); in Hash() 200 gc_counter_ = heap_->gc_count(); in InsertEntry() 203 Address not_mapped = ReadOnlyRoots(heap_).not_mapped_symbol().ptr(); in InsertEntry() [all …]
|
/external/v8/src/heap/cppgc-js/ |
D | unified-heap-marking-state.h | 25 explicit UnifiedHeapMarkingState(Heap& heap) : heap_(heap) {} in UnifiedHeapMarkingState() 33 Heap& heap_; 37 heap_.RegisterExternallyReferencedObject( in MarkAndPush()
|
/external/v8/src/profiler/ |
D | sampling-heap-profiler.h | 135 heap_(heap), in Observer() 141 USE(heap_); in Step() 142 DCHECK(heap_->gc_state() == Heap::NOT_IN_GC); in Step() 155 Heap* const heap_; variable 186 Heap* const heap_; variable
|
/external/v8/src/execution/ |
D | local-isolate.h | 45 OFFSET_OF(LocalIsolate, heap_)); in FromHeap() 48 LocalHeap* heap() { return &heap_; } in heap() 87 LocalHeap heap_; variable
|