/external/v8/src/heap/ |
D | heap-inl.h | 123 type* Heap::name() { return type::cast(roots_[k##camel_name##RootIndex]); } 128 Map* Heap::name##_map() { return Map::cast(roots_[k##Name##MapRootIndex]); } 133 String* Heap::name() { return String::cast(roots_[k##name##RootIndex]); } in STRUCT_LIST() 138 Symbol* Heap::name() { return Symbol::cast(roots_[k##name##RootIndex]); } 143 Symbol* Heap::name() { return Symbol::cast(roots_[k##name##RootIndex]); } 149 void Heap::set_##name(type* value) { \ 160 PagedSpace* Heap::paged_space(int idx) { 166 Space* Heap::space(int idx) { return space_[idx]; } in space() 168 Address* Heap::NewSpaceAllocationTopAddress() { in NewSpaceAllocationTopAddress() 172 Address* Heap::NewSpaceAllocationLimitAddress() { in NewSpaceAllocationLimitAddress() [all …]
|
D | heap.cc | 55 struct Heap::StrongRootsList { 63 IdleScavengeObserver(Heap& heap, intptr_t step_size) in IdleScavengeObserver() 71 Heap& heap_; 74 Heap::Heap() in Heap() function in v8::internal::Heap 153 current_gc_flags_(Heap::kNoGCFlags), in Heap() 186 size_t Heap::Capacity() { in Capacity() 192 size_t Heap::OldGenerationCapacity() { in OldGenerationCapacity() 199 size_t Heap::CommittedOldGenerationMemory() { in CommittedOldGenerationMemory() 206 size_t Heap::CommittedMemory() { in CommittedMemory() 213 size_t Heap::CommittedPhysicalMemory() { in CommittedPhysicalMemory() [all …]
|
D | scavenger.h | 19 explicit Scavenger(Heap* heap) : heap_(heap) {} in Scavenger() 29 static inline SlotCallbackResult CheckAndScavengeObject(Heap* heap, 40 Heap* heap() { return heap_; } in heap() 43 Heap* heap_; 52 explicit ScavengeVisitor(Heap* heap) : heap_(heap) {} in ScavengeVisitor() 60 Heap* heap_; 69 static inline void VisitPointer(Heap* heap, HeapObject* object, Object** p);
|
D | objects-visiting.cc | 222 static bool MustRecordSlots(Heap* heap) { in MustRecordSlots() 223 return heap->gc_state() == Heap::MARK_COMPACT && in MustRecordSlots() 233 Object* VisitWeakList(Heap* heap, Object* list, WeakObjectRetainer* retainer) { in VisitWeakList() 282 static void ClearWeakList(Heap* heap, Object* list) { in ClearWeakList() 304 static void VisitLiveObject(Heap*, JSFunction*, WeakObjectRetainer*) {} in VisitLiveObject() 306 static void VisitPhantomObject(Heap*, JSFunction*) {} in VisitPhantomObject() 320 static void VisitLiveObject(Heap*, Code*, WeakObjectRetainer*) {} in VisitLiveObject() 322 static void VisitPhantomObject(Heap*, Code*) {} in VisitPhantomObject() 340 static void VisitLiveObject(Heap* heap, Context* context, in VisitLiveObject() 346 if (heap->gc_state() == Heap::MARK_COMPACT) { in VisitLiveObject() [all …]
|
D | incremental-marking-job.cc | 17 void IncrementalMarkingJob::Start(Heap* heap) { in Start() 24 void IncrementalMarkingJob::ScheduleTask(Heap* heap) { in ScheduleTask() 33 void IncrementalMarkingJob::Task::Step(Heap* heap) { in Step() 45 Heap* heap = isolate()->heap(); in RunInternal() 50 Heap::IncrementalMarkingLimit::kNoLimit) { in RunInternal() 51 heap->StartIncrementalMarking(Heap::kNoGCFlags, in RunInternal()
|
D | array-buffer-tracker.h | 17 class Heap; variable 33 inline static void RegisterNew(Heap* heap, JSArrayBuffer* buffer); 34 inline static void Unregister(Heap* heap, JSArrayBuffer* buffer); 38 static void FreeDeadInNewSpace(Heap* heap); 68 explicit LocalArrayBufferTracker(Heap* heap) : heap_(heap) {} in LocalArrayBufferTracker() 95 Heap* heap_;
|
D | objects-visiting.h | 136 DCHECK(Heap::kMinObjectSizeInWords * kPointerSize <= object_size); in GetVisitorIdForSize() 144 base + (object_size >> kPointerSizeLog2) - Heap::kMinObjectSizeInWords, in GetVisitorIdForSize() 187 STATIC_ASSERT((generic - base + Heap::kMinObjectSizeInWords) == 10); in RegisterSpecializations() 262 INLINE(static void VisitPointers(Heap* heap, HeapObject* object, in INLINE() 272 inline static void VisitCodeEntry(Heap* heap, HeapObject* object, in VisitCodeEntry() 364 INLINE(static void VisitCodeEntry(Heap* heap, HeapObject* object, 366 INLINE(static void VisitEmbeddedPointer(Heap* heap, RelocInfo* rinfo)); 367 INLINE(static void VisitCell(Heap* heap, RelocInfo* rinfo)); 368 INLINE(static void VisitDebugTarget(Heap* heap, RelocInfo* rinfo)); 369 INLINE(static void VisitCodeTarget(Heap* heap, RelocInfo* rinfo)); [all …]
|
D | incremental-marking-job.h | 13 class Heap; variable 25 static void Step(Heap* heap); 37 void Start(Heap* heap); 41 void ScheduleTask(Heap* heap);
|
D | objects-visiting-inl.h | 217 Heap* heap, HeapObject* object, Address entry_address) { in VisitCodeEntry() 227 Heap* heap, RelocInfo* rinfo) { in VisitEmbeddedPointer() 242 void StaticMarkingVisitor<StaticVisitor>::VisitCell(Heap* heap, in VisitCell() 255 void StaticMarkingVisitor<StaticVisitor>::VisitDebugTarget(Heap* heap, in VisitDebugTarget() 267 void StaticMarkingVisitor<StaticVisitor>::VisitCodeTarget(Heap* heap, in VisitCodeTarget() 278 Heap* heap, RelocInfo* rinfo) { in VisitCodeAgeSequence() 306 Heap* heap = map->GetHeap(); in VisitMap() 329 Heap* heap = map->GetHeap(); in VisitWeakCell() 358 Heap* heap = array->GetHeap(); in VisitTransitionArray() 371 Heap* heap = map->GetHeap(); in VisitTransitionArray() [all …]
|
D | scavenge-job.h | 15 class Heap; variable 40 void ScheduleIdleTaskIfNeeded(Heap* heap, int bytes_allocated); 45 void RescheduleIdleTask(Heap* heap); 73 void ScheduleIdleTask(Heap* heap);
|
D | scavenge-job.cc | 20 Heap* heap = isolate()->heap(); in RunInternal() 80 void ScavengeJob::RescheduleIdleTask(Heap* heap) { in RescheduleIdleTask() 90 void ScavengeJob::ScheduleIdleTaskIfNeeded(Heap* heap, int bytes_allocated) { in ScheduleIdleTaskIfNeeded() 101 void ScavengeJob::ScheduleIdleTask(Heap* heap) { in ScheduleIdleTask()
|
D | memory-reducer.h | 16 class Heap; variable 113 explicit MemoryReducer(Heap* heap) in MemoryReducer() 139 Heap* heap() { return heap_; } in heap() 161 Heap* heap_;
|
D | object-stats.h | 20 explicit ObjectStats(Heap* heap) : heap_(heap) { ClearObjectStats(); } in ObjectStats() 94 Heap* heap() { return heap_; } in heap() 118 Heap* heap_; 135 ObjectStatsCollector(Heap* heap, ObjectStats* stats) in ObjectStatsCollector() 162 Heap* heap_;
|
/external/valgrind/exp-sgcheck/ |
D | pc_common.c | 139 } Heap; member 184 xe.XE.Heap.addr = a; in h_record_heap_error() 185 xe.XE.Heap.sszB = is_write ? -size : size; in h_record_heap_error() 186 xe.XE.Heap.vseg = vseg; in h_record_heap_error() 352 Addr a = xe->XE.Heap.addr; in pc_pp_Error() 353 Seg* vseg = xe->XE.Heap.vseg; in pc_pp_Error() 363 readwrite(xe->XE.Heap.sszB), in pc_pp_Error() 364 Word__abs(xe->XE.Heap.sszB) ); in pc_pp_Error() 373 readwrite(xe->XE.Heap.sszB), in pc_pp_Error() 374 Word__abs(xe->XE.Heap.sszB) ); in pc_pp_Error() [all …]
|
/external/v8/src/snapshot/ |
D | natives-common.cc | 15 FixedArray* NativesCollection<CORE>::GetSourceCache(Heap* heap) { in GetSourceCache() 21 FixedArray* NativesCollection<EXPERIMENTAL>::GetSourceCache(Heap* heap) { in GetSourceCache() 27 FixedArray* NativesCollection<EXTRAS>::GetSourceCache(Heap* heap) { in GetSourceCache() 33 FixedArray* NativesCollection<EXPERIMENTAL_EXTRAS>::GetSourceCache(Heap* heap) { in GetSourceCache()
|
/external/v8/src/ |
D | address-map.cc | 17 for (uint32_t i = 0; i < Heap::kStrongRootListLength; i++) { in RootIndexMap() 18 Heap::RootListIndex root_index = static_cast<Heap::RootListIndex>(i); in RootIndexMap() 36 CHECK(!Heap::RootIsImmortalImmovable(root_index)); in RootIndexMap()
|
D | objects-body-descriptors.h | 43 static inline void IteratePointers(Heap* heap, HeapObject* obj, 51 static inline void IteratePointer(Heap* heap, HeapObject* obj, int offset); 64 static inline void IterateBodyImpl(Heap* heap, HeapObject* obj, 96 Heap* heap = obj->GetHeap(); in IterateBody() 128 Heap* heap = obj->GetHeap(); in IterateBody()
|
/external/v8/src/profiler/ |
D | sampling-heap-profiler.h | 44 SamplingHeapProfiler(Heap* heap, StringsStorage* names, uint64_t rate, 123 Heap* heap() const { return heap_; } in heap() 144 Heap* const heap_; 161 SamplingAllocationObserver(Heap* heap, intptr_t step_size, uint64_t rate, in SamplingAllocationObserver() 174 DCHECK(heap_->gc_state() == Heap::NOT_IN_GC); in Step() 184 Heap* const heap_;
|
/external/v8/src/full-codegen/x64/ |
D | full-codegen-x64.cc | 139 __ PushRoot(Heap::kUndefinedValueRootIndex); in Generate() 145 __ CompareRoot(rcx, Heap::kRealStackLimitRootIndex); in Generate() 150 __ LoadRoot(rax, Heap::kUndefinedValueRootIndex); in Generate() 304 __ CompareRoot(rsp, Heap::kStackLimitRootIndex); in Generate() 320 __ LoadRoot(rax, Heap::kUndefinedValueRootIndex); in Generate() 439 void FullCodeGenerator::EffectContext::Plug(Heap::RootListIndex index) const { in Plug() 444 Heap::RootListIndex index) const { in Plug() 450 Heap::RootListIndex index) const { in Plug() 456 void FullCodeGenerator::TestContext::Plug(Heap::RootListIndex index) const { in Plug() 461 if (index == Heap::kUndefinedValueRootIndex || in Plug() [all …]
|
/external/v8/src/full-codegen/arm/ |
D | full-codegen-arm.cc | 153 __ LoadRoot(r2, Heap::kRealStackLimitRootIndex); in Generate() 159 __ LoadRoot(r9, Heap::kUndefinedValueRootIndex); in Generate() 311 __ LoadRoot(ip, Heap::kStackLimitRootIndex); in Generate() 332 __ LoadRoot(r0, Heap::kUndefinedValueRootIndex); in Generate() 477 void FullCodeGenerator::EffectContext::Plug(Heap::RootListIndex index) const { in Plug() 482 Heap::RootListIndex index) const { in Plug() 488 Heap::RootListIndex index) const { in Plug() 494 void FullCodeGenerator::TestContext::Plug(Heap::RootListIndex index) const { in Plug() 499 if (index == Heap::kUndefinedValueRootIndex || in Plug() 500 index == Heap::kNullValueRootIndex || in Plug() [all …]
|
/external/v8/src/full-codegen/arm64/ |
D | full-codegen-arm64.cc | 159 __ CompareRoot(x10, Heap::kRealStackLimitRootIndex); in Generate() 164 __ LoadRoot(x10, Heap::kUndefinedValueRootIndex); in Generate() 314 __ CompareRoot(jssp, Heap::kStackLimitRootIndex); in Generate() 332 __ LoadRoot(x0, Heap::kUndefinedValueRootIndex); in Generate() 468 void FullCodeGenerator::EffectContext::Plug(Heap::RootListIndex index) const { in Plug() 474 Heap::RootListIndex index) const { in Plug() 480 Heap::RootListIndex index) const { in Plug() 486 void FullCodeGenerator::TestContext::Plug(Heap::RootListIndex index) const { in Plug() 489 if (index == Heap::kUndefinedValueRootIndex || in Plug() 490 index == Heap::kNullValueRootIndex || in Plug() [all …]
|
/external/v8/src/full-codegen/mips/ |
D | full-codegen-mips.cc | 162 __ LoadRoot(a2, Heap::kRealStackLimitRootIndex); in Generate() 167 __ LoadRoot(t5, Heap::kUndefinedValueRootIndex); in Generate() 322 __ LoadRoot(at, Heap::kStackLimitRootIndex); in Generate() 341 __ LoadRoot(v0, Heap::kUndefinedValueRootIndex); in Generate() 470 void FullCodeGenerator::EffectContext::Plug(Heap::RootListIndex index) const { in Plug() 475 Heap::RootListIndex index) const { in Plug() 481 Heap::RootListIndex index) const { in Plug() 487 void FullCodeGenerator::TestContext::Plug(Heap::RootListIndex index) const { in Plug() 492 if (index == Heap::kUndefinedValueRootIndex || in Plug() 493 index == Heap::kNullValueRootIndex || in Plug() [all …]
|
/external/v8/src/full-codegen/mips64/ |
D | full-codegen-mips64.cc | 161 __ LoadRoot(a2, Heap::kRealStackLimitRootIndex); in Generate() 166 __ LoadRoot(t1, Heap::kUndefinedValueRootIndex); in Generate() 320 __ LoadRoot(at, Heap::kStackLimitRootIndex); in Generate() 341 __ LoadRoot(v0, Heap::kUndefinedValueRootIndex); in Generate() 470 void FullCodeGenerator::EffectContext::Plug(Heap::RootListIndex index) const { in Plug() 475 Heap::RootListIndex index) const { in Plug() 481 Heap::RootListIndex index) const { in Plug() 487 void FullCodeGenerator::TestContext::Plug(Heap::RootListIndex index) const { in Plug() 492 if (index == Heap::kUndefinedValueRootIndex || in Plug() 493 index == Heap::kNullValueRootIndex || in Plug() [all …]
|
/external/v8/src/full-codegen/s390/ |
D | full-codegen-s390.cc | 159 __ LoadRoot(r5, Heap::kRealStackLimitRootIndex); in Generate() 165 __ LoadRoot(ip, Heap::kUndefinedValueRootIndex); in Generate() 322 __ LoadRoot(ip, Heap::kStackLimitRootIndex); in Generate() 340 __ LoadRoot(r2, Heap::kUndefinedValueRootIndex); in Generate() 462 void FullCodeGenerator::EffectContext::Plug(Heap::RootListIndex index) const {} in Plug() 465 Heap::RootListIndex index) const { in Plug() 470 Heap::RootListIndex index) const { in Plug() 475 void FullCodeGenerator::TestContext::Plug(Heap::RootListIndex index) const { in Plug() 478 if (index == Heap::kUndefinedValueRootIndex || in Plug() 479 index == Heap::kNullValueRootIndex || in Plug() [all …]
|
/external/v8/src/full-codegen/ppc/ |
D | full-codegen-ppc.cc | 160 __ LoadRoot(r5, Heap::kRealStackLimitRootIndex); in Generate() 166 __ LoadRoot(ip, Heap::kUndefinedValueRootIndex); in Generate() 318 __ LoadRoot(ip, Heap::kStackLimitRootIndex); in Generate() 336 __ LoadRoot(r3, Heap::kUndefinedValueRootIndex); in Generate() 464 void FullCodeGenerator::EffectContext::Plug(Heap::RootListIndex index) const {} in Plug() 468 Heap::RootListIndex index) const { in Plug() 474 Heap::RootListIndex index) const { in Plug() 480 void FullCodeGenerator::TestContext::Plug(Heap::RootListIndex index) const { in Plug() 483 if (index == Heap::kUndefinedValueRootIndex || in Plug() 484 index == Heap::kNullValueRootIndex || in Plug() [all …]
|