• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // Copyright 2012 the V8 project authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
4 
5 #ifndef V8_HEAP_HEAP_H_
6 #define V8_HEAP_HEAP_H_
7 
8 #include <atomic>
9 #include <cmath>
10 #include <memory>
11 #include <unordered_map>
12 #include <unordered_set>
13 #include <vector>
14 
15 // Clients of this interface shouldn't depend on lots of heap internals.
16 // Do not include anything from src/heap here!
17 #include "include/v8-callbacks.h"
18 #include "include/v8-embedder-heap.h"
19 #include "include/v8-internal.h"
20 #include "include/v8-isolate.h"
21 #include "src/base/atomic-utils.h"
22 #include "src/base/enum-set.h"
23 #include "src/base/platform/condition-variable.h"
24 #include "src/base/platform/mutex.h"
25 #include "src/builtins/accessors.h"
26 #include "src/common/assert-scope.h"
27 #include "src/common/globals.h"
28 #include "src/heap/allocation-observer.h"
29 #include "src/heap/allocation-result.h"
30 #include "src/heap/heap-allocator.h"
31 #include "src/init/heap-symbols.h"
32 #include "src/objects/allocation-site.h"
33 #include "src/objects/fixed-array.h"
34 #include "src/objects/hash-table.h"
35 #include "src/objects/heap-object.h"
36 #include "src/objects/js-array-buffer.h"
37 #include "src/objects/objects.h"
38 #include "src/objects/smi.h"
39 #include "src/objects/visitors.h"
40 #include "src/roots/roots.h"
41 #include "src/utils/allocation.h"
42 #include "testing/gtest/include/gtest/gtest_prod.h"  // nogncheck
43 
44 namespace v8 {
45 
46 namespace debug {
47 using OutOfMemoryCallback = void (*)(void* data);
48 }  // namespace debug
49 
50 namespace internal {
51 
52 namespace heap {
53 class HeapTester;
54 class TestMemoryAllocatorScope;
55 }  // namespace heap
56 
57 namespace third_party_heap {
58 class Heap;
59 class Impl;
60 }  // namespace third_party_heap
61 
62 class IncrementalMarking;
63 class BackingStore;
64 class JSArrayBuffer;
65 class JSPromise;
66 class NativeContext;
67 
68 using v8::MemoryPressureLevel;
69 
70 class ArrayBufferCollector;
71 class ArrayBufferSweeper;
72 class BasicMemoryChunk;
73 class CodeLargeObjectSpace;
74 class CodeRange;
75 class CollectionBarrier;
76 class ConcurrentAllocator;
77 class ConcurrentMarking;
78 class CppHeap;
79 class GCIdleTimeHandler;
80 class GCIdleTimeHeapState;
81 class GCTracer;
82 template <typename T>
83 class GlobalHandleVector;
84 class IsolateSafepoint;
85 class HeapObjectAllocationTracker;
86 class HeapObjectsFilter;
87 class HeapStats;
88 class Isolate;
89 class JSFinalizationRegistry;
90 class LinearAllocationArea;
91 class LocalEmbedderHeapTracer;
92 class LocalHeap;
93 class MarkingBarrier;
94 class MemoryAllocator;
95 class MemoryChunk;
96 class MemoryMeasurement;
97 class MemoryReducer;
98 class MinorMarkCompactCollector;
99 class ObjectIterator;
100 class ObjectStats;
101 class Page;
102 class PagedSpace;
103 class ReadOnlyHeap;
104 class RootVisitor;
105 class SafepointScope;
106 class ScavengeJob;
107 class Scavenger;
108 class ScavengerCollector;
109 class SharedReadOnlySpace;
110 class Space;
111 class StressScavengeObserver;
112 class TimedHistogram;
113 class WeakObjectRetainer;
114 
115 enum ArrayStorageAllocationMode {
116   DONT_INITIALIZE_ARRAY_ELEMENTS,
117   INITIALIZE_ARRAY_ELEMENTS_WITH_HOLE
118 };
119 
120 enum class ClearRecordedSlots { kYes, kNo };
121 
122 enum class InvalidateRecordedSlots { kYes, kNo };
123 
124 enum class ClearFreedMemoryMode { kClearFreedMemory, kDontClearFreedMemory };
125 
126 enum ExternalBackingStoreType { kArrayBuffer, kExternalString, kNumTypes };
127 
128 enum class RetainingPathOption { kDefault, kTrackEphemeronPath };
129 
130 // These values are persisted to logs. Entries should not be renumbered and
131 // numeric values should never be reused. If you add new items here, update
132 // src/tools/metrics/histograms/enums.xml in chromium.
133 enum class GarbageCollectionReason : int {
134   kUnknown = 0,
135   kAllocationFailure = 1,
136   kAllocationLimit = 2,
137   kContextDisposal = 3,
138   kCountersExtension = 4,
139   kDebugger = 5,
140   kDeserializer = 6,
141   kExternalMemoryPressure = 7,
142   kFinalizeMarkingViaStackGuard = 8,
143   kFinalizeMarkingViaTask = 9,
144   kFullHashtable = 10,
145   kHeapProfiler = 11,
146   kTask = 12,
147   kLastResort = 13,
148   kLowMemoryNotification = 14,
149   kMakeHeapIterable = 15,
150   kMemoryPressure = 16,
151   kMemoryReducer = 17,
152   kRuntime = 18,
153   kSamplingProfiler = 19,
154   kSnapshotCreator = 20,
155   kTesting = 21,
156   kExternalFinalize = 22,
157   kGlobalAllocationLimit = 23,
158   kMeasureMemory = 24,
159   kBackgroundAllocationFailure = 25,
160 
161   kLastReason = kBackgroundAllocationFailure,
162 };
163 
164 static_assert(kGarbageCollectionReasonMaxValue ==
165                   static_cast<int>(GarbageCollectionReason::kLastReason),
166               "The value of kGarbageCollectionReasonMaxValue is inconsistent.");
167 
168 enum class YoungGenerationHandling {
169   kRegularScavenge = 0,
170   kFastPromotionDuringScavenge = 1,
171   // Histogram::InspectConstructionArguments in chromium requires us to have at
172   // least three buckets.
173   kUnusedBucket = 2,
174   // If you add new items here, then update the young_generation_handling in
175   // counters.h.
176   // Also update src/tools/metrics/histograms/histograms.xml in chromium.
177 };
178 
179 enum class GCIdleTimeAction : uint8_t;
180 
181 enum class SkipRoot {
182   kExternalStringTable,
183   kGlobalHandles,
184   kOldGeneration,
185   kStack,
186   kMainThreadHandles,
187   kUnserializable,
188   kWeak
189 };
190 
191 enum UnprotectMemoryOrigin {
192   kMainThread,
193   kMaybeOffMainThread,
194 };
195 
196 class StrongRootsEntry final {
StrongRootsEntry(const char * label)197   explicit StrongRootsEntry(const char* label) : label(label) {}
198 
199   // Label that identifies the roots in tooling.
200   const char* label;
201   FullObjectSlot start;
202   FullObjectSlot end;
203   StrongRootsEntry* prev;
204   StrongRootsEntry* next;
205 
206   friend class Heap;
207 };
208 
209 #ifdef DEBUG
210 struct CommentStatistic {
211   const char* comment;
212   int size;
213   int count;
ClearCommentStatistic214   void Clear() {
215     comment = nullptr;
216     size = 0;
217     count = 0;
218   }
219   // Must be small, since an iteration is used for lookup.
220   static const int kMaxComments = 64;
221 };
222 #endif
223 
224 using EphemeronRememberedSet =
225     std::unordered_map<EphemeronHashTable, std::unordered_set<int>,
226                        Object::Hasher>;
227 
228 class Heap {
229  public:
230   // Stores ephemeron entries where the EphemeronHashTable is in old-space,
231   // and the key of the entry is in new-space. Such keys do not appear in the
232   // usual OLD_TO_NEW remembered set.
233   EphemeronRememberedSet ephemeron_remembered_set_;
234   enum FindMementoMode { kForRuntime, kForGC };
235 
236   enum class HeapGrowingMode { kSlow, kConservative, kMinimal, kDefault };
237 
238   enum HeapState {
239     NOT_IN_GC,
240     SCAVENGE,
241     MARK_COMPACT,
242     MINOR_MARK_COMPACT,
243     TEAR_DOWN
244   };
245 
246   // Emits GC events for DevTools timeline.
247   class V8_NODISCARD DevToolsTraceEventScope {
248    public:
249     DevToolsTraceEventScope(Heap* heap, const char* event_name,
250                             const char* event_type);
251     ~DevToolsTraceEventScope();
252 
253    private:
254     Heap* heap_;
255     const char* event_name_;
256   };
257 
258   class ExternalMemoryAccounting {
259    public:
total()260     int64_t total() { return total_.load(std::memory_order_relaxed); }
limit()261     int64_t limit() { return limit_.load(std::memory_order_relaxed); }
low_since_mark_compact()262     int64_t low_since_mark_compact() {
263       return low_since_mark_compact_.load(std::memory_order_relaxed);
264     }
265 
ResetAfterGC()266     void ResetAfterGC() {
267       set_low_since_mark_compact(total());
268       set_limit(total() + kExternalAllocationSoftLimit);
269     }
270 
Update(int64_t delta)271     int64_t Update(int64_t delta) {
272       const int64_t amount =
273           total_.fetch_add(delta, std::memory_order_relaxed) + delta;
274       if (amount < low_since_mark_compact()) {
275         set_low_since_mark_compact(amount);
276         set_limit(amount + kExternalAllocationSoftLimit);
277       }
278       return amount;
279     }
280 
AllocatedSinceMarkCompact()281     int64_t AllocatedSinceMarkCompact() {
282       int64_t total_bytes = total();
283       int64_t low_since_mark_compact_bytes = low_since_mark_compact();
284 
285       if (total_bytes <= low_since_mark_compact_bytes) {
286         return 0;
287       }
288       return static_cast<uint64_t>(total_bytes - low_since_mark_compact_bytes);
289     }
290 
291    private:
set_total(int64_t value)292     void set_total(int64_t value) {
293       total_.store(value, std::memory_order_relaxed);
294     }
295 
set_limit(int64_t value)296     void set_limit(int64_t value) {
297       limit_.store(value, std::memory_order_relaxed);
298     }
299 
set_low_since_mark_compact(int64_t value)300     void set_low_since_mark_compact(int64_t value) {
301       low_since_mark_compact_.store(value, std::memory_order_relaxed);
302     }
303 
304     // The amount of external memory registered through the API.
305     std::atomic<int64_t> total_{0};
306 
307     // The limit when to trigger memory pressure from the API.
308     std::atomic<int64_t> limit_{kExternalAllocationSoftLimit};
309 
310     // Caches the amount of external memory registered at the last MC.
311     std::atomic<int64_t> low_since_mark_compact_{0};
312   };
313 
314   using PretenuringFeedbackMap =
315       std::unordered_map<AllocationSite, size_t, Object::Hasher>;
316 
317   // Taking this mutex prevents the GC from entering a phase that relocates
318   // object references.
relocation_mutex()319   base::Mutex* relocation_mutex() { return &relocation_mutex_; }
320 
321   // Support for context snapshots.  After calling this we have a linear
322   // space to write objects in each space.
323   struct Chunk {
324     uint32_t size;
325     Address start;
326     Address end;
327   };
328   using Reservation = std::vector<Chunk>;
329 
330 #if V8_OS_ANDROID
331   // Don't apply pointer multiplier on Android since it has no swap space and
332   // should instead adapt it's heap size based on available physical memory.
333   static const int kPointerMultiplier = 1;
334   static const int kHeapLimitMultiplier = 1;
335 #else
336   static const int kPointerMultiplier = kTaggedSize / 4;
337   // The heap limit needs to be computed based on the system pointer size
338   // because we want a pointer-compressed heap to have larger limit than
339   // an orinary 32-bit which that is contrained by 2GB virtual address space.
340   static const int kHeapLimitMultiplier = kSystemPointerSize / 4;
341 #endif
342 
343   static const size_t kMaxInitialOldGenerationSize =
344       256 * MB * kHeapLimitMultiplier;
345 
346   // These constants control heap configuration based on the physical memory.
347   static constexpr size_t kPhysicalMemoryToOldGenerationRatio = 4;
348   // Young generation size is the same for compressed heaps and 32-bit heaps.
349   static constexpr size_t kOldGenerationToSemiSpaceRatio =
350       128 * kHeapLimitMultiplier / kPointerMultiplier;
351   static constexpr size_t kOldGenerationToSemiSpaceRatioLowMemory =
352       256 * kHeapLimitMultiplier / kPointerMultiplier;
353   static constexpr size_t kOldGenerationLowMemory =
354       128 * MB * kHeapLimitMultiplier;
355   static constexpr size_t kNewLargeObjectSpaceToSemiSpaceRatio = 1;
356 #if ENABLE_HUGEPAGE
357   static constexpr size_t kMinSemiSpaceSize =
358       kHugePageSize * kPointerMultiplier;
359   static constexpr size_t kMaxSemiSpaceSize =
360       kHugePageSize * 16 * kPointerMultiplier;
361 #else
362   static constexpr size_t kMinSemiSpaceSize = 512 * KB * kPointerMultiplier;
363   static constexpr size_t kMaxSemiSpaceSize = 8192 * KB * kPointerMultiplier;
364 #endif
365 
366   STATIC_ASSERT(kMinSemiSpaceSize % (1 << kPageSizeBits) == 0);
367   STATIC_ASSERT(kMaxSemiSpaceSize % (1 << kPageSizeBits) == 0);
368 
369   static const int kTraceRingBufferSize = 512;
370   static const int kStacktraceBufferSize = 512;
371 
372   static const int kNoGCFlags = 0;
373   static const int kReduceMemoryFootprintMask = 1;
374   // GCs that are forced, either through testing configurations (requring
375   // --expose-gc) or through DevTools (using LowMemoryNotificaton).
376   static const int kForcedGC = 2;
377 
378   // The minimum size of a HeapObject on the heap.
379   static const int kMinObjectSizeInTaggedWords = 2;
380 
381   static const int kMinPromotedPercentForFastPromotionMode = 90;
382 
383   STATIC_ASSERT(static_cast<int>(RootIndex::kUndefinedValue) ==
384                 Internals::kUndefinedValueRootIndex);
385   STATIC_ASSERT(static_cast<int>(RootIndex::kTheHoleValue) ==
386                 Internals::kTheHoleValueRootIndex);
387   STATIC_ASSERT(static_cast<int>(RootIndex::kNullValue) ==
388                 Internals::kNullValueRootIndex);
389   STATIC_ASSERT(static_cast<int>(RootIndex::kTrueValue) ==
390                 Internals::kTrueValueRootIndex);
391   STATIC_ASSERT(static_cast<int>(RootIndex::kFalseValue) ==
392                 Internals::kFalseValueRootIndex);
393   STATIC_ASSERT(static_cast<int>(RootIndex::kempty_string) ==
394                 Internals::kEmptyStringRootIndex);
395 
396   // Calculates the maximum amount of filler that could be required by the
397   // given alignment.
398   V8_EXPORT_PRIVATE static int GetMaximumFillToAlign(
399       AllocationAlignment alignment);
400   // Calculates the actual amount of filler required for a given address at the
401   // given alignment.
402   V8_EXPORT_PRIVATE static int GetFillToAlign(Address address,
403                                               AllocationAlignment alignment);
404 
405   // Returns the size of the initial area of a code-range, which is marked
406   // writable and reserved to contain unwind information.
407   static size_t GetCodeRangeReservedAreaSize();
408 
409   [[noreturn]] void FatalProcessOutOfMemory(const char* location);
410 
411   // Checks whether the space is valid.
412   static bool IsValidAllocationSpace(AllocationSpace space);
413 
414   // Zapping is needed for verify heap, and always done in debug builds.
ShouldZapGarbage()415   static inline bool ShouldZapGarbage() {
416 #ifdef DEBUG
417     return true;
418 #else
419 #ifdef VERIFY_HEAP
420     return FLAG_verify_heap;
421 #else
422     return false;
423 #endif
424 #endif
425   }
426 
427   // Helper function to get the bytecode flushing mode based on the flags. This
428   // is required because it is not safe to acess flags in concurrent marker.
429   static inline base::EnumSet<CodeFlushMode> GetCodeFlushMode(Isolate* isolate);
430 
ZapValue()431   static uintptr_t ZapValue() {
432     return FLAG_clear_free_memory ? kClearedFreeMemoryValue : kZapValue;
433   }
434 
IsYoungGenerationCollector(GarbageCollector collector)435   static inline bool IsYoungGenerationCollector(GarbageCollector collector) {
436     return collector == GarbageCollector::SCAVENGER ||
437            collector == GarbageCollector::MINOR_MARK_COMPACTOR;
438   }
439 
YoungGenerationCollector()440   static inline GarbageCollector YoungGenerationCollector() {
441     return (FLAG_minor_mc) ? GarbageCollector::MINOR_MARK_COMPACTOR
442                            : GarbageCollector::SCAVENGER;
443   }
444 
CollectorName(GarbageCollector collector)445   static inline const char* CollectorName(GarbageCollector collector) {
446     switch (collector) {
447       case GarbageCollector::SCAVENGER:
448         return "Scavenger";
449       case GarbageCollector::MARK_COMPACTOR:
450         return "Mark-Compact";
451       case GarbageCollector::MINOR_MARK_COMPACTOR:
452         return "Minor Mark-Compact";
453     }
454     return "Unknown collector";
455   }
456 
CollectorName(v8::GCType gc_type)457   static inline const char* CollectorName(v8::GCType gc_type) {
458     switch (gc_type) {
459       case kGCTypeScavenge:
460         return "Scavenger";
461       case kGCTypeMarkSweepCompact:
462         return "Mark-Compact";
463       case kGCTypeMinorMarkCompact:
464         return "Minor Mark-Compact";
465       default:
466         break;
467     }
468     return "Unknown collector";
469   }
470 
471   // Copy block of memory from src to dst. Size of block should be aligned
472   // by pointer size.
473   static inline void CopyBlock(Address dst, Address src, int byte_size);
474 
475   // Executes generational and/or marking write barrier for a [start, end) range
476   // of non-weak slots inside |object|.
477   template <typename TSlot>
478   V8_EXPORT_PRIVATE void WriteBarrierForRange(HeapObject object, TSlot start,
479                                               TSlot end);
480 
481   V8_EXPORT_PRIVATE static void WriteBarrierForCodeSlow(Code host);
482 
483   V8_EXPORT_PRIVATE static void GenerationalBarrierSlow(HeapObject object,
484                                                         Address slot,
485                                                         HeapObject value);
486   V8_EXPORT_PRIVATE inline void RecordEphemeronKeyWrite(
487       EphemeronHashTable table, Address key_slot);
488   V8_EXPORT_PRIVATE static void EphemeronKeyWriteBarrierFromCode(
489       Address raw_object, Address address, Isolate* isolate);
490   V8_EXPORT_PRIVATE static void GenerationalBarrierForCodeSlow(
491       Code host, RelocInfo* rinfo, HeapObject value);
492   V8_EXPORT_PRIVATE static bool PageFlagsAreConsistent(HeapObject object);
493 
494   // Notifies the heap that is ok to start marking or other activities that
495   // should not happen during deserialization.
496   void NotifyDeserializationComplete();
497 
498   void NotifyBootstrapComplete();
499 
500   void NotifyOldGenerationExpansion(AllocationSpace space, MemoryChunk* chunk);
501 
502   inline Address* NewSpaceAllocationTopAddress();
503   inline Address* NewSpaceAllocationLimitAddress();
504   inline Address* OldSpaceAllocationTopAddress();
505   inline Address* OldSpaceAllocationLimitAddress();
506 
507   size_t NewSpaceSize();
508   size_t NewSpaceCapacity();
509 
510   // Move len non-weak tagged elements from src_slot to dst_slot of dst_object.
511   // The source and destination memory ranges can overlap.
512   V8_EXPORT_PRIVATE void MoveRange(HeapObject dst_object, ObjectSlot dst_slot,
513                                    ObjectSlot src_slot, int len,
514                                    WriteBarrierMode mode);
515 
516   // Copy len non-weak tagged elements from src_slot to dst_slot of dst_object.
517   // The source and destination memory ranges must not overlap.
518   template <typename TSlot>
519   void CopyRange(HeapObject dst_object, TSlot dst_slot, TSlot src_slot, int len,
520                  WriteBarrierMode mode);
521 
522   // Initialize a filler object to keep the ability to iterate over the heap
523   // when introducing gaps within pages. If slots could have been recorded in
524   // the freed area, then pass ClearRecordedSlots::kYes as the mode. Otherwise,
525   // pass ClearRecordedSlots::kNo. Clears memory if clearing slots.
526   V8_EXPORT_PRIVATE HeapObject CreateFillerObjectAt(
527       Address addr, int size, ClearRecordedSlots clear_slots_mode);
528 
529   void CreateFillerObjectAtBackground(Address addr, int size,
530                                       ClearFreedMemoryMode clear_memory_mode);
531 
532   template <typename T>
533   void CreateFillerForArray(T object, int elements_to_trim, int bytes_to_trim);
534 
535   bool CanMoveObjectStart(HeapObject object);
536 
537   bool IsImmovable(HeapObject object);
538 
539   V8_EXPORT_PRIVATE static bool IsLargeObject(HeapObject object);
540 
541   // Trim the given array from the left. Note that this relocates the object
542   // start and hence is only valid if there is only a single reference to it.
543   V8_EXPORT_PRIVATE FixedArrayBase LeftTrimFixedArray(FixedArrayBase obj,
544                                                       int elements_to_trim);
545 
546   // Trim the given array from the right.
547   V8_EXPORT_PRIVATE void RightTrimFixedArray(FixedArrayBase obj,
548                                              int elements_to_trim);
549   void RightTrimWeakFixedArray(WeakFixedArray obj, int elements_to_trim);
550 
551   // Converts the given boolean condition to JavaScript boolean value.
552   inline Oddball ToBoolean(bool condition);
553 
554   // Notify the heap that a context has been disposed.
555   V8_EXPORT_PRIVATE int NotifyContextDisposed(bool dependant_context);
556 
set_native_contexts_list(Object object)557   void set_native_contexts_list(Object object) {
558     native_contexts_list_.store(object.ptr(), std::memory_order_release);
559   }
560 
native_contexts_list()561   Object native_contexts_list() const {
562     return Object(native_contexts_list_.load(std::memory_order_acquire));
563   }
564 
set_allocation_sites_list(Object object)565   void set_allocation_sites_list(Object object) {
566     allocation_sites_list_ = object;
567   }
allocation_sites_list()568   Object allocation_sites_list() { return allocation_sites_list_; }
569 
set_dirty_js_finalization_registries_list(Object object)570   void set_dirty_js_finalization_registries_list(Object object) {
571     dirty_js_finalization_registries_list_ = object;
572   }
dirty_js_finalization_registries_list()573   Object dirty_js_finalization_registries_list() {
574     return dirty_js_finalization_registries_list_;
575   }
set_dirty_js_finalization_registries_list_tail(Object object)576   void set_dirty_js_finalization_registries_list_tail(Object object) {
577     dirty_js_finalization_registries_list_tail_ = object;
578   }
dirty_js_finalization_registries_list_tail()579   Object dirty_js_finalization_registries_list_tail() {
580     return dirty_js_finalization_registries_list_tail_;
581   }
582 
583   // Used in CreateAllocationSiteStub and the (de)serializer.
allocation_sites_list_address()584   Address allocation_sites_list_address() {
585     return reinterpret_cast<Address>(&allocation_sites_list_);
586   }
587 
588   // Traverse all the allocaions_sites [nested_site and weak_next] in the list
589   // and foreach call the visitor
590   void ForeachAllocationSite(
591       Object list, const std::function<void(AllocationSite)>& visitor);
592 
593   // Number of mark-sweeps.
ms_count()594   int ms_count() const { return ms_count_; }
595 
596   // Checks whether the given object is allowed to be migrated from it's
597   // current space into the given destination space. Used for debugging.
598   bool AllowedToBeMigrated(Map map, HeapObject object, AllocationSpace dest);
599 
600   void CheckHandleCount();
601 
602   // Print short heap statistics.
603   void PrintShortHeapStatistics();
604 
605   // Print statistics of freelists of old_space:
606   //  with FLAG_trace_gc_freelists: summary of each FreeListCategory.
607   //  with FLAG_trace_gc_freelists_verbose: also prints the statistics of each
608   //  FreeListCategory of each page.
609   void PrintFreeListsStats();
610 
611   // Dump heap statistics in JSON format.
612   void DumpJSONHeapStatistics(std::stringstream& stream);
613 
write_protect_code_memory()614   bool write_protect_code_memory() const { return write_protect_code_memory_; }
615 
code_space_memory_modification_scope_depth()616   uintptr_t code_space_memory_modification_scope_depth() {
617     return code_space_memory_modification_scope_depth_;
618   }
619 
increment_code_space_memory_modification_scope_depth()620   void increment_code_space_memory_modification_scope_depth() {
621     code_space_memory_modification_scope_depth_++;
622   }
623 
decrement_code_space_memory_modification_scope_depth()624   void decrement_code_space_memory_modification_scope_depth() {
625     code_space_memory_modification_scope_depth_--;
626   }
627 
628   void UnprotectAndRegisterMemoryChunk(MemoryChunk* chunk,
629                                        UnprotectMemoryOrigin origin);
630   V8_EXPORT_PRIVATE void UnprotectAndRegisterMemoryChunk(
631       HeapObject object, UnprotectMemoryOrigin origin);
632   void UnregisterUnprotectedMemoryChunk(MemoryChunk* chunk);
633   V8_EXPORT_PRIVATE void ProtectUnprotectedMemoryChunks();
634 
IncrementCodePageCollectionMemoryModificationScopeDepth()635   void IncrementCodePageCollectionMemoryModificationScopeDepth() {
636     code_page_collection_memory_modification_scope_depth_++;
637   }
638 
DecrementCodePageCollectionMemoryModificationScopeDepth()639   void DecrementCodePageCollectionMemoryModificationScopeDepth() {
640     code_page_collection_memory_modification_scope_depth_--;
641   }
642 
code_page_collection_memory_modification_scope_depth()643   uintptr_t code_page_collection_memory_modification_scope_depth() {
644     return code_page_collection_memory_modification_scope_depth_;
645   }
646 
gc_state()647   inline HeapState gc_state() const {
648     return gc_state_.load(std::memory_order_relaxed);
649   }
650   void SetGCState(HeapState state);
IsTearingDown()651   bool IsTearingDown() const { return gc_state() == TEAR_DOWN; }
force_oom()652   bool force_oom() const { return force_oom_; }
653 
ignore_local_gc_requests()654   bool ignore_local_gc_requests() const {
655     return ignore_local_gc_requests_depth_ > 0;
656   }
657 
IsInGCPostProcessing()658   inline bool IsInGCPostProcessing() { return gc_post_processing_depth_ > 0; }
659 
660   bool IsGCWithoutStack() const;
661 
662   // If an object has an AllocationMemento trailing it, return it, otherwise
663   // return a null AllocationMemento.
664   template <FindMementoMode mode>
665   inline AllocationMemento FindAllocationMemento(Map map, HeapObject object);
666 
667   // Performs GC after background allocation failure.
668   void CollectGarbageForBackground(LocalHeap* local_heap);
669 
670   //
671   // Support for the API.
672   //
673 
674   void CreateApiObjects();
675 
676   // Implements the corresponding V8 API function.
677   bool IdleNotification(double deadline_in_seconds);
678   bool IdleNotification(int idle_time_in_ms);
679 
680   V8_EXPORT_PRIVATE void MemoryPressureNotification(MemoryPressureLevel level,
681                                                     bool is_isolate_locked);
682   void CheckMemoryPressure();
683 
684   V8_EXPORT_PRIVATE void AddNearHeapLimitCallback(v8::NearHeapLimitCallback,
685                                                   void* data);
686   V8_EXPORT_PRIVATE void RemoveNearHeapLimitCallback(
687       v8::NearHeapLimitCallback callback, size_t heap_limit);
688   V8_EXPORT_PRIVATE void AutomaticallyRestoreInitialHeapLimit(
689       double threshold_percent);
690 
691   void AppendArrayBufferExtension(JSArrayBuffer object,
692                                   ArrayBufferExtension* extension);
693   void DetachArrayBufferExtension(JSArrayBuffer object,
694                                   ArrayBufferExtension* extension);
695 
safepoint()696   IsolateSafepoint* safepoint() { return safepoint_.get(); }
697 
698   V8_EXPORT_PRIVATE double MonotonicallyIncreasingTimeInMs() const;
699 
700 #if DEBUG
701   void VerifyNewSpaceTop();
702 #endif  // DEBUG
703 
704   void RecordStats(HeapStats* stats, bool take_snapshot = false);
705 
706   bool MeasureMemory(std::unique_ptr<v8::MeasureMemoryDelegate> delegate,
707                      v8::MeasureMemoryExecution execution);
708 
709   std::unique_ptr<v8::MeasureMemoryDelegate> MeasureMemoryDelegate(
710       Handle<NativeContext> context, Handle<JSPromise> promise,
711       v8::MeasureMemoryMode mode);
712 
713   // Check new space expansion criteria and expand semispaces if it was hit.
714   void CheckNewSpaceExpansionCriteria();
715 
716   void VisitExternalResources(v8::ExternalResourceVisitor* visitor);
717 
718   // An object should be promoted if the object has survived a
719   // scavenge operation.
720   inline bool ShouldBePromoted(Address old_address);
721 
722   void IncrementDeferredCount(v8::Isolate::UseCounterFeature feature);
723 
724   inline int NextScriptId();
725   inline int NextDebuggingId();
726   inline int GetNextTemplateSerialNumber();
727 
728   void SetSerializedObjects(FixedArray objects);
729   void SetSerializedGlobalProxySizes(FixedArray sizes);
730 
731   void SetBasicBlockProfilingData(Handle<ArrayList> list);
732 
733   // For post mortem debugging.
734   void RememberUnmappedPage(Address page, bool compacted);
735 
external_memory_hard_limit()736   int64_t external_memory_hard_limit() { return max_old_generation_size() / 2; }
737 
738   V8_INLINE int64_t external_memory();
739   V8_EXPORT_PRIVATE int64_t external_memory_limit();
740   V8_INLINE int64_t update_external_memory(int64_t delta);
741 
742   V8_EXPORT_PRIVATE size_t YoungArrayBufferBytes();
743   V8_EXPORT_PRIVATE size_t OldArrayBufferBytes();
744 
backing_store_bytes()745   uint64_t backing_store_bytes() const {
746     return backing_store_bytes_.load(std::memory_order_relaxed);
747   }
748 
749   void CompactWeakArrayLists();
750 
751   V8_EXPORT_PRIVATE void AddRetainedMap(Handle<NativeContext> context,
752                                         Handle<Map> map);
753 
754   // This event is triggered after object is moved to a new place.
755   void OnMoveEvent(HeapObject target, HeapObject source, int size_in_bytes);
756 
deserialization_complete()757   bool deserialization_complete() const { return deserialization_complete_; }
758 
759   // We can only invoke Safepoint() on the main thread local heap after
760   // deserialization is complete. Before that, main_thread_local_heap_ might be
761   // null.
CanSafepoint()762   V8_INLINE bool CanSafepoint() const { return deserialization_complete(); }
763 
764   bool HasLowAllocationRate();
765   bool HasHighFragmentation();
766   bool HasHighFragmentation(size_t used, size_t committed);
767 
768   void ActivateMemoryReducerIfNeeded();
769 
770   V8_EXPORT_PRIVATE bool ShouldOptimizeForMemoryUsage();
771 
HighMemoryPressure()772   bool HighMemoryPressure() {
773     return memory_pressure_level_.load(std::memory_order_relaxed) !=
774            MemoryPressureLevel::kNone;
775   }
776 
777   bool CollectionRequested();
778 
779   void CheckCollectionRequested();
780 
RestoreHeapLimit(size_t heap_limit)781   void RestoreHeapLimit(size_t heap_limit) {
782     // Do not set the limit lower than the live size + some slack.
783     size_t min_limit = SizeOfObjects() + SizeOfObjects() / 4;
784     set_max_old_generation_size(
785         std::min(max_old_generation_size(), std::max(heap_limit, min_limit)));
786   }
787 
788 #if V8_ENABLE_WEBASSEMBLY
789   // TODO(manoskouk): Inline this if STRONG_MUTABLE_MOVABLE_ROOT_LIST setters
790   // become public.
791   void EnsureWasmCanonicalRttsSize(int length);
792 #endif
793 
794   // ===========================================================================
795   // Initialization. ===========================================================
796   // ===========================================================================
797 
798   void ConfigureHeap(const v8::ResourceConstraints& constraints);
799   void ConfigureHeapDefault();
800 
801   // Prepares the heap, setting up for deserialization.
802   void SetUp(LocalHeap* main_thread_local_heap);
803 
804   // Sets read-only heap and space.
805   void SetUpFromReadOnlyHeap(ReadOnlyHeap* ro_heap);
806 
807   void ReplaceReadOnlySpace(SharedReadOnlySpace* shared_ro_space);
808 
809   // Sets up the heap memory without creating any objects.
810   void SetUpSpaces(LinearAllocationArea* new_allocation_info,
811                    LinearAllocationArea* old_allocation_info);
812 
813   // Prepares the heap, setting up for deserialization.
814   void InitializeMainThreadLocalHeap(LocalHeap* main_thread_local_heap);
815 
816   // (Re-)Initialize hash seed from flag or RNG.
817   void InitializeHashSeed();
818 
819   // Invoked once for the process from V8::Initialize.
820   static void InitializeOncePerProcess();
821 
822   // Bootstraps the object heap with the core set of objects required to run.
823   // Returns whether it succeeded.
824   bool CreateHeapObjects();
825 
826   // Create ObjectStats if live_object_stats_ or dead_object_stats_ are nullptr.
827   void CreateObjectStats();
828 
829   // Sets the TearDown state, so no new GC tasks get posted.
830   void StartTearDown();
831 
832   // Destroys all memory allocated by the heap.
833   void TearDown();
834 
835   // Returns whether SetUp has been called.
836   bool HasBeenSetUp() const;
837 
838   // ===========================================================================
839   // Getters for spaces. =======================================================
840   // ===========================================================================
841 
842   inline Address NewSpaceTop();
843 
new_space()844   NewSpace* new_space() { return new_space_; }
old_space()845   OldSpace* old_space() { return old_space_; }
shared_old_space()846   OldSpace* shared_old_space() { return shared_old_space_; }
code_space()847   CodeSpace* code_space() { return code_space_; }
map_space()848   MapSpace* map_space() { return map_space_; }
849   inline PagedSpace* space_for_maps();
lo_space()850   OldLargeObjectSpace* lo_space() { return lo_space_; }
code_lo_space()851   CodeLargeObjectSpace* code_lo_space() { return code_lo_space_; }
new_lo_space()852   NewLargeObjectSpace* new_lo_space() { return new_lo_space_; }
read_only_space()853   ReadOnlySpace* read_only_space() { return read_only_space_; }
854 
855   inline PagedSpace* paged_space(int idx);
856   inline Space* space(int idx);
857 
858   // ===========================================================================
859   // Getters to other components. ==============================================
860   // ===========================================================================
861 
tracer()862   GCTracer* tracer() { return tracer_.get(); }
863 
memory_allocator()864   MemoryAllocator* memory_allocator() { return memory_allocator_.get(); }
memory_allocator()865   const MemoryAllocator* memory_allocator() const {
866     return memory_allocator_.get();
867   }
868 
869   inline ConcurrentAllocator* concurrent_allocator_for_maps();
870 
871   inline Isolate* isolate();
872 
mark_compact_collector()873   MarkCompactCollector* mark_compact_collector() {
874     return mark_compact_collector_.get();
875   }
876 
minor_mark_compact_collector()877   MinorMarkCompactCollector* minor_mark_compact_collector() {
878     return minor_mark_compact_collector_.get();
879   }
880 
array_buffer_sweeper()881   ArrayBufferSweeper* array_buffer_sweeper() {
882     return array_buffer_sweeper_.get();
883   }
884 
885   // The potentially overreserved address space region reserved by the code
886   // range if it exists or empty region otherwise.
887   const base::AddressRegion& code_region();
888 
code_range()889   CodeRange* code_range() { return code_range_.get(); }
890 
891   // The base of the code range if it exists or null address.
892   inline Address code_range_base();
893 
main_thread_local_heap()894   LocalHeap* main_thread_local_heap() { return main_thread_local_heap_; }
895 
AsHeap()896   Heap* AsHeap() { return this; }
897 
898   // ===========================================================================
899   // Root set access. ==========================================================
900   // ===========================================================================
901 
902   // Shortcut to the roots table stored in the Isolate.
903   V8_INLINE RootsTable& roots_table();
904 
905 // Heap root getters.
906 #define ROOT_ACCESSOR(type, name, CamelName) inline type name();
907   MUTABLE_ROOT_LIST(ROOT_ACCESSOR)
908 #undef ROOT_ACCESSOR
909 
910   V8_INLINE void SetRootMaterializedObjects(FixedArray objects);
911   V8_INLINE void SetRootScriptList(Object value);
912   V8_INLINE void SetRootNoScriptSharedFunctionInfos(Object value);
913   V8_INLINE void SetMessageListeners(TemplateList value);
914   V8_INLINE void SetPendingOptimizeForTestBytecode(Object bytecode);
915 
916   StrongRootsEntry* RegisterStrongRoots(const char* label, FullObjectSlot start,
917                                         FullObjectSlot end);
918   void UnregisterStrongRoots(StrongRootsEntry* entry);
919   void UpdateStrongRoots(StrongRootsEntry* entry, FullObjectSlot start,
920                          FullObjectSlot end);
921 
922   void SetBuiltinsConstantsTable(FixedArray cache);
923   void SetDetachedContexts(WeakArrayList detached_contexts);
924 
925   // A full copy of the interpreter entry trampoline, used as a template to
926   // create copies of the builtin at runtime. The copies are used to create
927   // better profiling information for ticks in bytecode execution. Note that
928   // this is always a copy of the full builtin, i.e. not the off-heap
929   // trampoline.
930   // See also: FLAG_interpreted_frames_native_stack.
931   void SetInterpreterEntryTrampolineForProfiling(Code code);
932 
933   void EnqueueDirtyJSFinalizationRegistry(
934       JSFinalizationRegistry finalization_registry,
935       std::function<void(HeapObject object, ObjectSlot slot, Object target)>
936           gc_notify_updated_slot);
937 
938   MaybeHandle<JSFinalizationRegistry> DequeueDirtyJSFinalizationRegistry();
939 
940   // Called from Heap::NotifyContextDisposed to remove all
941   // FinalizationRegistries with {context} from the dirty list when the context
942   // e.g. navigates away or is detached. If the dirty list is empty afterwards,
943   // the cleanup task is aborted if needed.
944   void RemoveDirtyFinalizationRegistriesOnContext(NativeContext context);
945 
946   inline bool HasDirtyJSFinalizationRegistries();
947 
948   void PostFinalizationRegistryCleanupTaskIfNeeded();
949 
set_is_finalization_registry_cleanup_task_posted(bool posted)950   void set_is_finalization_registry_cleanup_task_posted(bool posted) {
951     is_finalization_registry_cleanup_task_posted_ = posted;
952   }
953 
is_finalization_registry_cleanup_task_posted()954   bool is_finalization_registry_cleanup_task_posted() {
955     return is_finalization_registry_cleanup_task_posted_;
956   }
957 
958   V8_EXPORT_PRIVATE void KeepDuringJob(Handle<HeapObject> target);
959   void ClearKeptObjects();
960 
961   // ===========================================================================
962   // Inline allocation. ========================================================
963   // ===========================================================================
964 
965   // Switch whether inline bump-pointer allocation should be used.
966   V8_EXPORT_PRIVATE void EnableInlineAllocation();
967   V8_EXPORT_PRIVATE void DisableInlineAllocation();
968 
969   // ===========================================================================
970   // Methods triggering GCs. ===================================================
971   // ===========================================================================
972 
973   // Performs garbage collection operation.
974   // Returns whether there is a chance that another major GC could
975   // collect more garbage.
976   V8_EXPORT_PRIVATE bool CollectGarbage(
977       AllocationSpace space, GarbageCollectionReason gc_reason,
978       const GCCallbackFlags gc_callback_flags = kNoGCCallbackFlags);
979 
980   // Performs a full garbage collection.
981   V8_EXPORT_PRIVATE void CollectAllGarbage(
982       int flags, GarbageCollectionReason gc_reason,
983       const GCCallbackFlags gc_callback_flags = kNoGCCallbackFlags);
984 
985   // Last hope GC, should try to squeeze as much as possible.
986   V8_EXPORT_PRIVATE void CollectAllAvailableGarbage(
987       GarbageCollectionReason gc_reason);
988 
989   // Precise garbage collection that potentially finalizes already running
990   // incremental marking before performing an atomic garbage collection.
991   // Only use if absolutely necessary or in tests to avoid floating garbage!
992   V8_EXPORT_PRIVATE void PreciseCollectAllGarbage(
993       int flags, GarbageCollectionReason gc_reason,
994       const GCCallbackFlags gc_callback_flags = kNoGCCallbackFlags);
995 
996   // Performs garbage collection operation for the shared heap.
997   V8_EXPORT_PRIVATE void CollectSharedGarbage(
998       GarbageCollectionReason gc_reason);
999 
1000   // Reports and external memory pressure event, either performs a major GC or
1001   // completes incremental marking in order to free external resources.
1002   void ReportExternalMemoryPressure();
1003 
1004   using GetExternallyAllocatedMemoryInBytesCallback =
1005       v8::Isolate::GetExternallyAllocatedMemoryInBytesCallback;
1006 
SetGetExternallyAllocatedMemoryInBytesCallback(GetExternallyAllocatedMemoryInBytesCallback callback)1007   void SetGetExternallyAllocatedMemoryInBytesCallback(
1008       GetExternallyAllocatedMemoryInBytesCallback callback) {
1009     external_memory_callback_ = callback;
1010   }
1011 
1012   // Invoked when GC was requested via the stack guard.
1013   void HandleGCRequest();
1014 
1015   // ===========================================================================
1016   // Iterators. ================================================================
1017   // ===========================================================================
1018 
1019   // None of these methods iterate over the read-only roots. To do this use
1020   // ReadOnlyRoots::Iterate. Read-only root iteration is not necessary for
1021   // garbage collection and is usually only performed as part of
1022   // (de)serialization or heap verification.
1023 
1024   // Iterates over the strong roots and the weak roots.
1025   void IterateRoots(RootVisitor* v, base::EnumSet<SkipRoot> options);
1026   void IterateRootsIncludingClients(RootVisitor* v,
1027                                     base::EnumSet<SkipRoot> options);
1028 
1029   // Iterates over entries in the smi roots list.  Only interesting to the
1030   // serializer/deserializer, since GC does not care about smis.
1031   void IterateSmiRoots(RootVisitor* v);
1032   // Iterates over weak string tables.
1033   void IterateWeakRoots(RootVisitor* v, base::EnumSet<SkipRoot> options);
1034   void IterateWeakGlobalHandles(RootVisitor* v);
1035   void IterateBuiltins(RootVisitor* v);
1036   void IterateStackRoots(RootVisitor* v);
1037 
1038   // ===========================================================================
1039   // Remembered set API. =======================================================
1040   // ===========================================================================
1041 
1042   // Used for query incremental marking status in generated code.
IsMarkingFlagAddress()1043   Address* IsMarkingFlagAddress() {
1044     return reinterpret_cast<Address*>(&is_marking_flag_);
1045   }
1046 
SetIsMarkingFlag(uint8_t flag)1047   void SetIsMarkingFlag(uint8_t flag) { is_marking_flag_ = flag; }
1048 
1049   void ClearRecordedSlot(HeapObject object, ObjectSlot slot);
1050   void ClearRecordedSlotRange(Address start, Address end);
1051   static int InsertIntoRememberedSetFromCode(MemoryChunk* chunk, Address slot);
1052 
1053 #ifdef DEBUG
1054   void VerifyClearedSlot(HeapObject object, ObjectSlot slot);
1055   void VerifySlotRangeHasNoRecordedSlots(Address start, Address end);
1056 #endif
1057 
1058   // ===========================================================================
1059   // Incremental marking API. ==================================================
1060   // ===========================================================================
1061 
GCFlagsForIncrementalMarking()1062   int GCFlagsForIncrementalMarking() {
1063     return ShouldOptimizeForMemoryUsage() ? kReduceMemoryFootprintMask
1064                                           : kNoGCFlags;
1065   }
1066 
1067   // Start incremental marking and ensure that idle time handler can perform
1068   // incremental steps.
1069   V8_EXPORT_PRIVATE void StartIdleIncrementalMarking(
1070       GarbageCollectionReason gc_reason,
1071       GCCallbackFlags gc_callback_flags = GCCallbackFlags::kNoGCCallbackFlags);
1072 
1073   // Starts incremental marking assuming incremental marking is currently
1074   // stopped.
1075   V8_EXPORT_PRIVATE void StartIncrementalMarking(
1076       int gc_flags, GarbageCollectionReason gc_reason,
1077       GCCallbackFlags gc_callback_flags = GCCallbackFlags::kNoGCCallbackFlags);
1078 
1079   void StartIncrementalMarkingIfAllocationLimitIsReached(
1080       int gc_flags,
1081       GCCallbackFlags gc_callback_flags = GCCallbackFlags::kNoGCCallbackFlags);
1082   void StartIncrementalMarkingIfAllocationLimitIsReachedBackground();
1083 
1084   void FinalizeIncrementalMarkingIfComplete(GarbageCollectionReason gc_reason);
1085   // Synchronously finalizes incremental marking.
1086   V8_EXPORT_PRIVATE void FinalizeIncrementalMarkingAtomically(
1087       GarbageCollectionReason gc_reason);
1088 
1089   void CompleteSweepingFull();
1090   void CompleteSweepingYoung(GarbageCollector collector);
1091 
1092   // Ensures that sweeping is finished for that object's page.
1093   void EnsureSweepingCompleted(HeapObject object);
1094 
incremental_marking()1095   IncrementalMarking* incremental_marking() const {
1096     return incremental_marking_.get();
1097   }
1098 
marking_barrier()1099   MarkingBarrier* marking_barrier() const { return marking_barrier_.get(); }
1100 
1101   // ===========================================================================
1102   // Concurrent marking API. ===================================================
1103   // ===========================================================================
1104 
concurrent_marking()1105   ConcurrentMarking* concurrent_marking() const {
1106     return concurrent_marking_.get();
1107   }
1108 
1109   // The runtime uses this function to notify potentially unsafe object layout
1110   // changes that require special synchronization with the concurrent marker.
1111   // The old size is the size of the object before layout change.
1112   // By default recorded slots in the object are invalidated. Pass
1113   // InvalidateRecordedSlots::kNo if this is not necessary or to perform this
1114   // manually.
1115   void NotifyObjectLayoutChange(
1116       HeapObject object, const DisallowGarbageCollection&,
1117       InvalidateRecordedSlots invalidate_recorded_slots =
1118           InvalidateRecordedSlots::kYes);
1119 
1120 #ifdef VERIFY_HEAP
1121   // This function checks that either
1122   // - the map transition is safe,
1123   // - or it was communicated to GC using NotifyObjectLayoutChange.
1124   V8_EXPORT_PRIVATE void VerifyObjectLayoutChange(HeapObject object,
1125                                                   Map new_map);
1126   // Checks that this is a safe map transition.
1127   V8_EXPORT_PRIVATE void VerifySafeMapTransition(HeapObject object,
1128                                                  Map new_map);
1129 #endif
1130 
1131   // ===========================================================================
1132   // Deoptimization support API. ===============================================
1133   // ===========================================================================
1134 
1135   // Setters for code offsets of well-known deoptimization targets.
1136   void SetConstructStubCreateDeoptPCOffset(int pc_offset);
1137   void SetConstructStubInvokeDeoptPCOffset(int pc_offset);
1138   void SetInterpreterEntryReturnPCOffset(int pc_offset);
1139 
1140   // Invalidates references in the given {code} object that are referenced
1141   // transitively from the deoptimization data. Mutates write-protected code.
1142   void InvalidateCodeDeoptimizationData(Code code);
1143 
1144   void DeoptMarkedAllocationSites();
1145 
1146   bool DeoptMaybeTenuredAllocationSites();
1147 
1148   // ===========================================================================
1149   // Embedder heap tracer support. =============================================
1150   // ===========================================================================
1151 
local_embedder_heap_tracer()1152   LocalEmbedderHeapTracer* local_embedder_heap_tracer() const {
1153     return local_embedder_heap_tracer_.get();
1154   }
1155 
1156   V8_EXPORT_PRIVATE void SetEmbedderHeapTracer(EmbedderHeapTracer* tracer);
1157   EmbedderHeapTracer* GetEmbedderHeapTracer() const;
1158 
1159   void RegisterExternallyReferencedObject(Address* location);
1160 
1161   EmbedderHeapTracer::TraceFlags flags_for_embedder_tracer() const;
1162 
1163   // ===========================================================================
1164   // Unified heap (C++) support. ===============================================
1165   // ===========================================================================
1166 
1167   V8_EXPORT_PRIVATE void AttachCppHeap(v8::CppHeap* cpp_heap);
1168   V8_EXPORT_PRIVATE void DetachCppHeap();
1169 
cpp_heap()1170   v8::CppHeap* cpp_heap() const { return cpp_heap_; }
1171 
1172   const cppgc::EmbedderStackState* overriden_stack_state() const;
1173 
1174   // ===========================================================================
1175   // Embedder roots optimizations. =============================================
1176   // ===========================================================================
1177 
1178   V8_EXPORT_PRIVATE void SetEmbedderRootsHandler(EmbedderRootsHandler* handler);
1179 
1180   EmbedderRootsHandler* GetEmbedderRootsHandler() const;
1181 
1182   // ===========================================================================
1183   // External string table API. ================================================
1184   // ===========================================================================
1185 
1186   // Registers an external string.
1187   inline void RegisterExternalString(String string);
1188 
1189   // Called when a string's resource is changed. The size of the payload is sent
1190   // as argument of the method.
1191   V8_EXPORT_PRIVATE void UpdateExternalString(String string, size_t old_payload,
1192                                               size_t new_payload);
1193 
1194   // Finalizes an external string by deleting the associated external
1195   // data and clearing the resource pointer.
1196   inline void FinalizeExternalString(String string);
1197 
1198   static String UpdateYoungReferenceInExternalStringTableEntry(
1199       Heap* heap, FullObjectSlot pointer);
1200 
1201   // ===========================================================================
1202   // Methods checking/returning the space of a given object/address. ===========
1203   // ===========================================================================
1204 
1205   // Returns whether the object resides in new space.
1206   static inline bool InYoungGeneration(Object object);
1207   static inline bool InYoungGeneration(MaybeObject object);
1208   static inline bool InYoungGeneration(HeapObject heap_object);
1209   static inline bool InFromPage(Object object);
1210   static inline bool InFromPage(MaybeObject object);
1211   static inline bool InFromPage(HeapObject heap_object);
1212   static inline bool InToPage(Object object);
1213   static inline bool InToPage(MaybeObject object);
1214   static inline bool InToPage(HeapObject heap_object);
1215 
1216   // Returns whether the object resides in old space.
1217   inline bool InOldSpace(Object object);
1218 
1219   // Checks whether an address/object is in the non-read-only heap (including
1220   // auxiliary area and unused area). Use IsValidHeapObject if checking both
1221   // heaps is required.
1222   V8_EXPORT_PRIVATE bool Contains(HeapObject value) const;
1223   // Same as above, but checks whether the object resides in any of the code
1224   // spaces.
1225   V8_EXPORT_PRIVATE bool ContainsCode(HeapObject value) const;
1226 
1227   // Checks whether an address/object is in the non-read-only heap (including
1228   // auxiliary area and unused area). Use IsValidHeapObject if checking both
1229   // heaps is required.
1230   V8_EXPORT_PRIVATE bool SharedHeapContains(HeapObject value) const;
1231 
1232   // Returns whether the object should be in the shared old space.
1233   V8_EXPORT_PRIVATE bool ShouldBeInSharedOldSpace(HeapObject value);
1234 
1235   // Checks whether an address/object in a space.
1236   // Currently used by tests, serialization and heap verification only.
1237   V8_EXPORT_PRIVATE bool InSpace(HeapObject value, AllocationSpace space) const;
1238 
1239   // Returns true when this heap is shared.
1240   V8_EXPORT_PRIVATE bool IsShared();
1241 
1242   // Slow methods that can be used for verification as they can also be used
1243   // with off-heap Addresses.
1244   V8_EXPORT_PRIVATE bool InSpaceSlow(Address addr, AllocationSpace space) const;
1245 
1246   static inline Heap* FromWritableHeapObject(HeapObject obj);
1247 
1248   // ===========================================================================
1249   // Object statistics tracking. ===============================================
1250   // ===========================================================================
1251 
1252   // Returns the number of buckets used by object statistics tracking during a
1253   // major GC. Note that the following methods fail gracefully when the bounds
1254   // are exceeded though.
1255   size_t NumberOfTrackedHeapObjectTypes();
1256 
1257   // Returns object statistics about count and size at the last major GC.
1258   // Objects are being grouped into buckets that roughly resemble existing
1259   // instance types.
1260   size_t ObjectCountAtLastGC(size_t index);
1261   size_t ObjectSizeAtLastGC(size_t index);
1262 
1263   // Retrieves names of buckets used by object statistics tracking.
1264   bool GetObjectTypeName(size_t index, const char** object_type,
1265                          const char** object_sub_type);
1266 
1267   // The total number of native contexts object on the heap.
1268   size_t NumberOfNativeContexts();
1269   // The total number of native contexts that were detached but were not
1270   // garbage collected yet.
1271   size_t NumberOfDetachedContexts();
1272 
1273   // ===========================================================================
1274   // Code statistics. ==========================================================
1275   // ===========================================================================
1276 
1277   // Collect code (Code and BytecodeArray objects) statistics.
1278   void CollectCodeStatistics();
1279 
1280   // ===========================================================================
1281   // GC statistics. ============================================================
1282   // ===========================================================================
1283 
1284   // Returns the maximum amount of memory reserved for the heap.
1285   V8_EXPORT_PRIVATE size_t MaxReserved();
MaxSemiSpaceSize()1286   size_t MaxSemiSpaceSize() { return max_semi_space_size_; }
InitialSemiSpaceSize()1287   size_t InitialSemiSpaceSize() { return initial_semispace_size_; }
MaxOldGenerationSize()1288   size_t MaxOldGenerationSize() { return max_old_generation_size(); }
1289 
1290   // Limit on the max old generation size imposed by the underlying allocator.
1291   V8_EXPORT_PRIVATE static size_t AllocatorLimitOnMaxOldGenerationSize();
1292 
1293   V8_EXPORT_PRIVATE static size_t HeapSizeFromPhysicalMemory(
1294       uint64_t physical_memory);
1295   V8_EXPORT_PRIVATE static void GenerationSizesFromHeapSize(
1296       size_t heap_size, size_t* young_generation_size,
1297       size_t* old_generation_size);
1298   V8_EXPORT_PRIVATE static size_t YoungGenerationSizeFromOldGenerationSize(
1299       size_t old_generation_size);
1300   V8_EXPORT_PRIVATE static size_t YoungGenerationSizeFromSemiSpaceSize(
1301       size_t semi_space_size);
1302   V8_EXPORT_PRIVATE static size_t SemiSpaceSizeFromYoungGenerationSize(
1303       size_t young_generation_size);
1304   V8_EXPORT_PRIVATE static size_t MinYoungGenerationSize();
1305   V8_EXPORT_PRIVATE static size_t MinOldGenerationSize();
1306   V8_EXPORT_PRIVATE static size_t MaxOldGenerationSize(
1307       uint64_t physical_memory);
1308 
1309   // Returns the capacity of the heap in bytes w/o growing. Heap grows when
1310   // more spaces are needed until it reaches the limit.
1311   size_t Capacity();
1312 
1313   // Returns the capacity of the old generation.
1314   V8_EXPORT_PRIVATE size_t OldGenerationCapacity();
1315 
1316   // Returns the amount of memory currently held alive by the unmapper.
1317   size_t CommittedMemoryOfUnmapper();
1318 
1319   // Returns the amount of memory currently committed for the heap.
1320   size_t CommittedMemory();
1321 
1322   // Returns the amount of memory currently committed for the old space.
1323   size_t CommittedOldGenerationMemory();
1324 
1325   // Returns the amount of executable memory currently committed for the heap.
1326   size_t CommittedMemoryExecutable();
1327 
1328   // Returns the amount of phyical memory currently committed for the heap.
1329   size_t CommittedPhysicalMemory();
1330 
1331   // Returns the maximum amount of memory ever committed for the heap.
MaximumCommittedMemory()1332   size_t MaximumCommittedMemory() { return maximum_committed_; }
1333 
1334   // Updates the maximum committed memory for the heap. Should be called
1335   // whenever a space grows.
1336   void UpdateMaximumCommitted();
1337 
1338   // Returns the available bytes in space w/o growing.
1339   // Heap doesn't guarantee that it can allocate an object that requires
1340   // all available bytes. Check MaxHeapObjectSize() instead.
1341   size_t Available();
1342 
1343   // Returns size of all objects residing in the heap.
1344   V8_EXPORT_PRIVATE size_t SizeOfObjects();
1345 
1346   // Returns size of all global handles in the heap.
1347   V8_EXPORT_PRIVATE size_t TotalGlobalHandlesSize();
1348 
1349   // Returns size of all allocated/used global handles in the heap.
1350   V8_EXPORT_PRIVATE size_t UsedGlobalHandlesSize();
1351 
1352   void UpdateSurvivalStatistics(int start_new_space_size);
1353 
IncrementPromotedObjectsSize(size_t object_size)1354   inline void IncrementPromotedObjectsSize(size_t object_size) {
1355     promoted_objects_size_ += object_size;
1356   }
promoted_objects_size()1357   inline size_t promoted_objects_size() { return promoted_objects_size_; }
1358 
IncrementSemiSpaceCopiedObjectSize(size_t object_size)1359   inline void IncrementSemiSpaceCopiedObjectSize(size_t object_size) {
1360     semi_space_copied_object_size_ += object_size;
1361   }
semi_space_copied_object_size()1362   inline size_t semi_space_copied_object_size() {
1363     return semi_space_copied_object_size_;
1364   }
1365 
SurvivedYoungObjectSize()1366   inline size_t SurvivedYoungObjectSize() {
1367     return promoted_objects_size_ + semi_space_copied_object_size_;
1368   }
1369 
IncrementNodesDiedInNewSpace()1370   inline void IncrementNodesDiedInNewSpace() { nodes_died_in_new_space_++; }
1371 
IncrementNodesCopiedInNewSpace()1372   inline void IncrementNodesCopiedInNewSpace() { nodes_copied_in_new_space_++; }
1373 
IncrementNodesPromoted()1374   inline void IncrementNodesPromoted() { nodes_promoted_++; }
1375 
IncrementYoungSurvivorsCounter(size_t survived)1376   inline void IncrementYoungSurvivorsCounter(size_t survived) {
1377     survived_last_scavenge_ = survived;
1378     survived_since_last_expansion_ += survived;
1379   }
1380 
1381   void UpdateNewSpaceAllocationCounter();
1382 
1383   V8_EXPORT_PRIVATE size_t NewSpaceAllocationCounter();
1384 
1385   // This should be used only for testing.
set_new_space_allocation_counter(size_t new_value)1386   void set_new_space_allocation_counter(size_t new_value) {
1387     new_space_allocation_counter_ = new_value;
1388   }
1389 
UpdateOldGenerationAllocationCounter()1390   void UpdateOldGenerationAllocationCounter() {
1391     old_generation_allocation_counter_at_last_gc_ =
1392         OldGenerationAllocationCounter();
1393     old_generation_size_at_last_gc_ = 0;
1394   }
1395 
OldGenerationAllocationCounter()1396   size_t OldGenerationAllocationCounter() {
1397     return old_generation_allocation_counter_at_last_gc_ +
1398            PromotedSinceLastGC();
1399   }
1400 
1401   size_t EmbedderAllocationCounter() const;
1402 
1403   // This should be used only for testing.
set_old_generation_allocation_counter_at_last_gc(size_t new_value)1404   void set_old_generation_allocation_counter_at_last_gc(size_t new_value) {
1405     old_generation_allocation_counter_at_last_gc_ = new_value;
1406   }
1407 
PromotedSinceLastGC()1408   size_t PromotedSinceLastGC() {
1409     size_t old_generation_size = OldGenerationSizeOfObjects();
1410     return old_generation_size > old_generation_size_at_last_gc_
1411                ? old_generation_size - old_generation_size_at_last_gc_
1412                : 0;
1413   }
1414 
gc_count()1415   int gc_count() const { return gc_count_; }
1416 
is_current_gc_forced()1417   bool is_current_gc_forced() const { return is_current_gc_forced_; }
1418 
1419   // Returns whether the currently in-progress GC should avoid increasing the
1420   // ages on any objects that live for a set number of collections.
ShouldCurrentGCKeepAgesUnchanged()1421   bool ShouldCurrentGCKeepAgesUnchanged() const {
1422     return is_current_gc_forced_ || is_current_gc_for_heap_profiler_;
1423   }
1424 
1425   // Returns the size of objects residing in non-new spaces.
1426   // Excludes external memory held by those objects.
1427   V8_EXPORT_PRIVATE size_t OldGenerationSizeOfObjects();
1428 
1429   // Returns the size of objects held by the EmbedderHeapTracer.
1430   V8_EXPORT_PRIVATE size_t EmbedderSizeOfObjects() const;
1431 
1432   // Returns the global size of objects (embedder + V8 non-new spaces).
1433   V8_EXPORT_PRIVATE size_t GlobalSizeOfObjects();
1434 
1435   // We allow incremental marking to overshoot the V8 and global allocation
1436   // limit for performace reasons. If the overshoot is too large then we are
1437   // more eager to finalize incremental marking.
1438   bool AllocationLimitOvershotByLargeMargin();
1439 
1440   // Return the maximum size objects can be before having to allocate them as
1441   // large objects. This takes into account allocating in the code space for
1442   // which the size of the allocatable space per V8 page may depend on the OS
1443   // page size at runtime. You may use kMaxRegularHeapObjectSize as a constant
1444   // instead if you know the allocation isn't in the code spaces.
1445   inline V8_EXPORT_PRIVATE int MaxRegularHeapObjectSize(
1446       AllocationType allocation);
1447 
1448   // ===========================================================================
1449   // Prologue/epilogue callback methods.========================================
1450   // ===========================================================================
1451 
1452   void AddGCPrologueCallback(v8::Isolate::GCCallbackWithData callback,
1453                              GCType gc_type_filter, void* data);
1454   void RemoveGCPrologueCallback(v8::Isolate::GCCallbackWithData callback,
1455                                 void* data);
1456 
1457   void AddGCEpilogueCallback(v8::Isolate::GCCallbackWithData callback,
1458                              GCType gc_type_filter, void* data);
1459   void RemoveGCEpilogueCallback(v8::Isolate::GCCallbackWithData callback,
1460                                 void* data);
1461 
1462   void CallGCPrologueCallbacks(GCType gc_type, GCCallbackFlags flags);
1463   void CallGCEpilogueCallbacks(GCType gc_type, GCCallbackFlags flags);
1464 
1465   // ===========================================================================
1466   // Allocation methods. =======================================================
1467   // ===========================================================================
1468 
1469   // Creates a filler object and returns a heap object immediately after it.
1470   V8_EXPORT_PRIVATE HeapObject PrecedeWithFiller(HeapObject object,
1471                                                  int filler_size);
1472 
1473   // Creates a filler object if needed for alignment and returns a heap object
1474   // immediately after it. If any space is left after the returned object,
1475   // another filler object is created so the over allocated memory is iterable.
1476   V8_WARN_UNUSED_RESULT HeapObject
1477   AlignWithFiller(HeapObject object, int object_size, int allocation_size,
1478                   AllocationAlignment alignment);
1479 
1480   // Allocate an external backing store with the given allocation callback.
1481   // If the callback fails (indicated by a nullptr result) then this function
1482   // will re-try the allocation after performing GCs. This is useful for
1483   // external backing stores that may be retained by (unreachable) V8 objects
1484   // such as ArrayBuffers, ExternalStrings, etc.
1485   //
1486   // The function may also proactively trigger GCs even if the allocation
1487   // callback does not fail to keep the memory usage low.
1488   V8_EXPORT_PRIVATE void* AllocateExternalBackingStore(
1489       const std::function<void*(size_t)>& allocate, size_t byte_length);
1490 
1491   // ===========================================================================
1492   // Allocation site tracking. =================================================
1493   // ===========================================================================
1494 
1495   // Updates the AllocationSite of a given {object}. The entry (including the
1496   // count) is cached on the local pretenuring feedback.
1497   inline void UpdateAllocationSite(
1498       Map map, HeapObject object, PretenuringFeedbackMap* pretenuring_feedback);
1499 
1500   // Merges local pretenuring feedback into the global one. Note that this
1501   // method needs to be called after evacuation, as allocation sites may be
1502   // evacuated and this method resolves forward pointers accordingly.
1503   void MergeAllocationSitePretenuringFeedback(
1504       const PretenuringFeedbackMap& local_pretenuring_feedback);
1505 
1506   // Adds an allocation site to the list of sites to be pretenured during the
1507   // next collection. Added allocation sites are pretenured independent of
1508   // their feedback.
1509   V8_EXPORT_PRIVATE void PretenureAllocationSiteOnNextCollection(
1510       AllocationSite site);
1511 
1512   // ===========================================================================
1513   // Allocation tracking. ======================================================
1514   // ===========================================================================
1515 
1516   // Adds {new_space_observer} to new space and {observer} to any other space.
1517   void AddAllocationObserversToAllSpaces(
1518       AllocationObserver* observer, AllocationObserver* new_space_observer);
1519 
1520   // Removes {new_space_observer} from new space and {observer} from any other
1521   // space.
1522   void RemoveAllocationObserversFromAllSpaces(
1523       AllocationObserver* observer, AllocationObserver* new_space_observer);
1524 
1525   // Check if the given object was recently allocated and its fields may appear
1526   // as uninitialized to background threads.
1527   // This predicate may be invoked from a background thread.
1528   inline bool IsPendingAllocation(HeapObject object);
1529   inline bool IsPendingAllocation(Object object);
1530 
1531   // Notifies that all previously allocated objects are properly initialized
1532   // and ensures that IsPendingAllocation returns false for them. This function
1533   // may be invoked only on the main thread.
1534   V8_EXPORT_PRIVATE void PublishPendingAllocations();
1535 
1536   // ===========================================================================
1537   // Heap object allocation tracking. ==========================================
1538   // ===========================================================================
1539 
1540   V8_EXPORT_PRIVATE void AddHeapObjectAllocationTracker(
1541       HeapObjectAllocationTracker* tracker);
1542   V8_EXPORT_PRIVATE void RemoveHeapObjectAllocationTracker(
1543       HeapObjectAllocationTracker* tracker);
has_heap_object_allocation_tracker()1544   bool has_heap_object_allocation_tracker() const {
1545     return !allocation_trackers_.empty();
1546   }
1547 
1548   // ===========================================================================
1549   // Retaining path tracking. ==================================================
1550   // ===========================================================================
1551 
1552   // Adds the given object to the weak table of retaining path targets.
1553   // On each GC if the marker discovers the object, it will print the retaining
1554   // path. This requires --track-retaining-path flag.
1555   void AddRetainingPathTarget(Handle<HeapObject> object,
1556                               RetainingPathOption option);
1557 
1558   // ===========================================================================
1559   // Stack frame support. ======================================================
1560   // ===========================================================================
1561 
1562   // Returns the Code object for a given interior pointer.
1563   Code GcSafeFindCodeForInnerPointer(Address inner_pointer);
1564 
1565   // Returns true if {addr} is contained within {code} and false otherwise.
1566   // Mostly useful for debugging.
1567   bool GcSafeCodeContains(Code code, Address addr);
1568 
1569   // Casts a heap object to a code object and checks if the inner_pointer is
1570   // within the object.
1571   Code GcSafeCastToCode(HeapObject object, Address inner_pointer);
1572 
1573   // Returns the map of an object. Can be used during garbage collection, i.e.
1574   // it supports a forwarded map. Fails if the map is not the code map.
1575   Map GcSafeMapOfCodeSpaceObject(HeapObject object);
1576 
1577 // =============================================================================
1578 #ifdef VERIFY_HEAP
1579   // Verify the heap is in its normal state before or after a GC.
1580   V8_EXPORT_PRIVATE void Verify();
1581   // Verify the read-only heap after all read-only heap objects have been
1582   // created.
1583   void VerifyReadOnlyHeap();
1584   void VerifyRememberedSetFor(HeapObject object);
1585 #endif
1586 
1587 #ifdef V8_ENABLE_ALLOCATION_TIMEOUT
1588   void V8_EXPORT_PRIVATE set_allocation_timeout(int allocation_timeout);
1589 #endif  // V8_ENABLE_ALLOCATION_TIMEOUT
1590 
1591 #ifdef DEBUG
1592   void VerifyCountersAfterSweeping();
1593   void VerifyCountersBeforeConcurrentSweeping();
1594   void VerifyCommittedPhysicalMemory();
1595 
1596   void Print();
1597   void PrintHandles();
1598 
1599   // Report code statistics.
1600   void ReportCodeStatistics(const char* title);
1601 #endif  // DEBUG
GetRandomMmapAddr()1602   void* GetRandomMmapAddr() {
1603     void* result = v8::internal::GetRandomMmapAddr();
1604 #if V8_TARGET_ARCH_X64
1605 #if V8_OS_DARWIN
1606     // The Darwin kernel [as of macOS 10.12.5] does not clean up page
1607     // directory entries [PDE] created from mmap or mach_vm_allocate, even
1608     // after the region is destroyed. Using a virtual address space that is
1609     // too large causes a leak of about 1 wired [can never be paged out] page
1610     // per call to mmap(). The page is only reclaimed when the process is
1611     // killed. Confine the hint to a 32-bit section of the virtual address
1612     // space. See crbug.com/700928.
1613     uintptr_t offset = reinterpret_cast<uintptr_t>(result) & kMmapRegionMask;
1614     result = reinterpret_cast<void*>(mmap_region_base_ + offset);
1615 #endif  // V8_OS_DARWIN
1616 #endif  // V8_TARGET_ARCH_X64
1617     return result;
1618   }
1619 
1620   void RegisterCodeObject(Handle<Code> code);
1621 
1622   static const char* GarbageCollectionReasonToString(
1623       GarbageCollectionReason gc_reason);
1624 
1625   // Calculates the nof entries for the full sized number to string cache.
1626   inline int MaxNumberToStringCacheSize() const;
1627 
1628   static Isolate* GetIsolateFromWritableObject(HeapObject object);
1629 
1630   // Ensure that we have swept all spaces in such a way that we can iterate
1631   // over all objects.
1632   void MakeHeapIterable();
1633 
1634  private:
1635   class AllocationTrackerForDebugging;
1636 
1637   using ExternalStringTableUpdaterCallback = String (*)(Heap* heap,
1638                                                         FullObjectSlot pointer);
1639 
1640   // External strings table is a place where all external strings are
1641   // registered.  We need to keep track of such strings to properly
1642   // finalize them.
1643   class ExternalStringTable {
1644    public:
ExternalStringTable(Heap * heap)1645     explicit ExternalStringTable(Heap* heap) : heap_(heap) {}
1646     ExternalStringTable(const ExternalStringTable&) = delete;
1647     ExternalStringTable& operator=(const ExternalStringTable&) = delete;
1648 
1649     // Registers an external string.
1650     inline void AddString(String string);
1651     bool Contains(String string);
1652 
1653     void IterateAll(RootVisitor* v);
1654     void IterateYoung(RootVisitor* v);
1655     void PromoteYoung();
1656 
1657     // Restores internal invariant and gets rid of collected strings. Must be
1658     // called after each Iterate*() that modified the strings.
1659     void CleanUpAll();
1660     void CleanUpYoung();
1661 
1662     // Finalize all registered external strings and clear tables.
1663     void TearDown();
1664 
1665     void UpdateYoungReferences(
1666         Heap::ExternalStringTableUpdaterCallback updater_func);
1667     void UpdateReferences(
1668         Heap::ExternalStringTableUpdaterCallback updater_func);
1669 
1670    private:
1671     void Verify();
1672     void VerifyYoung();
1673 
1674     Heap* const heap_;
1675 
1676     // To speed up scavenge collections young string are kept separate from old
1677     // strings.
1678     std::vector<Object> young_strings_;
1679     std::vector<Object> old_strings_;
1680   };
1681 
1682   struct StringTypeTable {
1683     InstanceType type;
1684     int size;
1685     RootIndex index;
1686   };
1687 
1688   struct ConstantStringTable {
1689     const char* contents;
1690     RootIndex index;
1691   };
1692 
1693   struct StructTable {
1694     InstanceType type;
1695     int size;
1696     RootIndex index;
1697   };
1698 
1699   struct GCCallbackTuple {
GCCallbackTupleGCCallbackTuple1700     GCCallbackTuple(v8::Isolate::GCCallbackWithData callback, GCType gc_type,
1701                     void* data)
1702         : callback(callback), gc_type(gc_type), data(data) {}
1703 
1704     bool operator==(const GCCallbackTuple& other) const;
1705 
1706     v8::Isolate::GCCallbackWithData callback;
1707     GCType gc_type;
1708     void* data;
1709   };
1710 
1711   static const int kInitialEvalCacheSize = 64;
1712   static const int kInitialNumberStringCacheSize = 256;
1713 
1714   static const int kRememberedUnmappedPages = 128;
1715 
1716   static const StringTypeTable string_type_table[];
1717   static const ConstantStringTable constant_string_table[];
1718   static const StructTable struct_table[];
1719 
1720   static const int kYoungSurvivalRateHighThreshold = 90;
1721   static const int kYoungSurvivalRateAllowedDeviation = 15;
1722   static const int kOldSurvivalRateLowThreshold = 10;
1723 
1724   static const int kMaxMarkCompactsInIdleRound = 7;
1725 
1726   static const int kInitialFeedbackCapacity = 256;
1727 
1728   Heap();
1729   ~Heap();
1730 
1731   Heap(const Heap&) = delete;
1732   Heap& operator=(const Heap&) = delete;
1733 
IsRegularObjectAllocation(AllocationType allocation)1734   static bool IsRegularObjectAllocation(AllocationType allocation) {
1735     return AllocationType::kYoung == allocation ||
1736            AllocationType::kOld == allocation;
1737   }
1738 
DefaultGetExternallyAllocatedMemoryInBytesCallback()1739   static size_t DefaultGetExternallyAllocatedMemoryInBytesCallback() {
1740     return 0;
1741   }
1742 
1743 #define ROOT_ACCESSOR(type, name, CamelName) inline void set_##name(type value);
ROOT_LIST(ROOT_ACCESSOR)1744   ROOT_LIST(ROOT_ACCESSOR)
1745 #undef ROOT_ACCESSOR
1746 
1747   void set_current_gc_flags(int flags) { current_gc_flags_ = flags; }
1748 
ShouldReduceMemory()1749   inline bool ShouldReduceMemory() const {
1750     return (current_gc_flags_ & kReduceMemoryFootprintMask) != 0;
1751   }
1752 
1753   int NumberOfScavengeTasks();
1754 
1755   // Checks whether a global GC is necessary
1756   GarbageCollector SelectGarbageCollector(AllocationSpace space,
1757                                           const char** reason);
1758 
1759   // Free all LABs in the heap.
1760   void FreeLinearAllocationAreas();
1761 
1762   // Free all shared LABs.
1763   void FreeSharedLinearAllocationAreas();
1764 
1765   // Free all shared LABs of main thread.
1766   void FreeMainThreadSharedLinearAllocationAreas();
1767 
1768   // Performs garbage collection in a safepoint.
1769   // Returns the number of freed global handles.
1770   size_t PerformGarbageCollection(
1771       GarbageCollector collector, GarbageCollectionReason gc_reason,
1772       const char* collector_reason,
1773       const GCCallbackFlags gc_callback_flags = kNoGCCallbackFlags);
1774 
1775   // Performs garbage collection in the shared heap.
1776   void PerformSharedGarbageCollection(Isolate* initiator,
1777                                       GarbageCollectionReason gc_reason);
1778 
1779   inline void UpdateOldSpaceLimits();
1780 
1781   bool CreateInitialMaps();
1782   void CreateInternalAccessorInfoObjects();
1783   void CreateInitialObjects();
1784 
1785   // Commits from space if it is uncommitted.
1786   void EnsureFromSpaceIsCommitted();
1787 
1788   // Uncommit unused semi space.
1789   V8_EXPORT_PRIVATE bool UncommitFromSpace();
1790 
1791   // Fill in bogus values in from space
1792   void ZapFromSpace();
1793 
1794   // Zaps the memory of a code object.
1795   V8_EXPORT_PRIVATE void ZapCodeObject(Address start_address,
1796                                        int size_in_bytes);
1797 
1798   // Initialize a filler object to keep the ability to iterate over the heap
1799   // when introducing gaps within pages. If the memory after the object header
1800   // of the filler should be cleared, pass in kClearFreedMemory. The default is
1801   // kDontClearFreedMemory.
1802   V8_EXPORT_PRIVATE HeapObject
1803   CreateFillerObjectAt(Address addr, int size,
1804                        ClearFreedMemoryMode clear_memory_mode =
1805                            ClearFreedMemoryMode::kDontClearFreedMemory);
1806 
1807   // Range write barrier implementation.
1808   template <int kModeMask, typename TSlot>
1809   V8_INLINE void WriteBarrierForRangeImpl(MemoryChunk* source_page,
1810                                           HeapObject object, TSlot start_slot,
1811                                           TSlot end_slot);
1812 
1813   // Deopts all code that contains allocation instruction which are tenured or
1814   // not tenured. Moreover it clears the pretenuring allocation site statistics.
1815   void ResetAllAllocationSitesDependentCode(AllocationType allocation);
1816 
1817   // Evaluates local pretenuring for the old space and calls
1818   // ResetAllTenuredAllocationSitesDependentCode if too many objects died in
1819   // the old space.
1820   void EvaluateOldSpaceLocalPretenuring(uint64_t size_of_objects_before_gc);
1821 
1822   // Record statistics after garbage collection.
1823   void ReportStatisticsAfterGC();
1824 
1825   // Flush the number to string cache.
1826   void FlushNumberStringCache();
1827 
1828   void ConfigureInitialOldGenerationSize();
1829 
1830   double ComputeMutatorUtilization(const char* tag, double mutator_speed,
1831                                    double gc_speed);
1832   bool HasLowYoungGenerationAllocationRate();
1833   bool HasLowOldGenerationAllocationRate();
1834   bool HasLowEmbedderAllocationRate();
1835 
1836   void ReduceNewSpaceSize();
1837 
1838   GCIdleTimeHeapState ComputeHeapState();
1839 
1840   bool PerformIdleTimeAction(GCIdleTimeAction action,
1841                              GCIdleTimeHeapState heap_state,
1842                              double deadline_in_ms);
1843 
1844   void IdleNotificationEpilogue(GCIdleTimeAction action,
1845                                 GCIdleTimeHeapState heap_state, double start_ms,
1846                                 double deadline_in_ms);
1847 
1848   void PrintMaxMarkingLimitReached();
1849   void PrintMaxNewSpaceSizeReached();
1850 
1851   int NextStressMarkingLimit();
1852 
1853   void AddToRingBuffer(const char* string);
1854   void GetFromRingBuffer(char* buffer);
1855 
1856   void CompactRetainedMaps(WeakArrayList retained_maps);
1857 
1858   void CollectGarbageOnMemoryPressure();
1859 
1860   void EagerlyFreeExternalMemory();
1861 
1862   bool InvokeNearHeapLimitCallback();
1863 
1864   void ComputeFastPromotionMode();
1865 
1866   // Attempt to over-approximate the weak closure by marking object groups and
1867   // implicit references from global handles, but don't atomically complete
1868   // marking. If we continue to mark incrementally, we might have marked
1869   // objects that die later.
1870   void FinalizeIncrementalMarkingIncrementally(
1871       GarbageCollectionReason gc_reason);
1872 
1873   void InvokeIncrementalMarkingPrologueCallbacks();
1874   void InvokeIncrementalMarkingEpilogueCallbacks();
1875 
1876   // ===========================================================================
1877   // Pretenuring. ==============================================================
1878   // ===========================================================================
1879 
1880   // Pretenuring decisions are made based on feedback collected during new space
1881   // evacuation. Note that between feedback collection and calling this method
1882   // object in old space must not move.
1883   void ProcessPretenuringFeedback();
1884 
1885   // Removes an entry from the global pretenuring storage.
1886   void RemoveAllocationSitePretenuringFeedback(AllocationSite site);
1887 
1888   // ===========================================================================
1889   // Actual GC. ================================================================
1890   // ===========================================================================
1891 
1892   // Code that should be run before and after each GC.  Includes some
1893   // reporting/verification activities when compiled with DEBUG set.
1894   void GarbageCollectionPrologue(GarbageCollectionReason gc_reason,
1895                                  const v8::GCCallbackFlags gc_callback_flags);
1896   void GarbageCollectionPrologueInSafepoint();
1897   void GarbageCollectionEpilogue(GarbageCollector collector);
1898   void GarbageCollectionEpilogueInSafepoint(GarbageCollector collector);
1899 
1900   // Performs a major collection in the whole heap.
1901   void MarkCompact();
1902   // Performs a minor collection of just the young generation.
1903   void MinorMarkCompact();
1904 
1905   // Code to be run before and after mark-compact.
1906   void MarkCompactPrologue();
1907   void MarkCompactEpilogue();
1908 
1909   // Performs a minor collection in new generation.
1910   void Scavenge();
1911   void EvacuateYoungGeneration();
1912 
1913   void UpdateYoungReferencesInExternalStringTable(
1914       ExternalStringTableUpdaterCallback updater_func);
1915 
1916   void UpdateReferencesInExternalStringTable(
1917       ExternalStringTableUpdaterCallback updater_func);
1918 
1919   void ProcessAllWeakReferences(WeakObjectRetainer* retainer);
1920   void ProcessYoungWeakReferences(WeakObjectRetainer* retainer);
1921   void ProcessNativeContexts(WeakObjectRetainer* retainer);
1922   void ProcessAllocationSites(WeakObjectRetainer* retainer);
1923   void ProcessDirtyJSFinalizationRegistries(WeakObjectRetainer* retainer);
1924   void ProcessWeakListRoots(WeakObjectRetainer* retainer);
1925 
1926   // ===========================================================================
1927   // GC statistics. ============================================================
1928   // ===========================================================================
1929 
OldGenerationSpaceAvailable()1930   inline size_t OldGenerationSpaceAvailable() {
1931     uint64_t bytes = OldGenerationSizeOfObjects() +
1932                      AllocatedExternalMemorySinceMarkCompact();
1933 
1934     if (old_generation_allocation_limit() <= bytes) return 0;
1935     return old_generation_allocation_limit() - static_cast<size_t>(bytes);
1936   }
1937 
1938   void UpdateTotalGCTime(double duration);
1939 
MaximumSizeScavenge()1940   bool MaximumSizeScavenge() { return maximum_size_scavenges_ > 0; }
1941 
1942   bool IsIneffectiveMarkCompact(size_t old_generation_size,
1943                                 double mutator_utilization);
1944   void CheckIneffectiveMarkCompact(size_t old_generation_size,
1945                                    double mutator_utilization);
1946 
1947   inline void IncrementExternalBackingStoreBytes(ExternalBackingStoreType type,
1948                                                  size_t amount);
1949 
1950   inline void DecrementExternalBackingStoreBytes(ExternalBackingStoreType type,
1951                                                  size_t amount);
1952 
1953   // ===========================================================================
1954   // Growing strategy. =========================================================
1955   // ===========================================================================
1956 
memory_reducer()1957   MemoryReducer* memory_reducer() { return memory_reducer_.get(); }
1958 
1959   // For some webpages RAIL mode does not switch from PERFORMANCE_LOAD.
1960   // This constant limits the effect of load RAIL mode on GC.
1961   // The value is arbitrary and chosen as the largest load time observed in
1962   // v8 browsing benchmarks.
1963   static const int kMaxLoadTimeMs = 7000;
1964 
1965   bool ShouldOptimizeForLoadTime();
1966 
old_generation_allocation_limit()1967   size_t old_generation_allocation_limit() const {
1968     return old_generation_allocation_limit_.load(std::memory_order_relaxed);
1969   }
1970 
set_old_generation_allocation_limit(size_t newlimit)1971   void set_old_generation_allocation_limit(size_t newlimit) {
1972     old_generation_allocation_limit_.store(newlimit, std::memory_order_relaxed);
1973   }
1974 
global_allocation_limit()1975   size_t global_allocation_limit() const { return global_allocation_limit_; }
1976 
max_old_generation_size()1977   size_t max_old_generation_size() {
1978     return max_old_generation_size_.load(std::memory_order_relaxed);
1979   }
1980 
set_max_old_generation_size(size_t value)1981   void set_max_old_generation_size(size_t value) {
1982     max_old_generation_size_.store(value, std::memory_order_relaxed);
1983   }
1984 
always_allocate()1985   bool always_allocate() { return always_allocate_scope_count_ != 0; }
1986 
1987   V8_EXPORT_PRIVATE bool CanExpandOldGeneration(size_t size);
1988   V8_EXPORT_PRIVATE bool CanExpandOldGenerationBackground(LocalHeap* local_heap,
1989                                                           size_t size);
1990   V8_EXPORT_PRIVATE bool CanPromoteYoungAndExpandOldGeneration(size_t size);
1991 
1992   bool ShouldExpandOldGenerationOnSlowAllocation(
1993       LocalHeap* local_heap = nullptr);
1994   bool IsRetryOfFailedAllocation(LocalHeap* local_heap);
1995   bool IsMainThreadParked(LocalHeap* local_heap);
1996 
1997   HeapGrowingMode CurrentHeapGrowingMode();
1998 
1999   double PercentToOldGenerationLimit();
2000   double PercentToGlobalMemoryLimit();
2001   enum class IncrementalMarkingLimit {
2002     kNoLimit,
2003     kSoftLimit,
2004     kHardLimit,
2005     kFallbackForEmbedderLimit
2006   };
2007   IncrementalMarkingLimit IncrementalMarkingLimitReached();
2008 
2009   bool ShouldStressCompaction() const;
2010 
UseGlobalMemoryScheduling()2011   bool UseGlobalMemoryScheduling() const {
2012     return FLAG_global_gc_scheduling && local_embedder_heap_tracer();
2013   }
2014 
2015   base::Optional<size_t> GlobalMemoryAvailable();
2016 
2017   void RecomputeLimits(GarbageCollector collector);
2018 
2019   // ===========================================================================
2020   // Idle notification. ========================================================
2021   // ===========================================================================
2022 
2023   bool RecentIdleNotificationHappened();
2024 
2025   // ===========================================================================
2026   // GC Tasks. =================================================================
2027   // ===========================================================================
2028 
2029   void ScheduleScavengeTaskIfNeeded();
2030 
2031   // ===========================================================================
2032   // Allocation methods. =======================================================
2033   // ===========================================================================
2034 
allocator()2035   HeapAllocator* allocator() { return &heap_allocator_; }
2036 
2037   // Allocates a JS Map in the heap.
2038   V8_WARN_UNUSED_RESULT AllocationResult
2039   AllocateMap(InstanceType instance_type, int instance_size,
2040               ElementsKind elements_kind = TERMINAL_FAST_ELEMENTS_KIND,
2041               int inobject_properties = 0);
2042 
2043   // Allocate an uninitialized object.  The memory is non-executable if the
2044   // hardware and OS allow.  This is the single choke-point for allocations
2045   // performed by the runtime and should not be bypassed (to extend this to
2046   // inlined allocations, use the Heap::DisableInlineAllocation() support).
2047   V8_WARN_UNUSED_RESULT V8_INLINE AllocationResult
2048   AllocateRaw(int size_in_bytes, AllocationType allocation,
2049               AllocationOrigin origin = AllocationOrigin::kRuntime,
2050               AllocationAlignment alignment = kTaggedAligned);
2051 
2052   // This method will try to allocate objects quickly (AllocationType::kYoung)
2053   // otherwise it falls back to a slower path indicated by the mode.
2054   enum AllocationRetryMode { kLightRetry, kRetryOrFail };
2055   template <AllocationRetryMode mode>
2056   V8_WARN_UNUSED_RESULT V8_INLINE HeapObject
2057   AllocateRawWith(int size, AllocationType allocation,
2058                   AllocationOrigin origin = AllocationOrigin::kRuntime,
2059                   AllocationAlignment alignment = kTaggedAligned);
2060 
2061   // Call AllocateRawWith with kRetryOrFail. Matches the method in LocalHeap.
2062   V8_WARN_UNUSED_RESULT inline Address AllocateRawOrFail(
2063       int size, AllocationType allocation,
2064       AllocationOrigin origin = AllocationOrigin::kRuntime,
2065       AllocationAlignment alignment = kTaggedAligned);
2066 
2067   // Allocates a heap object based on the map.
2068   V8_WARN_UNUSED_RESULT AllocationResult Allocate(Handle<Map> map,
2069                                                   AllocationType allocation);
2070 
2071   // Allocates a partial map for bootstrapping.
2072   V8_WARN_UNUSED_RESULT AllocationResult
2073   AllocatePartialMap(InstanceType instance_type, int instance_size);
2074 
2075   void FinalizePartialMap(Map map);
2076 
set_force_oom(bool value)2077   void set_force_oom(bool value) { force_oom_ = value; }
set_force_gc_on_next_allocation()2078   void set_force_gc_on_next_allocation() {
2079     force_gc_on_next_allocation_ = true;
2080   }
2081 
2082   // Helper for IsPendingAllocation.
2083   inline bool IsPendingAllocationInternal(HeapObject object);
2084 
2085   // ===========================================================================
2086   // Retaining path tracing ====================================================
2087   // ===========================================================================
2088 
2089   void AddRetainer(HeapObject retainer, HeapObject object);
2090   void AddEphemeronRetainer(HeapObject retainer, HeapObject object);
2091   void AddRetainingRoot(Root root, HeapObject object);
2092   // Returns true if the given object is a target of retaining path tracking.
2093   // Stores the option corresponding to the object in the provided *option.
2094   bool IsRetainingPathTarget(HeapObject object, RetainingPathOption* option);
2095   void PrintRetainingPath(HeapObject object, RetainingPathOption option);
2096   void UpdateRetainersAfterScavenge();
2097 
2098 #ifdef DEBUG
2099   V8_EXPORT_PRIVATE void IncrementObjectCounters();
2100 #endif  // DEBUG
2101 
2102   std::vector<Handle<NativeContext>> FindAllNativeContexts();
2103   std::vector<WeakArrayList> FindAllRetainedMaps();
memory_measurement()2104   MemoryMeasurement* memory_measurement() { return memory_measurement_.get(); }
2105 
allocation_type_for_in_place_internalizable_strings()2106   AllocationType allocation_type_for_in_place_internalizable_strings() const {
2107     return allocation_type_for_in_place_internalizable_strings_;
2108   }
2109 
2110   bool IsStressingScavenge();
2111 
2112   ExternalMemoryAccounting external_memory_;
2113 
2114   // This can be calculated directly from a pointer to the heap; however, it is
2115   // more expedient to get at the isolate directly from within Heap methods.
2116   Isolate* isolate_ = nullptr;
2117 
2118   HeapAllocator heap_allocator_;
2119 
2120   // These limits are initialized in Heap::ConfigureHeap based on the resource
2121   // constraints and flags.
2122   size_t code_range_size_ = 0;
2123   size_t max_semi_space_size_ = 0;
2124   size_t initial_semispace_size_ = 0;
2125   // Full garbage collections can be skipped if the old generation size
2126   // is below this threshold.
2127   size_t min_old_generation_size_ = 0;
2128   // If the old generation size exceeds this limit, then V8 will
2129   // crash with out-of-memory error.
2130   std::atomic<size_t> max_old_generation_size_{0};
2131   // TODO(mlippautz): Clarify whether this should take some embedder
2132   // configurable limit into account.
2133   size_t min_global_memory_size_ = 0;
2134   size_t max_global_memory_size_ = 0;
2135 
2136   size_t initial_max_old_generation_size_ = 0;
2137   size_t initial_max_old_generation_size_threshold_ = 0;
2138   size_t initial_old_generation_size_ = 0;
2139   bool old_generation_size_configured_ = false;
2140   size_t maximum_committed_ = 0;
2141   size_t old_generation_capacity_after_bootstrap_ = 0;
2142 
2143   // Backing store bytes (array buffers and external strings).
2144   // Use uint64_t counter since the counter could overflow the 32-bit range
2145   // temporarily on 32-bit.
2146   std::atomic<uint64_t> backing_store_bytes_{0};
2147 
2148   // For keeping track of how much data has survived
2149   // scavenge since last new space expansion.
2150   size_t survived_since_last_expansion_ = 0;
2151 
2152   // ... and since the last scavenge.
2153   size_t survived_last_scavenge_ = 0;
2154 
2155   // This is not the depth of nested AlwaysAllocateScope's but rather a single
2156   // count, as scopes can be acquired from multiple tasks (read: threads).
2157   std::atomic<size_t> always_allocate_scope_count_{0};
2158 
2159   // Stores the memory pressure level that set by MemoryPressureNotification
2160   // and reset by a mark-compact garbage collection.
2161   std::atomic<MemoryPressureLevel> memory_pressure_level_;
2162 
2163   std::vector<std::pair<v8::NearHeapLimitCallback, void*>>
2164       near_heap_limit_callbacks_;
2165 
2166   // For keeping track of context disposals.
2167   int contexts_disposed_ = 0;
2168 
2169   NewSpace* new_space_ = nullptr;
2170   OldSpace* old_space_ = nullptr;
2171   CodeSpace* code_space_ = nullptr;
2172   MapSpace* map_space_ = nullptr;
2173   OldLargeObjectSpace* lo_space_ = nullptr;
2174   CodeLargeObjectSpace* code_lo_space_ = nullptr;
2175   NewLargeObjectSpace* new_lo_space_ = nullptr;
2176   ReadOnlySpace* read_only_space_ = nullptr;
2177 
2178   OldSpace* shared_old_space_ = nullptr;
2179   MapSpace* shared_map_space_ = nullptr;
2180 
2181   std::unique_ptr<ConcurrentAllocator> shared_old_allocator_;
2182   std::unique_ptr<ConcurrentAllocator> shared_map_allocator_;
2183 
2184   // Map from the space id to the space.
2185   Space* space_[LAST_SPACE + 1];
2186 
2187   LocalHeap* main_thread_local_heap_ = nullptr;
2188 
2189   // List for tracking ArrayBufferExtensions
2190   ArrayBufferExtension* old_array_buffer_extensions_ = nullptr;
2191   ArrayBufferExtension* young_array_buffer_extensions_ = nullptr;
2192 
2193   // Determines whether code space is write-protected. This is essentially a
2194   // race-free copy of the {FLAG_write_protect_code_memory} flag.
2195   bool write_protect_code_memory_ = false;
2196 
2197   // Holds the number of open CodeSpaceMemoryModificationScopes.
2198   uintptr_t code_space_memory_modification_scope_depth_ = 0;
2199 
2200   // Holds the number of open CodePageCollectionMemoryModificationScopes.
2201   std::atomic<uintptr_t> code_page_collection_memory_modification_scope_depth_{
2202       0};
2203 
2204   std::atomic<HeapState> gc_state_{NOT_IN_GC};
2205 
2206   int gc_post_processing_depth_ = 0;
2207 
2208   // Returns the amount of external memory registered since last global gc.
2209   V8_EXPORT_PRIVATE uint64_t AllocatedExternalMemorySinceMarkCompact();
2210 
2211   // Starts marking when stress_marking_percentage_% of the marking start limit
2212   // is reached.
2213   std::atomic<int> stress_marking_percentage_{0};
2214 
2215   // Observer that causes more frequent checks for reached incremental
2216   // marking limit.
2217   AllocationObserver* stress_marking_observer_ = nullptr;
2218 
2219   // Observer that can cause early scavenge start.
2220   StressScavengeObserver* stress_scavenge_observer_ = nullptr;
2221 
2222   // The maximum percent of the marking limit reached wihout causing marking.
2223   // This is tracked when specyfing --fuzzer-gc-analysis.
2224   double max_marking_limit_reached_ = 0.0;
2225 
2226   // How many mark-sweep collections happened.
2227   unsigned int ms_count_ = 0;
2228 
2229   // How many gc happened.
2230   unsigned int gc_count_ = 0;
2231 
2232   // The number of Mark-Compact garbage collections that are considered as
2233   // ineffective. See IsIneffectiveMarkCompact() predicate.
2234   int consecutive_ineffective_mark_compacts_ = 0;
2235 
2236   static const uintptr_t kMmapRegionMask = 0xFFFFFFFFu;
2237   uintptr_t mmap_region_base_ = 0;
2238 
2239   // For post mortem debugging.
2240   int remembered_unmapped_pages_index_ = 0;
2241   Address remembered_unmapped_pages_[kRememberedUnmappedPages];
2242 
2243   // Limit that triggers a global GC on the next (normally caused) GC.  This
2244   // is checked when we have already decided to do a GC to help determine
2245   // which collector to invoke, before expanding a paged space in the old
2246   // generation and on every allocation in large object space.
2247   std::atomic<size_t> old_generation_allocation_limit_{0};
2248   size_t global_allocation_limit_ = 0;
2249 
2250   // Weak list heads, threaded through the objects.
2251   // List heads are initialized lazily and contain the undefined_value at start.
2252   // {native_contexts_list_} is an Address instead of an Object to allow the use
2253   // of atomic accessors.
2254   std::atomic<Address> native_contexts_list_;
2255   Object allocation_sites_list_;
2256   Object dirty_js_finalization_registries_list_;
2257   // Weak list tails.
2258   Object dirty_js_finalization_registries_list_tail_;
2259 
2260   std::vector<GCCallbackTuple> gc_epilogue_callbacks_;
2261   std::vector<GCCallbackTuple> gc_prologue_callbacks_;
2262 
2263   GetExternallyAllocatedMemoryInBytesCallback external_memory_callback_;
2264 
2265   int deferred_counters_[v8::Isolate::kUseCounterFeatureCount];
2266 
2267   size_t promoted_objects_size_ = 0;
2268   double promotion_ratio_ = 0.0;
2269   double promotion_rate_ = 0.0;
2270   size_t semi_space_copied_object_size_ = 0;
2271   size_t previous_semi_space_copied_object_size_ = 0;
2272   double semi_space_copied_rate_ = 0.0;
2273   int nodes_died_in_new_space_ = 0;
2274   int nodes_copied_in_new_space_ = 0;
2275   int nodes_promoted_ = 0;
2276 
2277   // This is the pretenuring trigger for allocation sites that are in maybe
2278   // tenure state. When we switched to the maximum new space size we deoptimize
2279   // the code that belongs to the allocation site and derive the lifetime
2280   // of the allocation site.
2281   unsigned int maximum_size_scavenges_ = 0;
2282 
2283   // Total time spent in GC.
2284   double total_gc_time_ms_ = 0.0;
2285 
2286   // Last time an idle notification happened.
2287   double last_idle_notification_time_ = 0.0;
2288 
2289   // Last time a garbage collection happened.
2290   double last_gc_time_ = 0.0;
2291 
2292   std::unique_ptr<GCTracer> tracer_;
2293   std::unique_ptr<MarkCompactCollector> mark_compact_collector_;
2294   std::unique_ptr<MinorMarkCompactCollector> minor_mark_compact_collector_;
2295   std::unique_ptr<ScavengerCollector> scavenger_collector_;
2296   std::unique_ptr<ArrayBufferSweeper> array_buffer_sweeper_;
2297 
2298   std::unique_ptr<MemoryAllocator> memory_allocator_;
2299   std::unique_ptr<IncrementalMarking> incremental_marking_;
2300   std::unique_ptr<ConcurrentMarking> concurrent_marking_;
2301   std::unique_ptr<GCIdleTimeHandler> gc_idle_time_handler_;
2302   std::unique_ptr<MemoryMeasurement> memory_measurement_;
2303   std::unique_ptr<MemoryReducer> memory_reducer_;
2304   std::unique_ptr<ObjectStats> live_object_stats_;
2305   std::unique_ptr<ObjectStats> dead_object_stats_;
2306   std::unique_ptr<ScavengeJob> scavenge_job_;
2307   std::unique_ptr<AllocationObserver> scavenge_task_observer_;
2308   std::unique_ptr<AllocationObserver> stress_concurrent_allocation_observer_;
2309   std::unique_ptr<LocalEmbedderHeapTracer> local_embedder_heap_tracer_;
2310   std::unique_ptr<MarkingBarrier> marking_barrier_;
2311   std::unique_ptr<AllocationTrackerForDebugging>
2312       allocation_tracker_for_debugging_;
2313 
2314   // This object controls virtual space reserved for code on the V8 heap. This
2315   // is only valid for 64-bit architectures where kRequiresCodeRange.
2316   //
2317   // Owned by the heap when !V8_COMPRESS_POINTERS_IN_SHARED_CAGE, otherwise is
2318   // process-wide.
2319   std::shared_ptr<CodeRange> code_range_;
2320 
2321   // The embedder owns the C++ heap.
2322   v8::CppHeap* cpp_heap_ = nullptr;
2323 
2324   EmbedderRootsHandler* embedder_roots_handler_ = nullptr;
2325 
2326   StrongRootsEntry* strong_roots_head_ = nullptr;
2327   base::Mutex strong_roots_mutex_;
2328 
2329   bool need_to_remove_stress_concurrent_allocation_observer_ = false;
2330 
2331   // This counter is increased before each GC and never reset.
2332   // To account for the bytes allocated since the last GC, use the
2333   // NewSpaceAllocationCounter() function.
2334   size_t new_space_allocation_counter_ = 0;
2335 
2336   // This counter is increased before each GC and never reset. To
2337   // account for the bytes allocated since the last GC, use the
2338   // OldGenerationAllocationCounter() function.
2339   size_t old_generation_allocation_counter_at_last_gc_ = 0;
2340 
2341   // The size of objects in old generation after the last MarkCompact GC.
2342   size_t old_generation_size_at_last_gc_{0};
2343 
2344   // The size of global memory after the last MarkCompact GC.
2345   size_t global_memory_at_last_gc_ = 0;
2346 
2347   // The feedback storage is used to store allocation sites (keys) and how often
2348   // they have been visited (values) by finding a memento behind an object. The
2349   // storage is only alive temporary during a GC. The invariant is that all
2350   // pointers in this map are already fixed, i.e., they do not point to
2351   // forwarding pointers.
2352   PretenuringFeedbackMap global_pretenuring_feedback_;
2353 
2354   std::unique_ptr<GlobalHandleVector<AllocationSite>>
2355       allocation_sites_to_pretenure_;
2356 
2357   char trace_ring_buffer_[kTraceRingBufferSize];
2358 
2359   // Used as boolean.
2360   uint8_t is_marking_flag_ = 0;
2361 
2362   // If it's not full then the data is from 0 to ring_buffer_end_.  If it's
2363   // full then the data is from ring_buffer_end_ to the end of the buffer and
2364   // from 0 to ring_buffer_end_.
2365   bool ring_buffer_full_ = false;
2366   size_t ring_buffer_end_ = 0;
2367 
2368   // Flag is set when the heap has been configured.  The heap can be repeatedly
2369   // configured through the API until it is set up.
2370   bool configured_ = false;
2371 
2372   // Currently set GC flags that are respected by all GC components.
2373   int current_gc_flags_ = Heap::kNoGCFlags;
2374 
2375   // Currently set GC callback flags that are used to pass information between
2376   // the embedder and V8's GC.
2377   GCCallbackFlags current_gc_callback_flags_ =
2378       GCCallbackFlags::kNoGCCallbackFlags;
2379 
2380   std::unique_ptr<IsolateSafepoint> safepoint_;
2381 
2382   bool is_current_gc_forced_ = false;
2383   bool is_current_gc_for_heap_profiler_ = false;
2384 
2385   ExternalStringTable external_string_table_;
2386 
2387   const AllocationType allocation_type_for_in_place_internalizable_strings_;
2388 
2389   base::Mutex relocation_mutex_;
2390 
2391   std::unique_ptr<CollectionBarrier> collection_barrier_;
2392 
2393   int ignore_local_gc_requests_depth_ = 0;
2394 
2395   int gc_callbacks_depth_ = 0;
2396 
2397   bool deserialization_complete_ = false;
2398 
2399   int max_regular_code_object_size_ = 0;
2400 
2401   bool fast_promotion_mode_ = false;
2402 
2403   // Used for testing purposes.
2404   bool force_oom_ = false;
2405   bool force_gc_on_next_allocation_ = false;
2406   bool delay_sweeper_tasks_for_testing_ = false;
2407 
2408   HeapObject pending_layout_change_object_;
2409 
2410   base::Mutex unprotected_memory_chunks_mutex_;
2411   std::unordered_set<MemoryChunk*> unprotected_memory_chunks_;
2412 
2413   std::unordered_map<HeapObject, HeapObject, Object::Hasher> retainer_;
2414   std::unordered_map<HeapObject, Root, Object::Hasher> retaining_root_;
2415   // If an object is retained by an ephemeron, then the retaining key of the
2416   // ephemeron is stored in this map.
2417   std::unordered_map<HeapObject, HeapObject, Object::Hasher>
2418       ephemeron_retainer_;
2419   // For each index inthe retaining_path_targets_ array this map
2420   // stores the option of the corresponding target.
2421   std::unordered_map<int, RetainingPathOption> retaining_path_target_option_;
2422 
2423   std::vector<HeapObjectAllocationTracker*> allocation_trackers_;
2424 
2425   bool is_finalization_registry_cleanup_task_posted_ = false;
2426 
2427   std::unique_ptr<third_party_heap::Heap> tp_heap_;
2428 
2429   // Classes in "heap" can be friends.
2430   friend class AlwaysAllocateScope;
2431   friend class ArrayBufferCollector;
2432   friend class ArrayBufferSweeper;
2433   friend class ConcurrentMarking;
2434   friend class EvacuateVisitorBase;
2435   friend class GCCallbacksScope;
2436   friend class GCTracer;
2437   friend class HeapAllocator;
2438   friend class HeapObjectIterator;
2439   friend class ScavengeTaskObserver;
2440   friend class IgnoreLocalGCRequests;
2441   friend class IncrementalMarking;
2442   friend class IncrementalMarkingRootMarkingVisitor;
2443   friend class IncrementalMarkingJob;
2444   friend class LargeObjectSpace;
2445   friend class LocalHeap;
2446   friend class MarkingBarrier;
2447   friend class OldLargeObjectSpace;
2448   friend class OptionalAlwaysAllocateScope;
2449   template <typename ConcreteVisitor, typename MarkingState>
2450   friend class MarkingVisitorBase;
2451   friend class MarkCompactCollector;
2452   friend class MarkCompactCollectorBase;
2453   friend class MinorMarkCompactCollector;
2454   friend class NewLargeObjectSpace;
2455   friend class NewSpace;
2456   friend class ObjectStatsCollector;
2457   friend class Page;
2458   friend class PagedSpace;
2459   friend class ReadOnlyRoots;
2460   friend class Scavenger;
2461   friend class ScavengerCollector;
2462   friend class StressConcurrentAllocationObserver;
2463   friend class Space;
2464   friend class Sweeper;
2465   friend class UnifiedHeapMarkingState;
2466   friend class heap::TestMemoryAllocatorScope;
2467   friend class third_party_heap::Heap;
2468   friend class third_party_heap::Impl;
2469 
2470   // The allocator interface.
2471   friend class Factory;
2472   friend class LocalFactory;
2473   template <typename IsolateT>
2474   friend class Deserializer;
2475 
2476   // The Isolate constructs us.
2477   friend class Isolate;
2478 
2479   // Used in cctest.
2480   friend class heap::HeapTester;
2481 };
2482 
2483 class HeapStats {
2484  public:
2485   static const int kStartMarker = 0xDECADE00;
2486   static const int kEndMarker = 0xDECADE01;
2487 
2488   intptr_t* start_marker;                  //  0
2489   size_t* ro_space_size;                   //  1
2490   size_t* ro_space_capacity;               //  2
2491   size_t* new_space_size;                  //  3
2492   size_t* new_space_capacity;              //  4
2493   size_t* old_space_size;                  //  5
2494   size_t* old_space_capacity;              //  6
2495   size_t* code_space_size;                 //  7
2496   size_t* code_space_capacity;             //  8
2497   size_t* map_space_size;                  //  9
2498   size_t* map_space_capacity;              // 10
2499   size_t* lo_space_size;                   // 11
2500   size_t* code_lo_space_size;              // 12
2501   size_t* global_handle_count;             // 13
2502   size_t* weak_global_handle_count;        // 14
2503   size_t* pending_global_handle_count;     // 15
2504   size_t* near_death_global_handle_count;  // 16
2505   size_t* free_global_handle_count;        // 17
2506   size_t* memory_allocator_size;           // 18
2507   size_t* memory_allocator_capacity;       // 19
2508   size_t* malloced_memory;                 // 20
2509   size_t* malloced_peak_memory;            // 21
2510   size_t* objects_per_type;                // 22
2511   size_t* size_per_type;                   // 23
2512   int* os_error;                           // 24
2513   char* last_few_messages;                 // 25
2514   char* js_stacktrace;                     // 26
2515   intptr_t* end_marker;                    // 27
2516 };
2517 
2518 // Disables GC for all allocations. It should not be used
2519 // outside heap, deserializer, and isolate bootstrap.
2520 // Use AlwaysAllocateScopeForTesting in tests.
2521 class V8_NODISCARD AlwaysAllocateScope {
2522  public:
2523   inline ~AlwaysAllocateScope();
2524 
2525  private:
2526   friend class AlwaysAllocateScopeForTesting;
2527   friend class Evacuator;
2528   friend class Heap;
2529   friend class HeapAllocator;
2530   friend class Isolate;
2531 
2532   explicit inline AlwaysAllocateScope(Heap* heap);
2533   Heap* heap_;
2534 };
2535 
2536 // Like AlwaysAllocateScope if the heap argument to the constructor is
2537 // non-null. No-op otherwise.
2538 //
2539 // This class exists because AlwaysAllocateScope doesn't compose with
2540 // base::Optional, since supporting that composition requires making
2541 // base::Optional a friend class, defeating the purpose of hiding its
2542 // constructor.
2543 class V8_NODISCARD OptionalAlwaysAllocateScope {
2544  public:
2545   inline ~OptionalAlwaysAllocateScope();
2546 
2547  private:
2548   friend class Heap;
2549 
2550   explicit inline OptionalAlwaysAllocateScope(Heap* heap);
2551   Heap* heap_;
2552 };
2553 
2554 class V8_NODISCARD AlwaysAllocateScopeForTesting {
2555  public:
2556   explicit inline AlwaysAllocateScopeForTesting(Heap* heap);
2557 
2558  private:
2559   AlwaysAllocateScope scope_;
2560 };
2561 
2562 // The CodeSpaceMemoryModificationScope can only be used by the main thread.
2563 class V8_NODISCARD CodeSpaceMemoryModificationScope {
2564  public:
2565   explicit inline CodeSpaceMemoryModificationScope(Heap* heap);
2566   inline ~CodeSpaceMemoryModificationScope();
2567 
2568  private:
2569   Heap* heap_;
2570 };
2571 
2572 // The CodePageCollectionMemoryModificationScope can be used by any thread. It
2573 // will not be enabled if a CodeSpaceMemoryModificationScope is already active.
2574 class V8_NODISCARD CodePageCollectionMemoryModificationScope {
2575  public:
2576   explicit inline CodePageCollectionMemoryModificationScope(Heap* heap);
2577   inline ~CodePageCollectionMemoryModificationScope();
2578 
2579  private:
2580   Heap* heap_;
2581 };
2582 
2583 // The CodePageMemoryModificationScope does not check if tansitions to
2584 // writeable and back to executable are actually allowed, i.e. the MemoryChunk
2585 // was registered to be executable. It can be used by concurrent threads.
2586 class V8_NODISCARD CodePageMemoryModificationScope {
2587  public:
2588   explicit inline CodePageMemoryModificationScope(BasicMemoryChunk* chunk);
2589   explicit inline CodePageMemoryModificationScope(Code object);
2590   inline ~CodePageMemoryModificationScope();
2591 
2592  private:
2593   BasicMemoryChunk* chunk_;
2594   bool scope_active_;
2595 
2596   // Disallow any GCs inside this scope, as a relocation of the underlying
2597   // object would change the {MemoryChunk} that this scope targets.
2598   DISALLOW_GARBAGE_COLLECTION(no_heap_allocation_)
2599 };
2600 
2601 class V8_NODISCARD IgnoreLocalGCRequests {
2602  public:
2603   explicit inline IgnoreLocalGCRequests(Heap* heap);
2604   inline ~IgnoreLocalGCRequests();
2605 
2606  private:
2607   Heap* heap_;
2608 };
2609 
2610 // Visitor class to verify interior pointers in spaces that do not contain
2611 // or care about intergenerational references. All heap object pointers have to
2612 // point into the heap to a location that has a map pointer at its first word.
2613 // Caveat: Heap::Contains is an approximation because it can return true for
2614 // objects in a heap space but above the allocation pointer.
2615 class VerifyPointersVisitor : public ObjectVisitorWithCageBases,
2616                               public RootVisitor {
2617  public:
2618   V8_INLINE explicit VerifyPointersVisitor(Heap* heap);
2619   void VisitPointers(HeapObject host, ObjectSlot start,
2620                      ObjectSlot end) override;
2621   void VisitPointers(HeapObject host, MaybeObjectSlot start,
2622                      MaybeObjectSlot end) override;
2623   void VisitCodePointer(HeapObject host, CodeObjectSlot slot) override;
2624   void VisitCodeTarget(Code host, RelocInfo* rinfo) override;
2625   void VisitEmbeddedPointer(Code host, RelocInfo* rinfo) override;
2626 
2627   void VisitRootPointers(Root root, const char* description,
2628                          FullObjectSlot start, FullObjectSlot end) override;
2629   void VisitRootPointers(Root root, const char* description,
2630                          OffHeapObjectSlot start,
2631                          OffHeapObjectSlot end) override;
2632 
2633  protected:
2634   V8_INLINE void VerifyHeapObjectImpl(HeapObject heap_object);
2635   V8_INLINE void VerifyCodeObjectImpl(HeapObject heap_object);
2636 
2637   template <typename TSlot>
2638   V8_INLINE void VerifyPointersImpl(TSlot start, TSlot end);
2639 
2640   virtual void VerifyPointers(HeapObject host, MaybeObjectSlot start,
2641                               MaybeObjectSlot end);
2642 
2643   Heap* heap_;
2644 };
2645 
2646 // Verify that all objects are Smis.
2647 class VerifySmisVisitor : public RootVisitor {
2648  public:
2649   void VisitRootPointers(Root root, const char* description,
2650                          FullObjectSlot start, FullObjectSlot end) override;
2651 };
2652 
2653 // Space iterator for iterating over all the paged spaces of the heap: Map
2654 // space, old space and code space. Returns each space in turn, and null when it
2655 // is done.
2656 class V8_EXPORT_PRIVATE PagedSpaceIterator {
2657  public:
PagedSpaceIterator(Heap * heap)2658   explicit PagedSpaceIterator(Heap* heap)
2659       : heap_(heap), counter_(FIRST_GROWABLE_PAGED_SPACE) {}
2660   PagedSpace* Next();
2661 
2662  private:
2663   Heap* heap_;
2664   int counter_;
2665 };
2666 
2667 class V8_EXPORT_PRIVATE SpaceIterator : public Malloced {
2668  public:
2669   explicit SpaceIterator(Heap* heap);
2670   virtual ~SpaceIterator();
2671 
2672   bool HasNext();
2673   Space* Next();
2674 
2675  private:
2676   Heap* heap_;
2677   int current_space_;  // from enum AllocationSpace.
2678 };
2679 
2680 // A HeapObjectIterator provides iteration over the entire non-read-only heap.
2681 // It aggregates the specific iterators for the different spaces as these can
2682 // only iterate over one space only.
2683 //
2684 // HeapObjectIterator ensures there is no allocation during its lifetime (using
2685 // an embedded DisallowGarbageCollection instance).
2686 //
2687 // HeapObjectIterator can skip free list nodes (that is, de-allocated heap
2688 // objects that still remain in the heap). As implementation of free nodes
2689 // filtering uses GC marks, it can't be used during MS/MC GC phases. Also, it is
2690 // forbidden to interrupt iteration in this mode, as this will leave heap
2691 // objects marked (and thus, unusable).
2692 //
2693 // See ReadOnlyHeapObjectIterator if you need to iterate over read-only space
2694 // objects, or CombinedHeapObjectIterator if you need to iterate over both
2695 // heaps.
2696 class V8_EXPORT_PRIVATE HeapObjectIterator {
2697  public:
2698   enum HeapObjectsFiltering { kNoFiltering, kFilterUnreachable };
2699 
2700   explicit HeapObjectIterator(Heap* heap,
2701                               HeapObjectsFiltering filtering = kNoFiltering);
2702   ~HeapObjectIterator();
2703 
2704   HeapObject Next();
2705 
2706  private:
2707   HeapObject NextObject();
2708 
2709   Heap* heap_;
2710   std::unique_ptr<SafepointScope> safepoint_scope_;
2711   HeapObjectsFiltering filtering_;
2712   HeapObjectsFilter* filter_;
2713   // Space iterator for iterating all the spaces.
2714   SpaceIterator* space_iterator_;
2715   // Object iterator for the space currently being iterated.
2716   std::unique_ptr<ObjectIterator> object_iterator_;
2717 
2718   DISALLOW_GARBAGE_COLLECTION(no_heap_allocation_)
2719 };
2720 
2721 // Abstract base class for checking whether a weak object should be retained.
2722 class WeakObjectRetainer {
2723  public:
2724   virtual ~WeakObjectRetainer() = default;
2725 
2726   // Return whether this object should be retained. If nullptr is returned the
2727   // object has no references. Otherwise the address of the retained object
2728   // should be returned as in some GC situations the object has been moved.
2729   virtual Object RetainAs(Object object) = 0;
2730 };
2731 
2732 // -----------------------------------------------------------------------------
2733 // Allows observation of heap object allocations.
2734 class HeapObjectAllocationTracker {
2735  public:
2736   virtual void AllocationEvent(Address addr, int size) = 0;
MoveEvent(Address from,Address to,int size)2737   virtual void MoveEvent(Address from, Address to, int size) {}
UpdateObjectSizeEvent(Address addr,int size)2738   virtual void UpdateObjectSizeEvent(Address addr, int size) {}
2739   virtual ~HeapObjectAllocationTracker() = default;
2740 };
2741 
2742 template <typename T>
2743 inline T ForwardingAddress(T heap_obj);
2744 
2745 // Address block allocator compatible with standard containers which registers
2746 // its allocated range as strong roots.
2747 class StrongRootBlockAllocator {
2748  public:
2749   using pointer = Address*;
2750   using const_pointer = const Address*;
2751   using reference = Address&;
2752   using const_reference = const Address&;
2753   using value_type = Address;
2754   using size_type = size_t;
2755   using difference_type = ptrdiff_t;
2756   template <class U>
2757   struct rebind;
2758 
StrongRootBlockAllocator(Heap * heap)2759   explicit StrongRootBlockAllocator(Heap* heap) : heap_(heap) {}
2760 
2761   Address* allocate(size_t n);
2762   void deallocate(Address* p, size_t n) noexcept;
2763 
2764  private:
2765   Heap* heap_;
2766 };
2767 
2768 // Rebinding to Address gives another StrongRootBlockAllocator.
2769 template <>
2770 struct StrongRootBlockAllocator::rebind<Address> {
2771   using other = StrongRootBlockAllocator;
2772 };
2773 
2774 // Rebinding to something other than Address gives a std::allocator that
2775 // is copy-constructable from StrongRootBlockAllocator.
2776 template <class U>
2777 struct StrongRootBlockAllocator::rebind {
2778   class other : public std::allocator<U> {
2779    public:
2780     // NOLINTNEXTLINE
2781     other(const StrongRootBlockAllocator&) {}
2782   };
2783 };
2784 
2785 class V8_EXPORT_PRIVATE V8_NODISCARD EmbedderStackStateScope final {
2786  public:
2787   enum Origin {
2788     kImplicitThroughTask,
2789     kExplicitInvocation,
2790   };
2791 
2792   // Only used for testing where the Origin is always an explicit invocation.
2793   static EmbedderStackStateScope ExplicitScopeForTesting(
2794       LocalEmbedderHeapTracer* local_tracer,
2795       EmbedderHeapTracer::EmbedderStackState stack_state);
2796 
2797   EmbedderStackStateScope(Heap* heap, Origin origin,
2798                           EmbedderHeapTracer::EmbedderStackState stack_state);
2799   ~EmbedderStackStateScope();
2800 
2801  private:
2802   EmbedderStackStateScope(LocalEmbedderHeapTracer* local_tracer,
2803                           EmbedderHeapTracer::EmbedderStackState stack_state);
2804 
2805   LocalEmbedderHeapTracer* const local_tracer_;
2806   const EmbedderHeapTracer::EmbedderStackState old_stack_state_;
2807 };
2808 
2809 }  // namespace internal
2810 }  // namespace v8
2811 
2812 #endif  // V8_HEAP_HEAP_H_
2813