/external/v8/src/heap/ |
D | scavenger.cc | 111 if (heap->new_space()->Contains(obj)) { in RecordCopiedObject() 112 heap->new_space()->RecordAllocation(obj); in RecordCopiedObject() 114 heap->new_space()->RecordPromotion(obj); in RecordCopiedObject() 128 target->address() + size == heap->new_space()->top() || in INLINE() 129 target->address() + size + kPointerSize == heap->new_space()->top()); in INLINE() 135 heap->new_space()->top())); in INLINE() 163 heap->new_space()->AllocateRaw(object_size, alignment); in SemiSpaceCopyObject() 171 heap->promotion_queue()->SetNewLimit(heap->new_space()->top()); in SemiSpaceCopyObject()
|
D | scavenge-job.cc | 28 size_t new_space_size = heap->new_space()->Size(); in RunInternal() 29 size_t new_space_capacity = heap->new_space()->Capacity(); in RunInternal()
|
D | array-buffer-tracker.cc | 80 for (Page* page : PageRange(heap->new_space()->FromSpaceStart(), in FreeDeadInNewSpace() 81 heap->new_space()->FromSpaceEnd())) { in FreeDeadInNewSpace()
|
D | mark-compact.cc | 154 VerifyMarking(heap->new_space()); in VerifyMarking() 227 VerifyEvacuation(heap->new_space()); in VerifyEvacuation() 343 VerifyMarkbitsAreClean(heap_->new_space()); in VerifyMarkbitsAreClean() 392 ClearMarkbitsInNewSpace(heap_->new_space()); in ClearMarkbits() 516 for (Page* p : *heap_->new_space()) { in EnsureNewSpaceCompleted() 1791 heap_->new_space()->AllocateRawSynchronized(size_in_bytes, alignment); in AllocateInNewSpace() 1793 if (!heap_->new_space()->AddFreshPageSynchronized()) { in AllocateInNewSpace() 1796 allocation = heap_->new_space()->AllocateRawSynchronized(size_in_bytes, in AllocateInNewSpace() 1863 page->heap()->new_space()->MovePageFromSpaceToSpace(page); in Move() 1937 NewSpace* space = heap()->new_space(); in DiscoverGreyObjectsInNewSpace() [all …]
|
D | incremental-marking.cc | 357 DeactivateIncrementalWriteBarrierForSpace(heap_->new_space()); in DeactivateIncrementalWriteBarrier() 383 ActivateIncrementalWriteBarrier(heap_->new_space()); in ActivateIncrementalWriteBarrier() 488 if (space == heap_->new_space()) { in Start() 944 if (space == heap_->new_space()) { in Stop() 1081 size_t oom_slack = heap()->new_space()->Capacity() + 64 * MB; in StepSizeToMakeProgress()
|
D | heap.cc | 687 (new_space()->CommittedMemory() * 100.0) / CommittedMemory())); in GarbageCollectionEpilogue() 729 UPDATE_COUNTERS_FOR_SPACE(new_space) in GarbageCollectionEpilogue() 744 new_space_top_after_last_gc_ = new_space()->top(); in GarbageCollectionEpilogue() 1196 allocation = new_space()->AllocateRawUnaligned(size); in ReserveSpace() 1322 int start_new_space_size = static_cast<int>(Heap::new_space()->Size()); in PerformGarbageCollection() 1346 CanExpandOldGeneration(new_space()->Size())) { in PerformGarbageCollection() 1549 reinterpret_cast<struct Entry*>(heap_->new_space()->ToSpaceEnd()); in Initialize() 1609 DCHECK(CanExpandOldGeneration(new_space()->Size())); in EvacuateYoungGeneration() 1617 PageRange range(new_space()->bottom(), new_space()->top()); in EvacuateYoungGeneration() 1627 if (!new_space()->Rebalance()) { in EvacuateYoungGeneration() [all …]
|
D | heap-inl.h | 60 SemiSpace::AssertValidRange(target->GetIsolate()->heap()->new_space()->top(), in insert() 189 return new_space_allocation_counter_ + new_space()->AllocatedSinceLastGC(); in NewSpaceAllocationCounter()
|
D | gc-tracer.cc | 184 current_.new_space_object_size = heap_->new_space()->Size(); in Start()
|
D | heap.h | 1021 NewSpace* new_space() { return new_space_; } in new_space() function
|
/external/tensorflow/tensorflow/contrib/model_pruning/python/layers/ |
D | core_layers.py | 217 new_space = [] 225 new_space.append(new_dim) 226 return tensor_shape.TensorShape([input_shape[0]] + new_space + 230 new_space = [] 238 new_space.append(new_dim) 240 new_space)
|
/external/v8/src/extensions/ |
D | statistics-extension.cc | 121 {heap->new_space()->Size(), "new_space_live_bytes"}, in GetCounters() 122 {heap->new_space()->Available(), "new_space_available_bytes"}, in GetCounters() 123 {heap->new_space()->CommittedMemory(), "new_space_commited_bytes"}, in GetCounters()
|
/external/v8/src/profiler/ |
D | sampling-heap-profiler.cc | 69 heap->new_space()->AddAllocationObserver(new_space_observer_.get()); in SamplingHeapProfiler() 72 if (space != heap->new_space()) { in SamplingHeapProfiler() 80 heap_->new_space()->RemoveAllocationObserver(new_space_observer_.get()); in ~SamplingHeapProfiler() 83 if (space != heap_->new_space()) { in ~SamplingHeapProfiler()
|
/external/v8/src/ |
D | string-stream.cc | 531 char* new_space = NewArray<char>(new_bytes); in grow() local 532 if (new_space == NULL) { in grow() 535 MemCopy(new_space, space_, *bytes); in grow() 538 space_ = new_space; in grow() 539 return new_space; in grow()
|
D | code-stub-assembler.cc | 799 bool const new_space = !(flags & kPretenured); in Allocate() local 801 new_space in Allocate() 1714 bool const new_space = !(flags & kPretenured); in AllocateConsString() local 1715 if (new_space) { in AllocateConsString()
|
D | api.cc | 6870 return !isolate->heap()->new_space()->Contains(*obj); in CanMakeExternal()
|
/external/libjpeg-turbo/ |
D | jmemmgr.c | 657 size_t new_space = (long) sptr->rows_in_array * in realize_virt_arrays() local 662 if (SIZE_MAX - maximum_space < new_space) in realize_virt_arrays() 664 maximum_space += new_space; in realize_virt_arrays() 669 size_t new_space = (long) bptr->rows_in_array * in realize_virt_arrays() local 674 if (SIZE_MAX - maximum_space < new_space) in realize_virt_arrays() 676 maximum_space += new_space; in realize_virt_arrays()
|
/external/tensorflow/tensorflow/contrib/bayesflow/python/ops/ |
D | layers_conv_variational.py | 326 new_space = [] 334 new_space.append(new_dim) 335 return tensor_shape.TensorShape([input_shape[0]] + new_space + 339 new_space = [] 347 new_space.append(new_dim) 349 new_space)
|
/external/tensorflow/tensorflow/python/layers/ |
D | convolutional.py | 200 new_space = [] 208 new_space.append(new_dim) 209 return tensor_shape.TensorShape([input_shape[0]] + new_space + 213 new_space = [] 221 new_space.append(new_dim) 223 new_space)
|
/external/blktrace/ |
D | blkparse.c | 326 int new_space, size; in resize_cpu_info() local 339 new_space = (new_count - ncpus) * sizeof(struct per_cpu_info); in resize_cpu_info() 340 memset(new_start, 0, new_space); in resize_cpu_info()
|