Lines Matching refs:alloc_size
184 size_t alloc_size, size_t* bytes_allocated, in TryToAllocate() argument
187 UNLIKELY(IsOutOfMemoryOnAllocation<kGrow>(allocator_type, alloc_size))) { in TryToAllocate()
194 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment); in TryToAllocate()
195 ret = bump_pointer_space_->AllocNonvirtual(alloc_size); in TryToAllocate()
197 *bytes_allocated = alloc_size; in TryToAllocate()
198 *usable_size = alloc_size; in TryToAllocate()
205 ret = rosalloc_space_->Alloc(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
208 ret = rosalloc_space_->AllocNonvirtual(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
215 ret = dlmalloc_space_->Alloc(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
218 ret = dlmalloc_space_->AllocNonvirtual(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
223 ret = non_moving_space_->Alloc(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
227 ret = large_object_space_->Alloc(self, alloc_size, bytes_allocated, usable_size); in TryToAllocate()
235 DCHECK_ALIGNED(alloc_size, space::BumpPointerSpace::kAlignment); in TryToAllocate()
236 if (UNLIKELY(self->TlabSize() < alloc_size)) { in TryToAllocate()
237 const size_t new_tlab_size = alloc_size + kDefaultTLABSize; in TryToAllocate()
251 ret = self->AllocTlab(alloc_size); in TryToAllocate()
253 *usable_size = alloc_size; in TryToAllocate()
291 inline bool Heap::IsOutOfMemoryOnAllocation(AllocatorType allocator_type, size_t alloc_size) { in IsOutOfMemoryOnAllocation() argument
292 size_t new_footprint = num_bytes_allocated_.LoadSequentiallyConsistent() + alloc_size; in IsOutOfMemoryOnAllocation()
303 << PrettySize(new_footprint) << " for a " << PrettySize(alloc_size) << " allocation"; in IsOutOfMemoryOnAllocation()