/* * Copyright (C) 2015 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef ART_RUNTIME_LINEAR_ALLOC_INL_H_ #define ART_RUNTIME_LINEAR_ALLOC_INL_H_ #include "linear_alloc.h" #include "base/gc_visited_arena_pool.h" #include "thread-current-inl.h" namespace art { inline void LinearAlloc::SetFirstObject(void* begin, size_t bytes) const { DCHECK(track_allocations_); if (ArenaAllocator::IsRunningOnMemoryTool()) { bytes += ArenaAllocator::kMemoryToolRedZoneBytes; } uint8_t* end = static_cast(begin) + bytes; Arena* arena = allocator_.GetHeadArena(); DCHECK_NE(arena, nullptr); // The object would either be in the head arena or the next one. if (UNLIKELY(begin < arena->Begin() || begin >= arena->End())) { arena = arena->Next(); } DCHECK(begin >= arena->Begin() && end <= arena->End()); down_cast(arena)->SetFirstObject(static_cast(begin), end); } inline void LinearAlloc::SetupForPostZygoteFork(Thread* self) { MutexLock mu(self, lock_); DCHECK(track_allocations_); allocator_.ResetCurrentArena(); } inline void* LinearAlloc::Realloc(Thread* self, void* ptr, size_t old_size, size_t new_size, LinearAllocKind kind) { MutexLock mu(self, lock_); if (track_allocations_) { if (ptr != nullptr) { // Realloc cannot be called on 16-byte aligned as Realloc doesn't guarantee // that. So the header must be immediately prior to ptr. TrackingHeader* header = reinterpret_cast(ptr) - 1; DCHECK_EQ(header->GetKind(), kind); old_size += sizeof(TrackingHeader); DCHECK_EQ(header->GetSize(), old_size); ptr = header; } else { DCHECK_EQ(old_size, 0u); } new_size += sizeof(TrackingHeader); void* ret = allocator_.Realloc(ptr, old_size, new_size); new (ret) TrackingHeader(new_size, kind); SetFirstObject(ret, new_size); return static_cast(ret) + 1; } else { return allocator_.Realloc(ptr, old_size, new_size); } } inline void* LinearAlloc::Alloc(Thread* self, size_t size, LinearAllocKind kind) { MutexLock mu(self, lock_); if (track_allocations_) { size += sizeof(TrackingHeader); TrackingHeader* storage = new (allocator_.Alloc(size)) TrackingHeader(size, kind); SetFirstObject(storage, size); return storage + 1; } else { return allocator_.Alloc(size); } } inline void* LinearAlloc::AllocAlign16(Thread* self, size_t size, LinearAllocKind kind) { MutexLock mu(self, lock_); DCHECK_ALIGNED(size, 16); if (track_allocations_) { size_t mem_tool_bytes = ArenaAllocator::IsRunningOnMemoryTool() ? ArenaAllocator::kMemoryToolRedZoneBytes : 0; uint8_t* ptr = allocator_.CurrentPtr() + sizeof(TrackingHeader); uintptr_t padding = RoundUp(reinterpret_cast(ptr), 16) - reinterpret_cast(ptr); DCHECK_LT(padding, 16u); size_t required_size = size + sizeof(TrackingHeader) + padding; if (allocator_.CurrentArenaUnusedBytes() < required_size + mem_tool_bytes) { // The allocator will require a new arena, which is expected to be // 16-byte aligned. static_assert(ArenaAllocator::kArenaAlignment >= 16, "Expecting sufficient alignment for new Arena."); required_size = size + RoundUp(sizeof(TrackingHeader), 16); } // Using ArenaAllocator's AllocAlign16 now would disturb the alignment by // trying to make header 16-byte aligned. The alignment requirements are // already addressed here. Now we want allocator to just bump the pointer. ptr = static_cast(allocator_.Alloc(required_size)); new (ptr) TrackingHeader(required_size, kind, /*is_16_aligned=*/true); SetFirstObject(ptr, required_size); return AlignUp(ptr + sizeof(TrackingHeader), 16); } else { return allocator_.AllocAlign16(size); } } inline size_t LinearAlloc::GetUsedMemory() const { MutexLock mu(Thread::Current(), lock_); return allocator_.BytesUsed(); } inline ArenaPool* LinearAlloc::GetArenaPool() { MutexLock mu(Thread::Current(), lock_); return allocator_.GetArenaPool(); } inline bool LinearAlloc::Contains(void* ptr) const { MutexLock mu(Thread::Current(), lock_); return allocator_.Contains(ptr); } } // namespace art #endif // ART_RUNTIME_LINEAR_ALLOC_INL_H_