/art/runtime/utils/ |
D | dex_cache_arrays_layout-inl.h | 38 RoundUp(types_offset_ + TypesSize(header.type_ids_size_), MethodsAlignment())), in DexCacheArraysLayout() 40 RoundUp(methods_offset_ + MethodsSize(header.method_ids_size_), StringsAlignment())), in DexCacheArraysLayout() 42 RoundUp(strings_offset_ + StringsSize(header.string_ids_size_), FieldsAlignment())), in DexCacheArraysLayout() 44 RoundUp(fields_offset_ + FieldsSize(header.field_ids_size_), MethodTypesAlignment())), in DexCacheArraysLayout() 46 RoundUp(method_types_offset_ + MethodTypesSize(header.proto_ids_size_), in DexCacheArraysLayout() 48 size_(RoundUp(call_sites_offset_ + CallSitesSize(num_call_sites), Alignment())) { in DexCacheArraysLayout()
|
/art/runtime/ |
D | image.cc | 62 CHECK_EQ(image_begin, RoundUp(image_begin, kPageSize)); in ImageHeader() 63 CHECK_EQ(oat_file_begin, RoundUp(oat_file_begin, kPageSize)); in ImageHeader() 64 CHECK_EQ(oat_data_begin, RoundUp(oat_data_begin, kPageSize)); in ImageHeader() 141 static const size_t kStartPos = RoundUp(sizeof(ImageHeader), kObjectAlignment); in VisitObjects() 145 pos += RoundUp(object->SizeOf(), kObjectAlignment); in VisitObjects()
|
/art/runtime/gc/space/ |
D | malloc_space.cc | 106 *growth_limit = RoundUp(*growth_limit, kPageSize); in CreateMemMap() 107 *capacity = RoundUp(*capacity, kPageSize); in CreateMemMap() 144 growth_limit = RoundUp(growth_limit, kPageSize); in SetGrowthLimit() 187 SetEnd(reinterpret_cast<uint8_t*>(RoundUp(reinterpret_cast<uintptr_t>(End()), kPageSize))); in CreateZygoteSpace() 192 size_t size = RoundUp(Size(), kPageSize); in CreateZygoteSpace() 204 SetGrowthLimit(RoundUp(size, kPageSize)); in CreateZygoteSpace()
|
D | bump_pointer_space.cc | 28 capacity = RoundUp(capacity, kPageSize); in Create() 102 return reinterpret_cast<mirror::Object*>(RoundUp(position, kAlignment)); in GetNextObject() 150 bytes = RoundUp(bytes, kAlignment); in AllocBlock() 234 *usable_size = RoundUp(num_bytes, kAlignment); in AllocationSizeNonvirtual()
|
D | bump_pointer_space-inl.h | 31 num_bytes = RoundUp(num_bytes, kAlignment); in Alloc() 48 num_bytes = RoundUp(num_bytes, kAlignment); in AllocThreadUnsafe()
|
D | region_space.cc | 274 const size_t bytes_allocated = RoundUp(BytesAllocated(), kRegionSize); in ShouldBeEvacuated() 304 size_t obj_alloc_size = RoundUp(obj_size, space::RegionSpace::kRegionSize); in ZeroLiveBytesForLargeObject() 380 num_expected_large_tails = RoundUp(r->BytesAllocated(), kRegionSize) / kRegionSize - 1; in SetFromSpace() 653 size_t alloc_size = RoundUp(obj_size, space::RegionSpace::kAlignment); in CheckLiveBytesAgainstRegionBitmap() 916 << (static_cast<float>(live_bytes_) / RoundUp(BytesAllocated(), kRegionSize)); in Dump() 943 prev_object_end = RoundUp(object_end, kAlignment); in GetLongestConsecutiveFreeBytes() 956 *usable_size = RoundUp(num_bytes, kAlignment); in AllocationSizeNonvirtual() 959 *usable_size = RoundUp(num_bytes, kRegionSize); in AllocationSizeNonvirtual()
|
/art/compiler/utils/ |
D | swap_space.cc | 110 size = RoundUp(size, 8U); in Alloc() 149 size_t next_part = std::max(RoundUp(min_size, kPageSize), RoundUp(kMininumMapSize, kPageSize)); in NewFileChunk() 177 size = RoundUp(size, 8U); in Free()
|
/art/dex2oat/linker/arm/ |
D | relative_patcher_thumb2_test.cc | 698 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmAlignment); in TestBakerFieldWide() 764 thunk_offset += RoundUp(expected_thunk.size(), kArmAlignment); in TestBakerFieldWide() 796 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmAlignment); in TestBakerFieldNarrow() 865 thunk_offset += RoundUp(expected_thunk.size(), kArmAlignment); in TestBakerFieldNarrow() 920 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmAlignment); in TEST_F() 941 - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArmAlignment) in TEST_F() 942 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArmAlignment) in TEST_F() 988 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmAlignment); in TEST_F() 995 const uint32_t bne = BneWWithOffset(kLiteralOffset1, RoundUp(raw_code1.size(), kArmAlignment)); in TEST_F() 1018 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmAlignment); in TEST_F() [all …]
|
/art/runtime/gc/accounting/ |
D | bitmap.cc | 48 const size_t bitmap_size = RoundUp( in AllocateMemMap() 49 RoundUp(num_bits, kBitsPerBitmapWord) / kBitsPerBitmapWord * sizeof(uintptr_t), kPageSize); in AllocateMemMap()
|
/art/runtime/mirror/ |
D | array.h | 78 size_t data_offset = RoundUp(OFFSETOF_MEMBER(Array, first_element_), component_size); in DataOffset() 79 DCHECK_EQ(RoundUp(data_offset, component_size), data_offset) in DataOffset() 86 constexpr size_t data_offset = RoundUp(kFirstElementOffset, kComponentSize); in DataOffset() 87 static_assert(RoundUp(data_offset, kComponentSize) == data_offset, "RoundUp fail"); in DataOffset()
|
/art/dex2oat/linker/arm64/ |
D | relative_patcher_arm64_test.cc | 1037 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArm64Alignment); in TestBakerField() 1090 thunk_offset += RoundUp(expected_thunk.size(), kArm64Alignment); in TestBakerField() 1127 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64Alignment); in TEST_F() 1142 1 * MB - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArm64Alignment) in TEST_F() 1143 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArm64Alignment) in TEST_F() 1187 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64Alignment); in TEST_F() 1194 const uint32_t cbnz_offset = RoundUp(raw_code1.size(), kArm64Alignment) - kLiteralOffset1; in TEST_F() 1216 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64Alignment); in TEST_F() 1231 1 * MB - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArm64Alignment) in TEST_F() 1232 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArm64Alignment) in TEST_F() [all …]
|
/art/libartbase/base/ |
D | scoped_arena_allocator.cc | 96 size_t rounded_bytes = RoundUp(bytes + kMemoryToolRedZoneBytes, 8); in AllocWithMemoryTool() 141 arena_stack_->top_ptr_ = mark_ptr_ + RoundUp(sizeof(ScopedArenaAllocator), 8); in Reset()
|
D | arena_allocator.h | 271 bytes = RoundUp(bytes, kAlignment); 311 const size_t aligned_ptr_size = RoundUp(ptr_size, kAlignment); 317 const size_t aligned_new_size = RoundUp(new_size, kAlignment);
|
D | bit_utils.h | 171 constexpr T RoundUp(T x, typename std::remove_reference<T>::type n) WARN_UNUSED; 174 constexpr T RoundUp(T x, typename std::remove_reference<T>::type n) { in RoundUp() function 192 return reinterpret_cast<T*>(RoundUp(reinterpret_cast<uintptr_t>(x), n)); in AlignUp() 504 return RoundUp(num_bits, kBitsPerByte) / kBitsPerByte; in BitsToBytesRoundUp()
|
/art/compiler/jni/quick/arm/ |
D | calling_convention_arm.cc | 202 fpr_double_index = (std::max(fpr_double_index * 2, RoundUp(fpr_index, 2))) / 2; in EntrySpills() 386 return RoundUp(total_size, kStackAlignment); in FrameSize() 391 return RoundUp(NumberOfOutgoingStackArgs() * kFramePointerSize + padding_, in OutArgSize()
|
/art/compiler/jni/quick/mips64/ |
D | calling_convention_mips64.cc | 203 return RoundUp(total_size, kStackAlignment); in FrameSize() 207 return RoundUp(NumberOfOutgoingStackArgs() * kFramePointerSize, kStackAlignment); in OutArgSize()
|
/art/compiler/jni/quick/x86/ |
D | calling_convention_x86.cc | 239 return RoundUp(total_size, kStackAlignment); in FrameSize() 244 return RoundUp(NumberOfOutgoingStackArgs() * kFramePointerSize, kStackAlignment); in OutArgSize()
|
/art/runtime/gc/allocator/ |
D | rosalloc.h | 549 return RoundUp(size, kThreadLocalBracketQuantumSize); in RoundToBracketSize() 551 return RoundUp(size, kBracketQuantumSize); in RoundToBracketSize() 563 return RoundUp(size, kThreadLocalBracketQuantumSize) / kThreadLocalBracketQuantumSize - 1; in SizeToIndex() 565 return (RoundUp(size, kBracketQuantumSize) - kMaxThreadLocalBracketSize) / kBracketQuantumSize in SizeToIndex() 580 bracket_size = RoundUp(size, kThreadLocalBracketQuantumSize); in SizeToIndexAndBracketSize() 583 bracket_size = RoundUp(size, kBracketQuantumSize); in SizeToIndexAndBracketSize() 874 return RoundUp(bytes, kPageSize); in UsableSize()
|
D | dlmalloc.cc | 72 start = reinterpret_cast<void*>(art::RoundUp(reinterpret_cast<uintptr_t>(start), art::kPageSize)); in DlmallocMadviseCallback()
|
/art/runtime/gc/ |
D | heap-inl.h | 83 byte_count = RoundUp(byte_count, space::BumpPointerSpace::kAlignment); in AllocObjectWithAllocator() 267 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment); in TryToAllocate() 347 alloc_size = RoundUp(alloc_size, space::RegionSpace::kAlignment); in TryToAllocate()
|
/art/compiler/jni/quick/x86_64/ |
D | calling_convention_x86_64.cc | 221 return RoundUp(total_size, kStackAlignment); in FrameSize() 225 return RoundUp(NumberOfOutgoingStackArgs() * kFramePointerSize, kStackAlignment); in OutArgSize()
|
/art/runtime/arch/arm/ |
D | quick_entrypoints_cc_arm.cc | 50 fpr_double_index = std::max(fpr_double_index, RoundUp(fpr_index, 2)); in quick_invoke_reg_setup()
|
/art/compiler/jni/quick/mips/ |
D | calling_convention_mips.cc | 362 return RoundUp(total_size, kStackAlignment); in FrameSize() 374 return RoundUp( in OutArgSize()
|
/art/compiler/jni/quick/arm64/ |
D | calling_convention_arm64.cc | 293 return RoundUp(total_size, kStackAlignment); in FrameSize() 298 return RoundUp(NumberOfOutgoingStackArgs() * kFramePointerSize, kStackAlignment); in OutArgSize()
|
/art/runtime/gc/collector/ |
D | immune_spaces.cc | 53 space_end = RoundUp(reinterpret_cast<uintptr_t>(image_space->GetImageEnd()), kPageSize); in CreateLargestImmuneRegion()
|