/art/libartbase/base/ |
D | arena_allocator_test.cc | 143 EXPECT_TRUE(IsAligned<ArenaAllocator::kAlignment>(allocation)) in TEST_F() 248 ASSERT_TRUE(IsAligned<ArenaAllocator::kAlignment>(original_allocation)); in TEST_F() 252 EXPECT_TRUE(IsAligned<ArenaAllocator::kAlignment>(realloc_allocation)); in TEST_F() 255 EXPECT_TRUE(IsAligned<ArenaAllocator::kAlignment>(after_alloc)); in TEST_F() 265 ASSERT_TRUE(IsAligned<ArenaAllocator::kAlignment>(original_allocation)); in TEST_F() 269 EXPECT_TRUE(IsAligned<ArenaAllocator::kAlignment>(realloc_allocation)); in TEST_F() 272 EXPECT_TRUE(IsAligned<ArenaAllocator::kAlignment>(after_alloc)); in TEST_F() 282 ASSERT_TRUE(IsAligned<ArenaAllocator::kAlignment>(original_allocation)); in TEST_F() 286 EXPECT_TRUE(IsAligned<ArenaAllocator::kAlignment>(realloc_allocation)); in TEST_F() 289 EXPECT_TRUE(IsAligned<ArenaAllocator::kAlignment>(after_alloc)); in TEST_F() [all …]
|
D | bit_utils.h | 196 constexpr bool IsAligned(T x) { in IsAligned() function 202 inline bool IsAligned(T* x) { in IsAligned() function 203 return IsAligned<n>(reinterpret_cast<const uintptr_t>(x)); in IsAligned() 217 CHECK(::art::IsAligned<alignment>(value)) << reinterpret_cast<const void*>(value) 220 DCHECK(::art::IsAligned<alignment>(value)) << reinterpret_cast<const void*>(value)
|
D | memory_region.h | 154 return IsAligned<kRuntimePointerSize>(address); in IsWordAligned()
|
D | bit_memory_region.h | 110 DCHECK(IsAligned<sizeof(uintptr_t)>(data_)); in LoadBits()
|
D | mem_map.cc | 1195 DCHECK(IsAligned<kPageSize>(begin)); in TryReadable() 1196 DCHECK(IsAligned<kPageSize>(end)); in TryReadable()
|
/art/libartbase/arch/ |
D | instruction_set.cc | 110 static_assert(IsAligned<kPageSize>(kArmStackOverflowReservedBytes), "ARM gap not page aligned"); 111 static_assert(IsAligned<kPageSize>(kArm64StackOverflowReservedBytes), "ARM64 gap not page aligned"); 112 static_assert(IsAligned<kPageSize>(kMipsStackOverflowReservedBytes), "Mips gap not page aligned"); 113 static_assert(IsAligned<kPageSize>(kMips64StackOverflowReservedBytes), 115 static_assert(IsAligned<kPageSize>(kX86StackOverflowReservedBytes), "X86 gap not page aligned"); 116 static_assert(IsAligned<kPageSize>(kX86_64StackOverflowReservedBytes),
|
/art/runtime/ |
D | verify_object.cc | 35 bool failed = !IsAligned<kObjectAlignment>(obj.Ptr()); in VerifyObjectImpl() 38 failed = failed || !IsAligned<kObjectAlignment>(c); in VerifyObjectImpl()
|
D | fault_handler.cc | 304 if (method_obj == nullptr || !IsAligned<sizeof(void*)>(method_obj)) { in IsInGeneratedCode() 319 if (!IsAligned<kObjectAlignment>(cls)) { in IsInGeneratedCode()
|
D | oat.cc | 103 if (!IsAligned<kPageSize>(executable_offset_)) { in IsValid() 125 if (!IsAligned<kPageSize>(executable_offset_)) { in GetValidationErrorMessage()
|
D | image.cc | 103 if (!IsAligned<kPageSize>(image_reservation_size_)) { in IsValid()
|
D | class_linker_test.cc | 388 CHECK(!IsAligned<4>(field->GetOffset().Uint32Value())); in AssertClass() 1081 EXPECT_TRUE(IsAligned<4>(data_offset)); // Check 4 byte alignment. in TEST_F() 1083 EXPECT_TRUE(IsAligned<8>(data_offset)); // Check 8 byte alignment. in TEST_F() 1093 EXPECT_TRUE(IsAligned<8>(data_offset)); // Longs require 8 byte alignment in TEST_F() 1098 EXPECT_TRUE(IsAligned<8>(data_offset)); // Doubles require 8 byte alignment in TEST_F() 1103 EXPECT_TRUE(IsAligned<4>(data_offset)); // Ints require 4 byte alignment in TEST_F() 1108 EXPECT_TRUE(IsAligned<2>(data_offset)); // Chars require 2 byte alignment in TEST_F() 1113 EXPECT_TRUE(IsAligned<2>(data_offset)); // Shorts require 2 byte alignment in TEST_F()
|
D | oat_file.cc | 428 IsAligned<alignof(IndexBssMapping)>(index_bss_mapping_offset) && in ReadIndexBssMapping() 508 if (!IsAligned<sizeof(uint32_t)>(data_bimg_rel_ro_begin_) || in Setup() 509 !IsAligned<sizeof(uint32_t)>(data_bimg_rel_ro_end_) || in Setup() 520 if (!IsAligned<kPageSize>(bss_begin_) || in Setup() 523 !IsAligned<alignof(GcRoot<mirror::Object>)>(bss_end_)) { in Setup() 760 if (UNLIKELY(!IsAligned<alignof(uint32_t)>(class_offsets_offset))) { in Setup()
|
D | stack_map.h | 236 DCHECK(IsAligned<kFrameSlotSize>(packed_value)); in PackValue()
|
/art/runtime/gc/accounting/ |
D | card_table-inl.h | 71 while (!IsAligned<sizeof(intptr_t)>(card_cur) && card_cur < card_end) { in Scan() 142 while (!IsAligned<sizeof(intptr_t)>(card_cur) && card_cur < card_end) { in ModifyCardsAtomic() 155 while (!IsAligned<sizeof(intptr_t)>(card_end) && card_end > card_cur) { in ModifyCardsAtomic()
|
/art/compiler/utils/arm/ |
D | assembler_arm_vixl.cc | 214 return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset); // VFP addressing mode. in CanHoldLoadOffsetThumb() 216 return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset); in CanHoldLoadOffsetThumb() 232 return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset); // VFP addressing mode. in CanHoldStoreOffsetThumb() 234 return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset); in CanHoldStoreOffsetThumb()
|
/art/dex2oat/linker/arm/ |
D | relative_patcher_thumb2_test.cc | 122 if (!IsAligned<2u>(bne_offset)) { in BneWWithOffset() 126 if (!IsAligned<2u>(target_offset)) { in BneWWithOffset() 489 ASSERT_TRUE(IsAligned<2u>(diff)); in TEST_F() 574 ASSERT_TRUE(IsAligned<kArmAlignment>(method_after_thunk_offset)); in TEST_F() 581 ASSERT_TRUE(IsAligned<kArmAlignment>(thunk_offset)); in TEST_F() 583 ASSERT_TRUE(IsAligned<2u>(diff)); in TEST_F() 617 ASSERT_TRUE(IsAligned<2u>(diff)); in TEST_F() 918 static_assert(IsAligned<kArmAlignment>(expected_thunk_offset), "Target offset must be aligned."); in TEST_F() 929 static_assert(IsAligned<kArmAlignment>(kLiteralOffset2 + kPcAdjustment), in TEST_F() 986 static_assert(IsAligned<kArmAlignment>(expected_thunk_offset), "Target offset must be aligned."); in TEST_F() [all …]
|
/art/tools/hiddenapi/ |
D | hiddenapi.cc | 621 CHECK(IsAligned<sizeof(uint32_t)>(data_.data())); in GetSizeField() 674 CHECK(IsAligned<kMapListAlignment>(old_dex_.Size())) in Encode() 762 CHECK(IsAligned<kMapListAlignment>(data_.data())); in AllocateMemory() 763 CHECK(IsAligned<kHiddenapiClassDataAlignment>(data_.data())); in AllocateMemory() 839 CHECK(IsAligned<kHiddenapiClassDataAlignment>(payload_offset)) in AppendHiddenapiClassData()
|
/art/dex2oat/linker/arm64/ |
D | relative_patcher_arm64_test.cc | 656 ASSERT_TRUE(IsAligned<4u>(diff)); in TEST_F() 745 ASSERT_TRUE(IsAligned<kArm64Alignment>(method_after_thunk_offset)); in TEST_F() 752 ASSERT_TRUE(IsAligned<kArm64Alignment>(thunk_offset)); in TEST_F() 754 ASSERT_TRUE(IsAligned<4u>(diff)); in TEST_F() 784 ASSERT_TRUE(IsAligned<4u>(diff)); in TEST_F() 855 bool unaligned = !IsAligned<8u>((adrp_offset) + 4u + static_cast<uint32_t>(pcrel_disp)); in TEST_F() 970 bool unaligned = !IsAligned<8u>((adrp_offset) + 4u + static_cast<uint32_t>(pcrel_disp)); in TEST_F()
|
/art/runtime/gc/ |
D | verification.cc | 133 return IsAligned<kObjectAlignment>(addr) && IsAddressInHeapSpace(addr, out_space); in IsValidHeapObjectAddress()
|
/art/runtime/mirror/ |
D | string-alloc-inl.h | 181 static_assert(IsAligned<sizeof(uint16_t)>(kObjectAlignment), in Alloc()
|
/art/compiler/utils/mips64/ |
D | assembler_mips64.h | 1072 if (type == kStoreDoubleword && IsAligned<kMips64DoublewordSize>(offset)) { 1154 if (!IsAligned<kMips64DoublewordSize>(offset)) { 1193 if (!IsAligned<kMips64DoublewordSize>(offset)) { 1243 if (!IsAligned<kMips64DoublewordSize>(offset)) { 1282 if (!IsAligned<kMips64DoublewordSize>(offset)) {
|
D | assembler_mips64.cc | 3107 if (!IsAligned<sizeof(uint64_t)>(first_literal_location)) { in PromoteBranches() 3390 bool doubleword_aligned = IsAligned<kMips64DoublewordSize>(offset); in AdjustBaseAndOffset() 3479 } else if (IsAligned<kMips64DoublewordSize>(offset)) { in AdjustBaseOffsetAndElementSizeShift() 3481 } else if (IsAligned<kMips64WordSize>(offset)) { in AdjustBaseOffsetAndElementSizeShift() 3483 } else if (IsAligned<kMips64HalfwordSize>(offset)) { in AdjustBaseOffsetAndElementSizeShift()
|
/art/runtime/gc/collector/ |
D | concurrent_copying.cc | 1163 && (!IsAligned<kPageSize>(ref) || heap_->GetLargeObjectsSpace() == nullptr)) { in TestAndSetMarkBitForRef() 1192 && (!IsAligned<kPageSize>(ref) || heap_->GetLargeObjectsSpace() == nullptr)) { in TestMarkBitmapForRef() 2193 if (!IsAligned<kPageSize>(to_ref)) { in ProcessMarkStackRef() 2546 DCHECK(IsAligned<kPageSize>(region_space_->Limit())); in CaptureRssAtPeak() 2553 DCHECK(IsAligned<kPageSize>(heap_->non_moving_space_->Limit())); in CaptureRssAtPeak() 2572 DCHECK(IsAligned<kPageSize>(map.BaseSize())); in CaptureRssAtPeak() 3551 if (!IsAligned<kPageSize>(ref)) { in MarkNonMoving()
|
/art/runtime/gc/space/ |
D | image_space.cc | 435 static_assert(IsAligned<kHeapReferenceSize>(kObjectAlignment), "Object alignment check."); in VisitClass() 446 static_assert(IsAligned<kHeapReferenceSize>(sizeof(mirror::Object)), "Size alignment check."); in VisitClass() 1691 static_assert(IsAligned<kObjectAlignment>(sizeof(ImageHeader)), "Header alignment check"); in DoRelocateSpaces()
|
/art/compiler/utils/mips/ |
D | assembler_mips.h | 974 if (IsAligned<kMipsDoublewordSize>(offset)) { 1061 if (IsAligned<kMipsDoublewordSize>(offset)) {
|