/external/v8/src/base/ |
D | bounded-page-allocator.cc | 18 CHECK(IsAligned(allocate_page_size, page_allocator->AllocatePageSize())); in BoundedPageAllocator() 19 CHECK(IsAligned(allocate_page_size_, commit_page_size_)); in BoundedPageAllocator() 32 CHECK(IsAligned(alignment, region_allocator_.page_size())); in AllocatePages() 50 CHECK(IsAligned(address, allocate_page_size_)); in AllocatePagesAt() 51 CHECK(IsAligned(size, allocate_page_size_)); in AllocatePagesAt() 65 CHECK(IsAligned(address, allocate_page_size_)); in ReserveForSharedMemoryMapping() 66 CHECK(IsAligned(size, commit_page_size_)); in ReserveForSharedMemoryMapping() 96 CHECK(IsAligned(address, allocate_page_size_)); in ReleasePages() 99 DCHECK(IsAligned(size - new_size, commit_page_size_)); in ReleasePages() 128 DCHECK(IsAligned(reinterpret_cast<Address>(address), commit_page_size_)); in SetPermissions() [all …]
|
D | region-allocator.cc | 32 CHECK(IsAligned(size(), page_size_)); in RegionAllocator() 33 CHECK(IsAligned(begin(), page_size_)); in RegionAllocator() 86 DCHECK(IsAligned(new_size, page_size_)); in Split() 126 DCHECK(IsAligned(size, page_size_)); in AllocateRegion() 134 DCHECK(IsAligned(region->begin(), page_size_)); in AllocateRegion() 164 DCHECK(IsAligned(requested_address, page_size_)); in AllocateRegionAt() 166 DCHECK(IsAligned(size, page_size_)); in AllocateRegionAt() 187 DCHECK(IsAligned(new_size, page_size_)); in AllocateRegionAt() 204 DCHECK(IsAligned(new_size, page_size_)); in TrimRegion()
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
D | TensorEvaluator.h | 41 IsAligned = Derived::IsAligned, enumerator 161 IsAligned = Derived::IsAligned, 233 IsAligned = true, 298 IsAligned = TensorEvaluator<ArgType, Device>::IsAligned, 366 …IsAligned = TensorEvaluator<LeftArgType, Device>::IsAligned & TensorEvaluator<RightArgType, Device… 446 …IsAligned = TensorEvaluator<Arg1Type, Device>::IsAligned & TensorEvaluator<Arg2Type, Device>::IsAl… 550 …IsAligned = TensorEvaluator<ThenArgType, Device>::IsAligned & TensorEvaluator<ElseArgType, Device>…
|
D | TensorAssign.h | 100 …IsAligned = TensorEvaluator<LeftArgType, Device>::IsAligned & TensorEvaluator<RightArgType, Device… 139 const int LhsStoreMode = TensorEvaluator<LeftArgType, Device>::IsAligned ? Aligned : Unaligned; 140 const int RhsLoadMode = TensorEvaluator<RightArgType, Device>::IsAligned ? Aligned : Unaligned;
|
D | TensorLayoutSwap.h | 119 IsAligned = TensorEvaluator<ArgType, Device>::IsAligned, 181 IsAligned = TensorEvaluator<ArgType, Device>::IsAligned,
|
D | TensorEvalTo.h | 104 IsAligned = TensorEvaluator<ArgType, Device>::IsAligned, 137 …(m_buffer + i, m_impl.template packet<TensorEvaluator<ArgType, Device>::IsAligned ? Aligned : Unal…
|
/external/webrtc/webrtc/system_wrappers/source/ |
D | aligned_array_unittest.cc | 19 bool IsAligned(const void* ptr, size_t alignment) { in IsAligned() function 29 ASSERT_TRUE(IsAligned(arr.Array(), 128)); in TEST() 31 ASSERT_TRUE(IsAligned(arr.Row(i), 128)); in TEST()
|
/external/angle/src/libANGLE/renderer/vulkan/shaders/src/ |
D | BufferUtils.comp | 11 // * IsAligned: if true, assumes the workgroup size divides the buffer size, so there is no 54 // size: used if !IsAligned 65 #if !IsAligned 68 #endif // IsAligned
|
/external/v8/src/utils/ |
D | allocation.cc | 169 DCHECK(IsAligned(size, page_allocator->AllocatePageSize())); in AllocatePages() 186 DCHECK(IsAligned(size, page_allocator->AllocatePageSize())); in FreePages() 194 DCHECK(IsAligned(new_size, page_allocator->CommitPageSize())); in ReleasePages() 219 DCHECK(IsAligned(size, page_allocator_->CommitPageSize())); in VirtualMemory() 228 DCHECK(IsAligned(address, alignment)); in VirtualMemory() 255 DCHECK(IsAligned(free_start, page_allocator_->CommitPageSize())); in Release()
|
/external/v8/src/objects/ |
D | field-index-inl.h | 16 DCHECK_IMPLIES(encoding == kWord32, IsAligned(offset, kInt32Size)); in ForInObjectOffset() 17 DCHECK_IMPLIES(encoding == kTagged, IsAligned(offset, kTaggedSize)); in ForInObjectOffset() 18 DCHECK_IMPLIES(encoding == kDouble, IsAligned(offset, kDoubleSize)); in ForInObjectOffset()
|
D | js-array-buffer.h | 243 STATIC_ASSERT(IsAligned(kByteOffsetOffset, kUIntptrSize)); 244 STATIC_ASSERT(IsAligned(kByteLengthOffset, kUIntptrSize)); 321 STATIC_ASSERT(IsAligned(kLengthOffset, kUIntptrSize)); 322 STATIC_ASSERT(IsAligned(kExternalPointerOffset, kSystemPointerSize)); 365 STATIC_ASSERT(IsAligned(kDataPointerOffset, kUIntptrSize));
|
D | foreign.h | 34 STATIC_ASSERT(IsAligned(kForeignAddressOffset, kTaggedSize)); 36 STATIC_ASSERT(IsAligned(kForeignAddressOffset, kExternalPointerSize));
|
D | field-index.h | 47 DCHECK(IsAligned(offset(), kTaggedSize)); in index() 79 DCHECK(IsAligned(first_inobject_property_offset, kTaggedSize)); in FieldIndex()
|
/external/v8/src/zone/ |
D | accounting-allocator.cc | 29 DCHECK(IsAligned(ZoneCompression::kReservationSize, in ReserveAddressSpace() 39 CHECK(IsAligned(memory.address(), ZoneCompression::kReservationAlignment)); in ReserveAddressSpace() 52 CHECK(IsAligned(reservation_start, ZoneCompression::kReservationAlignment)); in CreateBoundedAllocator()
|
D | zone-handle-set.h | 21 DCHECK(IsAligned(handle.address(), kPointerAlignment)); in ZoneHandleSet() 45 DCHECK(IsAligned(reinterpret_cast<Address>(value), kPointerAlignment)); in insert() 58 DCHECK(IsAligned(reinterpret_cast<Address>(list), kPointerAlignment)); in insert() 79 DCHECK(IsAligned(reinterpret_cast<Address>(new_list), kPointerAlignment)); in insert()
|
/external/tensorflow/tensorflow/core/framework/ |
D | tensor_interface.h | 47 virtual bool IsAligned() const = 0; 66 bool IsAligned() const override;
|
D | tensor.h | 282 bool IsAligned() const { in IsAligned() function 728 CHECK(IsAligned()); in bit_casted_tensor() 735 CHECK(IsAligned()); in bit_casted_tensor() 745 CHECK(IsAligned()); in reinterpret_last_dimension() 761 CHECK(IsAligned()); in reinterpret_last_dimension() 820 CHECK(IsAligned()); in bit_casted_shaped() 839 CHECK(IsAligned()); in shaped() 848 CHECK(IsAligned()); in bit_casted_shaped()
|
/external/libchrome/mojo/core/ |
D | options_validation.h | 35 CHECK(options && IsAligned<MOJO_ALIGNOF(Options)>(options)); in UserOptionsReader() 73 static bool IsAligned(const void* pointer) { in IsAligned() function
|
/external/v8/src/strings/ |
D | string-case.cc | 79 DCHECK(IsAligned(reinterpret_cast<Address>(dst), sizeof(word_t))); in FastAsciiConvert() 81 if (IsAligned(reinterpret_cast<Address>(src), sizeof(word_t))) { in FastAsciiConvert()
|
/external/compiler-rt/lib/sanitizer_common/ |
D | sanitizer_stacktrace.cc | 80 IsAligned((uptr)frame, sizeof(*frame)) && in FastUnwindStack() 88 !IsAligned((uptr)caller_frame, sizeof(uhwptr))) in FastUnwindStack()
|
/external/libchrome/mojo/public/cpp/bindings/lib/ |
D | buffer.cc | 19 DCHECK(IsAligned(data_)); in Buffer() 31 DCHECK(IsAligned(data_)); in Buffer()
|
D | validation_util.cc | 32 if (!IsAligned(data)) { in ValidateStructHeaderAndClaimMemory() 63 if (!IsAligned(data)) { in ValidateNonInlinedUnionHeaderAndClaimMemory()
|
D | serialization.h | 104 bool need_copy = !IsAligned(input_buffer); 107 DCHECK(IsAligned(aligned_input_buffer));
|
/external/vixl/src/ |
D | code-buffer-vixl.h | 139 bool Is16bitAligned() const { return IsAligned<2>(cursor_); } in Is16bitAligned() 141 bool Is32bitAligned() const { return IsAligned<4>(cursor_); } in Is32bitAligned()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | ops_util_test.cc | 334 EXPECT_EQ(sliced.IsAligned(), true); in TEST_F() 348 EXPECT_EQ(sliced.IsAligned(), false); in TEST_F() 372 EXPECT_EQ(sliced.IsAligned(), true); in TEST_F() 388 EXPECT_EQ(sliced.IsAligned(), false); in TEST_F()
|