Home
last modified time | relevance | path

Searched refs:IsAligned (Results 1 – 25 of 158) sorted by relevance

1234567

/external/v8/src/base/
Dbounded-page-allocator.cc18 CHECK(IsAligned(allocate_page_size, page_allocator->AllocatePageSize())); in BoundedPageAllocator()
19 CHECK(IsAligned(allocate_page_size_, commit_page_size_)); in BoundedPageAllocator()
32 CHECK(IsAligned(alignment, region_allocator_.page_size())); in AllocatePages()
50 CHECK(IsAligned(address, allocate_page_size_)); in AllocatePagesAt()
51 CHECK(IsAligned(size, allocate_page_size_)); in AllocatePagesAt()
65 CHECK(IsAligned(address, allocate_page_size_)); in ReserveForSharedMemoryMapping()
66 CHECK(IsAligned(size, commit_page_size_)); in ReserveForSharedMemoryMapping()
96 CHECK(IsAligned(address, allocate_page_size_)); in ReleasePages()
99 DCHECK(IsAligned(size - new_size, commit_page_size_)); in ReleasePages()
128 DCHECK(IsAligned(reinterpret_cast<Address>(address), commit_page_size_)); in SetPermissions()
[all …]
Dregion-allocator.cc32 CHECK(IsAligned(size(), page_size_)); in RegionAllocator()
33 CHECK(IsAligned(begin(), page_size_)); in RegionAllocator()
86 DCHECK(IsAligned(new_size, page_size_)); in Split()
126 DCHECK(IsAligned(size, page_size_)); in AllocateRegion()
134 DCHECK(IsAligned(region->begin(), page_size_)); in AllocateRegion()
164 DCHECK(IsAligned(requested_address, page_size_)); in AllocateRegionAt()
166 DCHECK(IsAligned(size, page_size_)); in AllocateRegionAt()
187 DCHECK(IsAligned(new_size, page_size_)); in AllocateRegionAt()
204 DCHECK(IsAligned(new_size, page_size_)); in TrimRegion()
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/
DTensorEvaluator.h41 IsAligned = Derived::IsAligned, enumerator
161 IsAligned = Derived::IsAligned,
233 IsAligned = true,
298 IsAligned = TensorEvaluator<ArgType, Device>::IsAligned,
366IsAligned = TensorEvaluator<LeftArgType, Device>::IsAligned & TensorEvaluator<RightArgType, Device…
446IsAligned = TensorEvaluator<Arg1Type, Device>::IsAligned & TensorEvaluator<Arg2Type, Device>::IsAl…
550IsAligned = TensorEvaluator<ThenArgType, Device>::IsAligned & TensorEvaluator<ElseArgType, Device>…
DTensorAssign.h100IsAligned = TensorEvaluator<LeftArgType, Device>::IsAligned & TensorEvaluator<RightArgType, Device…
139 const int LhsStoreMode = TensorEvaluator<LeftArgType, Device>::IsAligned ? Aligned : Unaligned;
140 const int RhsLoadMode = TensorEvaluator<RightArgType, Device>::IsAligned ? Aligned : Unaligned;
DTensorLayoutSwap.h119 IsAligned = TensorEvaluator<ArgType, Device>::IsAligned,
181 IsAligned = TensorEvaluator<ArgType, Device>::IsAligned,
DTensorEvalTo.h104 IsAligned = TensorEvaluator<ArgType, Device>::IsAligned,
137 …(m_buffer + i, m_impl.template packet<TensorEvaluator<ArgType, Device>::IsAligned ? Aligned : Unal…
/external/webrtc/webrtc/system_wrappers/source/
Daligned_array_unittest.cc19 bool IsAligned(const void* ptr, size_t alignment) { in IsAligned() function
29 ASSERT_TRUE(IsAligned(arr.Array(), 128)); in TEST()
31 ASSERT_TRUE(IsAligned(arr.Row(i), 128)); in TEST()
/external/angle/src/libANGLE/renderer/vulkan/shaders/src/
DBufferUtils.comp11 // * IsAligned: if true, assumes the workgroup size divides the buffer size, so there is no
54 // size: used if !IsAligned
65 #if !IsAligned
68 #endif // IsAligned
/external/v8/src/utils/
Dallocation.cc169 DCHECK(IsAligned(size, page_allocator->AllocatePageSize())); in AllocatePages()
186 DCHECK(IsAligned(size, page_allocator->AllocatePageSize())); in FreePages()
194 DCHECK(IsAligned(new_size, page_allocator->CommitPageSize())); in ReleasePages()
219 DCHECK(IsAligned(size, page_allocator_->CommitPageSize())); in VirtualMemory()
228 DCHECK(IsAligned(address, alignment)); in VirtualMemory()
255 DCHECK(IsAligned(free_start, page_allocator_->CommitPageSize())); in Release()
/external/v8/src/objects/
Dfield-index-inl.h16 DCHECK_IMPLIES(encoding == kWord32, IsAligned(offset, kInt32Size)); in ForInObjectOffset()
17 DCHECK_IMPLIES(encoding == kTagged, IsAligned(offset, kTaggedSize)); in ForInObjectOffset()
18 DCHECK_IMPLIES(encoding == kDouble, IsAligned(offset, kDoubleSize)); in ForInObjectOffset()
Djs-array-buffer.h243 STATIC_ASSERT(IsAligned(kByteOffsetOffset, kUIntptrSize));
244 STATIC_ASSERT(IsAligned(kByteLengthOffset, kUIntptrSize));
321 STATIC_ASSERT(IsAligned(kLengthOffset, kUIntptrSize));
322 STATIC_ASSERT(IsAligned(kExternalPointerOffset, kSystemPointerSize));
365 STATIC_ASSERT(IsAligned(kDataPointerOffset, kUIntptrSize));
Dforeign.h34 STATIC_ASSERT(IsAligned(kForeignAddressOffset, kTaggedSize));
36 STATIC_ASSERT(IsAligned(kForeignAddressOffset, kExternalPointerSize));
Dfield-index.h47 DCHECK(IsAligned(offset(), kTaggedSize)); in index()
79 DCHECK(IsAligned(first_inobject_property_offset, kTaggedSize)); in FieldIndex()
/external/v8/src/zone/
Daccounting-allocator.cc29 DCHECK(IsAligned(ZoneCompression::kReservationSize, in ReserveAddressSpace()
39 CHECK(IsAligned(memory.address(), ZoneCompression::kReservationAlignment)); in ReserveAddressSpace()
52 CHECK(IsAligned(reservation_start, ZoneCompression::kReservationAlignment)); in CreateBoundedAllocator()
Dzone-handle-set.h21 DCHECK(IsAligned(handle.address(), kPointerAlignment)); in ZoneHandleSet()
45 DCHECK(IsAligned(reinterpret_cast<Address>(value), kPointerAlignment)); in insert()
58 DCHECK(IsAligned(reinterpret_cast<Address>(list), kPointerAlignment)); in insert()
79 DCHECK(IsAligned(reinterpret_cast<Address>(new_list), kPointerAlignment)); in insert()
/external/tensorflow/tensorflow/core/framework/
Dtensor_interface.h47 virtual bool IsAligned() const = 0;
66 bool IsAligned() const override;
Dtensor.h282 bool IsAligned() const { in IsAligned() function
728 CHECK(IsAligned()); in bit_casted_tensor()
735 CHECK(IsAligned()); in bit_casted_tensor()
745 CHECK(IsAligned()); in reinterpret_last_dimension()
761 CHECK(IsAligned()); in reinterpret_last_dimension()
820 CHECK(IsAligned()); in bit_casted_shaped()
839 CHECK(IsAligned()); in shaped()
848 CHECK(IsAligned()); in bit_casted_shaped()
/external/libchrome/mojo/core/
Doptions_validation.h35 CHECK(options && IsAligned<MOJO_ALIGNOF(Options)>(options)); in UserOptionsReader()
73 static bool IsAligned(const void* pointer) { in IsAligned() function
/external/v8/src/strings/
Dstring-case.cc79 DCHECK(IsAligned(reinterpret_cast<Address>(dst), sizeof(word_t))); in FastAsciiConvert()
81 if (IsAligned(reinterpret_cast<Address>(src), sizeof(word_t))) { in FastAsciiConvert()
/external/compiler-rt/lib/sanitizer_common/
Dsanitizer_stacktrace.cc80 IsAligned((uptr)frame, sizeof(*frame)) && in FastUnwindStack()
88 !IsAligned((uptr)caller_frame, sizeof(uhwptr))) in FastUnwindStack()
/external/libchrome/mojo/public/cpp/bindings/lib/
Dbuffer.cc19 DCHECK(IsAligned(data_)); in Buffer()
31 DCHECK(IsAligned(data_)); in Buffer()
Dvalidation_util.cc32 if (!IsAligned(data)) { in ValidateStructHeaderAndClaimMemory()
63 if (!IsAligned(data)) { in ValidateNonInlinedUnionHeaderAndClaimMemory()
Dserialization.h104 bool need_copy = !IsAligned(input_buffer);
107 DCHECK(IsAligned(aligned_input_buffer));
/external/vixl/src/
Dcode-buffer-vixl.h139 bool Is16bitAligned() const { return IsAligned<2>(cursor_); } in Is16bitAligned()
141 bool Is32bitAligned() const { return IsAligned<4>(cursor_); } in Is32bitAligned()
/external/tensorflow/tensorflow/core/kernels/
Dops_util_test.cc334 EXPECT_EQ(sliced.IsAligned(), true); in TEST_F()
348 EXPECT_EQ(sliced.IsAligned(), false); in TEST_F()
372 EXPECT_EQ(sliced.IsAligned(), true); in TEST_F()
388 EXPECT_EQ(sliced.IsAligned(), false); in TEST_F()

1234567