Home
last modified time | relevance | path

Searched refs:DEVICE_MEMORY (Results 1 – 24 of 24) sorted by relevance

/external/tensorflow/tensorflow/core/framework/
Dmemory_types_test.cc77 MemoryTypeVector({HOST_MEMORY, DEVICE_MEMORY, DEVICE_MEMORY, HOST_MEMORY, in TEST()
78 HOST_MEMORY, HOST_MEMORY, DEVICE_MEMORY, DEVICE_MEMORY, in TEST()
79 DEVICE_MEMORY, HOST_MEMORY, HOST_MEMORY, HOST_MEMORY}), in TEST()
82 EXPECT_EQ(MemoryTypeVector({DEVICE_MEMORY, DEVICE_MEMORY, DEVICE_MEMORY, in TEST()
83 DEVICE_MEMORY, DEVICE_MEMORY, DEVICE_MEMORY, in TEST()
84 HOST_MEMORY, HOST_MEMORY, DEVICE_MEMORY}), in TEST()
90 MemoryTypeVector({HOST_MEMORY, HOST_MEMORY, DEVICE_MEMORY, HOST_MEMORY, in TEST()
94 EXPECT_EQ(MemoryTypeVector({DEVICE_MEMORY, DEVICE_MEMORY, DEVICE_MEMORY, in TEST()
96 HOST_MEMORY, HOST_MEMORY, DEVICE_MEMORY}), in TEST()
Dmemory_types.cc75 : DEVICE_MEMORY; in MTypeFromDType()
79 return DataTypeAlwaysOnHost(dtype) ? HOST_MEMORY : DEVICE_MEMORY; in MTypeFromDTypeIntsOnDevice()
123 inp_mtypes->resize(GetTotal(inp_names), DEVICE_MEMORY); in MemoryTypesForNode()
124 out_mtypes->resize(GetTotal(out_names), DEVICE_MEMORY); in MemoryTypesForNode()
139 inp_mtypes->resize(inp_dtypes.size(), DEVICE_MEMORY); in MemoryTypesForNode()
140 out_mtypes->resize(out_dtypes.size(), DEVICE_MEMORY); in MemoryTypesForNode()
Dtypes.h48 DEVICE_MEMORY = 0, enumerator
Dop_kernel.cc573 type, shape, DEVICE_MEMORY, allocator_attr); in forward_input_or_allocate_temp()
/external/swiftshader/src/Vulkan/
DVkSpecializationInfo.cpp25 allocate(sizeof(VkSpecializationInfo), REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY)); in SpecializationInfo()
34 allocate(entriesSize, REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY)); in SpecializationInfo()
42 void *data = allocate(specializationInfo->dataSize, REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY); in SpecializationInfo()
57 deallocate(const_cast<VkSpecializationMapEntry *>(info->pMapEntries), DEVICE_MEMORY); in operator ()()
58 deallocate(const_cast<void *>(info->pData), DEVICE_MEMORY); in operator ()()
59 deallocate(info, DEVICE_MEMORY); in operator ()()
DVkCommandPool.cpp32 vk::destroy(commandBuffer, DEVICE_MEMORY); in destroy()
47 DEVICE_MEMORY, DispatchableCommandBuffer::GetAllocationScope()); in allocateCommandBuffers()
58 vk::destroy(pCommandBuffers[j], DEVICE_MEMORY); in allocateCommandBuffers()
78 vk::destroy(pCommandBuffers[i], DEVICE_MEMORY); in freeCommandBuffers()
DVkDeviceMemory.cpp78 buffer = vk::allocate(size, REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY); in allocate()
90 vk::deallocate(buffer, DEVICE_MEMORY); in deallocate()
DVkQueue.cpp65 …vk::allocate(totalSize, vk::REQUIRED_MEMORY_ALIGNMENT, vk::DEVICE_MEMORY, vk::Fence::GetAllocation… in DeepCopySubmitInfo()
306 vk::deallocate(v.first, DEVICE_MEMORY); in garbageCollect()
DVkObject.hpp44 static constexpr VkAllocationCallbacks *DEVICE_MEMORY = nullptr; variable
DVkDeviceMemoryExternalAndroid.cpp467 result = vk::Image::Create(vk::DEVICE_MEMORY, &info, &Image, vk::Cast(device)); in GetAndroidHardwareBufferProperties()
474 vk::destroy(Image, vk::DEVICE_MEMORY); in GetAndroidHardwareBufferProperties()
/external/swiftshader/src/WSI/
DVkSurfaceKHR.cpp44 …= reinterpret_cast<VkImage *>(allocate(sizeof(VkImage), REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY)); in allocateImage()
53 deallocate(vkImagePtr, DEVICE_MEMORY); in allocateImage()
58 deallocate(vkImagePtr, DEVICE_MEMORY); in allocateImage()
68 allocate(sizeof(VkDeviceMemory), REQUIRED_MEMORY_ALIGNMENT, DEVICE_MEMORY)); in allocateAndBindImageMemory()
77 deallocate(vkDeviceMemoryPtr, DEVICE_MEMORY); in allocateAndBindImageMemory()
85 deallocate(vkDeviceMemoryPtr, DEVICE_MEMORY); in allocateAndBindImageMemory()
/external/tensorflow/tensorflow/core/common_runtime/
Dmemory_types.cc82 DEVICE_MEMORY); in ProcessMemoryTypes()
84 DEVICE_MEMORY); in ProcessMemoryTypes()
167 if (((sm == HOST_MEMORY) && (dm == DEVICE_MEMORY)) || in EnsureMemoryTypes()
168 ((sm == DEVICE_MEMORY) && (dm == HOST_MEMORY))) { in EnsureMemoryTypes()
Dmemory_types_test.cc70 EXPECT_EQ(memory_type, DEVICE_MEMORY); in TEST()
Dconstant_folding.cc532 (memory_type == DEVICE_MEMORY && is_int32)) { in ReplaceTensorWithConstant()
/external/tensorflow/tensorflow/compiler/jit/
Dxla_kernel_creator_test.cc119 EXPECT_EQ(DEVICE_MEMORY, kernel_->input_memory_types()[0]); in TEST_F()
124 EXPECT_EQ(DEVICE_MEMORY, kernel_->output_memory_types()[0]); in TEST_F()
Dcompilability_check_util.cc662 tensorflow::DEVICE_MEMORY); in GetInputMemoryTypes()
685 tensorflow::DEVICE_MEMORY); in GetOutputMemoryTypes()
Dpartially_decluster_pass.cc84 if (output_mtypes[e->src_output()] == DEVICE_MEMORY) { in FindNodesToDecluster()
Dbuild_xla_ops_pass.cc438 if (source_output_mem_types[e->src_output()] == DEVICE_MEMORY) { in PredicateInt32Inputs()
/external/tensorflow/tensorflow/core/kernels/
Dassign_op.h104 rhs.dtype(), rhs.shape(), DEVICE_MEMORY, attr); in Compute()
Dscatter_nd_op.cc197 MemoryType memory_type = DEVICE_MEMORY; in Compute()
202 memory_type = DEVICE_MEMORY; in Compute()
Dlist_kernels.cc466 DEVICE_MEMORY /* input is always on DEVICE_MEMORY */, attr); in Compute()
Dresource_variable_ops.cc453 DEVICE_MEMORY /* HOST_MEMORY is only reserved for special cases */, in Compute()
Dlist_kernels.h914 DEVICE_MEMORY /* input is always on DEVICE_MEMORY */, attr); in Compute()
/external/swiftshader/src/Device/
DRenderer.cpp175 …return vk::allocate(sizeof(Renderer), alignof(Renderer), vk::DEVICE_MEMORY, VK_SYSTEM_ALLOCATION_S… in operator new()
180 vk::deallocate(mem, vk::DEVICE_MEMORY); in operator delete()