/external/tensorflow/tensorflow/compiler/xla/python/ |
D | shared_device_buffer_test.cc | 34 client->backend().memory_allocator(), 0, nullptr)); in TEST() 37 EXPECT_EQ(buffer->allocator(), client->backend().memory_allocator()); in TEST() 51 client->backend().memory_allocator(), 0, nullptr)); in TEST() 55 client->backend().memory_allocator(), 0, nullptr)); in TEST() 60 client->backend().memory_allocator(), 0, nullptr)); in TEST() 66 EXPECT_EQ(tuple_buffer->allocator(), client->backend().memory_allocator()); in TEST() 81 client->backend().memory_allocator(), 0, nullptr)); in TEST() 85 client->backend().memory_allocator(), 0, nullptr)); in TEST() 90 client->backend().memory_allocator(), 0, in TEST() 95 client->backend().memory_allocator(), 0, nullptr)); in TEST() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | buffer_allocations.h | 53 se::DeviceMemoryAllocator* memory_allocator); 65 se::DeviceMemoryAllocator* memory_allocator() const { in memory_allocator() function 89 se::DeviceMemoryAllocator* memory_allocator, in BufferAllocations() argument 93 memory_allocator_(memory_allocator), in BufferAllocations()
|
D | buffer_allocations.cc | 42 se::DeviceMemoryAllocator* memory_allocator) { in Build() argument 45 num_buffers, device_ordinal, memory_allocator, buffer_assignment)); in Build() 82 buffer, memory_allocator->Allocate(device_ordinal, buffer_size)); in Build()
|
D | fft_thunk.cc | 32 int device_ordinal, se::DeviceMemoryAllocator* memory_allocator) in FftScratchAllocator() argument 33 : device_ordinal_(device_ordinal), memory_allocator_(memory_allocator) {} in FftScratchAllocator() 119 buffer_allocations.memory_allocator()); in ExecuteOnStream()
|
D | gpu_executable.cc | 306 se::DeviceMemoryAllocator* const memory_allocator = run_options->allocator(); in ExecuteAsyncOnStream() local 309 !memory_allocator->AllowsAsynchronousDeallocation(); in ExecuteAsyncOnStream() 366 assignment_.get(), executor->device_ordinal(), memory_allocator)); in ExecuteAsyncOnStream() 376 memory_allocator, device_ordinal); in ExecuteAsyncOnStream()
|
D | cudnn_fused_conv_rewriter_test.cc | 41 backend().default_stream_executor(), backend().memory_allocator()) in GetOptimizedHlo() 337 backend().default_stream_executor(), backend().memory_allocator()) in TEST_F()
|
D | fft_thunk.h | 41 se::DeviceMemoryAllocator* memory_allocator);
|
/external/v8/src/heap/ |
D | read-only-spaces.cc | 135 ->memory_allocator() in CreateReadOnlySpace() 243 heap->read_only_space()->TearDown(heap->memory_allocator()); in ReinstallReadOnlySpace() 285 void SharedReadOnlySpace::TearDown(MemoryAllocator* memory_allocator) { in TearDown() argument 294 void ReadOnlySpace::TearDown(MemoryAllocator* memory_allocator) { in TearDown() argument 296 memory_allocator->FreeReadOnlyPage(chunk); in TearDown() 321 void ReadOnlySpace::SetPermissionsForPages(MemoryAllocator* memory_allocator, in SetPermissionsForPages() argument 327 memory_allocator->page_allocator(NOT_EXECUTABLE); in SetPermissionsForPages() 377 auto* memory_allocator = heap()->memory_allocator(); in Seal() local 383 memory_allocator->UnregisterMemory(p); in Seal() 391 SetPermissionsForPages(memory_allocator, PageAllocator::kRead); in Seal() [all …]
|
D | new-spaces.cc | 60 heap()->memory_allocator()->Free<MemoryAllocator::kPooledAndQueue>( in EnsureCurrentCapacity() 71 heap()->memory_allocator()->AllocatePage<MemoryAllocator::kPooled>( in EnsureCurrentCapacity() 115 heap()->memory_allocator()->AllocatePage<MemoryAllocator::kPooled>( in Commit() 138 heap()->memory_allocator()->Free<MemoryAllocator::kPooledAndQueue>(chunk); in Uncommit() 143 heap()->memory_allocator()->unmapper()->FreeQueuedChunks(); in Uncommit() 171 heap()->memory_allocator()->AllocatePage<MemoryAllocator::kPooled>( in GrowTo() 194 heap()->memory_allocator()->Free<MemoryAllocator::kPooledAndQueue>(last); in RewindPages() 209 heap()->memory_allocator()->unmapper()->FreeQueuedChunks(); in ShrinkTo()
|
D | large-spaces.cc | 109 heap()->memory_allocator()->Free<MemoryAllocator::kFull>(page); in TearDown() 185 LargePage* page = heap()->memory_allocator()->AllocateLargePage( in AllocateLargePage() 304 heap()->memory_allocator()->PartialFreeMemory( in FreeUnmarkedObjects() 312 heap()->memory_allocator()->Free<MemoryAllocator::kPreFreeAndQueue>( in FreeUnmarkedObjects() 509 heap()->memory_allocator()->Free<MemoryAllocator::kPreFreeAndQueue>(page); in FreeDeadObjects() 521 heap()->memory_allocator()->unmapper()->FreeQueuedChunks(); in FreeDeadObjects()
|
D | read-only-spaces.h | 178 V8_EXPORT_PRIVATE virtual void TearDown(MemoryAllocator* memory_allocator); 235 void SetPermissionsForPages(MemoryAllocator* memory_allocator, 288 void TearDown(MemoryAllocator* memory_allocator) override;
|
D | paged-spaces.cc | 96 heap()->memory_allocator()->Free<MemoryAllocator::kFull>(chunk); in TearDown() 306 return heap()->memory_allocator()->AllocatePage(AreaSize(), this, in AllocatePage() 463 heap()->memory_allocator()->Free<MemoryAllocator::kPreFreeAndQueue>(page); in ReleasePage() 469 CHECK(heap()->memory_allocator()->IsMemoryChunkExecutable(page)); in SetReadable() 477 CHECK(heap()->memory_allocator()->IsMemoryChunkExecutable(page)); in SetReadAndExecutable() 485 CHECK(heap()->memory_allocator()->IsMemoryChunkExecutable(page)); in SetReadAndWritable()
|
D | memory-chunk.cc | 21 MemoryAllocator* memory_allocator = heap_->memory_allocator(); in DiscardUnusedMemory() local 23 memory_allocator->page_allocator(executable()); in DiscardUnusedMemory()
|
/external/libchrome/base/metrics/ |
D | persistent_histogram_allocator_unittest.cc | 40 allocator_ = GlobalHistogramAllocator::Get()->memory_allocator(); in CreatePersistentHistogramAllocator() 170 GlobalHistogramAllocator::Get()->memory_allocator()->Name()); in TEST_F() 176 GlobalHistogramAllocator::Get()->memory_allocator()->Name()); in TEST_F() 182 GlobalHistogramAllocator::Get()->memory_allocator()->Name()); in TEST_F() 261 const_cast<void*>(new_allocator->memory_allocator()->data()), in TEST_F() 262 new_allocator->memory_allocator()->size(), 0, 0, "", false)); in TEST_F() 308 const_cast<void*>(new_allocator->memory_allocator()->data()), in TEST_F() 309 new_allocator->memory_allocator()->size(), 0, 0, "", false)); in TEST_F()
|
D | persistent_histogram_allocator.cc | 259 : allocator_(allocator), memory_iter_(allocator->memory_allocator()) {} in Iterator() 914 PersistentMemoryAllocator* memory_allocator = in ReleaseForTesting() local 915 histogram_allocator->memory_allocator(); in ReleaseForTesting() 920 PersistentMemoryAllocator::Iterator iter(memory_allocator); in ReleaseForTesting() 964 memory_allocator()->SetMemoryState(PersistentMemoryAllocator::MEMORY_DELETED); in DeletePersistentLocation()
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
D | buffer_donation_test.cc | 73 auto memory_allocator = in RunAndCheck() local 78 run_options.set_allocator(memory_allocator.get()); in RunAndCheck() 85 argument_literal.shape(), memory_allocator.get(), in RunAndCheck() 96 memory_allocator.get()); in RunAndCheck()
|
/external/tensorflow/tensorflow/compiler/xrt/ |
D | xrt_state.cc | 96 shape, on_device_shape, backend->memory_allocator(), device_ordinal); in AllocateScopedShapedBuffer() 188 device_ordinal, backend->memory_allocator(), in CreateAndTransfer() 191 ->InitializeFromShapedBuffer(shaped_buffer, backend->memory_allocator(), in CreateAndTransfer() 211 device_ordinal, backend->memory_allocator(), in CreateUninitialized() 214 ->InitializeFromShapedBuffer(shaped_buffer, backend->memory_allocator(), in CreateUninitialized() 224 auto allocator = backend->memory_allocator(); in CreateFromBuffer() 318 InitializeFromShapedBuffer(shaped_buffer, backend->memory_allocator(), in SwapIn() 459 auto allocator = backend->memory_allocator(); in MakeTuple()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | cpu_executable.cc | 82 se::DeviceMemoryAllocator* memory_allocator, int device_ordinal, in CreateBufferTable() argument 122 TF_ASSIGN_OR_RETURN(owning_buffers[i], memory_allocator->Allocate( in CreateBufferTable() 313 se::DeviceMemoryAllocator* memory_allocator = run_options->allocator(); in ExecuteAsyncOnStream() local 319 CreateBufferTable(memory_allocator, stream->parent()->device_ordinal(), in ExecuteAsyncOnStream()
|
D | cpu_executable.h | 105 CreateBufferTable(se::DeviceMemoryAllocator* memory_allocator,
|
/external/tensorflow/tensorflow/compiler/jit/ |
D | xla_tensor.cc | 55 client->backend().memory_allocator(), in AllocateShapedBuffer() 63 client->backend().memory_allocator()->Allocate( in AllocateShapedBuffer()
|
D | xla_compile_on_demand_op.cc | 61 client, client->backend().memory_allocator(), in Run() 78 run_options.set_allocator(client->backend().memory_allocator()); in Run()
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | hlo_runner.cc | 102 literal.shape(), backend().memory_allocator(), in TransferLiteralToDevice() 312 argument->shape(), backend().memory_allocator(), device)); in ExecuteReplicated() 438 backend().memory_allocator())); in CreateExecutable() 443 backend().memory_allocator()); in CreateExecutable() 452 run_options.set_allocator(backend().memory_allocator()); in GetServiceRunOptionsForDevice()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/tests/ |
D | gpu_convolution_regression_test.cc | 31 backend().default_stream_executor(), backend().memory_allocator()); in CheckForHloText()
|
/external/tensorflow/tensorflow/stream_executor/gpu/ |
D | redzone_allocator.h | 45 RedzoneAllocator(Stream* stream, DeviceMemoryAllocator* memory_allocator,
|
/external/tensorflow/tensorflow/compiler/xla/client/ |
D | local_client.cc | 158 run_options.set_allocator(backend_->memory_allocator()); in RunHelper() 370 allocator = backend().memory_allocator(); in LiteralToShapedBuffer() 423 shape, backend().memory_allocator(), device_ordinal)); in TransferToLocalServer()
|