/external/tensorflow/tensorflow/lite/delegates/gpu/metal/ |
D | buffer_test.mm | 44 std::vector<float> gpu_data; 45 XCTAssertTrue(buffer.ReadData<float>(&gpu_data).ok()); 47 XCTAssertEqual(gpu_data.size(), data.size()); 48 for (int i = 0; i < gpu_data.size(); ++i) { 49 XCTAssertEqual(gpu_data[i], data[i]); 61 std::vector<half> gpu_data; 62 XCTAssertTrue(buffer.ReadData<half>(&gpu_data).ok()); 64 XCTAssertEqual(gpu_data.size(), data.size()); 65 for (int i = 0; i < gpu_data.size(); ++i) { 66 XCTAssertEqual(gpu_data[i], data[i]);
|
/external/tensorflow/tensorflow/lite/delegates/gpu/cl/ |
D | buffer_test.cc | 39 std::vector<float> gpu_data; in TEST_F() local 40 ASSERT_OK(buffer.ReadData<float>(env_.queue(), &gpu_data)); in TEST_F() 42 EXPECT_THAT(gpu_data, Pointwise(FloatNear(0.0f), data)); in TEST_F() 51 std::vector<half> gpu_data; in TEST_F() local 52 ASSERT_OK(buffer.ReadData<half>(env_.queue(), &gpu_data)); in TEST_F() 54 EXPECT_THAT(gpu_data, Pointwise(FloatNear(0.0f), data)); in TEST_F()
|
/external/eigen/unsupported/test/ |
D | cxx11_tensor_sycl.cpp | 80 DataType* gpu_data = static_cast<DataType*>(sycl_device.allocate(in1.size()*sizeof(DataType))); in test_sycl_mem_sync() local 82 TensorMap<Tensor<DataType, 1, DataLayout, IndexType>> gpu1(gpu_data, tensorRange); in test_sycl_mem_sync() 83 sycl_device.memcpyHostToDevice(gpu_data, in1.data(),(in1.size())*sizeof(DataType)); in test_sycl_mem_sync() 87 sycl_device.memcpyDeviceToHost(out.data(), gpu_data, out.size()*sizeof(DataType)); in test_sycl_mem_sync() 94 sycl_device.deallocate(gpu_data); in test_sycl_mem_sync() 106 DataType* gpu_data = static_cast<DataType*>(sycl_device.allocate(full_size * sizeof(DataType))); in test_sycl_mem_sync_offsets() local 107 TensorMap<tensor_type> gpu1(gpu_data, tensorRange); in test_sycl_mem_sync_offsets() 111 sycl_device.memcpyHostToDevice(gpu_data, in1.data(), full_size * sizeof(DataType)); in test_sycl_mem_sync_offsets() 112 sycl_device.memcpyDeviceToHost(out.data(), gpu_data + half_size, half_size * sizeof(DataType)); in test_sycl_mem_sync_offsets() 113 sycl_device.memcpyDeviceToHost(out.data() + half_size, gpu_data, half_size * sizeof(DataType)); in test_sycl_mem_sync_offsets() [all …]
|
D | cxx11_tensor_device_sycl.cpp | 49 DataType* gpu_data = static_cast<DataType*>(sycl_device.allocate(sizeDim1*sizeof(DataType))); in test_device_exceptions() local 50 sycl_device.memset(gpu_data, 1, sizeDim1*sizeof(DataType)); in test_device_exceptions() 52 TensorMap<Tensor<DataType, 1, DataLayout,IndexType>> in(gpu_data, tensorDims); in test_device_exceptions() 53 TensorMap<Tensor<DataType, 1, DataLayout,IndexType>> out(gpu_data, tensorDims); in test_device_exceptions() 58 sycl_device.deallocate(gpu_data); in test_device_exceptions()
|
D | cxx11_tensor_builtins_sycl.cpp | 64 DataType *gpu_data = static_cast<DataType *>( in test_unary_builtins_for_scalar() local 68 TensorMap<Tensor<DataType, 3, DataLayout, int64_t>> gpu(gpu_data, tensor_range); in test_unary_builtins_for_scalar() 70 sycl_device.memcpyHostToDevice(gpu_data, in.data(), in test_unary_builtins_for_scalar() 83 sycl_device.deallocate(gpu_data); in test_unary_builtins_for_scalar() 171 DataType *gpu_data = static_cast<DataType *>( in test_unary_builtins_return_bool() local 175 TensorMap<Tensor<DataType, 3, DataLayout, int64_t>> gpu(gpu_data, tensor_range); in test_unary_builtins_return_bool() 177 sycl_device.memcpyHostToDevice(gpu_data, in.data(), in test_unary_builtins_return_bool() 185 sycl_device.deallocate(gpu_data); in test_unary_builtins_return_bool()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/ |
D | convolution_transposed_thin.h | 88 float4* gpu_data = reinterpret_cast<float4*>(desc.data.data()); in UploadData() local 89 RearrangeWeightsData(weights, absl::MakeSpan(gpu_data, flt4_count)); in UploadData() 94 gpu_data[flt4_count] = bias_value; in UploadData() 96 half4* gpu_data = reinterpret_cast<half4*>(desc.data.data()); in UploadData() local 97 RearrangeWeightsData(weights, absl::MakeSpan(gpu_data, flt4_count)); in UploadData() 102 gpu_data[flt4_count] = bias_value; in UploadData()
|
D | conv_metal_simd.cc | 384 float* gpu_data = reinterpret_cast<float*>(result.data()); in ReorderWeightsForConv() local 386 gpu_data[i] = weights_gpu[i]; in ReorderWeightsForConv() 389 half* gpu_data = reinterpret_cast<half*>(result.data()); in ReorderWeightsForConv() local 391 gpu_data[i] = weights_gpu[i]; in ReorderWeightsForConv() 402 float* gpu_data = reinterpret_cast<float*>(result.data()); in ReorderBiasesForConv() local 404 gpu_data[i] = i < biases.shape.v ? biases.data[i] : 0.0f; in ReorderBiasesForConv() 407 half* gpu_data = reinterpret_cast<half*>(result.data()); in ReorderBiasesForConv() local 409 gpu_data[i] = i < biases.shape.v ? biases.data[i] : 0.0f; in ReorderBiasesForConv()
|
D | conv_generic.h | 248 float* gpu_data = reinterpret_cast<float*>(desc.data.data()); in UploadBias() local 250 gpu_data[i] = i < bias.shape.v ? bias.data[i] : 0.0f; in UploadBias() 253 half* gpu_data = reinterpret_cast<half*>(desc.data.data()); in UploadBias() local 255 gpu_data[i] = i < bias.shape.v ? bias.data[i] : 0.0f; in UploadBias()
|
/external/angle/src/libANGLE/ |
D | gen_extensions.py | 348 def get_ext_support(ext_name, gpu_data): argument 355 return ' | '.join([s(ext_name, support) for support in gpu_data]) 364 def format_md_gpu_info(gpu_data): argument 365 return _MD_CONFIG_INFO_TEMPLATE.format(**gpu_data) 462 md_gpu_info = [format_md_gpu_info(gpu_data) for gpu_data in gles_gpu_data]
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/task/ |
D | tensor_desc.h | 344 half* gpu_data = reinterpret_cast<half*>(data_.data()); in UploadData() local 345 DataFromLinear(src, *this, gpu_data); in UploadData() 347 T* gpu_data = reinterpret_cast<T*>(data_.data()); in UploadData() local 348 DataFromLinear(src, *this, gpu_data); in UploadData() 352 half* gpu_data = reinterpret_cast<half*>(data_.data()); in UploadData() local 353 DataFromBHWDC(src, shape_, *this, gpu_data); in UploadData() 355 T* gpu_data = reinterpret_cast<T*>(data_.data()); in UploadData() local 356 DataFromBHWDC(src, shape_, *this, gpu_data); in UploadData() 365 half* gpu_data = reinterpret_cast<half*>(data_.data()); in DownloadData() local 366 DataToBHWDC(gpu_data, shape_, *this, dst); in DownloadData() [all …]
|