/external/ruy/example/ |
D | example.cc | 23 const float rhs_data[] = {1, 2, 3, 4}; in ExampleMulFloat() local 31 rhs.set_data(rhs_data); in ExampleMulFloat() 47 const float rhs_data[] = {1, 2, 3, 4}; in ExampleMulFloatWithBiasAddAndClamp() local 56 rhs.set_data(rhs_data); in ExampleMulFloatWithBiasAddAndClamp() 75 const std::uint8_t rhs_data[] = {129, 130, 131, 132}; in ExampleMulUint8AsymmetricQuantized() local 84 rhs.set_data(rhs_data); in ExampleMulUint8AsymmetricQuantized() 104 const std::int8_t rhs_data[] = {1, 2, 3, 4}; in ExampleMulInt8PerChannelQuantized() local 114 rhs.set_data(rhs_data); in ExampleMulInt8PerChannelQuantized() 131 const std::int8_t rhs_data[] = {1, 2, 3, 4}; in ExampleMulInt8GetRawAccumulators() local 139 rhs.set_data(rhs_data); in ExampleMulInt8GetRawAccumulators()
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | cpu_backend_gemm_x86.h | 40 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Run() 50 quantization_flavor>::Run(lhs_params, lhs_data, rhs_params, rhs_data, in Run() 58 rhs_params, rhs_data, in Run() 69 const MatrixParams<float>& rhs_params, const float* rhs_data, 74 GemmImplUsingEigen::Run(lhs_params, lhs_data, rhs_params, rhs_data,
|
D | cpu_backend_gemm.h | 117 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Gemm() argument 145 rhs_params, rhs_data, in Gemm() 156 if (detail::CustomGemv(lhs_params, lhs_data, rhs_params, rhs_data, in Gemm() 163 quantization_flavor>::Run(lhs_params, lhs_data, rhs_params, rhs_data, in Gemm() 172 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Gemm() argument 184 rhs_params, rhs_data, in Gemm()
|
D | cpu_backend_gemm_test.cc | 256 const std::vector<RhsScalar>& rhs_data, in PerformGemmThenCompareResultsThenAgainWithClamping() argument 262 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in PerformGemmThenCompareResultsThenAgainWithClamping() 275 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in PerformGemmThenCompareResultsThenAgainWithClamping() 284 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in PerformGemmThenCompareResultsThenAgainWithClamping() 314 const std::vector<RhsScalar>& rhs_data, in BisectReasonableMultiplierExponent() argument 331 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in BisectReasonableMultiplierExponent() 336 bisect_mid + 1, bisect_max, lhs_params, lhs_data, rhs_params, rhs_data, in BisectReasonableMultiplierExponent() 340 bisect_min, bisect_mid, lhs_params, lhs_data, rhs_params, rhs_data, in BisectReasonableMultiplierExponent() 349 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in ReferenceGemm() argument 357 cpu_backend_gemm::detail::MakeRuyMatrix(rhs_params, rhs_data, &ruy_rhs); in ReferenceGemm() [all …]
|
D | cpu_backend_gemm_custom_gemv.h | 90 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Run() 103 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in CustomGemvTask() argument 110 rhs_data_(rhs_data), in CustomGemvTask() 149 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in CustomGemv() argument 168 Impl::Run(lhs_params, lhs_data, rhs_params, rhs_data, dst_params, dst_data, in CustomGemv() 180 tasks.emplace_back(lhs_params, lhs_data, rhs_params, rhs_data, dst_params, in CustomGemv() 324 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, 342 optimized_ops_preload_l1_keep(rhs_data + k); 372 Load16AndSubtractZeroPoint(rhs_data + in, rhs_params.zero_point); 429 Load8AndSubtractZeroPoint(rhs_data + in, rhs_params.zero_point); [all …]
|
D | cpu_backend_gemm_gemmlowp.h | 82 const MatrixParams<SrcScalar>& rhs_params, const SrcScalar* rhs_data, 91 gemmlowp_rhs(rhs_data, rhs_params.rows, rhs_params.cols); 139 const MatrixParams<SrcScalar>& rhs_params, const SrcScalar* rhs_data, 151 gemmlowp_rhs(rhs_data, rhs_params.rows, rhs_params.cols); 183 Run(lhs_params, lhs_data, rhs_params, rhs_data, dst_params, dst_data,
|
D | batch_matmul.cc | 305 const TfLiteTensor* rhs_data; in Prepare() local 307 GetInputSafe(context, node, kInputRHSTensor, &rhs_data)); in Prepare() 317 context, lhs_data, rhs_data, output, &real_multiplier)); in Prepare() 335 TF_LITE_ENSURE_EQ(context, rhs_data->params.zero_point, 0); in Prepare() 342 TF_LITE_ENSURE(context, rhs_data->type == kTfLiteFloat32 || in Prepare() 343 rhs_data->type == kTfLiteInt8 || in Prepare() 344 rhs_data->type == kTfLiteInt16); in Prepare() 348 rhs_data->type == kTfLiteInt8) || in Prepare() 349 lhs_data->type == rhs_data->type); in Prepare() 353 TF_LITE_ENSURE(context, NumDimensions(rhs_data) >= 2); in Prepare() [all …]
|
D | cpu_backend_gemm_eigen.cc | 39 const MatrixParams<float>& rhs_params, const float* rhs_data, in Run() argument 56 EigenMatrixMapColMajorConst eigen_rhs(rhs_data, rhs_params.rows, in Run()
|
D | cpu_backend_gemm_eigen.h | 30 const MatrixParams<float>& rhs_params, const float* rhs_data,
|
D | cpu_backend_gemm_ruy.h | 127 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, 135 MakeRuyMatrix(rhs_params, rhs_data, &ruy_rhs, context->use_caching());
|
D | batch_matmul_test.cc | 286 std::initializer_list<float> rhs_data, in ConstRHSBatchMatMulOpModel() argument 289 rhs_id_ = AddConstInput<float>(TensorType_FLOAT32, rhs_data, rhs_shape); in ConstRHSBatchMatMulOpModel()
|
/external/ruy/ruy/ |
D | kernel_avx512.cc | 295 std::int32_t rhs_data[32]; 306 _mm512_storeu_si512(reinterpret_cast<__m256i*>(rhs_data), 308 _mm512_storeu_si512(reinterpret_cast<__m256i*>(rhs_data + 16), 320 _mm512_set1_epi32(rhs_data[2 * col]); 322 _mm512_set1_epi32(rhs_data[2 * col + 1]); 737 std::int32_t rhs_data[2]; 741 _mm_storeu_si64(reinterpret_cast<__m128i*>(rhs_data), rhs_16_bit_dup); 754 const __m512i rhs_16_bit_dup_low = _mm512_set1_epi32(rhs_data[index]); 756 _mm512_set1_epi32(rhs_data[index + 1]); 979 const float* rhs_data = rhs_ptr; [all …]
|
D | perchannel_buffers_reallocation_test.cc | 60 const RhsScalar rhs_data[kMatrixSize * kMatrixSize] = {0}; in TestPerChannelBuffersReallocation() local 61 rhs.set_data(rhs_data); in TestPerChannelBuffersReallocation()
|
D | prepacked_cache_test.cc | 236 const float rhs_data[] = {1, 2}; in TestCachePolicies() local 244 rhs.set_data(rhs_data); in TestCachePolicies() 277 const float rhs_data[] = {1, 2}; in TEST() local 285 rhs.set_data(rhs_data); in TEST()
|
D | kernel_avx2_fma.cc | 262 std::int32_t rhs_data[16]; 273 _mm256_storeu_si256(reinterpret_cast<__m256i*>(rhs_data), 275 _mm256_storeu_si256(reinterpret_cast<__m256i*>(rhs_data + 8), 293 rhs_data)); // Load [0 1 2 3 4 5 6 7] 295 reinterpret_cast<const __m256i*>(rhs_data + 8)); // Load [8 - 15] 819 std::int32_t rhs_data[2]; 823 _mm_storeu_si64(reinterpret_cast<__m128i*>(rhs_data), rhs_16_bit_dup); 841 const std::int32_t low_rhs_value = rhs_data[0]; 842 const std::int32_t high_rhs_value = rhs_data[1];
|
D | kernel_x86.h | 610 const float* rhs_data = rhs_ptr; 613 __m256 rhs1 = _mm256_loadu_ps(rhs_data); // Load [0 1 2 3 4 5 6 7] 710 const float* rhs_data = rhs_ptr; 712 __m256 rhs1 = _mm256_loadu_ps(rhs_data); // Load [0 1 2 3 4 5 6 7] 827 const float* rhs_data = rhs_ptr; 829 const __m256 dup_rhs_element_j = _mm256_set1_ps(rhs_data[0]); 857 const float* rhs_data = rhs_ptr; 859 const __m256 dup_rhs_element_j = _mm256_set1_ps(rhs_data[0]);
|
D | kernel_avx.cc | 609 std::int32_t rhs_data[16]; 620 _mm256_storeu_si256(reinterpret_cast<__m256i*>(rhs_data), 622 _mm256_storeu_si256(reinterpret_cast<__m256i*>(rhs_data + 8), 646 rhs_data)); // Load [0 1 2 3 4 5 6 7] 648 reinterpret_cast<const __m256i*>(rhs_data + 8)); // Load [8 - 15] 1260 std::int32_t rhs_data[2]; 1264 _mm_storeu_si64(reinterpret_cast<__m128i*>(rhs_data), rhs_16_bit_dup); 1286 const std::int32_t low_rhs_value = rhs_data[0]; 1287 const std::int32_t high_rhs_value = rhs_data[1];
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | gemm_thunk.cc | 65 se::DeviceMemoryBase rhs_data = get_device_address(rhs_buffer_); in ExecuteOnStream() local 67 return RunGemm(config_, lhs_data, rhs_data, output_data, params.stream, in ExecuteOnStream() 118 se::DeviceMemory<Element> rhs_data(rhs_matrix.data); in DoGemmWithAlgorithm() local 136 /*leading dim of LHS=*/lhs_matrix.num_rows, rhs_data, in DoGemmWithAlgorithm() 153 /*leading dim of LHS=*/lhs_matrix.num_rows, lhs_stride, rhs_data, in DoGemmWithAlgorithm() 165 lhs_data, /*leading dim of LHS=*/lhs_matrix.num_rows, rhs_data, in DoGemmWithAlgorithm()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | batch_matmul.h | 54 const RuntimeShape& rhs_shape, const float* rhs_data, in BatchMatMul() argument 82 const float* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul() 109 const RuntimeShape& rhs_shape, const int8_t* rhs_data, in BatchMatMul() argument 159 const int8_t* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul() 201 const RuntimeShape& rhs_shape, const T* rhs_data, in BatchMatMul() argument 238 const T* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | batch_matmul.h | 29 const RuntimeShape& rhs_shape, const float* rhs_data, in BatchMatMul() argument 97 const float* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul() 116 const RuntimeShape& rhs_shape, const int8_t* rhs_data, in BatchMatMul() argument 207 const int8_t* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul() 273 const RuntimeShape& rhs_shape, const int8_t* rhs_data, in BatchMatMul() argument 354 const int8_t* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul()
|
/external/tensorflow/tensorflow/compiler/xla/client/lib/ |
D | tridiagonal_test.cc | 70 auto rhs_data = CreateR3Parameter<float>(rhs, 3, "rhs", &builder, &rhs_xla); in XLA_TEST_P() local 107 upper_diagonal_data.get(), rhs_data.get()})); in XLA_TEST_P()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | matmul_op_test.cc | 36 std::function<void(const Tensor& lhs_data, const Tensor& rhs_data, 100 void RunMatMulWithBias(const Tensor& lhs_data, const Tensor& rhs_data, in RunMatMulWithBias() argument 109 ops::Const(root.WithOpName("rhs"), Input::Initializer(rhs_data)), in RunMatMulWithBias() 120 const Tensor& lhs_data, const Tensor& rhs_data, const Tensor& bias_data, in RunMatMulWithBiasAndActivation() argument 128 ops::Const(root.WithOpName("rhs"), Input::Initializer(rhs_data)), in RunMatMulWithBiasAndActivation() 150 void RunFusedMatMulOp(const Tensor& lhs_data, const Tensor& rhs_data, in RunFusedMatMulOp() argument 163 ops::Const(root.WithOpName("rhs"), Input::Initializer(rhs_data)); in RunFusedMatMulOp()
|
/external/tensorflow/tensorflow/core/kernels/linalg/ |
D | tridiagonal_solve_op_gpu.cu.cc | 284 const Scalar* rhs_data = rhs.flat<Scalar>().data(); in ComputeWithGtsvBatched() local 306 CopyDeviceToDevice(context, rhs_data, output->flat<Scalar>().data(), in ComputeWithGtsvBatched()
|
/external/llvm-project/lldb/examples/python/ |
D | mach_o.py | 921 rhs_data = rhs_section.get_contents(rhs) 922 if lhs_data and rhs_data: 923 if lhs_data == rhs_data: 927 rhs_data_len = len(rhs_data) 947 elif lhs_data and not rhs_data: 952 elif not lhs_data and rhs_data:
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
D | array_elementwise_ops_test.cc | 277 std::unique_ptr<GlobalData> rhs_data = in XLA_TEST_F() local 287 ComputeAndCompareR1<uint64>(&b, expected, {lhs_data.get(), rhs_data.get()}); in XLA_TEST_F() 316 std::unique_ptr<GlobalData> rhs_data = in XLA_TEST_F() local 326 ComputeAndCompareR1<int64>(&b, expected, {lhs_data.get(), rhs_data.get()}); in XLA_TEST_F()
|