Home
last modified time | relevance | path

Searched refs:rhs_data (Results 1 – 25 of 30) sorted by relevance

12

/external/ruy/example/
Dexample.cc23 const float rhs_data[] = {1, 2, 3, 4}; in ExampleMulFloat() local
31 rhs.set_data(rhs_data); in ExampleMulFloat()
47 const float rhs_data[] = {1, 2, 3, 4}; in ExampleMulFloatWithBiasAddAndClamp() local
56 rhs.set_data(rhs_data); in ExampleMulFloatWithBiasAddAndClamp()
75 const std::uint8_t rhs_data[] = {129, 130, 131, 132}; in ExampleMulUint8AsymmetricQuantized() local
84 rhs.set_data(rhs_data); in ExampleMulUint8AsymmetricQuantized()
104 const std::int8_t rhs_data[] = {1, 2, 3, 4}; in ExampleMulInt8PerChannelQuantized() local
114 rhs.set_data(rhs_data); in ExampleMulInt8PerChannelQuantized()
131 const std::int8_t rhs_data[] = {1, 2, 3, 4}; in ExampleMulInt8GetRawAccumulators() local
139 rhs.set_data(rhs_data); in ExampleMulInt8GetRawAccumulators()
/external/tensorflow/tensorflow/lite/kernels/
Dcpu_backend_gemm_x86.h40 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Run()
50 quantization_flavor>::Run(lhs_params, lhs_data, rhs_params, rhs_data, in Run()
58 rhs_params, rhs_data, in Run()
69 const MatrixParams<float>& rhs_params, const float* rhs_data,
74 GemmImplUsingEigen::Run(lhs_params, lhs_data, rhs_params, rhs_data,
Dcpu_backend_gemm.h117 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Gemm() argument
145 rhs_params, rhs_data, in Gemm()
156 if (detail::CustomGemv(lhs_params, lhs_data, rhs_params, rhs_data, in Gemm()
163 quantization_flavor>::Run(lhs_params, lhs_data, rhs_params, rhs_data, in Gemm()
172 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Gemm() argument
184 rhs_params, rhs_data, in Gemm()
Dcpu_backend_gemm_test.cc256 const std::vector<RhsScalar>& rhs_data, in PerformGemmThenCompareResultsThenAgainWithClamping() argument
262 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in PerformGemmThenCompareResultsThenAgainWithClamping()
275 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in PerformGemmThenCompareResultsThenAgainWithClamping()
284 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in PerformGemmThenCompareResultsThenAgainWithClamping()
314 const std::vector<RhsScalar>& rhs_data, in BisectReasonableMultiplierExponent() argument
331 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in BisectReasonableMultiplierExponent()
336 bisect_mid + 1, bisect_max, lhs_params, lhs_data, rhs_params, rhs_data, in BisectReasonableMultiplierExponent()
340 bisect_min, bisect_mid, lhs_params, lhs_data, rhs_params, rhs_data, in BisectReasonableMultiplierExponent()
349 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in ReferenceGemm() argument
357 cpu_backend_gemm::detail::MakeRuyMatrix(rhs_params, rhs_data, &ruy_rhs); in ReferenceGemm()
[all …]
Dcpu_backend_gemm_custom_gemv.h90 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Run()
103 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in CustomGemvTask() argument
110 rhs_data_(rhs_data), in CustomGemvTask()
149 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in CustomGemv() argument
168 Impl::Run(lhs_params, lhs_data, rhs_params, rhs_data, dst_params, dst_data, in CustomGemv()
180 tasks.emplace_back(lhs_params, lhs_data, rhs_params, rhs_data, dst_params, in CustomGemv()
324 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data,
342 optimized_ops_preload_l1_keep(rhs_data + k);
372 Load16AndSubtractZeroPoint(rhs_data + in, rhs_params.zero_point);
429 Load8AndSubtractZeroPoint(rhs_data + in, rhs_params.zero_point);
[all …]
Dcpu_backend_gemm_gemmlowp.h82 const MatrixParams<SrcScalar>& rhs_params, const SrcScalar* rhs_data,
91 gemmlowp_rhs(rhs_data, rhs_params.rows, rhs_params.cols);
139 const MatrixParams<SrcScalar>& rhs_params, const SrcScalar* rhs_data,
151 gemmlowp_rhs(rhs_data, rhs_params.rows, rhs_params.cols);
183 Run(lhs_params, lhs_data, rhs_params, rhs_data, dst_params, dst_data,
Dbatch_matmul.cc305 const TfLiteTensor* rhs_data; in Prepare() local
307 GetInputSafe(context, node, kInputRHSTensor, &rhs_data)); in Prepare()
317 context, lhs_data, rhs_data, output, &real_multiplier)); in Prepare()
335 TF_LITE_ENSURE_EQ(context, rhs_data->params.zero_point, 0); in Prepare()
342 TF_LITE_ENSURE(context, rhs_data->type == kTfLiteFloat32 || in Prepare()
343 rhs_data->type == kTfLiteInt8 || in Prepare()
344 rhs_data->type == kTfLiteInt16); in Prepare()
348 rhs_data->type == kTfLiteInt8) || in Prepare()
349 lhs_data->type == rhs_data->type); in Prepare()
353 TF_LITE_ENSURE(context, NumDimensions(rhs_data) >= 2); in Prepare()
[all …]
Dcpu_backend_gemm_eigen.cc39 const MatrixParams<float>& rhs_params, const float* rhs_data, in Run() argument
56 EigenMatrixMapColMajorConst eigen_rhs(rhs_data, rhs_params.rows, in Run()
Dcpu_backend_gemm_eigen.h30 const MatrixParams<float>& rhs_params, const float* rhs_data,
Dcpu_backend_gemm_ruy.h127 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data,
135 MakeRuyMatrix(rhs_params, rhs_data, &ruy_rhs, context->use_caching());
Dbatch_matmul_test.cc286 std::initializer_list<float> rhs_data, in ConstRHSBatchMatMulOpModel() argument
289 rhs_id_ = AddConstInput<float>(TensorType_FLOAT32, rhs_data, rhs_shape); in ConstRHSBatchMatMulOpModel()
/external/ruy/ruy/
Dkernel_avx512.cc295 std::int32_t rhs_data[32];
306 _mm512_storeu_si512(reinterpret_cast<__m256i*>(rhs_data),
308 _mm512_storeu_si512(reinterpret_cast<__m256i*>(rhs_data + 16),
320 _mm512_set1_epi32(rhs_data[2 * col]);
322 _mm512_set1_epi32(rhs_data[2 * col + 1]);
737 std::int32_t rhs_data[2];
741 _mm_storeu_si64(reinterpret_cast<__m128i*>(rhs_data), rhs_16_bit_dup);
754 const __m512i rhs_16_bit_dup_low = _mm512_set1_epi32(rhs_data[index]);
756 _mm512_set1_epi32(rhs_data[index + 1]);
979 const float* rhs_data = rhs_ptr;
[all …]
Dperchannel_buffers_reallocation_test.cc60 const RhsScalar rhs_data[kMatrixSize * kMatrixSize] = {0}; in TestPerChannelBuffersReallocation() local
61 rhs.set_data(rhs_data); in TestPerChannelBuffersReallocation()
Dprepacked_cache_test.cc236 const float rhs_data[] = {1, 2}; in TestCachePolicies() local
244 rhs.set_data(rhs_data); in TestCachePolicies()
277 const float rhs_data[] = {1, 2}; in TEST() local
285 rhs.set_data(rhs_data); in TEST()
Dkernel_avx2_fma.cc262 std::int32_t rhs_data[16];
273 _mm256_storeu_si256(reinterpret_cast<__m256i*>(rhs_data),
275 _mm256_storeu_si256(reinterpret_cast<__m256i*>(rhs_data + 8),
293 rhs_data)); // Load [0 1 2 3 4 5 6 7]
295 reinterpret_cast<const __m256i*>(rhs_data + 8)); // Load [8 - 15]
819 std::int32_t rhs_data[2];
823 _mm_storeu_si64(reinterpret_cast<__m128i*>(rhs_data), rhs_16_bit_dup);
841 const std::int32_t low_rhs_value = rhs_data[0];
842 const std::int32_t high_rhs_value = rhs_data[1];
Dkernel_x86.h610 const float* rhs_data = rhs_ptr;
613 __m256 rhs1 = _mm256_loadu_ps(rhs_data); // Load [0 1 2 3 4 5 6 7]
710 const float* rhs_data = rhs_ptr;
712 __m256 rhs1 = _mm256_loadu_ps(rhs_data); // Load [0 1 2 3 4 5 6 7]
827 const float* rhs_data = rhs_ptr;
829 const __m256 dup_rhs_element_j = _mm256_set1_ps(rhs_data[0]);
857 const float* rhs_data = rhs_ptr;
859 const __m256 dup_rhs_element_j = _mm256_set1_ps(rhs_data[0]);
Dkernel_avx.cc609 std::int32_t rhs_data[16];
620 _mm256_storeu_si256(reinterpret_cast<__m256i*>(rhs_data),
622 _mm256_storeu_si256(reinterpret_cast<__m256i*>(rhs_data + 8),
646 rhs_data)); // Load [0 1 2 3 4 5 6 7]
648 reinterpret_cast<const __m256i*>(rhs_data + 8)); // Load [8 - 15]
1260 std::int32_t rhs_data[2];
1264 _mm_storeu_si64(reinterpret_cast<__m128i*>(rhs_data), rhs_16_bit_dup);
1286 const std::int32_t low_rhs_value = rhs_data[0];
1287 const std::int32_t high_rhs_value = rhs_data[1];
/external/tensorflow/tensorflow/compiler/xla/service/gpu/
Dgemm_thunk.cc65 se::DeviceMemoryBase rhs_data = get_device_address(rhs_buffer_); in ExecuteOnStream() local
67 return RunGemm(config_, lhs_data, rhs_data, output_data, params.stream, in ExecuteOnStream()
118 se::DeviceMemory<Element> rhs_data(rhs_matrix.data); in DoGemmWithAlgorithm() local
136 /*leading dim of LHS=*/lhs_matrix.num_rows, rhs_data, in DoGemmWithAlgorithm()
153 /*leading dim of LHS=*/lhs_matrix.num_rows, lhs_stride, rhs_data, in DoGemmWithAlgorithm()
165 lhs_data, /*leading dim of LHS=*/lhs_matrix.num_rows, rhs_data, in DoGemmWithAlgorithm()
/external/tensorflow/tensorflow/lite/kernels/internal/reference/
Dbatch_matmul.h54 const RuntimeShape& rhs_shape, const float* rhs_data, in BatchMatMul() argument
82 const float* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul()
109 const RuntimeShape& rhs_shape, const int8_t* rhs_data, in BatchMatMul() argument
159 const int8_t* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul()
201 const RuntimeShape& rhs_shape, const T* rhs_data, in BatchMatMul() argument
238 const T* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul()
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/
Dbatch_matmul.h29 const RuntimeShape& rhs_shape, const float* rhs_data, in BatchMatMul() argument
97 const float* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul()
116 const RuntimeShape& rhs_shape, const int8_t* rhs_data, in BatchMatMul() argument
207 const int8_t* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul()
273 const RuntimeShape& rhs_shape, const int8_t* rhs_data, in BatchMatMul() argument
354 const int8_t* rhs_ptr0 = rhs_data + (b0 * rhs_ext0); in BatchMatMul()
/external/tensorflow/tensorflow/compiler/xla/client/lib/
Dtridiagonal_test.cc70 auto rhs_data = CreateR3Parameter<float>(rhs, 3, "rhs", &builder, &rhs_xla); in XLA_TEST_P() local
107 upper_diagonal_data.get(), rhs_data.get()})); in XLA_TEST_P()
/external/tensorflow/tensorflow/core/kernels/
Dmatmul_op_test.cc36 std::function<void(const Tensor& lhs_data, const Tensor& rhs_data,
100 void RunMatMulWithBias(const Tensor& lhs_data, const Tensor& rhs_data, in RunMatMulWithBias() argument
109 ops::Const(root.WithOpName("rhs"), Input::Initializer(rhs_data)), in RunMatMulWithBias()
120 const Tensor& lhs_data, const Tensor& rhs_data, const Tensor& bias_data, in RunMatMulWithBiasAndActivation() argument
128 ops::Const(root.WithOpName("rhs"), Input::Initializer(rhs_data)), in RunMatMulWithBiasAndActivation()
150 void RunFusedMatMulOp(const Tensor& lhs_data, const Tensor& rhs_data, in RunFusedMatMulOp() argument
163 ops::Const(root.WithOpName("rhs"), Input::Initializer(rhs_data)); in RunFusedMatMulOp()
/external/tensorflow/tensorflow/core/kernels/linalg/
Dtridiagonal_solve_op_gpu.cu.cc284 const Scalar* rhs_data = rhs.flat<Scalar>().data(); in ComputeWithGtsvBatched() local
306 CopyDeviceToDevice(context, rhs_data, output->flat<Scalar>().data(), in ComputeWithGtsvBatched()
/external/llvm-project/lldb/examples/python/
Dmach_o.py921 rhs_data = rhs_section.get_contents(rhs)
922 if lhs_data and rhs_data:
923 if lhs_data == rhs_data:
927 rhs_data_len = len(rhs_data)
947 elif lhs_data and not rhs_data:
952 elif not lhs_data and rhs_data:
/external/tensorflow/tensorflow/compiler/xla/tests/
Darray_elementwise_ops_test.cc277 std::unique_ptr<GlobalData> rhs_data = in XLA_TEST_F() local
287 ComputeAndCompareR1<uint64>(&b, expected, {lhs_data.get(), rhs_data.get()}); in XLA_TEST_F()
316 std::unique_ptr<GlobalData> rhs_data = in XLA_TEST_F() local
326 ComputeAndCompareR1<int64>(&b, expected, {lhs_data.get(), rhs_data.get()}); in XLA_TEST_F()

12