/external/tensorflow/tensorflow/lite/kernels/ |
D | cpu_backend_gemm.h | 117 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Gemm() argument 122 ValidateParams(lhs_params, rhs_params, dst_params, params); in Gemm() 123 if (!IsValidGemm(lhs_params, rhs_params, dst_params)) { in Gemm() 140 rhs_params.order != Order::kColMajor || in Gemm() 152 rhs_params, rhs_data, in Gemm() 163 if (detail::CustomGemv(lhs_params, lhs_data, rhs_params, rhs_data, in Gemm() 170 quantization_flavor>::Run(lhs_params, lhs_data, rhs_params, rhs_data, in Gemm() 177 const MatrixParams<int16_t>& rhs_params, const int16_t* rhs_data, in Gemm() argument 182 ValidateParams(lhs_params, rhs_params, dst_params, params); in Gemm() 183 if (!IsValidGemm(lhs_params, rhs_params, dst_params)) { in Gemm() [all …]
|
D | cpu_backend_gemm_test.cc | 255 const MatrixParams<RhsScalar>& rhs_params, in PerformGemmThenCompareResultsThenAgainWithClamping() argument 262 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in PerformGemmThenCompareResultsThenAgainWithClamping() 275 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in PerformGemmThenCompareResultsThenAgainWithClamping() 284 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in PerformGemmThenCompareResultsThenAgainWithClamping() 313 const MatrixParams<RhsScalar>& rhs_params, in BisectReasonableMultiplierExponent() argument 331 Gemm(lhs_params, lhs_data.data(), rhs_params, rhs_data.data(), dst_params, in BisectReasonableMultiplierExponent() 336 bisect_mid + 1, bisect_max, lhs_params, lhs_data, rhs_params, rhs_data, in BisectReasonableMultiplierExponent() 340 bisect_min, bisect_mid, lhs_params, lhs_data, rhs_params, rhs_data, in BisectReasonableMultiplierExponent() 349 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in ReferenceGemm() argument 357 cpu_backend_gemm::detail::MakeRuyMatrix(rhs_params, rhs_data, &ruy_rhs); in ReferenceGemm() [all …]
|
D | cpu_backend_gemm_gemmlowp.h | 82 const MatrixParams<SrcScalar>& rhs_params, const SrcScalar* rhs_data, 91 gemmlowp_rhs(rhs_data, rhs_params.rows, rhs_params.cols); 116 &gemmlowp_dst, -lhs_params.zero_point, -rhs_params.zero_point, 123 &gemmlowp_dst, -lhs_params.zero_point, -rhs_params.zero_point, 139 const MatrixParams<SrcScalar>& rhs_params, const SrcScalar* rhs_data, 151 gemmlowp_rhs(rhs_data, rhs_params.rows, rhs_params.cols); 179 -lhs_params.zero_point, -rhs_params.zero_point, output_pipeline); 183 Run(lhs_params, lhs_data, rhs_params, rhs_data, dst_params, dst_data,
|
D | cpu_backend_gemm_x86.h | 40 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Run() 50 quantization_flavor>::Run(lhs_params, lhs_data, rhs_params, rhs_data, in Run() 58 rhs_params, rhs_data, in Run() 69 const MatrixParams<float>& rhs_params, const float* rhs_data, 74 GemmImplUsingEigen::Run(lhs_params, lhs_data, rhs_params, rhs_data,
|
D | cpu_backend_gemm_eigen.cc | 39 const MatrixParams<float>& rhs_params, const float* rhs_data, in Run() argument 56 EigenMatrixMapColMajorConst eigen_rhs(rhs_data, rhs_params.rows, in Run() 57 rhs_params.cols); in Run() 61 if (rhs_params.cols == 1) { in Run()
|
D | cpu_backend_gemm_params.h | 233 const MatrixParams<RhsScalar>& rhs_params, 244 const MatrixParams<RhsScalar>& rhs_params, 249 valid &= rhs_params.rows >= 1; 250 valid &= rhs_params.cols >= 1; 253 valid &= lhs_params.cols == rhs_params.rows; 254 valid &= rhs_params.cols == dst_params.cols;
|
D | cpu_backend_gemm_custom_gemv.h | 81 const MatrixParams<RhsScalar>& rhs_params, in IsSupportedGivenSufficientlyManyRows() 90 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in Run() 103 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in CustomGemvTask() argument 109 rhs_params_(rhs_params), in CustomGemvTask() 149 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, in CustomGemv() argument 159 if (!Impl::IsSupportedGivenSufficientlyManyRows(lhs_params, rhs_params, in CustomGemv() 168 Impl::Run(lhs_params, lhs_data, rhs_params, rhs_data, dst_params, dst_data, in CustomGemv() 180 tasks.emplace_back(lhs_params, lhs_data, rhs_params, rhs_data, dst_params, in CustomGemv() 311 const MatrixParams<RhsScalar>& rhs_params, 324 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, [all …]
|
D | cpu_backend_gemm_eigen.h | 30 const MatrixParams<float>& rhs_params, const float* rhs_data,
|
D | cpu_backend_gemm_ruy.h | 127 const MatrixParams<RhsScalar>& rhs_params, const RhsScalar* rhs_data, 135 MakeRuyMatrix(rhs_params, rhs_data, &ruy_rhs, context->use_caching());
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/integer_ops/ |
D | fully_connected.h | 73 cpu_backend_gemm::MatrixParams<InputScalar> rhs_params; in FullyConnectedPerChannel() local 74 rhs_params.rows = filter_cols; in FullyConnectedPerChannel() 75 rhs_params.cols = batches; in FullyConnectedPerChannel() 76 rhs_params.order = cpu_backend_gemm::Order::kColMajor; in FullyConnectedPerChannel() 77 rhs_params.zero_point = -input_offset; in FullyConnectedPerChannel() 78 rhs_params.cache_policy = in FullyConnectedPerChannel() 95 cpu_backend_gemm::Gemm(lhs_params, filter_data, rhs_params, input_data, in FullyConnectedPerChannel() 145 cpu_backend_gemm::MatrixParams<InputScalar> rhs_params; in FullyConnected() local 146 rhs_params.rows = filter_cols; in FullyConnected() 147 rhs_params.cols = batches; in FullyConnected() [all …]
|
D | transpose_conv.h | 79 cpu_backend_gemm::MatrixParams<InputScalar> rhs_params; in TransposeConvV2() local 80 rhs_params.order = cpu_backend_gemm::Order::kColMajor; in TransposeConvV2() 81 rhs_params.rows = input_depth; in TransposeConvV2() 82 rhs_params.cols = input_image_size; in TransposeConvV2() 83 rhs_params.zero_point = -params.input_offset; in TransposeConvV2() 91 cpu_backend_gemm::Gemm(lhs_params, hwoi_ordered_filter_data, rhs_params, in TransposeConvV2()
|
D | conv.h | 103 cpu_backend_gemm::MatrixParams<InputScalar> rhs_params; in ConvPerChannel() local 104 rhs_params.rows = gemm_input_rows; in ConvPerChannel() 105 rhs_params.cols = gemm_input_cols; in ConvPerChannel() 106 rhs_params.order = cpu_backend_gemm::Order::kColMajor; in ConvPerChannel() 107 rhs_params.zero_point = -input_offset; in ConvPerChannel() 122 cpu_backend_gemm::Gemm(lhs_params, filter_data, rhs_params, gemm_input_data, in ConvPerChannel()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | batch_matmul.h | 85 MatrixParams<float> rhs_params; in BatchMatMul() local 86 rhs_params.order = cpu_backend_gemm::Order::kColMajor; in BatchMatMul() 87 rhs_params.rows = accum_depth; in BatchMatMul() 88 rhs_params.cols = rhs_cols; in BatchMatMul() 108 cpu_backend_gemm::Gemm(lhs_params, lhs_ptr2, rhs_params, rhs_ptr2, in BatchMatMul() 195 MatrixParams<int8_t> rhs_params; in BatchMatMul() local 196 rhs_params.order = cpu_backend_gemm::Order::kColMajor; in BatchMatMul() 197 rhs_params.rows = accum_depth; in BatchMatMul() 198 rhs_params.cols = rhs_cols; in BatchMatMul() 227 cpu_backend_gemm::Gemm(lhs_params, lhs_ptr2, rhs_params, rhs_ptr2, in BatchMatMul() [all …]
|
D | optimized_ops.h | 277 cpu_backend_gemm::MatrixParams<float> rhs_params; in FullyConnected() local 278 rhs_params.order = cpu_backend_gemm::Order::kColMajor; in FullyConnected() 279 rhs_params.rows = input_rows; in FullyConnected() 280 rhs_params.cols = input_shape.FlatSize() / input_rows; in FullyConnected() 281 rhs_params.cache_policy = in FullyConnected() 283 TFLITE_DCHECK_EQ(input_shape.FlatSize(), rhs_params.rows * rhs_params.cols); in FullyConnected() 299 cpu_backend_gemm::Gemm(lhs_params, weights_data, rhs_params, input_data, in FullyConnected() 344 cpu_backend_gemm::MatrixParams<uint8> rhs_params; in FullyConnected() local 345 rhs_params.rows = filter_cols; in FullyConnected() 346 rhs_params.cols = batches; in FullyConnected() [all …]
|
D | sse_tensor_utils.cc | 392 MatrixParams<int8_t> rhs_params; in SseCpuBackendGemm() local 393 rhs_params.order = cpu_backend_gemm::Order::kColMajor; in SseCpuBackendGemm() 394 rhs_params.rows = n_input; in SseCpuBackendGemm() 395 rhs_params.cols = n_batch; in SseCpuBackendGemm() 406 cpu_backend_gemm::Gemm(lhs_params, input_to_gate_weights, rhs_params, input, in SseCpuBackendGemm()
|
D | neon_tensor_utils.cc | 1019 MatrixParams<int8_t> rhs_params; in NeonCpuBackendGemm() local 1020 rhs_params.order = cpu_backend_gemm::Order::kColMajor; in NeonCpuBackendGemm() 1021 rhs_params.rows = n_input; in NeonCpuBackendGemm() 1022 rhs_params.cols = n_batch; in NeonCpuBackendGemm() 1033 cpu_backend_gemm::Gemm(lhs_params, input_to_gate_weights, rhs_params, input, in NeonCpuBackendGemm()
|
/external/tensorflow/tensorflow/compiler/xla/client/lib/ |
D | comparators.cc | 64 std::vector<XlaOp> rhs_params; in CreateScalarComparisonComputation() local 78 rhs_params.emplace_back(rhs_param); in CreateScalarComparisonComputation() 101 generators[i].value()(lhs_params[i], rhs_params[i], {}), in CreateScalarComparisonComputation() 105 And(param_equal, EqTotalOrder(lhs_params[i], rhs_params[i])); in CreateScalarComparisonComputation()
|
D | approx_topk.cc | 75 std::vector<XlaOp> rhs_params; in BuildReductionComputation() local 78 rhs_params.reserve(num_operands); in BuildReductionComputation() 87 rhs_params.push_back(Parameter(reduction_builder.get(), param_number, in BuildReductionComputation() 97 comparator_args.push_back(rhs_params[i]); in BuildReductionComputation() 103 results.push_back(Select(pred, lhs_params[i], rhs_params[i])); in BuildReductionComputation()
|