/external/gemmlowp/meta/ |
D | legacy_operations_common.h | 20 Quantized8BitOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in Quantized8BitOperation() argument 23 : lhs_offset(lhs_offset), in Quantized8BitOperation() 30 std::int32_t lhs_offset; 39 FloatOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in FloatOperation() argument 41 : lhs_offset(lhs_offset), in FloatOperation() 46 std::int32_t lhs_offset; 53 Int32Operation(std::int32_t lhs_offset, std::int32_t rhs_offset) in Int32Operation() argument 54 : lhs_offset(lhs_offset), rhs_offset(rhs_offset) {} in Int32Operation() 57 std::int32_t lhs_offset;
|
D | legacy_multi_thread_gemv.h | 33 GemvQuantized8BitOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in GemvQuantized8BitOperation() argument 36 : Quantized8BitOperation(lhs_offset, rhs_offset, sum_offset, multiplier, in GemvQuantized8BitOperation() 43 gemv_q8(scratch, lhs, rhs, n, k, lhs_offset, rhs_offset, sum_offset, in ExecuteMatrixMatrix() 55 GemvFloatOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in GemvFloatOperation() argument 57 : FloatOperation(lhs_offset, rhs_offset, result_offset) {} in GemvFloatOperation() 63 gemv_f(scratch, lhs, rhs, n, k, lhs_offset, rhs_offset, result_offset, in ExecuteMatrixMatrix() 75 GemvInt32Operation(std::int32_t lhs_offset, std::int32_t rhs_offset) in GemvInt32Operation() argument 76 : Int32Operation(lhs_offset, rhs_offset) {} in GemvInt32Operation() 82 gemv_i32(scratch, lhs, rhs, n, k, lhs_offset, rhs_offset, result); in ExecuteMatrixMatrix() 102 std::int32_t k, std::int32_t lhs_offset, in multi_thread_gemv_q8() argument [all …]
|
D | legacy_multi_thread_gemm.h | 61 GemmQuantized8BitOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in GemmQuantized8BitOperation() argument 64 : Quantized8BitOperation(lhs_offset, rhs_offset, sum_offset, multiplier, in GemmQuantized8BitOperation() 81 gemm_q8_strided(scratch, lhs, rhs, m, n, k, lhs_offset, rhs_offset, in ExecuteCacheFriendlyMatrixMatrix() 93 GemmFloatOperation(std::int32_t lhs_offset, std::int32_t rhs_offset, in GemmFloatOperation() argument 95 : FloatOperation(lhs_offset, rhs_offset, result_offset) {} in GemmFloatOperation() 111 gemm_f_strided(scratch, lhs, rhs, m, n, k, lhs_offset, rhs_offset, in ExecuteCacheFriendlyMatrixMatrix() 123 GemmInt32Operation(std::int32_t lhs_offset, std::int32_t rhs_offset) in GemmInt32Operation() argument 124 : Int32Operation(lhs_offset, rhs_offset) {} in GemmInt32Operation() 140 gemm_i32_strided(scratch, lhs, rhs, m, n, k, lhs_offset, rhs_offset, result, in ExecuteCacheFriendlyMatrixMatrix() 162 std::int32_t lhs_offset, std::int32_t rhs_offset, in multi_thread_gemm_q8() argument [all …]
|
D | legacy_single_thread_gemm.h | 31 std::int32_t k, std::int32_t lhs_offset, in gemm_q8_strided() argument 58 result_offset + k * lhs_offset * rhs_offset; in gemm_q8_strided() 62 params.right_stream.multiplicative_sum_offset = lhs_offset; in gemm_q8_strided() 76 std::int32_t lhs_offset, std::int32_t rhs_offset, in gemv_q8() argument 102 result_offset + k * lhs_offset * rhs_offset; in gemv_q8() 106 params.right_stream.multiplicative_sum_offset = lhs_offset; in gemv_q8() 124 std::int32_t k, std::int32_t lhs_offset, in gemm_i32_strided() argument 150 params.left_stream.additive_sum_offset = k * lhs_offset * rhs_offset; in gemm_i32_strided() 154 params.right_stream.multiplicative_sum_offset = lhs_offset; in gemm_i32_strided() 165 std::int32_t lhs_offset, std::int32_t rhs_offset, in gemv_i32() argument [all …]
|
/external/gemmlowp/public/ |
D | gemmlowp.h | 38 const LhsOffset& lhs_offset, in GemmWithOutputPipelinePC() argument 42 context, lhs, rhs, result, lhs_offset, rhs_offset, output_pipeline); in GemmWithOutputPipelinePC() 57 int lhs_offset, int rhs_offset, in GemmWithOutputPipeline() argument 61 const OffsetColDup lhs_offset_vector(lhs_offset, lhs.rows()); in GemmWithOutputPipeline() 77 MatrixMap<Scalar, ResultOrder>* result, int lhs_offset, in Gemm() argument 81 context, lhs, rhs, result, lhs_offset, rhs_offset, in Gemm()
|
/external/gemmlowp/internal/ |
D | unpack.h | 98 const LhsOffset& lhs_offset, const RhsOffset& rhs_offset, in UnpackResultBlock() argument 113 LoadForBroadcasting<RegisterBlockType>(lhs_offset, src_row); in UnpackResultBlock() 134 const LhsOffset& lhs_offset, const RhsOffset& rhs_offset, in UnpackResult() argument 187 lhs_sums_of_each_slice, rhs_sums_of_each_slice, lhs_offset, in UnpackResult() 200 lhs_sums_of_each_slice, rhs_sums_of_each_slice, lhs_offset, in UnpackResult() 212 lhs_sums_of_each_slice, rhs_sums_of_each_slice, lhs_offset, in UnpackResult() 229 rhs_sums_of_each_slice, lhs_offset, rhs_offset, depth, r, c, in UnpackResult() 236 rhs_sums_of_each_slice, lhs_offset, rhs_offset, depth, r, c, in UnpackResult() 243 rhs_sums_of_each_slice, lhs_offset, rhs_offset, depth, r, c, in UnpackResult() 256 rhs_sums_of_each_slice, lhs_offset, rhs_offset, depth, r, c, in UnpackResult() [all …]
|
D | multi_thread_gemm.h | 456 lhs_offset(_lhs_offset), in GemmWithPackedRhsTask() 490 lhs_offset.block(curr_result_block.start_row, rs), in Run() 504 const LhsOffset& lhs_offset; member 612 const LhsOffset& lhs_offset, const RhsOffset& rhs_offset, in MultiThreadGemm() argument 635 lhs_offset, rhs_offset, in MultiThreadGemm() 682 lhs_offset, rhs_offset, block_params, output_pipeline)); in MultiThreadGemm()
|
D | dispatch_gemm_shape.h | 159 const LhsOffset& lhs_offset, const RhsOffset& rhs_offset, 177 Transpose(rhs_offset), Transpose(lhs_offset), 184 lhs_offset, rhs_offset, output_pipeline);
|
D | single_thread_gemm.h | 71 const LhsOffset& lhs_offset, const RhsOffset& rhs_offset, in SingleThreadGemm() argument 148 lhs_offset.block(r, rs), rhs_offset.block(c, cs), output_pipeline); in SingleThreadGemm()
|
/external/gemmlowp/doc/ |
D | low-precision.md | 40 - A int32 lhs_offset, that will be added to each entry of the lhs matrix. 46 1. Cast lhs entries from uint8 to int32 and add lhs_offset to each of them. 64 be interpreted during the matrix computation: - lhs_offset - rhs_offset - 70 1. Cast lhs entries from uint8 to int32 and add lhs_offset to each of them. 97 scheme requires adding the lhs_offset and rhs_offset to each of the lhs and rhs 102 lhs_offset and rhs_offset in registers, which would eat into the register space 105 One may then consider adding the lhs_offset and rhs_offset once and for all to 118 Adding lhs_offset to each entry of `lhs`, means adding `lhs_offset * P` to 124 Thus, as far as handling `lhs_offset` and `rhs_offset` goes, the matrix product 128 (lhs + lhs_offset * P) * (rhs + rhs_offset * Q) [all …]
|
D | public.md | 21 `lhs_offset`, `rhs_offset` to them, is explained in 39 int lhs_offset, int rhs_offset, 49 &uint8_result_matrix, lhs_offset, rhs_offset, output_pipeline); 94 * `lhs_offset`, `rhs_offset` are constants added to each matrix entry in the 142 This is a variant where `lhs_offset` and `rhs_offset` may be vectors instead of
|
D | quantization_example.cc | 332 const int lhs_offset = -lhs_qparams.zero_point; in main() local 364 &actual_uint8_result_map, lhs_offset, rhs_offset, output_pipeline); in main()
|
D | quantization.md | 283 mathematical trick to handle `lhs_offset`, `rhs_offset` that we alluded to 287 `lhs_offset`, `rhs_offset` were 0.
|
/external/gemmlowp/test/ |
D | test.cc | 133 MatrixMap<Scalar, ResultOrder>* result, int lhs_offset, in Gemm() 144 const OffsetColDup lhs_offset_vector(lhs_offset, rows); in Gemm() 173 MatrixMap<Scalar, ResultOrder>* result, int lhs_offset, in Gemm() 185 const OffsetColDup lhs_offset_vector(lhs_offset, rows); in Gemm() 210 MatrixMap<Scalar, ResultOrder>* result, int lhs_offset, in Gemm() 215 ResultOrder>(context, lhs, rhs, result, lhs_offset, in Gemm() 244 MatrixMap<Scalar, ResultOrder>* result, int lhs_offset, in Gemm() 253 lhs.cols(), lhs.data(), lhs_offset, lhs.stride(), rhs.data(), in Gemm() 270 MatrixMap<Scalar, ResultOrder>* result, int lhs_offset, in Gemm() 275 rhs.cols(), lhs.cols(), lhs.data(), lhs_offset, in Gemm() [all …]
|
D | correctness_meta_gemm.cc | 64 std::int32_t lhs_offset, std::int32_t rhs_offset, in check_result() argument 74 (static_cast<std::int32_t>(left[depth * i + k]) + lhs_offset) * in check_result() 107 std::int32_t lhs_offset, std::int32_t rhs_offset, in check_result_f() argument 115 (static_cast<std::int32_t>(left[depth * i + k]) + lhs_offset) * in check_result_f() 141 std::int32_t lhs_offset, std::int32_t rhs_offset) { in check_result_i32() argument 148 (static_cast<std::int32_t>(left[depth * i + k]) + lhs_offset) * in check_result_i32()
|
/external/gemmlowp/eight_bit_int_gemm/ |
D | eight_bit_int_gemm.cc | 78 const int lhs_offset = a_offset; in EightBitIntGemmImpl() local 99 context, lhs, rhs, &result, lhs_offset, rhs_offset, result_offset, \ in EightBitIntGemmImpl() 116 const int lhs_offset = a_offset; in EightBitIntGemmInt32Impl() local 136 context, lhs, rhs, &result, lhs_offset, rhs_offset, empty_pipeline); \ in EightBitIntGemmInt32Impl() 254 std::int32_t lhs_offset, std::int32_t rhs_offset, in MetaGemmQuantized8Bit() argument 264 scratch->buffer(), lhs, rhs, m, n, k, lhs_offset, in MetaGemmQuantized8Bit() 271 lhs_offset, sum_offset, multiplicative_offset, in MetaGemmQuantized8Bit() 280 std::int32_t lhs_offset, std::int32_t rhs_offset, in MetaGemmFloat() argument 288 scratch->buffer(), lhs, rhs, m, n, k, lhs_offset, in MetaGemmFloat() 294 lhs_offset, result_offset, result); in MetaGemmFloat()
|
/external/gemmlowp/meta/generators/ |
D | quantized_mul_kernels_common.py | 58 lhs_offset = _ReadParams(emitter, registers, lhs, kernel_m, 4) 71 self.lhs_offsets = _Duplicate(emitter, registers, kernel_m, lhs_offset) 79 for (row, lhs_offset) in zip(data, self.lhs_offsets): 81 emitter.EmitVAdd('s32', row_register, row_register, lhs_offset) 134 lhs_offset = _ReadParams(emitter, registers, lhs, kernel_m, 4) 136 self.lhs_offsets = _Duplicate(emitter, registers, kernel_m, lhs_offset) 144 for (row, lhs_offset) in zip(data, self.lhs_offsets): 146 emitter.EmitVAdd('s32', row_register, row_register, lhs_offset) 166 lhs_offset = _ReadParams(emitter, registers, lhs, kernel_m, 4) 171 self.lhs_offsets = _Duplicate(emitter, registers, kernel_m, lhs_offset) [all …]
|