/external/tensorflow/tensorflow/lite/kernels/ |
D | cpu_backend_context.cc | 46 CpuBackendContext::CpuInfo::~CpuInfo() { in ~CpuInfo() 52 bool CpuBackendContext::CpuInfo::EnsureInitialized() { in EnsureInitialized() 59 CpuBackendContext::CpuInfo::InitStatus 60 CpuBackendContext::CpuInfo::Initialize() { in Initialize() 68 bool CpuBackendContext::CpuInfo::Avx2Fma() { in Avx2Fma() 73 bool CpuBackendContext::CpuInfo::Avx() { in Avx() 77 bool CpuBackendContext::CpuInfo::Avx512() { in Avx512() 84 CpuBackendContext::CpuInfo::~CpuInfo() {} in ~CpuInfo() 86 bool CpuBackendContext::CpuInfo::EnsureInitialized() { in EnsureInitialized() 94 bool CpuBackendContext::CpuInfo::Avx2Fma() { return false; } in Avx2Fma() [all …]
|
D | cpu_backend_context.h | 33 class CpuBackendContext final : public TfLiteInternalBackendContext { 35 static CpuBackendContext* GetFromContext(TfLiteContext* context); 37 CpuBackendContext(); 38 ~CpuBackendContext() override; 118 CpuBackendContext(const CpuBackendContext&) = delete;
|
D | cpu_backend_threadpool.h | 38 CpuBackendContext* cpu_backend_context) { in Execute() 50 CpuBackendContext* cpu_backend_context) {
|
D | cpu_backend_gemm_x86.h | 43 CpuBackendContext* context) { in Run() 73 CpuBackendContext* context) {
|
D | fully_connected.cc | 516 CpuBackendContext::GetFromContext(context)); in EvalHybridImpl() 588 CpuBackendContext* cpu_backend_context = in EvalHybrid() 589 CpuBackendContext::GetFromContext(context); in EvalHybrid() 637 CpuBackendContext* cpu_backend_context) { in FullyConnectedInt8() 738 CpuBackendContext::GetFromContext(context)); in EvalQuantized() 744 CpuBackendContext::GetFromContext(context)); in EvalQuantized() 761 CpuBackendContext::GetFromContext(context)); in EvalQuantized() 798 CpuBackendContext::GetFromContext(context)); \ in EvalShuffledQuantized() 821 CpuBackendContext::GetFromContext(context)); in EvalShuffledQuantized() 886 CpuBackendContext::GetFromContext(context)); in EvalFloat() [all …]
|
D | cpu_backend_gemm_eigen.h | 33 CpuBackendContext* /* context */);
|
D | cpu_backend_gemm.h | 120 CpuBackendContext* context) { in Gemm() 175 CpuBackendContext* context) { in Gemm()
|
D | cpu_backend_threadpool_test.cc | 65 CpuBackendContext context; in TestGenerateArrayOfIncrementingInts()
|
D | lstm_eval.h | 170 bool* compute_row_sums, CpuBackendContext* context); 196 CpuBackendContext* context);
|
D | cpu_backend_gemm_gemmlowp.h | 87 CpuBackendContext* context) { 144 CpuBackendContext* context) {
|
D | cpu_backend_gemm_eigen.cc | 41 const GemmParams<float, float>& params, CpuBackendContext* /* context */) { in Run() argument
|
D | mirror_pad.cc | 201 CpuBackendContext* cpu_backend_context = in Eval() 202 CpuBackendContext::GetFromContext(context); in Eval()
|
/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | tensor_utils.h | 35 class CpuBackendContext; variable 47 CpuBackendContext* __restrict__ context); 55 bool* compute_row_sums, CpuBackendContext* context); 67 CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate() 108 int32_t* scratch, int16_t* output, CpuBackendContext* context); 139 int32_t* scratch, int8_t* output, CpuBackendContext* context);
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | neon_tensor_utils_impl.h | 52 CpuBackendContext* context); 60 bool* compute_row_sums, CpuBackendContext* context); 94 int32_t* scratch, int8_t* output, CpuBackendContext* context); 100 int32_t* scratch, int16_t* output, CpuBackendContext* context);
|
D | sse_tensor_utils_impl.h | 44 float* __restrict__ result, CpuBackendContext* context); 53 bool* compute_row_sums, CpuBackendContext* context);
|
D | sse_tensor_utils.h | 58 bool* compute_row_sums, CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate() 69 CpuBackendContext* __restrict__ context) { in MatrixBatchVectorMultiplyAccumulate() 103 int32_t* scratch, int16_t* output, CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate() 113 int32_t* scratch, int8_t* output, CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate()
|
D | neon_tensor_utils.h | 50 CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate() 60 bool* compute_row_sums, CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate() 94 int32_t* scratch, int16_t* output, CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate() 104 int32_t* scratch, int8_t* output, CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate()
|
D | batch_matmul.h | 31 CpuBackendContext* context) { in BatchMatMul() 121 bool* compute_row_sums, CpuBackendContext* context) { in BatchMatMul() 275 CpuBackendContext* context) { in BatchMatMul()
|
D | sse_tensor_utils.cc | 179 int32_t* scratch, CpuBackendContext* context) { in SseCpuBackendGemm() 223 float* __restrict__ result, CpuBackendContext* context) { in SseMatrixBatchVectorMultiplyAccumulate() 275 bool* compute_row_sums, CpuBackendContext* context) { in SseMatrixBatchVectorMultiplyAccumulate()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | portable_tensor_utils_impl.h | 29 class CpuBackendContext; variable 72 bool* compute_row_sums, CpuBackendContext* context); 78 CpuBackendContext* context); 112 int32_t* scratch, int16_t* output, CpuBackendContext* context); 118 int32_t* scratch, int8_t* output, CpuBackendContext* context);
|
D | portable_tensor_utils.h | 80 bool* compute_row_sums, CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate() 93 CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate() 127 int32_t* scratch, int16_t* output, CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate() 137 int32_t* scratch, int8_t* output, CpuBackendContext* context) { in MatrixBatchVectorMultiplyAccumulate()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/sparse_ops/ |
D | fully_connected.h | 82 int thread_end, const CpuBackendContext& cpu_backend_context) { in FullyConnectedSparseWeight1x4Impl() 122 int thread_end, const CpuBackendContext& cpu_backend_context_x) in FullyConnectedSparseWeight1x4Task() 157 const CpuBackendContext& cpu_backend_context; 170 CpuBackendContext* cpu_backend_context) { in FullyConnectedSparseWeight1x4()
|
/external/tensorflow/tensorflow/lite/experimental/kernels/ |
D | gru_cell.h | 39 tflite::CpuBackendContext* cpu_backend_context);
|
D | gru_cell.cc | 44 tflite::CpuBackendContext* cpu_backend_context) { in GruCell()
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/integer_ops/ |
D | fully_connected.h | 35 CpuBackendContext* cpu_backend_context) { in FullyConnected()
|