/external/tensorflow/tensorflow/lite/kernels/ |
D | transpose_conv.cc | 352 TfLiteTensor* scratch_buffer; in Prepare() local 355 &scratch_buffer)); in Prepare() 357 scratch_buffer->type = kTfLiteInt64; in Prepare() 359 scratch_buffer->type = kTfLiteInt32; in Prepare() 362 scratch_buffer->allocation_type = kTfLiteDynamic; in Prepare() 364 SetTensorToDynamic(scratch_buffer); in Prepare() 367 ResizeTensor(context, output_shape, scratch_buffer)); in Prepare() 439 TfLiteTensor* output, TfLiteTensor* scratch_buffer) { in EvalQuantized() argument 468 GetTensorData<int32_t>(scratch_buffer)); in EvalQuantized() 478 GetTensorData<int32>(col2im), GetTensorData<int32>(scratch_buffer), in EvalQuantized() [all …]
|
D | batch_matmul.cc | 155 TfLiteTensor* scratch_buffer; in InitializeTemporaries() local 157 context, GetTemporarySafe(context, node, /*index=*/0, &scratch_buffer)); in InitializeTemporaries() 166 scratch_buffer->type = op_context->lhs->type; in InitializeTemporaries() 167 scratch_buffer->allocation_type = kTfLiteArenaRw; in InitializeTemporaries() 168 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, scratch_buffer, in InitializeTemporaries() 178 TfLiteTensor* scratch_buffer; in InitializeTemporaries() local 180 context, GetTemporarySafe(context, node, /*index=*/1, &scratch_buffer)); in InitializeTemporaries() 192 scratch_buffer->allocation_type = kTfLiteArenaRwPersistent; in InitializeTemporaries() 194 scratch_buffer->allocation_type = kTfLiteArenaRw; in InitializeTemporaries() 196 scratch_buffer->type = op_context->rhs->type; in InitializeTemporaries() [all …]
|
D | unidirectional_sequence_lstm.cc | 962 TfLiteTensor* scratch_buffer; in Prepare() local 964 &scratch_buffer)); in Prepare() 965 scratch_buffer->type = input->type; in Prepare() 966 scratch_buffer->allocation_type = kTfLiteArenaRw; in Prepare() 980 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, scratch_buffer, in Prepare() 1326 TfLiteTensor* scratch_buffer; in Eval() local 1328 &scratch_buffer)); in Eval() 1345 /*output_offset=*/0, scratch_buffer, output_state, cell_state, in Eval() 1353 TfLiteTensor* scratch_buffer; in Eval() local 1356 GetTemporarySafe(context, node, kScratchBuffer, &scratch_buffer)); in Eval() [all …]
|
D | lstm_eval.h | 123 int output_offset, TfLiteTensor* scratch_buffer, TfLiteTensor* output_state, 161 TfLiteTensor* scratch_buffer, TfLiteTensor* input_sf,
|
D | lstm.cc | 1409 TfLiteTensor* scratch_buffer; in Prepare() local 1411 &scratch_buffer)); in Prepare() 1412 scratch_buffer->type = input->type; in Prepare() 1413 scratch_buffer->allocation_type = kTfLiteArenaRw; in Prepare() 1427 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, scratch_buffer, in Prepare() 1908 TfLiteTensor* scratch_buffer; in Eval() local 1910 GetTemporarySafe(context, node, 0, &scratch_buffer)); in Eval() 1928 /*output_offset=*/0, scratch_buffer, output_state, cell_state, in Eval()
|
/external/libgav1/libgav1/src/ |
D | tile_scratch_buffer.h | 147 std::unique_ptr<TileScratchBuffer> scratch_buffer(new (std::nothrow) in Get() 149 if (scratch_buffer == nullptr || !scratch_buffer->Init(bitdepth_)) { in Get() 152 return scratch_buffer; in Get() 157 void Release(std::unique_ptr<TileScratchBuffer> scratch_buffer) { in Release() argument 159 buffers_.Push(std::move(scratch_buffer)); in Release()
|
D | frame_scratch_buffer.h | 113 std::unique_ptr<FrameScratchBuffer> scratch_buffer(new (std::nothrow) in Get() 115 return scratch_buffer; in Get() 118 void Release(std::unique_ptr<FrameScratchBuffer> scratch_buffer) { in Release() argument 120 buffers_.Push(std::move(scratch_buffer)); in Release()
|
D | tile.h | 118 bool ProcessSuperBlockRow(int row4x4, TileScratchBuffer* scratch_buffer); 296 int row4x4_start, int column4x4_start, TileScratchBuffer* scratch_buffer, 299 TileScratchBuffer* scratch_buffer, 306 TileScratchBuffer* scratch_buffer); 310 TileScratchBuffer* scratch_buffer, ResidualPtr* residual); 312 void ClearBlockDecoded(TileScratchBuffer* scratch_buffer, int row4x4, 315 TileScratchBuffer* scratch_buffer, 787 TileScratchBuffer* const scratch_buffer, ResidualPtr* residual) in Block() 796 scratch_buffer(scratch_buffer), in Block() 937 TileScratchBuffer* const scratch_buffer; member [all …]
|
/external/libgav1/libgav1/src/tile/ |
D | tile.cc | 601 TileScratchBuffer* const scratch_buffer) { in ProcessSuperBlockRow() argument 603 assert(scratch_buffer != nullptr); in ProcessSuperBlockRow() 607 if (!ProcessSuperBlock(row4x4, column4x4, scratch_buffer, in ProcessSuperBlockRow() 627 int row4x4, TileScratchBuffer* scratch_buffer); 630 int row4x4, TileScratchBuffer* scratch_buffer); 645 std::unique_ptr<TileScratchBuffer> scratch_buffer = in ParseAndDecode() local 647 if (scratch_buffer == nullptr) { in ParseAndDecode() 656 row4x4, scratch_buffer.get())) { in ParseAndDecode() 661 tile_scratch_buffer_pool_->Release(std::move(scratch_buffer)); in ParseAndDecode() 668 std::unique_ptr<TileScratchBuffer> scratch_buffer = in Parse() local [all …]
|
D | prediction.cc | 537 if (!block.scratch_buffer->cfl_luma_buffer_valid) { in ChromaFromLumaPrediction() 541 block.scratch_buffer->cfl_luma_buffer, in ChromaFromLumaPrediction() 546 block.scratch_buffer->cfl_luma_buffer_valid = true; in ChromaFromLumaPrediction() 553 buffer_[plane].columns(), block.scratch_buffer->cfl_luma_buffer, in ChromaFromLumaPrediction() 618 prediction[0] = block.scratch_buffer->prediction_buffer[0]; in CompoundInterPrediction() 619 prediction[1] = block.scratch_buffer->prediction_buffer[1]; in CompoundInterPrediction() 622 prediction[0] = block.scratch_buffer->compound_prediction_buffer_8bpp[0]; in CompoundInterPrediction() 623 prediction[1] = block.scratch_buffer->compound_prediction_buffer_8bpp[1]; in CompoundInterPrediction() 735 candidate_column, block.scratch_buffer->prediction_buffer[index], in InterPrediction() 772 block.scratch_buffer->prediction_buffer[0], in InterPrediction() [all …]
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/integer_ops/ |
D | transpose_conv.h | 31 int32_t* scratch_buffer) { in TransposeConv() argument 63 memset(scratch_buffer, 0, num_elements * sizeof(int32_t)); in TransposeConv() 88 scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv() 104 int32_t acc = scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv() 130 std::int64_t* scratch_buffer) { in TransposeConv() argument 160 memset(scratch_buffer, 0, num_elements * sizeof(std::int64_t)); in TransposeConv() 185 scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv() 201 std::int64_t acc = scratch_buffer[Offset(output_shape, batch, out_y, in TransposeConv()
|
/external/libxaac/decoder/drc_src/ |
D | impd_drc_main_td_process.c | 126 FLOAT32 *scratch_buffer = (FLOAT32 *)p_obj_drc->pp_mem[1]; in impd_process_time_domain() local 130 scratch_buffer + i * (p_obj_drc->str_config.frame_size + 32); in impd_process_time_domain() 139 FLOAT32 *scratch_buffer = (FLOAT32 *)p_obj_drc->pp_mem[1]; in impd_process_time_domain() local 143 scratch_buffer + i * (p_obj_drc->str_config.frame_size + 32); in impd_process_time_domain() 159 FLOAT32 *scratch_buffer = (FLOAT32 *)p_obj_drc->pp_mem[1]; in impd_process_time_domain() local 163 scratch_buffer + i * (p_obj_drc->str_config.frame_size + 32); in impd_process_time_domain()
|
/external/mesa3d/src/gallium/drivers/radeonsi/ |
D | si_cp_dma.c | 266 if (!sctx->scratch_buffer || sctx->scratch_buffer->b.b.width0 < scratch_size) { in si_cp_dma_realign_engine() 267 si_resource_reference(&sctx->scratch_buffer, NULL); in si_cp_dma_realign_engine() 268 sctx->scratch_buffer = si_aligned_buffer_create(&sctx->screen->b, in si_cp_dma_realign_engine() 271 if (!sctx->scratch_buffer) in si_cp_dma_realign_engine() 277 si_cp_dma_prepare(sctx, &sctx->scratch_buffer->b.b, &sctx->scratch_buffer->b.b, size, size, in si_cp_dma_realign_engine() 280 va = sctx->scratch_buffer->gpu_address; in si_cp_dma_realign_engine()
|
D | si_state_shaders.c | 3565 uint64_t scratch_va = sctx->scratch_buffer->gpu_address; in si_update_scratch_buffer() 3583 if (shader->scratch_bo == sctx->scratch_buffer) { in si_update_scratch_buffer() 3588 assert(sctx->scratch_buffer); in si_update_scratch_buffer() 3599 si_resource_reference(&shader->scratch_bo, sctx->scratch_buffer); in si_update_scratch_buffer() 3709 if (!sctx->scratch_buffer || scratch_needed_size > sctx->scratch_buffer->b.b.width0) { in si_update_spi_tmpring_size() 3711 si_resource_reference(&sctx->scratch_buffer, NULL); in si_update_spi_tmpring_size() 3713 sctx->scratch_buffer = si_aligned_buffer_create( in si_update_spi_tmpring_size() 3718 if (!sctx->scratch_buffer) in si_update_spi_tmpring_size() 3722 si_context_add_resource_size(sctx, &sctx->scratch_buffer->b.b); in si_update_spi_tmpring_size() 4125 if (sctx->scratch_buffer) { in si_emit_scratch_state() [all …]
|
D | si_gfx_cs.c | 565 if (ctx->scratch_buffer) { in si_begin_new_gfx_cs() 566 si_context_add_resource_size(ctx, &ctx->scratch_buffer->b.b); in si_begin_new_gfx_cs()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | transpose_conv.h | 120 uint8_t* im2col_data, int32_t* scratch_buffer) { in TransposeConv() argument 155 memset(scratch_buffer, 0, num_elements * sizeof(int32_t)); in TransposeConv() 180 scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv() 196 int32_t acc = scratch_buffer[Offset(output_shape, batch, out_y, out_x, in TransposeConv()
|
/external/ComputeLibrary/src/runtime/CL/functions/ |
D | CLLSTMLayer.cpp | 78 …ICLTensor *scratch_buffer, ICLTensor *output_state_out, ICLTensor *cell_state_out, ICLTensor *outp… in configure() argument 82 …e_bias, cell_bias, output_gate_bias, output_state_in, cell_state_in, scratch_buffer, output_state_… in configure() 91 …ICLTensor *scratch_buffer, ICLTensor *output_state_out, ICLTensor *cell_state_out, ICLTensor *outp… in configure() argument 99 scratch_buffer, output_state_out, cell_state_out, output); in configure() 113 … scratch_buffer->info(), output_state_out->info(), cell_state_out->info(), output->info(), in configure() 401 _concat_scratch_buffer.configure(compile_context, scratch_inputs, scratch_buffer, Window::DimX); in configure() 413 …const ITensorInfo *scratch_buffer, const ITensorInfo *output_state_out, const ITensorInfo *cell_st… in validate() argument 421 scratch_buffer, output_state_out, cell_state_out, output); in validate() 430 … scratch_buffer, output_state_out, cell_state_out, output); in validate() 445 ARM_COMPUTE_RETURN_ERROR_ON(scratch_buffer->num_dimensions() > 2); in validate() [all …]
|
/external/ComputeLibrary/arm_compute/runtime/CL/functions/ |
D | CLLSTMLayer.h | 113 …ICLTensor *scratch_buffer, ICLTensor *output_state_out, ICLTensor *cell_state_out, ICLTensor *outp… 159 …ICLTensor *scratch_buffer, ICLTensor *output_state_out, ICLTensor *cell_state_out, ICLTensor *outp… 207 …const ITensorInfo *scratch_buffer, const ITensorInfo *output_state_out, const ITensorInfo *cell_st…
|
/external/ComputeLibrary/src/runtime/NEON/functions/ |
D | NELSTMLayer.cpp | 71 … ITensor *scratch_buffer, ITensor *output_state_out, ITensor *cell_state_out, ITensor *output, in configure() argument 79 scratch_buffer, output_state_out, cell_state_out, output); in configure() 93 … scratch_buffer->info(), output_state_out->info(), cell_state_out->info(), output->info(), in configure() 370 _concat_scratch_buffer.configure(scratch_inputs, scratch_buffer, Window::DimX); in configure() 382 …const ITensorInfo *scratch_buffer, const ITensorInfo *output_state_out, const ITensorInfo *cell_st… in validate() argument 390 scratch_buffer, output_state_out, cell_state_out, output); in validate() 399 … scratch_buffer, output_state_out, cell_state_out, output); in validate() 414 ARM_COMPUTE_RETURN_ERROR_ON(scratch_buffer->num_dimensions() > 2); in validate() 418 ARM_COMPUTE_RETURN_ERROR_ON(cell_bias->dimension(0) * 4 != scratch_buffer->dimension(0) in validate() 419 && cell_bias->dimension(0) * 3 != scratch_buffer->dimension(0)); in validate() [all …]
|
/external/pigweed/pw_protobuf/public/pw_protobuf/ |
D | encoder.h | 118 constexpr StreamEncoder(stream::Writer& writer, ByteSpan scratch_buffer) in StreamEncoder() argument 122 memory_writer_(scratch_buffer), in StreamEncoder() 650 constexpr StreamEncoder(StreamEncoder& parent, ByteSpan scratch_buffer) in StreamEncoder() argument 651 : status_(scratch_buffer.empty() ? Status::ResourceExhausted() in StreamEncoder() 655 memory_writer_(scratch_buffer), in StreamEncoder()
|
/external/ComputeLibrary/arm_compute/runtime/NEON/functions/ |
D | NELSTMLayer.h | 103 … ITensor *scratch_buffer, ITensor *output_state_out, ITensor *cell_state_out, ITensor *output, 150 …const ITensorInfo *scratch_buffer, const ITensorInfo *output_state_out, const ITensorInfo *cell_st…
|
/external/tensorflow/tensorflow/lite/tools/optimize/calibration/builtin_logging_ops/ |
D | lstm.cc | 289 int output_offset, TfLiteTensor* scratch_buffer, TfLiteTensor* output_state, in EvalCalibration() 315 float* scratch_buffer_ptr = GetTensorData<float>(scratch_buffer); in EvalCalibration() 558 TfLiteTensor* scratch_buffer; in lstm_eval() local 560 context, GetTemporarySafe(context, node, /*index=*/0, &scratch_buffer)); in lstm_eval() 625 /*output_offset=*/0, scratch_buffer, output_state, cell_state, output, in lstm_eval()
|
/external/tensorflow/tensorflow/lite/tools/optimize/calibration/custom_logging_ops/ |
D | lstm.cc | 365 int output_offset, TfLiteTensor* scratch_buffer, in EvalFloat() 392 float* scratch_buffer_ptr = GetTensorData<float>(scratch_buffer); in EvalFloat() 604 TfLiteTensor* scratch_buffer = GetTemporary(context, node, /*index=*/0); in lstm_eval() local 639 /*output_offset=*/0, scratch_buffer, activation_state, cell_state, in lstm_eval()
|
/external/libxaac/decoder/ |
D | ixheaacd_imdct.c | 220 WORD32 *ptr_scratch = &usac_data->scratch_buffer[0]; in ixheaacd_cal_fac_data() 313 WORD32 *scratch_mem = usac_data->scratch_buffer; in ixheaacd_fd_imdct_short() 440 WORD32 *scratch_mem = usac_data->scratch_buffer; in ixheaacd_fd_imdct_long()
|
D | ixheaacd_fwd_alias_cnx.c | 103 WORD32 *ptr_scratch = &usac_data->scratch_buffer[0]; in ixheaacd_fwd_alias_cancel_tool()
|