/external/tensorflow/tensorflow/lite/micro/examples/magic_wand/ |
D | gesture_predictor_test.cc | 26 int prediction; in TF_LITE_MICRO_TEST() local 29 prediction = PredictGesture(probabilities); in TF_LITE_MICRO_TEST() 30 TF_LITE_MICRO_EXPECT_EQ(prediction, kNoGesture); in TF_LITE_MICRO_TEST() 34 prediction = PredictGesture(probabilities); in TF_LITE_MICRO_TEST() 35 TF_LITE_MICRO_EXPECT_EQ(prediction, 0); in TF_LITE_MICRO_TEST() 41 int prediction; in TF_LITE_MICRO_TEST() local 44 prediction = PredictGesture(probabilities); in TF_LITE_MICRO_TEST() 45 TF_LITE_MICRO_EXPECT_EQ(prediction, kNoGesture); in TF_LITE_MICRO_TEST() 50 prediction = PredictGesture(probabilities); in TF_LITE_MICRO_TEST() 51 TF_LITE_MICRO_EXPECT_EQ(prediction, kNoGesture); in TF_LITE_MICRO_TEST() [all …]
|
/external/tensorflow/tensorflow/python/keras/engine/ |
D | partial_batch_padding_handler.py | 95 prediction = np.take(prediction_result, 99 if prediction.shape[0] == 1: 100 prediction = np.squeeze(prediction, axis=0) 101 return prediction 106 prediction = prediction_result[i] 107 prediction = np.take(prediction, np.nonzero( 108 padding_mask[:len(prediction)]), axis=0) 109 predictions.append(np.squeeze(prediction))
|
/external/libgav1/libgav1/src/dsp/x86/ |
D | obmc_sse4.cc | 40 uint8_t* const prediction, const ptrdiff_t prediction_stride, in OverlapBlendFromLeft2xH_SSE4_1() argument 43 uint8_t* pred = prediction; in OverlapBlendFromLeft2xH_SSE4_1() 71 uint8_t* const prediction, const ptrdiff_t prediction_stride, in OverlapBlendFromLeft4xH_SSE4_1() argument 74 uint8_t* pred = prediction; in OverlapBlendFromLeft4xH_SSE4_1() 109 uint8_t* const prediction, const ptrdiff_t prediction_stride, in OverlapBlendFromLeft8xH_SSE4_1() argument 112 uint8_t* pred = prediction; in OverlapBlendFromLeft8xH_SSE4_1() 133 void OverlapBlendFromLeft_SSE4_1(void* const prediction, in OverlapBlendFromLeft_SSE4_1() argument 138 auto* pred = static_cast<uint8_t*>(prediction); in OverlapBlendFromLeft_SSE4_1() 160 pred = static_cast<uint8_t*>(prediction) + x; in OverlapBlendFromLeft_SSE4_1() 188 uint8_t* const prediction, const ptrdiff_t prediction_stride, in OverlapBlendFromTop4xH_SSE4_1() argument [all …]
|
D | convolve_sse4.cc | 250 const int height, void* prediction, in Convolve2D_SSE4_1() argument 271 auto* dest = static_cast<uint8_t*>(prediction); in Convolve2D_SSE4_1() 409 const int height, void* prediction, in ConvolveVertical_SSE4_1() argument 416 auto* dest = static_cast<uint8_t*>(prediction); in ConvolveVertical_SSE4_1() 487 void* prediction, const ptrdiff_t pred_stride) { in ConvolveCompoundCopy_SSE4() argument 490 auto* dest = static_cast<uint16_t*>(prediction); in ConvolveCompoundCopy_SSE4() 545 const int width, const int height, void* prediction, in ConvolveCompoundVertical_SSE4_1() argument 552 auto* dest = static_cast<uint16_t*>(prediction); in ConvolveCompoundVertical_SSE4_1() 618 void* prediction, const ptrdiff_t pred_stride) { in ConvolveHorizontal_SSE4_1() argument 622 auto* dest = static_cast<uint8_t*>(prediction); in ConvolveHorizontal_SSE4_1() [all …]
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | obmc_neon.cc | 51 inline void OverlapBlend2xH_NEON(uint8_t* const prediction, in OverlapBlend2xH_NEON() argument 56 uint8_t* pred = prediction; in OverlapBlend2xH_NEON() 92 uint8_t* const prediction, const ptrdiff_t prediction_stride, in OverlapBlendFromLeft4xH_NEON() argument 95 uint8_t* pred = prediction; in OverlapBlendFromLeft4xH_NEON() 117 uint8_t* const prediction, const ptrdiff_t prediction_stride, in OverlapBlendFromLeft8xH_NEON() argument 120 uint8_t* pred = prediction; in OverlapBlendFromLeft8xH_NEON() 140 void OverlapBlendFromLeft_NEON(void* const prediction, in OverlapBlendFromLeft_NEON() argument 145 auto* pred = static_cast<uint8_t*>(prediction); in OverlapBlendFromLeft_NEON() 167 pred = static_cast<uint8_t*>(prediction) + x; in OverlapBlendFromLeft_NEON() 197 inline void OverlapBlendFromTop4x4_NEON(uint8_t* const prediction, in OverlapBlendFromTop4x4_NEON() argument [all …]
|
D | convolve_neon.cc | 744 void* const prediction, const ptrdiff_t pred_stride) { in Filter2DVertical() argument 745 auto* const dest = static_cast<uint8_t*>(prediction); in Filter2DVertical() 765 const int height, void* const prediction, in Convolve2D_NEON() argument 790 Filter2DVertical<8>(intermediate_result, width, height, taps, prediction, in Convolve2D_NEON() 793 Filter2DVertical<6>(intermediate_result, width, height, taps, prediction, in Convolve2D_NEON() 796 Filter2DVertical<4>(intermediate_result, width, height, taps, prediction, in Convolve2D_NEON() 799 Filter2DVertical<2>(intermediate_result, width, height, taps, prediction, in Convolve2D_NEON() 1486 void* const prediction, const ptrdiff_t pred_stride) { in ConvolveScale2D_NEON() argument 1593 prediction, pred_stride); in ConvolveScale2D_NEON() 1597 prediction, pred_stride); in ConvolveScale2D_NEON() [all …]
|
/external/libtextclassifier/native/lang_id/ |
D | lang-id-wrapper.cc | 76 const auto& prediction = langid_result.predictions[i]; in GetPredictions() local 77 if (prediction.second >= noise_threshold && prediction.first != "und") { in GetPredictions() 78 prediction_results.push_back({prediction.first, prediction.second}); in GetPredictions() 93 const auto& prediction = predictions[i]; in GetLanguageTags() local 94 if (threshold >= 0.f && prediction.second < threshold) { in GetLanguageTags() 102 detected_language_tags += prediction.first; in GetLanguageTags()
|
/external/tensorflow/tensorflow/lite/experimental/examples/lstm/ |
D | bidirectional_sequence_rnn_test.py | 142 prediction = tf.matmul(output, out_weights) + out_bias 143 output_class = tf.nn.softmax(prediction, name="OUTPUT_CLASS") 145 return x, prediction, output_class 147 def trainModel(self, x, prediction, output_class, sess): argument 160 tf.nn.softmax_cross_entropy_with_logits(logits=prediction, labels=y)) 212 x, prediction, output_class = self.buildModel( 218 return x, prediction, output_class, new_sess 285 x, prediction, output_class = self.buildModel( 287 self.trainModel(x, prediction, output_class, sess) 290 x, prediction, output_class, new_sess = self.saveAndRestoreModel( [all …]
|
D | unidirectional_sequence_lstm_test.py | 111 prediction = tf.matmul(outputs[-1], out_weights) + out_bias 112 output_class = tf.nn.softmax(prediction, name="OUTPUT_CLASS") 114 return x, prediction, output_class 116 def trainModel(self, x, prediction, output_class, sess): argument 129 tf.nn.softmax_cross_entropy_with_logits(logits=prediction, labels=y)) 170 x, prediction, output_class = self.buildModel(lstm_layer, is_dynamic_rnn) 175 return x, prediction, output_class, new_sess 245 x, prediction, output_class = self.buildModel( 247 self.trainModel(x, prediction, output_class, sess) 250 x, prediction, output_class, new_sess = self.saveAndRestoreModel( [all …]
|
D | unidirectional_sequence_rnn_test.py | 107 prediction = tf.matmul(outputs[-1], out_weights) + out_bias 108 output_class = tf.nn.softmax(prediction, name="OUTPUT_CLASS") 110 return x, prediction, output_class 112 def trainModel(self, x, prediction, output_class, sess): argument 125 tf.nn.softmax_cross_entropy_with_logits(logits=prediction, labels=y)) 165 x, prediction, output_class = self.buildModel(rnn_layer, is_dynamic_rnn) 170 return x, prediction, output_class, new_sess 236 x, prediction, output_class = self.buildModel( 238 self.trainModel(x, prediction, output_class, sess) 241 x, prediction, output_class, new_sess = self.saveAndRestoreModel( [all …]
|
D | bidirectional_sequence_lstm_test.py | 121 prediction = tf.matmul(output, out_weights) + out_bias 122 output_class = tf.nn.softmax(prediction, name="OUTPUT_CLASS") 124 return x, prediction, output_class 126 def trainModel(self, x, prediction, output_class, sess): argument 139 tf.nn.softmax_cross_entropy_with_logits(logits=prediction, labels=y)) 184 x, prediction, output_class = self.buildModel(fw_lstm_layer, bw_lstm_layer, 190 return x, prediction, output_class, new_sess 260 x, prediction, output_class = self.buildModel(self.buildLstmLayer(), 262 self.trainModel(x, prediction, output_class, sess) 265 x, prediction, output_class, new_sess = self.saveAndRestoreModel( [all …]
|
/external/libgav1/libgav1/src/dsp/ |
D | convolve.cc | 63 const int width, const int height, void* prediction, in ConvolveScale2D_C() argument 90 auto* dest = static_cast<Pixel*>(prediction); in ConvolveScale2D_C() 147 void* prediction, const ptrdiff_t pred_stride) { in ConvolveCompoundScale2D_C() argument 176 auto* dest = static_cast<uint16_t*>(prediction); in ConvolveCompoundScale2D_C() 232 const int height, void* prediction, in ConvolveCompound2D_C() argument 260 auto* dest = static_cast<uint16_t*>(prediction); in ConvolveCompound2D_C() 314 const int width, const int height, void* prediction, in Convolve2D_C() argument 339 auto* dest = static_cast<Pixel*>(prediction); in Convolve2D_C() 394 const int height, void* prediction, in ConvolveHorizontal_C() argument 403 auto* dest = static_cast<Pixel*>(prediction); in ConvolveHorizontal_C() [all …]
|
D | obmc.cc | 33 void OverlapBlendVertical_C(void* const prediction, in OverlapBlendVertical_C() argument 37 auto* pred = static_cast<Pixel*>(prediction); in OverlapBlendVertical_C() 56 void OverlapBlendHorizontal_C(void* const prediction, in OverlapBlendHorizontal_C() argument 61 auto* pred = static_cast<Pixel*>(prediction); in OverlapBlendHorizontal_C()
|
/external/tensorflow/tensorflow/lite/examples/label_image/ |
D | get_top_n_impl.h | 33 void get_top_n(T* prediction, int prediction_size, size_t num_results, in get_top_n() argument 47 value = prediction[i]; in get_top_n() 50 value = (prediction[i] + 128) / 256.0; in get_top_n() 53 value = prediction[i] / 255.0; in get_top_n()
|
/external/python/google-api-python-client/samples/prediction/ |
D | README | 1 Before you can run the prediction sample prediction.py, you must load some csv 7 api: prediction
|
/external/libavc/common/arm/ |
D | ih264_intra_pred_luma_4x4_a9q.s | 26 @* Contains function definitions for intra 4x4 Luma prediction . 62 @* Perform Intra prediction for luma_4x4 mode:vertical 65 @* Perform Intra prediction for luma_4x4 mode:vertical ,described in sec 8.3.1.2.1 135 @* Perform Intra prediction for luma_4x4 mode:horizontal 138 @* Perform Intra prediction for luma_4x4 mode:horizontal ,described in sec 8.3.1.2.2 217 @* Perform Intra prediction for luma_4x4 mode:DC 220 @* Perform Intra prediction for luma_4x4 mode:DC ,described in sec 8.3.1.2.3 356 @* Perform Intra prediction for luma_4x4 mode:Diagonal_Down_Left 359 @* Perform Intra prediction for luma_4x4 mode:Diagonal_Down_Left ,described in sec 8.3.1.2.4 438 @* Perform Intra prediction for luma_4x4 mode:Diagonal_Down_Right [all …]
|
D | ih264_intra_pred_luma_8x8_a9q.s | 26 @* Contains function definitions for intra 8x8 Luma prediction . 67 @* Reference sample filtering process for Intra_8x8 sample prediction 70 @* Perform Reference sample filtering process for Intra_8x8 sample prediction ,described in sec 8.… 152 @* Perform Intra prediction for luma_8x8 mode:vertical 155 @* Perform Intra prediction for luma_8x8 mode:vertical ,described in sec 8.3.2.2.2 225 @* Perform Intra prediction for luma_8x8 mode:horizontal 228 @* Perform Intra prediction for luma_8x8 mode:horizontal ,described in sec 8.3.2.2.2 305 @* Perform Intra prediction for luma_8x8 mode:DC 308 @* Perform Intra prediction for luma_8x8 mode:DC ,described in sec 8.3.2.2.3 416 @* Perform Intra prediction for luma_8x8 mode:Diagonal_Down_Left [all …]
|
D | ih264_intra_pred_luma_16x16_a9q.s | 26 @* Contains function definitions for intra 16x16 Luma prediction . 66 @* Perform Intra prediction for luma_16x16 mode:vertical 69 @* Perform Intra prediction for luma_16x16 mode:Vertical ,described in sec 8.3.3.1 147 @* Perform Intra prediction for luma_16x16 mode:horizontal 150 @* Perform Intra prediction for luma_16x16 mode:horizontal ,described in sec 8.3.3.2 225 @* Perform Intra prediction for luma_16x16 mode:DC 228 @* Perform Intra prediction for luma_16x16 mode:DC ,described in sec 8.3.3.3 349 @* Perform Intra prediction for luma_16x16 mode:PLANE 352 @* Perform Intra prediction for luma_16x16 mode:PLANE ,described in sec 8.3.3.4
|
D | ih264_intra_pred_chroma_a9q.s | 26 @* Contains function definitions for intra chroma prediction . 65 @* Perform Intra prediction for chroma_8x8 mode:DC 68 @* Perform Intra prediction for chroma_8x8 mode:DC ,described in sec 8.3.4.1 197 @* Perform Intra prediction for chroma_8x8 mode:Horizontal 200 @* Perform Intra prediction for chroma_8x8 mode:Horizontal ,described in sec 8.3.4.2 273 @* Perform Intra prediction for chroma_8x8 mode:vertical 276 @*Perform Intra prediction for chroma_8x8 mode:vertical ,described in sec 8.3.4.3 345 @* Perform Intra prediction for chroma_8x8 mode:PLANE 348 @* Perform Intra prediction for chroma_8x8 mode:PLANE ,described in sec 8.3.4.4
|
/external/arm-neon-tests/ |
D | InitCache.s | 5 ; and program flow prediction 39 ; Cortex-A8 program flow prediction 43 ORR r0, r0, #(0x1 <<11) ; Enable all forms of branch prediction 44 ;BIC r0, r0, #(0x1 << 11) ; Disable all forms of branch prediction
|
/external/libtextclassifier/native/actions/ |
D | ngram-model.cc | 218 const auto prediction = Eval(message_unicode); in EvalConversation() local 219 if (prediction.first) { in EvalConversation() 220 return prediction; in EvalConversation() 222 score = std::max(score, prediction.second); in EvalConversation()
|
/external/tensorflow/tensorflow/core/kernels/boosted_trees/ |
D | boosted_trees.proto | 167 // prediction paths 3) Leaf node IDs. 169 // Return the logits and associated feature splits across prediction paths for 171 // compute DFCs in Python, by subtracting each child prediction from its 172 // parent prediction and associating this change with its respective feature 176 // Return the node_id for each leaf node we reach in our prediction path. 178 // TODO(crawles): return 4) Node IDs for ensemble prediction path
|
/external/libgav1/libgav1/src/tile/ |
D | prediction.cc | 628 void* prediction[2]; in CompoundInterPrediction() local 632 prediction[0] = block.scratch_buffer->prediction_buffer[0]; in CompoundInterPrediction() 633 prediction[1] = block.scratch_buffer->prediction_buffer[1]; in CompoundInterPrediction() 636 prediction[0] = block.scratch_buffer->compound_prediction_buffer_8bpp[0]; in CompoundInterPrediction() 637 prediction[1] = block.scratch_buffer->compound_prediction_buffer_8bpp[1]; in CompoundInterPrediction() 648 prediction[0], prediction[1], in CompoundInterPrediction() 654 DistanceWeightedPrediction(prediction[0], prediction[1], prediction_width, in CompoundInterPrediction() 661 dsp_.average_blend(prediction[0], prediction[1], prediction_width, in CompoundInterPrediction() 839 uint8_t* const prediction = GetStartPoint(buffer_, plane, x, y, bitdepth); in ObmcBlockPrediction() local 841 dsp_.obmc_blend[blending_direction](prediction, prediction_stride, width, in ObmcBlockPrediction() [all …]
|
/external/tensorflow/tensorflow/python/keras/integration_test/ |
D | vectorized_map_test.py | 34 prediction = layer(inp) 35 loss = tf.nn.l2_loss(label - prediction)
|
/external/tensorflow/tensorflow/core/kernels/ |
D | in_topk_op_gpu.cu.cc | 53 T prediction = ldg(predictions + i); in ComputePredictionMaskKernel() local 57 if (!Eigen::numext::isfinite(prediction) || in ComputePredictionMaskKernel() 61 mask[i] = prediction > target_prediction ? 1 : 0; in ComputePredictionMaskKernel()
|