/external/tensorflow/tensorflow/examples/speech_commands/ |
D | models.py | 92 is_training, runtime_settings=None): argument 127 is_training) 129 return create_conv_model(fingerprint_input, model_settings, is_training) 132 is_training) 135 is_training, runtime_settings) 138 is_training) 141 is_training) 160 def create_single_fc_model(fingerprint_input, model_settings, is_training): argument 185 if is_training: 197 if is_training: [all …]
|
/external/tensorflow/tensorflow/python/ops/ |
D | nn_fused_batchnorm_test.py | 91 is_training=False) 174 is_training=True) 230 is_training=True): argument 240 if is_training and exponential_avg_factor == 1.0: 254 is_training=is_training) 271 is_training=is_training) 293 is_training=True, argument 306 if is_training and exponential_avg_factor == 1.0: 320 is_training=is_training) 324 if is_training: [all …]
|
D | nn_fused_batchnorm_d9m_test.py | 83 for is_training in [False, True]: 91 is_training=is_training, 95 if is_training: 106 is_training=is_training, 111 if is_training: 125 for is_training in [False, True]: 136 is_training=is_training, 140 not is_training):
|
D | nn_fused_batchnorm_deterministic_test.py | 85 for is_training in [False, True]: 93 is_training=is_training, 97 if is_training: 108 is_training=is_training, 113 if is_training: 127 for is_training in [False, True]: 138 is_training=is_training, 142 not is_training):
|
/external/tensorflow/tensorflow/core/kernels/ |
D | fused_batch_norm_ex_op_test.cc | 140 const TensorFormat data_format, bool is_training, in RunFusedBatchNorm() argument 164 ops::FusedBatchNormV3::IsTraining(is_training) in RunFusedBatchNorm() 198 ops::FusedBatchNormGradV3::IsTraining(is_training) in RunFusedBatchNorm() 228 const TensorFormat data_format, bool is_training, in RunFusedBatchNormEx() argument 278 .Attr("is_training", is_training) in RunFusedBatchNormEx() 285 if (is_training) { in RunFusedBatchNormEx() 302 .Attr("is_training", is_training) in RunFusedBatchNormEx() 330 .Attr("is_training", is_training) in RunFusedBatchNormEx() 358 TensorFormat data_format, bool is_training, in VerifyTensorsNear() argument 401 run_default(y_backprop, input, scale, offset, is_training ? empty : mean, in VerifyTensorsNear() [all …]
|
D | fused_batch_norm_op_test.cc | 225 bool is_training, in FusedBatchNormInference() argument 248 .Input(is_training ? empty : other) // mean in FusedBatchNormInference() 249 .Input(is_training ? empty : other) // variance in FusedBatchNormInference() 253 .Attr("is_training", is_training) in FusedBatchNormInference() 261 static Graph* FusedBatchNormGrad(int n, int h, int w, int c, bool is_training, in FusedBatchNormGrad() argument 294 .Attr("is_training", is_training) in FusedBatchNormGrad()
|
/external/tensorflow/tensorflow/core/kernels/mkl/ |
D | mkl_fused_batch_norm_op_test.cc | 48 const float exponential_avg_factor, const bool is_training, Tensor* output, 81 const bool is_training, const GraphRunner& run, in VerifyTensorsClose() argument 96 if (is_training && (exponential_avg_factor == 1.0)) { in VerifyTensorsClose() 113 is_training, &output, &batch_mean, &batch_var); in VerifyTensorsClose() 115 is_training, &mkl_output, &mkl_batch_mean, &mkl_batch_var); in VerifyTensorsClose() 226 const bool is_training) { in VerifyFusedBatchNorm() argument 231 const bool is_training, Tensor* output, in VerifyFusedBatchNorm() 246 attr = attr.IsTraining(is_training); in VerifyFusedBatchNorm() 277 const bool is_training, Tensor* output, in VerifyFusedBatchNorm() 294 .Attr("is_training", is_training) in VerifyFusedBatchNorm() [all …]
|
/external/tensorflow/tensorflow/core/api_def/base_api/ |
D | api_def_FusedBatchNormGradV3.pbtxt | 24 When is_training is True, a 1D Tensor for the computed batch 25 mean to be reused in gradient computation. When is_training is 33 When is_training is True, a 1D Tensor for the computed batch 35 gradient computation. When is_training is False, a 1D Tensor 43 When is_training is True, a 1D Tensor for some intermediate results to be reused 44 in gradient computation. When is_training is False, a dummy empty Tensor will be 105 name: "is_training"
|
D | api_def_FusedBatchNormGrad.pbtxt | 24 When is_training is True, a 1D Tensor for the computed batch 25 mean to be reused in gradient computation. When is_training is 33 When is_training is True, a 1D Tensor for the computed batch 35 gradient computation. When is_training is False, a 1D Tensor 91 name: "is_training"
|
D | api_def_FusedBatchNormGradV2.pbtxt | 24 When is_training is True, a 1D Tensor for the computed batch 25 mean to be reused in gradient computation. When is_training is 33 When is_training is True, a 1D Tensor for the computed batch 35 gradient computation. When is_training is False, a 1D Tensor 97 name: "is_training"
|
D | api_def_CudnnRNNV2.pbtxt | 32 is_training: Indicates whether this operation is used for inference or 35 is only produced if is_training is true. 37 only produced if is_training is true. It is output on host memory rather than
|
/external/tensorflow/tensorflow/security/advisory/ |
D | tfsa-2021-072.md | 21 is_training = False 26 data_format=data_format, is_training=is_training) 44 is_training = False 49 data_format=data_format, is_training=is_training)
|
D | tfsa-2021-043.md | 21 is_training = False 27 data_format=data_format, is_training=is_training)
|
/external/tensorflow/tensorflow/compiler/mlir/tensorflow/tests/ |
D | gpu_fusion.mlir | 9 … {T = "tfdtype$DT_FLOAT", data_format = "NHWC", epsilon = 0.001 : f32, is_training = false} : (ten… 18 … {T = "tfdtype$DT_FLOAT", data_format = "NHWC", epsilon = 0.001 : f32, is_training = false} : (ten… 31 … {T = "tfdtype$DT_FLOAT", data_format = "NHWC", epsilon = 0.001 : f32, is_training = false} : (ten… 43 … {T = "tfdtype$DT_FLOAT", data_format = "NHWC", epsilon = 0.001 : f32, is_training = true} : (tens…
|
D | layout_optimization_layout_assignment_gpu_cc_60.mlir | 80 is_training = true 105 is_training = true 131 is_training = true 157 is_training = true
|
D | layout_optimization_layout_assignment_gpu_cc_70.mlir | 161 is_training = true 186 is_training = true 212 is_training = true 238 is_training = true
|
/external/tensorflow/tensorflow/cc/gradients/ |
D | nn_grad.cc | 400 bool is_training) { in FusedBatchNormGradAttrs() argument 404 result.is_training_ = is_training; in FusedBatchNormGradAttrs() 412 bool is_training, std::vector<Output>* grad_outputs)>; 434 bool is_training; in BaseFusedBatchNormGrad() local 439 GetNodeAttr(op.node()->attrs(), "is_training", &is_training)); in BaseFusedBatchNormGrad() 448 if (is_training) { in BaseFusedBatchNormGrad() 450 data_format, is_training, grad_outputs); in BaseFusedBatchNormGrad() 475 target_data_format, is_training, grad_outputs)); in BaseFusedBatchNormGrad() 493 StringPiece data_format, bool is_training, in FusedBatchNormV3Grad() 499 epsilon, data_format, is_training)); in FusedBatchNormV3Grad()
|
/external/tensorflow/tensorflow/python/compiler/tensorrt/test/ |
D | quantization_mnist_test.py | 190 def _Run(self, is_training, use_trt, batch_size, num_epochs, model_dir): argument 226 if is_training: 257 model_dir=model_dir if is_training else None, 260 if is_training: 280 is_training=False, 289 is_training=False,
|
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v2/ |
D | FusedBatchNormV3.pbtxt | 95 name: "is_training" 198 name: "is_training" 302 name: "is_training"
|
D | FusedBatchNorm.pbtxt | 73 name: "is_training" 159 name: "is_training"
|
D | FusedBatchNormV2.pbtxt | 84 name: "is_training" 181 name: "is_training"
|
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v1/ |
D | FusedBatchNorm.pbtxt | 73 name: "is_training" 159 name: "is_training"
|
D | FusedBatchNormV2.pbtxt | 84 name: "is_training" 181 name: "is_training"
|
D | CudnnRNNV3.pbtxt | 118 name: "is_training" 243 name: "is_training" 382 name: "is_training"
|
/external/tensorflow/tensorflow/compiler/tests/ |
D | fused_batchnorm_test.py | 120 is_training=False) 176 is_training=True) 276 is_training=True) 326 is_training=False) 330 grad, x, scale, mean, var, data_format=data_format, is_training=False)
|