/external/tensorflow/tensorflow/core/kernels/mkl/ |
D | mkl_fused_batch_norm_op_test.cc | 49 Tensor* batch_mean, Tensor* batch_var)>; 107 Tensor batch_var; in VerifyTensorsClose() local 113 is_training, &output, &batch_mean, &batch_var); in VerifyTensorsClose() 121 ASSERT_EQ(batch_var.dtype(), mkl_batch_var.dtype()); in VerifyTensorsClose() 122 ASSERT_EQ(batch_var.shape(), mkl_batch_var.shape()); in VerifyTensorsClose() 126 test::ExpectClose(batch_var, mkl_batch_var, 1e-5); in VerifyTensorsClose() 232 Tensor* batch_mean, Tensor* batch_var) { in VerifyFusedBatchNorm() argument 270 *batch_var = output_tensors[2]; in VerifyFusedBatchNorm() 278 Tensor* batch_mean, Tensor* batch_var) { in VerifyFusedBatchNorm() argument 335 batch_var); in VerifyFusedBatchNorm() [all …]
|
/external/tensorflow/tensorflow/compiler/mlir/tensorflow/tests/ |
D | layout_optimization_layout_assignment_gpu_cc_60.mlir | 74 %y, %batch_mean, %batch_var, %reserve_1, %reserve_2, %reserve_3 99 %y, %batch_mean, %batch_var, %reserve_1, %reserve_2, %reserve_3
|
D | layout_optimization_layout_assignment_gpu_cc_70.mlir | 155 %y, %batch_mean, %batch_var, %reserve_1, %reserve_2, %reserve_3 180 %y, %batch_mean, %batch_var, %reserve_1, %reserve_2, %reserve_3
|
D | layout_optimization_layout_assignment_to_nhwc.mlir | 58 %y, %batch_mean, %batch_var, %reserve_1, %reserve_2, %reserve_3
|
D | layout_optimization_layout_assignment_to_nchw.mlir | 171 %y, %batch_mean, %batch_var, %reserve_1, %reserve_2, %reserve_3
|
D | layout_optimization_move_transposes_end.mlir | 186 %2, %batch_mean, %batch_var, %reserve_1, %reserve_2, %reserve_3
|
/external/tensorflow/tensorflow/core/kernels/ |
D | fused_batch_norm_op.cc | 775 Tensor* batch_mean, Tensor* batch_var, Tensor* saved_mean, in operator ()() 833 f(context->eigen_device<GPUDevice>(), batch_var->flat<U>()); in operator ()() 927 auto batch_var_ptr = StreamExecutorUtil::AsDeviceMemory<U>(*batch_var); in operator ()() 953 if (!batch_var->SharesBufferWith(estimated_variance) && in operator ()() 1361 Tensor* batch_var = nullptr; in ComputeWithReservedSpace() local 1363 {4}, 2, scale.shape(), &batch_var)); in ComputeWithReservedSpace() 1375 batch_mean, batch_var, saved_mean, saved_maybe_inv_var, in ComputeWithReservedSpace() 1381 batch_mean, batch_var, saved_mean, saved_maybe_inv_var, in ComputeWithReservedSpace()
|
/external/tensorflow/tensorflow/core/grappler/costs/ |
D | virtual_scheduler_test.cc | 898 auto batch_var = batch_norm.batch_variance; in CreateGrapplerItemWithBatchNorm() local 901 auto z2 = ops::Add(s.WithOpName("z2"), batch_var, batch_var); in CreateGrapplerItemWithBatchNorm() 902 auto z3 = ops::Add(s.WithOpName("z3"), batch_var, batch_var); in CreateGrapplerItemWithBatchNorm() 909 auto z4 = ops::NoOp(s.WithControlDependencies(batch_var).WithOpName("z4")); in CreateGrapplerItemWithBatchNorm() 2235 auto batch_var = batch_norm.batch_variance; in CreateGrapplerItemWithInterDeviceTransfers() local 2244 ops::Identity(s.WithOpName("batch_var1").WithDevice(kCPU1), batch_var); in CreateGrapplerItemWithInterDeviceTransfers()
|
/external/tensorflow/tensorflow/compiler/xla/stream_executor/cuda/ |
D | cuda_dnn.h | 288 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, 303 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, 547 DeviceMemory<U>* batch_mean, DeviceMemory<U>* batch_var,
|
D | cuda_dnn.cc | 5417 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, in DoBatchNormalizationForward() argument 5426 activation_mode, y, batch_mean, batch_var, saved_mean, saved_inv_var, in DoBatchNormalizationForward() 5441 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, in DoBatchNormalizationForward() argument 5450 activation_mode, y, batch_mean, batch_var, saved_mean, saved_inv_var, in DoBatchNormalizationForward() 5466 DeviceMemory<U>* batch_mean, DeviceMemory<U>* batch_var, in DoBatchNormalizationForwardImpl() argument 5534 CHECK_EQ(batch_mean->is_null(), batch_var->is_null()) in DoBatchNormalizationForwardImpl() 5539 if (!batch_mean->is_null() && !batch_var->is_null()) { in DoBatchNormalizationForwardImpl() 5542 stream->ThenMemZero(batch_var, batch_var->size()); in DoBatchNormalizationForwardImpl() 5545 batch_var_opaque = batch_var->opaque(); in DoBatchNormalizationForwardImpl()
|
/external/tensorflow/tensorflow/python/ops/ |
D | nn_fused_batchnorm_test.py | 119 batch_mean, batch_var = nn_impl.moments( 122 y = self._batch_norm(x, batch_mean, batch_var, offset, scale, epsilon) 133 batch_var_corrected = batch_var * sample_size / (
|
/external/tensorflow/tensorflow/stream_executor/rocm/ |
D | rocm_dnn.h | 292 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, 307 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, 651 DeviceMemory<U>* batch_mean, DeviceMemory<U>* batch_var,
|
D | rocm_dnn.cc | 1550 void* batch_mean, void* batch_var, in SetBatchNormForwardArgs() argument 1556 k_batchnorm_op_idx, &alpha, &beta, scale, offset, batch_mean, batch_var, in SetBatchNormForwardArgs() 3646 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, in DoBatchNormalizationForward() argument 3654 batch_var, saved_mean, saved_inv_var, is_training); in DoBatchNormalizationForward() 3666 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, in DoBatchNormalizationForward() argument 3674 batch_var, saved_mean, saved_inv_var, is_training); in DoBatchNormalizationForward() 3688 DeviceMemory<U>* batch_mean, DeviceMemory<U>* batch_var, in DoBatchNormalizationForwardImpl() argument 3707 exponential_average_factor, batch_mean->opaque(), batch_var->opaque(), in DoBatchNormalizationForwardImpl()
|
/external/tensorflow/tensorflow/compiler/xla/stream_executor/ |
D | stream.cc | 337 DeviceMemory<float> *batch_mean, DeviceMemory<float> *batch_var, in ThenBatchNormalizationForward() argument 347 activation_mode, y, batch_mean, batch_var, saved_mean, saved_inv_var, in ThenBatchNormalizationForward() 391 DeviceMemory<float> *batch_mean, DeviceMemory<float> *batch_var, in ThenBatchNormalizationForward() argument 401 activation_mode, y, batch_mean, batch_var, saved_mean, saved_inv_var, in ThenBatchNormalizationForward()
|
D | dnn.h | 1171 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, in DoBatchNormalizationForward() argument 1191 DeviceMemory<float>* batch_mean, DeviceMemory<float>* batch_var, in DoBatchNormalizationForward() argument
|
D | stream.h | 261 DeviceMemory<float> *batch_mean, DeviceMemory<float> *batch_var, 288 DeviceMemory<float> *batch_mean, DeviceMemory<float> *batch_var,
|
/external/tensorflow/tensorflow/compiler/xla/mlir_hlo/tests/Dialect/lhlo/ |
D | ops.mlir | 876 %batch_var: memref<8xf32>) -> () { 877 …"lmhlo.batch_norm_training"(%arg0, %arg1, %arg2, %output, %batch_mean, %batch_var) {epsilon = 1.00…
|
/external/tensorflow/tensorflow/python/kernel_tests/variables/ |
D | resource_variable_ops_test.py | 1093 batch_var = resource_variable_ops.ResourceVariable(array_ops.ones((2, 2))) 1100 batch_var.batch_scatter_update(batch_slices1)
|
/external/tensorflow/tensorflow/compiler/xla/client/ |
D | xla_builder.h | 924 XlaOp batch_var, XlaOp grad_output, float epsilon, 1492 XlaOp batch_var, XlaOp grad_output, float epsilon, 2693 XlaOp batch_var, XlaOp grad_output, float epsilon,
|
D | xla_builder.cc | 2843 XlaOp batch_var, XlaOp grad_output, in BatchNormGrad() argument 2851 TF_ASSIGN_OR_RETURN(const Shape* batch_var_shape, GetShapePtr(batch_var)); in BatchNormGrad() 2864 {operand, scale, batch_mean, batch_var, grad_output}); in BatchNormGrad() 4989 const XlaOp batch_mean, const XlaOp batch_var, in BatchNormGrad() argument 4992 return operand.builder()->BatchNormGrad(operand, scale, batch_mean, batch_var, in BatchNormGrad()
|
/external/tensorflow/tensorflow/compiler/xla/mlir_hlo/stablehlo/dialect/ |
D | StablehloOps.td | 1272 AllElementTypesMatch<["operand", "batch_mean", "batch_var"]>, 1273 AllShapesMatch<["scale", "offset", "batch_mean", "batch_var"]>]> { 1292 1DTensorOf<[HLO_Float]>:$batch_var);
|
/external/tensorflow/tensorflow/compiler/xla/mlir_hlo/include/mlir-hlo/Dialect/mhlo/IR/ |
D | hlo_ops.td | 1384 AllElementTypesMatch<["operand", "batch_mean", "batch_var"]>, 1385 AllShapesMatch<["scale", "offset", "batch_mean", "batch_var"]>]> { 1404 1DTensorOf<[HLO_Float]>:$batch_var);
|
/external/tensorflow/tensorflow/compiler/xla/mlir_hlo/include/mlir-hlo/Dialect/lhlo/IR/ |
D | lhlo_ops.td | 779 Arg<LHLO_Buffer, "", [MemWrite]>:$batch_var,
|
/external/tensorflow/tensorflow/compiler/xla/g3doc/ |
D | operation_semantics.md | 337 | `batch_var` | `XlaOp` | 1 dimensional array (\\(\sigma^2\\)) | 339 The `batch_mean` and `batch_var` are moments calculated across the batch and
|