/external/tensorflow/tensorflow/core/util/autotune_maps/ |
D | autotune_serialize_test.cc | 62 /*data_format=*/TensorFormat::FORMAT_NCHW, in TEST() 75 /*data_format=*/TensorFormat::FORMAT_NCHW, in TEST() 93 /*data_format=*/TensorFormat::FORMAT_NCHW, in TEST() 142 /*data_format=*/TensorFormat::FORMAT_NCHW, in TEST()
|
/external/tensorflow/tensorflow/core/util/ |
D | tensor_format.h | 42 FORMAT_NCHW = 1, enumerator 119 case FORMAT_NCHW: in GetTensorSpatialDims() 148 case FORMAT_NCHW: in GetTensorDimsFromSpatialDims() 176 case FORMAT_NCHW: in GetTensorBatchDimIndex() 201 case FORMAT_NCHW: in GetTensorFeatureDimIndex() 235 case FORMAT_NCHW: in GetTensorSpatialDimIndex() 328 } else if (format == FORMAT_NCHW || format == FORMAT_NCHW_VECT_C) { in GetTensorDimIndex()
|
D | tensor_format.cc | 44 case FORMAT_NCHW: in ToString() 82 *format = FORMAT_NCHW; in FormatFromString()
|
D | tensor_format_test.cc | 29 EnumStringPair(FORMAT_NHWC), EnumStringPair(FORMAT_NCHW), 119 (format == FORMAT_NCHW || in GetTensorDimMap()
|
D | mkl_util.h | 165 FORMAT_NCHW = 1, enumerator 202 } else if (format == MklTensorFormat::FORMAT_NCHW) { 553 TensorFormat data_format = FORMAT_NCHW; in IsTensorInNCHWFormat() 1048 if (format == MklTensorFormat::FORMAT_NCHW) return memory::format_tag::nchw; in MklTensorFormatToMklDnnDataFormat() 1063 if (format == FORMAT_NCHW) return MklTensorFormat::FORMAT_NCDHW; in TFDataFormatToMklDnn3DDataFormat() 1075 if (format == FORMAT_NCHW) return MklTensorFormat::FORMAT_NCHW; in TFDataFormatToMklDnnDataFormat() 1089 if (format == MklTensorFormat::FORMAT_NCHW || in MklDnnDataFormatToTFDataFormat() 1091 return FORMAT_NCHW; in MklDnnDataFormatToTFDataFormat()
|
D | mkl_util_test.cc | 34 MklTensorFormat::FORMAT_NCHW); in TEST()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | spacetodepth_op_gpu.cu.cc | 174 struct SpaceToDepthOpFunctor<GPUDevice, T, FORMAT_NCHW> { 239 template struct functor::SpaceToDepthOpFunctor<GPUDevice, float, FORMAT_NCHW>; 244 FORMAT_NCHW>; 249 template struct functor::SpaceToDepthOpFunctor<GPUDevice, uint8, FORMAT_NCHW>; 253 template struct functor::SpaceToDepthOpFunctor<GPUDevice, int32, FORMAT_NCHW>;
|
D | depthtospace_op_gpu.cu.cc | 177 struct DepthToSpaceOpFunctor<GPUDevice, T, FORMAT_NCHW> { 242 template struct functor::DepthToSpaceOpFunctor<GPUDevice, float, FORMAT_NCHW>; 247 FORMAT_NCHW>; 252 template struct functor::DepthToSpaceOpFunctor<GPUDevice, int32, FORMAT_NCHW>;
|
D | bias_op.cc | 62 } else if (data_format == FORMAT_NCHW) { in GetBiasValueDims() 114 if (data_format_ == FORMAT_NCHW) { in Compute() 135 if (data_format_ == FORMAT_NCHW && input.shape().dims() > 2) { in Compute() 187 if (data_format_ == FORMAT_NCHW) { in Compute() 202 if (data_format_ == FORMAT_NCHW) { in Compute() 395 data_format_ = FORMAT_NCHW; in BiasGradOp() 413 if (data_format_ == FORMAT_NCHW) { in ComputeWithReduceSum()
|
D | depthtospace_op.cc | 120 functor::DepthToSpaceOpFunctor<Device, int32, FORMAT_NCHW> functor; in Compute() 124 } else if (data_format_ == FORMAT_NCHW) { in Compute() 125 functor::DepthToSpaceOpFunctor<Device, T, FORMAT_NCHW> functor; in Compute()
|
D | fused_batch_norm_op.cu.cc | 129 static_assert(tensor_format == FORMAT_NHWC || tensor_format == FORMAT_NCHW, 376 } else if (tensor_format == FORMAT_NCHW) { in operator ()() 381 LAUNCH(FORMAT_NCHW, false, Activation::kIdentity, c, inner); in operator ()() 383 LAUNCH(FORMAT_NCHW, false, Activation::kRelu, c, inner); in operator ()() 385 LAUNCH(FORMAT_NCHW, true, Activation::kIdentity, c, inner); in operator ()() 387 LAUNCH(FORMAT_NCHW, true, Activation::kRelu, c, inner); in operator ()()
|
D | spacetodepth_op.cc | 135 functor::SpaceToDepthOpFunctor<Device, int32, FORMAT_NCHW> functor; in Compute() 138 } else if (data_format_ == FORMAT_NCHW) { in Compute() 140 functor::SpaceToDepthOpFunctor<Device, RT, FORMAT_NCHW> functor; in Compute()
|
D | cudnn_pooling_gpu.cc | 52 ShapeFromFormat(FORMAT_NCHW, tensor_in.shape(), in Compute() 66 ShapeFromFormat(FORMAT_NCHW, out_shape, data_format), in Compute() 147 ShapeFromFormat(FORMAT_NCHW, tensor_in_shape, data_format); in Compute() 158 ShapeFromFormat(FORMAT_NCHW, out_backprop.shape(), data_format); in Compute()
|
D | pooling_ops_common.cc | 250 ShapeFromFormat(FORMAT_NCHW, tensor_in.shape(), in Compute() 263 ShapeFromFormat(FORMAT_NCHW, tensor_out_shape, in Compute() 278 case FORMAT_NCHW: in Compute() 484 ShapeFromFormat(FORMAT_NCHW, tensor_in_shape, data_format); in Compute() 495 ShapeFromFormat(FORMAT_NCHW, out_backprop.shape(), data_format); in Compute() 530 transformed_input_data_format = FORMAT_NCHW; in Compute() 569 case FORMAT_NCHW: in Compute()
|
D | conv_ops_3d.cc | 357 : FORMAT_NCHW; in launch() 363 if (data_format == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in launch() 366 FORMAT_NCHW, in_batch, {{in_planes, in_rows, in_cols}}, in_depth); in launch() 440 compute_data_format == FORMAT_NCHW ? FORMAT_OIHW : FORMAT_OHWI; in launch() 468 ShapeFromFormat(FORMAT_NCHW, in_batch, in launch() 523 if (data_format == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in launch()
|
D | fused_batch_norm_op.cc | 130 if (tensor_format == FORMAT_NCHW) { in operator ()() 220 if (tensor_format == FORMAT_NCHW) { in operator ()() 271 if (tensor_format == FORMAT_NCHW) { in operator ()() 336 if (tensor_format == FORMAT_NCHW) { in operator ()() 372 if (tensor_format == FORMAT_NCHW) { in operator ()() 519 if (tensor_format == FORMAT_NCHW) { in operator ()() 808 : FORMAT_NCHW; in operator ()() 871 } else if (tensor_format == FORMAT_NHWC && compute_format == FORMAT_NCHW) { in operator ()() 984 if (tensor_format == FORMAT_NHWC && compute_format == FORMAT_NCHW) { in operator ()() 1027 : FORMAT_NCHW; in operator ()() [all …]
|
D | lrn_op.cc | 243 ShapeFromFormat(FORMAT_NCHW, in.shape(), FORMAT_NHWC), in launch() 253 ShapeFromFormat(FORMAT_NCHW, output->shape(), FORMAT_NHWC), in launch() 536 ShapeFromFormat(FORMAT_NCHW, in_grads.shape(), in launch() 546 ShapeFromFormat(FORMAT_NCHW, in_image.shape(), in launch() 556 ShapeFromFormat(FORMAT_NCHW, out_image.shape(), in launch() 567 ShapeFromFormat(FORMAT_NCHW, output->shape(), FORMAT_NHWC), in launch()
|
D | conv_grad_input_ops.cc | 226 : FORMAT_NCHW; in operator ()() 299 if (compute_data_format == FORMAT_NCHW) { in operator ()() 310 if (data_format == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in operator ()() 427 if (data_format == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in operator ()()
|
D | conv_ops_fused_impl.h | 331 params.data_format == FORMAT_NHWC || params.data_format == FORMAT_NCHW, 441 ShapeFromFormat(FORMAT_NCHW, in_batch, in_rows, in_cols, in_depths); 480 compute_in_nhwc ? FORMAT_NHWC : FORMAT_NCHW; 559 ShapeFromFormat(FORMAT_NCHW, out_batch, out_rows,
|
D | conv_grad_ops_3d.cc | 1381 : FORMAT_NCHW; in Compute() 1436 compute_data_format == FORMAT_NCHW ? FORMAT_OIHW : FORMAT_OHWI; in Compute() 1456 if (data_format_ == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in Compute() 1564 if (data_format_ == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in Compute() 1777 : FORMAT_NCHW; in Compute() 1834 compute_data_format == FORMAT_NCHW ? FORMAT_OIHW : FORMAT_OHWI; in Compute() 1848 if (data_format_ == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in Compute() 1867 if (data_format_ == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in Compute()
|
D | conv_ops.cc | 903 : FORMAT_NCHW; in operator ()() 988 if (data_format == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in operator ()() 992 ShapeFromFormat(FORMAT_NCHW, in_batch, in_rows, in_cols, in_depths); in operator ()() 1080 if (compute_data_format == FORMAT_NCHW) { in operator ()() 1151 if (data_format == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in operator ()()
|
D | conv_grad_filter_ops.cc | 845 : FORMAT_NCHW; in operator ()() 909 if (data_format == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in operator ()() 930 if (data_format == FORMAT_NHWC && compute_data_format == FORMAT_NCHW) { in operator ()() 1007 compute_data_format == FORMAT_NCHW ? FORMAT_OIHW : FORMAT_OHWI; in operator ()()
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | depthtospace_op.cc | 47 data_format = FORMAT_NCHW; in Compile() 53 OP_REQUIRES(ctx, data_format == FORMAT_NCHW || data_format == FORMAT_NHWC, in Compile() 171 DCHECK(data_format == FORMAT_NCHW && data_format_ == FORMAT_NCHW_VECT_C); in Compile()
|
D | spacetodepth_op.cc | 47 data_format = FORMAT_NCHW; in Compile() 53 OP_REQUIRES(ctx, data_format == FORMAT_NCHW || data_format == FORMAT_NHWC, in Compile() 175 DCHECK(data_format == FORMAT_NCHW && data_format_ == FORMAT_NCHW_VECT_C); in Compile()
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | mkl_tfconversion_pass.cc | 183 data_format == ToString(FORMAT_NCHW))) { in InsertConversionNodeOnEdge() 269 data_format == ToString(FORMAT_NCHW))) { in InsertInputConversionNode()
|