/external/tensorflow/tensorflow/python/framework/ |
D | common_shapes.py | 214 stride_b, stride_d, stride_r, stride_c = op.get_attr("strides") 216 stride_b, stride_r, stride_c, stride_d = op.get_attr("strides") 226 filter_cols, stride_r, stride_c, 272 stride_b, stride_r, stride_c, stride_d = op.get_attr("strides") 276 if stride_r != stride_c: 335 stride_b, stride_r, stride_c, stride_d = op.get_attr("strides") 339 if stride_r != stride_c: 390 stride_b, stride_d, stride_r, stride_c = op.get_attr("strides") 393 stride_b, stride_r, stride_c, stride_d = op.get_attr("strides") 413 ksize_c, stride_r, stride_c, [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | conv_ops_using_gemm.cc | 445 const int64 stride_c = GetTensorDim(strides_, data_format_, 'C'); in Conv2DUsingGemmOp() local 447 context, stride_n == 1 && stride_c == 1, in Conv2DUsingGemmOp()
|
D | depthwise_conv_op.cc | 280 const int64 stride_c = GetTensorDim(strides_, data_format_, 'C'); in DepthwiseConv2dNativeOp() local 287 context, (stride_n == 1 && stride_c == 1), in DepthwiseConv2dNativeOp()
|
D | depthwise_conv_grad_op.cc | 554 const int64 stride_c = GetTensorDim(strides_, data_format_, 'C'); in DepthwiseConv2dNativeBackpropInputOp() local 561 context, (stride_n == 1 && stride_c == 1), in DepthwiseConv2dNativeBackpropInputOp() 1028 const int64 stride_c = GetTensorDim(strides_, data_format_, 'C'); in DepthwiseConv2dNativeBackpropFilterOp() local 1035 context, (stride_n == 1 && stride_c == 1), in DepthwiseConv2dNativeBackpropFilterOp()
|
D | mkl_conv_ops.h | 561 int stride_c = GetTensorDim(strides_, data_format_, 'C'); in MklConvBackpropCommonOp() local 565 context, (stride_n == 1 && stride_c == 1), in MklConvBackpropCommonOp()
|
D | conv_ops_fused_image_transform.cc | 637 const int64 stride_c = GetTensorDim(strides_, FORMAT_NHWC, 'C'); in FusedResizeConv2DUsingGemmOp() local 639 context, stride_n == 1 && stride_c == 1, in FusedResizeConv2DUsingGemmOp()
|
D | mkl_conv_ops.cc | 432 const int64 stride_c = GetTensorDim(strides_, data_format_, 'C'); in MklConvOp() local 434 context, stride_n == 1 && stride_c == 1, in MklConvOp() 876 const int64 stride_c = GetTensorDim(strides_, data_format_, 'C'); in MklConvOp() local 878 context, stride_n == 1 && stride_c == 1, in MklConvOp()
|
D | conv_ops.cc | 305 const int64 stride_c = GetTensorDim(strides, data_format, 'C'); in InitConv2DParameters() local 309 stride_n == 1 && stride_c == 1, in InitConv2DParameters()
|
D | pooling_ops_3d.cc | 677 const int32 stride_c = GetTensorDim(stride_, data_format_, 'C'); in MaxPooling3dGradGradOp() local 678 OP_REQUIRES(context, ksize_c == 1 && stride_c == 1, in MaxPooling3dGradGradOp()
|
D | conv_grad_filter_ops.cc | 454 int stride_c = GetTensorDim(strides_, data_format_, 'C'); in Conv2DSlowBackpropFilterOp() local 458 context, (stride_n == 1 && stride_c == 1), in Conv2DSlowBackpropFilterOp()
|
D | conv_grad_input_ops.cc | 591 int stride_c = GetTensorDim(strides_, data_format_, 'C'); in Conv2DSlowBackpropInputOp() local 595 context, (stride_n == 1 && stride_c == 1), in Conv2DSlowBackpropInputOp()
|
/external/tensorflow/tensorflow/stream_executor/cuda/ |
D | cuda_blas.cc | 2363 int64 stride_c, int batch_count) { in DoBlasGemmStridedBatched() argument 2382 ldb, stride_b, &beta, GpuMemoryMutable(c), CUDA_R_16F, ldc, stride_c, in DoBlasGemmStridedBatched() 2400 reinterpret_cast<__half *>(GpuMemoryMutable(c) + batch * stride_c); in DoBlasGemmStridedBatched() 2419 float beta, DeviceMemory<float> *c, int ldc, int64 stride_c, in DoBlasGemmStridedBatched() argument 2425 GpuMemoryMutable(c), ldc, stride_c, batch_count); in DoBlasGemmStridedBatched() 2432 double beta, DeviceMemory<double> *c, int ldc, int64 stride_c, in DoBlasGemmStridedBatched() argument 2438 GpuMemoryMutable(c), ldc, stride_c, batch_count); in DoBlasGemmStridedBatched() 2447 int64 stride_c, int batch_count) { in DoBlasGemmStridedBatched() argument 2453 GpuComplex(GpuMemoryMutable(c)), ldc, stride_c, batch_count); in DoBlasGemmStridedBatched() 2462 int64 stride_c, int batch_count) { in DoBlasGemmStridedBatched() argument [all …]
|
/external/tensorflow/tensorflow/stream_executor/ |
D | blas.h | 1131 int64 stride_c, int batch_count) = 0; 1136 float beta, DeviceMemory<float> *c, int ldc, int64 stride_c, 1142 double beta, DeviceMemory<double> *c, int ldc, int64 stride_c, 1150 int64 stride_c, int batch_count) = 0; 1157 int64 stride_c, int batch_count) = 0; 2033 DeviceMemory<Eigen::half> *c, int ldc, int64 stride_c, int batch_count); \ 2039 int64 stride_c, int batch_count); \ 2045 DeviceMemory<double> *c, int ldc, int64 stride_c, int batch_count); \ 2052 int64 stride_c, int batch_count); \ 2059 int ldc, int64 stride_c, int batch_count); \
|
D | stream.cc | 4557 float beta, DeviceMemory<Eigen::half> *c, int ldc, int64 stride_c, in ThenBlasGemmStridedBatched() argument 4562 PARAM(stride_c), PARAM(batch_count)); in ThenBlasGemmStridedBatched() 4571 c, ldc, stride_c, batch_count); in ThenBlasGemmStridedBatched() 4578 float beta, DeviceMemory<float> *c, int ldc, int64 stride_c, in ThenBlasGemmStridedBatched() argument 4583 PARAM(stride_c), PARAM(batch_count)); in ThenBlasGemmStridedBatched() 4592 c, ldc, stride_c, batch_count); in ThenBlasGemmStridedBatched() 4599 double beta, DeviceMemory<double> *c, int ldc, int64 stride_c, in ThenBlasGemmStridedBatched() argument 4604 PARAM(stride_c), PARAM(batch_count)); in ThenBlasGemmStridedBatched() 4613 c, ldc, stride_c, batch_count); in ThenBlasGemmStridedBatched() 4622 int64 stride_c, int batch_count) { in ThenBlasGemmStridedBatched() argument [all …]
|
D | stream.h | 1455 int64 stride_c, int batch_count); 1460 float beta, DeviceMemory<float> *c, int ldc, int64 stride_c, 1466 double beta, DeviceMemory<double> *c, int ldc, int64 stride_c, 1474 int64 stride_c, int batch_count); 1481 int64 stride_c, int batch_count);
|
/external/tensorflow/tensorflow/stream_executor/rocm/ |
D | rocm_blas.cc | 2277 int64 stride_c, int batch_count) { in DoBlasGemmStridedBatched() argument 2287 float beta, DeviceMemory<float> *c, int ldc, int64 stride_c, in DoBlasGemmStridedBatched() argument 2298 double beta, DeviceMemory<double> *c, int ldc, int64 stride_c, in DoBlasGemmStridedBatched() argument 2311 int64 stride_c, int batch_count) { in DoBlasGemmStridedBatched() argument 2323 int64 stride_c, int batch_count) { in DoBlasGemmStridedBatched() argument
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | optimized_ops.h | 2247 int stride_c = n; in Conv() local 2250 stride_a, b, stride_b, 0.0f, c, stride_c); in Conv()
|