/external/tensorflow/tensorflow/core/kernels/image/ |
D | resize_bilinear_op_gpu.cu.cc | 58 const float in_x = (static_cast<float>(x) + 0.5f) * width_scale - 0.5f; in ResizeBilinearKernel_faster() local 59 const int left_x_index = in_x > 0.0 ? floorf(in_x) : 0; in ResizeBilinearKernel_faster() 61 (in_x < in_width - 1) ? ceilf(in_x) : in_width - 1; in ResizeBilinearKernel_faster() 62 const float x_lerp = in_x - floorf(in_x); in ResizeBilinearKernel_faster() 136 const float in_x = (static_cast<float>(x) + 0.5f) * width_scale - 0.5f; in ResizeBilinearKernel() local 137 const int left_x_index = in_x > 0.0 ? floorf(in_x) : 0; in ResizeBilinearKernel() 139 (in_x < in_width - 1) ? ceilf(in_x) : in_width - 1; in ResizeBilinearKernel() 140 const float x_lerp = in_x - floorf(in_x); in ResizeBilinearKernel() 261 int in_x = in_x_start; in ResizeBilinearDeterministicGradKernel() local 262 while (out_x < out_x_center + 1 && in_x < resized_width) { in ResizeBilinearDeterministicGradKernel() [all …]
|
D | crop_and_resize_op_gpu.cu.cc | 90 const float in_x = in CropAndResizeKernel() local 94 if (in_x < 0 || in_x > image_width_minus_one) { in CropAndResizeKernel() 104 const int left_x_index = floorf(in_x); in CropAndResizeKernel() 105 const int right_x_index = ceilf(in_x); in CropAndResizeKernel() 106 const float x_lerp = in_x - left_x_index; in CropAndResizeKernel() 132 const int closest_x_index = roundf(in_x); in CropAndResizeKernel() 183 const float in_x = (crop_width > 1) in CropAndResizeBackpropImageKernel() local 186 if (in_x < 0 || in_x > image_width - 1) { in CropAndResizeBackpropImageKernel() 195 const int left_x_index = floorf(in_x); in CropAndResizeBackpropImageKernel() 196 const int right_x_index = ceilf(in_x); in CropAndResizeBackpropImageKernel() [all …]
|
D | crop_and_resize_op.cc | 277 const float in_x = (crop_width > 1) in operator ()() local 280 if (in_x < 0 || in_x > image_width - 1) { in operator ()() 286 const int left_x_index = floorf(in_x); in operator ()() 287 const int right_x_index = ceilf(in_x); in operator ()() 288 const float x_lerp = in_x - left_x_index; in operator ()() 307 const float in_x = (crop_width > 1) in operator ()() local 310 if (in_x < 0 || in_x > image_width - 1) { in operator ()() 316 const int closest_x_index = roundf(in_x); in operator ()() 508 const float in_x = (crop_width > 1) in operator ()() local 511 if (in_x < 0 || in_x > image_width - 1) { in operator ()() [all …]
|
D | resize_nearest_neighbor_op_gpu.cu.cc | 55 const int in_x = in ResizeNearestNeighborNHWC() local 60 const int idx = (in_y * in_width + in_x) * channels + c; in ResizeNearestNeighborNHWC() 85 const int in_x = in LegacyResizeNearestNeighborNHWC() local 89 const int idx = (in_y * in_width + in_x) * channels + c; in LegacyResizeNearestNeighborNHWC() 104 int in_x = n % in_width; in ResizeNearestNeighborBackwardNHWC() local 117 floorf((static_cast<float>(in_x) + 0.5f) * width_scale)), in ResizeNearestNeighborBackwardNHWC() 135 int in_x = n % in_width; in LegacyResizeNearestNeighborBackwardNHWC() local 146 min((align_corners) ? static_cast<int>(roundf(in_x * width_scale)) in LegacyResizeNearestNeighborBackwardNHWC() 147 : static_cast<int>(floorf(in_x * width_scale)), in LegacyResizeNearestNeighborBackwardNHWC()
|
D | resize_nearest_neighbor_op.cc | 161 Eigen::Index in_x = std::min( in operator ()() local 167 in_x = std::max(static_cast<Eigen::Index>(0), in_x); in operator ()() 169 std::copy_n(&input(b, in_y, in_x, 0), channels, &output(b, y, x, 0)); in operator ()() 187 Eigen::Index in_x = std::min( in operator ()() local 193 in_x = std::max(static_cast<Eigen::Index>(0), in_x); in operator ()() 195 std::copy_n(&input(bs, in_y, in_x, 0), channels, &output(bs, y, x, 0)); in operator ()()
|
D | resize_area_op.cc | 165 const float in_x = x * st.width_scale; in Compute() local 169 int64_t v = std::floor(in_x); in Compute() 173 v < in_x ? (v + 1 > in_x1 ? st.width_scale : v + 1 - in_x) in Compute() 180 v < in_x ? (v + 1 > in_x1 ? st.width_scale : v + 1 - in_x) in Compute()
|
/external/ComputeLibrary/src/core/helpers/ |
D | ScaleHelpers.h | 132 float in_x = (x + 0.5f) * wr - 0.5f; in pixel_area_c1u8_clamp() local 136 int x_from = std::floor(x * wr - 0.5f - in_x); in pixel_area_c1u8_clamp() 138 int x_to = std::ceil((x + 1) * wr - 0.5f - in_x); in pixel_area_c1u8_clamp() 142 in_x = std::max(-1.f, std::min(in_x, static_cast<float>(width))); in pixel_area_c1u8_clamp() 146 x_from = ((in_x + x_from) < -1) ? -1 : x_from; in pixel_area_c1u8_clamp() 148 x_to = ((in_x + x_to) > width) ? (width - in_x) : x_to; in pixel_area_c1u8_clamp() 152 const int xi = std::floor(in_x); in pixel_area_c1u8_clamp()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | transpose_conv.h | 71 for (int in_x = 0; in_x < input_width; ++in_x) { in TransposeConv() local 74 const int out_x_origin = (in_x * stride_width) - pad_width; in TransposeConv() 87 input_shape, batch, in_y, in_x, in_channel)]; in TransposeConv() 162 for (int in_x = 0; in_x < input_width; ++in_x) { in TransposeConv() local 165 const int out_x_origin = (in_x * stride_width) - pad_width; in TransposeConv() 178 input_shape, batch, in_y, in_x, in_channel)]; in TransposeConv()
|
D | conv.h | 75 const int in_x = in_x_origin + dilation_width_factor * filter_x; in Conv() local 79 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in Conv() 88 input_data[Offset(input_shape, batch, in_y, in_x, in Conv() 164 const int in_x = in_x_origin + dilation_width_factor * filter_x; in Conv() local 168 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in Conv() 178 input_data[Offset(input_shape, batch, in_y, in_x, in Conv() 252 const int in_x = in_x_origin + dilation_width_factor * filter_x; in HybridConvPerChannel() local 257 if ((in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in HybridConvPerChannel() 260 input_shape, batch, in_y, in_x, in HybridConvPerChannel()
|
D | pooling.h | 64 const int in_x = in_x_origin + filter_x; in AveragePool() local 67 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in AveragePool() 122 const int in_x = in_x_origin + filter_x; in AveragePool() local 125 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in AveragePool() 177 const int in_x = in_x_origin + filter_x; in L2Pool() local 180 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in L2Pool() 230 const int in_x = in_x_origin + filter_x; in MaxPool() local 234 input_data[Offset(input_shape, batch, in_y, in_x, channel)]); in MaxPool() 284 const int in_x = in_x_origin + filter_x; in MaxPool() local 288 input_data[Offset(input_shape, batch, in_y, in_x, channel)]); in MaxPool()
|
D | conv3d_transpose.h | 66 for (int in_x = 0; in_x < input_width; ++in_x) { in Conv3DTranspose() local 70 const int out_x_origin = (in_x * stride_width) - pad_width; in Conv3DTranspose() 90 input_shape, batch, in_d, in_y, in_x, in_channel)]; in Conv3DTranspose()
|
D | depthwiseconv_uint8.h | 162 const int in_x = in Run() local 168 if ((in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in Run() 171 input_data[Offset(input_shape, b, in_y, in_x, ic)]; in Run() 249 const int in_x = in RunPerChannel() local 255 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in RunPerChannel() 259 input_shape, batch, in_y, in_x, in_channel)]; in RunPerChannel()
|
D | depthwiseconv_float.h | 67 const int in_x = in_x_origin + dilation_width_factor * filter_x; in DepthwiseConv() local 72 if ((in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in DepthwiseConv() 75 input_data[Offset(input_shape, b, in_y, in_x, ic)]; in DepthwiseConv()
|
D | conv3d.h | 70 const int in_x = in Conv3D() local 75 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in Conv3D() 86 input_shape, batch, in_d, in_y, in_x, in_channel)]; in Conv3D()
|
/external/tensorflow/tensorflow/core/kernels/linalg/ |
D | banded_triangular_solve_op.cc | 62 static void Run(const Tensor& in_x, const Tensor& in_y, bool lower, in Run() 68 int num_bands = in_x.dim_size(1); in Run() 69 int matrix_size = in_x.dim_size(2); in Run() 74 auto matrix = ConstTensorSliceToEigenMatrix(in_x, x_batch_index); in Run() 168 static void Launch(OpKernelContext* context, const Tensor& in_x, in Launch() 174 in_x.dim_size(1) * in_x.dim_size(2) * in_y.dim_size(2); in Launch() 183 auto matrix = ConstMatrixMap(in_x.flat<Scalar>().data(), in_x.dim_size(1), in Launch() 184 in_x.dim_size(2)); in Launch() 189 min_abs_pivot = matrix.row(in_x.dim_size(1) - 1).cwiseAbs().minCoeff(); in Launch() 196 [&in_x, &in_y, adjoint, lower, &bcast, out](int64_t start, in Launch() [all …]
|
D | matrix_triangular_solve_op_impl.h | 79 static void Run(const Tensor& in_x, const Tensor& in_y, bool lower, in Run() 88 auto matrix = ConstTensorSliceToEigenMatrix(in_x, x_batch_index); in Run() 115 static void Launch(OpKernelContext* context, const Tensor& in_x, 121 in_x.dim_size(1) * in_x.dim_size(1) * in_y.dim_size(2) / 2; 131 [&in_x, &in_y, adjoint, lower, &bcast, out](int start, int limit) { 133 in_x, in_y, lower, adjoint, bcast, out, start, limit); 258 static void Launch(OpKernelContext* context, const Tensor& in_x, 263 const uint64 m = in_x.dim_size(1); 354 auto* a_base_ptr = in_x.template flat<Scalar>().data();
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/integer_ops/ |
D | depthwise_conv.h | 74 const int in_x = in_x_origin + dilation_width_factor * filter_x; in DepthwiseConvPerChannel() local 79 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in DepthwiseConvPerChannel() 83 input_shape, batch, in_y, in_x, in_channel)]; in DepthwiseConvPerChannel() 171 const int in_x = in_x_origin + dilation_width_factor * filter_x; in DepthwiseConvPerChannel() local 176 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in DepthwiseConvPerChannel() 180 input_shape, batch, in_y, in_x, in_channel)]; in DepthwiseConvPerChannel() 255 const int in_x = in_x_origin + dilation_width_factor * filter_x; in DepthwiseConvHybridPerChannel() local 260 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in DepthwiseConvHybridPerChannel() 264 input_shape, batch, in_y, in_x, in_channel)]; in DepthwiseConvHybridPerChannel()
|
D | transpose_conv.h | 70 for (int in_x = 0; in_x < input_width; ++in_x) { in TransposeConv() local 73 const int out_x_origin = (in_x * stride_width) - pad_width; in TransposeConv() 86 input_shape, batch, in_y, in_x, in_channel)]; in TransposeConv() 168 for (int in_x = 0; in_x < input_width; ++in_x) { in TransposeConv() local 171 const int out_x_origin = (in_x * stride_width) - pad_width; in TransposeConv() 184 input_shape, batch, in_y, in_x, in_channel)]; in TransposeConv()
|
D | conv.h | 81 const int in_x = in_x_origin + dilation_width_factor * filter_x; in ConvPerChannel() local 85 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in ConvPerChannel() 95 input_data[Offset(input_shape, batch, in_y, in_x, in ConvPerChannel() 192 const int in_x = in_x_origin + dilation_width_factor * filter_x; in ConvPerChannel() local 196 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in ConvPerChannel() 206 input_data[Offset(input_shape, batch, in_y, in_x, in ConvPerChannel()
|
D | pooling.h | 64 const int in_x = in_x_origin + filter_x; in AveragePool() local 67 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in AveragePool() 126 const int in_x = in_x_origin + filter_x; in MaxPool() local 130 input_data[Offset(input_shape, batch, in_y, in_x, channel)]); in MaxPool() 182 const int in_x = in_x_origin + filter_x; in AveragePool() local 185 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in AveragePool() 244 const int in_x = in_x_origin + filter_x; in MaxPool() local 248 input_data[Offset(input_shape, batch, in_y, in_x, channel)]); in MaxPool()
|
/external/tensorflow/tensorflow/lite/kernels/perception/ |
D | dense_image_warp.cc | 51 for (int in_x = 0; in_x < width; ++in_x) { in DenseImageWarp() local 53 in_y - flow_data[Offset(flow_shape, batch, in_y, in_x, 0)]; in DenseImageWarp() 55 in_x - flow_data[Offset(flow_shape, batch, in_y, in_x, 1)]; in DenseImageWarp() 82 output_data[Offset(input_shape, batch, in_y, in_x, c)] = interp; in DenseImageWarp()
|
/external/eigen/unsupported/test/ |
D | cxx11_tensor_gpu.cu | 721 Tensor<Scalar, 1> in_x(6); in test_gpu_zeta() local 727 in_x(0) = Scalar(1); in test_gpu_zeta() 728 in_x(1) = Scalar(1.5); in test_gpu_zeta() 729 in_x(2) = Scalar(4); in test_gpu_zeta() 730 in_x(3) = Scalar(-10.5); in test_gpu_zeta() 731 in_x(4) = Scalar(10000.5); in test_gpu_zeta() 732 in_x(5) = Scalar(3); in test_gpu_zeta() 748 std::size_t bytes = in_x.size() * sizeof(Scalar); in test_gpu_zeta() 757 gpuMemcpy(d_in_x, in_x.data(), bytes, gpuMemcpyHostToDevice); in test_gpu_zeta() 789 Tensor<Scalar, 1> in_x(7); in test_gpu_polygamma() local [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | matmul_op_impl.h | 86 static void Run(const OpKernelContext* context, const Tensor& in_x, in Run() 91 auto Tx = in_x.tensor<Scalar, 3>(); in Run() 131 static void Run(const OpKernelContext* context, const Tensor& in_x, 140 auto Tx = in_x.flat_inner_dims<Scalar, 2>(); 145 auto Tx = in_x.tensor<Scalar, 3>(); 187 static void Run(const Tensor& in_x, const Tensor& in_y, bool adj_x, 196 auto x = ConstTensorSliceToEigenMatrix(in_x, x_batch_index); 236 static void Launch(OpKernelContext* context, const Tensor& in_x, 246 in_x.dim_size(1) * in_x.dim_size(2) * out->dim_size(2); 248 std::min(in_x.dim_size(1), in_x.dim_size(2)), out->dim_size(2)); [all …]
|
/external/deqp/external/openglcts/modules/glesext/tessellation_shader/ |
D | esextcTessellationShaderUtils.hpp | 52 _ivec4(int in_x, int in_y, int in_z, int in_w) in _ivec4() 54 x = in_x; in _ivec4() 92 _vec2(float in_x, float in_y) in _vec2() 94 x = in_x; in _vec2() 134 _vec4(float in_x, float in_y, float in_z, float in_w) in _vec4() 136 x = in_x; in _vec4()
|
/external/ComputeLibrary/src/core/NEON/kernels/ |
D | NESpaceToDepthLayerKernel.cpp | 126 … const size_t in_x = id.x() * _block_shape + (channel_id / channel_size) % _block_shape; in run() local 129 Coordinates input_coords{ in_x, in_y, z, batch_id }; in run() 145 … const size_t in_x = id.y() * _block_shape + (channel_id / channel_size) % _block_shape; in run() local 148 Coordinates input_coords{ z, in_x, in_y, batch_id }; in run()
|