Home
last modified time | relevance | path

Searched refs:in_x (Results 1 – 25 of 78) sorted by relevance

1234

/external/tensorflow/tensorflow/core/kernels/image/
Dresize_bilinear_op_gpu.cu.cc58 const float in_x = (static_cast<float>(x) + 0.5f) * width_scale - 0.5f; in ResizeBilinearKernel_faster() local
59 const int left_x_index = in_x > 0.0 ? floorf(in_x) : 0; in ResizeBilinearKernel_faster()
61 (in_x < in_width - 1) ? ceilf(in_x) : in_width - 1; in ResizeBilinearKernel_faster()
62 const float x_lerp = in_x - floorf(in_x); in ResizeBilinearKernel_faster()
136 const float in_x = (static_cast<float>(x) + 0.5f) * width_scale - 0.5f; in ResizeBilinearKernel() local
137 const int left_x_index = in_x > 0.0 ? floorf(in_x) : 0; in ResizeBilinearKernel()
139 (in_x < in_width - 1) ? ceilf(in_x) : in_width - 1; in ResizeBilinearKernel()
140 const float x_lerp = in_x - floorf(in_x); in ResizeBilinearKernel()
261 int in_x = in_x_start; in ResizeBilinearDeterministicGradKernel() local
262 while (out_x < out_x_center + 1 && in_x < resized_width) { in ResizeBilinearDeterministicGradKernel()
[all …]
Dcrop_and_resize_op_gpu.cu.cc90 const float in_x = in CropAndResizeKernel() local
94 if (in_x < 0 || in_x > image_width_minus_one) { in CropAndResizeKernel()
104 const int left_x_index = floorf(in_x); in CropAndResizeKernel()
105 const int right_x_index = ceilf(in_x); in CropAndResizeKernel()
106 const float x_lerp = in_x - left_x_index; in CropAndResizeKernel()
132 const int closest_x_index = roundf(in_x); in CropAndResizeKernel()
183 const float in_x = (crop_width > 1) in CropAndResizeBackpropImageKernel() local
186 if (in_x < 0 || in_x > image_width - 1) { in CropAndResizeBackpropImageKernel()
195 const int left_x_index = floorf(in_x); in CropAndResizeBackpropImageKernel()
196 const int right_x_index = ceilf(in_x); in CropAndResizeBackpropImageKernel()
[all …]
Dcrop_and_resize_op.cc277 const float in_x = (crop_width > 1) in operator ()() local
280 if (in_x < 0 || in_x > image_width - 1) { in operator ()()
286 const int left_x_index = floorf(in_x); in operator ()()
287 const int right_x_index = ceilf(in_x); in operator ()()
288 const float x_lerp = in_x - left_x_index; in operator ()()
307 const float in_x = (crop_width > 1) in operator ()() local
310 if (in_x < 0 || in_x > image_width - 1) { in operator ()()
316 const int closest_x_index = roundf(in_x); in operator ()()
508 const float in_x = (crop_width > 1) in operator ()() local
511 if (in_x < 0 || in_x > image_width - 1) { in operator ()()
[all …]
Dresize_nearest_neighbor_op_gpu.cu.cc55 const int in_x = in ResizeNearestNeighborNHWC() local
60 const int idx = (in_y * in_width + in_x) * channels + c; in ResizeNearestNeighborNHWC()
85 const int in_x = in LegacyResizeNearestNeighborNHWC() local
89 const int idx = (in_y * in_width + in_x) * channels + c; in LegacyResizeNearestNeighborNHWC()
104 int in_x = n % in_width; in ResizeNearestNeighborBackwardNHWC() local
117 floorf((static_cast<float>(in_x) + 0.5f) * width_scale)), in ResizeNearestNeighborBackwardNHWC()
135 int in_x = n % in_width; in LegacyResizeNearestNeighborBackwardNHWC() local
146 min((align_corners) ? static_cast<int>(roundf(in_x * width_scale)) in LegacyResizeNearestNeighborBackwardNHWC()
147 : static_cast<int>(floorf(in_x * width_scale)), in LegacyResizeNearestNeighborBackwardNHWC()
Dresize_nearest_neighbor_op.cc161 Eigen::Index in_x = std::min( in operator ()() local
167 in_x = std::max(static_cast<Eigen::Index>(0), in_x); in operator ()()
169 std::copy_n(&input(b, in_y, in_x, 0), channels, &output(b, y, x, 0)); in operator ()()
187 Eigen::Index in_x = std::min( in operator ()() local
193 in_x = std::max(static_cast<Eigen::Index>(0), in_x); in operator ()()
195 std::copy_n(&input(bs, in_y, in_x, 0), channels, &output(bs, y, x, 0)); in operator ()()
Dresize_area_op.cc165 const float in_x = x * st.width_scale; in Compute() local
169 int64_t v = std::floor(in_x); in Compute()
173 v < in_x ? (v + 1 > in_x1 ? st.width_scale : v + 1 - in_x) in Compute()
180 v < in_x ? (v + 1 > in_x1 ? st.width_scale : v + 1 - in_x) in Compute()
/external/ComputeLibrary/src/core/helpers/
DScaleHelpers.h132 float in_x = (x + 0.5f) * wr - 0.5f; in pixel_area_c1u8_clamp() local
136 int x_from = std::floor(x * wr - 0.5f - in_x); in pixel_area_c1u8_clamp()
138 int x_to = std::ceil((x + 1) * wr - 0.5f - in_x); in pixel_area_c1u8_clamp()
142 in_x = std::max(-1.f, std::min(in_x, static_cast<float>(width))); in pixel_area_c1u8_clamp()
146 x_from = ((in_x + x_from) < -1) ? -1 : x_from; in pixel_area_c1u8_clamp()
148 x_to = ((in_x + x_to) > width) ? (width - in_x) : x_to; in pixel_area_c1u8_clamp()
152 const int xi = std::floor(in_x); in pixel_area_c1u8_clamp()
/external/tensorflow/tensorflow/lite/kernels/internal/reference/
Dtranspose_conv.h71 for (int in_x = 0; in_x < input_width; ++in_x) { in TransposeConv() local
74 const int out_x_origin = (in_x * stride_width) - pad_width; in TransposeConv()
87 input_shape, batch, in_y, in_x, in_channel)]; in TransposeConv()
162 for (int in_x = 0; in_x < input_width; ++in_x) { in TransposeConv() local
165 const int out_x_origin = (in_x * stride_width) - pad_width; in TransposeConv()
178 input_shape, batch, in_y, in_x, in_channel)]; in TransposeConv()
Dconv.h75 const int in_x = in_x_origin + dilation_width_factor * filter_x; in Conv() local
79 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in Conv()
88 input_data[Offset(input_shape, batch, in_y, in_x, in Conv()
164 const int in_x = in_x_origin + dilation_width_factor * filter_x; in Conv() local
168 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in Conv()
178 input_data[Offset(input_shape, batch, in_y, in_x, in Conv()
252 const int in_x = in_x_origin + dilation_width_factor * filter_x; in HybridConvPerChannel() local
257 if ((in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in HybridConvPerChannel()
260 input_shape, batch, in_y, in_x, in HybridConvPerChannel()
Dpooling.h64 const int in_x = in_x_origin + filter_x; in AveragePool() local
67 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in AveragePool()
122 const int in_x = in_x_origin + filter_x; in AveragePool() local
125 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in AveragePool()
177 const int in_x = in_x_origin + filter_x; in L2Pool() local
180 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in L2Pool()
230 const int in_x = in_x_origin + filter_x; in MaxPool() local
234 input_data[Offset(input_shape, batch, in_y, in_x, channel)]); in MaxPool()
284 const int in_x = in_x_origin + filter_x; in MaxPool() local
288 input_data[Offset(input_shape, batch, in_y, in_x, channel)]); in MaxPool()
Dconv3d_transpose.h66 for (int in_x = 0; in_x < input_width; ++in_x) { in Conv3DTranspose() local
70 const int out_x_origin = (in_x * stride_width) - pad_width; in Conv3DTranspose()
90 input_shape, batch, in_d, in_y, in_x, in_channel)]; in Conv3DTranspose()
Ddepthwiseconv_uint8.h162 const int in_x = in Run() local
168 if ((in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in Run()
171 input_data[Offset(input_shape, b, in_y, in_x, ic)]; in Run()
249 const int in_x = in RunPerChannel() local
255 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in RunPerChannel()
259 input_shape, batch, in_y, in_x, in_channel)]; in RunPerChannel()
Ddepthwiseconv_float.h67 const int in_x = in_x_origin + dilation_width_factor * filter_x; in DepthwiseConv() local
72 if ((in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in DepthwiseConv()
75 input_data[Offset(input_shape, b, in_y, in_x, ic)]; in DepthwiseConv()
Dconv3d.h70 const int in_x = in Conv3D() local
75 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in Conv3D()
86 input_shape, batch, in_d, in_y, in_x, in_channel)]; in Conv3D()
/external/tensorflow/tensorflow/core/kernels/linalg/
Dbanded_triangular_solve_op.cc62 static void Run(const Tensor& in_x, const Tensor& in_y, bool lower, in Run()
68 int num_bands = in_x.dim_size(1); in Run()
69 int matrix_size = in_x.dim_size(2); in Run()
74 auto matrix = ConstTensorSliceToEigenMatrix(in_x, x_batch_index); in Run()
168 static void Launch(OpKernelContext* context, const Tensor& in_x, in Launch()
174 in_x.dim_size(1) * in_x.dim_size(2) * in_y.dim_size(2); in Launch()
183 auto matrix = ConstMatrixMap(in_x.flat<Scalar>().data(), in_x.dim_size(1), in Launch()
184 in_x.dim_size(2)); in Launch()
189 min_abs_pivot = matrix.row(in_x.dim_size(1) - 1).cwiseAbs().minCoeff(); in Launch()
196 [&in_x, &in_y, adjoint, lower, &bcast, out](int64_t start, in Launch()
[all …]
Dmatrix_triangular_solve_op_impl.h79 static void Run(const Tensor& in_x, const Tensor& in_y, bool lower, in Run()
88 auto matrix = ConstTensorSliceToEigenMatrix(in_x, x_batch_index); in Run()
115 static void Launch(OpKernelContext* context, const Tensor& in_x,
121 in_x.dim_size(1) * in_x.dim_size(1) * in_y.dim_size(2) / 2;
131 [&in_x, &in_y, adjoint, lower, &bcast, out](int start, int limit) {
133 in_x, in_y, lower, adjoint, bcast, out, start, limit);
258 static void Launch(OpKernelContext* context, const Tensor& in_x,
263 const uint64 m = in_x.dim_size(1);
354 auto* a_base_ptr = in_x.template flat<Scalar>().data();
/external/tensorflow/tensorflow/lite/kernels/internal/reference/integer_ops/
Ddepthwise_conv.h74 const int in_x = in_x_origin + dilation_width_factor * filter_x; in DepthwiseConvPerChannel() local
79 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in DepthwiseConvPerChannel()
83 input_shape, batch, in_y, in_x, in_channel)]; in DepthwiseConvPerChannel()
171 const int in_x = in_x_origin + dilation_width_factor * filter_x; in DepthwiseConvPerChannel() local
176 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in DepthwiseConvPerChannel()
180 input_shape, batch, in_y, in_x, in_channel)]; in DepthwiseConvPerChannel()
255 const int in_x = in_x_origin + dilation_width_factor * filter_x; in DepthwiseConvHybridPerChannel() local
260 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in DepthwiseConvHybridPerChannel()
264 input_shape, batch, in_y, in_x, in_channel)]; in DepthwiseConvHybridPerChannel()
Dtranspose_conv.h70 for (int in_x = 0; in_x < input_width; ++in_x) { in TransposeConv() local
73 const int out_x_origin = (in_x * stride_width) - pad_width; in TransposeConv()
86 input_shape, batch, in_y, in_x, in_channel)]; in TransposeConv()
168 for (int in_x = 0; in_x < input_width; ++in_x) { in TransposeConv() local
171 const int out_x_origin = (in_x * stride_width) - pad_width; in TransposeConv()
184 input_shape, batch, in_y, in_x, in_channel)]; in TransposeConv()
Dconv.h81 const int in_x = in_x_origin + dilation_width_factor * filter_x; in ConvPerChannel() local
85 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in ConvPerChannel()
95 input_data[Offset(input_shape, batch, in_y, in_x, in ConvPerChannel()
192 const int in_x = in_x_origin + dilation_width_factor * filter_x; in ConvPerChannel() local
196 (in_x >= 0) && (in_x < input_width) && (in_y >= 0) && in ConvPerChannel()
206 input_data[Offset(input_shape, batch, in_y, in_x, in ConvPerChannel()
Dpooling.h64 const int in_x = in_x_origin + filter_x; in AveragePool() local
67 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in AveragePool()
126 const int in_x = in_x_origin + filter_x; in MaxPool() local
130 input_data[Offset(input_shape, batch, in_y, in_x, channel)]); in MaxPool()
182 const int in_x = in_x_origin + filter_x; in AveragePool() local
185 input_data[Offset(input_shape, batch, in_y, in_x, channel)]; in AveragePool()
244 const int in_x = in_x_origin + filter_x; in MaxPool() local
248 input_data[Offset(input_shape, batch, in_y, in_x, channel)]); in MaxPool()
/external/tensorflow/tensorflow/lite/kernels/perception/
Ddense_image_warp.cc51 for (int in_x = 0; in_x < width; ++in_x) { in DenseImageWarp() local
53 in_y - flow_data[Offset(flow_shape, batch, in_y, in_x, 0)]; in DenseImageWarp()
55 in_x - flow_data[Offset(flow_shape, batch, in_y, in_x, 1)]; in DenseImageWarp()
82 output_data[Offset(input_shape, batch, in_y, in_x, c)] = interp; in DenseImageWarp()
/external/eigen/unsupported/test/
Dcxx11_tensor_gpu.cu721 Tensor<Scalar, 1> in_x(6); in test_gpu_zeta() local
727 in_x(0) = Scalar(1); in test_gpu_zeta()
728 in_x(1) = Scalar(1.5); in test_gpu_zeta()
729 in_x(2) = Scalar(4); in test_gpu_zeta()
730 in_x(3) = Scalar(-10.5); in test_gpu_zeta()
731 in_x(4) = Scalar(10000.5); in test_gpu_zeta()
732 in_x(5) = Scalar(3); in test_gpu_zeta()
748 std::size_t bytes = in_x.size() * sizeof(Scalar); in test_gpu_zeta()
757 gpuMemcpy(d_in_x, in_x.data(), bytes, gpuMemcpyHostToDevice); in test_gpu_zeta()
789 Tensor<Scalar, 1> in_x(7); in test_gpu_polygamma() local
[all …]
/external/tensorflow/tensorflow/core/kernels/
Dmatmul_op_impl.h86 static void Run(const OpKernelContext* context, const Tensor& in_x, in Run()
91 auto Tx = in_x.tensor<Scalar, 3>(); in Run()
131 static void Run(const OpKernelContext* context, const Tensor& in_x,
140 auto Tx = in_x.flat_inner_dims<Scalar, 2>();
145 auto Tx = in_x.tensor<Scalar, 3>();
187 static void Run(const Tensor& in_x, const Tensor& in_y, bool adj_x,
196 auto x = ConstTensorSliceToEigenMatrix(in_x, x_batch_index);
236 static void Launch(OpKernelContext* context, const Tensor& in_x,
246 in_x.dim_size(1) * in_x.dim_size(2) * out->dim_size(2);
248 std::min(in_x.dim_size(1), in_x.dim_size(2)), out->dim_size(2));
[all …]
/external/deqp/external/openglcts/modules/glesext/tessellation_shader/
DesextcTessellationShaderUtils.hpp52 _ivec4(int in_x, int in_y, int in_z, int in_w) in _ivec4()
54 x = in_x; in _ivec4()
92 _vec2(float in_x, float in_y) in _vec2()
94 x = in_x; in _vec2()
134 _vec4(float in_x, float in_y, float in_z, float in_w) in _vec4()
136 x = in_x; in _vec4()
/external/ComputeLibrary/src/core/NEON/kernels/
DNESpaceToDepthLayerKernel.cpp126 … const size_t in_x = id.x() * _block_shape + (channel_id / channel_size) % _block_shape; in run() local
129 Coordinates input_coords{ in_x, in_y, z, batch_id }; in run()
145 … const size_t in_x = id.y() * _block_shape + (channel_id / channel_size) % _block_shape; in run() local
148 Coordinates input_coords{ z, in_x, in_y, batch_id }; in run()

1234