/external/pytorch/aten/src/ATen/native/ |
D | ReflectionPad.cpp | 133 int64_t output_h = input_h + pad_top + pad_bottom; in TORCH_META_FUNC() local 152 TORCH_CHECK(output_w >= 1 || output_h >=1 || output_d >= 1, in TORCH_META_FUNC() 155 " Calculated output D: ", output_d, " H: ", output_h, " W: ", output_w); in TORCH_META_FUNC() 158 …set_output_raw_strided(0, {input.size(0), nplane, output_d, output_h, output_w}, {}, input.options… in TORCH_META_FUNC() 160 set_output_raw_strided(0, {nplane, output_d, output_h, output_w}, {}, input.options()); in TORCH_META_FUNC() 194 int64_t output_h = input_h + pad_top + pad_bottom; in TORCH_META_FUNC() local 199 TORCH_CHECK(output_h == grad_output.size(dim_h), "grad_output height unexpected." in TORCH_META_FUNC() 200 " Expected: ", output_h, ", Got: ", grad_output.size(dim_h)); in TORCH_META_FUNC() 238 int64_t output_h = input_h + pad_t + pad_b; in reflection_pad2d_out_template() local 251 TORCH_CHECK(output_w >= 1 || output_h >= 1, in reflection_pad2d_out_template() [all …]
|
/external/rust/android-crates-io/crates/grpcio-sys/grpc/third_party/upb/bazel/ |
D | amalgamate.py | 42 self.output_h = open(h_out, "w") 51 self.output_h.write("// Ruby is still using proto3 enum semantics for proto2\n") 52 self.output_h.write("#define UPB_DISABLE_PROTO2_ENUM_CHECKING\n") 54 self.output_h.write("/* Amalgamated source file */\n") 58 self._process_file(port_def, self.output_h) 64 self._process_file(port_undef, self.output_h) 105 self._process_file(h_file, self.output_h)
|
/external/armnn/third-party/stb/ |
D | stb_image_resize.h | 231 … unsigned char *output_pixels, int output_w, int output_h, int output_stride_in_bytes, 235 … float *output_pixels, int output_w, int output_h, int output_stride_in_bytes, 256 … unsigned char *output_pixels, int output_w, int output_h, int output_stride_in_bytes, 270 … unsigned char *output_pixels, int output_w, int output_h, int output_stride_in_bytes, 310 … unsigned char *output_pixels, int output_w, int output_h, int output_stride_in_bytes, 316 … stbir_uint16 *output_pixels , int output_w, int output_h, int output_stride_in_bytes, 322 … float *output_pixels , int output_w, int output_h, int output_stride_in_bytes, 353 … void *output_pixels, int output_w, int output_h, int output_stride_in_bytes, 361 … void *output_pixels, int output_w, int output_h, int output_stride_in_bytes, 371 … void *output_pixels, int output_w, int output_h, int output_stride_in_bytes, [all …]
|
/external/executorch/backends/xnnpack/_passes/ |
D | convert_to_upsample_bilinear2d.py | 27 output_h = output_shape[-2] 30 isinstance(output_h, int) and isinstance(output_w, int), 41 args=(input_node, [output_h, output_w], align_corners, None),
|
/external/tensorflow/tensorflow/python/ops/ |
D | cudnn_rnn_grad.py | 32 output_h=op.outputs[1], 57 output_h=op.outputs[1], 85 output_h=op.outputs[1],
|
/external/pytorch/aten/src/ATen/native/cuda/ |
D | ReflectionPad.cu | 310 int output_h = input_h + pad_t + pad_b; in reflection_pad2d_out_template() local 313 TORCH_CHECK(output_w >= 1 || output_h >= 1, in reflection_pad2d_out_template() 315 "output H: ", output_h, " W: ", output_w); in reflection_pad2d_out_template() 318 output.resize_({nplane, output_h, output_w}); in reflection_pad2d_out_template() 320 output.resize_({nbatch, nplane, output_h, output_w}); in reflection_pad2d_out_template() 328 int64_t output_plane_size = output_h * output_w; in reflection_pad2d_out_template() 390 int output_h = input_h + pad_t + pad_b; in reflection_pad2d_backward_out_template() local 395 TORCH_CHECK(output_h == grad_output_.size(dim_h), "grad_output height " in reflection_pad2d_backward_out_template() 396 "unexpected. Expected: ", output_h, ", Got: ", grad_output_.size(dim_h)); in reflection_pad2d_backward_out_template() 400 int64_t output_plane_size = output_h * output_w; in reflection_pad2d_backward_out_template()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | cudnn_rnn_ops.cc | 797 Tensor* output, Tensor* output_h, Tensor* output_c, in DoForward() argument 825 auto output_h_data = AsDeviceMemory<T>(output_h); in DoForward() 876 const Tensor* output, const Tensor* output_h, const Tensor* output_c, in DoBackward() argument 907 auto output_h_data = AsDeviceMemory<T>(output_h); in DoBackward() 1526 Tensor* output_h = nullptr; in ComputeAndReturnAlgorithm() local 1529 &output_h, &output_c)); in ComputeAndReturnAlgorithm() 1546 input_h, input_c, params, output, output_h, in ComputeAndReturnAlgorithm() 1560 input_c, params, is_training_, output, output_h, output_c, in ComputeAndReturnAlgorithm() 1573 Tensor* output, Tensor* output_h, in MaybeAutotune() argument 1589 Tensor** output, Tensor** output_h, in AllocateOutputs() argument [all …]
|
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v1/ |
D | CudnnRNNV3.pbtxt | 28 name: "output_h" 153 name: "output_h" 285 name: "output_h"
|
D | CudnnRNNBackpropV3.pbtxt | 28 name: "output_h" 174 name: "output_h" 327 name: "output_h"
|
D | CudnnRNN.pbtxt | 24 name: "output_h"
|
D | CudnnRNNV2.pbtxt | 24 name: "output_h"
|
D | CudnnRNNBackprop.pbtxt | 24 name: "output_h"
|
D | CudnnRNNBackpropV2.pbtxt | 24 name: "output_h"
|
/external/angle/src/common/ |
D | gen_uniform_type_table.py | 321 output_h = template_h.format( 325 out_file.write(output_h)
|
/external/executorch/backends/apple/mps/runtime/operations/ |
D | PadOps.mm | 72 int64_t output_h = padding_dim > 1 ? input_h + pad_t + pad_b : 0; 76 output_w >= 1 || output_h >= padding_dim - 1, 78 "output H: %lld, W: %lld", input_h, input_w, output_h, output_w);
|
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v2/ |
D | CudnnRNNV3.pbtxt | 28 name: "output_h" 153 name: "output_h"
|
D | CudnnRNNBackpropV3.pbtxt | 28 name: "output_h" 174 name: "output_h"
|
D | CudnnRNN.pbtxt | 24 name: "output_h"
|
D | CudnnRNNV2.pbtxt | 24 name: "output_h"
|
D | CudnnRNNBackprop.pbtxt | 24 name: "output_h"
|
/external/tensorflow/tensorflow/core/api_def/base_api/ |
D | api_def_CudnnRNNBackprop.pbtxt | 28 output_h: The same shape has input_h. 31 output_h_backprop: A 3-D tensor with the same shape as output_h in the forward
|
D | api_def_CudnnRNNBackpropV2.pbtxt | 31 output_h: The same shape has input_h. 34 output_h_backprop: A 3-D tensor with the same shape as output_h in the forward
|
D | api_def_CudnnRNNBackpropV3.pbtxt | 35 output_h: The same shape has input_h. 38 output_h_backprop: A 3-D tensor with the same shape as output_h in the forward
|
/external/pytorch/aten/src/ATen/native/mps/operations/ |
D | Pad.mm | 96 int64_t output_h = padding_dim > 1 ? input_h + pad_t + pad_b : 0; 103 TORCH_CHECK(output_w >= 1 || output_h >= padding_dim - 1, 110 output_h, 170 outputSizes.insert(outputSizes.begin(), output_h); 196 TORCH_CHECK(output_h == grad_output_.size(dim_h), 198 output_h,
|
/external/openthread/third_party/mbedtls/repo/tests/scripts/ |
D | generate_psa_wrappers.py | 251 if options.output_h: 252 generator.write_h_file(options.output_h)
|