Home
last modified time | relevance | path

Searched refs:dst_shape (Results 1 – 24 of 24) sorted by relevance

/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/default/
Dconvolution_selector.cc37 const Convolution2DAttributes& attr, const BHWC& dst_shape, in SelectConvolutionAdreno() argument
44 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionAdreno()
50 const Convolution2DAttributes& attr, const BHWC& dst_shape, in SelectConvolutionWinogradAdreno() argument
54 CreateConvPowerVRWino4x4To6x6(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionWinogradAdreno()
60 const BHWC& dst_shape, const GpuInfo& gpu_info, in SelectConvolutionDynamicWeightsAdreno() argument
64 gpu_info, op_def, attr, weights_shape, &dst_shape); in SelectConvolutionDynamicWeightsAdreno()
70 const Convolution2DAttributes& attr, const BHWC& dst_shape, in SelectConvolutionNVidia() argument
76 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionNVidia()
89 const Convolution2DAttributes& attr, const BHWC& dst_shape, in SelectConvolutionMali() argument
94 CreateConvBuffer1x1(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionMali()
[all …]
Dfully_connected_selector.cc31 BHWC dst_shape = BHWC(batch_size, 1, 1, attr.weights.shape.o); in SelectFullyConnectedGeneric() local
32 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedGeneric()
44 BHWC dst_shape = BHWC(batch_size, 1, 1, attr.weights.shape.o); in SelectFullyConnectedAdreno() local
45 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedAdreno()
73 BHWC dst_shape = BHWC(batch_size, 1, 1, attr.weights.shape.o); in SelectFullyConnectedMali() local
75 CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedMali()
89 BHWC dst_shape = BHWC(batch_size, 1, 1, attr.weights.shape.o); in SelectFullyConnected() local
98 CreateConvolutionMetal(op_def, dst_shape, conv_attr, gpu_info); in SelectFullyConnected()
/external/tensorflow/tensorflow/python/ops/ragged/
Dragged_tensor_shape.py516 def _broadcast_to_ragged_shape(rt_input, dst_shape, broadcast_inner_dimensions): argument
520 rt_input.row_splits.dtype != dst_shape.dim_size_dtype):
527 dst_shape = dst_shape.with_dim_size_dtype(dtypes.int64)
530 if rt_input.shape.ndims is None or dst_shape.rank is None:
532 if rt_input.shape.ndims > dst_shape.rank:
535 rt_input.ragged_rank >= dst_shape.num_partitioned_dimensions):
539 src_shape = src_shape.broadcast_to_rank(dst_shape.rank)
542 if dst_shape.rank > rt_input.shape.ndims:
543 if rt_input.shape.ndims < dst_shape.num_inner_dimensions + 1:
545 rt_input, array_ops.concat([[-1], dst_shape.inner_dim_sizes], axis=0))
[all …]
/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/
Dconv_powervr.h112 const BHWC* dst_shape = nullptr);
115 const GpuInfo& gpu_info, const BHWC* dst_shape = nullptr);
118 const BHWC* dst_shape = nullptr);
122 const BHWDC* dst_shape = nullptr);
145 const BHWC* dst_shape);
150 const BHWC* dst_shape);
155 const BHWC* dst_shape);
159 const Convolution2DAttributes& attr, const BHWC* dst_shape);
164 const BHWDC* dst_shape);
169 const BHWC* dst_shape = nullptr);
[all …]
Dconv_metal.cc620 int GetGroupsCount(const BHWC& dst_shape, const int3& wg_size, in GetGroupsCount() argument
622 const int dst_slices = DivideRoundUp(dst_shape.c, 4); in GetGroupsCount()
624 int grid_x = DivideRoundUp(dst_shape.w, block_size.x); in GetGroupsCount()
625 int grid_y = DivideRoundUp(dst_shape.h, block_size.y); in GetGroupsCount()
632 int GetGroupsCountForLinearWH(const BHWC& dst_shape, const int3& wg_size, in GetGroupsCountForLinearWH() argument
634 const int dst_slices = DivideRoundUp(dst_shape.c, 4); in GetGroupsCountForLinearWH()
636 int grid_x = DivideRoundUp(dst_shape.w, block_size.x); in GetGroupsCountForLinearWH()
637 int grid_y = DivideRoundUp(dst_shape.h, block_size.y); in GetGroupsCountForLinearWH()
644 int GetGroupsCountForLinearWHS(const BHWC& dst_shape, const int3& wg_size, in GetGroupsCountForLinearWHS() argument
646 const int dst_slices = DivideRoundUp(dst_shape.c, 4); in GetGroupsCountForLinearWHS()
[all …]
Dconv_powervr.cc172 const GpuInfo& gpu_info, const BHWC* dst_shape) in ConvPowerVR() argument
178 conv_params_(GuessBestParams(gpu_info, definition, attr, dst_shape)) {} in ConvPowerVR()
183 const BHWC* dst_shape) in ConvPowerVR() argument
190 dst_shape)) {} in ConvPowerVR()
194 const GpuInfo& gpu_info, const BHWC* dst_shape) in ConvPowerVR() argument
200 conv_params_(GuessBestParams(gpu_info, definition, attr, dst_shape)) {} in ConvPowerVR()
219 const GpuInfo& gpu_info, const BHWDC* dst_shape) in ConvPowerVR() argument
227 conv_params_(GuessBestParams(gpu_info, definition, attr, dst_shape)) {} in ConvPowerVR()
1061 bool different_weights_for_height, const BHWC* dst_shape) { in GuessBestParams() argument
1092 if (dst_shape) { in GuessBestParams()
[all …]
Dconv_metal.h83 const OperationDef& definition, const BHWC& dst_shape,
87 const OperationDef& definition, const BHWC& dst_shape,
96 const BHWC& dst_shape,
101 const OperationDef& definition, const BHWC& dst_shape,
Dconv_buffer_1x1.h84 const BHWC* dst_shape);
189 const BHWC* dst_shape = nullptr);
Dconv_buffer_1x1.cc405 const BHWC* dst_shape) { in CreateConvBuffer1x1DynamicWeights() argument
409 if (dst_shape) { in CreateConvBuffer1x1DynamicWeights()
411 GetBestParams(gpu_info, definition, *dst_shape, src_depth, dst_depth); in CreateConvBuffer1x1DynamicWeights()
/external/tensorflow/tensorflow/lite/delegates/gpu/common/
Dwinograd_util.cc114 OHWI dst_shape; in RearrangeWeightsToWinograd4x4To6x6Weights() local
115 dst_shape.o = src_weights.shape.o; in RearrangeWeightsToWinograd4x4To6x6Weights()
116 dst_shape.h = 6; in RearrangeWeightsToWinograd4x4To6x6Weights()
117 dst_shape.w = 6; in RearrangeWeightsToWinograd4x4To6x6Weights()
118 dst_shape.i = src_weights.shape.i; in RearrangeWeightsToWinograd4x4To6x6Weights()
119 dst_weights->shape = dst_shape; in RearrangeWeightsToWinograd4x4To6x6Weights()
120 dst_weights->data.resize(dst_shape.DimensionsProduct()); in RearrangeWeightsToWinograd4x4To6x6Weights()
144 const int f_index = dst_shape.LinearIndex({d, y, x, s}); in RearrangeWeightsToWinograd4x4To6x6Weights()
/external/tensorflow/tensorflow/lite/delegates/gpu/common/transformations/
Dglobal_pooling_to_reduce_op.cc35 const BHWC& dst_shape) { in IsGlobalPooling() argument
36 return dst_shape.w == 1 && dst_shape.h == 1 && attr.kernel.w == src_shape.w && in IsGlobalPooling()
43 const BHWC& src_shape, const BHWC& dst_shape) { in IsGlobalAveragePooling() argument
46 IsGlobalPooling(attr, src_shape, dst_shape); in IsGlobalAveragePooling()
/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/
Dconvolution_selector.h32 const Convolution2DAttributes& attr, const BHWC& dst_shape,
36 const Convolution2DAttributes& attr, const BHWC& dst_shape,
41 const BHWC& dst_shape, const GpuInfo& gpu_info, const OperationDef& op_def,
Doperation_selector.cc45 const BHWC& dst_shape) { in IsRecommendedForWinograd4x4To6x6() argument
46 const int tiles_x = DivideRoundUp(dst_shape.w, 4); in IsRecommendedForWinograd4x4To6x6()
47 const int tiles_y = DivideRoundUp(dst_shape.h, 4); in IsRecommendedForWinograd4x4To6x6()
196 auto dst_shape = outputs[0]->tensor.shape; in GPUOperationFromNode() local
197 if (dst_shape.b != 1) { in GPUOperationFromNode()
229 attr, weights_shape, dst_shape, gpu_info, conv_def, hints, in GPUOperationFromNode()
/external/tensorflow/tensorflow/lite/delegates/gpu/cl/kernels/
Dcl_test.cc73 auto dst_shape = dst_sizes[i]; in ExecuteGPUOperation() local
74 if (dst_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation()
78 RETURN_IF_ERROR(CreateTensor(*creation_context.context, dst_shape, in ExecuteGPUOperation()
121 auto dst_shape = dst_sizes[i]; in ExecuteGPUOperation() local
122 if (dst_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation()
126 RETURN_IF_ERROR(CreateTensor(*creation_context.context, dst_shape, in ExecuteGPUOperation()
/external/tensorflow/tensorflow/lite/delegates/gpu/metal/kernels/
Dtest_util.cc83 auto dst_shape = dst_sizes[i]; in ExecuteGPUOperation() local
84 if (dst_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation()
88 RETURN_IF_ERROR(CreateTensor(device_.device(), dst_shape, in ExecuteGPUOperation()
Dconv_test.mm243 auto dst_shape = CalculateOutputShape(src_shape, attr);
247 conv_shape.b = dst_shape.b;
250 conv_shape.c = dst_shape.c;
270 auto gpu_op0 = CreateConvolutionMetal(op_def, dst_shape, attr, env->GetGpuInfo());
273 env->ExecuteGPUOperation(src_tensor, std::move(op0_ptr), dst_shape, &output0));
299 BHWC output3_shape = dst_shape;
/external/tensorflow/tensorflow/lite/delegates/gpu/cl/
Dinference_context.cc117 const auto dst_shape = outputs[0]->tensor.shape; in IsGenericAdd() local
120 if (dst_shape.b != src_shape.b && src_shape.b == 1) { in IsGenericAdd()
123 if (dst_shape.h != src_shape.h && src_shape.h == 1) { in IsGenericAdd()
126 if (dst_shape.w != src_shape.w && src_shape.w == 1) { in IsGenericAdd()
129 if (dst_shape.c != src_shape.c && src_shape.c == 1) { in IsGenericAdd()
723 const auto dst_shape = BHWC(gpu_tensor.Batch(), gpu_tensor.Height(), in GetOutputTensor() local
726 result->shape = dst_shape; in GetOutputTensor()
727 result->data.resize(dst_shape.DimensionsProduct()); in GetOutputTensor()
/external/tensorflow/tensorflow/compiler/xla/pjrt/
Dutils.cc171 const Shape& sharded_shape, Shape* dst_shape) { in DetermineArgumentLayoutsFromCompileOptions() argument
173 dst_shape, [&](Shape* subshape, const ShapeIndex& idx) { in DetermineArgumentLayoutsFromCompileOptions()
/external/tensorflow/tensorflow/lite/delegates/gpu/metal/
Dinference_context.cc97 const auto dst_shape = outputs[0]->tensor.shape; in IsGenericAdd() local
100 if (dst_shape.b != src_shape.b && src_shape.b == 1) { in IsGenericAdd()
103 if (dst_shape.h != src_shape.h && src_shape.h == 1) { in IsGenericAdd()
106 if (dst_shape.w != src_shape.w && src_shape.w == 1) { in IsGenericAdd()
109 if (dst_shape.c != src_shape.c && src_shape.c == 1) { in IsGenericAdd()
/external/tensorflow/tensorflow/compiler/mlir/hlo/lib/Dialect/mhlo/transforms/
Dlegalize_to_linalg.cc738 ArrayRef<int64_t> dst_shape = in matchAndRewrite() local
744 dst_shape.size()); in matchAndRewrite()
750 while (curr_src_dim < src_shape.size() && curr_dst_dim < dst_shape.size()) { in matchAndRewrite()
751 int64_t dst_size = dst_shape[curr_dst_dim]; in matchAndRewrite()
763 if (curr_dst_dim == dst_shape.size() - 1 || in matchAndRewrite()
764 dst_shape[curr_dst_dim + 1] != 1) { in matchAndRewrite()
777 if (curr_src_dim != src_shape.size() || curr_dst_dim != dst_shape.size()) in matchAndRewrite()
/external/tensorflow/tensorflow/core/kernels/
Dconv_grad_input_ops.cc286 TensorShape dst_shape = in operator ()() local
293 TF_RETURN_IF_ERROR(ctx->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in operator ()()
Dconv_ops_3d.cc435 TensorShape dst_shape = in launch() local
443 OP_REQUIRES_OK(ctx, ctx->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in launch()
Dconv_grad_ops_3d.cc1335 TensorShape dst_shape = in Compute() local
1344 context->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in Compute()
1831 TensorShape dst_shape = in Compute() local
1840 context->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in Compute()
Dconv_ops.cc912 TensorShape dst_shape = in operator ()() local
919 TF_RETURN_IF_ERROR(ctx->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in operator ()()