/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/default/ |
D | convolution_selector.cc | 37 const Convolution2DAttributes& attr, const BHWC& dst_shape, in SelectConvolutionAdreno() argument 44 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionAdreno() 50 const Convolution2DAttributes& attr, const BHWC& dst_shape, in SelectConvolutionWinogradAdreno() argument 54 CreateConvPowerVRWino4x4To6x6(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionWinogradAdreno() 60 const BHWC& dst_shape, const GpuInfo& gpu_info, in SelectConvolutionDynamicWeightsAdreno() argument 64 gpu_info, op_def, attr, weights_shape, &dst_shape); in SelectConvolutionDynamicWeightsAdreno() 70 const Convolution2DAttributes& attr, const BHWC& dst_shape, in SelectConvolutionNVidia() argument 76 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionNVidia() 89 const Convolution2DAttributes& attr, const BHWC& dst_shape, in SelectConvolutionMali() argument 94 CreateConvBuffer1x1(gpu_info, op_def, attr, &dst_shape); in SelectConvolutionMali() [all …]
|
D | fully_connected_selector.cc | 31 BHWC dst_shape = BHWC(batch_size, 1, 1, attr.weights.shape.o); in SelectFullyConnectedGeneric() local 32 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedGeneric() 44 BHWC dst_shape = BHWC(batch_size, 1, 1, attr.weights.shape.o); in SelectFullyConnectedAdreno() local 45 ConvPowerVR conv = CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedAdreno() 73 BHWC dst_shape = BHWC(batch_size, 1, 1, attr.weights.shape.o); in SelectFullyConnectedMali() local 75 CreateConvPowerVR(gpu_info, op_def, attr, &dst_shape); in SelectFullyConnectedMali() 89 BHWC dst_shape = BHWC(batch_size, 1, 1, attr.weights.shape.o); in SelectFullyConnected() local 98 CreateConvolutionMetal(op_def, dst_shape, conv_attr, gpu_info); in SelectFullyConnected()
|
/external/tensorflow/tensorflow/python/ops/ragged/ |
D | ragged_tensor_shape.py | 516 def _broadcast_to_ragged_shape(rt_input, dst_shape, broadcast_inner_dimensions): argument 520 rt_input.row_splits.dtype != dst_shape.dim_size_dtype): 527 dst_shape = dst_shape.with_dim_size_dtype(dtypes.int64) 530 if rt_input.shape.ndims is None or dst_shape.rank is None: 532 if rt_input.shape.ndims > dst_shape.rank: 535 rt_input.ragged_rank >= dst_shape.num_partitioned_dimensions): 539 src_shape = src_shape.broadcast_to_rank(dst_shape.rank) 542 if dst_shape.rank > rt_input.shape.ndims: 543 if rt_input.shape.ndims < dst_shape.num_inner_dimensions + 1: 545 rt_input, array_ops.concat([[-1], dst_shape.inner_dim_sizes], axis=0)) [all …]
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/ |
D | conv_powervr.h | 112 const BHWC* dst_shape = nullptr); 115 const GpuInfo& gpu_info, const BHWC* dst_shape = nullptr); 118 const BHWC* dst_shape = nullptr); 122 const BHWDC* dst_shape = nullptr); 145 const BHWC* dst_shape); 150 const BHWC* dst_shape); 155 const BHWC* dst_shape); 159 const Convolution2DAttributes& attr, const BHWC* dst_shape); 164 const BHWDC* dst_shape); 169 const BHWC* dst_shape = nullptr); [all …]
|
D | conv_metal.cc | 620 int GetGroupsCount(const BHWC& dst_shape, const int3& wg_size, in GetGroupsCount() argument 622 const int dst_slices = DivideRoundUp(dst_shape.c, 4); in GetGroupsCount() 624 int grid_x = DivideRoundUp(dst_shape.w, block_size.x); in GetGroupsCount() 625 int grid_y = DivideRoundUp(dst_shape.h, block_size.y); in GetGroupsCount() 632 int GetGroupsCountForLinearWH(const BHWC& dst_shape, const int3& wg_size, in GetGroupsCountForLinearWH() argument 634 const int dst_slices = DivideRoundUp(dst_shape.c, 4); in GetGroupsCountForLinearWH() 636 int grid_x = DivideRoundUp(dst_shape.w, block_size.x); in GetGroupsCountForLinearWH() 637 int grid_y = DivideRoundUp(dst_shape.h, block_size.y); in GetGroupsCountForLinearWH() 644 int GetGroupsCountForLinearWHS(const BHWC& dst_shape, const int3& wg_size, in GetGroupsCountForLinearWHS() argument 646 const int dst_slices = DivideRoundUp(dst_shape.c, 4); in GetGroupsCountForLinearWHS() [all …]
|
D | conv_powervr.cc | 172 const GpuInfo& gpu_info, const BHWC* dst_shape) in ConvPowerVR() argument 178 conv_params_(GuessBestParams(gpu_info, definition, attr, dst_shape)) {} in ConvPowerVR() 183 const BHWC* dst_shape) in ConvPowerVR() argument 190 dst_shape)) {} in ConvPowerVR() 194 const GpuInfo& gpu_info, const BHWC* dst_shape) in ConvPowerVR() argument 200 conv_params_(GuessBestParams(gpu_info, definition, attr, dst_shape)) {} in ConvPowerVR() 219 const GpuInfo& gpu_info, const BHWDC* dst_shape) in ConvPowerVR() argument 227 conv_params_(GuessBestParams(gpu_info, definition, attr, dst_shape)) {} in ConvPowerVR() 1061 bool different_weights_for_height, const BHWC* dst_shape) { in GuessBestParams() argument 1092 if (dst_shape) { in GuessBestParams() [all …]
|
D | conv_metal.h | 83 const OperationDef& definition, const BHWC& dst_shape, 87 const OperationDef& definition, const BHWC& dst_shape, 96 const BHWC& dst_shape, 101 const OperationDef& definition, const BHWC& dst_shape,
|
D | conv_buffer_1x1.h | 84 const BHWC* dst_shape); 189 const BHWC* dst_shape = nullptr);
|
D | conv_buffer_1x1.cc | 405 const BHWC* dst_shape) { in CreateConvBuffer1x1DynamicWeights() argument 409 if (dst_shape) { in CreateConvBuffer1x1DynamicWeights() 411 GetBestParams(gpu_info, definition, *dst_shape, src_depth, dst_depth); in CreateConvBuffer1x1DynamicWeights()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/ |
D | winograd_util.cc | 114 OHWI dst_shape; in RearrangeWeightsToWinograd4x4To6x6Weights() local 115 dst_shape.o = src_weights.shape.o; in RearrangeWeightsToWinograd4x4To6x6Weights() 116 dst_shape.h = 6; in RearrangeWeightsToWinograd4x4To6x6Weights() 117 dst_shape.w = 6; in RearrangeWeightsToWinograd4x4To6x6Weights() 118 dst_shape.i = src_weights.shape.i; in RearrangeWeightsToWinograd4x4To6x6Weights() 119 dst_weights->shape = dst_shape; in RearrangeWeightsToWinograd4x4To6x6Weights() 120 dst_weights->data.resize(dst_shape.DimensionsProduct()); in RearrangeWeightsToWinograd4x4To6x6Weights() 144 const int f_index = dst_shape.LinearIndex({d, y, x, s}); in RearrangeWeightsToWinograd4x4To6x6Weights()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/transformations/ |
D | global_pooling_to_reduce_op.cc | 35 const BHWC& dst_shape) { in IsGlobalPooling() argument 36 return dst_shape.w == 1 && dst_shape.h == 1 && attr.kernel.w == src_shape.w && in IsGlobalPooling() 43 const BHWC& src_shape, const BHWC& dst_shape) { in IsGlobalAveragePooling() argument 46 IsGlobalPooling(attr, src_shape, dst_shape); in IsGlobalAveragePooling()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/selectors/ |
D | convolution_selector.h | 32 const Convolution2DAttributes& attr, const BHWC& dst_shape, 36 const Convolution2DAttributes& attr, const BHWC& dst_shape, 41 const BHWC& dst_shape, const GpuInfo& gpu_info, const OperationDef& op_def,
|
D | operation_selector.cc | 45 const BHWC& dst_shape) { in IsRecommendedForWinograd4x4To6x6() argument 46 const int tiles_x = DivideRoundUp(dst_shape.w, 4); in IsRecommendedForWinograd4x4To6x6() 47 const int tiles_y = DivideRoundUp(dst_shape.h, 4); in IsRecommendedForWinograd4x4To6x6() 196 auto dst_shape = outputs[0]->tensor.shape; in GPUOperationFromNode() local 197 if (dst_shape.b != 1) { in GPUOperationFromNode() 229 attr, weights_shape, dst_shape, gpu_info, conv_def, hints, in GPUOperationFromNode()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/cl/kernels/ |
D | cl_test.cc | 73 auto dst_shape = dst_sizes[i]; in ExecuteGPUOperation() local 74 if (dst_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation() 78 RETURN_IF_ERROR(CreateTensor(*creation_context.context, dst_shape, in ExecuteGPUOperation() 121 auto dst_shape = dst_sizes[i]; in ExecuteGPUOperation() local 122 if (dst_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation() 126 RETURN_IF_ERROR(CreateTensor(*creation_context.context, dst_shape, in ExecuteGPUOperation()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/metal/kernels/ |
D | test_util.cc | 83 auto dst_shape = dst_sizes[i]; in ExecuteGPUOperation() local 84 if (dst_shape.b != 1 && !op_def.IsBatchSupported()) { in ExecuteGPUOperation() 88 RETURN_IF_ERROR(CreateTensor(device_.device(), dst_shape, in ExecuteGPUOperation()
|
D | conv_test.mm | 243 auto dst_shape = CalculateOutputShape(src_shape, attr); 247 conv_shape.b = dst_shape.b; 250 conv_shape.c = dst_shape.c; 270 auto gpu_op0 = CreateConvolutionMetal(op_def, dst_shape, attr, env->GetGpuInfo()); 273 env->ExecuteGPUOperation(src_tensor, std::move(op0_ptr), dst_shape, &output0)); 299 BHWC output3_shape = dst_shape;
|
/external/tensorflow/tensorflow/lite/delegates/gpu/cl/ |
D | inference_context.cc | 117 const auto dst_shape = outputs[0]->tensor.shape; in IsGenericAdd() local 120 if (dst_shape.b != src_shape.b && src_shape.b == 1) { in IsGenericAdd() 123 if (dst_shape.h != src_shape.h && src_shape.h == 1) { in IsGenericAdd() 126 if (dst_shape.w != src_shape.w && src_shape.w == 1) { in IsGenericAdd() 129 if (dst_shape.c != src_shape.c && src_shape.c == 1) { in IsGenericAdd() 723 const auto dst_shape = BHWC(gpu_tensor.Batch(), gpu_tensor.Height(), in GetOutputTensor() local 726 result->shape = dst_shape; in GetOutputTensor() 727 result->data.resize(dst_shape.DimensionsProduct()); in GetOutputTensor()
|
/external/tensorflow/tensorflow/compiler/xla/pjrt/ |
D | utils.cc | 171 const Shape& sharded_shape, Shape* dst_shape) { in DetermineArgumentLayoutsFromCompileOptions() argument 173 dst_shape, [&](Shape* subshape, const ShapeIndex& idx) { in DetermineArgumentLayoutsFromCompileOptions()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/metal/ |
D | inference_context.cc | 97 const auto dst_shape = outputs[0]->tensor.shape; in IsGenericAdd() local 100 if (dst_shape.b != src_shape.b && src_shape.b == 1) { in IsGenericAdd() 103 if (dst_shape.h != src_shape.h && src_shape.h == 1) { in IsGenericAdd() 106 if (dst_shape.w != src_shape.w && src_shape.w == 1) { in IsGenericAdd() 109 if (dst_shape.c != src_shape.c && src_shape.c == 1) { in IsGenericAdd()
|
/external/tensorflow/tensorflow/compiler/mlir/hlo/lib/Dialect/mhlo/transforms/ |
D | legalize_to_linalg.cc | 738 ArrayRef<int64_t> dst_shape = in matchAndRewrite() local 744 dst_shape.size()); in matchAndRewrite() 750 while (curr_src_dim < src_shape.size() && curr_dst_dim < dst_shape.size()) { in matchAndRewrite() 751 int64_t dst_size = dst_shape[curr_dst_dim]; in matchAndRewrite() 763 if (curr_dst_dim == dst_shape.size() - 1 || in matchAndRewrite() 764 dst_shape[curr_dst_dim + 1] != 1) { in matchAndRewrite() 777 if (curr_src_dim != src_shape.size() || curr_dst_dim != dst_shape.size()) in matchAndRewrite()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | conv_grad_input_ops.cc | 286 TensorShape dst_shape = in operator ()() local 293 TF_RETURN_IF_ERROR(ctx->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in operator ()()
|
D | conv_ops_3d.cc | 435 TensorShape dst_shape = in launch() local 443 OP_REQUIRES_OK(ctx, ctx->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in launch()
|
D | conv_grad_ops_3d.cc | 1335 TensorShape dst_shape = in Compute() local 1344 context->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in Compute() 1831 TensorShape dst_shape = in Compute() local 1840 context->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in Compute()
|
D | conv_ops.cc | 912 TensorShape dst_shape = in operator ()() local 919 TF_RETURN_IF_ERROR(ctx->allocate_temp(DataTypeToEnum<T>::value, dst_shape, in operator ()()
|