/external/XNNPACK/src/operators/ |
D | constant-pad-nd.c | 116 if (num_dims > XNN_MAX_TENSOR_DIMS) { in setup_constant_pad_nd() 120 xnn_operator_type_to_string(constant_pad_op->type), num_dims, XNN_MAX_TENSOR_DIMS); in setup_constant_pad_nd() 134 size_t normalized_pre_paddings[XNN_MAX_TENSOR_DIMS]; in setup_constant_pad_nd() 135 size_t normalized_input_shape[XNN_MAX_TENSOR_DIMS]; in setup_constant_pad_nd() 136 size_t normalized_output_shape[XNN_MAX_TENSOR_DIMS]; in setup_constant_pad_nd() 137 for (size_t i = 0; i < XNN_MAX_TENSOR_DIMS; i++) { in setup_constant_pad_nd() 151 normalized_pre_paddings[XNN_MAX_TENSOR_DIMS - 1 - num_squeezed_dims] = pre_padding; in setup_constant_pad_nd() 152 normalized_input_shape[XNN_MAX_TENSOR_DIMS - 1 - num_squeezed_dims] = input_dim; in setup_constant_pad_nd() 153 …normalized_output_shape[XNN_MAX_TENSOR_DIMS - 1 - num_squeezed_dims] = pre_padding + input_dim + p… in setup_constant_pad_nd() 163 normalized_input_shape[XNN_MAX_TENSOR_DIMS - num_squeezed_dims] *= input_dim; in setup_constant_pad_nd() [all …]
|
D | binary-elementwise-nd.c | 871 if (max(num_input1_dims, num_input2_dims) > XNN_MAX_TENSOR_DIMS) { in setup_binary_elementwise_nd() 875 …ype_to_string(binary_elementwise_op->type), num_input1_dims, num_input2_dims, XNN_MAX_TENSOR_DIMS); in setup_binary_elementwise_nd() 880 size_t compressed_input1_shape[XNN_MAX_TENSOR_DIMS]; in setup_binary_elementwise_nd() 881 size_t compressed_input2_shape[XNN_MAX_TENSOR_DIMS]; in setup_binary_elementwise_nd() 882 size_t compressed_output_shape[XNN_MAX_TENSOR_DIMS]; in setup_binary_elementwise_nd() 883 for (size_t i = 0; i < XNN_MAX_TENSOR_DIMS; i++) { in setup_binary_elementwise_nd() 996 …binary_elementwise_op->context.elementwise_binary.a_stride[XNN_MAX_TENSOR_DIMS - 1 - i] = a_stride… in setup_binary_elementwise_nd() 999 …binary_elementwise_op->context.elementwise_binary.b_stride[XNN_MAX_TENSOR_DIMS - 1 - i] = b_stride… in setup_binary_elementwise_nd() 1001 …binary_elementwise_op->context.elementwise_binary.y_stride[XNN_MAX_TENSOR_DIMS - 1 - i] = y_stride… in setup_binary_elementwise_nd()
|
D | transpose-nd.c | 28 size_t tmp[XNN_MAX_TENSOR_DIMS]; in reorder_array() 124 if (num_dims > XNN_MAX_TENSOR_DIMS) { in setup_transpose_nd() 127 xnn_operator_type_to_string(transpose_op->type), num_dims, XNN_MAX_TENSOR_DIMS); in setup_transpose_nd() 189 size_t normalized_shape[XNN_MAX_TENSOR_DIMS]; in setup_transpose_nd() 190 size_t normalized_perm[XNN_MAX_TENSOR_DIMS]; in setup_transpose_nd() 195 size_t loop_order[XNN_MAX_TENSOR_DIMS]; in setup_transpose_nd()
|
/external/XNNPACK/test/ |
D | constant-pad-operator-tester.h | 25 assert(input_shape.size() <= XNN_MAX_TENSOR_DIMS); in input_shape() 48 assert(pre_paddings.size() <= XNN_MAX_TENSOR_DIMS); in pre_paddings() 66 assert(post_paddings.size() <= XNN_MAX_TENSOR_DIMS); in post_paddings() 114 std::array<size_t, XNN_MAX_TENSOR_DIMS> input_dims; in TestX8() 115 std::array<size_t, XNN_MAX_TENSOR_DIMS> input_pre_paddings; in TestX8() 116 std::array<size_t, XNN_MAX_TENSOR_DIMS> input_post_paddings; in TestX8() 117 std::array<size_t, XNN_MAX_TENSOR_DIMS> output_dims; in TestX8() 123 input_dims[XNN_MAX_TENSOR_DIMS - num_dims() + i] = input_dim(i); in TestX8() 124 input_pre_paddings[XNN_MAX_TENSOR_DIMS - num_dims() + i] = pre_padding(i); in TestX8() 125 input_post_paddings[XNN_MAX_TENSOR_DIMS - num_dims() + i] = post_padding(i); in TestX8() [all …]
|
D | binary-elementwise-operator-tester.h | 40 assert(input1_shape.size() <= XNN_MAX_TENSOR_DIMS); in input1_shape() 82 assert(input2_shape.size() <= XNN_MAX_TENSOR_DIMS); in input2_shape() 214 std::array<size_t, XNN_MAX_TENSOR_DIMS> input1_dims; in TestQS8() 215 std::array<size_t, XNN_MAX_TENSOR_DIMS> input2_dims; in TestQS8() 216 std::array<size_t, XNN_MAX_TENSOR_DIMS> output_dims; in TestQS8() 222 for (size_t i = 0; i < XNN_MAX_TENSOR_DIMS; i++) { in TestQS8() 232 std::array<size_t, XNN_MAX_TENSOR_DIMS> input1_strides; in TestQS8() 233 std::array<size_t, XNN_MAX_TENSOR_DIMS> input2_strides; in TestQS8() 234 std::array<size_t, XNN_MAX_TENSOR_DIMS> output_strides; in TestQS8() 236 for (size_t i = XNN_MAX_TENSOR_DIMS; i != 0; i--) { in TestQS8() [all …]
|
D | subgraph-binary-tester.h | 31 shape_dist = std::uniform_int_distribution<size_t>(0, XNN_MAX_TENSOR_DIMS); in BinaryTest() 48 …t2_num_dims = std::uniform_int_distribution<size_t>(input1_shape.size(), XNN_MAX_TENSOR_DIMS)(rng); in SetUp() 68 for (size_t i = 0; i < XNN_MAX_TENSOR_DIMS; i++) { in SetUp() 95 size_t NumElements(std::array<size_t, XNN_MAX_TENSOR_DIMS>& dims) in NumElements() 112 std::array<size_t, XNN_MAX_TENSOR_DIMS> input1_dims; 113 std::array<size_t, XNN_MAX_TENSOR_DIMS> input2_dims; 114 std::array<size_t, XNN_MAX_TENSOR_DIMS> output_dims;
|
D | static-constant-pad.cc | 31 std::array<size_t, XNN_MAX_TENSOR_DIMS> pre_paddings; in TEST_F() 32 std::array<size_t, XNN_MAX_TENSOR_DIMS> post_paddings; in TEST_F() 83 std::array<size_t, XNN_MAX_TENSOR_DIMS> pre_paddings; in TEST_F() 84 std::array<size_t, XNN_MAX_TENSOR_DIMS> post_paddings; in TEST_F() 133 std::array<size_t, XNN_MAX_TENSOR_DIMS> pre_paddings; in TEST_F() 134 std::array<size_t, XNN_MAX_TENSOR_DIMS> post_paddings; in TEST_F() 185 std::array<size_t, XNN_MAX_TENSOR_DIMS> pre_paddings; in TEST_F() 186 std::array<size_t, XNN_MAX_TENSOR_DIMS> post_paddings; in TEST_F() 259 std::array<size_t, XNN_MAX_TENSOR_DIMS> pre_paddings; in TEST_F() 260 std::array<size_t, XNN_MAX_TENSOR_DIMS> post_paddings; in TEST_F() [all …]
|
D | transpose-normalization-tester.h | 46 assert(shape.size() <= XNN_MAX_TENSOR_DIMS); in shape() 52 assert(perm.size() <= XNN_MAX_TENSOR_DIMS); in perm() 58 assert(input_stride.size() <= XNN_MAX_TENSOR_DIMS); in input_stride() 64 assert(output_stride.size() <= XNN_MAX_TENSOR_DIMS); in output_stride()
|
D | subgraph-unary-tester.h | 28 shape_dist = std::uniform_int_distribution<size_t>(min_dim, XNN_MAX_TENSOR_DIMS); in UnaryTest()
|
D | transpose-operator-tester.h | 47 assert(shape.size() <= XNN_MAX_TENSOR_DIMS); in shape() 55 assert(perm.size() <= XNN_MAX_TENSOR_DIMS); in perm()
|
D | global-average-pooling-1d.cc | 29 shape_dist = std::uniform_int_distribution<size_t>(2, XNN_MAX_TENSOR_DIMS); in GlobalAveragePooling1DTest()
|
D | global-average-pooling-2d.cc | 29 shape_dist = std::uniform_int_distribution<size_t>(1, XNN_MAX_TENSOR_DIMS); in GlobalAveragePooling2DTest()
|
D | concatenate2.cc | 28 shape_dist = std::uniform_int_distribution<size_t>(1, XNN_MAX_TENSOR_DIMS); in Concatenate2Test()
|
D | even-split2.cc | 28 shape_dist = std::uniform_int_distribution<size_t>(1, XNN_MAX_TENSOR_DIMS); in EvenSplit2Test()
|
D | concatenate3.cc | 28 shape_dist = std::uniform_int_distribution<size_t>(1, XNN_MAX_TENSOR_DIMS); in Concatenate3Test()
|
D | fully-connected.cc | 36 auto shape_dist = std::uniform_int_distribution<size_t>(2, XNN_MAX_TENSOR_DIMS); in FullyConnectedTestBase()
|
D | even-split3.cc | 28 shape_dist = std::uniform_int_distribution<size_t>(1, XNN_MAX_TENSOR_DIMS); in EvenSplit3Test()
|
D | concatenate4.cc | 28 shape_dist = std::uniform_int_distribution<size_t>(1, XNN_MAX_TENSOR_DIMS); in Concatenate4Test()
|
/external/XNNPACK/src/ |
D | tensor.c | 42 if (num_dims > XNN_MAX_TENSOR_DIMS) { in xnn_define_tensor_value() 44 XNN_MAX_TENSOR_DIMS); in xnn_define_tensor_value() 101 if (num_dims > XNN_MAX_TENSOR_DIMS) { in xnn_define_quantized_tensor_value() 104 XNN_MAX_TENSOR_DIMS); in xnn_define_quantized_tensor_value() 197 if (num_dims > XNN_MAX_TENSOR_DIMS) { in xnn_define_channelwise_quantized_tensor_value() 200 XNN_MAX_TENSOR_DIMS); in xnn_define_channelwise_quantized_tensor_value()
|
/external/XNNPACK/src/xnnpack/ |
D | subgraph.h | 43 size_t dim[XNN_MAX_TENSOR_DIMS]; 248 size_t pre_paddings[XNN_MAX_TENSOR_DIMS]; 249 size_t post_paddings[XNN_MAX_TENSOR_DIMS]; 266 size_t perm[XNN_MAX_TENSOR_DIMS]; 315 size_t pre_paddings[XNN_MAX_TENSOR_DIMS]; 316 size_t post_paddings[XNN_MAX_TENSOR_DIMS];
|
D | compute.h | 75 size_t input_stride[XNN_MAX_TENSOR_DIMS]; 76 size_t output_stride[XNN_MAX_TENSOR_DIMS]; 824 size_t a_stride[XNN_MAX_TENSOR_DIMS - 1]; 826 size_t b_stride[XNN_MAX_TENSOR_DIMS - 1]; 828 size_t y_stride[XNN_MAX_TENSOR_DIMS - 1]; 1047 size_t input_stride[XNN_MAX_TENSOR_DIMS - 1]; 1049 size_t output_stride[XNN_MAX_TENSOR_DIMS - 1]; 1050 size_t pre_paddings[XNN_MAX_TENSOR_DIMS]; 1052 size_t input_size[XNN_MAX_TENSOR_DIMS];
|
/external/XNNPACK/src/subgraph/ |
D | static-reshape.c | 246 if (num_dims > XNN_MAX_TENSOR_DIMS) { in xnn_define_static_reshape() 249 … xnn_node_type_to_string(xnn_node_type_static_reshape), num_dims, (size_t) XNN_MAX_TENSOR_DIMS); in xnn_define_static_reshape()
|
D | static-transpose.c | 155 if (num_dims > XNN_MAX_TENSOR_DIMS) { in xnn_define_static_transpose() 158 xnn_node_type_to_string(xnn_node_type_static_transpose), num_dims, XNN_MAX_TENSOR_DIMS); in xnn_define_static_transpose()
|
D | static-constant-pad.c | 74 …(opdata->pre_paddings, node->params.static_pad.pre_paddings, sizeof(size_t) * XNN_MAX_TENSOR_DIMS); in create_constant_pad_operator() 75 …pdata->post_paddings, node->params.static_pad.post_paddings, sizeof(size_t) * XNN_MAX_TENSOR_DIMS); in create_constant_pad_operator()
|
/external/tensorflow/tensorflow/lite/delegates/xnnpack/ |
D | xnnpack_delegate.cc | 3896 XNN_MAX_TENSOR_DIMS, in VisitPadNode() 3916 XNN_MAX_TENSOR_DIMS, in VisitPadNode() 3944 std::array<size_t, XNN_MAX_TENSOR_DIMS> pre_paddings{}; in VisitPadNode() 3945 std::array<size_t, XNN_MAX_TENSOR_DIMS> post_paddings{}; in VisitPadNode() 3979 XNN_MAX_TENSOR_DIMS, in VisitPreluNode() 3998 XNN_MAX_TENSOR_DIMS, in VisitPreluNode() 4157 XNN_MAX_TENSOR_DIMS, in VisitReshapeNode() 4177 XNN_MAX_TENSOR_DIMS, in VisitReshapeNode() 4183 std::array<size_t, XNN_MAX_TENSOR_DIMS> new_shape; in VisitReshapeNode() 4412 std::array<size_t, XNN_MAX_TENSOR_DIMS> perm; in VisitTransposeNode()
|