Home
last modified time | relevance | path

Searched refs:num_split (Results 1 – 25 of 69) sorted by relevance

123

/external/tensorflow/tensorflow/core/kernels/
Dsplit_op_test.cc29 static Graph* MakeGraph(int split_dim, int num_split, in MakeGraph() argument
33 in_shape.set_dim(split_dim, in_shape.dim_size(split_dim) * num_split); in MakeGraph()
41 .Attr("num_split", num_split) in MakeGraph()
46 #define BM_SPLIT_1D(num_split, chunk_size) \ argument
47 static void BM_Split_1d_##num_split##_##chunk_size( \
50 strings::Printf("1-D %d chunks of %d each", num_split, chunk_size); \
52 auto g = MakeGraph(/* split_dim = */ 0, num_split, {chunk_size}); \
55 num_split * chunk_size); \
57 BENCHMARK(BM_Split_1d_##num_split##_##chunk_size)->UseRealTime();
59 #define BM_SPLIT_2D(split_dim, num_split, chunk_size0, chunk_size1) \ argument
[all …]
Dsplit_op.cc58 const int32 num_split = num_outputs(); in ComputeEasyCases() local
67 context, num_split > 0, in ComputeEasyCases()
69 "Number of ways to split should be > 0, but got ", num_split)); in ComputeEasyCases()
71 OP_REQUIRES(context, input_shape.dim_size(split_dim) % num_split == 0, in ComputeEasyCases()
76 ") ", "and num_split ", num_split)); in ComputeEasyCases()
78 if (num_split == 1) { in ComputeEasyCases()
95 const int64 delta = input_shape.dim_size(0) / num_split; in ComputeEasyCases()
96 for (int i = 0; i < num_split; ++i) { in ComputeEasyCases()
138 const ReshapeResultType& reshape_result, int32 num_split, in operator ()() argument
145 (num_split >= 4 && in operator ()()
[all …]
Dsplit_v_op.cc55 const int32 num_split = context->num_outputs(); in ComputeEasyCases() local
71 split_tensor.dims() == 1 && split_tensor.NumElements() == num_split, in ComputeEasyCases()
85 context, num_split > 0, in ComputeEasyCases()
87 "Number of ways to split should be > 0, but got ", num_split)); in ComputeEasyCases()
98 if (num_split == 1) { in ComputeEasyCases()
151 for (int i = 0; i < num_split; ++i) { in ComputeEasyCases()
206 const int num_split = split_start_points.size(); in operator ()() local
208 (num_split >= 4 && in operator ()()
209 input_element_count >= std::min(num_threads, num_split) * 4096 && in operator ()()
210 input_element_count < num_split * 180 * 1024); in operator ()()
[all …]
Dsplit_v_op_test.cc75 #define BM_SPLITV_1D(num_split, total_size) \ argument
76 static void BM_SplitV_1d_##num_split##_##total_size( \
79 strings::Printf("1-D %d chunks totaling %d", num_split, total_size); \
82 GenerateRandomIntsWithSum(total_size, num_split), \
88 BENCHMARK(BM_SplitV_1d_##num_split##_##total_size)->UseRealTime();
90 #define BM_SPLITV_2D(split_dim, num_split, total_size0, total_size1) \ argument
92 BM_SplitV_2d_##split_dim##_##num_split##_##total_size0##_##total_size1( \
97 num_split, split_dim, total_size0, total_size1); \
101 GenerateRandomIntsWithSum(total_size_vec[split_dim], num_split), \
108 BM_SplitV_2d_##split_dim##_##num_split##_##total_size0##_##total_size1) \
Dsplit_lib_gpu.cu.cc77 const int32 num_split = output_ptr_data.size; in SplitOpKernel() local
82 eigen_assert(split_dim_size % num_split == 0); in SplitOpKernel()
85 int32 piece_size = split_dim_size / num_split; in SplitOpKernel()
173 const int32 num_split = output_ptr_data.size; in SplitVOpKernel_fixed() local
180 int32 piece_size = suffix_dim_size / num_split; in SplitVOpKernel_fixed()
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/
Dsplit_op.cc37 const int32 num_split = num_outputs(); in Compile() local
57 ctx, num_split > 0, in Compile()
59 "Number of ways to split should be > 0, but got ", num_split)); in Compile()
62 ctx, input_shape.dim_size(split_dim) % num_split == 0, in Compile()
67 "and num_split ", num_split)); in Compile()
71 const int32 slice_size = input_shape.dim_size(split_dim) / num_split; in Compile()
88 for (int i = 0; i < num_split; ++i) { in Compile()
104 const int32 num_split = num_outputs(); in Compile() local
128 ctx, num_split > 0, in Compile()
130 "Number of ways to split should be > 0, but got ", num_split)); in Compile()
[all …]
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v1/
DSparseSplit.pbtxt22 number_attr: "num_split"
27 number_attr: "num_split"
32 number_attr: "num_split"
35 name: "num_split"
DSplit.pbtxt14 number_attr: "num_split"
17 name: "num_split"
DSplitV.pbtxt18 number_attr: "num_split"
21 name: "num_split"
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v2/
DSparseSplit.pbtxt22 number_attr: "num_split"
27 number_attr: "num_split"
32 number_attr: "num_split"
35 name: "num_split"
DSplit.pbtxt14 number_attr: "num_split"
17 name: "num_split"
DSplitV.pbtxt18 number_attr: "num_split"
21 name: "num_split"
/external/tensorflow/tensorflow/python/kernel_tests/
Dsparse_split_op_test.py83 sp_input=self._SparseTensor_4x6(), num_split=2, axis=axis))
100 sp_input=self._SparseTensor_5x7(), num_split=3, axis=axis))
116 sp_input=self._SparseTensor_5x7(), num_split=4, axis=axis)
138 sp_input=self._SparseTensor_5x7(), num_split=2, axis=axis))
152 sp_input=self._SparseTensor_5x7(), num_split=3, axis=axis)
170 sp_input=self._SparseTensor_4x6(), num_split=4, axis=axis))
191 sp_input=self._SparseTensor_4x6(), num_split=3, axis=axis))
210 sp_input=self._SparseTensor_4x6(), num_split=6, axis=axis))
236 sp_input=sp_input, num_split=2, axis=axis)
248 sp_input=self._SparseTensor_4x6(), num_split=3, axis=axis))
[all …]
Dsplit_op_test.py132 num_split = np.random.randint(16, 25)
134 num_split = np.random.randint(2, 8)
135 size_splits = np.random.randint(2, 8, num_split, dtype=np.int32)
142 for i in range(num_split):
159 num_split = 1000
160 size_splits = np.random.randint(1, 3, num_split, dtype=np.int32)
168 for i in range(num_split):
290 num_split = np.random.randint(9, 15)
292 num_split = np.random.randint(2, 8)
293 shape[split_dim] = np.random.randint(2, 5) * num_split
[all …]
/external/tensorflow/tensorflow/core/api_def/base_api/
Dapi_def_SparseSplit.pbtxt45 name: "num_split"
50 summary: "Split a `SparseTensor` into `num_split` tensors along one dimension."
52 If the `shape[split_dim]` is not an integer multiple of `num_split`. Slices
53 `[0 : shape[split_dim] % num_split]` gets one extra dimension.
54 For example, if `split_dim = 1` and `num_split = 2` and the input is
Dapi_def_Split.pbtxt22 `values.shape[split_dim] / num_split`.
26 name: "num_split"
32 summary: "Splits a tensor into `num_split` tensors along one dimension."
Dapi_def_SplitV.pbtxt33 summary: "Splits a tensor into `num_split` tensors along one dimension."
/external/tensorflow/tensorflow/compiler/mlir/tfr/examples/pad/
Dops_defs.py67 num_split=2)
72 num_split=2)
78 num_split=2)
83 num_split=2)
114 num_split=3)
123 num_split=2)
128 num_split=2)
/external/tensorflow/tensorflow/core/util/sparse/
Dsparse_tensor.h173 const int num_split, std::vector<SparseTensor>* result);
496 const int split_dim, const int num_split, in Split() argument
501 output_indices.reserve(num_split); in Split()
502 output_values.reserve(num_split); in Split()
503 output_shapes.reserve(num_split); in Split()
507 output_indices_t.reserve(num_split); in Split()
508 output_values_t.reserve(num_split); in Split()
512 std::vector<int> num_values(num_split, 0); in Split()
515 const int split_size = split_dim_size / num_split; in Split()
517 if (!(num_split > 0 && num_split <= split_dim_size)) { in Split()
[all …]
/external/tensorflow/tensorflow/java/src/test/java/org/tensorflow/
DGraphOperationTest.java182 private static int split(int[] values, int num_split) { in split() argument
187 .setAttr("num_split", num_split) in split()
193 private static int splitWithInputList(int[] values, int num_split, String name) { in splitWithInputList() argument
198 .setAttr("num_split", num_split) in splitWithInputList()
/external/tensorflow/tensorflow/compiler/mlir/tfr/python/
Dtfr_gen_test.py81 z, _ = array_ops.Split(axis=0, value=x, num_split=2)
88 z = array_ops.Split(axis=0, value=x, num_split=2)
145 value=lhs, size_splits=[rhs, -1], axis=0, num_split=2)
147 value=lhs, size_splits=[rhs, rhs], axis=1, num_split=2)
/external/mesa3d/src/gallium/auxiliary/gallivm/
Dlp_bld_pack.c650 int num_split = src_type.width * src_type.length / 128; in lp_build_pack2() local
662 assert(num_split <= LP_MAX_VECTOR_WIDTH / 128); in lp_build_pack2()
664 for (i = 0; i < num_split / 2; i++) { in lp_build_pack2()
675 for (i = 0; i < num_split / 2; i++) { in lp_build_pack2()
680 tmpres[i+num_split/2] = lp_build_intrinsic_binary(builder, intrinsic, in lp_build_pack2()
684 tmpres[i+num_split/2] = LLVMBuildBitCast(builder, tmpres[i+num_split/2], in lp_build_pack2()
688 res = lp_build_concat(gallivm, tmpres, ndst_type, num_split); in lp_build_pack2()
/external/tensorflow/tensorflow/compiler/mlir/tensorflow/utils/
Dxla_sharding_util.cc49 mlir::LogicalResult CreateSplitOp(const int num_split, in CreateSplitOp() argument
74 if (shape[split_dimension] % num_split != 0) { in CreateSplitOp()
80 split_dimension, num_split)); in CreateSplitOp()
83 shape[split_dimension] = shape[split_dimension] / num_split; in CreateSplitOp()
92 llvm::SmallVector<mlir::Type, 4> output_types(num_split, output_type); in CreateSplitOp()
97 builder->getIntegerAttr(builder->getIntegerType(32), num_split)); in CreateSplitOp()
/external/tensorflow/tensorflow/compiler/mlir/lite/transforms/
Dlegalize_tf.cc307 auto num_split = rewriter.getI32IntegerAttr(tf_split_op.num_split()); in matchAndRewrite() local
311 tf_split_op.value(), num_split); in matchAndRewrite()
320 auto num_split = rewriter.getI32IntegerAttr(tf_splitv_op.num_split()); in matchAndRewrite() local
324 tf_splitv_op.size_splits(), tf_splitv_op.split_dim(), num_split); in matchAndRewrite()
/external/tensorflow/tensorflow/compiler/mlir/tensorflow/tests/mlir2graphdef/
Dderived_size_attr.mlir5 // CHECK: key: "num_split"

123