/external/tensorflow/tensorflow/core/kernels/ |
D | split_op_test.cc | 29 static Graph* MakeGraph(int split_dim, int num_split, in MakeGraph() argument 33 in_shape.set_dim(split_dim, in_shape.dim_size(split_dim) * num_split); in MakeGraph() 41 .Attr("num_split", num_split) in MakeGraph() 46 #define BM_SPLIT_1D(num_split, chunk_size) \ argument 47 static void BM_Split_1d_##num_split##_##chunk_size( \ 50 strings::Printf("1-D %d chunks of %d each", num_split, chunk_size); \ 52 auto g = MakeGraph(/* split_dim = */ 0, num_split, {chunk_size}); \ 55 num_split * chunk_size); \ 57 BENCHMARK(BM_Split_1d_##num_split##_##chunk_size)->UseRealTime(); 59 #define BM_SPLIT_2D(split_dim, num_split, chunk_size0, chunk_size1) \ argument [all …]
|
D | split_op.cc | 58 const int32 num_split = num_outputs(); in ComputeEasyCases() local 67 context, num_split > 0, in ComputeEasyCases() 69 "Number of ways to split should be > 0, but got ", num_split)); in ComputeEasyCases() 71 OP_REQUIRES(context, input_shape.dim_size(split_dim) % num_split == 0, in ComputeEasyCases() 76 ") ", "and num_split ", num_split)); in ComputeEasyCases() 78 if (num_split == 1) { in ComputeEasyCases() 95 const int64 delta = input_shape.dim_size(0) / num_split; in ComputeEasyCases() 96 for (int i = 0; i < num_split; ++i) { in ComputeEasyCases() 138 const ReshapeResultType& reshape_result, int32 num_split, in operator ()() argument 145 (num_split >= 4 && in operator ()() [all …]
|
D | split_v_op.cc | 55 const int32 num_split = context->num_outputs(); in ComputeEasyCases() local 71 split_tensor.dims() == 1 && split_tensor.NumElements() == num_split, in ComputeEasyCases() 85 context, num_split > 0, in ComputeEasyCases() 87 "Number of ways to split should be > 0, but got ", num_split)); in ComputeEasyCases() 98 if (num_split == 1) { in ComputeEasyCases() 151 for (int i = 0; i < num_split; ++i) { in ComputeEasyCases() 206 const int num_split = split_start_points.size(); in operator ()() local 208 (num_split >= 4 && in operator ()() 209 input_element_count >= std::min(num_threads, num_split) * 4096 && in operator ()() 210 input_element_count < num_split * 180 * 1024); in operator ()() [all …]
|
D | split_v_op_test.cc | 75 #define BM_SPLITV_1D(num_split, total_size) \ argument 76 static void BM_SplitV_1d_##num_split##_##total_size( \ 79 strings::Printf("1-D %d chunks totaling %d", num_split, total_size); \ 82 GenerateRandomIntsWithSum(total_size, num_split), \ 88 BENCHMARK(BM_SplitV_1d_##num_split##_##total_size)->UseRealTime(); 90 #define BM_SPLITV_2D(split_dim, num_split, total_size0, total_size1) \ argument 92 BM_SplitV_2d_##split_dim##_##num_split##_##total_size0##_##total_size1( \ 97 num_split, split_dim, total_size0, total_size1); \ 101 GenerateRandomIntsWithSum(total_size_vec[split_dim], num_split), \ 108 BM_SplitV_2d_##split_dim##_##num_split##_##total_size0##_##total_size1) \
|
D | split_lib_gpu.cu.cc | 77 const int32 num_split = output_ptr_data.size; in SplitOpKernel() local 82 eigen_assert(split_dim_size % num_split == 0); in SplitOpKernel() 85 int32 piece_size = split_dim_size / num_split; in SplitOpKernel() 173 const int32 num_split = output_ptr_data.size; in SplitVOpKernel_fixed() local 180 int32 piece_size = suffix_dim_size / num_split; in SplitVOpKernel_fixed()
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | split_op.cc | 37 const int32 num_split = num_outputs(); in Compile() local 57 ctx, num_split > 0, in Compile() 59 "Number of ways to split should be > 0, but got ", num_split)); in Compile() 62 ctx, input_shape.dim_size(split_dim) % num_split == 0, in Compile() 67 "and num_split ", num_split)); in Compile() 71 const int32 slice_size = input_shape.dim_size(split_dim) / num_split; in Compile() 88 for (int i = 0; i < num_split; ++i) { in Compile() 104 const int32 num_split = num_outputs(); in Compile() local 128 ctx, num_split > 0, in Compile() 130 "Number of ways to split should be > 0, but got ", num_split)); in Compile() [all …]
|
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v1/ |
D | SparseSplit.pbtxt | 22 number_attr: "num_split" 27 number_attr: "num_split" 32 number_attr: "num_split" 35 name: "num_split"
|
D | Split.pbtxt | 14 number_attr: "num_split" 17 name: "num_split"
|
D | SplitV.pbtxt | 18 number_attr: "num_split" 21 name: "num_split"
|
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v2/ |
D | SparseSplit.pbtxt | 22 number_attr: "num_split" 27 number_attr: "num_split" 32 number_attr: "num_split" 35 name: "num_split"
|
D | Split.pbtxt | 14 number_attr: "num_split" 17 name: "num_split"
|
D | SplitV.pbtxt | 18 number_attr: "num_split" 21 name: "num_split"
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | sparse_split_op_test.py | 83 sp_input=self._SparseTensor_4x6(), num_split=2, axis=axis)) 100 sp_input=self._SparseTensor_5x7(), num_split=3, axis=axis)) 116 sp_input=self._SparseTensor_5x7(), num_split=4, axis=axis) 138 sp_input=self._SparseTensor_5x7(), num_split=2, axis=axis)) 152 sp_input=self._SparseTensor_5x7(), num_split=3, axis=axis) 170 sp_input=self._SparseTensor_4x6(), num_split=4, axis=axis)) 191 sp_input=self._SparseTensor_4x6(), num_split=3, axis=axis)) 210 sp_input=self._SparseTensor_4x6(), num_split=6, axis=axis)) 236 sp_input=sp_input, num_split=2, axis=axis) 248 sp_input=self._SparseTensor_4x6(), num_split=3, axis=axis)) [all …]
|
D | split_op_test.py | 132 num_split = np.random.randint(16, 25) 134 num_split = np.random.randint(2, 8) 135 size_splits = np.random.randint(2, 8, num_split, dtype=np.int32) 142 for i in range(num_split): 159 num_split = 1000 160 size_splits = np.random.randint(1, 3, num_split, dtype=np.int32) 168 for i in range(num_split): 290 num_split = np.random.randint(9, 15) 292 num_split = np.random.randint(2, 8) 293 shape[split_dim] = np.random.randint(2, 5) * num_split [all …]
|
/external/tensorflow/tensorflow/core/api_def/base_api/ |
D | api_def_SparseSplit.pbtxt | 45 name: "num_split" 50 summary: "Split a `SparseTensor` into `num_split` tensors along one dimension." 52 If the `shape[split_dim]` is not an integer multiple of `num_split`. Slices 53 `[0 : shape[split_dim] % num_split]` gets one extra dimension. 54 For example, if `split_dim = 1` and `num_split = 2` and the input is
|
D | api_def_Split.pbtxt | 22 `values.shape[split_dim] / num_split`. 26 name: "num_split" 32 summary: "Splits a tensor into `num_split` tensors along one dimension."
|
D | api_def_SplitV.pbtxt | 33 summary: "Splits a tensor into `num_split` tensors along one dimension."
|
/external/tensorflow/tensorflow/compiler/mlir/tfr/examples/pad/ |
D | ops_defs.py | 67 num_split=2) 72 num_split=2) 78 num_split=2) 83 num_split=2) 114 num_split=3) 123 num_split=2) 128 num_split=2)
|
/external/tensorflow/tensorflow/core/util/sparse/ |
D | sparse_tensor.h | 173 const int num_split, std::vector<SparseTensor>* result); 496 const int split_dim, const int num_split, in Split() argument 501 output_indices.reserve(num_split); in Split() 502 output_values.reserve(num_split); in Split() 503 output_shapes.reserve(num_split); in Split() 507 output_indices_t.reserve(num_split); in Split() 508 output_values_t.reserve(num_split); in Split() 512 std::vector<int> num_values(num_split, 0); in Split() 515 const int split_size = split_dim_size / num_split; in Split() 517 if (!(num_split > 0 && num_split <= split_dim_size)) { in Split() [all …]
|
/external/tensorflow/tensorflow/java/src/test/java/org/tensorflow/ |
D | GraphOperationTest.java | 182 private static int split(int[] values, int num_split) { in split() argument 187 .setAttr("num_split", num_split) in split() 193 private static int splitWithInputList(int[] values, int num_split, String name) { in splitWithInputList() argument 198 .setAttr("num_split", num_split) in splitWithInputList()
|
/external/tensorflow/tensorflow/compiler/mlir/tfr/python/ |
D | tfr_gen_test.py | 81 z, _ = array_ops.Split(axis=0, value=x, num_split=2) 88 z = array_ops.Split(axis=0, value=x, num_split=2) 145 value=lhs, size_splits=[rhs, -1], axis=0, num_split=2) 147 value=lhs, size_splits=[rhs, rhs], axis=1, num_split=2)
|
/external/mesa3d/src/gallium/auxiliary/gallivm/ |
D | lp_bld_pack.c | 650 int num_split = src_type.width * src_type.length / 128; in lp_build_pack2() local 662 assert(num_split <= LP_MAX_VECTOR_WIDTH / 128); in lp_build_pack2() 664 for (i = 0; i < num_split / 2; i++) { in lp_build_pack2() 675 for (i = 0; i < num_split / 2; i++) { in lp_build_pack2() 680 tmpres[i+num_split/2] = lp_build_intrinsic_binary(builder, intrinsic, in lp_build_pack2() 684 tmpres[i+num_split/2] = LLVMBuildBitCast(builder, tmpres[i+num_split/2], in lp_build_pack2() 688 res = lp_build_concat(gallivm, tmpres, ndst_type, num_split); in lp_build_pack2()
|
/external/tensorflow/tensorflow/compiler/mlir/tensorflow/utils/ |
D | xla_sharding_util.cc | 49 mlir::LogicalResult CreateSplitOp(const int num_split, in CreateSplitOp() argument 74 if (shape[split_dimension] % num_split != 0) { in CreateSplitOp() 80 split_dimension, num_split)); in CreateSplitOp() 83 shape[split_dimension] = shape[split_dimension] / num_split; in CreateSplitOp() 92 llvm::SmallVector<mlir::Type, 4> output_types(num_split, output_type); in CreateSplitOp() 97 builder->getIntegerAttr(builder->getIntegerType(32), num_split)); in CreateSplitOp()
|
/external/tensorflow/tensorflow/compiler/mlir/lite/transforms/ |
D | legalize_tf.cc | 307 auto num_split = rewriter.getI32IntegerAttr(tf_split_op.num_split()); in matchAndRewrite() local 311 tf_split_op.value(), num_split); in matchAndRewrite() 320 auto num_split = rewriter.getI32IntegerAttr(tf_splitv_op.num_split()); in matchAndRewrite() local 324 tf_splitv_op.size_splits(), tf_splitv_op.split_dim(), num_split); in matchAndRewrite()
|
/external/tensorflow/tensorflow/compiler/mlir/tensorflow/tests/mlir2graphdef/ |
D | derived_size_attr.mlir | 5 // CHECK: key: "num_split"
|