/external/tensorflow/tensorflow/core/kernels/ |
D | gather_functor_batched_gpu.cu.h | 35 int64 outer_size, in GatherOpKernel() argument 59 batch_i = entries_count / outer_size; in GatherOpKernel() 60 outer_i = entries_count - batch_i * outer_size; in GatherOpKernel() 78 (batch_i * outer_size + outer_i) * gather_dim_size + gather_i in GatherOpKernel() 103 const int64 outer_size = params.dimension(1); 119 outer_size, gather_dim_size, indices_size, slice_size, out_size));
|
D | gather_functor_batched.h | 45 const SliceIndex outer_size = static_cast<SliceIndex>(params.dimension(1)); in HandleCopiesBatched() local 62 const int64 r_start = start % (outer_size * indices_size); in HandleCopiesBatched() 64 start / (outer_size * indices_size)); in HandleCopiesBatched() 77 if (++o_next >= outer_size) { in HandleCopiesBatched() 123 batch_size * outer_size * indices_size, slice_elems * sizeof(T), work); in HandleCopiesBatched() 138 const int64 outer_size = params.dimension(1); in operator() local 143 batch_size * outer_size * indices_size * slice_size > in operator()
|
D | gather_op.cc | 138 int64 outer_size = 1; in Compute() local 147 outer_size *= params.dim_size(i); in Compute() 166 {batch_size, outer_size, gather_dim_size, inner_size}); in Compute() 168 {batch_size, outer_size, N / batch_size, inner_size}); in Compute() 174 params.shaped<T, 3>({outer_size, gather_dim_size, inner_size}); in Compute() 175 auto out_flat = out->shaped<T, 3>({outer_size, N, inner_size}); in Compute()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | concatenation.h | 49 int64_t outer_size = 1; in Concatenation() local 51 outer_size *= output_shape.Dims(i); in Concatenation() 61 for (int k = 0; k < outer_size; k++) { in Concatenation() 100 int64_t outer_size = 1; in ConcatenationWithScaling() local 102 outer_size *= output_shape.Dims(i); in ConcatenationWithScaling() 113 for (int k = 0; k < outer_size; k++) { in ConcatenationWithScaling()
|
D | l2normalization.h | 35 const int outer_size = variable 39 for (int i = 0; i < outer_size; ++i) { 61 const int outer_size = in L2Normalization() local 65 for (int i = 0; i < outer_size; ++i) { in L2Normalization()
|
D | gather.h | 40 int outer_size = 1; in Gather() local 42 outer_size *= input_shape.Dims(i); in Gather() 50 for (int outer = 0; outer < outer_size; ++outer) { in Gather()
|
D | arg_min_max.h | 37 int outer_size = 1; in ArgMinMax() local 40 outer_size *= input1_shape.Dims(i); in ArgMinMax() 49 for (int outer = 0; outer < outer_size; ++outer) { in ArgMinMax()
|
D | softmax.h | 34 const int outer_size = in Softmax() local 39 for (int i = 0; i < outer_size; ++i) { in Softmax() 87 const int outer_size = in Softmax() local 92 for (int i = 0; i < outer_size; ++i) { in Softmax() 171 const int outer_size = in SoftmaxInt16() local 176 for (int i = 0; i < outer_size; ++i) { in SoftmaxInt16()
|
D | reference_ops.h | 325 int outer_size = 1; in Pack() local 327 outer_size *= output_shape.Dims(i); in Pack() 333 TFLITE_DCHECK_EQ((**input_shapes).FlatSize(), copy_size * outer_size); in Pack() 336 for (int k = 0; k < outer_size; k++) { in Pack() 352 int outer_size = 1; in Unpack() local 360 outer_size *= input_shape.Dims(i); in Unpack() 366 TFLITE_DCHECK_EQ(output_shape.FlatSize(), copy_size * outer_size); in Unpack() 369 for (int k = 0; k < outer_size; k++) { in Unpack() 391 int outer_size = 1; in PackWithScaling() local 393 outer_size *= output_shape.Dims(i); in PackWithScaling() [all …]
|
/external/perfetto/src/base/ |
D | paged_memory.cc | 56 size_t outer_size = rounded_up_size + GuardSize() * 2; in Allocate() local 58 void* ptr = VirtualAlloc(nullptr, outer_size, MEM_RESERVE, PAGE_NOACCESS); in Allocate() 64 void* ptr = mmap(nullptr, outer_size, PROT_READ | PROT_WRITE, in Allocate() 113 const size_t outer_size = RoundUpToSysPageSize(size_) + GuardSize() * 2; in ~PagedMemory() local 114 int res = munmap(start, outer_size); in ~PagedMemory()
|
/external/tensorflow/tensorflow/lite/micro/kernels/ |
D | pack.cc | 43 int outer_size = 1; in PackImpl() local 45 outer_size *= output_dims->data[i]; in PackImpl() 55 TFLITE_DCHECK_EQ(input_size, copy_size * outer_size); in PackImpl() 62 for (int k = 0; k < outer_size; ++k) { in PackImpl()
|
D | unpack.cc | 46 int outer_size = 1; in UnpackImpl() local 48 outer_size *= input_dims->data[i]; in UnpackImpl() 58 TFLITE_DCHECK_EQ(output_size, copy_size * outer_size); in UnpackImpl() 65 for (int k = 0; k < outer_size; ++k) { in UnpackImpl()
|
D | split.cc | 45 int64_t outer_size = 1; in SplitImpl() local 47 outer_size *= input_dims->data[i]; in SplitImpl() 56 for (int k = 0; k < outer_size; ++k) { in SplitImpl()
|
D | split_v.cc | 47 int64_t outer_size = 1; in SplitImpl() local 49 outer_size *= input_dims->data[i]; in SplitImpl() 58 for (int k = 0; k < outer_size; ++k) { in SplitImpl()
|
D | l2norm.cc | 125 const int outer_size = in Eval() local 128 data.input_zero_point, outer_size, depth, in Eval()
|
/external/perfetto/src/profiling/memory/ |
D | shared_ring_buffer.cc | 105 size_t outer_size = kMetaPageSize + size_ * 2 + kGuardSize; in ~SharedRingBuffer() local 106 munmap(meta_, outer_size); in ~SharedRingBuffer() 155 size_t outer_size = kMetaPageSize + size * 2 + kGuardSize; in Initialize() local 157 mmap(nullptr, outer_size, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0)); in Initialize() 175 munmap(region, outer_size); in Initialize()
|
/external/tensorflow/tensorflow/lite/micro/kernels/ceva/ |
D | ceva_tflm_lib.h | 275 const int outer_size, const int depth); 277 int32_t outer_size, int32_t depth, 321 int outer_size, int copy_size, 324 int* const outer_size, 328 int outer_size, int copy_size, 331 int outer_size, int copy_size, 334 int outer_size, int copy_size, 515 const float* input_ptr, int outer_size, 518 const int8_t* input_ptr, int outer_size,
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/integer_ops/ |
D | l2normalization.h | 23 inline void L2Normalization(int32_t input_zero_point, int32_t outer_size, in L2Normalization() argument 32 for (int outer_index = 0; outer_index < outer_size; ++outer_index) { in L2Normalization()
|
D | log_softmax.h | 25 int32_t diff_min, int32_t outer_size, int32_t depth, in LogSoftmax() argument 43 for (int outer_index = 0; outer_index < outer_size; ++outer_index) { in LogSoftmax()
|
/external/tensorflow/tensorflow/lite/micro/kernels/cmsis_nn/ |
D | softmax.cc | 50 const int outer_size = in SoftmaxQuantized() local 55 arm_softmax_s8(tflite::micro::GetTensorData<int8_t>(input), outer_size, in SoftmaxQuantized()
|
/external/tensorflow/tensorflow/lite/micro/kernels/xtensa/ |
D | softmax.cc | 69 const int outer_size = in SoftmaxHifimini() local 74 for (int i = 0; i < outer_size; ++i) { in SoftmaxHifimini() 219 const int outer_size = in EvalHifi4() local 227 for (int i = 0; i < outer_size; ++i) { in EvalHifi4()
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | l2norm.cc | 136 const int outer_size = in Eval() local 138 reference_integer_ops::L2Normalization(input->params.zero_point, outer_size, in Eval()
|
/external/ruy/example/ |
D | parametrized_example.cc | 108 int outer_size = in run() local 110 return outer_size * layout.stride(); in run()
|
/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | logsoftmax_quantized_test.cc | 230 const int outer_size = in RunOneLogSoftmaxTest() local 235 reverse_scaling_right_shift, diff_min, outer_size, inner_size, input_data, in RunOneLogSoftmaxTest()
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | array_ops_test.py | 479 for outer_size in (1, 2): 483 outer_size=outer_size, 486 np.arange(outer_size * middle_size * 3, dtype=np.float32), 487 newshape=(outer_size, middle_size, 3)) 496 for outer_size in (1, 2): 500 outer_size=outer_size, 503 np.arange(outer_size * middle_size * 4, dtype=np.float32), 504 newshape=(outer_size, middle_size, 4)) 513 for outer_size in list(range(50)) + [100000]: 517 outer_size=outer_size, [all …]
|