/external/tensorflow/tensorflow/core/kernels/batching_util/ |
D | basic_batch_scheduler_test.cc | 46 Status ScheduleTask(size_t task_size, BatchScheduler<FakeTask>* scheduler) { in ScheduleTask() argument 47 std::unique_ptr<FakeTask> task(new FakeTask(task_size)); in ScheduleTask()
|
D | adaptive_shared_batch_scheduler_test.cc | 47 Status ScheduleTask(size_t task_size, BatchScheduler<FakeTask>* scheduler) { in ScheduleTask() argument 48 std::unique_ptr<FakeTask> task(new FakeTask(task_size)); in ScheduleTask() 398 int task_size = std::min(remaining_size, max_size); in TEST() local 399 output_tasks->emplace_back(new FakeTask(task_size)); in TEST() 400 remaining_size -= task_size; in TEST()
|
D | serial_device_batch_scheduler_test.cc | 44 Status ScheduleTask(size_t task_size, BatchScheduler<FakeTask>* scheduler) { in ScheduleTask() argument 45 std::unique_ptr<FakeTask> task(new FakeTask(task_size)); in ScheduleTask()
|
D | adaptive_shared_batch_scheduler.h | 684 [task_size = task->size()] { in Schedule() 687 {{"batching_input_task_size", task_size}}); in Schedule()
|
D | shared_batch_scheduler_test.cc | 45 Status ScheduleTask(size_t task_size, BatchScheduler<FakeTask>* scheduler) { in ScheduleTask() argument 46 std::unique_ptr<FakeTask> task(new FakeTask(task_size)); in ScheduleTask()
|
/external/gemmlowp/meta/ |
D | multi_thread_transform.h | 37 const int task_size = Util::EstimateComputeCost(params.kernel); in PrepareTransform1DTasks() local 39 (task_size - kTransformTaskOverhead) / kMinTransformTaskSize; in PrepareTransform1DTasks()
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/task/ |
D | util.cc | 84 int task_size) { in GetRecommendedBlockSizeForConv() argument 86 task_size / static_cast<float>(gpu_info.GetComputeUnitsCount()); in GetRecommendedBlockSizeForConv()
|
D | util.h | 60 int task_size);
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/tasks/ |
D | conv_buffer_1x1.cc | 103 int task_size = shape.w * shape.b * shape.h * dst_depth; in GetBestParams() local 105 GetRecommendedBlockSizeForConv(gpu_info, definition.precision, task_size); in GetBestParams()
|
D | conv_powervr.cc | 1093 int task_size = dst_shape->w * dst_shape->b * dst_shape->h * dst_depth; in GuessBestParams() local 1095 static_cast<float>(task_size) / gpu_info.GetComputeUnitsCount(); in GuessBestParams() 1196 int task_size = dst_shape->w * dst_shape->b * dst_shape->h * dst_depth; in GuessBestParams() local 1198 gpu_info, definition.precision, task_size); in GuessBestParams()
|
/external/llvm-project/openmp/runtime/src/ |
D | kmp_tasking.cpp | 3961 size_t task_size; in __kmp_task_dup_alloc() local 3968 task_size = taskdata_src->td_size_alloc; in __kmp_task_dup_alloc() 3972 task_size)); in __kmp_task_dup_alloc() 3974 taskdata = (kmp_taskdata_t *)__kmp_fast_allocate(thread, task_size); in __kmp_task_dup_alloc() 3976 taskdata = (kmp_taskdata_t *)__kmp_thread_malloc(thread, task_size); in __kmp_task_dup_alloc() 3978 KMP_MEMCPY(taskdata, taskdata_src, task_size); in __kmp_task_dup_alloc()
|
/external/mesa3d/src/gallium/frontends/lavapipe/ |
D | lvp_device.c | 1033 …uint32_t task_size = sizeof(struct lvp_queue_work) + pSubmits[i].commandBufferCount * sizeof(struc… in lvp_QueueSubmit() local 1034 struct lvp_queue_work *task = malloc(task_size); in lvp_QueueSubmit()
|
/external/libabigail/tests/data/test-read-dwarf/ |
D | PR25007-sdhci.ko.abi | 3406 …<var-decl name='task_size' type-id='type-id-211' visibility='default' filepath='/ws/android/kernel…
|
/external/libabigail/tests/data/test-diff-filter/ |
D | test-PR27569-v1.abi | 1615 …<var-decl name="task_size" type-id="7359adad" visibility="default" filepath="include/linux/mm_type…
|
D | test-PR27569-v0.abi | 1615 …<var-decl name="task_size" type-id="7359adad" visibility="default" filepath="include/linux/mm_type…
|