/external/tensorflow/tensorflow/core/kernels/batching_util/ |
D | batch_scheduler.h | 78 class Batch { 80 Batch() = default; 81 virtual ~Batch(); // Blocks until the batch is closed. 128 TF_DISALLOW_COPY_AND_ASSIGN(Batch); 191 Batch<TaskType>::~Batch() { in ~Batch() 196 void Batch<TaskType>::AddTask(std::unique_ptr<TaskType> task) { in AddTask() 206 std::unique_ptr<TaskType> Batch<TaskType>::RemoveTask() { in RemoveTask() 220 int Batch<TaskType>::num_tasks() const { in num_tasks() 228 bool Batch<TaskType>::empty() const { in empty() 236 const TaskType& Batch<TaskType>::task(int i) const { in task() [all …]
|
D | shared_batch_scheduler.h | 163 std::function<void(std::unique_ptr<Batch<TaskType>>)> 231 std::function<void(std::unique_ptr<Batch<TaskType>>)>; 259 std::unique_ptr<Batch<TaskType>> ScheduleBatch(); 262 void ProcessBatch(std::unique_ptr<Batch<TaskType>> batch); 309 std::deque<std::unique_ptr<Batch<TaskType>>> batches_ GUARDED_BY(mu_); 394 std::function<void(std::unique_ptr<Batch<TaskType>>)> in AddQueue() 452 std::unique_ptr<Batch<TaskType>> batch_to_process; in ThreadLogic() 515 batches_.emplace_back(new Batch<TaskType>); in Queue() 591 std::unique_ptr<Batch<TaskType>> Queue<TaskType>::ScheduleBatch() { in ScheduleBatch() 594 std::unique_ptr<Batch<TaskType>> batch_to_schedule; in ScheduleBatch() [all …]
|
D | shared_batch_scheduler_test.cc | 74 [&queue_0_callback_called](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 84 [&queue_1_callback_called](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 138 &callback_data](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 201 &third_batch_processed](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 260 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 303 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 351 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 363 &queue_1_first_batch_proceed](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 414 &proceed](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 469 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() [all …]
|
D | batch_scheduler_test.cc | 41 Batch<FakeTask> batch; in TEST() 84 Batch<FakeTask> batch; in TEST() 98 Batch<FakeTask>* batch = new Batch<FakeTask>; in TEST()
|
D | basic_batch_scheduler_benchmark_test.cc | 126 void ProcessBatch(std::unique_ptr<Batch<BenchmarkBatchTask>> batch); 168 [this](std::unique_ptr<Batch<BenchmarkBatchTask>> batch) { in ResetState() 176 std::unique_ptr<Batch<BenchmarkBatchTask>> batch) { in ProcessBatch() 204 void ProcessBatch(std::unique_ptr<Batch<BenchmarkBatchTask>> batch); 297 [this](std::unique_ptr<Batch<BenchmarkBatchTask>> batch) { in ResetState() 311 std::unique_ptr<Batch<BenchmarkBatchTask>> batch) { in ProcessBatch()
|
D | adaptive_shared_batch_scheduler_test.cc | 105 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 143 auto queue_callback = [&env](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 210 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 274 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 314 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 350 std::unique_ptr<Batch<FakeTask>> batch) { in TEST()
|
D | serial_device_batch_scheduler_test.cc | 102 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 151 &scheduler](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 211 [&mu, &processed_batches](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 269 std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 320 &finish_processing](std::unique_ptr<Batch<FakeTask>> batch) { in TEST() 357 std::unique_ptr<Batch<FakeTask>> batch) { in TEST()
|
D | basic_batch_scheduler.h | 185 std::function<void(std::unique_ptr<Batch<TaskType>>)> 216 std::function<void(std::unique_ptr<Batch<TaskType>>)> in Create()
|
D | serial_device_batch_scheduler.h | 113 using BatchProcessor = std::function<void(std::unique_ptr<Batch<TaskType>>)>; 241 class SDBSBatch : public Batch<TaskType> { 409 callback(std::unique_ptr<Batch<TaskType>>( in ProcessBatches()
|
D | adaptive_shared_batch_scheduler.h | 134 using BatchProcessor = std::function<void(std::unique_ptr<Batch<TaskType>>)>; 272 class ASBSBatch : public Batch<TaskType> { 470 callback(std::unique_ptr<Batch<TaskType>>( in CallbackWrapper()
|
D | basic_batch_scheduler_test.cc | 60 auto callback = [&callback_called](std::unique_ptr<Batch<FakeTask>> batch) { in TEST()
|
/external/skqp/src/gpu/ccpr/ |
D | GrCCStroker.h | 85 struct Batch { struct 86 Batch(GrTAllocator<InstanceTallies>* alloc, const InstanceTallies& startNonScissorIndices, in Batch() argument 96 void appendStrokeMeshesToBuffers(int numSegmentsLog2, const Batch&, argument 104 const GrCCCoverageProcessor&, const Batch&, 110 SkSTArray<32, Batch> fBatches;
|
D | GrCCFiller.h | 85 struct Batch { struct 104 SkSTArray<32, Batch, true> fBatches; argument
|
/external/skia/src/gpu/ccpr/ |
D | GrCCStroker.h | 85 struct Batch { struct 86 Batch(GrTAllocator<InstanceTallies>* alloc, const InstanceTallies& startNonScissorIndices, in Batch() argument 96 void appendStrokeMeshesToBuffers(int numSegmentsLog2, const Batch&, argument 104 const GrCCCoverageProcessor&, const Batch&, 110 SkSTArray<32, Batch> fBatches;
|
D | GrCCFiller.h | 85 struct Batch { struct 104 SkSTArray<32, Batch, true> fBatches; argument
|
/external/tensorflow/tensorflow/core/api_def/base_api/ |
D | api_def_Unbatch.pbtxt | 3 summary: "Reverses the operation of Batch for a single output Tensor." 11 batched_tensor: The possibly transformed output of Batch. The size of the first 14 batch_index: The matching batch_index obtained from Batch. 15 id: The id scalar emitted by Batch.
|
D | api_def_Batch.pbtxt | 2 graph_op_name: "Batch" 25 max_batch_size: Batch sizes will never be bigger than this. 36 id: always contains a scalar with a unique ID for this invocation of Batch.
|
D | api_def_UnbatchGrad.pbtxt | 5 Acts like Batch but using the given batch_index index of batching things as they 13 id: The id scalar emitted by Batch.
|
D | api_def_RecvTPUEmbeddingActivations.pbtxt | 27 the arguments to TPUEmbeddingEnqueue(Integer/Sparse/SparseTensor)Batch. The
|
D | api_def_BatchNormWithGlobalNormalization.pbtxt | 53 summary: "Batch normalization."
|
D | api_def_FusedBatchNorm.pbtxt | 94 summary: "Batch normalization."
|
/external/compiler-rt/lib/sanitizer_common/ |
D | sanitizer_allocator.h | 320 typedef typename SizeClassMap::TransferBatch Batch; typedef 353 NOINLINE Batch* AllocateBatch(AllocatorStats *stat, AllocatorCache *c, in AllocateBatch() 357 Batch *b = region->free_list.Pop(); in AllocateBatch() 364 NOINLINE void DeallocateBatch(AllocatorStats *stat, uptr class_id, Batch *b) { in DeallocateBatch() 513 LFStack<Batch> free_list; 538 NOINLINE Batch* PopulateFreeList(AllocatorStats *stat, AllocatorCache *c, in PopulateFreeList() 541 Batch *b = region->free_list.Pop(); in PopulateFreeList() 581 b = (Batch*)c->Allocate(this, SizeClassMap::ClassID(sizeof(Batch))); in PopulateFreeList() 583 b = (Batch*)(region_beg + beg_idx); in PopulateFreeList() 713 typedef typename SizeClassMap::TransferBatch Batch; typedef [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | batch_kernels.cc | 296 using Batch = serving::Batch<BatchTask>; typedef in tensorflow::BatchResource 300 static Status ValidateBatch(const Batch& batch) { in ValidateBatch() 330 Status ConcatInputTensors(const Batch& batch, OpKernelContext* context, in ConcatInputTensors() 408 Batch* batch) const { in SplitOutputTensors() 473 void ProcessFuncBatch(std::unique_ptr<Batch> batch) const { in ProcessFuncBatch() 556 void ProcessBatch(std::unique_ptr<Batch> batch) const { in ProcessBatch() 631 static Status EmitIndexTensor(OpKernelContext* context, const Batch& batch, in EmitIndexTensor() 662 auto process_batch_callback = [this](std::unique_ptr<Batch> batch) { in LookupOrCreateBatcherQueue() 1167 .emplace(batch_key, Batch{missing_tensors, context, done}) in Compute() 1215 struct Batch { struct in tensorflow::UnbatchGradResource [all …]
|
/external/tensorflow/tensorflow/core/api_def/java_api/ |
D | api_def_Batch.pbtxt | 2 graph_op_name: "Batch"
|
/external/libcxx/benchmarks/ |
D | algorithms.bench.cpp | 120 const size_t Batch = CountElements ? Copies.size() * Quantity : Copies.size(); in runOpOnCopies() local 121 while (state.KeepRunningBatch(Batch)) { in runOpOnCopies()
|