/external/tensorflow/tensorflow/core/common_runtime/ |
D | base_collective_executor.cc | 45 int64 num_chunks) { in AlignedChunkElts() argument 46 DCHECK_GT(num_chunks, 0); in AlignedChunkElts() 47 int64 base_chunk_elts = (total_elts + (num_chunks - 1)) / num_chunks; in AlignedChunkElts() 57 << "total_elts=" << total_elts << " num_chunks=" << num_chunks in AlignedChunkElts() 69 << "total_elts=" << total_elts << " num_chunks=" << num_chunks in AlignedChunkElts() 81 CollectiveAdapterImpl(Tensor* output, int64 num_chunks, Allocator* allocator, in CollectiveAdapterImpl() argument 86 num_chunks_(num_chunks), in CollectiveAdapterImpl() 184 CollectiveAdapter* MakeCollectiveAdapter(Tensor* output, int num_chunks, in MakeCollectiveAdapter() argument 189 return new CollectiveAdapterImpl<float>(output, num_chunks, allocator, in MakeCollectiveAdapter() 193 return new CollectiveAdapterImpl<double>(output, num_chunks, allocator, in MakeCollectiveAdapter() [all …]
|
D | base_collective_executor.h | 77 int64 num_chunks); 87 CollectiveAdapter* MakeCollectiveAdapter(Tensor* output, int num_chunks,
|
/external/webrtc/webrtc/modules/audio_processing/vad/ |
D | voice_activity_detector_unittest.cc | 72 size_t num_chunks = 0; in TEST() local 79 ++num_chunks; in TEST() 82 mean_probability /= num_chunks; in TEST() 107 size_t num_chunks = 0; in TEST() local 114 ++num_chunks; in TEST() 117 mean_probability /= num_chunks; in TEST()
|
/external/perfetto/src/tracing/core/ |
D | shared_memory_abi_unittest.cc | 84 const size_t num_chunks = in TEST_P() local 88 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P() 112 (page_size() - sizeof(SharedMemoryABI::PageHeader)) / num_chunks; in TEST_P() 151 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P() 156 ASSERT_EQ(chunk_idx == num_chunks - 1, abi.is_page_complete(page_idx)); in TEST_P() 162 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P() 171 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P() 178 ASSERT_EQ(chunk_idx == num_chunks - 1, abi.is_page_free(page_idx)); in TEST_P()
|
D | shared_memory_abi.cc | 56 size_t num_chunks = SharedMemoryABI::kNumChunksForLayout[i]; in InitChunkSizes() local 57 size_t size = num_chunks == 0 ? 0 : GetChunkSize(page_size, num_chunks); in InitChunkSizes() 145 const size_t num_chunks = GetNumChunksForLayout(page_layout); in GetChunkUnchecked() local 146 PERFETTO_DCHECK(chunk_idx < num_chunks); in GetChunkUnchecked() 167 const size_t num_chunks = GetNumChunksForLayout(layout); in TryAcquireChunk() local 170 if (chunk_idx >= num_chunks) in TryAcquireChunk() 220 const uint32_t num_chunks = GetNumChunksForLayout(layout); in GetFreeChunks() local 222 for (uint32_t i = 0; i < num_chunks; i++) { in GetFreeChunks()
|
D | trace_writer_impl_unittest.cc | 83 size_t num_chunks = SharedMemoryABI::GetNumChunksForLayout(page_layout); in TEST_P() local 84 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P()
|
/external/webrtc/webrtc/modules/audio_processing/test/ |
D | audioproc_float.cc | 145 int num_chunks = 0; in main() local 147 trace_to_stderr.SetTimeSeconds(num_chunks * 1.f / kChunksPerSecond); in main() 148 ++num_chunks; in main() 157 exec_time_us * 1e-6, num_chunks * 1.f / kChunksPerSecond, in main() 158 exec_time_us * 1.f / num_chunks, 1.f * proc_time.max.Microseconds(), in main()
|
/external/tensorflow/tensorflow/core/debug/ |
D | debug_io_utils.cc | 62 const uint64 wall_time_us, const size_t num_chunks, in PrepareChunkEventProto() argument 84 metadata.set_num_chunks(num_chunks); in PrepareChunkEventProto() 156 const size_t num_chunks = cutoffs.size(); in WrapStringTensorAsEvents() local 158 for (size_t i = 0; i < num_chunks; ++i) { in WrapStringTensorAsEvents() 160 num_chunks, i, tensor_proto->dtype(), in WrapStringTensorAsEvents() 205 const size_t num_chunks = in WrapTensorAsEvents() local 209 for (size_t i = 0; i < num_chunks; ++i) { in WrapTensorAsEvents() 212 (i == num_chunks - 1) ? (total_length - pos) : chunk_size_ub; in WrapTensorAsEvents() 214 num_chunks, i, tensor_proto.dtype(), in WrapTensorAsEvents() 253 const size_t num_chunks = in PublishEncodedGraphDefInChunks() local [all …]
|
D | debugger_event_metadata.proto | 9 int32 num_chunks = 3; field
|
/external/libdrm/amdgpu/ |
D | amdgpu_cs.c | 214 cs.in.num_chunks = ibs_request->number_of_ibs; in amdgpu_cs_submit_one() 237 i = cs.in.num_chunks++; in amdgpu_cs_submit_one() 270 i = cs.in.num_chunks++; in amdgpu_cs_submit_one() 303 i = cs.in.num_chunks++; in amdgpu_cs_submit_one() 704 int num_chunks, in amdgpu_cs_submit_raw() argument 711 if (num_chunks == 0) in amdgpu_cs_submit_raw() 715 chunk_array = alloca(sizeof(uint64_t) * num_chunks); in amdgpu_cs_submit_raw() 716 for (i = 0; i < num_chunks; i++) in amdgpu_cs_submit_raw() 721 cs.in.num_chunks = num_chunks; in amdgpu_cs_submit_raw()
|
/external/perfetto/include/perfetto/tracing/core/ |
D | shared_memory_abi.h | 462 const uint32_t num_chunks = GetNumChunksForLayout(layout); in is_page_complete() local 463 if (num_chunks == 0) in is_page_complete() 466 (kAllChunksComplete & ((1 << (num_chunks * kChunkShift)) - 1)); in is_page_complete() 553 const uint32_t num_chunks = GetNumChunksForLayout(page_layout); in GetUsedChunks() local 555 for (uint32_t i = 0; i < num_chunks; i++) { in GetUsedChunks()
|
/external/mesa3d/src/gallium/winsys/amdgpu/drm/ |
D | amdgpu_cs.c | 1302 unsigned num_chunks = 0; in amdgpu_cs_submit_ib() local 1308 chunks[num_chunks].chunk_id = AMDGPU_CHUNK_ID_IB; in amdgpu_cs_submit_ib() 1309 chunks[num_chunks].length_dw = sizeof(struct drm_amdgpu_cs_chunk_ib) / 4; in amdgpu_cs_submit_ib() 1310 chunks[num_chunks].chunk_data = (uintptr_t)&cs->ib[IB_MAIN]; in amdgpu_cs_submit_ib() 1311 num_chunks++; in amdgpu_cs_submit_ib() 1315 chunks[num_chunks].chunk_id = AMDGPU_CHUNK_ID_FENCE; in amdgpu_cs_submit_ib() 1316 chunks[num_chunks].length_dw = sizeof(struct drm_amdgpu_cs_chunk_fence) / 4; in amdgpu_cs_submit_ib() 1317 chunks[num_chunks].chunk_data = (uintptr_t)&acs->fence_chunk; in amdgpu_cs_submit_ib() 1318 num_chunks++; in amdgpu_cs_submit_ib() 1343 chunks[num_chunks].chunk_id = AMDGPU_CHUNK_ID_DEPENDENCIES; in amdgpu_cs_submit_ib() [all …]
|
D | amdgpu_bo.c | 640 for (unsigned i = 0; i < backing->num_chunks; ++i) in sparse_dump() 664 for (unsigned idx = 0; idx < backing->num_chunks; ++idx) { in sparse_backing_alloc() 713 best_backing->num_chunks = 1; in sparse_backing_alloc() 730 sizeof(*best_backing->chunks) * (best_backing->num_chunks - best_idx - 1)); in sparse_backing_alloc() 731 best_backing->num_chunks--; in sparse_backing_alloc() 766 unsigned high = backing->num_chunks; in sparse_backing_free() 778 assert(low >= backing->num_chunks || end_page <= backing->chunks[low].begin); in sparse_backing_free() 784 if (low < backing->num_chunks && end_page == backing->chunks[low].begin) { in sparse_backing_free() 787 sizeof(*backing->chunks) * (backing->num_chunks - low - 1)); in sparse_backing_free() 788 backing->num_chunks--; in sparse_backing_free() [all …]
|
/external/autotest/client/tests/ebizzy/ |
D | ebizzy.py | 23 def run_once(self, args='', num_chunks=1000, chunk_size=512000, argument 29 args2 = '-m -n %s -P -R -s %s -S %s -t %s' % (num_chunks, chunk_size,
|
/external/tensorflow/tensorflow/python/debug/lib/ |
D | grpc_debug_server.py | 252 num_chunks = debugger_plugin_metadata["numChunks"] 255 if num_chunks <= 1: 263 tensor_chunks[tensor_key] = [None] * num_chunks 307 num_chunks = int(event.graph_def[index_bar_1 + 1 : index_bar_2]) 309 graph_def_chunks[graph_def_hash_device_timestamp] = [None] * num_chunks
|
/external/mesa3d/src/intel/vulkan/tests/ |
D | state_pool_test_helper.h | 39 const unsigned num_chunks = STATES_PER_THREAD / chunk_size; in alloc_states() local 45 for (unsigned c = 0; c < num_chunks; c++) { in alloc_states()
|
/external/tensorflow/tensorflow/python/distribute/ |
D | all_reduce.py | 325 num_chunks = num_devices * num_subchunks 332 splits, split_pad_len = _padded_split(input_tensors[d], num_chunks) 337 new_partial_reductions = [None for _ in range(0, num_chunks)] 395 num_chunks = len(chunks_by_dev[0]) 396 if 0 != num_chunks % num_devices: 399 num_subchunks = int(num_chunks / num_devices) 402 passed_values = [None for _ in range(0, num_chunks)]
|
D | all_reduce_test.py | 121 num_chunks = 2 * len(input_tensors) 124 self.assertEqual(len(otl), num_chunks) 126 self.assertEqual(ot.shape, [tlen/num_chunks])
|
/external/pdfium/third_party/libpng16/ |
D | pngset.c | 1327 unsigned int num_chunks, old_num_chunks; in png_set_keep_unknown_chunks() local 1374 num_chunks = (unsigned int)/*SAFE*/(sizeof chunks_to_ignore)/5U; in png_set_keep_unknown_chunks() 1389 num_chunks = num_chunks_in; in png_set_keep_unknown_chunks() 1398 if (num_chunks + old_num_chunks > UINT_MAX/5) in png_set_keep_unknown_chunks() 1412 5 * (num_chunks + old_num_chunks))); in png_set_keep_unknown_chunks() 1435 for (i=0; i<num_chunks; ++i) in png_set_keep_unknown_chunks() 1442 num_chunks = 0; in png_set_keep_unknown_chunks() 1450 ++num_chunks; in png_set_keep_unknown_chunks() 1455 if (num_chunks == 0) in png_set_keep_unknown_chunks() 1465 num_chunks = 0; in png_set_keep_unknown_chunks() [all …]
|
/external/libpng/ |
D | pngset.c | 1376 unsigned int num_chunks, old_num_chunks; in png_set_keep_unknown_chunks() local 1424 num_chunks = (unsigned int)/*SAFE*/(sizeof chunks_to_ignore)/5U; in png_set_keep_unknown_chunks() 1439 num_chunks = (unsigned int)num_chunks_in; in png_set_keep_unknown_chunks() 1448 if (num_chunks + old_num_chunks > UINT_MAX/5) in png_set_keep_unknown_chunks() 1462 5 * (num_chunks + old_num_chunks))); in png_set_keep_unknown_chunks() 1485 for (i=0; i<num_chunks; ++i) in png_set_keep_unknown_chunks() 1492 num_chunks = 0; in png_set_keep_unknown_chunks() 1500 ++num_chunks; in png_set_keep_unknown_chunks() 1505 if (num_chunks == 0) in png_set_keep_unknown_chunks() 1515 num_chunks = 0; in png_set_keep_unknown_chunks() [all …]
|
/external/mesa3d/src/gallium/winsys/radeon/drm/ |
D | radeon_drm_cs.c | 637 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush() 646 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush() 652 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush() 660 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush() 664 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush() 668 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush() 672 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush()
|
/external/tensorflow/tensorflow/python/keras/saving/ |
D | hdf5_format.py | 857 num_chunks = 1 858 chunked_data = np.array_split(data_npy, num_chunks) 862 num_chunks += 1 863 chunked_data = np.array_split(data_npy, num_chunks) 865 if num_chunks > 1:
|
/external/u-boot/drivers/spi/ |
D | fsl_espi.c | 253 int num_blks, num_chunks, max_tran_len, tran_len; in spi_xfer() local 303 num_chunks = DIV_ROUND_UP(data_len, max_tran_len); in spi_xfer() 304 while (num_chunks--) { in spi_xfer()
|
D | atcspi200_spi.c | 196 int num_blks, num_chunks, max_tran_len, tran_len; in __atcspi200_spi_xfer() local 236 num_chunks = DIV_ROUND_UP(data_len, max_tran_len); in __atcspi200_spi_xfer() 239 while (num_chunks--) { in __atcspi200_spi_xfer()
|
/external/bsdiff/ |
D | split_patch_writer_unittest.cc | 19 void SetUpForSize(size_t num_chunks, in SetUpForSize() argument 22 fake_patches_.resize(num_chunks); in SetUpForSize()
|