Home
last modified time | relevance | path

Searched refs:num_chunks (Results 1 – 25 of 44) sorted by relevance

12

/external/tensorflow/tensorflow/core/common_runtime/
Dbase_collective_executor.cc45 int64 num_chunks) { in AlignedChunkElts() argument
46 DCHECK_GT(num_chunks, 0); in AlignedChunkElts()
47 int64 base_chunk_elts = (total_elts + (num_chunks - 1)) / num_chunks; in AlignedChunkElts()
57 << "total_elts=" << total_elts << " num_chunks=" << num_chunks in AlignedChunkElts()
69 << "total_elts=" << total_elts << " num_chunks=" << num_chunks in AlignedChunkElts()
81 CollectiveAdapterImpl(Tensor* output, int64 num_chunks, Allocator* allocator, in CollectiveAdapterImpl() argument
86 num_chunks_(num_chunks), in CollectiveAdapterImpl()
184 CollectiveAdapter* MakeCollectiveAdapter(Tensor* output, int num_chunks, in MakeCollectiveAdapter() argument
189 return new CollectiveAdapterImpl<float>(output, num_chunks, allocator, in MakeCollectiveAdapter()
193 return new CollectiveAdapterImpl<double>(output, num_chunks, allocator, in MakeCollectiveAdapter()
[all …]
Dbase_collective_executor.h77 int64 num_chunks);
87 CollectiveAdapter* MakeCollectiveAdapter(Tensor* output, int num_chunks,
/external/webrtc/webrtc/modules/audio_processing/vad/
Dvoice_activity_detector_unittest.cc72 size_t num_chunks = 0; in TEST() local
79 ++num_chunks; in TEST()
82 mean_probability /= num_chunks; in TEST()
107 size_t num_chunks = 0; in TEST() local
114 ++num_chunks; in TEST()
117 mean_probability /= num_chunks; in TEST()
/external/perfetto/src/tracing/core/
Dshared_memory_abi_unittest.cc84 const size_t num_chunks = in TEST_P() local
88 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P()
112 (page_size() - sizeof(SharedMemoryABI::PageHeader)) / num_chunks; in TEST_P()
151 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P()
156 ASSERT_EQ(chunk_idx == num_chunks - 1, abi.is_page_complete(page_idx)); in TEST_P()
162 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P()
171 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P()
178 ASSERT_EQ(chunk_idx == num_chunks - 1, abi.is_page_free(page_idx)); in TEST_P()
Dshared_memory_abi.cc56 size_t num_chunks = SharedMemoryABI::kNumChunksForLayout[i]; in InitChunkSizes() local
57 size_t size = num_chunks == 0 ? 0 : GetChunkSize(page_size, num_chunks); in InitChunkSizes()
145 const size_t num_chunks = GetNumChunksForLayout(page_layout); in GetChunkUnchecked() local
146 PERFETTO_DCHECK(chunk_idx < num_chunks); in GetChunkUnchecked()
167 const size_t num_chunks = GetNumChunksForLayout(layout); in TryAcquireChunk() local
170 if (chunk_idx >= num_chunks) in TryAcquireChunk()
220 const uint32_t num_chunks = GetNumChunksForLayout(layout); in GetFreeChunks() local
222 for (uint32_t i = 0; i < num_chunks; i++) { in GetFreeChunks()
Dtrace_writer_impl_unittest.cc83 size_t num_chunks = SharedMemoryABI::GetNumChunksForLayout(page_layout); in TEST_P() local
84 for (size_t chunk_idx = 0; chunk_idx < num_chunks; chunk_idx++) { in TEST_P()
/external/webrtc/webrtc/modules/audio_processing/test/
Daudioproc_float.cc145 int num_chunks = 0; in main() local
147 trace_to_stderr.SetTimeSeconds(num_chunks * 1.f / kChunksPerSecond); in main()
148 ++num_chunks; in main()
157 exec_time_us * 1e-6, num_chunks * 1.f / kChunksPerSecond, in main()
158 exec_time_us * 1.f / num_chunks, 1.f * proc_time.max.Microseconds(), in main()
/external/tensorflow/tensorflow/core/debug/
Ddebug_io_utils.cc62 const uint64 wall_time_us, const size_t num_chunks, in PrepareChunkEventProto() argument
84 metadata.set_num_chunks(num_chunks); in PrepareChunkEventProto()
156 const size_t num_chunks = cutoffs.size(); in WrapStringTensorAsEvents() local
158 for (size_t i = 0; i < num_chunks; ++i) { in WrapStringTensorAsEvents()
160 num_chunks, i, tensor_proto->dtype(), in WrapStringTensorAsEvents()
205 const size_t num_chunks = in WrapTensorAsEvents() local
209 for (size_t i = 0; i < num_chunks; ++i) { in WrapTensorAsEvents()
212 (i == num_chunks - 1) ? (total_length - pos) : chunk_size_ub; in WrapTensorAsEvents()
214 num_chunks, i, tensor_proto.dtype(), in WrapTensorAsEvents()
253 const size_t num_chunks = in PublishEncodedGraphDefInChunks() local
[all …]
Ddebugger_event_metadata.proto9 int32 num_chunks = 3; field
/external/libdrm/amdgpu/
Damdgpu_cs.c214 cs.in.num_chunks = ibs_request->number_of_ibs; in amdgpu_cs_submit_one()
237 i = cs.in.num_chunks++; in amdgpu_cs_submit_one()
270 i = cs.in.num_chunks++; in amdgpu_cs_submit_one()
303 i = cs.in.num_chunks++; in amdgpu_cs_submit_one()
704 int num_chunks, in amdgpu_cs_submit_raw() argument
711 if (num_chunks == 0) in amdgpu_cs_submit_raw()
715 chunk_array = alloca(sizeof(uint64_t) * num_chunks); in amdgpu_cs_submit_raw()
716 for (i = 0; i < num_chunks; i++) in amdgpu_cs_submit_raw()
721 cs.in.num_chunks = num_chunks; in amdgpu_cs_submit_raw()
/external/perfetto/include/perfetto/tracing/core/
Dshared_memory_abi.h462 const uint32_t num_chunks = GetNumChunksForLayout(layout); in is_page_complete() local
463 if (num_chunks == 0) in is_page_complete()
466 (kAllChunksComplete & ((1 << (num_chunks * kChunkShift)) - 1)); in is_page_complete()
553 const uint32_t num_chunks = GetNumChunksForLayout(page_layout); in GetUsedChunks() local
555 for (uint32_t i = 0; i < num_chunks; i++) { in GetUsedChunks()
/external/mesa3d/src/gallium/winsys/amdgpu/drm/
Damdgpu_cs.c1302 unsigned num_chunks = 0; in amdgpu_cs_submit_ib() local
1308 chunks[num_chunks].chunk_id = AMDGPU_CHUNK_ID_IB; in amdgpu_cs_submit_ib()
1309 chunks[num_chunks].length_dw = sizeof(struct drm_amdgpu_cs_chunk_ib) / 4; in amdgpu_cs_submit_ib()
1310 chunks[num_chunks].chunk_data = (uintptr_t)&cs->ib[IB_MAIN]; in amdgpu_cs_submit_ib()
1311 num_chunks++; in amdgpu_cs_submit_ib()
1315 chunks[num_chunks].chunk_id = AMDGPU_CHUNK_ID_FENCE; in amdgpu_cs_submit_ib()
1316 chunks[num_chunks].length_dw = sizeof(struct drm_amdgpu_cs_chunk_fence) / 4; in amdgpu_cs_submit_ib()
1317 chunks[num_chunks].chunk_data = (uintptr_t)&acs->fence_chunk; in amdgpu_cs_submit_ib()
1318 num_chunks++; in amdgpu_cs_submit_ib()
1343 chunks[num_chunks].chunk_id = AMDGPU_CHUNK_ID_DEPENDENCIES; in amdgpu_cs_submit_ib()
[all …]
Damdgpu_bo.c640 for (unsigned i = 0; i < backing->num_chunks; ++i) in sparse_dump()
664 for (unsigned idx = 0; idx < backing->num_chunks; ++idx) { in sparse_backing_alloc()
713 best_backing->num_chunks = 1; in sparse_backing_alloc()
730 sizeof(*best_backing->chunks) * (best_backing->num_chunks - best_idx - 1)); in sparse_backing_alloc()
731 best_backing->num_chunks--; in sparse_backing_alloc()
766 unsigned high = backing->num_chunks; in sparse_backing_free()
778 assert(low >= backing->num_chunks || end_page <= backing->chunks[low].begin); in sparse_backing_free()
784 if (low < backing->num_chunks && end_page == backing->chunks[low].begin) { in sparse_backing_free()
787 sizeof(*backing->chunks) * (backing->num_chunks - low - 1)); in sparse_backing_free()
788 backing->num_chunks--; in sparse_backing_free()
[all …]
/external/autotest/client/tests/ebizzy/
Debizzy.py23 def run_once(self, args='', num_chunks=1000, chunk_size=512000, argument
29 args2 = '-m -n %s -P -R -s %s -S %s -t %s' % (num_chunks, chunk_size,
/external/tensorflow/tensorflow/python/debug/lib/
Dgrpc_debug_server.py252 num_chunks = debugger_plugin_metadata["numChunks"]
255 if num_chunks <= 1:
263 tensor_chunks[tensor_key] = [None] * num_chunks
307 num_chunks = int(event.graph_def[index_bar_1 + 1 : index_bar_2])
309 graph_def_chunks[graph_def_hash_device_timestamp] = [None] * num_chunks
/external/mesa3d/src/intel/vulkan/tests/
Dstate_pool_test_helper.h39 const unsigned num_chunks = STATES_PER_THREAD / chunk_size; in alloc_states() local
45 for (unsigned c = 0; c < num_chunks; c++) { in alloc_states()
/external/tensorflow/tensorflow/python/distribute/
Dall_reduce.py325 num_chunks = num_devices * num_subchunks
332 splits, split_pad_len = _padded_split(input_tensors[d], num_chunks)
337 new_partial_reductions = [None for _ in range(0, num_chunks)]
395 num_chunks = len(chunks_by_dev[0])
396 if 0 != num_chunks % num_devices:
399 num_subchunks = int(num_chunks / num_devices)
402 passed_values = [None for _ in range(0, num_chunks)]
Dall_reduce_test.py121 num_chunks = 2 * len(input_tensors)
124 self.assertEqual(len(otl), num_chunks)
126 self.assertEqual(ot.shape, [tlen/num_chunks])
/external/pdfium/third_party/libpng16/
Dpngset.c1327 unsigned int num_chunks, old_num_chunks; in png_set_keep_unknown_chunks() local
1374 num_chunks = (unsigned int)/*SAFE*/(sizeof chunks_to_ignore)/5U; in png_set_keep_unknown_chunks()
1389 num_chunks = num_chunks_in; in png_set_keep_unknown_chunks()
1398 if (num_chunks + old_num_chunks > UINT_MAX/5) in png_set_keep_unknown_chunks()
1412 5 * (num_chunks + old_num_chunks))); in png_set_keep_unknown_chunks()
1435 for (i=0; i<num_chunks; ++i) in png_set_keep_unknown_chunks()
1442 num_chunks = 0; in png_set_keep_unknown_chunks()
1450 ++num_chunks; in png_set_keep_unknown_chunks()
1455 if (num_chunks == 0) in png_set_keep_unknown_chunks()
1465 num_chunks = 0; in png_set_keep_unknown_chunks()
[all …]
/external/libpng/
Dpngset.c1376 unsigned int num_chunks, old_num_chunks; in png_set_keep_unknown_chunks() local
1424 num_chunks = (unsigned int)/*SAFE*/(sizeof chunks_to_ignore)/5U; in png_set_keep_unknown_chunks()
1439 num_chunks = (unsigned int)num_chunks_in; in png_set_keep_unknown_chunks()
1448 if (num_chunks + old_num_chunks > UINT_MAX/5) in png_set_keep_unknown_chunks()
1462 5 * (num_chunks + old_num_chunks))); in png_set_keep_unknown_chunks()
1485 for (i=0; i<num_chunks; ++i) in png_set_keep_unknown_chunks()
1492 num_chunks = 0; in png_set_keep_unknown_chunks()
1500 ++num_chunks; in png_set_keep_unknown_chunks()
1505 if (num_chunks == 0) in png_set_keep_unknown_chunks()
1515 num_chunks = 0; in png_set_keep_unknown_chunks()
[all …]
/external/mesa3d/src/gallium/winsys/radeon/drm/
Dradeon_drm_cs.c637 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush()
646 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush()
652 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush()
660 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush()
664 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush()
668 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush()
672 cs->cst->cs.num_chunks = 3; in radeon_drm_cs_flush()
/external/tensorflow/tensorflow/python/keras/saving/
Dhdf5_format.py857 num_chunks = 1
858 chunked_data = np.array_split(data_npy, num_chunks)
862 num_chunks += 1
863 chunked_data = np.array_split(data_npy, num_chunks)
865 if num_chunks > 1:
/external/u-boot/drivers/spi/
Dfsl_espi.c253 int num_blks, num_chunks, max_tran_len, tran_len; in spi_xfer() local
303 num_chunks = DIV_ROUND_UP(data_len, max_tran_len); in spi_xfer()
304 while (num_chunks--) { in spi_xfer()
Datcspi200_spi.c196 int num_blks, num_chunks, max_tran_len, tran_len; in __atcspi200_spi_xfer() local
236 num_chunks = DIV_ROUND_UP(data_len, max_tran_len); in __atcspi200_spi_xfer()
239 while (num_chunks--) { in __atcspi200_spi_xfer()
/external/bsdiff/
Dsplit_patch_writer_unittest.cc19 void SetUpForSize(size_t num_chunks, in SetUpForSize() argument
22 fake_patches_.resize(num_chunks); in SetUpForSize()

12