Home
last modified time | relevance | path

Searched refs:block_count (Results 1 – 25 of 155) sorted by relevance

1234567

/external/tensorflow/tensorflow/core/util/
Dgpu_launch_config.h119 int block_count = -1; member
136 const int block_count = in GetGpuLaunchConfig() local
142 config.block_count = block_count; in GetGpuLaunchConfig()
162 int block_count = 0; in GetGpuLaunchConfig() local
167 &block_count, &thread_per_block, func, dynamic_shared_memory_size, in GetGpuLaunchConfig()
172 &block_count, &thread_per_block, func, dynamic_shared_memory_size, in GetGpuLaunchConfig()
177 block_count = in GetGpuLaunchConfig()
178 std::min(block_count, DivUp(work_element_count, thread_per_block)); in GetGpuLaunchConfig()
182 config.block_count = block_count; in GetGpuLaunchConfig()
197 int block_count = 0; in GetGpuLaunchConfigFixedBlockSize() local
[all …]
Dgpu_kernel_helper_test.cu.cc184 TF_CHECK_OK(GpuLaunchKernel(SetOutbufZero, cfg.block_count, \ in TEST_F()
189 TF_CHECK_OK(GpuLaunchKernel(Count1D, cfg.block_count, cfg.thread_per_block, \ in TEST_F()
197 TF_CHECK_OK(GpuLaunchKernel(SetOutbufZero, cfg.block_count, \ in TEST_F()
202 TF_CHECK_OK(GpuLaunchKernel(Count1D, cfg.block_count, cfg.thread_per_block, \ in TEST_F()
226 a.block_count.x == b.block_count.x && in operator ==()
227 a.block_count.y == b.block_count.y && in operator ==()
228 a.block_count.z == b.block_count.z && in operator ==()
241 TF_EXPECT_OK(GpuLaunchKernel(SetOutbufZero, cfg1d.block_count, \ in TEST_F()
246 TF_EXPECT_OK(GpuLaunchKernel(Count2D, cfg.block_count, cfg.thread_per_block, \ in TEST_F()
254 TF_EXPECT_OK(GpuLaunchKernel(SetOutbufZero, cfg1d.block_count, \ in TEST_F()
[all …]
/external/tensorflow/tensorflow/compiler/xla/service/gpu/
Dlaunch_dimensions.cc131 int64_t block_count = in CalculateLaunchDimensions() local
140 if (capped_block_count < block_count) { in CalculateLaunchDimensions()
142 block_count = capped_block_count; in CalculateLaunchDimensions()
143 VLOG(2) << "Update the # of blocks to " << block_count in CalculateLaunchDimensions()
151 int64_t capped_block_count = block_count; in CalculateLaunchDimensions()
161 if (capped_block_count < block_count) { in CalculateLaunchDimensions()
163 block_count = capped_block_count; in CalculateLaunchDimensions()
167 block_count >= gpu_device_info.block_dim_limit_x) { in CalculateLaunchDimensions()
169 "Kernel launch needs more blocks (", block_count, in CalculateLaunchDimensions()
177 block_count, threads_per_block_x, threads_per_block_y, num_elements); in CalculateLaunchDimensions()
[all …]
/external/cronet/net/disk_cache/blockfile/
Dblock_files.cc185 bool BlockHeader::NeedToGrowBlockFile(int block_count) const { in NeedToGrowBlockFile()
190 if (i >= block_count - 1 && header_->empty[i]) in NeedToGrowBlockFile()
203 bool BlockHeader::CanAllocate(int block_count) const { in CanAllocate()
204 DCHECK_GT(block_count, 0); in CanAllocate()
205 for (int i = block_count - 1; i < kMaxNumBlocks; i++) { in CanAllocate()
311 bool BlockFiles::CreateBlock(FileType block_type, int block_count, in CreateBlock() argument
318 if (block_count < 1 || block_count > kMaxNumBlocks) in CreateBlock()
324 MappedFile* file = FileForNewBlock(block_type, block_count); in CreateBlock()
332 if (!file_header.CreateMapBlock(block_count, &index)) in CreateBlock()
335 Addr address(block_type, block_count, file_header.FileId(), index); in CreateBlock()
[all …]
Dblock_files.h59 bool NeedToGrowBlockFile(int block_count) const;
63 bool CanAllocate(int block_count) const;
113 bool CreateBlock(FileType block_type, int block_count, Addr* block_address);
137 MappedFile* FileForNewBlock(FileType block_type, int block_count);
/external/skia/tests/
DSkBlockAllocatorTest.cpp31 static int block_count(const SkSBlockAllocator<N>& pool) { in block_count() function
338 REPORTER_ASSERT(r, block_count(pool) == 4); // 3 added plus the implicit head in DEF_TEST()
347 REPORTER_ASSERT(r,block_count(pool) == 1); in DEF_TEST()
381 REPORTER_ASSERT(r, block_count(pool) == 3); in DEF_TEST()
386 REPORTER_ASSERT(r, block_count(pool) == 2); in DEF_TEST()
391 REPORTER_ASSERT(r, block_count(pool) == 1); in DEF_TEST()
404 REPORTER_ASSERT(r, block_count(pool) == 1); in DEF_TEST()
413 REPORTER_ASSERT(r, block_count(pool) == 3); in DEF_TEST()
435 REPORTER_ASSERT(r, block_count(pool) == 1); in DEF_TEST()
452 REPORTER_ASSERT(r, block_count(pool) == 1); in DEF_TEST()
[all …]
/external/tensorflow/tensorflow/core/kernels/
Dbias_op_gpu.cu.cc90 TF_CHECK_OK(GpuLaunchKernel(BiasNHWCKernel<T>, config.block_count, in compute()
97 TF_CHECK_OK(GpuLaunchKernel(BiasNCHWKernel<T>, config.block_count, in compute()
237 config.block_count, config.thread_per_block, in compute()
242 int group_size = (config.block_count + bias_size - 1) / bias_size; in compute()
243 config.block_count = group_size * bias_size; in compute()
248 config.block_count, config.thread_per_block, in compute()
258 BiasGradNHWC_Naive<T>, config.block_count, config.thread_per_block, 0, in compute()
261 TF_CHECK_OK(GpuLaunchKernel(BiasGradNCHW_Naive<T>, config.block_count, in compute()
Ddepthtospace_op_gpu.cu.cc165 D2S_NHWC<T>, config.block_count, config.thread_per_block, 0, d.stream(), in operator ()()
201 D2S_NCHW_LOOP<T, 2>, config.block_count, config.thread_per_block, in operator ()()
208 D2S_NCHW_LOOP<T, 3>, config.block_count, config.thread_per_block, in operator ()()
215 D2S_NCHW_LOOP<T, 4>, config.block_count, config.thread_per_block, in operator ()()
230 D2S_NCHW<T>, config.block_count, config.thread_per_block, 0, d.stream(), in operator ()()
Dspacetodepth_op_gpu.cu.cc162 S2D_NHWC<T>, config.block_count, config.thread_per_block, 0, d.stream(), in operator ()()
198 S2D_NCHW_LOOP<T, 2>, config.block_count, config.thread_per_block, in operator ()()
205 S2D_NCHW_LOOP<T, 3>, config.block_count, config.thread_per_block, in operator ()()
212 S2D_NCHW_LOOP<T, 4>, config.block_count, config.thread_per_block, in operator ()()
227 S2D_NCHW<T>, config.block_count, config.thread_per_block, 0, d.stream(), in operator ()()
Dinplace_ops_functor_gpu.cu.cc54 DoParallelConcatOpKernel<T>, cfg.block_count, cfg.thread_per_block, 0, in DoParallelConcatUpdate()
124 cfg.block_count, cfg.thread_per_block, 0, in DoInplaceOp()
129 TF_CHECK_OK(GpuLaunchKernel(DoInplaceOpKernel<T, I_ADD>, cfg.block_count, in DoInplaceOp()
135 TF_CHECK_OK(GpuLaunchKernel(DoInplaceOpKernel<T, I_SUB>, cfg.block_count, in DoInplaceOp()
158 cfg.block_count, cfg.thread_per_block, 0, in DoInplaceOp()
Dconv_2d_gpu.h498 config.block_count, config.thread_per_block,
504 config.block_count, config.thread_per_block,
535 config.block_count, config.thread_per_block,
549 config.block_count, config.thread_per_block,
586 PadInputCustomKernelNHWC<T, NDIMS>, config.block_count,
592 PadInputCustomKernelNCHW<T, NDIMS>, config.block_count,
1037 config.block_count, config.thread_per_block, 0,
1085 int block_count;
1087 block_count = (total_size + config.thread_per_block - 1) /
1090 block_count = config.block_count;
[all …]
Ddilation_ops_gpu.cu.cc200 DilationKernel<T>, config.block_count, config.thread_per_block, 0, in operator ()()
233 TF_CHECK_OK(GpuLaunchKernel(SetZero<T>, config.block_count, in operator ()()
241 DilationBackpropInputKernel<T>, config.block_count, in operator ()()
275 TF_CHECK_OK(GpuLaunchKernel(SetZero<T>, config.block_count, in operator ()()
283 DilationBackpropFilterKernel<T>, config.block_count, in operator ()()
Dcwise_op_clip_gpu.cu.cc74 UnaryClipCustomKernel<T>, config.block_count, config.thread_per_block, in operator ()()
90 BinaryRightClipCustomKernel<T>, config.block_count, in operator ()()
106 BinaryLeftClipCustomKernel<T>, config.block_count, in operator ()()
Dsplit_lib_gpu.cu.cc202 TF_CHECK_OK(GpuLaunchKernel(SplitOpKernel<T>, config.block_count, in Run()
218 TF_CHECK_OK(GpuLaunchKernel(SplitVOpKernel_fixed<T>, config.block_count, in Run()
232 split_v_kernel<T, IntType, true>, config.block_count, in Run()
237 split_v_kernel<T, IntType, false>, config.block_count, in Run()
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/
DTensorDeviceThreadPool.h316 ParallelForAsyncContext(Index block_count, in ParallelForAsyncContext()
319 : count(block_count), in ParallelForAsyncContext()
358 Index block_count = divup(n, block_size); in CalculateParallelForBlock() local
363 static_cast<double>(block_count) / in CalculateParallelForBlock()
364 (divup<int>(block_count, numThreads()) * numThreads()); in CalculateParallelForBlock()
368 for (Index prev_block_count = block_count; in CalculateParallelForBlock()
391 block_count = coarser_block_count; in CalculateParallelForBlock()
398 return {block_size, block_count}; in CalculateParallelForBlock()
/external/webrtc/modules/desktop_capture/
Ddesktop_capturer_differ_wrapper.cc63 const int block_count = (width - 1) / kBlockSize; in CompareRow() local
64 const int last_block_width = width - block_count * kBlockSize; in CompareRow()
73 for (int x = 0; x < block_count; x++) { in CompareRow()
101 first_dirty_x_block = block_count; in CompareRow()
108 block_count * kBlockSize + left, bottom)); in CompareRow()
/external/python/cpython2/Mac/Demo/applescript/Disk_Copy/
DStandard_Suite.py222 class block_count(aetools.NProperty): class
285 block_count = block_count variable
346 'block_count' : block_count,
373 'block_count' : block_count,
459 'Xblk' : block_count,
/external/armnn/python/pyarmnn/examples/common/
Daudio_capture.py49 self.block_count = 0
87 self.block_count, remainder_samples = divmod(
92 self.block_count += 1
110 if self.current_block == self.block_count:
/external/tensorflow/tensorflow/core/kernels/image/
Dadjust_hue_op_gpu.cu.cc36 const int block_count = in operator ()() local
39 block_count, threads_per_block, 0, stream, in operator ()()
Dadjust_saturation_op_gpu.cu.cc37 const int block_count = in operator ()() local
40 block_count, threads_per_block, 0, stream, in operator ()()
/external/libchrome/base/trace_event/
Dmalloc_dump_provider.cc37 size_t block_count; member
53 crt_heap_info->block_count++; in WinHeapMemoryDumpImpl()
131 allocated_objects_count = main_heap_info.block_count; in OnMemoryDump()
/external/arm-trusted-firmware/drivers/brcm/emmc/
Demmc_pboot_hal_memory_drv.c197 size_t block_count = 0, blocks = 0; in emmc_erase() local
207 while (block_count < num_of_blocks) { in emmc_erase()
208 blocks = ((num_of_blocks - block_count) > erase_group) ? in emmc_erase()
209 erase_group : (num_of_blocks - block_count); in emmc_erase()
210 err = emmc_block_erase(mem_addr + block_count, blocks); in emmc_erase()
214 block_count += blocks; in emmc_erase()
/external/tensorflow/tensorflow/compiler/xla/stream_executor/
Ddevice_description.cc144 int64_t *block_count) { in CalculateDimensionality() argument
146 *block_count = port::MathUtil::CeilOfRatio(element_count, *threads_per_block); in CalculateDimensionality()
147 if (*block_count == 1) { in CalculateDimensionality()
/external/tensorflow/tensorflow/core/kernels/sparse/
Dkernels_gpu.cu.cc158 config.block_count, config.thread_per_block, 0, in operator ()()
163 config.block_count, config.thread_per_block, 0, in operator ()()
234 config.block_count, config.thread_per_block, 0, in operator ()()
250 GpuLaunchKernel(COOMatrixToSparseTensorKernel3D, config.block_count, in operator ()()
320 CSRSparseMatrixBatchMulVecKernel3D<T>, config.block_count, in CSRSparseMatrixBatchMulVecImpl()
435 config.block_count, config.thread_per_block, 0, in CSRSparseMatrixSoftmaxGPUImpl()
456 config.block_count, config.thread_per_block, in CSRSparseMatrixSoftmaxGPUImpl()
630 CSRSparseMatrixSoftmaxGradKernel2D<T>, config.block_count, in CSRSparseMatrixSoftmaxGradGPUImpl()
661 CSRSparseMatrixSoftmaxGradKernel3D<T>, config.block_count, in CSRSparseMatrixSoftmaxGradGPUImpl()
/external/libjpeg-turbo/
Drdtarga.c63 int block_count; /* # of pixels remaining in RLE block */ member
142 if (--sinfo->block_count < 0) { /* decrement pixels remaining in block */ in read_rle_pixel()
146 sinfo->block_count = 0; /* then read new block header */ in read_rle_pixel()
148 sinfo->block_count = i & 0x7F; /* number of pixels after this one */ in read_rle_pixel()
375 source->block_count = source->dup_pixel_count = 0; in start_input_tga()

1234567