/external/tensorflow/tensorflow/core/util/ |
D | gpu_launch_config.h | 119 int block_count = -1; member 136 const int block_count = in GetGpuLaunchConfig() local 142 config.block_count = block_count; in GetGpuLaunchConfig() 162 int block_count = 0; in GetGpuLaunchConfig() local 167 &block_count, &thread_per_block, func, dynamic_shared_memory_size, in GetGpuLaunchConfig() 172 &block_count, &thread_per_block, func, dynamic_shared_memory_size, in GetGpuLaunchConfig() 177 block_count = in GetGpuLaunchConfig() 178 std::min(block_count, DivUp(work_element_count, thread_per_block)); in GetGpuLaunchConfig() 182 config.block_count = block_count; in GetGpuLaunchConfig() 197 int block_count = 0; in GetGpuLaunchConfigFixedBlockSize() local [all …]
|
D | gpu_kernel_helper_test.cu.cc | 184 TF_CHECK_OK(GpuLaunchKernel(SetOutbufZero, cfg.block_count, \ in TEST_F() 189 TF_CHECK_OK(GpuLaunchKernel(Count1D, cfg.block_count, cfg.thread_per_block, \ in TEST_F() 197 TF_CHECK_OK(GpuLaunchKernel(SetOutbufZero, cfg.block_count, \ in TEST_F() 202 TF_CHECK_OK(GpuLaunchKernel(Count1D, cfg.block_count, cfg.thread_per_block, \ in TEST_F() 226 a.block_count.x == b.block_count.x && in operator ==() 227 a.block_count.y == b.block_count.y && in operator ==() 228 a.block_count.z == b.block_count.z && in operator ==() 241 TF_EXPECT_OK(GpuLaunchKernel(SetOutbufZero, cfg1d.block_count, \ in TEST_F() 246 TF_EXPECT_OK(GpuLaunchKernel(Count2D, cfg.block_count, cfg.thread_per_block, \ in TEST_F() 254 TF_EXPECT_OK(GpuLaunchKernel(SetOutbufZero, cfg1d.block_count, \ in TEST_F() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | launch_dimensions.cc | 131 int64_t block_count = in CalculateLaunchDimensions() local 140 if (capped_block_count < block_count) { in CalculateLaunchDimensions() 142 block_count = capped_block_count; in CalculateLaunchDimensions() 143 VLOG(2) << "Update the # of blocks to " << block_count in CalculateLaunchDimensions() 151 int64_t capped_block_count = block_count; in CalculateLaunchDimensions() 161 if (capped_block_count < block_count) { in CalculateLaunchDimensions() 163 block_count = capped_block_count; in CalculateLaunchDimensions() 167 block_count >= gpu_device_info.block_dim_limit_x) { in CalculateLaunchDimensions() 169 "Kernel launch needs more blocks (", block_count, in CalculateLaunchDimensions() 177 block_count, threads_per_block_x, threads_per_block_y, num_elements); in CalculateLaunchDimensions() [all …]
|
/external/cronet/net/disk_cache/blockfile/ |
D | block_files.cc | 185 bool BlockHeader::NeedToGrowBlockFile(int block_count) const { in NeedToGrowBlockFile() 190 if (i >= block_count - 1 && header_->empty[i]) in NeedToGrowBlockFile() 203 bool BlockHeader::CanAllocate(int block_count) const { in CanAllocate() 204 DCHECK_GT(block_count, 0); in CanAllocate() 205 for (int i = block_count - 1; i < kMaxNumBlocks; i++) { in CanAllocate() 311 bool BlockFiles::CreateBlock(FileType block_type, int block_count, in CreateBlock() argument 318 if (block_count < 1 || block_count > kMaxNumBlocks) in CreateBlock() 324 MappedFile* file = FileForNewBlock(block_type, block_count); in CreateBlock() 332 if (!file_header.CreateMapBlock(block_count, &index)) in CreateBlock() 335 Addr address(block_type, block_count, file_header.FileId(), index); in CreateBlock() [all …]
|
D | block_files.h | 59 bool NeedToGrowBlockFile(int block_count) const; 63 bool CanAllocate(int block_count) const; 113 bool CreateBlock(FileType block_type, int block_count, Addr* block_address); 137 MappedFile* FileForNewBlock(FileType block_type, int block_count);
|
/external/skia/tests/ |
D | SkBlockAllocatorTest.cpp | 31 static int block_count(const SkSBlockAllocator<N>& pool) { in block_count() function 338 REPORTER_ASSERT(r, block_count(pool) == 4); // 3 added plus the implicit head in DEF_TEST() 347 REPORTER_ASSERT(r,block_count(pool) == 1); in DEF_TEST() 381 REPORTER_ASSERT(r, block_count(pool) == 3); in DEF_TEST() 386 REPORTER_ASSERT(r, block_count(pool) == 2); in DEF_TEST() 391 REPORTER_ASSERT(r, block_count(pool) == 1); in DEF_TEST() 404 REPORTER_ASSERT(r, block_count(pool) == 1); in DEF_TEST() 413 REPORTER_ASSERT(r, block_count(pool) == 3); in DEF_TEST() 435 REPORTER_ASSERT(r, block_count(pool) == 1); in DEF_TEST() 452 REPORTER_ASSERT(r, block_count(pool) == 1); in DEF_TEST() [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | bias_op_gpu.cu.cc | 90 TF_CHECK_OK(GpuLaunchKernel(BiasNHWCKernel<T>, config.block_count, in compute() 97 TF_CHECK_OK(GpuLaunchKernel(BiasNCHWKernel<T>, config.block_count, in compute() 237 config.block_count, config.thread_per_block, in compute() 242 int group_size = (config.block_count + bias_size - 1) / bias_size; in compute() 243 config.block_count = group_size * bias_size; in compute() 248 config.block_count, config.thread_per_block, in compute() 258 BiasGradNHWC_Naive<T>, config.block_count, config.thread_per_block, 0, in compute() 261 TF_CHECK_OK(GpuLaunchKernel(BiasGradNCHW_Naive<T>, config.block_count, in compute()
|
D | depthtospace_op_gpu.cu.cc | 165 D2S_NHWC<T>, config.block_count, config.thread_per_block, 0, d.stream(), in operator ()() 201 D2S_NCHW_LOOP<T, 2>, config.block_count, config.thread_per_block, in operator ()() 208 D2S_NCHW_LOOP<T, 3>, config.block_count, config.thread_per_block, in operator ()() 215 D2S_NCHW_LOOP<T, 4>, config.block_count, config.thread_per_block, in operator ()() 230 D2S_NCHW<T>, config.block_count, config.thread_per_block, 0, d.stream(), in operator ()()
|
D | spacetodepth_op_gpu.cu.cc | 162 S2D_NHWC<T>, config.block_count, config.thread_per_block, 0, d.stream(), in operator ()() 198 S2D_NCHW_LOOP<T, 2>, config.block_count, config.thread_per_block, in operator ()() 205 S2D_NCHW_LOOP<T, 3>, config.block_count, config.thread_per_block, in operator ()() 212 S2D_NCHW_LOOP<T, 4>, config.block_count, config.thread_per_block, in operator ()() 227 S2D_NCHW<T>, config.block_count, config.thread_per_block, 0, d.stream(), in operator ()()
|
D | inplace_ops_functor_gpu.cu.cc | 54 DoParallelConcatOpKernel<T>, cfg.block_count, cfg.thread_per_block, 0, in DoParallelConcatUpdate() 124 cfg.block_count, cfg.thread_per_block, 0, in DoInplaceOp() 129 TF_CHECK_OK(GpuLaunchKernel(DoInplaceOpKernel<T, I_ADD>, cfg.block_count, in DoInplaceOp() 135 TF_CHECK_OK(GpuLaunchKernel(DoInplaceOpKernel<T, I_SUB>, cfg.block_count, in DoInplaceOp() 158 cfg.block_count, cfg.thread_per_block, 0, in DoInplaceOp()
|
D | conv_2d_gpu.h | 498 config.block_count, config.thread_per_block, 504 config.block_count, config.thread_per_block, 535 config.block_count, config.thread_per_block, 549 config.block_count, config.thread_per_block, 586 PadInputCustomKernelNHWC<T, NDIMS>, config.block_count, 592 PadInputCustomKernelNCHW<T, NDIMS>, config.block_count, 1037 config.block_count, config.thread_per_block, 0, 1085 int block_count; 1087 block_count = (total_size + config.thread_per_block - 1) / 1090 block_count = config.block_count; [all …]
|
D | dilation_ops_gpu.cu.cc | 200 DilationKernel<T>, config.block_count, config.thread_per_block, 0, in operator ()() 233 TF_CHECK_OK(GpuLaunchKernel(SetZero<T>, config.block_count, in operator ()() 241 DilationBackpropInputKernel<T>, config.block_count, in operator ()() 275 TF_CHECK_OK(GpuLaunchKernel(SetZero<T>, config.block_count, in operator ()() 283 DilationBackpropFilterKernel<T>, config.block_count, in operator ()()
|
D | cwise_op_clip_gpu.cu.cc | 74 UnaryClipCustomKernel<T>, config.block_count, config.thread_per_block, in operator ()() 90 BinaryRightClipCustomKernel<T>, config.block_count, in operator ()() 106 BinaryLeftClipCustomKernel<T>, config.block_count, in operator ()()
|
D | split_lib_gpu.cu.cc | 202 TF_CHECK_OK(GpuLaunchKernel(SplitOpKernel<T>, config.block_count, in Run() 218 TF_CHECK_OK(GpuLaunchKernel(SplitVOpKernel_fixed<T>, config.block_count, in Run() 232 split_v_kernel<T, IntType, true>, config.block_count, in Run() 237 split_v_kernel<T, IntType, false>, config.block_count, in Run()
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
D | TensorDeviceThreadPool.h | 316 ParallelForAsyncContext(Index block_count, in ParallelForAsyncContext() 319 : count(block_count), in ParallelForAsyncContext() 358 Index block_count = divup(n, block_size); in CalculateParallelForBlock() local 363 static_cast<double>(block_count) / in CalculateParallelForBlock() 364 (divup<int>(block_count, numThreads()) * numThreads()); in CalculateParallelForBlock() 368 for (Index prev_block_count = block_count; in CalculateParallelForBlock() 391 block_count = coarser_block_count; in CalculateParallelForBlock() 398 return {block_size, block_count}; in CalculateParallelForBlock()
|
/external/webrtc/modules/desktop_capture/ |
D | desktop_capturer_differ_wrapper.cc | 63 const int block_count = (width - 1) / kBlockSize; in CompareRow() local 64 const int last_block_width = width - block_count * kBlockSize; in CompareRow() 73 for (int x = 0; x < block_count; x++) { in CompareRow() 101 first_dirty_x_block = block_count; in CompareRow() 108 block_count * kBlockSize + left, bottom)); in CompareRow()
|
/external/python/cpython2/Mac/Demo/applescript/Disk_Copy/ |
D | Standard_Suite.py | 222 class block_count(aetools.NProperty): class 285 block_count = block_count variable 346 'block_count' : block_count, 373 'block_count' : block_count, 459 'Xblk' : block_count,
|
/external/armnn/python/pyarmnn/examples/common/ |
D | audio_capture.py | 49 self.block_count = 0 87 self.block_count, remainder_samples = divmod( 92 self.block_count += 1 110 if self.current_block == self.block_count:
|
/external/tensorflow/tensorflow/core/kernels/image/ |
D | adjust_hue_op_gpu.cu.cc | 36 const int block_count = in operator ()() local 39 block_count, threads_per_block, 0, stream, in operator ()()
|
D | adjust_saturation_op_gpu.cu.cc | 37 const int block_count = in operator ()() local 40 block_count, threads_per_block, 0, stream, in operator ()()
|
/external/libchrome/base/trace_event/ |
D | malloc_dump_provider.cc | 37 size_t block_count; member 53 crt_heap_info->block_count++; in WinHeapMemoryDumpImpl() 131 allocated_objects_count = main_heap_info.block_count; in OnMemoryDump()
|
/external/arm-trusted-firmware/drivers/brcm/emmc/ |
D | emmc_pboot_hal_memory_drv.c | 197 size_t block_count = 0, blocks = 0; in emmc_erase() local 207 while (block_count < num_of_blocks) { in emmc_erase() 208 blocks = ((num_of_blocks - block_count) > erase_group) ? in emmc_erase() 209 erase_group : (num_of_blocks - block_count); in emmc_erase() 210 err = emmc_block_erase(mem_addr + block_count, blocks); in emmc_erase() 214 block_count += blocks; in emmc_erase()
|
/external/tensorflow/tensorflow/compiler/xla/stream_executor/ |
D | device_description.cc | 144 int64_t *block_count) { in CalculateDimensionality() argument 146 *block_count = port::MathUtil::CeilOfRatio(element_count, *threads_per_block); in CalculateDimensionality() 147 if (*block_count == 1) { in CalculateDimensionality()
|
/external/tensorflow/tensorflow/core/kernels/sparse/ |
D | kernels_gpu.cu.cc | 158 config.block_count, config.thread_per_block, 0, in operator ()() 163 config.block_count, config.thread_per_block, 0, in operator ()() 234 config.block_count, config.thread_per_block, 0, in operator ()() 250 GpuLaunchKernel(COOMatrixToSparseTensorKernel3D, config.block_count, in operator ()() 320 CSRSparseMatrixBatchMulVecKernel3D<T>, config.block_count, in CSRSparseMatrixBatchMulVecImpl() 435 config.block_count, config.thread_per_block, 0, in CSRSparseMatrixSoftmaxGPUImpl() 456 config.block_count, config.thread_per_block, in CSRSparseMatrixSoftmaxGPUImpl() 630 CSRSparseMatrixSoftmaxGradKernel2D<T>, config.block_count, in CSRSparseMatrixSoftmaxGradGPUImpl() 661 CSRSparseMatrixSoftmaxGradKernel3D<T>, config.block_count, in CSRSparseMatrixSoftmaxGradGPUImpl()
|
/external/libjpeg-turbo/ |
D | rdtarga.c | 63 int block_count; /* # of pixels remaining in RLE block */ member 142 if (--sinfo->block_count < 0) { /* decrement pixels remaining in block */ in read_rle_pixel() 146 sinfo->block_count = 0; /* then read new block header */ in read_rle_pixel() 148 sinfo->block_count = i & 0x7F; /* number of pixels after this one */ in read_rle_pixel() 375 source->block_count = source->dup_pixel_count = 0; in start_input_tga()
|