Home
last modified time | relevance | path

Searched refs:shared_memory_size (Results 1 – 3 of 3) sorted by relevance

/external/tensorflow/tensorflow/core/kernels/
Dbias_op_gpu.cu.cc211 int32 shared_memory_size = 0; in compute() local
213 shared_memory_size = bias_size * sizeof(typename AccumulatorType<T>::type); in compute()
216 if (shared_memory_size <= max_shared_memory_size) { in compute()
219 <<<config.block_count, config.thread_per_block, shared_memory_size, in compute()
Dtopk_op_gpu.h383 constexpr auto shared_memory_size = 48 << 10; // 48 KB
386 num_shards = shared_memory_size / heap_size - 1;
402 auto shared_memory_size = (num_shards + 1) * k * sizeof(Entry<T>);
405 shared_memory_size, stream, input, length, k,
Ddepthwise_conv_op_gpu.h635 const int shared_memory_size = in LaunchDepthwiseConv2dGPUSmall() local
640 num_outputs, device, kernel, shared_memory_size, in LaunchDepthwiseConv2dGPUSmall()
642 kernel<<<config.block_count, block_dim, shared_memory_size, in LaunchDepthwiseConv2dGPUSmall()
1594 const int shared_memory_size =
1596 if (shared_memory_size > device.sharedMemPerBlock()) {
1626 num_out_backprop, device, kernel, shared_memory_size,
1628 kernel<<<config.block_count, block_dim, shared_memory_size,