| /external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
| D | TensorContractionGpu.h | 24 const Index m_size, const Index n_size, const Index k_size) { in EigenContractionKernelInternal() 513 const Index m_size, const Index n_size, const Index k_size) { in EigenContractionKernel() 538 const Index n_size, const Index k_size, in EigenFloatContractionKernelInternal16x16() 778 const Index n_size, const Index k_size, in EigenFloatContractionKernelInternal() 1145 const Index m_size, const Index n_size, const Index k_size) { in EigenFloatContractionKernel() 1192 const Index m_size, const Index n_size, const Index k_size) { in EigenFloatContractionKernel16x16()
|
| /external/eigen/unsupported/test/ |
| D | cxx11_tensor_contract_gpu.cu | 26 void test_gpu_contraction(int m_size, int k_size, int n_size) in test_gpu_contraction() 90 void test_scalar(int m_size, int k_size, int n_size) in test_scalar()
|
| D | cxx11_tensor_contract_sycl.cpp | 37 IndexType k_size, IndexType n_size) { in test_sycl_contraction() 155 IndexType k_size, IndexType n_size) { in test_no_out_of_bounds() 257 void test_scalar(const Device &sycl_device, IndexType m_size, IndexType k_size, in test_scalar() 321 IndexType k_size, IndexType n_size, IndexType m_batch, in contraction_batch() 395 IndexType k_size, IndexType n_size) { in contraction_rhs_transposed() 464 IndexType k_size, IndexType n_size) { in contraction_lhs_transposed() 530 IndexType k_size, IndexType n_size) { in contraction_both_transposed()
|
| D | cxx11_tensor_scan_gpu.cu | 25 void test_gpu_cumsum(int m_size, int k_size, int n_size) in test_gpu_cumsum()
|
| D | cxx11_tensor_scan_sycl.cpp | 27 IndexType k_size, IndexType n_size, int consume_dim, in test_sycl_cumsum()
|
| /external/ComputeLibrary/src/core/NEON/kernels/arm_gemm/ |
| D | gemm_interleaved_pretransposed_2d.hpp | 466 unsigned int k_size = (current.kmax() - current.k0()); in get_B_pretransposed_array_size() local 490 unsigned int k_size = (current.kmax() - current.k0()); in pretranspose_B_array() local
|
| D | gemm_hybrid.hpp | 232 const unsigned int k_size = roundup(kmax-k0, strategy::k_unroll()); in pretranspose_B_array() local
|
| D | gemm_hybrid_quantized_inline.hpp | 241 const unsigned int k_size = roundup(kmax-k0, strategy::k_unroll()); in pretranspose_B_array() local
|
| D | gemm_hybrid_quantized.hpp | 291 const unsigned int k_size = roundup(kmax-k0, strategy::k_unroll()); in pretranspose_B_array() local
|
| D | gemm_hybrid_indirect.hpp | 645 unsigned int k_size = kmax - k0; in pretranspose_B_array() local
|
| D | gemm_interleaved.hpp | 1062 unsigned int k_size = (current.kmax() - current.k0()); in pretranspose_B_array() local
|
| /external/XNNPACK/src/ |
| D | operator-run.c | 965 const size_t k_size = context->input_size[3]; in xnn_compute_pad_5d() local
|
| /external/rust/crates/ash/src/vk/ |
| D | definitions.rs | 35228 pub k_size: u32, field 35284 pub fn k_size(mut self, k_size: u32) -> Self { in k_size() method
|