/external/cblas/examples/ |
D | cblas_example2.c | 12 int rout=-1,info=0,m,n,k,lda,ldb,ldc; in main() local 47 m=INVALID; n=0; k=0; lda=1; ldb=1; ldc=1; in main() 49 &ALPHA, A, &lda, B, &ldb, &BETA, C, &ldc ); in main() 52 m=0; n=INVALID; k=0; lda=1; ldb=1; ldc=1; in main() 55 &ALPHA, A, &lda, B, &ldb, &BETA, C, &ldc ); in main() 59 m=2; n=0; k=0; lda=1; ldb=1; ldc=2; in main() 61 &ALPHA, A, &lda, B, &ldb, &BETA, C, &ldc ); in main() 65 m=0; n=0; k=2; lda=1; ldb=1; ldc=1; in main() 67 &ALPHA, A, &lda, B, &ldb, &BETA, C, &ldc ); in main()
|
/external/cblas/testing/ |
D | c_zblas3.c | 16 CBLAS_TEST_ZOMPLEX *b, int *ldb, CBLAS_TEST_ZOMPLEX *beta, in F77_zgemm() argument 51 B[i*LDB+j].real=b[j*(*ldb)+i].real; in F77_zgemm() 52 B[i*LDB+j].imag=b[j*(*ldb)+i].imag; in F77_zgemm() 60 B[i*LDB+j].real=b[j*(*ldb)+i].real; in F77_zgemm() 61 B[i*LDB+j].imag=b[j*(*ldb)+i].imag; in F77_zgemm() 85 b, *ldb, beta, c, *ldc ); in F77_zgemm() 88 b, *ldb, beta, c, *ldc ); in F77_zgemm() 92 CBLAS_TEST_ZOMPLEX *b, int *ldb, CBLAS_TEST_ZOMPLEX *beta, in F77_zhemm() argument 126 B[i*LDB+j].real=b[j*(*ldb)+i].real; in F77_zhemm() 127 B[i*LDB+j].imag=b[j*(*ldb)+i].imag; in F77_zhemm() [all …]
|
D | c_cblas3.c | 16 CBLAS_TEST_COMPLEX *b, int *ldb, CBLAS_TEST_COMPLEX *beta, in F77_cgemm() argument 51 B[i*LDB+j].real=b[j*(*ldb)+i].real; in F77_cgemm() 52 B[i*LDB+j].imag=b[j*(*ldb)+i].imag; in F77_cgemm() 60 B[i*LDB+j].real=b[j*(*ldb)+i].real; in F77_cgemm() 61 B[i*LDB+j].imag=b[j*(*ldb)+i].imag; in F77_cgemm() 85 b, *ldb, beta, c, *ldc ); in F77_cgemm() 88 b, *ldb, beta, c, *ldc ); in F77_cgemm() 92 CBLAS_TEST_COMPLEX *b, int *ldb, CBLAS_TEST_COMPLEX *beta, in F77_chemm() argument 126 B[i*LDB+j].real=b[j*(*ldb)+i].real; in F77_chemm() 127 B[i*LDB+j].imag=b[j*(*ldb)+i].imag; in F77_chemm() [all …]
|
D | c_sblas3.c | 13 int *k, float *alpha, float *a, int *lda, float *b, int *ldb, in F77_sgemm() argument 43 B[i*LDB+j]=b[j*(*ldb)+i]; in F77_sgemm() 50 B[i*LDB+j]=b[j*(*ldb)+i]; in F77_sgemm() 68 b, *ldb, *beta, c, *ldc ); in F77_sgemm() 71 b, *ldb, *beta, c, *ldc ); in F77_sgemm() 74 float *alpha, float *a, int *lda, float *b, int *ldb, in F77_ssymm() argument 104 B[i*LDB+j]=b[j*(*ldb)+i]; in F77_ssymm() 120 cblas_ssymm( CblasColMajor, side, uplo, *m, *n, *alpha, a, *lda, b, *ldb, in F77_ssymm() 123 cblas_ssymm( UNDEFINED, side, uplo, *m, *n, *alpha, a, *lda, b, *ldb, in F77_ssymm() 176 float *alpha, float *a, int *lda, float *b, int *ldb, in F77_ssyr2k() argument [all …]
|
D | c_dblas3.c | 15 int *k, double *alpha, double *a, int *lda, double *b, int *ldb, in F77_dgemm() argument 45 B[i*LDB+j]=b[j*(*ldb)+i]; in F77_dgemm() 52 B[i*LDB+j]=b[j*(*ldb)+i]; in F77_dgemm() 71 b, *ldb, *beta, c, *ldc ); in F77_dgemm() 74 b, *ldb, *beta, c, *ldc ); in F77_dgemm() 77 double *alpha, double *a, int *lda, double *b, int *ldb, in F77_dsymm() argument 107 B[i*LDB+j]=b[j*(*ldb)+i]; in F77_dsymm() 123 cblas_dsymm( CblasColMajor, side, uplo, *m, *n, *alpha, a, *lda, b, *ldb, in F77_dsymm() 126 cblas_dsymm( UNDEFINED, side, uplo, *m, *n, *alpha, a, *lda, b, *ldb, in F77_dsymm() 179 double *alpha, double *a, int *lda, double *b, int *ldb, in F77_dsyr2k() argument [all …]
|
/external/eigen/Eigen/src/Core/products/ |
D | SelfadjointMatrixMatrix_BLAS.h | 58 BlasIndex m, n, lda, ldb, ldc; \ 70 ldb = convert_index<BlasIndex>(rhsStride); \ 81 ldb = convert_index<BlasIndex>(b_tmp.outerStride()); \ 84 …, &numext::real_ref(alpha), (const BLASTYPE*)a, &lda, (const BLASTYPE*)b, &ldb, &numext::real_ref(… 104 BlasIndex m, n, lda, ldb, ldc; \ 117 ldb = convert_index<BlasIndex>(rhsStride); \ 144 ldb = convert_index<BlasIndex>(b_tmp.outerStride()); \ 147 …, &numext::real_ref(alpha), (const BLASTYPE*)a, &lda, (const BLASTYPE*)b, &ldb, &numext::real_ref(… 175 BlasIndex m, n, lda, ldb, ldc; \ 186 ldb = convert_index<BlasIndex>(lhsStride); \ [all …]
|
D | TriangularSolverMatrix_BLAS.h | 56 … BlasIndex m = convert_index<BlasIndex>(size), n = convert_index<BlasIndex>(otherSize), lda, ldb; \ 60 ldb = convert_index<BlasIndex>(otherStride);\ 83 …nsa, &diag, &m, &n, &numext::real_ref(alpha), (const BLASTYPE*)a, &lda, (BLASTYPE*)_other, &ldb); \ 109 … BlasIndex m = convert_index<BlasIndex>(otherSize), n = convert_index<BlasIndex>(size), lda, ldb; \ 113 ldb = convert_index<BlasIndex>(otherStride);\ 136 …nsa, &diag, &m, &n, &numext::real_ref(alpha), (const BLASTYPE*)a, &lda, (BLASTYPE*)_other, &ldb); \
|
D | GeneralMatrixMatrix_BLAS.h | 69 BlasIndex m, n, k, lda, ldb, ldc; \ 85 ldb = convert_index<BlasIndex>(rhsStride); \ 100 ldb = convert_index<BlasIndex>(b_tmp.outerStride()); \ 103 …, &numext::real_ref(alpha), (const BLASTYPE*)a, &lda, (const BLASTYPE*)b, &ldb, &numext::real_ref(…
|
/external/eigen/blas/ |
D | level3_impl.h | 13 …const RealScalar *pa, const int *lda, const RealScalar *pb, const int *ldb, const RealScalar *pbet… in EIGEN_BLAS_FUNC() 54 else if(*ldb<std::max(1,(OP(*opb)==NOTR)?*k:*n)) info = 10; in EIGEN_BLAS_FUNC() 74 func[code](*m, *n, *k, a, *lda, b, *ldb, c, *ldc, alpha, blocking, 0); in EIGEN_BLAS_FUNC() 79 … const RealScalar *palpha, const RealScalar *pa, const int *lda, RealScalar *pb, const int *ldb) in EIGEN_BLAS_FUNC() 154 else if(*ldb<std::max(1,*m)) info = 11; in EIGEN_BLAS_FUNC() 166 func[code](*m, *n, a, *lda, b, *ldb, blocking); in EIGEN_BLAS_FUNC() 171 func[code](*n, *m, a, *lda, b, *ldb, blocking); in EIGEN_BLAS_FUNC() 175 matrix(b,*m,*n,*ldb) *= alpha; in EIGEN_BLAS_FUNC() 184 … const RealScalar *palpha, const RealScalar *pa, const int *lda, RealScalar *pb, const int *ldb) in EIGEN_BLAS_FUNC() 259 else if(*ldb<std::max(1,*m)) info = 11; in EIGEN_BLAS_FUNC() [all …]
|
/external/eigen/Eigen/src/misc/ |
D | lapacke.h | 275 const float* b, lapack_int ldb, float* x, 281 const double* b, lapack_int ldb, double* x, 288 const lapack_complex_float* b, lapack_int ldb, 296 const lapack_complex_double* b, lapack_int ldb, 305 const float* c, const float* b, lapack_int ldb, 315 const double* c, const double* b, lapack_int ldb, 326 const lapack_complex_float* b, lapack_int ldb, 337 const lapack_complex_double* b, lapack_int ldb, 346 lapack_int ldb ); 350 lapack_int ldb ); [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | mkl_matmul_op.cc | 151 const int ldb, float* c, const int ldc) { in MklBlasGemm() argument 166 b, &ldb, a, &lda, &beta, c, &ldc); in MklBlasGemm() 171 ldb, beta, c, ldc); in MklBlasGemm() 182 const int ldb, double* c, const int ldc) { in MklBlasGemm() argument 187 ldb, beta, c, ldc); in MklBlasGemm() 194 const complex64* b, const int ldb, complex64* c, in MklBlasGemm() argument 201 reinterpret_cast<const MKL_Complex8*>(b), ldb, &beta, in MklBlasGemm() 210 const complex128* b, const int ldb, complex128* c, in MklBlasGemm() argument 217 reinterpret_cast<const MKL_Complex16*>(b), ldb, &beta, in MklBlasGemm()
|
D | quantized_matmul_op.cc | 37 int offset_a, int offset_b, int lda, int ldb, int ldc) { in GemmlowpMultiply() argument 50 ldb); in GemmlowpMultiply() 132 const size_t ldb = b.dim_size(1); in Compute() local 142 c_data, m, n, k, -offset_a, -offset_b, lda, ldb, ldc); in Compute() 153 ldb, ldc); in Compute() 157 ldb, ldc); in Compute() 163 ldb, ldc); in Compute() 167 lda, ldb, ldc); in Compute() 173 lda, b_data, offset_b, ldb, c_data, shift_c, offset_c, mult_c, ldc); in Compute()
|
D | gemm_functors.h | 65 size_t k, const T1* a, size_t lda, const T2* b, size_t ldb, in operator() 70 const size_t b_l_stride = ldb; in operator() 99 size_t k, const T1* a, size_t lda, const T2* b, size_t ldb, in operator() 120 size_t ldb, float* c, size_t ldc) { in operator() 122 lda, b, ldb, 0.0f, c, ldc); in operator()
|
/external/tensorflow/tensorflow/stream_executor/cuda/ |
D | cuda_blas.cc | 1564 const DeviceMemory<Eigen::half> &b, int ldb, float beta, in DoBlasGemm() argument 1572 a.opaque(), lda, b.opaque(), ldb, beta, c->opaque(), ldc); in DoBlasGemm() 1585 if (ldb < static_cast<int64>(k)) { in DoBlasGemm() 1586 LOG(WARNING) << "GEMM ldb (" << ldb << ") was smaller than k (" << k in DoBlasGemm() 1590 if (ldb < static_cast<int64>(n)) { in DoBlasGemm() 1612 SE_CUDA_DATA_HALF, lda, GpuMemory(b), SE_CUDA_DATA_HALF, ldb, &beta, in DoBlasGemm() 1625 const DeviceMemory<float> &b, int ldb, float beta, in DoBlasGemm() argument 1632 a.opaque(), lda, b.opaque(), ldb, beta, c->opaque(), ldc); in DoBlasGemm() 1645 if (ldb < static_cast<int64>(k)) { in DoBlasGemm() 1646 LOG(WARNING) << "GEMM ldb (" << ldb << ") was smaller than k (" << k in DoBlasGemm() [all …]
|
/external/cblas/include/ |
D | cblas.h | 419 const int lda, const float *B, const int ldb, 424 const float *B, const int ldb, const float beta, 433 const float *B, const int ldb, const float beta, 439 float *B, const int ldb); 444 float *B, const int ldb); 449 const int lda, const double *B, const int ldb, 454 const double *B, const int ldb, const double beta, 463 const double *B, const int ldb, const double beta, 469 double *B, const int ldb); 474 double *B, const int ldb); [all …]
|
/external/tensorflow/tensorflow/stream_executor/rocm/ |
D | rocm_blas.cc | 1463 int lda, const DeviceMemory<Eigen::half> &b, int ldb, in DoBlasGemm() argument 1470 a.opaque(), lda, b.opaque(), ldb, beta, c->opaque(), ldc); in DoBlasGemm() 1483 if (ldb < static_cast<int64>(k)) { in DoBlasGemm() 1484 LOG(WARNING) << "GEMM ldb (" << ldb << ") was smaller than k (" << k in DoBlasGemm() 1488 if (ldb < static_cast<int64>(n)) { in DoBlasGemm() 1500 reinterpret_cast<const rocblas_half *>(GpuMemory(b)), ldb, in DoBlasGemm() 1508 const DeviceMemory<float> &b, int ldb, float beta, in DoBlasGemm() argument 1515 a.opaque(), lda, b.opaque(), ldb, beta, c->opaque(), ldc); in DoBlasGemm() 1528 if (ldb < static_cast<int64>(k)) { in DoBlasGemm() 1529 LOG(WARNING) << "GEMM ldb (" << ldb << ") was smaller than k (" << k in DoBlasGemm() [all …]
|
/external/tensorflow/tensorflow/stream_executor/ |
D | blas.h | 957 int lda, const DeviceMemory<Eigen::half> &b, int ldb, 963 const DeviceMemory<float> &b, int ldb, float beta, 968 const DeviceMemory<double> &b, int ldb, double beta, 974 const DeviceMemory<std::complex<float>> &b, int ldb, 981 const DeviceMemory<std::complex<double>> &b, int ldb, 988 int lda, const DeviceMemory<Eigen::half> &b, int ldb, float beta, 994 const DeviceMemory<float> &b, int ldb, float beta, DeviceMemory<float> *c, 999 const DeviceMemory<double> &b, int ldb, double beta, 1006 const DeviceMemory<std::complex<float>> &b, int ldb, 1013 const DeviceMemory<std::complex<double>> &b, int ldb, [all …]
|
D | stream.cc | 3501 const DeviceMemory<Eigen::half> &b, int ldb, in ThenBlasGemm() argument 3505 PARAM(alpha), PARAM(a), PARAM(lda), PARAM(b), PARAM(ldb), in ThenBlasGemm() 3513 alpha, a, lda, b, ldb, beta, c, ldc); in ThenBlasGemm() 3519 const DeviceMemory<float> &b, int ldb, float beta, in ThenBlasGemm() argument 3522 PARAM(alpha), PARAM(a), PARAM(lda), PARAM(b), PARAM(ldb), in ThenBlasGemm() 3529 alpha, a, lda, b, ldb, beta, c, ldc); in ThenBlasGemm() 3535 const DeviceMemory<double> &b, int ldb, in ThenBlasGemm() argument 3538 PARAM(alpha), PARAM(a), PARAM(lda), PARAM(b), PARAM(ldb), in ThenBlasGemm() 3545 alpha, a, lda, b, ldb, beta, c, ldc); in ThenBlasGemm() 3554 int ldb, std::complex<float> beta, in ThenBlasGemm() argument [all …]
|
D | stream.h | 1253 const DeviceMemory<Eigen::half> &b, int ldb, 1259 const DeviceMemory<float> &b, int ldb, 1264 const DeviceMemory<double> &b, int ldb, 1272 int ldb, std::complex<float> beta, 1280 int ldb, std::complex<double> beta, 1288 const DeviceMemory<Eigen::half> &b, int ldb, 1296 const DeviceMemory<float> &b, int ldb, 1303 const DeviceMemory<double> &b, int ldb, 1311 const DeviceMemory<std::complex<float>> &b, int ldb, 1318 const DeviceMemory<std::complex<double>> &b, int ldb, [all …]
|
/external/gemmlowp/eight_bit_int_gemm/ |
D | eight_bit_int_gemm.cc | 74 const std::uint8_t* b, std::int32_t b_offset, int ldb, in EightBitIntGemmImpl() argument 92 MatrixMap<const std::uint8_t, RhsOrder> rhs(b, k, n, ldb); in EightBitIntGemmImpl() 114 std::int32_t b_offset, int ldb, std::int32_t* c, in EightBitIntGemmInt32Impl() argument 127 MatrixMap<const std::uint8_t, RhsOrder> rhs(b, k, n, ldb); in EightBitIntGemmInt32Impl() 221 int m, int n, int k, int lda, int ldb, int ldc, in CanHandleMetaFastpath() argument 234 if (!IsColumnMajorOrVector(transpose_b, ldb, k, n)) { in CanHandleMetaFastpath() 307 std::int32_t b_offset, int ldb, std::uint8_t* c, in EightBitIntGemm() argument 315 ldb, ldc, bit_depth)) { in EightBitIntGemm() 325 b_offset, ldb, c, c_offset, c_mult_int, \ in EightBitIntGemm() 345 std::int32_t ldb, float* c, float c_offset, in EightBitIntGemm() argument [all …]
|
/external/eigen/Eigen/src/SparseLU/ |
D | SparseLU_gemm_kernel.h | 26 …ex m, Index n, Index d, const Scalar* A, Index lda, const Scalar* B, Index ldb, Scalar* C, Index l… in sparselu_gemm() argument 53 c += B[k+j*ldb] * A[i+k*lda]; in sparselu_gemm() 67 const Scalar* Bc0 = B+(j+0)*ldb; in sparselu_gemm() 68 const Scalar* Bc1 = B+(j+1)*ldb; in sparselu_gemm() 172 const Scalar* Bc0 = B+(n-1)*ldb; in sparselu_gemm() 261 …if(rd==1) MapVector(C+j*ldc+ib,actual_b) += B[0+d_end+j*ldb] * ConstMapVector(A+(d_end+0)*ld… in sparselu_gemm() 263 …else if(rd==2) MapVector(C+j*ldc+ib,actual_b) += B[0+d_end+j*ldb] * ConstMapVector(A+(d_end+0)*ld… in sparselu_gemm() 264 … + B[1+d_end+j*ldb] * ConstMapVector(A+(d_end+1)*lda+ib, actual_b); in sparselu_gemm() 266 …else MapVector(C+j*ldc+ib,actual_b) += B[0+d_end+j*ldb] * ConstMapVector(A+(d_end+0)*ld… in sparselu_gemm() 267 … + B[1+d_end+j*ldb] * ConstMapVector(A+(d_end+1)*lda+ib, actual_b) in sparselu_gemm() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | triangular_solve_thunk.cc | 82 const int ldb = m_; in ExecuteOnStream() local 101 &b_data_typed, ldb) in ExecuteOnStream() 111 &b_data_typed, ldb) in ExecuteOnStream() 122 lda, &b_data_typed, ldb) in ExecuteOnStream() 133 lda, &b_data_typed, ldb) in ExecuteOnStream()
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/Hexagon/ |
D | csr-func-usedef.ll | 6 declare i8* @llvm.hexagon.circ.ldb(i8*, i8*, i32, i32) #1 26 %0 = call i8* @llvm.hexagon.circ.ldb(i8* %p0.082, i8* nonnull %element_load0, i32 %or, i32 2) 27 %1 = call i8* @llvm.hexagon.circ.ldb(i8* %p1.079, i8* nonnull null, i32 0, i32 1) 28 %2 = call i8* @llvm.hexagon.circ.ldb(i8* %p2.078, i8* nonnull %element_load2, i32 %or5, i32 3) 29 %3 = call i8* @llvm.hexagon.circ.ldb(i8* %2, i8* nonnull %element_load5, i32 %or5, i32 1) 30 %4 = call i8* @llvm.hexagon.circ.ldb(i8* %p3.077, i8* nonnull %element_load3, i32 %or7, i32 1)
|
/external/llvm/test/CodeGen/Hexagon/ |
D | csr-func-usedef.ll | 6 declare i8* @llvm.hexagon.circ.ldb(i8*, i8*, i32, i32) #1 26 %0 = call i8* @llvm.hexagon.circ.ldb(i8* %p0.082, i8* nonnull %element_load0, i32 %or, i32 2) 27 %1 = call i8* @llvm.hexagon.circ.ldb(i8* %p1.079, i8* nonnull null, i32 0, i32 1) 28 %2 = call i8* @llvm.hexagon.circ.ldb(i8* %p2.078, i8* nonnull %element_load2, i32 %or5, i32 3) 29 %3 = call i8* @llvm.hexagon.circ.ldb(i8* %2, i8* nonnull %element_load5, i32 %or5, i32 1) 30 %4 = call i8* @llvm.hexagon.circ.ldb(i8* %p3.077, i8* nonnull %element_load3, i32 %or7, i32 1)
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/ARC/ |
D | ldst.ll | 24 ; CHECK: ldb %r0, [%r0,4000] 45 ; CHECK: ldb.x %r0, [%r0,4000] 67 ; CHECK: ldb.x %r0, [%r0,16] 194 ; CHECK: ldb %r0, [@cccc+16] 213 ; CHECK-DAG: ldb %r[[REG0:[0-9]+]], [%r0,0] 214 ; CHECK-DAG: ldb %r[[REG1:[0-9]+]], [%r0,1] 215 ; CHECK-DAG: ldb %r[[REG2:[0-9]+]], [%r0,2] 216 ; CHECK-DAG: ldb %r[[REG3:[0-9]+]], [%r0,3] 227 ; CHECK-DAG: ldb %r[[REG0:[0-9]+]], [%r0,0] 228 ; CHECK-DAG: ldb %r[[REG1:[0-9]+]], [%r0,1]
|