/external/eigen/test/ |
D | vectorization_logic.cpp | 81 PacketSize = internal::packet_traits<Scalar>::size enumerator 86 typedef Matrix<Scalar,PacketSize,1> Vector1; in run() 89 typedef Matrix<Scalar,PacketSize,PacketSize> Matrix11; in run() 90 typedef Matrix<Scalar,2*PacketSize,2*PacketSize> Matrix22; in run() 91 … Matrix<Scalar,(Matrix11::Flags&RowMajorBit)?16:4*PacketSize,(Matrix11::Flags&RowMajorBit)?4*Packe… in run() 92 … Matrix<Scalar,(Matrix11::Flags&RowMajorBit)?16:4*PacketSize,(Matrix11::Flags&RowMajorBit)?4*Packe… in run() 93 typedef Matrix<Scalar,4*PacketSize,16,ColMajor> Matrix44c; in run() 94 typedef Matrix<Scalar,4*PacketSize,16,RowMajor> Matrix44r; in run() 97 (PacketSize==8 ? 4 : PacketSize==4 ? 2 : PacketSize==2 ? 1 : /*PacketSize==1 ?*/ 1), in run() 98 (PacketSize==8 ? 2 : PacketSize==4 ? 2 : PacketSize==2 ? 2 : /*PacketSize==1 ?*/ 1) in run() [all …]
|
D | packetmath.cpp | 54 for (int i=0; i<PacketSize; ++i) \ 55 ref[i] = REFOP(data1[i], data1[i+PacketSize]); \ 56 …l::pstore(data2, POP(internal::pload<Packet>(data1), internal::pload<Packet>(data1+PacketSize))); \ 57 VERIFY(areApprox(ref, data2, PacketSize) && #POP); \ 61 for (int i=0; i<PacketSize; ++i) \ 64 VERIFY(areApprox(ref, data2, PacketSize) && #POP); \ 89 for (int i=0; i<PacketSize; ++i) \ 92 VERIFY(areApprox(ref, data2, PacketSize) && #POP); \ 104 const int PacketSize = internal::packet_traits<Scalar>::size; in packetmath() local 107 const int size = PacketSize*4; in packetmath() [all …]
|
D | product_extra.cpp | 117 const int PacketSize = internal::packet_traits<Scalar>::size; in zero_sized_objects() local 118 const int PacketSize1 = PacketSize>1 ? PacketSize-1 : 1; in zero_sized_objects() 140 Matrix<Scalar,PacketSize,0> a; in zero_sized_objects() 142 Matrix<Scalar,PacketSize,1> res; in zero_sized_objects() 143 VERIFY_IS_APPROX( (res=a*b), MatrixType::Zero(PacketSize,1) ); in zero_sized_objects() 144 VERIFY_IS_APPROX( (res=a.lazyProduct(b)), MatrixType::Zero(PacketSize,1) ); in zero_sized_objects() 156 Matrix<Scalar,PacketSize,Dynamic> a(PacketSize,0); in zero_sized_objects() 158 Matrix<Scalar,PacketSize,1> res; in zero_sized_objects() 159 VERIFY_IS_APPROX( (res=a*b), MatrixType::Zero(PacketSize,1) ); in zero_sized_objects() 160 VERIFY_IS_APPROX( (res=a.lazyProduct(b)), MatrixType::Zero(PacketSize,1) ); in zero_sized_objects()
|
/external/eigen/test/eigen2/ |
D | eigen2_packetmath.cpp | 22 for (int i=0; i<PacketSize; ++i) \ 23 ref[i] = REFOP(data1[i], data1[i+PacketSize]); \ 24 ei_pstore(data2, POP(ei_pload(data1), ei_pload(data1+PacketSize))); \ 25 VERIFY(areApprox(ref, data2, PacketSize) && #POP); \ 46 const int PacketSize = ei_packet_traits<Scalar>::size; in packetmath() local 48 const int size = PacketSize*4; in packetmath() 51 EIGEN_ALIGN_128 Packet packets[PacketSize*2]; in packetmath() 60 VERIFY(areApprox(data1, data2, PacketSize) && "aligned load/store"); in packetmath() 62 for (int offset=0; offset<PacketSize; ++offset) in packetmath() 65 VERIFY(areApprox(data1+offset, data2, PacketSize) && "ei_ploadu"); in packetmath() [all …]
|
/external/eigen/bench/ |
D | benchVecAdd.cpp | 94 const int PacketSize = internal::packet_traits<Scalar>::size; in benchVec() local 97 for (int i=0; i<size; i+=PacketSize*8) in benchVec() 128 …internal::pstore(&a[i+2*PacketSize], internal::padd(internal::ploadu(&a[i+2*PacketSize]), internal… in benchVec() 129 …internal::pstore(&a[i+3*PacketSize], internal::padd(internal::ploadu(&a[i+3*PacketSize]), internal… in benchVec() 130 …internal::pstore(&a[i+4*PacketSize], internal::padd(internal::ploadu(&a[i+4*PacketSize]), internal… in benchVec() 131 …internal::pstore(&a[i+5*PacketSize], internal::padd(internal::ploadu(&a[i+5*PacketSize]), internal… in benchVec() 132 …internal::pstore(&a[i+6*PacketSize], internal::padd(internal::ploadu(&a[i+6*PacketSize]), internal… in benchVec() 133 …internal::pstore(&a[i+7*PacketSize], internal::padd(internal::ploadu(&a[i+7*PacketSize]), internal… in benchVec()
|
/external/eigen/Eigen/src/SparseLU/ |
D | SparseLU_gemm_kernel.h | 33 PacketSize = packet_traits<Scalar>::size, in sparselu_gemm() enumerator 38 SM = PM*PacketSize // step along M in sparselu_gemm() 44 …eigen_internal_assert(((lda%PacketSize)==0) && ((ldc%PacketSize)==0) && (i0==internal::first_align… in sparselu_gemm() 62 …Index actual_b_end2 = (actual_b/PacketSize)*PacketSize; // actual number of rows suitable for ve… in sparselu_gemm() 109 c0 = pload<Packet>(C0+i+(I)*PacketSize); \ in sparselu_gemm() 110 c1 = pload<Packet>(C1+i+(I)*PacketSize); \ in sparselu_gemm() 113 a0 = pload<Packet>(A0+i+(I+1)*PacketSize); \ in sparselu_gemm() 116 a1 = pload<Packet>(A1+i+(I+1)*PacketSize); \ in sparselu_gemm() 119 if(RK==4) a2 = pload<Packet>(A2+i+(I+1)*PacketSize); \ in sparselu_gemm() 122 if(RK==4) a3 = pload<Packet>(A3+i+(I+1)*PacketSize); \ in sparselu_gemm() [all …]
|
D | SparseLU_kernel_bmod.h | 66 const Index PacketSize = internal::packet_traits<Scalar>::size; in run() local 67 Index ldl = internal::first_multiple(nrow, PacketSize); in run() 69 Index aligned_offset = internal::first_aligned(tempv.data()+segsize, PacketSize); in run() 70 …Index aligned_with_B_offset = (PacketSize-internal::first_aligned(B.data(), PacketSize))%PacketSiz… in run()
|
D | SparseLU_panel_bmod.h | 69 const Index PacketSize = internal::packet_traits<Scalar>::size; in panel_bmod() local 104 Index ldu = internal::first_multiple<Index>(u_rows, PacketSize); in panel_bmod() 147 Index ldl = internal::first_multiple<Index>(nrow, PacketSize); in panel_bmod() 148 Index offset = (PacketSize-internal::first_aligned(B.data(), PacketSize)) % PacketSize; in panel_bmod()
|
/external/eigen/Eigen/src/Jacobi/ |
D | Jacobi.h | 305 enum { PacketSize = packet_traits<Scalar>::size }; in apply_rotation_in_the_plane() enumerator 324 ((incrx==1 && incry==1) || PacketSize == 1)) in apply_rotation_in_the_plane() 330 Index alignedEnd = alignedStart + ((size-alignedStart)/PacketSize)*PacketSize; in apply_rotation_in_the_plane() 349 for(Index i=alignedStart; i<alignedEnd; i+=PacketSize) in apply_rotation_in_the_plane() 355 px += PacketSize; in apply_rotation_in_the_plane() 356 py += PacketSize; in apply_rotation_in_the_plane() 361 … Index peelingEnd = alignedStart + ((size-alignedStart)/(Peeling*PacketSize))*(Peeling*PacketSize); in apply_rotation_in_the_plane() 362 for(Index i=alignedStart; i<peelingEnd; i+=Peeling*PacketSize) in apply_rotation_in_the_plane() 365 Packet xi1 = ploadu<Packet>(px+PacketSize); in apply_rotation_in_the_plane() 367 Packet yi1 = pload <Packet>(py+PacketSize); in apply_rotation_in_the_plane() [all …]
|
/external/eigen/Eigen/src/Core/ |
D | Reverse.h | 85 PacketSize = internal::packet_traits<Scalar>::size, 89 OffsetRow = ReverseRow && IsColMajor ? PacketSize : 1, 90 OffsetCol = ReverseCol && IsRowMajor ? PacketSize : 1, 164 …urn internal::preverse(m_matrix.template packet<LoadMode>( m_matrix.size() - index - PacketSize )); 170 …ast_derived().template writePacket<LoadMode>(m_matrix.size() - index - PacketSize, internal::preve…
|
D | Redux.h | 31 PacketSize = packet_traits<typename Derived::Scalar>::size, enumerator 41 MaySliceVectorize = MightVectorize && int(InnerMaxSize)>=3*PacketSize 59 …lingLimit = EIGEN_UNROLLING_LIMIT * (int(Traversal) == int(DefaultTraversal) ? 1 : int(PacketSize)) 124 PacketSize = packet_traits<typename Derived::Scalar>::size, 295 PacketSize = packet_traits<Scalar>::size, 297 VectorizedSize = (Size / PacketSize) * PacketSize 302 … Scalar res = func.predux(redux_vec_unroller<Func, Derived, 0, Size / PacketSize>::run(mat,func));
|
D | Assign.h | 43 PacketSize = packet_traits<typename Derived::Scalar>::size enumerator 50 …MayInnerVectorize = MightVectorize && int(InnerSize)!=Dynamic && int(InnerSize)%int(PacketSize)==0 58 && (int(InnerMaxSize)==Dynamic || int(InnerMaxSize)>=3*PacketSize) 78 UnrollingLimit = EIGEN_UNROLLING_LIMIT * (Vectorized ? int(PacketSize) : 1), 110 EIGEN_DEBUG_VAR(PacketSize) in debug()
|
D | GeneralProduct.h | 395 PacketSize = internal::packet_traits<Scalar>::size 397 …internal::plain_array<Scalar,EIGEN_SIZE_MIN_PREFER_FIXED(Size,MaxSize)+(ForceAlignment?PacketSize:…
|
/external/eigen/Eigen/src/Core/products/ |
D | SelfadjointMatrixVector.h | 47 const Index PacketSize = sizeof(Packet)/sizeof(Scalar); in run() local 98 size_t alignedEnd = alignedStart + ((endi-alignedStart)/(PacketSize))*(PacketSize); in run() 126 for (size_t i=alignedStart; i<alignedEnd; i+=PacketSize) in run() 128 Packet A0i = ploadu<Packet>(a0It); a0It += PacketSize; in run() 129 Packet A1i = ploadu<Packet>(a1It); a1It += PacketSize; in run() 130 … Packet Bi = ploadu<Packet>(rhsIt); rhsIt += PacketSize; // FIXME should be aligned in most cases in run() 136 pstore(resIt,Xi); resIt += PacketSize; in run()
|
D | GeneralBlockPanelKernel.h | 1128 enum { PacketSize = packet_traits<Scalar>::size }; 1134 eigen_assert( (StorageOrder==RowMajor) || ((Pack1%PacketSize)==0 && Pack1<=4*PacketSize) ); 1148 if(Pack1>=1*PacketSize) A = ploadu<Packet>(&lhs(i+0*PacketSize, k)); 1149 if(Pack1>=2*PacketSize) B = ploadu<Packet>(&lhs(i+1*PacketSize, k)); 1150 if(Pack1>=3*PacketSize) C = ploadu<Packet>(&lhs(i+2*PacketSize, k)); 1151 if(Pack1>=4*PacketSize) D = ploadu<Packet>(&lhs(i+3*PacketSize, k)); 1152 if(Pack1>=1*PacketSize) { pstore(blockA+count, cj.pconj(A)); count+=PacketSize; } 1153 if(Pack1>=2*PacketSize) { pstore(blockA+count, cj.pconj(B)); count+=PacketSize; } 1154 if(Pack1>=3*PacketSize) { pstore(blockA+count, cj.pconj(C)); count+=PacketSize; } 1155 if(Pack1>=4*PacketSize) { pstore(blockA+count, cj.pconj(D)); count+=PacketSize; } [all …]
|
D | CoeffBasedProduct.h | 130 PacketSize = internal::packet_traits<Scalar>::size, 288 enum { PacketSize = packet_traits<typename Lhs::Scalar>::size }; 291 …product_coeff_vectorized_unroller<UnrollingIndex-PacketSize, Lhs, Rhs, Packet>::run(row, col, lhs,… 321 enum { PacketSize = packet_traits<typename Lhs::Scalar>::size }; 325 …product_coeff_vectorized_unroller<UnrollingIndex-PacketSize, Lhs, Rhs, Packet>::run(row, col, lhs,…
|
D | SelfadjointMatrixMatrix.h | 79 enum { PacketSize = packet_traits<Scalar>::size }; enumerator
|
/external/eigen/bench/btl/libs/eigen3/ |
D | eigen3_interface.hh | 117 const int PacketSize = sizeof(Packet)/sizeof(Scalar); in triassign() local 126 alignedEnd = alignedStart + ((size-alignedStart)/(2*PacketSize))*(PacketSize*2); in triassign() 138 for (int index = alignedStart; index<alignedEnd; index+=PacketSize) in triassign()
|
/external/eigen/Eigen/src/Core/util/ |
D | Memory.h | 469 static const Index PacketSize = packet_traits<Scalar>::size; in first_aligned() local 470 static const Index PacketAlignedMask = PacketSize-1; in first_aligned() 472 if(PacketSize==1) in first_aligned() 486 … return std::min<Index>( (PacketSize - (Index((size_t(array)/sizeof(Scalar))) & PacketAlignedMask)) in first_aligned()
|
/external/webrtc/webrtc/modules/rtp_rtcp/source/ |
D | rtcp_utility_unittest.cc | 103 TEST_F(RtcpParseCommonHeaderTest, PacketSize) { in TEST_F() argument
|
/external/eigen/Eigen/src/Cholesky/ |
D | LLT.h | 65 PacketSize = internal::packet_traits<Scalar>::size, enumerator 66 AlignmentMask = int(PacketSize)-1,
|
/external/eigen/Eigen/src/Eigen2Support/ |
D | SVD.h | 37 PacketSize = internal::packet_traits<Scalar>::size, enumerator 38 AlignmentMask = int(PacketSize)-1,
|
/external/eigen/doc/ |
D | InsideEigenExample.dox | 360 As we said at the beginning, vectorization works with blocks of 4 floats. Here, \a PacketSize is 4.
|