• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // This file is part of Eigen, a lightweight C++ template library
2 // for linear algebra.
3 //
4 // Copyright (C) 2008-2014 Gael Guennebaud <gael.guennebaud@inria.fr>
5 //
6 // This Source Code Form is subject to the terms of the Mozilla
7 // Public License v. 2.0. If a copy of the MPL was not distributed
8 // with this file, You can obtain one at http://mozilla.org/MPL/2.0/.
9 
10 #ifndef EIGEN_SPARSEASSIGN_H
11 #define EIGEN_SPARSEASSIGN_H
12 
13 namespace Eigen {
14 
15 template<typename Derived>
16 template<typename OtherDerived>
17 Derived& SparseMatrixBase<Derived>::operator=(const EigenBase<OtherDerived> &other)
18 {
19   internal::call_assignment_no_alias(derived(), other.derived());
20   return derived();
21 }
22 
23 template<typename Derived>
24 template<typename OtherDerived>
25 Derived& SparseMatrixBase<Derived>::operator=(const ReturnByValue<OtherDerived>& other)
26 {
27   // TODO use the evaluator mechanism
28   other.evalTo(derived());
29   return derived();
30 }
31 
32 template<typename Derived>
33 template<typename OtherDerived>
34 inline Derived& SparseMatrixBase<Derived>::operator=(const SparseMatrixBase<OtherDerived>& other)
35 {
36   // by default sparse evaluation do not alias, so we can safely bypass the generic call_assignment routine
37   internal::Assignment<Derived,OtherDerived,internal::assign_op<Scalar,typename OtherDerived::Scalar> >
38           ::run(derived(), other.derived(), internal::assign_op<Scalar,typename OtherDerived::Scalar>());
39   return derived();
40 }
41 
42 template<typename Derived>
43 inline Derived& SparseMatrixBase<Derived>::operator=(const Derived& other)
44 {
45   internal::call_assignment_no_alias(derived(), other.derived());
46   return derived();
47 }
48 
49 namespace internal {
50 
51 template<>
52 struct storage_kind_to_evaluator_kind<Sparse> {
53   typedef IteratorBased Kind;
54 };
55 
56 template<>
57 struct storage_kind_to_shape<Sparse> {
58   typedef SparseShape Shape;
59 };
60 
61 struct Sparse2Sparse {};
62 struct Sparse2Dense  {};
63 
64 template<> struct AssignmentKind<SparseShape, SparseShape>           { typedef Sparse2Sparse Kind; };
65 template<> struct AssignmentKind<SparseShape, SparseTriangularShape> { typedef Sparse2Sparse Kind; };
66 template<> struct AssignmentKind<DenseShape,  SparseShape>           { typedef Sparse2Dense  Kind; };
67 template<> struct AssignmentKind<DenseShape,  SparseTriangularShape> { typedef Sparse2Dense  Kind; };
68 
69 
70 template<typename DstXprType, typename SrcXprType>
71 void assign_sparse_to_sparse(DstXprType &dst, const SrcXprType &src)
72 {
73   typedef typename DstXprType::Scalar Scalar;
74   typedef internal::evaluator<DstXprType> DstEvaluatorType;
75   typedef internal::evaluator<SrcXprType> SrcEvaluatorType;
76 
77   SrcEvaluatorType srcEvaluator(src);
78 
79   const bool transpose = (DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit);
80   const Index outerEvaluationSize = (SrcEvaluatorType::Flags&RowMajorBit) ? src.rows() : src.cols();
81   if ((!transpose) && src.isRValue())
82   {
83     // eval without temporary
84     dst.resize(src.rows(), src.cols());
85     dst.setZero();
86     dst.reserve((std::max)(src.rows(),src.cols())*2);
87     for (Index j=0; j<outerEvaluationSize; ++j)
88     {
89       dst.startVec(j);
90       for (typename SrcEvaluatorType::InnerIterator it(srcEvaluator, j); it; ++it)
91       {
92         Scalar v = it.value();
93         dst.insertBackByOuterInner(j,it.index()) = v;
94       }
95     }
96     dst.finalize();
97   }
98   else
99   {
100     // eval through a temporary
101     eigen_assert(( ((internal::traits<DstXprType>::SupportedAccessPatterns & OuterRandomAccessPattern)==OuterRandomAccessPattern) ||
102               (!((DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit)))) &&
103               "the transpose operation is supposed to be handled in SparseMatrix::operator=");
104 
105     enum { Flip = (DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit) };
106 
107 
108     DstXprType temp(src.rows(), src.cols());
109 
110     temp.reserve((std::max)(src.rows(),src.cols())*2);
111     for (Index j=0; j<outerEvaluationSize; ++j)
112     {
113       temp.startVec(j);
114       for (typename SrcEvaluatorType::InnerIterator it(srcEvaluator, j); it; ++it)
115       {
116         Scalar v = it.value();
117         temp.insertBackByOuterInner(Flip?it.index():j,Flip?j:it.index()) = v;
118       }
119     }
120     temp.finalize();
121 
122     dst = temp.markAsRValue();
123   }
124 }
125 
126 // Generic Sparse to Sparse assignment
127 template< typename DstXprType, typename SrcXprType, typename Functor>
128 struct Assignment<DstXprType, SrcXprType, Functor, Sparse2Sparse>
129 {
130   static void run(DstXprType &dst, const SrcXprType &src, const internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
131   {
132     assign_sparse_to_sparse(dst.derived(), src.derived());
133   }
134 };
135 
136 // Generic Sparse to Dense assignment
137 template< typename DstXprType, typename SrcXprType, typename Functor>
138 struct Assignment<DstXprType, SrcXprType, Functor, Sparse2Dense>
139 {
140   static void run(DstXprType &dst, const SrcXprType &src, const Functor &func)
141   {
142     if(internal::is_same<Functor,internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> >::value)
143       dst.setZero();
144 
145     internal::evaluator<SrcXprType> srcEval(src);
146     resize_if_allowed(dst, src, func);
147     internal::evaluator<DstXprType> dstEval(dst);
148 
149     const Index outerEvaluationSize = (internal::evaluator<SrcXprType>::Flags&RowMajorBit) ? src.rows() : src.cols();
150     for (Index j=0; j<outerEvaluationSize; ++j)
151       for (typename internal::evaluator<SrcXprType>::InnerIterator i(srcEval,j); i; ++i)
152         func.assignCoeff(dstEval.coeffRef(i.row(),i.col()), i.value());
153   }
154 };
155 
156 // Specialization for "dst = dec.solve(rhs)"
157 // NOTE we need to specialize it for Sparse2Sparse to avoid ambiguous specialization error
158 template<typename DstXprType, typename DecType, typename RhsType, typename Scalar>
159 struct Assignment<DstXprType, Solve<DecType,RhsType>, internal::assign_op<Scalar,Scalar>, Sparse2Sparse>
160 {
161   typedef Solve<DecType,RhsType> SrcXprType;
162   static void run(DstXprType &dst, const SrcXprType &src, const internal::assign_op<Scalar,Scalar> &)
163   {
164     Index dstRows = src.rows();
165     Index dstCols = src.cols();
166     if((dst.rows()!=dstRows) || (dst.cols()!=dstCols))
167       dst.resize(dstRows, dstCols);
168 
169     src.dec()._solve_impl(src.rhs(), dst);
170   }
171 };
172 
173 struct Diagonal2Sparse {};
174 
175 template<> struct AssignmentKind<SparseShape,DiagonalShape> { typedef Diagonal2Sparse Kind; };
176 
177 template< typename DstXprType, typename SrcXprType, typename Functor>
178 struct Assignment<DstXprType, SrcXprType, Functor, Diagonal2Sparse>
179 {
180   typedef typename DstXprType::StorageIndex StorageIndex;
181   typedef typename DstXprType::Scalar Scalar;
182   typedef Array<StorageIndex,Dynamic,1> ArrayXI;
183   typedef Array<Scalar,Dynamic,1> ArrayXS;
184   template<int Options>
185   static void run(SparseMatrix<Scalar,Options,StorageIndex> &dst, const SrcXprType &src, const internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
186   {
187     Index dstRows = src.rows();
188     Index dstCols = src.cols();
189     if((dst.rows()!=dstRows) || (dst.cols()!=dstCols))
190       dst.resize(dstRows, dstCols);
191 
192     Index size = src.diagonal().size();
193     dst.makeCompressed();
194     dst.resizeNonZeros(size);
195     Map<ArrayXI>(dst.innerIndexPtr(), size).setLinSpaced(0,StorageIndex(size)-1);
196     Map<ArrayXI>(dst.outerIndexPtr(), size+1).setLinSpaced(0,StorageIndex(size));
197     Map<ArrayXS>(dst.valuePtr(), size) = src.diagonal();
198   }
199 
200   template<typename DstDerived>
201   static void run(SparseMatrixBase<DstDerived> &dst, const SrcXprType &src, const internal::assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
202   {
203     dst.diagonal() = src.diagonal();
204   }
205 
206   static void run(DstXprType &dst, const SrcXprType &src, const internal::add_assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
207   { dst.diagonal() += src.diagonal(); }
208 
209   static void run(DstXprType &dst, const SrcXprType &src, const internal::sub_assign_op<typename DstXprType::Scalar,typename SrcXprType::Scalar> &/*func*/)
210   { dst.diagonal() -= src.diagonal(); }
211 };
212 } // end namespace internal
213 
214 } // end namespace Eigen
215 
216 #endif // EIGEN_SPARSEASSIGN_H
217