• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // This file is part of Eigen, a lightweight C++ template library
2 // for linear algebra.
3 //
4 // Copyright (C) 2008-2014 Gael Guennebaud <gael.guennebaud@inria.fr>
5 //
6 // This Source Code Form is subject to the terms of the Mozilla
7 // Public License v. 2.0. If a copy of the MPL was not distributed
8 // with this file, You can obtain one at http://mozilla.org/MPL/2.0/.
9 
10 #ifndef EIGEN_SPARSE_CWISE_BINARY_OP_H
11 #define EIGEN_SPARSE_CWISE_BINARY_OP_H
12 
13 namespace Eigen {
14 
15 // Here we have to handle 3 cases:
16 //  1 - sparse op dense
17 //  2 - dense op sparse
18 //  3 - sparse op sparse
19 // We also need to implement a 4th iterator for:
20 //  4 - dense op dense
21 // Finally, we also need to distinguish between the product and other operations :
22 //                configuration      returned mode
23 //  1 - sparse op dense    product      sparse
24 //                         generic      dense
25 //  2 - dense op sparse    product      sparse
26 //                         generic      dense
27 //  3 - sparse op sparse   product      sparse
28 //                         generic      sparse
29 //  4 - dense op dense     product      dense
30 //                         generic      dense
31 //
32 // TODO to ease compiler job, we could specialize product/quotient with a scalar
33 //      and fallback to cwise-unary evaluator using bind1st_op and bind2nd_op.
34 
35 template<typename BinaryOp, typename Lhs, typename Rhs>
36 class CwiseBinaryOpImpl<BinaryOp, Lhs, Rhs, Sparse>
37   : public SparseMatrixBase<CwiseBinaryOp<BinaryOp, Lhs, Rhs> >
38 {
39   public:
40     typedef CwiseBinaryOp<BinaryOp, Lhs, Rhs> Derived;
41     typedef SparseMatrixBase<Derived> Base;
42     EIGEN_SPARSE_PUBLIC_INTERFACE(Derived)
CwiseBinaryOpImpl()43     CwiseBinaryOpImpl()
44     {
45       EIGEN_STATIC_ASSERT((
46                 (!internal::is_same<typename internal::traits<Lhs>::StorageKind,
47                                     typename internal::traits<Rhs>::StorageKind>::value)
48             ||  ((internal::evaluator<Lhs>::Flags&RowMajorBit) == (internal::evaluator<Rhs>::Flags&RowMajorBit))),
49             THE_STORAGE_ORDER_OF_BOTH_SIDES_MUST_MATCH);
50     }
51 };
52 
53 namespace internal {
54 
55 
56 // Generic "sparse OP sparse"
57 template<typename XprType> struct binary_sparse_evaluator;
58 
59 template<typename BinaryOp, typename Lhs, typename Rhs>
60 struct binary_evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>, IteratorBased, IteratorBased>
61   : evaluator_base<CwiseBinaryOp<BinaryOp, Lhs, Rhs> >
62 {
63 protected:
64   typedef typename evaluator<Lhs>::InnerIterator  LhsIterator;
65   typedef typename evaluator<Rhs>::InnerIterator  RhsIterator;
66   typedef CwiseBinaryOp<BinaryOp, Lhs, Rhs> XprType;
67   typedef typename traits<XprType>::Scalar Scalar;
68   typedef typename XprType::StorageIndex StorageIndex;
69 public:
70 
71   class InnerIterator
72   {
73   public:
74 
75     EIGEN_STRONG_INLINE InnerIterator(const binary_evaluator& aEval, Index outer)
76       : m_lhsIter(aEval.m_lhsImpl,outer), m_rhsIter(aEval.m_rhsImpl,outer), m_functor(aEval.m_functor)
77     {
78       this->operator++();
79     }
80 
81     EIGEN_STRONG_INLINE InnerIterator& operator++()
82     {
83       if (m_lhsIter && m_rhsIter && (m_lhsIter.index() == m_rhsIter.index()))
84       {
85         m_id = m_lhsIter.index();
86         m_value = m_functor(m_lhsIter.value(), m_rhsIter.value());
87         ++m_lhsIter;
88         ++m_rhsIter;
89       }
90       else if (m_lhsIter && (!m_rhsIter || (m_lhsIter.index() < m_rhsIter.index())))
91       {
92         m_id = m_lhsIter.index();
93         m_value = m_functor(m_lhsIter.value(), Scalar(0));
94         ++m_lhsIter;
95       }
96       else if (m_rhsIter && (!m_lhsIter || (m_lhsIter.index() > m_rhsIter.index())))
97       {
98         m_id = m_rhsIter.index();
99         m_value = m_functor(Scalar(0), m_rhsIter.value());
100         ++m_rhsIter;
101       }
102       else
103       {
104         m_value = 0; // this is to avoid a compilation warning
105         m_id = -1;
106       }
107       return *this;
108     }
109 
110     EIGEN_STRONG_INLINE Scalar value() const { return m_value; }
111 
112     EIGEN_STRONG_INLINE StorageIndex index() const { return m_id; }
113     EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
114     EIGEN_STRONG_INLINE Index row() const { return Lhs::IsRowMajor ? m_lhsIter.row() : index(); }
115     EIGEN_STRONG_INLINE Index col() const { return Lhs::IsRowMajor ? index() : m_lhsIter.col(); }
116 
117     EIGEN_STRONG_INLINE operator bool() const { return m_id>=0; }
118 
119   protected:
120     LhsIterator m_lhsIter;
121     RhsIterator m_rhsIter;
122     const BinaryOp& m_functor;
123     Scalar m_value;
124     StorageIndex m_id;
125   };
126 
127 
128   enum {
129     CoeffReadCost = evaluator<Lhs>::CoeffReadCost + evaluator<Rhs>::CoeffReadCost + functor_traits<BinaryOp>::Cost,
130     Flags = XprType::Flags
131   };
132 
133   explicit binary_evaluator(const XprType& xpr)
134     : m_functor(xpr.functor()),
135       m_lhsImpl(xpr.lhs()),
136       m_rhsImpl(xpr.rhs())
137   {
138     EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
139     EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
140   }
141 
142   inline Index nonZerosEstimate() const {
143     return m_lhsImpl.nonZerosEstimate() + m_rhsImpl.nonZerosEstimate();
144   }
145 
146 protected:
147   const BinaryOp m_functor;
148   evaluator<Lhs> m_lhsImpl;
149   evaluator<Rhs> m_rhsImpl;
150 };
151 
152 // dense op sparse
153 template<typename BinaryOp, typename Lhs, typename Rhs>
154 struct binary_evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>, IndexBased, IteratorBased>
155   : evaluator_base<CwiseBinaryOp<BinaryOp, Lhs, Rhs> >
156 {
157 protected:
158   typedef typename evaluator<Rhs>::InnerIterator  RhsIterator;
159   typedef CwiseBinaryOp<BinaryOp, Lhs, Rhs> XprType;
160   typedef typename traits<XprType>::Scalar Scalar;
161   typedef typename XprType::StorageIndex StorageIndex;
162 public:
163 
164   class InnerIterator
165   {
166     enum { IsRowMajor = (int(Rhs::Flags)&RowMajorBit)==RowMajorBit };
167   public:
168 
169     EIGEN_STRONG_INLINE InnerIterator(const binary_evaluator& aEval, Index outer)
170       : m_lhsEval(aEval.m_lhsImpl), m_rhsIter(aEval.m_rhsImpl,outer), m_functor(aEval.m_functor), m_value(0), m_id(-1), m_innerSize(aEval.m_expr.rhs().innerSize())
171     {
172       this->operator++();
173     }
174 
175     EIGEN_STRONG_INLINE InnerIterator& operator++()
176     {
177       ++m_id;
178       if(m_id<m_innerSize)
179       {
180         Scalar lhsVal = m_lhsEval.coeff(IsRowMajor?m_rhsIter.outer():m_id,
181                                         IsRowMajor?m_id:m_rhsIter.outer());
182         if(m_rhsIter && m_rhsIter.index()==m_id)
183         {
184           m_value = m_functor(lhsVal, m_rhsIter.value());
185           ++m_rhsIter;
186         }
187         else
188           m_value = m_functor(lhsVal, Scalar(0));
189       }
190 
191       return *this;
192     }
193 
194     EIGEN_STRONG_INLINE Scalar value() const { eigen_internal_assert(m_id<m_innerSize); return m_value; }
195 
196     EIGEN_STRONG_INLINE StorageIndex index() const { return m_id; }
197     EIGEN_STRONG_INLINE Index outer() const { return m_rhsIter.outer(); }
198     EIGEN_STRONG_INLINE Index row() const { return IsRowMajor ? m_rhsIter.outer() : m_id; }
199     EIGEN_STRONG_INLINE Index col() const { return IsRowMajor ? m_id : m_rhsIter.outer(); }
200 
201     EIGEN_STRONG_INLINE operator bool() const { return m_id<m_innerSize; }
202 
203   protected:
204     const evaluator<Lhs> &m_lhsEval;
205     RhsIterator m_rhsIter;
206     const BinaryOp& m_functor;
207     Scalar m_value;
208     StorageIndex m_id;
209     StorageIndex m_innerSize;
210   };
211 
212 
213   enum {
214     CoeffReadCost = evaluator<Lhs>::CoeffReadCost + evaluator<Rhs>::CoeffReadCost + functor_traits<BinaryOp>::Cost,
215     // Expose storage order of the sparse expression
216     Flags = (XprType::Flags & ~RowMajorBit) | (int(Rhs::Flags)&RowMajorBit)
217   };
218 
219   explicit binary_evaluator(const XprType& xpr)
220     : m_functor(xpr.functor()),
221       m_lhsImpl(xpr.lhs()),
222       m_rhsImpl(xpr.rhs()),
223       m_expr(xpr)
224   {
225     EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
226     EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
227   }
228 
229   inline Index nonZerosEstimate() const {
230     return m_expr.size();
231   }
232 
233 protected:
234   const BinaryOp m_functor;
235   evaluator<Lhs> m_lhsImpl;
236   evaluator<Rhs> m_rhsImpl;
237   const XprType &m_expr;
238 };
239 
240 // sparse op dense
241 template<typename BinaryOp, typename Lhs, typename Rhs>
242 struct binary_evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>, IteratorBased, IndexBased>
243   : evaluator_base<CwiseBinaryOp<BinaryOp, Lhs, Rhs> >
244 {
245 protected:
246   typedef typename evaluator<Lhs>::InnerIterator  LhsIterator;
247   typedef CwiseBinaryOp<BinaryOp, Lhs, Rhs> XprType;
248   typedef typename traits<XprType>::Scalar Scalar;
249   typedef typename XprType::StorageIndex StorageIndex;
250 public:
251 
252   class InnerIterator
253   {
254     enum { IsRowMajor = (int(Lhs::Flags)&RowMajorBit)==RowMajorBit };
255   public:
256 
257     EIGEN_STRONG_INLINE InnerIterator(const binary_evaluator& aEval, Index outer)
258       : m_lhsIter(aEval.m_lhsImpl,outer), m_rhsEval(aEval.m_rhsImpl), m_functor(aEval.m_functor), m_value(0), m_id(-1), m_innerSize(aEval.m_expr.lhs().innerSize())
259     {
260       this->operator++();
261     }
262 
263     EIGEN_STRONG_INLINE InnerIterator& operator++()
264     {
265       ++m_id;
266       if(m_id<m_innerSize)
267       {
268         Scalar rhsVal = m_rhsEval.coeff(IsRowMajor?m_lhsIter.outer():m_id,
269                                         IsRowMajor?m_id:m_lhsIter.outer());
270         if(m_lhsIter && m_lhsIter.index()==m_id)
271         {
272           m_value = m_functor(m_lhsIter.value(), rhsVal);
273           ++m_lhsIter;
274         }
275         else
276           m_value = m_functor(Scalar(0),rhsVal);
277       }
278 
279       return *this;
280     }
281 
282     EIGEN_STRONG_INLINE Scalar value() const { eigen_internal_assert(m_id<m_innerSize); return m_value; }
283 
284     EIGEN_STRONG_INLINE StorageIndex index() const { return m_id; }
285     EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
286     EIGEN_STRONG_INLINE Index row() const { return IsRowMajor ? m_lhsIter.outer() : m_id; }
287     EIGEN_STRONG_INLINE Index col() const { return IsRowMajor ? m_id : m_lhsIter.outer(); }
288 
289     EIGEN_STRONG_INLINE operator bool() const { return m_id<m_innerSize; }
290 
291   protected:
292     LhsIterator m_lhsIter;
293     const evaluator<Rhs> &m_rhsEval;
294     const BinaryOp& m_functor;
295     Scalar m_value;
296     StorageIndex m_id;
297     StorageIndex m_innerSize;
298   };
299 
300 
301   enum {
302     CoeffReadCost = evaluator<Lhs>::CoeffReadCost + evaluator<Rhs>::CoeffReadCost + functor_traits<BinaryOp>::Cost,
303     // Expose storage order of the sparse expression
304     Flags = (XprType::Flags & ~RowMajorBit) | (int(Lhs::Flags)&RowMajorBit)
305   };
306 
307   explicit binary_evaluator(const XprType& xpr)
308     : m_functor(xpr.functor()),
309       m_lhsImpl(xpr.lhs()),
310       m_rhsImpl(xpr.rhs()),
311       m_expr(xpr)
312   {
313     EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
314     EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
315   }
316 
317   inline Index nonZerosEstimate() const {
318     return m_expr.size();
319   }
320 
321 protected:
322   const BinaryOp m_functor;
323   evaluator<Lhs> m_lhsImpl;
324   evaluator<Rhs> m_rhsImpl;
325   const XprType &m_expr;
326 };
327 
328 template<typename T,
329          typename LhsKind   = typename evaluator_traits<typename T::Lhs>::Kind,
330          typename RhsKind   = typename evaluator_traits<typename T::Rhs>::Kind,
331          typename LhsScalar = typename traits<typename T::Lhs>::Scalar,
332          typename RhsScalar = typename traits<typename T::Rhs>::Scalar> struct sparse_conjunction_evaluator;
333 
334 // "sparse .* sparse"
335 template<typename T1, typename T2, typename Lhs, typename Rhs>
336 struct binary_evaluator<CwiseBinaryOp<scalar_product_op<T1,T2>, Lhs, Rhs>, IteratorBased, IteratorBased>
337   : sparse_conjunction_evaluator<CwiseBinaryOp<scalar_product_op<T1,T2>, Lhs, Rhs> >
338 {
339   typedef CwiseBinaryOp<scalar_product_op<T1,T2>, Lhs, Rhs> XprType;
340   typedef sparse_conjunction_evaluator<XprType> Base;
341   explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
342 };
343 // "dense .* sparse"
344 template<typename T1, typename T2, typename Lhs, typename Rhs>
345 struct binary_evaluator<CwiseBinaryOp<scalar_product_op<T1,T2>, Lhs, Rhs>, IndexBased, IteratorBased>
346   : sparse_conjunction_evaluator<CwiseBinaryOp<scalar_product_op<T1,T2>, Lhs, Rhs> >
347 {
348   typedef CwiseBinaryOp<scalar_product_op<T1,T2>, Lhs, Rhs> XprType;
349   typedef sparse_conjunction_evaluator<XprType> Base;
350   explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
351 };
352 // "sparse .* dense"
353 template<typename T1, typename T2, typename Lhs, typename Rhs>
354 struct binary_evaluator<CwiseBinaryOp<scalar_product_op<T1,T2>, Lhs, Rhs>, IteratorBased, IndexBased>
355   : sparse_conjunction_evaluator<CwiseBinaryOp<scalar_product_op<T1,T2>, Lhs, Rhs> >
356 {
357   typedef CwiseBinaryOp<scalar_product_op<T1,T2>, Lhs, Rhs> XprType;
358   typedef sparse_conjunction_evaluator<XprType> Base;
359   explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
360 };
361 
362 // "sparse ./ dense"
363 template<typename T1, typename T2, typename Lhs, typename Rhs>
364 struct binary_evaluator<CwiseBinaryOp<scalar_quotient_op<T1,T2>, Lhs, Rhs>, IteratorBased, IndexBased>
365   : sparse_conjunction_evaluator<CwiseBinaryOp<scalar_quotient_op<T1,T2>, Lhs, Rhs> >
366 {
367   typedef CwiseBinaryOp<scalar_quotient_op<T1,T2>, Lhs, Rhs> XprType;
368   typedef sparse_conjunction_evaluator<XprType> Base;
369   explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
370 };
371 
372 // "sparse && sparse"
373 template<typename Lhs, typename Rhs>
374 struct binary_evaluator<CwiseBinaryOp<scalar_boolean_and_op, Lhs, Rhs>, IteratorBased, IteratorBased>
375   : sparse_conjunction_evaluator<CwiseBinaryOp<scalar_boolean_and_op, Lhs, Rhs> >
376 {
377   typedef CwiseBinaryOp<scalar_boolean_and_op, Lhs, Rhs> XprType;
378   typedef sparse_conjunction_evaluator<XprType> Base;
379   explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
380 };
381 // "dense && sparse"
382 template<typename Lhs, typename Rhs>
383 struct binary_evaluator<CwiseBinaryOp<scalar_boolean_and_op, Lhs, Rhs>, IndexBased, IteratorBased>
384   : sparse_conjunction_evaluator<CwiseBinaryOp<scalar_boolean_and_op, Lhs, Rhs> >
385 {
386   typedef CwiseBinaryOp<scalar_boolean_and_op, Lhs, Rhs> XprType;
387   typedef sparse_conjunction_evaluator<XprType> Base;
388   explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
389 };
390 // "sparse && dense"
391 template<typename Lhs, typename Rhs>
392 struct binary_evaluator<CwiseBinaryOp<scalar_boolean_and_op, Lhs, Rhs>, IteratorBased, IndexBased>
393   : sparse_conjunction_evaluator<CwiseBinaryOp<scalar_boolean_and_op, Lhs, Rhs> >
394 {
395   typedef CwiseBinaryOp<scalar_boolean_and_op, Lhs, Rhs> XprType;
396   typedef sparse_conjunction_evaluator<XprType> Base;
397   explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
398 };
399 
400 // "sparse ^ sparse"
401 template<typename XprType>
402 struct sparse_conjunction_evaluator<XprType, IteratorBased, IteratorBased>
403   : evaluator_base<XprType>
404 {
405 protected:
406   typedef typename XprType::Functor BinaryOp;
407   typedef typename XprType::Lhs LhsArg;
408   typedef typename XprType::Rhs RhsArg;
409   typedef typename evaluator<LhsArg>::InnerIterator  LhsIterator;
410   typedef typename evaluator<RhsArg>::InnerIterator  RhsIterator;
411   typedef typename XprType::StorageIndex StorageIndex;
412   typedef typename traits<XprType>::Scalar Scalar;
413 public:
414 
415   class InnerIterator
416   {
417   public:
418 
419     EIGEN_STRONG_INLINE InnerIterator(const sparse_conjunction_evaluator& aEval, Index outer)
420       : m_lhsIter(aEval.m_lhsImpl,outer), m_rhsIter(aEval.m_rhsImpl,outer), m_functor(aEval.m_functor)
421     {
422       while (m_lhsIter && m_rhsIter && (m_lhsIter.index() != m_rhsIter.index()))
423       {
424         if (m_lhsIter.index() < m_rhsIter.index())
425           ++m_lhsIter;
426         else
427           ++m_rhsIter;
428       }
429     }
430 
431     EIGEN_STRONG_INLINE InnerIterator& operator++()
432     {
433       ++m_lhsIter;
434       ++m_rhsIter;
435       while (m_lhsIter && m_rhsIter && (m_lhsIter.index() != m_rhsIter.index()))
436       {
437         if (m_lhsIter.index() < m_rhsIter.index())
438           ++m_lhsIter;
439         else
440           ++m_rhsIter;
441       }
442       return *this;
443     }
444 
445     EIGEN_STRONG_INLINE Scalar value() const { return m_functor(m_lhsIter.value(), m_rhsIter.value()); }
446 
447     EIGEN_STRONG_INLINE StorageIndex index() const { return m_lhsIter.index(); }
448     EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
449     EIGEN_STRONG_INLINE Index row() const { return m_lhsIter.row(); }
450     EIGEN_STRONG_INLINE Index col() const { return m_lhsIter.col(); }
451 
452     EIGEN_STRONG_INLINE operator bool() const { return (m_lhsIter && m_rhsIter); }
453 
454   protected:
455     LhsIterator m_lhsIter;
456     RhsIterator m_rhsIter;
457     const BinaryOp& m_functor;
458   };
459 
460 
461   enum {
462     CoeffReadCost = evaluator<LhsArg>::CoeffReadCost + evaluator<RhsArg>::CoeffReadCost + functor_traits<BinaryOp>::Cost,
463     Flags = XprType::Flags
464   };
465 
466   explicit sparse_conjunction_evaluator(const XprType& xpr)
467     : m_functor(xpr.functor()),
468       m_lhsImpl(xpr.lhs()),
469       m_rhsImpl(xpr.rhs())
470   {
471     EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
472     EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
473   }
474 
475   inline Index nonZerosEstimate() const {
476     return (std::min)(m_lhsImpl.nonZerosEstimate(), m_rhsImpl.nonZerosEstimate());
477   }
478 
479 protected:
480   const BinaryOp m_functor;
481   evaluator<LhsArg> m_lhsImpl;
482   evaluator<RhsArg> m_rhsImpl;
483 };
484 
485 // "dense ^ sparse"
486 template<typename XprType>
487 struct sparse_conjunction_evaluator<XprType, IndexBased, IteratorBased>
488   : evaluator_base<XprType>
489 {
490 protected:
491   typedef typename XprType::Functor BinaryOp;
492   typedef typename XprType::Lhs LhsArg;
493   typedef typename XprType::Rhs RhsArg;
494   typedef evaluator<LhsArg> LhsEvaluator;
495   typedef typename evaluator<RhsArg>::InnerIterator  RhsIterator;
496   typedef typename XprType::StorageIndex StorageIndex;
497   typedef typename traits<XprType>::Scalar Scalar;
498 public:
499 
500   class InnerIterator
501   {
502     enum { IsRowMajor = (int(RhsArg::Flags)&RowMajorBit)==RowMajorBit };
503 
504   public:
505 
506     EIGEN_STRONG_INLINE InnerIterator(const sparse_conjunction_evaluator& aEval, Index outer)
507       : m_lhsEval(aEval.m_lhsImpl), m_rhsIter(aEval.m_rhsImpl,outer), m_functor(aEval.m_functor), m_outer(outer)
508     {}
509 
510     EIGEN_STRONG_INLINE InnerIterator& operator++()
511     {
512       ++m_rhsIter;
513       return *this;
514     }
515 
516     EIGEN_STRONG_INLINE Scalar value() const
517     { return m_functor(m_lhsEval.coeff(IsRowMajor?m_outer:m_rhsIter.index(),IsRowMajor?m_rhsIter.index():m_outer), m_rhsIter.value()); }
518 
519     EIGEN_STRONG_INLINE StorageIndex index() const { return m_rhsIter.index(); }
520     EIGEN_STRONG_INLINE Index outer() const { return m_rhsIter.outer(); }
521     EIGEN_STRONG_INLINE Index row() const { return m_rhsIter.row(); }
522     EIGEN_STRONG_INLINE Index col() const { return m_rhsIter.col(); }
523 
524     EIGEN_STRONG_INLINE operator bool() const { return m_rhsIter; }
525 
526   protected:
527     const LhsEvaluator &m_lhsEval;
528     RhsIterator m_rhsIter;
529     const BinaryOp& m_functor;
530     const Index m_outer;
531   };
532 
533 
534   enum {
535     CoeffReadCost = evaluator<LhsArg>::CoeffReadCost + evaluator<RhsArg>::CoeffReadCost + functor_traits<BinaryOp>::Cost,
536     // Expose storage order of the sparse expression
537     Flags = (XprType::Flags & ~RowMajorBit) | (int(RhsArg::Flags)&RowMajorBit)
538   };
539 
540   explicit sparse_conjunction_evaluator(const XprType& xpr)
541     : m_functor(xpr.functor()),
542       m_lhsImpl(xpr.lhs()),
543       m_rhsImpl(xpr.rhs())
544   {
545     EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
546     EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
547   }
548 
549   inline Index nonZerosEstimate() const {
550     return m_rhsImpl.nonZerosEstimate();
551   }
552 
553 protected:
554   const BinaryOp m_functor;
555   evaluator<LhsArg> m_lhsImpl;
556   evaluator<RhsArg> m_rhsImpl;
557 };
558 
559 // "sparse ^ dense"
560 template<typename XprType>
561 struct sparse_conjunction_evaluator<XprType, IteratorBased, IndexBased>
562   : evaluator_base<XprType>
563 {
564 protected:
565   typedef typename XprType::Functor BinaryOp;
566   typedef typename XprType::Lhs LhsArg;
567   typedef typename XprType::Rhs RhsArg;
568   typedef typename evaluator<LhsArg>::InnerIterator LhsIterator;
569   typedef evaluator<RhsArg> RhsEvaluator;
570   typedef typename XprType::StorageIndex StorageIndex;
571   typedef typename traits<XprType>::Scalar Scalar;
572 public:
573 
574   class InnerIterator
575   {
576     enum { IsRowMajor = (int(LhsArg::Flags)&RowMajorBit)==RowMajorBit };
577 
578   public:
579 
580     EIGEN_STRONG_INLINE InnerIterator(const sparse_conjunction_evaluator& aEval, Index outer)
581       : m_lhsIter(aEval.m_lhsImpl,outer), m_rhsEval(aEval.m_rhsImpl), m_functor(aEval.m_functor), m_outer(outer)
582     {}
583 
584     EIGEN_STRONG_INLINE InnerIterator& operator++()
585     {
586       ++m_lhsIter;
587       return *this;
588     }
589 
590     EIGEN_STRONG_INLINE Scalar value() const
591     { return m_functor(m_lhsIter.value(),
592                        m_rhsEval.coeff(IsRowMajor?m_outer:m_lhsIter.index(),IsRowMajor?m_lhsIter.index():m_outer)); }
593 
594     EIGEN_STRONG_INLINE StorageIndex index() const { return m_lhsIter.index(); }
595     EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
596     EIGEN_STRONG_INLINE Index row() const { return m_lhsIter.row(); }
597     EIGEN_STRONG_INLINE Index col() const { return m_lhsIter.col(); }
598 
599     EIGEN_STRONG_INLINE operator bool() const { return m_lhsIter; }
600 
601   protected:
602     LhsIterator m_lhsIter;
603     const evaluator<RhsArg> &m_rhsEval;
604     const BinaryOp& m_functor;
605     const Index m_outer;
606   };
607 
608 
609   enum {
610     CoeffReadCost = evaluator<LhsArg>::CoeffReadCost + evaluator<RhsArg>::CoeffReadCost + functor_traits<BinaryOp>::Cost,
611     // Expose storage order of the sparse expression
612     Flags = (XprType::Flags & ~RowMajorBit) | (int(LhsArg::Flags)&RowMajorBit)
613   };
614 
615   explicit sparse_conjunction_evaluator(const XprType& xpr)
616     : m_functor(xpr.functor()),
617       m_lhsImpl(xpr.lhs()),
618       m_rhsImpl(xpr.rhs())
619   {
620     EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
621     EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
622   }
623 
624   inline Index nonZerosEstimate() const {
625     return m_lhsImpl.nonZerosEstimate();
626   }
627 
628 protected:
629   const BinaryOp m_functor;
630   evaluator<LhsArg> m_lhsImpl;
631   evaluator<RhsArg> m_rhsImpl;
632 };
633 
634 }
635 
636 /***************************************************************************
637 * Implementation of SparseMatrixBase and SparseCwise functions/operators
638 ***************************************************************************/
639 
640 template<typename Derived>
641 template<typename OtherDerived>
642 Derived& SparseMatrixBase<Derived>::operator+=(const EigenBase<OtherDerived> &other)
643 {
644   call_assignment(derived(), other.derived(), internal::add_assign_op<Scalar,typename OtherDerived::Scalar>());
645   return derived();
646 }
647 
648 template<typename Derived>
649 template<typename OtherDerived>
650 Derived& SparseMatrixBase<Derived>::operator-=(const EigenBase<OtherDerived> &other)
651 {
652   call_assignment(derived(), other.derived(), internal::assign_op<Scalar,typename OtherDerived::Scalar>());
653   return derived();
654 }
655 
656 template<typename Derived>
657 template<typename OtherDerived>
658 EIGEN_STRONG_INLINE Derived &
659 SparseMatrixBase<Derived>::operator-=(const SparseMatrixBase<OtherDerived> &other)
660 {
661   return derived() = derived() - other.derived();
662 }
663 
664 template<typename Derived>
665 template<typename OtherDerived>
666 EIGEN_STRONG_INLINE Derived &
667 SparseMatrixBase<Derived>::operator+=(const SparseMatrixBase<OtherDerived>& other)
668 {
669   return derived() = derived() + other.derived();
670 }
671 
672 template<typename Derived>
673 template<typename OtherDerived>
674 Derived& SparseMatrixBase<Derived>::operator+=(const DiagonalBase<OtherDerived>& other)
675 {
676   call_assignment_no_alias(derived(), other.derived(), internal::add_assign_op<Scalar,typename OtherDerived::Scalar>());
677   return derived();
678 }
679 
680 template<typename Derived>
681 template<typename OtherDerived>
682 Derived& SparseMatrixBase<Derived>::operator-=(const DiagonalBase<OtherDerived>& other)
683 {
684   call_assignment_no_alias(derived(), other.derived(), internal::sub_assign_op<Scalar,typename OtherDerived::Scalar>());
685   return derived();
686 }
687 
688 template<typename Derived>
689 template<typename OtherDerived>
690 EIGEN_STRONG_INLINE const typename SparseMatrixBase<Derived>::template CwiseProductDenseReturnType<OtherDerived>::Type
691 SparseMatrixBase<Derived>::cwiseProduct(const MatrixBase<OtherDerived> &other) const
692 {
693   return typename CwiseProductDenseReturnType<OtherDerived>::Type(derived(), other.derived());
694 }
695 
696 template<typename DenseDerived, typename SparseDerived>
697 EIGEN_STRONG_INLINE const CwiseBinaryOp<internal::scalar_sum_op<typename DenseDerived::Scalar,typename SparseDerived::Scalar>, const DenseDerived, const SparseDerived>
698 operator+(const MatrixBase<DenseDerived> &a, const SparseMatrixBase<SparseDerived> &b)
699 {
700   return CwiseBinaryOp<internal::scalar_sum_op<typename DenseDerived::Scalar,typename SparseDerived::Scalar>, const DenseDerived, const SparseDerived>(a.derived(), b.derived());
701 }
702 
703 template<typename SparseDerived, typename DenseDerived>
704 EIGEN_STRONG_INLINE const CwiseBinaryOp<internal::scalar_sum_op<typename SparseDerived::Scalar,typename DenseDerived::Scalar>, const SparseDerived, const DenseDerived>
705 operator+(const SparseMatrixBase<SparseDerived> &a, const MatrixBase<DenseDerived> &b)
706 {
707   return CwiseBinaryOp<internal::scalar_sum_op<typename SparseDerived::Scalar,typename DenseDerived::Scalar>, const SparseDerived, const DenseDerived>(a.derived(), b.derived());
708 }
709 
710 template<typename DenseDerived, typename SparseDerived>
711 EIGEN_STRONG_INLINE const CwiseBinaryOp<internal::scalar_difference_op<typename DenseDerived::Scalar,typename SparseDerived::Scalar>, const DenseDerived, const SparseDerived>
712 operator-(const MatrixBase<DenseDerived> &a, const SparseMatrixBase<SparseDerived> &b)
713 {
714   return CwiseBinaryOp<internal::scalar_difference_op<typename DenseDerived::Scalar,typename SparseDerived::Scalar>, const DenseDerived, const SparseDerived>(a.derived(), b.derived());
715 }
716 
717 template<typename SparseDerived, typename DenseDerived>
718 EIGEN_STRONG_INLINE const CwiseBinaryOp<internal::scalar_difference_op<typename SparseDerived::Scalar,typename DenseDerived::Scalar>, const SparseDerived, const DenseDerived>
719 operator-(const SparseMatrixBase<SparseDerived> &a, const MatrixBase<DenseDerived> &b)
720 {
721   return CwiseBinaryOp<internal::scalar_difference_op<typename SparseDerived::Scalar,typename DenseDerived::Scalar>, const SparseDerived, const DenseDerived>(a.derived(), b.derived());
722 }
723 
724 } // end namespace Eigen
725 
726 #endif // EIGEN_SPARSE_CWISE_BINARY_OP_H
727