mirror of
https://gitlab.com/libeigen/eigen.git
synced 2026-04-10 11:34:33 +08:00
939 lines
36 KiB
C++
939 lines
36 KiB
C++
// This file is part of Eigen, a lightweight C++ template library
|
|
// for linear algebra.
|
|
//
|
|
// Copyright (C) 2008-2014 Gael Guennebaud <gael.guennebaud@inria.fr>
|
|
//
|
|
// This Source Code Form is subject to the terms of the Mozilla
|
|
// Public License v. 2.0. If a copy of the MPL was not distributed
|
|
// with this file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
|
|
|
#ifndef EIGEN_SPARSE_CWISE_BINARY_OP_H
|
|
#define EIGEN_SPARSE_CWISE_BINARY_OP_H
|
|
|
|
// IWYU pragma: private
|
|
#include "./InternalHeaderCheck.h"
|
|
|
|
namespace Eigen {
|
|
|
|
// Here we have to handle 3 cases:
|
|
// 1 - sparse op dense
|
|
// 2 - dense op sparse
|
|
// 3 - sparse op sparse
|
|
// We also need to implement a 4th iterator for:
|
|
// 4 - dense op dense
|
|
// Finally, we also need to distinguish between the product and other operations :
|
|
// configuration returned mode
|
|
// 1 - sparse op dense product sparse
|
|
// generic dense
|
|
// 2 - dense op sparse product sparse
|
|
// generic dense
|
|
// 3 - sparse op sparse product sparse
|
|
// generic sparse
|
|
// 4 - dense op dense product dense
|
|
// generic dense
|
|
//
|
|
// TODO: to ease compiler job, we could specialize product/quotient with a scalar
|
|
// and fallback to cwise-unary evaluator using bind1st_op and bind2nd_op.
|
|
|
|
template <typename BinaryOp, typename Lhs, typename Rhs>
|
|
class CwiseBinaryOpImpl<BinaryOp, Lhs, Rhs, Sparse> : public SparseMatrixBase<CwiseBinaryOp<BinaryOp, Lhs, Rhs> > {
|
|
public:
|
|
typedef CwiseBinaryOp<BinaryOp, Lhs, Rhs> Derived;
|
|
typedef SparseMatrixBase<Derived> Base;
|
|
EIGEN_SPARSE_PUBLIC_INTERFACE(Derived)
|
|
EIGEN_STATIC_ASSERT(((!internal::is_same<typename internal::traits<Lhs>::StorageKind,
|
|
typename internal::traits<Rhs>::StorageKind>::value) ||
|
|
((internal::evaluator<Lhs>::Flags & RowMajorBit) ==
|
|
(internal::evaluator<Rhs>::Flags & RowMajorBit))),
|
|
THE_STORAGE_ORDER_OF_BOTH_SIDES_MUST_MATCH)
|
|
};
|
|
|
|
namespace internal {
|
|
|
|
// The default evaluator performs an "arithmetic" operation on two input arrays.
|
|
// Given input arrays 'lhs' and 'rhs' and binary functor 'func',
|
|
// the sparse destination array 'dst' is evaluated as follows:
|
|
// if lhs(i,j) and rhs(i,j) are present, dst(i,j) = func(lhs(i,j), rhs(i,j))
|
|
// if lhs(i,j) is present and rhs(i,j) is null, dst(i,j) = func(lhs(i,j), 0)
|
|
// if lhs(i,j) is null and rhs(i,j) is present, dst(i,j) = func(0, rhs(i,j))
|
|
|
|
// Generic "sparse OP sparse"
|
|
template <typename XprType>
|
|
struct binary_sparse_evaluator;
|
|
|
|
template <typename BinaryOp, typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>, IteratorBased, IteratorBased>
|
|
: evaluator_base<CwiseBinaryOp<BinaryOp, Lhs, Rhs> > {
|
|
protected:
|
|
typedef typename evaluator<Lhs>::InnerIterator LhsIterator;
|
|
typedef typename evaluator<Rhs>::InnerIterator RhsIterator;
|
|
typedef CwiseBinaryOp<BinaryOp, Lhs, Rhs> XprType;
|
|
typedef typename traits<XprType>::Scalar Scalar;
|
|
typedef typename XprType::StorageIndex StorageIndex;
|
|
|
|
public:
|
|
class InnerIterator {
|
|
public:
|
|
EIGEN_STRONG_INLINE InnerIterator(const binary_evaluator& aEval, Index outer)
|
|
: m_lhsIter(aEval.m_lhsImpl, outer),
|
|
m_rhsIter(aEval.m_rhsImpl, outer),
|
|
m_functor(aEval.m_functor),
|
|
m_value(Scalar(0)) {
|
|
this->operator++();
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE InnerIterator& operator++() {
|
|
if (m_lhsIter && m_rhsIter && (m_lhsIter.index() == m_rhsIter.index())) {
|
|
m_id = m_lhsIter.index();
|
|
m_value = m_functor(m_lhsIter.value(), m_rhsIter.value());
|
|
++m_lhsIter;
|
|
++m_rhsIter;
|
|
} else if (m_lhsIter && (!m_rhsIter || (m_lhsIter.index() < m_rhsIter.index()))) {
|
|
m_id = m_lhsIter.index();
|
|
m_value = m_functor(m_lhsIter.value(), Scalar(0));
|
|
++m_lhsIter;
|
|
} else if (m_rhsIter && (!m_lhsIter || (m_lhsIter.index() > m_rhsIter.index()))) {
|
|
m_id = m_rhsIter.index();
|
|
m_value = m_functor(Scalar(0), m_rhsIter.value());
|
|
++m_rhsIter;
|
|
} else {
|
|
m_id = -1;
|
|
}
|
|
return *this;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE Scalar value() const { return m_value; }
|
|
|
|
EIGEN_STRONG_INLINE StorageIndex index() const { return m_id; }
|
|
EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
|
|
EIGEN_STRONG_INLINE Index row() const { return Lhs::IsRowMajor ? m_lhsIter.row() : index(); }
|
|
EIGEN_STRONG_INLINE Index col() const { return Lhs::IsRowMajor ? index() : m_lhsIter.col(); }
|
|
|
|
EIGEN_STRONG_INLINE operator bool() const { return m_id >= 0; }
|
|
|
|
protected:
|
|
LhsIterator m_lhsIter;
|
|
RhsIterator m_rhsIter;
|
|
const BinaryOp& m_functor;
|
|
Scalar m_value;
|
|
StorageIndex m_id;
|
|
};
|
|
|
|
enum {
|
|
CoeffReadCost =
|
|
int(evaluator<Lhs>::CoeffReadCost) + int(evaluator<Rhs>::CoeffReadCost) + int(functor_traits<BinaryOp>::Cost),
|
|
Flags = XprType::Flags
|
|
};
|
|
|
|
explicit binary_evaluator(const XprType& xpr) : m_functor(xpr.functor()), m_lhsImpl(xpr.lhs()), m_rhsImpl(xpr.rhs()) {
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
|
|
}
|
|
|
|
inline Index nonZerosEstimate() const { return m_lhsImpl.nonZerosEstimate() + m_rhsImpl.nonZerosEstimate(); }
|
|
|
|
protected:
|
|
const BinaryOp m_functor;
|
|
evaluator<Lhs> m_lhsImpl;
|
|
evaluator<Rhs> m_rhsImpl;
|
|
};
|
|
|
|
// dense op sparse
|
|
template <typename BinaryOp, typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>, IndexBased, IteratorBased>
|
|
: evaluator_base<CwiseBinaryOp<BinaryOp, Lhs, Rhs> > {
|
|
protected:
|
|
typedef typename evaluator<Rhs>::InnerIterator RhsIterator;
|
|
typedef CwiseBinaryOp<BinaryOp, Lhs, Rhs> XprType;
|
|
typedef typename traits<XprType>::Scalar Scalar;
|
|
typedef typename XprType::StorageIndex StorageIndex;
|
|
|
|
public:
|
|
class InnerIterator {
|
|
enum { IsRowMajor = (int(Rhs::Flags) & RowMajorBit) == RowMajorBit };
|
|
|
|
public:
|
|
EIGEN_STRONG_INLINE InnerIterator(const binary_evaluator& aEval, Index outer)
|
|
: m_lhsEval(aEval.m_lhsImpl),
|
|
m_rhsIter(aEval.m_rhsImpl, outer),
|
|
m_functor(aEval.m_functor),
|
|
m_value(0),
|
|
m_id(-1),
|
|
m_innerSize(aEval.m_expr.rhs().innerSize()) {
|
|
this->operator++();
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE InnerIterator& operator++() {
|
|
++m_id;
|
|
if (m_id < m_innerSize) {
|
|
Scalar lhsVal = m_lhsEval.coeff(IsRowMajor ? m_rhsIter.outer() : m_id, IsRowMajor ? m_id : m_rhsIter.outer());
|
|
if (m_rhsIter && m_rhsIter.index() == m_id) {
|
|
m_value = m_functor(lhsVal, m_rhsIter.value());
|
|
++m_rhsIter;
|
|
} else
|
|
m_value = m_functor(lhsVal, Scalar(0));
|
|
}
|
|
|
|
return *this;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE Scalar value() const {
|
|
eigen_internal_assert(m_id < m_innerSize);
|
|
return m_value;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE StorageIndex index() const { return m_id; }
|
|
EIGEN_STRONG_INLINE Index outer() const { return m_rhsIter.outer(); }
|
|
EIGEN_STRONG_INLINE Index row() const { return IsRowMajor ? m_rhsIter.outer() : m_id; }
|
|
EIGEN_STRONG_INLINE Index col() const { return IsRowMajor ? m_id : m_rhsIter.outer(); }
|
|
|
|
EIGEN_STRONG_INLINE operator bool() const { return m_id < m_innerSize; }
|
|
|
|
protected:
|
|
const evaluator<Lhs>& m_lhsEval;
|
|
RhsIterator m_rhsIter;
|
|
const BinaryOp& m_functor;
|
|
Scalar m_value;
|
|
StorageIndex m_id;
|
|
StorageIndex m_innerSize;
|
|
};
|
|
|
|
enum {
|
|
CoeffReadCost =
|
|
int(evaluator<Lhs>::CoeffReadCost) + int(evaluator<Rhs>::CoeffReadCost) + int(functor_traits<BinaryOp>::Cost),
|
|
Flags = XprType::Flags
|
|
};
|
|
|
|
explicit binary_evaluator(const XprType& xpr)
|
|
: m_functor(xpr.functor()), m_lhsImpl(xpr.lhs()), m_rhsImpl(xpr.rhs()), m_expr(xpr) {
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
|
|
}
|
|
|
|
inline Index nonZerosEstimate() const { return m_expr.size(); }
|
|
|
|
protected:
|
|
const BinaryOp m_functor;
|
|
evaluator<Lhs> m_lhsImpl;
|
|
evaluator<Rhs> m_rhsImpl;
|
|
const XprType& m_expr;
|
|
};
|
|
|
|
// sparse op dense
|
|
template <typename BinaryOp, typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>, IteratorBased, IndexBased>
|
|
: evaluator_base<CwiseBinaryOp<BinaryOp, Lhs, Rhs> > {
|
|
protected:
|
|
typedef typename evaluator<Lhs>::InnerIterator LhsIterator;
|
|
typedef CwiseBinaryOp<BinaryOp, Lhs, Rhs> XprType;
|
|
typedef typename traits<XprType>::Scalar Scalar;
|
|
typedef typename XprType::StorageIndex StorageIndex;
|
|
|
|
public:
|
|
class InnerIterator {
|
|
enum { IsRowMajor = (int(Lhs::Flags) & RowMajorBit) == RowMajorBit };
|
|
|
|
public:
|
|
EIGEN_STRONG_INLINE InnerIterator(const binary_evaluator& aEval, Index outer)
|
|
: m_lhsIter(aEval.m_lhsImpl, outer),
|
|
m_rhsEval(aEval.m_rhsImpl),
|
|
m_functor(aEval.m_functor),
|
|
m_value(0),
|
|
m_id(-1),
|
|
m_innerSize(aEval.m_expr.lhs().innerSize()) {
|
|
this->operator++();
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE InnerIterator& operator++() {
|
|
++m_id;
|
|
if (m_id < m_innerSize) {
|
|
Scalar rhsVal = m_rhsEval.coeff(IsRowMajor ? m_lhsIter.outer() : m_id, IsRowMajor ? m_id : m_lhsIter.outer());
|
|
if (m_lhsIter && m_lhsIter.index() == m_id) {
|
|
m_value = m_functor(m_lhsIter.value(), rhsVal);
|
|
++m_lhsIter;
|
|
} else
|
|
m_value = m_functor(Scalar(0), rhsVal);
|
|
}
|
|
|
|
return *this;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE Scalar value() const {
|
|
eigen_internal_assert(m_id < m_innerSize);
|
|
return m_value;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE StorageIndex index() const { return m_id; }
|
|
EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
|
|
EIGEN_STRONG_INLINE Index row() const { return IsRowMajor ? m_lhsIter.outer() : m_id; }
|
|
EIGEN_STRONG_INLINE Index col() const { return IsRowMajor ? m_id : m_lhsIter.outer(); }
|
|
|
|
EIGEN_STRONG_INLINE operator bool() const { return m_id < m_innerSize; }
|
|
|
|
protected:
|
|
LhsIterator m_lhsIter;
|
|
const evaluator<Rhs>& m_rhsEval;
|
|
const BinaryOp& m_functor;
|
|
Scalar m_value;
|
|
StorageIndex m_id;
|
|
StorageIndex m_innerSize;
|
|
};
|
|
|
|
enum {
|
|
CoeffReadCost =
|
|
int(evaluator<Lhs>::CoeffReadCost) + int(evaluator<Rhs>::CoeffReadCost) + int(functor_traits<BinaryOp>::Cost),
|
|
Flags = XprType::Flags
|
|
};
|
|
|
|
explicit binary_evaluator(const XprType& xpr)
|
|
: m_functor(xpr.functor()), m_lhsImpl(xpr.lhs()), m_rhsImpl(xpr.rhs()), m_expr(xpr) {
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
|
|
}
|
|
|
|
inline Index nonZerosEstimate() const { return m_expr.size(); }
|
|
|
|
protected:
|
|
const BinaryOp m_functor;
|
|
evaluator<Lhs> m_lhsImpl;
|
|
evaluator<Rhs> m_rhsImpl;
|
|
const XprType& m_expr;
|
|
};
|
|
|
|
template <typename T, typename LhsKind = typename evaluator_traits<typename T::Lhs>::Kind,
|
|
typename RhsKind = typename evaluator_traits<typename T::Rhs>::Kind,
|
|
typename LhsScalar = typename traits<typename T::Lhs>::Scalar,
|
|
typename RhsScalar = typename traits<typename T::Rhs>::Scalar>
|
|
struct sparse_conjunction_evaluator;
|
|
|
|
// "sparse .* sparse"
|
|
template <typename T1, typename T2, typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<scalar_product_op<T1, T2>, Lhs, Rhs>, IteratorBased, IteratorBased>
|
|
: sparse_conjunction_evaluator<CwiseBinaryOp<scalar_product_op<T1, T2>, Lhs, Rhs> > {
|
|
typedef CwiseBinaryOp<scalar_product_op<T1, T2>, Lhs, Rhs> XprType;
|
|
typedef sparse_conjunction_evaluator<XprType> Base;
|
|
explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
|
|
};
|
|
// "dense .* sparse"
|
|
template <typename T1, typename T2, typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<scalar_product_op<T1, T2>, Lhs, Rhs>, IndexBased, IteratorBased>
|
|
: sparse_conjunction_evaluator<CwiseBinaryOp<scalar_product_op<T1, T2>, Lhs, Rhs> > {
|
|
typedef CwiseBinaryOp<scalar_product_op<T1, T2>, Lhs, Rhs> XprType;
|
|
typedef sparse_conjunction_evaluator<XprType> Base;
|
|
explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
|
|
};
|
|
// "sparse .* dense"
|
|
template <typename T1, typename T2, typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<scalar_product_op<T1, T2>, Lhs, Rhs>, IteratorBased, IndexBased>
|
|
: sparse_conjunction_evaluator<CwiseBinaryOp<scalar_product_op<T1, T2>, Lhs, Rhs> > {
|
|
typedef CwiseBinaryOp<scalar_product_op<T1, T2>, Lhs, Rhs> XprType;
|
|
typedef sparse_conjunction_evaluator<XprType> Base;
|
|
explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
|
|
};
|
|
|
|
// "sparse ./ dense"
|
|
template <typename T1, typename T2, typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<scalar_quotient_op<T1, T2>, Lhs, Rhs>, IteratorBased, IndexBased>
|
|
: sparse_conjunction_evaluator<CwiseBinaryOp<scalar_quotient_op<T1, T2>, Lhs, Rhs> > {
|
|
typedef CwiseBinaryOp<scalar_quotient_op<T1, T2>, Lhs, Rhs> XprType;
|
|
typedef sparse_conjunction_evaluator<XprType> Base;
|
|
explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
|
|
};
|
|
|
|
// "sparse && sparse"
|
|
template <typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<scalar_boolean_and_op<bool>, Lhs, Rhs>, IteratorBased, IteratorBased>
|
|
: sparse_conjunction_evaluator<CwiseBinaryOp<scalar_boolean_and_op<bool>, Lhs, Rhs> > {
|
|
typedef CwiseBinaryOp<scalar_boolean_and_op<bool>, Lhs, Rhs> XprType;
|
|
typedef sparse_conjunction_evaluator<XprType> Base;
|
|
explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
|
|
};
|
|
// "dense && sparse"
|
|
template <typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<scalar_boolean_and_op<bool>, Lhs, Rhs>, IndexBased, IteratorBased>
|
|
: sparse_conjunction_evaluator<CwiseBinaryOp<scalar_boolean_and_op<bool>, Lhs, Rhs> > {
|
|
typedef CwiseBinaryOp<scalar_boolean_and_op<bool>, Lhs, Rhs> XprType;
|
|
typedef sparse_conjunction_evaluator<XprType> Base;
|
|
explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
|
|
};
|
|
// "sparse && dense"
|
|
template <typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<scalar_boolean_and_op<bool>, Lhs, Rhs>, IteratorBased, IndexBased>
|
|
: sparse_conjunction_evaluator<CwiseBinaryOp<scalar_boolean_and_op<bool>, Lhs, Rhs> > {
|
|
typedef CwiseBinaryOp<scalar_boolean_and_op<bool>, Lhs, Rhs> XprType;
|
|
typedef sparse_conjunction_evaluator<XprType> Base;
|
|
explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
|
|
};
|
|
|
|
// The conjunction "^" evaluator performs a logical "and" or set "intersection" operation on two input arrays.
|
|
// Given input arrays 'lhs' and 'rhs' and binary functor 'func',
|
|
// the sparse destination array 'dst' is evaluated as follows:
|
|
// if lhs(i,j) and rhs(i,j) are present, dst(i,j) = func(lhs(i,j), rhs(i,j))
|
|
// if lhs(i,j) is present and rhs(i,j) is null, dst(i,j) is null
|
|
// if lhs(i,j) is null and rhs(i,j) is present, dst(i,j) is null
|
|
|
|
// "sparse ^ sparse"
|
|
template <typename XprType>
|
|
struct sparse_conjunction_evaluator<XprType, IteratorBased, IteratorBased> : evaluator_base<XprType> {
|
|
protected:
|
|
typedef typename XprType::Functor BinaryOp;
|
|
typedef typename XprType::Lhs LhsArg;
|
|
typedef typename XprType::Rhs RhsArg;
|
|
typedef typename evaluator<LhsArg>::InnerIterator LhsIterator;
|
|
typedef typename evaluator<RhsArg>::InnerIterator RhsIterator;
|
|
typedef typename XprType::StorageIndex StorageIndex;
|
|
typedef typename traits<XprType>::Scalar Scalar;
|
|
|
|
public:
|
|
class InnerIterator {
|
|
public:
|
|
EIGEN_STRONG_INLINE InnerIterator(const sparse_conjunction_evaluator& aEval, Index outer)
|
|
: m_lhsIter(aEval.m_lhsImpl, outer), m_rhsIter(aEval.m_rhsImpl, outer), m_functor(aEval.m_functor) {
|
|
while (m_lhsIter && m_rhsIter && (m_lhsIter.index() != m_rhsIter.index())) {
|
|
if (m_lhsIter.index() < m_rhsIter.index())
|
|
++m_lhsIter;
|
|
else
|
|
++m_rhsIter;
|
|
}
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE InnerIterator& operator++() {
|
|
++m_lhsIter;
|
|
++m_rhsIter;
|
|
while (m_lhsIter && m_rhsIter && (m_lhsIter.index() != m_rhsIter.index())) {
|
|
if (m_lhsIter.index() < m_rhsIter.index())
|
|
++m_lhsIter;
|
|
else
|
|
++m_rhsIter;
|
|
}
|
|
return *this;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE Scalar value() const { return m_functor(m_lhsIter.value(), m_rhsIter.value()); }
|
|
|
|
EIGEN_STRONG_INLINE StorageIndex index() const { return m_lhsIter.index(); }
|
|
EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
|
|
EIGEN_STRONG_INLINE Index row() const { return m_lhsIter.row(); }
|
|
EIGEN_STRONG_INLINE Index col() const { return m_lhsIter.col(); }
|
|
|
|
EIGEN_STRONG_INLINE operator bool() const { return (m_lhsIter && m_rhsIter); }
|
|
|
|
protected:
|
|
LhsIterator m_lhsIter;
|
|
RhsIterator m_rhsIter;
|
|
const BinaryOp& m_functor;
|
|
};
|
|
|
|
enum {
|
|
CoeffReadCost = int(evaluator<LhsArg>::CoeffReadCost) + int(evaluator<RhsArg>::CoeffReadCost) +
|
|
int(functor_traits<BinaryOp>::Cost),
|
|
Flags = XprType::Flags
|
|
};
|
|
|
|
explicit sparse_conjunction_evaluator(const XprType& xpr)
|
|
: m_functor(xpr.functor()), m_lhsImpl(xpr.lhs()), m_rhsImpl(xpr.rhs()) {
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
|
|
}
|
|
|
|
inline Index nonZerosEstimate() const {
|
|
return (std::min)(m_lhsImpl.nonZerosEstimate(), m_rhsImpl.nonZerosEstimate());
|
|
}
|
|
|
|
protected:
|
|
const BinaryOp m_functor;
|
|
evaluator<LhsArg> m_lhsImpl;
|
|
evaluator<RhsArg> m_rhsImpl;
|
|
};
|
|
|
|
// "dense ^ sparse"
|
|
template <typename XprType>
|
|
struct sparse_conjunction_evaluator<XprType, IndexBased, IteratorBased> : evaluator_base<XprType> {
|
|
protected:
|
|
typedef typename XprType::Functor BinaryOp;
|
|
typedef typename XprType::Lhs LhsArg;
|
|
typedef typename XprType::Rhs RhsArg;
|
|
typedef evaluator<LhsArg> LhsEvaluator;
|
|
typedef typename evaluator<RhsArg>::InnerIterator RhsIterator;
|
|
typedef typename XprType::StorageIndex StorageIndex;
|
|
typedef typename traits<XprType>::Scalar Scalar;
|
|
|
|
public:
|
|
class InnerIterator {
|
|
enum { IsRowMajor = (int(RhsArg::Flags) & RowMajorBit) == RowMajorBit };
|
|
|
|
public:
|
|
EIGEN_STRONG_INLINE InnerIterator(const sparse_conjunction_evaluator& aEval, Index outer)
|
|
: m_lhsEval(aEval.m_lhsImpl), m_rhsIter(aEval.m_rhsImpl, outer), m_functor(aEval.m_functor), m_outer(outer) {}
|
|
|
|
EIGEN_STRONG_INLINE InnerIterator& operator++() {
|
|
++m_rhsIter;
|
|
return *this;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE Scalar value() const {
|
|
return m_functor(
|
|
m_lhsEval.coeff(IsRowMajor ? m_outer : m_rhsIter.index(), IsRowMajor ? m_rhsIter.index() : m_outer),
|
|
m_rhsIter.value());
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE StorageIndex index() const { return m_rhsIter.index(); }
|
|
EIGEN_STRONG_INLINE Index outer() const { return m_rhsIter.outer(); }
|
|
EIGEN_STRONG_INLINE Index row() const { return m_rhsIter.row(); }
|
|
EIGEN_STRONG_INLINE Index col() const { return m_rhsIter.col(); }
|
|
|
|
EIGEN_STRONG_INLINE operator bool() const { return m_rhsIter; }
|
|
|
|
protected:
|
|
const LhsEvaluator& m_lhsEval;
|
|
RhsIterator m_rhsIter;
|
|
const BinaryOp& m_functor;
|
|
const Index m_outer;
|
|
};
|
|
|
|
enum {
|
|
CoeffReadCost = int(evaluator<LhsArg>::CoeffReadCost) + int(evaluator<RhsArg>::CoeffReadCost) +
|
|
int(functor_traits<BinaryOp>::Cost),
|
|
Flags = XprType::Flags
|
|
};
|
|
|
|
explicit sparse_conjunction_evaluator(const XprType& xpr)
|
|
: m_functor(xpr.functor()), m_lhsImpl(xpr.lhs()), m_rhsImpl(xpr.rhs()) {
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
|
|
}
|
|
|
|
inline Index nonZerosEstimate() const { return m_rhsImpl.nonZerosEstimate(); }
|
|
|
|
protected:
|
|
const BinaryOp m_functor;
|
|
evaluator<LhsArg> m_lhsImpl;
|
|
evaluator<RhsArg> m_rhsImpl;
|
|
};
|
|
|
|
// "sparse ^ dense"
|
|
template <typename XprType>
|
|
struct sparse_conjunction_evaluator<XprType, IteratorBased, IndexBased> : evaluator_base<XprType> {
|
|
protected:
|
|
typedef typename XprType::Functor BinaryOp;
|
|
typedef typename XprType::Lhs LhsArg;
|
|
typedef typename XprType::Rhs RhsArg;
|
|
typedef typename evaluator<LhsArg>::InnerIterator LhsIterator;
|
|
typedef evaluator<RhsArg> RhsEvaluator;
|
|
typedef typename XprType::StorageIndex StorageIndex;
|
|
typedef typename traits<XprType>::Scalar Scalar;
|
|
|
|
public:
|
|
class InnerIterator {
|
|
enum { IsRowMajor = (int(LhsArg::Flags) & RowMajorBit) == RowMajorBit };
|
|
|
|
public:
|
|
EIGEN_STRONG_INLINE InnerIterator(const sparse_conjunction_evaluator& aEval, Index outer)
|
|
: m_lhsIter(aEval.m_lhsImpl, outer), m_rhsEval(aEval.m_rhsImpl), m_functor(aEval.m_functor), m_outer(outer) {}
|
|
|
|
EIGEN_STRONG_INLINE InnerIterator& operator++() {
|
|
++m_lhsIter;
|
|
return *this;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE Scalar value() const {
|
|
return m_functor(m_lhsIter.value(), m_rhsEval.coeff(IsRowMajor ? m_outer : m_lhsIter.index(),
|
|
IsRowMajor ? m_lhsIter.index() : m_outer));
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE StorageIndex index() const { return m_lhsIter.index(); }
|
|
EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
|
|
EIGEN_STRONG_INLINE Index row() const { return m_lhsIter.row(); }
|
|
EIGEN_STRONG_INLINE Index col() const { return m_lhsIter.col(); }
|
|
|
|
EIGEN_STRONG_INLINE operator bool() const { return m_lhsIter; }
|
|
|
|
protected:
|
|
LhsIterator m_lhsIter;
|
|
const evaluator<RhsArg>& m_rhsEval;
|
|
const BinaryOp& m_functor;
|
|
const Index m_outer;
|
|
};
|
|
|
|
enum {
|
|
CoeffReadCost = int(evaluator<LhsArg>::CoeffReadCost) + int(evaluator<RhsArg>::CoeffReadCost) +
|
|
int(functor_traits<BinaryOp>::Cost),
|
|
Flags = XprType::Flags
|
|
};
|
|
|
|
explicit sparse_conjunction_evaluator(const XprType& xpr)
|
|
: m_functor(xpr.functor()), m_lhsImpl(xpr.lhs()), m_rhsImpl(xpr.rhs()) {
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
|
|
}
|
|
|
|
inline Index nonZerosEstimate() const { return m_lhsImpl.nonZerosEstimate(); }
|
|
|
|
protected:
|
|
const BinaryOp m_functor;
|
|
evaluator<LhsArg> m_lhsImpl;
|
|
evaluator<RhsArg> m_rhsImpl;
|
|
};
|
|
|
|
template <typename T, typename LhsKind = typename evaluator_traits<typename T::Lhs>::Kind,
|
|
typename RhsKind = typename evaluator_traits<typename T::Rhs>::Kind,
|
|
typename LhsScalar = typename traits<typename T::Lhs>::Scalar,
|
|
typename RhsScalar = typename traits<typename T::Rhs>::Scalar>
|
|
struct sparse_disjunction_evaluator;
|
|
|
|
// The disjunction "v" evaluator performs a logical "or" or set "union" operation on two input arrays.
|
|
// Given input arrays 'lhs' and 'rhs' and binary functor 'func',
|
|
// the sparse destination array 'dst' is evaluated as follows:
|
|
// if lhs(i,j) and rhs(i,j) are present, dst(i,j) = func(lhs(i,j), rhs(i,j))
|
|
// if lhs(i,j) is present and rhs(i,j) is null, dst(i,j) = lhs(i,j)
|
|
// if lhs(i,j) is null and rhs(i,j) is present, dst(i,j) = rhs(i,j)
|
|
|
|
// "sparse v sparse"
|
|
template <typename XprType>
|
|
struct sparse_disjunction_evaluator<XprType, IteratorBased, IteratorBased> : evaluator_base<XprType> {
|
|
protected:
|
|
typedef typename XprType::Functor BinaryOp;
|
|
typedef typename XprType::Lhs LhsArg;
|
|
typedef typename XprType::Rhs RhsArg;
|
|
typedef typename evaluator<LhsArg>::InnerIterator LhsIterator;
|
|
typedef typename evaluator<RhsArg>::InnerIterator RhsIterator;
|
|
typedef typename XprType::StorageIndex StorageIndex;
|
|
typedef typename traits<XprType>::Scalar Scalar;
|
|
|
|
public:
|
|
class InnerIterator {
|
|
public:
|
|
EIGEN_STRONG_INLINE InnerIterator(const sparse_disjunction_evaluator& aEval, Index outer)
|
|
: m_lhsIter(aEval.m_lhsImpl, outer),
|
|
m_rhsIter(aEval.m_rhsImpl, outer),
|
|
m_functor(aEval.m_functor),
|
|
m_value(Scalar(0)) {
|
|
this->operator++();
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE InnerIterator& operator++() {
|
|
if (m_lhsIter && m_rhsIter && (m_lhsIter.index() == m_rhsIter.index())) {
|
|
m_id = m_lhsIter.index();
|
|
m_value = m_functor(m_lhsIter.value(), m_rhsIter.value());
|
|
++m_lhsIter;
|
|
++m_rhsIter;
|
|
} else if (m_lhsIter && (!m_rhsIter || (m_lhsIter.index() < m_rhsIter.index()))) {
|
|
m_id = m_lhsIter.index();
|
|
m_value = m_lhsIter.value();
|
|
++m_lhsIter;
|
|
} else if (m_rhsIter && (!m_lhsIter || (m_lhsIter.index() > m_rhsIter.index()))) {
|
|
m_id = m_rhsIter.index();
|
|
m_value = m_rhsIter.value();
|
|
++m_rhsIter;
|
|
} else {
|
|
m_id = -1;
|
|
}
|
|
return *this;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE Scalar value() const { return m_value; }
|
|
|
|
EIGEN_STRONG_INLINE StorageIndex index() const { return m_id; }
|
|
EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
|
|
EIGEN_STRONG_INLINE Index row() const { return LhsArg::IsRowMajor ? m_lhsIter.row() : index(); }
|
|
EIGEN_STRONG_INLINE Index col() const { return LhsArg::IsRowMajor ? index() : m_lhsIter.col(); }
|
|
|
|
EIGEN_STRONG_INLINE operator bool() const { return m_id >= 0; }
|
|
|
|
protected:
|
|
LhsIterator m_lhsIter;
|
|
RhsIterator m_rhsIter;
|
|
const BinaryOp& m_functor;
|
|
Scalar m_value;
|
|
StorageIndex m_id;
|
|
};
|
|
|
|
enum {
|
|
CoeffReadCost = int(evaluator<LhsArg>::CoeffReadCost) + int(evaluator<RhsArg>::CoeffReadCost) +
|
|
int(functor_traits<BinaryOp>::Cost),
|
|
Flags = XprType::Flags
|
|
};
|
|
|
|
explicit sparse_disjunction_evaluator(const XprType& xpr)
|
|
: m_functor(xpr.functor()), m_lhsImpl(xpr.lhs()), m_rhsImpl(xpr.rhs()) {
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
|
|
}
|
|
|
|
inline Index nonZerosEstimate() const { return m_lhsImpl.nonZerosEstimate() + m_rhsImpl.nonZerosEstimate(); }
|
|
|
|
protected:
|
|
const BinaryOp m_functor;
|
|
evaluator<LhsArg> m_lhsImpl;
|
|
evaluator<RhsArg> m_rhsImpl;
|
|
};
|
|
|
|
// "dense v sparse"
|
|
template <typename XprType>
|
|
struct sparse_disjunction_evaluator<XprType, IndexBased, IteratorBased> : evaluator_base<XprType> {
|
|
protected:
|
|
typedef typename XprType::Functor BinaryOp;
|
|
typedef typename XprType::Lhs LhsArg;
|
|
typedef typename XprType::Rhs RhsArg;
|
|
typedef evaluator<LhsArg> LhsEvaluator;
|
|
typedef typename evaluator<RhsArg>::InnerIterator RhsIterator;
|
|
typedef typename XprType::StorageIndex StorageIndex;
|
|
typedef typename traits<XprType>::Scalar Scalar;
|
|
|
|
public:
|
|
class InnerIterator {
|
|
enum { IsRowMajor = (int(RhsArg::Flags) & RowMajorBit) == RowMajorBit };
|
|
|
|
public:
|
|
EIGEN_STRONG_INLINE InnerIterator(const sparse_disjunction_evaluator& aEval, Index outer)
|
|
: m_lhsEval(aEval.m_lhsImpl),
|
|
m_rhsIter(aEval.m_rhsImpl, outer),
|
|
m_functor(aEval.m_functor),
|
|
m_value(0),
|
|
m_id(-1),
|
|
m_innerSize(aEval.m_expr.rhs().innerSize()) {
|
|
this->operator++();
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE InnerIterator& operator++() {
|
|
++m_id;
|
|
if (m_id < m_innerSize) {
|
|
Scalar lhsVal = m_lhsEval.coeff(IsRowMajor ? m_rhsIter.outer() : m_id, IsRowMajor ? m_id : m_rhsIter.outer());
|
|
if (m_rhsIter && m_rhsIter.index() == m_id) {
|
|
m_value = m_functor(lhsVal, m_rhsIter.value());
|
|
++m_rhsIter;
|
|
} else
|
|
m_value = lhsVal;
|
|
}
|
|
|
|
return *this;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE Scalar value() const {
|
|
eigen_internal_assert(m_id < m_innerSize);
|
|
return m_value;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE StorageIndex index() const { return m_id; }
|
|
EIGEN_STRONG_INLINE Index outer() const { return m_rhsIter.outer(); }
|
|
EIGEN_STRONG_INLINE Index row() const { return IsRowMajor ? m_rhsIter.outer() : m_id; }
|
|
EIGEN_STRONG_INLINE Index col() const { return IsRowMajor ? m_id : m_rhsIter.outer(); }
|
|
|
|
EIGEN_STRONG_INLINE operator bool() const { return m_id < m_innerSize; }
|
|
|
|
protected:
|
|
const evaluator<LhsArg>& m_lhsEval;
|
|
RhsIterator m_rhsIter;
|
|
const BinaryOp& m_functor;
|
|
Scalar m_value;
|
|
StorageIndex m_id;
|
|
StorageIndex m_innerSize;
|
|
};
|
|
|
|
enum {
|
|
CoeffReadCost = int(evaluator<LhsArg>::CoeffReadCost) + int(evaluator<RhsArg>::CoeffReadCost) +
|
|
int(functor_traits<BinaryOp>::Cost),
|
|
Flags = XprType::Flags
|
|
};
|
|
|
|
explicit sparse_disjunction_evaluator(const XprType& xpr)
|
|
: m_functor(xpr.functor()), m_lhsImpl(xpr.lhs()), m_rhsImpl(xpr.rhs()), m_expr(xpr) {
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
|
|
}
|
|
|
|
inline Index nonZerosEstimate() const { return m_expr.size(); }
|
|
|
|
protected:
|
|
const BinaryOp m_functor;
|
|
evaluator<LhsArg> m_lhsImpl;
|
|
evaluator<RhsArg> m_rhsImpl;
|
|
const XprType& m_expr;
|
|
};
|
|
|
|
// "sparse v dense"
|
|
template <typename XprType>
|
|
struct sparse_disjunction_evaluator<XprType, IteratorBased, IndexBased> : evaluator_base<XprType> {
|
|
protected:
|
|
typedef typename XprType::Functor BinaryOp;
|
|
typedef typename XprType::Lhs LhsArg;
|
|
typedef typename XprType::Rhs RhsArg;
|
|
typedef typename evaluator<LhsArg>::InnerIterator LhsIterator;
|
|
typedef evaluator<RhsArg> RhsEvaluator;
|
|
typedef typename XprType::StorageIndex StorageIndex;
|
|
typedef typename traits<XprType>::Scalar Scalar;
|
|
|
|
public:
|
|
class InnerIterator {
|
|
enum { IsRowMajor = (int(LhsArg::Flags) & RowMajorBit) == RowMajorBit };
|
|
|
|
public:
|
|
EIGEN_STRONG_INLINE InnerIterator(const sparse_disjunction_evaluator& aEval, Index outer)
|
|
: m_lhsIter(aEval.m_lhsImpl, outer),
|
|
m_rhsEval(aEval.m_rhsImpl),
|
|
m_functor(aEval.m_functor),
|
|
m_value(0),
|
|
m_id(-1),
|
|
m_innerSize(aEval.m_expr.lhs().innerSize()) {
|
|
this->operator++();
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE InnerIterator& operator++() {
|
|
++m_id;
|
|
if (m_id < m_innerSize) {
|
|
Scalar rhsVal = m_rhsEval.coeff(IsRowMajor ? m_lhsIter.outer() : m_id, IsRowMajor ? m_id : m_lhsIter.outer());
|
|
if (m_lhsIter && m_lhsIter.index() == m_id) {
|
|
m_value = m_functor(m_lhsIter.value(), rhsVal);
|
|
++m_lhsIter;
|
|
} else
|
|
m_value = rhsVal;
|
|
}
|
|
|
|
return *this;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE Scalar value() const {
|
|
eigen_internal_assert(m_id < m_innerSize);
|
|
return m_value;
|
|
}
|
|
|
|
EIGEN_STRONG_INLINE StorageIndex index() const { return m_id; }
|
|
EIGEN_STRONG_INLINE Index outer() const { return m_lhsIter.outer(); }
|
|
EIGEN_STRONG_INLINE Index row() const { return IsRowMajor ? m_lhsIter.outer() : m_id; }
|
|
EIGEN_STRONG_INLINE Index col() const { return IsRowMajor ? m_id : m_lhsIter.outer(); }
|
|
|
|
EIGEN_STRONG_INLINE operator bool() const { return m_id < m_innerSize; }
|
|
|
|
protected:
|
|
LhsIterator m_lhsIter;
|
|
const evaluator<RhsArg>& m_rhsEval;
|
|
const BinaryOp& m_functor;
|
|
Scalar m_value;
|
|
StorageIndex m_id;
|
|
StorageIndex m_innerSize;
|
|
};
|
|
|
|
enum {
|
|
CoeffReadCost = int(evaluator<LhsArg>::CoeffReadCost) + int(evaluator<RhsArg>::CoeffReadCost) +
|
|
int(functor_traits<BinaryOp>::Cost),
|
|
Flags = XprType::Flags
|
|
};
|
|
|
|
explicit sparse_disjunction_evaluator(const XprType& xpr)
|
|
: m_functor(xpr.functor()), m_lhsImpl(xpr.lhs()), m_rhsImpl(xpr.rhs()), m_expr(xpr) {
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost);
|
|
EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost);
|
|
}
|
|
|
|
inline Index nonZerosEstimate() const { return m_expr.size(); }
|
|
|
|
protected:
|
|
const BinaryOp m_functor;
|
|
evaluator<LhsArg> m_lhsImpl;
|
|
evaluator<RhsArg> m_rhsImpl;
|
|
const XprType& m_expr;
|
|
};
|
|
|
|
// when DupFunc is wrapped with scalar_dup_op, use disjunction evaluator
|
|
template <typename T1, typename T2, typename DupFunc, typename Lhs, typename Rhs>
|
|
struct binary_evaluator<CwiseBinaryOp<scalar_disjunction_op<DupFunc, T1, T2>, Lhs, Rhs>, IteratorBased, IteratorBased>
|
|
: sparse_disjunction_evaluator<CwiseBinaryOp<scalar_disjunction_op<DupFunc, T1, T2>, Lhs, Rhs> > {
|
|
typedef CwiseBinaryOp<scalar_disjunction_op<DupFunc, T1, T2>, Lhs, Rhs> XprType;
|
|
typedef sparse_disjunction_evaluator<XprType> Base;
|
|
explicit binary_evaluator(const XprType& xpr) : Base(xpr) {}
|
|
};
|
|
} // namespace internal
|
|
|
|
/***************************************************************************
|
|
* Implementation of SparseMatrixBase and SparseCwise functions/operators
|
|
***************************************************************************/
|
|
|
|
template <typename Derived>
|
|
template <typename OtherDerived>
|
|
Derived& SparseMatrixBase<Derived>::operator+=(const EigenBase<OtherDerived>& other) {
|
|
call_assignment(derived(), other.derived(), internal::add_assign_op<Scalar, typename OtherDerived::Scalar>());
|
|
return derived();
|
|
}
|
|
|
|
template <typename Derived>
|
|
template <typename OtherDerived>
|
|
Derived& SparseMatrixBase<Derived>::operator-=(const EigenBase<OtherDerived>& other) {
|
|
call_assignment(derived(), other.derived(), internal::sub_assign_op<Scalar, typename OtherDerived::Scalar>());
|
|
return derived();
|
|
}
|
|
|
|
template <typename Derived>
|
|
template <typename OtherDerived>
|
|
EIGEN_STRONG_INLINE Derived& SparseMatrixBase<Derived>::operator-=(const SparseMatrixBase<OtherDerived>& other) {
|
|
return derived() = derived() - other.derived();
|
|
}
|
|
|
|
template <typename Derived>
|
|
template <typename OtherDerived>
|
|
EIGEN_STRONG_INLINE Derived& SparseMatrixBase<Derived>::operator+=(const SparseMatrixBase<OtherDerived>& other) {
|
|
return derived() = derived() + other.derived();
|
|
}
|
|
|
|
template <typename Derived>
|
|
template <typename OtherDerived>
|
|
Derived& SparseMatrixBase<Derived>::operator+=(const DiagonalBase<OtherDerived>& other) {
|
|
call_assignment_no_alias(derived(), other.derived(),
|
|
internal::add_assign_op<Scalar, typename OtherDerived::Scalar>());
|
|
return derived();
|
|
}
|
|
|
|
template <typename Derived>
|
|
template <typename OtherDerived>
|
|
Derived& SparseMatrixBase<Derived>::operator-=(const DiagonalBase<OtherDerived>& other) {
|
|
call_assignment_no_alias(derived(), other.derived(),
|
|
internal::sub_assign_op<Scalar, typename OtherDerived::Scalar>());
|
|
return derived();
|
|
}
|
|
|
|
template <typename Derived>
|
|
template <typename OtherDerived>
|
|
EIGEN_STRONG_INLINE const typename SparseMatrixBase<Derived>::template CwiseProductDenseReturnType<OtherDerived>::Type
|
|
SparseMatrixBase<Derived>::cwiseProduct(const MatrixBase<OtherDerived>& other) const {
|
|
return typename CwiseProductDenseReturnType<OtherDerived>::Type(derived(), other.derived());
|
|
}
|
|
|
|
template <typename DenseDerived, typename SparseDerived>
|
|
EIGEN_STRONG_INLINE const
|
|
CwiseBinaryOp<internal::scalar_sum_op<typename DenseDerived::Scalar, typename SparseDerived::Scalar>,
|
|
const DenseDerived, const SparseDerived>
|
|
operator+(const MatrixBase<DenseDerived>& a, const SparseMatrixBase<SparseDerived>& b) {
|
|
return CwiseBinaryOp<internal::scalar_sum_op<typename DenseDerived::Scalar, typename SparseDerived::Scalar>,
|
|
const DenseDerived, const SparseDerived>(a.derived(), b.derived());
|
|
}
|
|
|
|
template <typename SparseDerived, typename DenseDerived>
|
|
EIGEN_STRONG_INLINE const
|
|
CwiseBinaryOp<internal::scalar_sum_op<typename SparseDerived::Scalar, typename DenseDerived::Scalar>,
|
|
const SparseDerived, const DenseDerived>
|
|
operator+(const SparseMatrixBase<SparseDerived>& a, const MatrixBase<DenseDerived>& b) {
|
|
return CwiseBinaryOp<internal::scalar_sum_op<typename SparseDerived::Scalar, typename DenseDerived::Scalar>,
|
|
const SparseDerived, const DenseDerived>(a.derived(), b.derived());
|
|
}
|
|
|
|
template <typename DenseDerived, typename SparseDerived>
|
|
EIGEN_STRONG_INLINE const
|
|
CwiseBinaryOp<internal::scalar_difference_op<typename DenseDerived::Scalar, typename SparseDerived::Scalar>,
|
|
const DenseDerived, const SparseDerived>
|
|
operator-(const MatrixBase<DenseDerived>& a, const SparseMatrixBase<SparseDerived>& b) {
|
|
return CwiseBinaryOp<internal::scalar_difference_op<typename DenseDerived::Scalar, typename SparseDerived::Scalar>,
|
|
const DenseDerived, const SparseDerived>(a.derived(), b.derived());
|
|
}
|
|
|
|
template <typename SparseDerived, typename DenseDerived>
|
|
EIGEN_STRONG_INLINE const
|
|
CwiseBinaryOp<internal::scalar_difference_op<typename SparseDerived::Scalar, typename DenseDerived::Scalar>,
|
|
const SparseDerived, const DenseDerived>
|
|
operator-(const SparseMatrixBase<SparseDerived>& a, const MatrixBase<DenseDerived>& b) {
|
|
return CwiseBinaryOp<internal::scalar_difference_op<typename SparseDerived::Scalar, typename DenseDerived::Scalar>,
|
|
const SparseDerived, const DenseDerived>(a.derived(), b.derived());
|
|
}
|
|
|
|
} // end namespace Eigen
|
|
|
|
#endif // EIGEN_SPARSE_CWISE_BINARY_OP_H
|