|  | // This file is part of Eigen, a lightweight C++ template library | 
|  | // for linear algebra. | 
|  | // | 
|  | // Copyright (C) 2008-2014 Gael Guennebaud <gael.guennebaud@inria.fr> | 
|  | // | 
|  | // This Source Code Form is subject to the terms of the Mozilla | 
|  | // Public License v. 2.0. If a copy of the MPL was not distributed | 
|  | // with this file, You can obtain one at http://mozilla.org/MPL/2.0/. | 
|  |  | 
|  | #ifndef EIGEN_SPARSEASSIGN_H | 
|  | #define EIGEN_SPARSEASSIGN_H | 
|  |  | 
|  | // IWYU pragma: private | 
|  | #include "./InternalHeaderCheck.h" | 
|  |  | 
|  | namespace Eigen { | 
|  |  | 
|  | template <typename Derived> | 
|  | template <typename OtherDerived> | 
|  | Derived &SparseMatrixBase<Derived>::operator=(const EigenBase<OtherDerived> &other) { | 
|  | internal::call_assignment_no_alias(derived(), other.derived()); | 
|  | return derived(); | 
|  | } | 
|  |  | 
|  | template <typename Derived> | 
|  | template <typename OtherDerived> | 
|  | Derived &SparseMatrixBase<Derived>::operator=(const ReturnByValue<OtherDerived> &other) { | 
|  | // TODO use the evaluator mechanism | 
|  | other.evalTo(derived()); | 
|  | return derived(); | 
|  | } | 
|  |  | 
|  | template <typename Derived> | 
|  | template <typename OtherDerived> | 
|  | inline Derived &SparseMatrixBase<Derived>::operator=(const SparseMatrixBase<OtherDerived> &other) { | 
|  | // by default sparse evaluation do not alias, so we can safely bypass the generic call_assignment routine | 
|  | internal::Assignment<Derived, OtherDerived, internal::assign_op<Scalar, typename OtherDerived::Scalar>>::run( | 
|  | derived(), other.derived(), internal::assign_op<Scalar, typename OtherDerived::Scalar>()); | 
|  | return derived(); | 
|  | } | 
|  |  | 
|  | template <typename Derived> | 
|  | inline Derived &SparseMatrixBase<Derived>::operator=(const Derived &other) { | 
|  | internal::call_assignment_no_alias(derived(), other.derived()); | 
|  | return derived(); | 
|  | } | 
|  |  | 
|  | namespace internal { | 
|  |  | 
|  | template <> | 
|  | struct storage_kind_to_evaluator_kind<Sparse> { | 
|  | typedef IteratorBased Kind; | 
|  | }; | 
|  |  | 
|  | template <> | 
|  | struct storage_kind_to_shape<Sparse> { | 
|  | typedef SparseShape Shape; | 
|  | }; | 
|  |  | 
|  | struct Sparse2Sparse {}; | 
|  | struct Sparse2Dense {}; | 
|  |  | 
|  | template <> | 
|  | struct AssignmentKind<SparseShape, SparseShape> { | 
|  | typedef Sparse2Sparse Kind; | 
|  | }; | 
|  | template <> | 
|  | struct AssignmentKind<SparseShape, SparseTriangularShape> { | 
|  | typedef Sparse2Sparse Kind; | 
|  | }; | 
|  | template <> | 
|  | struct AssignmentKind<DenseShape, SparseShape> { | 
|  | typedef Sparse2Dense Kind; | 
|  | }; | 
|  | template <> | 
|  | struct AssignmentKind<DenseShape, SparseTriangularShape> { | 
|  | typedef Sparse2Dense Kind; | 
|  | }; | 
|  |  | 
|  | template <typename DstXprType, typename SrcXprType> | 
|  | void assign_sparse_to_sparse(DstXprType &dst, const SrcXprType &src) { | 
|  | typedef typename DstXprType::Scalar Scalar; | 
|  | typedef internal::evaluator<DstXprType> DstEvaluatorType; | 
|  | typedef internal::evaluator<SrcXprType> SrcEvaluatorType; | 
|  |  | 
|  | SrcEvaluatorType srcEvaluator(src); | 
|  |  | 
|  | constexpr bool transpose = (DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit); | 
|  | const Index outerEvaluationSize = (SrcEvaluatorType::Flags & RowMajorBit) ? src.rows() : src.cols(); | 
|  |  | 
|  | Index reserveSize = 0; | 
|  | for (Index j = 0; j < outerEvaluationSize; ++j) | 
|  | for (typename SrcEvaluatorType::InnerIterator it(srcEvaluator, j); it; ++it) reserveSize++; | 
|  |  | 
|  | if ((!transpose) && src.isRValue()) { | 
|  | // eval without temporary | 
|  | dst.resize(src.rows(), src.cols()); | 
|  | dst.setZero(); | 
|  | dst.reserve(reserveSize); | 
|  | for (Index j = 0; j < outerEvaluationSize; ++j) { | 
|  | dst.startVec(j); | 
|  | for (typename SrcEvaluatorType::InnerIterator it(srcEvaluator, j); it; ++it) { | 
|  | Scalar v = it.value(); | 
|  | dst.insertBackByOuterInner(j, it.index()) = v; | 
|  | } | 
|  | } | 
|  | dst.finalize(); | 
|  | } else { | 
|  | // eval through a temporary | 
|  | eigen_assert((((internal::traits<DstXprType>::SupportedAccessPatterns & OuterRandomAccessPattern) == | 
|  | OuterRandomAccessPattern) || | 
|  | (!((DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit)))) && | 
|  | "the transpose operation is supposed to be handled in SparseMatrix::operator="); | 
|  |  | 
|  | enum { Flip = (DstEvaluatorType::Flags & RowMajorBit) != (SrcEvaluatorType::Flags & RowMajorBit) }; | 
|  |  | 
|  | DstXprType temp(src.rows(), src.cols()); | 
|  |  | 
|  | temp.reserve(reserveSize); | 
|  | for (Index j = 0; j < outerEvaluationSize; ++j) { | 
|  | temp.startVec(j); | 
|  | for (typename SrcEvaluatorType::InnerIterator it(srcEvaluator, j); it; ++it) { | 
|  | Scalar v = it.value(); | 
|  | temp.insertBackByOuterInner(Flip ? it.index() : j, Flip ? j : it.index()) = v; | 
|  | } | 
|  | } | 
|  | temp.finalize(); | 
|  |  | 
|  | dst = temp.markAsRValue(); | 
|  | } | 
|  | } | 
|  |  | 
|  | // Generic Sparse to Sparse assignment | 
|  | template <typename DstXprType, typename SrcXprType, typename Functor> | 
|  | struct Assignment<DstXprType, SrcXprType, Functor, Sparse2Sparse> { | 
|  | static void run(DstXprType &dst, const SrcXprType &src, | 
|  | const internal::assign_op<typename DstXprType::Scalar, typename SrcXprType::Scalar> & /*func*/) { | 
|  | assign_sparse_to_sparse(dst.derived(), src.derived()); | 
|  | } | 
|  | }; | 
|  |  | 
|  | // Generic Sparse to Dense assignment | 
|  | template <typename DstXprType, typename SrcXprType, typename Functor, typename Weak> | 
|  | struct Assignment<DstXprType, SrcXprType, Functor, Sparse2Dense, Weak> { | 
|  | static void run(DstXprType &dst, const SrcXprType &src, const Functor &func) { | 
|  | if (internal::is_same<Functor, | 
|  | internal::assign_op<typename DstXprType::Scalar, typename SrcXprType::Scalar>>::value) | 
|  | dst.setZero(); | 
|  |  | 
|  | internal::evaluator<SrcXprType> srcEval(src); | 
|  | resize_if_allowed(dst, src, func); | 
|  | internal::evaluator<DstXprType> dstEval(dst); | 
|  |  | 
|  | const Index outerEvaluationSize = (internal::evaluator<SrcXprType>::Flags & RowMajorBit) ? src.rows() : src.cols(); | 
|  | for (Index j = 0; j < outerEvaluationSize; ++j) | 
|  | for (typename internal::evaluator<SrcXprType>::InnerIterator i(srcEval, j); i; ++i) | 
|  | func.assignCoeff(dstEval.coeffRef(i.row(), i.col()), i.value()); | 
|  | } | 
|  | }; | 
|  |  | 
|  | // Specialization for dense ?= dense +/- sparse and dense ?= sparse +/- dense | 
|  | template <typename DstXprType, typename Func1, typename Func2> | 
|  | struct assignment_from_dense_op_sparse { | 
|  | template <typename SrcXprType, typename InitialFunc> | 
|  | static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void run(DstXprType &dst, const SrcXprType &src, | 
|  | const InitialFunc & /*func*/) { | 
|  | #ifdef EIGEN_SPARSE_ASSIGNMENT_FROM_DENSE_OP_SPARSE_PLUGIN | 
|  | EIGEN_SPARSE_ASSIGNMENT_FROM_DENSE_OP_SPARSE_PLUGIN | 
|  | #endif | 
|  |  | 
|  | call_assignment_no_alias(dst, src.lhs(), Func1()); | 
|  | call_assignment_no_alias(dst, src.rhs(), Func2()); | 
|  | } | 
|  |  | 
|  | // Specialization for dense1 = sparse + dense2; -> dense1 = dense2; dense1 += sparse; | 
|  | template <typename Lhs, typename Rhs, typename Scalar> | 
|  | static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE | 
|  | std::enable_if_t<internal::is_same<typename internal::evaluator_traits<Rhs>::Shape, DenseShape>::value> | 
|  | run(DstXprType &dst, const CwiseBinaryOp<internal::scalar_sum_op<Scalar, Scalar>, const Lhs, const Rhs> &src, | 
|  | const internal::assign_op<typename DstXprType::Scalar, Scalar> & /*func*/) { | 
|  | #ifdef EIGEN_SPARSE_ASSIGNMENT_FROM_SPARSE_ADD_DENSE_PLUGIN | 
|  | EIGEN_SPARSE_ASSIGNMENT_FROM_SPARSE_ADD_DENSE_PLUGIN | 
|  | #endif | 
|  |  | 
|  | // Apply the dense matrix first, then the sparse one. | 
|  | call_assignment_no_alias(dst, src.rhs(), Func1()); | 
|  | call_assignment_no_alias(dst, src.lhs(), Func2()); | 
|  | } | 
|  |  | 
|  | // Specialization for dense1 = sparse - dense2; -> dense1 = -dense2; dense1 += sparse; | 
|  | template <typename Lhs, typename Rhs, typename Scalar> | 
|  | static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE | 
|  | std::enable_if_t<internal::is_same<typename internal::evaluator_traits<Rhs>::Shape, DenseShape>::value> | 
|  | run(DstXprType &dst, | 
|  | const CwiseBinaryOp<internal::scalar_difference_op<Scalar, Scalar>, const Lhs, const Rhs> &src, | 
|  | const internal::assign_op<typename DstXprType::Scalar, Scalar> & /*func*/) { | 
|  | #ifdef EIGEN_SPARSE_ASSIGNMENT_FROM_SPARSE_SUB_DENSE_PLUGIN | 
|  | EIGEN_SPARSE_ASSIGNMENT_FROM_SPARSE_SUB_DENSE_PLUGIN | 
|  | #endif | 
|  |  | 
|  | // Apply the dense matrix first, then the sparse one. | 
|  | call_assignment_no_alias(dst, -src.rhs(), Func1()); | 
|  | call_assignment_no_alias(dst, src.lhs(), add_assign_op<typename DstXprType::Scalar, typename Lhs::Scalar>()); | 
|  | } | 
|  | }; | 
|  |  | 
|  | #define EIGEN_CATCH_ASSIGN_DENSE_OP_SPARSE(ASSIGN_OP, BINOP, ASSIGN_OP2)                                        \ | 
|  | template <typename DstXprType, typename Lhs, typename Rhs, typename Scalar>                                   \ | 
|  | struct Assignment<                                                                                            \ | 
|  | DstXprType, CwiseBinaryOp<internal::BINOP<Scalar, Scalar>, const Lhs, const Rhs>,                         \ | 
|  | internal::ASSIGN_OP<typename DstXprType::Scalar, Scalar>, Sparse2Dense,                                   \ | 
|  | std::enable_if_t<internal::is_same<typename internal::evaluator_traits<Lhs>::Shape, DenseShape>::value || \ | 
|  | internal::is_same<typename internal::evaluator_traits<Rhs>::Shape, DenseShape>::value>>  \ | 
|  | : assignment_from_dense_op_sparse<DstXprType,                                                             \ | 
|  | internal::ASSIGN_OP<typename DstXprType::Scalar, typename Lhs::Scalar>, \ | 
|  | internal::ASSIGN_OP2<typename DstXprType::Scalar, typename Rhs::Scalar>> {} | 
|  |  | 
|  | EIGEN_CATCH_ASSIGN_DENSE_OP_SPARSE(assign_op, scalar_sum_op, add_assign_op); | 
|  | EIGEN_CATCH_ASSIGN_DENSE_OP_SPARSE(add_assign_op, scalar_sum_op, add_assign_op); | 
|  | EIGEN_CATCH_ASSIGN_DENSE_OP_SPARSE(sub_assign_op, scalar_sum_op, sub_assign_op); | 
|  |  | 
|  | EIGEN_CATCH_ASSIGN_DENSE_OP_SPARSE(assign_op, scalar_difference_op, sub_assign_op); | 
|  | EIGEN_CATCH_ASSIGN_DENSE_OP_SPARSE(add_assign_op, scalar_difference_op, sub_assign_op); | 
|  | EIGEN_CATCH_ASSIGN_DENSE_OP_SPARSE(sub_assign_op, scalar_difference_op, add_assign_op); | 
|  |  | 
|  | // Specialization for "dst = dec.solve(rhs)" | 
|  | // NOTE we need to specialize it for Sparse2Sparse to avoid ambiguous specialization error | 
|  | template <typename DstXprType, typename DecType, typename RhsType, typename Scalar> | 
|  | struct Assignment<DstXprType, Solve<DecType, RhsType>, internal::assign_op<Scalar, Scalar>, Sparse2Sparse> { | 
|  | typedef Solve<DecType, RhsType> SrcXprType; | 
|  | static void run(DstXprType &dst, const SrcXprType &src, const internal::assign_op<Scalar, Scalar> &) { | 
|  | Index dstRows = src.rows(); | 
|  | Index dstCols = src.cols(); | 
|  | if ((dst.rows() != dstRows) || (dst.cols() != dstCols)) dst.resize(dstRows, dstCols); | 
|  |  | 
|  | src.dec()._solve_impl(src.rhs(), dst); | 
|  | } | 
|  | }; | 
|  |  | 
|  | struct Diagonal2Sparse {}; | 
|  |  | 
|  | template <> | 
|  | struct AssignmentKind<SparseShape, DiagonalShape> { | 
|  | typedef Diagonal2Sparse Kind; | 
|  | }; | 
|  |  | 
|  | template <typename DstXprType, typename SrcXprType, typename Functor> | 
|  | struct Assignment<DstXprType, SrcXprType, Functor, Diagonal2Sparse> { | 
|  | typedef typename DstXprType::StorageIndex StorageIndex; | 
|  | typedef typename DstXprType::Scalar Scalar; | 
|  |  | 
|  | template <int Options, typename AssignFunc> | 
|  | static void run(SparseMatrix<Scalar, Options, StorageIndex> &dst, const SrcXprType &src, const AssignFunc &func) { | 
|  | dst.assignDiagonal(src.diagonal(), func); | 
|  | } | 
|  |  | 
|  | template <typename DstDerived> | 
|  | static void run(SparseMatrixBase<DstDerived> &dst, const SrcXprType &src, | 
|  | const internal::assign_op<typename DstXprType::Scalar, typename SrcXprType::Scalar> & /*func*/) { | 
|  | dst.derived().diagonal() = src.diagonal(); | 
|  | } | 
|  |  | 
|  | template <typename DstDerived> | 
|  | static void run(SparseMatrixBase<DstDerived> &dst, const SrcXprType &src, | 
|  | const internal::add_assign_op<typename DstXprType::Scalar, typename SrcXprType::Scalar> & /*func*/) { | 
|  | dst.derived().diagonal() += src.diagonal(); | 
|  | } | 
|  |  | 
|  | template <typename DstDerived> | 
|  | static void run(SparseMatrixBase<DstDerived> &dst, const SrcXprType &src, | 
|  | const internal::sub_assign_op<typename DstXprType::Scalar, typename SrcXprType::Scalar> & /*func*/) { | 
|  | dst.derived().diagonal() -= src.diagonal(); | 
|  | } | 
|  | }; | 
|  | }  // end namespace internal | 
|  |  | 
|  | }  // end namespace Eigen | 
|  |  | 
|  | #endif  // EIGEN_SPARSEASSIGN_H |