| // This file is part of Eigen, a lightweight C++ template library |
| // for linear algebra. |
| // |
| // Copyright (C) 2011 Benoit Jacob <jacob.benoit.1@gmail.com> |
| // Copyright (C) 2011-2014 Gael Guennebaud <gael.guennebaud@inria.fr> |
| // Copyright (C) 2011-2012 Jitse Niesen <jitse@maths.leeds.ac.uk> |
| // |
| // This Source Code Form is subject to the terms of the Mozilla |
| // Public License v. 2.0. If a copy of the MPL was not distributed |
| // with this file, You can obtain one at http://mozilla.org/MPL/2.0/. |
| // SPDX-License-Identifier: MPL-2.0 |
| |
| #ifndef EIGEN_COREEVALUATORS_H |
| #define EIGEN_COREEVALUATORS_H |
| |
| // IWYU pragma: private |
| #include "./InternalHeaderCheck.h" |
| |
| namespace Eigen { |
| |
| namespace internal { |
| |
| // This class returns the evaluator kind from the expression storage kind. |
| // Default assumes index based accessors |
| template <typename StorageKind> |
| struct storage_kind_to_evaluator_kind { |
| using Kind = IndexBased; |
| }; |
| |
| // This class returns the evaluator shape from the expression storage kind. |
| // It can be Dense, Sparse, Triangular, Diagonal, SelfAdjoint, Band, etc. |
| template <typename StorageKind> |
| struct storage_kind_to_shape; |
| |
| template <> |
| struct storage_kind_to_shape<Dense> { |
| using Shape = DenseShape; |
| }; |
| template <> |
| struct storage_kind_to_shape<SolverStorage> { |
| using Shape = SolverShape; |
| }; |
| template <> |
| struct storage_kind_to_shape<PermutationStorage> { |
| using Shape = PermutationShape; |
| }; |
| template <> |
| struct storage_kind_to_shape<TranspositionsStorage> { |
| using Shape = TranspositionsShape; |
| }; |
| |
| // Evaluators have to be specialized with respect to various criteria such as: |
| // - storage/structure/shape |
| // - scalar type |
| // - etc. |
| // Therefore, we need specialization of evaluator providing additional template arguments for each kind of evaluators. |
| // We currently distinguish the following kind of evaluators: |
| // - unary_evaluator for expressions taking only one arguments (CwiseUnaryOp, CwiseUnaryView, Transpose, |
| // MatrixWrapper, ArrayWrapper, Reverse, Replicate) |
| // - binary_evaluator for expression taking two arguments (CwiseBinaryOp) |
| // - ternary_evaluator for expression taking three arguments (CwiseTernaryOp) |
| // - product_evaluator for linear algebra products (Product); special case of binary_evaluator because it requires |
| // additional tags for dispatching. |
| // - mapbase_evaluator for Map, Block, Ref |
| // - block_evaluator for Block (special dispatching to a mapbase_evaluator or unary_evaluator) |
| |
| template <typename T, typename Arg1Kind = typename evaluator_traits<typename T::Arg1>::Kind, |
| typename Arg2Kind = typename evaluator_traits<typename T::Arg2>::Kind, |
| typename Arg3Kind = typename evaluator_traits<typename T::Arg3>::Kind, |
| typename Arg1Scalar = typename traits<typename T::Arg1>::Scalar, |
| typename Arg2Scalar = typename traits<typename T::Arg2>::Scalar, |
| typename Arg3Scalar = typename traits<typename T::Arg3>::Scalar> |
| struct ternary_evaluator; |
| |
| template <typename T, typename LhsKind = typename evaluator_traits<typename T::Lhs>::Kind, |
| typename RhsKind = typename evaluator_traits<typename T::Rhs>::Kind, |
| typename LhsScalar = typename traits<typename T::Lhs>::Scalar, |
| typename RhsScalar = typename traits<typename T::Rhs>::Scalar> |
| struct binary_evaluator; |
| |
| template <typename T, typename Kind = typename evaluator_traits<typename T::NestedExpression>::Kind, |
| typename Scalar = typename T::Scalar> |
| struct unary_evaluator; |
| |
| // evaluator_traits<T> contains traits for evaluator<T> |
| |
| template <typename T> |
| struct evaluator_traits_base { |
| // by default, get evaluator kind and shape from storage |
| using Kind = typename storage_kind_to_evaluator_kind<typename traits<T>::StorageKind>::Kind; |
| using Shape = typename storage_kind_to_shape<typename traits<T>::StorageKind>::Shape; |
| }; |
| |
| // Default evaluator traits |
| template <typename T> |
| struct evaluator_traits : public evaluator_traits_base<T> {}; |
| |
| template <typename T, typename Shape = typename evaluator_traits<T>::Shape> |
| struct evaluator_assume_aliasing : std::false_type {}; |
| |
| // By default, we assume a unary expression: |
| template <typename T> |
| struct evaluator : public unary_evaluator<T> { |
| using Base = unary_evaluator<T>; |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit evaluator(const T& xpr) : Base(xpr) {} |
| }; |
| |
| // TODO: Think about const-correctness |
| template <typename T> |
| struct evaluator<const T> : evaluator<T> { |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit evaluator(const T& xpr) : evaluator<T>(xpr) {} |
| }; |
| |
| // ---------- base class for all evaluators ---------- |
| |
| template <typename ExpressionType> |
| struct evaluator_base { |
| // TODO: find a way to avoid propagating all these traits. They are currently only needed to handle |
| // outer,inner indices. |
| using ExpressionTraits = traits<ExpressionType>; |
| |
| enum { Alignment = 0 }; |
| // Spell out deleted copy operations instead of inheriting from an empty helper: |
| // an extra base can kill EBO and make complex evaluators larger than they should be. |
| EIGEN_DEVICE_FUNC constexpr evaluator_base() = default; |
| |
| evaluator_base(const evaluator_base&) = delete; |
| evaluator_base& operator=(const evaluator_base&) = delete; |
| }; |
| |
| // -------------------- Matrix and Array -------------------- |
| // |
| // evaluator<PlainObjectBase> is a common base class for the |
| // Matrix and Array evaluators. |
| // Here we directly specialize evaluator. This is not really a unary expression, and it is, by definition, dense, |
| // so no need for more sophisticated dispatching. |
| |
| // this helper permits to completely eliminate m_outerStride if it is known at compiletime. |
| template <typename Scalar, int OuterStride> |
| class plainobjectbase_evaluator_data { |
| public: |
| EIGEN_DEVICE_FUNC constexpr plainobjectbase_evaluator_data(const Scalar* ptr, Index outerStride) : data(ptr) { |
| #ifndef EIGEN_INTERNAL_DEBUGGING |
| EIGEN_UNUSED_VARIABLE(outerStride); |
| #endif |
| eigen_internal_assert(outerStride == OuterStride); |
| } |
| EIGEN_DEVICE_FUNC constexpr Index outerStride() const noexcept { return OuterStride; } |
| const Scalar* data; |
| }; |
| |
| template <typename Scalar> |
| class plainobjectbase_evaluator_data<Scalar, Dynamic> { |
| public: |
| EIGEN_DEVICE_FUNC constexpr plainobjectbase_evaluator_data(const Scalar* ptr, Index outerStride) |
| : data(ptr), m_outerStride(outerStride) {} |
| EIGEN_DEVICE_FUNC constexpr Index outerStride() const { return m_outerStride; } |
| const Scalar* data; |
| |
| protected: |
| Index m_outerStride; |
| }; |
| |
| template <typename Derived> |
| struct evaluator<PlainObjectBase<Derived>> : evaluator_base<Derived> { |
| using PlainObjectType = PlainObjectBase<Derived>; |
| using Scalar = typename PlainObjectType::Scalar; |
| using CoeffReturnType = typename PlainObjectType::CoeffReturnType; |
| |
| enum { |
| IsRowMajor = PlainObjectType::IsRowMajor, |
| IsVectorAtCompileTime = PlainObjectType::IsVectorAtCompileTime, |
| RowsAtCompileTime = PlainObjectType::RowsAtCompileTime, |
| ColsAtCompileTime = PlainObjectType::ColsAtCompileTime, |
| |
| CoeffReadCost = NumTraits<Scalar>::ReadCost, |
| Flags = traits<Derived>::EvaluatorFlags, |
| Alignment = traits<Derived>::Alignment |
| }; |
| enum { |
| // We do not need to know the outer stride for vectors |
| OuterStrideAtCompileTime = IsVectorAtCompileTime ? 0 |
| : int(IsRowMajor) ? ColsAtCompileTime |
| : RowsAtCompileTime |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr evaluator() : m_d(0, OuterStrideAtCompileTime) { |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr explicit evaluator(const PlainObjectType& m) |
| : m_d(m.data(), IsVectorAtCompileTime ? 0 : m.outerStride()) { |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE constexpr CoeffReturnType coeff(Index row, Index col) const { |
| return coeff(getIndex(row, col)); |
| } |
| |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE constexpr CoeffReturnType coeff(Index index) const { return m_d.data[index]; } |
| |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE constexpr Scalar& coeffRef(Index row, Index col) { |
| return coeffRef(getIndex(row, col)); |
| } |
| |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE constexpr Scalar& coeffRef(Index index) { |
| return const_cast<Scalar*>(m_d.data)[index]; |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| return packet<LoadMode, PacketType>(getIndex(row, col)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| return ploadt<PacketType, LoadMode>(m_d.data + index); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index row, Index col, const PacketType& x) { |
| writePacket<StoreMode, PacketType>(getIndex(row, col), x); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index index, const PacketType& x) { |
| pstoret<Scalar, PacketType, StoreMode>(const_cast<Scalar*>(m_d.data) + index, x); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| return packetSegment<LoadMode, PacketType>(getIndex(row, col), begin, count); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| return ploadtSegment<PacketType, LoadMode>(m_d.data + index, begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index row, Index col, const PacketType& x, Index begin, |
| Index count) { |
| writePacketSegment<StoreMode, PacketType>(getIndex(row, col), x, begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index index, const PacketType& x, Index begin, |
| Index count) { |
| pstoretSegment<Scalar, PacketType, StoreMode>(const_cast<Scalar*>(m_d.data) + index, x, begin, count); |
| } |
| |
| protected: |
| plainobjectbase_evaluator_data<Scalar, OuterStrideAtCompileTime> m_d; |
| |
| private: |
| EIGEN_DEVICE_FUNC constexpr Index getIndex(Index row, Index col) const { |
| return IsRowMajor ? row * m_d.outerStride() + col : row + col * m_d.outerStride(); |
| } |
| }; |
| |
| template <typename Scalar, int Rows, int Cols, int Options, int MaxRows, int MaxCols> |
| struct evaluator<Matrix<Scalar, Rows, Cols, Options, MaxRows, MaxCols>> |
| : evaluator<PlainObjectBase<Matrix<Scalar, Rows, Cols, Options, MaxRows, MaxCols>>> { |
| using XprType = Matrix<Scalar, Rows, Cols, Options, MaxRows, MaxCols>; |
| |
| EIGEN_DEVICE_FUNC constexpr evaluator() = default; |
| |
| EIGEN_DEVICE_FUNC constexpr explicit evaluator(const XprType& m) : evaluator<PlainObjectBase<XprType>>(m) {} |
| }; |
| |
| template <typename Scalar, int Rows, int Cols, int Options, int MaxRows, int MaxCols> |
| struct evaluator<Array<Scalar, Rows, Cols, Options, MaxRows, MaxCols>> |
| : evaluator<PlainObjectBase<Array<Scalar, Rows, Cols, Options, MaxRows, MaxCols>>> { |
| using XprType = Array<Scalar, Rows, Cols, Options, MaxRows, MaxCols>; |
| |
| EIGEN_DEVICE_FUNC constexpr evaluator() = default; |
| |
| EIGEN_DEVICE_FUNC constexpr explicit evaluator(const XprType& m) : evaluator<PlainObjectBase<XprType>>(m) {} |
| }; |
| |
| // -------------------- Transpose -------------------- |
| |
| template <typename ArgType> |
| struct unary_evaluator<Transpose<ArgType>, IndexBased> : evaluator_base<Transpose<ArgType>> { |
| using XprType = Transpose<ArgType>; |
| |
| enum { |
| CoeffReadCost = evaluator<ArgType>::CoeffReadCost, |
| Flags = evaluator<ArgType>::Flags ^ RowMajorBit, |
| Alignment = evaluator<ArgType>::Alignment |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit unary_evaluator(const XprType& t) |
| : m_argImpl(t.nestedExpression()) {} |
| |
| using Scalar = typename XprType::Scalar; |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| return m_argImpl.coeff(col, row); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return m_argImpl.coeff(index); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index row, Index col) { |
| return m_argImpl.coeffRef(col, row); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE typename XprType::Scalar& coeffRef(Index index) { |
| return m_argImpl.coeffRef(index); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| return m_argImpl.template packet<LoadMode, PacketType>(col, row); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| return m_argImpl.template packet<LoadMode, PacketType>(index); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index row, Index col, const PacketType& x) { |
| m_argImpl.template writePacket<StoreMode, PacketType>(col, row, x); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index index, const PacketType& x) { |
| m_argImpl.template writePacket<StoreMode, PacketType>(index, x); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(col, row, begin, count); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(index, begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index row, Index col, const PacketType& x, Index begin, |
| Index count) { |
| m_argImpl.template writePacketSegment<StoreMode, PacketType>(col, row, x, begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index index, const PacketType& x, Index begin, |
| Index count) { |
| m_argImpl.template writePacketSegment<StoreMode, PacketType>(index, x, begin, count); |
| } |
| |
| protected: |
| evaluator<ArgType> m_argImpl; |
| }; |
| |
| // -------------------- CwiseNullaryOp -------------------- |
| // Like Matrix and Array, this is not really a unary expression, so we directly specialize evaluator. |
| // Likewise, there is not need to more sophisticated dispatching here. |
| |
| template <typename Scalar, typename NullaryOp, bool has_nullary = has_nullary_operator<NullaryOp>::value, |
| bool has_unary = has_unary_operator<NullaryOp>::value, |
| bool has_binary = has_binary_operator<NullaryOp>::value> |
| struct nullary_wrapper { |
| template <typename IndexType> |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar operator()(const NullaryOp& op, IndexType i, |
| IndexType j) const { |
| return op(i, j); |
| } |
| template <typename IndexType> |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar operator()(const NullaryOp& op, IndexType i) const { |
| return op(i); |
| } |
| |
| template <typename T, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE T packetOp(const NullaryOp& op, IndexType i, IndexType j) const { |
| return op.template packetOp<T>(i, j); |
| } |
| template <typename T, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE T packetOp(const NullaryOp& op, IndexType i) const { |
| return op.template packetOp<T>(i); |
| } |
| }; |
| |
| template <typename Scalar, typename NullaryOp> |
| struct nullary_wrapper<Scalar, NullaryOp, true, false, false> { |
| template <typename IndexType> |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar operator()(const NullaryOp& op, IndexType = 0, |
| IndexType = 0) const { |
| return op(); |
| } |
| template <typename T, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE T packetOp(const NullaryOp& op, IndexType = 0, IndexType = 0) const { |
| return op.template packetOp<T>(); |
| } |
| }; |
| |
| template <typename Scalar, typename NullaryOp> |
| struct nullary_wrapper<Scalar, NullaryOp, false, false, true> { |
| template <typename IndexType> |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar operator()(const NullaryOp& op, IndexType i, |
| IndexType j = 0) const { |
| return op(i, j); |
| } |
| template <typename T, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE T packetOp(const NullaryOp& op, IndexType i, IndexType j = 0) const { |
| return op.template packetOp<T>(i, j); |
| } |
| }; |
| |
| // We need the following specialization for vector-only functors assigned to a runtime vector, |
| // for instance, using linspace and assigning a RowVectorXd to a MatrixXd or even a row of a MatrixXd. |
| // In this case, i==0 and j is used for the actual iteration. |
| template <typename Scalar, typename NullaryOp> |
| struct nullary_wrapper<Scalar, NullaryOp, false, true, false> { |
| template <typename IndexType> |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar operator()(const NullaryOp& op, IndexType i, |
| IndexType j) const { |
| eigen_assert(i == 0 || j == 0); |
| return op(i + j); |
| } |
| template <typename T, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE T packetOp(const NullaryOp& op, IndexType i, IndexType j) const { |
| eigen_assert(i == 0 || j == 0); |
| return op.template packetOp<T>(i + j); |
| } |
| |
| template <typename IndexType> |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar operator()(const NullaryOp& op, IndexType i) const { |
| return op(i); |
| } |
| template <typename T, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE T packetOp(const NullaryOp& op, IndexType i) const { |
| return op.template packetOp<T>(i); |
| } |
| }; |
| |
| template <typename Scalar, typename NullaryOp> |
| struct nullary_wrapper<Scalar, NullaryOp, false, false, false> {}; |
| |
| template <typename NullaryOp, typename PlainObjectType> |
| struct evaluator<CwiseNullaryOp<NullaryOp, PlainObjectType>> |
| : evaluator_base<CwiseNullaryOp<NullaryOp, PlainObjectType>> { |
| using XprType = CwiseNullaryOp<NullaryOp, PlainObjectType>; |
| using PlainObjectTypeCleaned = remove_all_t<PlainObjectType>; |
| |
| enum { |
| CoeffReadCost = functor_traits<NullaryOp>::Cost, |
| |
| Flags = (evaluator<PlainObjectTypeCleaned>::Flags & |
| (HereditaryBits | (functor_has_linear_access<NullaryOp>::value ? LinearAccessBit : 0) | |
| (functor_traits<NullaryOp>::PacketAccess ? PacketAccessBit : 0))) | |
| (functor_traits<NullaryOp>::IsRepeatable ? 0 : EvalBeforeNestingBit), |
| Alignment = AlignedMax |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr explicit evaluator(const XprType& n) : m_functor(n.functor()), m_wrapper() { |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| template <typename IndexType> |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(IndexType row, IndexType col) const { |
| return m_wrapper(m_functor, row, col); |
| } |
| |
| template <typename IndexType> |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(IndexType index) const { |
| return m_wrapper(m_functor, index); |
| } |
| |
| template <int LoadMode, typename PacketType, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(IndexType row, IndexType col) const { |
| return m_wrapper.template packetOp<PacketType>(m_functor, row, col); |
| } |
| |
| template <int LoadMode, typename PacketType, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(IndexType index) const { |
| return m_wrapper.template packetOp<PacketType>(m_functor, index); |
| } |
| |
| template <int LoadMode, typename PacketType, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(IndexType row, IndexType col, Index /*begin*/, |
| Index /*count*/) const { |
| return packet<LoadMode, PacketType, IndexType>(row, col); |
| } |
| |
| template <int LoadMode, typename PacketType, typename IndexType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(IndexType index, Index /*begin*/, |
| Index /*count*/) const { |
| return packet<LoadMode, PacketType, IndexType>(index); |
| } |
| |
| protected: |
| const NullaryOp m_functor; |
| const nullary_wrapper<CoeffReturnType, NullaryOp> m_wrapper; |
| }; |
| |
| // -------------------- CwiseUnaryOp -------------------- |
| |
| template <typename UnaryOp, typename ArgType> |
| struct unary_evaluator<CwiseUnaryOp<UnaryOp, ArgType>, IndexBased> : evaluator_base<CwiseUnaryOp<UnaryOp, ArgType>> { |
| using XprType = CwiseUnaryOp<UnaryOp, ArgType>; |
| |
| enum { |
| CoeffReadCost = int(evaluator<ArgType>::CoeffReadCost) + int(functor_traits<UnaryOp>::Cost), |
| |
| Flags = evaluator<ArgType>::Flags & |
| (HereditaryBits | LinearAccessBit | (functor_traits<UnaryOp>::PacketAccess ? PacketAccessBit : 0)), |
| Alignment = evaluator<ArgType>::Alignment |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit unary_evaluator(const XprType& op) : m_d(op) { |
| EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<UnaryOp>::Cost); |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| return m_d.func()(m_d.argImpl.coeff(row, col)); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return m_d.func()(m_d.argImpl.coeff(index)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| return m_d.func().packetOp(m_d.argImpl.template packet<LoadMode, PacketType>(row, col)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| return m_d.func().packetOp(m_d.argImpl.template packet<LoadMode, PacketType>(index)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| return m_d.func().packetOp(m_d.argImpl.template packetSegment<LoadMode, PacketType>(row, col, begin, count)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| return m_d.func().packetOp(m_d.argImpl.template packetSegment<LoadMode, PacketType>(index, begin, count)); |
| } |
| |
| protected: |
| // this helper permits to completely eliminate the functor if it is empty |
| struct Data { |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Data(const XprType& xpr) |
| : op(xpr.functor()), argImpl(xpr.nestedExpression()) {} |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE const UnaryOp& func() const { return op; } |
| UnaryOp op; |
| evaluator<ArgType> argImpl; |
| }; |
| |
| Data m_d; |
| }; |
| |
| // ----------------------- Casting --------------------- |
| |
| template <typename SrcType, typename DstType, typename ArgType> |
| struct unary_evaluator<CwiseUnaryOp<core_cast_op<SrcType, DstType>, ArgType>, IndexBased> { |
| using CastOp = core_cast_op<SrcType, DstType>; |
| using XprType = CwiseUnaryOp<CastOp, ArgType>; |
| |
| // Use the largest packet type by default |
| using SrcPacketType = typename packet_traits<SrcType>::type; |
| static constexpr int SrcPacketSize = unpacket_traits<SrcPacketType>::size; |
| static constexpr int SrcPacketBytes = SrcPacketSize * sizeof(SrcType); |
| |
| enum { |
| CoeffReadCost = int(evaluator<ArgType>::CoeffReadCost) + int(functor_traits<CastOp>::Cost), |
| PacketAccess = functor_traits<CastOp>::PacketAccess, |
| ActualPacketAccessBit = PacketAccess ? PacketAccessBit : 0, |
| Flags = evaluator<ArgType>::Flags & (HereditaryBits | LinearAccessBit | ActualPacketAccessBit), |
| IsRowMajor = (evaluator<ArgType>::Flags & RowMajorBit), |
| Alignment = evaluator<ArgType>::Alignment |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit unary_evaluator(const XprType& xpr) |
| : m_argImpl(xpr.nestedExpression()), m_rows(xpr.rows()), m_cols(xpr.cols()) { |
| EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<CastOp>::Cost); |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| |
| template <typename DstPacketType> |
| using AltSrcScalarOp = std::enable_if_t<(unpacket_traits<DstPacketType>::size < SrcPacketSize && |
| !find_packet_by_size<SrcType, unpacket_traits<DstPacketType>::size>::value), |
| bool>; |
| template <typename DstPacketType> |
| using SrcPacketArgs1 = |
| std::enable_if_t<(find_packet_by_size<SrcType, unpacket_traits<DstPacketType>::size>::value), bool>; |
| template <typename DstPacketType> |
| using SrcPacketArgs2 = std::enable_if_t<(unpacket_traits<DstPacketType>::size) == (2 * SrcPacketSize), bool>; |
| template <typename DstPacketType> |
| using SrcPacketArgs4 = std::enable_if_t<(unpacket_traits<DstPacketType>::size) == (4 * SrcPacketSize), bool>; |
| template <typename DstPacketType> |
| using SrcPacketArgs8 = std::enable_if_t<(unpacket_traits<DstPacketType>::size) == (8 * SrcPacketSize), bool>; |
| |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE bool check_array_bounds(Index row, Index col, Index begin, Index count) const { |
| return IsRowMajor ? (col + count + begin <= cols()) : (row + count + begin <= rows()); |
| } |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE bool check_array_bounds(Index index, Index begin, Index count) const { |
| return index + count + begin <= size(); |
| } |
| |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE SrcType srcCoeff(Index row, Index col, Index offset) const { |
| Index actualRow = IsRowMajor ? row : row + offset; |
| Index actualCol = IsRowMajor ? col + offset : col; |
| return m_argImpl.coeff(actualRow, actualCol); |
| } |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE SrcType srcCoeff(Index index, Index offset) const { |
| Index actualIndex = index + offset; |
| return m_argImpl.coeff(actualIndex); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE DstType coeff(Index row, Index col) const { |
| return cast<SrcType, DstType>(srcCoeff(row, col, 0)); |
| } |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE DstType coeff(Index index) const { |
| return cast<SrcType, DstType>(srcCoeff(index, 0)); |
| } |
| |
| template <int LoadMode, typename PacketType = SrcPacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType srcPacket(Index row, Index col, Index offset) const { |
| constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| Index packetOffset = offset * PacketSize; |
| Index actualRow = IsRowMajor ? row : row + packetOffset; |
| Index actualCol = IsRowMajor ? col + packetOffset : col; |
| eigen_assert(check_array_bounds(actualRow, actualCol, 0, PacketSize) && "Array index out of bounds"); |
| return m_argImpl.template packet<LoadMode, PacketType>(actualRow, actualCol); |
| } |
| template <int LoadMode, typename PacketType = SrcPacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType srcPacket(Index index, Index offset) const { |
| constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| Index packetOffset = offset * PacketSize; |
| Index actualIndex = index + packetOffset; |
| eigen_assert(check_array_bounds(actualIndex, 0, PacketSize) && "Array index out of bounds"); |
| return m_argImpl.template packet<LoadMode, PacketType>(actualIndex); |
| } |
| template <int LoadMode, typename PacketType = SrcPacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType srcPacketSegment(Index row, Index col, Index begin, Index count, |
| Index offset) const { |
| constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| Index packetOffset = offset * PacketSize; |
| Index actualRow = IsRowMajor ? row : row + packetOffset; |
| Index actualCol = IsRowMajor ? col + packetOffset : col; |
| eigen_assert(check_array_bounds(actualRow, actualCol, begin, count) && "Array index out of bounds"); |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(actualRow, actualCol, begin, count); |
| } |
| template <int LoadMode, typename PacketType = SrcPacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType srcPacketSegment(Index index, Index begin, Index count, |
| Index offset) const { |
| constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| Index packetOffset = offset * PacketSize; |
| Index actualIndex = index + packetOffset; |
| eigen_assert(check_array_bounds(actualIndex, begin, count) && "Array index out of bounds"); |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(actualIndex, begin, count); |
| } |
| |
| template <int NumPackets, int LoadMode, typename PacketType = SrcPacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketBlock<PacketType, NumPackets> srcPacketSegmentHelper(Index row, Index col, |
| Index begin, |
| Index count) const { |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| PacketBlock<PacketType, NumPackets> packets; |
| for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); |
| Index offset = begin / SrcPacketSize; |
| Index actualBegin = begin % SrcPacketSize; |
| for (; offset < NumPackets; offset++) { |
| Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); |
| packets.packet[offset] = srcPacketSegment<SrcLoadMode>(row, col, actualBegin, actualCount, offset); |
| if (count == actualCount) break; |
| actualBegin = 0; |
| count -= actualCount; |
| } |
| return packets; |
| } |
| template <int NumPackets, int LoadMode, typename PacketType = SrcPacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketBlock<PacketType, NumPackets> srcPacketSegmentHelper(Index index, |
| Index begin, |
| Index count) const { |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| PacketBlock<PacketType, NumPackets> packets; |
| for (Index i = 0; i < NumPackets; i++) packets.packet[i] = pzero(PacketType()); |
| Index offset = begin / SrcPacketSize; |
| Index actualBegin = begin % SrcPacketSize; |
| for (; offset < NumPackets; offset++) { |
| Index actualCount = numext::mini(SrcPacketSize - actualBegin, count); |
| packets.packet[offset] = srcPacketSegment<SrcLoadMode>(index, actualBegin, actualCount, offset); |
| if (count == actualCount) break; |
| actualBegin = 0; |
| count -= actualCount; |
| } |
| return packets; |
| } |
| |
| // There is no source packet type with equal or fewer elements than DstPacketType. |
| // This is problematic as the evaluation loop may attempt to access data outside the bounds of the array. |
| // For example, consider the cast utilizing pcast<Packet4f,Packet2d> with an array of size 4: {0.0f,1.0f,2.0f,3.0f}. |
| // The first iteration of the evaluation loop will load 16 bytes: {0.0f,1.0f,2.0f,3.0f} and cast to {0.0,1.0}, which |
| // is acceptable. The second iteration will load 16 bytes: {2.0f,3.0f,?,?}, which is outside the bounds of the array. |
| template <int LoadMode, typename DstPacketType, AltSrcScalarOp<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index row, Index col) const { |
| constexpr int DstPacketSize = unpacket_traits<DstPacketType>::size; |
| constexpr int SrcBytesIncrement = DstPacketSize * sizeof(SrcType); |
| constexpr int SrcLoadMode = plain_enum_min(SrcBytesIncrement, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>(srcPacketSegment<SrcLoadMode>(row, col, 0, DstPacketSize, 0)); |
| } |
| // Use the source packet type with the same size as DstPacketType, if it exists |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs1<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index row, Index col) const { |
| constexpr int DstPacketSize = unpacket_traits<DstPacketType>::size; |
| using SizedSrcPacketType = typename find_packet_by_size<SrcType, DstPacketSize>::type; |
| constexpr int SrcBytesIncrement = DstPacketSize * sizeof(SrcType); |
| constexpr int SrcLoadMode = plain_enum_min(SrcBytesIncrement, LoadMode); |
| return pcast<SizedSrcPacketType, DstPacketType>(srcPacket<SrcLoadMode, SizedSrcPacketType>(row, col, 0)); |
| } |
| // unpacket_traits<DstPacketType>::size == 2 * SrcPacketSize |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs2<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index row, Index col) const { |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>(srcPacket<SrcLoadMode>(row, col, 0), |
| srcPacket<SrcLoadMode>(row, col, 1)); |
| } |
| // unpacket_traits<DstPacketType>::size == 4 * SrcPacketSize |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs4<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index row, Index col) const { |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>(srcPacket<SrcLoadMode>(row, col, 0), srcPacket<SrcLoadMode>(row, col, 1), |
| srcPacket<SrcLoadMode>(row, col, 2), |
| srcPacket<SrcLoadMode>(row, col, 3)); |
| } |
| // unpacket_traits<DstPacketType>::size == 8 * SrcPacketSize |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs8<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index row, Index col) const { |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>( |
| srcPacket<SrcLoadMode>(row, col, 0), srcPacket<SrcLoadMode>(row, col, 1), srcPacket<SrcLoadMode>(row, col, 2), |
| srcPacket<SrcLoadMode>(row, col, 3), srcPacket<SrcLoadMode>(row, col, 4), srcPacket<SrcLoadMode>(row, col, 5), |
| srcPacket<SrcLoadMode>(row, col, 6), srcPacket<SrcLoadMode>(row, col, 7)); |
| } |
| |
| // packetSegment variants |
| template <int LoadMode, typename DstPacketType, AltSrcScalarOp<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index row, Index col, Index begin, |
| Index count) const { |
| constexpr int DstPacketSize = unpacket_traits<DstPacketType>::size; |
| constexpr int SrcBytesIncrement = DstPacketSize * sizeof(SrcType); |
| constexpr int SrcLoadMode = plain_enum_min(SrcBytesIncrement, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>(srcPacketSegment<SrcLoadMode>(row, col, begin, count, 0)); |
| } |
| // Use the source packet type with the same size as DstPacketType, if it exists |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs1<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index row, Index col, Index begin, |
| Index count) const { |
| constexpr int DstPacketSize = unpacket_traits<DstPacketType>::size; |
| using SizedSrcPacketType = typename find_packet_by_size<SrcType, DstPacketSize>::type; |
| constexpr int SrcBytesIncrement = DstPacketSize * sizeof(SrcType); |
| constexpr int SrcLoadMode = plain_enum_min(SrcBytesIncrement, LoadMode); |
| return pcast<SizedSrcPacketType, DstPacketType>( |
| srcPacketSegment<SrcLoadMode, SizedSrcPacketType>(row, col, begin, count, 0)); |
| } |
| // unpacket_traits<DstPacketType>::size == 2 * SrcPacketSize |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs2<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index row, Index col, Index begin, |
| Index count) const { |
| constexpr int NumPackets = 2; |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| PacketBlock<SrcPacketType, NumPackets> packets = |
| srcPacketSegmentHelper<NumPackets, SrcLoadMode>(row, col, begin, count); |
| return pcast<SrcPacketType, DstPacketType>(packets.packet[0], packets.packet[1]); |
| } |
| // unpacket_traits<DstPacketType>::size == 4 * SrcPacketSize |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs4<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index row, Index col, Index begin, |
| Index count) const { |
| constexpr int NumPackets = 4; |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| PacketBlock<SrcPacketType, NumPackets> packets = |
| srcPacketSegmentHelper<NumPackets, SrcLoadMode>(row, col, begin, count); |
| return pcast<SrcPacketType, DstPacketType>(packets.packet[0], packets.packet[1], packets.packet[2], |
| packets.packet[3]); |
| } |
| // unpacket_traits<DstPacketType>::size == 8 * SrcPacketSize |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs8<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index row, Index col, Index begin, |
| Index count) const { |
| constexpr int NumPackets = 8; |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| PacketBlock<SrcPacketType, NumPackets> packets = |
| srcPacketSegmentHelper<NumPackets, SrcLoadMode>(row, col, begin, count); |
| return pcast<SrcPacketType, DstPacketType>(packets.packet[0], packets.packet[1], packets.packet[2], |
| packets.packet[3], packets.packet[4], packets.packet[5], |
| packets.packet[6], packets.packet[7]); |
| } |
| |
| // Analogous routines for linear access. |
| template <int LoadMode, typename DstPacketType, AltSrcScalarOp<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index index) const { |
| constexpr int DstPacketSize = unpacket_traits<DstPacketType>::size; |
| constexpr int SrcBytesIncrement = DstPacketSize * sizeof(SrcType); |
| constexpr int SrcLoadMode = plain_enum_min(SrcBytesIncrement, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>(srcPacketSegment<SrcLoadMode>(index, 0, DstPacketSize, 0)); |
| } |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs1<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index index) const { |
| constexpr int DstPacketSize = unpacket_traits<DstPacketType>::size; |
| using SizedSrcPacketType = typename find_packet_by_size<SrcType, DstPacketSize>::type; |
| constexpr int SrcBytesIncrement = DstPacketSize * sizeof(SrcType); |
| constexpr int SrcLoadMode = plain_enum_min(SrcBytesIncrement, LoadMode); |
| return pcast<SizedSrcPacketType, DstPacketType>(srcPacket<SrcLoadMode, SizedSrcPacketType>(index, 0)); |
| } |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs2<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index index) const { |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>(srcPacket<SrcLoadMode>(index, 0), srcPacket<SrcLoadMode>(index, 1)); |
| } |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs4<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index index) const { |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>(srcPacket<SrcLoadMode>(index, 0), srcPacket<SrcLoadMode>(index, 1), |
| srcPacket<SrcLoadMode>(index, 2), srcPacket<SrcLoadMode>(index, 3)); |
| } |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs8<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packet(Index index) const { |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>(srcPacket<SrcLoadMode>(index, 0), srcPacket<SrcLoadMode>(index, 1), |
| srcPacket<SrcLoadMode>(index, 2), srcPacket<SrcLoadMode>(index, 3), |
| srcPacket<SrcLoadMode>(index, 4), srcPacket<SrcLoadMode>(index, 5), |
| srcPacket<SrcLoadMode>(index, 6), srcPacket<SrcLoadMode>(index, 7)); |
| } |
| |
| // packetSegment variants |
| template <int LoadMode, typename DstPacketType, AltSrcScalarOp<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index index, Index begin, Index count) const { |
| constexpr int DstPacketSize = unpacket_traits<DstPacketType>::size; |
| constexpr int SrcBytesIncrement = DstPacketSize * sizeof(SrcType); |
| constexpr int SrcLoadMode = plain_enum_min(SrcBytesIncrement, LoadMode); |
| return pcast<SrcPacketType, DstPacketType>(srcPacketSegment<SrcLoadMode>(index, begin, count, 0)); |
| } |
| // Use the source packet type with the same size as DstPacketType, if it exists |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs1<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index index, Index begin, Index count) const { |
| constexpr int DstPacketSize = unpacket_traits<DstPacketType>::size; |
| using SizedSrcPacketType = typename find_packet_by_size<SrcType, DstPacketSize>::type; |
| constexpr int SrcBytesIncrement = DstPacketSize * sizeof(SrcType); |
| constexpr int SrcLoadMode = plain_enum_min(SrcBytesIncrement, LoadMode); |
| return pcast<SizedSrcPacketType, DstPacketType>( |
| srcPacketSegment<SrcLoadMode, SizedSrcPacketType>(index, begin, count, 0)); |
| } |
| // unpacket_traits<DstPacketType>::size == 2 * SrcPacketSize |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs2<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index index, Index begin, Index count) const { |
| constexpr int NumPackets = 2; |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| PacketBlock<SrcPacketType, NumPackets> packets = |
| srcPacketSegmentHelper<NumPackets, SrcLoadMode>(index, begin, count); |
| return pcast<SrcPacketType, DstPacketType>(packets.packet[0], packets.packet[1]); |
| } |
| // unpacket_traits<DstPacketType>::size == 4 * SrcPacketSize |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs4<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index index, Index begin, Index count) const { |
| constexpr int NumPackets = 4; |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| PacketBlock<SrcPacketType, NumPackets> packets = |
| srcPacketSegmentHelper<NumPackets, SrcLoadMode>(index, begin, count); |
| return pcast<SrcPacketType, DstPacketType>(packets.packet[0], packets.packet[1], packets.packet[2], |
| packets.packet[3]); |
| } |
| // unpacket_traits<DstPacketType>::size == 8 * SrcPacketSize |
| template <int LoadMode, typename DstPacketType, SrcPacketArgs8<DstPacketType> = true> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE DstPacketType packetSegment(Index index, Index begin, Index count) const { |
| constexpr int NumPackets = 8; |
| constexpr int SrcLoadMode = plain_enum_min(SrcPacketBytes, LoadMode); |
| PacketBlock<SrcPacketType, NumPackets> packets = |
| srcPacketSegmentHelper<NumPackets, SrcLoadMode>(index, begin, count); |
| return pcast<SrcPacketType, DstPacketType>(packets.packet[0], packets.packet[1], packets.packet[2], |
| packets.packet[3], packets.packet[4], packets.packet[5], |
| packets.packet[6], packets.packet[7]); |
| } |
| |
| constexpr EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE Index rows() const { return m_rows; } |
| constexpr EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE Index cols() const { return m_cols; } |
| constexpr EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE Index size() const { return m_rows * m_cols; } |
| |
| protected: |
| const evaluator<ArgType> m_argImpl; |
| const variable_if_dynamic<Index, XprType::RowsAtCompileTime> m_rows; |
| const variable_if_dynamic<Index, XprType::ColsAtCompileTime> m_cols; |
| }; |
| |
| // -------------------- CwiseTernaryOp -------------------- |
| |
| // this is a ternary expression |
| template <typename TernaryOp, typename Arg1, typename Arg2, typename Arg3> |
| struct evaluator<CwiseTernaryOp<TernaryOp, Arg1, Arg2, Arg3>> |
| : public ternary_evaluator<CwiseTernaryOp<TernaryOp, Arg1, Arg2, Arg3>> { |
| using XprType = CwiseTernaryOp<TernaryOp, Arg1, Arg2, Arg3>; |
| using Base = ternary_evaluator<CwiseTernaryOp<TernaryOp, Arg1, Arg2, Arg3>>; |
| |
| EIGEN_DEVICE_FUNC constexpr explicit evaluator(const XprType& xpr) : Base(xpr) {} |
| }; |
| |
| template <typename TernaryOp, typename Arg1, typename Arg2, typename Arg3> |
| struct ternary_evaluator<CwiseTernaryOp<TernaryOp, Arg1, Arg2, Arg3>, IndexBased, IndexBased> |
| : evaluator_base<CwiseTernaryOp<TernaryOp, Arg1, Arg2, Arg3>> { |
| using XprType = CwiseTernaryOp<TernaryOp, Arg1, Arg2, Arg3>; |
| |
| enum { |
| CoeffReadCost = int(evaluator<Arg1>::CoeffReadCost) + int(evaluator<Arg2>::CoeffReadCost) + |
| int(evaluator<Arg3>::CoeffReadCost) + int(functor_traits<TernaryOp>::Cost), |
| |
| Arg1Flags = evaluator<Arg1>::Flags, |
| Arg2Flags = evaluator<Arg2>::Flags, |
| Arg3Flags = evaluator<Arg3>::Flags, |
| SameType = std::is_same<typename Arg1::Scalar, typename Arg2::Scalar>::value && |
| std::is_same<typename Arg1::Scalar, typename Arg3::Scalar>::value, |
| StorageOrdersAgree = (int(Arg1Flags) & RowMajorBit) == (int(Arg2Flags) & RowMajorBit) && |
| (int(Arg1Flags) & RowMajorBit) == (int(Arg3Flags) & RowMajorBit), |
| Flags0 = (int(Arg1Flags) | int(Arg2Flags) | int(Arg3Flags)) & |
| (HereditaryBits | |
| (int(Arg1Flags) & int(Arg2Flags) & int(Arg3Flags) & |
| ((StorageOrdersAgree ? LinearAccessBit : 0) | |
| (functor_traits<TernaryOp>::PacketAccess && StorageOrdersAgree && SameType ? PacketAccessBit : 0)))), |
| Flags = (Flags0 & ~RowMajorBit) | (Arg1Flags & RowMajorBit), |
| Alignment = plain_enum_min(plain_enum_min(evaluator<Arg1>::Alignment, evaluator<Arg2>::Alignment), |
| evaluator<Arg3>::Alignment) |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr explicit ternary_evaluator(const XprType& xpr) : m_d(xpr) { |
| EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<TernaryOp>::Cost); |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| return m_d.func()(m_d.arg1Impl.coeff(row, col), m_d.arg2Impl.coeff(row, col), m_d.arg3Impl.coeff(row, col)); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return m_d.func()(m_d.arg1Impl.coeff(index), m_d.arg2Impl.coeff(index), m_d.arg3Impl.coeff(index)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| return m_d.func().packetOp(m_d.arg1Impl.template packet<LoadMode, PacketType>(row, col), |
| m_d.arg2Impl.template packet<LoadMode, PacketType>(row, col), |
| m_d.arg3Impl.template packet<LoadMode, PacketType>(row, col)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| return m_d.func().packetOp(m_d.arg1Impl.template packet<LoadMode, PacketType>(index), |
| m_d.arg2Impl.template packet<LoadMode, PacketType>(index), |
| m_d.arg3Impl.template packet<LoadMode, PacketType>(index)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| return m_d.func().packetOp(m_d.arg1Impl.template packetSegment<LoadMode, PacketType>(row, col, begin, count), |
| m_d.arg2Impl.template packetSegment<LoadMode, PacketType>(row, col, begin, count), |
| m_d.arg3Impl.template packetSegment<LoadMode, PacketType>(row, col, begin, count)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| return m_d.func().packetOp(m_d.arg1Impl.template packetSegment<LoadMode, PacketType>(index, begin, count), |
| m_d.arg2Impl.template packetSegment<LoadMode, PacketType>(index, begin, count), |
| m_d.arg3Impl.template packetSegment<LoadMode, PacketType>(index, begin, count)); |
| } |
| |
| protected: |
| // this helper permits to completely eliminate the functor if it is empty |
| struct Data { |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Data(const XprType& xpr) |
| : op(xpr.functor()), arg1Impl(xpr.arg1()), arg2Impl(xpr.arg2()), arg3Impl(xpr.arg3()) {} |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE const TernaryOp& func() const { return op; } |
| TernaryOp op; |
| evaluator<Arg1> arg1Impl; |
| evaluator<Arg2> arg2Impl; |
| evaluator<Arg3> arg3Impl; |
| }; |
| |
| Data m_d; |
| }; |
| |
| template <typename Arg1, typename Arg2, typename Scalar, typename CmpLhsType, typename CmpRhsType, ComparisonName cmp> |
| struct scalar_boolean_select_spec { |
| using DummyTernaryOp = scalar_boolean_select_op<Scalar, Scalar, bool>; |
| using DummyArg3 = CwiseBinaryOp<scalar_cmp_op<Scalar, Scalar, cmp, false>, CmpLhsType, CmpRhsType>; |
| using DummyXprType = CwiseTernaryOp<DummyTernaryOp, Arg1, Arg2, DummyArg3>; |
| |
| // only use the typed comparison if it is vectorized |
| static constexpr bool UseTyped = functor_traits<scalar_cmp_op<Scalar, Scalar, cmp, true>>::PacketAccess; |
| using CondScalar = std::conditional_t<UseTyped, Scalar, bool>; |
| |
| using TernaryOp = scalar_boolean_select_op<Scalar, Scalar, CondScalar>; |
| using Arg3 = CwiseBinaryOp<scalar_cmp_op<Scalar, Scalar, cmp, UseTyped>, CmpLhsType, CmpRhsType>; |
| using XprType = CwiseTernaryOp<TernaryOp, Arg1, Arg2, Arg3>; |
| |
| using Base = ternary_evaluator<XprType>; |
| }; |
| |
| // specialization for expressions like (a < b).select(c, d) to enable full vectorization |
| template <typename Arg1, typename Arg2, typename Scalar, typename CmpLhsType, typename CmpRhsType, ComparisonName cmp> |
| struct evaluator<CwiseTernaryOp<scalar_boolean_select_op<Scalar, Scalar, bool>, Arg1, Arg2, |
| CwiseBinaryOp<scalar_cmp_op<Scalar, Scalar, cmp, false>, CmpLhsType, CmpRhsType>>> |
| : public scalar_boolean_select_spec<Arg1, Arg2, Scalar, CmpLhsType, CmpRhsType, cmp>::Base { |
| using Helper = scalar_boolean_select_spec<Arg1, Arg2, Scalar, CmpLhsType, CmpRhsType, cmp>; |
| using Base = typename Helper::Base; |
| using DummyXprType = typename Helper::DummyXprType; |
| using Arg3 = typename Helper::Arg3; |
| using XprType = typename Helper::XprType; |
| |
| EIGEN_DEVICE_FUNC constexpr explicit evaluator(const DummyXprType& xpr) |
| : Base(XprType(xpr.arg1(), xpr.arg2(), Arg3(xpr.arg3().lhs(), xpr.arg3().rhs()))) {} |
| }; |
| |
| // -------------------- CwiseBinaryOp -------------------- |
| |
| // this is a binary expression |
| template <typename BinaryOp, typename Lhs, typename Rhs> |
| struct evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>> : public binary_evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>> { |
| using XprType = CwiseBinaryOp<BinaryOp, Lhs, Rhs>; |
| using Base = binary_evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>>; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit evaluator(const XprType& xpr) : Base(xpr) {} |
| }; |
| |
| template <typename BinaryOp, typename Lhs, typename Rhs> |
| struct binary_evaluator<CwiseBinaryOp<BinaryOp, Lhs, Rhs>, IndexBased, IndexBased> |
| : evaluator_base<CwiseBinaryOp<BinaryOp, Lhs, Rhs>> { |
| using XprType = CwiseBinaryOp<BinaryOp, Lhs, Rhs>; |
| |
| enum { |
| CoeffReadCost = |
| int(evaluator<Lhs>::CoeffReadCost) + int(evaluator<Rhs>::CoeffReadCost) + int(functor_traits<BinaryOp>::Cost), |
| |
| LhsFlags = evaluator<Lhs>::Flags, |
| RhsFlags = evaluator<Rhs>::Flags, |
| SameType = std::is_same<typename Lhs::Scalar, typename Rhs::Scalar>::value, |
| StorageOrdersAgree = (int(LhsFlags) & RowMajorBit) == (int(RhsFlags) & RowMajorBit), |
| Flags0 = (int(LhsFlags) | int(RhsFlags)) & |
| (HereditaryBits | |
| (int(LhsFlags) & int(RhsFlags) & |
| ((StorageOrdersAgree ? LinearAccessBit : 0) | |
| (functor_traits<BinaryOp>::PacketAccess && StorageOrdersAgree && SameType ? PacketAccessBit : 0)))), |
| Flags = (Flags0 & ~RowMajorBit) | (LhsFlags & RowMajorBit), |
| Alignment = plain_enum_min(evaluator<Lhs>::Alignment, evaluator<Rhs>::Alignment) |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit binary_evaluator(const XprType& xpr) : m_d(xpr) { |
| EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<BinaryOp>::Cost); |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| return m_d.func()(m_d.lhsImpl.coeff(row, col), m_d.rhsImpl.coeff(row, col)); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return m_d.func()(m_d.lhsImpl.coeff(index), m_d.rhsImpl.coeff(index)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| return m_d.func().packetOp(m_d.lhsImpl.template packet<LoadMode, PacketType>(row, col), |
| m_d.rhsImpl.template packet<LoadMode, PacketType>(row, col)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| return m_d.func().packetOp(m_d.lhsImpl.template packet<LoadMode, PacketType>(index), |
| m_d.rhsImpl.template packet<LoadMode, PacketType>(index)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| return m_d.func().packetOp(m_d.lhsImpl.template packetSegment<LoadMode, PacketType>(row, col, begin, count), |
| m_d.rhsImpl.template packetSegment<LoadMode, PacketType>(row, col, begin, count)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| return m_d.func().packetOp(m_d.lhsImpl.template packetSegment<LoadMode, PacketType>(index, begin, count), |
| m_d.rhsImpl.template packetSegment<LoadMode, PacketType>(index, begin, count)); |
| } |
| |
| protected: |
| // this helper permits to completely eliminate the functor if it is empty |
| struct Data { |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Data(const XprType& xpr) |
| : op(xpr.functor()), lhsImpl(xpr.lhs()), rhsImpl(xpr.rhs()) {} |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE const BinaryOp& func() const { return op; } |
| BinaryOp op; |
| evaluator<Lhs> lhsImpl; |
| evaluator<Rhs> rhsImpl; |
| }; |
| |
| Data m_d; |
| }; |
| |
| // -------------------- CwiseUnaryView -------------------- |
| |
| template <typename UnaryOp, typename ArgType, typename StrideType> |
| struct unary_evaluator<CwiseUnaryView<UnaryOp, ArgType, StrideType>, IndexBased> |
| : evaluator_base<CwiseUnaryView<UnaryOp, ArgType, StrideType>> { |
| using XprType = CwiseUnaryView<UnaryOp, ArgType, StrideType>; |
| |
| enum { |
| CoeffReadCost = int(evaluator<ArgType>::CoeffReadCost) + int(functor_traits<UnaryOp>::Cost), |
| |
| Flags = (evaluator<ArgType>::Flags & (HereditaryBits | LinearAccessBit | DirectAccessBit)), |
| |
| Alignment = 0 // FIXME: clarify why alignment is lost for CwiseUnaryView. |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr explicit unary_evaluator(const XprType& op) : m_d(op) { |
| EIGEN_INTERNAL_CHECK_COST_VALUE(functor_traits<UnaryOp>::Cost); |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| |
| using Scalar = typename XprType::Scalar; |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| return m_d.func()(m_d.argImpl.coeff(row, col)); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return m_d.func()(m_d.argImpl.coeff(index)); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index row, Index col) { |
| return m_d.func()(m_d.argImpl.coeffRef(row, col)); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index index) { |
| return m_d.func()(m_d.argImpl.coeffRef(index)); |
| } |
| |
| protected: |
| // this helper permits to completely eliminate the functor if it is empty |
| struct Data { |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Data(const XprType& xpr) |
| : op(xpr.functor()), argImpl(xpr.nestedExpression()) {} |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE const UnaryOp& func() const { return op; } |
| UnaryOp op; |
| evaluator<ArgType> argImpl; |
| }; |
| |
| Data m_d; |
| }; |
| |
| // -------------------- Map -------------------- |
| |
| // FIXME: consider using Derived::PlainObject for PlainObjectType. |
| // but that might complicate template specialization |
| template <typename Derived, typename PlainObjectType> |
| struct mapbase_evaluator; |
| |
| template <typename Derived, typename PlainObjectType> |
| struct mapbase_evaluator : evaluator_base<Derived> { |
| using XprType = Derived; |
| using PointerType = typename XprType::PointerType; |
| using Scalar = typename XprType::Scalar; |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| enum { |
| IsRowMajor = XprType::IsRowMajor, |
| ColsAtCompileTime = XprType::ColsAtCompileTime, |
| CoeffReadCost = NumTraits<Scalar>::ReadCost |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit mapbase_evaluator(const XprType& map) |
| : m_data(const_cast<PointerType>(map.data())), |
| m_innerStride(map.innerStride()), |
| m_outerStride(map.outerStride()) { |
| EIGEN_STATIC_ASSERT(check_implication((evaluator<Derived>::Flags & PacketAccessBit) != 0, |
| inner_stride_at_compile_time<Derived>::value == 1), |
| PACKET_ACCESS_REQUIRES_TO_HAVE_INNER_STRIDE_FIXED_TO_1); |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| return m_data[col * colStride() + row * rowStride()]; |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return m_data[index * m_innerStride.value()]; |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index row, Index col) { |
| return m_data[col * colStride() + row * rowStride()]; |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index index) { |
| return m_data[index * m_innerStride.value()]; |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| PointerType ptr = m_data + row * rowStride() + col * colStride(); |
| return ploadt<PacketType, LoadMode>(ptr); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| return ploadt<PacketType, LoadMode>(m_data + index * m_innerStride.value()); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index row, Index col, const PacketType& x) { |
| PointerType ptr = m_data + row * rowStride() + col * colStride(); |
| pstoret<Scalar, PacketType, StoreMode>(ptr, x); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index index, const PacketType& x) { |
| pstoret<Scalar, PacketType, StoreMode>(m_data + index * m_innerStride.value(), x); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| PointerType ptr = m_data + row * rowStride() + col * colStride(); |
| return ploadtSegment<PacketType, LoadMode>(ptr, begin, count); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| return ploadtSegment<PacketType, LoadMode>(m_data + index * m_innerStride.value(), begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index row, Index col, const PacketType& x, Index begin, |
| Index count) { |
| PointerType ptr = m_data + row * rowStride() + col * colStride(); |
| pstoretSegment<Scalar, PacketType, StoreMode>(ptr, x, begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index index, const PacketType& x, Index begin, |
| Index count) { |
| pstoretSegment<Scalar, PacketType, StoreMode>(m_data + index * m_innerStride.value(), x, begin, count); |
| } |
| |
| protected: |
| EIGEN_DEVICE_FUNC constexpr Index rowStride() const noexcept { |
| return XprType::IsRowMajor ? m_outerStride.value() : m_innerStride.value(); |
| } |
| EIGEN_DEVICE_FUNC constexpr Index colStride() const noexcept { |
| return XprType::IsRowMajor ? m_innerStride.value() : m_outerStride.value(); |
| } |
| |
| PointerType m_data; |
| const variable_if_dynamic<Index, XprType::InnerStrideAtCompileTime> m_innerStride; |
| const variable_if_dynamic<Index, XprType::OuterStrideAtCompileTime> m_outerStride; |
| }; |
| |
| template <typename PlainObjectType, int MapOptions, typename StrideType> |
| struct evaluator<Map<PlainObjectType, MapOptions, StrideType>> |
| : public mapbase_evaluator<Map<PlainObjectType, MapOptions, StrideType>, PlainObjectType> { |
| using XprType = Map<PlainObjectType, MapOptions, StrideType>; |
| using Scalar = typename XprType::Scalar; |
| // TODO: should check for smaller packet types once we can handle multi-sized packet types |
| using PacketScalar = typename packet_traits<Scalar>::type; |
| |
| enum { |
| InnerStrideAtCompileTime = StrideType::InnerStrideAtCompileTime == 0 |
| ? int(PlainObjectType::InnerStrideAtCompileTime) |
| : int(StrideType::InnerStrideAtCompileTime), |
| OuterStrideAtCompileTime = StrideType::OuterStrideAtCompileTime == 0 |
| ? int(PlainObjectType::OuterStrideAtCompileTime) |
| : int(StrideType::OuterStrideAtCompileTime), |
| HasNoInnerStride = InnerStrideAtCompileTime == 1, |
| HasNoOuterStride = StrideType::OuterStrideAtCompileTime == 0, |
| HasNoStride = HasNoInnerStride && HasNoOuterStride, |
| IsDynamicSize = PlainObjectType::SizeAtCompileTime == Dynamic, |
| |
| PacketAccessMask = bool(HasNoInnerStride) ? ~int(0) : ~int(PacketAccessBit), |
| LinearAccessMask = |
| bool(HasNoStride) || bool(PlainObjectType::IsVectorAtCompileTime) ? ~int(0) : ~int(LinearAccessBit), |
| Flags = int(evaluator<PlainObjectType>::Flags) & (LinearAccessMask & PacketAccessMask), |
| |
| Alignment = int(MapOptions) & int(AlignedMask) |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr explicit evaluator(const XprType& map) |
| : mapbase_evaluator<XprType, PlainObjectType>(map) {} |
| }; |
| |
| // -------------------- Ref -------------------- |
| |
| template <typename PlainObjectType, int RefOptions, typename StrideType> |
| struct evaluator<Ref<PlainObjectType, RefOptions, StrideType>> |
| : public mapbase_evaluator<Ref<PlainObjectType, RefOptions, StrideType>, PlainObjectType> { |
| using XprType = Ref<PlainObjectType, RefOptions, StrideType>; |
| |
| enum { |
| Flags = evaluator<Map<PlainObjectType, RefOptions, StrideType>>::Flags, |
| Alignment = evaluator<Map<PlainObjectType, RefOptions, StrideType>>::Alignment |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit evaluator(const XprType& ref) |
| : mapbase_evaluator<XprType, PlainObjectType>(ref) {} |
| }; |
| |
| // -------------------- Block -------------------- |
| |
| template <typename ArgType, int BlockRows, int BlockCols, bool InnerPanel, |
| bool HasDirectAccess = has_direct_access<ArgType>::value> |
| struct block_evaluator; |
| |
| template <typename ArgType, int BlockRows, int BlockCols, bool InnerPanel> |
| struct evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>> |
| : block_evaluator<ArgType, BlockRows, BlockCols, InnerPanel> { |
| using XprType = Block<ArgType, BlockRows, BlockCols, InnerPanel>; |
| using Scalar = typename XprType::Scalar; |
| // TODO: should check for smaller packet types once we can handle multi-sized packet types |
| using PacketScalar = typename packet_traits<Scalar>::type; |
| |
| enum { |
| CoeffReadCost = evaluator<ArgType>::CoeffReadCost, |
| |
| RowsAtCompileTime = traits<XprType>::RowsAtCompileTime, |
| ColsAtCompileTime = traits<XprType>::ColsAtCompileTime, |
| MaxRowsAtCompileTime = traits<XprType>::MaxRowsAtCompileTime, |
| MaxColsAtCompileTime = traits<XprType>::MaxColsAtCompileTime, |
| |
| ArgTypeIsRowMajor = (int(evaluator<ArgType>::Flags) & RowMajorBit) != 0, |
| IsRowMajor = (MaxRowsAtCompileTime == 1 && MaxColsAtCompileTime != 1) ? 1 |
| : (MaxColsAtCompileTime == 1 && MaxRowsAtCompileTime != 1) ? 0 |
| : ArgTypeIsRowMajor, |
| HasSameStorageOrderAsArgType = (IsRowMajor == ArgTypeIsRowMajor), |
| InnerSize = IsRowMajor ? int(ColsAtCompileTime) : int(RowsAtCompileTime), |
| InnerStrideAtCompileTime = HasSameStorageOrderAsArgType ? int(inner_stride_at_compile_time<ArgType>::value) |
| : int(outer_stride_at_compile_time<ArgType>::value), |
| OuterStrideAtCompileTime = HasSameStorageOrderAsArgType ? int(outer_stride_at_compile_time<ArgType>::value) |
| : int(inner_stride_at_compile_time<ArgType>::value), |
| MaskPacketAccessBit = (InnerStrideAtCompileTime == 1 || HasSameStorageOrderAsArgType) ? PacketAccessBit : 0, |
| |
| FlagsLinearAccessBit = (RowsAtCompileTime == 1 || ColsAtCompileTime == 1 || |
| (InnerPanel && (evaluator<ArgType>::Flags & LinearAccessBit))) |
| ? LinearAccessBit |
| : 0, |
| FlagsRowMajorBit = XprType::Flags & RowMajorBit, |
| Flags0 = evaluator<ArgType>::Flags & ((HereditaryBits & ~RowMajorBit) | DirectAccessBit | MaskPacketAccessBit), |
| Flags = Flags0 | FlagsLinearAccessBit | FlagsRowMajorBit, |
| |
| PacketAlignment = unpacket_traits<PacketScalar>::alignment, |
| Alignment0 = (InnerPanel && (OuterStrideAtCompileTime != Dynamic) && (OuterStrideAtCompileTime != 0) && |
| (((OuterStrideAtCompileTime * int(sizeof(Scalar))) % int(PacketAlignment)) == 0)) |
| ? int(PacketAlignment) |
| : 0, |
| Alignment = plain_enum_min(evaluator<ArgType>::Alignment, Alignment0) |
| }; |
| using block_evaluator_type = block_evaluator<ArgType, BlockRows, BlockCols, InnerPanel>; |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit evaluator(const XprType& block) |
| : block_evaluator_type(block) { |
| EIGEN_INTERNAL_CHECK_COST_VALUE(CoeffReadCost); |
| } |
| }; |
| |
| // no direct-access => dispatch to a unary evaluator |
| template <typename ArgType, int BlockRows, int BlockCols, bool InnerPanel> |
| struct block_evaluator<ArgType, BlockRows, BlockCols, InnerPanel, /*HasDirectAccess*/ false> |
| : unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>> { |
| using XprType = Block<ArgType, BlockRows, BlockCols, InnerPanel>; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit block_evaluator(const XprType& block) |
| : unary_evaluator<XprType>(block) {} |
| }; |
| |
| template <typename ArgType, int BlockRows, int BlockCols, bool InnerPanel> |
| struct unary_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, IndexBased> |
| : evaluator_base<Block<ArgType, BlockRows, BlockCols, InnerPanel>> { |
| using XprType = Block<ArgType, BlockRows, BlockCols, InnerPanel>; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit unary_evaluator(const XprType& block) |
| : m_argImpl(block.nestedExpression()), |
| m_startRow(block.startRow()), |
| m_startCol(block.startCol()), |
| m_linear_offset(ForwardLinearAccess |
| ? (ArgType::IsRowMajor |
| ? block.startRow() * block.nestedExpression().cols() + block.startCol() |
| : block.startCol() * block.nestedExpression().rows() + block.startRow()) |
| : 0) {} |
| |
| using Scalar = typename XprType::Scalar; |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| enum { |
| RowsAtCompileTime = XprType::RowsAtCompileTime, |
| ForwardLinearAccess = (InnerPanel || int(XprType::IsRowMajor) == int(ArgType::IsRowMajor)) && |
| bool(evaluator<ArgType>::Flags & LinearAccessBit) |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| return m_argImpl.coeff(m_startRow.value() + row, m_startCol.value() + col); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return linear_coeff_impl(index, bool_constant<ForwardLinearAccess>()); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index row, Index col) { |
| return m_argImpl.coeffRef(m_startRow.value() + row, m_startCol.value() + col); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index index) { |
| return linear_coeffRef_impl(index, bool_constant<ForwardLinearAccess>()); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| return m_argImpl.template packet<LoadMode, PacketType>(m_startRow.value() + row, m_startCol.value() + col); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| EIGEN_IF_CONSTEXPR(ForwardLinearAccess) |
| return m_argImpl.template packet<LoadMode, PacketType>(m_linear_offset.value() + index); |
| else return packet<LoadMode, PacketType>(RowsAtCompileTime == 1 ? 0 : index, RowsAtCompileTime == 1 ? index : 0); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index row, Index col, const PacketType& x) { |
| return m_argImpl.template writePacket<StoreMode, PacketType>(m_startRow.value() + row, m_startCol.value() + col, x); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index index, const PacketType& x) { |
| EIGEN_IF_CONSTEXPR(ForwardLinearAccess) |
| return m_argImpl.template writePacket<StoreMode, PacketType>(m_linear_offset.value() + index, x); |
| else return writePacket<StoreMode, PacketType>(RowsAtCompileTime == 1 ? 0 : index, |
| RowsAtCompileTime == 1 ? index : 0, x); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(m_startRow.value() + row, m_startCol.value() + col, |
| begin, count); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| EIGEN_IF_CONSTEXPR(ForwardLinearAccess) |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(m_linear_offset.value() + index, begin, count); |
| else return packetSegment<LoadMode, PacketType>(RowsAtCompileTime == 1 ? 0 : index, |
| RowsAtCompileTime == 1 ? index : 0, begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index row, Index col, const PacketType& x, Index begin, |
| Index count) { |
| return m_argImpl.template writePacketSegment<StoreMode, PacketType>(m_startRow.value() + row, |
| m_startCol.value() + col, x, begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index index, const PacketType& x, Index begin, |
| Index count) { |
| EIGEN_IF_CONSTEXPR(ForwardLinearAccess) |
| return m_argImpl.template writePacketSegment<StoreMode, PacketType>(m_linear_offset.value() + index, x, begin, |
| count); |
| else return writePacketSegment<StoreMode, PacketType>(RowsAtCompileTime == 1 ? 0 : index, |
| RowsAtCompileTime == 1 ? index : 0, x, begin, count); |
| } |
| |
| protected: |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType |
| linear_coeff_impl(Index index, std::true_type /* ForwardLinearAccess */) const { |
| return m_argImpl.coeff(m_linear_offset.value() + index); |
| } |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType |
| linear_coeff_impl(Index index, std::false_type /* not ForwardLinearAccess */) const { |
| return coeff(RowsAtCompileTime == 1 ? 0 : index, RowsAtCompileTime == 1 ? index : 0); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& linear_coeffRef_impl( |
| Index index, std::true_type /* ForwardLinearAccess */) { |
| return m_argImpl.coeffRef(m_linear_offset.value() + index); |
| } |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& linear_coeffRef_impl( |
| Index index, std::false_type /* not ForwardLinearAccess */) { |
| return coeffRef(RowsAtCompileTime == 1 ? 0 : index, RowsAtCompileTime == 1 ? index : 0); |
| } |
| |
| evaluator<ArgType> m_argImpl; |
| const variable_if_dynamic<Index, (ArgType::RowsAtCompileTime == 1 && BlockRows == 1) ? 0 : Dynamic> m_startRow; |
| const variable_if_dynamic<Index, (ArgType::ColsAtCompileTime == 1 && BlockCols == 1) ? 0 : Dynamic> m_startCol; |
| const variable_if_dynamic<Index, ForwardLinearAccess ? Dynamic : 0> m_linear_offset; |
| }; |
| |
| // TODO: This evaluator does not actually use the child evaluator; |
| // all action is via the data() as returned by the Block expression. |
| |
| template <typename ArgType, int BlockRows, int BlockCols, bool InnerPanel> |
| struct block_evaluator<ArgType, BlockRows, BlockCols, InnerPanel, /* HasDirectAccess */ true> |
| : mapbase_evaluator<Block<ArgType, BlockRows, BlockCols, InnerPanel>, |
| typename Block<ArgType, BlockRows, BlockCols, InnerPanel>::PlainObject> { |
| using XprType = Block<ArgType, BlockRows, BlockCols, InnerPanel>; |
| using Scalar = typename XprType::Scalar; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit block_evaluator(const XprType& block) |
| : mapbase_evaluator<XprType, typename XprType::PlainObject>(block) { |
| eigen_internal_assert((internal::is_constant_evaluated() || |
| (std::uintptr_t(block.data()) % plain_enum_max(1, evaluator<XprType>::Alignment)) == 0) && |
| "data is not aligned"); |
| } |
| }; |
| |
| // -------------------- Replicate -------------------- |
| |
| template <typename ArgType, int RowFactor, int ColFactor> |
| struct unary_evaluator<Replicate<ArgType, RowFactor, ColFactor>> |
| : evaluator_base<Replicate<ArgType, RowFactor, ColFactor>> { |
| using XprType = Replicate<ArgType, RowFactor, ColFactor>; |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| enum { Factor = (RowFactor == Dynamic || ColFactor == Dynamic) ? Dynamic : RowFactor * ColFactor }; |
| using ArgTypeNested = typename nested_eval<ArgType, Factor>::type; |
| using ArgTypeNestedCleaned = remove_all_t<ArgTypeNested>; |
| |
| enum { |
| CoeffReadCost = evaluator<ArgTypeNestedCleaned>::CoeffReadCost, |
| LinearAccessMask = XprType::IsVectorAtCompileTime ? LinearAccessBit : 0, |
| Flags = (evaluator<ArgTypeNestedCleaned>::Flags & (HereditaryBits | LinearAccessMask) & ~RowMajorBit) | |
| (traits<XprType>::Flags & RowMajorBit), |
| |
| Alignment = evaluator<ArgTypeNestedCleaned>::Alignment |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit unary_evaluator(const XprType& replicate) |
| : m_arg(replicate.nestedExpression()), |
| m_argImpl(m_arg), |
| m_rows(replicate.nestedExpression().rows()), |
| m_cols(replicate.nestedExpression().cols()) {} |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| // try to avoid using modulo; this is a pure optimization strategy |
| const Index actual_row = traits<XprType>::RowsAtCompileTime == 1 ? 0 : RowFactor == 1 ? row : row % m_rows.value(); |
| const Index actual_col = traits<XprType>::ColsAtCompileTime == 1 ? 0 : ColFactor == 1 ? col : col % m_cols.value(); |
| |
| return m_argImpl.coeff(actual_row, actual_col); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| // try to avoid using modulo; this is a pure optimization strategy |
| const Index actual_index = traits<XprType>::RowsAtCompileTime == 1 |
| ? (ColFactor == 1 ? index : index % m_cols.value()) |
| : (RowFactor == 1 ? index : index % m_rows.value()); |
| |
| return m_argImpl.coeff(actual_index); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| const Index actual_row = traits<XprType>::RowsAtCompileTime == 1 ? 0 : RowFactor == 1 ? row : row % m_rows.value(); |
| const Index actual_col = traits<XprType>::ColsAtCompileTime == 1 ? 0 : ColFactor == 1 ? col : col % m_cols.value(); |
| |
| return m_argImpl.template packet<LoadMode, PacketType>(actual_row, actual_col); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| const Index actual_index = traits<XprType>::RowsAtCompileTime == 1 |
| ? (ColFactor == 1 ? index : index % m_cols.value()) |
| : (RowFactor == 1 ? index : index % m_rows.value()); |
| |
| return m_argImpl.template packet<LoadMode, PacketType>(actual_index); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| const Index actual_row = traits<XprType>::RowsAtCompileTime == 1 ? 0 : RowFactor == 1 ? row : row % m_rows.value(); |
| const Index actual_col = traits<XprType>::ColsAtCompileTime == 1 ? 0 : ColFactor == 1 ? col : col % m_cols.value(); |
| |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(actual_row, actual_col, begin, count); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| const Index actual_index = traits<XprType>::RowsAtCompileTime == 1 |
| ? (ColFactor == 1 ? index : index % m_cols.value()) |
| : (RowFactor == 1 ? index : index % m_rows.value()); |
| |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(actual_index, begin, count); |
| } |
| |
| protected: |
| ArgTypeNested m_arg; |
| evaluator<ArgTypeNestedCleaned> m_argImpl; |
| const variable_if_dynamic<Index, ArgType::RowsAtCompileTime> m_rows; |
| const variable_if_dynamic<Index, ArgType::ColsAtCompileTime> m_cols; |
| }; |
| |
| // -------------------- MatrixWrapper and ArrayWrapper -------------------- |
| // |
| // evaluator_wrapper_base<T> is a common base class for the |
| // MatrixWrapper and ArrayWrapper evaluators. |
| |
| template <typename XprType> |
| struct evaluator_wrapper_base : evaluator_base<XprType> { |
| using ArgType = remove_all_t<typename XprType::NestedExpressionType>; |
| enum { |
| CoeffReadCost = evaluator<ArgType>::CoeffReadCost, |
| Flags = evaluator<ArgType>::Flags, |
| Alignment = evaluator<ArgType>::Alignment |
| }; |
| |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE explicit evaluator_wrapper_base(const ArgType& arg) : m_argImpl(arg) {} |
| |
| using Scalar = typename ArgType::Scalar; |
| using CoeffReturnType = typename ArgType::CoeffReturnType; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| return m_argImpl.coeff(row, col); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return m_argImpl.coeff(index); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index row, Index col) { |
| return m_argImpl.coeffRef(row, col); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index index) { return m_argImpl.coeffRef(index); } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| return m_argImpl.template packet<LoadMode, PacketType>(row, col); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| return m_argImpl.template packet<LoadMode, PacketType>(index); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index row, Index col, const PacketType& x) { |
| m_argImpl.template writePacket<StoreMode>(row, col, x); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index index, const PacketType& x) { |
| m_argImpl.template writePacket<StoreMode>(index, x); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(row, col, begin, count); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| return m_argImpl.template packetSegment<LoadMode, PacketType>(index, begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index row, Index col, const PacketType& x, Index begin, |
| Index count) { |
| m_argImpl.template writePacketSegment<StoreMode>(row, col, x, begin, count); |
| } |
| |
| template <int StoreMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index index, const PacketType& x, Index begin, |
| Index count) { |
| m_argImpl.template writePacketSegment<StoreMode>(index, x, begin, count); |
| } |
| |
| protected: |
| evaluator<ArgType> m_argImpl; |
| }; |
| |
| template <typename TArgType> |
| struct unary_evaluator<MatrixWrapper<TArgType>> : evaluator_wrapper_base<MatrixWrapper<TArgType>> { |
| using XprType = MatrixWrapper<TArgType>; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit unary_evaluator(const XprType& wrapper) |
| : evaluator_wrapper_base<MatrixWrapper<TArgType>>(wrapper.nestedExpression()) {} |
| }; |
| |
| template <typename TArgType> |
| struct unary_evaluator<ArrayWrapper<TArgType>> : evaluator_wrapper_base<ArrayWrapper<TArgType>> { |
| using XprType = ArrayWrapper<TArgType>; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit unary_evaluator(const XprType& wrapper) |
| : evaluator_wrapper_base<ArrayWrapper<TArgType>>(wrapper.nestedExpression()) {} |
| }; |
| |
| // -------------------- Reverse -------------------- |
| |
| // defined in Reverse.h: |
| template <typename PacketType, bool ReversePacket> |
| struct reverse_packet_cond; |
| |
| template <typename ArgType, int Direction> |
| struct unary_evaluator<Reverse<ArgType, Direction>> : evaluator_base<Reverse<ArgType, Direction>> { |
| using XprType = Reverse<ArgType, Direction>; |
| using Scalar = typename XprType::Scalar; |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| enum { |
| IsRowMajor = XprType::IsRowMajor, |
| IsColMajor = !IsRowMajor, |
| ReverseRow = (Direction == Vertical) || (Direction == BothDirections), |
| ReverseCol = (Direction == Horizontal) || (Direction == BothDirections), |
| ReversePacket = (Direction == BothDirections) || ((Direction == Vertical) && IsColMajor) || |
| ((Direction == Horizontal) && IsRowMajor), |
| |
| CoeffReadCost = evaluator<ArgType>::CoeffReadCost, |
| |
| // let's enable LinearAccess only with vectorization because of the product overhead |
| // FIXME: consider enabling DirectAccess with negative strides. |
| Flags0 = evaluator<ArgType>::Flags, |
| LinearAccess = |
| ((Direction == BothDirections) && (int(Flags0) & PacketAccessBit)) || |
| ((ReverseRow && XprType::ColsAtCompileTime == 1) || (ReverseCol && XprType::RowsAtCompileTime == 1)) |
| ? LinearAccessBit |
| : 0, |
| |
| Flags = int(Flags0) & (HereditaryBits | PacketAccessBit | LinearAccess), |
| |
| Alignment = 0 // FIXME: in some rare cases, Alignment could be preserved. |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit unary_evaluator(const XprType& reverse) |
| : m_argImpl(reverse.nestedExpression()), |
| m_rows(ReverseRow ? reverse.nestedExpression().rows() : 1), |
| m_cols(ReverseCol ? reverse.nestedExpression().cols() : 1) {} |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index col) const { |
| return m_argImpl.coeff(ReverseRow ? m_rows.value() - row - 1 : row, ReverseCol ? m_cols.value() - col - 1 : col); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return m_argImpl.coeff(m_rows.value() * m_cols.value() - index - 1); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index row, Index col) { |
| return m_argImpl.coeffRef(ReverseRow ? m_rows.value() - row - 1 : row, ReverseCol ? m_cols.value() - col - 1 : col); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index index) { |
| return m_argImpl.coeffRef(m_rows.value() * m_cols.value() - index - 1); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index row, Index col) const { |
| static constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| static constexpr int OffsetRow = ReverseRow && IsColMajor ? PacketSize : 1; |
| static constexpr int OffsetCol = ReverseCol && IsRowMajor ? PacketSize : 1; |
| using reverse_packet = reverse_packet_cond<PacketType, ReversePacket>; |
| |
| Index actualRow = ReverseRow ? m_rows.value() - row - OffsetRow : row; |
| Index actualCol = ReverseCol ? m_cols.value() - col - OffsetCol : col; |
| |
| return reverse_packet::run(m_argImpl.template packet<LoadMode, PacketType>(actualRow, actualCol)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packet(Index index) const { |
| static constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| |
| Index actualIndex = m_rows.value() * m_cols.value() - index - PacketSize; |
| |
| return preverse(m_argImpl.template packet<LoadMode, PacketType>(actualIndex)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index row, Index col, const PacketType& x) { |
| static constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| static constexpr int OffsetRow = ReverseRow && IsColMajor ? PacketSize : 1; |
| static constexpr int OffsetCol = ReverseCol && IsRowMajor ? PacketSize : 1; |
| using reverse_packet = reverse_packet_cond<PacketType, ReversePacket>; |
| |
| Index actualRow = ReverseRow ? m_rows.value() - row - OffsetRow : row; |
| Index actualCol = ReverseCol ? m_cols.value() - col - OffsetCol : col; |
| |
| m_argImpl.template writePacket<LoadMode>(actualRow, actualCol, reverse_packet::run(x)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacket(Index index, const PacketType& x) { |
| static constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| |
| Index actualIndex = m_rows.value() * m_cols.value() - index - PacketSize; |
| |
| m_argImpl.template writePacket<LoadMode>(actualIndex, preverse(x)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index row, Index col, Index begin, Index count) const { |
| static constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| static constexpr int OffsetRow = ReverseRow && IsColMajor ? PacketSize : 1; |
| static constexpr int OffsetCol = ReverseCol && IsRowMajor ? PacketSize : 1; |
| using reverse_packet = reverse_packet_cond<PacketType, ReversePacket>; |
| |
| Index actualRow = ReverseRow ? m_rows.value() - row - OffsetRow : row; |
| Index actualCol = ReverseCol ? m_cols.value() - col - OffsetCol : col; |
| Index actualBegin = ReversePacket ? (PacketSize - count - begin) : begin; |
| |
| return reverse_packet::run( |
| m_argImpl.template packetSegment<LoadMode, PacketType>(actualRow, actualCol, actualBegin, count)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE PacketType packetSegment(Index index, Index begin, Index count) const { |
| static constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| |
| Index actualIndex = m_rows.value() * m_cols.value() - index - PacketSize; |
| Index actualBegin = PacketSize - count - begin; |
| |
| return preverse(m_argImpl.template packetSegment<LoadMode, PacketType>(actualIndex, actualBegin, count)); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index row, Index col, const PacketType& x, Index begin, |
| Index count) { |
| static constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| static constexpr int OffsetRow = ReverseRow && IsColMajor ? PacketSize : 1; |
| static constexpr int OffsetCol = ReverseCol && IsRowMajor ? PacketSize : 1; |
| using reverse_packet = reverse_packet_cond<PacketType, ReversePacket>; |
| |
| Index actualRow = ReverseRow ? m_rows.value() - row - OffsetRow : row; |
| Index actualCol = ReverseCol ? m_cols.value() - col - OffsetCol : col; |
| Index actualBegin = ReversePacket ? (PacketSize - count - begin) : begin; |
| |
| m_argImpl.template writePacketSegment<LoadMode>(actualRow, actualCol, reverse_packet::run(x), actualBegin, count); |
| } |
| |
| template <int LoadMode, typename PacketType> |
| EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void writePacketSegment(Index index, const PacketType& x, Index begin, |
| Index count) { |
| static constexpr int PacketSize = unpacket_traits<PacketType>::size; |
| |
| Index actualIndex = m_rows.value() * m_cols.value() - index - PacketSize; |
| Index actualBegin = PacketSize - count - begin; |
| |
| m_argImpl.template writePacketSegment<LoadMode>(actualIndex, preverse(x), actualBegin, count); |
| } |
| |
| protected: |
| evaluator<ArgType> m_argImpl; |
| |
| // If we do not reverse rows, then we do not need to know the number of rows; same for columns |
| // Nonetheless, in this case it is important to set to 1 such that the coeff(index) method works fine for vectors. |
| const variable_if_dynamic<Index, ReverseRow ? ArgType::RowsAtCompileTime : 1> m_rows; |
| const variable_if_dynamic<Index, ReverseCol ? ArgType::ColsAtCompileTime : 1> m_cols; |
| }; |
| |
| // -------------------- Diagonal -------------------- |
| |
| template <typename ArgType, int DiagIndex> |
| struct evaluator<Diagonal<ArgType, DiagIndex>> : evaluator_base<Diagonal<ArgType, DiagIndex>> { |
| using XprType = Diagonal<ArgType, DiagIndex>; |
| |
| enum { |
| CoeffReadCost = evaluator<ArgType>::CoeffReadCost, |
| |
| Flags = static_cast<unsigned int>(evaluator<ArgType>::Flags & (HereditaryBits | DirectAccessBit) & ~RowMajorBit) | |
| LinearAccessBit, |
| |
| Alignment = 0 |
| }; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE explicit evaluator(const XprType& diagonal) |
| : m_argImpl(diagonal.nestedExpression()), m_index(diagonal.index()) {} |
| |
| using Scalar = typename XprType::Scalar; |
| using CoeffReturnType = typename XprType::CoeffReturnType; |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index row, Index) const { |
| return m_argImpl.coeff(row + rowOffset(), row + colOffset()); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE CoeffReturnType coeff(Index index) const { |
| return m_argImpl.coeff(index + rowOffset(), index + colOffset()); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index row, Index) { |
| return m_argImpl.coeffRef(row + rowOffset(), row + colOffset()); |
| } |
| |
| EIGEN_DEVICE_FUNC constexpr EIGEN_STRONG_INLINE Scalar& coeffRef(Index index) { |
| return m_argImpl.coeffRef(index + rowOffset(), index + colOffset()); |
| } |
| |
| protected: |
| evaluator<ArgType> m_argImpl; |
| const variable_if_dynamicindex<Index, XprType::DiagIndex> m_index; |
| |
| private: |
| EIGEN_DEVICE_FUNC constexpr Index rowOffset() const { return m_index.value() > 0 ? 0 : -m_index.value(); } |
| EIGEN_DEVICE_FUNC constexpr Index colOffset() const { return m_index.value() > 0 ? m_index.value() : 0; } |
| }; |
| |
| //---------------------------------------------------------------------- |
| // deprecated code |
| //---------------------------------------------------------------------- |
| |
| // -------------------- EvalToTemp -------------------- |
| |
| // expression class for evaluating nested expression to a temporary |
| |
| template <typename ArgType> |
| class EvalToTemp; |
| |
| template <typename ArgType> |
| struct traits<EvalToTemp<ArgType>> : public traits<ArgType> {}; |
| |
| template <typename ArgType> |
| class EvalToTemp : public dense_xpr_base<EvalToTemp<ArgType>>::type { |
| public: |
| using Base = typename dense_xpr_base<EvalToTemp>::type; |
| EIGEN_GENERIC_PUBLIC_INTERFACE(EvalToTemp) |
| |
| explicit EvalToTemp(const ArgType& arg) : m_arg(arg) {} |
| |
| const ArgType& arg() const { return m_arg; } |
| |
| constexpr Index rows() const noexcept { return m_arg.rows(); } |
| |
| constexpr Index cols() const noexcept { return m_arg.cols(); } |
| |
| private: |
| const ArgType& m_arg; |
| }; |
| |
| template <typename ArgType> |
| struct evaluator<EvalToTemp<ArgType>> : public evaluator<typename ArgType::PlainObject> { |
| using XprType = EvalToTemp<ArgType>; |
| using PlainObject = typename ArgType::PlainObject; |
| using Base = evaluator<PlainObject>; |
| |
| EIGEN_DEVICE_FUNC constexpr explicit evaluator(const XprType& xpr) : m_result(xpr.arg()) { |
| internal::construct_at<Base>(this, m_result); |
| } |
| |
| // This constructor is used when nesting an EvalTo evaluator in another evaluator |
| EIGEN_DEVICE_FUNC constexpr evaluator(const ArgType& arg) : m_result(arg) { |
| internal::construct_at<Base>(this, m_result); |
| } |
| |
| protected: |
| PlainObject m_result; |
| }; |
| |
| } // namespace internal |
| |
| } // end namespace Eigen |
| |
| #endif // EIGEN_COREEVALUATORS_H |