diff --git a/aidge_core/unit_tests/test_parameters.py b/aidge_core/unit_tests/test_parameters.py index e4d2cb4faca3dda64cff6aea541c30787c23d0ad..342e4d9666910e90960b762277e8914c68df28b5 100644 --- a/aidge_core/unit_tests/test_parameters.py +++ b/aidge_core/unit_tests/test_parameters.py @@ -27,8 +27,8 @@ class test_attributes(unittest.TestCase): out_channels = 8 k_dims = [2, 2] conv_op = aidge_core.Conv2D(in_channels , out_channels, k_dims).get_operator() - self.assertEqual(conv_op.get_attr("InChannels"), in_channels) - self.assertEqual(conv_op.get_attr("OutChannels"), out_channels) + self.assertEqual(conv_op.in_channels(), in_channels) + self.assertEqual(conv_op.out_channels(), out_channels) self.assertEqual(conv_op.get_attr("KernelDims"), k_dims) def test_fc(self): diff --git a/include/aidge/operator/Softmax.hpp b/include/aidge/operator/Softmax.hpp index d48dbc2b60e46eb5c074b8adae065383e29b1769..1868dc6e3df48401ef3f8a126b07572e2f45144d 100644 --- a/include/aidge/operator/Softmax.hpp +++ b/include/aidge/operator/Softmax.hpp @@ -30,16 +30,16 @@ class Softmax_Op : public OperatorTensor, public Registrable<Softmax_Op, std::string, std::shared_ptr<OperatorImpl>(const Softmax_Op&)>, - public StaticAttributes<SoftmaxAttr, int> { + public StaticAttributes<SoftmaxAttr, std::size_t> { public: static const std::string Type; Softmax_Op() = delete; - using Attributes_ = StaticAttributes<SoftmaxAttr, int>; + using Attributes_ = StaticAttributes<SoftmaxAttr, std::size_t>; template <SoftmaxAttr e> using attr = typename Attributes_::template attr<e>; - Softmax_Op(int axis) + Softmax_Op(std::size_t axis) : OperatorTensor(Type, 1, 0, 1), Attributes_(attr<SoftmaxAttr::AxisIdx>(axis)) {} @@ -76,7 +76,7 @@ public: } }; -inline std::shared_ptr<Node> Softmax(int axis, const std::string& name = "") { +inline std::shared_ptr<Node> Softmax(std::size_t axis, const std::string& name = "") { return std::make_shared<Node>(std::make_shared<Softmax_Op>(axis), name); } } // namespace Aidge diff --git a/include/aidge/operator/Transpose.hpp b/include/aidge/operator/Transpose.hpp index 16ac2794a283d817f6a4e1586349e55ec626167e..31420110f19761442b67e9701aeca566976aee1b 100644 --- a/include/aidge/operator/Transpose.hpp +++ b/include/aidge/operator/Transpose.hpp @@ -26,9 +26,9 @@ #include "aidge/utils/Types.h" namespace Aidge { -class Transpose_OpImpl : public OperatorImpl { +class TransposeImpl : public OperatorImpl { public: - Transpose_OpImpl(const Operator& op, const std::string& backend = ""): OperatorImpl(op, backend) {} + TransposeImpl(const Operator& op, const std::string& backend = ""): OperatorImpl(op, backend) {} void forward() override; }; @@ -47,11 +47,11 @@ class Transpose_Op : public OperatorTensor, template <TransposeAttr e> using attr = typename Attributes_::template attr<e>; - Transpose_Op(const std::vector<DimSize_t> &output_dims_order) + Transpose_Op(const std::vector<DimSize_t> &outputDimsOrder) : OperatorTensor(Type, 1, 0, 1), - Attributes_(attr<TransposeAttr::OutputDimsOrder>(output_dims_order)) + Attributes_(attr<TransposeAttr::OutputDimsOrder>(outputDimsOrder)) { - mImpl = std::make_shared<Transpose_OpImpl>(*this); + mImpl = std::make_shared<TransposeImpl>(*this); } /** @@ -66,7 +66,7 @@ class Transpose_Op : public OperatorTensor, SET_IMPL_MACRO(Transpose_Op, *this, op.backend()); } else { - mImpl = std::make_shared<Transpose_OpImpl>(*this); + mImpl = std::make_shared<TransposeImpl>(*this); } } @@ -90,9 +90,9 @@ class Transpose_Op : public OperatorTensor, } }; -inline std::shared_ptr<Node> Transpose(const std::vector<DimSize_t> &output_dims_order, +inline std::shared_ptr<Node> Transpose(const std::vector<DimSize_t> &outputDimsOrder, const std::string& name = "") { - return std::make_shared<Node>(std::make_shared<Transpose_Op>(output_dims_order), name); + return std::make_shared<Node>(std::make_shared<Transpose_Op>(outputDimsOrder), name); } } // namespace Aidge diff --git a/src/graph/GraphView.cpp b/src/graph/GraphView.cpp index 0bc918995c55a914b29987506578491e2c86fae5..163ea35c716cd6948c998f1b08f9f07d28fe1940 100644 --- a/src/graph/GraphView.cpp +++ b/src/graph/GraphView.cpp @@ -119,9 +119,11 @@ void Aidge::GraphView::save(const std::string& path, bool verbose, bool showProd } } - if (node_ptr == mRootNode || node_ptr->type() != "Producer" || showProducers) { - fmt::print(fp.get(), "{}_{}({}){}\n", node_ptr->type(), namePtrTable.at(node_ptr), - givenName, nodeCls); + if (node_ptr->type() != "Producer" || showProducers) { + // if (node_ptr == mRootNode) { + fmt::print(fp.get(), "{}_{}({}){}\n", node_ptr->type(), namePtrTable.at(node_ptr), + givenName, nodeCls); + // } } } @@ -1412,10 +1414,9 @@ std::shared_ptr<Aidge::GraphView> Aidge::GraphView::cloneCallback(NodePtr(*clone } std::shared_ptr<Aidge::GraphView> Aidge::getConnectedGraphView(std::shared_ptr<Node> node) { - std::vector<NodePtr> foundNodes; - foundNodes.push_back(node); + std::vector<NodePtr> foundNodes{node}; - for (size_t curNodeIdx = 0; curNodeIdx < foundNodes.size(); ++curNodeIdx) { + for (std::size_t curNodeIdx = 0; curNodeIdx < foundNodes.size(); ++curNodeIdx) { NodePtr curNode = foundNodes[curNodeIdx]; for (auto childs : curNode->getOrderedChildren()) { diff --git a/src/operator/Transpose.cpp b/src/operator/Transpose.cpp index 08c4770e3fb43fe819a924dd963356401c3ce801..20b2e5a15508368a7a3ca3bbf80bd4174d98ae4e 100644 --- a/src/operator/Transpose.cpp +++ b/src/operator/Transpose.cpp @@ -23,7 +23,7 @@ #include "aidge/utils/Registrar.hpp" #include "aidge/utils/Types.h" -void Aidge::Transpose_OpImpl::forward() { +void Aidge::TransposeImpl::forward() { const Transpose_Op& op = dynamic_cast<const Transpose_Op&>(mOp); const auto inputDims = op.getInput(0)->dims(); const auto outputDims = op.getOutput(0)->dims(); @@ -83,7 +83,7 @@ void Aidge::Transpose_Op::setBackend(const std::string& name, Aidge::DeviceIdx_t SET_IMPL_MACRO(Transpose_Op, *this, name); } else { - mImpl = std::make_shared<Transpose_OpImpl>(*this); + mImpl = std::make_shared<TransposeImpl>(*this); } mOutputs[0]->setBackend(name, device); }