diff --git a/src/operator/ConvImpl.cpp b/src/operator/ConvImpl.cpp index d23a9968ffb424b4639e0fcd2629a3a1cc2e11c3..eae5f109f6af8298b90cc8e505ff44eff51bab5c 100644 --- a/src/operator/ConvImpl.cpp +++ b/src/operator/ConvImpl.cpp @@ -12,15 +12,18 @@ #include "aidge/backend/cpu/operator/ConvImpl.hpp" #include "aidge/backend/cpu/operator/ConvImpl_kernels.hpp" -#include <cassert> +#include <memory> +#include <vector> #include "aidge/backend/cpu/data/GetCPUPtr.h" #include "aidge/operator/Conv.hpp" +#include "aidge/utils/ErrorHandling.hpp" +#include "aidge/utils/Types.h" namespace Aidge { template <> -void Aidge::ConvImpl1D_cpu::forward() { +void ConvImpl1D_cpu::forward() { const auto& op_ = static_cast<const Conv_Op<1>&>(mOp); // FIXME: uncomment the following code once memory handling will work @@ -53,7 +56,8 @@ void Aidge::ConvImpl1D_cpu::forward() { ); } -template <> void ConvImpl1D_cpu::backward() { +template <> +void ConvImpl1D_cpu::backward() { const auto &op = dynamic_cast<const Conv1D_Op &>(mOp); const auto &outputGrad = op.getOutput(0)->grad(); AIDGE_ASSERT(outputGrad, "{}: missing ouput #0 gradient", op.type()); @@ -97,7 +101,7 @@ template <> void ConvImpl1D_cpu::backward() { } template <> -void Aidge::ConvImpl2D_cpu::forward() { +void ConvImpl2D_cpu::forward() { const auto& op_ = dynamic_cast<const Conv_Op<2>&>(mOp); // FIXME: uncomment the following code once memory handling will work @@ -130,7 +134,8 @@ void Aidge::ConvImpl2D_cpu::forward() { } -template <> void ConvImpl2D_cpu::backward() { +template <> +void ConvImpl2D_cpu::backward() { const auto &op = dynamic_cast<const Conv2D_Op &>(mOp); const auto &outputGrad = op.getOutput(0)->grad(); AIDGE_ASSERT(outputGrad, "{}: missing ouput #0 gradient", op.type());