diff --git a/unit_tests/optimizer/Test_SGD.cpp b/unit_tests/optimizer/Test_SGD.cpp index 3f1380779b90115652fe54249a77a41d58f15b13..14986a71125ca8fafe64ea2496f75ed3dfc010ef 100644 --- a/unit_tests/optimizer/Test_SGD.cpp +++ b/unit_tests/optimizer/Test_SGD.cpp @@ -9,13 +9,15 @@ * ********************************************************************************/ -#include <catch2/catch_test_macros.hpp> #include <cstddef> // std::size_t #include <memory> #include <random> // std::random_device, std::mt19937, std::uniform_int_distribution #include <set> #include <vector> +#include <catch2/catch_test_macros.hpp> +#include <fmt/core.h> + #include "aidge/data/Tensor.hpp" #include "aidge/backend/cpu/data/TensorImpl.hpp" #include "aidge/learning/learningRate/LRScheduler.hpp" @@ -81,9 +83,7 @@ TEST_CASE("[learning/SGD] update", "[Optimizer][SGD]") { tensors[i] = std::make_shared<Tensor>(dims); tensors[i]->setBackend("cpu"); tensors[i]->getImpl()->setRawPtr(val_tensors[i].get(), size_tensors[i]); - optim_tensors[i] = std::make_shared<Tensor>(dims); - optim_tensors[i]->setBackend("cpu"); - optim_tensors[i]->getImpl()->copy(val_tensors[i].get(), size_tensors[i]); + optim_tensors[i] = std::make_shared<Tensor>(tensors[i]->clone()); // optim_tensors[i]->initGrad(); grad_tensors[i] = std::make_shared<Tensor>(dims);