diff --git a/include/aidge/learning/optimizer/Optimizer.hpp b/include/aidge/learning/optimizer/Optimizer.hpp index 9e621875beb1cfd58bf8474753c536b8c4e5183c..195d64965d3ba4eb89c9c4d0ca2155cb719f76f3 100644 --- a/include/aidge/learning/optimizer/Optimizer.hpp +++ b/include/aidge/learning/optimizer/Optimizer.hpp @@ -49,7 +49,7 @@ public: virtual void setParameters(const std::vector<std::shared_ptr<Tensor>>& parameters) { mParameters = parameters; for (const auto& param : parameters) { - param->initGradient(); // create gradient and set it to zeros + param->initGrad(); // create gradient and set it to zeros } } diff --git a/unit_tests/optimizer/Test_SGD.cpp b/unit_tests/optimizer/Test_SGD.cpp index 17f946ae1630c2423a37f703c7923a40e5fe66bf..df9924d557d89d0483d018ce08951cf573e233d7 100644 --- a/unit_tests/optimizer/Test_SGD.cpp +++ b/unit_tests/optimizer/Test_SGD.cpp @@ -77,7 +77,7 @@ TEST_CASE("[learning/SGD] update", "[Optimizer][SGD]") { optim_tensors[i] = std::make_shared<Tensor>(dims); optim_tensors[i]->setBackend("cpu"); optim_tensors[i]->getImpl()->copy(val_tensors[i].get(), size_tensors[i]); - optim_tensors[i]->initGradient(); + optim_tensors[i]->initGrad(); grad_tensors[i] = std::make_shared<Tensor>(dims); grad_tensors[i]->setBackend("cpu");