[mlpack-git] master: Initialize the gradient storage with zero. (2c9cb2c)

gitdub at big.cc.gt.atl.ga.us gitdub at big.cc.gt.atl.ga.us
Fri Jan 9 07:30:24 EST 2015


Repository : https://github.com/mlpack/mlpack

On branch  : master
Link       : https://github.com/mlpack/mlpack/compare/ef989e99696f41f1145c191636a9677cb33ae908...2c9cb2c2a51d74b42465aae892f29e1e4b842156

>---------------------------------------------------------------

commit 2c9cb2c2a51d74b42465aae892f29e1e4b842156
Author: Marcus Edel <marcus.edel at fu-berlin.de>
Date:   Fri Jan 9 13:23:45 2015 +0100

    Initialize the gradient storage with zero.


>---------------------------------------------------------------

2c9cb2c2a51d74b42465aae892f29e1e4b842156
 src/mlpack/methods/ann/ffnn.hpp | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/src/mlpack/methods/ann/ffnn.hpp b/src/mlpack/methods/ann/ffnn.hpp
index 6bb30c3..907782e 100644
--- a/src/mlpack/methods/ann/ffnn.hpp
+++ b/src/mlpack/methods/ann/ffnn.hpp
@@ -67,6 +67,7 @@ class FFNN
       ResetActivations(network);
       std::get<0>(
             std::get<0>(network)).InputLayer().InputActivation() = input;
+
       FeedForward(network, target, error);
     }
 
@@ -406,7 +407,7 @@ class FFNN
     Layer(std::tuple<Tp...>& t)
     {
       gradients.push_back(new MatType(std::get<I>(t).Weights().n_rows,
-          std::get<I>(t).Weights().n_cols));
+          std::get<I>(t).Weights().n_cols, arma::fill::zeros));
 
       Layer<I + 1, Tp...>(t);
     }



More information about the mlpack-git mailing list