[mlpack-git] master: Use the logistic function and the RMSProp optimizer because the combination tends to be more stable and in most cases it converges faster using the specified architecture. (952cc39)

gitdub at big.cc.gt.atl.ga.us gitdub at big.cc.gt.atl.ga.us
Fri Sep 4 08:47:05 EDT 2015


Repository : https://github.com/mlpack/mlpack

On branch  : master
Link       : https://github.com/mlpack/mlpack/compare/df9446c5a37c0de0955b6e1f20983cdcf611074c...ef290321115f6dfd21522ba7ccec5f08b52d7631

>---------------------------------------------------------------

commit 952cc39ac15c5851c82fd3603ef2489465b3ade1
Author: Marcus Edel <marcus.edel at fu-berlin.de>
Date:   Fri Sep 4 13:58:44 2015 +0200

    Use the logistic function and the RMSProp optimizer because the combination tends to be more stable and in most cases it converges faster using the specified architecture.


>---------------------------------------------------------------

952cc39ac15c5851c82fd3603ef2489465b3ade1
 src/mlpack/tests/convolutional_network_test.cpp | 18 +++++++++---------
 1 file changed, 9 insertions(+), 9 deletions(-)

diff --git a/src/mlpack/tests/convolutional_network_test.cpp b/src/mlpack/tests/convolutional_network_test.cpp
index 943db53..5a2840c 100644
--- a/src/mlpack/tests/convolutional_network_test.cpp
+++ b/src/mlpack/tests/convolutional_network_test.cpp
@@ -87,18 +87,18 @@ void BuildVanillaNetwork()
    * +---+        +---+        +---+        +---+        +---+    +---+
    */
 
-  ConvLayer<AdaDelta> convLayer0(1, 8, 5, 5);
-  BiasLayer2D<AdaDelta, ZeroInitialization> biasLayer0(8);
+  ConvLayer<RMSPROP> convLayer0(1, 8, 5, 5);
+  BiasLayer2D<RMSPROP, ZeroInitialization> biasLayer0(8);
   BaseLayer2D<PerformanceFunction> baseLayer0;
   PoolingLayer<> poolingLayer0(2);
 
-  ConvLayer<AdaDelta> convLayer1(8, 12, 5, 5);
-  BiasLayer2D<AdaDelta, ZeroInitialization> biasLayer1(12);
+  ConvLayer<RMSPROP> convLayer1(8, 12, 5, 5);
+  BiasLayer2D<RMSPROP, ZeroInitialization> biasLayer1(12);
   BaseLayer2D<PerformanceFunction> baseLayer1;
   PoolingLayer<> poolingLayer1(2);
 
-  LinearMappingLayer<AdaDelta> linearLayer0(192, 10);
-  BiasLayer<AdaDelta> biasLayer2(10);
+  LinearMappingLayer<RMSPROP> linearLayer0(192, 10);
+  BiasLayer<RMSPROP> biasLayer2(10);
   SoftmaxLayer<> softmaxLayer0;
 
   OneHotLayer outputLayer;
@@ -110,7 +110,7 @@ void BuildVanillaNetwork()
   CNN<decltype(modules), decltype(outputLayer)>
       net(modules, outputLayer);
 
-  Trainer<decltype(net)> trainer(net, 100, 1, 0.7);
+  Trainer<decltype(net)> trainer(net, 50, 1, 0.7);
   trainer.Train(input, Y, input, Y);
 
   BOOST_REQUIRE_LE(trainer.ValidationError(), 0.7);
@@ -121,7 +121,7 @@ void BuildVanillaNetwork()
  */
 BOOST_AUTO_TEST_CASE(VanillaNetworkTest)
 {
-  BuildVanillaNetwork<RectifierFunction>();
+  BuildVanillaNetwork<LogisticFunction>();
 }
 
 /**
@@ -203,7 +203,7 @@ void BuildVanillaDropoutNetwork()
   CNN<decltype(modules), decltype(outputLayer)>
       net(modules, outputLayer);
 
-  Trainer<decltype(net)> trainer(net, 100, 1, 0.7);
+  Trainer<decltype(net)> trainer(net, 50, 1, 0.7);
   trainer.Train(input, Y, input, Y);
 
   BOOST_REQUIRE_LE(trainer.ValidationError(), 0.7);



More information about the mlpack-git mailing list