[mlpack-git] master: const correctness and adjust format (83f3752)

gitdub at mlpack.org gitdub at mlpack.org
Wed Mar 2 00:38:09 EST 2016


Repository : https://github.com/mlpack/mlpack
On branch  : master
Link       : https://github.com/mlpack/mlpack/compare/a2e57d617e952f1ea2fda8a23e1c6bd2f78beb6d...1bedf15126f6bd0bc93a3233914ac95486a3c0da

>---------------------------------------------------------------

commit 83f3752ba27a82100dcf724502782487cfa3668c
Author: stereomatchingkiss <stereomatchingkiss at gmail.com>
Date:   Wed Mar 2 13:38:09 2016 +0800

    const correctness and adjust format


>---------------------------------------------------------------

83f3752ba27a82100dcf724502782487cfa3668c
 src/mlpack/methods/ann/layer/conv_layer.hpp          | 12 ++++++------
 src/mlpack/methods/ann/layer/dropout_layer.hpp       |  6 +++---
 src/mlpack/methods/ann/layer/lstm_layer.hpp          | 10 +++++-----
 src/mlpack/methods/ann/layer/pooling_layer.hpp       | 14 +++-----------
 src/mlpack/methods/ann/layer/recurrent_layer.hpp     | 16 ++++++++--------
 src/mlpack/methods/ann/layer/softmax_layer.hpp       |  6 +++---
 src/mlpack/methods/ann/layer/sparse_bias_layer.hpp   |  2 +-
 src/mlpack/methods/ann/layer/sparse_input_layer.hpp  |  4 ++--
 src/mlpack/methods/ann/layer/sparse_output_layer.hpp |  2 +-
 9 files changed, 32 insertions(+), 40 deletions(-)

diff --git a/src/mlpack/methods/ann/layer/conv_layer.hpp b/src/mlpack/methods/ann/layer/conv_layer.hpp
index 266742a..21e3f6f 100644
--- a/src/mlpack/methods/ann/layer/conv_layer.hpp
+++ b/src/mlpack/methods/ann/layer/conv_layer.hpp
@@ -160,27 +160,27 @@ class ConvLayer
   }  
 
   //! Get the weights.
-  OutputDataType& Weights() const { return weights; }
+  OutputDataType const& Weights() const { return weights; }
   //! Modify the weights.
   OutputDataType& Weights() { return weights; }
 
   //! Get the input parameter.
-  InputDataType& InputParameter() const { return inputParameter; }
+  InputDataType const& InputParameter() const { return inputParameter; }
   //! Modify the input parameter.
   InputDataType& InputParameter() { return inputParameter; }
 
   //! Get the output parameter.
-  OutputDataType& OutputParameter() const { return outputParameter; }
+  OutputDataType const& OutputParameter() const { return outputParameter; }
   //! Modify the output parameter.
   OutputDataType& OutputParameter() { return outputParameter; }
 
   //! Get the delta.
-  OutputDataType& Delta() const { return delta; }
+  OutputDataType const& Delta() const { return delta; }
   //! Modify the delta.
   OutputDataType& Delta() { return delta; }
 
   //! Get the gradient.
-  OutputDataType& Gradient() const { return gradient; }
+  OutputDataType const& Gradient() const { return gradient; }
   //! Modify the gradient.
   OutputDataType& Gradient() { return gradient; }
   
@@ -190,7 +190,7 @@ class ConvLayer
   template<typename Archive>
   void Serialize(Archive& ar, const unsigned int /* version */)
   {    			
-	ar & data::CreateNVP(weights, "weights");
+    ar & data::CreateNVP(weights, "weights");
   }
 
  private:
diff --git a/src/mlpack/methods/ann/layer/dropout_layer.hpp b/src/mlpack/methods/ann/layer/dropout_layer.hpp
index eba1a2e..b191583 100644
--- a/src/mlpack/methods/ann/layer/dropout_layer.hpp
+++ b/src/mlpack/methods/ann/layer/dropout_layer.hpp
@@ -147,17 +147,17 @@ class DropoutLayer
   }
 
   //! Get the input parameter.
-  InputDataType& InputParameter() const { return inputParameter; }
+  InputDataType const& InputParameter() const { return inputParameter; }
   //! Modify the input parameter.
   InputDataType& InputParameter() { return inputParameter; }
 
   //! Get the output parameter.
-  OutputDataType& OutputParameter() const { return outputParameter; }
+  OutputDataType const& OutputParameter() const { return outputParameter; }
   //! Modify the output parameter.
   OutputDataType& OutputParameter() { return outputParameter; }
 
   //! Get the detla.
-  OutputDataType& Delta() const { return delta; }
+  OutputDataType const& Delta() const { return delta; }
   //! Modify the delta.
   OutputDataType& Delta() { return delta; }
 
diff --git a/src/mlpack/methods/ann/layer/lstm_layer.hpp b/src/mlpack/methods/ann/layer/lstm_layer.hpp
index 1ff6278..d424676 100644
--- a/src/mlpack/methods/ann/layer/lstm_layer.hpp
+++ b/src/mlpack/methods/ann/layer/lstm_layer.hpp
@@ -248,27 +248,27 @@ class LSTMLayer
   }
 
   //! Get the peephole weights.
-  OutputDataType& Weights() const { return peepholeWeights; }
+  OutputDataType const& Weights() const { return peepholeWeights; }
   //! Modify the peephole weights.
   OutputDataType& Weights() { return peepholeWeights; }
 
   //! Get the input parameter.
-  InputDataType& InputParameter() const { return inputParameter; }
+  InputDataType const& InputParameter() const { return inputParameter; }
   //! Modify the input parameter.
   InputDataType& InputParameter() { return inputParameter; }
 
   //! Get the output parameter.
-  OutputDataType& OutputParameter() const { return outputParameter; }
+  OutputDataType const& OutputParameter() const { return outputParameter; }
   //! Modify the output parameter.
   OutputDataType& OutputParameter() { return outputParameter; }
 
   //! Get the delta.
-  OutputDataType& Delta() const { return delta; }
+  OutputDataType const& Delta() const { return delta; }
   //! Modify the delta.
   OutputDataType& Delta() { return delta; }
 
   //! Get the peephole gradient.
-  OutputDataType& Gradient() const { return peepholeGradient; }
+  OutputDataType const& Gradient() const { return peepholeGradient; }
   //! Modify the peephole gradient.
   OutputDataType& Gradient() { return peepholeGradient; }
 
diff --git a/src/mlpack/methods/ann/layer/pooling_layer.hpp b/src/mlpack/methods/ann/layer/pooling_layer.hpp
index d091e63..78e929f 100644
--- a/src/mlpack/methods/ann/layer/pooling_layer.hpp
+++ b/src/mlpack/methods/ann/layer/pooling_layer.hpp
@@ -133,17 +133,17 @@ class PoolingLayer
   }
 
   //! Get the input parameter.
-  InputDataType& InputParameter() const { return inputParameter; }
+  InputDataType const& InputParameter() const { return inputParameter; }
   //! Modify the input parameter.
   InputDataType& InputParameter() { return inputParameter; }
 
   //! Get the output parameter.
-  InputDataType& OutputParameter() const { return outputParameter; }
+  InputDataType const& OutputParameter() const { return outputParameter; }
   //! Modify the output parameter.
   InputDataType& OutputParameter() { return outputParameter; }
 
   //! Get the delta.
-  OutputDataType& Delta() const { return delta; }
+  OutputDataType const& Delta() const { return delta; }
   //! Modify the delta.
   OutputDataType& Delta() { return delta; }   
   
@@ -155,14 +155,6 @@ class PoolingLayer
   {    				
   }
 
-  /**
-   * Serialize the layer
-   */
-  template<typename Archive>
-  void Serialize(Archive& ar, const unsigned int /* version */)
-  {    				
-  }
-
  private:
   /**
    * Apply pooling to the input and store the results.
diff --git a/src/mlpack/methods/ann/layer/recurrent_layer.hpp b/src/mlpack/methods/ann/layer/recurrent_layer.hpp
index be28680..55f3acd 100644
--- a/src/mlpack/methods/ann/layer/recurrent_layer.hpp
+++ b/src/mlpack/methods/ann/layer/recurrent_layer.hpp
@@ -100,32 +100,32 @@ class RecurrentLayer
   }
 
   //! Get the weights.
-  OutputDataType& Weights() const { return weights; }
+  OutputDataType const& Weights() const { return weights; }
   //! Modify the weights.
   OutputDataType& Weights() { return weights; }
 
   //! Get the input parameter.
-  InputDataType& InputParameter() const { return inputParameter; }
+  InputDataType const& InputParameter() const { return inputParameter; }
   //! Modify the input parameter.
   InputDataType& InputParameter() { return inputParameter; }
 
   //! Get the input parameter.
-  InputDataType& RecurrentParameter() const { return recurrentParameter; }
+  InputDataType const& RecurrentParameter() const { return recurrentParameter; }
   //! Modify the input parameter.
   InputDataType& RecurrentParameter() { return recurrentParameter; }
 
   //! Get the output parameter.
-  OutputDataType& OutputParameter() const { return outputParameter; }
+  OutputDataType const& OutputParameter() const { return outputParameter; }
   //! Modify the output parameter.
   OutputDataType& OutputParameter() { return outputParameter; }
 
   //! Get the delta.
-  OutputDataType& Delta() const { return delta; }
+  OutputDataType const& Delta() const { return delta; }
   //! Modify the delta.
   OutputDataType& Delta() { return delta; }
 
   //! Get the gradient.
-  OutputDataType& Gradient() const { return gradient; }
+  OutputDataType const& Gradient() const { return gradient; }
   //! Modify the gradient.
   OutputDataType& Gradient() { return gradient; }
   
@@ -135,8 +135,8 @@ class RecurrentLayer
   template<typename Archive>
   void Serialize(Archive& ar, const unsigned int /* version */)
   {    		
-	ar & data::CreateNVP(recurrentParameter, "recurrentParameter");	
-	ar & data::CreateNVP(weights, "weights");
+    ar & data::CreateNVP(recurrentParameter, "recurrentParameter");
+    ar & data::CreateNVP(weights, "weights");
   }
 
  private:
diff --git a/src/mlpack/methods/ann/layer/softmax_layer.hpp b/src/mlpack/methods/ann/layer/softmax_layer.hpp
index 5d282ba..0875084 100644
--- a/src/mlpack/methods/ann/layer/softmax_layer.hpp
+++ b/src/mlpack/methods/ann/layer/softmax_layer.hpp
@@ -69,17 +69,17 @@ class SoftmaxLayer
   }
 
   //! Get the input parameter.
-  InputDataType& InputParameter() const { return inputParameter; }
+  InputDataType const& InputParameter() const { return inputParameter; }
   //! Modify the input parameter.
   InputDataType& InputParameter() { return inputParameter; }
 
   //! Get the output parameter.
-  OutputDataType& OutputParameter() const { return outputParameter; }
+  OutputDataType const& OutputParameter() const { return outputParameter; }
   //! Modify the output parameter.
   OutputDataType& OutputParameter() { return outputParameter; }
 
   //! Get the delta.
-  InputDataType& Delta() const { return delta; }
+  InputDataType const& Delta() const { return delta; }
   //! Modify the delta.
   InputDataType& Delta() { return delta; }
   
diff --git a/src/mlpack/methods/ann/layer/sparse_bias_layer.hpp b/src/mlpack/methods/ann/layer/sparse_bias_layer.hpp
index 4d83f1b..c166dda 100644
--- a/src/mlpack/methods/ann/layer/sparse_bias_layer.hpp
+++ b/src/mlpack/methods/ann/layer/sparse_bias_layer.hpp
@@ -123,7 +123,7 @@ class SparseBiasLayer
   template<typename Archive>
   void Serialize(Archive& ar, const unsigned int /* version */)
   {    			
-	ar & data::CreateNVP(weights, "weights");
+    ar & data::CreateNVP(weights, "weights");
   }
 
  private:
diff --git a/src/mlpack/methods/ann/layer/sparse_input_layer.hpp b/src/mlpack/methods/ann/layer/sparse_input_layer.hpp
index faa697d..d5384da 100644
--- a/src/mlpack/methods/ann/layer/sparse_input_layer.hpp
+++ b/src/mlpack/methods/ann/layer/sparse_input_layer.hpp
@@ -114,7 +114,7 @@ class SparseInputLayer
   OutputDataType& Delta() { return delta; }
 
   //! Get the gradient.
-  OutputDataType& Gradient() const { return gradient; }
+  OutputDataType const& Gradient() const { return gradient; }
   //! Modify the gradient.
   OutputDataType& Gradient() { return gradient; }
   
@@ -124,7 +124,7 @@ class SparseInputLayer
   template<typename Archive>
   void Serialize(Archive& ar, const unsigned int /* version */)
   {    			
-	ar & data::CreateNVP(weights, "weights");
+    ar & data::CreateNVP(weights, "weights");
   }
 
  private:
diff --git a/src/mlpack/methods/ann/layer/sparse_output_layer.hpp b/src/mlpack/methods/ann/layer/sparse_output_layer.hpp
index a27569d..3e647d6 100644
--- a/src/mlpack/methods/ann/layer/sparse_output_layer.hpp
+++ b/src/mlpack/methods/ann/layer/sparse_output_layer.hpp
@@ -162,7 +162,7 @@ class SparseOutputLayer
   template<typename Archive>
   void Serialize(Archive& ar, const unsigned int /* version */)
   {    	
-	ar & data::CreateNVP(weights, "weights");
+    ar & data::CreateNVP(weights, "weights");
   }
 
  private:




More information about the mlpack-git mailing list