[mlpack-git] master: implement serialize (30d652d)
gitdub at mlpack.org
gitdub at mlpack.org
Sat Feb 27 02:37:15 EST 2016
Repository : https://github.com/mlpack/mlpack
On branch : master
Link : https://github.com/mlpack/mlpack/compare/a2e57d617e952f1ea2fda8a23e1c6bd2f78beb6d...1bedf15126f6bd0bc93a3233914ac95486a3c0da
>---------------------------------------------------------------
commit 30d652d6db07ddd962f970769b5a585aad121622
Author: stereomatchingkiss <stereomatchingkiss at gmail.com>
Date: Sat Feb 27 15:37:15 2016 +0800
implement serialize
>---------------------------------------------------------------
30d652d6db07ddd962f970769b5a585aad121622
src/mlpack/methods/ann/layer/softmax_layer.hpp | 8 ++++++++
src/mlpack/methods/ann/layer/sparse_bias_layer.hpp | 10 ++++++++++
src/mlpack/methods/ann/layer/sparse_input_layer.hpp | 10 ++++++++++
src/mlpack/methods/ann/layer/sparse_output_layer.hpp | 12 ++++++++++++
4 files changed, 40 insertions(+)
diff --git a/src/mlpack/methods/ann/layer/softmax_layer.hpp b/src/mlpack/methods/ann/layer/softmax_layer.hpp
index 12e0146..bb1ff89 100644
--- a/src/mlpack/methods/ann/layer/softmax_layer.hpp
+++ b/src/mlpack/methods/ann/layer/softmax_layer.hpp
@@ -97,6 +97,14 @@ class SoftmaxLayer
//! Modify the delta.
InputDataType& Delta() { return delta; }
+ /**
+ * Serialize the layer
+ */
+ template<typename Archive>
+ void Serialize(Archive& ar, const unsigned int /* version */)
+ {
+ }
+
private:
//! Locally-stored delta object.
OutputDataType delta;
diff --git a/src/mlpack/methods/ann/layer/sparse_bias_layer.hpp b/src/mlpack/methods/ann/layer/sparse_bias_layer.hpp
index 2723a98..effe912 100644
--- a/src/mlpack/methods/ann/layer/sparse_bias_layer.hpp
+++ b/src/mlpack/methods/ann/layer/sparse_bias_layer.hpp
@@ -131,6 +131,16 @@ class SparseBiasLayer
//! Modify the gradient.
InputDataType& Gradient() { return gradient; }
+ /**
+ * Serialize the layer
+ */
+ template<typename Archive>
+ void Serialize(Archive& ar, const unsigned int /* version */)
+ {
+ ar & data::CreateNVP(lambda, "lambda");
+ ar & data::CreateNVP(weights, "weights");
+ }
+
private:
//! Locally-stored number of output units.
size_t outSize;
diff --git a/src/mlpack/methods/ann/layer/sparse_input_layer.hpp b/src/mlpack/methods/ann/layer/sparse_input_layer.hpp
index efdb777..fbb0d67 100644
--- a/src/mlpack/methods/ann/layer/sparse_input_layer.hpp
+++ b/src/mlpack/methods/ann/layer/sparse_input_layer.hpp
@@ -133,6 +133,16 @@ class SparseInputLayer
//! Modify the gradient.
OutputDataType& Gradient() { return gradient; }
+ /**
+ * Serialize the layer
+ */
+ template<typename Archive>
+ void Serialize(Archive& ar, const unsigned int /* version */)
+ {
+ ar & data::CreateNVP(lambda, "lambda");
+ ar & data::CreateNVP(weights, "weights");
+ }
+
private:
//! Locally-stored number of input units.
size_t inSize;
diff --git a/src/mlpack/methods/ann/layer/sparse_output_layer.hpp b/src/mlpack/methods/ann/layer/sparse_output_layer.hpp
index 950081a..1d5cf70 100644
--- a/src/mlpack/methods/ann/layer/sparse_output_layer.hpp
+++ b/src/mlpack/methods/ann/layer/sparse_output_layer.hpp
@@ -143,6 +143,18 @@ class SparseOutputLayer
return rho;
}
+ /**
+ * Serialize the layer
+ */
+ template<typename Archive>
+ void Serialize(Archive& ar, const unsigned int /* version */)
+ {
+ ar & data::CreateNVP(rhoCap, "rhoCap");
+ ar & data::CreateNVP(lambda, "lambda");
+ ar & data::CreateNVP(beta, "beta");
+ ar & data::CreateNVP(weights, "weights");
+ }
+
//! Get the weights.
OutputDataType const& Weights() const { return weights; }
//! Modify the weights.
More information about the mlpack-git
mailing list