[mlpack-git] [mlpack] improve speed of SparseAutoencoder and make it more flexible (#451)

Marcus Edel notifications at github.com
Fri Dec 4 13:00:43 EST 2015


Serialization for the ANN code would be absolutely great. In fact, someone asks me some time ago, to implement a feature to save the trained model, so I think the serialization function will come in handy. Opening another pull request to discuss the implementation details sounds great. In most cases it shouldn't be more than saving the weights.


Reusing the existing ann code to implement an autoencoder should be straightforward. We have to implement a new performance function to calculate the reconstruction error, the regularization cost and the KL divergence cost terms to monitor the training process and because we don't have a target value we use the already provided input.

I don't have a good solution for warning the user if the layer combinations doesn't make sense at all. As @rcurtin already said using static_assert could be a good way to solve the problem, but I'm not sure if it's the best. Every time someone implements a new layer, he needs to think about in which context this layer makes sense, maybe this effort is negligible. So maybe it's enough, if we just define some default models. If a user is interested in a special layer combination, I would say he knows what combination makes sense. I guess this is what @rcurtin suggested as another solution.

---
Reply to this email directly or view it on GitHub:
https://github.com/mlpack/mlpack/pull/451#issuecomment-162036050
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://mailman.cc.gatech.edu/pipermail/mlpack-git/attachments/20151204/bc6c7dc4/attachment.html>


More information about the mlpack-git mailing list