Brain-like approaches to unsupervised learning of hidden representations
-- a comparative study
International Conference on Artificial Neural Networks (ICANN), 2020
- BDLSSL
Abstract
Unsupervised learning of hidden representations has been one of the most vibrant research directions in machine learning in recent years. In this work we study the brain-like Bayesian Confidence Propagating Neural Network (BCPNN) model, recently extended to extract sparse distributed high-dimensional representations. The saliency and separability of the hidden representations when trained on MNIST dataset is studied using an external classifier, and compared with other unsupervised learning methods that include restricted Boltzmann machines and autoencoders.
View on arXivComments on this paper
