v1v2 (latest)
Brain-like approaches to unsupervised learning of hidden representations
-- a comparative study
International Conference on Artificial Neural Networks (ICANN), 2020
- BDLSSL
Abstract
Unsupervised learning of hidden representations has been one of the most vibrant research directions in machine learning in recent years. In this work we study the brain-like Bayesian Confidence Propagating Neural Network (BCPNN) model, recently extended to extract sparse distributed high-dimensional representations. The usefulness and class-dependent separability of the hidden representations when trained on MNIST and Fashion-MNIST datasets is studied using an external linear classifier and compared with other unsupervised learning methods that include restricted Boltzmann machines and autoencoders.
View on arXivComments on this paper
