Abstract
Autoassociative neural networks (ANNs) have been proposed as a nonlinear extension of principal component analysis (PCA), which is commonly used to identify linear variation patterns in high-dimensional data. While principal component scores represent uncorrelated features, standard backpropagation methods for training ANNs provide no guarantee of producing distinct features, which is important for interpretability and for discovering the nature of the variation patterns in the data. Here, we present an alternating nonlinear PCA method, which encourages learning of distinct features in ANNs. A new measure motivated by the condition of orthogonal loadings in PCA is proposed for measuring the extent to which the nonlinear principal components represent distinct variation patterns. We demonstrate the effectiveness of our method using a simulated point cloud data set as well as a subset of the MNIST handwritten digits data. The results show that standard ANNs consistently mix the true variation sources in the low-dimensional representation learned by the model, whereas our alternating method produces solutions where the patterns are better separated in the low-dimensional space.
Original language | English (US) |
---|---|
Article number | 7707373 |
Pages (from-to) | 156-166 |
Number of pages | 11 |
Journal | IEEE Transactions on Neural Networks and Learning Systems |
Volume | 29 |
Issue number | 1 |
DOIs | |
State | Published - Jan 2018 |
Funding
Manuscript received December 30, 2015; revised June 27, 2016; accepted September 26, 2016. Date of publication October 26, 2016; date of current version December 29, 2017. This work was supported by the National Science Foundation under Grant 1265713.
Keywords
- Autoassociative neural network (ANN)
- distinct features
- nonlinear principal component analysis (NLPCA)
- tangent vector
- variation pattern
ASJC Scopus subject areas
- Software
- Computer Science Applications
- Computer Networks and Communications
- Artificial Intelligence