TY - GEN
T1 - Variational capsule encoder
AU - RaviPrakash, Harish
AU - Anwar, Syed Muhammad
AU - Bagci, Ulas
N1 - Funding Information:
ACKNOWLEDGMENT This project is partially supported by NIH R01-CA246704 and R01-CA240639, and Florida Department of Health 20K04.
Publisher Copyright:
© 2020 IEEE
PY - 2020
Y1 - 2020
N2 - We propose a novel capsule network based variational encoder architecture, called Bayesian capsules (B-Caps), to modulate the mean and standard deviation of the sampling distribution in the latent space. We hypothesized that this approach can learn a better representation of features in the latent space than traditional approaches. Our hypothesis was tested by using the learned latent variables for image reconstruction task, where for MNIST and Fashion-MNIST datasets, different classes were separated successfully in the latent space using our proposed model. Our experimental results have shown improved reconstruction and classification performances for both datasets adding credence to our hypothesis. We also showed that by increasing the latent space dimension, the proposed B-Caps was able to learn a better representation when compared to the traditional variational auto-encoders (VAE). Hence our results indicate the strength of capsule networks in representation learning which has never been examined under the VAE settings before.
AB - We propose a novel capsule network based variational encoder architecture, called Bayesian capsules (B-Caps), to modulate the mean and standard deviation of the sampling distribution in the latent space. We hypothesized that this approach can learn a better representation of features in the latent space than traditional approaches. Our hypothesis was tested by using the learned latent variables for image reconstruction task, where for MNIST and Fashion-MNIST datasets, different classes were separated successfully in the latent space using our proposed model. Our experimental results have shown improved reconstruction and classification performances for both datasets adding credence to our hypothesis. We also showed that by increasing the latent space dimension, the proposed B-Caps was able to learn a better representation when compared to the traditional variational auto-encoders (VAE). Hence our results indicate the strength of capsule networks in representation learning which has never been examined under the VAE settings before.
KW - Capsule network
KW - Data-driven sampling
KW - Deep learning
KW - VAE
UR - http://www.scopus.com/inward/record.url?scp=85110427820&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85110427820&partnerID=8YFLogxK
U2 - 10.1109/ICPR48806.2021.9411953
DO - 10.1109/ICPR48806.2021.9411953
M3 - Conference contribution
AN - SCOPUS:85110427820
T3 - Proceedings - International Conference on Pattern Recognition
SP - 5820
EP - 5827
BT - Proceedings of ICPR 2020 - 25th International Conference on Pattern Recognition
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 25th International Conference on Pattern Recognition, ICPR 2020
Y2 - 10 January 2021 through 15 January 2021
ER -