Sparse Coding, Canonical Correlation and Dictionary Learning are Matrix Factorization operations. They are used in a variety of ways in building Deep Neural architectures. Here are a few I noticed in   the 500 submissions at the ICLR 2017 conference that are in the open review process.
- Understanding Neural Sparse Coding with Matrix Factorization Thomas Moreau, Joan Bruna
 - Energy-Based Spherical Sparse Coding Bailey Kong, Charless C. Fowlkes
 - Support Regularized Sparse Coding and Its Fast Encoder Yingzhen Yang, Jiahui Yu, Pushmeet Kohli, Jianchao Yang, Thomas S. Huang
 - Transformational Sparse Coding Dimitrios C. Gklezakos, Rajesh P. N. Rao
 - NEUROGENESIS-INSPIRED DICTIONARY LEARNING: ONLINE MODEL ADAPTION IN A CHANGING WORLD Sahil Garg, Irina Rish, Guillermo Cecchi, Aurelie Lozano
 - Deep Variational Canonical Correlation Analysis Weiran Wang, Xinchen Yan, Honglak Lee, Karen Livescu
 - Differentiable Canonical Correlation Analysis Matthias Dorfer, Jan Schlüter, Gerhard Widmer
 - Deep Generalized Canonical Correlation Analysis Adrian Benton, Huda Khayrallah, Biman Gujral, Drew Reisinger, Sheng Zhang, Raman Arora
 
Credit photo: NASA/JPL/University of Arizona

No comments:
Post a Comment