The proceedings for ICML 2015 which is to take place in Lille, is out. Here is a small sample of papers that we mentioned before or are of interest to the general themes covered on Nuit Blanche:
Towards a Lower Sample Complexity for Robust One-bit Compressed Sensing
Large-scale log-determinant computation through stochastic Chebyshev expansions
Multi-view Sparse Co-clustering via Proximal Alternating Linearized Minimization
Learning Word Representations with Hierarchical Sparse Coding
Theory of Dual-sparse Regularized Randomized Reduction
Streaming Sparse Principal Component Analysis
Multi-view Sparse Co-clustering via Proximal Alternating Linearized Minimization
Inferring Graphs from Cascades: A Sparse Recovery Framework
Swept Approximate Message Passing for Sparse Estimation
Sparse Variational Inference for Generalized GP Models
A Deterministic Analysis of Noisy Sparse Subspace Clustering for Dimensionality-reduced Data
Sparse Subspace Clustering with Missing Entries
Theory of Dual-sparse Regularized Randomized Reduction
Statistical and Algorithmic Perspectives on Randomized Sketching for Ordinary Least-Squares
The Power of Randomization: Distributed Submodular Maximization on Massive Datasets
A Unified Framework for Outlier-Robust PCA-like Algorithms
A Stochastic PCA and SVD Algorithm with an Exponential Convergence Rate
Pushing the Limits of Affine Rank Minimization by Adapting Probabilistic PCA
Stay on path: PCA along graph paths
Credit: Dawn RC3 Image 20
This image of Ceres is part of a sequence taken by NASA's Dawn spacecraft on May 7, 2015, from a distance of 8,400 miles (13,600 kilometers).
Image credit: NASA/JPL-Caltech/UCLA/MPS/DLR/IDA
Image credit: NASA/JPL-Caltech/UCLA/MPS/DLR/IDA
Liked this entry ? subscribe to Nuit Blanche's feed, there's more where that came from. You can also subscribe to Nuit Blanche by Email, explore the Big Picture in Compressive Sensing or the Matrix Factorization Jungle and join the conversations on compressive sensing, advanced matrix factorization and calibration issues on Linkedin.
No comments:
Post a Comment