Zhilin successfully defended his Ph.D. dissertation on 12/12/12. Congratulations Dr. Zhang !

Masaaki points to Compressed Sensing as a New Trend in Control

Suresh makes a summary of his thoughts about NIPS in:

In his NIPS ruminations I Suresh mentioned this about the Earth Mover's distance:

Kernel distances: Ever since I discovered the kernel distance (as in, found it, not invented it) I've been fascinated by how it behaves more or less like the earth mover distance, but is so much easier to compute. Scott Aaronson (at his NIPS invited talk) made this joke about how nature loves ℓ 2. The kernel distance is "essentially" the ℓ variant of EMD (which makes so many things easier). There's been a series of papers by Sriperumbudur et al. on this topic, and in a series of works they have shown that (a) the kernel distance captures the notion of "distance covariance" that has become popular in statistics as a way of testing independence of distributions. (b) as an estimator of distance between distributions, the kernel distance has more efficient estimators than (say) the EMD because its estimator can be computed in closed form instead of needing an algorithm that solves a transportation problem and (c ) the kernel that optimizes the efficient of the two-sample estimator can also be determined (the NIPS paper).

We've seen EMD recently. In Learning Manifolds in the Wild by Chinmay Hegde, Aswin C. Sankaranarayanan, Richard Baraniuk made the case that one could learn manifolds through the use of the Earth Mover’s Distance on top of keypoint descriptors. Does that mean that a combination of the faster FREAKs and kernel distance might provide for a speedier way of learning manifold from images and videos ? [Update: the answer seems to be Yes]

As an aside, one of the commenters pointed out to a manuscript on the Earth Mover's distance by Cedric Villani where you get to learn (page 33) the historical nature of the concept from Monge. It is quite fascinating that indeed it has to do with moving soil from one place to another.

On the PSD matrices remark, I am reminded of the work by Frédéric Barbaresco on Applications Radar de la Géométrie de l'information associée aux matrices de covariances : traitements spatio-temporels. It's in French but understandable, there is also a video of him on Applications of Information Geometry to Radar Signal Processing (Interactions between symmetric cone and Information Geomtries: Bruhat-Tits and Siegel Spaces, Models for High Resolution Autoregressive Doppler Imagery). As another aside, I wonder if one could multiplex the various modailities of radar.

Sergey talks about Algebraic topology – now in compressed sensingRich talks about NuMax – A Convex Approach for Learning Near-Isometric Linear Embeddings, we featured here earlier.

Larry has New Names For Statistical Methods

Bob points out that Even humans can make extreme octave errors

Danny talks about Collaborative filtering:

- Collaborative Filtering - 3rd Generation [or] winning the kdd cup in 5 minutes!
- Collaborative filtering - 3rd generation - part 2
- Some interesting feedback for the 3rd generation cf discussed in this blog

Mathblogging features a new blog in Mathematical Instruments: Haggis the Sheep

John mentions the The Lindy effect

Laurent features the ERBlet transform (on WITS: Where is the starlet)

Emmanuel talks about a Paper: Inverting and Visualizing Features for Object Detection

Andrew ia announcing Sublinear.Info!, that page is now part of the highly technical reference page

OpenPicus has a 20% discount megasaale until december 31st.

**Join the CompressiveSensing subreddit or the Google+ Community and post there !**

Liked this entry ? subscribe to Nuit Blanche's feed, there's more where that came from. You can also subscribe to Nuit Blanche by Email, explore the Big Picture in Compressive Sensing or the Matrix Factorization Jungle and join the conversations on compressive sensing, advanced matrix factorization and calibration issues on Linkedin.

## 1 comment:

Thank you Igor!

Post a Comment