Wednesday, April 15, 2015

Paris Machine Learning Meetup #8, Season 2: Deep Learning and more...

 
This is the where the meetup will be streamed starting at around 7:15PM Paris time.


The theme of today's meetup will about (mostly) Deep Learning. It will be held in conjunction with the Deep Learning Paris meetup . We should also have the Kiev Deep Learning Meetup as a guest audience as well. All the presentation slides should be available below by the time the meetup starts. Our host and sponsor in Paris will be Criteo. Here is the tentative schedule/program that starts at 19h15 Paris time.


and the presentation slides:
Presentation of Criteo by Damien Lefortier

+ Yoshua Bengio, Title: Deep Learning Theory (remote from the London Machine Learning meetup)

Although neural networks have long been considered lacking in theory and much remains to be done, theoretical evidence is mounting and will be discussed, to support distributed representations, depth of representation, the non-convexity of the training objective, and the probabilistic interpretation of learning algorithms (especially of the auto-encoder type, which were lacking one). The talk will focus on the intuitions behind these theoretical results.

+ Sander Dieleman and Ira Korshunova, Ghent University

Title: Classifying plankton with deep neural networks by the Deep Sea team from Reservoir Lab
 ( ppt version
Deep learning has become a very popular approach for solving computer vision problems in recent years. In this talk we'll demonstrate how this approach can be applied in practice. We'll show how our team of 7 built a model for the automated classification of plankton based on convolutional neural networks. Using this model, we placed 1st in the National Data Science Bowl competition on Kaggle.
+ Gabriella Contardo, LIP6, UPMC 
Learning to build representations from partial information: Application to cold-start recommendation
Most of the successful machine learning algorithms rely on data representation, i.e a way to disentangle and extract useful information from data, which will help the model in its objective task. Classical approaches build representations based on fully observed data. But in many cases, one wants to build representations ''on the fly'', based on a partially observed information. As an example, learning representations over users can be done by progressively gathering information about their profiles. This paper presents an inductive representation-based model to tackle the twofold more general problem of (i) selecting the right information to collect for building relevant representations, (ii) updating these representations based on new incoming information. It is developed in this paper to design static interview for the cold-start collaborative filtering problem but it can also be used to go smoothly to the warm context where all information has been gathered.

+ Guillaume Wenzek
Sentiment Analysis With Recursive Neural Tensor Network / Analyse de sentiment à l'aide de réseaux de neurones récursifs
Sentiment analysis is one of the hardest NLP (Natural Language Processing) task, due to complex linguistic structures such as negation or double-negation. Socher et al. introduced a method that combines a classic NLP tool, a syntaxic parser, with a special kind of neural networks. We will review this method and introduce a few improvements in order to train on a corpus with fewer annotations than the Stanford Sentiment Treebank used in the paper.


 
 
Join the CompressiveSensing subreddit or the Google+ Community and post there !
Liked this entry ? subscribe to Nuit Blanche's feed, there's more where that came from. You can also subscribe to Nuit Blanche by Email, explore the Big Picture in Compressive Sensing or the Matrix Factorization Jungle and join the conversations on compressive sensing, advanced matrix factorization and calibration issues on Linkedin.

No comments:

Printfriendly