This is the where the meetup will be streamed starting at around 7:15PM Paris time.
- 19h15 Introduction: Franck Bardol, Igor Carron, Charles Ollion, Tony Pinville
- 19h30 Yoshua Bengio (remote from London, 18:30PM Local time)
- 19h50 Q&A from Paris and Kiev (tentative) - after that Yoshua continues with the London crowd -
- 20h00 Sander Dieleman and Ira Korshunova followed by Q&A Paris and Kiev (tentative)
- 20h20 Gabriella Contardo, followed by Q&A Paris
- 20h50 Guillaume Wenzek, followed by Q&A Paris
and the presentation slides:
+ Presentation of Criteo by Damien Lefortier
+ Yoshua Bengio, Title: Deep Learning Theory (remote from the London Machine Learning meetup)
Although neural networks have long been considered lacking in theory and much remains to be done, theoretical evidence is mounting and will be discussed, to support distributed representations, depth of representation, the non-convexity of the training objective, and the probabilistic interpretation of learning algorithms (especially of the auto-encoder type, which were lacking one). The talk will focus on the intuitions behind these theoretical results.
+ Sander Dieleman and Ira Korshunova, Ghent University
Title: Classifying plankton with deep neural networks by the Deep Sea team from Reservoir Lab
( ppt version )
Deep learning has become a very popular approach for solving computer vision problems in recent years. In this talk we'll demonstrate how this approach can be applied in practice. We'll show how our team of 7 built a model for the automated classification of plankton based on convolutional neural networks. Using this model, we placed 1st in the National Data Science Bowl competition on Kaggle.+ Gabriella Contardo, LIP6, UPMC
Learning to build representations from partial information: Application to cold-start recommendation
Most of the successful machine learning algorithms rely on data representation, i.e a way to disentangle and extract useful information from data, which will help the model in its objective task. Classical approaches build representations based on fully observed data. But in many cases, one wants to build representations ''on the fly'', based on a partially observed information. As an example, learning representations over users can be done by progressively gathering information about their profiles. This paper presents an inductive representation-based model to tackle the twofold more general problem of (i) selecting the right information to collect for building relevant representations, (ii) updating these representations based on new incoming information. It is developed in this paper to design static interview for the cold-start collaborative filtering problem but it can also be used to go smoothly to the warm context where all information has been gathered.
+ Guillaume Wenzek
Sentiment Analysis With Recursive Neural Tensor Network / Analyse de sentiment à l'aide de réseaux de neurones récursifs
Sentiment analysis is one of the hardest NLP (Natural Language Processing) task, due to complex linguistic structures such as negation or double-negation. Socher et al. introduced a method that combines a classic NLP tool, a syntaxic parser, with a special kind of neural networks. We will review this method and introduce a few improvements in order to train on a corpus with fewer annotations than the Stanford Sentiment Treebank used in the paper.
Liked this entry ? subscribe to Nuit Blanche's feed, there's more where that came from. You can also subscribe to Nuit Blanche by Email, explore the Big Picture in Compressive Sensing or the Matrix Factorization Jungle and join the conversations on compressive sensing, advanced matrix factorization and calibration issues on Linkedin.
No comments:
Post a Comment