Automatic analysis of the acoustic environment of a preterm infant in a neonatal intensive care unit
Seminar Room, D5-007
Wed July 20, 2016
At the TALP Research Center from Universitat Politècnica de Catalunya (UPC, Barcelona) we are looking for a PhD student to work in Deep Learning in Speech Technologies. The PhD position is a FPI-MINECO contract in the context of the Project: DeepVoice: Deep Learning Technologies for Speech and Audio Processing. This project proposes the development of new deep learning methods for speech and audio processing including machine translation, exploring new applications and continuing the initial work of the research team and the international community.
The PhD position is funded including salary, tuition, travel and equipment for students of any nationality. We offer a 4-year contract (expected start date is third trimester 2016). The gross salary of a PhD student, in accordance with the agreement for Spanish official PhD contracts is 16,422 €/year. Additionally, the TALP research group provides excellent opportunities for professional and personal development, research stays at the most competitive international research labs, participation in top international conferences.
The applicants must have a MS degree in Electrical Engineering, Mathematics, Computer Science, or Computational Linguistic, or equivalent (300 ECTS including the BS degree). We encourage candidates with strong computer science foundation, natural language processing and/or deep learning experience.
Speaker 1: Joan Bruna (UC Berkeley) Talk1: Convolutional Neural Networks against the Curse of Dimensionality
Speaker 2: Oriol Vinyals (Google DeepMind) Talk 2: Recurrent Neural Nets and Differentiable Memory Mechanism by Oriol Vinyals
Joan Bruna Bio: Joan graduated cum-laude from Universitat Politècnica de Catalunya in both Mathematics and Telecommunications Engineering, before graduating in Applied Mathematics from ENS Cachan (France). He then became a Sr. Research Engineer in an Image Processing startup, developing real-time video processing algorithms. In 2013 he obtained his PhD in Applied Mathematics at École Polytechnique (France). After a postdoctoral stay at the Computer Science department of Courant Institute, NYU, he became a Postdoctoral fellow at Facebook AI Research. Since Jan 2015 he is an Assistant Professor at UC Berkeley, Statistics Department. His research interests include invariant signal representations, deep learning, stochastic processes, and its applications to computer vision.
Talk 1 abstract: Convolutional Neural Networks are a powerful class of non-linear representations that have shown through numerous supervised learning tasks their ability to extract rich information from images, speech and text, with excellent statistical generalization. These are examples of truly high-dimensional signals, in which classical statistical models suffer from the curse of dimensionality, referring to their inability to generalize well unless provided with exponentially large amounts of training data. In this talk we will start by studying statistical models defined from wavelet scattering networks, a class of CNNs where the convolutional filter banks are given by complex, multi-resolution wavelet families. The reasons for such success lie on their ability to preserve discriminative information while being stable with respect to high-dimensional deformations, providing a framework that partially extends to trained CNNs. We will give conditions under which signals can be recovered from their scattering coefficients, and will discuss a family of Gibbs processes defined by CNN sufficient statistics, from which one can sample image and auditory textures. Although the scattering recovery is non-convex and corresponds to a generalized phase recovery problem, gradient descent algorithms show good empirical performance and enjoy weak convergence properties. We will discuss connections with non-linear compressed sensing, applications to texture synthesis, inverse problems such as super-resolution, as well as an application to sentence modeling, where convolutions are generalized using associative trees to generate robust sentence representations.
Oriol Vinyals Bio: Oriol is a Research Scientist at Google DeepMind, working on Deep Learning. Oriol holds a Ph.D. in EECS from University of California, Berkeley, a Masters degree from University of California, San Diego, and a double degree in Mathematics and Telecommunication Engineering from UPC, Barcelona. He is a recipient of the 2011 Microsoft Research PhD Fellowship. He was an early adopter of the new deep learning wave at Berkeley, and in his thesis he focused on non-convex optimization and recurrent neural networks. At Google Brain and Google DeepMind he continues working on his areas of interest, which include artificial intelligence, with particular emphasis on machine learning, language, and vision.
Talk 2 abstract: This past year, RNNs have seen a lot of attention as powerful models that are able to decode sequences from signals. The key component of such methods are the use of a recurrent neural network architecture that is trained end-to-end to optimize the probability of the output sequence given those signals. In this talk, I’ll define the architecture and review some recent successes in my group on machine translation, image understanding, and beyond. On the second part of the talk, I will introduce a new paradigm — differentiable memory — that has enabled learning programs (e.g., planar Traveling Salesman Problem) using training instances via a powerful extension of RNNs with memory. This effectively turns a machine learning model into a “differentiable computer”. I will conclude the talk giving a few examples (e.g., AlphaGo) on how these recent Machine Learning advances have been the main catalyst in Artificial Intelligence in the past years.
- 27/04/2016 - Hermann Ney - Human Language Technology and Machine Learning
- 10/02/2016 - Ali M. Naderi PhD dissertation
- 13/07/2015 - Xavier Lluís PhD dissertation
- TALP organizes 2015 RTTH Summer School in Barcelona
- 27th European Summer School in Logic, Language, and Information
- TweetMT 2015 -- Tweet Translation Workshop at SEPLN 2015
- 13/03/2015 - Tatyana Polyakova PhD dissertation
- 05/02/2015 SKATER 2nd Workshop at Barcelona
- Workshop on Adaptive Natural Language Processing at IJCAI 2015
- The Massive Open On-line Course on Machine Translation that will start on October 23
- 22/09/2014 - Prof. David Zhang - Biometrics Computing
- 16/09/2014 - Eva Martínez - Document-Level Machine Translation as a Re-translation Process
Job Offers and Research Positions
Scientific awards Best: Publications, Thesis, Demos, ...
Framework to promote research related with Natural Language Processing or Speech Processing.
TALP PhD dissertation
Information about Courses, Conferences, workshops, talks, ...
Call For Papers
Relevant Conference, workshops or special issues call for papers announcement
- Relevant Publications
Professors, students and colaborators visiting TALP.
New members incorporation and former members carreer.
TALP research and innovation projects
Selected articles and reports about the TALP Research Center taken from the international press