20/07/2016 - Anna Raboshchuck PhD Dissertation

  Automatic analysis of the acoustic environment of a preterm infant in a neonatal intensive care unit

Anna Raboshchuck

Seminar Room, D5-007

Wed July 20, 2016


  Climent Nadeu

PhD scholarship on Deep Learning applied to Speech Technologies

At the TALP Research Center from Universitat Politècnica de Catalunya (UPC, Barcelona) we are looking for a PhD student to work in Deep Learning in Speech Technologies. The PhD position is a FPI-MINECO contract in the context of the Project: DeepVoice: Deep Learning Technologies for Speech and Audio Processing. This project proposes the development of new deep learning methods for speech and audio processing including machine translation, exploring new applications and continuing the initial work of the research team and the international community.


The PhD position is funded including salary, tuition, travel and equipment for students of any nationality. We offer a 4-year contract (expected start date is third trimester 2016). The gross salary of a PhD student, in accordance with the agreement for Spanish official PhD contracts is 16,422 €/year. Additionally, the TALP research group provides excellent opportunities for professional and personal development, research stays at the most competitive international research labs, participation in top international conferences.


The applicants must have a MS degree in Electrical Engineering, Mathematics, Computer Science, or Computational Linguistic, or equivalent (300 ECTS including the BS degree).  We encourage candidates with strong computer science foundation, natural language processing and/or deep learning experience.


You should apply by sending a brief CV (2-pages) and an application letter to This email address is being protected from spambots. You need JavaScript enabled to view it. , by 15th June 2016. Subject of email should be **PhD Application**

Joan Bruna & Oriol Vinyals Seminar on Deep Learning

Date:  Monday, 23 May, 2016 - 11:00
Venue: Aula Master, Campus Nord, UPC

Speaker 1: Joan Bruna (UC Berkeley) Talk1: Convolutional Neural Networks against the Curse of Dimensionality

Speaker 2: Oriol Vinyals (Google DeepMind) Talk 2: Recurrent Neural Nets and Differentiable Memory Mechanism by Oriol Vinyals

Joan Bruna Bio: Joan graduated cum-laude from Universitat Politècnica de Catalunya in both Mathematics and Telecommunications Engineering, before graduating in Applied Mathematics from ENS Cachan (France). He then became a Sr. Research Engineer in an Image Processing startup, developing real-time video processing algorithms. In 2013 he obtained his PhD in Applied Mathematics at École Polytechnique (France). After a postdoctoral stay at the Computer Science department of Courant Institute, NYU, he became a Postdoctoral fellow at Facebook AI Research. Since Jan 2015 he is an Assistant Professor at UC Berkeley, Statistics Department. His research interests include invariant signal representations, deep learning, stochastic processes, and its applications to computer vision.

Talk 1 abstract: Convolutional Neural Networks are a powerful class of non-linear representations that have shown through numerous supervised learning tasks their ability to extract rich information from images, speech and text, with excellent statistical generalization. These are examples of truly high-dimensional signals, in which classical statistical models suffer from the curse of dimensionality, referring to their inability to generalize well unless provided with exponentially large amounts of training data. In this talk we will start by studying statistical models defined from wavelet scattering networks, a class of CNNs where the convolutional filter banks are given by complex, multi-resolution wavelet families. The reasons for such success lie on their ability to preserve discriminative information while being stable with respect to high-dimensional deformations, providing a framework that partially extends to trained CNNs. We will give conditions under which signals can be recovered from their scattering coefficients, and will discuss a family of Gibbs processes defined by  CNN sufficient statistics, from which one can sample image and auditory textures. Although the scattering recovery is non-convex and corresponds to a generalized phase recovery problem, gradient descent algorithms show good empirical performance and enjoy weak convergence properties. We will discuss connections with non-linear compressed sensing, applications to texture synthesis, inverse problems such as super-resolution, as well as an application to sentence modeling, where convolutions are generalized using associative trees to generate robust sentence representations.

Oriol Vinyals Bio: Oriol is a Research Scientist at Google DeepMind, working on Deep Learning. Oriol holds a Ph.D. in EECS from University of California, Berkeley, a Masters degree from University of California, San Diego, and a double degree in Mathematics and Telecommunication Engineering from UPC, Barcelona. He is a recipient of the 2011 Microsoft Research PhD Fellowship. He was an early adopter of the new deep learning wave at Berkeley, and in his thesis he focused on non-convex optimization and recurrent neural networks. At Google Brain and Google DeepMind he continues working on his areas of interest, which include artificial intelligence, with particular emphasis on machine learning, language, and vision.

Talk 2 abstract: This past year, RNNs have seen a lot of attention as powerful models that are able to decode sequences from signals. The key component of such methods are the use of a recurrent neural network architecture that is trained end-to-end to optimize the probability of the output sequence given those signals. In this talk, I’ll define the architecture and review some recent successes in my group on machine translation, image understanding, and beyond. On the second part of the talk, I will introduce a new paradigm — differentiable memory —  that has enabled learning programs (e.g., planar Traveling Salesman Problem) using training instances via a powerful extension of RNNs with memory. This effectively turns a machine learning model into a “differentiable computer”. I will conclude the talk giving a few examples (e.g., AlphaGo) on how these recent Machine Learning advances have been the main catalyst in Artificial Intelligence in the past years.

Organisers of the event:

UPC, TelecomBCN, FIB, CFIS and BSC


Seminar on Deep Learning


  • Join us

    Job Offers and Research Positions

  • Awards

    Scientific awards Best: Publications, Thesis, Demos, ...

  • Seminars

    Framework to promote research related with Natural Language Processing or Speech Processing.

  • Thesis

    TALP PhD dissertation

  • Events

    Information about Courses, Conferences, workshops, talks, ...

  • Call For Papers

    Relevant Conference, workshops or special issues call for papers announcement

  • Relevant Publications
  • Members

    Professors, students and colaborators visiting TALP.

    New members incorporation and former members carreer.

  • Projects

    TALP research and innovation projects

  • Press

    Selected articles and reports about the TALP Research Center taken from the international press

  • Formation

Additional information