As part of my onboarding at Dolby, I had the pleasure to be working in San Francisco. In order to share my recent experiences with my colleagues, I have been updating these slides and I presented some of my recent work at Dolby and Adobe headquarters.
During the last summer, I have been a research intern at Telefónica Research (Barcelona). The article “Training neural audio classifiers with few data” is the outcome of this short (but intense!) collaboration with Joan Serrà, where we explored how to train deep learning models with just 1, 2 or 10 audios per class. Check it out on arXiv, and reproduce our results running our code! These slides are the extended version of what I will be presenting next week in ICASSP! See you in Brighton 🙂
A few weeks ago Olga Slizovskaya and I were invited to give a talk to the Centre for Digital Music (C4DM) @ Queen Mary Universtity of London – one of the most renowned music technology research institutions in Europe, and possibly in the world. It’s been an honor, and a pleasure to share our thoughts (and some beers) with you!
The talk was centered in our recent work on music audio tagging, which is available on arXiv, where we study how non-trained (randomly weighted) convolutional neural networks perform as feature extractors for (music) audio classification tasks.