Stable Audio allows you creating custom-length audio just by describing it. It is powered by a generative audio model based on diffusion. You can generate and download audio in 44.1 kHz stereo. You also have a nice interface, no need to be a hacker! And the audio you create can be used in your commercial projects. I’ve been experimenting with it during the last weeks, and here some ideas on how to use it!
Now that I’m a doctor, I can be called to be part of a jury for helping to evaluate PhD students. Today was my first experience doing that, and it was double special because the other jury member was Axel Roebel — with whom I started my research career.
Thanks to Javier Nistal, Gaël Richard and Steffan Lattner for inviting me!
To facilitate my online teaching activities, I collected on this website the educational material I have been preparing throughout the years.
I thought it could be useful to share it! It is ready to use and includes a quiz and a lab. For now, I cover music/audio classification and the basics of deep learning for music/audio.
Although I’m now a researcher at Dolby Laboratories, I’m still collaborating with some universities in Barcelona — where I’ll keep teaching deep learning for music and audio. In this context, and given the importance of the gradient vanishing/explode problem in deep neural networks, this week I’ll be teaching recurrent neural networks to the Master in Sound and Music Computing students of the Universitat Pompeu Fabra.
As part of my onboarding at Dolby, I had the pleasure to be working in San Francisco. In order to share my recent experiences with my colleagues, I have been updating these slides and I presented some of my recent work at Dolby and Adobe headquarters.
I hope this update makes this tutorial-like presentation more understandable to everyone!