[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[AUDITORY] Announcing OpenL3, a competitive and open deep audio embedding!



(Apologies for cross-posting)


Hello everyone!


We're excited to announce the release of OpenL3, an open-source deep audio embedding based on the self-supervised L3-Net. OpenL3 is an improved version of L3-Net, and outperforms VGGish and SoundNet (and the original L3-Net) on several sound recognition tasks. Most importantly, OpenL3 is open source and readily available for everyone to use: if you have TensorFlow installed just run pip install openl3 and you're good to go!


Full details are provided in our paper, which will be presented at ICASSP 2019:


Look, Listen and Learn More: Design Choices for Deep Audio Embeddings

J. Cramer, H.-H. Wu, J. Salamon, and J. P. Bello.

IEEE Int. Conf. on Acoustics, Speech and Signal Proc. (ICASSP), pp 3852-3856, Brighton, UK, May 2019.


If you're attending ICASSP 2019 and would like to discuss OpenL3 with us please stop by our poster on Friday, May 17 between 13:30-15:30 (session MLSP-P17: Deep Learning V, Poster Area G, paper 2149).


We look forward to seeing what the community does with OpenL3!


Cheers,

Jason Cramer, Ho-Hsiang Wu, Justin Salamon and Juan Pablo Bello.