Paper ID | MLSP-11.3 | ||
Paper Title | UNSUPERVISED DISCRIMINATIVE LEARNING OF SOUNDS FOR AUDIO EVENT CLASSIFICATION | ||
Authors | Sascha Hornauer, Ke Li, Stella Yu, University of California, Berkeley, United States; Shabnam Ghaffarzadegan, Liu Ren, Robert Bosch LLC, United States | ||
Session | MLSP-11: Self-supervised Learning for Speech Processing | ||
Location | Gather.Town | ||
Session Time: | Tuesday, 08 June, 16:30 - 17:15 | ||
Presentation Time: | Tuesday, 08 June, 16:30 - 17:15 | ||
Presentation | Poster | ||
Topic | Machine Learning for Signal Processing: [MLR-SSUP] Self-supervised and semi-supervised learning | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | Recent progress in network-based audio event classification has shown the benefit of pre-training models on visual data such as ImageNet. While this process allows knowledge transfer across different domains, training a model on large-scale visual datasets is time consuming. On several audio event classification benchmarks, we show a fast and effective alternative that pre-trains the model unsupervised, only on audio data and yet delivers on-par performance with ImageNet pre-training. Furthermore, we show that our discriminative audio learning can be used to transfer knowledge across audio datasets and optionally include ImageNet pre-training. |