Exploiting Parallel Audio Recordings to Enforce Device Invariance in CNN-based Acoustic Scene Classification

Research output: Chapter in Book/Report/Conference proceedingConference proceedingspeer-review

Abstract

Distribution mismatches between the data seen at training and at application time remain a major challenge in all application areas of machine learning. We study this problem in the context of ma-chine listening (Task 1b of the DCASE 2019 Challenge). We pro-pose a novel approach to learn domain-invariant classifiers in an end-to-end fashion by enforcing equal hidden layer representations for domain-parallel samples, i.e. time-aligned recordings from different recording devices. No classification labels are needed for our domain adaptation (DA) method, which makes the data collection process cheaper. We show that our method improves the tar-get domain accuracy for both a toy dataset and an urban acoustic scenes dataset. We further compare our method to Maximum Mean Discrepancy-based DA and find it more robust to the choice of DA parameters. Our submission, based on this method, to DCASE 2019Task 1b gave us the 4th place in the team ranking.
Original languageEnglish
Title of host publicationProceedings of the Detection and Classification of Acoustic Scenes and Events 2019 Workshop (DCASE2019)
Number of pages5
Publication statusPublished - 2019

Fields of science

  • 202002 Audiovisual media
  • 102 Computer Sciences
  • 102001 Artificial intelligence
  • 102003 Image processing
  • 102015 Information systems

JKU Focus areas

  • Digital Transformation

Cite this