Abstract
Musical onset detection is one of the most elementary tasks in music analysis, but still only solved imperfectly for polyphonic music
signals. Interpreted as a computer vision problem in spectrograms,
Convolutional Neural Networks (CNNs) seem to be an ideal fit. On
a dataset of about 100 minutes of music with 26k annotated onsets,
we show that CNNs outperform the previous state-of-the-art while
requiring less manual preprocessing. Investigating their inner workings, we find two key advantages over hand-designed methods: Using separate detectors for percussive and harmonic onsets, and combining results from many minor variations of the same scheme. The
results suggest that even for well-understood signal processing tasks,
machine learning can be superior to knowledge engineering.
| Originalsprache | Englisch |
|---|---|
| Titel | Proceedings of the 39th IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP) |
| Seitenumfang | 6 |
| Publikationsstatus | Veröffentlicht - Mai 2014 |
Wissenschaftszweige
- 202002 Audiovisuelle Medien
- 102 Informatik
- 102001 Artificial Intelligence
- 102003 Bildverarbeitung
- 102015 Informationssysteme
JKU-Schwerpunkte
- Computation in Informatics and Mathematics
- TNF Allgemein
Dieses zitieren
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver