Abstract
In this paper we present preliminary work examining the relationship
between the formation of expectations and the realization of
musical performances, paying particular attention to expressive tempo
and dynamics. To compute features that re
ect what a listener is expecting
to hear, we employ a computational model of auditory expectation
called the Information Dynamics of Music model (IDyOM). We then
explore how well these expectancy features { when combined with score
descriptors using the Basis-Function modeling approach { can predict
expressive tempo and dynamics in a dataset of Mozart piano sonata
performances. Our results suggest that using expectancy features significantly
improves the predictions for tempo.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 10th International Workshop on Machine Learning and Music (MML 2017) |
| Number of pages | 6 |
| Publication status | Published - Oct 2017 |
Fields of science
- 202002 Audiovisual media
- 102 Computer Sciences
- 102001 Artificial intelligence
- 102003 Image processing
- 102015 Information systems
JKU Focus areas
- Computation in Informatics and Mathematics
- Engineering and Natural Sciences (in general)
Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver