Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
149 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sounding Out Reconstruction Error-Based Evaluation of Generative Models of Expressive Performance (2401.00471v1)

Published 31 Dec 2023 in cs.SD and eess.AS

Abstract: Generative models of expressive piano performance are usually assessed by comparing their predictions to a reference human performance. A generative algorithm is taken to be better than competing ones if it produces performances that are closer to a human reference performance. However, expert human performers can (and do) interpret music in different ways, making for different possible references, and quantitative closeness is not necessarily aligned with perceptual similarity, raising concerns about the validity of this evaluation approach. In this work, we present a number of experiments that shed light on this problem. Using precisely measured high-quality performances of classical piano music, we carry out a listening test indicating that listeners can sometimes perceive subtle performance difference that go unnoticed under quantitative evaluation. We further present tests that indicate that such evaluation frameworks show a lot of variability in reliability and validity across different reference performances and pieces. We discuss these results and their implications for quantitative evaluation, and hope to foster a critical appreciation of the uncertainties involved in quantitative assessments of such performances within the wider music information retrieval (MIR) community.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (29)
  1. Michel Bernays and Caroline Traube. 2010. Expression of piano timbre: gestural control, perception and verbalization. In Proceedings of CIM09: The 5th Conference on Interdisciplinary Musicology.
  2. Michel Bernays and Caroline Traube. 2013. Expressive production of piano timbre: touch and playing techniques for timbre control in piano performance. In Proceedings of the 10th Sound and Music Computing Conference (SMC2013). KTH Royal Institute of Technology Stockholm, Sweden, 341–346.
  3. Michel Bernays and Caroline Traube. 2014. Investigating pianists’ individuality in the performance of five timbral nuances through patterns of articulation, touch, dynamics, and pedaling. Frontiers in Psychology 5 (2014), 157.
  4. Roberto Bresin and Anders Friberg. 2013. Evaluation of Computer Systems for Expressive Music Performance. In Guide to Computing for Expressive Music Performance, Alexis Kirke and Eduardo R Miranda (Eds.). Springer-Verlag, London, UK, 181–203.
  5. Computational Models of Expressive Music Performance: A Comprehensive and Critical Review. Frontiers in Digital Humanities 5 (2018), 25. https://doi.org/10.3389/fdigh.2018.00025
  6. On the characterization of expressive performance in classical music: First results of the con espressione game. (2020).
  7. An Evaluation of Linear and Non-linear Models of Expressive Dynamics in Classical Piano and Symphonic Music. Machine Learning 106, 6 (2017), 887–909.
  8. Elaine Chew. 2012. About Time: Strategies of Performance Revealed in Graphs. Visions of Research in Music Education 20 (2012).
  9. Perceptual Smoothness of Tempo in Expressively Performed Music. Music Perception 23, 3 (2006), 195–214.
  10. Expressive Performance Rendering with Probabilistic Models. In Guide to Computing for Expressive Music Performance, Alexis Kirke and Eduardo R Miranda (Eds.). Springer, London, UK, 75–98.
  11. The match file format: Encoding Alignments between Scores and Performances. In Proceedings of the Music Encoding Conference (MEC). Halifax, Canada.
  12. Werner Goebl. 1999. The Vienna 4x22 Piano Corpus. https://doi.org/10.21939/4X22
  13. Werner Goebl. 2001. Melody lead in piano performance: Expressive device or artifact? The Journal of the Acoustical Society of America 110, 1 (2001), 563–572. https://doi.org/10.1121/1.1376133 arXiv:https://doi.org/10.1121/1.1376133
  14. Werner Goebl and Caroline Palmer. 2009. Synchronization of timing and motion among performing musicians. Music Perception 26, 5 (2009), 427–438.
  15. Graham Grindlay and David Helmbold. 2006. Modeling, analyzing, and synthesizing expressive piano performance with graphical models. Machine Learning 65, 2-3 (June 2006), 361–387.
  16. Analysis of Objective Descriptors for Music Performance Assessment. In Proceedings of the International Conference on Music Perception and Cognition (ICMPC) (2018-01-01). Montreal, Canada. http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2018/06/Gururani-et-al.-2018-Analysis-of-Objective-Descriptors-for-Music-Perfor.pdf
  17. VirtuosoNet: A Hierarchical RNN-based System for Modeling Expressive Piano Performance. In Proceedings of the 20th International Society for Music Information Retrieval Conference (ISMIR 2019). Delft, The Netherlands, 908–9015.
  18. Statistical Approach to Automatic Expressive Rendition of Polyphonic Piano Music. In Guide to Computing for Expressive Music Performance, Alexis Kirke and Eduardo R Miranda (Eds.). Springer, London, UK, 145–179.
  19. Music Performance Analysis: A Survey. In Proceedings of the International Society for Music Information Retrieval Conference (ISMIR) (2019-01-01). Delft. http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2019/06/Lerch-et-al.-2019-Music-Performance-Analysis-A-Survey.pdf
  20. The clustering of expressive timing within a phrase in classical piano performances by Gaussian Mixture Models. In Proceedings of the 11th International Symposium on Computer Music Multidisciplinary Research (CMMR 2015). Plymouth, UK, 322–345.
  21. Performance Error Detection and Post-Processing for Fast and Accurate Symbolic Music Alignment. In Proceedings of the 18th International Society for Music Information Retrieval Conference (ISMIR 2018). Suzhou, China, 347–353.
  22. Assessment of Student Music Performances Using Deep Neural Networks. Applied Sciences 8, 4 (2018), 507. https://doi.org/10.3390/app8040507
  23. Automatic Note-Level Score-to-Performance Alignments in the ASAP Dataset. Transactions of the International Society for Music Information Retrieval (TISMIR) (2023). https://doi.org/10.5334/tismir.149
  24. Friederich Platz and Reinhard Kopiez. 2012. When the Eye Listens: A Meta-analysis of How Audio-visual Presentation Enhances the Appreciation of Music Performance. Music Perception 30, 1 (2012), 71–83.
  25. B. Repp. 1997. The aesthetic quality of a quantitatively average music performance: Two preliminary experiments. Music Perception 14, 4 (1997), 419–444.
  26. Gaussian process regression for rendering music performance. In Proceedings of the 10th International Conference on Music Perception and Cognition (ICMPC 10). Sapporo, Japan.
  27. A Tobudic and G Widmer. 2003. Playing Mozart phrase by phrase. In Proceedings of the Fifth International Conference on Case-Based Reasoning (ICCBR-03). Springer-Verlag, 552–566.
  28. Examining Rater Precision in Music Performance Assessment: An Analysis of Rating Scale Structure Using the Multifaceted Rasch Partial Credit Model. Music Perception 33, 5 (June 2016), 662–678.
  29. ATEPP: A Dataset of Automatically Transcribed Expressive Piano Performance. In Proceedings of the 23rd International Society for Music Information Retrieval Conference (ISMIR).
Citations (1)

Summary

We haven't generated a summary for this paper yet.