- The paper introduces a regression-based method to precisely model piano note onsets, offsets, and pedal events.
- It leverages convolutional layers and bidirectional GRUs on log mel spectrograms to outperform traditional frame-based techniques.
- The system achieves a 96.72% onset F1 score and 91.86% pedal F1 score on the MAESTRO dataset, enabling robust high-resolution transcription.
High-Resolution Piano Transcription with Pedals by Regressing Onset and Offset Times
This paper addresses the task of Automatic Music Transcription (AMT) for piano recordings, focusing on improving transcription resolution and robustness. AMT involves converting audio recordings into symbolic representations like MIDI files. The authors introduce a novel method that enhances the resolution of piano transcription by regressing precise onset and offset times for notes and pedal events, diverging from traditional frame-wise methods constrained by frame hop sizes.
Methodology
The proposed system leverages neural networks to perform regression-based modeling of onsets and offsets, offering enhancements over binary classification models. Traditional methods often suffer from limited temporal resolution and are sensitive to misaligned labels. By regressing the time differences between the center of a frame and the nearest onset or offset, this approach maintains precise onset and offset information.
The architecture comprises convolutional layers followed by bidirectional GRUs to handle log mel spectrogram inputs. This allows for capturing both temporal and frequency-domain information effectively. The system employs loss functions based on binary cross-entropy over regression targets rather than classifying presence/absence in each frame.
Numerical Results
The system exhibits strong performance on the MAESTRO dataset, achieving an onset F1 score of 96.72%, surpassing previous onsets and frames methods. The paper highlights robustness against label misalignments, suggesting that the regression-based approach can handle variations in label alignment more effectively than frame-based methods.
Additionally, the inclusion of a novel sustain pedal detection mechanism within the framework is noteworthy. The pedal transcription demonstrates an F1 score of 91.86%, providing a benchmark for future evaluations.
Implications and Future Work
This regression-based approach opens new avenues for high-resolution music transcription, presenting implications for music information retrieval, performance analysis, and intelligent music editing. The architecture is adaptable for future developments in multi-instrument transcription and can potentially improve real-time transcription systems.
However, limitations such as dependency on audio quality and the need for system modifications for real-time applications are acknowledged. Future work may focus on expanding this framework to accommodate diverse musical contexts and instruments.
The release of the system’s source code encourages further validation and adaptation by the research community, fostering advancements in digital music processing.