| Paper ID | AUD-10.5 | ||
| Paper Title | JOINT MULTI-PITCH DETECTION AND SCORE TRANSCRIPTION FOR POLYPHONIC PIANO MUSIC | ||
| Authors | Lele Liu, Veronica Morfi, Emmanouil Benetos, Queen Mary University of London, United Kingdom | ||
| Session | AUD-10: Music Information Retrieval and Music Language Processing 2: Singing Voice | ||
| Location | Gather.Town | ||
| Session Time: | Wednesday, 09 June, 14:00 - 14:45 | ||
| Presentation Time: | Wednesday, 09 June, 14:00 - 14:45 | ||
| Presentation | Poster | ||
| Topic | Audio and Acoustic Signal Processing: [AUD-MIR] Music Information Retrieval and Music Language Processing | ||
| IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
| Abstract | Research on automatic music transcription has largely focused on multi-pitch detection; there is limited discussion on how to obtain a machine- or human-readable score transcription. In this paper, we propose a method for joint multi-pitch detection and score transcription for polyphonic piano music. The outputs of our system include both a piano-roll representation (a descriptive transcription) and a symbolic musical notation (a prescriptive transcription). Unlike traditional methods that further convert MIDI transcriptions into musical scores, we use a multitask model combined with a Convolutional Recurrent Neural Network and Sequence-to-sequence models with attention mechanisms. We propose a Reshaped score representation that outperforms a LilyPond representation in terms of both prediction accuracy and time/memory resources, and compare different input audio spectrograms. We also create a new synthesized dataset for score transcription research. Experimental results show that the joint model outperforms a single-task model in score transcription. | ||