optical music recognition
Recently Published Documents


TOTAL DOCUMENTS

89
(FIVE YEARS 23)

H-INDEX

11
(FIVE YEARS 1)

2021 ◽  
Vol 10 (4) ◽  
pp. 80-90
Author(s):  
Kyeongmin Oh ◽  
Yoseop Hong ◽  
Geongyeong Baek ◽  
Chanjun Chun

Author(s):  
Carlos de la Fuente ◽  
Jose J. Valero-Mas ◽  
Francisco J. Castellanos ◽  
Jorge Calvo-Zaragoza

AbstractOptical Music Recognition (OMR) and Automatic Music Transcription (AMT) stand for the research fields that aim at obtaining a structured digital representation from sheet music images and acoustic recordings, respectively. While these fields have traditionally evolved independently, the fact that both tasks may share the same output representation poses the question of whether they could be combined in a synergistic manner to exploit the individual transcription advantages depicted by each modality. To evaluate this hypothesis, this paper presents a multimodal framework that combines the predictions from two neural end-to-end OMR and AMT systems by considering a local alignment approach. We assess several experimental scenarios with monophonic music pieces to evaluate our approach under different conditions of the individual transcription systems. In general, the multimodal framework clearly outperforms the single recognition modalities, attaining a relative improvement close to $$40\%$$ 40 % in the best case. Our initial premise is, therefore, validated, thus opening avenues for further research in multimodal OMR-AMT transcription.


2021 ◽  
Author(s):  
Aozhi Liu ◽  
Lipei Zhang ◽  
Yaqi Mei ◽  
Baoqiang Han ◽  
Zifeng Cai ◽  
...  

2021 ◽  
Vol 31 (3) ◽  
pp. 199-205
Author(s):  
Hobin Kim ◽  
Minhoon Lee ◽  
Mikyeong Moon ◽  
Seung-Min Park

2021 ◽  
Vol 11 (8) ◽  
pp. 3621
Author(s):  
María Alfaro-Contreras ◽  
Jose J. Valero-Mas

State-of-the-art Optical Music Recognition (OMR) techniques follow an end-to-end or holistic approach, i.e., a sole stage for completely processing a single-staff section image and for retrieving the symbols that appear therein. Such recognition systems are characterized by not requiring an exact alignment between each staff and their corresponding labels, hence facilitating the creation and retrieval of labeled corpora. Most commonly, these approaches consider an agnostic music representation, which characterizes music symbols by their shape and height (vertical position in the staff). However, this double nature is ignored since, in the learning process, these two features are treated as a single symbol. This work aims to exploit this trademark that differentiates music notation from other similar domains, such as text, by introducing a novel end-to-end approach to solve the OMR task at a staff-line level. We consider two Convolutional Recurrent Neural Network (CRNN) schemes trained to simultaneously extract the shape and height information and to propose different policies for eventually merging them at the actual neural level. The results obtained for two corpora of monophonic early music manuscripts prove that our proposal significantly decreases the recognition error in figures ranging between 14.4% and 25.6% in the best-case scenarios when compared to the baseline considered.


2021 ◽  
pp. 661-675
Author(s):  
Antonio Ríos-Vila ◽  
David Rizo ◽  
Jorge Calvo-Zaragoza

2021 ◽  
pp. 708-722
Author(s):  
Enrique Mas-Candela ◽  
María Alfaro-Contreras ◽  
Jorge Calvo-Zaragoza

2021 ◽  
pp. 59-73
Author(s):  
Juan C. López-Gutiérrez ◽  
Jose J. Valero-Mas ◽  
Francisco J. Castellanos ◽  
Jorge Calvo-Zaragoza

Sign in / Sign up

Export Citation Format

Share Document