music recognition
Recently Published Documents


TOTAL DOCUMENTS

163
(FIVE YEARS 40)

H-INDEX

17
(FIVE YEARS 4)

2021 ◽  
Vol 10 (4) ◽  
pp. 80-90
Author(s):  
Kyeongmin Oh ◽  
Yoseop Hong ◽  
Geongyeong Baek ◽  
Chanjun Chun

2021 ◽  
Vol 11 (24) ◽  
pp. 11663
Author(s):  
Eugenio Brusa ◽  
Cristiana Delprete ◽  
Luigi Gianpio Di Maggio

Today’s deep learning strategies require ever-increasing computational efforts and demand for very large amounts of labelled data. Providing such expensive resources for machine diagnosis is highly challenging. Transfer learning recently emerged as a valuable approach to address these issues. Thus, the knowledge learned by deep architectures in different scenarios can be reused for the purpose of machine diagnosis, minimizing data collecting efforts. Existing research provides evidence that networks pre-trained for image recognition can classify machine vibrations in the time-frequency domain by means of transfer learning. So far, however, there has been little discussion about the potentials included in networks pre-trained for sound recognition, which are inherently suited for time-frequency tasks. This work argues that deep architectures trained for music recognition and sound detection can perform machine diagnosis. The YAMNet convolutional network was designed to serve extremely efficient mobile applications for sound detection, and it was originally trained on millions of data extracted from YouTube clips. That framework is employed to detect bearing faults for the CWRU dataset. It is shown that transferring knowledge from sound and music recognition to bearing fault detection is successful. The maximum accuracy is achieved using a few hundred data for fine-tuning the fault diagnosis model.


Author(s):  
Carlos de la Fuente ◽  
Jose J. Valero-Mas ◽  
Francisco J. Castellanos ◽  
Jorge Calvo-Zaragoza

AbstractOptical Music Recognition (OMR) and Automatic Music Transcription (AMT) stand for the research fields that aim at obtaining a structured digital representation from sheet music images and acoustic recordings, respectively. While these fields have traditionally evolved independently, the fact that both tasks may share the same output representation poses the question of whether they could be combined in a synergistic manner to exploit the individual transcription advantages depicted by each modality. To evaluate this hypothesis, this paper presents a multimodal framework that combines the predictions from two neural end-to-end OMR and AMT systems by considering a local alignment approach. We assess several experimental scenarios with monophonic music pieces to evaluate our approach under different conditions of the individual transcription systems. In general, the multimodal framework clearly outperforms the single recognition modalities, attaining a relative improvement close to $$40\%$$ 40 % in the best case. Our initial premise is, therefore, validated, thus opening avenues for further research in multimodal OMR-AMT transcription.


2021 ◽  
Author(s):  
Aozhi Liu ◽  
Lipei Zhang ◽  
Yaqi Mei ◽  
Baoqiang Han ◽  
Zifeng Cai ◽  
...  

2021 ◽  
Vol 31 (3) ◽  
pp. 199-205
Author(s):  
Hobin Kim ◽  
Minhoon Lee ◽  
Mikyeong Moon ◽  
Seung-Min Park

2021 ◽  
Vol 11 (9) ◽  
pp. 3890
Author(s):  
Antonio Ríos-Vila ◽  
Miquel Esplà-Gomis ◽  
David Rizo ◽  
Pedro J. Ponce de León ◽  
José M. Iñesta

Optical music recognition is a research field whose efforts have been mainly focused, due to the difficulties involved in its processes, on document and image recognition. However, there is a final step after the recognition phase that has not been properly addressed or discussed, and which is relevant to obtaining a standard digital score from the recognition process: the step of encoding data into a standard file format. In this paper, we address this task by proposing and evaluating the feasibility of using machine translation techniques, using statistical approaches and neural systems, to automatically convert the results of graphical encoding recognition into a standard semantic format, which can be exported as a digital score. We also discuss the implications, challenges and details to be taken into account when applying machine translation techniques to music languages, which are very different from natural human languages. This needs to be addressed prior to performing experiments and has not been reported in previous works. We also describe and detail experimental results, and conclude that applying machine translation techniques is a suitable solution for this task, as they have proven to obtain robust results.


2021 ◽  
Vol 11 (8) ◽  
pp. 3621
Author(s):  
María Alfaro-Contreras ◽  
Jose J. Valero-Mas

State-of-the-art Optical Music Recognition (OMR) techniques follow an end-to-end or holistic approach, i.e., a sole stage for completely processing a single-staff section image and for retrieving the symbols that appear therein. Such recognition systems are characterized by not requiring an exact alignment between each staff and their corresponding labels, hence facilitating the creation and retrieval of labeled corpora. Most commonly, these approaches consider an agnostic music representation, which characterizes music symbols by their shape and height (vertical position in the staff). However, this double nature is ignored since, in the learning process, these two features are treated as a single symbol. This work aims to exploit this trademark that differentiates music notation from other similar domains, such as text, by introducing a novel end-to-end approach to solve the OMR task at a staff-line level. We consider two Convolutional Recurrent Neural Network (CRNN) schemes trained to simultaneously extract the shape and height information and to propose different policies for eventually merging them at the actual neural level. The results obtained for two corpora of monophonic early music manuscripts prove that our proposal significantly decreases the recognition error in figures ranging between 14.4% and 25.6% in the best-case scenarios when compared to the baseline considered.


Sign in / Sign up

Export Citation Format

Share Document