Abstract

Optical Music Recognition (OMR) aims to recognize andunderstand written music scores. With the help of Deep Learning, researchers were able to significantly improve the state-of-the-art in this research area. However, Deep Learning requires a substantial amount of annotated data for supervised training. Various datasets have been collected in the past, but without a common standard that defines data formats and terminology, combining them is a challenging task. In this paper we present our approach towards unifying multiple datasets into the largest currently available body of over 90000 musical symbols that belong to 79 classes, containing both handwritten and printed music symbols. A universal music symbol classifier, trained on such a dataset using Deep Learning, can achieve an accuracy that exceeds 98%.

Reference

Pacha, A., & Eidenberger, H. (2017). Towards a Universal Music Symbol Classifier. In 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR). International Workshop on Graphics Recognition, New York, Non-EU. https://doi.org/10.1109/icdar.2017.265