Musical Instruments Sound Classification using GMM


DOI:
https://doi.org/10.31039/ljss.2021.1.37Keywords:
Musical Instrument Sound Classification, mel frequency cepstral coefficient, Gaussian Mixture ModelAbstract
Classification is the task of assigning objects to one of several predefined categories. In today’s decade classifying the musical signal from large data is a major task; the proposed work classifies the music into their respective classes. In this paper, the sound of the musical instruments classified automatically from the musical signals. Mel frequency cepstral coefficient is used as a feature extractor and the machine learning model namely Gaussian Mixture Model is used for classification. This system tested in ten different classes of musical instrument sound from two different instrument families such as Woodwind and Brass instruments. In this proposed work, the result yields satisfactory accuracy in the classification of musical instruments sound.
References
Dhanalakshmi, P., Palanivel, Sengottayan & Ramalingam, Vivekanandan. (2011). Classification of audio signals using AANN and GMM. Applied Soft Computing. 11. Pp. 716-723, 10.1016/j.asoc.2009.12.033.
Thiruvengatanadhan, R., Music Genre Classification using GMM (2018). International Research Journal of Engineering and Technology (IRJET), Volume: 05 Issue: 10.
Bhalke, D.G., Rao, C.B.R. & Bormane, D.S. Automatic Musical Instrument Classification using Fractional Fourier Transform based- MFCC features and Counter Propagation Neural Network. J Intell Inf Syst 46, 425–446 (2016). https://doi.org/10.1007/s10844-015-0360-9
Prabavathy, S, Rathikarani, V., & Dhanalakshmi, P, Classification of Musical Instruments using SVM and KNN, International Journal of Innovative Technology and Exploring Engineering (IJITEE), ISSN: 2278-3075, Volume-9 Issue-7, May 2020.
Shreevathsa, P. K., Harshith, M., A. R. M. & Ashwini, "Music Instrument Recognition using Machine Learning Algorithms," 2020 International Conference on Computation, Automation and Knowledge Management (ICCAKM), 2020, pp. 161-166, doi: 10.1109/ICCAKM46823.2020.9051514.
Dhanalakshmi, P. (2010). Classification of Audio for Retrieval Applications. [Doctoral Dissertation]. Faculty of Computer Science and Engineering, Annamalai University, Chidambaram, Tamilnadu.
Zokaee, Sara & Faez, Karim. (2012). Human Identification Based on Electrocardiogram and Palmprint. 2. 261-266.
Rafael Iriya & Miguel Arjona Ramírez. Gaussian Mixture Models with Class-Dependent Features for Speech Emotion Recognition. IEEE Workshop on Statistical Signal Processing, pp. 480-483, 2014.
Tang, H., Chu, S. M., Hasegawa-Johnson, M. & Huang, T. S., Partially Supervised Speaker Clustering. IEEE transactions on Pattern Analysis and Machine Intelligence. vol. 34, no. 5, pp. 959-971, 2012.
Krishna, A.G. & Sreenivas, Tv. (2004). Music instrument recognition: from isolated notes to solo phrases. International Conference on Acoustics, Speech, and Signal Processing, ICASSP.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2021 Prabhu Kumar Aurchana, S. Prabavathy

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
You are free to:
Share: copy and redistribute the material in any medium or format. The licensor cannot revoke these freedoms as long as you follow the license terms. Under the following terms: Attribution-NonCommercial-NoDerivatives-No additional restrictions.
Authors retain copyright and agree to license their articles with a Creative Commons Attribution-NonCommercial-