Musical Instruments Sound Classification using GMM
Keywords:Musical Instrument Sound Classification, mel frequency cepstral coefficient, Gaussian Mixture Model
Classification is the task of assigning objects to one of several predefined categories. In today’s decade classifying the musical signal from large data is a major task; the proposed work classifies the music into their respective classes. In this paper, the sound of the musical instruments classified automatically from the musical signals. Mel frequency cepstral coefficient is used as a feature extractor and the machine learning model namely Gaussian Mixture Model is used for classification. This system tested in ten different classes of musical instrument sound from two different instrument families such as Woodwind and Brass instruments. In this proposed work, the result yields satisfactory accuracy in the classification of musical instruments sound.
Dhanalakshmi, P., Palanivel, Sengottayan & Ramalingam, Vivekanandan. (2011). Classification of audio signals using AANN and GMM. Applied Soft Computing. 11. Pp. 716-723, 10.1016/j.asoc.2009.12.033.
Thiruvengatanadhan, R., Music Genre Classification using GMM (2018). International Research Journal of Engineering and Technology (IRJET), Volume: 05 Issue: 10.
Bhalke, D.G., Rao, C.B.R. & Bormane, D.S. Automatic Musical Instrument Classification using Fractional Fourier Transform based- MFCC features and Counter Propagation Neural Network. J Intell Inf Syst 46, 425–446 (2016). https://doi.org/10.1007/s10844-015-0360-9
Prabavathy, S, Rathikarani, V., & Dhanalakshmi, P, Classification of Musical Instruments using SVM and KNN, International Journal of Innovative Technology and Exploring Engineering (IJITEE), ISSN: 2278-3075, Volume-9 Issue-7, May 2020.
Shreevathsa, P. K., Harshith, M., A. R. M. & Ashwini, "Music Instrument Recognition using Machine Learning Algorithms," 2020 International Conference on Computation, Automation and Knowledge Management (ICCAKM), 2020, pp. 161-166, doi: 10.1109/ICCAKM46823.2020.9051514.
Dhanalakshmi, P. (2010). Classification of Audio for Retrieval Applications. [Doctoral Dissertation]. Faculty of Computer Science and Engineering, Annamalai University, Chidambaram, Tamilnadu.
Zokaee, Sara & Faez, Karim. (2012). Human Identification Based on Electrocardiogram and Palmprint. 2. 261-266.
Rafael Iriya & Miguel Arjona Ramírez. Gaussian Mixture Models with Class-Dependent Features for Speech Emotion Recognition. IEEE Workshop on Statistical Signal Processing, pp. 480-483, 2014.
Tang, H., Chu, S. M., Hasegawa-Johnson, M. & Huang, T. S., Partially Supervised Speaker Clustering. IEEE transactions on Pattern Analysis and Machine Intelligence. vol. 34, no. 5, pp. 959-971, 2012.
Krishna, A.G. & Sreenivas, Tv. (2004). Music instrument recognition: from isolated notes to solo phrases. International Conference on Acoustics, Speech, and Signal Processing, ICASSP.
How to Cite
Copyright (c) 2021 Prabhu Kumar Aurchana, S. Prabavathy
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
CC Attribution-NonCommercial-NoDerivatives 4.0