Performance Comparison of Naive Bayes and Support Vector Machine Methods in Music Genre Classification Based on Audio Signal Feature Extraction Using Mel-Frequency Cepstral Coefficients (MFCC)
DOI:
https://doi.org/10.30871/jaic.v10i1.11770Keywords:
Naïve Bayes, Support Vector Machine, Music Genre Classification, Mel-Frequency Cepstral Coefficients, Audio Signal ProcessingAbstract
Music genre classification has gained increasing attention with the emergence of digital music platforms. One of the relevant features extracted from audio signals is Mel-Frequency Cepstral Coefficients (MFCC), which is widely recognized as an effective technique. MFCC features are extracted at the frame level and aggregated at the clip level to represent each music track, making them suitable for audio-based classification tasks. This study applies Naïve Bayes and Support Vector Machine (SVM) algorithms for classification using the GTZAN dataset consisting of 1,000 audio files from 10 music genres, each with a duration of 30 seconds. The performance of these methods is evaluated using accuracy, precision, recall, and F1-score. The results show that SVM demonstrates superior performance, achieving an accuracy of 95.25% compared to 50.37% for Naïve Bayes. This performance gap can be attributed to SVM’s ability to model non-linear decision boundaries and effectively handle high-dimensional MFCC feature spaces. The main contribution of this study lies in the systematic evaluation of multiple SVM kernel configurations and parameter settings, providing empirical insights into the robustness of classical machine learning methods for MFCC-based music genre classification. This study concludes that SVM is better than Naive Bayes in music genre classification with MFCC feature extraction.
Downloads
References
[1] Y. Zhang and T. Li, “Music genre classification with parallel convolutional neural networks and capuchin search algorithm,” Scientific Reports, vol. 15, no. 1, p. 9580, 2025.
[2] R. A. Oladejo, A. A. Abayomi-Alli, O. Arogundade, A. A. Adeyanju and M. O. Lawrence, “ Artificial Intelligence in Music Genre Classification: A Systematic Review of Techniques, Applications, and Emerging Trends,” Cureus Journal Of Computer Science, 2025.
[3] A. Alamsyah, F. Ardiansyah and A. Kholiq, “Music Genre Classification Using Mel Frequency Cepstral Coefficients and Artificial Neural Networks: A Novel Approach,” Sceintific Journal of Informatics, vol. 11, no. 4, pp. 937-948, 2024.
[4] P. Ghosh, S. Mahapatra, S. Jana and R. K. Jha, “A Study on Music Genre Classification using Machine Learning,” International Journal of Engineering Business and Social Science, vol. 1, no. 4, pp. 257-268, 2023.
[5] T. Pratiwi, A. Sunyoto and D. Ariatmanto, “Music Genre Classification Using K-Nearest Neighbor and Mel-Frequency Cepstral Coefficients,” Sinkron : Jurnal Dan Penelitian Teknik Informatika, vol. 8, no. 2, pp. 861-867, 2024.
[6] L. Abdoune, M. Fezari and A. Dib, “Indoor Sound Classification with Support Vector Machines: State of the Art and Experimentation,” International Journal of Computational Methods and Experimental Measurements, vol. 12, no. 3, pp. 269-279, 2024.
[7] M. Müller, Fundamentals of Music Processing : : Audio, Analysis, Algorithms, Applications, Switzerland: Springer, 2015.
[8] N. Asanah and I. Pratama, “Deep Learning Approach for Music Genre Classification using Multi Feature Audio Representations,” Sistemasi: Jurnal Sistem Informasi, vol. 14, no. 5, pp. 2045-2054, 2025.
[9] G. Tzanetakis and P. R. Cook, “Musical Genre Classification of Audio Signals,” IEEE Transactions on Speech and Audio Processing, vol. 10, no. 5, pp. 293 - 302, 2002.
[10] S. Aggarwal and N. Aggarwal, “Classification of Audio Data using Support Vector Machine,” International Journal of Computer Science and Technology, vol. 2, no. 3, pp. 398-405, 2011.
[11] P. Sawaengsawangarom, S. Phongoen and P. Wongchaisuwat, “Deep Learning for Music Genre Classification: A case study of Thai music,” in 2025 International Conference on Multimedia Retrieval, 2025.
[12] Ardiansyah, B. Yuliadi and R. Sahara, “Music Genre Classification using Naïve Bayes Algorithm,” International Journal of Computer Trends and Technology (IJCTT), vol. 62, no. 1, pp. 50-57, 2018.
[13] J. Han, M. Kamber and J. Pei, Data Mining: Concepts and Techniques (3rd Edition). The Morgan Kaufmann Series in Data Management Systems., Elsevier, 2012.
[14] C.-W. Hsu, C.-C. Chang and C.-J. Lin, “A Practical Guide to Support Vector Classification,” 2003.
[15] T. Li, M. Ogihara and Q. Li, “A Comparative Study on Content-Based Music Genre Classification,” in Proceedings of the 26th annual international ACM SIGIR conference on Research and development in informaion retrieval, 2003.
[16] D. Barchiesi, D. Giannoulis, D. Stowell and M. D. Plumbley, “Acoustic Scene Classification: Classifying environments from the sounds they produce,” IEEE Signal Processing Magazine, vol. 32, no. 3, pp. 16-34, 2015.
[17] T. C. Ba, T. D. T. Le and L. T. Van, Music genre classification using deep neural networks and data augmentation, Elsevier, 2025.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 Kevin Putrayudha Naserwan, Kanda Januar Miraswan, Meylani Utari

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License (Attribution-ShareAlike 4.0 International (CC BY-SA 4.0) ) that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See The Effect of Open Access).








