Classification of Post-School Tendencies of Madrasah Aliyah Students Using the K-Nearest Neighbor Algorithm
DOI:
https://doi.org/10.30871/jaic.v10i2.12354Keywords:
Data Mining, K-Nearest Neighbor, Classification, Madrasah Aliyah, Graduation PredictionAbstract
This research develops a classification model to identify post-graduation tendencies of Madrasah Aliyah students using the K-Nearest Neighbor (KNN) algorithm with academic report card scores as input features. The dataset includes 76 students, using average scores from semesters 1 to 5 as predictors and students’ post-school tendencies as the target variable. Data preprocessing involved normalization and splitting the dataset into training and testing subsets, while similarity between instances was measured using Euclidean Distance with k = 5. The experimental results achieved an accuracy of 87.50%, indicating that KNN performs well on small-scale academic datasets. This study contributes by specifically applying KNN to classify post-school tendencies in a Madrasah Aliyah context using limited academic features, an area that has not been extensively explored in previous educational data mining studies. The proposed model can assist schools in providing data-driven academic counseling and decision support within the Madrasah Aliyah environment.
Downloads
References
[1] T. Setiyorini and R. T. Asmono, “Implementation of Gain Ratio and K-Nearest Neighbor for Classification of Student Performance,” J. Pilar Nusa Mandiri, vol. 16, no. 1, pp. 19–24, 2020, doi: 10.33480/pilar.v16i1.813.
[2] “View of Comparative Analysis of Classification Methods of KNN and Naïve Bayes to Determine Stress Level of Junior High School Students.pdf.”
[3] “View of Penerapan Data Mining Metode K-Nearest Neighbor Untuk Memprediksi Kelulusan Siswa Sekolah Menengah Pertama.pdf.”
[4] A. Muhaimin, M. Amin Hariyadi, and M. I. Imamudin, “Klasifikasi Prestasi Akademik Siswa Berdasarkan Nilai Rapor dan Kedisiplinan dengan Metode K-Nearest Neighbor,” J. Ilmu Komput. dan Sist. Inf., vol. 7, no. 1, pp. 193–202, 2024, doi: 10.55338/jikomsi.v7i1.2865.
[5] D. R. Sari, V. Julianto, and H. Rhomadona, “Prediction Of Student Graduation Using The K-Nearest Neighbor Method Case Study in Politeknik Negeri Tanah Laut,” J. Ilm. Inform., vol. 8, no. 1, pp. 74–88, 2023, doi: 10.35316/jimi.v8i1.74-88.
[6] “View of Penerapan K-Nearest Neighbor Untuk Klasifikasi Tingkat Kelulusan Pada Siswa Sma Negeri 11 Kota Bengkulu.pdf.”
[7] L. H. Jufri and D. Dasari, “Evaluation of the implementation of machine learning algorithm K-Nearest Neighbors (KNN) using rapid miner on junior high school student learning outcomes,” J. Gantang, vol. 8, no. 2, pp. 193–197, 2023, doi: 10.31629/jg.v8i2.6590.
[8] A. Qurotul, E. Tasia, N. Nazira, P. F. Pratama, M. R. Anugrah, and J. Adhiva, “Implementasi Algoritma Naïve Bayes Classifier ( NBC ) untuk Klasifikasi Penyakit Ginjal Kronik,” vol. 4, no. September, pp. 72–76, 2022, doi: 10.30865/json.v4i1.4781.
[9] Z. Fatah et al., “Klasifikasi Data Mining Untuk Memprediksi Kelulusan Mahasiswa,” vol. 2, no. 1, pp. 29–37, 2025.
[10] N. B. Putri and A. W. Wijayanto, “Analisis Komparasi Algoritma Klasifikasi Data Mining Dalam Klasifikasi Website Phishing Comparative Analysis Of Data Mining Classification Algorithm In Phishing Website Classification,” vol. 11, no. 28, pp. 59–66, 2022, doi: 10.34010/komputika.v11i1.4350.
[11] P. Studi et al., “Data Mining Klasifikasi Penduduk Miskin Menggunakan Metode Support Vektor Machine,” vol. 8, no. April, 2025.
[12] I. M. Karo Karo, R. Romia, S. Dewi, and P. M. Fadilah, “Hoax Detection on Indonesian Tweets using Naïve Bayes Classifier with TF-IDF,” J. Inf. Syst. Res., vol. 4, no. 3, pp. 914–919, 2023, doi: 10.47065/josh.v4i3.3317.
[13] “View of Penerapan K-Nearest Neighbor Untuk Klasifikasi Tingkat Kelulusan Pada Siswa.pdf.”
[14] R. I. Sulthoni, L. A. Muharom, and M. Rahman, “Jurnal Smart Teknologi Analisis Tingkat Kepuasan Siswa Dalam Pembelajaran Hybyrid Menggunakan Algoritma K – Nearest Neighbor ( KNN ) Student Satisfaction Level Analysis In Hybyrid Learning Using K – Nearest Neighbor ( KNN ) Algorithm Jurnal Smart Teknologi,” vol. 4, no. 4, pp. 406–411, 2023.
[15] “View of Analisis Perbandingan Algoritma ID3 dan KNN Pada Klasifikasi Emosi Teks Berita Berbahasa Indonesia.pdf.”
[16] “Tampilan Penerapan Algoritma K-Nearest Neighbor untuk Klasifikasi Kemampuan Lulusan Siswa Dalam Bersaing untuk Mendapatkan Pekerjaan (Studi Kasus_ SMK ‘SORE’ Tulungagung).pdf,” 2017, Fakultas Ilmu Komputer, Universitas Brawijaya, Malang. doi: e-ISSN: 2548-964X.
[17] R. Dirantara, F. Sugandi, and U. D. Wacana, “Total data yang diprediksi dengan benar Akurasi = Total pengujian yang dilaksanakan x 100 %,” vol. 4307, no. 1, pp. 552–556, 2025.
[18] N. F. Munazhif, G. J. Yanris, and M. N. S. Hasibuan, “Implementation of the K-Nearest Neighbor (kNN) Method to Determine Outstanding Student Classes,” SinkrOn, vol. 8, no. 2, pp. 719–732, 2023, doi: 10.33395/sinkron.v8i2.12227.
[19] D. Zapata-Medina, A. Espinosa-Bedoya, and J. A. Jiménez-Builes, “Improving the Automatic Detection of Dropout Risk in Middle and High School Students: A Comparative Study of Feature Selection Techniques,” Mathematics, vol. 12, no. 12, 2024, doi: 10.3390/math12121776.
[20] “View of Analisis Klasifikasi Kelulusan Mahasiswa menggunakan Algoritma K-Nearest Neighbor di Fakultas Teknik Universitas Negeri Manado.pdf.”
[21] Roysid HA, Maulana A, Pujianto U. Can K-nearest neighbor method be used to predict success in Indonesia state university student selection. Jurnal Ilmiah Kursor. 2018 Dec 10;9(4).
[22] “Tampilan Optimasi Penyaluran Bantuan Langsung Tunai Menggunakan Algoritma K-Nearest Neighbor di Desa Penedagador.pdf.”
[23] W. Apriliah et al., “Prediksi Kemungkinan Diabetes pada Tahap Awal Menggunakan Algoritma Klasifikasi Random Forest,” vol. 10, pp. 163–171, 2021.
[24] I. G. Ayu, N. Lestari, N. Made, R. Masita, and K. G. Meiliana, “Effectiveness of AdaBoost and XGBoost Algorithms in Sentiment Analysis of Movie Reviews,” vol. 9, no. 2, pp. 258–264, 2025.
[25] A. Perbandingan et al., “Jurnal Teknologi Terpadu,” vol. 8, no. 2, pp. 121–126, 2022.
[26] M. Salsabil, N. Lutvi, and A. Eviyanti, “Implementasi Data Mining Dalam Melakukan Prediksi Penyakit Diabetes Menggunakan Metode Random Forest Dan Xgboost,” J. Ilm. Komputasi, vol. 23, no. 1, pp. 51–58, 2024, doi: 10.32409/jikstik.23.1.3507.
[27] A. Alhafiz and R. K. R, “Implementation of the K-Nearest Neighbor Algorithm for Birth Rate Prediction,” vol. 9, no. 4, pp. 1441–1450, 2025.
[28] H. Putra, “Comparative Study of Logistic Regression , Random Forest , and XGBoost for Bank Loan Approval Classification,” vol. 9, no. 5, pp. 2822–2835, 2025.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 Eka Asfira Atha Yanuha Fiftani, Ifnu Wisma Dwi Prastya, Ita Aristia Sa’ida

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License (Attribution-ShareAlike 4.0 International (CC BY-SA 4.0) ) that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See The Effect of Open Access).








