Classification of Post-School Tendencies of Madrasah Aliyah Students Using the K-Nearest Neighbor Algorithm

Authors

  • Eka Asfira Atha Yanuha Fiftani Universitas Nahdlatul Ulama Sunan Giri
  • Ifnu Wisma Dwi Prastya Universitas Nahdlatul Ulama Sunan Giri
  • Ita Aristia Sa’ida Universitas Nahdlatul Ulama Sunan Giri

DOI:

https://doi.org/10.30871/jaic.v10i2.12354

Keywords:

Data Mining, K-Nearest Neighbor, Classification, Madrasah Aliyah, Graduation Prediction

Abstract

This research develops a classification model to identify post-graduation tendencies of Madrasah Aliyah students using the K-Nearest Neighbor (KNN) algorithm with academic report card scores as input features. The dataset includes 76 students, using average scores from semesters 1 to 5 as predictors and students’ post-school tendencies as the target variable. Data preprocessing involved normalization and splitting the dataset into training and testing subsets, while similarity between instances was measured using Euclidean Distance with k = 5. The experimental results achieved an accuracy of 87.50%, indicating that KNN performs well on small-scale academic datasets. This study contributes by specifically applying KNN to classify post-school tendencies in a Madrasah Aliyah context using limited academic features, an area that has not been extensively explored in previous educational data mining studies. The proposed model can assist schools in providing data-driven academic counseling and decision support within the Madrasah Aliyah environment.

Downloads

Download data is not yet available.

References

[1] T. Setiyorini and R. T. Asmono, “Implementation of Gain Ratio and K-Nearest Neighbor for Classification of Student Performance,” J. Pilar Nusa Mandiri, vol. 16, no. 1, pp. 19–24, 2020, doi: 10.33480/pilar.v16i1.813.

[2] “View of Comparative Analysis of Classification Methods of KNN and Naïve Bayes to Determine Stress Level of Junior High School Students.pdf.”

[3] “View of Penerapan Data Mining Metode K-Nearest Neighbor Untuk Memprediksi Kelulusan Siswa Sekolah Menengah Pertama.pdf.”

[4] A. Muhaimin, M. Amin Hariyadi, and M. I. Imamudin, “Klasifikasi Prestasi Akademik Siswa Berdasarkan Nilai Rapor dan Kedisiplinan dengan Metode K-Nearest Neighbor,” J. Ilmu Komput. dan Sist. Inf., vol. 7, no. 1, pp. 193–202, 2024, doi: 10.55338/jikomsi.v7i1.2865.

[5] D. R. Sari, V. Julianto, and H. Rhomadona, “Prediction Of Student Graduation Using The K-Nearest Neighbor Method Case Study in Politeknik Negeri Tanah Laut,” J. Ilm. Inform., vol. 8, no. 1, pp. 74–88, 2023, doi: 10.35316/jimi.v8i1.74-88.

[6] “View of Penerapan K-Nearest Neighbor Untuk Klasifikasi Tingkat Kelulusan Pada Siswa Sma Negeri 11 Kota Bengkulu.pdf.”

[7] L. H. Jufri and D. Dasari, “Evaluation of the implementation of machine learning algorithm K-Nearest Neighbors (KNN) using rapid miner on junior high school student learning outcomes,” J. Gantang, vol. 8, no. 2, pp. 193–197, 2023, doi: 10.31629/jg.v8i2.6590.

[8] A. Qurotul, E. Tasia, N. Nazira, P. F. Pratama, M. R. Anugrah, and J. Adhiva, “Implementasi Algoritma Naïve Bayes Classifier ( NBC ) untuk Klasifikasi Penyakit Ginjal Kronik,” vol. 4, no. September, pp. 72–76, 2022, doi: 10.30865/json.v4i1.4781.

[9] Z. Fatah et al., “Klasifikasi Data Mining Untuk Memprediksi Kelulusan Mahasiswa,” vol. 2, no. 1, pp. 29–37, 2025.

[10] N. B. Putri and A. W. Wijayanto, “Analisis Komparasi Algoritma Klasifikasi Data Mining Dalam Klasifikasi Website Phishing Comparative Analysis Of Data Mining Classification Algorithm In Phishing Website Classification,” vol. 11, no. 28, pp. 59–66, 2022, doi: 10.34010/komputika.v11i1.4350.

[11] P. Studi et al., “Data Mining Klasifikasi Penduduk Miskin Menggunakan Metode Support Vektor Machine,” vol. 8, no. April, 2025.

[12] I. M. Karo Karo, R. Romia, S. Dewi, and P. M. Fadilah, “Hoax Detection on Indonesian Tweets using Naïve Bayes Classifier with TF-IDF,” J. Inf. Syst. Res., vol. 4, no. 3, pp. 914–919, 2023, doi: 10.47065/josh.v4i3.3317.

[13] “View of Penerapan K-Nearest Neighbor Untuk Klasifikasi Tingkat Kelulusan Pada Siswa.pdf.”

[14] R. I. Sulthoni, L. A. Muharom, and M. Rahman, “Jurnal Smart Teknologi Analisis Tingkat Kepuasan Siswa Dalam Pembelajaran Hybyrid Menggunakan Algoritma K – Nearest Neighbor ( KNN ) Student Satisfaction Level Analysis In Hybyrid Learning Using K – Nearest Neighbor ( KNN ) Algorithm Jurnal Smart Teknologi,” vol. 4, no. 4, pp. 406–411, 2023.

[15] “View of Analisis Perbandingan Algoritma ID3 dan KNN Pada Klasifikasi Emosi Teks Berita Berbahasa Indonesia.pdf.”

[16] “Tampilan Penerapan Algoritma K-Nearest Neighbor untuk Klasifikasi Kemampuan Lulusan Siswa Dalam Bersaing untuk Mendapatkan Pekerjaan (Studi Kasus_ SMK ‘SORE’ Tulungagung).pdf,” 2017, Fakultas Ilmu Komputer, Universitas Brawijaya, Malang. doi: e-ISSN: 2548-964X.

[17] R. Dirantara, F. Sugandi, and U. D. Wacana, “Total data yang diprediksi dengan benar Akurasi = Total pengujian yang dilaksanakan x 100 %,” vol. 4307, no. 1, pp. 552–556, 2025.

[18] N. F. Munazhif, G. J. Yanris, and M. N. S. Hasibuan, “Implementation of the K-Nearest Neighbor (kNN) Method to Determine Outstanding Student Classes,” SinkrOn, vol. 8, no. 2, pp. 719–732, 2023, doi: 10.33395/sinkron.v8i2.12227.

[19] D. Zapata-Medina, A. Espinosa-Bedoya, and J. A. Jiménez-Builes, “Improving the Automatic Detection of Dropout Risk in Middle and High School Students: A Comparative Study of Feature Selection Techniques,” Mathematics, vol. 12, no. 12, 2024, doi: 10.3390/math12121776.

[20] “View of Analisis Klasifikasi Kelulusan Mahasiswa menggunakan Algoritma K-Nearest Neighbor di Fakultas Teknik Universitas Negeri Manado.pdf.”

[21] Roysid HA, Maulana A, Pujianto U. Can K-nearest neighbor method be used to predict success in Indonesia state university student selection. Jurnal Ilmiah Kursor. 2018 Dec 10;9(4).

[22] “Tampilan Optimasi Penyaluran Bantuan Langsung Tunai Menggunakan Algoritma K-Nearest Neighbor di Desa Penedagador.pdf.”

[23] W. Apriliah et al., “Prediksi Kemungkinan Diabetes pada Tahap Awal Menggunakan Algoritma Klasifikasi Random Forest,” vol. 10, pp. 163–171, 2021.

[24] I. G. Ayu, N. Lestari, N. Made, R. Masita, and K. G. Meiliana, “Effectiveness of AdaBoost and XGBoost Algorithms in Sentiment Analysis of Movie Reviews,” vol. 9, no. 2, pp. 258–264, 2025.

[25] A. Perbandingan et al., “Jurnal Teknologi Terpadu,” vol. 8, no. 2, pp. 121–126, 2022.

[26] M. Salsabil, N. Lutvi, and A. Eviyanti, “Implementasi Data Mining Dalam Melakukan Prediksi Penyakit Diabetes Menggunakan Metode Random Forest Dan Xgboost,” J. Ilm. Komputasi, vol. 23, no. 1, pp. 51–58, 2024, doi: 10.32409/jikstik.23.1.3507.

[27] A. Alhafiz and R. K. R, “Implementation of the K-Nearest Neighbor Algorithm for Birth Rate Prediction,” vol. 9, no. 4, pp. 1441–1450, 2025.

[28] H. Putra, “Comparative Study of Logistic Regression , Random Forest , and XGBoost for Bank Loan Approval Classification,” vol. 9, no. 5, pp. 2822–2835, 2025.

Downloads

Published

2026-04-23

How to Cite

[1]
E. A. A. Yanuha Fiftani, I. W. Dwi Prastya, and I. A. Sa’ida, “Classification of Post-School Tendencies of Madrasah Aliyah Students Using the K-Nearest Neighbor Algorithm”, JAIC, vol. 10, no. 2, pp. 1843–1852, Apr. 2026.

Issue

Section

Articles

Similar Articles

1 2 3 4 5 > >> 

You may also start an advanced similarity search for this article.