Hyperparameter Tuning Modeling for Socioeconomic based Academic Analysis

Authors

  • Triyan Agung Laksono Universitas Teknologi Digital Indonesia (UTDI), Daerah Istimewa Yogyakarta, Indonesia
  • Sri Redjeki Universitas Teknologi Digital Indonesia (UTDI), Daerah Istimewa Yogyakarta, Indonesia
  • Tutut Herawan Department of information Systems, Faculty of Computer Science and Information Technology, Universiti Malaya

Keywords:

Academic performance; Socioeconomic analysis; Machine learning; Hyperparameter tuning; Gradient Boosting Machine.

Abstract

This study analyzes the impact of hyperparameter tuning in improving the performance of predictive models for academic success based on socioeconomic data. To analyze their predictive capabilities, this research focuses on two ML algorithms, Gradient Boosting Machine (GBM) and Random Forest (RF). Using a UCI Machine Learning Repository dataset, this study implements grid search for hyperparameter tuning, optimizing parameters such as learning speed and number of estimators. Results show that GBM consistently outperforms RF, with higher average accuracy (78.64% vs. 77.45%), lower standard deviation (0.0077 vs. 0.0091), and better stability. Paired t-test results (p-value = 0.0081) confirmed the statistical significance of the superiority of GBM. This research contributes to the field by integrating socioeconomic factors into academic performance prediction models, providing valuable insights for data-driven educational decision-making.

Downloads

Download data is not yet available.

References

A. Selvitopu and M. Kaya, “A Meta-Analytic Review of the Effect of Socioeconomic Status on Academic Performance,” Journal of Education, vol. 203, no. 4, pp. 768–780, Oct. 2023, doi: 10.1177/00220574211031978.

S. Boateng, D. Asare, P. T. Manu, E. A. Sefah, and J. Adomako, “Relationship Between Students’ Home Background and Their Academic Performance: A Case of Some Selected Senior High School Students in Rural Districts in Ashanti Region, Ghana,” Journal of Education, vol. 201, no. 3, pp. 153–161, Oct. 2021, doi: 10.1177/0022057420904370.

B. Vadivel, S. Alam, I. Nikpoo, and B. Ajanil, “The Impact of Low Socioeconomic Background on a Child’s Educational Achievements,” Education Research International, vol. 2023, pp. 1–11, Jan. 2023, doi: 10.1155/2023/6565088.

W. Tomaszewski, N. Xiang, and M. Western, “Student engagement as a mediator of the effects of socio‐economic status on academic performance among secondary school students in Australia,” British Educational Res J, vol. 46, no. 3, pp. 610–630, Jun. 2020, doi: 10.1002/berj.3599.

F. Zhang, Y. Jiang, H. Ming, Y. Ren, L. Wang, and S. Huang, “Family socio‐economic status and children’s academic achievement: The different roles of parental academic involvement and subjective social mobility,” Brit J of Edu Psychol, vol. 90, no. 3, pp. 561–579, Sep. 2020, doi: 10.1111/bjep.12374.

M. Mwita, J. Mbelwa, J. Agbinya, and A. Elikana Sam, “The Effect of Hyperparameter Optimization on the Estimation of Performance Metrics in Network Traffic Prediction using the Gradient Boosting Machine Model,” Eng. Technol. Appl. Sci. Res., vol. 13, no. 3, pp. 10714–10720, Jun. 2023, doi: 10.48084/etasr.5548.

M. Tsiakmaki, G. Kostopoulos, S. Kotsiantis, and O. Ragos, “Implementing AutoML in Educational Data Mining for Prediction Tasks,” Applied Sciences, vol. 10, no. 1, p. 90, Dec. 2019, doi: 10.3390/app10010090.

I. Gligorea, M. U. Yaseen, M. Cioca, H. Gorski, and R. Oancea, “An Interpretable Framework for an Efficient Analysis of Students’ Academic Performance,” Sustainability, vol. 14, no. 14, p. 8885, Jul. 2022, doi: 10.3390/su14148885.

S. Jayaprakash, S. Krishnan, and V. Jaiganesh, “Predicting Students Academic Performance using an Improved Random Forest Classifier,” in 2020 International Conference on Emerging Smart Computing and Informatics (ESCI), Pune, India: IEEE, Mar. 2020, pp. 238–243. doi: 10.1109/ESCI48226.2020.9167547.

M. Adnan, A. A. S. Alarood, M. I. Uddin, and I. Ur Rehman, “Utilizing grid search cross-validation with adaptive boosting for augmenting performance of machine learning models,” PeerJ Computer Science, vol. 8, p. e803, Feb. 2022, doi: 10.7717/peerj-cs.803.

E. Elgeldawi, A. Sayed, A. R. Galal, and A. M. Zaki, “Hyperparameter Tuning for Machine Learning Algorithms Used for Arabic Sentiment Analysis,” Informatics, vol. 8, no. 4, p. 79, Nov. 2021, doi: 10.3390/informatics8040079.

E. Ahmed, “Student Performance Prediction Using Machine Learning Algorithms,” Applied Computational Intelligence and Soft Computing, vol. 2024, no. 1, p. 4067721, Jan. 2024, doi: 10.1155/2024/4067721.

S. Lee and J. Y. Chung, “The Machine Learning-Based Dropout Early Warning System for Improving the Performance of Dropout Prediction,” Applied Sciences, vol. 9, no. 15, p. 3093, Jul. 2019, doi: 10.3390/app9153093.

A. Behr, M. Giese, H. D. Teguim K, and K. Theune, “Early Prediction of University Dropouts – A Random Forest Approach,” Jahrbücher für Nationalökonomie und Statistik, vol. 240, no. 6, pp. 743–789, Oct. 2020, doi: 10.1515/jbnst-2019-0006.

M. Arifin, W. Widowati, and F. Farikhin, “Optimization of Hyperparameters in Machine Learning for Enhancing Predictions of Student Academic Performance,” ISI, vol. 28, no. 3, pp. 575–582, Jun. 2023, doi: 10.18280/isi.280305.

F. Abbas et al., “Optimizing Machine Learning Algorithms for Landslide Susceptibility Mapping along the Karakoram Highway, Gilgit Baltistan, Pakistan: A Comparative Study of Baseline, Bayesian, and Metaheuristic Hyperparameter Optimization Techniques,” Sensors, vol. 23, no. 15, p. 6843, Aug. 2023, doi: 10.3390/s23156843.

S. Rajendran, S. Chamundeswari, and A. A. Sinha, “Predicting the academic performance of middle- and high-school students using machine learning algorithms,” Social Sciences & Humanities Open, vol. 6, no. 1, p. 100357, 2022, doi: 10.1016/j.ssaho.2022.100357.

L. Zahedi, F. G. Mohammadi, and M. H. Amini, “HyP-ABC: A Novel Automated Hyper-Parameter Tuning Algorithm Using Evolutionary Optimization,” Jun. 11, 2021. doi: 10.36227/techrxiv.14714508.v2.

N. Ali et al., “Academic performance of children in relation to gender, parenting styles, and socioeconomic status: What attributes are important,” PLoS ONE, vol. 18, no. 11, p. e0286823, Nov. 2023, doi: 10.1371/journal.pone.0286823.

M. Ogunsanya, J. Isichei, and S. Desai, “Grid search hyperparameter tuning in additive manufacturing processes,” Manufacturing Letters, vol. 35, pp. 1031–1042, Aug. 2023, doi: 10.1016/j.mfglet.2023.08.056.

L. Yang and A. Shami, “On hyperparameter optimization of machine learning algorithms: Theory and practice,” Neurocomputing, vol. 415, pp. 295–316, Nov. 2020, doi: 10.1016/j.neucom.2020.07.061.

J. Wu, X.-Y. Chen, H. Zhang, L.-D. Xiong, H. Lei, and S.-H. Deng, “Hyperparameter Optimization for Machine Learning Models Based on Bayesian Optimization,” vol. 17, no. 1, 2019.

A. Alhassan, B. Zafar, and A. Mueen, “Predict Students’ Academic Performance based on their Assessment Grades and Online Activity Data,” IJACSA, vol. 11, no. 4, 2020, doi: 10.14569/IJACSA.2020.0110425.

S. Batool, J. Rashid, M. W. Nisar, J. Kim, T. Mahmood, and A. Hussain, “A Random Forest Students’ Performance Prediction (RFSPP) Model Based on Students’ Demographic Features,” in 2021 Mohammad Ali Jinnah University International Conference on Computing (MAJICC), Karachi, Pakistan: IEEE, Jul. 2021, pp. 1–4. doi: 10.1109/MAJICC53071.2021.9526239.

M. V. M. Valentim Realinho, “Predict Students’ Dropout and Academic Success.” UCI Machine Learning Repository, 2021. doi: 10.24432/C5MC89.

J. L. Rastrollo-Guerrero, J. A. Gómez-Pulido, and A. Durán-Domínguez, “Analyzing and Predicting Students’ Performance by Means of Machine Learning: A Review,” Applied Sciences, vol. 10, no. 3, p. 1042, Feb. 2020, doi: 10.3390/app10031042.

M. L. Ruiz-Rodriguez, J. Andres Sandoval-Bringas, and M. A. Carreno-Leon, “Classification of student success using Random Forest and Neural Networks,” in 2020 3rd International Conference of Inclusive Technology and Education (CONTIE), Baja California Sur, Mexico: IEEE, Oct. 2020, pp. 98–103. doi: 10.1109/CONTIE51334.2020.00027.

K. Kiatkarun and P. Phunchongharn, “Automatic Hyper-Parameter Tuning for Gradient Boosting Machine,” in 2020 1st International Conference on Big Data Analytics and Practices (IBDAP), Bangkok, Thailand: IEEE, Sep. 2020, pp. 1–6. doi: 10.1109/IBDAP50342.2020.9245609.

B. C. Perez, M. C. A. M. Bink, K. L. Svenson, G. A. Churchill, and M. P. L. Calus, “Prediction performance of linear models and gradient boosting machine on complex phenotypes in outbred mice,” G3 Genes|Genomes|Genetics, vol. 12, no. 4, p. jkac039, Apr. 2022, doi: 10.1093/g3journal/jkac039.

N. Gorgolis, I. Hatzilygeroudis, Z. Istenes, and L. N. G. Gyenne, “Hyperparameter Optimization of LSTM Network Models through Genetic Algorithm,” in 2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA), PATRAS, Greece: IEEE, Jul. 2019, pp. 1–4. doi: 10.1109/IISA.2019.8900675.

M. Zhang, C. Krintz, M. Mock, and R. Wolski, “Seneca: Fast and Low Cost Hyperparameter Search for Machine Learning Models,” in 2019 IEEE 12th International Conference on Cloud Computing (CLOUD), Milan, Italy: IEEE, Jul. 2019, pp. 404–408. doi: 10.1109/CLOUD.2019.00071.

Published

2024-12-20

How to Cite

Laksono, T. A. ., Redjeki, S. ., & Herawan, T. (2024). Hyperparameter Tuning Modeling for Socioeconomic based Academic Analysis. Journal of Information Systems Research and Practice, 2(5), 2–16. Retrieved from https://vmis.um.edu.my/index.php/JISRP/article/view/57829