TY - JOUR
T1 - A comparative analysis of ensemble autoML machine learning prediction accuracy of STEM student grade prediction
T2 - a multi-class classification prospective
AU - Rimal, Yagya Nath
AU - Sharma, Navneet
AU - Alsadoon, Abeer
AU - Abbas, Sayyed Khawar
PY - 2025
Y1 - 2025
N2 - Statisticians in the early stages in research aiming to predict the relationship between dependent and independent variables for better classification. This association may exhibit either a negative or positive correlation to target features, albeit with varying degrees of reliability. This paper tries to fulfill the research gap between the selection of appropriate Automl model for multiclass student grade classification for the predication bachelor degree grade. Consequently, this study endeavors to assess the predictive accuracy of an ensemble AutoML (Automated Machine Learning) model for science, technology, engineering, and management students letter grading. This assessment is based on their subject grades from high school through to internal evaluation of bachelor's degree, to predict bachelor's degree (final) outcomes when the target variables are in multiclass letter grading in a modern system. The ensemble AutoML approach is employed to forecast upcoming grades. Nine out of 78 recommended Automl models undergo fine-tuning and cross-validation for performance metrics, evaluating the best-optimized hyperparameters and assessing their performance after best-optimized hyperparameters. The study analyzed the performance of various models in classifying STEM students, focusing on their accuracy and prediction error rates and miss classification between train and predicated values. The GBM_4_AutoML_1 model scored the highest at 0.28 (28%), followed by StackedEnsemble_BestOfFamily_5 at 0.31 (31%), DRF at 0.28 (28%), XRT at 0.30 (30%), DeepLearning_grid at 0.56 (56%), and GLM at 0.35 (35%). Furthermore, the confusion matrices when the optimized model of the GBM scored 100% matched the true and predicated student grades. The history scoring of each model tuned recommended hyperparameters to achieve the best model. The feature importance of dependent and independent features was analyzed comprehensively to true and predicated and internal multiclass grading classification of STEM student and contrasted to provide a detailed explanation of their respective performance.
AB - Statisticians in the early stages in research aiming to predict the relationship between dependent and independent variables for better classification. This association may exhibit either a negative or positive correlation to target features, albeit with varying degrees of reliability. This paper tries to fulfill the research gap between the selection of appropriate Automl model for multiclass student grade classification for the predication bachelor degree grade. Consequently, this study endeavors to assess the predictive accuracy of an ensemble AutoML (Automated Machine Learning) model for science, technology, engineering, and management students letter grading. This assessment is based on their subject grades from high school through to internal evaluation of bachelor's degree, to predict bachelor's degree (final) outcomes when the target variables are in multiclass letter grading in a modern system. The ensemble AutoML approach is employed to forecast upcoming grades. Nine out of 78 recommended Automl models undergo fine-tuning and cross-validation for performance metrics, evaluating the best-optimized hyperparameters and assessing their performance after best-optimized hyperparameters. The study analyzed the performance of various models in classifying STEM students, focusing on their accuracy and prediction error rates and miss classification between train and predicated values. The GBM_4_AutoML_1 model scored the highest at 0.28 (28%), followed by StackedEnsemble_BestOfFamily_5 at 0.31 (31%), DRF at 0.28 (28%), XRT at 0.30 (30%), DeepLearning_grid at 0.56 (56%), and GLM at 0.35 (35%). Furthermore, the confusion matrices when the optimized model of the GBM scored 100% matched the true and predicated student grades. The history scoring of each model tuned recommended hyperparameters to achieve the best model. The feature importance of dependent and independent features was analyzed comprehensively to true and predicated and internal multiclass grading classification of STEM student and contrasted to provide a detailed explanation of their respective performance.
KW - Deep learning
KW - Distributed random forest
KW - Extremely randomized trees
KW - Gradient linear models
KW - Mean squared error
KW - Root mean squared error
KW - School leaving certificate
KW - Stacked ensembled ALL gradient boosting
UR - http://www.scopus.com/inward/record.url?scp=86000328045&partnerID=8YFLogxK
UR - https://go.openathens.net/redirector/westernsydney.edu.au?url=https://doi.org/10.1007/s11042-024-20554-8
U2 - 10.1007/s11042-024-20554-8
DO - 10.1007/s11042-024-20554-8
M3 - Article
AN - SCOPUS:86000328045
SN - 1380-7501
JO - Multimedia Tools and Applications
JF - Multimedia Tools and Applications
ER -