NAWOA-XGBoost: A Novel Model for Early Prediction of Academic Potential in Computer Science Students

Reading time: 5 minute
...

📝 Original Info

  • Title: NAWOA-XGBoost: A Novel Model for Early Prediction of Academic Potential in Computer Science Students
  • ArXiv ID: 2512.04751
  • Date: 2025-12-04
  • Authors: Researchers from original ArXiv paper

📝 Abstract

Whale Optimization Algorithm (WOA) suffers from limited global search ability, slow convergence, and tendency to fall into local optima, restricting its effectiveness in hyperparameter optimization for machine learning models. To address these issues, this study proposes a Nonlinear Adaptive Whale Optimization Algorithm (NAWOA), which integrates strategies such as Good Nodes Set initialization, Leader-Followers Foraging, Dynamic Encircling Prey, Triangular Hunting, and a nonlinear convergence factor to enhance exploration, exploitation, and convergence stability. Experiments on 23 benchmark functions demonstrate NAWOA's superior optimization capability and robustness. Based on this optimizer, an NAWOA-XGBoost model was developed to predict academic potential using data from 495 Computer Science undergraduates at Macao Polytechnic University (2009-2019). Results show that NAWOA-XGBoost outperforms traditional XGBoost and WOA-XGBoost across key metrics, including Accuracy (0.8148), Macro F1 (0.8101), AUC (0.8932), and G-Mean (0.8172), demonstrating strong adaptability on multi-class imbalanced datasets.

💡 Deep Analysis

Deep Dive into NAWOA-XGBoost: A Novel Model for Early Prediction of Academic Potential in Computer Science Students.

Whale Optimization Algorithm (WOA) suffers from limited global search ability, slow convergence, and tendency to fall into local optima, restricting its effectiveness in hyperparameter optimization for machine learning models. To address these issues, this study proposes a Nonlinear Adaptive Whale Optimization Algorithm (NAWOA), which integrates strategies such as Good Nodes Set initialization, Leader-Followers Foraging, Dynamic Encircling Prey, Triangular Hunting, and a nonlinear convergence factor to enhance exploration, exploitation, and convergence stability. Experiments on 23 benchmark functions demonstrate NAWOA’s superior optimization capability and robustness. Based on this optimizer, an NAWOA-XGBoost model was developed to predict academic potential using data from 495 Computer Science undergraduates at Macao Polytechnic University (2009-2019). Results show that NAWOA-XGBoost outperforms traditional XGBoost and WOA-XGBoost across key metrics, including Accuracy (0.8148), Macro

📄 Full Content

NAWOA-XGBOOST: A NOVEL MODEL FOR EARLY PREDICTION OF ACADEMIC POTENTIAL IN COMPUTER SCIENCE STUDENTS ∗ Junhao Wei, Yanzhao Gu, Ran Zhang, Mingjing Huang, Jinhong Song, Yanxiao Li, Wenxuan Zhu, Yapeng Wang Faculty of Applied Sciences Macao Polytechnic University Macao, China {p2312195,p2311998,p2512396,P2314640,p2315937,P2525981,p2525620,yapengwang}@mpu.edu.mo Zikun Li School of Economics and Management South China Normal University 18520610821@163.com Zhiwen Wang School of Nursing Peking University wzwjing@sina.com Xu Yang*, Ngai Cheong* Faculty of Applied Sciences Macao Polytechnic University {xuyang,ncheong}@mpu.edu.mo ABSTRACT Whale Optimization Algorithm (WOA) suffers from limited global search ability, slow convergence, and tendency to fall into local optima, restricting its effectiveness in hyperparameter optimization for machine learning models. To address these issues, this study proposes a Nonlinear Adaptive Whale Optimization Algorithm (NAWOA), which integrates strategies such as Good Nodes Set initializa- tion, Leader-Followers Foraging, Dynamic Encircling Prey, Triangular Hunting, and a nonlinear convergence factor to enhance exploration, exploitation, and convergence stability. Experiments on 23 benchmark functions demonstrate NAWOA’s superior optimization capability and robustness. Based on this optimizer, an NAWOA-XGBoost model was developed to predict academic potential using data from 495 Computer Science undergraduates at Macao Polytechnic University (2009-2019). Results show that NAWOA-XGBoost outperforms traditional XGBoost and WOA-XGBoost across key metrics, including Accuracy (0.8148), Macro F1 (0.8101), AUC (0.8932), and G-Mean (0.8172), demonstrating strong adaptability on multi-class imbalanced datasets. Keywords Whale Optimization Algorithm · XGBoost · data mining · education technology 1 Introduction With the rapid development of Educational Data Mining (EDM) [1], predicting students’ future academic potential based on their early learning performance has become an important research direction for quality assurance, resource allocation, and personalized instruction in higher education. In the field of Computer Science (CS), students’ academic records and behavioral data during their first and second years often provide strong indications of their foundational knowledge, learning capability, and future development potential. Therefore, establishing a reliable early academic potential prediction model is of significant importance. In recent years, machine learning methods have been extensively applied to educational prediction tasks. Among these methods, XGBoost has been regarded as a suitable model due to its strong generalization ability and superior performance on structured data. However, the performance of XGBoost highly depends on its hyperparameter settings. Manual tuning is inefficient and prone to local optima, which restricts further improvements in predictive performance. To address this limitation, researchers started to pay attention on metaheuristic algorithms. Metaheuristic algorithms, which are famous of their strong optimization abilities, have been introduced to path planning [2][3][4], engineering design, antenna design [5], and optimizing the hyperparameters of machine learning models. Among them, the Whale Optimization Algorithm (WOA) has attracted attention due to its simplicity, efficiency, and global search capability. ∗Citation: Authors. Title. Pages.... DOI:000000/11111. arXiv:2512.04751v2 [cs.CE] 5 Dec 2025 Running Title for Header Nevertheless, the conventional WOA tends to suffer from premature convergence and loss of population diversity in later iterations, limiting its ability to effectively explore complex parameter spaces. Previous studies have integrated the classical WOA with XGBoost and achieved promising prediction results [6]. However, there remains considerable room for improvement. Building upon this foundation, the present study proposes a Nonlinear Adaptive Whale Optimization Algorithm (NAWOA), which incorporates strategies such as Good Nodes Set initialization, Leader-Followers Foraging, and Dynamic Encircling Prey to enhance global exploration ability and optimization stability. Experiments conducted on 23 benchmark functions validate the strong optimization capability of the proposed algorithm [7][8]. Based on this enhanced optimizer, we construct an NAWOA-XGBoost academic potential prediction model and evaluate it using a real dataset of 495 undergraduate students majoring in Computer Science at Macao Polytechnic University (MPU) from 2009 to 2019. The dataset was preprocessed and provided by the institution, enabling direct use for training and comparative analysis. Experimental results indicate that NAWOA can more effectively optimize the hyperparameters of XGBoost, thereby enhancing model stability and predictive performance in multi-class educational prediction tasks. The proposed model provides a valuable reference for early acad

…(Full text truncated)…

📸 Image Gallery

TSWOA_matrix.jpg TSWOA_matrix.webp WOA_matrix.jpg WOA_matrix.webp XGB_matrix.jpg XGB_matrix.webp ini.png ini.webp

Reference

This content is AI-processed based on ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut