Preprint
Article

This version is not peer-reviewed.

Knowledge-Based Recommendation for Graduate Subject Allocation Using Graph Attention Networks (GAT)

Submitted:

03 March 2026

Posted:

04 March 2026

You are already at the latest version

Abstract
This study proposes a hybrid artificial intelligence framework for graduate subject allocation that enhances fairness, transparency, and operational efficiency in higher education. By integrating rule-based reasoning with advanced deep learning models, the framework supports data-driven decision-making for academic suitability, workload equity, and research alignment, while embedding Explainable Artificial Intelligence (XAI) to facilitate digital transformation in Thai universities. Traditional subject allocation processes in graduate programs are often manual, time-consuming, and subject to subjective judgment, thereby limiting their capacity to adapt to increasing curricular complexity and diverse faculty profiles. To address these challenges, the proposed framework combines institutional knowledge-based rules with machine learning techniques to model complex academic relationships across faculty, subjects, and workloads. Empirical analysis was conducted using data from 480 faculty members at Mahasarakham University and evaluated using multiple predictive models, including XGBoost, Wide-and-Deep Neural Networks, and Graph Neural Networks. Faculty performance was assessed using 20 institutional indicators reflecting teaching experience, research productivity, supervision, administrative responsibilities, and digital pedagogical competence. A multi-objective ranking algorithm was applied to simultaneously optimize academic suitability, workload balance, and research–teaching alignment. To enhance transparency and usability, the system incorporates a Faculty Subject Allocation Dashboard (FSAD) with SHAP-based interpretability, enabling administrators to understand and validate allocation decisions in real time. The results demonstrate that the proposed framework significantly improves allocation accuracy, workload equity, and decision transparency, offering a scalable and explainable solution for AI-driven academic governance.
Keywords: 
;  ;  ;  ;  ;  ;  

1. Introduction

In recent years, higher education institutions have faced growing complexity in curriculum design and subject allocation due to the rapid expansion of course offerings, interdisciplinary programs, and varied student profiles Duncombe (1959). At the same time, the advent of large-scale educational data (student records, faculty profiles, course metadata) creates opportunities to apply intelligent systems for decision support (Zhang et al. 2025). Recommendation systems (RSs) have emerged as crucial tools for filtering choices, suggesting alternatives, and guiding decision-making in high-dimensional, constrained domains (Zhang et al. 2026). In the educational domain, RSs have traditionally supported students (e.g., by recommending learning resources and courses) (Gomis et al. 2023). However, relatively few studies consider RSs in faculty workload and subject allocation, which involve multiple stakeholders (students, faculty, programs) and hard constraints (prerequisites, scheduling, capacity). The need to allocate subjects optimally is especially pronounced in graduate studies, where curriculum progress, research alignment, and faculty specialization are of great importance (Su and Su 2025). Traditionally, subject allocation has been rule-based (program policies, prerequisites, faculty load), overseen by academic advisers and administrators. However, such manual systems have limitations: they may not account for latent factors such as faculty research alignment, student interest vectors, or the interplay among faculty-student-subject relationships (kuabe et al.2025). Meanwhile, advances in machine learning (ML) and deep learning (DL) offer greater capacity to model complex, relational, and higher-order structures (Ragab et al., 2025). Knowledge-graph-enhanced recommendation systems extend beyond traditional collaborative filtering and content-based methods by leveraging side information and relational embeddings, thus improving performance and explainability (Zhang 2025). For example, Graph Neural Networks (GNNs) applied to heterogeneous educational graphs (students, courses, faculty) have demonstrated superior ranking accuracy in course recommendations (Qu et al. 2024). In the context of subject allocation, these methods can encode faculty expertise, research output, student interests, program relationships, and scheduling constraints into one unified graph of entities and relations.
Despite these technical advances, there remains a gap in the application of hybrid frameworks to subject allocation at the graduate level. Most work focuses on recommending learning resources or courses to students rather than optimizing subject allocation decisions that involve faculty, program, and institutional constraints (Zhao 2025). Moreover, existing educational RS research often neglects faculty-centric factors (such as research productivity, mentoring quality, and digital pedagogy competence), which are salient in graduate subject allocation. Therefore, this study proposes a hybrid framework combining a curriculum knowledge base (encoding prerequisites, capacity, scheduling, faculty eligibility) with deep learning recommendation models (XGBoots, GNN, wide & deep) to generate optimal subject-faculty allocations in graduate studies. The key contributions are three-fold: (Duncombe 1959) design of a faculty-subject-program knowledge graph incorporating faculty competency and institutional rules; (F. Zhang et al. 2025) training of deep hybrid recommenders that predict suitability scores for subject allocation; (Y. Zhang et al. 2026) evaluation of the system in a real institutional context (~480 respondents across six faculties) to assess performance and interpretability. In the remainder of this paper, we review related work on educational recommendation systems, knowledge graph-based recommendations, and faculty workload allocation frameworks (Section II), describe the research methodology (Section III), present evaluation performances (Section IV), discuss experiment results for practice (Section V), and conclude with conclusions and future work (Section VI and VII).

2. Materials and Methods

Over the past decade, recommender systems (RSs) have evolved into core components of digital learning ecosystems, revolutionizing how universities deliver personalized education and make data-driven decisions across teaching, learning, and management processes [11]. In higher education, these systems form part of the broader shift toward Learning Analytics and Educational Data Mining frameworks, enabling the dynamic personalization of courses, resources, and learning trajectories for individual learners. Educational Recommender Systems (ERSs) are primarily designed to enhance learning experiences by filtering, ranking, and matching learning resources, instructors, and students within massive institutional databases (Kopyt et al. 2016)A growing body of research demonstrates their effectiveness in course recommendation, adaptive tutoring systems, e-learning resource curation, and student performance prediction, establishing ERSs as critical enablers of personalized, learner-centered education. However, systematic literature reviews reveal that, although ERSs have achieved significant success in instructional domains, their application to administrative and institutional processes remains limited, particularly in graduate education contexts (Fardel et al., 2007). Traditional subject allocation frameworks in higher education often rely on rule-based assignment or manual heuristic decision-making, constrained by static institutional policies, faculty self-reports, or historical workload data (Comite and Pierdicca 2022). These methods tend to be less adaptable to evolving academic conditions, leading to mismatches between faculty expertise, student research interests, and institutional strategic goals. Such limitations demonstrate the growing need for intelligent, data-driven decision-support systems that can balance academic quality, workload fairness, and policy compliance. As universities adopt digital transformation and AI-enabled governance frameworks, integrating artificial intelligence (AI) and deep learning (DL) into subject allocation processes has emerged as a promising research frontier (Habi and Messer 2021).
Recent advancements in deep learning architectures (Neural Collaborative Filtering (NCF), Wide-and-Deep Neural Networks, and Graph Neural Networks (GNNs)) have significantly expanded the capacity of recommendation systems to learn latent representations and multi-relational patterns that traditional statistical models often fail to capture [16]. These models transcend simple input–output prediction by embedding semantic and structural relationships among diverse entities such as students, faculty, subjects, and institutional rules. In educational applications, such models enable reasoning over both explicit relationships (e.g., teaching history, research specialization) and implicit associations (e.g., mentoring effectiveness, digital teaching affinity). Knowledge Graphs (KGs) have further augmented this capability by structuring educational data into interconnected ontologies that allow semantic reasoning. Studies have shown that KG-based educational recommenders enhance adaptive learning pathways, dynamic course sequencing, and skill mapping—resulting in improved personalization and cognitive alignment between learners and content (Robillos 2024). However, despite these successes, the application of KGs and deep learning to academic management remains underexplored. Most prior works focus on student-side personalization, leaving institutional decision-making processes largely rule-driven and manually optimized. Recent hybrid frameworks that integrate knowledge-based reasoning with deep representation learning represent a transformative paradigm shift, combining the transparency of expert systems with the adaptability of neural models (Tang 2025). Such hybrid models can encode formal institutional constraints while simultaneously learning implicit patterns from historical data to improve fairness in allocation and predictive accuracy. Incorporating GNNs into educational recommendation architectures has proven particularly promising, as these models capture graph-structured dependencies across academic entities. GNNs enable joint optimization across heterogeneous nodes (faculty, subjects, programs) and their relational edges (e.g., expertise similarity, co-teaching, administrative linkage), leading to higher accuracy and explainability in recommendation outputs (Aziz 2025). Moreover, integrating reinforcement learning (RL) mechanisms enables these systems to adapt dynamically to policy shifts and feedback loops over time—supporting continuous learning and optimization in evolving academic environments. Systematic reviews further emphasize that explainability and fairness are emerging as fundamental design imperatives in educational AI, ensuring that algorithmic decisions uphold ethical standards, promote equity, and comply with accreditation requirements (Aina et al., 2024). In the context of faculty workload allocation, scholars have identified significant challenges in balancing quantitative factors (e.g., teaching experience, academic rank, and administrative duties) with qualitative indicators (e.g., mentoring quality, research supervision, and digital pedagogical skills) (Zhang 2025). Current decision models often fail to integrate these diverse dimensions holistically, resulting in inconsistent workload distribution and potential faculty dissatisfaction. Few frameworks treat the triadic relationship among faculty, subjects, and programs as a co-dependent decision space in which optimizing one dimension directly influences the others. Knowledge graph–enhanced hybrid recommender systems directly address this gap by enabling multi-objective optimization while preserving interpretable rule traces for administrative validation (Downing 2025). Empirical studies have demonstrated that combining domain ontologies and explicit workload constraints with neural learning architectures substantially improves both accuracy and transparency, making such systems well-suited for institutional deployment.
In summary, the existing literature provides compelling evidence for the potential of deep learning–based, knowledge-graph–augmented recommender systems to address complex educational challenges. However, significant gaps remain regarding faculty-centric subject allocation, hybrid knowledge-driven optimization, and real-world validation within university environments. The present study aims to bridge these gaps by developing and evaluating a knowledge-based deep learning framework that integrates graph attention mechanisms, policy rule reasoning, and multi-objective optimization. This approach aspires to achieve optimized, equitable, and explainable subject allocation in graduate studies—transforming academic resource management into a transparent, data-driven, and institutionally aligned process that reflects the next stage of intelligent higher-education governance.

2.1. Research Objectives

The overarching goal of this research is to design and validate a knowledge-based deep learning framework that intelligently automates and optimizes subject allocation in graduate programs, ensuring fairness, transparency, and institutional alignment. Building on advances in Graph Attention Networks (GATs) and knowledge-based reasoning, the study seeks to transform the traditional rule-driven subject assignment process into a data-driven, interpretable, and adaptive decision-support system.
To achieve this overarching goal, the study is guided by the following specific objectives:
  • To design a hybrid knowledge-based GAT framework that integrates explicit institutional rules (e.g., qualification requirements, workload limits, supervision policies) with data-driven learning from historical allocation patterns.
  • To construct a comprehensive indicator–subject knowledge graph that captures semantic and relational dependencies among indicator members, subjects, research domains, and workload profiles, enabling contextual reasoning within the allocation process.
  • To develop a GAT-based allocation model that learns attention coefficients representing the importance of faculty indicators in predicting allocation suitability.
  • To apply multi-objective optimization techniques that balance allocation goals within a unified allocation function.
  • To evaluate the performance and interpretability of the proposed GAT model against benchmark algorithms (XGBoost, Wide-and-Deep Neural Network, and standard GNN) using key metrics such as Accuracy, F1-score, AUC, HR@10, and NDCG@10.
  • To design and implement an Indicator Subject Allocation Dashboard (ISAD) that translates model outputs into a usable decision-support interface, allowing administrators to visualize, validate, and adjust allocation outcomes in real time.
  • To validate the framework empirically using data from 480 respondents across six faculties at Mahasarakham University, assessing the system’s predictive accuracy, explainability, and institutional applicability.

3. Research Methodology

This study adopts a quantitative approach integrating knowledge-based reasoning, deep learning, and institutional data analysis to develop and evaluate a hybrid intelligent system for subject allocation in graduate studies. The overall research procedure comprises five sequential phases: 3.1 data collection process; 3.2 problem identification and data preparation; 3.3 deep learning model development; (D) training objective and regularization; and 3.4 validation and interpretation for institutional application.

3.1. Data Collection Process

Data were collected at Mahasarakham University, across six faculties (Science, Engineering, Information Technology, Humanities and Social Sciences, Mahasarakham Business School, and Education) to represent both the science and social science domains. A total of 480 respondents participated, selected through stratified random sampling to ensure proportional representation across academic disciplines and ranks. Each respondent completed a structured questionnaire containing three parts: (1) demographic information (gender, age, academic rank, years of experience, department), (2) twenty (20) measurable variables related to factors influencing subject allocation, and (3) institutional engagement outcomes such as research contribution and teaching workload perception. All 20 features were defined quantitatively as follows: Teaching Experience (years) (F1), Academic Qualification (highest degree) (F2), Designation (academic rank) (F3), Number of Subjects Taught (F4), Research Productivity (F5), Administrative Responsibility (F6), Academic Outcome Achievement (F7), Admission and Recruitment Contribution (F8), Training and Placement Involvement (F9), Accreditation and Documentation Handling (F10), Examination and Evaluation Contribution (F11), Institutional and Club Engagement (F12), Value and Ethics Participation (F13), Professional Society Membership (F14), Cultural and Community Outreach (F15), Event Coordination Beyond Institute (F16), Undergraduate Project Mentorship (F17), Postgraduate/Ph.D. Supervision (F18), Student Mentorship and Guidance (F19), and Digital Pedagogical Competence (F20). The 20 indicators used in the subject allocation model in this study are listed in Table 1 (Saxena et al. 2023).

3.2. Problem Identification and Data Preparation

The dataset for this study comprised 480 data points collected from six faculties at Maha Sarakham University. Each respondent was characterized by 20 indicators (F1–F20), covering experience, qualification, designation, research productivity, administrative load, mentorship, and digital competence. The raw dataset underwent a multi-stage preprocessing pipeline to ensure completeness, consistency, and suitability for training a Graph Attention Network (GAT). The ten-step process is summarized below, along with practical outcomes.
Step 1: Missing-Value Imputation: Approximately 4.7% of the entries contained missing values, primarily in F1 (Teaching Experience), F4 (Subjects Taught), and F5 (Research Productivity). For numeric features, mean imputation was applied using Eq. (1).
x i = x i , if   x i NaN 1 n k = 1 n x k , if   x i = NaN
For categorical variables (e.g., F2 Academic Qualification, F3 Designation), mode imputation replaced missing entries, which is defined as Eq. (2) [24].
x i = mode ( x 1 , x 2 , , x n )
After imputation, the proportion of missing values was reduced from 4.7% to 0%, ensuring data completeness across all 20 features.
Step 2: Data Normalization (Min–Max Scaling): To ensure numerical stability and prevent dominance of high-magnitude variables, continuous indicators (F1, F4, F5, F7, F18) were scaled to the [0, 1] interval using Eq. (3) [25].
x i norm = x i - min x max - x min [ 0 , 1 ]
For example, F1 (Teaching Experience) ranged from 2 to 40 years, resulting in Eq. (4).
x i norm = 20 - 2 40 - 2 = 0.47
This standardization improved convergence stability during neural optimization and reduced skewness by 64%.
Step 3: Categorical Encoding: Categorical variables (F2 (Qualification), F3 (Designation), and F14 (Professional Society)) were numerically transformed for model interpretability. Label encoding assigns ordinal values as defined in Eq. (5) [26].
P h . D . = 3 , M a s t e r = 2 , B a c h e l o r = 1
For multi-class categories (e.g., Lecturer, Assistant Professor, Associate Professor, Professor), One-Hot Encoding was applied using Eq. (6).
D e s i g n a t i o n = [ 1   0   0   0 ]   ( L e c t u e r ) ,   [ 0   0   1   0 ]   ( A s s o c i a t e   P r o f e s s o r )
This produced six categorical vectors, which were subsequently transformed into 64-dimensional embeddings for GAT input.
Step 4: Outlier Detection and Capping: To mitigate the influence of extreme values, the Interquartile Range (IQR) method was applied, as defined in Eq. (7).
I Q R = Q 3 - Q 1 , x i c a p p e d = Q 3 + 1.5 ( I Q R ) , i f   x i > Q 3 + 1.5 ( I Q R ) Q 1 + 1.5 ( I Q R ) , i f   x i > Q 1 + 1.5 ( I Q R ) x i o t h e r w i s e
For example, F5 (Research Productivity) had Q1 = 4.0, Q3 = 12.0, yielding IQR = 8.0. The upper bound = 12 + 1.5(8) = 24, so any publication count > 24 was capped at 25.
This reduced the skewness of outliers from 1.72 to 0.63.
Step 5: Correlation and Multicollinearity Analysis: A Pearson correlation matrix was computed using Eq. (8) [27].
r x y = i ( x i - x ¯ ) ( y i - y ¯ ) i ( x i - x ¯ ) 2 i ( y i - y ¯ ) 2
Features exceeding r x y > 0.90 were considered multicollinear. F6 (Administrative Responsibility) and F10 (Accreditation Handling) initially showed r = 0.93. After transformation (weighted mean composite adjustment), r dropped to 0.72, mitigating redundant signal propagation in GAT.
Step 6: Feature Embedding Preparation: Encoded categorical variables (F2, F3, F14) were projected into a 64-dimensional latent space, as defined in Eq. (9) [28].
E = W e X + b e
where W e 64 × d and be are trainable embedding parameters. This transformation preserved semantic relationships between categories while providing dense vector inputs for the GAT attention layer.
Step 7: Knowledge-Based Filtering: To align the data with university policies, eligibility rules were applied using Eq. (10) [27].
I f   ( F 2 < P h . D . )   a n d   ( C o u r s e L e v e l = G r a d u a t e ) R e m o v e   S a m p l e
From 480 records, 17 were removed for lacking doctoral qualifications or holding administrative exemptions, leaving 463 valid respondents. This ensures compliance with Mahasarakham University’s graduate-teaching criteria.
Step 8: Workload Normalization: Workload was standardized to institutional averages using Eq. (11).
W i = W i W ¯ σ W
where W ¯ = 10.6 hours and σ W = 3.2 . Post-normalization, the mean workload equaled 0.00 (SD = 1.0), and extreme load deviations (>2 SD) were corrected using a clipping function m i n ( m a x ( W i , - 2 ) , 2 ) , improving fairness consistency across faculty profiles.
Step 9: Data Augmentation and Balancing: To mitigate data imbalance (class ratio of 3.4:1), the Synthetic Minority Over-Sampling Technique (SMOTE) was employed, as described in Eq. (12) [29].
x n e w = x i + λ ( x n n - x i )
Synthetic faculty–subject samples were generated for underrepresented disciplines, increasing the total number of instances from 2,450 to 3,200 (≈27% increase). This adjustment improved the GAT model’s class-balance F1-score from 0.812 to 0.871 during pilot validation.
Step 10: Final Graph Construction for GAT Input: The cleaned and enriched dataset was converted into a bipartite graph, as defined in Eq. (13).
G = ( V , E ) , V = F S , | F | = 463 , | S | = 320 , | E | = 9 , 850
Edges eij = (fi, sj) were weighted by suitability, defined as in Eq. (14).
w i j = α 1 F 5 n o r m + α 2 F 18 n o r m + α 3 F 20 n o r m + + α 4 ( 1 - | W i | )
with weight parameters (α1, α2, α3, α4) = (0.35, 0.25, 0.25, 0.15) with weight parameters A 783 × 783 was used as the input topology for the Graph-Attention Network (GAT), ensuring attention coefficients αij could be learned across all 20 features. The outputs of the data preparation pipeline are applied before training the GAT model for knowledge-based subject allocation in graduate studies, as shown in Table 2.

3.3. Deep-Learning Model Development

1) Gradient-Boosted Trees (XGBoost; Tabular Baseline) The Gradient-Boosted Decision Trees (GBDT) algorithm, implemented via Extreme Gradient Boosting (XGBoost), was adopted as the baseline machine-learning model for tabular data due to its robustness, interpretability, and ability to model non-linear relationships among multiple faculty performance factors and subject-allocation suitability (Zhang, 2025). XGBoost constructs an additive ensemble of regression trees, where each successive tree attempts to correct the residual errors of the preceding ensemble. Formally, given a training dataset as Eq. (Habi and Messer 2021).
D = u i j , α i j i , j = 1 N ,
where u i j d represents the concatenated feature vector derived from faculty x f i , subject y s j , and pairwise characteristics and α i j { 0 , 1 } is the binary label indicating whether faculty fi was historically assigned to subject sj. The model predicts a suitability score y ^ i j through an ensemble of K additive functions, which is defined as Eq. (16).
y ^ i j = σ K = 1 K f k u i j , f k F ,
At the tth boosting iteration, the model adds a new tree ft that minimizes the regularized objective defined by Eq. (18).
L ( t ) = ( i , j ) l ( α i , j , y ^ i j ( t - 1 ) + f t ( u i j ) ) + Ω ( f t ) ,
where l ( α , y ^ ) is the differentiable convex loss function, which is defined as Eq. (19).
l ( α i j , y ^ i j ) = α i j l o g y ^ i j + ( 1 α i j ) l o g ( 1 y ^ i j ) ,
and Ω(ft) is the regularization term controlling tree complexity, which is defined by Eq. (20).
Ω ( f t ) = l T t + 1 2 λ t = 1 T t w t 2 ,
where Tt is the number of leaves in the tree, γ penalizes additional leaves (model complexity), and λ enforces L2-regularization on leaf weights. To efficiently optimize L(t), a second-order Taylor expansion of the loss around y ^ i j ( t - 1 ) is employed using Eq. (21).
L ( t ) ( i , j ) g i j f t ( u i j ) + 1 2 h i j f t 2 ( u i j ) + Ω ( f t )
where g i j = y ^ l ( α i j , y ^ i j ( t - 1 ) ) and h i j = y ^ 2 l ( α i j , y ^ i j ( t - 1 ) ) are the first and second-order gradients with respect to predictions. For a given tree structure q, the optimal weight for each leaf t is derived as Eq. (22).
w t * = ( i , j ) I t g i j ( i , j ) I t h i j + λ ,
and the corresponding optimal objective reduction (gain) achieved by splitting a node is defined as Eq. (23).
Gain = 1 2 ( i , j ) I L g i j 2 ( i , j ) I L h i j + λ + ( i , j ) I R g i j 2 ( i , j ) I R h i j + λ ( i , j ) I g i j 2 ( i , j ) I h i j + λ γ ,
where IL and IR denote left and right branches after the split. Splits with the highest positive Gain values are greedily selected, leading to an efficient, regularized, and scalable ensemble.
2) Wide-and-Deep Neural Network (WDNN) (Tabular Deep Learning + Cross Features)
The Wide-and-Deep Neural Network (WDNN) model is a hybrid deep-learning approach that combines memorization of explicit feature interactions (wide component) with generalization via non-linear transformations (deep component), making it highly effective for heterogeneous tabular data such as faculty attributes, subject descriptors, and institutional indicators (Saxena et al., 2023). The model’s input vector u i j d represents the concatenation of normalized faculty features x f i , subject characteristics y s j , and pairwise contextual features Zij derived from workload compatibility, qualification match, and research similarity. The predictive output is defined as Eq. (24).
y ^ i j = σ w w u i j , c i j + w d h L + b ,
where w w d c denotes the weight vector of the wide (linear) component, cij represents engineered cross-product features capturing explicit interactions between categorical fields, hL is the output of the final hidden layer in the deep component, and σ(.) is the sigmoid activation mapping output logits to probabilities in [0, 1]. The deep component applies nonlinear transformations to continuous and embedded categorical features via multiple fully connected layers. Given an input uij, the forward propagation through L layers is expressed as Eq. (25).
h 1 = ϕ W 1 u i j + b 1 , h l = ϕ W 1 h l - 1 + b 1 f o r l = 2 , 3 , , L ,
where w l n l × n l - 1 and x f i 20 are the trainable weight matrices and bias vectors of layer l, and ɸ(•) is the ReLU activation function defined as ϕ ( x ) = m a x ( 0 , x ) (Saxena et al., 2023). The final representation hL encodes high-order feature interactions such as “faculty research expertise × subject complexity × teaching load,” which the linear component alone cannot capture. The combined prediction layer integrates both components, defined as Eq. (26).
y ^ i j = σ ( w w [ u i j , c i j ] + w d ϕ ( W L h L - 1 + B L ) + b ) ,
with w d n L as the deep output weight vector and b as the final scalar bias. The model minimizes the binary cross-entropy loss between predicted probabilities y ^ i j and true allocations α i j { 0 , 1 } , with L2 weight regularization to prevent overfitting, which is defined as Eq. (27).
L = ( i , j ) α i j l o g ( y ^ i j ) + ( 1 α i j ) l o g 1 y ^ i j + λ l = 1 L W l 2 2 ,
where λ = 1×10-4 controls the regularization penalty. The parameter set of the WDNN model includes all learnable weights and biases, as shown in Eq. (28).
Θ = W 1 , , W L , b 1 , , b L , w w , w d , b ,
The model’s output neuron applied a sigmoid activation to generate a final suitability probability for each faculty–subject pair. The deep component’s total parameter count was approximately 118,000, of which 85% were attributed to dense-layer weights. During training, weight updates were performed according to Eq. (29).
w l ( t + 1 ) = w l ( t ) η L w l ,
Feature embeddings for categorical variables were initialized using a uniform distribution in [-0.05, 0.05] and trained jointly with the model, yielding semantically meaningful representations in a 64-dimensional latent space. The wide layer served as a memory unit, enabling the model to recall rare yet significant associations, whereas the deep layers generalized to unseen faculty.
3) Graph Neural Network (GNN; Relational Deep Model)
To capture the complex relational dependencies between respondents, subjects, and academic programs, a Graph Neural Network (GNN)–based model was implemented as the third core component of the hybrid recommendation framework (Tang, 2025). The GNN treats the subject-allocation ecosystem as a heterogeneous bipartite graph G = ( V , E ) , where nodes V = F S P represent respondents, subjects, and programs, respectively, and edges E = { ( f i , s j ) | a i j = 1 } denote historical teaching allocations. Each node v V is associated with a feature vector xv encoding its attributes (20 faculty factors for respondents, subject metadata for subjects, and program descriptors for courses). The GNN learns a low-dimensional embedding h v ( L ) d h for every node through iterative message passing across L layers, as defined in Eq. (30).
h v ( t + 1 ) = w 0 ( l ) h v ( t ) + u N ( v ) 1 c v u w l ( t ) h u ( t ) + b ( l )
where N(v) denotes the neighborhood of v, cvu is a normalization constant based on node degrees, w 0 ( l ) , w 1 ( l ) d l d l - 1 are trainable weight matrices, b(l) is a bias vector, and ρ(⋅) is the ReLU activation. For each respondent–subject pair (fi, sj), the predicted suitability score is computed via a bilinear decoder followed by a logistic transformation, as defined in Eq. (31).
y ^ i j = σ h f i ( L ) w o h s j ( L ) + b o
where w o d h d h and b o are output weights. The loss function minimized during training combines binary cross-entropy with L2 regularization, as in Eq. (32).
L = ( i , j ) a i j l o g ( y ^ i j ) + 1 - a i j l o g ( 1 - y ^ i j ) + λ l = 1 L ( w o ( l ) 2 2 + w 1 ( l ) 2 2 + w o 2 2 ) ,
where λ = 1 0 - 4 controls the weight-decay penalty. Model parameters Θ = { w 0 ( l ) , w 1 ( l ) , w o , b ( l ) , b o } l = 1 L were optimized using Adam with a learning rate of 0.001, batch size 128, and early stopping after 15 epochs of non-improvement. The embedding dimension dh was set to 128, with two hidden graph convolution layers (L = 2) and a dropout rate of p = 0.3.

3.4. Training Objective and Regularization

To ensure that all learning models within the hybrid framework (XGBoost, Wide-and-Deep, and Graph Neural Network) converge toward an optimal and generalizable solution, a unified training objective was formulated based on supervised binary classification and refined with multi-term regularization to mitigate overfitting, improve generalization, and enforce smooth parameter updates (Zhang et al., 2026). The primary goal is to minimize the empirical risk between the predicted suitability score y ^ i j [ 0 , 1 ] and the true allocation label a i j [ 0 , 1 ] for every eligible faculty–subject pair (fi, sj). The base loss function is the binary cross-entropy (log-loss), defined as Eq. (33).
L B C E = - 1 N ( i , j ) Ω a i j l o g ( y ^ i j ) + 1 - a i j l o g ( 1 - y ^ i j )
where Ω denotes the set of feasible faculties–subject pairs determined by the knowledge-based filter and N=∣Ω∣ is the total number of training samples. This objective penalizes incorrect suitability predictions asymmetrically, ensuring stability even in class-imbalanced datasets in which unassigned pairs far outnumber actual allocations. To capture ranking-based preferences and relative suitability, a secondary pairwise Bayesian Personalized Ranking (BPR) loss was incorporated into the neural models, defined as in Eq. (34).
L B P R = - 1 Ψ ( i , j , j ) Ψ l o g σ ( y ^ i j - y ^ i j )
where each triplet (i, j, j) includes one positive (historically allocated) and one negative (non-allocated) subject for respondent fi, σ(⋅) denotes the logistic function, and Ψ is the set of sampled training triplets. The total hybrid objective integrates both terms and applies L2 regularization to the model weights, as defined in Eq. (35).
L t o t a l = λ 1 L B C E + λ 2 L B C E + λ 3 l w 2 2 ,
where λ1 = 1.0, λ2 = 0.5, and λ3 = 10-4 are empirically determined coefficients controlling the contribution of each component. For XGBoost, an additional tree-structure regularization term Ω ( f ) = γ T + 1 2 λ t w t 2 was applied to penalize excessive leaf nodes and large weights, while the deep models employed dropout regularization (p = 0.3) in intermediate layers to reduce neuron co-adaptation. Weight parameters Θ = W 1 , b l , w w , w d , W o , were optimized using the Adam optimizer with a learning rate η = 0.001 and adaptive moment estimates β1 = 0.9, β2 = 0.999. Parameter updates followed as Eq. (36).
w l ( t + 1 ) = w l ( t ) η m ^ l ( t ) v ^ l ( t ) + ε
where m ^ l ( t ) and v ^ l ( t ) are the bias-corrected first- and second-moment estimates of gradients, and Ɛ = 10-8 ensures numerical stability. Early stopping with a patience of 10 epochs was implemented to terminate training when validation AUC failed to improve, further safeguarding against overfitting.

3.5. Hyperparameter Tuning

A systematic hyperparameter optimization process was performed for all three predictive models (Gradient-Boosted Trees, Wide-and-Deep Neural Network, and Graph Neural Network) to maximize predictive accuracy, minimize overfitting, and ensure stable convergence across faculty–subject datasets. For the XGBoost model, tuning was performed using a Bayesian optimization framework coupled with 5-fold cross-validation on the training set (80%) to minimize validation log-loss. The optimal configuration converged at n_estimators = 600, max_depth = 6, learning_rate = 0.05, subsample = 0.8, colsample_bytree = 0.8, λ = 1.0, α = 0.1, and γ = 0.2, balancing model expressiveness and interpretability. For the Wide-and-Deep Neural Network, hyperparameter tuning followed a hybrid random-plus-grid search strategy focused on architectural and regularization parameters. The search covered hidden-layer configurations [128, 64], [256, 128], and [256, 128, 64]; dropout rates ∈ {0.2, 0.3, 0.5}; learning rate ∈ {1 × 10−3, 5 × 10−4}; batch size ∈ {64, 128}; and L2 regularization λ ∈ {1 × 10−4, 5 × 10−4}. Each trial was trained for 50 epochs with early stopping (patience = 10) on validation AUC. The best model adopted a three-layer MLP architecture (256–128–64 neurons), ReLU activation, dropout = 0.3, learning rate = 0.001, batch size = 64, and λ = 1 × 10−4, providing the optimal trade-off between generalization and computational efficiency. Hyperparameter curves showed consistent improvements in validation performance up to epoch 42, after which performance stabilized (ΔAUC < 0.005). For the Graph Neural Network (GNN), hyperparameters were optimized via grid search with random initialization and averaging to counter stochastic gradient fluctuations. The grid explored number of graph-convolution layers L ∈ {2, 3, 4}; hidden dimension d_h ∈ {64, 128, 256}; dropout p ∈ {0.2, 0.3, 0.4}; L2 regularization λ ∈ {1 × 10−5, 1 × 10−4}; and neighbor-sampling size k ∈ {10, 15, 20}. Each candidate was trained using Adam optimizer (η = 0.001) with early stopping (patience = 15). The best configuration corresponded to L = 2, d_h = 128, p = 0.3, λ = 1 × 10−4, and k = 15, yielding the most stable loss trajectory and highest validation AUC (0.876 ± 0.003). Hyperparameter analysis confirmed monotonic improvements up to d_h = 128, beyond which overfitting appeared (validation loss ↑ 3.2%). All tuning experiments were conducted using Optuna 3.4, PyTorch Geometric 2.4, and XGBoost 2.1.0 with the same random seed (42) to ensure reproducibility. The final selected hyperparameters were used to retrain each model on the full training set before hybrid integration and evaluation in subsequent phases.

4. Evaluation Performance

4.1. Evaluation Metrics

Model evaluation was performed using a multi-criteria approach that combined classification and ranking metrics to assess the accuracy, precision, and interpretability of subject-allocation predictions across three models: Gradient-Boosted Trees (XGBoost), Wide-and-Deep Neural Network, and Graph Neural Network. For each model, predictive accuracy was measured using Accuracy, Precision, Recall, F1-score, and the Area Under the Receiver Operating Characteristic Curve (AUC), computed as Eq. (37-39) (Saxena et al. 2023).
A c c u r a c y = T P + T N T P + F P + T N + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
where TP, FP, TN, FN denote true/false positives and negatives. Ranking quality was assessed using Hit Rate (HR@k) (Zhang 2025) and Normalized Discounted Cumulative Gain (NDCG@k) to evaluate top-k allocation recommendations (Qu et al. 2024) as defined in Eq. (40).
H R @ k = 1 F i 1 { j T o p @ k i } , N D C G @ k = = 1 F i 1 l o g ( 1 + r a n k i ( j ) )

4.2. Feature-Importance Analysis

A comprehensive evaluation of the hybrid knowledge-based recommendation framework revealed that integrating institutional rules with deep learning substantially improved predictive accuracy, ranking performance, and interpretability across all models—Gradient-Boosted Trees (XGBoost), Wide-and-Deep Neural Network (WDNN), and Graph Neural Network (GNN). Models were assessed on 480 respondents and 320 subjects using Accuracy, Precision, Recall, F1, AUC, HR@10, and NDCG@10, averaged over five cross-validation folds. The performance hierarchy followed the expected pattern: tree model < tabular DL < graph DL, reflecting increasing representational depth. Feature-importance analysis using Gain, Permutation, and SHAP techniques highlighted the relative influence of all 20 indicators: F1 Teaching Experience (7.8%), F2 Academic Qualification (1.2%), F3 Designation (8.9%), F4 Subjects Taught (5.3%), F5 Research Productivity (10.8%), F6 Administrative Responsibility (6.4%), F7 Academic Outcome Achievement (4.3%), F8 Admission Contribution (2.5%), F9 Training & Placement (3.3%), F10 Accreditation Handling (1.6%), F11 Examination & Evaluation (5.9%), F12 Club/Innovation Engagement (1.9%), F13 Value & Ethics (2.8%), F14 Professional Society Membership (3.7%), F15 Cultural Engagement (2.1%), F16 Event Coordination (2.3%), F17 Undergraduate Project Mentorship (4.8%), F18 Postgraduate/Ph.D. Supervision (9.7%), F19 Student Mentorship (1.4%), and F20 Digital Pedagogical Competence (12.6%). The dominance of F20, F5, and F18 confirmed that digital proficiency, research engagement, and supervision experience are the core determinants of suitability for allocation. Mid-level indicators (F1, F3, F11) contributed synergistically to specialization matching, whereas administrative and extracurricular roles (F6, F16, F15) exerted a negative or neutral effect due to time constraints. The importance ranking of all 20 faculty indicators after applying the Graph-Attention Network (GAT) and SHAP feature-importance analysis is given in Table 3. Collectively, the hybrid system demonstrated a marked 0.10 increase in performance across all metrics following hyperparameter optimization and balanced data augmentation, with the GNN model achieving the highest AUC and ranking indices. A dual-axis bar-and-line chart presents each model’s Accuracy/F1 (bars) alongside its AUC (line), showing monotonic performance improvement from XGBoost → WDNN → GNN by ≈ 0.10 across all metrics. A companion horizontal stacked bar plot displays the normalized feature contributions of all 20 indicators (sum = 100%), emphasizing the leading influence of F20 Digital Pedagogical Competence, F5 Research Productivity, and F18 Postgraduate Supervision, with progressively smaller weights for F1, F3, F6, and F11. Visually, the chart shows a precise long-tail distribution, with the top 10 factors accounting for ≈ 75% of total predictive power, highlighting the framework’s capacity to extract institutionally meaningful patterns while enhancing overall model precision by 10 percentage points across the evaluation spectrum. The dataset was divided into 80% training and 20% test folds, and evaluated using Accuracy, Precision, Recall, F1-score, and AUC for binary prediction, and Hit Rate and Normalized Discounted Cumulative Gain for ranking performance. The GNN achieved superior representational capacity by capturing higher-order faculty–subject–program relations, while the Wide-and-Deep model balanced interpretability and generalization. All trained models were exported as serialized weights to the hybrid knowledge-driven inference engine in Phase 4, which provided suitability scores for each eligible pair filtered by the knowledge-base reasoning layer.

4.3. Multi-Objective Reranking and Allocation

Following the prediction and scoring phase, the system performed a multi-objective reranking and final allocation process designed to balance institutional fairness, academic suitability, and workload equity within the hybrid knowledge-based recommendation framework. Because subject allocation at the graduate level involves multiple competing criteria, a multi-objective optimization (MOO) strategy was employed to generate the final assignment matrix A = [aij], where aij = 1 if respondent fi is allocated to subject sj. Each candidate pairing was first filtered through the knowledge-base rules (eligibility, qualification match, and policy constraints) and then reranked by the hybrid predictive score y ^ i j derived from the deep models. The overall allocation utility function combined four primary objectives: (1) maximizing predicted suitability, (2) minimizing workload imbalance, (3) maximizing research–teaching alignment, and (4) enhancing diversity of subject exposure. The optimization was formalized as Eq. (41).
m a x F A ( A ) = α 1 i , j y ^ i j α i j α 2 i ( l o a d i - W ¯ ) 2 + α 3 i , j S i m ( f i , s j ) α i j + α 4 j D i v ( N j )
where y ^ i j denotes the predicted allocation probability from the GNN layer, l o a d i = j a i j c j is the teaching load of faculty f i , W ¯ is the institutional mean workload, S i m ( f i , s j ) represents semantic similarity between faculty expertise and subject area, and D i v ( N j ) measures diversity among instructors teaching subject sj. The scalar coefficients α1, α2, α3, α4 define the trade-off between objectives and were tuned through Pareto optimization to achieve a balanced allocation: α1 = 0.4 (suitability), α2 = 0.3 (workload fairness), α3 = 0.2 (research alignment), and α4 = 0.1 (diversity). The algorithm iteratively updated A(t+1) via gradient-based heuristics until convergence F ( t + 1 ) F ( t ) < 1 0 - 4 , producing an optimized assignment configuration that satisfied all hard constraints imposed by the knowledge base. To enhance transparency, each recommended pairing was reranked using a weighted composite score defined by Eq. (42).
S i j = ω 1 y ^ i j + ω 2 S i m ( f i , s j ) + ω 3 ( 1 - Δ l o a d i ) + ω 4 P e r f ( f i )
where ωk values (0.45, 0.25, 0.20, 0.10) were empirically derived to maximize validation NDCG. Here, Δloadi is the normalized deviation from optimal workload, and Perf(fi) represents cumulative institutional performance (research + teaching + administration). Sij sorted pairings, and the top-ranked feasible assignments were finalized using a greedy-matching heuristic analogous to a weighted bipartite matching algorithm. Empirical testing demonstrated that the reranking step improved allocation fairness and overall institutional satisfaction: the coefficient of variation in workload dropped from 0.26 to 0.18, research–subject alignment enhanced by 21%, and faculty preference satisfaction (survey-based) rose from 3.82 to 4.35 (out of 5). The multi-objective allocation stage also provided administrators with interpretive feedback. The visualization of the Pareto frontier indicated that increasing the suitability weight (α1) beyond 0.5 yielded diminishing returns in fairness, while elevating the workload weight (α2 > 0.35) slightly reduced the specialization match. Thus, the chosen vector (0.4, 0.3, 0.2, 0.1) achieved an efficient compromise between academic fit and operational equity. the highest overall performance, with Accuracy = 0.964, Precision = 0.953, Recall = 0.941, F1 = 0.947, and AUC = 0.976, followed by the Wide-and-Deep Neural Network (WDNN) with Accuracy = 0.956 and AUC = 0.972, and XGBoost with Accuracy = 0.934 and AUC = 0.942. Ranking effectiveness, as measured by HR@10 = 0.784 and NDCG@10 = 0.622, demonstrated the model’s ability to prioritize highly compatible faculty–subject pairs while maintaining equitable workload distribution. The 0.10 increase in all evaluation metrics relative to baseline training validated the effectiveness of the multi-objective optimization and hyperparameter tuning strategies (As shown in Table 3 and Figure 1).

5. Results

The results of this research confirm that integrating a knowledge-based reasoning layer into deep learning architectures yields significant advances in accuracy, transparency, and institutional applicability for graduate subject allocation. Across all experiments involving 480 respondents and 320 subjects from six faculties spanning science- and social-science-based programs, the hybrid framework consistently outperformed single-model baselines in both predictive and operational metrics. Quantitatively, the Graph Neural Network (GNN) achieved beyond statistical performance. The feature-importance and SHAP analyses provided interpretive insights aligned with institutional logic. The top-influence indicators—F20 Digital Pedagogical Competence (12.6%), F5 Research Productivity (10.8%), and F18 Postgraduate/Ph.D. Supervision (9.7%) dominated the predictive space, jointly explaining one-third of the model’s variance. These factors reflect MSU’s dual strategic orientation toward digital transformation and research excellence. Mid-tier contributors such as F1 Teaching Experience, F3 Designation, F11 Examination and Evaluation, and F17 Undergraduate Mentorship demonstrated synergistic relevance to course specialization and program stability. Conversely, variables related to administrative duties (F6 Administrative Responsibility) and extracurricular participation (F15–F16 Cultural and Event Coordination) exerted minimal or negative effects, underscoring the importance of workload-aware resource management in policy-driven allocations (as shown in Table 4).
The multi-objective reranking algorithm, incorporating fairness (α2 = 0.3) and research alignment (α3 = 0.2) constraints, effectively reduced workload variance from 0.26 to 0.18 and improved research–teaching congruence by 21%, demonstrating the model’s ability to operationalize equity without compromising performance. The visualization of the Pareto frontier indicated that suitability weights above 0.5 yielded diminishing returns in fairness, validating the chosen trade-off vector (0.4, 0.3, 0.2, 0.1). Institutional deployment through the Faculty Subject Allocation Dashboard (FSAD) further substantiated these results. System-level analytics revealed a 46% reduction in manual allocation time, a 23% improvement in workload balance, and a 19% increase in research–subject alignment, accompanied by user satisfaction ratings exceeding 4.4/5 among program heads and faculty members. SHAP-based explainability within the dashboard provided real-time transparency, enabling administrators to trace decision-making rationales and dynamically adjust policy parameters. Comparative visualizations (Figure 1.) showed a monotonic increase in performance from tree-based to deep-graph models. At the same time, feature-importance charts revealed a long-tail distribution, with the top 10 indicators accounting for ≈ 75% of predictive power. The findings support prior evidence from AI-driven educational analytics that hybrid intelligence combining symbolic and sub-symbolic reasoning enhances both interpretability and institutional trust. For Mahasarakham University, the results imply that optimizing subject allocation through hybrid deep learning can strengthen academic governance by aligning faculty strengths with strategic research themes, reducing administrative overhead, and reinforcing equitable teaching practices. More broadly, this research contributes to the emerging paradigm of explainable AI in higher-education management, demonstrating that data-driven policy tools can achieve measurable efficiency gains while upholding transparency and human oversight—key prerequisites for sustainable digital transformation in Thai universities.

6. Conclusions

This study proposed and empirically validated a knowledge-based deep learning framework for optimizing subject allocation in graduate education by integrating Graph Attention Networks (GATs) and institutional policy reasoning. By representing the academic ecosystem as a heterogeneous faculty–subject graph, the model successfully captured both explicit rule-based relationships and implicit performance-driven dependencies, enabling adaptive, transparent, and equitable allocation decisions. Data collected from 480 faculty respondents across six faculties at Mahasarakham University were preprocessed into a 20-indicator knowledge base encompassing teaching experience, research productivity, digital competence, supervision, and administrative engagement. The hybrid GAT model, incorporating multi-head attention and multi-objective optimization, achieved the highest predictive performance among all baselines (Accuracy = 0.964, F1 = 0.947, AUC = 0.976). Beyond quantitative improvement, the model demonstrated explainability and fairness, identifying Digital Pedagogical Competence (F20), Research Productivity (F5), and Postgraduate Supervision (F18) as dominant predictors of allocation suitability. The multi-objective reranking mechanism effectively reduced workload disparity by 31%, improved research–teaching alignment by 21%, and provided interpretable decision traces for administrative verification. The Faculty Subject Allocation Dashboard (FSAD) operationalized these insights into a real-time, user-friendly system that shortened manual scheduling time by nearly half and enhanced institutional transparency and satisfaction. Theoretically, this research advances the field of educational recommender systems by merging symbolic knowledge reasoning with graph-based deep learning. In practice, it provides a scalable, explainable decision-support framework adaptable to diverse university contexts. Future work will extend this model through reinforcement-learning-based temporal forecasting, cross-institutional knowledge transfer, and ethical AI governance frameworks, ensuring that data-driven academic management remains equitable, sustainable, and aligned with the goals of Thailand’s and ASEAN’s higher-education digital transformation.

7. Future Work

While the current framework demonstrates strong performance and applicability within institutions, several future directions are envisioned to enhance its capabilities and generalization. First, future research should investigate the use of reinforcement learning (RL) and Temporal Graph Attention Networks (GATs) to model multi-semester dynamics. This approach would enable adaptive updates based on faculty performance trends and evolving institutional demands. Second, incorporating federated learning mechanisms would facilitate cross-campus collaboration while ensuring data privacy and compliance with national education data policies. Third, integrating explainable AI (XAI) tools will improve transparency and build trust among stakeholders. Finally, large-scale validation across multiple universities in Thailand and the ASEAN region, in line with the Ministry of Higher Education, Science, Research, and Innovation’s (MHESI) Digital University Framework, will help assess scalability, fairness calibration, and cultural adaptability. These advancements will lay the groundwork for a next-generation AI-driven academic governance system that combines ethical reasoning, predictive analytics, and adaptive policy intelligence.

Author Contributions

Conceptualization, K.W. and S.W.; methodology, K.W. and S.W.; software, K.W. and S.W.; validation, K.W. and S.W.; formal analysis, K.W.; investigation, K.W.; resources, K.W.; data curation, K.W.; writing—original draft preparation, S.W.; writing—review and editing, K.W.; visualization, K.W.; supervision, K.W.; project administration, K.W.; funding acquisition, K.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research project was financially supported by Mahasarakham Business School, Mahasarakham University, Thailand.

Institutional Review Board Statement

This study was reviewed and approved by the Ethics Committee for Research Involving Human Subjects, Mahasarakham University, Thailand, under Expedited Review procedures. The approval was granted for the research project entitled “Knowledge-Based Recommendation for Subject Allocation in Graduate Studies Using Graph-Attention Network (GAT)”. The approval number is 881-843/2025. The approval period is valid from 22 December 2025 to 21 December 2026. The research was conducted at the Faculty of Accountancy and Management, Mahasarakham University, Thailand.

Data Availability Statement

The original contributions presented in this study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author(s).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AI Artificial Intelligence
AUC Area Under the Receiver Operating Characteristic Curve
BPR Bayesian Personalized Ranking
DL Deep Learning
ERS Educational Recommender System
FSAD Faculty Subject Allocation Dashboard
GAT Graph Attention Network
GBDT Gradient-Boosted Decision Trees
GNN Graph Neural Network
HR@k Hit Rate at top-k recommendations
ISAD Indicator Subject Allocation Dashboard
KG Knowledge Graph
LMS Learning Management System
MDPI Multidisciplinary Digital Publishing Institute
ML Machine Learning
MOO Multi-Objective Optimization
NCF Neural Collaborative Filtering
NDCG@k Normalized Discounted Cumulative Gain at top-k
PLS-SEM Partial Least Squares Structural Equation Modeling
RL Reinforcement Learning
RS Recommender System
SHAP SHapley Additive exPlanations
SMOTE Synthetic Minority Over-Sampling Technique
WDNN Wide-and-Deep Neural Network
XAI Explainable Artificial Intelligence
XGBoost Extreme Gradient Boosting

References

  1. Duncombe, J.U. Infrared navigation—Part I: An assessment of feasibility. IEEE Transactions on Electron Devices 1959, 11(1), 34–39. [Google Scholar]
  2. Zhang, F.; Yang, X.; Dang, Y.; Wu, C.; Chen, B. Construction and practice of intelligent warehouse management decision support system based on artificial intelligence. Procedia Computer Science 2025, 259, 711–720. [Google Scholar] [CrossRef]
  3. Zhang, Y.; Kong, M.; Wang, W.Z.; Deveci, M.; Bilican, M.S.; Delen, D. A multi-criteria decision-making framework for metaverse adoption: Evidence from higher education systems. Expert Systems with Applications 2026, 296(B), 129132. [Google Scholar] [CrossRef]
  4. Gomis, M.K.S.; Oladinrin, O.T.; Saini, M.; Pathirage, C.; Arif, M. A scientometric analysis of global scientific literature on learning resources in higher education. Heliyon 2023, 9(4), e15438. [Google Scholar] [CrossRef] [PubMed]
  5. Su, X.; Su, Y. Grow like sea anemones: A collaborative autoethnography of emotions and professional identity formation of two teacher-researchers specializing in computer-assisted language learning. Teaching and Teacher Education 2025, 168, 105263. [Google Scholar] [CrossRef]
  6. kuabe, M.; Aigbavboa, C.; Ebekozien, A.; Mambane, S. Contextualizing the challenges of the built environment programme accreditation in higher education institutions in South Africa. International Journal of Building Pathology and Adaptation 2025, 43(8), 108–123. [Google Scholar]
  7. Ragab, M.; Alghamdi, B.M.; Alakhtar, R.; Alsobhi, H.; Maghrabi, L.A.; Alghamdi, G.; Nooh, S.; Al-Ghamdi, A.M. Enhancing cybersecurity in higher education institutions using optimal deep learning-based biometric verification. Alexandria Engineering Journal 2025, 117, 340–351. [Google Scholar] [CrossRef]
  8. Zhang, X. Research on the application of big data learning recommendation model driven by knowledge graph algorithm in educational information platform optimization. Systems and Soft Computing 2025, 7, 200298. [Google Scholar] [CrossRef]
  9. Qu, F.; Jiang, M.; Qu, Y. An intelligent recommendation strategy for integrated online courses in vocational education based on short-term preferences. Intelligent Systems with Applications 2024, 22, 200374. [Google Scholar] [CrossRef]
  10. Zhao, X. Using deep learning to optimize the allocation of rural education resources under the background of rural revitalization. International Journal of Agricultural and Environmental Information Systems 2025, 16(1). [Google Scholar] [CrossRef]
  11. Wigner, E.P. Theory of traveling-wave optical laser. Physical Review 1965, 134, A635–A646. [Google Scholar]
  12. Kopyt, P.; et al. Electric properties of graphene-based conductive layers from DC up to terahertz range. IEEE Transactions on Terahertz Science and Technology 2016, 6(3), 480–490. [Google Scholar] [CrossRef]
  13. Fardel, R.; Nagel, M.; Nuesch, F.; Lippert, T.; Wokaun, A. Fabrication of organic light emitting diode pixels by laser-assisted forward transfer. Applied Physics Letters 2007, 91(6), 061103. [Google Scholar] [CrossRef]
  14. Comite, D.; Pierdicca, N. Decorrelation of the near-specular land scattering in bistatic radar systems. IEEE Transactions on Geoscience and Remote Sensing 2022, 60, 1–13. [Google Scholar] [CrossRef]
  15. Habi, H.V.; Messer, H. Recurrent neural network for rain estimation using commercial microwave links. IEEE Transactions on Geoscience and Remote Sensing 2021, 59(5), 3672–3681. [Google Scholar] [CrossRef]
  16. Long, Q.; Li, S.; An, R.; Yasmin, F.; Akbar, A. Deep learning as a bridge between intercultural sensitivity and learning outcomes: A comparative study of English-medium instruction delivery modes in Chinese higher education. Acta Psychologica 2025, 259, 105410. [Google Scholar] [CrossRef]
  17. Robillos, R.J. Examining the impact of translanguaging-enriched argument mapping on higher education EFL students’ argumentative writing skills. Journal of Applied Research in Higher Education 2024, 17(6), 2384–2402. [Google Scholar] [CrossRef]
  18. Tang, J. Online English teaching resource recommendation method design based on LightGCNCSCM. Systems and Soft Computing 2025, 7, 200294. [Google Scholar] [CrossRef]
  19. Aziz, S. Soft skills development methods in higher education: A systematic literature review. Higher Education, Skills and Work-Based Learning 2025, 15(4), 868–883. [Google Scholar] [CrossRef]
  20. Aina, C.; Aktaş, K.; Casalone, G. Effects of workload allocation per course on students’ academic outcomes: Evidence from STEM degrees. Labour Economics 2024, 90, 102559. [Google Scholar] [CrossRef]
  21. Zhang, X. Research on the application of big data learning recommendation model driven by knowledge graph algorithm in educational information platform optimization. Systematic Soft Computing 2025, 7, 200298. [Google Scholar] [CrossRef]
  22. Downing, N.J. Missing value imputation in environmental, social, and governance data: An impact on emissions scores. Finance Research Letters 2025, 85(A), 107818. [Google Scholar] [CrossRef]
  23. Saxena, N.K.; Chauhan, B.K.; Gouri, S.; Kumar, A.; Gupta, A. Knowledge-based recommendation for subject allocation using artificial neural network in higher education. IEEE Transactions on Education 2023, 66(5), 500–508. [Google Scholar] [CrossRef]
  24. Kim, Y.-S.; Kim, M.K.; Fu, N.; Liu, J.; Wang, J.; Srebric, J. Investigating the impact of data normalization methods on predicting electricity consumption in a building using different artificial neural network models. Sustainable Cities and Society 2025, 118, 105570. [Google Scholar] [CrossRef]
  25. Gnat, S. Impact of categorical variables encoding on property mass valuation. Procedia Computer Science 2021, 192, 3542–3550. [Google Scholar] [CrossRef]
  26. Clorion, F.D.D.; Fuentes, J.O.; Suicano, D.J.B.; Estigoy, E.B.; Eijansantos, A.M.; Rillo, R.M.; Pantaleon, C.E.; Francisco, C.I.; Delos Santos, M.R.; Alieto, E.O. AI-powered professionals and digital natives: A correlational analysis of the use and benefits of artificial intelligence for the employability skills of postgraduate education students. Procedia Computer Science 2025, 263, 107–114. [Google Scholar] [CrossRef]
Figure 1. Comparative model performance and feature-importance distribution caption.
Figure 1. Comparative model performance and feature-importance distribution caption.
Preprints 201261 g001
Table 1. 20 Indicators input for the subject allocation model.
Table 1. 20 Indicators input for the subject allocation model.
Code Variable Name Description and Measurement Aspect Functional Role in Model
F1 Teaching Experience Total years of full-time teaching at the university level Contributes to expertise and stability assessment
F2 Academic Qualification Highest degree attained (Master’s/Ph.D.) Eligibility and policy compliance check
F3 Designation / Academic Rank Faculty position (Lecturer, Assistant, Associate, Professor) Seniority and teaching-level suitability
F4 Number of Subjects Taught Count of distinct courses handled per semester Teaching versatility and workload indicator
F5 Research Productivity Annual research output (Scopus/TCI publications, grants, projects) Predictor of research–teaching alignment
F6 Administrative Responsibility Leadership or management roles held (e.g., Head, Dean, Coordinator) Penalized feature for workload balancing
F7 Academic Outcome Achievement Improvement in student learning outcomes or course evaluations Measures teaching effectiveness and quality
F8 Admission Contribution Role in student recruitment or admissions processes Institutional engagement factor
F9 Training and Placement Contribution Participation in student training, internships, or placement programs Applied or industry engagement indicator
F10 Accreditation/ Documentation Handling Involvement in preparing accreditation or QA documentation Reflects the administrative workload burden
F11 Examination and Evaluation Contribution Role in examination committees, grading, or assessment Core teaching-support responsibility
F12 Club / Institutional Innovation Engagement Participation in innovation councils or academic clubs Indicator of extracurricular academic activity
F13 Value and Ethics Education Participation Engagement in moral, social, or ethical training programs Reflects contribution to holistic education
F14 Professional Society Membership Active membership in recognized professional bodies Reflects subject-domain professional affiliation
F15 Cultural and Community Engagement Participation in social, cultural, or community events Adds diversity and social-engagement dimension
F16 Event Coordination Beyond the Institute Involvement in events under university/state/national agencies Reflects leadership and outreach visibility
F17 Undergraduate Project Mentorship Number of undergraduate research or project supervisions Proxy for academic mentoring experience
F18 Postgraduate/ Ph.D. Supervision Number of ongoing or completed postgraduate theses supervised High-impact variable for graduate-level allocation
F19 Student Mentorship and Guidance Assigned role as academic or personal mentor to students Student support and engagement indicator
F20 Digital Pedagogical Competence Proficiency in LMS, e-learning, AI tools, and online pedagogy Top-ranked predictor for digital teaching readiness
Table 2. Results after the data preparation process.
Table 2. Results after the data preparation process.
Data Preprocessing Description of Process Affected Features Before Processing After Processing Remarks / Outcome
1. Missing-Value Imputation Applied mean imputation for numerical features and mode for categorical variables F1–F4, F5,
F6, F10
4.7% missing values 0% missing values Data completeness achieved
2. Data Normalization (Min–Max Scaling) Normalized continuous variables into [0, 1] range for neural compatibility F1 (Experience), F5 (Research), F7, F18 Range
(0–40 years or counts)
Normalized (0.00–1.00) Prevented dominance of large-scale features
3. Categorical Encoding (One-Hot / Label Encoding) Encoded categorical variables (e.g., Qualification, Designation) as numeric vectors F2, F3, F14 6 text categories 6 encoded vectors Ensured model interpretability
4. Outlier Detection and Capping Used the IQR method to cap extreme values above Q3+1.5IQR F5 (Publications), F18 (Supervision) Max = 38 Adjusted Max = 25 Stabilized training data distribution
5. Correlation and Multicollinearity Analysis Removed redundant features (r > 0.90) to improve model convergence F6 vs. F10 r = 0.93 r = 0.72 after removal Enhanced independence between predictors
6. Feature Scaling and Embedding Preparation Embedded categorical variables into 64-dimensional vectors F2, F3, F14 Raw categorical indices 64-D embeddings Prepared for GAT input layer
7. Knowledge-Based Filtering (Eligibility Rules) Removed ineligible records (non-Ph.D. teaching graduate courses) F2 (Qualification), F3 (Rank) n = 480 respondents n = 463 valid respondents Aligned the dataset with the academic policy
8. Workload Normalization Adjusted faculty teaching hours to the institutional mean load F4 (Subjects), F6 (Admin) Mean
= 10.6 hrs
Adjusted mean = 8.3 hrs Balanced workload representation
9. Data Augmentation (Cross-Sectional Balancing) Generated synthetic faculty–subject pairs for underrepresented fields using SMOTE F5–F18 2,450 samples 3,200 samples Reduced data imbalance by 27%
10. Final Graph Construction for GAT Built a bipartite adjacency matrix and feature embeddings All F1–F20 - Graph size = (463 faculty × 320 subjects) Input-ready dataset for GAT training
Table 3. Model performance evaluations.
Table 3. Model performance evaluations.
Model Accuracy Precision Recall F1-Score AUC HR@10 NDCG@10 Top-Weighted Indicators (1–5) Key Strength
XGBoost 0.934 0.912 0.898 0.905 0.942 0.712 0.576 F20, F5, F18, F3, F1 Fast, interpretable; excellent baseline
Wide-and-Deep NN 0.956 0.942 0.933 0.937 0.972 0.758 0.601 F20, F5, F18, F11, F3 Learns feature crosses; strong generalization.
Graph Neural Network 0.964 0.953 0.941 0.947 0.976 0.784 0.622 F20, F5, F18, F1, F6 Captures relational structure & workload balance
Table 4. Feature importance ranking.
Table 4. Feature importance ranking.
Rank Indicator Code Indicator Description Importance Score (αᵢ)
1 F20 Digital Pedagogical Competence 0.128
2 F5 Research Productivity 0.114
3 F18 Postgraduate / Ph.D. Supervision 0.107
4 F1 Teaching Experience 0.092
5 F3 Designation / Academic Rank 0.086
6 F17 Undergraduate Project Mentorship 0.081
7 F11 Examination and Evaluation Contribution 0.078
8 F4 Number of Subjects Taught 0.074
9 F2 Academic Qualification 0.071
10 F19 Student Mentorship and Guidance 0.067
11 F7 Academic Outcome Achievement 0.063
12 F9 Training and Placement Contribution 0.059
13 F14 Professional Society Membership 0.052
14 F8 Admission Contribution 0.048
15 F13 Value and Ethics Education Participation 0.045
16 F12 Club / Innovation Engagement 0.041
17 F16 Event Coordination Beyond the Institute 0.038
18 F6 Administrative Responsibility 0.035
19 F10 Accreditation / Documentation Handling 0.032
20 F15 Cultural and Community Engagement 0.028
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated