Preprint
Article

This version is not peer-reviewed.

Towards Reliable LLM Grading Through Self-Consistency and Selective Human Review: Higher Accuracy, Less Work

Submitted:

01 December 2025

Posted:

02 December 2025

Read the latest preprint version here

Abstract
Large language models (LLMs) show promise for grading open-ended assessments but still exhibit inconsistent accuracy, systematic biases, and limited reliability across assignments. To address these concerns, we introduce SURE (Selective Uncertainty-based Re-Evaluation), a human-in-the-loop pipeline that combines repeated LLM prompting, uncertainty-based flagging, and selective human regrading. Three LLMs – gpt-4.1-nano, gpt-5-nano, and the open-source gpt-oss-20b – graded answers of 46 students to 130 programming questions across five assignments. Each student answer was scored 20 times to derive majority-voted predictions and self-consistency-based certainty estimates. We simulated human regrading by flagging low-certainty cases and replacing them with scores from four human graders. We used the first assignment as a training set for tuning certainty thresholds and to explore LLM output diversification via sampling parameters, rubric shuffling, varied personas, multilingual prompts, and post-hoc ensembles. We then evaluated effectiveness and efficiency of SURE on the other four assignments using a fixed certainty threshold. Across assignments, fully automated grading with a single-prompt resulted in substantial underscoring, and majority-voting based on 20 prompts improved but did not eliminate this bias. Low certainty (i.e., high output diversity) was diagnostic of incorrect LLM scores, enabling targeted human regrading that improved grading accuracy while reducing manual grading time by 40–90%. Aggregating responses from all three LLMs in an ensemble improved certainty based flagging and most consistently approached human-level accuracy, with 70–90% of the grades students would receive falling inside human grader ranges. These findings demonstrate that self-consistency-based uncertainty estimation and selective human oversight can substantially improve the reliability and efficiency of AI-assisted grading.
Keywords: 
;  ;  ;  ;  
Subject: 
Social Sciences  -   Education

1. Introduction

Studies find that large language models (LLMs) can grade coding tasks (Korthals et al. 2025), short answers to open questions (Flodén 2025,Grévisse 2024), and student essays (Ishida et al. 2024,Yavuz et al. 2025) with amicable performance. Grading open forms of assessment manually can be time-consuming and boring, but they are often better measures of student ability than multiple-choice (Grévisse 2024,Polat 2020). Using LLMs for grading would allow teachers to rely on open forms of assessment while saving considerable time (Grévisse 2024,Korthals et al. 2025). Teachers could use this time to improve educational materials or tutoring students, making LLM grading a win-win for students and teachers alike. However, despite promising early findings, it remains unclear whether LLMs can grade accurately and fairly across assignments, courses, and programs, or if they are only accurate for specific cases. Consequently, multiple authors advise against fully automating the grading process at this point (Grévisse 2024,Kortemeyer Nöhl 2025,Schneider et al. 2024). This sentiment is mirrored by the European Union’s artificial intelligence (AI) act, which classifies AI grading as "high risk" and mandates human oversight (European Parliament and Council of the European Union 2024). To address this, we introduce SURE (Selective Uncertainty-based Re-Evaluation), a human-in-the-loop pipeline that combines automated LLM grading with uncertainty-based flagging and human review.
Specifically, we propose repeatedly prompting LLMs to score the same student answer to obtain a distribution of candidate scores, from which to derive a predicted score (e.g., mean, median, mode, etc.) and certainty estimate (e.g, standard deviation, entropy, etc.). Then, any low-certainty scores (e.g., falling below a threshold) can be flagged and later manually graded by a teacher. Repeatedly sampling from LLMs and aggregating their outputs serves two purposes. First, aggregating results across prompts may improve grading accuracy: instead of relying on a single score, multiple samples might allow the model’s judgments to converge toward a more reliable estimate. Second, by examining the variability across samples, we aim to quantify uncertainty and flag questionable grades. We assume that when a grading task falls well within the LLM’s training distribution, it will consistently assign the correct score – or at least do so on average across repeated samples. In contrast, when a task is underrepresented, ambiguous, or absent from the training data, we expect greater variation in the scores, as the model may hallucinate or explore multiple plausible solutions rather than settling on a single, well-defined answer. This idea parallels the "self-consistency" approach introduced by X. Wang et al. (2023), who showed that aggregating answers from multiple reasoning paths not only improves overall accuracy but that the level of agreement among samples can serve as a measure for uncertainty.
Combining automated assessment with human review of uncertain cases (Horton et al. 2025) has been suggested for other high-risk applications, such as medical diagnosis (Strong et al. 2025) and financial fraud detection (Alves et al. 2025). Similarly, Kortemeyer Nöhl (2025) evaluated a related procedure for grading: they obtained ten independent LLM-generated scores per student response, averaged these scores, and compared the mean to predictions from item response theory (IRT) to estimate grading uncertainty and identify responses requiring human review. They found that uncertainty based thresholding improved the LLM grading accuracy for physics exams (Kortemeyer Nöhl 2025).
The effectiveness of the SURE pipeline we propose depends critically on the diversity of LLM outputs that arises from repeated prompting. If repeated scores are always identical, uncertainty estimates become meaningless, and we cannot reliably distinguish between easy to grade cases and such that require human review. We explored several strategies for increasing output diversity by influencing the stochasticity and variability of LLM responses:
First, we varied temperature and top-p parameters to control token-level randomness: lower values make outputs more deterministic, while higher values encourage more varied responses (Holtzman et al. 2019,OpenAI n.d. 5,Peeperkorn et al. 2024). For reasoning models that do not expose these parameters, we instead varied text verbosity, which affects the length of responses OpenAI (2025).
Second, we introduced several prompt perturbations designed to elicit different reasoning paths without changing the prompt content. Specifically, we explored shuffling the order of rubric criteria, instructing LLMs to adopt different grader personas (e.g., strict vs lenient), and prompting them in different languages. Prior work shows that LLMs are sensitive not only to a prompt’s semantic content but also to its phrasing and presentation (Fröhling et al. 2025,Lu et al. 2022,Q. Wang et al. 2025), and that leveraging such diversification can improve uncertainty estimation (Chen Mueller 2024).These effects may be especially pronounced in cases where the model has not converged on a stable reasoning path, and may help reveal those instances in which its grading is unreliable.
Third, we investigated LLM ensembles – aggregating the outputs of multiple models rather than relying on a single one – to increase output diversity and reduce model-specific biases. This approach builds on the idea of ensemble learning, where combining several imperfect predictors often yields more robust performance, as seen in methods such as bagging and random forests (Hastie et al. 2009). Similar ideas are now being explored for LLMs (Tekin et al. 2024,J. Wang et al. 2024,Yang et al. 2025) and they might be particularly useful for estimating (un)certainty: because different LLMs are trained on distinct data and optimization objectives, their outputs might vary in informative ways when evaluating the same student response. Aggregating these diverse perspectives might stabilize majority-voted scores, especially when certain models are better suited to specific response types. For example, in a mixed programming assignment, a model fine-tuned for Python (Van Rossum Drake Jr 1995) might reliably evaluate Python-based answers, while another model optimized for R (Team 2022) could perform better on R-related questions; combining their outputs would yield more stable majority-voted scores than relying on either model alone. Moreover, if the outputs from multiple LLMs are consistent (high certainty) this may provide an even stronger indicator that the score assigned by the LLM was correct and does not require human review.

2. Materials and Methods

2.1. Data

We used data from 46 graduate students enrolled in an introductory programming course described in Korthals et al. (2025). The five graded assignments consisted of 130 coding challenges and open questions about coding. Each question was scored between 0 and 1 in 0.25 increments based on subtractive rubric criteria (e.g., "subtract 0.5 points if the student did not set their working directory correctly"). Four of the authors independently graded the student answers to obtain reliable ground truth scores 1. We built a Dash app (Hossain 2019) that let graders score answers question-by-question in random student order while recording the time between opening and submitting each score, enabling estimates of potential time savings. Missing answers were automatically assigned zero points and zero seconds. Because graders occasionally left the app open during breaks, any recorded time above the 99th percentile was treated as unrealistic and replaced with the mean time recorded by the other graders for that same answer. To compare our current findings with those from our earlier study, we additionally prompted gpt-4o (OpenAI et al. 2024) following the procedure in Korthals et al. (2025) to grade all student responses.
We defined the ground truth for each student answer as the score most frequently awarded by the four human graders. For ties, all tied scores were treated as valid ground-truth values. For each grader, we marked a score as correct if it matched any ground-truth value and incorrect otherwise. The relative frequency of correct scores across assignments, questions, or students provides a measure of grading accuracy. Because we previously found substantial underscoring by gpt-4o (Korthals et al. 2025), we also examined grading bias by calculating the signed deviation from the closest ground-truth value of each score. Averaging these deviations across grouping variables yields a measure of systematic over- or underscoring. We later computed the same accuracy and bias metrics for all LLM graders to compare their performance under fully automated and human-in-the-loop SURE grading with that of the human graders.
We had no strong a-priori hypotheses and wanted to explore many different LLM configurations and certainty thresholds. To ensure conservative estimates of the effectiveness and efficiency of the proposed SURE pipeline we used the first programming assignment (47 questions) as a training set for exploration and to select a fixed certainty threshold and single diversification strategy. Afterwards, we used the remaining four programming assignments (34, 27, 8, and 14 questions respectively)2 to compare the performance of SURE to manual and fully automated grading.

2.2. SURE Pipeline

2.2.1. Repeated Prompting for Score Prediction and Uncertainty Estimation

We graded each student answer with 20 repeated prompts, selected the most frequent score as the predicted score and its relative frequency as the models certainty. With 20 prompts, certainty can vary between 0 and 1 in increments of 0.05. In rare cases ( 0.01 % of processed prompts), LLM runs produced fewer valid responses due to API or parsing errors, resulting in coarser certainty estimates (smallest number of valid iterations for a student answer was 16). We landed on 20 iterations because it struck a balance between costs and precision – more prompts might improve certainty estimation, but would also increase API costs and the time to process them. Estimating certainty this way focuses only on the peak (the most frequent score) of the distribution of plausible scores and disregards potentially informative aspects of its shape, such as skewness, spread, or multimodality. While this simplification ignores some potential uncertainty cues, we used it here to get a first estimate of the effectiveness of the proposed pipeline and to maintain an uncertainty metric that is intuitively interpretable (i.e., "What percentage of runs produced this grade?").

2.2.2. Flagging Low-Certainty Scores and Simulating Human Regrading

Predicted scores with certainties below a selected threshold are flagged for human review. Thus, the threshold determines the trade-off between human effort and grading accuracy: a lower threshold reduces effort but risks overlooking incorrect scores, while a higher threshold increases human workload but (thereby) improves accuracy. Flagged cases can be understood as true positives (TP), false positives (FP), true negatives (TN), and false negatives (FN):
  • TP: LLM is uncertain (flagged) and incorrect — a useful flag
  • FP: LLM is uncertain (flagged) but correct — unnecessary teacher effort
  • TN: LLM is confident (unflagged) and correct — ideal automatic grading
  • FN: LLM is confident (unflagged) but incorrect — undetected error
In the training set, we selected the threshold ( τ ) that maximized the F 1 score. Similar to Kortemeyer Nöhl (2025), we chose the F 1 metric because it balances accuracy with human workload. In other words, it tries to catch as many incorrect scores (TP), while avoiding unnecessary flags (FP) that would inflate teacher effort.
F 1 = 2 · T P 2 · T P + F P + F N τ * = arg max τ [ 0 , 1 ] F 1 ( τ )
To assess the effectiveness of the SURE pipeline, we simulated manual regrading by replacing predicted scores of flagged cases with randomly sampled scores from one of the human graders and recalculating correctness (matching ground truth) and bias (signed deviation from closest ground truth). Sampling randomly from the human graders accounts for the fact that they also sometimes make mistakes, and regrading could even result in less accurate grading when uncertain yet correct scores are replaced with human scores which deviated from the ground truth.

2.3. LLM Configurations and Diversification Strategies

We prompted three LLMs to respond with structured JSON outputs (see Listing A1 and Listing A2): gpt-4.1-nano, a closed-source non-reasoning model (OpenAI n.d. 2); gpt-5-nano, a closed-source reasoning model (OpenAI n.d. 3); and gpt-oss-20b, an open-source reasoning model (OpenAI n.d. 4). LLMs were queried through the OpenAI batch API (OpenAI n.d. 1) (gpt-4.1-nano and gpt-5-nano) or on a private AzureML (Microsoft n.d.) compute cluster owned by the university (gpt-oss-20b).
To explore how parameter settings and prompt perturbations affect grading behavior, we created 48 distinct prompting conditions: 24 for gpt-4.1-nano, 16 for gpt-5-nano, and 8 for gpt-oss-20b (see prompting conditions in Table 1). Each condition was used to grade every student answer in the first programming assignment (training set) 20 times, resulting in a total of 2,075,520 prompts (48 conditions × 20 iterations × 47 questions × 46 students).

2.3.1. Parameter Variations

For gpt-4.1.-nano, we varied t e m p e r a t u r e { 0.0 , 1.0 } and t o p _ p { 0.1 , 1.0 } (the latter only when t e m p e r a t u r e = 1.0 ). For gpt-5-nano, we varied t e x t _ v e r b o s i t y { l o w , m e d i u m } . For gpt-oss-20bs we fixed t e x t _ v e r b o s i t y = m e d i u m to reduce runtime on the compute cluster.

2.3.2. Prompt Perturbations

For all three LLMs we applied three prompt perturbation techniques:
  • Shuffled rubrics – Each rubric consisted of a list of subtractive grading criteria (Korthals et al. 2025). When this intervention was active, we randomly sampled 20 criteria orderings from all possible permutations. For rubrics with fewer than four criteria ( 3 ! = 6 < 20 ), permutations were repeated equally until reaching 20. Otherwise, criteria followed the original order.
  • Grader personas – We defined four personas: strict, lenient, meticulous, and sloppy (see Listing A3). When enabled, we sampled each persona 5 times to add a persona to each of the 20 prompts. Otherwise, prompts contained no persona.
  • Multilingual prompting – We used gpt-5-nano to translate all prompt components (base prompt, questions, rubrics, and persona snippets) into German, Spanish, French, Japanese, and Chinese, and verified the translations by back-translating to English via DeepL (DeepL n.d.). When this intervention was active, we sampled equally across the six languages (including English). Otherwise, all 20 prompts were in English.

2.3.3. Post-Hoc Ensembles

We created LLM ensembles post-hoc by resampling the outputs of eight matched triplets taken from 24 prompting conditions. Specifically, we constructed eight ensembles that varied the three prompt-perturbation techniques: for gpt-oss-20b we used all eight conditions, for gpt-4.1-nano we used the eight conditions with t e m p e r a t u r e = 1.0 and t o p _ p = 1.0 , and for gpt-5-nano we used the eight conditions with t e x t _ v e r b o s i t y = m e d i u m . These ensembles are not fully independent from their reference conditions and we used bootstrap resampling with 3000 iterations to break ties: For each student answer, within each ensemble condition mapping onto three prompting conditions, we first pooled all 60 available responses (20 from each LLM) and then drew B = 3000 bootstrap resamples (with replacement) of size 20. For each resample, we computed the majority voted predicted score and its certainty. We then aggregated over all B resamples by taking the most frequent score as the ensemble’s final predicted score and the mean certainty as its certainty estimate.

2.4. Grading Procedures

We examined three LLM-based grading procedures:
  • Majority-voting (MV): Fully automated grading based on the most frequent scores assigned to each student answer across 20 repeated grading iterations. We computed these for all 56 conditions (prompting + post-hoc ensembles).
  • Single-prompt (SP): Fully automated grading based on a single score we sampled from the 20 iterations for each student answer. We applied this only to the 48 prompting conditions, ensembles per definition aggregate the outputs from multiple prompts.
  • SURE: Human-in-the-loop grading based on majority-voting with simulated human regrading of flagged scores. We assessed SURE for all 56 conditions (prompting + post-hoc ensembles). In the training set we tuned separate uncertainty thresholds by maximizing the F 1 score for each of the 56 conditions. In the test set we used the median of these 56 thresholds as a fixed certainty threshold.
Like for the four human graders, we classified the resulting scores as correct or incorrect and computed their deviation from the closest ground-truth scores. We used these metrics to evaluate and compare the performance of the three LLM-based grading procedures against one another and against fully manual grading.

2.5. Research Questions

We investigated the following research questions:
  • RQ1: Can majority-voting improve the accuracy of fully automated LLM grading?
  • RQ2: Can SURE improve the accuracy over fully automated LLM grading?
  • RQ3: Can diversification strategies (token sampling, prompt perturbations, LLM ensembles) improve the SURE protocol?
  • RQ4: How effective (accuracy) and efficient (time spent grading) is SURE compared to fully manual grading?
RQ3 was only investigated on the training set. All other research questions were investigated both on the training and the test set.

2.6. Exploratory Analyses on the Training Set

2.6.1. Grading Procedures and Diversification Strategies

Table 2 illustrates the the data structure for grading procedures and outcome variables. Each row represents the outcome of a grading procedure applied to a specific student answer. The variable correct is binary (1 = correct, 0 = incorrect), error is the signed deviation from the nearest ground-truth value (negative = under-grading, positive = over-grading). Each unique student–question–condition combination contains two rows for ensemble conditions or three for prompting conditions – one per procedure.
Table 3 illustrates the condition-level manipulations. The variable llm has four levels (gpt-4.1-nano, gpt-5-nano, gpt-oss-20b, ensemble). All other predictors were binary indicators: temperature (0 = 0.0 or absent, 1 = 1.0), topp (0 = 0.1 or absent, 1 = 1.0), verbosity (0 = low or absent, 1 = medium), shuffling (0 = no shuffle, 1 = shuffle), persona (0 = none, 1 = varied), and languages (0 = English, 1 = varied).
We fit a Bayesian hierarchical logistic regression model to predict the log-odds of scoring student answers correctly based on grading procedure, LLM, model parameters, prompt perturbations and all meaningful two-way interactions: For procedure the reference category was set to majority-voting. For llm the reference category was set to gpt-4.1-nano. Consequently, the model intercept represents the condition majority-voting with gpt-4.1-nano at t e m p = 0 , t o p p = 1 , v e r b = 0 , and without any prompt perturbations ( s h u f = p e r s = l a n g = 0 ). Terms for temp and topp were only included for gpt-4.1-nano (with top_p nested within temperature), and terms for verb were only included for gpt-5-nano. We added all two-way interactions while preserving the conditional nesting of temperature, top_p and text_verbosity. We also added random intercepts for questions and students to account for repeated measures and clustering.
We used the Bambi package (Capretto et al. 2022) to fit models in Python (Van Rossum Drake Jr 1995). We used four MCMC chains with 1000 tuning and 1000 sampling iterations, yielding 4000 posterior draws in total. We used Bambi’s default weakly informative priors (Westfall 2017). We interpreted all coefficients whose 95% highest density intervals (HDIs) excluded zero.
This analysis addressed the first three research questions at the level of individual student answers:
  • A negative coefficient for single-prompt grading would indicate that majority-grading improves fully automated grading (RQ1).
  • A positive coefficient for SURE that is larger than those for SP and MV would indicate that the proposed pipeline improves accuracy over automated grading (RQ2).
  • Positive coefficients for any of the diversification strategies – particularly in combination with SURE – would indicate that diversification strategies are beneficial (RQ3).

2.6.2. Comparing Single-Prompt, Majority-Voting, SURE and Manual Grading

Based on the results of the regression model outlined in the previous section, we selected the following model configurations without any prompt perturbations and compared their performance under different grading procedures (SP, MV, SURE) with manual grading:
  • gpt-4.1-nano with temperature and top_p set to 1.0.
  • gpt-5-nano with text_verbosity set to medium.
  • gpt-oss-20b with text_verbosity set to medium.
  • ensemble based on the three selected LLM configurations.
Accuracy and bias at the level of student answers. To compare SURE and manual grading at the level of student answers (RQ4) we selected data after simulating human regrading. Table 4 illustrates the the data structure. Each row shows the score awarded by a specific grader (human or LLM with SURE) to a specific student answer.
We used this data to run two Bayesian hierarchical regression models: a logistic model for correct and a Gaussian model for error. For both, we set intercepts to zero and tested the effect of the grader predictor with random intercepts for students and questions. We estimated the models using four MCMC chains, each with 1000 tuning and sampling iterations. Then, we compared the posterior estimates for each pairwise comparisons (e.g., grader-1 vs. grader-2, grader-1 vs. gpt-5-nano, etc.) to assess whether one grader was better (more accurate, less biased) than another. Specifically, for each pairwise comparison we computed the percentage of samples for grader A that were greater (accuracy) or closer to zero (bias) than grader B. If this analysis would reveal that SURE grading achieves or exceeds the performance of some or all human graders, this would indicate that manual grading may be replaced with it.
Alignment at the level of assignment grades. The previously described analysis was done on the student answer level; however, even small differences (e.g., some grader slightly underscoring student answers) may accumulate at the level of overall grades, which is a phenomenon we observed for single-prompt LLM grading in Korthals et al. (2025). Therefore, we also compared the grading performance of fully automated LLM grading (SP, MV), human-in-the-loop SURE grading, and manual grading (RQ1, RQ2, RQ4) at the assignment level. For this, we computed grades on the Dutch 10-point grading scale. In the first assignment, students had to answer 30 basic R questions (worth 8 points) and either 9 Python or 7 advanced R questions (worth 2 points):
grade = sum ( R ) 30 · 8 + max sum ( Python ) 9 · 2 , sum ( R _ adv) 7 · 2
We applied this formula to each human grader’s scores and to the ground truth scores obtained from the four graders. The minimum and maximum across graders defined each student’s human grade range, and the minimum and maximum ground-truth scores defined the ground truth grade range; together, we consider these the target ranges in which grades from fully automated grading (SP, MV) or human-in-the-loop SURE grading should fall to be considered equivalent to manual grading. We calculated the grades students would have received from the three LLMs and the ensemble under each of the three grading procedures (SP, MV, SURE). For each, we computed the proportion of LLM grades that fall inside the target ranges and the maximum and median deviation from the closest target range boundary as metrics for alignment. If the majority of LLM awarded grades falls inside the target ranges and if the maximum and median deviation are small, this would indicate that a given grading procedure may be accurate enough to replace fully manual grading. We considered these metrics descriptively and inspected plots to address RQ1, RQ2, and RQ4:
  • If majority-voted grades would be more aligned with target ranges than grades based on single-prompts this would lend support that majority-voting improves fully automated gradign (RQ1)
  • If grades from SURE would be more aligned than fully automated grades (SP and MV) this would indicate the benefit of SURE (RQ2).
  • By assessing the proportion of SURE grades that fall inside the target ranges and the maximum and median deviations from target range boundaries we assess whether SURE may be suitable to replace manual grading (RQ4).
Time savings from SURE. Even if the previous analyses would indicate that SURE grading achieves human performance, it might not make sense to replace fully manual grading with it. This is because the performance of SURE grading could be entirely driven by manual regrading. Consider a scenario in which almost all cases are flagged and regraded; if this would be the case, SURE grading would essentially be manual grading and provide not meaningful time savings and not be more efficient (RQ4). To address this we calculated the time it took human graders to manually score all student answers and compared it with the time it would have taken them if they only had to score flagged cases. We present the absolute grading times under manual and SURE grading and time savings as a percentage and discuss whether they support replacing manual grading or not.

2.7. Planned Analyses on the Test Set

We used the test set to repeat the analyses for the four LLM configurations without prompt perturbations that we selected based on the first analysis in the training set:
  • gpt-4.1-nano with temperature and top_p set to 1.0.
  • gpt-5-nano with text_verbosity set to medium.
  • gpt-oss-20b with text_verbosity set to medium.
  • ensemble based on the three selected LLM configurations.
We prompted them to score each student answer in assignments 2-5 20 times and created the post-hoc LLM ensemble by resampling the three prompting conditions. For flagging we used a fixed threshold, which we determined by taking the median ( τ = 0.7 ) of the 56 tuned thresholds from the training set. We then repeated the analyses described in Section 2.6.2. For the question-level analyses, we considered data from all four assignments (2-5) and used Bayesian regression to predict correctness, and grading bias, from grader (humans and LLMs with SURE) with random intercepts for questions and students. In contrast, we computed the alignment with assignment grades and time savings for each assignment separately. Assignments 2 and 3 used the same grading logic with different numbers of R, advanced R, and Python questions; for assignment 4 we only considered R questions; and for assignment 5 (the exam) students had to complete 14 out of 16 available questions (R or Python):
grade a s s 2 = sum ( R ) 18 · 8 + max sum ( Python ) 7 · 2 , sum ( R _ adv) 2 · 2
grade a s s 3 = sum ( R ) 25 · 8 + max sum ( Python ) 7 · 2 , sum ( R _ adv) 2 · 2
grade a s s 4 = sum ( R ) 8 · 10
grade a s s 5 = sum ( Q ) 14 · 10

3. Results

3.1. Exploratory Findings on the Training Set

3.1.1. Descriptive Findings

Figure 1 shows how certainty thresholds were tuned by maximizing F 1 scores for each of the 56 conditions. Most optimal thresholds (Δ) lie between 0.6 and 0.85. Optimal thresholds and the average F1 trajectory for the ensemble condition (black triangles and line) are markedly shifted to the left and more peaked than those for the other models. For the test set we fixed the threshold at 0.7, which was the median optimal threshold across all conditions (dashed vertical line). Figure 2 shows the certainty of correct and incorrect scores and their relationship to tuned certainty thresholds for all conditions of a given LLM. It suggests that certainty is diagnostic of correctness: most correct scores cluster at 100% certainty (all iterations agree), while incorrect scores are more widely distributed at lower levels of certainty. The ensemble stands out with a distinctly bimodal distribution and concentrated thresholds, suggesting that mixing multiple LLMs can help separate cases suited for automated grading from those that aren’t.
Figure 3 displays the observed grading accuracy of different models, grading procedures, and prompting configurations, aggregated across students and questions for the first assignment. Different LLMs are displayed on the x-axis and visual inspection suggests that the reasoning models (gpt-5-nano, and gpt-oss-20b) and the ensemble clearly outperformed gpt-4.1-nano even when using only a single prompt (circles). Both majority-voting (squares) and SURE (triangles) appear to improve the accuracy of all LLMs, with gpt-oss-20b and the ensemble even reaching human level accuracy (grey band) and gpt-4.1-nano achieving the greatest relative gains, potentially because more cases were flagged for this LLM. Effects of prompt perturbations (colors) are difficult to assess visually, but it seems like multilingual prompting may have hurt the accuracy of the three LLMs, particularly gpt-4.1-nano and especially for single-prompt grading. We also see that gpt-5-nano with a single-prompt is on par with gpt-4o (dotted line) used in Korthals et al. (2025), while majority-voting improves the accuracy beyond it. Based on visual inspection alone, the Figure 1, Figure 2 and Figure 3 suggest that self-consistency (X. Wang et al. 2023) based certainty estimation can work for weaker as well as stronger LLMs and that the performance of the proposed SURE procedure may be comparable to fully manual grading.

3.1.2. Grading Procedures and Diversification Strategies

We used Bayesian logistic regression to predict the log-odds of scoring student answers correctly based on the grading procedure, LLM, sampling parameters, and prompt perturbation techniques and all meaningful two-way interactions (predictors that were varied together). The four MCMC chains each with 1000 warmup and sampling iterations mixed well (all R ^ 1.01 ). Table 5) shows the coefficients and 95% HDI of all coefficients. In the following we only interpret those whose 95% HDI excludes zero.
At the intercept (majority-voting, gpt-4.1-nano, temp=0, topp=1, verb=0, shuf=0, pers=0, lang=0; M = 1.601 , 95% HDI [ 1.199 , 1.980 ] ) the probability to score a student answer correctly is estimated to be about 83%. Relying oly on a single-prompt ( M = 0.230 , 95% HDI [ 0.300 , 0.161 ] ) reduces that probability, while human-in-the-loop SURE ( M = 1.311 , 95% HDI [ 1.219 , 1.400 ] ) increases it. Utilizing gpt-5-nano ( M = 1.437 , 95% HDI [ 1.319 , 1.557 ] ), gpt-oss-20b ( M = 1.958 , 95% HDI [ 1.821 , 2.100 ] ), or the LLM ensemble ( M = 1.989 , 95% HDI [ 1.830 , 2.147 ] ), instead of gpt-4.1-nano also increased the probability of scoring student answers correctly. These results at the level of individual student answers are consistent with the earlier visual inspection of the grading accuracy at the assignment level (Figure 3): majority-voting and SURE are better than single-prompt grading, and the reasoning models (gpt-5-nano, gpt-oss-20b) and the ensemble outperformed gpt-4.1-nano.
We obtained negative coefficients for interactions between SURE : gpt-5-nano ( M = 0.632 , 95% HDI [ 0.751 , 0.513 ] ), SURE : gpt-oss-20b ( M = 0.686 , 95% HDI [ 0.831 , 0.542 ] ), and SURE : ensemble ( M = 0.742 , 95% HDI [ 0.880 , 0.601 ] ). This reflects that the relative performance gain for gpt-4.1-nano from SURE is greater than that for the other LLMs, whose baseline accuracy (single-prompt / majority-voting) is already greater and closer to the ceiling.
We found a positive coefficient for the main effect of topp(llm=gpt-4.1-nano; temp=1) ( M = 0.176 , 95% HDI [ 0.075 , 0.275 ] ) and a negative interaction for topp(llm=gpt-4.1-nano; temp=1) : single-prompt ( M = 0.158 , 95% HDI [ 0.233 , 0.082 ] ). This indicates that prompting gpt-4.1-nano with temperature and top_p set to 1 is beneficial but only for majority-voting and SURE. Figure 3 clearly shows how majority-voting (squares) for gpt-4.1-nano with lower temperature and top_p is only slightly beneficial, while a large jump in accuracy can be seen for gpt-4.1-nano with higher temperature and top_p. Together with the regression results, this indicates that token-level variability may help stabilize majority voted scores, potentially because more plausible scores are explored, while deterministic sampling results in getting stuck in a local minimum similar to relying on a single-prompt.
None of the prompt perturbation techniques improved the probability to score student answers correctly. On the contrary, we obtained negative coefficients for lang ( M = 0.080 , 95% HDI [ 0.167 , 0.002 ] ), and the interactions between lang : single-prompt ( M = 0.527 , 95% HDI [ 0.580 , 0.475 ] ), lang : topp(llm=gpt-4.1-nano; temp=1) ( M = 0.169 , 95% HDI [ 0.260 , 0.073 ] ), and lang : shuf ( M = 0.082 , 95% HDI [ 0.14 , 0.024 ] ). These indicate that multilingual prompting was detrimental, particularly when relying only on a single-prompt, simultaneously shuffling rubrics, and using gpt-4.1-nano with increased token-level sampling variability.
We also found positive coefficients for the interactions between lang : gpt-5-nano ( M = 0.295 , 95% HDI [ 0.192 , 0.397 ] ), and lang : gpt-oss-20b ( M = 0.283 , 95% HDI [ 0.160 , 0.397 ] ), suggesting that multilingual prompting was less detrimental for more recent the reasoning models. This is in line with Figure 3, which clearly shows how multilingual prompting was very detrimental for single-prompt grading with gpt-4.1-nano but less so for the other LLMs and grading procedures.
Finally, for random intercepts we found moderate variability for students ( M σ = 0.355 , 95% HDI [ 0.282 , 0.435 ] ), and considerable variability for questions ( M σ = 0.355 , 95% HDI [ 0.282 , 0.435 ] ). This indicates that some students are easier to score than others, which raises concerns for potentially biased grading, and suggests that LLMs are worse at scoring certain questions, which is in line with our earlier findings (Korthals et al. 2025) and exactly what we want to address with SURE grading.
With respect to the research questions, this regression model and Figure 2 and Figure 3 suggest that majority-grading improves fully automated grading (RQ1), SURE improves accuracy over automated grading (RQ2), and only higher temperature and top_p and ensembling are effective diversification strategies (RQ3). Based on these results, we decided to focus only on four LLM configurations for all other analyses:
  • gpt-4.1-nano with temperature and top_p set to 1.0.
  • gpt-5-nano with default "medium" text_verbosity.
  • gpt-oss-20b with default "medium" text_verbosity.
  • ensemble based on the three selected LLM configurations.

3.1.3. Comparing Single-Prompt, Majority-Voting, SURE and Manual Grading

Accuracy and bias at the level of student answers. We fit a Bayesian logistic regression with random intercepts for students and questions to estimate the log-odds that each of the four human graders and human-in-the-loop SURE grading with four LLM graders (gpt-4.1-nano at temp=topp=1, gpt-5-nano with verb=1, gpt-oss-20b, and the ensemble; without prompt perturbations) would score student answers correctly. At first we obtained R ^ values around 1.05, so we increased the sampling to 2000 tuning and 2000 sampling iterations. After this change convergence between the four MCMC chains improved ( R ^ 1.01 ). The model was fit without an intercept, which means that once the coefficients are transformed from log-odds to probabilities, each one directly represents that grader’s estimated probability of assigning a correct score. Below we report estimated coefficients as log-odds but interpret the results at the probability level.
Figure 4 displays the posterior means and 95% HDIs (left panel) and the pairwise probability that a grader was more accurate than another (right panel). With about 92% estimated probability to score student answers correctly, grader 2 ( M = 2.482 , 95% HDI [ 1.53 , 3.476 ] ) was the most accurate human grader. They were followed by grader 3 ( M = 1.92 , 95% HDI [ 0.968 , 2.854 ] ; 87 % ), grader 1 ( M = 1.734 , 95% HDI [ 0.823 , 2.716 ] ; 85 % ) and grader 4 ( M = 1.524 , 95% HDI [ 0.617 , 2.505 ] ; 82 % ). Under human-in-the-loop SURE grading with tuned certainty thresholds, both gpt-oss-20b ( M = 1.693 , 95% HDI [ 0.732 , 2.632 ] ; 84 % ) and the LLM ensemble ( M = 1.595 , 95% HDI [ 0.644 , 2.543 ] ; 83 % ) reached accuracies comparable to the mid-range of human graders. In contrast gpt-5-nano ( M = 1.343 , 95% HDI [ 0.379 , 2.281 ] ; 79 % ) and particularly gpt-4.1-nano ( M = 0.755 , 95% HDI [ 0.143 , 1.718 ] ; 68 % ) performed worse, with all human graders likely outperforming them.
We fit a similar Bayesian linear regression to predict grading bias with four MCM chains making 1000 tuning and sampling draws (all R ^ = 1.0 ). This regression revealed a more pronounced difference between human graders and human-in-the-loop SURE grading: Human grader 4 ( M = 0 , 95% HDI [ 0.005 , 0.005 ] ) and grader 2 ( M = 0.003 , 95% HDI [ 0.008 , 0.002 ] ) were unbiased (HDI include zero), while grader 3 ( M = 0.009 , 95% HDI [ 0.014 , 0.004 ] ) was underscoring and grader 1 ( M = 0.008 , 95% HDI [ 0.002 , 0.013 ] ) was overscoring. In contrast, despite SURE all LLM graders underscored students (negative bias) and were likely more biased than most human graders: gpt-oss-20b ( M = 0.009 , 95% HDI [ 0.014 , 0.004 ] ); ensemble ( M = 0.011 , 95% HDI [ 0.016 , 0.006 ] ); gpt-4.1-nano ( M = 0.015 , 95% HDI [ 0.02 , 0.01 ] ); gpt-5-nano ( M = 0.016 , 95% HDI [ 0.021 , 0.011 ] ).
Figure 5. Posterior estimates of grading bias for human graders and LLMs under SURE with tuned certainty thresholds in the training set. The left panel shows posterior means and 95% HDIs for grading bias (deviation from closest ground-truth). The right panel displays pairwise dominance probabilities, indicating for each row–column pair the posterior probability that the grader in the row is less biased (closer to zero) than the grader in the column.
Figure 5. Posterior estimates of grading bias for human graders and LLMs under SURE with tuned certainty thresholds in the training set. The left panel shows posterior means and 95% HDIs for grading bias (deviation from closest ground-truth). The right panel displays pairwise dominance probabilities, indicating for each row–column pair the posterior probability that the grader in the row is less biased (closer to zero) than the grader in the column.
Preprints 187686 g005
Alignment at the level of assignment grades. The previous analysis was conducted at the level of individual student answers, for which differences between manual grading and SURE were relatively small. However, even small differences may accumulate at the level of assignment grades. To evaluate this, we computed target grade ranges (human and ground-truth grades) and the grade each student would have received under a given LLM and grading procedure.
Figure 6 reveals considerable disagreement between human graders (wide grey areas) in some cases but also perfect agreement in one case (black horizontal line instead of grey area). It also shows a pronounced negative grading bias for single-prompt grading (circles are often far below the grey target ranges), which is consistent with earlier findings using gpt-4o (Korthals et al. 2025). The plot also shows that majority-voting and SURE pull student grades closer to or even inside the target ranges for most students. Interestingly, there are also individual cases where majority-voting or SURE decreased the alignment with human grade ranges. This reflects that sometimes single prompts can "accidentally" be right and that human (re)graders sometimes make mistakes and are less correct than LLMs, which is something we also observed in Korthals et al. (2025). The starkest difference emerged for gpt-4.1-nano: Table 6 shows that for single-prompt grading only 7 % of grades fell inside the target range and the maximum grade deviation reached 1.9 grade points. In contrast, with SURE, 61 % of grades fell inside the target ranges and the maximum deviation was only 0.4 grade points. We see a similar pattern for all other LLMs, with SURE resulting in maximum and median grade point deviations of 0.5 and 0.1 grade points for all LLMs respectively. This finding suggests that SURE can result in grades matching human accuracy, particularly for stronger LLMs and the LLM ensemble.
Time savings from SURE. While the previous analysis indicates that SURE can achieve good alignment with human grades, this could be driven by a large flagging rate which would mean that SURE is largely manual grading anyway and offers no substantial time-savings. We assessed this by comparing the time spent for manual grading with the time graders would spend under SURE. Table 7 shows that manually grading the first assignment took between 3 and 6.5 hours. In contrast, the SURE procedure would yield time savings up to 90%, making it highly efficient. This table also shows that time savings would be less for gpt-4.1-nano than for the other LLMs and the ensemble, reflecting the lower baseline accuracy and greater flagging rate for this LLM. Together with the big performance increases for gpt-4.1-nano, this provides further evidence that self-consistency based uncertainty estimation is effective for both weaker non-reasoning models (gpt-4.1-nano) and reasoning models (gpt-5-nano, gpt-oss-20b).

3.1.4. Summary of Training Set Results

To summarize, in the training set we found evidence that self-consistency (X. Wang et al. 2023) based (un)certainty can distinguish between cases that can be graded automatically and those that should be reviewed by a human grader. Notably, aggregating the outputs from several LLMs in an ensemble visibly improved the separability of incorrect and correct scores.
We also found that SURE based on optimal thresholds can result in greatly improved alignment with human graders both at the level of individual student answers and assignment grades, while saving teachers more than 80% of the grading time spent on this assignment.
However, these results were based on optimally set flagging thresholds tuned on an efficiency ( F 1 score) objective. Therefore, they might be overly optimistic estimates of the effectiveness and efficiency of the proposed SURE pipeline. In practice, thresholds would have to be set in advance, which might be problematic if the obtained certainties are highly assignment specific. Therefore, we simulated SURE with a fixed certainty threshold (0.7) in the test set (assignments 2-5). These assignments include more complex programming tasks (e.g., manipulating data, creating plots) which we previously found to be graded less accurately with gpt-4o (Korthals et al. 2025) and therefore likely will include exactly such cases that we want to identify and flag for human reviews.

3.2. Test Set Validation

3.2.1. Descriptive Findings

Figure 7 shows how correct and incorrect scores were distributed across certainty levels in the test set (assignments 2–5). As in the training set, most correct scores cluster at 100% certainty, indicating full agreement across repeated prompts. However, there are also more correct scores at lower certainty values, which increases the overlap between the distributions of correct and incorrect scores. This overlap implies more unwanted flags (false positives; blue bars to the left of the threshold) and more unflagged incorrect scores (false negatives; red bars to the right of the threshold). Once again, the ensemble stands out with a distinctly bimodal certainty distribution that is neatly separated by the fixed threshold at 0.7. In contrast, the same fixed threshold appears to be too low for the individual LLMs, resulting in many incorrect scores that remain unflagged.
Figure 8 shows assignment-level grading accuracy in the test set for all LLMs and grading procedures. As in the training set, SURE (triangles) and, to a lesser extent, majority-voting (squares) improve accuracy over single-prompt grading (circles) for all models. For all four assignments, SURE with the ensemble reaches human-level grading accuracy (grey band). In contrast, gpt-4.1.-nano clearly reached lower grading accuracy for all assignments even with SURE, and gpt-oss-20b, and gpt-5-nano approached human performance only for assignments 4 and 5. This likely reflects the inadequate thresholds, which resulted in too many unflagged incorrect cases.

3.2.2. Comparing Single-Prompt, Majority-Voting, SURE and Manual Grading

Accuracy and bias at the level of student answers. Similar to the training set, we ran a Bayesian logistic regression to assess the accuracy of the four human graders and the four human-in-the-loop SURE grading procedures (gpt-4.1-nano, gpt-5-nano, gpt-oss-20b, ensemble) at the level of individual student answers in the test set (assignments 2–5). The model included random intercepts for students and questions and was fit without an intercept, so that each coefficient directly represents the log-odds of assigning a correct score for a specific grader. We estimated the model with four MCMC chains, 1000 tuning 1000 and sampling iterations per chain, and all R ^ 1.01 indicated good convergence.
With an estimated probability of correctly scoring student answers of about 95%, human grader 3 ( M = 2.927 , 95% HDI [ 2.530 , 3.369 ] ) was the most accurate grader. However, this estimate is based only on assignment 5 (the only test-set assignment graded by grader 3), which makes it less generalizable than the estimates for the other human graders and the LLMs under SURE. Grader 3 is followed by grader 1 ( M = 2.788 , 95% HDI [ 2.532 , 3.028 ] ; 94 % ), the ensemble ( M = 2.59 , 95% HDI [ 2.341 , 2.831 ] ; 93 % ), grader 4 ( M = 2.558 , 95% HDI [ 2.309 , 2.8 ] ; 93 % ), grader 2 ( M = 2.276 , 95% HDI [ 2.037 , 2.515 ] ; 91 % ), gpt-oss-20b ( M = 2.058 , 95% HDI [ 1.817 , 2.283 ] ; 89 % ), gpt-5-nano ( M = 1.928 , 95% HDI [ 1.701 , 2.174 ] ; 87 % ), and gpt-4.1-nano ( M = 1.263 , 95% HDI [ 1.046 , 1.494 ] ; 78 % ).
Figure 9 visualizes these results and highlights the relative ranking of graders in terms of pairwise dominance. The heatmap shows that grader 3 and grader 1 clearly outperform the two other human graders and all LLMs under SURE with very high posterior probability, and that grader 3 is likely more accurate than grader 1 as well. The ensemble occupies a distinct middle position: it is almost certainly more accurate than gpt-4.1-nano, gpt-5-nano, gpt-oss-20b, and grader 2, and is roughly comparable to grader 4 (pairwise probability close to 0.5). Together, the posterior means and pairwise dominance structure indicate that, in the test set, SURE with the ensemble achieves accuracy similar to mid-range human graders while SURE was less effective for the individual LLMs and did not achieve accuracies comparable to manual grading.
For grading bias, we find very similar results and even more evidence that SURE with the LLM ensemble rivals human performance: With zero being included in the 95% HDI, Human grader 4 ( M = 0.003 , 95% HDI [ 0.006 , 0.012 ] ) and the ensemble ( M = 0.007 , 95% HDI [ 0.016 , 0.001 ] ) may be considered unbiased. Like on the training set, human grader 1 was the only grader with a positive bias (overscoring; M = 0.016 , 95% HDI [ 0.007 , 0.024 ] ). All other graders were negatively biased (underscoring): human grader 3 ( M = 0.016 , 95% HDI [ 0.032 , 0.002 ] ); gpt-oss-20b ( M = 0.027 , 95% HDI [ 0.036 , 0.018 ] ); gpt-5-nano ( M = 0.037 , 95% HDI [ 0.045 , 0.027 ] ); human grader 2 ( M = 0.039 , 95% HDI [ 0.048 , 0.031 ] ); gpt-4.1-nano ( M = 0.045 , 95% HDI [ 0.054 , 0.036 ] ).
In the accuracy analysis, human graders 1 and 3 were clearly the strongest performers, with the ensemble under SURE grading occupying a solid mid-range position. In terms of bias, however, the picture shifts: Figure 10 shows that the ensemble is much closer to zero than most human graders, with an HDI that includes zero and a posterior mean comparable to the nearly unbiased grader 4. In contrast, graders 1 and 3 – despite being the most accurate – exhibit clear positive and negative bias respectively. The pairwise-dominance heatmap shows that the ensemble with SURE was very likely less biased than human graders 1, 2, and 3.
Alignment at the level of assignment grades. Like for assignment 1, we also assessed the accuracy of grades by computing human and ground-truth target ranges, and the grades students would receive under different LLM grading procedures. In contrast to the previous analysis at the level of student answers where we used data across assignments, we did this separately for each of the four test set assignments. The previous analyses suggest that SURE achieved human like performance for the ensemble and for brevity, we only show its results in Figure 11, while Table 8 shows performance metrics (percentage of grades inside target ranges, median and maximum grade point deviation from target range boundaries) for all LLMs.
The figure shows that the ensemble even under fully automated grading with majority-voting frequently produced grades that fall inside the target ranges; however, for some students this procedure resulted in severe underscoring with maximum and median grade point deviations up to 3.2 and 0.3 respectively. While the median deviation of less than half a grade point might be acceptable in practice, underscoring a student by more than three full grades is not. This highlights the importance of evaluating LLM grading not only at the level of averages but also at the level of individual students.
In contrast, SURE grading with the ensemble resulted in between 84% and 91% of grades falling inside target ranges, with 0.4 maximum and 0.1 median grade point deviation for all assignments. Notably, for assignment 4, we observed 15/46 students for which all three human graders and the ensemble with SURE agreed perfectly.
For the individual LLMs, Table 8 shows a similar pattern: SURE reduced the severity of grading errors, yielding maximum and median grade point deviations of 1.6 and 0.6 for gpt-4.1-nano, 1.9 and 0.3 for gpt-5-nano, and 1.4 and 0.1 for gpt-oss-20b. This indicates that even with suboptimal certainty thresholds, SURE can meaningfully improve alignment with human grading. However, the remaining deviations and relatively low proportions of grades falling inside the human target ranges suggest that these models still produce errors that are too large to justify replacing manual grading in practice. These results highlight the benefit of using LLM ensembles for uncertainty based flagging, but also put into question whether it is possible to set proper flagging threshold in advance.
Time savings from SURE. Table 9 shows that the time savings achieved by SURE in the test set varied substantially across LLMs and assignments. Notably, the ensemble yielded comparatively modest time savings – typically between 26% and 58% – while the individual LLMs often saved considerably more time, in some cases exceeding 80%. This pattern is consistent with earlier results showing that the fixed certainty threshold ( τ = 0.7 ) was well calibrated for the ensemble but too lenient for the individual LLMs: the weaker models produced many incorrect but high-certainty predictions that went unflagged, reducing the amount of manual regrading and thereby inflating time savings at the cost of lower accuracy. Conversely, the ensemble flagged a larger proportion of cases for review, which reduced automation but produced human-level accuracy.
Importantly, this outcome is not necessarily a limitation of the proposed SURE approach. For assignments that contain questions the LLMs struggle to grade reliably – such as those in assignments 2-5 – we explicitly want them to be flagged which necessarily results in more manual effort. At the same time, the table also shows that when the proportion of flagged responses becomes very large, the resulting time savings may be too small to justify deploying such a pipeline in practice. Thus, while the ensemble achieved the highest grading accuracy, it did so by relying more heavily on human review in the test set, illustrating the trade-off between efficiency and reliability inherent to certainty-based flagging.

3.3. Summary of Results

Across both the training and test sets, our results support the core idea of the proposed SURE pipeline. Repeated prompting produced a certainty measure that was strongly diagnostic of correctness, and combining models in an ensemble yielded a distinctly bimodal certainty distribution that separated clearly between high- and low-confidence predictions. On the training set, tuning certainty thresholds for each condition showed that self-consistency–based flagging can substantially improve grading accuracy for both weaker and stronger LLMs. With optimally chosen thresholds, SURE brought assignment-level accuracy and grade alignment close to or within the range of human graders, while reducing manual grading time by more than 80%.
The test set analysis, which used a single fixed threshold of τ = 0.7 and omitted prompt perturbations, provides a more conservative but still encouraging picture. For the ensemble, the fixed threshold aligned well with its bimodal certainty distribution: SURE achieved human-level accuracy at the level of individual answers, near-unbiased grading, and high alignment with human assignment grades (84–91% of grades inside target ranges, with very small maximum and median deviations). However, these gains came with only moderate time savings (typically 26–58%), reflecting that many responses were still routed to human graders. For the individual LLMs, the same fixed threshold was too lenient, leading to more unflagged incorrect answers, lower accuracy, stronger negative bias, and larger grade deviations, even though SURE still improved performance relative to single-prompt and majority-voting baselines.
Together, these findings highlight a central trade-off of certainty-based flagging. When thresholds are well calibrated – most clearly for the ensemble – the pipeline can match mid-range human graders in accuracy and bias while still reducing manual effort. At the same time, the test set results show that thresholds are assignment- and model-sensitive: overly aggressive automation can save time but harms reliability, whereas conservative thresholds preserve human-level accuracy at the cost of smaller time savings.

4. Discussion

We evaluated a human-in-the-loop LLM-based grading approach aimed at improving the reliability of AI-supported assessment. Three LLMs were prompted repeatedly to score the responses of 46 students to programming tasks, yielding majority-voted scores and certainty estimates. Low-certainty predictions were flagged, and we simulated manual regrading by randomly sampling scores from four human graders. Ground-truth scores were defined from the human graders, enabling us to evaluate alignment for each human grader and each AI-based grading procedure (single-prompt, majority-voting, and SURE). The first assignment served as a training set to assess diversification strategies – rubric shuffling, grader-persona instructions, multilingual prompting, and post-hoc LLM ensembles – and to tune certainty thresholds. Four held-out assignments were then used to validate findings using a fixed threshold. Across assignments, fully automated grading based on a single prompt produced substantially lower grades than human graders. Majority-voting improved alignment but still resulted in lower grades in most cases. Certainty was strongly diagnostic of incorrect LLM scores, and the SURE procedure generally decreased grading error while reducing manual grading time by roughly 40–90%, depending on model and assignment. Performance varied across LLMs, but the ensemble most consistently approached human-level accuracy, with approximately 70–90% of its assignment-level grades falling inside the range of human graders.
Our findings align with concerns raised in recent studies showing that, despite promising early results, fully automated LLM grading can exhibit systematic biases and inconsistencies on closer inspection (Johnson Zhang 2024,Y. Wang et al. 2025). Human oversight is frequently emphasized and mandated by the EU AI act (European Parliament and Council of the European Union 2024). The procedure we present here is only one of several human-in-the-loop approaches that have been proposed in response to this. These include CoTAL, which iteratively refines prompts through cycles of human and LLM grading (Cohn et al. 2025), and AVALON, which compares LLM scores with student self-evaluation to escalate ambiguous cases for instructor review (Armfield et al. 2025). Notably, Kortemeyer Nöhl (2025) introduced an approach conceptually similar to ours but requiring IRT modeling to derive uncertainty estimates. Our method complements these human-in-the-loop designs by being comparatively lightweight: it requires no iterative alignment or psychometric modeling and relies only on self-consistency (X. Wang et al. 2023), yet still allows large portions of an assignment to be graded automatically when the model is confident.
However, several limitations apply to this study and should be addressed in future work investigating SURE or similar approaches. One limitation concerns the way we derived predicted scores and certainty estimates. We relied on majority-voted scores and their relative frequency, which worked well under the coarse 0–1 rubric used here. For more continuous grading schemes, however, summary statistics such as the mean and standard deviation may provide more appropriate and informative indicators. Moreover, by focusing solely on the frequency of the modal score, we ignored other potentially useful uncertainty cues contained in the full distribution of predictions. Future studies should investigate these additional signals – such as distributional spread, entropy, or multimodality – as they may improve the accuracy of flagging low-certainty cases and thereby enhance both the effectiveness and the efficiency of the SURE procedure.
Setting an appropriate certainty threshold poses a second challenge for generalizability, and this issue was already evident within our own study when moving from the training to the test set, as the chosen threshold worked well for the ensemble but not for the individual LLMs. The clearer separation between correct and incorrect scores in the ensemble’s bimodal certainty distribution likely contributed to its robustness under a fixed threshold. Although we derived tuned thresholds for all 56 diversification conditions in the training set, we used their median as a single threshold for validation. It is possible that applying LLM-specific tuned thresholds to the test set would have yielded better performance, but we avoided this approach to avoid overfitting. In practical use, however, such “overfitting” may not be inherently problematic – teachers could tune thresholds on past assignments and apply them to future iterations, particularly when assignment formats and grading schemes remain consistent across years. Another practical consideration is how to tune thresholds. While we – like (Kortemeyer Nöhl 2025) – tuned thresholds using the F 1 score to balance accuracy and teacher effort, other objectives (e.g., maximizing sensitivity to ensure maximal accuracy) may be more appropriate in some settings. Future research may also investigate alternatives to tuned thresholds. For example deriving them dynamically from the empirical distribution of certainties or utilizing heuristics such as regrading all predictions below a specific certainty percentile or regrading a fixed proportion of least certain responses.
The ensemble’s robust performance under a fixed threshold, supported by its bimodal certainty distribution, underscores the promise of diversification through mixtures of LLMs. Our ensemble combined three OpenAI models, but it is plausible that even greater gains could be achieved by incorporating more heterogeneous models—such as LLMs developed by different providers, models fine-tuned for specific domains, or models trained to pursue complementary reasoning strategies. The goal for ensembles in future studies should be to include models that are each well-suited to the grading task yet maximally distinct in their inductive biases, thereby amplifying the benefits of the mixture-of-experts approach. Notably, gpt-oss-20b performed surprisingly well in our study, which is particularly encouraging because it is open-source and can be deployed on a teacher’s laptop or a university compute cluster, reducing costs and ensuring data privacy. We also found preliminary evidence that self-consistency–based uncertainty estimation may generalize across different categories of models: the older, non-reasoning gpt-4.1-nano, the proprietary reasoning gpt-5-nano, and the open-source reasoning gpt-oss-20b all exhibited meaningful links between certainty and correctness. In fact, gpt-4.1-nano—despite being substantially weaker in the single-prompt and majority-voting conditions—benefited markedly from repeated prompting, simply flagging more cases and achieving accuracies under regrading that were considerably closer to the stronger models and even to human graders. These patterns suggest that self-consistency (X. Wang et al. 2023) could be a broadly applicable mechanism for uncertainty estimation, though confirming this will require systematic evaluation across a wider range of architectures and tasks. Given these considerations, future work should continue investigating open-source models, while also examining state-of-the-art proprietary models, especially since our use of nano models for cost reasons likely underestimates the performance achievable with more capable systems.
Taken together, our findings illustrate both the promise and the complexity of incorporating LLMs into high-stakes assessment. Repeated prompting, uncertainty estimation, and selective human oversight enabled even relatively small models to approach human-level performance, while substantially reducing manual grading effort. At the same time, our results highlight that reliable AI-supported grading requires careful design choices, such as appropriate uncertainty metrics, robust thresholding strategies, and well-calibrated ensembles. As LLM capabilities continue to improve, and as open-source models become more widely accessible, there is significant potential for human–AI collaboration to make grading more accurate, efficient, and scalable. Realizing this potential will depend on continued empirical work across diverse courses, grading schemes, and model families, as well as sustained attention to fairness, transparency, and teacher agency. By building on the methods and insights presented here, future research can help shape AI-assisted assessment practices that genuinely support educators while maintaining the rigor and integrity of human judgment.

Author Contributions

Conceptualization, L.K. and E.A.; methodology, L.K.; software, L.K. and H.R.; validation, L.K.; formal analysis, L.K.; investigation, L.K.; data curation, L.K., H.R., G.G. and E.A.; writing—original draft preparation, L.K.; writing—review and editing, L.K., H.R., R.G., I.V., G.G. and E.A.; visualization, L.K.; supervision, L.K., H.R., R.G. and I.V.; project administration, L.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of the Faculty of Social and Behavioral Sciences at the University of Amsterdam (FMG-11904, July 1st 2025).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets presented in this article are not readily available because assignment will be reused by teachers in future courses and student grades were considered too sensitive.

Acknowledgments

During the preparation of this manuscript, the authors used ChatGPT 5.1 voice mode to discuss and summarize ideas and ChatGPT 5.1 to revise parts of the manuscript. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

    The following abbreviations are used in this manuscript:
LLMs Large language models
AI Artificial intelligence
SURE Selective Uncertainty-based Re-Evaluation
IRT Item Response Theory
TP True positive
FP False positive
TN True negative
FN False negative
SP single-prompt
MV majority-voting

Appendix A. LLM Prompts

We assembled the grading prompt based on the prompting condition (e.g., varying grading personas) and the respective student answer. For brevity, we only include the English prompt components, which we translated into different languages for conditions that used multilingual prompting. The base prompt included placeholders for format, personality, question, submission, and rubric (Listing 1):
Listing 1. Base prompt for LLM grading
Preprints 187686 i001
The format placeholder was replaced with a structured json response format (Listing 2):
Listing 2. JSON response format used in the format placeholder
Preprints 187686 i002
The personality placeholder was replaced with an empty string in conditions without personas and otherwise with one of the four grading personas (Listing 3):
Listing 3. Grading personas used for the personality placeholder
Preprints 187686 i003
The question, submission, and rubric placeholders were replaced with the respective text for each student answer that was scored by the LLMs.

References

  1. Alves, J. V., Leitão, D., Jesus, S., Sampaio, M. O. P., Liébana, J., Saleiro, P., Figueiredo, M. A. T., & Bizarro, P. (2025, April). A benchmarking framework and dataset for learning to defer in human-AI decision-making. Scientific Data, 12(1), 506. Available online: https://www.nature.com/articles/s41597-025-04664-y (accessed on 3 November 2025). Publisher: Nature Publishing Group . [CrossRef]
  2. Armfield, D., Chen, E., Omonkulov, A., Tang, X., Lin, J., Thiessen, E., & Koedinger, K. 2025. Avalon: A Human-in-the-Loop LLM Grading System with Instructor Calibration and Student Self-assessment. In A. I. Cristea, E. Walker, Y. Lu, O. C. Santos, & S. Isotani (Eds.), Artificial Intelligence in Education. Posters and Late Breaking Results, Workshops and Tutorials, Industry and Innovation Tracks, Practitioners, Doctoral Consortium, Blue Sky, and WideAIED (pp. 111–118). Cham: Springer Nature Switzerland. [CrossRef]
  3. Capretto, T., Piho, C., Kumar, R., Westfall, J., Yarkoni, T., & Martin, O. A. (2022, January). Bambi: A simple interface for fitting Bayesian linear models in Python. arXiv. Available online: http://arxiv.org/abs/2012.10754 (accessed on 26 November 2025). arXiv:2012.10754 [stat] . [CrossRef]
  4. Chen, J., & Mueller, J. (2024, August). Quantifying Uncertainty in Answers from any Language Model and Enhancing their Trustworthiness. IIn L.-W. Ku, A. Martins, & V. Srikumar (Eds.), Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (pp. 5186–5200). Bangkok, Thailand: Association for Computational Linguistics. Available online:https://aclanthology.org/2024.acl-long.283/ (accessed on 3 November 2025). [CrossRef]
  5. Cohn, C., S, A. T., Mohammed, N., & Biswas, G. (2025, April). CoTAL: Human-in-the-Loop Prompt Engineering for Generalizable Formative Assessment Scoring. Available online: https://arxiv.org/abs/2504.02323v3 (accessed on 28 November 2025).
  6. DeepL. (n.d.). DeepL Translate: The world’s most accurate translator. Available online: https://www.deepl.com/translator (accessed on 3 November 2025).
  7. European Parliament and Council of the European Union. (2024, July). Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act) (Text with EEA relevance). Official Journal of the European Union, 2024/1689. Available online: https://eur-lex.europa.eu/eli/reg/2024/1689/oj (accessed on 28 August 2025).
  8. Flodén, J. 2025. Grading exams using large language models: A comparison between human and AI grading of exams in higher education using ChatGPT. British Educational Research Journal 51(1), 201–224. Available online: https://onlinelibrary.wiley.com/doi/abs/10.1002/berj.4069 (accessed on 21 August 2025). [CrossRef]
  9. Fröhling, L., Demartini, G., & Assenmacher, D. (2025, August). Personas with Attitudes: Controlling LLMs for Diverse Data Annotation. In A. Calabrese, C. de Kock, D. Nozza, F. M. Plaza-del Arco, Z. Talat, & F. Vargas (Eds.), Proceedings of the The 9th Workshop on Online Abuse and Harms (WOAH) (pp. 468–481). Vienna, AustriaAssociation for Computational Linguistics. Available online: https://aclanthology.org/2025.woah-1.43/ (accessed on 3 November 2025).
  10. Grévisse, C. (2024, September). LLM-based automatic short answer grading in undergraduate medical education. BMC Medical Education 24(1), 1060. Available online: (accessed on 21 August 2025). [CrossRef]
  11. Hastie, T., Tibshirani, R., & Friedman, J. 2009. Ensemble Learning. In T. Hastie, R. Tibshirani, & J. Friedman (Eds.), The Elements of Statistical Learning: Data Mining, Inference, and Prediction (pp. 605–624). New York, NYSpringer. Available online: (accessed on 28 November 2025). [CrossRef]
  12. Holtzman, A., Buys, J., Du, L., Forbes, M., & Choi, Y. (2019, September). The Curious Case of Neural Text Degeneration. Available online: https://openreview.net/forum?id=rygGQyrFvH (accessed on 3 November 2025).
  13. Horton, P., Florea, A., & Stringfield, B. (2025, December). Conformal validation: A deferral policy using uncertainty quantification with a human-in-the-loop for model validation. Machine Learning with Applications, 22, 100733. Available online: https://www.sciencedirect.com/science/article/pii/S2666827025001161 (accessed on 3 November 2025). [CrossRef]
  14. Hossain, S. (2019, June). Visualization of Bioinformatics Data with Dash Bio. SciPy 2019. Available online: https://proceedings.scipy.org/articles/Majora-7ddc1dd1-012 (accessed on 3 November 2025). [CrossRef]
  15. Ishida, T., Liu, T., Wang, H., & Cheung, W. K. (2024, May). Large Language Models as Partners in Student Essay Evaluation. arXiv. Available online: http://arxiv.org/abs/2405.18632 (accessed on 22 November 2024). Number: arXiv:2405.18632 arXiv:2405.18632 . [CrossRef]
  16. Johnson, M., & Zhang, M. (2024, December). Examining the responsible use of zero-shot AI approaches to scoring essays. Scientific Reports 14(1), 30064. Available online: https://www.nature.com/articles/s41598-024-79208-2 (accessed on 28 November 2025). Publisher: Nature Publishing Group . [CrossRef]
  17. Kortemeyer, G., & Nöhl, J. (2025, April). Assessing confidence in AI-assisted grading of physics exams through psychometrics: An exploratory study. Physical Review Physics Education Research, 21(1), 010136. Available online: https://link.aps.org/doi/10.1103/PhysRevPhysEducRes.21.010136 (accessed on 3 November 2025). Publisher: American Physical Society . [CrossRef]
  18. Korthals, L., Rosenbusch, H., Grasman, R., & Visser, I. (2025). Grading University Students with LLMs: Performance and Acceptance of a Canvas-Based Automation. In A. I. Cristea, E. Walker, Y. Lu, O. C. Santos, & S. Isotani (Eds.), Artificial Intelligence in Education. Posters and Late Breaking Results, Workshops and Tutorials, Industry and Innovation Tracks, Practitioners, Doctoral Consortium, Blue Sky, and WideAIED (pp. 36–43).Springer Nature Switzerland. [CrossRef]
  19. Lu, Y., Bartolo, M., Moore, A., Riedel, S., & Stenetorp, P. (2022, May). Fantastically Ordered Prompts and Where to Find Them: Overcoming Few-Shot Prompt Order Sensitivity. In S. Muresan, P. Nakov, & A. Villavicencio (Eds.), Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (pp. 8086–8098). Dublin, Ireland: Association for Computational Linguistics. Available online: https://aclanthology.org/2022.acl-long.556/ (accessed on 3 November 2025). [CrossRef]
  20. Microsoft. (n.d.). Azure Machine Learning - ML as a Service | Microsoft Azure. Available online: https://azure.microsoft.com/en-us/products/machine-learning (accessed on 3 November 2025).
  21. OpenAI. (n.d.-a). Batch API - OpenAI API. Available online: https://platform.openai.com (accessed on 3 November 2025).
  22. OpenAI. (n.d.-b). GPT-4.1 nano. Available online: https://platform.openai.com/docs/models/gpt-4.1-nano (accessed on 3 November 2025).
  23. OpenAI. (n.d.-c). GPT-5 nano. Available online: https://platform.openai.com/docs/models/gpt-5-nano (accessed on 3 November 2025).
  24. OpenAI. (n.d.-d). gpt-oss-20b. Available online: https://platform.openai.com/docs/models/gpt-oss-20b (accessed on 3 November 2025).
  25. OpenAI. (n.d.-e). How should I set the temperature parameter? Available online: https://platform.openai.com/docs/faq/how-should-i-set-the-temperature-parameter (accessed on 3 November 2025).
  26. OpenAI. 2025. Using GPT-5. Available online: https://platform.openai.com (accessed on 3 November 2025).
  27. OpenAI, Hurst, A., Lerer, A., Goucher, A. P., Perelman, A., Ramesh, A., & et al. (2024, October). GPT-4o System Card. arXiv. Available online: http://arxiv.org/abs/2410.21276 (accessed on 19 February 2025). [CrossRef]
  28. Peeperkorn, M., Kouwenhoven, T., Brown, D., & Jordanous, A. (2024, May). Is Temperature the Creativity Parameter of Large Language Models? arXiv. Available online: http://arxiv.org/abs/2405.00492 (accessed on 3 November 2025). arXiv:2405.00492 [cs] . [CrossRef]
  29. Polat, M. 2020. Analysis of Multiple-Choice versus Open-Ended Questions in Language Tests According to Different Cognitive Domain Levels. Novitas-ROYAL (Research on Youth and Language), 14(2), 76–96. Available online: https://eric.ed.gov/?id=EJ1272114 (accessed on 3 November 2025). Publisher: Children’s Research Center-Turkey ERIC Number: EJ1272114.
  30. Schneider, J., Schenk, B., & Niklaus, C. (2024, July). Towards LLM-based Autograding for Short Textual Answers. arXiv. Available online: http://arxiv.org/abs/2309.11508 (accessed on 3 November 2025). arXiv:2309.11508 [cs] . [CrossRef]
  31. Strong, J., Men, Q., & Noble, A. (2025, February). Trustworthy and Practical AI for Healthcare: A Guided Deferral System with Large Language Models. arXiv. Available online: http://arxiv.org/abs/2406.07212 (accessed on 3 November 2025). arXiv:2406.07212 [cs] . [CrossRef]
  32. Team, R. C. 2022. R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing. Available online: https://www.R-project.org/ (accessed on).
  33. Tekin, S. F., Ilhan, F., Huang, T., Hu, S., & Liu, L. (2024, November). LLM-TOPLA: Efficient LLM Ensemble by Maximising Diversity. In Y. Al-Onaizan, M. Bansal, & Y.-N. Chen (Eds.), Findings of the Association for Computational Linguistics: EMNLP 2024 (pp. 11951–11966). Miami, Florida, USA: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.findings-emnlp.698/ (accessed on 3 November 2025). [CrossRef]
  34. Van Rossum, G., & Drake Jr, F. L. 1995. Python reference manual. Centrum voor Wiskunde en Informatica Amsterdam.
  35. Wang, J., Wang, J., Athiwaratkun, B., Zhang, C., & Zou, J. (2024, June). Mixture-of-Agents Enhances Large Language Model Capabilities. arXiv. Available online: http://arxiv.org/abs/2406.04692 (accessed on 3 November 2025). arXiv:2406.04692 [cs] . [CrossRef]
  36. Wang, Q., Pan, S., Linzen, T., & Black, E. (2025, November). Multilingual Prompting for Improving LLM Generation Diversity. In C. Christodoulopoulos, T. Chakraborty, C. Rose, & V. Peng (Eds.), Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing (pp. 6378–6400). Suzhou, China: Association for Computational Linguistics. Available online: https://aclanthology.org/2025.emnlp-main.324/ (accessed on 3 November 2025).
  37. Wang, X., Wei, J., Schuurmans, D., Le, Q., Chi, E., Narang, S., Chowdhery, A., & Zhou, D. (2023, March). Self-Consistency Improves Chain of Thought Reasoning in Language Models. arXiv. Available online: http://arxiv.org/abs/2203.11171 (accessed on 5 September 2025). arXiv:2203.11171 [cs] . [CrossRef]
  38. Wang, Y., Huang, J., Du, L., Guo, Y., Liu, Y., & Wang, R. (2025, December). Evaluating large language models as raters in large-scale writing assessments: A psychometric framework for reliability and validity. Computers and Education: Artificial Intelligence, 9, 100481. Available online: https://www.sciencedirect.com/science/article/pii/S2666920X25001213 (accessed on 28 November 2025). [CrossRef]
  39. Westfall, J. (2017, February). Statistical details of the default priors in the Bambi library. arXiv. Available online: http://arxiv.org/abs/1702.01201 (accessed on 3 November 2025). arXiv:1702.01201 [stat] . [CrossRef]
  40. Yang, H., Li, M., Zhou, H., Xiao, Y., Fang, Q., Zhou, S., & Zhang, R. (2025, July). Large Language Model Synergy for Ensemble Learning in Medical Question Answering: Design and Evaluation Study. Journal of Medical Internet Research, 27, e70080. Available online: https://pmc.ncbi.nlm.nih.gov/articles/PMC12337233/ (accessed on 3 November 2025). [CrossRef]
  41. Yavuz, F., Çelik, O., & Yavaş Çelik, G. 2025. Utilizing large language models for EFL essay grading: An examination of reliability and validity in rubric-based assessments. British Journal of Educational Technology, 56(1), 150–166. Available online: https://onlinelibrary.wiley.com/doi/abs/10.1111/bjet.13494 (accessed on 3 November 2025). _eprint: https://bera-journals.onlinelibrary.wiley.com/doi/pdf/10.1111/bjet.13494 . [CrossRef]
1
Graders 1, 2, and 4 graded all assignments, while grader 3 only graded the first and last assignments.
2
Assignment 4 originally had 10 questions, but we removed two because they required graders to run a program for counting lines or opening a link to evaluate a Dash app, for which our LLM grading setup was not suited.
Figure 1. Threshold tuning based on F 1 scores. Each ⋄ indicates the threshold that maximized F 1 for one of the 56 conditions. The dashed vertical line indicates the median threshold across all conditions which we used as a fixed threshold for flagging in the test set.
Figure 1. Threshold tuning based on F 1 scores. Each ⋄ indicates the threshold that maximized F 1 for one of the 56 conditions. The dashed vertical line indicates the median threshold across all conditions which we used as a fixed threshold for flagging in the test set.
Preprints 187686 g001
Figure 2. Certainty distributions for Correct and Incorrect scores across models. Proportions are normalized within each category, such that the bars for Correct and Incorrect each sum to one. Histograms use bins of 5% certainty. Dashed lines show tuned thresholds for the 56 conditions.
Figure 2. Certainty distributions for Correct and Incorrect scores across models. Proportions are normalized within each category, such that the bars for Correct and Incorrect each sum to one. Histograms use bins of 5% certainty. Dashed lines show tuned thresholds for the 56 conditions.
Preprints 187686 g002
Figure 3. Assignment-level grading accuracy across models and conditions in the training set. SURE (Δ) consistently achieves the highest accuracies, with gpt-oss-20b and the ensemble reaching human performance (grey band).
Figure 3. Assignment-level grading accuracy across models and conditions in the training set. SURE (Δ) consistently achieves the highest accuracies, with gpt-oss-20b and the ensemble reaching human performance (grey band).
Preprints 187686 g003
Figure 4. Posterior estimates of grading accuracy for human graders and LLMs under SURE with tuned certainty thresholds in the training set. The left panel shows posterior means and 95% HDIs for the probability (transformed log-odds) of scoring a student answer correctly. The right panel displays pairwise dominance probabilities, indicating for each row–column pair the posterior probability that the grader in the row is more accurate than the grader in the column.
Figure 4. Posterior estimates of grading accuracy for human graders and LLMs under SURE with tuned certainty thresholds in the training set. The left panel shows posterior means and 95% HDIs for the probability (transformed log-odds) of scoring a student answer correctly. The right panel displays pairwise dominance probabilities, indicating for each row–column pair the posterior probability that the grader in the row is more accurate than the grader in the column.
Preprints 187686 g004
Figure 6. Alignment between manually assigned human grades (grey areas), fully automated LLM grades based on single-prompt (∘), majority-voting (□), and human-in-the-loop LLM grading with SURE (Δ) in the training set. Alignment is markedly improved by SURE.
Figure 6. Alignment between manually assigned human grades (grey areas), fully automated LLM grades based on single-prompt (∘), majority-voting (□), and human-in-the-loop LLM grading with SURE (Δ) in the training set. Alignment is markedly improved by SURE.
Preprints 187686 g006
Figure 7. Certainty distributions for Correct and Incorrect scores in the test set. Proportions are normalized within each category, such that the bars for Correct and Incorrect each sum to one. Histograms use bins of 5% certainty. The vertical dashed line marks the fixed certainty threshold ( τ = 0.7 ) used for flagging in the test set.
Figure 7. Certainty distributions for Correct and Incorrect scores in the test set. Proportions are normalized within each category, such that the bars for Correct and Incorrect each sum to one. Histograms use bins of 5% certainty. The vertical dashed line marks the fixed certainty threshold ( τ = 0.7 ) used for flagging in the test set.
Preprints 187686 g007
Figure 8. Assignment-level grading accuracy in the test set. SURE (Δ) consistently achieves the highest accuracies, with the ensemble reaching human performance (grey band).
Figure 8. Assignment-level grading accuracy in the test set. SURE (Δ) consistently achieves the highest accuracies, with the ensemble reaching human performance (grey band).
Preprints 187686 g008
Figure 9. Posterior estimates of grading accuracy for human graders and LLMs under SURE with tuned certainty thresholds in the test set. The left panel shows posterior means and 95% HDIs for the probability (transformed log-odds) of scoring a student answer correctly. The right panel displays pairwise dominance probabilities, indicating for each row–column pair the posterior probability that the grader in the row is more accurate than the grader in the column.
Figure 9. Posterior estimates of grading accuracy for human graders and LLMs under SURE with tuned certainty thresholds in the test set. The left panel shows posterior means and 95% HDIs for the probability (transformed log-odds) of scoring a student answer correctly. The right panel displays pairwise dominance probabilities, indicating for each row–column pair the posterior probability that the grader in the row is more accurate than the grader in the column.
Preprints 187686 g009
Figure 10. Posterior estimates of grading bias for human graders and LLMs under SURE with tuned certainty thresholds in the test set. The left panel shows posterior means and 95% HDIs for grading bias (deviation from closest ground-truth). The right panel displays pairwise dominance probabilities, indicating for each row–column pair the posterior probability that the grader in the row is less biased (closer to zero) than the grader in the column.
Figure 10. Posterior estimates of grading bias for human graders and LLMs under SURE with tuned certainty thresholds in the test set. The left panel shows posterior means and 95% HDIs for grading bias (deviation from closest ground-truth). The right panel displays pairwise dominance probabilities, indicating for each row–column pair the posterior probability that the grader in the row is less biased (closer to zero) than the grader in the column.
Preprints 187686 g010
Figure 11. Alignment between manually assigned human grades (grey areas), fully automated LLM grades based on single-prompt (∘), majority-voting (□), and human-in-the-loop LLM grading with SURE (Δ) in the test set. Alignment is markedly improved by SURE.
Figure 11. Alignment between manually assigned human grades (grey areas), fully automated LLM grades based on single-prompt (∘), majority-voting (□), and human-in-the-loop LLM grading with SURE (Δ) in the test set. Alignment is markedly improved by SURE.
Preprints 187686 g011
Table 1. Model configurations and diversification settings.
Table 1. Model configurations and diversification settings.
LLM temperature top_p text verbosity shuffled rubrics varied personas varied languages n conditions
Prompting conditions
gpt-4.1-nano 0 / 1 0.1 / 1 - no / yes no / yes no / yes 24
gpt-5-nano - - low / medium no / yes no / yes no / yes 16
gpt-oss-20b - - medium no / yes no / yes no / yes 8
Post-hoc conditions
ensemble 1 (gpt-4.1-nano) 1 (gpt-4.1-nano) medium (gpt-5-nano & gpt-oss-20b) no / yes no / yes no / yes 8
LLMs in prompting conditions were queried via API. The post-hoc ensemble was derived from the existing outputs of matched conditions of all three LLMs.
Table 2. Illustrative grading procedure dataset: each row represents the outcome of a specific grading procedure applied to a student’s answer. Condition 1 illustrates all three grading procedures for prompting conditions. Condition 2000 illustrates ensemble conditions, for which single-prompt grading is missing as it does not make practical sense.
Table 2. Illustrative grading procedure dataset: each row represents the outcome of a specific grading procedure applied to a student’s answer. Condition 1 illustrates all three grading procedures for prompting conditions. Condition 2000 illustrates ensemble conditions, for which single-prompt grading is missing as it does not make practical sense.
student question condition procedure correct error
1 #R23 1 SP 0 -0.5
1 #R23 1 MV 0 0.25
1 #R23 1 SURE 1 0
2000 #R23 1 MV 0 0.25
2000 #R23 1 SURE 1 0
Table 3. Illustrative rows from the condition-level dataset. The first three rows show variations of temperature and rubric shuffling for gpt-4.1-nano; the last row shows an ensemble condition without prompt perturbations.
Table 3. Illustrative rows from the condition-level dataset. The first three rows show variations of temperature and rubric shuffling for gpt-4.1-nano; the last row shows an ensemble condition without prompt perturbations.
condition llm temp topp verb shuf pers lang
1 gpt-4.1-nano 0 1 0 0 0 0
2 gpt-4.1-nano 1 1 0 0 0 0
3 gpt-4.1-nano 1 1 0 1 0 0
2000 ensemble 1 1 1 0 0 0
Table 4. Illustrative grader dataset: each row shows the score a student answer received from a human grader (grader-1, grader-2, grader-3, grader-4) or the human-in-the-loop SURE protocol with a given LLM (gpt-4.1-nano, gpt-5-nano, gpt-oss-20b, ensemble). Values for illustration purposes and do not show real data.
Table 4. Illustrative grader dataset: each row shows the score a student answer received from a human grader (grader-1, grader-2, grader-3, grader-4) or the human-in-the-loop SURE protocol with a given LLM (gpt-4.1-nano, gpt-5-nano, gpt-oss-20b, ensemble). Values for illustration purposes and do not show real data.
student question grader correct error
1 #R23 grader-1 0 0.25
1 #R23 grader-2 0 -0.5
1 #R23 grader-3 1 0
1 #R23 grader-4 1 0
1 #R23 gpt-4.1-nano 0 -0.75
1 #R23 gpt-5-nano 0 0.25
1 #R23 gpt-oss-20b 1 0
1 #R23 ensemble 1 0
Table 5. Regression Coefficients.
Table 5. Regression Coefficients.
Coefficient Mean 2.5% HDI 97.5% HDI
HDI excludes zero
Intercept 1.601 1.199 1.980
procedure[single-prompt] -0.230 -0.300 -0.161
procedure[SURE] 1.311 1.219 1.400
llm[gpt-5-nano] 1.437 1.319 1.557
llm[gpt-oss-20b] 1.958 1.821 2.100
llm[ensemble] 1.989 1.830 2.147
topp(llm=gpt-4.1-nano; temp=1) 0.176 0.075 0.275
languages -0.080 -0.167 -0.002
procedure[SURE] : llm[ensemble] -0.742 -0.880 -0.601
procedure[SURE] : llm[gpt-5-nano] -0.632 -0.751 -0.513
procedure[SURE] : llm[gpt-oss-20b] -0.686 -0.831 -0.542
topp(llm=gpt-4.1-nano; temp=1) : procedure[single-prompt] -0.158 -0.233 -0.082
languages : procedure[single-prompt] -0.527 -0.580 -0.475
languages : llm[gpt-5-nano] 0.295 0.192 0.397
languages : llm[gpt-oss-20b] 0.283 0.160 0.397
languages : shuffle_rubrics -0.082 -0.140 -0.024
languages : topp(llm=gpt-4.1-nano; temp=1) -0.169 -0.260 -0.073
1|student_sigma 0.355 0.282 0.435
1|question_sigma 1.253 1.004 1.531
HDI includes zero
temp(llm=gpt-4.1-nano) -0.002 -0.100 0.099
verb(llm=gpt-5-nano) 0.067 -0.076 0.195
shuffle_rubrics 0.067 -0.013 0.152
personalities -0.003 -0.086 0.077
procedure[single-prompt] : llm[ensemble-3.5] -0.018 -1.987 1.846
procedure[single-prompt] : llm[gpt-5-nano] -0.098 -0.189 0.004
procedure[single-prompt] : llm[gpt-oss-20b] 0.111 -0.002 0.230
temp(llm=gpt-4.1-nano) : procedure[single-prompt] : -0.003 -0.072 0.075
temp(llm=gpt-4.1-nano) : procedure[SURE] 0.010 -0.078 0.113
temp(llm=gpt-4.1-nano) : shuffle_rubrics -0.034 -0.125 0.060
temp(llm=gpt-4.1-nano) : personalities 0.011 -0.085 0.101
temp(llm=gpt-4.1-nano) : languages 0.021 -0.076 0.111
topp(llm=gpt-4.1-nano; temp=1) : procedure[SURE] 0.035 -0.065 0.130
topp(llm=gpt-4.1-nano; temp=1) : shuffle_rubrics -0.030 -0.125 0.063
topp(llm=gpt-4.1-nano; temp=1) : personalities 0.000 -0.099 0.088
verb(llm=gpt-5-nano) : procedure[single-prompt] : -0.048 -0.161 0.070
verb(llm=gpt-5-nano) : procedure[SURE] -0.035 -0.175 0.112
verb(llm=gpt-5-nano) : shuffle_rubrics -0.081 -0.196 0.037
verb(llm=gpt-5-nano) : personalities -0.062 -0.173 0.062
verb(llm=gpt-5-nano) : languages 0.047 -0.065 0.164
shuffle_rubrics : procedure[SURE] 0.032 -0.032 0.099
shuffle_rubrics : procedure[single-prompt] : -0.014 -0.065 0.036
shuffle_rubrics : llm[ensemble-3.5] -0.127 -0.269 0.022
shuffle_rubrics : llm[gpt-5-nano] 0.004 -0.099 0.121
shuffle_rubrics : llm[gpt-oss-20b] -0.030 -0.153 0.092
shuffle_rubrics : personalities -0.007 -0.062 0.051
personalities : procedure[SURE] -0.009 -0.077 0.053
personalities : procedure[single-prompt] : 0.002 -0.053 0.053
personalities : llm[ensemble-3.5] 0.123 -0.023 0.260
personalities : llm[gpt-5-nano] 0.015 -0.093 0.121
personalities : llm[gpt-oss-20b] 0.064 -0.059 0.181
personalities : languages -0.023 -0.080 0.037
languages : procedure[SURE] -0.036 -0.106 0.022
languages : llm[ensemble-3.5] 0.101 -0.046 0.240
1|condition_sigma 0.027 0.000 0.052
Table 6. Alignment of LLM grades with human grade target ranges under different procedures for assignment 1.
Table 6. Alignment of LLM grades with human grade target ranges under different procedures for assignment 1.
LLM Grading Procedure % in Target Range Maximum Grade Deviation Median Grade Deviation
Assignment 1
gpt-4.1-nano SP 6.522 1.9 0.85
MV 8.696 1.2 0.5
SURE 60.870 0.4 0.1
gpt-5-nano SP 19.565 1.1 0.40
MV 47.826 0.7 0.1
SURE 60.870 0.5 0.1
gpt-oss-20b SP 54.348 0.7 0.1
MV 52.174 0.6 0.1
SURE 73.913 0.3 0
ensemble MV 45.652 0.7 0.1
SURE 73.913 0.4 0
Table 7. Manual grading time and regrading time after SURE (minutes) with time savings for assignment 1.
Table 7. Manual grading time and regrading time after SURE (minutes) with time savings for assignment 1.
Regrading (min) and time savings (%)
Grader Manual (min) gpt-4.1-nano gpt-5-nano gpt-oss-20b Ensemble
Assignment 1
Grader 1 186 85 (54%) 22 (88%) 19 (90%) 22 (88%)
Grader 2 195 95 (51%) 25 (87%) 24 (88%) 26 (87%)
Grader 3 399 203 (49%) 56 (87%) 57 (86%) 68 (83%)
Grader 4 238 115 (52%) 30 (87%) 33 (86%) 38 (84%)
Table 8. Alignment of LLM grades with human grade target ranges under different procedures for assignments 2–5.
Table 8. Alignment of LLM grades with human grade target ranges under different procedures for assignments 2–5.
LLM Grading Procedure % in Target Range Maximum Grade Deviation Median Grade Deviation
Assignment 2
gpt-4.1-nano SP 4.444 2.8 1.1
MV 8.889 2.8 1.0
SURE 42.222 1.2 0.2
gpt-5-nano SP 33.333 1.8 0.3
MV 46.667 1.7 0.3
SURE 73.333 1.2 0.1
gpt-oss-20b SP 55.556 1.1 0.2
MV 57.778 1.2 0.2
SURE 75.556 1.0 0.1
ensemble MV 55.556 1.4 0.2
SURE 91.111 0.4 0.1
Assignment 3
gpt-4.1-nano SP 4.348 2.8 0.85
MV 15.217 2.4 0.7
SURE 71.739 0.6 0.1
gpt-5-nano SP 8.696 2.7 0.8
MV 13.043 1.8 0.4
SURE 50 1.2 0.2
gpt-oss-20b SP 26.087 1.5 0.3
MV 32.609 1.7 0.3
SURE 52.174 1.4 0.1
ensemble MV 26.087 1.8 0.3
SURE 89.13 0.3 0.1
Assignment 4
gpt-4.1-nano SP 4.348 3.1 1.55
MV 2.174 2.5 1.3
SURE 19.565 1.6 0.6
gpt-5-nano SP 19.565 4.4 0.8
MV 34.783 3.8 0.3
SURE 73.913 1.9 0.0
gpt-oss-20b SP 54.348 3.5 0.3
MV 47.826 3.2 0.3
SURE 60.87 1.3 0.0
ensemble MV 54.348 3.2 0.3
SURE 91.304 0.4 0.0
Assignment 5
gpt-4.1-nano SP 21.739 2.1 0.65
MV 10.87 2.3 0.65
SURE 54.348 0.9 0.2
gpt-5-nano SP 32.609 2.6 0.45
MV 47.826 1.2 0.2
SURE 69.565 0.9 0.1
gpt-oss-20b SP 58.696 1.5 0.2
MV 76.087 0.8 0.1
SURE 80.435 0.4 0.0
ensemble MV 47.826 0.9 0.2
SURE 84.783 0.4 0.0
Table 9. Manual grading time and regrading time after SURE (minutes) with time savings for assignments 2–5.
Table 9. Manual grading time and regrading time after SURE (minutes) with time savings for assignments 2–5.
Regrading (min) and time savings (%)
Grader Manual (min) gpt-4.1-nano gpt-5-nano gpt-oss-20b Ensemble
Assignment 2
Grader 1 137 59 (57%) 34 (75%) 22 (84%) 62 (55%)
Grader 2 224 102 (54%) 50 (78%) 30 (87%) 95 (58%)
Grader 4 194 93 (52%) 39 (80%) 27 (86%) 82 (58%)
Assignment 3
Grader 1 323 168 (48%) 70 (78%) 51 (84%) 163 (50%)
Grader 2 380 214 (44%) 91 (76%) 67 (82%) 201 (47%)
Grader 4 253 147 (42%) 67 (74%) 49 (81%) 142 (44%)
Assignment 4
Grader 1 125 57 (54%) 66 (47%) 27 (78%) 89 (29%)
Grader 2 145 65 (55%) 87 (40%) 30 (79%) 107 (26%)
Grader 4 94 39 (59%) 50 (47%) 21 (78%) 69 (27%)
Assignment 5
Grader 1 162 91 (44%) 44 (73%) 31 (81%) 89 (45%)
Grader 2 185 96 (48%) 58 (69%) 37 (80%) 99 (46%)
Grader 3 294 160 (46%) 89 (70%) 51 (83%) 161 (45%)
Grader 4 192 99 (48%) 55 (71%) 36 (81%) 105 (45%)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated