Submitted:
13 April 2026
Posted:
15 April 2026
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Background and Related Work
2.1. Assessment Validity Under Conditions of Routine AI Support
2.2. Governance, Policy, and the Limits of Prohibition
2.3. Why Detection-Led Responses Are Inadequate
2.4. Assessment Redesign, AUTHENTICITY, and Transparency
2.5. Positioning of the Present Study
3. Why Product-Only Assessment Has Become Fragile
3.1. Outsourcing of Cognitive Work and Unearned Fluency
3.2. Hallucinated Content and Fabricated Academic Signals
3.3. Shortcut Solutions in Quantitative and Computational Tasks
3.4. Equity, Access, and Inconsistency
3.5. Erosion of Construct Validity
3.6. Design Requirements Derived from These Failures
4. A Process-Based Framework for AI-Aware Assessment
4.1. Overview
4.2. Core Principles
4.3. The Process-Based Assessment Cycle
4.4. Redesign Patterns Toolkit
4.5. Failure Modes and Redesign Responses
| Failure mode | Threat to validity | High-value redesign response | Typical evidence |
|---|---|---|---|
| Outsourced prose and unearned fluency | Product quality no longer reflects underlying reasoning | Staged drafting + oral validation | Issue framing, outline, draft notes, mini exams |
| Fabricated references or claims | False academic signals may be rewarded | Source annotation + verification log | Annotated bibliography, checking notes, correction memo |
| Quantitative shortcutting | Correct-looking output without interpretive competence | Verification-by-design + checkpoint | Commented code, model choice note, diagnostic explanation |
| Silent AI orchestration advantages | Grades may reflect AI literacy or access rather than outcomes | Transparent AI-use rules + rubric alignment | Disclosure note, prompt summary, decision log |
| Construct drift | Assessment begins to measure unintended constructs | Rebalance toward process evidence and contextual performance | Checkpoint artefacts, targeted questioning, applied response |
5. Practical Toolkit: Prompts, Rubrics, and Transparency Mechanisms
5.1. Designing an AI-Aware Task Brief
5.2. Reasoning-Focused Rubric Architecture
5.3. Transparency Mechanisms
5.4. Moderation, Workload, and Reliability
5.5. Minimal Viable Redesign
| Criterion | What excellent performance shows | Why it matters in AI-aware assessment | Suggested weighting |
|---|---|---|---|
| Conceptual understanding | Accurate framing of the problem, concepts, and relevant theory | Reduces the risk that fluency masks misunderstanding | 20-25% |
| Method or approach | Appropriate selection and justified use of method, evidence, or procedure | Shows judgment rather than copied procedure | 15-20% |
| Transparency of process and AI use | Clear account of stages, AI support, and author decisions | Makes the evidential chain visible | 10-15% |
| Interpretation and justification | Explains results, assumptions, trade-offs, and limitations | Rewards reasoning rather than surface polish | 20-25% |
| Verification behavior | Checks outputs, sources, calculations, or claims systematically | Counters hallucinations and shortcutting | 15-20% |
| Communication and disciplinary quality | Presents the work coherently and appropriately for the field | Maintains standards without over-rewarding polish alone | 10-15% |
6. Applied Illustrations from Higher Education
6.1. Illustration 1: Writing-Intensive Essay in Economics or Business
6.2. Illustration 2: Quantitative Statistics Assignment
6.3. Illustration 3: Team-Based Applied Project in Tourism or Management
6.4. Cross-Case Lesson
| Task type | Main GenAI risk | Redesign pattern | Validation moment | Main competence preserved |
|---|---|---|---|---|
| Policy essay | Generated argument and fabricated references | Issue framing + argument map + final essay | Short oral defence | Causal reasoning and evidence evaluation |
| Statistics report | Generated code and shallow interpretation | Pre-analysis note + code log + report | In-class model checkpoint | Method selection and result interpretation |
| Team consultancy project | Uneven participation hidden by AI-assisted production | Project charter + decision log + report | Individual reflective defence | Applied judgment and accountable collaboration |
7. Institutional Integration and Policy Alignment
8. Discussion
9. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Giannakos, M.; Cukurova, M.; Papamitsiou, Z.; et al. The promise and challenges of generative AI in education. Behav. Inf. Technol. 2025, 44, 2518–2544. [Google Scholar] [CrossRef]
- UNESCO. Guidance for Generative AI in Education and Research; UNESCO: Paris, France, 2023. [Google Scholar] [CrossRef]
- OECD. OECD Digital Education Outlook 2023: Towards an Effective Digital Education Ecosystem; OECD Publishing: Paris, France, 2023. [Google Scholar] [CrossRef]
- Lodge, J.M.; Howard, S.; Bearman, M.; Dawson, P.; Associates. Assessment Reform for the Age of Artificial Intelligence; Tertiary Education Quality and Standards Agency (TEQSA): Melbourne, Australia, 2023. [Google Scholar]
- Lodge, J.M. The Evolving Risk to Academic Integrity Posed by Generative Artificial Intelligence: Options for Immediate Action; Tertiary Education Quality and Standards Agency (TEQSA): Melbourne, Australia, 2024. [Google Scholar]
- Ullah, M.; Bin Naeem, S.; Kamel Boulos, M.N. Assessing the Guidelines on the Use of Generative Artificial Intelligence Tools in Universities: A Survey of the World’s Top 50 Universities. Big Data Cogn. Comput. 2024, 8, 194. [Google Scholar] [CrossRef]
- Luo, J. A critical review of GenAI policies in higher education assessment: a call to reconsider the “originality” of students’ work. Assessment & Evaluation in Higher Education 2024, 49, 651–664. [Google Scholar] [CrossRef]
- Corbin, T.; Dawson, P.; Liu, D. Talk is cheap: Why structural assessment changes are needed for a time of GenAI. Assessment & Evaluation in Higher Education 2025, 50, 1087–1097. [Google Scholar] [CrossRef]
- Perkins, M.; Furze, L.; Roe, J.; MacVaugh, J. The Artificial Intelligence Assessment Scale (AIAS): A framework for ethical integration of generative AI in educational assessment. Journal of University Teaching & Learning Practice 2024, 21, 6. [Google Scholar] [CrossRef]
- Perkins, M.; Roe, J.; Furze, L. Reimagining the Artificial Intelligence Assessment Scale (AIAS): A refined framework for educational assessment. Journal of University Teaching & Learning Practice 2025, 22, 7. [Google Scholar] [CrossRef]
- Karunaratne, T.; Linblad, L. Imagining assessment futures through artificial intelligence in higher education teachers’ perspectives. Discover Education 2025, 4, 532. [Google Scholar] [CrossRef]
- Perez-Perez, I.; Gonzalez-Afonso, M.C.; Plasencia-Carballo, Z.; Perez-Jorge, D. Transparency mechanisms for generative AI use in higher education assessment: A systematic scoping review (2022-2026). Computers 2026, 15, 111. [Google Scholar] [CrossRef]
- Ilieva, G.; Yankova, T.; Ruseva, M.; Kabaivanov, S. A framework for generative AI-driven assessment in higher education. Information 2025, 16, 472. [Google Scholar] [CrossRef]
- Ahangama, N. Designing assessments in the generative AI era: A tailored assessment framework for ICT tertiary education. International Journal of Educational Technology in Higher Education 2026, 23, 9. [Google Scholar] [CrossRef]
- Chase, A.-M.; Galvin, K. Thinking to learn: Managing the risks of outsourcing to GenAI. Assessment & Evaluation in Higher Education 2026, 1–20. [Google Scholar] [CrossRef]
- Rivera-Galicia, L.F.; Montero, J.-M.; García-Pérez, C.; Senra-Díaz, E. (Eds.) Teaching Innovations in Economics: Integrating Artificial Intelligence and Emerging Technologies; Springer: Cham, Switzerland, 2026. [Google Scholar] [CrossRef]
- Mir Fernández, C.; Pablo Martí, F. A critical framework for pedagogical evaluation in generative environments: Integrating heuristic serendipity and assisted materiality in higher education. In Teaching Innovations in Economics: Integrating Artificial Intelligence and Emerging Technologies; Rivera-Galicia, L.F., Montero, J.-M., García-Pérez, C., Senra-Díaz, E., Eds.; Springer: Cham, Switzerland, 2026; pp. 3–23. [Google Scholar] [CrossRef]
- Cabrera, A.; García-Pérez, C.; Rivera-Galicia, L.F.; Senra-Díaz, E. Artificial intelligence applied to teaching and research in welfare economics, inequality, and poverty. In Teaching Innovations in Economics: Integrating Artificial Intelligence and Emerging Technologies; Rivera-Galicia, L.F., Montero, J.-M., García-Pérez, C., Senra-Díaz, E., Eds.; Springer: Cham, Switzerland, 2026; pp. 85–102. [Google Scholar] [CrossRef]
- Giménez Baldazo, M. When AI takes a seat at the desk: Innovating business education. In Teaching Innovations in Economics: Integrating Artificial Intelligence and Emerging Technologies; Rivera-Galicia, L.F., Montero, J.-M., García-Pérez, C., Senra-Díaz, E., Eds.; Springer: Cham, Switzerland, 2026; pp. 347–364. [Google Scholar] [CrossRef]
- Mir, C.; Pablo-Martí, F. Evaluación en tiempos de IA. Perspectivas SCCS 2025, 2505, July. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).