Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

ChatGPT: The End of Online Exam Integrity?

Version 1 : Received: 22 April 2024 / Approved: 22 April 2024 / Online: 22 April 2024 (10:29:15 CEST)

How to cite: Susnjak, T. ChatGPT: The End of Online Exam Integrity?. Preprints 2024, 2024041388. https://doi.org/10.20944/preprints202404.1388.v1 Susnjak, T. ChatGPT: The End of Online Exam Integrity?. Preprints 2024, 2024041388. https://doi.org/10.20944/preprints202404.1388.v1

Abstract

Background: This study addresses the burgeoning impact of Large Language Models (LLMs) such as ChatGPT on the integrity of online examinations and seeks to demonstrate latent and advanced LLM reasoning capabilities. The research devises an iterative self-reflective strategy for invoking critical thinking and higher-order reasoning in LLMs based on multimodal exam questions using visuals and text. Methods: The proposed strategy was demonstrated and evaluated on real exam questions by subject experts and the performance of ChatGPT (GPT-4) with vision was estimated on an additional dataset of 600 text descriptions of multimodal exam questions, focusing on its response to complex, visual, and textual prompts. Results: The results indicate that latent multi-hop reasoning capabilities can be invoked using the proposed self-reflective strategy, which can effectively steer ChatGPT towards correct answers by integrating critical thinking from each modality into the final response. Meanwhile, ChatGPT demonstrated considerable proficiency in being able to answer multimodal exam questions across 12 subjects. Conclusions: This work challenges findings from recent studies which claim that existing LLMs are unable to reason on complex multimodal tasks, and therefore this study underscores the need for enhanced online exam security measures like proctoring software, that can mitigate the potential misuse of AI technologies in educational settings.

Keywords

AI impact on education; ChatGPT in exam cheating; multimodal online assessments; GPT-4 vision evaluation; educational integrity; Generative AI and examination security; large language models and assessments; critical thinking in LLMs; multi-hop reasoning in LLMs

Subject

Social Sciences, Education

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.