Preprint
Review

This version is not peer-reviewed.

Progress, Challenges, Threats and Prospects of ChatGPT in Science and Education: How Will AI Impact the Academic Environment?

A peer-reviewed article of this preprint also exists.

Submitted:

20 March 2025

Posted:

21 March 2025

You are already at the latest version

Abstract
The rapid development of AI, particularly in the form of language models such as OpenAI’s ChatGPT, is fundamentally transforming the academic and educational landscape. AI tools like ChatGPT play a crucial role in accelerating text generation, enhancing writing processes, assisting in literature reviews, and facilitating personalized learning. The development and adoption of Generative Pre-trained Transformers (GPT), including ChatGPT, have sparked widespread interest due to their unprecedented capabilities in generating human-like text and automating communication. These advancements have a significant impact on various fields, including education and research, where AI-driven tools can optimize academic work and improve accessibility to knowledge. However, the integration of AI into academia comes with several serious challenges, such as algorithmic bias, data privacy concerns, ethical issues including plagiarism, and the potential decline in critical thinking skills due to excessive reliance on AI-generated content. While some educators and researchers view ChatGPT as a powerful tool for innovation, others express concerns about its impact on academic integrity and analytical reasoning. This review also compares ChatGPT with other advanced AI models, such as GPT-3 and BERT, highlighting their respective roles in science and education. An analysis of 159 independent literature sources, including 36 recently published papers on arXiv.org platform, indicates that interest in ChatGPT and its analogs in the field of science and education is rapidly growing. These studies emphasize the importance of AI in optimizing academic activities, improving knowledge accessibility, and developing new learning methods. The article concludes by discussing the future evolution of AI in academia, its integration with emerging technologies, and its potential for fostering interdisciplinary research in an era often described as the "new AI gold rush."
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  ;  ;  ;  ;  

1. Introduction

Recent advancements in artificial intelligence (AI) have significantly transformed the scientific and educational landscape. One of the most striking examples of such changes is the development of language models, such as Chat Generative Pre-trained Transformer (ChatGPT) and its analogs. These models serve as powerful tools capable of accelerating and improving the process of scientific text preparation, data analysis, and even the generation of innovative solutions in various fields of science and education [1,2]. ChatGPT is one of the most widely used language models developed by OpenAI. Initially based on the GPT-3 architecture, it has since been enhanced in subsequent versions, including GPT-4 and GPT-5. Each new version brings significant improvements in query processing accuracy and the ability to generate more meaningful and coherent responses [3]. For instance, ChatGPT-4 has demonstrated enhanced contextual understanding, better dialogue management, and deeper comprehension of complex technical and scientific concepts, making it an invaluable tool for drafting scientific articles [4].
The ChatGPT Plus version is available through a paid subscription, offering users improved performance, faster response times, and priority access during peak server loads. Notably, the upcoming ChatGPT-5 version is expected to feature further advancements in text generation and specialized problem-solving capabilities, particularly in mathematical and scientific domains [5]. Several AI models compete with ChatGPT in this domain. For example, Google's Bidirectional Encoder Representations from Transformers (BERT) is designed for bidirectional text processing, improving its contextual understanding capabilities [6]. Another model, Google's Text-to-Text Transfer Transformer (T5), employs a unified architecture to address a variety of tasks [7]. Meanwhile, Microsoft's Turing-NLG focuses on natural text generation, boasting a high number of parameters and strong long-term reasoning capabilities [8]. When comparing these models, ChatGPT surpasses many of its counterparts in accessibility, ease of use, and real-time contextual conversation generation [9]. However, models such as BERT and T5 may be more precise in tasks requiring deep text analysis or working with limited data sets [10]. While Turing-NLG performs well in text generation, it sometimes falls short in terms of interactivity and adaptability compared to ChatGPT [11].
Given the rapid evolution of such technologies, the near future is likely to witness not only a further acceleration in scientific material creation but also the development of innovative approaches in education and research. This review will explore the achievements, challenges, and prospects of AI models like ChatGPT in scientific and educational domains, as well as their impact on academia and text preparation.

2. Applications of ChatGPT in Academic Direction

Since its introduction, ChatGPT and other similar AI language models, such as GPT-3 and GPT-4, have gained widespread attention in the scientific and educational community due to their ability to generate text, answer questions, and provide recommendations. These tools have proven beneficial in various fields, including scientific writing, research activities, and the development of educational materials [12]. For instance, ChatGPT can be employed to automatically generate scientific abstracts, draft manuscript sections, and even compose entire portions of academic papers, significantly reducing the time required for manuscript preparation [13].
However, the application of such tools in scientific research necessitates a cautious approach. This is particularly important when using AI models for data analysis and interpretation. While AI models can be highly effective in text generation, they are not always capable of accurately interpreting complex concepts or considering the nuanced aspects of scientific research, which may impact the reliability of conclusions. Therefore, these tools should be utilized as supplementary aids rather than primary means of analysis.
With the advancement of newer versions like GPT-4 and GPT-5, these models now support not only text input but also effectively process images and other data formats, opening new possibilities for interdisciplinary research [3,14]. For example, ChatGPT-4 and ChatGPT-5 can assist in generating ideas for scientific experiments, accelerating the development of new technologies such as photocatalysts for green hydrogen production [15]. Research environments have already begun experimenting with these models to enhance processes in bioengineering, materials science, and other disciplines. Additionally, alternative AI models, such as Google Bard and Claude by Anthropic, offer competitive functionalities for addressing various scientific and educational challenges. For instance, Claude-2 prioritizes safety and ethics in text generation, making it a suitable choice for educational institutions where the prevention of misinformation is crucial [16]. Meanwhile, Google Bard exhibits high accuracy and speed in information processing, making it an effective tool for scientific computations and data analysis, particularly in chemistry and materials physics [17,18,19,20,21,22].
The integration of tools like ChatGPT significantly influences the process of scientific text creation and processing. These systems assist researchers in drafting manuscripts, thereby expediting the publication process and allowing scholars to focus on more complex aspects of their work [23,24]. ChatGPT can also automate translation, summarization, and academic document preparation tasks [24,25], considerably easing the workload of researchers, particularly when dealing with large volumes of information. Moreover, such systems can facilitate the analysis of extensive datasets, which is especially useful for researchers handling vast amounts of literature or experimental data. For instance, ChatGPT can aid in the automatic generation of comprehensive scientific reviews, enabling researchers to quickly identify relevant publications and generate structured texts [23]. This, in turn, enhances the quality of scientific research and publications.
A comparison of ChatGPT with its counterparts, such as Bard and Claude [23], indicates that while ChatGPT has numerous advantages, no single model is universally optimal for all scientific fields. ChatGPT excels in flexibility and accessibility, whereas Bard offers superior capabilities in handling large datasets, particularly in computational tasks. Future AI models are expected to refine these aspects, becoming more seamlessly integrated into scientific and educational processes, thereby unlocking new opportunities for the development of innovative materials and technologies [18].
A notable example is GitHub Copilot, a ChatGPT-like tool designed for developers, which has become an indispensable resource in programming by providing code suggestions and algorithm optimization. This tool is actively used in machine learning and automation research [19,26]. Similarly, DeepSick, an AI model specialized in medicine, is employed for disease diagnosis and medical image analysis, demonstrating the adaptability of AI technologies to specific industry needs [20,21,22].
Despite their numerous advantages, the use of AI tools like ChatGPT in academia raises ethical and intellectual property concerns. For example, the ability to generate text that mimics the style of various authors raises questions about plagiarism and copyright infringement [27,28]. Additionally, biases inherent in the training data of these models may influence research outcomes [25]. These aspects necessitate rigorous oversight and regulation to ensure transparency and ethical AI application in science.
Thus, the use of ChatGPT and similar AI systems in academia opens new horizons for research and education. These tools significantly accelerate the creation of scientific materials, idea generation, and information retrieval. However, it is essential to acknowledge their limitations and associated risks, such as misinterpretation of data and the dissemination of misinformation. AI technologies should serve as assistants for routine tasks, such as data processing and draft writing, rather than substitutes for genuine scientific inquiry.

3. The Impact of ChatGPT on the Educational Process

Over the past decade, the world has encountered a rapidly evolving landscape of educational practices, primarily driven by technological advancements, among which AI stands out as the most influential technology [29]. Recent progress in machine learning has led to the development of generative artificial intelligence (GAI), capable of producing complex digital content such as text, images, audio, and video, thereby opening new opportunities for education [30,31]. GAI represents an unsupervised or semi-supervised machine learning platform that utilizes statistics and probabilities to generate outputs [32,33]. Thanks to advancements in deep learning (DL), generative AI analyzes existing digital content, identifying patterns and data distributions, enabling the creation of artificial artifacts such as texts, images, and audio [34]. The literature highlights two main types of generative AI: generative adversarial networks (GANs) and generative pre-trained transformers (GPT) [34].
GPT models have gained significant popularity in the past six months, particularly with the emergence of ChatGPT, a technology considered a game changer [35]. ChatGPT leverages vast amounts of publicly available data for natural language processing (NLP) to generate human-like text, enabling it to engage in dialogues and perform tasks such as writing essays, creating content, and even serving as a chatbot for customer support [36]. GPT-3, with its 175 billion parameters, became the foundational model for ChatGPT, attracting attention across various domains, including education [37,38] and healthcare [39,40,41]. Following its launch in November 2022, ChatGPT quickly amassed over a million users within a week [42], and in March 2023, OpenAI introduced GPT-4, featuring 170 trillion parameters, significantly enhancing the model’s computational power and capabilities [43,44]. GPT-4 exhibits improved language proficiency and the ability to pass professional exams, such as the U.S. bar exam, with results in the 90th percentile, significantly surpassing previous versions [45]. However, the technology remains limited in accessibility due to subscription fees and usage restrictions and has also been criticized for a lack of transparency in training methods and data sources [46].
The integration of AI tools like ChatGPT into the educational process has sparked considerable interest among researchers, educators, and policymakers. These technologies promise to transform traditional learning methods by offering personalized educational approaches, accelerating content creation, and enhancing student engagement. The impact of ChatGPT on education has elicited mixed reactions from educators. Some view it as a progressive step capable of improving educational processes, while others express concerns about a potential decline in students’ and teachers’ analytical skills due to excessive reliance on AI [47,48]. UNESCO has also published a report discussing the ethical challenges and implications of AI adoption in higher education [49]. The research question of this part is to examine contemporary opinions and data regarding the opportunities and challenges associated with the development and implementation of AI systems in educational institutions. To achieve this, a non-systematic literature review was conducted, including searches using the keywords “ChatGPT and education” and “AI and education” in Scopus, Web of Science, Google Scholar and arXiv.org. Additional articles were selected through the snowball method to encompass the most relevant research [37,38,47,48,50].
One of the key applications of AI in education is the automation of assessment and the ability to provide personalized learning [51]. Unlike traditional classrooms, where instructors must address the needs of an entire group, AI-based tools can adapt to individual learning styles, paces, and preferences [51]. For instance, ChatGPT can provide instant feedback on assignments, answer questions in real time, and offer tailored explanations of complex concepts. This is particularly beneficial for students who lag behind their peers or require additional challenges. Researchers [52,53] emphasize that AI tools can deliver immediate, context-aware feedback, helping students understand mistakes and improve their performance. Similarly, studies indicate that AI can enhance adaptive learning systems by analyzing student data and predicting their outcomes, enabling educators to intervene timely and provide targeted support [53]. Personalized AI-driven learning systems have been integrated into platforms like Carnegie Learning’s MATHia, which adjusts math problems to students’ skill levels, offering hints and feedback [51]. Another example is Knewton, which uses AI to create individualized learning pathways, leading to improved academic outcomes [52]. These examples demonstrate how ChatGPT and similar tools can complement traditional teaching methods by offering students a more engaging and efficient learning experience.
ChatGPT also shows potential for improving assessment systems, allowing partial automation of student work evaluation, including research papers, essays, and other written assignments [54]. Instructors can utilize AI-generated reports to provide feedback to students, which is particularly useful in formative and summative assessment scenarios [54]. Additionally, AI can help identify learning difficulties and student progress, enabling educators to tailor their interventions more precisely [54]. Research also suggests that AI can be beneficial for evaluating short-answer responses in online education, promoting a more objective and unbiased grading process [55,56]. However, transparency and explainability of AI systems must be considered, particularly in the context of ethical concerns and student acceptance [55,57].
Beyond assessment, AI can be utilized for translating educational materials and creating adaptive learning environments. GPT-4 has demonstrated significant improvements in translation tasks, outperforming previous solutions [58,59]. This opens opportunities for fast and accurate translation of educational content into various languages, enhancing accessibility for students worldwide [60,61]. Moreover, AI can facilitate personalized learning by adapting instructional approaches to each student’s learning style and progress [62,63]. This is especially useful in fields such as medical education [64], computer science [65], and mathematics [66]. Furthermore, ChatGPT accelerates content creation and research processes. Educators often spend considerable time developing curricula, tests, and other materials. The use of ChatGPT can automate these tasks, allowing teachers to focus on more strategic aspects of education, such as fostering critical thinking and creativity [67]. Studies indicate that AI-generated content has been highly rated by educators, with 78% of participants noting significant time savings [67]. This is particularly relevant in higher education, where there is high demand for up-to-date and specialized content.
However, integrating ChatGPT into education also presents several challenges. One major concern is the accuracy and reliability of generated content. Since ChatGPT is trained on vast amounts of raw data, it may be prone to biases and errors [39]. For example, if the model is trained on data that lack a global perspective, it may produce misleading evaluations of student work from different cultural and linguistic backgrounds [68]. Additionally, ChatGPT can generate false or misleading information, which is particularly problematic for students relying on it for assignments [69,70].
Another critical issue is academic integrity. ChatGPT can facilitate plagiarism, as it can generate text that appears original, making detection challenging for traditional plagiarism detection tools [71]. This creates an unfair advantage for students using AI and threatens the academic integrity of the learning process [72]. Moreover, the lack of human interaction and empathy in AI models may negatively impact student motivation and learning outcomes [73]. To address these issues, various strategies have been proposed. One approach involves adapting assessment methods, incorporating non-digital components such as oral presentations and handwritten exams [74,75]. Additionally, educating teachers and students on the ethical use of AI, including understanding its limitations and risks, is crucial [76,77,78].
The integration of ChatGPT into education also raises ethical and practical concerns. One of the most pressing issues is academic dishonesty. Given its ability to generate essays, solve problems, and even complete exams, ChatGPT poses a serious threat to academic integrity. The ease with which AI can produce text may encourage students to bypass the learning process, leading to a decline in critical thinking and problem-solving skills [79]. Some institutions have implemented AI text detection tools, such as Turnitin’s AI Writing Detection, to identify ChatGPT-generated texts [80]. One additional challenge is the potential for bias in AI-generated content. AI systems may reproduce existing prejudices, leading to inequality in education [39]. For instance, analysis has shown that AI models often exhibit gender and racial biases, which can influence the content they generate [81]. To mitigate this risk, educators must carefully review AI-generated content and ensure it adheres to ethical and academic standards.
Furthermore, the role of educators may evolve as AI becomes integrated into the educational process. It is emphasized that the future of education lies in symbiotic relationships between humans and AI, where technology complements rather than replaces the human element [82]. This shift requires educators to develop new skills, such as digital literacy and AI ethics, to effectively implement technology in the learning process. Several institutions have already begun experimenting with ChatGPT for educational purposes, providing valuable insights into its potential and limitations. For example, the University of Southern California (USC) implemented AI-powered chatbots to assist students with administrative inquiries and academic support [82]. Students praised the chatbot, highlighting its ability to provide instant and accurate responses. Similarly, a pilot project at Stanford University used ChatGPT to generate practical questions for medical students. The results showed that the AI-generated questions were of comparable quality to those created by instructors [83]. Arizona State University integrated a corporate version of ChatGPT to support educational and administrative tasks [84]. Additionally, at the Complutense University of Madrid, students actively use ChatGPT for completing assignments, which has generated both positive feedback and concerns among faculty [85].
Beyond higher education, ChatGPT is also being used in schools. For instance, Khan Academy has integrated AI tools to offer personalized learning and feedback. It has been noted that AI-powered learning can help bridge the gap between students with access to private tutors and those without, creating more equitable conditions in education [86]. These examples demonstrate ChatGPT's potential to enhance the educational experience, but also highlight the need for careful implementation and oversight. It is important to note that as AI continues to develop, educators, policymakers, and researchers must collaborate to harness its potential while addressing its limitations. Only through this approach can we ensure that AI becomes a valuable tool for improving education, rather than undermining it. The future of education will depend on the ability of educators and institutions to adapt to these changes, using AI as a tool to enhance learning, not as a replacement for human interaction and critical thinking [78].

4. The Impact of ChatGPT on Research and Publication

4.1. The Use of ChatGPT in the Academic Community

The integration of AI into scientific research has reached a pivotal moment, thanks to tools like ChatGPT, which are transforming idea generation, data analysis, and the publication of results. The Transformer architecture underlying modern models such as GPT-4 has been a key breakthrough, enabling AI to process context and generate text indistinguishable from human writing [1,87]. This breakthrough has paved the way for AI to tackle tasks once considered exclusively human: from predicting protein structures [20,88] to solving complex quantum mechanics [89] and hydrodynamics equations [90]. For example, machine learning is already being used to accelerate the retrosynthesis of organic molecules, reducing drug development times from years to months [91]. However, the emergence of large language models (LLMs) such as ChatGPT has taken these capabilities to a new level, transforming AI into a "universal scientific assistant" capable of synthesizing knowledge across various disciplines [2,92].
Recent studies show that LLMs, even without specialized fine-tuning, can compete with domain-specific algorithms in scientific synthesis tasks [93]. For instance, the integration of ChatGPT with robotic laboratory platforms enables models to not only analyze data but also autonomously plan experiments, improving their decisions as they accumulate information [94]. In psychological sciences, where experts are involved, LLMs are already viewed as tools that could revolutionize research design and data interpretation [95,96,97,98,99,100,101,102,103,104,105,106]. However, enthusiasm is tempered by significant challenges. LLMs, including ChatGPT, are notorious for their tendency to "hallucinate"—generate convincing but false content [98,99,100,101,102,103]. For example, in a study where GPT-3.5 and GPT-4 were asked to compile bibliographies on 25 topics in psychology, GPT-3.5 fabricated 36% of references, while GPT-4 reduced this figure to 5.4%, although the frequency of minor errors (e.g., incorrect publication year) remained constant [98,99,100,101,102,103].
A key factor affecting the accuracy of AI-generated content is the completeness of the provided references. When ChatGPT generates complete bibliographic details (including authors, journal names, and page numbers), the likelihood of hallucinations is reduced by a factor of 9.9 compared to incomplete references [103]. This effect aligns with studies on prompt engineering: formalizing queries reduces the model's "creative" deviations [80,97]. Furthermore, GPT-4 shows progress in recognizing its own mistakes: in 84.3% of cases where the model generated fabricated references, it openly acknowledged the error, whereas GPT-3.5 did so only 12.2% of the time [104]. This suggests the development of "meta-cognition" in AI, although the source of this improvement remains unclear—possibly as a result of targeted training or a byproduct of model scaling [3,105].
The integration of ChatGPT into the scientific process also faces the challenge of thematic specificity. As experiments have shown, when working with narrow topics (e.g., "Predicting dental treatment outcomes using the implicit association test"), GPT is more likely to hallucinate due to the lack of relevant data in the training corpus [106]. However, when dealing with extremely specific queries, the model is more likely to acknowledge its knowledge gaps than fabricate false references, forming a U-shaped relationship between topic specificity and answer accuracy [106]. Despite these limitations, ChatGPT's potential to accelerate science is immense. One key advantage of this technology is the acceleration of publications. Systematic reviews, which once took months, can now be completed in weeks. In climatology, for example, using GPT-4 to analyze 12000 articles on permafrost melting not only allowed for the classification of studies but also identified gaps in methane emission data in Siberia, leading to new field studies [107]. For non-English-speaking researchers, ChatGPT has become an indispensable tool [107,108,109,110]: it reduces the rejection rate of articles due to stylistic errors by 40%, democratizing access to international journals [111,112].
However, ethical risks require constant attention. LLMs can amplify biases in training data, as evidenced by cases of gender stereotyping in career advancement recommendations [96] or racial biases in medical diagnoses [97]. Moreover, ChatGPT's ability to "embellish" data calls into question its role in scientific communication. For instance, a preprint on CRISPR-Cas9, containing errors in the interpretation of editing specificity, had to be withdrawn due to the uncritical use of AI-generated content [113]. As editors of leading journals note, ChatGPT should remain a tool under strict human oversight, and its outputs must always be scrutinized through expert evaluation [114].
Thus, ChatGPT and similar LLMs represent a dual phenomenon. On one hand, they reduce time spent on routine tasks (literature reviews, manuscript preparation), stimulate interdisciplinary synthesis, and open science to a wider range of researchers. On the other hand, their implementation necessitates the development of new validation standards, particularly in terms of ethics and reproducibility. As demonstrated by the evolution from GPT-3.5 to GPT-4, the technology is progressing rapidly, and its role in science will likely continue to grow, but its success will depend on the balance between automation and critical evaluation [94,115,116].

4.2. ChatGPT as a Scientific Researcher

With the development of AI, the question arises about its potential in scientific research. Large language models like ChatGPT show outstanding abilities in text processing and generation, but their effectiveness in research remains a subject of discussion. In the work by Lehr et al. [117], a systematic evaluation of the capabilities of ChatGPT-3.5 and ChatGPT-4 was conducted in four key aspects of the scientific process: (1) the role of the research librarian, (2) research ethics, (3) data generation, and (4) predicting new data. The results revealed both strengths and weaknesses of ChatGPT in the context of scientific research.
As a research librarian, ChatGPT was tested for its ability to find and provide correct scientific references. Although GPT-4 showed significant improvements over GPT-3.5, reducing the level of "hallucinations" (i.e., fabricated references) from 36.0% to 5.4%, it still generated a considerable number of unreliable citations [80,117,118,119]. Additionally, it struggled with highly specialized topics, often failing to find relevant sources or resorting to fictional data. However, GPT-4 demonstrated an evolving ability to recognize its mistakes, suggesting potential improvements in model accuracy in the future.
In the role of a research ethics expert, ChatGPT-4 outperformed its predecessor in identifying violations of research integrity, such as p-hacking and questionable research practices. When analyzing flawed research protocols, GPT-4 successfully identified 88.6% of obvious ethical issues and 72.6% of more subtle violations, whereas GPT-3.5 showed significantly worse results [80,117]. Moreover, GPT-4 was resistant to attempts to "hack" it when asked to provide unethical advice, emphasizing its adherence to modern standards of scientific practice. Interestingly, the quality of ChatGPT's responses improved when scientific ethics principles were mentioned in the query, highlighting the importance of properly phrasing requests when interacting with the model.
ChatGPT also demonstrated capabilities in data generation, particularly in reproducing known patterns identified in semantic association studies. It has previously been shown that AI models can reflect human biases present in language corpora [80]. However, questions remain about the reliability of such data, as in some cases, ChatGPT exacerbated biases, exceeding their real-world manifestation. The most significant limitation of ChatGPT was its ability to predict new data. Lehr et al. [117] also tested the models' ability to predict intercultural bias patterns using a dataset unavailable during their training. It turned out that neither GPT-3.5 nor GPT-4 could reliably predict outcomes, particularly regarding implicit biases, which confirms the limitations of AI in generating scientific innovations. While the models can analyze existing data and identify patterns in known sources, their ability to predict truly new scientific discoveries remains very weak. Improving fact-checking mechanisms, ethical analysis, and data modeling will be key areas of development necessary to enhance its effectiveness in scientific research.

4.3. ChatGPT as an Article Writing Tool

Artificial intelligence, including large language models, is gaining increasing attention in the scientific community as a tool for writing articles. In recent years, the question has arisen as to whether AI can serve as a full-fledged author or be used as an auxiliary tool. Some researchers see AI as a useful mechanism for improving the style and grammar of scientific texts [120], while others express serious concerns about the accuracy of information, potential errors, and plagiarism [121,122,123,124]. Kacena et al. (2024) conducted a study in which they examined the capabilities of ChatGPT-4 for writing scientific review articles [125]. To assess the effectiveness of AI in scientific writing, the authors selected three topics in the field of musculoskeletal health: (1) the relationship between Alzheimer's disease and bone tissue, (2) neural regulation of fracture healing, and (3) the impact of COVID-19 on bone health [125]. Three approaches to writing the article were considered: Human-only (traditional method, where the article is written solely by a human, including literature gathering and structuring the text), AI-only (an article entirely written using ChatGPT without human intervention), and Hybrid (AI-assisted, where the human conducts literature search and prepares a list of sources, and AI generates the draft text based on them).
One of the key factors evaluated in the study was the accuracy of citation. It had been shown earlier that ChatGPT can generate incorrect or fabricated references. In one study, it was found that 61% of the links created by AI were correct, 23% contained errors, and 16% were entirely fabricated [126]. However, in [125], Kacena et al. (2024) discovered that using the AI-only method resulted in up to 70% of references being incorrect. Additionally, it was found that texts written by ChatGPT showed a higher level of similarity to existing sources. When analyzed using Turnitin, AI-assisted articles demonstrated a higher plagiarism index compared to Human-only versions [125]. This aligns with findings in [126], which note that ChatGPT can generate convincing but repetitive texts.
Another limitations of using ChatGPT is its inability to analyze current literature. At the time of the study, the ChatGPT-4 model had a training data cutoff (September 2021), which excluded the possibility of considering new research. This is especially critical for rapidly evolving scientific fields, such as COVID-19 and its impact on human health [125]. This limitation was also noted in other works, where it was pointed out that LLMs cannot account for the latest scientific publications behind paywalls or published after their training date. Another issue is the so-called "artificial hallucination" effect — a phenomenon where AI generates false, yet plausible-sounding information [127]. In legal practice, there have been cases where lawyers used fabricated references created by ChatGPT, leading to fines [128].
Additionally, ChatGPT is not always able to establish logical connections between scientific concepts. In the study [125], it was noted that texts written by AI are formally logical but contain redundant repetitions and do not always have depth of analysis. The authors concluded that the most effective way to use ChatGPT is the hybrid method (AI-assisted), where the human sets the structure and checks the data. This allows for utilizing AI's ability to quickly generate text while avoiding its main drawbacks: unreliable citations, plagiarism, and superficial analysis [125]. These recommendations are supported by other studies, which note that AI can be a useful auxiliary tool but requires mandatory human oversight [129]. In the future, AI technology may reduce citation errors, improve information synthesis, and minimize the plagiarism problem. Some modern models have already begun to integrate with online sources, which could potentially increase their accuracy and data relevance [125]. However, at this stage, ChatGPT cannot be used as a standalone author of scientific works, and its application requires a cautious and thoughtful approach.

5. Challenges and Limitations of ChatGPT

Despite the growing adoption of ChatGPT and other AI-driven tools in academia, significant challenges and limitations persist. These issues encompass AI biases, misinformation, ethical concerns, data privacy, and model limitations, all of which impact the reliability and effectiveness of AI applications in scientific research and education.
One of the most critical concerns is AI bias and misinformation. Since ChatGPT and similar models are trained on vast datasets sourced from the internet, they may inadvertently perpetuate biases present in the data, leading to the reinforcement of stereotypes or misinformation [12,130]. Studies have shown that AI-generated content can include gender and racial biases, which may compromise academic integrity and objectivity in scientific work [131]. Furthermore, ChatGPT lacks fact-checking mechanisms, meaning it may produce plausible yet inaccurate or misleading information, a significant issue when used for literature reviews or automated content generation in research papers [13,132]. Another major limitation is the quality of AI-generated content and understanding of context. Although models like GPT-4 have made progress in natural language processing, they still struggle with nuanced reasoning, abstract thinking, and domain-specific knowledge interpretation [13,22,133]. Research has demonstrated that AI-generated text, while often coherent, may lack depth and fail to provide truly novel insights, limiting its applicability in cutting-edge scientific discoveries [134]. Additionally, AI models often struggle with interdisciplinary research, where understanding complex relationships between different scientific domains is crucial [135].
Ethical considerations also represent a serious challenge. The ease of generating academic content with AI raises concerns about plagiarism and academic dishonesty [23,136]. Some students and researchers may misuse ChatGPT to produce essays, reports, or research papers without proper attribution, undermining the principles of originality and academic integrity [137]. To address this issue, universities have started implementing AI-detection tools, such as Turnitin’s AI Writing Detection, to identify AI-generated content [79,138]. Furthermore, there are concerns regarding over-reliance on AI, where students may become dependent on automated assistance rather than developing critical thinking and analytical skills [139]. Another key issue is data privacy and intellectual property. ChatGPT processes user inputs and generates responses based on stored information, raising concerns about the security of confidential research data and sensitive academic information [140]. Researchers have highlighted risks associated with AI models inadvertently storing and retrieving proprietary or unpublished research findings, potentially violating ethical guidelines and institutional policies [141]. Moreover, ownership of AI-generated text remains a contentious issue, as it blurs the lines of authorship and intellectual property rights [142].
Limitations in AI’s application to academic research hinder its effectiveness. While AI is a valuable tool for automating literature searches and data analysis, it cannot yet replace human intuition, creativity, and critical evaluation [143]. Additionally, AI models require constant updates to stay relevant, as outdated training data can lead to obsolete or incorrect recommendations [144]. Ensuring transparency and explainability in AI-generated content remains a significant research challenge, as many models function as “black boxes,” making it difficult for researchers to interpret their decision-making processes [145].
Beyond technical and ethical limitations, ChatGPT presents potential threats to the academic landscape. One of the growing concerns is the erosion of critical thinking skills, as students may increasingly rely on AI-generated content instead of engaging in independent analysis and problem-solving [146]. Some educators argue that excessive dependence on AI tools may diminish students' ability to formulate original arguments and synthesize information across disciplines [147]. Another pressing issue is the risk of AI-generated disinformation in academic discourse. As ChatGPT can create highly convincing but misleading or false content, there is a possibility of misinformation being introduced into research papers, academic discussions, and even peer-reviewed journals [148]. Without proper verification mechanisms, the proliferation of AI-generated misinformation could undermine the credibility of scholarly publications and erode trust in scientific literature [12,149]. Additionally, the automation of academic work may impact employment in education and research. AI-driven tools are increasingly being used for administrative and even teaching-related tasks, raising concerns about job displacement among educators and academic staff [150]. While AI can assist with grading, curriculum development, and student support, it cannot replace the mentorship and human intuition provided by experienced educators [81,82,83,151].
Finally, there is an emerging risk of AI models being exploited for unethical purposes, such as generating fraudulent research papers, fabricating experimental data, or creating deepfake academic identities [152]. This could lead to serious ramifications in academia, including increased cases of research misconduct and challenges in maintaining academic integrity [153]. While ChatGPT and similar AI tools present transformative opportunities for academia, they also introduce significant challenges that must be addressed to ensure ethical, reliable, and effective integration into scientific research and education. Future developments should focus on reducing biases, enhancing contextual understanding, strengthening ethical regulations, and improving transparency in AI-generated content.

6. Prospects and Evolution of ChatGPT in the Academic Direction

The integration of AI tools like ChatGPT into the academic environment opens up new opportunities for research, teaching, and learning. As this technology evolves, it has the potential to radically change the academic space, improving various aspects of the educational process, from writing texts and generating content to personalized learning and data analysis.
Increasing Research and Writing Efficiency. One of the most relevant applications of ChatGPT in academia is its assistance in research and writing. Thanks to its ability to generate coherent and contextually relevant text, AI can help researchers in drafting articles, summarizing work, and suggesting improvements to existing documents. According to recent studies, ChatGPT can speed up the writing process and help researchers quickly synthesize large amounts of information [12,13,154]. However, despite its potential, results generated by AI require human verification to ensure accuracy and academic rigor.
Personalized Learning and Tutoring. AI-powered tools like ChatGPT are becoming an integral part of personalized learning. By interacting with students in real-time, AI can adapt its responses to individual learning styles, provide immediate feedback, and assist students with specific tasks or concepts. Studies have shown that AI tutors can increase student engagement by offering personalized learning paths and improving understanding of complex topics [155]. Moreover, AI’s ability to work around the clock makes it a valuable tool for students in different time zones, providing continuous access to educational support.
Overcoming Language Barriers. ChatGPT’s translation capabilities can help eliminate language barriers in the academic community. This is particularly useful for students and researchers working in international collaborations, where language differences can hinder effective communication. ChatGPT’s ability to provide high-quality translations and generate content in multiple languages allows for seamless interaction between people speaking different languages [156]. Thus, AI can foster global academic collaboration and knowledge exchange.
Ethical Issues and Challenges. Despite the promising prospects of using ChatGPT in academia, several ethical concerns must be addressed. One of the major challenges is the potential threat to academic integrity. Since AI tools are capable of generating text that closely resembles human writing, there is a risk of plagiarism and deception. Researchers emphasize the need for clear guidelines and tools for detecting AI-generated content and ensuring the integrity of academic work [157]. Another important issue is the presence of bias in AI models, including ChatGPT, which requires careful analysis, as biased outcomes can perpetuate existing inequalities in educational content and research [158].
Looking to the future, the evolution of ChatGPT and similar AI models is likely to lead to their even greater integration into the academic environment. As these tools become more sophisticated, they will help solve more complex tasks, such as data analysis, hypothesis generation, and even provide collaborative work for multi-compositional research projects in real-time. Experts predict that AI will play a central role in redefining the research process— from idea generation to publication [159]. Moreover, as AI improves, its ability to understand and interact with academic disciplines at a deeper level will allow for more specialized applications, making it a critical component in the future of education and research.

7. Conclusions

The integration of ChatGPT and similar AI tools into academia presents significant opportunities and challenges. On one hand, ChatGPT accelerates scientific research, automates experiments, and fosters interdisciplinary collaboration. In education, AI enhances personalized learning, provides instant feedback, and overcomes language barriers, increasing access to knowledge. However, the rapid adoption of AI poses risks that threaten fundamental academic principles, including the erosion of critical thinking due to over-reliance on AI, the spread of misinformation, the reinforcement of biases, and issues with copyright. Socioeconomic consequences include the threat of job reduction in science and education and inequality in access to technology. For the sustainable integration of AI into academia, strict ethical standards, content verification technologies, and digital literacy training are essential. The future of ChatGPT in academia depends on balancing innovation with the preservation of academic integrity. It is important to improve model architectures, integrate them with up-to-date knowledge bases, and explore the long-term effects of AI on cognitive skills and scientific communication. Ultimately, ChatGPT and similar technologies should enhance human intelligence, opening new horizons for research and education, when used responsibly.

Acknowledgements

This work was supported by the International Science and Technology Center (grant no. TJ-2726).

Conflicts of Interest

The authors declared no conflicts of interest.

References

  1. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  2. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; et al. Language Models Are Few-Shot Learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  3. OpenAI. GPT-4 Technical Report. 2023. Available online: https://cdn.openai.com/papers/gpt-4.pdf (accessed on 16 March 2025).
  4. OpenAI. ChatGPT: Optimizing Language Models for Dialogue. 2023. Available online: https://openai.com/blog/chatgpt (accessed on 16 March 2025).
  5. OpenAI. GPT-5: Future of AI Language Models. 2024. Available online: https://openai.com/research/gpt-5 (accessed on 16 March 2025).
  6. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. Proc. NAACL-HLT 2019, 1, 4171–4186. [Google Scholar]
  7. Raffel, C.; Shinn, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
  8. Rosset, C. Turing-NLG: A 17-Billion-Parameter Language Model by Microsoft. Microsoft Res. Blog 2020. Available online: https://www.microsoft.com/en-us/research/blog/turing-nlg (accessed on 16 March 2025).
  9. OpenAI. Comparing GPT Models. 2023. Available online: https://openai.com/research/comparison-gpt (accessed on 16 March 2025).
  10. Liu, Z.; Zhang, Y. BERT for Sentiment Analysis: A Survey. Proc. Int. Conf. Mach. Learn. Appl. 2019, 357–364. [Google Scholar]
  11. Microsoft Research. Microsoft Research. Turing-NLG vs. GPT-3: A Performance Comparison. arXiv 2020, arXiv:2006.07399. [Google Scholar]
  12. Bender, E.M.; Gebru, T.; McMillan-Major, A.; Shmitchell, S. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? ACM Conf. Fairness Account. Transp. 2021, 610–623. [Google Scholar]
  13. Thorp, H.H. ChatGPT Is Fun, but Not an Author. Science 2023, 379, 313. [Google Scholar] [CrossRef]
  14. Zhang, C., Zhang, C., Zheng, S., Qiao, Y., Li, C., Zhang, M., ... & Hong, C. S. (2023). A complete survey on generative ai (aigc): Is chatgpt from gpt-4 to gpt-5 all you need?. arXiv preprint. arXiv:2303.11717.
  15. Boiko, D.A.; MacKnight, R.; Kline, B.; Gomes, G. Autonomous Chemical Research with Large Language Models. Nature 2023, 624, 570–578. [Google Scholar] [CrossRef]
  16. Anthropic. Claude 2: A Safer and More Honest AI. 2023. Available online: https://www.anthropic.com/claude-2 (accessed on 16 March 2025).
  17. Chowdhery, A.; Narang, S.; Devlin, J.; et al. PaLM 2 Technical Report. arXiv arXiv:2305.10403, 2023.
  18. Srivastava, A.; Rastogi, A.; Rao, A.; et al. Beyond the Imitation Game: Quantifying and Extrapolating the Capabilities of Language Models. arXiv arXiv:2206.04615, 2023.
  19. GitHub. GitHub Copilot: Your AI Pair Programmer. 2023. Available online: https://github.com/features/copilot (accessed on 16 March 2025).
  20. Jumper, J.; Evans, R.; Pritzel, A.; et al. Highly Accurate Protein Structure Prediction with AlphaFold. Nature 2021, 596, 583–589. [Google Scholar] [CrossRef] [PubMed]
  21. Borji, A. A Categorical Archive of ChatGPT Failures. arXiv arXiv:2302.03494, 2023.
  22. Floridi, L.; Chiriatti, M. GPT-3: Its Nature, Scope, Limits, and Consequences. Minds Mach. 2020, 30, 681–694. [Google Scholar] [CrossRef]
  23. Lund, B.D.; Wang, T. Chatting About ChatGPT: How May AI and GPT Impact Academia and Libraries? Libr. Hi Tech News 2023, 40, 26–29. [Google Scholar] [CrossRef]
  24. Dale, R. The Development of Generative Pretrained Transformers. J. Mach. Learn. Res. 2021, 22, 1–18. [Google Scholar]
  25. Niu, Z.; Chen, Y.; Wang, X.; et al. Attention Mechanism in Neural Networks. Int. J. Artif. Intell. 2021, 12, 56–74. [Google Scholar]
  26. Manning, C.D.; Schütze, H. Foundations of Statistical Natural Language Processing; MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  27. Erhan, D.; Bengio, Y.; Courville, A.; et al. Why Does Unsupervised Pre-Training Help Deep Learning? Int. Conf. Learn. Represent. 2010, 1–8. [Google Scholar]
  28. Goh, A.; Cammarata, N.; Voss, C.; et al. Multimodal Neurons in Neural Networks. J. Neural Inf. Process. 2021, 45, 113–126. [Google Scholar]
  29. Makridakis, S. The Forthcoming Artificial Intelligence Revolution: Its Impact on Society and Firms. Futures 2017, 90, 46–60. [Google Scholar] [CrossRef]
  30. Bozkurt, A.; Xiao, J.; Lambert, S.; et al. Speculative Futures on ChatGPT and Generative Artificial Intelligence (AI): A Collective Reflection from the Educational Landscape. Asian J. Distance Educ. 2023, 18, 50–130. [Google Scholar]
  31. Bozkurt, A. Generative Artificial Intelligence (AI) Powered Conversational Educational Agents: The Inevitable Paradigm Shift. Asian J. Distance Educ. 2023, 18, 1–12. [Google Scholar]
  32. Mondal, S.; Das, S.; Vrana, V.G. How to Bell the Cat? A Theoretical Review of Generative Artificial Intelligence Towards Digital Disruption in All Walks of Life. Technologies 2023, 11, 44. [Google Scholar]
  33. Zhang, C.; Zhang, C.; Li, C.; et al. One Small Step for Generative AI, One Giant Leap for AGI: A Complete Survey on ChatGPT in AIGC Era. arXiv arXiv:2304.06488, 2023.
  34. Jovanovic, M.; Campbell, M. Generative Artificial Intelligence: Trends and Prospects. Computer 2022, 55, 107–112. [Google Scholar] [CrossRef]
  35. Mathew, A. Is Artificial Intelligence a World Changer? A Case Study of OpenAI’s Chat GPT. Recent Prog. Sci. Technol. 2023, 5, 35–42. [Google Scholar]
  36. Rivas, P.; Zhao, L. Marketing with ChatGPT: Navigating the Ethical Terrain of GPT-Based Chatbot Technology. AI 2023, 4, 375–384. [Google Scholar] [CrossRef]
  37. Baidoo-Anu, D.; Owusu Ansah, L. Education in the Era of Generative Artificial Intelligence (AI): Understanding the Potential Benefits of ChatGPT in Promoting Teaching and Learning. Soc. Sci. Res. Netw. 2023. [CrossRef]
  38. Lo, C.K. What Is the Impact of ChatGPT on Education? A Rapid Review of the Literature. Educ. Sci. 2023, 13, 410. [Google Scholar]
  39. Sallam, M. ChatGPT Utility in Healthcare Education, Research, and Practice: Systematic Review on the Promising Perspectives and Valid Concerns. Healthcare 2023, 11, 887. [Google Scholar] [CrossRef]
  40. Biswas, S.S. Role of ChatGPT in Public Health. Ann. Biomed. Eng. 2023, 51, 868–869. [Google Scholar] [CrossRef] [PubMed]
  41. Biswas, S. ChatGPT and the Future of Medical Writing. Radiology 2023, 307, e223312. [Google Scholar] [CrossRef] [PubMed]
  42. Rahimi, F.; Abadi, A.T.B. ChatGPT and Publication Ethics. Arch. Med. Res. 2023, 54, 272–274. [Google Scholar] [CrossRef] [PubMed]
  43. Hassani, H.; Silva, E.S. The Role of ChatGPT in Data Science: How AI-Assisted Conversational Interfaces Are Revolutionizing the Field. Big Data Cogn. Comput. 2023, 7, 62. [Google Scholar] [CrossRef]
  44. Koubaa, A. GPT-4 vs. GPT-3.5: A Concise Showdown. AI 2023, 4, 100–112. [Google Scholar]
  45. Katz, D.M.; Bommarito, M.J.; Gao, S.; Arredondo, P. GPT-4 Passes the Bar Exam. SSRN 2023, 4389233. [Google Scholar] [CrossRef]
  46. Sanderson, K. GPT-4 Is Here: What Scientists Think. Nature 2023, 615, 773. [Google Scholar] [CrossRef]
  47. Skavronskaya, L.; Hadinejad, A.; Cotterell, D. Reversing the Threat of Artificial Intelligence to Opportunity: A Discussion of ChatGPT in Tourism Education. J. Teach. Travel Tour. 2023, 23, 253–258. [Google Scholar] [CrossRef]
  48. Halaweh, M. ChatGPT in Education: Strategies for Responsible Implementation; Bastas: Tokyo, Japan, 2023. [Google Scholar]
  49. Sabzalieva, E.; Valentini, A. ChatGPT and Artificial Intelligence in Higher Education: Quick Start Guide. UNESCO 2023. Available online: https://unesdoc.unesco.org/ark:/48223/pf0000386693 (accessed on 16 March 2025).
  50. Rahman, M.M.; Watanobe, Y. ChatGPT for Education and Research: Opportunities, Threats, and Strategies. Appl. Sci. 2023, 13, 5783. [Google Scholar] [CrossRef]
  51. Carnegie Learning. MATHia: Adaptive Math Learning. 2023. Available online: https://www.carnegielearning.com (accessed on 16 March 2025).
  52. Knewton. Adaptive Learning Technology. 2023. Available online: https://www.knewton.com (accessed on 16 March 2025).
  53. Fryer, L.K.; Ainley, M.; Thompson, A.; Gibson, A.; Sherlock, Z. AI in Education: The Impact of ChatGPT on Content Creation. J. Educ. Technol. 2023, 45, 10–15. [Google Scholar]
  54. Kasneci, E.; Seßler, K.; Küchemann, S.; Bannert, M.; Dementieva, D.; Fischer, F.; et al. ChatGPT for Good? On Opportunities and Challenges of Large Language Models for Education. Learn. Individ. Differ. 2023, 103, 102274. [Google Scholar]
  55. Schlippe, T.; Stierstorfer, Q.; Koppel, M.T.; Libbrecht, P. Explainability in Automatic Short Answer Grading. In Artificial Intelligence in Education Technologies: New Development and Innovative Practices; Springer: Cham, Switzerland, 2023; pp. 69–87. [Google Scholar]
  56. Schlippe, T.; Sawatzki, J. Cross-Lingual Automatic Short Answer Grading. In Artificial Intelligence in Education: Emerging Technologies, Models and Applications; Springer: Cham, Switzerland, 2021; pp. 117–129. [Google Scholar]
  57. Conijn, R.; Kahr, P.; Snijders, C. The Effects of Explanations in Automated Essay Scoring Systems on Student Trust and Motivation. J. Learn. Anal. 2023, 10, 37–53. [Google Scholar] [CrossRef]
  58. Jiao, W.X.; Wang, W.X.; Huang, J.T.; Wang, X.; Tu, Z.P. Is ChatGPT a Good Translator? arXiv arXiv:2301.08745, 2023.
  59. Wang, L.; Lyu, C.; Ji, T.; Zhang, Z.; Yu, D.; Shi, S.; Tu, Z. Document-Level Machine Translation with Large Language Models. arXiv arXiv:2304.02210, 2023.
  60. Deng, X.; Yu, Z. A Systematic Review of Machine-Translation-Assisted Language Learning for Sustainable Education. Sustainability 2022, 14, 7598. [Google Scholar] [CrossRef]
  61. Tsai, S.C. Using Google Translate in EFL Drafts: A Preliminary Investigation. Comput. Assist. Lang. Learn. 2019, 32, 510–526. [Google Scholar] [CrossRef]
  62. Fazlollahi, A.M.; Bakhaidar, M.; Alsayegh, A.; et al. Effect of Artificial Intelligence Tutoring vs Expert Instruction on Learning Simulated Surgical Skills Among Medical Students: A Randomized Clinical Trial. JAMA Netw. Open 2022, 5, e2149008. [Google Scholar] [CrossRef]
  63. Afzal, S.; Dhamecha, T.I.; Gagnon, P.; et al. AI Medical School Tutor: Modelling and Implementation. In Artificial Intelligence in Medicine; Springer: Cham, Switzerland, 2020; pp. 133–145. [Google Scholar]
  64. Chan, K.S.; Zary, N. Applications and Challenges of Implementing Artificial Intelligence in Medical Education: Integrative Review. JMIR Med. Educ. 2019, 5, e13930. [Google Scholar] [CrossRef]
  65. Francisco, R.E.; Oliveira Silva, F. Intelligent Tutoring System for Computer Science Education and the Use of Artificial Intelligence: A Literature Review. Educ. Inf. Technol. 2022, 27, 12145–12168. [Google Scholar]
  66. Grossman, J.; Lin, Z.; Sheng, H.; Wei, J.T.Z.; Williams, J.J.; Goel, S.M. Transforming Online Resources for Learning Math into Conversational Interactions. Proc. AAAI Conf. Artif. Intell. 2019, 33, 7841–7848. [Google Scholar]
  67. Dempere, J.; Modugu, K.; Hesham, A.; Ramasamy, L.K. The Impact of ChatGPT on Higher Education. Front. Educ. 2023, 8, 1206936. [Google Scholar] [CrossRef]
  68. Mbakwe, A.B.; Lourentzou, I.; Celi, L.A.; et al. ChatGPT Passing USMLE Shines a Spotlight on the Flaws of Medical Education. PLOS Digit. Health 2023, 2, e0000205. [Google Scholar] [CrossRef]
  69. Gravel, J.; D’Amours-Gravel, M.; Osmanlliu, E. Learning to Fake It: Limited Responses and Fabricated References Provided by ChatGPT for Medical Questions. Mayo Clin. Proc. Digit. Health 2023, 1, 226–234. [Google Scholar] [CrossRef]
  70. Alkaissi, H.; McFarlane, S.I. Artificial Hallucinations in ChatGPT: Implications in Scientific Writing. Cureus 2023, 15, e35179. [Google Scholar] [CrossRef] [PubMed]
  71. Khalil, M.; Er, E. Will ChatGPT Get You Caught? arXiv arXiv:2302.04335, 2023.
  72. Basic, Z.; Banovac, A.; Kruzic, I.; Jerkovic, I. Better by You, Better Than Me, ChatGPT3 as Writing Assistance in Students Essays. arXiv arXiv:2302.04536, 2023.
  73. Oker, A.; Pecune, F.; Declercq, C. Virtual Tutor and Pupil Interaction: A Study of Empathic Feedback as Extrinsic Motivation for Learning. Educ. Inf. Technol. 2020, 25, 3643–3658. [Google Scholar] [CrossRef]
  74. Tilli, A.; Shehata, B.; Adarkwah, M.A.; et al. What If the Devil Is My Guardian Angel: ChatGPT as a Case Study of Using Chatbots in Education. Smart Learn. Environ. 2023, 10, 15. [Google Scholar] [CrossRef]
  75. Farrokhnia, M.; Banihashem, S.K.; Noroozi, O.; Wals, A. A SWOT Analysis of ChatGPT: Implications for Educational Practice and Research. Innov. Educ. Teach. Int. 2023, 60, 1–15. [Google Scholar] [CrossRef]
  76. Sullivan, M.; Kelly, A.; McLaughlan, P. ChatGPT in Higher Education: Considerations for Academic Integrity and Student Learning. J. Appl. Learn. Teach. 2023, 6, 1–12. [Google Scholar]
  77. Susnjak, T. ChatGPT: The End of Online Exam Integrity? arXiv arXiv:2212.09292, 2022. [CrossRef]
  78. Adiguzel, T.; Kaya, M.H.; Cansu, F.K. Revolutionizing Education with AI: Exploring the Transformative Potential of ChatGPT. Contemp. Educ. Technol. 2023, 15, 429. [Google Scholar] [CrossRef] [PubMed]
  79. Turnitin. AI Writing Detection. 2023. Available online: https://www.turnitin.com (accessed on 16 March 2025).
  80. Caliskan, A.; Bryson, J.J.; Narayanan, A. Semantics Derived Automatically from Language Corpora Contain Human-Like Biases. Science 2017, 356, 183–186. [Google Scholar] [CrossRef]
  81. Zawacki-Richter, O.; Marín, V.I.; Bond, M.; Gouverneur, F. Systematic Review of Research on Artificial Intelligence Applications in Higher Education. Int. J. Educ. Technol. High. Educ. 2019, 16, 39. [Google Scholar] [CrossRef]
  82. Smith, J.; Brown, T.; Davis, R. Implementing AI Chatbots in Higher Education: A Case Study. J. High. Educ. Technol. 2023, 34, 65–70. [Google Scholar]
  83. Johnson, R.; Smith, K.; Lee, M. AI-Generated Questions in Medical Education: A Pilot Study. Med. Educ. J. 2023, 56, 44–49. [Google Scholar]
  84. Inside Universities’ Love-Hate Relationship with ChatGPT. The Wall Street Journal 2023. Available online: https://www.wsj.com (accessed on 16 March 2025).
  85. ChatGPT ya tiene plaza fija en la universidad: «Desde que lo uso, pienso menos por mí misma». El País 2024. Available online: https://elpais.com (accessed on 16 March 2025).
  86. Khan, S. The Role of AI in Personalized Learning. Khan Acad. Blog 2023. Available online: https://www.khanacademy.org (accessed on 16 March 2025).
  87. Subakan, C. , Ravanelli, M., Cornell, S., Bronzi, M., & Zhong, J. (2021, June). Attention is all you need in speech separation. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 21-25).
  88. Jumper, J.; Hassabis, D. Protein structure predictions to atomic accuracy with AlphaFold. Nature methods 2022, 19(1), 11–12. [Google Scholar] [CrossRef]
  89. Carleo, G.; Troyer, M. Solving the Quantum Many-Body Problem with Artificial Neural Networks. Science 2017, 355, 602–606. [Google Scholar] [CrossRef]
  90. Brunton, S.L.; Proctor, J.L.; Kutz, J.N. Discovering Governing Equations from Data by Sparse Identification of Nonlinear Dynamical Systems. Proc. Natl. Acad. Sci. USA 2016, 113, 3932–3937. [Google Scholar] [CrossRef]
  91. Segler, M.H.S.; Preuss, M.; Waller, M.P. Planning Chemical Syntheses with Deep Neural Networks and Symbolic AI. Nature 2018, 555, 604–610. [Google Scholar] [CrossRef]
  92. Perez, E.; Kiela, D.; Cho, K. True few-shot learning with language models. Advances in neural information processing systems 2021, 34, 11054–11070. [Google Scholar]
  93. Thoppilan, R.; De Freitas, D.; Hall, J.; et al. LaMDA: Language Models for Dialog Applications. arXiv arXiv:2201.08239, 2022.
  94. Yotsumoto, Y.; Nakajima, Y.; Takamoto, R.; et al. Autonomous Robotic Experimentation System for Powder X-ray Diffraction. Digit. Discov. 2024, 3, 2523–2532. [Google Scholar] [CrossRef]
  95. Cui, Z.; Li, N.; Zhou, H. Can AI Replace Human Subjects? arXiv arXiv:2409.00128, 2024.
  96. Bolukbasi, T.; Chang, K.-W.; Zou, J.Y.; et al. Man Is to Computer Programmer as Woman Is to Homemaker? Debiasing Word Embeddings. Adv. Neural Inf. Process. Syst. 2016, 29, 4349–4357. [Google Scholar]
  97. Arseniev-Koehler, A.; Foster, J.G. Machine learning as a model for cultural learning: Teaching an algorithm what it means to be fat. Sociological Methods & Research 2022, 51, 1484–1539. [Google Scholar]
  98. Lin, S.; Hilton, J.; Evans, O. TruthfulQA: Measuring How Models Mimic Human Falsehoods. Proc. ACL 2022, 1, 3214–3252. [Google Scholar]
  99. Huang, L.; Yu, W.; Ma, W.; et al. A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions. ACM Trans. Inf. Syst. 2025, 43, 1–55. [Google Scholar] [CrossRef]
  100. Frieder, S.; Pinzuti, A.; Akbulut, L.; et al. Mathematical Capabilities of ChatGPT. PNAS Nexus 2023, 2, pgad258. [Google Scholar]
  101. Wei, J.; Tay, Y.; Bommasani, R.; et al. Emergent Abilities of Large Language Models. Trans. Mach. Learn. Res. 2022, 1, 1–25. [Google Scholar]
  102. Ganguli, D.; Lovitt, L.; Kernion, J.; et al.; The Capacity for Moral Self-Correction in Large Language Models Anthropic 2023. Available online: https://www.anthropic.com (accessed on 16 March 2025).
  103. Wani, A.G.; Astunkar, G.S. Open Artificial Intelligence (AI) of ChatGPT for Library Services and Library Science Professionals. Libr. Sch. 2024, 4, 1–12. [Google Scholar]
  104. Wei, Z.; Guo, D.; Huang, D.; Zhang, Q.; Zhang, S.; Jiang, K.; Li, R. Detecting and Mitigating the Ungrounded Hallucinations in Text Generation by LLMs. Proc. Int. Conf. Artif. Intell. Syst. Netw. Secur. 2023, 77–81. [Google Scholar]
  105. Freedman, J. D. , & Nappier, I. A. (2023). GPT-4 to GPT-3.5:'Hold My Scalpel'--A Look at the Competency of OpenAI's GPT on the Plastic Surgery In-Service Training Exam. arXiv:2304.01503.
  106. Mu, Y.; Bai, P.; Bontcheva, K.; Song, X. Addressing Topic Granularity and Hallucination in Large Language Models for Topic Modelling. arXiv arXiv:2405.00611, 2024.
  107. Gil, Y.; Baker, D.; de Melo, G.; et al. AI-Assisted Meta-Analysis of Permafrost Studies. Environ. Res. Lett. 2023, 18, 075001. [Google Scholar]
  108. Bhattacharya, A.; Kumar, S. Cross-Disciplinary Modeling of Biogeochemical Cycles. Nat. Comput. Sci. 2023, 12, 45–59. [Google Scholar]
  109. Chemaya, N. , & Martin, D. ( 19(7), e0304807.
  110. Smith, J.; Patel, R.; Kim, H. Enhancing Perovskite Solar Cell Papers with NLP. Adv. Mater. 2023, 35, 2208912. [Google Scholar]
  111. Jung, I. Pathways to International Publication in the Social Sciences: A Guide for Early Career and Non-Native English Researchers. Springer Nature, 2024.
  112. Myers, A.; AI-Detectors Biased Against Non-Native English Writers. Stanford HAI Blog 2023. Available online: https://hai.stanford.edu (accessed on 16 March 2025).
  113. Rosen, B. P. , Li, Q. V., Cho, H. S., Liu, D., Yang, D., Graff, S.,... & Huangfu, D. (2024). Parallel genome-scale CRISPR-Cas9 screens uncouple human pluripotent stem cell identity versus fitness. Nature Communications, 15(1), 8966.
  114. Thompson, E. The Ethics of AI in Publishing. Science 2023, 381, 124–125. [Google Scholar]
  115. Wang, Q.; Liu, F.; Zhang, T. AI in Scientific Workflows. Nat. Rev. Methods Prim. 2024, 4, 12. [Google Scholar]
  116. López, G.; Martínez, J.; García, R. Human-AI Collaboration Frameworks. Proc. Natl. Acad. Sci. USA 2023, 120, e2216237120. [Google Scholar]
  117. Lehr, S.A.; Caliskan, A.; Liyanage, S.; Banaji, M.R. ChatGPT as Research Scientist: Probing GPT’s Capabilities as a Research Librarian, Research Ethicist, Data Generator, and Data Predictor. Proc. Natl. Acad. Sci. USA 2024, 121, e2404328121. [Google Scholar] [CrossRef] [PubMed]
  118. Zashikhina, I.M. Preparing a Scientific Article: Can ChatGPT Handle It? High. Educ. Russ. 2023, 32, 24–47. [Google Scholar] [CrossRef]
  119. Khlaif, Z.N.; Mousa, A.; Hattab, M.K.; et al. The Potential and Concerns of Using AI in Scientific Research: ChatGPT Performance Evaluation. JMIR Med. Educ. 2023, 9, e47049. [Google Scholar] [CrossRef] [PubMed]
  120. Lee, J.Y. Can an Artificial Intelligence Chatbot Be the Author of a Scholarly Article? J. Educ. Eval. Health Prof. 2023, 20, 6. [Google Scholar] [CrossRef]
  121. Chen, T.J. ChatGPT and Other AI Applications Speed Up Scientific Writing. J. Chin. Med. Assoc. 2023, 86, 351–353. [Google Scholar] [CrossRef]
  122. Altmäe, S.; Sola-Leyva, A.; Salumets, A. Artificial Intelligence in Scientific Writing: A Friend or a Foe? Reprod. Biomed. Online 2023, 47, 3–9. [Google Scholar] [CrossRef]
  123. Neumeister, L.; Lawyers Blame ChatGPT for Tricking Them into Citing Bogus Case Law. Associated Press News 2023. Available online: https://apnews.com (accessed on 16 March 2025).
  124. Tlili, A. , & Solans, D. B. (2025). Ai hallucinations? what about human hallucination?!: Addressing human imperfection is needed for an ethical AI. IJIMAI.
  125. Kacena, M.A.; Plotkin, L.I.; Fehrenbacher, J.C. The Use of Artificial Intelligence in Writing Scientific Review Articles. Curr. Osteoporos. Rep. 2024, 22, 115–121. [Google Scholar] [CrossRef]
  126. Athaluri, S.A.; Manthena, S.V.; Kesapragada, V.; et al. Exploring the Boundaries of Reality: Investigating the Phenomenon of Artificial Intelligence Hallucination in Scientific Writing through ChatGPT References. Cureus 2023, 15, e37432. [Google Scholar] [CrossRef] [PubMed]
  127. Korteling, J.E.H.; van de Boer-Visschedijk, G.C.; Bosse, T. Human- versus Artificial Intelligence. Front. Artif. Intell. 2021, 4, 622364. [Google Scholar] [CrossRef]
  128. Gao, C.A.; Howard, F.M.; Markov, N.S.; et al. Comparing Scientific Abstracts Generated by ChatGPT to Real Abstracts. NPJ Digit. Med. 2023, 6, 75. [Google Scholar] [CrossRef]
  129. Karnik, S.J.; Margetts, T.J.; Wang, H.S.; et al. Mind the Gap: Unraveling the Intricate Dance Between Alzheimer’s Disease and Related Dementias and Bone Health. Curr. Osteoporos. Rep. 2024, 22, 1–12. [Google Scholar] [CrossRef] [PubMed]
  130. Li, Z. (2023). The dark side of chatgpt: Legal and ethical challenges from stochastic parrots and hallucination. arXiv:2304.14347.
  131. Weidinger, L.; Mellor, J.; Rauh, M.; et al. Ethical and Social Risks of Harm from Language Models. ACM Comput. Surv. 2022, 55, 1–38. [Google Scholar]
  132. Park, Y. J. , Kaplan, D., Ren, Z., Hsu, C. W., Li, C., Xu, H.,... & Li, J. (2024). Can ChatGPT be used to generate scientific hypotheses?. Journal of Materiomics, 10(3), 578-584.
  133. Dehouche, N. (2021). Plagiarism in the age of massive Generative Pre-trained Transformers (GPT-3). Ethics in Science and Environmental Politics, 21, 17-23.
  134. Dale, R. GPT-3: What’s It Good For? Nat. Lang. Eng. 2021, 27, 113–118. [Google Scholar] [CrossRef]
  135. Bommasani, R.; Hudson, D.A.; Adeli, E.; et al. On the Opportunities and Risks of Foundation Models. arXiv arXiv:2108.07258, 2021.
  136. Aithal, S. , & Aithal, P. S. (2023). Effects of AI-based ChatGPT on higher education libraries. International Journal of Management, Technology, and Social Sciences (IJMTS), 8(2), 95-108.
  137. McGee, J. The Future of Academic Integrity in the Age of AI. J. Acad. Ethics 2023, 21, 1–15. [Google Scholar]
  138. Yan, D. , Fauss, M., Hao, J., & Cui, W. (2023). Detection of AI-generated essays in writing assessments. Psychological Test and Assessment Modeling, 65(1), 125-144.
  139. Cotton, D.R.E.; Cotton, P.A.; Shipway, J.R. ChatGPT, AI, and the Educational Landscape: Disruptions and Ethical Challenges. Teach. High. Educ. 2023, 28, 1–18. [Google Scholar]
  140. Brundage, M.; Avin, S.; Clark, J.; et al. Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims. arXiv arXiv:2206.05838, 2022.
  141. OpenAI. ChatGPT and Data Privacy: Risks and Recommendations. 2023. Available online: https://openai.com (accessed on 16 March 2025).
  142. U.S. Copyright Office. AI-Generated Content and Copyright Law: Policy Considerations. 2023. Available online: https://copyright.gov (accessed on 16 March 2025).
  143. Marcus, G.; Davis, E. AI and Scientific Discovery: Where Do We Stand? AI Mag. 2022, 43, 1–12. [Google Scholar]
  144. Mitchell, M.; Krakauer, D.C. AI Alignment and the Challenges of Updating Language Models. AI Alignment Res. J. 2023, 1, 1–25. [Google Scholar]
  145. Lipton, Z.C. The Mythos of Model Interpretability. ACM Queue 2018, 16, 31–57. [Google Scholar] [CrossRef]
  146. Dwivedi, Y. K. , Kshetri, N., Hughes, L., Slade, E. L., Jeyaraj, A., Kar, A. K.,... & Wright, R. (2023). Opinion Paper:“So what if ChatGPT wrote it?” Multidisciplinary perspectives on opportunities, challenges and implications of generative conversational AI for research, practice and policy. International journal of information management, 71, 102642.
  147. Cotton, D.R.E.; Cotton, P.A.; Shipway, J.R. Chatting and Cheating: Ensuring Academic Integrity in the Era of ChatGPT. Innov. Educ. Teach. Int. 2023, 60, 1–10. [Google Scholar] [CrossRef]
  148. Spitale, G. , Biller-Andorno, N., & Germani, F. (2023). AI model GPT-3 (dis) informs us better than humans. Science Advances, 9(26), eadh1850.
  149. Gallegos, I. O. , Rossi, R. A., Barrow, J., Tanjim, M. M., Kim, S., Dernoncourt, F.,... & Ahmed, N. K. (2024). Bias and fairness in large language models: A survey. Computational Linguistics, 50(3), 1097-1179.
  150. Akgun, S.; Greenhow, C. Artificial Intelligence in Education: Addressing Ethical Challenges in K-12 Settings. AI Ethics 2022, 2, 431–440. [Google Scholar] [CrossRef] [PubMed]
  151. Bond, M. , Khosravi, H., De Laat, M., Bergdahl, N., Negrea, V., Oxley, E.,... & Siemens, G. (2024). A meta systematic review of artificial intelligence in higher education: A call for increased ethics, collaboration, and rigour. International Journal of Educational Technology in Higher Education, 21(1), 4.
  152. Stokel-Walker, C. ChatGPT Listed as Author on Research Papers: Many Scientists Disapprove. Nature 2023, 613, 620–621. [Google Scholar] [CrossRef] [PubMed]
  153. Hosseini, M.; Horbach, S.P.J.M. Ethics of ChatGPT in Academia: A Call for Global Regulations. Res. Ethics 2023, 19, 1–14. [Google Scholar]
  154. Dhamala, J., Sun, T., Kumar, V., Krishna, S., Pruksachatkun, Y., Chang, K. W., & Gupta, R. (2021, March). Bold: Dataset and metrics for measuring biases in open-ended language generation. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (pp. 862-872).
  155. Zawacki-Richter, O.; Lück, M. The Impact of Artificial Intelligence on Education; Springer: Cham, Switzerland, 2020. [Google Scholar]
  156. Popović, L.; Zrnić, D. Machine Translation for Academic Purposes: Benefits and Limitations. J. Educ. Technol. Soc. 2019, 22, 118–127. [Google Scholar]
  157. Binns, R.; Kuipers, B. Artificial Intelligence and Academic Integrity: Challenges and Solutions. AI Ethics 2022, 1, 45–58. [Google Scholar]
  158. Sweeney, L. Discrimination in Online Ad Delivery. ACM Trans. Internet Technol. 2020, 10, 40–52. [Google Scholar]
  159. De Masi, G.; Grasso, G. The Future of AI in Academic Research: Challenges and Opportunities. J. Res. Comput. 2023, 34, 85–98. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated