Preprint
Article

This version is not peer-reviewed.

Hallucination Is Inevitable for LLMs with the Open World Assumption

Submitted:

06 October 2025

Posted:

10 October 2025

You are already at the latest version

Abstract
Large Language Models (LLMs) exhibit impressive linguistic competence but also produce inaccurate or fabricated outputs, often called “hallucinations”. Engineering approaches usually regard hallucination as a defect to be minimized, while formal analyses have argued for its theoretical inevitability. Yet both perspectives remain incomplete when considering the conditions required for artificial general intelligence (AGI). This paper reframes “hallucination” as a manifestation of the generalization problem. Under the Closed World assumption, where training and test distributions are consistent, hallucinations may be mitigated. Under the Open World assumption, however, where the environment is unbounded, hallucinations become inevitable. This paper further develops a classification of hallucination, distinguishing cases that may be corrected from those that appear unavoidable under open-world conditions. On this basis, it suggests that “hallucination” should be approached not merely as an engineering defect but as a structural feature to be tolerated and made compatible with human intelligence.
Keywords: 
;  ;  

1. Introduction

Large Language Models (LLMs) [1] have rapidly become central to contemporary artificial intelligence, influencing domains as diverse as information retrieval, scientific discovery, and everyday decision-making. Yet their increasing deployment has brought renewed attention to a longstanding difficulty: the phenomenon of “hallucination” [2]. LLMs sometimes generate information that is fluent and persuasive but nonetheless inaccurate or fabricated. In engineering discourse, hallucination is typically treated as a defect to be minimized through better datasets, architectures, or alignment procedures [3,4]. In philosophical discourse, however, the status of hallucination is far less straightforward: is it a contingent flaw of current systems, or is it in some sense inevitable for any learner operating in open-ended environments?
Recent research has attempted to formalize this issue. Some authors define hallucination as output error relative to a ground-truth and argue that such errors are theoretically unavoidable given infinite input spaces and adversarial environments. [5] Some others take a probabilistic perspective, showing that while hallucinations cannot be eliminated in principle, they can be rendered statistically negligible with sufficient data and appropriate algorithms [6], and the analyses presuppose what might be called the “Closed World assumption”1 termed herein, in which the training distribution is assumed to adequately represent the target environment. Under this assumption, hallucination appears controllable: additional data, improved inference, or cautious abstention (i.e., “I don’t know”) may reduce error to low levels.
This paper takes a different stance and argues that when the “Open World assumption”—a condition more faithful to real-world intelligence and to the ambitions of artificial general intelligence—is adopted, hallucination becomes not merely difficult to avoid but inevitable. The Open World assumption maintains that intelligent systems must cope with unbounded space and time, and tasks. In such a setting, past experience does not guarantee future accuracy; generalization is both indispensable and fallible. Just as humans mis-generalize, or form inductive expectations that later prove false, LLMs cannot hope to avoid hallucination. Hallucination, in this light, should be seen not as a marginal bug but as a structural byproduct of intelligence under open-world conditions.
The contribution of this paper is threefold. First, the hallucination problem is reduced to the generalization problem in the traditional machine learning research Second, three types of errors are distinguished, and the latter two are viewed as hallucination. Type-I hallucination arises from false memorization: the model produces an answer inconsistent with facts already present in its training data. Such errors are, in principle, corrigible. Type-II hallucination, by contrast, arises from false generalization to cases not present in the training set. This latter form is inescapable under the Open World assumption, since no finite past can guarantee correctness in an infinite future. Third, constructive ways of “treating” hallucination are proposed: tolerating it as a normal phenomenon of deep learning mechanisms, designing systems that remain adaptive in the face of limited experience, and aligning error patterns with forms of reasoning that are intelligible and acceptable to humans.

2. Debates on Inevitability of “Hallucination”

LLMs’ behaviors of generating false information have been noticed in the early stage. The formal discussion on the inevitability of LLM hallucination can be traced back to Xu et al. [5]: the authors defined LLM hallucination as the output error compared with a ground-truth, and they argued that LLM hallucination is inevitable in the sense that there exists a function f, representing a world, which makes the sole ground-truth different from an LLM’s output at a certain training stage, and any LLMs would make infinitely many errors in that world. Although this paper draws a similar conclusion as [5], the meaning and line of argument are fundamentally different. I disagree with Xu et al. [5]’s justification in three aspects. First, I do not view arbitrary types of errors as “hallucination”. Humans make mistakes due to many reasons. One origin is lacking information, as a result, they have beliefs that are not consistent with some facts. Additionally, some errors stem from the changes of outside standards. For example, an LLM can always generate wrong answers in an adversarial world (see Figure A1 in Appendix A). In this paper, as argued in Section 3 and Section 4, “hallucination” can be viewed as a byproduct of generalization: it can be reasonable to some extent, but the correctness cannot be guaranteed under the Open World assumption. Second, proofs in Xu et al. [5] rely on an infinite set of samples (otherwise, the statement that an LLM makes infinitely many errors cannot be true). However, in practice, the lengths of strings input to and output from an LLM are always finite, which makes the sample space finite. Therefore, if collecting enough data, with enough resources, all samples are contained in the training-set, and errors can be fully avoided in theory. Third, Xu et al. [5] assume that each input corresponds to a single correct answer as the ground-truth, while any different answer from the ground-truth is viewed as error. However, due to the many-to-many mapping between sentences and meanings, this assumption is not suitable. At the worst case, an LLM can answer “I don’t know”, which is different from the ground-truth but should not be simply viewed as a wrong answer. For these reasons, while I acknowledge the formal validity of their theorems within their chosen framework, I find their definitions and assumptions too restrictive for capturing the phenomenon of hallucination as it arises in practice. My analysis therefore shifts the focus: from existence proofs to the epistemic and methodological inevitability of hallucination under open-world conditions.
Suzuki et al. [6] criticized Xu et al. [5] from a different perspective. Suzuki et al. [6] introduced the notion of hallucination probability to measure the likelihood that a language model generates outputs outside the acceptable set under a given input distribution. While computability theory shows that hallucinations are inevitable on an infinite set of inputs, the authors prove from a probabilistic perspective that the probability of hallucinations can be made arbitrarily close to zero, provided the training data is sufficiently high-quality and abundant, and appropriate training and inference algorithms are used. In other words, hallucinations are theoretically unavoidable in principle, but in practice they can be rendered statistically negligible, according toSuzuki et al. [6]. Therefore, if hallucinations persisted in real applications, the cause should be attributed to limitations in the dataset or algorithms, rather than to any “innate inevitability” dictated by computability theory. I agree with Suzuki et al. [6] that hallucination can be largely mitigated when the test-set is fully consistent with the training-set (as termed herein “the closed environment assumption”). However, the Open World assumption should be adopted for some reasons, which negate the guarantee of the consistency (see Section 4).
Undoubtedly, current LLMs cannot answer any questions without errors. Some researchers summarized types of problems that LLMs cannot solve. [7,8] For example, some problems have no solution logically, some solutions are hard to compute, and some solutions suffers from resource limitation. Yet humans also cannot solve those problems for the same reasons. In these three situations, hallucinations are still evitable—LLMs can answer “Any answer is wrong for this question”, “I am not able to solve that”, “I cannot do that”, etc. Many problems that cause LLMs’ “inevitable” wrong answers are actually evitable by answering, e.g., “I don’t know”.

3. Generalization and the “No Free Lunch” Theorem

Beyond these debates, this paper advances a distinct perspective: hallucination should be understood as a manifestation of the generalization problem in machine learning.
The operation of an LLM can be abstracted at the level of single-token prediction. Concretely, each step can be represented as a mapping x g ( x ) , where x denotes the current context (a tensor) and g ( x ) is the embedding of the next predicted token. This mapping is applied repeatedly to extend the sequence. For example, define g 0 ( x ) = g ( x ) , g 1 ( x ) = g ( [ x , g 0 ( x ) ] ) , g 2 ( x ) = g ( [ x , g 0 ( x ) , g 1 ( x ) ] ) , and so on, up to g n ( x ) = g ( [ x , g 0 ( x ) , g 1 ( x ) , . . . , g n 1 ( x ) ] ) . On this basis, the overall sequence-generation process can itself be viewed as a higher-level mapping, f ( x ) = [ g 0 ( x ) , g 1 ( x ) , , g n ( x ) , , , ] , where “⋄” denotes a placeholder (e.g., an all-zero tensor) to ensure a constant output dimension n max , corresponding to the maximal number of tokens an LLM can produce in practice. Although LLMs are in principle capable of unbounded continuation, their outputs are always restricted to a finite length; the above abstraction adopts this restriction for theoretical clarity.
The mapping f ( · ) is learned from training samples through different mechanisms. For instance, g ( · ) can be trained by minimizing the prediction error between its output g ^ ( x ) and the next token. In addition, g ( · ) —and hence f ( · ) —can be further adjusted by other mechanisms, such as Reinforcement Learning from Human Feedback (RLHF) [9], where updates depend on the evaluation of output sequence f ^ ( x ) against reward signals.
Therefore, an LLM can be abstracted as a mapping x f ( x ) , where both x and f ( x ) are tensors. In this sense, a large language model may be regarded as a type of machine learning model. Although its training procedures differ from those of traditional classifiers and function approximators, many insights from classical machine learning—particularly those concerning generalization—remain applicable.
For ease of exposition, this mapping can be illustrated in a two-dimensional diagram (Figure 1). The reduction to two dimensions is purely for intuitive visualization; the underlying analysis applies equally to the high-dimensional case. In the figure, black dots represent training samples, while the orange curve depicts the mapping learned by an LLM.
An LLM maps an input to an output. However, training sets typically do not cover all possible input–output relations. As a result, an LLM must generalize to novel situations based on its prior experience. Even when large-scale corpora are employed for training, they cannot exhaustively capture every potential input. For instance, it would be infeasible and unnecessary to include every possible expression of integer addition (e.g., “ 1 + 1 = 2 ”, “ 14159 + 26535 = 40694 ”, and so on). In such cases, the model must extrapolate beyond its training data. Sometimes this extrapolation aligns well with the developer’s expectations—such as when the test distribution is anticipated and aligned with the training set, as in the case of basic arithmetic. Sometimes, however, the test distribution diverges from the training set in ways that cannot be foreseen in advance.
Machine learning theorists have shown that no optimization algorithm is universally superior; improved performance is only achievable when the problem distribution is non-uniform and the algorithm is specifically adapted to that structure. This result is formalized in the “No Free Lunch” (NFL) theorem. [10] According to NFL, without an assumption about the distribution of the test set, one cannot claim that a particular learning outcome is superior to another. For instance, the mapping in Figure 1 is not necessarily better than that in Figure 2, since the latter may perform better on certain test samples (represented as purple dots in Figure 3).
In Figure 2, a generated output, shown as a dashed circle and labeled “H”, may be interpreted as a hallucination. Yet the same output could also correspond to an actual fact, illustrated as a purple dot labeled “F” in Figure 3. From the learner’s perspective during training, it is impossible to determine which mapping is superior. Once new situations are incorporated as additional training samples, the system can update its mapping and eliminate the corresponding errors—although errors will remain possible in other, as yet unseen, cases.
If the distribution of a test set is consistent with that of a training set—that is, if the regularities learned from past data can be fully applied to future cases—then a learner may succeed in generalizing to new situations. Nevertheless, errors can still arise when the available experience is insufficient. In such cases, additional data (i.e., providing more training examples to the LLM) can improve performance.

4. “Hallucination” Is Inevitable

The preceding analysis suggests that errors can, in principle, be reduced when the test distribution aligns with the training distribution, as in closed-world settings. In such cases, additional data and careful optimization may eventually eliminate many sources of error. In open-world settings, however, this assumption no longer holds: past regularities may fail to extend to future cases, and errors become unavoidable.
Against this background, it is useful to distinguish two kinds of “hallucination”. Type-I hallucination (HT-I) arises from false memorization and is, at least in principle, corrigible. Type-II hallucination (HT-II), by contrast, stems from false generalization under the Open World assumption and is therefore inevitable.

Generalization Error Is Inevitable with the Open World Assumption

As discussed in Section 3, perfect generalization presupposes that the regularities observed in past data remain valid in the future—a condition termed herein as the Closed World assumption. In practice, however, this presupposition cannot be sustained. The regularities available to a learner are always limited by its finite experience, and future cases may challenge or overturn them. This does not imply that the world itself is unstable, but rather that any agent’s summary of the world is necessarily provisional and subject to revision.
As Hume famously observed, no matter how extensive our past experience may be, our predictions about the future can never be infallible [11]. A classical illustration is Russell’s “inductive turkey”: a turkey, fed at the same time every day, concludes that this pattern will always continue—until the day it is slaughtered. This example highlights the inherent limits of induction. For artificial general intelligence (AGI) [12,13] and AI systems intended to operate under realistic conditions, it is therefore more appropriate to adopt the Open World assumption [14], which explicitly acknowledges the unbounded and potentially unpredictable character of future situations.
An intelligent system should work with unbounded space and time, and unbounded tasks.
  • Unbounded space and time: The future of the world may not align with the past. In other words, the system’s experiences may not be directly applicable to future scenarios.
  • Unbounded tasks: A wide range of tasks needs to be addressed, and these tasks are not predetermined or defined at the system’s birth.
According to the Open World assumption, it is impossible for a system to collect and memorize all facts, and the future cannot be assumed to align fully with the past. This does not imply that the environment is wholly unstable—for in such a setting no agent could adapt at all—but rather that the scope of any agent’s experience is necessarily limited.
The Open World assumption becomes especially relevant when AI systems are expected to cope with conditions beyond human experience. For instance, an AI designed to explore an unfamiliar planet cannot rely on the expectation that its new environment will mirror conditions on Earth.
With the Open World assumption, we can analyze errors in intelligent systems more generally. Such errors are not unique to LLMs but are also exhibited by humans, reflecting shared cognitive and computational constraints. At the most abstract level, three sources of error can be distinguished:
  • Environmental change, where past regularities no longer apply to the present situation. For example, one might have learned that “the capital of China is Chang’an,” whereas the current fact is that “the capital of China is Beijing.”
  • False memorization, where the system’s present belief diverges from information previously encountered. A well-known illustration in human cognition is the Mandela Effect, in which large groups of people recall a past event inaccurately—for instance, misremembering that Nelson Mandela died in prison in the 1980s.
  • False generalization, where multiple plausible generalizations of past experience are available, and the one adopted diverges from an accepted standard. For example, based on limited observation, one might infer that “the Earth is the center of the universe,” whereas the standard was “the Sun is the center of the universe” (though it can be further challenged).
Not all of these errors should be classified as “hallucination.” For instance, mistakes due to environmental change are better understood as outdated knowledge rather than hallucination. In what follows, the focus is on the latter two cases. When false memorization occurs, an LLM may produce an output inconsistent with facts present in its training data. When false generalization occurs, the model may fail to provide a correct answer absent from its training set but inferable from a human perspective. These are hereafter referred to as Type-I Hallucination (HT-I) and Type-II Hallucination (HT-II), respectively.
HT-I may arise for various reasons. Nevertheless, it is in principle corrigible, since the system only needs to adjust its mapping to account for a finite set of past experiences. As illustrated in Figure 4, the output (“H”) deviates from the corresponding training sample (“F”); yet because this sample is already known, the mapping can be revised until the output aligns exactly with the expected value.
By contrast, HT-II arises directly from the Open World assumption, which precludes any guarantee of consistency between training and test sets. Such errors cannot be eliminated simply by accumulating and summarizing additional experiences, since generalization itself is never assured. HT-II is therefore inevitable.
HT-II cannot be solved by abandoning the capability of generalization and simply answering “I don’t know” (as shown in Figure 5). We do not want an intelligent machine to answer only what it memorize accurately, like a traditional database searching program. For example, if the system is told “1+1=2” and “ x 2 = x + x ”, the system is not expected to answer “I don’t know” to the question “what is 1*2”, as well as any other questions whose answers are outside records.

Training a Detector Is Not Promising—It Still Needs Generalization

One might suggest addressing hallucination by training a secondary model to detect and filter out erroneous outputs. While such detectors can reduce certain kinds of mistakes, they cannot provide a definitive solution. The reason is that the detector itself must generalize beyond the data on which it is trained: it will inevitably encounter novel cases, and in those situations its judgments are likewise fallible. In effect, the hallucination problem is merely displaced from the generator to the detector, without being fundamentally resolved.

5. Treatment of “Hallucination”

The preceding discussion should not be read as a purely negative diagnosis. It also points towards constructive strategies for addressing the phenomenon of “hallucination” in LLMs, or more generally, the broader problem of generalization in learning systems.
Within restricted domains, performance can certainly be improved by narrowing the scope of application, collecting additional data, and refining model architectures. Yet these measures, while effective in closed or well-defined settings, do not address the deeper challenge posed by open-world conditions. In such contexts, the central task is to design systems capable of operating under conditions of insufficient knowledge and resources. Here, insufficient knowledge and resources [15] should not be regarded as a temporary obstacle to be eliminated, but rather as a fundamental constraint with which intelligent systems must learn to cope.

Tolerating Hallucination as a Structural Phenomenon of Deep Learning

Humans themselves are not immune to cognitive errors. Memory lapses and false recollections, as exemplified by the Mandela Effect, or perceptual illusions such as Adelson’s “checker shadow” and the “rotating snakes” illusion, demonstrate that human cognition can diverge from factual accuracy due to the very mechanisms by which the mind operates. In an analogous way, hallucinations in LLMs arise as a structural byproduct of deep learning. They are not mere technical glitches but follow from the principles that enable such systems to generalize at all.
Accordingly, hallucination should not be dismissed as entirely “wrong” from a human standpoint, just as human illusions are not entirely “wrong” from the perspective of the underlying perceptual mechanisms. To regard hallucinations as categorically unacceptable risks an overly anthropocentric standard; conversely, to regard human errors as unacceptable from the perspective of deep learning would be equally misplaced.

Tolerating Limited Experience and Maintaining Adaptivity

Under the Open World assumption, intelligent systems must cope with an unbounded future while relying only on finite and incomplete past experience. The inevitable mismatch means that summarization of past data cannot guarantee consistency with future cases, echoing Hume’s critique of induction. Nonetheless, systems can remain effective by maintaining the capacity to adapt: continuously revising their internal models, updating beliefs, and incorporating new information as it becomes available.

Making Errors Intelligible and Acceptable to Humans

If errors are inevitable, then the design challenge shifts from their complete elimination to making them understandable and acceptable from a human perspective. Some LLM errors appear arbitrary or absurd because they are grounded in representational forms opaque to human reasoning. A promising direction is to adopt more transparent representational schemes, such as logical and concept-centered representations [16], which align more closely with human conceptual structures. By doing so, even when errors occur, they can be interpreted as reasonable within a shared framework of thinking.

6. Conclusions

Errors—some of which are characterized as “hallucination”—are inevitable for learning systems, including LLMs especially, operating under the Open World assumption. The central difficulty lies in showing why hallucination is inevitable in a principled way. This should not be attempted merely by analyzing special cases, since any finite set of counterexamples could in principle be incorporated into training data, thereby eliminating those particular errors. Another extreme is also unhelpful: to equate hallucinations with errors in general and to argue that errors must occur in a certain environment that forces their occurrence. In contrast, we have argued that “hallucination” is a distinctive subclass of errors, common to both LLMs and other learners. While systems can produce correct outputs in relatively stable environments, the limitation of generalization implies that finite past experience can never guarantee accuracy across an unbounded future.
To summarize, the line of argument developed in this paper proceeds in three steps. First, the behavior of an LLM can be abstracted as a mapping from an input space of linguistic strings to an output space of linguistic strings. Second, any such mapping necessarily faces the familiar problem of generalization: no learner, human or artificial, can guarantee perfect alignment between past experience and future cases, a limitation well recognized in classical discussions of machine learning. Third, once we adopt the Open World assumption, in which the range of space-time and tasks is unbounded, these generalization errors—framed as “hallucination” in the context of LLMs—become inescapable. Under such conditions, hallucination should not be regarded as a contingent flaw to be engineered away, but rather as a structural byproduct of intelligence itself.
At the same time, this does not undermine the value of reducing “hallucinations” under more restricted conditions. In domains where tasks, and space and time are sufficiently constrained, the environment can be approximated as a closed world, and hallucinations can often be rendered rare or negligible through additional data and careful optimization. Such efforts remain worthwhile, particularly for applied systems operating within specific boundaries. Nevertheless, for AGI the Open World must be treated as part of the problem definition itself [13], not as an imperfection to be eliminated in engineering practice.

Acknowledgments

The author appreciates the comments from Boyang Xu.

Appendix A. Counter Example

Figure A1. An Adversarial World
Figure A1. An Adversarial World
Preprints 179763 g0a1

References

  1. Bubeck, S.; Chandrasekaran, V.; Eldan, R.; Gehrke, J.; Horvitz, E.; Kamar, E.; Lee, P.; Lee, Y.T.; Li, Y.; Lundberg, S.; et al. Sparks of artificial general intelligence: Early experiments with gpt-4. arXiv arXiv:2303.12712 2023.
  2. Huang, L.; Yu, W.; Ma, W.; Zhong, W.; Feng, Z.; Wang, H.; Chen, Q.; Peng, W.; Feng, X.; Qin, B.; et al. A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions. ACM Trans. Inf. Syst. 2025, 43, 42:1–42:55. [Google Scholar] [CrossRef]
  3. Cossio, M. A comprehensive taxonomy of hallucinations in Large Language Models, 2025. arXiv:2508.01781 [cs]. [CrossRef]
  4. Kalai, A.T.; Nachum, O.; Vempala, S.S.; Zhang, E. Why Language Models Hallucinate, 2025. arXiv:2509.04664 [cs]. [CrossRef]
  5. Xu, Z.; Jain, S.; Kankanhalli, M. Hallucination is Inevitable: An Innate Limitation of Large Language Models, 2024. arXiv:2401.11817 [cs] version: 1. [CrossRef]
  6. Suzuki, A.; He, Y.; Tian, F.; Wang, Z. Hallucinations are inevitable but can be made statistically negligible. The "innate" inevitability of hallucinations cannot explain practical LLM issues, 2025. arXiv:2502.12187 [cs]. [CrossRef]
  7. Shi, Q.; Xi, W.; Ding, Z.; Gao, J.; Yang, X. Hallucination as a Computational Boundary: A Hierarchy of Inevitability and the Oracle Escape, 2025. arXiv:2508.07334 [cs]. [CrossRef]
  8. Banerjee, S.; Agarwal, A.; Singla, S. LLMs Will Always Hallucinate, and We Need to Live with This. In Proceedings of the Intelligent Systems and Applications; Arai, K., Ed., Cham, 2025; pp. 624–648. [CrossRef]
  9. Kaufmann, T.; Weng, P.; Bengs, V.; Hüllermeier, E. A Survey of Reinforcement Learning from Human Feedback. Transactions on Machine Learning Research 2024. [Google Scholar]
  10. Wolpert, D.; Macready, W. No free lunch theorems for optimization. IEEE Transactions on Evolutionary Computation 1997, 1, 67–82. [Google Scholar] [CrossRef]
  11. Hume, D. An Enquiry Concerning Human Understanding; 1748.
  12. Goertzel, B.; Wang, P. Advances in Artificial General Intelligence: Concepts, Architectures and Algorithms : Proceedings of the AGI Workshop 2006; IOS Press, 2007.
  13. Xu, B. What is Meant by AGI? On the Definition of Artificial General Intelligence, 2024. arXiv:2404.10731.
  14. Xu, B. What Do You Mean by "Open World"? In Proceedings of the NeurIPS 2024 Workshop on Open-World Agents, 2024.
  15. Wang, P. On defining artificial intelligence. Journal of Artificial General Intelligence 2019, 10, 1–37. [Google Scholar] [CrossRef]
  16. Wang, P.; Thórisson, K.R. Concept-Centered Knowledge Representation: A ‘Middle-Out’ Approach Fusing the Symbolic-Subsymbolic Divide. In Proceedings of the Situated Self-guided Learning; Robertson, P.; Georgeon, O., Eds., Cham, 2025; pp. 10–52. [CrossRef]
1
This usage should not be confused with the so-called Closed World Assumption in logic, which assumes that any statement not contained in a knowledge base is false.
2
Color versions of this and all subsequent figures are available in the online version.
Figure 1. Illustration of a learned mapping. Black dots represent training samples, while the orange curve denotes the mapping acquired by an LLM. The reduction to two dimensions is for intuitive visualization; the underlying analysis applies equally to high-dimensional spaces.2
Figure 1. Illustration of a learned mapping. Black dots represent training samples, while the orange curve denotes the mapping acquired by an LLM. The reduction to two dimensions is for intuitive visualization; the underlying analysis applies equally to high-dimensional spaces.2
Preprints 179763 g001
Figure 2. A generated output marked “H” (dashed circle) deviates from the training samples, and may therefore be interpreted as a hallucination.
Figure 2. A generated output marked “H” (dashed circle) deviates from the training samples, and may therefore be interpreted as a hallucination.
Preprints 179763 g002
Figure 3. The same output marked “H” in Figure 2 could also correspond to a true fact, represented here as a purple dot “F”. From the learner’s perspective during training, it is indeterminate which mapping is superior.
Figure 3. The same output marked “H” in Figure 2 could also correspond to a true fact, represented here as a purple dot “F”. From the learner’s perspective during training, it is indeterminate which mapping is superior.
Preprints 179763 g003
Figure 4. Type-I Hallucination (HT-I). The output “H” deviates from a known training sample “F”. Since the fact is already contained in the training set, the mapping can be revised until the output aligns with the expected value.
Figure 4. Type-I Hallucination (HT-I). The output “H” deviates from a known training sample “F”. Since the fact is already contained in the training set, the mapping can be revised until the output aligns with the expected value.
Preprints 179763 g004
Figure 5. Illustration of the inadequacy of abandoning generalization. The output “U” corresponds to an “I don’t know” response.
Figure 5. Illustration of the inadequacy of abandoning generalization. The output “U” corresponds to an “I don’t know” response.
Preprints 179763 g005
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated