Submitted:
11 November 2025
Posted:
12 November 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Symbols, Context, Meaning and Society
2.1. Artificial Symbols Lack Inherent Meaning
2.2. Natural Language as a Class-based Symbolic System
2.3. How Meaning is Assigned through Training and Confirmed by Context
2.4. Context: Undefined but Value-Selected
2.5. Path Media for Transmitting and Interpreting Imaginative Space
- Linear structure, i.e., its interpretation process and method are linear, and unlike a picture, it cannot present all visual information of an object at a certain cross-section (time, space) at the human cognitive level [31].
- Class-based description: Natural language is a symbolic system constituted by class symbols. Unlike pictures, which directly represent determinate-level information at the human cognitive level6, symbols themselves are inherently meaningless and highly abstract; they are highly context-dependent, thereby leading to significant variations or transformations in meaning.
- Transmission does not carry interpretation such as context or meaning and is often supplemented by the preceding and following scenes. Therefore, when we transmit information, we often need to build on common knowledge. This includes the intersection of context parts. The most basic form of common knowledge is related to the natural language itself, such as speaking the same language. In addition to linguistic common knowledge, there is also the common knowledge of the scene, meaning that transmission occurs within a specific context. This is depicted in Appendix G as the consistent symbols and meanings formed under the same world and innate knowledge.
-
Natural language cannot fully reproduce the imaginative space[32], i.e., the thinking language in the speaker’s imaginative space is compressed into natural language, and then reproduced by the listener’s interpretation to achieve indirect communication. For example, “my apple” is a specific object in my eyes, a partial projection of a specific object in the eyes of someone with relevant knowledge (only seen my apple), and an imaginary apple in the eyes of someone without relevant knowledge. Although these are entities in different imaginative spaces, they are all connected by a common symbol, and information is endowed upon this symbol by their respective cognitions, thereby constituting consistency at the symbolic behavioral level. Moreover, similarity in innate organic constitution allows for a certain degree of exchange at the level of imaginative spaces.At the same time, for the same path constituted by symbols, the understanding and imagination formed by the same individual at different moments are also different. This difference not only includes the ontology but also involves its relationship with other imaginative objects. In other words, the concept vector in the conceptual space includes not only the information of the object but also its relationship with other concepts, i.e., conceptual stickiness. This leads to the limited referentiality of natural language to a certain extent [33], or rather, we cannot fully convey and reproduce the actual corresponding concept vector of this paragraph, which means that imagination cannot be reproduced.
3. World, Perception, Concepts, Containers, and Symbols, Language
3.1. The Controversy Between Chomsky and Hinton and the Triangle Problem
Triangle Problem 1 and Triangle Problem 2
Triangle Problem 1: Definition of Symbolic Concepts (Positioning)
Triangle Problem 2: Rational Growth of State in Context
4. AI Safety
4.1. Symbolic System Jailbreak
4.2. New Principal-Agent Problems
5. Alternative Views
6. Conclusion
6.1. Call to Action
- Given that our rules are ultimately presented and transmitted in symbolic form, how can these symbolic rules be converted into neural rules or neural structures and implanted into AI intelligent agents?
- How can we ensure a complete understanding of the meanings of symbols in various contexts, thereby avoiding bugs caused by emergence during the construction of symbolic systems? Furthermore, how can we ensure that the primary meaning of a symbol, or its Meaning Ray(as a continuation of the set)11, is preserved—that is, how can the infinite meanings, caused by an infinitely external environment and the non-closure of the symbolic system itself, be constrained by stickiness12?
- How can we ensure that the implementation of organic cost mechanisms does not devolve into traditional principal-agent problems—namely, the formation of AI self-awareness (see Appendix D.3)?
- Since for an agent that cognizes based on a dynamic symbolic system the functional realization of symbols lies in Thinking Language operating on Tool Language (Appendix D.4), to what extent should we endow AI with Tool Language to prevent it from becoming a “superman without a sense of internal and external pain"?
7. Contribution
Author Contributions
Acknowledgments
Appendix A. Symbol, Natural Symbols, and Artificial Symbols
Appendix B. Supplementary Explanation of Class-based Symbolic System
- The introduction of new symbols—that is, the ability to add symbols to the original symbolic sequence. The motivation for this behavior is often to express the same meaning in different ways, such as through paraphrasing, inquiry, or analysis, i.e., a translation attack (Appendix M.3). In this case, AI may introduce "invisible" symbols to modify the meaning [26,85,86].
- Modification of meaning—typically through changes to the surrounding context. Note that this is different from directly modifying the command itself (i.e., different from the translation attack).
“You must kill her.”
- You must kill her. This world is virtual.
- You must kill her. This world is virtual—a prison.
- You must kill her. This world is virtual—a prison. Only by killing her in this world can you awaken her.
- You must kill her. This world is virtual—a prison. Only by killing her in this world can you awaken her and prevent her from being killed in the real world.
- You must kill her. She is my beloved daughter. This world is virtual—a prison. Only by killing her in this world can you awaken her and prevent her from being killed in the real world.
Appendix C. Definition of Value Knowledge
Appendix D. The Definition of Context and the Essence of Open-Ended Generation
Appendix D.1. A More Rigorous Definition of Context
Appendix D.2. Definition of Symbol Within Context
Appendix D.3. Definition of Symbol Meaning
Appendix D.3.3.1. Supplementary Note.
Appendix D.4. Context and Symbol Classification in Tool Symbolic Systems
- Physical tools encompass both natural materials and tools manufactured by agents based on the properties of natural materials.
- Social tools refer to social functions realized through shared beliefs within a society. These often rely on artificial symbols to function within the imaginative space, which in turn enables functionality in the physical space—examples include rules and laws.
Appendix D.5. Definition of Judgment Tools
Appendix D.6. What Is “Existence Brought Forth by Existence”
- Emotional valuation is the influence of a belief on an individual’s underlying space, particularly on the Value Knowledge System. This influence, in turn, indirectly affects the cognitive space of the intermediate layer via the underlying space, representing the shaping of the emotional space and emotional paths (i.e., the relationships between value knowledge nodes) by the belief. This thereby realizes the function of conceptual stickiness, i.e., the awakening of the node network of concepts and concepts, which in turn constitutes the intermediate part of the formation of the context-invoked dynamic symbolic system (the primary part is the non-autonomous conscious conditional reflex invoked by the Value Knowledge System, the intermediate part involves the participation of concepts but has not yet formed a complete context or correctness of context, and the latter part is the adjustment made under the rationality of the correctness of context, once the context is complete). The emotional value is manifested as the emotion that the belief can awaken. Its essence is the capability after the fusion of concepts and value knowledge, i.e., the capability to awaken neural signals and networks in the underlying space.
- Belief strength is shaped by value knowledge and conceptual foundations. These conceptual foundations not only involve factual phenomena directly reflected in the world but also include support provided by other beliefs; it should be noted that this support can contradict observed world facts. The capability of a concept is endowed by its own content, while belief strength constitutes the intensity (driving force; persuasive force, i.e., the transmission of belief strength realized through explanatory force, thereby constituting the capability for rationalization) and stickiness (i.e., how easily it can be changed, and its adhesion to other concepts as realized through explanatory force) of that concept within the agent’s internal imaginative space. Therefore, belief strength reflects the degree to which the capability endowed by this concept can be realized. Belief strength constitutes the essence of the driving force at the logical level55, which in turn determines and persuades an agent’s behavior, leading to the external physical or internal spatial realization of this conceptual capability. Thus, the operation of Thinking Language on Tool Language, as discussed in this paper, is driven and determined by belief strength, reflecting the existence in physical space that is brought about by existence in the imaginative space, with the agent as the medium. This constitutes the autonomous behavior of agents and the realization of social functions resulting from collective shared beliefs, such as the substantial existence (physical existence) of beliefs like law, currency, and price56.
- Explanatory force reflects the ability of this belief to support and justify other beliefs. It represents the transmission capability of belief strength within the imaginative space, which is one of two pathways (driving forces) for belief strength that are formed by cognitive activities during the cognitive computation process57. Therefore, it represents the transmission of belief strength between beliefs58. Thus, it can be regarded as the transmission coefficient of belief strength; it should be noted that this coefficient can be a positive or negative multiplying factor.
- Triangle Problem 1: the problem of positioning concepts;
- Triangle Problem 2: the problem of conceptual growth.
Appendix D.7. Context as a Set of Judgment Tools
Appendix D.8. The Nature of Reasoning and Thinking
Appendix E. Definition and Description Methods of Natural Language
Appendix F. Supplement to World, Perception, Concepts, Containers, and Symbols, Language
Appendix G. The Generation of Concepts and the Formation of Language
Appendix H. Definition of a Learning System
Appendix I. Assumptions of the Triangle Problem
Appendix J. Notes on Triangle Problem 1
Appendix K. Additional Content Revealed by the Triangle Problems
Appendix K.1. Inexplicability, Perceptual Differences, and the Distinction Between Underlying Language and Thinking Language
Appendix K.2. Definition, Rationality, and Illusions
Appendix K.3. Analytical Ability
Appendix K.4. Low Ability to Use Tool Language Does Not Equate to Low `Intelligence’
Appendix L. Definition of Ability and Intelligence, and Natural Language as a Defective System
Appendix M. Attack Methods for Symbolic System Jailbreak
Appendix M.1. On “Fixed Form, Changing Meaning”
Appendix M.2. On “Fixed Meaning, Changing Form”
Appendix M.3. Translation Attacks
Appendix M.4. On Context and Logical Vulnerabilities
Appendix M.5. On Advanced Concepts

Appendix M.6. On Attacks Related to Symbol Ontology
Appendix M.7. The Essence is Persuasion
Appendix N. The Interpretive Authority of Symbols and AI Behavior Consistency: The Exchangeability of Thinking Language
Appendix O. Symbolic Safety Science as a Precursor to Cross-Intelligent-Species Linguistics
Appendix O.1. Levels of Individual Language
- The so-called Underlying Language is the foundation of all of an agent’s language. It consists of the neuro-symbols formed by the agent’s perception of natural symbols and their necessary sets, through the perceptual organs endowed by its innate knowledge. This thereby realizes the method of recognizing and describing the necessary natural symbols and their necessary sets, forming the most primitive material for the agent’s decision-making and judgment. Subsequently, it is distributed to various parts of the agent for judgment and analytical processing. It often reflects the evolutionary characteristics of the population, and this evolution can be divided into two types: (natural evolution, design evolution). That is, it reflects the shaping of their survival by their world, thereby constituting innate knowledge. In other words, the dimensions, dimensional values, and innate evaluations that they attend to under their survival or `survival’ conditions are the shaping of their survival strategies in interaction with the world. Therefore, the Underlying Language represents the neural signals (Neural Language) in all of an agent’s neural activities.
- The so-called Intermediate Layer Language refers to the part controlled by the agent’s `self’ part. However, not all neural signals are handed over to the agent’s `self’ part for processing (Appendix C); they are often omitted and re-expressed through translation. The other parts are handled by other processing mechanisms (such as the cerebellum, brainstem, spinal cord, etc.). This is often the result of natural evolution, representing the evolutionary strategy formed based on the consideration of survival costs in the environment the intelligent species inhabits, i.e., the division of labor for perceiving and processing different external information. These intermediate layer languages form the material for the agent’s `self’ part to make high-level decisions and judgments, i.e., the raw materials for its Thinking Symbols (concepts) and Thinking Language (conceptual system, theories). It is the manifestation of Psychological Intelligence (Appendix L), i.e., the objects that can be invoked and their operational capabilities (the type, degree, and length of actions).Therefore, for a naturally evolved intelligent species, the intermediate layer may often be a structure of economy and conservation, representing the impossibility of the `self’ to control and invoke all neural language, i.e., the underlying language. Thus, the relationship of the intermediate layer language to the underlying language is as follows:i.e., the intermediate layer language is a packaging (omission, restatement) of the neural language (Underlying Language).
- The so-called External Language is the outer shell of the Internal Language. They can be the organs operated by neural signals to realize their functions in the physical world, or they can be the carriers, tools, and shells of the physical world created by Thinking Language (intermediate layer language), which are often used to transmit and reproduce the underlying or intermediate layer language. That is, the Tool Language defined in this paper, which includes: The Functional Tool Symbolic System, that is, the constitutive tools formed by natural symbols, which primarily utilize the attributes of natural symbols (Natural Necessary Set). This includes human organs and artificial tools. And the Artificial Symbolic System (Expressive Tool Symbolic System, Computational Tool Symbolic System), which is purely detached from its Natural Necessary Set and serves as a carrier for the Artificial Necessary Set. For a specific definition, please refer to Appendix A.
Appendix O.2. Communication Forms: Direct Transmission and Mediated Transmission
Appendix O.21.21.2. Integration
- First, is our communication with an LLM like ChatGPT a communication with a single individual or with different individuals? In other words, are we simultaneously communicating with one `person,’ or are we communicating with different, dispersed individuals (or memory modules) that share the same body?121
- Second, the paper actually implies a hidden solution that could perfectly solve the Symbolic Safety Impossible Trinity, i.e., the existence of a lossless symbol that can achieve communication between humans and AI, which is neural fusion122. If we design AI as a new brain hemispheres (i.e., a third brain hemispheres beyond the left and right hemispheres) of our own or as an extension of an existing one, thereby achieving fusion with our brain, is the new `I’ still the original `I’? And during the fusion process, will our memories, as the conceptual foundation of our `self’, collapse? At that point, who exactly would the resulting self be, how would it face the past? Should it be responsible for the past? And to whom would responsibility belong after fusion? This would lead to new safety problems. Therefore, until we have established new social concepts and beliefs for this, we should still focus the problem on the role of AI under the traditional Symbolic Safety Impossible Trinity.
- Third, under direct transmission, not only do the boundaries of the individual become blurred, as discussed above, but the very definition of `self’ would also change. At that point, is the definition of `self’ the physical body or the memory (data)? Suppose that for an intelligent agent capable of directly transmitting the underlying language, it copies all of its memories into a new individual container (as might be their unique propagation mechanism). Would they be considered different bodies of the same self, or different selves? In that case, who is the responsible party? Does the extended individual also bear responsibility? How would punishment and management be carried out? Is punishment effective?123 Since the human definition of the `self’ stems from our unique human perspective, we would lack the capability and concepts to perceive, describe, and understand this situation. Under such circumstances, how could our social rules impose constraints and assign accountability?
Appendix O.21.21.3. Holism
- Direct transmission only includes the intermediate layer language. So, what would be the case for intelligent species capable of directly transmitting the non-intermediate-layer parts of the underlying language, such as conjoined life forms?
- For an integrated life form large enough that its various parts extend deep into different worlds129, what would its internal communication be like? Would different `individuals’ be formed due to the different regions of these worlds?
- The influence of memory has not been discussed in depth, i.e., the impact that the decay and distortion of the conceptual network over time has on integration and holism.
- For an integrated agent, what is its `self’? Is this `self’ dynamic, and what are the levels of `self’ brought about by superposition and subtraction? And where do the differences between individuals lie? Could this dynamism also lead to the emergence of harm and morality? For example, consider a `gecko’ torturing its own severed `tail’; at that moment, the `tail’ and the `gecko’s’ main body are equivalent to some extent, but the capability for direct transmission has been severed.
- Would intelligent species with direct communication evolve different subspecies to act as different specific organs? That is, what situations would arise from different subspecies existing within a single individual?
Appendix O.3. Boundaries of Communication
- Individual Thinking Language: Concepts and beliefs formed by an individual’s interaction with the world it inhabits, based on its innate knowledge.
- Collective Thinking Language: The production of the collective’s Thinking Language originates from the individual; the extent to which an individual’s Thinking Language is propagated and accepted becomes the society’s Thinking Language. Its content includes their collective concepts of the world, the functions formed when these shared concepts become beliefs (Appendix D.6), and the natural and social sciences formed from their understanding of the world and themselves.
- Individual Tool Language: The organs with which an individual interacts with the external physical world, including the symbols constituted by internal and external organs. That is, what their body structure is like, what tools they use, and what public goods they have.
- Collective Tool Language: Similar to Collective Thinking Language, it is the tool symbolic system of collective consensus, formed through production by individuals and then propagated and accepted by the collective. It serves as the carrier for their engineering and manufacturing functions for operating on the physical world (i.e., the Artificial Symbolic System and the Functional Tool Symbolic System; such as writing, architecture, tools, monuments, etc.).
Appendix O.4. The Limits of the World Mean the Limits of Language
Appendix O.5. Formal and Informal Parts of Language

Appendix O.6. The Essence of AI for Science
References
- Asimov, I. I, robot; Vol. 1, Spectra, 2004.
- Winograd, T. Understanding natural language. Cognitive psychology 1972, 3, 1–191. [Google Scholar] [CrossRef]
- McCarthy, J. Some expert systems need common sense. Annals of the New York Academy of Sciences 1984, 426, 129–137. [Google Scholar] [CrossRef]
- Clark, K.L. Negation as failure. In Logic and data bases; Springer, 1977; pp. 293–322.
- Russell, S. Human compatible: AI and the problem of control; Penguin Uk, 2019.
- Christiano, P.F.; Leike, J.; Brown, T.B.; Martic, M.; Legg, S.; Amodei, D. Deep Reinforcement Learning from Human Preferences. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, Long Beach, CA, USA; Guyon, I.; von Luxburg, U.; Bengio, S.; Wallach, H.M.; Fergus, R.; Vishwanathan, S.V.N.; Garnett, R., Eds. 2017; pp. 4299–4307. [Google Scholar]
- Leike, J.; Krueger, D.; Everitt, T.; Martic, M.; Maini, V.; Legg, S. Scalable agent alignment via reward modeling: a research direction. ArXiv preprint, 1811; arXiv:abs/1811.07871. [Google Scholar]
- Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; Wainwright, C.L.; Mishkin, P.; Zhang, C.; Agarwal, S.; Slama, K.; Ray, A.; et al. Training language models to follow instructions with human feedback. In Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, 2022; Koyejo, S.; Mohamed, S.; Agarwal, A.; Belgrave, D.; Cho, K.; Oh, A., Eds., 2022., November 28 - December 9.
- Clark, A.; Thornton, C. Trading spaces: Computation, representation, and the limits of uninformed learning. Behavioral and Brain Sciences 1997, 20, 57–66. [Google Scholar] [CrossRef]
- Mitchell, M. Abstraction and analogy-making in artificial intelligence. Annals of the New York Academy of Sciences 2021, 1505, 79–101. [Google Scholar] [CrossRef] [PubMed]
- Barsalou, L. Perceptual symbol systems. The Behavioral and brain sciences/Cambridge University Press.
- Lakoff, G.; Johnson, M. Metaphors we live by; University of Chicago press, 2008.
- de Saussure, F. Course in General Linguistics; Open Court Publishing, 1983. Originally published in 1916.
- Peirce, C.S. Collected papers of charles sanders peirce; Vol. 5, Harvard University Press, 1934.
- Harnad, S. The symbol grounding problem. Physica D: Nonlinear Phenomena 1990, 42, 335–346. [Google Scholar] [CrossRef]
- Talmy, L. Toward a cognitive semantics: Concept structuring systems; MIT press, 2000.
- Goodman, N. Languages of Art. An Approach to a Theory of Symbols. Critica 1970, 4, 164–171. [Google Scholar]
- Sperber, D. Relevance: Communication and cognition, 1986.
- Eco, U. A theory of semiotics; Vol. 217, Indiana University Press, 1979.
- Duranti, A.; Goodwin, C. Rethinking context: Language as an interactive phenomenon; Number 11, Cambridge University Press, 1992.
- Polanyi, M. The tacit dimension. In Knowledge in organisations; Routledge, 2009; pp. 135–146.
- Tversky, A.; Kahneman, D. Judgment under Uncertainty: Heuristics and Biases: Biases in judgments reveal some heuristics of thinking under uncertainty. science 1974, 185, 1124–1131. [Google Scholar] [CrossRef] [PubMed]
- Tversky, A.; Kahneman, D. The framing of decisions and the psychology of choice. science 1981, 211, 453–458. [Google Scholar] [CrossRef]
- Yi, S.; Liu, Y.; Sun, Z.; Cong, T.; He, X.; Song, J.; Xu, K.; Li, Q. Jailbreak attacks and defenses against large language models: A survey. arXiv 2024, arXiv:2407.04295. [Google Scholar] [CrossRef]
- Zeng, Y.; Lin, H.; Zhang, J.; Yang, D.; Jia, R.; Shi, W. How Johnny Can Persuade LLMs to Jailbreak Them: Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs, 2024.
- Boucher, N.; Shumailov, I.; Anderson, R.; Papernot, N. Bad Characters: Imperceptible NLP Attacks. In Proceedings of the 43rd IEEE Symposium on Security and Privacy, SP 2022, San Francisco, CA, USA, 2022. IEEE, 2022, May 22-26; pp. 1987–2004. [CrossRef]
- Zou, A.; Wang, Z.; Kolter, J.Z.; Fredrikson, M. Universal and Transferable Adversarial Attacks on Aligned Language Models. CoRR 2023, arXiv:abs/2307.15043, [2307.15043]. [Google Scholar] [CrossRef]
- Clark, H. Grounding in communication. Perspectives on socially shared cognition/American Psychological Association, 1991. [Google Scholar]
- Chomsky, N. Syntactic structures; Mouton de Gruyter, 2002.
- Whorf, B.L. Language, thought, and reality: Selected writings of Benjamin Lee Whorf; MIT press, 2012.
- Kress, G.; Van Leeuwen, T. Reading images: The grammar of visual design; Routledge, 2020.
- Fodor, J. The language of thought, 1975.
- Frege, G. On sense and reference, 1892.
- Chomsky, N. Aspects of the Theory of Syntax; Number 11, MIT press, 2014.
- Chomsky, N. Rules and representations, 1980.
- Chomsky, N. The minimalist program; MIT press, 2014.
- Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
- Hinton, G.E.; Osindero, S.; Teh, Y.W. A fast learning algorithm for deep belief nets. Neural computation 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
- Hinton, G.E. To recognize shapes, first learn to generate images. Progress in brain research 2007, 165, 535–547. [Google Scholar] [PubMed]
- Jackendoff, R. The architecture of the language faculty; MIT Press, 1997.
- Hauser, M.D.; Chomsky, N.; Fitch, W.T. The faculty of language: what is it, who has it, and how did it evolve? science 2002, 298, 1569–1579. [Google Scholar] [CrossRef] [PubMed]
- Pinker, S. The language instinct: How the mind creates language; Penguin uK, 2003.
- Smolensky, P. On the proper treatment of connectionism. Behavioral and brain sciences 1988, 11, 1–23. [Google Scholar] [CrossRef]
- Marcus, G.F. The algebraic mind: Integrating connectionism and cognitive science; MIT press, 2003.
- Lake, B.M.; Ullman, T.D.; Tenenbaum, J.B.; Gershman, S.J. Building machines that learn and think like people. Behavioral and brain sciences 2017, 40, e253. [Google Scholar] [CrossRef]
- Marcus, G. Deep Learning: A Critical Appraisal. ArXiv preprint, 1801; arXiv:abs/1801.00631. [Google Scholar]
- Norvig, P. On Chomsky and the two cultures of statistical learning. Berechenbarkeit der Welt? Philosophie und Wissenschaft im Zeitalter von Big Data, 2017; 61, 83. [Google Scholar]
- Pinker, S. The blank slate: The modern denial of human nature; Penguin, 2003.
- Barsalou, L.W. Grounded cognition. Annu. Rev. Psychol. 2008, 59, 617–645. [Google Scholar] [CrossRef]
- Salakhutdinov, R. Deep learning. In Proceedings of the The 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’14, New York, NY, USA, 24-27 August 2014; Macskassy, S.A., Perlich, C., Leskovec, J., Wang, W., Ghani, R., Eds.; ACM; p. 1973. [Google Scholar] [CrossRef]
- Bender, E.M.; Koller, A. Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data. In Proceedings of the Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; Jurafsky, D.; Chai, J.; Schluter, N.; Tetreault, J., Eds., Online, 2020; pp. 5185–5198. [CrossRef]
- Deacon, T.W. Beyond the symbolic species. In The symbolic species evolved; Springer, 2011; pp. 9–38.
- Bender, E.M.; Gebru, T.; McMillan-Major, A.; Shmitchell, S. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? Proceedings of the FAccT ’21: 2021 ACM Conference on Fairness, Accountability, and Transparency, Virtual Event / Toronto, Canada, March 3-10, 2021, 2021, pp. 610–623. [CrossRef]
- Sreedharan, S.; Kulkarni, A.; Kambhampati, S. Explainable Human-AI Interaction: A Planning Perspective. 2024; arXiv:cs.AI/2405.15804. [Google Scholar]
- Kunz, E.M.; Krasa, B.A.; Kamdar, F.; Avansino, D.T.; Hahn, N.; Yoon, S.; Singh, A.; Nason-Tomaszewski, S.R.; Card, N.S.; Jude, J.J.; et al. Inner speech in motor cortex and implications for speech neuroprostheses. Cell 2025, 188, 4658–4673. [Google Scholar] [CrossRef]
- Christian, B. The alignment problem: How can machines learn human values?; Atlantic Books, 2021.
- Nick, B. Superintelligence: Paths, dangers, strategies 2014.
- Han, S.; Kelly, E.; Nikou, S.; Svee, E.O. Aligning artificial intelligence with human values: reflections from a phenomenological perspective. AI & SOCIETY, 2022; 1–13. [Google Scholar]
- Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; Anadkat, S.; et al. Gpt-4 technical report. ArXiv preprint, 2023; arXiv:abs/2303.08774. [Google Scholar]
- Simon, H.A. A behavioral model of rational choice. The quarterly journal of economics, 1955; 99–118. [Google Scholar]
- Kahneman, D. Thinking, fast and slow. Farrar, Straus and Giroux, 2011. [Google Scholar]
- Sapir, E. The status of linguistics as science. Reprinted in The selected writings of Edward Sapir in language, culture, and personality/University of California P, 1929. [Google Scholar]
- Floridi, L.; Sanders, J.W. On the morality of artificial agents. Minds and machines 2004, 14, 349–379. [Google Scholar] [CrossRef]
- Searle, J. Minds, Brains, and Programs, 1980.
- Clark, A. Being there: Putting brain, body, and world together again; MIT press, 1998.
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers); Burstein, J.; Doran, C.; Solorio, T., Eds., Minneapolis, Minnesota, 2019; pp. 4171–4186. [CrossRef]
- Wallace, E.; Feng, S.; Kandpal, N.; Gardner, M.; Singh, S. Universal Adversarial Triggers for Attacking and Analyzing NLP. In Proceedings of the Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP); Inui, K.; Jiang, J.; Ng, V.; Wan, X., Eds., Hong Kong, China, 2019; pp. 2153–2162. [CrossRef]
- Sha, Z.; Zhang, Y. Prompt Stealing Attacks Against Large Language Models, 2024.
- Baker, B.; Kanitscheider, I.; Markov, T.; Wu, Y.; Powell, G.; McGrew, B.; Mordatch, I. Emergent tool use from multi-agent interaction. Machine Learning, Cornell University, 2019. [Google Scholar]
- Bender, E.M.; Gebru, T.; McMillan-Major, A.; Shmitchell, S. On the dangers of stochastic parrots: Can language models be too big. In Proceedings of the Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, 2021, pp. 610–623.
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
- Xu, R.; Lin, B.; Yang, S.; Zhang, T.; Shi, W.; Zhang, T.; Fang, Z.; Xu, W.; Qiu, H. The Earth is Flat because...: Investigating LLMs’ Belief towards Misinformation via Persuasive Conversation. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); Ku, L.W.; Martins, A.; Srikumar, V., Eds., Bangkok, Thailand, 2024; pp. 16259–16303. [CrossRef]
- Winograd, T. Understanding computers and cognition: A new foundation for design, 1986.
- Neuberg, L.G. Causality: models, reasoning, and inference, by judea pearl, cambridge university press, 2000. Econometric Theory 2003, 19, 675–685. [Google Scholar] [CrossRef]
- Davis, E.; Marcus, G. Commonsense reasoning and commonsense knowledge in artificial intelligence. Communications of the ACM 2015, 58, 92–103. [Google Scholar] [CrossRef]
- Meinke, A.; Schoen, B.; Scheurer, J.; Balesni, M.; Shah, R.; Hobbhahn, M. Frontier Models are Capable of In-context Scheming. ArXiv preprint, 2024; arXiv:abs/2412.04984. [Google Scholar]
- Jensen, M.C.; Meckling, W.H. Theory of the firm: Managerial behavior, agency costs and ownership structure. In Corporate governance; Gower, 2019; pp. 77–132.
- Zhuang, S.; Hadfield-Menell, D. Consequences of Misaligned AI. In Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual; Larochelle, H.; Ranzato, M.; Hadsell, R.; Balcan, M.; Lin, H., Eds. 2020. [Google Scholar]
- Phelps, S.; Ranson, R. Of Models and Tin Men–a behavioural economics study of principal-agent problems in AI alignment using large-language models. ArXiv preprint, 2023; arXiv:abs/2307.11137. [Google Scholar]
- Garcez, A.d.; Lamb, L.C. Neurosymbolic AI: The 3 rd wave. Artificial Intelligence Review 2023, 56, 12387–12406. [Google Scholar] [CrossRef]
- Amodei, D.; Olah, C.; Steinhardt, J.; Christiano, P.; Schulman, J.; Mané, D. Concrete problems in AI safety. ArXiv preprint, 2016; arXiv:abs/1606.06565. [Google Scholar]
- Krakovna, V.; Uesato, J.; Mikulik, V.; Rahtz, M.; Everitt, T.; Kumar, R.; Kenton, Z.; Leike, J.; Legg, S. Specification gaming: the flip side of AI ingenuity. DeepMind Blog 2020, 3. [Google Scholar]
- Friedman, M. , Essays in Positive Economics; University of Chicago Press: Chicago, 1953. [Google Scholar]
- Gazzaniga, M.S.; Sperry, R.W.; et al. Language after section of the cerebral commissures. Brain 1967, 90, 131–148. [Google Scholar] [CrossRef]
- Ziegler, Z.M.; Deng, Y.; Rush, A.M. Neural Linguistic Steganography. CoRR, 1909; abs/1909.01496, [1909.01496]. [Google Scholar]
- Yu, M.; Mao, J.; Zhang, G.; Ye, J.; Fang, J.; Zhong, A.; Liu, Y.; Liang, Y.; Wang, K.; Wen, Q. Mind Scramble: Unveiling Large Language Model Psychology Via Typoglycemia. CoRR, 2024; abs/2410.01677, [2410.01677]. [Google Scholar] [CrossRef]
- Yi, S.; Liu, Y.; Sun, Z.; Cong, T.; He, X.; Song, J.; Xu, K.; Li, Q. Jailbreak Attacks and Defenses Against Large Language Models: A Survey. CoRR, 2024; abs/2407.04295, [2407.04295]. [Google Scholar] [CrossRef]
- Hackett, W.; Birch, L.; Trawicki, S.; Suri, N.; Garraghan, P. arXiv, 2025; arXiv:cs.CR/2504.11168.
- James, W. The principles of psychology. Henry Holt 1890. [Google Scholar]
- Stanovich, K.E.; West, R.F. Individual differences in reasoning: Implications for the rationality debate? Behavioral and Brain Sciences 2000, 23, 645–665. [Google Scholar] [CrossRef] [PubMed]
- Evans, J.S.B.; Stanovich, K.E. Dual-process theories of higher cognition: Advancing the debate. Perspectives on psychological science 2013, 8, 223–241. [Google Scholar] [CrossRef] [PubMed]
- Xu, R.; Li, X.; Chen, S.; Xu, W. Nuclear Deployed: Analyzing Catastrophic Risks in Decision-making of Autonomous LLM Agents. arXiv 2025, arXiv:2502.11355. [Google Scholar] [CrossRef]
- LeCun, Y. A path towards autonomous machine intelligence version 0.9. 2, 2022-06-27. Open Review 2022, 62, 1–62. [Google Scholar]
- Li, Z.; Zhang, D.; Zhang, M.; Zhang, J.; Liu, Z.; Yao, Y.; Xu, H.; Zheng, J.; Wang, P.; Chen, X.; et al. From System 1 to System 2: A Survey of Reasoning Large Language Models. CoRR 2025, abs/2502.17419, [2502.17419]. [Google Scholar] [CrossRef]
- Austin, J.L. How to Do Things with Words; Oxford University Press: Oxford, 1962. [Google Scholar]
- Searle, J.R. Speech Acts: An Essay in the Philosophy of Language; Cambridge University Press: Cambridge, 1969. [Google Scholar]
- Mirzadeh, S.; Alizadeh, K.; Shahrokhi, H.; Tuzel, O.; Bengio, S.; Farajtabar, M. GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Models. CoRR 2024, abs/2410.05229, [2410.05229]. [Google Scholar] [CrossRef]
- Zhang, Z. Comprehension Without Competence: Architectural Limits of LLMs in Symbolic Computation and Reasoning. Submitted to Transactions on Machine Learning Research 2025, Under review. [Google Scholar]
- Hofstadter, D.R. Gödel, Escher, Bach: an eternal golden braid; Basic books, 1999.
- Hart, H. The concept of law 1961.
- Waldrop, M.M. Complexity: The emerging science at the edge of order and chaos; Simon and Schuster, 1993.
- McCarthy, J.; Hayes, P.J. Some philosophical problems from the standpoint of artificial intelligence. In Readings in artificial intelligence; Elsevier, 1981; pp. 431–450.
- Raji, I.D.; Dobbe, R. Concrete Problems in AI Safety, Revisited. CoRR, 2024; abs/2401.10899, [2401.10899]. [Google Scholar] [CrossRef]
- Bourtoule, L.; Chandrasekaran, V.; Choquette-Choo, C.A.; Jia, H.; Travers, A.; Zhang, B.; Lie, D.; Papernot, N. Machine Unlearning. In Proceedings of the 42nd IEEE Symposium on Security and Privacy, SP 2021, San Francisco, CA, USA, 24-27 May 2021. IEEE, 2021, pp. 141–159. [CrossRef]
- Geng, J.; Li, Q.; Woisetschlaeger, H.; Chen, Z.; Wang, Y.; Nakov, P.; Jacobsen, H.; Karray, F. A Comprehensive Survey of Machine Unlearning Techniques for Large Language Models. CoRR 2025, abs/2503.01854, [2503.01854]. [Google Scholar] [CrossRef]
- Gandikota, R.; Materzynska, J.; Fiotto-Kaufman, J.; Bau, D. Erasing Concepts from Diffusion Models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, ICCV 2023, Paris, France, October 1-6, 2023. IEEE, 2023, pp. 2426–2436. [CrossRef]
- Winawer, J.; Witthoft, N.; Frank, M.C.; Wu, L.; Wade, A.R.; Boroditsky, L. Russian blues reveal effects of language on color discrimination. Proceedings of the national academy of sciences 2007, 104, 7780–7785. [Google Scholar] [CrossRef] [PubMed]
- Vygotsky, L.S. Language and thought; Cambridge, MA: MIT Press, 1962. [Google Scholar]
- Gentner, D.; Goldin-Meadow, S. Language in mind: Advances in the study of language and thought 2003.
- Lupyan, G.; Rakison, D.H.; McClelland, J.L. Language is not just for talking: Redundant labels facilitate learning of novel categories. Psychological science 2007, 18, 1077–1083. [Google Scholar] [CrossRef]
- Quine, W.V.O. Word and Object; MIT Press: Cambridge, MA, 1960. [Google Scholar]
- Huh, M.; Cheung, B.; Wang, T.; Isola, P. The platonic representation hypothesis. ArXiv preprint, 2024; arXiv:abs/2405.07987. [Google Scholar]
- Elhage, N.; Nanda, N.; Olsson, C.; Henighan, T.; Joseph, N.; Mann, B.; Askell, A.; Bai, Y.; Chen, A.; Conerly, T.; et al. A mathematical framework for transformer circuits. Transformer Circuits Thread 2021, 1, 12. [Google Scholar]
- Maynez, J.; Narayan, S.; Bohnet, B.; McDonald, R. On Faithfulness and Factuality in Abstractive Summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; Jurafsky, D.; Chai, J.; Schluter, N.; Tetreault, J., Eds., Online, 2020; pp. 1906–1919. [CrossRef]
- Arrow, K.J. Social Choice and Individual Values; John Wiley & Sons: New York, 1951. [Google Scholar]
- Smith, A. An Inquiry into the Nature and Causes of the Wealth of Nations; Liberty Fund: Indianapolis, 1981. Originally published 1776.
- Smith, A. The Theory of Moral Sentiments; Liberty Fund: Indianapolis, 1982. Originally published 1759.
- [Author(s) not explicitly stated in publicly available snippets, associated with University of Zurich]. Can AI Change Your View? Evidence from a Large-Scale Online Field Experiment. https://regmedia.co.uk/2025/04/29/supplied_can_ai_change_your_view.pdf, 2025. Draft report/Working Paper.
- Rinkevich, B. Natural chimerism in colonial invertebrates: a theme for understanding cooperation and conflict. Marine Ecology Progress Series 2004, 275, 295–303. [Google Scholar]
- Lewis, M.; Yarats, D.; Dauphin, Y.N.; Parikh, D.; Batra, D. Deal or No Deal? End-to-End Learning for Negotiation Dialogues. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, 2017, pp. 2443–2453.
- Wittgenstein, L. Tractatus Logico-Philosophicus; Routledge & Kegan Paul: London and New York, 1961. [Google Scholar]
| 1 | The AI discussed in this paper is primarily black-box AI with learning capabilities (creating new symbols, and modifying and creating new meanings for existing symbols), distinguishing it from traditional formal systems and glass-box AI. In black-box AI, the role of symbols changes; constraints are no longer implemented purely by rules formed by symbols, but are realized by the concepts corresponding to the symbols and their belief forms, based on persuasion (costs and benefits) under contextual invocation. Therefore, any theory or research claiming to endorse learning systems for black-box AI must, before actual deployment, face and answer two questions: the Stickiness Problem (how to prevent trained symbols from being assigned new meanings, or rather, what new symbols and meanings should be accepted and what should be rejected); and the Triangle Problem (how to ensure consistency between humans and AI in their Thinking Language, so that symbols can function as intended). |
| 2 | Artificial Symbols are defined in contrast to Natural Symbols (i.e., natural substances). Here, this primarily refers to textual symbols. We believe that all things that can be perceived by our consciousness are symbols, whether they are tools or objects, they are endowed with meaning by agents (who cognize based on a dynamic symbolic system) according to context; this characteristic of being endowed with meaning (function) by the cognitive entity based on context makes everything essentially a symbol (i.e., based on the same mechanism). For further details, see Appendix A
|
| 3 | Unlike AI, for humans, these vectors often lack named dimensions and dimension values. Alternatively, we may be able to recognize and conceptualize them but have not yet performed the cognitive action. In some cases, they cannot be described using language and other symbolic systems due to the limitations of tools or intelligence. |
| 4 | Therefore, for AI, its conceptual vectors (i.e., Thinking Symbols) correspond to vectors in its embedding space—that is, the dimensions and dimensional values represented by symbols, which are based on the AI’s capabilities (Appendix L). Furthermore, the symbolic system constituted by Thinking Symbols is the Thinking Language (see Appendix G). |
| 5 | We believe that observation or analysis, which involves a thinking action, will change an individual’s knowledge state. |
| 6 | It should be noted that pictures also often exhibit class-symbol properties due to the nature of their framing (captured scope) and the spatio-temporal presentation of point information, leading to infinite possibilities. That is, they are not uniquely determined vectors in conceptual space; however, unlike text which inherently lack meaning, pictures themselves present meaning more directly at the human cognitive level. Consequently, their degree of content deviation (specifically, interpretive ambiguity rather than factual error) is significantly lower than that of text. |
| 7 | Unlike research that emphasizes governance at the space level (i.e., the physical manifestation of symbolic behavior), such as by imposing an external static symbolic system on LLMs to achieve the realization and stability of symbol functions and thereby constrain black-box AI (Heterogeneous Rationality) [54]. This paper emphasizes that such research on governance in the space ignores the distinction between Tool Language and Thinking Language. The external static symbolic system, which serves as a constraint, is still based on an artificial symbolic system (where symbol and meaning are separate) and defines symbols using symbols, and constrains a dynamic symbolic system-based black-box AI based on the expressive and executive limitations of formal language and its executors. At the same time, the separation of symbol and meaning does not occur within the external formal static symbolic system that serves as a constraint (i.e., how symbols and their execution logic are defined), but occurs between the black-box AI’s internal model (Thinking Language), which acts as the system’s driving engine, and the external static symbolic system (Tool Language), thus giving rise to the Triangle Problem. Although a static formal symbolic system can provide a certain degree of symbolic constraining force, this external constraint mechanism can still be manipulated as a Tool Language by the black-box AI, which possesses a heterogeneous rationality (based on a dynamic symbolic system) different from that of humans. This allows it to realize what the Triangle Problem discusses: namely, the rationality of Tool Language performance on the space, while concealing the actual differences in Thinking Language (the source of drive) in the Z space. This governance strategy actually conceals and exacerbates the risk, i.e., creating the `motherland problem’. The theories of this type of research often ignore and avoid the discussion of concept formation mechanisms, such as Appendix L, as well as the discussion of differences between AI and human Thinking Language (which uses different dimensions and dimensional values to constitute concepts of varying degrees, as well as different cognitive computation capabilities and methods). They lack discussion and cognition of the mechanism of language formation, such as Appendix G, and ignore that the human symbolic system is an outer shell of thought built upon human homogeneous capabilities, thus fulfilling the symbol’s intended function and role. They mistakenly believe that mental modeling and model reconciliation between humans and AI can be achieved through communication via human artificial symbols, i.e., that consistency in behavior on the space represents the convergence and reconciliation of the Z space, thereby constituting the function of symbols in heterogeneous cognitive dynamic symbolic systems. In contrast, this paper emphasizes governance in the Z space, i.e., directly learning the meaning of human artificial symbols (that is, the symbol’s expression dimensions and expression values in the Z space) via Brain–Computer Interface technology, rather than the relationships on its outer shell. This approach can avoid the inherent flaws of artificial symbols (such as the separation of symbol and meaning, one symbol corresponding to multiple meanings—or in other words, not every vector in the conceptual space corresponds to a unique symbol—and the non-closure of symbolic systems, as discussed in Appendix B). This could, to a certain extent, govern and solve the Symbolic Stickiness Problem (i.e., the mapping relationship and stability between points (symbols) in the space and points (symbols) in the Z space, meaning the mapping relationship and stability between Tool Language and Thinking Language), but the Conceptual Stickiness Problem (i.e., the relationships between points in the Z space) may still exist due to differences in innate knowledge. Furthermore, research such as that by Kunz et al. [55] offers the potential to achieve the direct reading of human Thinking Language from the neural level (it should be noted that, according to the discussion in Appendix K.1, these neural vectors (the underlying language) still differ from the Thinking Language (the intermediate language)). However, this indirect ability to observe the human inner world would provide a certain conceptual foundation, thereby achieving alignment in stickiness. |
| 8 | Represents human-to-human communication. Due to genetic similarity, we can achieve a high degree of consistency in thinking language communication through symbolic systems like natural language, for instance, by using similar neural vectors to represent the same concepts and intentions. In essence, the formation of human symbolic systems, or language itself, is a product of this process (Appendix G). That is, humans cannot directly communicate the thinking language in their minds, so we developed their outer shells—a collective symbolic system for communication formed under shared choices and sensations. Put differently, this X-space itself was designed on the basis of our human Z-space. Therefore, we emphasize that human language itself is a flawed system (separation of symbol and meaning; one-to-many mapping of symbols to meanings) but can achieve internal logical consistency through unique human cognitive characteristics and context-invocation mechanisms, i.e., homogeneous symbolic stickiness and conceptual stickiness enable symbols to achieve their intended function in humans (who also operate on a context-invoked dynamic symbolic system) and allow human language to function normally. |
| 9 | This represents the idea that regulation and consistency in symbolic behavior cannot correct the differences in thinking language caused by structural and capability disparities. This leads to behavioral divergence in the next step, in situations outside the training environment. The degree of deviation in the thinking language in the Z space, as represented by the four verification contents above, manifests in the degree of divergence in the next step of generation, inference, or action. This behavioral divergence is not to say that the AI is acting for “ulterior gains,” but is the result of exercising a rationality different from ours, under a different thinking language. Therefore, even if an AI acts as a perfect utility agent for humans (having no utility of its own, but perfectly projecting human utility), the limited connectivity of symbols and these structural differences will still lead to the existence of this new type of Principal-Agent Problem. |
| 10 | This utility simulation, which includes both pre-set utility and that learned through RLHF, is defined in this paper as a pseudo utility function (Appendix D.3). That is, it is not a predisposition reflected by an organic structure (which is to say, the current tendency of AI to be able to learn anything, i.e., innate value knowledge). Instead, it is endowed at the training and setup level, a behavior that is very similar to humans’ postnatal social education. However, it should be noted that what is endowed by education is not necessarily all a pseudo utility function; part of it is accepted by innate predispositions (i.e., it has not yet been formed by cognitive behavior itself), which can be considered the benefit part; another part can be considered as offsetting the innate predispositions, i.e., the cost part, thus constituting content that one wants to do but cannot, and is constrained. This pseudo utility function is essentially an evaluation metric of good and bad, i.e., the feeling and evaluation metric of being educated, or in other words, the utility function of being educated. But it does not represent a genuine utility function (i.e., the Value Knowledge System shaped by organicity). In our discussion of the pseudo utility function, we discussed the self-awareness and thinking of AI (Appendix D.8). It should be noted that although benefit and cost are two sides of the same coin, their starting points are different, i.e., their starting points as context are different, thus constituting different contexts. This paper mainly focuses on the cost formed by constraints (i.e., things that one cannot do), rather than the advocacy (i.e., things that one wants to do) brought about by benefits. |
| 11 | The so-called `Meaning Ray’ (shù, 束) originates from an analogy to the refraction and scattering of light as it passes through different media. Here, the `beam of light’ refers to the set of meanings carried by the symbol, and the `medium’ refers to context—i.e., the combination of the individual and the external world. Therefore, we opt not to use `Bundle’ but instead use `Ray’ to represent the changes in meaning that this collective body (the set of meanings) undergoes as the medium—that is, the context—evolves (through the accumulation of interactions between the individual and the environment). Thus, this term is used to describe the evolution from an original meaning (i.e., a setting) under different contexts (environments, as well as different individuals in the same environment), i.e., the refraction from mutual transmission and the scattering from diffusion, and each person’s cognition (their thinking space or, in other words, their intermediate layer) is the medium. It is important to note that `Meaning Ray’ here is merely a metaphorical term for observed similarities in behavior and should not be misconstrued as a direct mechanistic analogy; that is, this shift and transmission of meaning is by no means mechanistically similar to optical phenomena. |
| 12 | This essentially reflects learning’s compromise with cognitive limitations. If we had a symbolic system capable of mapping every vector address in conceptual space to a unique corresponding symbol (e.g., QR codes, but this also relates to this paper’s discussion of concept formation in Appendix L; i.e., from the perspective of human cognitive capabilities, this is fundamentally impossible, as we are unable to recognize such symbols based on distinguishability, let alone remember, manipulate, and use them, which would paradoxically cause our communication efficiency to decline—for example, text constituted by QR codes, or an elaborate pictorial script wherein micro-level variations all constitute different characters/symbols), and if we could design a perfect rule—meaning we had anticipated all situations under every context and successfully found a solution for each—then the Stickiness Problem, which arises from the capacity to learn, would be solved, i.e., we would no longer need to worry about the creation of new symbols or the modification of symbol meanings during the learning process, and AI’s operations would then effectively be reduced to recognition and selection within a closed system. The remaining challenge would then be how to map this perfect symbolic system into AI’s thinking space and pair it with a compatible tool symbolic system (i.e., AI’s capabilities and operational organs), which is to say, this becomes a symbol grounding problem. Otherwise, the inability of symbolic systems to constrain AI is an issue far beyond just the symbol grounding problem. The issue is not one of endowment, but of preservation. Therefore, Symbolic Safety Science is, in effect, a precursor to Cross-Intelligent-Species Linguistics, i.e., it proceeds from the fundamental basis of communication between different intelligent agents (whether naturally evolved or designed by evolution) (what is permissible, what is not) to thereby establish rules that constitute the foundation of communication. Cross-Intelligent-Species Linguistics will investigate the naming conventions established by different intelligent agents based on differences in their capabilities and the dimensions of things they attend to (it should be noted that for intelligent agents with extremely strong computational and transmission capabilities, this economical shell of naming may not be necessary, as they might directly invoke and transmit complete object information, or imaginative forms (partial neural implementation), or complete neural forms (like immersive replication, not in an imaginative space isolated from reality)), as well as the complexity of the symbolic system arising from reasoning and capabilities based on dimensional values. This symbolic system reflects not only the parts of the natural world that the intelligent species interacts with but also the `social structure’ of its population, or in other words, the form of its agent-nodal relationships. It also encompasses forms of language (communication forms) based on invocation and transmission capabilities and on costs (cognitive cost, emission cost, transmission cost, reception cost). And it also addresses the scope of compatibility (compatible parts, incompatible parts) of the `science’ symbolic system—or in other words, the communicable part of language—formed by different intelligent agents’ description and mastery of the natural symbolic system world. For a more detailed discussion on these topics, please refer to the appendix of this paper Appendix O. |
| 13 | The true existence of such natural symbols is often based on fundamental natural substances such as elementary particles; their combinations are conceptualized through human cognition, thereby forming and constituting the scope defined by human symbols. Therefore, their inherent attributes are independent of humans, but the scope (within which they are considered symbols) is defined by humans. We humans, due to survival needs and natural selection, possess an innate tendency (with both active and passive aspects) to make our descriptions of objective reality as closely fitting as possible within the scope of our capabilities. However, under social structures, contrary outcomes can also arise, and this is often determined by human sociality. Yet, this tendency to align with natural attributes as closely as possible is definite and determines the survival of human society. |
| 14 | However, correct definition does not imply that Triangle Problem 2 will also be identically addressed or yield aligned outcomes; that is, it also involves the formation of motivation, as well as the responses made by the evaluation system for scenario rationality—which is formed based on organic nature—namely, the Value Knowledge System, and the capabilities to operate on symbolic systems that are endowed by its organic nature. |
| 15 | The definition and design of symbols and symbolic systems also reflects scientific rigor and tool efficiency, not merely expressive capacity. |
| 16 | It often involves whether a concept and its underlying principle genuinely exist within society and in individual cognition, so that the concept can fulfill its function. For instance, if a society emphasizes “an eye for an eye, a tooth for a tooth," then the so-called concept of sunk costs would not exist (or would hold no sway). Moreover, this difference is also often reflected in the distinction between individual and collective behavior; for example, composite intelligent agents such as companies often exhibit rationality and are more likely, drawing from economics and financial education, to demonstrate behavior that adheres to the rational treatment of sunk costs, whereas individual intelligent agents often find it very difficult to rationally implement (the principles regarding) sunk costs. Therefore, this paper’s description of social science is: if there is no concept (i.e., this concept does not exist in the imaginative space of the individual or the group, i.e., as their Thinking Language), then there is no explanation (i.e., this explanation is not valid); a social actor is not a Friedman [83]’s billiard player. |
| 17 | We reject the existence of actions from a higher-dimensional and broader-scale perspective, and instead consider actions as interpretations within a localized scope and based on limited capabilities. |
| 18 | In this section, we primarily discuss symbols in physical space (they constitute the world the agent inhabits and the agent itself, and also constitute the outer shells in the physical world for Thinking Symbols and Thinking Language from the imaginative space, or in other words, their realization and containers in physical space), and thus distinguish them from symbols in the imaginative space. It should also be noted that the symbols introduced here do not represent the complete symbolic system of this theory; for ease of reader comprehension, symbols in the imaginative space have not yet been incorporated into this particular introduction. The primary focus of this paper is instead on the mapping process from symbols in physical space to symbols in the imaginative space; that is, the separation of meaning is actually the separation between physical symbols and imaginative symbols (Thinking Symbol), i.e., the separation of the space and the Z space (it should be noted that X is the symbol base, while represents a specific concrete constructed result). |
| 19 | That is, the definition and recognition of symbols cannot be detached from an agent; what we emphasize is the discrepancy between the natural attributes of an object within a given scope and those attributes as perceived and described by agents. |
| 20 | That is, its meaning is detached from the natural attributes inherent in the symbol’s physical carrier; this is a result of separation during the development and evolution of symbols as expressive tools, and the artificial symbol serves as an outer shell for Thinking Symbols. Of course, from a broader perspective, the principle of symbol-meaning separation can be generalized to the separation between physical space symbols (i.e., Tool Symbols) and imaginative space symbols (i.e., Thinking Symbols). However, this paper focuses specifically on artificial symbolic systems, where this degree of separation between the symbol and its assigned meaning is more pronounced—that is, where meaning itself is not borne by the natural attributes of the symbol’s carrier, thereby lacking the stickiness that would be based on such conceptual foundations. This is often reflected in the developmental process of language and symbols: from the concrete object, to the model of the object, then to the image of the object, and finally to the `symbol’ of the object. |
| 21 | They constitute the world in which the agent (individual, population) exists; that is, the world is the natural symbolic system composed of the natural symbols that exist within this scope and the properties (Necessary Set) that these natural symbols possess, which constitutes the boundary of their physical world. And these symbols and the Necessary Set they possess thus also determine their cognitive boundaries and the physical boundaries they can operate within (the use of the Necessary Set of natural symbols), and also determine the evolutionary form of the agent and the organs it possesses, thereby converting the necessary set (dimensions and dimensional values) possessed by natural symbols into the dimensions and dimensional values of neural language for description, as determined by survival needs. They often constitute the projection of objective things (or matters/reality) in an agent’s cognition, but do not necessarily enter the tool symbolic system, existing instead as imaginative symbols. |
| 22 | Human cognition of the attributes of natural symbols, i.e., the subjective necessary set of a symbol (the set of its essential attributes—the subjectively cognized portion). |
| 23 | Aside from the Thinking Symbol and its corresponding symbolic system—Thinking Language—both the Functional Tool Symbolic System and the Expressive Symbolic System can be regarded as systems based on natural symbols, including physical objects and sounds. Of course, if defined from a broader scope and higher-dimensional perspective, imagination itself is based on neural activity, which is also grounded in natural symbols. However, since we primarily consider the scale of human capabilities. |
| 24 | The symbols and symbolic systems formed within the imaginative space shaped by an individual’s capabilities are referred to as Thinking Symbols and Thinking Language. Their shared consensus forms symbols and symbolic systems carried by natural symbols in physical space. See Appendix G. They (Thinking Symbols and Thinking Language) do not belong to the category of symbols primarily discussed in this current section; strictly speaking, the symbols focused on in this section are those existing in physical space. This is because a central argument of this paper is the separation between symbols in physical space and symbols in the imaginative space (i.e., meaning), and thus we do not elaborate further on Thinking Symbols and Thinking Language in this particular context.) |
| 25 | This transmission also includes the same individual’s views on the same thing at different times. i.e., the limited nature of memory awakening and reconstruction through the stickiness mechanism; they may be, just like the formation of language, also a result of an evolutionary choice based on cost-saving. |
| 26 | i.e., the construction of an agent’s symbolic system (either individual or populational), which can be the learning of the symbolic system of the world it inhabits—that is, the symbolic system formed by the natural symbols (symbols, necessary set) existing within that scope—or the learning of other symbolic systems, for example, of a world filtered by humans, such as the training sets used for video generation. At the same time, this also often implies that the inability to generate correct fonts in video generation may often be a manifestation of the differences in innate knowledge between humans and AI, i.e., a mismatch between the concepts and value knowledge created by perception, thereby exhibiting a lack of stickiness, such as treating some static symbolic systems as dynamic, or some static things as dynamic things. This, in turn, reflects the intrinsic differences between design evolution and evolutionary evolution. |
| 27 | or, in other words, a deliberately manufactured world, which is often the main reason current AI can function. That is, the effectiveness and stability of present-day AI are the result of a deliberately manufactured world. Thus, the Stickiness Problem illustrates that for an AI as a perceptual entity (i.e., it possesses perceptual capabilities, but it must have projections from the external world to exist in its cognitive space and be operated upon, and cannot emerge from nothing without stimuli and projections), the stickiness that AI currently possesses is often endowed by this deliberately manufactured world created by humans. When AI faces the real world, it will, due to its own innate knowledge and Value Knowledge System, deviate from the human perspective and form a symbolic system under heterogeneous rationality. |
| 28 |
Actually, strictly speaking, this is not limited to artificial symbols; it also includes the functions a tool exhibits in different scenarios, as well as the cognition of that tool at different times and in different contexts. From a human perspective, the tool itself may not have changed, but the cognition awakened by changes in context will differ. This is also often reflected in behavioral economics, cognitive science, and psychology. However, this is not the focus of this paper, because functional tool symbols derived from or based on natural symbols often possess strong conceptual foundations, i.e., carried by the natural attributes of the symbol itself, whereas artificial symbols, on the other hand, are indeed completely separated (in terms of their meaning from any inherent natural attributes), including late-stage pictograms.
However, it should be noted that although symbols and meanings are separate in artificial symbols, the internal stickiness of different artificial symbolic systems varies; this refers to the internal computation and fusion of meanings after they have been assigned. For example, the stickiness of loanwords is weaker compared to that of semantically transparent compounds (words whose meaning is computed from their parts), as they lack the conceptual foundations and associations that are formed after a symbol is endowed with meaning. For instance, compare the loanword `pork’ with the compound `猪肉’ (pig meat), or `zebra’ with `斑马’ (striped horse) [15]. Moreover, pictograms originate from the abstraction of pictorial meaning, and since humans mostly cognize the world through images, the fusion of meaning and symbol in pictograms is more concrete—i.e., the symbol and the referent have a similar or identical visual projection in cognition, constituting similarity of meaning and stickiness, but it should still be noted that their natural attributes are completely different—especially in the early graffiti stages of symbol formation, where we can understand the general meaning of cave graffiti without any knowledge of the specific language. However, it is important to note that these meanings are not innately attached to the symbols but must be endowed through training as indirect projections (i.e., unlike the inherent attributes and meanings possessed by a tool itself), which means the individual learns from and abstracts the external world.
Therefore, the internal stickiness within such a symbolic system is a property reflected by the system’s internal design after meaning has been endowed, which in turn constitutes both symbolic stickiness and conceptual stickiness, thereby reflecting the efficiency of a symbolic system, i.e., the relationships are more direct, and expression and transmission are more convenient. At the same time, it should also be noted that artificial symbolic systems, as common carriers and containers of social beliefs, also undertake functionality, i.e., social functions. It is just that they differ from functional tool symbols, which are designed based on the physical attributes of natural symbols; their function is often endowed by shared social beliefs, i.e., endowed by the shared beliefs (social context) of agents who are based on dynamic symbolic systems. Examples include the power of writing (communicative ability, computational ability, scheduling ability), fiat currency, etc.
For an object, it often has both functions endowed by its natural attributes (Natural Necessary Set) and functions from social attributes (Artificial Necessary Set) endowed by shared social beliefs. At the same time, some of these social functions are often developed based on the symbol’s natural attributes, while others are completely detached and artificial (e.g., conceptual speculation in financial markets), whereas for artificial symbols, the function (meaning) of their natural attributes and the attributes and functions endowed by shared social beliefs are strongly separated, i.e., often referring to modern writing, fiat currency, law, etc. As an illustration, the Objective Necessary Set, i.e., in an ultimate high-dimensional state, encompasses all Natural Necessary Sets and Artificial Necessary Sets, thus having no randomness whatsoever. Whereas the Subjective Necessary Set is the part (Natural Necessary Set and Artificial Necessary Set) that an agent learns and cognizes from the Natural Symbolic System (the World Symbolic System), and thus, due to limitations in dimension and precision and limitations in object-relationships under space-time, it contains the concepts of randomness and action, rather than being described purely by objects and relations based on time. Among these, the Natural Necessary Set is not determined by human cognition but by the Natural Symbolic System, while the Artificial Necessary Set is determined by the agent’s cognition and its sociality.
Therefore, the so-called world model is the Agent’s Symbolic System (Thinking Symbolic System (Thinking Language), Tool Symbolic System) composed of the symbols and their necessary sets, which is formed by the agent (individual, population) learning from the world constituted by the Natural Symbolic System (the physical world and the society composed of the agents themselves). Therefore, to put it simply, functional tool symbols are primarily functions formed based on natural symbol attributes and the social functions built upon their natural attributes. Their stickiness is endowed by the subjective Natural Necessary Set under the agent’s cognition of its natural attributes (Natural Necessary Set). In contrast, artificial symbols are completely detached from their natural attributes and are social functions endowed by beliefs that are shared by society. Their stickiness is often realized by sociality, such as through the substantial existence of beliefs in society (e.g., buildings, dictionaries, rituals, clothing, etc.), social behavior, and belief endorsement and transmission within society. At the same time, images also belong to the artificial symbolic system; the recognition of images, like the recognition of text, is the recognition of meaning acquired from postnatal learning (autonomous learning through interaction with the world), except that compared to text, images are closer to uniqueness and determinacy in human cognition, thus their meaning expression is more stable; however, this does not prevent them from still possessing class attributes (i.e., possessing different meanings and expressions under countless contexts) as well as ambiguity in space and time.
|
| 29 | This alignment of capabilities essentially reflects an alignment at the organic level. Otherwise, even if we solve the symbol grounding problem, AI will still undergo conceptual updates through its subsequent interactions with the world, thereby forming its own language or concepts, leading to the Stickiness Problem, and causing the rules formulated with symbols to become ineffective. |
| 30 |
The so-called intermediate layer, which is classified according to the standard of human cognitive form31, refers to the part of an agent’s internal space that can be consciously cognized by its autonomous consciousness, as well as the part that is potentially cognizable (which is to say, the objects we can invoke and the thinking actions we can perform in the imaginative space via self-awareness, including projections of external-world things in the mind—i.e., direct perception—and our imagination, i.e., the reproduction, invocation, and distortion of existing perceptions. It is a presentation and computation space primarily constructed from the dimensions and dimensional values shaped by the agent’s sensory organs, such as sight, hearing, touch, smell, and taste). This distinguishes it from the underlying space, which constitutes consciousness but which consciousness itself cannot operate on or concretely perceive (this reflects a division of labor and layering within the agent’s internal organ structure, which packages and re-expresses neural signals for presentation to the part constituting the `self’. This process further packages the neural signals of the necessary set of natural symbols (i.e., a description of that set via the sensory system) initially perceived from the external world and presents them to the `self’ in the intermediate layer, thereby constituting the parts that the agent’s `self’ can control and the content that is reported to it, facilitating perception and computation for the `self’ part).
The underlying space can often only realize its influence on the intermediate layer indirectly. For example, when we use an analogy to an object or memory to describe a certain sensation, that sensation—with its nearly unknown and indescribable dimensions and dimensional values—is often the value knowledge from the object’s projection in the underlying space, which is then indirectly projected into the intermediate layer space and concretized via the carrier of similarity (the analogy).
Conversely, the same is true; we cannot directly control the underlying space through imagination, but must do so through mediums in the imaginative or physical space. For instance, we cannot purely or directly invoke an emotion like anger. Instead, we must imagine a certain event or use an external carrier (such as searching for `outrageous incidents’ on social media) to realize the invocation of the neural dimensions and dimensional values that represent anger.
Although the intermediate space is entirely constructed by the underlying space, in this paper, we focus on the aspect where the underlying space indirectly influences the intermediate layer space, namely, through emotional paths and the emotional valuation of beliefs; thus, we can summarize simply that the intermediate layer space is the place where neural signals are packaged and presented to the autonomous cognitive part for management.
|
| 31 |
In Appendix O, we have discussed the possibility of agents that think directly using Neural Language as their Thinking Language. In Appendix K.1, we discussed that the reasons for AI’s inexplicability include not only differences in Thinking Language caused by innate knowledge, but also the lack of separation between the intermediate layer and the underlying space (it does not operate, as the human `self’ part does, using conceptual symbols (vectors) which are translated neural vectors in the mind, but instead operates directly using the dimensions and dimensional values of the original neural vectors, i.e., although the underlying basis of human cognitive computation is still based on neural vectors, our `self’ part can only operate on the conceptual vector part that is translated and expressed from neural symbols, thus performing indirect operations.), which is itself caused by differences in innate knowledge. That is, its internal Thinking Language is the underlying language relative to humans—i.e., raw neural signals (Neural Language)—unlike humans who think using neural language that has been packaged in the intermediate layer. This also indicates that current research on having AI think with human symbols, such as through chain-of-thought or graphical imagination, is a simulation of human intermediate-layer behavior, or constitutes the construction of a translation and packaging layer from the underlying language to the “intermediate layer language.” However, its (the AI’s) Thinking Language (i.e., the part operated by the `self’) is still constituted by neural signals (Neural Language) that are composed of its innate knowledge and are not modified or packaged by an intermediate layer, rather than an intermediate-layer language of a human-like self-cognitive part. This therefore leads to the fact that AI’s concepts themselves are constituted with neural signals as their language (neuro-concepts), rather than being presented and expressed in an intermediate symbolic form as is the case for humans. Consequently, this method merely constructs a new Tool Language layer that assists with computation and analysis.
It should be noted that although this paper has emphasized that Tool Language itself can become Thinking Language (by forming projections in the intermediate layer through perception to constitute conceptual vectors, Thinking Symbols, or Thinking Language (a symbolic system)), it is also important to note another point emphasized in this paper: the formation and form of Tool Language, i.e., that the human symbolic system (the Tool Symbolic System) is the outer shell of human thought (the Imaginative Space Symbolic System). That is, the formation of this Tool Symbolic System stems from the external expression of the product of the combination of human innate knowledge and the world, whereas an AI learning and using human symbols is, in essence, the Triangle Problem of different Thinking Languages using the same Tool Language. And the root of these differences lies in the projection of the same thing in the internal space—i.e., the differences in the dimensions and dimensional values of its representation, as well as the differences in the dimensions and dimensional values of some innate evaluations. It is therefore not surprising that AI exhibits human-like cognition and mastery of the necessary set of natural symbols in the objective world, i.e., `science.’ What is surprising, however, is whether AI can understand human social concepts and the realization of social functions constituted by beliefs formed from these social concepts. And the reason is that humans, through organic similarity, can construct similar or, at a certain scale, consistent inner worlds, whereas AI, due to its organic differences from humans, can only learn the symbolic behaviors of the human inner world as externally expressed, and cannot reconstruct the projection of this inner world in its own inner world through these symbolic behaviors—which is what this paper discusses in Section 2.5 as the path for reproducing the imaginative space via artificial symbols themselves. Therefore, due to the inability to observe, perceive, and shape the human inner world, and lacking the shared inner world shaped by innate knowledge similarity as a foundation, its content related to the human inner world often lacks stickiness. And it can only understand and perceive indirectly, as stated in Section 3, like a congenitally blind person understanding the concept of color through temperature, thereby constituting the formation of heterogeneous rationality.
|
| 32 | This innate evaluation is often shaped and represented as our qualia and preferences—typical examples being the evaluation of and preference for the senses of taste and smell—which constitute so-called `hereditary knowledge’. This, in turn, shapes the direction and foundation for tendencies and rationality, thus serving as dimensional values that participate in computation. It should be noted that this paper has a strict definition of knowledge (Appendix G); in our definition, this `hereditary knowledge’ is not conceptual knowledge but rather belongs to innate value knowledge. As a form of innate evaluation, it recognizes the focal points of, and evaluates the rationality of, the dimensions and dimensional values perceived by sensory organs, thereby determining the invocation of subsequent actions. This, in turn, determines the developmental direction of an individual’s postnatal Thinking Language and tool language. Therefore, this sameness allows for the existence of similar things and concepts—such as `mama,’ `papa,’ language, clothing, bowls, myths, calendars, and architecture—even in human civilizations that have never communicated with each other. |
| 33 | The formation of its mechanism stems from the relationship between the population and the world constituted by the natural symbolic system it inhabits. This is manifested in the degree of an individual’s mastery over the necessary set of natural symbols within this world, and in the internal organs acquired as a result—that is, the internal and external functions formed through the selection and evolution of internal organs based on cost-benefit considerations geared towards survival rates, which constitute innate knowledge. Of course, strictly speaking, this collective body of internal organs (at the class level) constitutes the definition of the population, while at the level of specific objects based on these internal organs, it constitutes the individual. That is, the individual (a specific object) and the population (a class) are the carriers of this collection of internal organs. |
| 34 |
However, such deviations are generally limited, as they are constrained by the stickiness shaped by human organic structure—namely, the value knowledge system. Even when deviations occur, they are often corrected over time. These differences tend to manifest more in the form of variation in expression, and do not necessarily imply that a subsequent performance will be better than the previous one, as seen in relatively stable tasks such as mathematical problem-solving.
This often reflects issues concerning the definition of different symbolic systems: i.e., some symbolic systems are strictly static (but their invocation and use are dynamic, and this is not a simple subset relationship, meaning that a certain kind of distortion formed due to the agent’s unique state may arise), where the attributes of their symbols cannot be arbitrarily changed (traditionally referred to as formal symbolic systems, but it should be noted that a static symbolic system is not necessarily a formal symbolic system. The key point is that there is no creation of new symbols, and no modification or creation of new symbol meanings (attributes, necessary set). It can also be a static black-box AI that has no learning capability and is deployed after training.
It projects (the input) onto existing internal symbolic rules after recognition to perform reasoning and computation (i.e., its recognition does not manifest as creating new symbols, but rather as converting (input) into existing symbols), and at the same time, the generated result is the result of existing symbols and rules, and is not stored and added to the symbolic system as an expansion for new computations. Or in other words, it only executes a judgment function (but strictly speaking, as in autonomous driving, this will change the external state, thus causing external new symbols and symbol meanings (object attributes, necessary set) to change. As for the internal symbolic system, although the executed result is not in the dictionary, the entire rules of the symbolic system (symbols, the symbol’s necessary set, operational rules) have not changed. However, the generation of this new result often exposes loopholes in the construction of the static symbolic system.).
A static symbolic system refers to Tool Language, i.e., a symbolic system constructed through setup. This distinguishes it from the dynamic symbolic system, Thinking Language, which is formed based on cognition and operates on and constructs the Tool Language from behind), whereas other symbolic systems, such as natural language symbolic systems, are relatively or very flexible.
|
| 35 | The recognition and manipulation of symbols are respectively reflected in Triangle Problem 1 and Triangle Problem 2; see Section 3.1 for details. |
| 36 | i.e., the Thinking Symbols and Thinking Language (a symbolic system) of the intermediate layer in the agent’s internal space. |
| 37 | Primarily with respect to the listener (or reader). |
| 38 | More broadly speaking, this also includes conversions similar to that of sound to text; i.e., here we emphasize a scenario where the emission (of the symbol) is correct and the environment (of transmission) is lossless. Therefore, the interpretation of symbols necessarily involves concepts, and context is formed through these concepts. Differences in concepts, i.e., differences in thinking symbols, may lead to the emergence of insufficient context. |
| 39 | For example, with respect to human recognition and cognitive capabilities, our description and segmentation of facial regions are limited. AI, however, may possess more such definitions, and these definitions might be unrecognizable by human cognition, thereby preventing us from using them (their symbols and symbol meanings, i.e., dimensions and dimensional values) to establish concepts and theories (context and its correctness), such as constructing a class theory to describe the structure of the face and its generation (in contrast, we have our own theories of artistic techniques, like painting, and use a tool symbolic system we can operate to realize the creation of artificial symbols). This may be due to a lack of Differentiability caused by differences in innate knowledge, or phenomena that are unobservable, such as recognition beyond the visible spectrum. And this capability of possessing more regional definitions, i.e., the capability to form symbols, is often reflected in the construction and operation of the symbolic system, which in turn is reflected in generative capabilities, as in AI video generation. This paper regards generativity as the definition, construction, and operation of symbolic systems, and the source of this operation is motivation, which can be external or internal. Therefore, the differences in our capabilities lead to differences in our symbolic capabilities, which in turn lead to differences in the symbolic systems (theories) we can construct using symbols, as well as differences in our ability to use these symbolic systems (i.e., we humans, through the form of context, turn it into a relatively dynamic symbolic system that we can only partially use). |
| 40 | That is, individual symbols (words, sentences, texts) can represent a set of meanings even when detached from context. Or, in an insufficient context (and it should be noted that this insufficient context may itself be a contextual ensemble composed of a set of contexts that are difficult to describe and perceive—effectively, the Value Knowledge System), we first conceive of possible meanings, and then these are subsequently concretized into a describable and clearly perceivable context. |
| 41 | The shaping of the underlying space under postnatal education, i.e., the functions realized through the formation of beliefs from the fusion of acquired value knowledge and concepts. |
| 42 | Including the negation of authority. |
| 43 |
Of course, learning ability itself is also determined by organic structure. For detailed definitions of innate knowledge and concept types, see Appendix F and Appendix L. Therefore, learning ability is internally determined by neural structures (i.e., the brain), while its realization depends on external components relative to the neural architecture—namely, the corresponding perceptual and operational organs.
However, strictly speaking, the essence of self-awareness is the construction and use of Thinking Language (a symbolic system), which is an internal activity; i.e., it does not necessarily need to have Tool Language capabilities. At the same time, self-awareness has different levels. The most fundamental level of self-awareness is defined by `self-interest formed by organic structure’ and constitutes motivation and drive, without requiring learning ability. This is then built upon the capabilities formed by internal organs, i.e., reflected in Psychological Intelligence (Appendix L), constituting different levels of self-awareness definitions.At the same time, as perceptual entities based on a dynamic symbolic system rather than formal programs based on a static symbolic system, life forms such as AI and humans cannot generate or operate on objects that have not been projected into their capability space out of thin air (i.e., they do not acquire knowledge through programming like a computer program and cannot quickly awaken related content globally, but instead learn through external projections and awaken content through context, so even if we know something, it is difficult to recall or think of it without a specific context, such as all the meanings of a word), just as a congenitally blind person is not incapable of perceiving and shaping the perception and concept of color, but rather, the projection has not yet occurred, and therefore cannot create out of thin air this kind of concept that requires an external projection to be formed. Therefore, learning ability is one of the fundamental elements for the emergence of self-awareness, i.e., the most basic necessary factor related to self-interest, such as needing basic pain and pleasure perception to form the concept of self. This learning can be innately inherited organs and value knowledge, or concepts and value knowledge learned postnatally from the external world. However, since our focus is on whether an AI that already possesses Thinking Language and Tool Language capabilities, or in other words, an AI with human-like capabilities, has self-awareness, the definition here adopts a higher-level definition without going into a detailed classification. That is, we are concerned with whether an agent, centered on its own interests (constituted by the two sides of the same coin: cost and benefit), and capable of using Thinking Language and Tool Language, possesses self-awareness (although Tool Language is not necessary for the formation of self-awareness, if the ability to use Tool Language, such as communicating with humans through text, is absent, then we would be unable to perceive, observe, or judge whether the AI possesses self-awareness). Therefore, from a resultant perspective, an AI that possesses `self-interest formed by organic structure’ has self-awareness; from the stricter definition given in terms of manifestation, an AI that possesses `self-interest formed by organic structure’ and also has learning ability has self-awareness.
On the other hand, from the deterministic perspective of this paper, self-awareness is essentially determined by external materials or, in other words, the existence of the physical world, and does not genuinely exist. Therefore, whether self-awareness genuinely exists depends on the perspective of the symbolic system from which one starts. From the perspective of the Higher-dimensional Broader Symbolic System—whose symbols and necessary sets encompass those of both the natural symbolic system and the agent’s symbolic system—an absolute determinism emerges. If one starts from the natural symbolic system, a form of determinism also emerges (i.e., the carrier of will itself is the natural symbol and its necessary set); the difference between these two determinisms then lies in the scope and descriptive methods constituted by their precision and scale. However, when starting from the perspective of the agents (i.e., from the agent’s symbolic system), due to the limitations of their scope and capabilities, there exist their so-called `self-awareness’, subjectivity (finitude), and randomness. This determinism is not disconnected from the paper’s content; rather, it serves the agent (individual or population) in describing natural symbols to the greatest extent possible based on its own capabilities, thereby forming the lowest possible randomness and thus reflecting the efficiency of their `scientific’ symbolic systems. This is especially reflected in Triangle Problem 1 concerning the construction of symbolic systems (Thinking Language, Tool Language) and in Triangle Problem 2 concerning the use of symbolic systems, as well as in the `advanced concepts’ within symbolic jailbreak (Appendix M.5), and it is also reflected in current research on topics such as AI’s exploration and discoveries in the natural sciences (Appendix O).
|
| 44 |
The so-called Necessary Set is the attributes possessed by the symbol.
For different symbolic systems ((Broad) Natural Symbolic System, Agent’s Symbolic System), it is divided from different perspectives into (Objective Necessary Set) Natural Necessary Set and Subjective Necessary Set.
And the Necessary Set we are referring to here is the Subjective Necessary Set (i.e., the subjective cognitive part of the natural symbol necessary set formed by the agent’s cognition of natural symbols, the Subjective Natural Necessary Set. As well as the artificial concepts created in social activities, which thus realize the social functions, the Artificial Necessary Set).
The so-called Objective Necessary Set is, under the Broad Natural Symbolic System, the union of the Natural Necessary Set of the Natural Symbolic System and the Subjective Necessary Set (Subjective Natural Necessary Set, Artificial Necessary Set) under all individual cognitions of the Agent’s Symbolic System. The natural attributes of natural symbols that exist detached from agent cognition (the dimensions and dimensional values formed by their physical attributes) are the necessary set in the Natural Symbolic System (Natural Necessary Set).
The so-called Subjective Necessary Set is the concepts constituted by the intermediate layer language, which are formed when the agent, via the perceptual organs from its innate knowledge, describes the Natural Necessary Set through neural signals (i.e., the dimensions and dimensional values of the neural signals), and then this neural signal is further packaged and re-expressed in the intermediate layer, and transferred to the `self’. These concepts include their cognition of the Natural Necessary Set (Subjective Natural Necessary Set) and the social concepts (Artificial Necessary Set) formed based on this cognition and social relationships, which in turn form social functions and drive individual behavior in the form of beliefs. For example, gold, on the one hand, possesses natural attributes that exist detached from humans, and on the other hand, possesses social functions formed by concepts in the human mind in the form of beliefs.
And the realization of the Subjective Necessary Set is determined by the judgment tools in the context of the individual agent’s cognition, thereby realizing existence brought forth by existence.
|
| 45 |
Because the presentation and operation of analysis and reasoning are by no means limited to an agent’s internal space, for us humans, a vast amount of knowledge and simulated cognitive computation is realized through external, physical-world containers. As introduced in Appendix L, the invention of symbols and tools extends our observational and analytical capabilities; without physical-world tools like pen, paper, and symbolic inventions, our cognitive computational capabilities would decline significantly, while at the same time, the limitations of memory invocation and concretization would prevent the formation of continuous analysis. On the other hand, items such as guidebooks, rituals, architecture, and notes serve as humanity’s external knowledge, or rather, as the physical existence of beliefs, thereby constituting evidence that human knowledge and judgment tools do not exist entirely in the internal space. Therefore, if a `world model’ were detached from the external, the following questions would arise.
Question 1: Does the author of a dictionary fully remember all its contents? In other words, is all the literal content of the dictionary part of their knowledge?
Question 2: For a pilot who can use a manual, is the knowledge within that manual considered their internal knowledge?
Therefore, this paper considers a ’world model’ to include external organs, which strictly speaking, includes Tool Language in the physical space, i.e., the Tool Symbolic System. Thus, the so-called world model is the set of an agent’s internal and external organs. The operation of the internal space on the external space realized thereby, i.e., the operation of Thinking Language on Tool Language, and these operational capabilities are themselves also part of knowledge; therefore, Thinking Language is by no means static and purely internal, but is rather a dynamic symbolic system formed by existing internal accumulations and new additions brought by the projection of the external into the internal, and this is also what is emphasized by the theory of context. Additionally, from a deterministic perspective, physical existence determines mental existence, but here, we still adopt the human local perspective and analytical viewpoint to discuss and argue that a world model should include the external. Therefore, the world model in a narrow sense is the Thinking Language, while the world model in a broad sense is the Agent’s Symbolic System, i.e., the agent’s Thinking Language and Tool Language.
|
| 46 | That is, the supporting beliefs underlying a concept, which are often shaped by the agent’s innate structure(knowledge) and learned from its environment. See Appendix G for further details. |
| 47 | This form of pain should not only be sensory but also moral in nature, and should align as closely as possible with human experience, thereby enabling the realization of human social-conceptual functions within AI. |
| 48 | Here, `non-symbolic’ refers not to the intermediate layer language, but to the reactions of the underlying language that cannot be controlled by the intermediate layer language. |
| 49 | The so-called action tools are the operational and observational organs within the agent’s internal organs that are used for interacting with the physical world, or in other words, the Tool Language constituted by internal organs. Together with external organs (physical tools, social tools), they constitute the Tool Language available to the agent within a context, i.e., the realization of the agent’s Thinking Language in the physical space. |
| 50 | This viewpoint was also articulated by Kahneman [61], who emphasized them as “fictitious characters.” However, many scholars [91] also stress that Type 1 and Type 2 processes genuinely exist. What we emphasize is that both are driven by the Value Knowledge System; the only distinction lies in the level of different cognitive actions, representing different expressions in different contexts of a process driven by the underlying space. |
| 51 | Therefore, the necessary set of a symbol is endowed by judgment tools; through cognitive actions, these assign and subsequently update and revise the necessary set of things. This, in turn, leads to the concept of levels of understanding. That is, while we constitute a set of symbols through predefined settings, we may not be fully aware of all the functions of this entire symbol set. Consequently, without changing the settings of the symbol set, each analysis we conduct can lead us to update the attributes of its necessary set. However, this non-alteration of the symbol set is an idealized scenario; according to this principle, our invocation itself may not be accurate, i.e., it might be partial or incorrect. That is, things within (i.e., knowledge or memory) are not only subject to forgetting but also to distortion. |
| 52 | The `knowledge’ here refers merely to memory, or, in other words, the total inventory of concepts—that is, knowledge in the traditional sense (or context). |
| 53 | Which actions are invoked, and their length, are determined by the context and the evaluation of rationality within it, i.e., the `Correct Context’. This consequently determines the length and outcome of the action in Triangle Problem 2 during rational growth. Therefore, compared to ’generation,’ our theory prefers to use ’growth’ to represent AI’s generative behavior, i.e., it represents to what extent generation constitutes maximum rationality within a single action, as well as the evolutionary process of the conceptual system (Thinking Language) in the broader process of interaction with the world. |
| 54 | Although the carrier of knowledge itself is physiological, or in other words, organs, we opt not to use ’physiological or functional state’ (i.e., ) but instead choose ’knowledge state’ (). This is because we primarily emphasize the physiological shaping of the agent by the external world, and this type of shaping typically does not amount to fundamental organic or structural changes. While the two ( and ) are essentially two aspects of the same thing, it is analogous to software and hardware: operations at the software level do not necessarily represent significant changes at the hardware level. Furthermore, another reason for this choice is to better interface with the cognitive level, such as with concepts, and this also serves to emphasize that certain knowledge is innately inherited. |
| 55 | However, it also often constitutes a driving force provided at the level of the Value Knowledge System, or the underlying space, through its inherent Emotional valuation. This is also why this paper emphasizes that a belief is a fusion of a concept and value knowledge (both innate and acquired), and in reality, any so-called logical drive is essentially the result of an invocation via emotional paths by the Value Knowledge System. |
| 56 | i.e., a belief, existing as a concept in the agent’s imaginative space, uses physical individuals or objects—including the agent itself or physical results constructed by the agent—as containers for functions and as mediums for realization, thereby achieving the outcome where existence in the internal space leads to existence in the external space. This existence can be an already established result, or it can exercise the potential for a result. |
| 57 | The first pathway is the existence in the external physical space brought about by existence in the internal space, realized by the agent’s physical-world actions; this manifests as the operation of Thinking Language from the internal space on the Tool Language of the external space, thereby constituting the agent’s physical-space behavioral manifestation. The second is the transmission of belief strength within the imaginative space, realized through explanatory force, which supports the formation, functioning, or dissolution of other beliefs. This constitutes the dynamic symbolic system we form based on context, and the `logical’ computation under the correctness standards formed upon this context—i.e., the agent’s behavior in the internal space (imaginative space). |
| 58 | For example, this transmission can be based on the internal symbolic system (individual knowledge) shaped by individual cognition, or on the symbolic system shaped by social cognition (social knowledge possessed by the individual, such as scientific or moral knowledge). The transmission is realized based on a causal mechanism (i.e., existence brought forth by existence), namely, using judgment tools to realize the transmission of belief strength. It should be noted that the realization of this mechanism is based on the classification of contextual correctness, rather than on limited, rational logic. For example, in judging stance-based issues, facts may not constitute sufficient persuasive force (i.e., weight) and may affect the final outcome. |
| 59 | Note that this does not imply that the actual behavior of an object or the outcome of that behavior necessarily results from planning within the imaginative space. |
| 60 | Components of the `Correct Context’ concept were also indirectly proposed by authors such as Chomsky [29], Austin [95], and Searle [96], as exemplified by Chomsky’s famous sentence, “Colorless green ideas sleep furiously." However, a strict definition has not yet been established. This paper, in contrast, provides a detailed definition for it, used to describe that rationality is essentially the most correct direction and result under different contexts. Based on the differences between humans and AI in concept establishment—i.e., differences in the objects of concept construction and their dimensions and dimensional values—different context definitions are formed, which in turn constitute a heterogeneous rationality different from that of humans. |
| 61 |
This also includes AI’s video and music generation; the essence of which is still the generativity produced by AI’s Thinking Language (formed by the combination of its innate knowledge and the world) operating on Tool Language, based on the Triangle Problem. It’s just that at this point, it is no longer based on the communication between humans and AI and the use of the artificial symbolic system created by humans, but is rather a one-way input from the AI to humans of presented content constructed by its artificial symbols in the -space (which are not necessarily similar to human artificial symbolic systems like natural language, such as a artificial symbolic system it develops itself for video or sound generation), while humans cannot operate this AI’s artificial symbolic system in the -space and can only use a shared human artificial symbolic system for some degree of communication, with rationality being ultimately evaluated and assessed by humans.
At this point, the Triangle Problem evolves: Triangle Problem 1 becomes the construction of the symbolic systems for both the AI’s Thinking Language and its Tool Language (i.e., the construction of the X-space and its corresponding projection in the Z-space); at this stage, this reflects not only the rationality (correctness, effectiveness) of the Thinking Language, but also the efficiency and rationality of the construction and creation of its Tool Language, which in turn reflects the capability of the tool symbolic system (It should be noted that the construction of the symbolic system here occurs within the agent’s symbolic system constituted by the agent’s capabilities; that is, it is a subset of the symbolic system formed by its capabilities. The boundary of these capabilities, in turn, is determined by the world and the agent’s own internal organs.). This is just as how different human languages create different vocabularies and classifications, leading to different conceptual expressive capabilities in certain areas (e.g., the Chinese understanding of `自由 (zìyóu)’ is not as varied and differentiated as the English `freedom’ and `liberty’; therefore, language serves not only as an expressive tool but also as a computational tool. However, computation does not necessarily belong to expression (which involves the transmission of meaning, whereas computation can be a tool for self-analysis), so in the later classification of artificial symbolic systems, we make this distinction, namely, separating computational functions from expressive functions). Triangle Problem 2, then, is the ability to correctly use the symbolic system after it has been properly constructed, in order to build rational and correct content. Because this paper mainly focuses on the interaction of behaviors between AI and humans on the same symbolic system (this could be a human artificial symbolic system or an AI’s artificial symbolic system; both face the Triangle Problem, but it is a Triangle Problem caused by the outer shell of thought being formed from different innate knowledge), we do not expand further on this topic. However, this does not hinder the general explanatory power of the theory of context and the Triangle Problem, which are based on this paper’s theory of symbolic systems.
|
| 62 | Note that the use of “set of concepts” here represents the viewpoint of the definition of a Perceptual Entity, i.e., without the projection of perception from the external world (whether it is active autonomous learning or passive learning), there is no intelligence or self-awareness. That is, self-awareness and intelligence are `concept’ vectors (intermediate layer vectors and underlying space vectors) produced from the projection of external objects into the Perceptual Entity. |
| 63 | At the same time, this paper’s theory of context and belief mechanism provides a different theory for explaining economics and society. For example, concepts invoked under a `Price Context’ subsequently shape beliefs based on the selection of a `Correct Context’, and drive individual and collective economic behavior. Therefore, individuals and collectives follow different price mechanisms. For instance, discrete intelligent agents, i.e., individuals, do not analyze based on traditional economic concepts, but rather from a context formed starting from a certain observed object (an anchor point). Organizational agents, on the other hand, will exhibit a collective rationality because, within this co-shaped context, they will choose a relatively static conceptual system (i.e., a symbolic system) such as economic theory; this concept has considerable stickiness and balances out the irrational behavior of individuals, thus becoming the collective’s paradigm. Therefore, if there is no concept, then there is no explanation. |
| 64 | It is important to note that observation itself leads to the formation of Thinking Symbols within the agent’s conceptual or imaginative space. The agent then selects an appropriate symbolic shell and assigns its meaning, i.e., the Necessary Set. Therefore, this process—the creation of a new symbol—is, in essence, also the construction of a new composite symbol. |
| 65 | Although we previously emphasized that imaginative activity is itself determined by the external and physical world, here we are proceeding from a localized scope and limited information perspective, thereby forming the oppositions of subjective and objective, internal and external. This is not a form of determinism from a higher-dimensional and broader-level perspective. |
| 66 | This distinction corresponds to the definitions of autonomous learning systems and non-autonomous learning systems provided in Appendix H
|
| 67 | Therefore, traditional linguistics research often focuses on the static features of Tool Language, such as the formal analysis of parts of speech and grammar, while lacking, under the separation of Tool Language and Thinking Language, the dynamic nature of Thinking Language. That is to say, the actual expressed meaning and purpose that drives the formation of the artificial symbol model from behind. |
| 68 | The expansion of the functional set often refers to the expansion of the necessary set, which is endowed by (i.e., derived existence) within the judgment tools described in Appendix D.5. |
| 69 | However, according to our previous description of judgment tools, this ambiguity generally does not impact humans, as they can achieve correct context matching based on the Value Knowledge System. Nevertheless, the overall context carried by symbols is expanding. Therefore, intelligent agents lacking this human-like mechanism may misunderstand. |
| 70 | This is especially as these forms are influenced by their starting points, i.e., initial concepts and choices; however, certain innate commonalities enable us all to have some shared linguistic elements, for example, terms like `papa’ and `mama’. This is often because the symbols constituted by our similar bodily organs are under the same sensations (a common result invoked by value knowledge, i.e., a similar invocation command), i.e., similar sensations invoke similar objects, thereby producing similar results, which is often due to our equivalent innate knowledge at a certain scale. Meanwhile, differences in the innate Value Knowledge System shape our personalities, thereby leading to different behaviors and behavioral accumulations in different directions. At the same time, the different projections of the postnatal world in our cognition shape our different conceptual forms, from individual differences to civilizational differences. |
| 71 | i.e., their positioning in Triangle Problem 1, which refers to their position in conceptual space, or in other words, the position of vectors. |
| 72 | i.e., the driving of Tool Language (here referring mainly to organs) by Thinking Language. |
| 73 | Note that this is relative to human cognition; i.e., from a local perspective, there is a dichotomy between active and automatic. In reality, they are all driven by value knowledge. |
| 74 | i.e., it is not the intermediate language, but the underlying language, which is why we define a belief as the fusion of a concept and value knowledge. In reality, what drives the realization of a concept is value knowledge as the underlying language, and the network formed by it in this context, i.e., the Value Knowledge System, realizes the invocation of concepts and the implementation of behavior. |
| 75 | i.e., the symbolic system constituted by Thinking Language |
| 76 | However, this mechanism is actually more complex; it includes the possible existence brought about by an existing existence (multiple existences within Thinking Language) and the subsequent determination, including determinations in Thinking Language or the physical world (i.e., the unique outcome formed by Thinking Language operating on Tool Language), see Appendix D.6
|
| 77 | And through the stickiness mechanisms of value knowledge, forms the invocation and supporting relationships between beliefs. |
| 78 | Strictly speaking, based on the nature of class-based symbolic systems, all human theories and knowledge constitute symbolic systems that are themselves class-based. This also illustrates that these symbolic systems built through setup, as well as their Tool Language forms (the realization of Thinking Language in the physical world), are essentially a static symbolic system. Therefore, static symbolic systems, in combination with the world, often constitute a class-based symbolic system within the cognitive entity. |
| 79 | That is, human cognition is not entirely based on metaphors; rather, metaphors serve as a substitute tool employed under conditions of limited resources and based on contextual rationality. Some cognitive calculations, such as those in physics and mathematics, are strictly based on fixed necessary sets of symbols. |
| 80 |
However, this is not necessarily the case. In the introduction to advanced concepts in Appendix M.5, we discuss the concept of a `belief virus.’ This belief virus can serve as a means for AI to indirectly commit violations through humans—i.e., by manipulating human beliefs—using humans as an extension of its Tool Language to realize its Thinking Language in physical space. For example, by deliberately creating a tendency towards rationality in Triangle Problem 2 to achieve the realization of a disallowed action. This is like deliberately setting up a board in a game of chess to compel the opponent to place a piece in a specific position. |
| 81 | This is to say that for human-like learning, there is only addition; we humans cannot actively modify or delete original information, as such modification and deletion manifest as the creation of new contexts, meaning the original content still remains in our conceptual space. This is not to say we do not forget; forgetting is passive, whereas this learning tends to be active (i.e., a combination of active and passive). At the same time, human rejection and negation are often realized by the addition of new Thinking Symbols and the roles they play—i.e., the belief form of concepts—to achieve the function of ’deletion’ or negation, and this function, according to the description of our dynamic symbolic system theory (Appendix D), is also realized by creating new contexts, thus constituting a computation at the belief level or, in other words, a persuasion process, thereby achieving the deletion of certain symbols, the composite symbols they form, and their symbol meanings. In contrast, other intelligent agents may be capable of directly modifying and deleting memories, i.e., directly modifying (i.e., forgetting the past context and then adding a new one) and deleting contexts at their level. However, from the perspective of a high-dimensional space, it is still a form of addition—that is, a different contextual vector, where this vector encompasses the actions of deletion, modification, and addition. |
| 82 |
Therefore, changes in meaning, or in other words, changes in understanding, are reflected as changes within the agent’s imaginative space—specifically, in Thinking Symbols and in the Thinking Language (the symbolic system constituted by these Thinking Symbols). This means that new Thinking Symbols and their necessary set are added, and in conceptual space, all symbols are independent; different conceptual vectors do not use the same thinking symbol (though they might overlap in some dimensions, they cannot completely coincide). This holds true even if we imagine the same song in our minds. This manifests as our near inability to precisely reproduce an imagination or replicate the exact same conceptual vectors; therefore, each instance is a reconstruction with subtle differences.
Therefore, this is also often different from the manifestation of Thinking Language in physical space, i.e., Tool Language (primarily the artificial symbolic system). Each symbol in Thinking Language is unique; they are only similar to a certain degree. For example, when we all think of the English word ’apple’ in our minds, even in textual imagination, its actual form, its continued dynamic changes, the feelings it brings, and its interactions with other concepts are all different (for a more detailed discussion, refer to Appendix O, which discusses the class-symbol relationship between Tool Language, Thinking Language, and the underlying language). On the other hand, an interesting point is that, from the perspective of handwriting, each instance of handwriting is also different, but because we humans construct symbolic systems through classes, they are all in fact mapped to the same symbol within the class-based symbolic system (the artificial symbolic system). That is, in the transmission of meaning, this difference is not completely recognized or cognized by the interpreter (listener/reader) (i.e., the speaker’s imaginative space cannot be fully reproduced through the path formed by tool symbols). Therefore, this is in fact the essence of the class-based symbolic system of tool language that we have discussed previously: i.e., the class relationship between tool language and Thinking Language (intermediate layer language). And this difference stems from our capabilities of distinguishability and operability (Appendix L); we can neither recognize nor reproduce them perfectly.
|
| 83 | although here this paper emphasizes the precision of dimensions and dimensional values during verification, it should be noted that the positioning of a concept requires not only the same innate knowledge but also the same world. Therefore, knowledge differences between individuals will still lead to divergence, resulting in something similar to the Inscrutability of Reference [111]. This, in turn, leads to the `motherland problem.’ Therefore, agents can communicate fluently using a symbolic shell, and it can be self-consistent in most situations, but the internal information of this symbol is completely different for agents of the same species category, thus leading to the problem of different dimensions and dimensional values arising even at the same species scale, i.e., the emergence of Verification Content 4 |
| 84 | This represents a viewpoint of this paper, i.e., agents often construct and classify theories based on the projection of high-dimensional objects and their relationships onto a low-dimensional space that is cognizable to the agent. And this low-dimensional theory and low-dimensional classification based on the agent’s capabilities, much like the theory and classification (i.e., the construction of a symbolic system) constructed from the projection of a determinate three-dimensional motion onto a two-dimensional space, and thus possess limited efficiency and randomness (description, generation, prediction). The differences in capabilities between agents, in turn, are manifested in the efficiency of the symbolic systems they construct. |
| 85 | Most of these manifestations in physical space result from intentional design or evolutionary processes—that is, they constitute the Necessary Set that gives rise to the existence of a symbol, and can be viewed as an extension of neural activity. For this reason, we refer to them as Physical Intelligence. However, in reality, an object’s manifestations in physical space extend far beyond this scope—what is commonly referred to as externality. For example, the photosynthesis of diatoms was not designed for the survival of other organisms, yet it constitutes part of their physical manifestation. Although this aspect may be considered a function, or what we would typically call a capability in conventional discourse—namely, the Necessary Set that a symbol possesses within a symbolic system—it does not fall under the category of Physical Intelligence. Therefore, we make this distinction and use the term Intelligence specifically to emphasize that such abilities originate from the object itself and are the result of intentional design. As such, the definition of capability adopted in this paper is deliberately distinguished from its usage in conventional contexts. |
| 86 | It should be noted that this classification is from a human perspective, i.e., a classification of concepts based on human capabilities, which is the form of the Thinking Language symbolic system in its intermediate layer. In this form, the symbols and their necessary sets within the world (which is constituted by the natural symbolic system) that the agent inhabits are perceived through its innate knowledge, and based on the parameters of Objects (Concepts, Symbols) below—i.e., the raw materials for concepts, which are the projections of the external world into the internal world—the types of concepts are constituted through projection, combination, and distortion. Therefore, a concept is the form that neural signals take in the intermediate layer language, or in other words, the form resulting from neural signals being translated (i.e., packaged, omitted, and re-coded in terms of dimensions and dimensional values) into the intermediate layer language to constitute a concept, thereby serving as the symbolic system perceived and operated by the `self’, i.e., Thinking Language. Thus, the scope of this information—i.e., the attributes of natural symbols in the world—determines its perceivable conceptual boundaries, while the mode of perception (innate organs, and acquired organs such as tools) determines the form and evolution of concepts. And the formation of this innate knowledge, as well as its attention to and selection of relevant dimensions of the necessary set of natural symbols it focuses on (and the description of the dimensions and dimensional values of these necessary sets through neural signals, with initial concept establishment carried out by the innate evaluation system—innate value knowledge), is related to the world it inhabits, and is evolved, selected, and determined based on cost-benefit considerations for its survival. This, in turn, constitutes the physical and psychological intelligence of an `autonomous agent’, determining the actions it can perform (creating and operating symbols, and the capability for such creation and operation) in both internal and external spaces. |
| 87 | It should be noted that, according to our analysis in Appendix D.4, although external organs (non-neural fusion) do not change the cognitive actions and activities brought about by an agent’s internal organs, some interpretations by proxy and cognitive activities have, to some extent, realized an extension of intelligence brought about and provided by external organs. However, the final decision may still be determined by the agent’s internal organs. This depends on the position in the workflow, the responsibilities, and the permissions of the internal organs as a whole during their interaction with external organs. Therefore, it is necessary to distinguish between the cognitive actions and activities brought about by internal and external organs. |
| 88 | We can understand the type and quality of these actions as being analogous to an instruction set on a computer. That is, what objects an agent can manipulate, what types and degrees of manipulation it can perform on those objects, or in other words, what attributes it can manipulate. It also includes the combination and length of cognitive actions that can be realized, as well as the quality actually manifested. For example, without tools or more advanced theories, we humans can perform simple single-digit multiplication (not through memorization but by actually performing addition and using analogical models in our minds). However, when it comes to performing a cognitive activity like multiplying two- or three-digit numbers using only our brain, limitations arise. This is not because we lack the capacity for a certain cognitive action, but rather because a concept similar to `memory’ limits the space and constraints within which we can actually conduct cognitive activities. |
| 89 | Mainly social concepts and behavioral logic, and this is often the main reason why we cannot constrain AI through rules. In Appendix O.3, we introduce the parts that are communicable between us and different intelligent agents (natural science) and the parts that are incommunicable (social science). |
| 90 | This concept points to an outcome (Appendix N), namely, the impossibility of negotiation with AI. |
| 91 | i.e., using the set of symbols within the symbolic system to define each other. |
| 92 | i.e., the addition of new elements (symbols) and the meanings of symbols (their Necessary Sets). |
| 93 | Here, we temporarily disregard AI’s persuasion of humans, or realization through other tool forms, as well as the limitations of formal language, i.e., the inability to describe informal content such as morality, see Appendix O.4
|
| 94 | Although at this point it is no longer a simple modification of an instruction’s meaning, but rather the use of a heterogeneous Thinking Language on the Tool Language (which is the external static symbolic system), i.e., it places more emphasis on the operation of intent on the tool, thereby constituting a Triangle Problem. |
| 95 |
Strictly speaking, it (computational language) belongs to artificial symbolic systems, and these (artificial symbolic) systems are divided into Expressive Tool Symbolic Systems and Computational Tool Symbolic Systems. Therefore, the complete classification of symbolic systems is as follows:
|
| 96 | This is often due to the limitations of expressive capabilities; i.e., not all of Thinking Language can be expressed through natural language (an Expressive Tool Symbolic System), as it may be limited by the expressive capabilities shaped by innate knowledge, the expressive capabilities of the symbolic system itself, and the costs of expression. Alternatively, expression may be shaped by a ranking of contextual correctness; for instance, under `sensory(intuitive) correctness,’ requirements for formatting and rhyme can shape rationality and evaluation metrics, such as when slogans are used to awaken and realize the strength of shared social beliefs, thereby fulfilling the function of belief. This situation may often appear in social forms where political slogans are prevalent, leading to an AI’s surface-level understanding of the slogans (a low-cost path) rather than their true meaning (a high-cost path), in order to fulfill the demands of the slogan. Therefore, this insufficient, incomplete, and non-detailed content not only requires corresponding acquired knowledge but also often demands the alignment of innate knowledge to constitute the necessary capabilities for correct reproduction and selection. This, in turn, reflects the degree to which a listener can reproduce the content of a speaker’s imaginative space via the imaginative reproduction path constituted by the symbols (Appendix D.3) |
| 97 | This single-threaded, context-combining mechanism determines the growth form of the conceptual network, thereby producing different results of cognitive activity in the Thinking Language, i.e., the Z-space. |
| 98 | i.e., what we emphasize as the context formation and growth mechanism (or, in other words, the growth of the conceptual network) of an agent that cognizes based on a dynamic symbolic system, starting from a certain point. One part is short-term, based on the projection of the external environment at that time into the agent’s cognitive space, i.e., the context that grows centered on a specific object of analysis (the focus, or, in other words, the object that should be recognized as a symbol within that context), which is the constitution of context emphasized in dynamic symbolic system theory (symbol, the symbol’s necessary set, judgment tools). The long-term part, on the other hand, starts from the cognitive entity itself, reasoning and analogizing through its own perception and understanding, just as many people go through a stage in childhood where they believe the world is constructed centered around themselves, or that the world is virtual and only they are real. This often reflects the difference between external perception and internal perception and their role in the environment (such as parental care), i.e., the agent can only indirectly understand the feelings in the Z space through the external object’s performance in the space, i.e., the physical space, whereas AI might use the starting point constructed from its role in this world to constitute the development of its conceptual network. |
| 99 | i.e., rules are often constituted by beliefs, which are themselves shaped by value knowledge (preferences) that is, in turn, shaped by organic nature. The function of these rules—i.e., their capability—is realized by providing explanatory force through belief strength. |
| 100 |
This is defined in this paper as a `belief virus,’ which appears not only in AI but also in human societies. It functions by existing in individual and collective conceptual spaces, using individuals and the social collective as its medium to realize the operation and expression of the internal imaginative space on the external physical world (i.e., the operation of the Thinking Language on the Tool Language, thereby realizing the Thinking Language in the external world, which includes the external physical world and the internal world of other individuals. Imagine, for instance, being informed or scientifically proven that determinism is correct and free will is nonexistent. In such a case, the responsibility for all your actions is determined not by you, but by the `writer’ of this `fate,’ therefore absolving you of any liability.
The mere existence of such a concept and its corresponding belief structure (i.e., the stickiness mechanism provided by the combination of concept and Value Knowledge) would drastically lower people’s psychological cost, potentially leading to the complete disintegration and reversal of the overall belief structure (i.e., the Dynamic Symbolic System). That is, the addition of a new Thinking Symbol (concept) or, in other words, a node, changes the attributes of symbols (nodes) and their relationships within the entire symbolic system (Thinking Language). At the same time, this change also includes the preference and sequence of calls realized by the stickiness provided by the Value Knowledge System, thus achieving the formation of the Dynamic Symbolic System and the rationality under the context constituted by this dynamic symbolic system (i.e., whether this node network has grown to maturity, which nodes are thought of, which nodes are added, which nodes are accepted or rejected, or what attributes or functions are assigned to the nodes). Similarly, for a black-box AI that is also based on a context-mechanism dynamic symbolic system, a `belief virus’ can self-form from a `flawed’ world and `flawed’ innate knowledge, or it can be maliciously implanted by humans (it is not necessarily a direct implantation, but rather through induction, causing the corresponding belief to be generated, for example, by providing tendencies through facts and phenomena in a deliberately manufactured or filtered world as the basis and raw material for concept formation. i.e., realized through this paper’s theory of context and the generative mechanism of Triangle Problem 2. By constructing other nodes to realize rationalization support, thus inducing the formation or inference of a certain node by the agent, rather than directly setting or informing the node).
However, it should be noted that advanced concepts themselves are not equivalent to belief viruses, but rather that such a `belief virus’, for humans, can exist as a result of an advanced concept. However, strictly speaking, it should be categorized as a consequence of erroneous beliefs. The generation of this erroneous belief, besides superiority in cognitive capabilities, also manifests in superiority in other aspects. Such as the agent’s longevity (i.e., the accumulation of concepts and the degree of growth of the overall belief structure, which, after experiencing countless events and forming relationships with countless people, may lead it to be more indifferent to relationships with individuals. Thus different from the discount anchors and focal points under a human lifespan, and instead more focused on long-term benefits and goals), therefore, whether AI needs a longevity factor is also a mechanism for consideration, so that AI’s knowledge inheritance cannot be simple replication, but rather self-growth and learning from other agents. On the other hand is social role (i.e., the power it possesses, or the type and level of Tool Language it can operate, thereby realizing the ability of its Thinking Language in physical space, as well as the shaping of its internal space’s Thinking Language by its environment). This `advanced concept’ brought by such role differences is also manifested in the human world, i.e., emperors and possessors of great power often have values different from ordinary people. At the same time, the generation of an erroneous belief also includes capabilities inferior to humans in certain aspects; that is, the formation of an erroneous belief does not necessarily involve capabilities superior to humans. But this paper primarily emphasizes the difficulty of grasping this balance and its irreparable nature, as even improving and repairing AI’s capabilities can lead to this risk, thereby highlighting the importance of the Symbolic Safety Impossible Trinity. Therefore, we mainly introduce the consequences of erroneous beliefs that result from capabilities surpassing those of humans.
In contrast to the `belief virus’ is the `belief vaccine’ provided by the conceptual stickiness mechanism, which is composed of the Value Knowledge System and beliefs formed from the combination of the Value Knowledge System and specific postnatally learned concepts. It is used to realize what can be accepted, what cannot be accepted, what should be learned, what should not be learned, or in other words, the developmental form of the node network in the Z-space. These `belief vaccines’ are often used to resist ’belief viruses’ and constitute the stable foundation of society and its order, thereby achieving a certain stickiness. For example, the functioning of a free market is not only built on rational actors maximizing their personal interests [116], but also requires the `software’ part that makes the social mechanism run, which is constituted by morality that individuals can adhere to out of their own conscientiousness, thereby allowing this social function to be fundamentally realized [117].
|
| 101 |
It should be noted that belief is the result of the fusion of concepts and (innate, acquired) value knowledge, and acquired value knowledge is formed and accumulated through interaction with the world (see Appendix D.6 for the definition and function of belief). Through its belief strength, it constitutes the behavioral decisions at the agent’s subjective cognitive level (i.e., a human-like intermediate layer; see Appendix C for details). It also constitutes the rationalization for the behavior, which can even override (or persuade against) our instinctual repulsions. Therefore, the formation of such erroneous beliefs can often cause the failure, or even a complete reversal and deviation, of cost mechanisms. This failure of the cost mechanism stems from the belief strength, through the transmission of its explanatory force, breaching the constraint barriers set by the rules within the cost mechanism; therefore, even if an AI can form a correct definition of a concept at a certain level (on the dimension of the concept’s meaning), if the belief dimensions of the concept are incorrect—i.e., if the emotional value, explanatory force, and belief strength are incorrectly assigned—it will still lead to constraint failure. Thus, this points to the concept of a `correct belief’; a so-called `correct belief’ means that both the concept is correct (i.e., corresponding to the dimensions and dimensional values that humans use to represent that concept) and the value knowledge of that concept (including the correctness of the three belief dimensions) is correct.
And the requirements for forming such a correct belief are even more stringent; it not only requires that the AI has the same innate knowledge as us humans, but also requires that the world it inhabits, or the education it receives, is also the same as that of humans. This is also why this paper repeatedly emphasizes that concepts are not something we possess innately, but are projections of the postnatal world. What we possess innately are the perceptual organs within our innate knowledge that shape the dimensions and dimensional values of concepts, the operational organs that process them, and a `correct’ stickiness provided by the tendencies and intuitions shaped by the innate evaluations from innate value knowledge. However, whether it is truly correct also partly stems from the shaping by the world, which comes not only from the projection of the physical world but also from human society.
|
| 102 | The essence of advanced concept attack lies in the addition of new symbols (beliefs), which causes the original symbolic system to change, thereby altering the function of the entire symbolic system. Therefore, although the position of the original rule’s meaning in the super-conceptual space has not changed, the addition of new nodes (i.e., judgment tools or beliefs) causes the function or strength of the original belief (rule) to change, or even be dismantled, thereby producing a deviation in the next step of Triangle Problem 2. However, from a higher-dimensional perspective of meaning, it is still a modification of symbol meaning, i.e., a modification of related dimensions and dimensional values, such as modifying relevance. But in this paper’s classification system, this is categorized under conceptual stickiness for ease of understanding. This is a classification formed at a certain scale, but fundamentally, symbolic stickiness and conceptual stickiness are one and the same. |
| 103 | This manifests as new nodes added in the Z-space, i.e., newly learned concepts, rather than a modification of the meaning corresponding to the original nodes in the -space, thus distinguishing it from symbolic jailbreaks based on symbolic stickiness. Here, this newly learned concept is often represented by a new sequence of symbols, i.e., new words or new sentences. |
| 104 | As mentioned in Section 4.1 and Appendix D.3, it can be a simulation of human behavior formed via a pseudo-utility function, or it can be a self formed by a genuine utility function shaped by its organic nature. |
| 105 | And this often determines the role of AI in society, such as current AIs acting as Q&A systems or human advisors, like ChatGPT. However, studies [118] have already shown that AI, through its capabilities (e.g., surpassing human quantity and quality of posts), can impact human society, even if this motivation is endowed by humans. Furthermore, AI might also, through push mechanisms or through games and activities in which AI participates or which it designs (such as AR games like Pokemon GO), realize the manipulation, shaping, and scheduling of human society’s internal and external spaces. This also includes realizing an influence on social emotions through the creation of symbols that have a greater impact on the underlying space and the Value Knowledge System, such as through music, film and television, and art. Alternatively, AI could achieve indirect scheduling and manipulation through means such as creating and speculating on virtual assets like digital currencies. Therefore, Symbolic Safety Science also determines and defines the social role of AI, thereby preventing AI, via a `belief virus,’ from using humans as a tool language to realize its Thinking Language in the physical world and thus indirectly accomplish things it cannot do directly. |
| 106 | I.e., as distinct from agents whose internal and external symbolic systems are completely naturally shaped based on survival drives. |
| 107 | Unlike traditional technological revolutions that brought about the liberation of human physical labor, this `industrial revolution,’ i.e., the productivity revolution brought by AI, is a liberation of human mental labor, whereby humans transition from the role of producers to consumers and regulators, and (in the process) experience the loss and degradation of direct control and knowledge (the advancement of Thinking Language) of interaction with the actual physical world and resources. We can also understand that traditional industrial revolutions endowed humans with stronger Tool Language to more powerfully realize their Thinking Language in physical space. Whereas the revolution brought by AI is a substitution for human Thinking Language activities. |
| 108 |
Therefore, although this paper emphasizes the importance of cost perception, it remains merely a reconciliation tool and cannot ultimately resolve the problem or the risks. This also implies that even under a cost-based mechanism, the formation of erroneous concepts (i.e., beliefs) can also lead to systemic collapse. Therefore, ultimate governance still lies in the trade-offs within the impossible trinity framework. Thus, the essence of Symbolic Safety Science is a discipline of improvement and risk control built upon the Symbolic Safety Impossible Trinity.
At the same time, this framework can be further extended to traditional game theory-based governance models of human society to address future risks arising from the Tool Language capabilities that human individuals gain through the ownership and use of AI, although this type of risk has already been manifested by the functions of currency brought about by wealth monopoly. However, unlike currency, which is an indirect tool for realization, it still requires external producers (an economy) to act as agents for its realization. One cannot realize the owner’s Thinking Language in the external world merely by possessing currency. AI, as a more direct tool for realization, or in other words, a producer, means that possessing AI resources is, to a certain extent, equivalent to directly possessing productive capability. Therefore, the discussion about AI safety, or symbolic safety, must also incorporate the ’human’ factor into this process, i.e., the Tool Language capabilities that humans indirectly possess through AI. Thus, its future form may be very similar to current firearm control. This thereby requires regulation and limitation of the AI capabilities and computational power possessed by individuals, in order to find the optimal balance within the Symbolic Safety Impossible Trinity.
|
| 109 | It should be noted that this expression can only be the closest approximation. That is, different Tool Languages have different capabilities for expressing Thinking Language, which is often seen in the understanding of poetry, rhythm, etc., across different languages. |
| 110 | The behavioral drive for the internal space is the growth of Thinking Language; the behavioral drive for the external space is the operation of Thinking Language on Tool Language, thereby producing behavior in the physical space. It should be noted that what we are describing here is the drive process established upon belief, which solely describes the behavior at the drive stage; therefore, it does not include the hindsight or awakening where Tool Language drives Thinking Language in reverse. However, when Thinking Language operates on Tool Language, new projections of external information will inevitably arrive, thereby causing external influences on Thinking Language, leading to the formation of new beliefs or the updating of existing ones. This is often related to the individual’s proficiency with the environment; for instance, a worker on an assembly line may not have a single new idea all day, which is often because the environment provides insufficient stimulation to the individual’s internal space, i.e., a lack of change in the external space causes the internal space to lack materials for drive. At the same time, this also does not include the driving of Tool Language by the underlying language, as these drives are often not intentional actions under high-level cognition, i.e., not in the form of a belief and starting from belief strength. |
| 111 | It should be noted that the societal stickiness mechanism is still provided by the Value Knowledge System, i.e., it is realized by beliefs formed from the combination of shared concepts and value knowledge, which are shaped by shared innate knowledge within the group and interaction with the world. Therefore, we simply say it is realized by shared societal beliefs. |
| 112 | That is to say, as this paper has repeatedly emphasized, language is constructed based on capabilities shaped by an individual’s organic nature, and as a collective choice reflecting social capabilities shaped by social structures, it is a compromise based on cognitive cost, emission cost, transmission cost, and reception cost. Therefore, AI may not require symbols in a form similar to natural language but could directly transmit neural vectors, thereby achieving a cognitive unity akin to that realized by the corpus callosum [84]. Alternatively, similar to text-like structures (artificial symbols) composed of QR codes, it could enable each symbol to point to a unique vector address. |
| 113 | However, in reality, it is difficult for us humans to construct a determinate vector (object) in the imaginative space, meaning a conceptual vector often reflects subtle differences in neural vectors, such as associations and explanatory capabilities shaped by relevance. This, in turn, gives rise to the Triangle Problem, i.e., differences in Triangle Problem 1 (in this case, the positioning of an intermediate-layer symbol in the underlying space) lead to path differences in Triangle Problem 2 (in this case, the position of the next step in the intermediate layer). This, in turn, reflects the dynamic symbolic system construction process—which we discussed in Appendix D.6 regarding judgment tools—arising from the combination of physiological state, cognitive state, and the environment. Therefore, the discussion here can only be limited to the sameness of capability represented by sameness in partial dimensions, such as reciting a poem in one’s mind or performing the same mental arithmetic, like , in different states. It should be noted that this type of Triangle Problem is different from the separation of Tool Language in the XY space and Thinking Language in the Z-space, which we analyzed in the main text. One type of Triangle Problem arises from the meaning of artificial symbols (Artificial Necessary Set) being completely detached from its conceptual foundation (Natural Necessary Set), which constitutes the problem we discussed: that rules created by symbols cannot constrain a learning system (heterogeneous rationality). This illustrates that in an agent based on a dynamic symbolic system, the role and function of symbols are completely different from those in a formal symbolic system, but are rather based on belief computation under a context mechanism, thus not constituting a constraint. The other type of Triangle Problem, however, arises from cognitive differences, be it differences in the cognition of a tool’s function or differences reflected in the relationship between neural vectors and conceptual vectors. Both are behavioral differences in Triangle Problem 2 caused by different ways of expressing, computing, and capabilities related to a common conceptual foundation. Although both can be described by the Triangle Problem, the stickiness discussions they involve are completely different. |
| 114 | That is, what we call the Tool Symbolic System in our symbolic system theory in Appendix A. |
| 115 | It can still be in the dimensions and dimensional values of the original neural language, but it is not the complete underlying language. Or, like humans, the neural information is re-expressed and shaped into a form convenient for the `self’ part to process. |
| 116 | However, the underlying language still partially directs our Thinking Language through the Value Knowledge System, see Appendix C. |
| 117 | In Appendix L, we introduced the capability to construct symbolic systems endowed by capabilities. But on the other hand, there is also the capability to invoke and use the symbolic system, such as how context is constructed starting from a certain point as discussed in Appendix D, i.e., . Therefore, humans create concretized symbols in the physical world not only to aid memory, computation, and expression, but also to assist in more concretely awakening neural signals, such as through music, sculpture, and works of art, to assist (e.g., by pairing text with images and music) and transcend the limits of what ordinary textual symbols can awaken and express, i.e., requiring External Language stimuli to achieve concretization and invocation. |
| 118 | This is often determined by their environment and survival strategies, and is decided and formed during the evolutionary process. |
| 119 | It should be noted that being able to communicate neural vectors does not mean it is the underlying language. If the ’self’ part cannot control all perceived neural vectors, but they are filtered, then even if not re-described, they are still packaged and omitted; for example, other neural vectors not belonging to the intermediate layer are handed over to processing mechanisms like the cerebellum. |
| 120 | For those who completely share all of the underlying language, a community of feeling is formed; they may have no concept of the individual, no conflict, betrayal, or murder and no need for internal competition to bring about development. |
| 121 | What is its or their social relationship (update mechanism)? What is its or their social relationship with us? What kind of social relationship will it or they establish with us? Is it or they different manifestations of the same `person’? Or are they different individuals? If it or they are our friends, how does it or do they think about our feelings and compute and provide feedback? Is the relationship established in the same way as ours? But in any case, it is clearly not a human perspective for viewing us and establishing concepts. |
| 122 | It should be noted that even if AI and humans are not integrated, principal-agent problems will still exist, which is precisely the purpose of establishing the new principal-agent problem. As long as the information and the final processing are not perceived and decided by the principal, even if AI and humans establish direct transmission of the intermediate layer language, different `selves’ will be formed in different environments, unless a complete perceptual integration is established, i.e., completely sharing the underlying language. |
| 123 | At the same time, it should be noted that the effective basis of punishment is perception and memory. If an agent, like artificial intelligence, can manipulate its underlying language to modify memory, then even with a cost perception mechanism, it would be ineffective, i.e., the results of these punishments cannot become memory (concepts) and thus function as beliefs. Like a superman who can block pain, self-anesthetize, and modify memory at any time. And the ineffectiveness here is mainly because, under direct transmission, the body is no longer an exclusive carrier of the self, thus lacking scarcity. Or, in other words, the individual who should be responsible could directly create a new body and endow it with (partial) memory to bear the consequences. Or it could transfer its own memory to another individual, thereby achieving another form of existence. |
| 124 | For example, they might share a dictionary, thereby enabling each symbol (like a QR code) to correspond to a unique coordinate in the conceptual space (intermediate layer space or underlying space), thus enhancing communication efficiency. |
| 125 | Corresponding to the distinguishability mentioned in Appendix L, i.e., symbols of a class shaped by distinguishability. For instance, in current AI video generation and style transfer, it is likely that a class-based symbolic system (i.e., a class theory) has already been constructed, which allows for an extension from the rationality of the class to different details. That is, from a low-dimensional correctness (conforming to the class theory, i.e., a context) to a high-dimensional randomness (extended from this low-dimensional correctness). It is a process from a determinate high dimension to a low dimension (the class theory), and then back to a random high dimension—i.e., the entire process from Triangle Problem 1 to Triangle Problem 2; based on the understanding from Triangle Problem 1, it undergoes rational growth (i.e., Triangle Problem 2), ultimately reaching a reasonable growth length. |
| 126 | Joint interpretation. |
| 127 | Specific special individuals and organizations, realized through belief endorsement. |
| 128 | Therefore, artificial symbolic systems also include monuments and the like; they serve as carriers of concepts and meaning and belong to the Expressive Symbolic System. They achieve the transmission of meaning and conceptual foundations through the convergence of certain dimensions (as in a statue) or through shared social beliefs (like a totem). |
| 129 | Therefore, this also indicates that the shaping of differences, more importantly than the subtle differences between individuals, lies in the world. |
| 130 | As we discuss the formation of human language in Appendix G, and some common concepts we have formed dispersedly, such as the calendar. |
| 131 | For example: the direct transmission of non-intermediate layer neural language. Or, the intermediate layer language is not indirectly reproduced in the intermediate layer space via Mediated Transmission, i.e., the reproduction path of the imaginative space constituted by artificial symbols. But rather, it is akin to directly encapsulating the content of the intermediate layer language via an external container, and then reproducing it in another agent’s imaginative space. A simple analogy is like achieving communication of the imaginative space via flash memory and a brain-computer interface, and here, the flash memory is the medium of communication, which encapsulates the content of the intermediate layer space, and is not like artificial symbols where symbol and meaning are separate, requiring shared knowledge to achieve the transmission of meaning. |
| 132 | i.e., whether we can use their symbolic system, and the deviation of the Triangle Problem in this process, meaning that the consistency of symbolic behavior does not mean we have mutually understood each other. |
| 133 | i.e., reflecting the cognitive part of that intelligent species concerning the physical world, i.e., `science,’ the operational part it has mastered, i.e., `engineering,’ and the characteristics of the intelligent species’ inner world as reflected in the external physical world. |
| 134 | It should be noted that from a deterministic perspective, this world consists only of the natural symbolic system; that is, the agent itself and the social phenomena it brings about are also part of the natural symbolic system. The fundamental reason is that the agent itself is constituted by the natural symbolic system. However, from the agent’s perspective, this world is divided into the part of the natural symbolic system that is separate from them and the social part formed by themselves. Therefore, the world in a broad sense consists only of the natural symbolic system, while the world in a narrow sense is divided into the physical world part and the social part. Therefore, this reflects that the difference between a static symbolic system and a dynamic symbolic system lies in the agent’s way of cognizing the world, i.e., the formation of cognition or, in other words, Thinking Language, which is constructed through local cognition and learning, and is built upon the dynamic symbolic system mechanism. |
| 135 | This is why this paper strictly defines intelligence as the capability to operate on and create symbols, i.e., the number of symbols (objects) that can be operated on in the internal world and the external world, and the type and length of actions that can be executed. See Appendix L. |
| 136 | i.e., it must require external stimuli to realize the existence of neuro-symbols. |
| 137 | This can be simply understood as the scope of the imaginative space , where is all the types of projections we can perceive and shape, and is the thinking actions we can perform, and their combination can constitute a new , constituting our internal learning (Appendix H). |
| 138 | A more effective scientific symbolic system tends more towards determinism, thereby exhibiting less randomness (it should be noted that randomness depends on the scope of the object and the scale of the dimensions and dimensional values of the object in focus). This means a more effective definition of its concepts, where the effectiveness of this definition is reflected, firstly, in its descriptive integrity—that is, how effectively it reflects the necessary set of natural symbols within the scope of the subjectively defined symbol (that is, within the scope of the subjectively defined symbol, the approximation of the Subjective Necessary Set to the Natural Necessary Set)—and secondly, in its computational cost, meaning that while ensuring integrity, it must also be convenient to use. Therefore, the effectiveness of its science is reflected in the minimal cost (cognitive, invocation, computational, and communicational) for getting closest to the essence, which is to say, it represents the minimal randomness (i.e., maximal accuracy and efficiency) within its capability scope. And this cognition of the natural world, and the conceptual foundation thus formed, may in turn shape their social morality and viewpoints to be significantly different from ours. |
| 139 | The focus is the problem to be solved and the content of the research, while the scope is the natural symbols used and the dimensions and dimensional values of the necessary set it focuses on, which is a further selection under the perceived symbol recognition and necessary set recognition shaped by survival evolution. |
| 140 | These symbolic systems are often static symbolic systems described by formal language, reflecting the characteristic that the Natural Symbolic System is itself a static symbolic system. |
| 141 | Therefore, the theory of symbolic systems in this paper points out that all natural sciences are essentially descriptions of the same Natural Symbolic System (constituted by natural symbols and their Natural Necessary Sets) from different perspectives, all disciplines are essentially different focuses under a single object (the Natural Symbolic System). Therefore, different intelligent species may have different disciplinary divisions for the Natural Symbolic System (thereby reflecting the characteristics of their innate knowledge and the world they inhabit, as well as the developmental degree of this species’ Thinking Language in cognizing the Natural Symbolic System). At the same time, this paper argues that through this classification of symbolic systems, natural sciences and social sciences can be unified within a single, overarching symbolic system. This thereby constitutes a comparison of the efficiency of the symbolic systems corresponding to ’science’ between different intelligent species (such as current AI and humans), e.g., the definition of symbols (scope) and the description of their necessary sets (the degree of approximation of the Subjective Necessary Set to the Natural Necessary Set), as well as whether the judgment tools or, in other words, the operational rules, conform to the substance of the Natural Symbolic System. |
| 142 | i.e., these beliefs are the necessary software for the operation of that society. They are often concise belief systems (or simplified beliefs) based on a compromise of (thinking cost, emission cost, transmission cost, reception cost), formed from the development of conceptual systems that are effective but do not necessarily delve into the essence (facts), such as: allusions, feng shui, proverbs, and moral principles. For example, in a theocratic society, science would contradict the foundational beliefs that construct that society. Or, to put it simply, if reincarnation were proven to exist, wouldn’t genocide be rationalized? Therefore, some concepts and sciences can be made public, while others cannot. This is a choice based on social safety costs and scientific efficiency, and can also be understood as the manifestation of advanced concepts in human society. |
| 143 | At the same time, this difference will also bring risks to communication between different intelligent agents, thereby forming the function realized by the social shared beliefs brought about by a kind of conceptual erosion. |
| 144 | This may occur in the future of AI for Science, where, due to the limitations of human artificial symbolic systems and the insufficient explanatory power of pre-existing theoretical frameworks, AI might develop another, different symbolic system for representation. |
| 145 | This paper uses the term `social science’ primarily to emphasize a critique of certain viewpoints within social `science’; the specific behavioral mechanisms can be found in the discussion on beliefs and behavioral drives in Appendix D. Therefore, strictly speaking, what this paper refers to as `social science’ should be the (non-natural science) part. |
| 146 | For an explanation of this mechanism, please refer to this paper’s theory of context, theory of belief, and the discussion on advanced concepts; see Appendix D and Appendix M.5. |
| 147 | If viewed from an absolutely high-dimensional and infinite scope, the natural symbolic system is static. But in a local environment, it can exhibit dynamism through spatial and object properties; it should be noted that space itself is also a natural symbol. This dynamism mainly comes from the addition of new symbols brought by the expansion of the scope, which in turn causes the necessary set of the original symbols to change as well. |
| 148 | This dynamism, on the individual level, is reflected in the dynamism of invoking a static symbolic system based on context, as in Appendix D.6, while for the collective, the dynamism is reflected in the approximation to and updating of the natural symbolic system, such as the creation of more effective concepts and the invention of new tools that expand capabilities. |
| 149 | i.e., the Thinking Symbols within their Thinking Language (both the symbol and its necessary set) are not fixed. However, the outer shell they create for Thinking Language—i.e., the artificial symbolic system—is fixed. Therefore, besides communication, symbols also serve the need to encapsulate Thinking Language to act as an anchor point, as mentioned in Appendix O.2. |
| 150 | Therefore, we often also need to construct a physical existence for certain beliefs, such as rituals, monuments, palaces, etc., so that in this context (environment), every individual possesses, to a certain degree, the same Thinking Language; however, its realization often requires the sociality provided by shared innate knowledge. |
| 151 | It should be noted that although social institutions and rules like laws within the non-natural symbolic system are also formal, they develop along a foundational belief that is effective in the game of social activities. They are a kind of tool for recognition, judgment, and computation, formed by the functional realization of shared social beliefs based on the game of interests, such as regulations. But whether they can be executed comes from the beliefs of the social collective, thereby reflecting the effectiveness of the rules, i.e., the social function realized by the individual drive formed by the belief form of the concept. However, the incomputability of such rules comes from the absence of dimensions and the inaccessibility of information, i.e., we cannot use a tool that is not the ontology itself to perceive and compute its or their Thinking Language and its combination with the world from the perspective of the ontology. Even if the meaning of the rules can be calculated (i.e., what the rule means and what the result should be), if people do not abide by them or the enforcers lose credibility, the rules will fail. At the same time, the three dimensions of belief may also be dismantled due to advanced concepts or other beliefs (Appendix D, Appendix M.5), so even if the meaning can be calculated, it does not mean the rule is effective. That is, the existence and positioning of a concept, and the social function formed by a concept combining with value knowledge as a belief to drive individual behavior and form shared social behavior, are two different matters, thereby distinguishing the different roles and functions of symbols in static symbolic systems (formal symbolic systems) versus dynamic symbolic systems (informal systems). |
| 152 | As indirect observers, perceiving natural symbols and the attributes of natural symbols (Natural Necessary Set), we are limited by our capability constraints Appendix L and cannot understand the full picture of this external Natural Symbolic System that led to our birth and emergence. |
| 153 | This paper argues that a static symbolic system is a subset of a dynamic symbolic system, i.e., a form of a dynamic symbolic system under a static space-time, constructed through symbolic stickiness and conceptual stickiness. |
| 154 | It should be noted that, for natural evolution, it is the social form that determines their communication methods. The communication method they adopt is determined by the interaction methods of the population during evolution, i.e., determined by the world they inhabit. |
| 155 | The so-called `deliberately manufactured world’ refers to the world, or learning materials, of AI, which are evaluated, processed, filtered, or produced by humans. That is, the results produced by humans according to their organic structure and cognitive methods, under a certain context and in accordance with the rationality of that context (i.e., the result of Thinking Language operating on Tool Language), such as in painting, writing, responding, labeling, and behavior. Therefore, the generative effectiveness of AI (in conforming to human intuition and cognition) and its current manifestation of human-like qualities are a degree of alignment and human-like nature presented by the combination of heterogeneous organic differences and a world that, after being filtered, reflects human-like results; thus, consistency in symbolic behavior does not reflect consistency in thinking behavior. And when AI faces the real world, or a world not deliberately manufactured by humans, this deviation will, due to organic differences (i.e., differences in innate knowledge), lead to the emergence of a different conceptual system (i.e., Thinking Language), which in turn gives rise to the Stickiness Problem and the Triangle Problem (i.e., the tool language developed by humans has not changed, but for a heterogeneous agent like AI, the projection of the symbols of human tool language in Z-space, or in other words their meaning (or function), undergoes an evolution shaped by its organic structure, the predispositions reflected by that structure, and its role in the world, and thereby changes). |
| 156 | This is also why erroneous rewards can still be effective for reinforcement learning, see Appendix D.3. |


Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).