Submitted:
19 March 2025
Posted:
21 March 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Conceptual Foundations
2.1. The Elusive Nature of Intelligence
2.2. General Intelligence vs. Modularity
2.3. Introspection in Large Language Models (LLMs)
2.4. Emergent Properties: Genuine or Illusory?
2.5. Conclusion of Conceptual Challenges
3. Philosophical Critiques
3.1. Searle’s Chinese Room Argument
3.2. The Test-Passing Illusion
3.3. Anthropomorphism and Bias in Evaluation
3.4. Implications for AGI Development
4. Critique of AGI Benchmarks
4.1. Benchmark Limitations and Psychometric Gaps
4.2. Frontier Benchmarks, Emergent Abilities, and Ongoing Challenges
4.3. Moving Toward Adaptive, Psychometrically Informed Tests
- Adaptive Testing: Dynamically adjusting question difficulty based on performance, akin to how human IQ tests find an individual’s ceiling.
- Construct Validity: Designing tasks to isolate cognitive skills like abstraction, causal reasoning, and ethical judgment rather than conflating these with domain knowledge.
- Standardized Administration: Ensuring no training-data overlap and consistent testing conditions to rule out data leaks or resource inequalities.
- Reliability Checks: Re-testing models under different initialization or prompt conditions to gauge performance stability.
5. Psychometric Perspectives
5.1. Validity, Reliability, and Human Comparison
5.2. Sensitivity and Specificity in AI Tests
6. Ethical Implications
6.1. Risks of Overclaiming and Societal Misdirection
6.2. Transparency, Accountability, and Benchmark Honesty
6.3. On the Horizon
7. Conclusion
7.1. Summary of Key Findings
- Conceptual Foundations: Intelligence may be less a single capacity than a mosaic of specialized abilities, complicating efforts to replicate or measure “general” intelligence in machines.
- LLM Introspection and Emergent Behaviors: While large-scale models display intriguing phenomena—such as apparent self-reflection or sudden skill acquisition—these might be measurement artifacts or sophisticated pattern usage rather than genuine general intelligence.
- Philosophical Critiques: Searle’s Chinese Room highlights the gap between simulating understanding and possessing it. The test-passing illusion warns against conflating performance metrics with true cognition.
- Benchmark Limitations: Existing AGI benchmarks often lack psychometric rigor and may be contaminated by training data, overstating model competencies.
- Ethical Dimensions: Overclaims about AGI risk misallocation of resources, erosion of trust, and misguided policy. Transparent, accountable practices are essential for responsible AI advancement.
7.2. Toward a More Rigorous AGI Paradigm
7.3. Future Directions
- Adopt Psychometric Standards: Develop adaptive, validated, contamination-resistant benchmarks with strong construct validity and reliability measures.
- Investigate Mechanisms in LLMs: Probe how introspection-like and emergent behaviors arise, distinguishing illusions from substantive progress toward general cognition.
- Incorporate Interdisciplinary Insights: Use philosophy, cognitive science, and neuroscience to refine definitions of intelligence and create ethically robust AI.
- Ensure Ethical Transparency: Avoid overstating capabilities, reveal evaluation details, and engage with regulatory bodies when deploying high-stakes systems.
References
- Goertzel, B., & Pennachin, C. (Eds.). (2007). Artificial General Intelligence. Springer.
- Legg, S., & Hutter, M. (2007). Universal Intelligence: A Definition of Machine Intelligence. Minds and Machines, 17(4), 391–444. [CrossRef]
- Silver, D., Schrittwieser, J., Simonyan, K., et al. (2018). A General Reinforcement Learning Algorithm that Masters Chess, Shogi, and Go Through Self-Play. Science, 362(6419), 1140–1144. [CrossRef]
- Marcus, G. (2022). Artificial Intelligence: A Guide to the Future. [Publisher Information].
- Russell, S. (2019). Human Compatible: Artificial Intelligence and the Problem of Control. Viking Press.
- Hernandez-Orallo, J. (2017). The Measure of All Minds: Evaluating Natural and Artificial Intelligence. Cambridge University Press.
- Gottfredson, L. S. (1997). Mainstream Science on Intelligence: An Editorial With 52 Signatories, History, and Bibliography. Intelligence, 24(1), 13–23. [CrossRef]
- Sternberg, R. J. (2000). The Modularity of Mind: Debates in Cognitive Science. In Handbook of Intelligence (pp. 15–37). Cambridge University Press.
- Marcus, G. (2022). The Next Decade in AI: Four Steps Towards Robust Artificial Intelligence. AI Magazine, 43(1), 49–63. [CrossRef]
- Searle, J. R. (1980). Minds, Brains, and Programs. Behavioral and Brain Sciences, 3(3), 417–457.
- Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021). On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, 610–623.
- Brundage, M., Avin, S., Wang, J., Belfield, H., Krueger, G., Hadfield, G., & Dafoe, A. (2018). The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation. arXiv preprint arXiv:1802.07228.
- Whittlestone, J., Nyrup, R., Alexandrova, A., & Dihal, K. (2019). Ethical and Societal Implications of Algorithms, Data, and Artificial Intelligence: A Roadmap for Research. Minds and Machines, 29(3), 395–410.
- Spearman, C. (1904). General Intelligence, Objectively Determined and Measured. The American Journal of Psychology, 15(2), 201–293. [CrossRef]
- Gardner, H. (1983). Frames of Mind: The Theory of Multiple Intelligences. Basic Books.
- Anderson, M. L. (2010). Neural reuse: A fundamental organizational principle of the brain. Behavioral and Brain Sciences, 33(4), 245–313. [CrossRef]
- Gazzaniga, M. S. (2018). Cognitive Neuroscience: The Biology of the Mind. W.W. Norton & Company.
- Silver, D., et al. (2017). Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm. Nature, 550, 354–359.
- Brown, T., et al. (2020). Language Models are Few-Shot Learners. Advances in Neural Information Processing Systems, 33, 1877–1901.
- Lake, B. M., Ullman, T. D., Tenenbaum, J. B., & Gershman, S. J. (2017). Building Machines that Learn and Think Like People. Behavioral and Brain Sciences, 40, e253.
- Burkart, J. M., Schubiger, M. N., & van Schaik, C. P. (2017). The evolution of general intelligence. Behavioral and Brain Sciences, 40, e195.
- [Placeholder: Example research on LLM introspection or self-reflection; replace with a real citation as needed.].
- [Placeholder: Example research on privileged access in LLMs; replace with a real citation as needed.].
- [Placeholder: Example reference discussing emergent abilities, e.g., DeepChecks blog or relevant arXiv.].
- [Placeholder: E.g., a Wired article or academic paper calling emergent claims a “mirage.”].
- Nass, C., & Moon, Y. (1994). Machines and Mindlessness: Social Responses to Computers. Journal of Social Issues, 50(1), 81–103. [CrossRef]
- Tihanyi, N., Bisztray, T., Dubniczky, R. A., Toth, R., Borsos, B., Cherif, B., Ferrag, M. A., Muzsai, L., Jain, R., Marinelli, R., Cordeiro, L. C., Debbah, M., Mavroeidis, V., & Josang, A. (2024). Dynamic Intelligence Assessment: Benchmarking LLMs on the Road to AGI with a Focus on Model Confidence. arXiv preprint arXiv:2410.15490.
- WJARR Editorial. (2024). A Critical Review Towards Artificial General Intelligence: Challenges, Ethical Considerations, and the Path Forward. World Journal of Advanced Research and Reviews, 8(1), 1–12.
- ResearchGate Authors. (2024). Opinion Paper: The Frustrating Quest to Define AGI. ResearchGate Publications.
- [Placeholder: Reference for FrontierMath or other advanced new benchmark to avoid contamination.].
- Cronbach, L. J., & Meehl, P. E. (1971). Construct Validity in Psychological Tests. Psychological Bulletin, 52(4), 281–302. [CrossRef]
- Bostrom, N. (2014). Superintelligence: Paths, Dangers, Strategies. Oxford University Press.
- Floridi, L., & Cowls, J. (2020). A Unified Framework of Five Principles for AI in Society. Harvard Data Science Review, 2(1).
- Altman, D. G., & Bland, J. M. (1994). Diagnostic Tests 1: Sensitivity and Specificity. British Medical Journal, 308(6943), 1552. [CrossRef]
- Pepe, M. S. (2003). The Statistical Evaluation of Medical Tests for Classification and Prediction. Oxford University Press.
- Price, C. J., & Friston, K. J. (2002). Degeneracy and cognitive anatomy. Trends in Cognitive Sciences, 6(10), 416-421.
- Fodor, J. A. (2008). The modularity of mind: An essay on faculty psychology. In J. E. Adler & L. J. Rips (Eds.), Reasoning: Studies of human inference and its foundations (pp. 878–914). Cambridge University Press.
- Turing, A. M. (1950). Computing Machinery and Intelligence. Mind, 59(236), 433–460.
- Weizenbaum, J. (1966). ELIZA—A Computer Program for the Study of Natural Language Communication Between Man and Machine. Communications of the ACM, 9(1), 36–45. [CrossRef]
- Madhavan, P., & Wiegmann, D. A. (2007). Similarities and differences between human–human and human–automation trust: An integrative review. Theoretical Issues in Ergonomics Science, 8(4), 277–301. [CrossRef]
- Riek, L. D. (2017). Healthcare robotics. Communications of the ACM, 60(11), 68–78.
- de Visser, E. J., Peeters, M. M., Jung, M. F., Kohn, S., Shaw, T. H., Pak, R., & Neerincx, M. A. (2017). Trust in automation: An updated review of trust across disciplines. Human Factors, 59(5), 834–861. [CrossRef]
- Waytz, A., Cacioppo, J., & Epley, N. (2010). Who sees human? The stability and importance of individual differences in anthropomorphism. Perspectives on Psychological Science, 5(3), 219–232. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).