Submitted:
21 February 2024
Posted:
23 February 2024
You are already at the latest version
Abstract
Keywords:
1. Introduction
- (i)
- Identify and characterize the potential universals, and, more importantly,
- (ii)
- Establish relations between them.
Does one imply the other? Or are they independent? These questions can only be addressed from a common framework with which to compare them. This is what we (hope to) provide in this work.What is the relation between mechanistic-universality and emergent-univerality?
2. The basic structures of universality and recursion
Basic structure of universality
Given a set C and a relation R landing in C,
a finite u is universal if for all .
In this case, R consists of applying a finite set of rules (or ‘mould’) a potentially unbounded number of times to generate the elements of a family. Note that this use of the term ‘grammar’ is broader than the usual one. A paradigmatic example are so-called generative grammars (Section 3.1), such as and , with S the start symbol, A and B terminal symbols and 0 the empty symbol, where these two rules (the ‘mould’) generate the (context-free) language (the set C; see Figure 2c).Basic structure of recusionLet u be universal for set C with relation R. u is recursive if R reads ‘can be applied to a base case a finite but unbounded number of times’. We call such u a grammar.
3. Grammars
3.1. Generative grammars
3.2. Universal Grammar
3.3. Greenberg Universals
An example of a statement of the second type—which are the vast majority—is:A language never has more gender categories in nonsingular numbers than in the singular.
If a language has the category of gender, then it has the category of number.
3.4. On the relation between UG and GU
4. Writing of languages
or the ideographic writing system Blissymbolics. These writing systems do not obey the so-called rebus principle, by which symbols represent phonemes, instead of spoken words [70]. Since the phonemes of a language constitute a finite alphabet P, and spoken words can be identified with finite strings of phonemes (i.e. elements of ), it follows that in writing systems using the rebus principle, a finite number of symbols suffices to represent an unbounded number of words. This is in opposition to writing systems that do not (exclusively) use the rebus principle, where a new lexical item6 generally leads to the introduction of a new symbol.The basic structure of universality is instantiated by letting C be the set of possible lexical items, R the relation of written representation, and u the finite repertoire of symbols (we shall be more specific below). Note that ‘possible’ is an important qualification in this definition, as every natural language has a finite number of lexical items and hence a finite representation, but not necessarily a finite representation of any possible lexical item. While we cannot characterize (or even grasp) the set of possible lexical items, we posit that it be unbounded, and that suffices for our argument.A writing system for a natural language is universal if it can represent any possible lexical item with a finite number of symbols.
5. Emergent language universals
5.1. Zipf’s law: Statement and origin
5.2. On the universality of Zipf’s law and Turing machines
6. Summary and Outlook
Appendix A. Infinite sets with a finite description
| Language L | Attribute | |
| With structure | With grammar | Computable |
| Finite description | Grammar | T’s transition rules |
| Without structure | Without grammar | Non-computable |
| Infinite ‘description’ | Listing all elements in L | Specifying for all |
Appendix B. Communicative conflicts leading to Zipf’s law
Start from a given sample with distribution . For every , compute under the following conditions:
- (i)
- For all n it must hold that (where is fixed);
- (ii)
- is maximal for all n.
References
- Maynard Smith, J.; Szathmáry, E. The Major Transitions in Evolution; Oxford University Press: Oxford, UK, 1995. [Google Scholar]
- Hauser, M.D.; Chomsky, N.; Fitch, W.T. The faculty of language: what is it, who has it, and how did it evolve? Science 2002, 298, 1569–1579. [Google Scholar] [CrossRef] [PubMed]
- Bickerton, D. Language and species; University of Chicago Press, 1990. [Google Scholar]
- Nowak, M.A.; Komarova, N.L.; Niyogi, P. Computational and evolutionary aspects of language. Nature 2002, 417, 611–617. [Google Scholar] [CrossRef]
- Jackendoff, R. Précis of foundations of language: Brain, meaning, grammar, evolution. Behav. Brain Sci. 2003, 26, 651–665. [Google Scholar] [CrossRef]
- Harley, T.A. The psychology of language: From data to theory; Psychology press, 2013. [Google Scholar]
- Zanette, D.H. Analytical approach to bit-string models of language evolution. Int. J. Mod. Phys. C 2008, 19, 569–581. [Google Scholar] [CrossRef]
- Solé, R.V.; Corominas-Murtra, B.; Fortuny, J. Diversity, competition, extinction: the ecophysics of language change. J. R. Soc. Interface 2010, 7, 1647–1664. [Google Scholar] [CrossRef] [PubMed]
- Cavalli-Sforza, L.L. Genes, peoples, and languages. Proc. Natl. Acad. Sci. USA 1997, 94, 7719–7724. [Google Scholar] [CrossRef] [PubMed]
- Steels, L. Evolving grounded communication for robots. Trends in cognitive sciences 2003, 7, 308–312. [Google Scholar] [CrossRef] [PubMed]
- Steels, L. The Talking Heads experiment: Origins of words and meanings; Language Science Press, 2015; Volume 1. [Google Scholar]
- Lake, B.M.; Ullman, T.D.; Tenenbaum, J.B.; Gershman, S.J. Building machines that learn and think like people. Behav. Brain Sci. 2017, 40, e253. [Google Scholar] [CrossRef] [PubMed]
- Solé, R.V.; Corominas-Murtra, B.; Valverde, S.; Steels, L. Language networks: Their structure, function, and evolution. Complexity 2010, 15, 20–26. [Google Scholar] [CrossRef]
- Borge-Holthoefer, J.; Arenas, A. Semantic networks: Structure and dynamics. Entropy 2010, 12, 1264–1302. [Google Scholar] [CrossRef]
- Cong, J.; Liu, H. Approaching human language with complex networks. Phys. Life Rev. 2014, 11, 598–618. [Google Scholar] [CrossRef]
- Stella, M.; Beckage, N.M.; Brede, M.; Domenico, M.D. Multiplex model of mental lexicon reveals explosive learning in humans. Sci. Rep. 2018, 8, 2259. [Google Scholar] [CrossRef]
- Seoane, L.F.; Solé, R. The morphospace of language networks. Sci. Rep. 2018, 8, 10465. [Google Scholar] [CrossRef]
- Baronchelli, A.; Felici, M.; Loreto, V.; Caglioti, E.; Steels, L. Sharp transition towards shared vocabularies in multi-agent systems. J. Stat. Mech. Theory Exp. 2006, 2006, P06014. [Google Scholar] [CrossRef]
- Corominas-Murtra, B.; Fortuny, J.; Solé, R. Emergence of zipf’s law in the evolution of communication. Phys. Rev. E 2011, 83, 036115. [Google Scholar] [CrossRef]
- Luque, J.; Luque, B.; Lacasa, L. Scaling and universality in the human voice. J. R. Soc. Interface 2015, 12, 20141344. [Google Scholar] [CrossRef]
- Torre, I.G.; Luque, B.; Lacasa, L.; Luque, J.; Hernández-Fernández, A. Emergence of linguistic laws in human voice. Sci. Rep. 2017, 7, 43862. [Google Scholar] [CrossRef]
- DeGiuli, E. Random language model. Phys. Rev. Lett. 2019, 122, 128301. [Google Scholar] [CrossRef]
- Seoane, L.F.; Solé, R. Criticality in pareto optimal grammars? Entropy 2020, 22, 165. [Google Scholar] [CrossRef]
- Nowak, M.A.; Krakauer, D.C. The evolution of language. Proc. Natl. Acad. Sci. USA 1999, 96, 8028–8033. [Google Scholar] [CrossRef]
- Pagel, M. Human language as a culturally transmitted replicator. Nat. Rev. Genet. 2009, 10, 405–415. [Google Scholar] [CrossRef]
- Számadó, S.; Szathmáry, E. Selective scenarios for the emergence of natural language. Trends Ecol. Evol. 2006, 21, 555–561. [Google Scholar] [CrossRef]
- Tomasello, M. Origins of human communication; MIT Press, 2010. [Google Scholar]
- Pulvermüller, F. Words in the brain’s language. Behav. Brain Sci. 1999, 22, 253–279. [Google Scholar] [CrossRef]
- Christiansen, M.H.; Chater, N. Language as shaped by the brain. Behav. Brain Sci. 2008, 31, 489–509. [Google Scholar] [CrossRef]
- Berwick, R.C.; Friederici, A.D.; Chomsky, N.; Bolhuis, J.J. Evolution, brain, and the nature of language. Trends Cogn. Sci. 2013, 17, 89–98. [Google Scholar] [CrossRef]
- Dehaene, S.; Roumi, F.A.; Lakretz, Y.; Planton, S.; Sablé-Meyer, M. Symbols and mental programs: a hypothesis about human singularity. Trends Cogn. Sci. 2022. [Google Scholar] [CrossRef]
- Kozen, D.C. Automata and Computability; Springer, 1997. [Google Scholar]
- Prusinkiewicz, P.; Lindenmayer, A. The algorithmic beauty of plants; Springer Science & Business Media, 2012. [Google Scholar]
- Stanley, H.E. Scaling, universality, and renormalization: Three pillars of modern critical phenomena. Rev. Mod. Phys. 1999, 71, S358. [Google Scholar] [CrossRef]
- Solé, R.; Manrubia, S.C.; Luque, B.; Delgado, J.; Bascompte, J. Phase transitions and complex systems: Simple, nonlinear models capture complex systems at the edge of chaos. Complexity 1996, 1, 13–26. [Google Scholar] [CrossRef]
- Drossel, B. Biological evolution and statistical physics. Adv. Phys. 2001, 50, 209–295. [Google Scholar] [CrossRef]
- Szathmáry, E. Toward major evolutionary transitions theory 2.0. Proc. Natl. Acad. Sci. USA 2015, 112, 10104–10111. [Google Scholar] [CrossRef]
- Chomsky, N. Cartesian Linguistics: A Chapter in the History of Rationalist Thought; Cambridge University Press: New York, NY, USA; London, UK,, 1966. [Google Scholar]
- Chomsky, N. Language and problems of knowledge : the Managua lectures; MIT Press: London, UK, 1988; Old accession number 11728. [Google Scholar]
- Greenberg, J.H. Universals of Language; The M.I.T. Press, 1963. [Google Scholar]
- Deutsch, D. The beginning of infinity. Begin. Infin. Allen Lane 2011. [Google Scholar]
- De las Cuevas, G. Universality everywhere implies undecidability everywhere. FQXi Essay 2020. [Google Scholar]
- De las Cuevas, G.; Cubitt, T.S. Simple universal models capture all classical spin physics. Science 2016, 351, 1180–1183. [Google Scholar] [CrossRef]
- Gonda, T.; Reinhart, T.; Stengele, S.; De les Coves, G. A Framework for Universality in Physics, Computer Science and Beyond. arXiv 2023, arXiv:2307.06851. [Google Scholar]
- Yanofsky, N.S. A universal approach to self-referential paradoxes, incompleteness and fixed points. Bull. Symbolic Logic 2003, 9, 362–386. [Google Scholar] [CrossRef]
- Lawvere, F.W. Diagonal arguments and cartesian closed categories. Lect. Notes Math. 1969, 92, 134–145. [Google Scholar]
- Priest, G. Beyond the limits of thought; Clarendon Press: Oxford, UK, 2001. [Google Scholar]
- Batterman, R.W. The Devil in the Details: Asymptotic Reasoning in Explanation, Reduction, and Emergence; Oxford University Press: New York, NY, USA, 2002. [Google Scholar]
- De les Coves, G. Quantum Theory: Ideals, Infinities and Pluralities. In Open Systems: Physics, Metaphysics, and Methodology; Cuffaro, M., Hartmann, S., Eds.; Oxford University Press, 2024. [Google Scholar]
- Moore, A.W. The Infinite, 3rd ed.; Routledge, 2019. [Google Scholar]
- Chomsky, N. Syntactic Structures; Mouton and Co.: The Hague, The Netherlands, 1957. [Google Scholar]
- Hopcroft, J.E.; Ullman, J.D. Introduction to Automata Theory, Languages, and Computation; Addison-Wesley Publishing Company, 1979. [Google Scholar]
- Cottingham, J. Rationalism; Paladin: London, UK, 1984. [Google Scholar]
- Cottingham, J. The Rationalists; Oxford University Press: New York, NY, USA, 1988. [Google Scholar]
- Chomsky, N. Reflections on Language; Temple Smith, 1975. [Google Scholar]
- Radford, A. Syntactic theory and the acquisition of English syntax; Blackwell: Oxford, UK, 1990. [Google Scholar]
- Christiansen, M.H.; Collins, C.; Edelman, S. Language Universals; Oxford University Press, 2009. [Google Scholar]
- Francez, N.; Wintner, S. Unification Grammars; Cambridge University Press, 2011. [Google Scholar]
- Chomsky, N. Aspects of the theory of syntax; MIT Press, 1965. [Google Scholar]
- Baker, M.C. The atoms of language; Basic Books, 2001. [Google Scholar]
- Nowak, M.A. Evolutionary dynamics: exploring the equations of life; Harvard University Press, 2006. [Google Scholar]
- Anderson, S.R. Languages: A very short introduction; Oxford University Press, 2012. [Google Scholar]
- Moro, A. Impossible languages; MIT Press: 2016.
- Solé, R.V. La lógica de los monstruos; Tusquets, 2016. [Google Scholar]
- Deutsch, D. Constructor Theory. Synthese 2012, 190, 18. [Google Scholar] [CrossRef]
- Chomsky, N. The Minimalist Program; MIT Press: Cambridge, MA, USA, 1995. [Google Scholar]
- Adger, D. Core Syntax: A Minimalist Approach; Oxford University Press: Oxford, UK, 2003. [Google Scholar]
- Dehaene, S. Reading in the Brain: The Science and Evolution of a Human Invention; Viking, 2009. [Google Scholar]
- Tuson, J. Introducció a la lingüística; Aula, 2009. [Google Scholar]
- Robinson, A. Writing and script: A very short introduction; Oxford University Press, 2009. [Google Scholar]
- Yates, F. The art of memory; The Bodley Head, 1966. [Google Scholar]
- Antognazza, M.R. Leibniz: A very short Introduction; Oxford University Press: 2016.
- Sklar, L. Physics and Chance: Philosophical Issues in the Foundations of Statistical Mechanics; Cambridge University Press: New York, NY, USA, 1993. [Google Scholar]
- Zipf, G.K. Human Behaviour and the Principle of Least Effort; Addison-Wesley, 1949. [Google Scholar]
- Piantadosi, S.T. Zipf’s word frequency law in natural language: a critical review and future directions. Psychon. Bull. Rev. 2014, 21, 1112–1130. [Google Scholar] [CrossRef]
- Ferrer-i Cancho, R.; Solé, R. Two regimes in the frequency of words and the origins of complex lexicons: Zipf’s law revisited. J. Quant. Linguist. 2001, 8, 165–173. [Google Scholar] [CrossRef]
- Harremoes, P.; Topsoe, F. Maximum entropy fundamentals. Entropy 2001, 3, 191–226. [Google Scholar] [CrossRef]
- Ferrer i Cancho, R.; Solé, R.V. Least effort and the origins of scaling in human language. PNAS 2003, 10, 788. [Google Scholar] [CrossRef] [PubMed]
- Corominas-Murtra, B.; Fortuny, J.; Solé, R. Emergence of zipf’s law in the evolution of communication. Phys. Rev. E 2011, 83, 03611. [Google Scholar] [CrossRef] [PubMed]
- Gabaix, X. Zipf’s law for cities: An explanation. Q. J. Econ. 1999, 114, 739–767. [Google Scholar] [CrossRef]
- Corominas-Murtra, B.; Hanel, R.; Thurner, S. Understanding scaling through history-dependent processes with collapsing sample space. Proc. Natl. Acad. Sci. USA 2015, 112, 5348–5353. [Google Scholar] [CrossRef] [PubMed]
- Corominas-Murtra, B.; Solé, R.V. On the universality of Zipf’s Law. Phys. Rev. E 2010, 82, 011102. [Google Scholar] [CrossRef] [PubMed]
- Corominas-Murtra, B.; Seoane, L.F.; Solé, R. Zipf’s Law, unbounded complexity and open-ended evolution. J. R. Soc. Interface 2018, 15, 20180395. [Google Scholar] [CrossRef] [PubMed]
- Cover, T.M.; Thomas, J.A. Elements of information theory, 2nd ed.; Wiley, 2006. [Google Scholar]
- Solé, R.V.; Goodwin, B. Signs of Life: How complexity pervades biology; Basic Books, 2000. [Google Scholar]
- Stengele, S.; Drexel, D.; De las Cuevas, G. Classical spin Hamiltonians are context-sensitive languages. Proc. R. Soc. A 2023, 479, 20220553. [Google Scholar] [CrossRef]
- Reinhart, T.; De las Cuevas, G. The grammar of the Ising model: A new complexity hierarchy. arXiv 2022, arXiv:2208.08301. [Google Scholar]
- Hinzen, W.; Sheehan, M. The philosophy of universal grammar; Oxford University Press, 2013. [Google Scholar]
| 1 | This is the one implicitly considered in [42]. |
| 2 | For some perspectives on the role of infinities in physics, as well as the impossibility to conceive of certain pluralities as a unity, see [49]. |
| 3 | The latter meaning of ‘all-encompassing’ is a frozen metaphor, where ‘frozen metaphor’ is itself a frozen metaphor. |
| 4 | The fact that we can conceive of objects that do not admit a finite description is remarkable in itself. However, in all fairness, perhaps we can only conceive their negation. More in Appendix A. |
| 5 | The same caveat as for the Universal Grammar applies: Is it the set of actual or possible natural languages? We assume it is the actual, but the aim is to reach the possible. |
| 6 | We adopt a broad definition of lexical item as a basic unit of the lexicon of a spoken language. This includes words, morphemes or phrasal verbs, among other basic units. |
| 7 | One can only wonder: What would the world be like if such a language were invented? |
| 8 | This L is not to be confused with a formal language such as the ones in Section 3.1. |
| 9 | There is a regime shift (around rank values of ) that defines a crossover between common and infrequent words in human language, where the exponent turns to [76]. |
| 10 | The representation r is implicitly included in the definition of L. |
| 11 | Note that instead of imposing constraints on the system to calculate the maximum entropy à la Jaynes, we are saying that the maximum entropy distribution (e.g., the uniform one) is unattainable due to an internal conflict of the system. Consequently, the system experiences a non-equilibrium tension, making this problem distinct from a conventional maximum entropy problem with constraints. |
| 12 | We are neglecting the issue of tuning in music systems, a problem stemming from the mathematics of frequency ratios. Depending on the selected approach to the so-called comma, F♯ may not correspond to the same frequency as G♭, and the same discrepancy applies to other notes. As a matter of fact, this issue can be viewed as the challenge of establishing a discrete code within a continuum of frequencies, where the code must be compatible with the mathematics of frequency ratios. |
| 13 | This S is not to be confused with the set of spoken words of Section 4. |






Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).