Formal semantics of language and the Richard-Berry paradox
The classical logical antinomy known as Richard-Berry paradox is combined with plausible assumptions about the size i.e. the descriptional complexity of Turing machines formalizing certain sentences, to show that formalization of language leads to contradiction.
š” Research Summary
The paper revisits the classic RichardāBerry paradox ā the definition of āthe smallest natural number that cannot be defined in fewer than twenty wordsā ā and recasts it in the language of algorithmic information theory. The author first sketches the historical development of formal semantics, from Frege and Russell through Tarski, Carnap and Montague, and notes its modern relevance to computational linguistics and AI.
The core of the argument is a twoāstep construction. First, the author defines the formal complexity of an English text as the length (in bits) of the smallest Turingāmachine program that computes exactly that text. If a text admits several formalizations, the one with minimal program size is taken as the reference machine. This notion mirrors KolmogorovāChaitin complexity.
Two plausible assumptions are then introduced:
-
Unboundedness ā for every integer n there exists an English sentence whose formal complexity exceeds n. This reflects the intuition that natural language can express arbitrarily intricate ideas.
-
Logarithmic overhead ā consider the family of sentences
t(n) : āthe first text whose formal complexity is not less than n.ā
The description of t(n) consists of a fixed part (independent of n) plus a representation of the integer n. Since an integer can be encoded in ālogāāÆnā bits, the formal complexity of t(n) exceeds that of t(20) by at most a quantity proportional to logāÆn.
With these definitions, the author reproduces the paradoxical reasoning. Suppose a āfirst textā satisfying (2) exists. Two subācases arise:
-
(a) If the formal complexity of t(20) is less than 20, then t(20) itself provides a definition of āthe first text ā¦ā, contradicting the very statement that its complexity is at least 20.
-
(b) If the formal complexity of t(20) is kāÆā„āÆ20, the logarithmicāoverhead assumption guarantees a sufficiently large KāÆ>āÆk such that the formal complexity of t(K) is <āÆK. Consequently, t(K) would be a definition of āthe first text whose complexity is not less than Kā, again violating its own definition.
If, on the other hand, no such āfirst textā exists, then every English sentence would have formal complexity <āÆ20, which directly contradicts the unboundedness assumption.
Thus, ordering English sentences by the size of their minimal Turingāmachine descriptions and then referring to āthe first sentence whose description is at least n words longā reproduces the selfāreferential inconsistency of the RichardāBerry paradox. The conclusion is that a complete computational formalization of natural languageāi.e., a mapping that assigns to every sentence a unique, minimal Turingāmachine descriptionāis logically impossible. The paper therefore highlights a fundamental limitation of formal semantics: the very act of trying to capture all of natural language within a fixed formal system inevitably leads to paradox.
Comments & Academic Discussion
Loading comments...
Leave a Comment