Section I: Languages & Corpora#
The goal of Section I is to establish the hierarchy of logical relations that govern Strings, Characters, Alphabets, Words, Languages, Sentences and Corpora. As each of these entities is introduced and defined, a new level of relations will be revealed and codified. Palindromic symmetries will manifest on each level, in slightly different but related forms. Each type of symmetry will involve, in some form or another, the concept of String Inversion, to be defined shortly.
The essence of a Palindrome lies in binding together the syntactical symmetries at every linguistic layer into a semantic whole. Indeed, it will be seen the symmetrical structure required by Palindromes in turn requires these linguistic layers to have explicit relations and specific synactical properties, regardless of their semantic interpretation. These symmetries, in turn, guide the formal development in seeking the machinery capable of expressing them.
A Character will form the first layer of the hierarchy. Characters will be regarded as a type of String; they will be seen as “atoms” or “units” of Strings. The elementary operation of Concatenation will be used to build up Strings of greater complexity starting from Character. This much agrees with formal language theory and context free grammar constructions, although the details will differ.
Where the current formalization will start to diverge is at the next level of the semantic hiearchy. A Word will be considered another type of String. Colloquially, a Word can be understood as a String with semantic content, but as this formal system will establish, a Word is not distinguished solely by its semantic content, and in fact, it’s semantic interpretation is dependent on prior syntactical conditions that differentiate Words from Strings; This fact is often not made explicit in formal treatmeants of language, but the study of palindromes makes this distinction critical. In the current work, Words will be treated as linguistic entities that are distinguishable by their non-zero length and lack of Delimiters.
Finally, a Sentence will also be considered a type of String. A Sentence will be regarded as a sequence of Words that have been delimited together and selected as bearing semantic content. A specialization of concatenation in the form of Limitations will be introduced to describe this structure of Sentences. This operation will in turn define a subdomain within the set of all finite Strings that is constructed from the permutation of delimited Words. This set will allow proofs to leverage induction to establish results over the entire set of all Sentences.
Section I will elaborate the necessary syntactic conditions for a String to be distinguished as a formal Word or a formal Sentence, without taking into account the semantic content that is assigned to either entity through colloquial use. In other words, this section seeks to formally disentangle the syntactical functions of Words, Sentences and Strings.
Section I.I: Formalization#
Note
All of the terminology presented in this section will be elaborated upon in the subsequent sections.
Conventions#
General conventions adopted throughout the course of this work are given below.
\(N_n\) will represent the set of natural numbers starting at 1 and ending at n,
The cardinality of a set \(A\) will be denoted \(\lvert A \rvert\)
∎ will be used to denote the ending of all definitions, examples and proofs.
The terms “set” and “class” are used interchangeably.
Constants#
Terms#
Important
The exact meaning of these symbols should be attended with care. \(\mathfrak{a}, \mathfrak{b}, \mathfrak{c}, ...\) represent Characters of the Alphabet and thus are all unique, each one representing a different linguistic element. When Character symbols are used with subscripts, \(\mathfrak{a}_1, \mathfrak{a}_2, \mathfrak{a}_3, ...\), they are being referenced in their capacity to be ordered within a String. With this notation, it is not necessarily implied \(\mathfrak{a}_1\) and \(\mathfrak{a}_2\) are unequal Character-wise, but that they are differentiated only by their relative order in a String.
Likewise, when Character Variables are used with subscripts, it is meant to refer to the capacity of a Character Variable to be indeterminate at a determinate position within a String.
Moreover, the range of a Character Variable is understood to be the Alphabet \(\Sigma\) from which it is being drawn.
Relations#
Sets#
Section I.II: Strings#
All non-Empty Characters belong to the Alphabet,
Important
The Delimiter belongs to the Alphabet.
The aggregate of the Alphabet and the Empty Character is referred to as the Total Alphabet and is denoted,
A Character is the basic unit of a String. In order to construct a String or set of Strings, an Alphabet must be selected. A String is regarded as a linguistic artifact or inscription that is defined entirely by its Characters and their ordering. In order to construct more complicated Strings through the sequencing of Characters, the operation of Concatenation must be defined.
Concatenation#
Important
Many of the results of regular expressions and automata theory are taken as given and will not be proved, such as the associativity of concatenation (i.e. \((ut)v = u(vt)\)), the closure of concatenation over \(S\) (i.e., concatenating two Strings will always yield a String), etc.
Example Let \(s_1 = \mathfrak{abc}\) and \(s_2 = \mathfrak{def}\). The concatenation of these two Strings \({s_1}{s_2}\) is written,
Using the Inductive Clause, this concatenation can be grouped into simpler concatenations as follows,
By Character Comprehension Axiom, all Characters are Strings and concatenation is closed under \(S\), therefore, \(\mathfrak{ef} \in S\). As each nested concatenation is evaluated, the Induction clause in Concatenation ensures the next level of concatenation is a String.
As a result, \({s_1}{s_2} = \mathfrak{abcdef}\) and \({s_1}{s_2} \in S\)
∎
String Length#
The length of a String is defined as its number of non-Empty Characters.
Example Let \(s_1 = \mathfrak{abc}\varepsilon\mathfrak{def}\). Using Concatenation, this can be grouped as \(s_1 = (\mathfrak{abc}\varepsilon\mathfrak{de})(\mathfrak{f})\).
Applying String Length to \(\mathfrak{f}\) where \(u = \mathfrak{f}\) and \(v = \varepsilon\),
Note
This same logic generalizes to all Alphabetic Characters,
Applying String Length with \(u = \mathfrak{abc}\varepsilon\mathfrak{de}\) and \(v = \mathfrak{f}\),
The first term on the righthand side can be evaluated by applying String Length with \(u = \mathfrak{abc}\varepsilon\mathfrak{d}\) and \(v = \mathfrak{e}\),
Continuing in this fashion, the result is calculated,
∎
The definition of String length allows an important shorthand to be defined. This notation introduces nothing new into the system, but significantly improves the readability of proofs.
Note
The notation \(s[i]\) is borrowed directly from string slicing in computer science.
The following example shows how the definition of Character indexing “skips” over the physical index of Empty Characters and assigns a logical index to any non-Empty Characters in a String.
Example Let \(s_1 = \mathfrak{ab}\varepsilon\mathfrak{c}\). By String Length, \(l(s_1) = 3\).
Consider \(s_1[3]\). Apply the definition of Character Indices with \(u_1 =\mathfrak{ab}\varepsilon\) and \(v_1 = \mathfrak{c}\). \(i = l(s_1)\) and \(v_1 \neq \varepsilon\), therefore, by the Induction clause, \(s[3] = \mathfrak{c}\).
Consider \(s_1[2]\). Apply the definition of Character Indices with \(u_1 =\mathfrak{ab}\varepsilon\) and \(v_1 = \mathfrak{c}\). At this step, \(v_1 \neq \varepsilon\) but \(i \neq l(s_1)\), so the \(s_1[i] = u_1[i]\). Note \(l(u_1) = 2\).
To find \(u_1[i]\), let \(u_1 = {u_2}{v_2}\) where \(u_2 = \mathfrak{ab}\) and \(v_2 = \varepsilon\). At this step, \(i = l(u_1)\), but \(v_2 = \varepsilon\), therefore \(u_1[i] = u_2[i]\). Note \(l(u_2) = 2\).
To find \(u_2[i]\), let \(u_2 = {u_3}{v_3}\) where \(u_3 = \mathfrak{a}\) and \(v_3 = \mathfrak{b}\). At this step, \(i = l(u_2)\) and \(v_3 \neq \varepsilon\), therefore \(u_2[i] = v_3 = \mathfrak{b}\).
From this, it follows, \(s_1[2] = u_1[2] = u_2[2] = v_3 = \mathfrak{b}\).
∎
The first theorem confirms the well known result that String Length sums over concatenation within the formal system.
Note
The proof of Theorem 1.2.1 is a standard proof included in virtually every textbook on the subject of regular expressions, automata, formal language, etc. For this reason, the proof has been omitted from the main body of the work. A proof by induction is presented in Appendix, Omitted Proofs for completionists.
String Equality#
Two Strings are said to be equal if they have the same length and their corresponding Alphabetic Characters (\(\iota \in \Sigma\)) are equal.
Example Let \(s_1 = \mathfrak{ab}\) and \(s_2 = \mathfrak{a}\varepsilon\mathfrak{b}\). Apply String Length,
Now, \(N_n = { 1, 2 }\). Using Character Indices,
Therefore, \(\forall i \in N_n: s_1[i] = s_2[1]\). It follows from these facts and application of String Equality,
∎
Containment#
The notion of containment is the formal explication of the colloquial relation of “being a substring of”.
Example Let \(s_1 = \mathfrak{abcdef}\). Then the truth of the following propositions can be verified using the given values of \(w_1\) and \(w_2\) in the definition of Containment.
\(\mathfrak{ab} \subset_s s_1\), where \(w_1 = \varepsilon\) and \(w_2 = \mathfrak{cdef}\).
\(\mathfrak{cde} \subset_s s_1\), where \(w_1 = \mathfrak{ab}\) and \(w_2 = \mathfrak{f}\).
\(\neg (\mathfrak{g} \subset_s s_1)\), for any \(w_1, w_2\)
∎
Note
This is another standard theorem in formal theory of strings. See Appendix, Omitted Proofs for a proof.
String Inversion#
Note
Many of the theorems in this section are also standard results in the formal theory of strings, but the mechanics of the proofs are materially different from their usual constructions, due to the definition of String Inversion as the reversal of logical Character indices within the String, rather than an induction on a recursive definition of reversal (which canonical proofs found in textbooks typically follow). Since these proofs illustrate the simplification affected by Character indices, they have been retained in the main body of the work.
Example Let \(s_1 = \mathfrak{abc}\). Let \(s_2 = {s_1}^{-1}\). The inverse can be constructed through its Character Indices by applying String Inversion,
Concatenating the results,
∎
Proof Let \(s \in S\). Let \(t = s^{-1}\). Let \(n = l(s)\). From String Inversion,
Let \(u = t^{-1}\). Applying String Inversion again,
Plugging \(i = n - j + 1\) into (2) and substituting into (4),
Moreover, from (1) and (3), it follows,
By the String Equality, (5) and (6) together imply,
Therefore,
∎
Proof Let \(s,t \in S\). Let \(u = st\). Let \(m = l(s)\) and \(n = l(t)\). Let \(u = st\). By Theorem 1.2.1,
Let \(v = u^{-1} = (st)^{-1}\). Let \(w = (t)^{-1}(s)^{-1}\). By repeated application of String Inversion,
Using these results and applying Theorem 1.2.1 to \(w\),
From (1) and (2), it follows,
Let \(i \in N_{m+n}\).
Case 1: \(i \leq i \leq n\)
By String Inversion,
By assumption \(i \leq n\) or \(n - i \geq 0\), therefore,
Increasing the LHS of this inequality does not affect the truth of its assertion,
From this, \(u = st\) and \(l(s) = m\), it follows that \(u[m + n - i + 1]\) is an index in \(t\),
Consider \(w[i]\). Since \(l((t)^{-1}) = n\) and \(i \leq n\), it follows that \(w[i] = (t^{-1})[i]\). By String Inversion,
Combining (4) and (5),
Applying String Equality, (3) and (6) imply,
Case 2: \(n + 1 \leq i \leq m + n\)
By String Inversion,
v[i] = u[m + n - i + 1]
By assumption \(i \geq n + 1\) or \(n - i + 1 \leq 0\), therefore,
From this, \(u = st\) and \(l(s) = m\), it follows that \(u[m + n - i + 1]\) is an index in \(s\),
Consider \(w[i]\). Since \(l((t)^{-1}) = n\) and \(i \geq n\), it follows that \(w[i] = (s^{-1})[i - n]\). By String Inversion,
Combining (7) and (8),
Applying String Equality, (3) and (6) imply,
In both cases, the theorem is proved. Summarizing and generalizing,
∎
Proof Let \(s,t \in S\).
(\(\rightarrow\)) Assume \(t \subset_s s\). Then by Containment, there exists \(w_1, w_2 \in S\) such that,
Consider \(s^{-1}\). Applying Theorem 1.2.4 twice, this becomes,
Therefore, there exists \(u_1 = {w_2}^{-1}\) and \(u_2 = {w_1}^{-1}\) such that \(s^{-1} = (u_1)(t^{-1})(u_2)\) and by the definition of Containment,
(\(\leftarrow\)) The proof is identical to (\(\rightarrow\)).
Therefore,
∎
Section I.III: Words#
Important
To reiterate the introduction to this section, the current formal system does not seek to describe a generative grammar. Its theorems cannot be used as schema for generating grammatical sentences. The intent of this analysis is to treat Words as interpretted constructs embedded in a syntactical structure that is independent of their specific interpretations.
A Word is a type of String constructed through concatenation that has been assigned by semantic content. A Language is the aggregate of all Words.
Or equivalently,
Word Classes#
Note
\(R\) may be defined equivalently through set builder notation,
Example The following table lists some reflective English words.
Word |
---|
mom |
dad |
noon |
racecar |
madam |
level |
civic |
∎
Important
A Word is invertible if and only if its inverse belongs to the Language.
Example The following table lists some English words and their inverses (where applicable).
Word |
Inverse |
---|---|
time |
emit |
saw |
was |
raw |
war |
dog |
god |
pool |
loop |
cat |
x |
you |
x |
help |
x |
door |
x |
book |
x |
∎
Note
Invertible Words are often called semiordnilaps in other fields of study.
Proof Let \(\alpha \in L\).
(\(\rightarrow\)) Assume \(\alpha \in I\). By the definition of invertible Words,
By Theorem 1.2.3,
Therefore, by assumption,
By the definition of invertible Words,
(\(\leftarrow\)) Assume \({\alpha}^{-1} \in L\) such that \({\alpha}^{-1} \in I\). By the definition of invertible Words,
By Theorem 1.2.3,
Since \({\alpha}^{-1} \in L\) by assumption, it follows immediately from the definition of invertible Words,
Summarizing and generalizing,
∎
Proof Let \(\alpha in R\) and \(l(\alpha) = n\). By the definition of Reflective Words,
Since \(\alpha \in L\) by assumption, it follows \(\alpha in I\). In other words,
But this is exactly the definition of the subset relation in set theory, therefore,
∎
Limitations#
Note
A Limitation, though notationally complex, can be understood as shorthand for the iterated concatenation of words and Delimiters. is the presence of the Delimiter in the Induction clause. In other words, a Limitation inserts Delimiters inbetween each Word in the Lexicon over which the index is ranging.
Example Let \(L = L_english\). Consider calculating the Limitation of the following Phrase,
Apply the Basis clause Limitations ,
The Limitation can then be built up recursively using the Induction clause,
So the Limitation of the Phrase is shown to be,
Important
The result of a Limitation is a String. Since a Limitation is shorthand for alternating concatenation of Characters and Delimiters, the closure of Limitations over \(S\) is guaranteed by the closure of concatenation over \(S\)
∎
This subsection closes with a definition that will be used to quantify a theorem regarding Word Length.
Warning
The type of each set defined in this section should be carefully analyzed.
A Phrase is an ordered set of Words.
A Lexicon is the set of all Phrases of a fixed Word Length.
A Dialect is the set of Strings formed by delimiting every Phrase in every Lexicon of a Language.
Example Let \(L = \{ \text{hakuna}, \text{matata} \}\). Then, the first few Lexicons are given below,
The Dialect is the union of all delimited Phrases in all Lexicons of the Language,
∎
Canonization#
Canonization is a function defined over \(s \in S\) that produces the canonical form of a String by removing all instances of the Empty Character from it.
Example Let \(s_1 = (\mathfrak{a})(\varepsilon)(\mathfrak{b})\).
Let \(u_1 = (\mathfrak{a})(\varepsilon)\) and \(v_1 = \mathfrak{b}\). Note \(v_1 \in \Sigma\) and \(s_1 = (u_1)(v_1)\). By the Induction clause of Canonization,
Let \(u_2 = \mathfrak{a}\) and \(v_2 = \varepsilon\). Note \(u_1 = (u_2)(v_2)\). By the Induction clause,
Let \(u_3 = (\varepsilon)\) and \(v_3 = \mathfrak{a}\). Note \(v_3 \in \Sigma\) and \(u_2 = (u_3)(v_3)\). By the Induction clause,
By the Basis clause,
Putting the recursion together,
By the Basis clause of Concatenation, this becomes,
∎
Canonization provides a method of “cleaning” \(S\) of troublesome Strings, such as \(\mathfrak{a}\varepsilon\mathkfra{b}\) that prevent the assertion of uniqueness within the semantic domains of \(L\) and \(C\). The Canon provides a domain within \(S\) where the uniqueness of the Limitation can be established.
Proof Let \(n \in \mathbb{N}\) and \(p \in L_n\) such that,
The proof will proceed by induction on \(n\).
Basis: Assume \(n = 1\). By Basis clause of Limitations,
Induction: Assume for \(k \geq 1\), these exists a unique String \(s_k\) such that,
By Induction clause of Limitations,
By inductive hypothesis,
Therefore, by induction,
∎
Proof Let \(s \in S\). The proof proceeds by induction on \(s\).
Basis Let \(s = \varepsilon\). By the definition Canonization,
Let \(t = \pi(\varepsilon)\). Consider,
Induction Assume \(\pi(\pi(t)) = \pi(t)\) for some \(t \in S\). Let \(s = (t)(\iota)\) where \(\iota \in \Sigma_e\). Either \(\iota = \varepsilon\) or \(\iota \neq \varepsilon\).
Case I: \(\iota = \varepsilon\)
By the Induction clause of Canonization,
By the Basis clause of Concatenation,
Therefore, by inductive hypothesis,
Case II \(iota \neq \varepsilon\)
By the Induction clause of Canonization,
Now the String \(u = \pi(t)\) belongs to the Canon, \(u \in \mathbb{S}\), and must therefore be a String free of \(\varepsilon\). Likewise, \(\iota \neq \varepsilon\) by assumption. Therefore, \(u\iota\) is also a String free of \(\varepsilon\). From this and the definition of Canonization, it follows \(\pi(u\iota) = u\iota\),
Consider,
Therefore,
And the induction is established. Summarizing and generalizing,
∎
Proof Let \(s \in S\).
(\(\leftarrow\)) Assume \(s = \pi(s)\). By the definition of Canon, any String that is the result of Canonization belongs to the Canon, therefore \(s \in \mathbb{S}\).
(\(\rightarrow\)) Assume \(s \in \mathbb{S}\). By the definition of Canon, there must exist a \(t \in S\) such that \(\pi(t) = s\). Consider \(\pi(\pi(t))\). By Theorem 1.3.4,
Substituting \(\pi(t) = s\),
Therefore, the equivalence is established.
∎
Proof Let \(t \in S\). The proof will proceed by induction on \(t\).
Basis: Let \(s \in \mathbb{S}\). Let \(t = \varepsilon\). By the Basis clause of Canonization and the definition of \(Canon <palindromics-definition-1-3-7>\), \(t \in \mathbb{S}\)
Consider \(st\). By the Basis clause of \(Concatenation <palindromics-definition-1-2-1>\), \(st = s\varepsilon = s\). But \(s \in mathbb{S}\) by assumption, thus \(st \in mathbb{S}\).
Induction. Assume \(u \in \mathbb{S}\) such that \(su \in \mathbb{S}\). By Theorem 1.3.5,
Let \(t = (u)(\iota)\) where \(\iota \in \Sigma\). Consider \(st\),
Where the last equality follows from the associativity of concatenation. By inductive hypothesis, \(su \in \mathbb{S}\). Moreover, \(\iota \in \mathbb{S}\) since \(\pi(\iota) = \iota\). Therefore, by definition of \(Canonization <palindromics-definition-1-3-7>\)
Substituting in (1) and (2)
By Theorem 1.3.5,
Thus, the induction is complete. Summarizing and generalizing,
Section I.IV: Sentences#
A Sentence is a Limitation of Words over a Phrase in the Language’s Lexicon for any value of \(n \geq 1\).
Warning
This statement should not be interpretted as a schema for generating grammatical sentences. In general, Limitations are not grammatical. However, all grammatical sentences are Limitations.
In other words, this statement should be interpretted as a necessary syntactic pre-condition a Sentence must satisfy before it may be assigned semantic content.
A Corpus is the aggregate of all Sentences.
Note
The value of \(n\) in the preceding equation will be further specified after several definitions and theorems. It will be shown to be directly and necessarily related to the Word structure of \(\zeta\).
The full semantic hierarchy has now been formalized. The hierarchy is summarized as follows,
Strings: \(\iota, \alpha, \zeta\)
Sets: \(\Sigma, L, C\)
Character Membership: \(\iota \in \Sigma\)
Word Membership: \(\alpha \in L\)
Sentence Membership: \(\zeta \in C\)
These observations can be rendered into English,
All Characters, Words and Sentences are Strings.
The Alphabet, Languages and Corpus are sets of Strings.
All non-Empty Characters belong to an Alphabet.
All Words belong to the Language.
All Sentences belong to the Corpus.
Word Length#
Important
The Induction clause of Word Length relies on the Discovery Axiom and the Measureable Axiom to ensure for any Strings \(u \in L\), \(\neg(\sigma \subset_s u)\) and \(u \neq \varepsilon\).
Important
While Word Length will be primarily used on \(\zeta \in C\), it is important to note the definition is defined over all \(s \in S\). In other words, Word Length is a property of Strings, as can be seen in the example, “blargafaful buttons”.
Example Let \(ᚠ = \text{truth is beauty}\).
Let \(u_1 = \text{truth}\) and \(v_1 = \text{is beauty}\). Then \(u_1 \in L_{\text{english}}\) and \(ᚠ = (u_1)(\sigma)(v_1)\). Apply the Induction clause of Word Length,
Let \(u_2 = \text{is}\) and \(v_2 = \text{beauty}\).
Important
A selection of \(u_2 = \text{i}\) or \(u_2 = \text{is be}\) would not satisfy the condition \(s = {u}{\sigma}{v}\) in the Induction clause, which requires \(u\) and \(v\) to be delimited with \(\sigma\).
Then \(u_2 \in L_{\text{english}}\) and \(v_1 = (u_2)(\sigma)(v_2)\). Apply the Induction clause of Word Length,
Finally, note \(v_2 \in L_{\text{english}}\) and apply the Basis clause to \(v_2\),
Putting the recursion together,
∎
Example Let \(ᚠ = \text{palindromes vorpal semiordinlap}\)
Let \(u_1 = \text{palindromes}\) and \(v_1 = \text{vorpal semiordinlap}\). Then \(u_1 \in L_{\text{english}}\) and \(ᚠ = (u_1)(\sigma)(v_1)\). Apply the Induction clause of Word Length,
Let \(u_2 = \text{vorpal}\) and \(\v_2 = \text{semiordinlap}\). Then \(u_2 \notin L_{\text{english}}\) and \(v_1 = (u_2)(\sigma)(v_2)\). Apply the Induction clause of Word Length,
Finally, note \(v_2 \in L_{\text{english}}\) and apply the Basis clause to \(v_2\),
Putting the recursion together,
∎
Important
As these examples demonstrate, the Word Length of a String is always relative to a given a Language. A subscript will be used to denote whether a Word Length is relative to a particular language,
Whereas,
Example Let \(L = L_{\text{english}}\). Let \(ᚠ = \text{observe how system into system runs}\). Consider \(ᚠ[[3]]\).
Let \(u_1 = \text{observe}\) and \(v_1 = \text{how system into system runs}\). Then \(ᚠ = (u_1)(\sigma)(v_1)\), \(u_1 \in L\) and \(3 > 1\). Therefore, by the Induction clause of Word Indices,
At the next step, let \(u_2 = \text{how}\) and \(v_2 = \text{system into system runs}\). Then \(v_1 = (u_2)(\sigma)(v_2)\), \(u_2 \in L\) and \(2 > 1\),
At the next step, let \(u_3 = \text{system}\) and \(v_3 = \text{into system runs}\). Then \(v_2 = (u_3)(\sigma)(v_3)\), \(u_3 \in L\) but \(1 = 1\), therefore,
∎
Example Let \(ᚠ = \text{the gobberwarts with my blurglecruncheon}\). Consider \(ᚠ[2]\).
Let \(u_1 = \text{"the"}\) and \(v_1 = \text{gobberwarts with my blurglecruncheon}\). Then \(ᚠ = (u_1)(\sigma)(v_1)\), \(u_1 \in L\) and \(2 > 1\). Therefore, by the Induction clause of Word Indices,
At the next step, let \(u_2 = \text{gobberwarts}\) and \(v_2 = \text{with my blurglecruncheon}\). Then \(v_1 = (u_2)(\sigma)(v_2)\) but \(u_2 \notin L\) and \(1 = 1\), so by the first condition of the Induction clause,
At the next step, let \(u_3 = \text{with}\) and \(v_3 = \text{my blurglecruncheon}\). Then \(v_2 = (u_3)(\sigma)(v_3)\), \(u_3 \in L\) and \(1 = 1\). So, the second condition of the Induction clause,
∎
The next theorems will not be required for the final postulates, but they are given to indicate the type of results that may be established regarding the concept of Word Length. For the curious reader, the details can be found in Appendix I.II: Omitted Proofs.
Note
Theorem 1.4.1 and Theorem 1.4.2 demonstrate Word Length is fundamentally different than String Length with respect to the operation of concatenation. In Theorem 1.2.1, it was shown String Length sums over concatenation. Theorem 1.4.1 shows the corresponding property is not necessarily true for Word Length. This is an artifact of the potential destruction of semantic content that may occur upon concatenation.
The edge case of compound Words (e.g. daylight) makes the proof Theorem 1.4.2 particularly interesting.
Sentence Axioms#
The following theorem is proved in Appendix I.II: Omitted Proofs, as it is not required for the results in Section III. This theorem demonstrates the relationship between a Limitation and Word Length that was pointed out in the introduction of this subsection.
Note
The next theorem can be seen as a specialiation of Theorem 1.2.4 for the subdomain of the Corpus.
Proof Let \(\zeta \in C\). Let \(n = \Lambda(\zeta)\). Let \(s\),
Consider \(s^{-1}\),
From String Inversion and the fact \(l(\varsigma) = 1\), it follows \(\sigma^{-1} = \sigma\). Using this fact, the application of Theorem 1.2.4 \(n\) times yields,
Reindex the terms on the RHS to match Limitation with \(j = n - i + 1\). Then, as \(i\) goes from \(1 \to n\), \(j\) goes \(n \to 1\) and visa versa,
Combining (1) and (2) and generalizing,
∎
Proof Let \(s, t \in D\). That is, assume, for some \(n, m \in \mathbb{N}\),
where \(n = \Lambda(s)\) and \(m = \Lambda(t)\).
The proof proceeds by induction on \(n\).
Basis: Assume \(n = 1\).
Then, by the Basis clause of Limitations, \(s = \alpha\) for some \(\alpha \in L\). By the Discovery Axiom, \(\neg(\sigma \subset_s \alpha)\).
Consider \(u = (\alpha)(\sigma)(t)\). By the Basis clause of Word Length,
Induction Assume for any \(u \in D\) with \(\Lambda(u) = n\),
Let \(s \in D\) such that \(\Lambda(s) = n + 1\). By the Induction clause of the Dialects and Limitations,
By the Induction clause of Word Length,
From this and \(\Lambda(s) = n + 1\), it is concluded \(\Lambda(v) = n\) and therefore satisfies the inductive hypothesis.
Consider \(\Lambda((s)(\sigma)(t))\).
But from (1), this reduces to,
Therefore, putting everything together, the Induction is complete,
Summarizing and generalizing,
∎
Important
Theorem 1.4.5 only applies to Strings quantified over the Dialect. If the theorem were quantified over the Corpus, i.e. semantic Sentences, then the inductive hypothesis would fail at the step where the induced String is decomposed,
To see this, note that when a Sentence has it’s first Word partitioned from it, there is no guarantee the resultant will also be a semantic Sentence, e.g. “we are the stuffed men” is a Sentence, but “are the stuffed men” is not a Sentence. Therefore, the theorem must be induced over the Dialect.
This may seem a strong restriction, but as the next two theorems establish, this result still applies to the Corpus.
Proof Let \(\zeta \in C\). By definition of a Sentence,
By the definition of a Dialect, \(\zeta \in D\).
Therefore, \(\zeta \in C \implies \zeta \in D\). This is exactly the definition of a subset,
∎
Proof Let \(\zeta, \xi \in C\).
By Theorem 1.4.6, \(C \subseteq D\). By definition of subsets,
Therefore, by Theorem 1.4.5,
forall zeta, xi in C: Lamdbda((zeta)(sigma)(xi)) = Lambda(zeta) + Lambda(xi)
∎
Proof Let \(\zeta in C\). By Theorem 1.4.3,
By the Word Comprehension Axiom and Canonization Axiom,
By the definition of Canonization,
By the definition of Limitation, \(\Pi\) produces Strings through Concatenation. By Theorem 1.3.4, the Canon is closed over Concatenation. From this, it must be the case \(\zeta \in \mathbb{S}\). Therefore,
This is exactly the definition of subsets,
∎
Sentence Classes#
Proof Let \(\zeta in C\).
(\(\rightarrow\)) Assume \(\zeta \in K\). By the definition of Invertible Sentences,
By :ref:` <palindromics-theorem-1-2-3>`,
By assumption, \(\zeta \in C\), therefore, by the definition of Invertible Sentences,
(\(\leftarrow\)) Assume \({\zeta}^{-1} \in K\), which implies \({\zeta}^{-1} \in C\). By assumption \(\zeta \in C\). Therefore, definition of Invertible Sentences,
Summarizing and generalizing,
∎
Proof Let \(\zeta \in K\). By the definition of Invertible Sentences,
By Theorem 1.4.4, this can be written,
where,
By the Word Comprehension Axiom,
From this, it can be concluded every \({\zeta[[i]]}^{-1}\) in \(p\) must belong to \(L\), and each of those Words has an inverse that is also in \(L\).
By the definition of Invertible Words, the inverse of a Word can only belong to a Language if and only if the Word is invertible.
Therefore,
By Theorem 1.3.1,
Generalizing,
∎
Proof Let \(\zeta \in K\), let \(n = \Lambda(\zeta)\) and let \(i \in N_n\).
By Theorem 1.4.6 and assumption,
By Theorem 1.3.1,
Consider,
By Theorem 1.4.4,
And by definition of Sentences and Limitations,
Therefore,
By the \(Theoreom 1.4.8 <palindromics-theorem-1-4-8>\), \(C \subset \mathbb{S}\). By Theorem 1.3.3, Limitations are unique over the Canon, thus the only way two Limitations that belong to the Corpus can be equal to \(\zeta^{-1}\) is when,
Summarizing and generalizing,
Section I.V: Summary#
The analysis requires one more piece of formal machinery before it can codify the phenomenon of palindromes. However, even without the later results, Theorem 1.4.10 and Theorem 1.4.11 are particularly compelling results that demonstrate the efficacy of the current formal system and its ability to generate novel, if intuitively obvious, theorems.
The deductive path from Theorem 1.4.10 to Theorem 1.4.11 follows a “propagation of inversion” up the semantic hierarchy, from Characters to Words to Sentences.
First, String Inversion was defined as a operation performed on the Characters within a String,
Where \(t\) is the inverse of \(s\), \(t^{-1} = s\). This in turn defined an equivalence class over involutive Words in Reflective Words,
alpha in R equiv alpha = {alpha}^{-1}
Moreover, it created a semi-group in Invertible Words,
This inversion makes its way to the top layer of the semantic hierarchy with Invertible Sentences,
The class \(K\) then imposes a condition on all Sentences that belong to it, namely that its Words must be also invertible,
The inversion then “propagates” up a level in the semantic hierarchy and results in a directly analogous condition on the Word-level to the Character-level symmetry,
Important
The direction of implication in Theorem 1.4.10 and Theorem 1.4.11 is unidirectional. In other words, while invertibility implies the previous two equations, invertibility cannot be concluded on the basis of the previous two equations. This is an artifact of the formal system’s inability to formalize the grammar of Sentences.