The Church-Turing thesis gives a universally applicable definition of algorithm, but we don’t have an equivalently comprehensive definition of information
To start with an example,
Intuitively, contains less information than because it’s more repetitive
Our idea is that the length of an object’s shortest possible description is a measure of its information
We restrict our attention to binary strings and binary descriptions, and then look at what kinds of descriptions we may allow
Definition: The minimal description of a binary string is the lexicographically first shortest string where TM on input halts with on its tape
The descriptive complexity of is
This definition amounts to measuring the possible compression of
Theorem:
We can always use the identity Turing machine to create a description of
Theorem:
We can construct a Turing machine “On input , run on until it halts and produces , and then output ”
A description of is then , and its length is
Theorem: , i.e. concatenating strings leads to a greater bound than
To prove this theorem, we must show that for any there is some and where
The book leaves this as an exercise :/
Definition: Consider a description language to be any computable function and define the minimal description of with respect to , written , as the first string where in standard string order, with
Theorem: For any description language , a fixed constant exists where , i.e. our Turing machine definition of descriptive complexity is reasonably optimal
This seems impressive but the proof is trivial, since we can define “On input : Output ” and its length is at most greater than
Definition: is c-compressible if , otherwise is incompressible by c, and if then is incompressible
Theorem: Incompressible strings of every length exist
The number of binary strings of length is and the number of descriptions of length less than is , so there is at least one incompressible string of length
Corollary: At least strings of length are incompressible by
This follows from the same logic
Incompressible strings have many properties we’d expect in randomly chosen strings, such as a roughly equal number of 0s and 1s and a longest run length of about
We can prove something even more general
A property of strings is a function that maps strings to and a property holds for almost all strings if the fraction of strings of length on which it is approaches 0 as grows
Theorem: Let be a computable property that holds for almost all strings, then for any , the property is on only finite many strings that are incompressible by
“On input , find the th string where and output ”
For any such string , let be the position or index of on a list of all strings that fail to have property , in which case is a short description of with length
For any , select such that at most a fraction of strings of length or less fail to satisfy
Let be a string of length that fails to have property
, so
Thus every sufficiently long that fails to satisfy by virtue is compressible by and our theorem follows
Unfortunately, this is all confined to theory because is not computable, no algorithm can decide in general if a string is incompressible, and there is not even an infinite subset of Turing-recognizable incompressible strings
Theorem: For some constant , for every , the minimal description is incompressible by
To prove this, consider “On , run on and reject if the output is not of form . Run on and halt with its output on the tape.”
Let and suppose is -compressible for some string
This means
But then is a description of with length , which is a contradiction