The Church-Turing thesis gives a universally applicable definition of algorithm, but we don’t have an equivalently comprehensive definition of information

To start with an example,

Intuitively, contains less information than because it’s more repetitive

Our idea is that the length of an object’s shortest possible description is a measure of its information

We restrict our attention to binary strings and binary descriptions, and then look at what kinds of descriptions we may allow

Definition: The minimal description of a binary string is the lexicographically first shortest string where TM on input halts with on its tape

The descriptive complexity of is

This definition amounts to measuring the possible compression of

Theorem:
We can always use the identity Turing machine to create a description of

Theorem:

We can construct a Turing machine “On input , run on until it halts and produces , and then output

A description of is then , and its length is

Theorem: , i.e. concatenating strings leads to a greater bound than

To prove this theorem, we must show that for any there is some and where

The book leaves this as an exercise :/

Definition: Consider a description language to be any computable function and define the minimal description of with respect to , written , as the first string where in standard string order, with

Theorem: For any description language , a fixed constant exists where , i.e. our Turing machine definition of descriptive complexity is reasonably optimal

This seems impressive but the proof is trivial, since we can define “On input : Output ” and its length is at most greater than

Definition: is c-compressible if , otherwise is incompressible by c, and if then is incompressible

Theorem: Incompressible strings of every length exist

The number of binary strings of length is and the number of descriptions of length less than is , so there is at least one incompressible string of length

Corollary: At least strings of length are incompressible by

This follows from the same logic

Incompressible strings have many properties we’d expect in randomly chosen strings, such as a roughly equal number of 0s and 1s and a longest run length of about

We can prove something even more general

A property of strings is a function that maps strings to and a property holds for almost all strings if the fraction of strings of length on which it is approaches 0 as grows

Theorem: Let be a computable property that holds for almost all strings, then for any , the property is on only finite many strings that are incompressible by

“On input , find the th string where and output

For any such string , let be the position or index of on a list of all strings that fail to have property , in which case is a short description of with length

For any , select such that at most a fraction of strings of length or less fail to satisfy

Let be a string of length that fails to have property

, so
Thus every sufficiently long that fails to satisfy by virtue is compressible by and our theorem follows

Unfortunately, this is all confined to theory because is not computable, no algorithm can decide in general if a string is incompressible, and there is not even an infinite subset of Turing-recognizable incompressible strings

Theorem: For some constant , for every , the minimal description is incompressible by

To prove this, consider “On , run on and reject if the output is not of form . Run on and halt with its output on the tape.”

Let and suppose is -compressible for some string
This means
But then is a description of with length , which is a contradiction