Claude Shannon telecommunication |

## Information theory

- Information is
- symbol sequences
- (like these syllables)
- that can be counted,
- reduced to zeros
- and ones, stored, and
- communicated
- without regard for
- meaning or importance.
- Claude Shannon figured
- an arrangement of
- switches could resolve
- any logical
- or arithmetic
- problem. He laid down
- essential limits
- on information
- storage, transmission,
- and processing.

## Information entropy

- The cost of writing information is,
- minimum, one digital bit for a true or false.
- In practice, much more than that.
- How many interviews, how many reviews
- to make a few improvements to a manual?
- How many times must I repeat myself?

## Success of English

- English pronunciations and spellings are irrational
- and weird, idioms and euphemisms are mysterious,
- but English has succeeded world-wide because
- it is the one language that you may butcher,
- mangle, and contort freely, and still
- be understood—the ideal language for poetry.

One can say a lot about information without concern for meaning or importance. Even a poet concerned mainly about meaning might count syllables. Claude Shannon’s ideas contributed to many fields including digital circuit design, data compression, sampling theory, cryptography, artificial intelligence, natural-language processing, genetics, and game theory.

See also in

The book of science:Cipher—Polybius, Julius Caesar, Giovan Battista Bellaso, Antoine and Bonaventure Rossignol, Charles Wheatstone, Frank Miller, Felix Delastelle, Fritz Nebel, Lester S. Hill, John F. ByrneBoolean algebra—George BooleGame theory—John von Neumann, Oskar MorgensternComputer—Charles Babbage, Alan Turing, John von NeumannReadings on wikipedia: