Information Contained in Letter of Alphabet

From ProofWiki
Jump to navigation Jump to search

Theorem

Let $\psi$ be a letter of the English alphabet.

To a first degree of approximation, the quantity of information contained in $\psi$ is $4 \cdotp 7$.

That is, there is approximately $4 \cdotp 7$ times as much information conveyed by transmission of a single letter as a single bit.


Further Analysis

It needs to be pointed out this analysis of the quantity of information in a letter of the English alphabet is appropriate only if each letter is equally likely to occur.

In practice this is unlikely to occur, and the concept of entropy, also known as uncertainty, is used instead.


Proof

Let it be assumed for the purposes of this exercise that each letter has an equal probability of occurring as an element of a message.

There are $26$ letters of the English alphabet.

The amount of information contained in $\psi$ is therefore:

$\map I \psi = \dfrac {\lg 26} {\lg 2} \approx 4 \cdotp 7$

where $\lg$ denotes the binary logarithm.

$\blacksquare$




Sources