![]() |
|
information, n. Knowledge communicated or received concerning a particular fact or circumstance.
A Computer Scientist's definiton:
Information resolves uncertainty.
Information is simply that which cannot be predicted. The less predictable a message is, the more information it conveys!
Suppose you're faced with N equally probable choices, and I give you a fact that narrows it down to M choices. Then you've been given:
log2(N/M) bits of information
Examples:
log2(107/104) = ~9.966 bits
![]() |
i2 = log2(36/1) = 5.170 bits
i3 = log2(36/2) = 4.170 bits i4 = log2(36/3) = 3.585 bits i5 = log2(36/4) = 3.170 bits i6 = log2(36/5) = 2.848 bits i7 = log2(36/6) = 2.585 bits i8 = log2(36/5) = 2.848 bits i9 = log2(36/4) = 3.170 bits i10 = log2(36/3) = 3.585 bits i11 = log2(36/2) = 4.170 bits i12 = log2(36/1) = 5.170 bits |
The average information provided by the sum of 2 dice is: 3.274 bits.
The average information from events of given process is called its Entropy.
2 = 10011 | 3 = 0101 | 4 = 011 | 5 = 001 |
6 = 111 | 7 = 101 | 8 = 110 | 9 = 000 |
10 = 1000 | 11 = 0100 | 12 = 10010 |
2 = 10011 | 3 = 0101 | 4 = 011 | 5 = 001 |
6 = 111 | 7 = 101 | 8 = 110 | 9 = 000 |
10 = 1000 | 11 = 0100 | 12 = 10010 |
(Click in box to decode)
1001100101011110011100101
2
1001100101011110011100101
2 5
1001100101011110011100101
2 5 3
1001100101011110011100101
2 5 3 6
1001100101011110011100101
2 5 3 6 5
1001100101011110011100101
2 5 3 6 5 8
1001100101011110011100101
2 5 3 6 5 8 3
1001100101011110011100101
A simple greedy algorithm for approximating a minimum encoding
Once the *tree* is constructed, label its edges consistently and follow the paths from the largest *meta* item to each of the real item to find the encoding.
2 = 10011 | 3 = 0101 | 4 = 011 | 5 = 001 |
6 = 111 | 7 = 101 | 8 = 110 | 9 = 000 |
10 = 1000 | 11 = 0100 | 12 = 10010 |
How does this code compare to the information content?
Pretty close. Recall that the lower bound was 3.274 bits.
However, an efficient encoding (as defined by having an average code size close to the information content) is not always what we want!
If all choices are equally likely (or we have no reason to expect otherwise), then a fixed-size code is often used. Such a code should use at least enough bits to represent the information content.
BCD | |
0 - | 0000 |
1 - | 0001 |
2 - | 0010 |
3 - | 0011 |
4 - | 0100 |
5 - | 0101 |
6 - | 0110 |
7 - | 0111 |
8 - | 1000 |
9 - | 1001 |
ex. Decimal digits 10 = {0,1,2,3,4,5,6,7,8,9}
4-bit BCD (binary code decimal)
log2(10/1) = 3.322 < 4 bits
ex. ~84 English characters = {A-Z (26), a-z (26), 0-9 (10),
punctuation (8), math (9), financial (5)}
7-bit ASCII (American Standard Code for Information Interchange)
log2(84/1) = 6.392 < 7 bits
  | 0000 | 0001 | 0010 | 0011 | 0100 | 0101 | 0110 | 0111 | 1000 | 1001 | 1010 | 1011 | 1100 | 1101 | 1110 | 1111 |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
000 | NUL | SOH | STX | ETX | EOT | ACK | ENQ | BEL | BS | HT | LF | VT | FF | CR | SO | SI |
001 | DLE | DC1 | DC2 | DC3 | DC4 | NAK | SYN | ETB | CAN | EM | SUB | ESC | FS | GS | RS | US |
010 | ! | " | # | $ | % | & | ' | ( | ) | * | + | , | - | . | / | |
011 | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | : | ; | < | = | > | ? |
100 | @ | A | B | C | D | E | F | G | H | I | J | K | L | M | N | O |
101 | P | Q | R | S | T | U | V | W | X | Y | Z | [ | \ | ] | ^ | _ |
110 | ` | a | b | c | d | e | f | g | h | i | j | k | l | m | n | o |
111 | p | q | r | s | t | u | v | w | x | y | z | { | | | } | ~ | DEL |
It is straightforward to encode positive integers as a sequence of bits. Each bit is assigned a weight. Ordered from right to left, these weights are increasing powers of 2. The value of an n-bit number encoded in this fashion is given by the following formula:
211 | 210 | 29 | 28 | 27 | 26 | 25 | 24 | 23 | 22 | 21 | 20 |
---|---|---|---|---|---|---|---|---|---|---|---|
0 | 1 | 1 | 1 | 1 | 1 | 0 | 1 | 1 | 1 | 1 | 1 |
20 = 1 | 25 = 32 |
21 = 2 | 26 = 64 |
22 = 4 | 27 = 128 |
23 = 8 | 28 = 256 |
24 = 16 | 29 = 512 |
210 = Kilo (1024) |
220 = Mega (1024*1024) |
230 = Giga (1024*1024*1024) |
240 = Tera (1024*1024*1024*1024) |
250 = Peta (1024*1024*1024*1024*1024) |
260 = Exa (1024*1024*1024*1024*1024*1024) |
A "Giga" something or other
Oftentimes we will find it convenient to cluster groups of bits together for a more compact written representation. Clustering by 3 bits is called Octal, and it is often indicated with a leading zero, 0. Octal is not that common today.
Clusters of 4 bits are used most frequently. This representation is called hexadecimal. The hexadecimal digits include 0-9, and A-F, and each digit position represents a power of 16. Commonly indicated with a leading "0x".