Introduction
So far, our journey has been in the abstract realm of pure logic. Gödel, Church, and Turing explored the absolute limits of what could be computed. But their work didn’t address a critical, real-world problem: once you have a computation, how do you send its result from one place to another reliably?
This was the world of engineers, of crackly telephone lines, static-filled radio waves, and garbled telegraph signals. The prevailing wisdom was that noise was an unavoidable enemy that would always corrupt a message. The more you transmitted, the more errors you’d get.
Then, in 1948, a brilliant and playful engineer at Bell Labs named Claude Shannon published a paper that changed everything. This was a man who not only laid the foundations for the digital age but also spent his spare time building juggling machines and estimating the game-tree complexity of chess (a figure now known as the Shannon number, roughly ). His landmark paper was called “A Mathematical Theory of Communication,” and it single-handedly created the field of information theory.
The Big Idea: Separating Information from Meaning
Shannon’s first stroke of genius was to completely ignore the meaning of a message. To him, it didn’t matter if you were sending a Shakespearean sonnet, a stock market quote, or complete gibberish. He wanted to quantify the information itself.
He proposed a radical new unit of measure: the bit.
A bit (a portmanteau of “binary digit”) is the most fundamental unit of information. It represents the answer to a single yes/no question, reducing uncertainty by exactly half.
- Is the light on or off?
- Is the coin heads or tails?
- Is the voltage high or low?
By encoding any message into a long string of these binary digits (0s and 1s), Shannon could treat all information the same, regardless of its content.
Entropy: The Measure of Surprise
Shannon then asked, “How much information is in a message?” His answer was profound: the amount of information is the measure of its surprise.
If I tell you “the sun will rise tomorrow,” you’re not surprised. It’s a predictable event. The information content is very low. But if I tell you “a blizzard will hit Miami tomorrow,” you are very surprised. That message contains a lot of information.
Shannon gave this concept a precise mathematical form and called it Entropy, denoted by . For a source of information that can produce several symbols, the entropy is the average amount of information (in bits) per symbol.
The formula for entropy is:
Let’s break that down:
-
is the source of information (e.g., a coin flip).
-
is one of the possible outcomes (e.g., “heads”).
-
is the probability of that outcome happening.
-
is the magic part. It quantifies the “surprise” of that outcome in bits. The logarithm base 2 is used because we are measuring in bits.
-
The (sigma) means we sum this value up for all possible outcomes.
Example: A Fair Coin vs. A Rigged Coin
-
Fair Coin: Heads () and Tails (). The entropy is bit. Each flip gives you exactly one bit of information because the outcome is maximally uncertain.
-
Rigged Coin: Heads () and Tails (). The entropy is bits. The outcome is more predictable, so on average, each flip gives you less than half a bit of new information.
The Channel Capacity Theorem
This was all leading to Shannon’s masterpiece. He modeled a communication system: a source sends a message through a noisy channel to a receiver.
He defined a property of the channel itself, called its Channel Capacity (). This is the maximum rate, in bits per second, at which information can be sent through the channel with an arbitrarily small probability of error.
He then gave the world the stunning Shannon-Hartley Theorem, which calculates this capacity:
- : The channel capacity in bits per second.
- : The bandwidth of the channel (e.g., the range of frequencies a radio can use), measured in Hertz.
- : The average power of the signal you are sending.
- : The average power of the noise in the channel.
- : The famous Signal-to-Noise Ratio (SNR).
This elegant formula connected the physical properties of a wire or an antenna (, , ) to the abstract, logical concept of information ().
Shannon’s theorem made a shocking promise: It doesn’t matter how noisy your channel is. As long as your transmission rate is below the channel capacity , you can invent an encoding scheme that allows you to transmit information with zero errors.
This was revolutionary. Noise was no longer an insurmountable barrier; it was just a variable in an equation. It defined a speed limit, not a hard wall. This insight is the theoretical foundation for every piece of modern digital communication, from Wi-Fi and 4G to data compression (like .zip files) and the error-correcting codes that allow spacecraft to send clear images from billions of miles away.
With the theory of information laid out, the stage was set. The world now understood the limits of computation and the rules of communication. All that was needed was a blueprint for a machine that could actually do it all. That blueprint would come from John von Neumann.