Information Theory

From Shannon's 1948 paper to quantum error correction — the mathematical theory of communication, compression, and the fundamental limits of knowledge.

Shannon's Communication Model

InformationSourceEncoderChannelDecoderDestinationNoisemessagecodewordreceiveddecodedChannel Capacity: C = max I(X;Y)

About This Course

In 1948, Claude Shannon published “A Mathematical Theory of Communication” — a paper that created an entirely new science. Shannon showed that information can be measured in bits, that every communication channel has a maximum rate at which information can be transmitted reliably, and that clever coding can approach this limit with vanishingly small error probability.

Information theory has since grown far beyond communication engineering. It now underpins data compression, cryptography, machine learning, statistical inference, thermodynamics, quantum computing, and even our understanding of black holes. The entropy formula\(H = -\sum p_i \log p_i\) is as fundamental to science as \(E = mc^2\).

This course covers the full sweep: from Shannon's original theorems through practical coding algorithms, continuous-channel theory, Kolmogorov complexity, the physics of information (Maxwell's demon, Landauer's principle, black holes), quantum information theory, and modern applications in AI and cryptography. Every chapter includes interactive Python simulations and detailed mathematical derivations.

The Central Equations

Shannon Entropy

\( H(X) = -\sum_{i=1}^{n} p_i \log_2 p_i \)

Channel Capacity

\( C = \max_{p(x)} I(X;Y) \)

Shannon-Hartley (AWGN)

\( C = B\log_2\!\left(1 + \frac{S}{N}\right) \)

Von Neumann Entropy

\( S(\rho) = -\text{Tr}(\rho \ln \rho) \)

Course Structure

Part I

Foundations

Shannon 1948

Shannon entropy, the source coding theorem, mutual information, and channel capacity — the four pillars of information theory.

Part II

Coding Theory

Practical Codes

From Huffman trees to turbo codes — the algorithms that make digital communication and storage possible.

Part III

Continuous Channels

Analog → Digital

Differential entropy, the Gaussian channel, Shannon’s capacity formula, MIMO systems, and the water-filling solution.

Part IV

Algorithmic Complexity

Kolmogorov 1965

Kolmogorov complexity, the minimum description length principle, and deep connections to Gödel’s incompleteness and Turing’s halting problem.

Part V

Information & Physics

It from Bit

Maxwell’s demon, Landauer’s erasure principle, Bekenstein-Hawking entropy, and the black hole information paradox.

Part VI

Quantum Information

Qubits & Entanglement

Von Neumann entropy, quantum channels, the Holevo bound, and quantum error correction — the information theory of the quantum world.

Part VII

Modern Applications

21st Century

Data compression algorithms, the information bottleneck in deep learning, modern cryptography, and network information theory.

Timeline Highlights

1928Hartley defines information as logarithm of number of possible messages
1948Shannon publishes "A Mathematical Theory of Communication" — information theory is born
1949Shannon proves the noisy channel coding theorem
1950Hamming invents the first error-correcting code
1952Huffman invents optimal prefix-free codes
1961Landauer shows that erasing a bit dissipates at least kT ln 2 energy
1965Kolmogorov and Chaitin independently define algorithmic complexity
1973Bekenstein proposes black hole entropy — information meets gravity
1993Turbo codes approach Shannon limit within 0.5 dB
1995Shor's quantum error-correcting code
2012Polar codes provably achieve channel capacity (Arıkan, 2009; adopted by 5G)
2017Information bottleneck theory connects deep learning to information theory

Recommended Reading

  • Elements of Information Theory — Thomas Cover & Joy Thomas (2nd ed., 2006)
  • Information Theory, Inference, and Learning Algorithms — David MacKay (2003, free online)
  • A Mathematical Theory of Communication — Claude Shannon (1948, original paper)
  • Quantum Computation and Quantum Information — Nielsen & Chuang (2010)
  • The Information — James Gleick (2011, popular history)