Mathematical Psychology
About

Channel Capacity

Shannon's channel capacity theorem establishes the maximum rate at which information can be reliably transmitted through a noisy communication channel, with deep implications for understanding human information processing limits.

C = max I(X;Y)

Claude Shannon's 1948 channel coding theorem proved that every noisy communication channel has a definite capacity C — measured in bits per channel use — such that information can be transmitted at any rate below C with arbitrarily low error probability, but not at any rate above C. This result, one of the most profound in twentieth-century mathematics, established the theoretical limits of reliable communication and provided the conceptual foundation for analyzing human information processing as a capacity-limited channel.

The Channel Capacity Formula

Shannon's Channel Capacity C = max_{p(x)} I(X;Y)

I(X;Y) = H(Y) − H(Y|X)
= Σ_x Σ_y p(x,y) · log₂[p(x,y) / (p(x) · p(y))]

The maximum is taken over all possible input distributions p(x)

Channel capacity is defined as the maximum mutual information between the input X and output Y of a channel, optimized over all possible input probability distributions. For a binary symmetric channel with crossover probability p, the capacity is C = 1 − H(p) = 1 + p·log₂(p) + (1−p)·log₂(1−p). For an additive white Gaussian noise channel with bandwidth W and signal-to-noise ratio S/N, the capacity is C = W·log₂(1 + S/N), known as the Shannon-Hartley theorem.

The Noisy Channel Coding Theorem

Shannon's coding theorem has two parts. The direct theorem states that for any rate R < C, there exist codes that achieve probability of error as close to zero as desired. The converse theorem states that for R > C, the error probability is bounded away from zero regardless of the code used. The proof is existential — Shannon showed that random codes achieve capacity with high probability — but it launched the field of coding theory devoted to constructing practical codes approaching capacity.

Channel Capacity and Human Performance

Shannon's framework inspired psychologists to model the human observer as a noisy channel. Garner and Hake (1951) and Miller (1956) measured human channel capacity for absolute judgments of unidimensional stimuli, finding consistent limits around 2–3 bits — the capacity to distinguish about 5–9 categories. This convergence across modalities suggested a fundamental architectural constraint on human information processing, though the analogy between electronic and biological channels has important limitations.

Psychological Significance

The concept of channel capacity has been applied broadly in psychology. In absolute identification tasks, observers are asked to identify stimuli varying on a single dimension, and information transmission (mutual information between stimulus and response) asymptotes at roughly 2.5 bits regardless of the number of stimuli. In studies of attention, dual-task paradigms can be analyzed as parallel channels whose joint capacity is constrained. The channel capacity framework also connects to Hick's law: reaction time increases linearly with stimulus entropy, as if the observer processes information at a fixed rate determined by channel capacity.

Despite its power, the channel metaphor has limitations when applied to human cognition. Unlike electronic channels, human processing is influenced by meaning, context, chunking, and learning — factors that effectively recode information and alter the apparent channel capacity. Shannon himself cautioned against overextending information theory into domains where its assumptions do not strictly hold.

Interactive Calculator

Each row provides an event label and its probability. The calculator computes Shannon entropy H = −Σ p(x)·log₂ p(x), information content per event, and redundancy.

Click Calculate to see results, or Animate to watch the statistics update one record at a time.

Related Topics

References

  1. Shannon, C. E. (1948). A mathematical theory of communication. The Bell System Technical Journal, 27(3), 379–423. doi:10.1002/j.1538-7305.1948.tb01338.x
  2. Shannon, C. E., & Weaver, W. (1949). The Mathematical Theory of Communication. University of Illinois Press.
  3. Garner, W. R., & Hake, H. W. (1951). The amount of information in absolute judgments. Psychological Review, 58(6), 446–459. doi:10.1037/h0054482
  4. Miller, G. A. (1956). The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychological Review, 63(2), 81–97. doi:10.1037/h0043158
  5. Cover, T. M., & Thomas, J. A. (2006). Elements of Information Theory (2nd ed.). Wiley-Interscience. doi:10.1002/047174882X

External Links