Mathematical Psychology
About

Perceptron

The perceptron is the simplest artificial neural network, a single-layer model that learns linear decision boundaries through error-corrective learning.

y = σ(Σwᵢxᵢ + b)

The perceptron, introduced by Frank Rosenblatt in 1958, is the foundational model of artificial neural networks and connectionist approaches to cognition. It takes a weighted sum of inputs, adds a bias term, and passes the result through a step function to produce a binary classification.

Perceptron Learning Rule ŷ = step(Σ wᵢ xᵢ + b)
Δwᵢ = η · (y − ŷ) · xᵢ

Convergence theorem: If data is linearly separable,
the perceptron learning rule converges in finite steps.

The Perceptron Convergence Theorem

Rosenblatt proved that if the training data is linearly separable, the perceptron learning rule is guaranteed to find a separating hyperplane in a finite number of steps. This was an exciting result, but Minsky and Papert's 1969 book Perceptrons demonstrated that single-layer perceptrons cannot learn XOR or any other non-linearly separable function, temporarily diminishing enthusiasm for neural network approaches.

Legacy

The limitations of single-layer perceptrons were overcome by multilayer networks trained with backpropagation. Nevertheless, the perceptron remains important as a pedagogical model, as the building block of larger networks, and as a model of individual neurons that perform linear discrimination. The perceptron learning rule is mathematically equivalent to the delta rule used in many connectionist models of learning.

Interactive Calculator

Each row provides a training example: x1, x2, and label (0 or 1). The calculator trains a single-layer perceptron using the perceptron learning rule with learning rate η=0.1.

Click Calculate to see results, or Animate to watch the statistics update one record at a time.

Related Topics

References

  1. Rosenblatt, F. (1958). The perceptron: A probabilistic model for information storage and organization in the brain. Psychological Review, 65(6), 386–408. https://doi.org/10.1037/h0042519
  2. Minsky, M., & Papert, S. (1969). Perceptrons: An introduction to computational geometry. MIT Press. https://doi.org/10.7551/mitpress/11301.001.0001
  3. Rumelhart, D. E., Hinton, G. E., & Williams, R. J. (1986). Learning representations by back-propagating errors. Nature, 323(6088), 533–536. https://doi.org/10.1038/323533a0

External Links