Perceptron

Quick Reference

Core Concept

Learn to classify by making mistakes and correcting them. Like teaching a child to sort toys: watch, correct errors, repeat until they get it right.

The Algorithm

Four-Step Loop:

Repeat until all points correctly classified or convergence fails.

Decision Function

h(x; θ, θ0) = sign(θTx + θ0)

Components:

Update Rule

When point (x(i), y(i)) is misclassified:

θ ← θ + y(i)x(i)
θ0 ← θ0 + y(i)

Geometric Intuition:

No calculus. No gradient descent. Pure geometry.

Convergence Theorem

Guarantee: If data is linearly separable with margin γ and bounded inputs ||x|| ≤ R, then Perceptron converges in at most (R/γ)² updates.

Key Insight: Wider margin = faster convergence.

Critical Limitation

Only works for linearly separable data.

If no straight line can separate classes, Perceptron never converges.

Classic Failure: XOR Problem

Quick Facts