© 2026 Greg T. Chism · MIT License

Early Stopping — Interactive Explorer

Watch training and validation loss diverge — find the optimal stopping point before overfitting takes hold


Model Complexity
Low capacity — less prone to overfitting
Training Parameters
η 0.010
Learning rate — controls weight update size
Patience (epochs)
P 5
Stop after 5 epochs without val improvement
Dataset Noise
Label noise in the training set
Simulation
Speed Med
Early Stopping Criterion
stop if e − e* ≥ P
Key Concepts
What early stopping is: A regularization technique that halts training when the model begins to overfit — validation loss stops improving while training loss keeps falling. The weights at the best validation epoch are restored.
What patience means: The number of consecutive epochs without validation improvement before training stops. Higher patience tolerates longer plateaus; lower patience halts sooner and risks underfitting.
Why val loss diverges: Once a model memorizes training noise, its weights no longer generalize. Training loss keeps falling (fitting noise), but validation loss rises. The widening gap is the generalization error.
When to stop: At the epoch where validation loss is lowest — marked by the vertical line. More patience gives the model more chances to escape temporary plateaus before deciding to stop.
What's happening?
Press Play or Step to begin training. Watch training loss (green) and validation loss (orange) — when they diverge, overfitting has begun.
Epoch 0
Phase
Training — epoch 0
Loss curve renders here Press Play to begin training
Train loss Val loss Best val epoch Early stopping point Patience window Overtrained zone
Training Status
0
Epoch
Best Val Epoch
Train Loss
Val Loss
Generalization Gap
val − train (proportion)
TrainVal
Patience Counter 0 / 5
Weight Norm
Weight norm chart renders here
Growing weight norms signal memorization — the model is fitting noise rather than signal.