Machine Learning

Overfitting

When a model learns the training data too well — including its noise and random fluctuations — and performs poorly on new, unseen data. The model essentially memorizes rather than generalizes.

Why It Matters

Overfitting is one of the most common pitfalls in ML. A model that overfits is useless in production because it cannot handle real-world data it has not seen before.

Example

A model that scores 99% accuracy on training data but only 60% on test data — it memorized the training examples instead of learning general patterns.

Think of it like...

Like a student who memorizes every answer in a practice exam but fails the real exam because the questions are slightly different — they learned the answers, not the concepts.

Related Terms