Representation Learning
The process of automatically discovering useful features or representations from raw data, rather than manually engineering them. Deep learning excels at learning hierarchical representations.
Why It Matters
Good representations make downstream tasks easier. A model that learns useful representations of images can be adapted for detection, classification, or generation.
Example
A vision model automatically learning to represent images as combinations of edges, textures, shapes, and object parts — without being told what features to look for.
Think of it like...
Like learning to see the world not as individual pixels but as meaningful objects, textures, and scenes — the right representation makes understanding much easier.
Related Terms
Feature Engineering
The process of selecting, transforming, and creating input variables (features) from raw data to improve model performance. It requires domain knowledge to identify what information is most useful for the model.
Embedding
A numerical representation of data (text, images, etc.) as a vector of numbers in a high-dimensional space. Similar items are placed closer together in this space, enabling machines to understand semantic relationships.
Latent Space
A compressed, lower-dimensional representation of data learned by a model. Points in latent space capture the essential features of the data, and nearby points represent similar data items.
Self-Supervised Learning
A training approach where the model generates its own labels from the data, typically by masking or hiding parts of the input and learning to predict them. No human-annotated labels are needed.
Transfer Learning
A technique where a model trained on one task is repurposed as the starting point for a model on a different but related task. Instead of training from scratch, you leverage knowledge the model has already acquired.