Self-Supervised Learning
Self-supervised learning is the approach behind most of today's large language models and it has quietly become one of the most important developments in modern AI. The core idea is clever: instead of requiring humans to label data, the system creates its own learning tasks from the data itself. For a language model, the task might be predicting the next word in a sentence, or filling in a word that has been deliberately hidden. The text itself provides both the question and the answer - no human labelling needed. This matters enormously for practical reasons. Labelling data is expensive, slow, and often requires specialist knowledge. Self-supervised learning sidesteps that bottleneck by using the vast quantities of unlabelled text, images, and other content available on the internet. GPT, BERT and their successors were all trained this way. The model learns rich, general-purpose representations of language or images that can then be fine-tuned for specific tasks with relatively small amounts of labelled data. The breakthrough was realising that prediction - simply guessing what comes next - forces a model to develop deep representations of structure, meaning and context. It is an elegant solution to what was previously one of AI's biggest practical obstacles.