Back to timeline

BERT

Google introduces BERT, a bidirectional Transformer model that achieves breakthrough results across NLP benchmarks and transforms the field.

Architecture

What Happened

Google AI Language researchers published "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding." BERT (Bidirectional Encoder Representations from Transformers) introduced a new approach to pre-training language representations by jointly conditioning on both left and right context in all layers.

Why It Matters

BERT shattered records across 11 NLP benchmarks and fundamentally changed how the field approached language understanding tasks. It demonstrated that bidirectional pre-training was significantly more effective than left-to-right approaches like GPT. Google integrated BERT into its search engine in 2019, affecting an estimated 10% of all English-language queries. BERT became the most widely used model in NLP research and industry applications.

Technical Details