Back to timeline

GPT-3

OpenAI unveils GPT-3 with 175 billion parameters, demonstrating remarkable few-shot learning abilities and launching the era of large language models.

Model Release

What Happened

OpenAI published "Language Models are Few-Shot Learners," introducing GPT-3 — a 175-billion parameter autoregressive language model. GPT-3 demonstrated that scaling language models to sufficient size enabled strong performance on downstream tasks with only a few examples provided in the prompt (few-shot learning), without any gradient updates or fine-tuning.

Why It Matters

GPT-3 was a paradigm shift. It showed that sufficiently large language models could perform tasks from simple demonstrations in the prompt — an ability now called "in-context learning." This kicked off the scaling era in AI, with labs racing to build ever-larger models. OpenAI also made GPT-3 available via API, creating the first commercial large language model platform and spawning an entire ecosystem of AI-powered applications.

Technical Details