The 175 Billion Parameter Surprise (2020)

In 2020, the release of GPT-3 marked a significant milestone in the field of artificial intelligence, showcasing the power of scale in achieving generality.

What happened: On September 22, 2020, Microsoft announced that it had licensed GPT-3 exclusively, developed by Tom Brown, Benjamin Mann, Ilya Sutskever, and Sam Altman at OpenAI. This massive language model, with 175 billion parameters, demonstrated unprecedented capabilities in writing essays, translating languages, answering questions, and generating code, all without task-specific training. GPT-3 - Wikipedia

Why it matters: GPT-3’s release transformed the AI industry’s trajectory, proving that scale itself could be a path to generality. This breakthrough launched a wave of foundation-model startups and products, emphasizing the importance of large-scale models in advancing AI research and applications. Language Models are Few-Shot Learners (original paper)

Further reading: