The Machine That Taught Itself to Play

In 2015, a groundbreaking algorithm called the deep Q-network (DQN) emerged, marking a pivotal moment in the history of artificial intelligence.

What happened: In February 2015, a team of researchers at DeepMind, including Volodymyr Mnih, Koray Kavukcuoglu, David Silver, and Demis Hassabis, published their work on the deep Q-network (DQN) in the journal Nature. This algorithm was the first to successfully combine deep learning with reinforcement learning, enabling it to achieve human-level performance across a wide range of tasks. Notably, DQN could learn directly from high-dimensional sensory input, such as raw pixels from video games, without any task-specific engineering. Nature

Why it matters: The DQN’s ability to learn from unstructured data and perform complex tasks autonomously was a significant breakthrough. It not only reignited interest in reinforcement learning but also laid the groundwork for subsequent achievements like AlphaGo, which defeated a world champion at the game of Go just one year later. This milestone demonstrated the potential of deep reinforcement learning to solve real-world problems and opened up new avenues for research and application in robotics, healthcare, finance, and more.

Further reading: