Building AlphaGo from scratch – Eric Jang
Summary
Eric Jang rebuilt AlphaGo from scratch and explained in detail the application of Monte Carlo Tree Search and deep learning in Go, demonstrating the feasibility of reproducing a powerful Go AI at low cost nowadays.
View Cached Full Text
Cached at: 05/15/26, 05:08 PM
Similar Articles
@dwarkesh_sp: New blackboard lecture w @ericjang11 He walks through how to build AlphaGo from scratch, but with modern AI tools. Some…
A blackboard lecture by Eric Jang walks through building AlphaGo from scratch with modern AI tools, covering RL, MCTS, self-play, and connecting to LLM training, along with a discussion on automated AI research.
10 years of AlphaGo: The turning point for AI | Thore Graepel & Pushmeet Kohli
This article reviews the history of AlphaGo's historic victory over Lee Sedol in 2016, analyzes its technical principles combining deep learning and search, and explores the profound impact of this event on the development of AI.
From games to biology and beyond: 10 years of AlphaGo’s impact
DeepMind reflects on the 10th anniversary of AlphaGo, highlighting its role in kickstarting the modern AI era and its subsequent impact on scientific research and the pursuit of AGI.
@0xLogicrw: Former OpenAI post-training core member Jiayi Weng proposed a new reinforcement learning paradigm called "Heuristic Learning" in his personal capacity and open-sourced all experimental code. He used Codex (GPT-5.4) to repeatedly play the Atari game Breakout, but GPT-5.4 was never retrained...
Former OpenAI researcher Jiayi Weng proposed a new paradigm called "Heuristic Learning", which uses large language models to generate and iteratively modify Python code to solve reinforcement learning tasks. Knowledge is stored in interpretable code rather than neural network parameters, effectively avoiding catastrophic forgetting. It has achieved excellent results on Atari and MuJoCo benchmarks and the code has been open-sourced.
@berryxia: Moonshot AI founder Yang Zhilin recently released a 40-minute video. Born in 1992, valedictorian of Tsinghua CS undergrad, PhD from CMU, co-author of Transformer-XL and XLNet, former researcher at Google Brain and Meta, he calmly deconstructs Kimi K2 in front of the camera...
Moonshot AI founder Yang Zhilin released a 40-minute video detailing the training process of the Kimi K2 model, which cost only $4.6 million. In an 8-model real-time programming competition, Kimi K2 took first place, defeating GPT-5.5 and others, demonstrating how a small team can overturn the traditional compute-stacking paradigm through architecture optimization.