3.1 AlphaGo vs. Lee Sedol

The March 2016 match between the AlphaGo computer program and the legendary Go player Lee Sedol was more than just a sporting event. It was a historical fork in the road, a moment when humanity received undeniable proof: artificial intelligence had reached a level capable not only of calculation, but also of demonstrating something indistinguishable from creative intuition in one of the most complex intellectual spheres created by humans.

Context: Why Go is Not Chess

To understand the scale of the achievement, one must grasp the complexity of Go.

Simple Rules, Unimaginable Complexity. The rules of Go can be explained in a minute: two players take turns placing black and white stones on a 19x19 board, the goal is to surround territory. However, the number of possible positions in Go (~10¹⁷⁰) exceeds the number of atoms in the observable universe (~10⁸⁰). This makes a complete enumeration of all options absolutely impossible even for the most powerful supercomputers.

Intuition vs. Calculation. In chess, a strong player can calculate sequences of moves many steps ahead. In Go, this is impossible due to the branching width. Top-level players, like Lee Sedol (holder of the highest rank, 9th dan), make decisions based on strategic intuition, a "sense of shape," and experience accumulated over decades of recognizing thousands of patterns. They "see" strong and weak groups of stones, "feel" the tension on the board.

History of AI Defeats. Until 2016, the best computer programs in Go played at the level of a strong amateur. Experts predicted it would take at least another 10 years before AI could defeat a world champion. It was believed that this would require a breakthrough in the field of artificial intuitive thinking, not just an increase in computing power.

AlphaGo's Architecture: A Brilliant Synthesis

Google's DeepMind subsidiary's AlphaGo was not just a "fast calculator." Its strength lay in the synergy of several deep neural networks that mimicked different aspects of human thinking in Go.

Policy Network — "Intuition." This deep convolutional neural network (similar to image recognition networks) was trained on 30 million moves from games played by strong human players. It looked at a board position as an image and predicted which moves a professional player would make in that situation. It learned to recognize patterns and discard obviously weak moves, narrowing the search field from hundreds of possible moves to a few dozen of the most promising ones.

Value Network — "Strategic Vision." This was a key innovation. Instead of calculating a position to the end (which is impossible), this neural network learned to evaluate the probability of winning from any given board position. It answered the question: "If the game continues from this position, how likely is it that black (or white) will win?" Training was done on millions of games AlphaGo simulated against itself. Thus, the value network learned to feel the strategic potential of a position, not to count specific points.

Monte Carlo Tree Search (MCTS) — "Thought Experiment." The algorithm combined the power of both networks. To choose a move, it conducted thousands of quick "thought experiments" (simulations) from the current position:

  1. Selection: At each simulation step, the Policy Network suggested which moves were interesting to explore.
  2. Expansion: The algorithm delved into interesting branches.
  3. Simulation: The game was quickly played out to the end (often with random moves or using a simplified Policy Network).
  4. Backpropagation: The result of the simulation (win/loss) was used to update the evaluation of moves along the path to the root.

The final move was chosen as the most explored and statistically promising. This was not a brute-force calculation, but a balanced exploration guided by intuition (Policy Network) and strategic assessment (Value Network).

The Climax: Move 37, Game 2

After AlphaGo's victory in the first game, in the second game, Lee Sedol was playing an aggressive, complex game. On the 37th move, AlphaGo made a move that stunned the entire Go world.

Expert Reaction: A commentator, a 9th dan professional, upon seeing the move, first thought it was a mistake or "interference." The move was made on the fifth line from the edge, which in classical strategy is considered too distant and "empty" for the early middle game. No high-level human would have made it in that position.

Consequences: Dozens of moves later, the genius of the move became apparent. It created a subtle, pressuring influence on the entire center and right flank of the board, linking seemingly unrelated groups of stones. It was a move that did not pursue an immediate tactical goal, but built a strategic advantage for dozens of moves ahead. Lee Sedol later admitted that it was after this move that he realized he was playing against a being of a different order.

Philosophical and Practical Significance

AlphaGo's 4-1 victory was an event of planetary scale.

Paradigm Shift: It proved that deep learning and reinforcement learning allow machines to master areas requiring intuitive, pattern-based thinking, without explicit rule programming.

A New Tool for Humanity: Go players themselves began studying AlphaGo's games, discovering new, non-obvious strategies and concepts. AI became a coach and a source of inspiration for humans in their own discipline.

The Path to AGI: AlphaGo became a prototype for systems that can learn in complex, uncertain environments. Its successor, AlphaZero, learned from scratch to play not only Go but also chess and shogi, achieving superhuman level in just a few hours, simply by playing against itself. This showed a path toward creating universal learning algorithms.

The AlphaGo vs. Lee Sedol match will remain in history as the "Sputnik moment" for AI — a vivid demonstration that machines have moved into a new league of intellectual ability, forcing humanity for the first time to seriously contemplate partnership and competition with non-biological intelligence.

Previous: 3.0 Introduction Next: 3.2 AlphaFold Discovery