Back to Journal
BigChessAIDeep BlueTechnology

Artificial Intelligence and Chess: From Deep Blue to BigChess Engines

Marat Fatalov
By Marat Fatalov
32 min read
Artificial Intelligence and Chess: From Deep Blue to BigChess Engines

Artificial Intelligence and Chess: From Deep Blue to BigChess Engines

The story of chess AI is a story of machine intelligence learning to surpass human mastery — and then confronting the humbling difficulty of a new game it has never seen before.


Introduction: The Oldest Benchmark

Chess has been the benchmark problem for artificial intelligence since the field was invented. When Alan Turing and Claude Shannon wrote the first theoretical papers on computer chess in the late 1940s, they were not primarily interested in chess. They were interested in whether machines could exhibit intelligent behavior, and chess — demanding, rule-governed, stratified by skill, with measurable outcomes — was the cleanest available test case.

For half a century, chess AI served as a progress marker for machine intelligence. The program that could beat a club player proved machines could reason tactically. The program that could beat a grandmaster proved something harder to admit. The program that could beat the world champion, in 1997, settled the question for that particular benchmark: machines had surpassed humans at chess.

But the story did not end there. The methods that produced Deep Blue were replaced by methods that produced AlphaZero, which changed not just the strength of chess engines but the nature of their intelligence. And now, chess AI confronts a genuinely new problem: BigChess, with its 10×10 board, its Clone piece, its expanded pawn mechanics, and its absence of the large game databases that modern AI depends on. The machine that learned chess has to learn chess again, from the beginning, with new pieces.


Part I: The Early History of Chess Computing (1940s–1980s)

Turing and Shannon: The Theoretical Foundations

Alan Turing proposed a chess-playing algorithm in 1948 — before he had a computer to run it on. His "Turochamp" program was a paper algorithm that he and David Champernowne verified by hand, working through the calculations manually and making moves according to the program's rules. The process was extraordinarily slow — a single move could take half an hour to compute by hand — but the program played recognizable chess. It evaluated captures, checks, and basic tactical threats, and it demonstrated that a mechanical rule-following process could produce chess moves that were not entirely random.

Claude Shannon published his foundational paper "Programming a Computer for Playing Chess" in 1950. Shannon identified the central challenge of chess programming: the game tree — the branching structure of all possible move sequences from any position — grows exponentially with depth. At a branching factor of approximately 35 (the average number of legal moves in a chess position), a three-move sequence involves examining 35³ = 42,875 positions. A six-move sequence involves more than 1.8 billion. Full game-length evaluation was computationally impossible; selective search was unavoidable.

Shannon proposed two approaches: "Type A" programs that evaluate all moves to a fixed depth (brute force within limits), and "Type B" programs that selectively extend promising lines and prune unpromising ones. The tension between these approaches — brute force depth versus selective intelligence — defined chess programming debate for the next four decades.

The First Programs and Early Tournaments (1950s–1970s)

The first actual computer chess programs appeared in the early 1950s, running on early computers like the Manchester Mark 1. They played very weak chess by human standards — easily beaten by beginners — but they were genuinely playing chess, evaluating legal moves and selecting among them according to a heuristic function.

The key evaluation functions in early programs were material balance (counting piece values), king safety, pawn structure, and piece mobility. These are the same factors a human player considers, but the computer evaluated them through explicit arithmetic rather than pattern recognition. The result was chess that was tactically alert (not hanging pieces or walking into simple mates) but strategically incoherent (unable to form and execute plans).

Progress was steady but incremental through the 1960s and 1970s. Programs improved as hardware improved and as more sophisticated evaluation functions were developed. The Chess 4.x series, developed at Northwestern University, was the first computer program to achieve a national chess master rating, crossing the 2000 mark in human ELO terms by 1977. This was genuinely impressive — a machine competitive with the top 1% of human tournament players — but still far from the grandmaster level.

The Hardware Arms Race (1980s)

By the 1980s, it was becoming clear that chess programming strength was closely correlated with hardware speed. Programs that could search deeper — examine more positions per second — played better chess, even with relatively simple evaluation functions. This led to dedicated chess hardware: special-purpose computers designed specifically to evaluate chess positions as rapidly as possible.

Belle, designed by Ken Thompson and Joe Condon at Bell Labs, was one of the first dedicated chess computers. It could evaluate 160,000 positions per second — orders of magnitude faster than general-purpose computers of the era — and achieved a USCF rating of approximately 2250, placing it in the top ranks of American tournament players below the elite grandmaster level.

Deep Thought, the direct predecessor of Deep Blue, followed in the late 1980s. Developed at Carnegie Mellon University by Feng-hsiung Hsu and others, Deep Thought could evaluate 700,000 positions per second and achieved a FIDE rating above 2550 — solidly grandmaster level. The question was no longer whether a computer could beat a grandmaster but whether one could beat the best grandmaster in the world.


Part II: Deep Blue vs. Kasparov — The Match That Changed Everything (1996–1997)

The 1996 Match: A Preview

In February 1996, the world's reigning chess champion, Garry Kasparov, played a six-game match against IBM's Deep Blue — an upgraded successor to Deep Thought — in Philadelphia. The match attracted massive media attention. For the first time, the question of whether a computer could defeat the world's best human chess player was being seriously tested under standard match conditions.

Kasparov won the match 4–2. Deep Blue won the first game — the first time a computer had won a game against a reigning world champion under standard conditions — but Kasparov recovered, studied his opponent's patterns, and adjusted his play to exploit the computer's weaknesses. He found that Deep Blue could be confused by positions with long-term strategic elements it could not evaluate properly, and he steered toward those positions in the games that followed.

The result was encouraging for human pride: the champion had studied his mechanical opponent and adapted. But the match had also demonstrated that the gap was narrowing rapidly. Deep Blue had won a game against Kasparov under full tournament conditions. The rematch, already scheduled, would be decisive.

The 1997 Match: Man vs. Machine

The rematch in May 1997 in New York was covered as a major cultural event. Television networks broadcast live updates. Magazine covers depicted Kasparov facing the IBM logo. The question was framed, perhaps hyperbolically but not inaccurately, as a contest between human intelligence and artificial intelligence for primacy in the game that had defined intellectual achievement for fifteen centuries.

Deep Blue won the match 3.5–2.5. Kasparov won the first game, Deep Blue won the second, then three draws, and Deep Blue won the decisive final game. The match was controversial — Kasparov disputed a key move in the second game, claiming it reflected a depth of positional understanding inconsistent with brute-force computation (he suspected human intervention), and IBM declined to release the complete game logs. The controversy simmered for years.

But the result stood. Deep Blue had defeated the world chess champion in a regulation match. The benchmark had been reached. Machines had surpassed human mastery at chess.

The Aftermath: Complexity of Feelings

The public reaction to Deep Blue's victory was complex. Many people felt a genuine unease — not because chess had been "solved" (Deep Blue was not close to solving chess in the mathematical sense) but because a form of human cognitive achievement had been surpassed by a machine. If computers could beat humans at chess, what exactly did that say about human intelligence?

Chess players, somewhat perversely, largely moved on. Kasparov continued playing chess. Grandmasters continued training. The world championship continued. What changed was the role of computers: rather than being opponents to be feared, they became tools. Engine-assisted preparation became standard. Players used engines to check their opening analysis, verify their tactical calculations, and identify their weaknesses. The machine became the training partner of every serious chess player in the world.


Part III: From Deep Blue to AlphaZero — A Revolution in Method (2017)

How Deep Blue Worked

Deep Blue was, at its core, a very fast search machine with a sophisticated hand-crafted evaluation function. Its ability to analyze hundreds of millions of positions per second allowed it to examine game trees to depths of ten to twenty plies in typical middlegame positions, using alpha-beta pruning to eliminate branches that could not possibly lead to better outcomes. Its evaluation function — the heuristic that assigned a score to each position — was crafted by grandmasters working with IBM engineers over years, encoding positional knowledge as explicit rules.

This approach — deep brute-force search with hand-crafted evaluation — produced formidable results. But it had fundamental limitations. The evaluation function was as good as the knowledge encoded in it, and there were positions whose subtlety exceeded what explicit rules could capture. More importantly, it was brittle: it played much stronger in the types of positions its evaluation function was designed for and weaker in positions with unusual structures or long-term considerations outside its encoded knowledge.

The Neural Network Revolution

The decisive methodological shift arrived in December 2017 when Google DeepMind published a paper describing AlphaZero — a chess-playing system that had learned chess entirely by playing against itself, using deep neural networks and reinforcement learning, with no human-provided knowledge beyond the rules of the game.

AlphaZero was given the rules of chess. It played itself millions of times. Through reinforcement learning — rewarding configurations that led to victories and penalizing those that led to losses — it developed its own evaluation function, encoded not as explicit rules but as the weights of a deep neural network. After approximately four hours of self-play training, it defeated Stockfish 8 — then the world's strongest traditional engine — in a 100-game match, winning 28, drawing 72, and losing none.

The chess world was stunned, not just by the results but by how AlphaZero played. Its games looked different from Stockfish's. Where Stockfish's play, despite its strength, sometimes had a mechanical quality — evaluating positions correctly but without the narrative coherence of human chess — AlphaZero played with what grandmasters described as genuine strategic intuition. It sacrificed material for long-term positional advantages in ways that looked counterintuitive by conventional evaluation standards but proved correct under deep calculation. It played, in the words of Kasparov himself, "a bit like a human — a superior human."

AlphaZero's Significance

AlphaZero's significance was not just its strength. It was its method. Deep Blue had proved that machines could surpass humans at chess through brute force plus human expertise encoded in evaluation functions. AlphaZero proved that machines could exceed even the best human-designed evaluation functions by learning their own evaluation from scratch, given only the game rules and millions of iterations of self-play.

This method — learning evaluation functions rather than encoding them — generalizes across games. AlphaZero was the same system that had previously mastered Go (as AlphaGo) and Shogi (Japanese chess). It learned each game the same way: self-play plus reinforcement learning. The implication was profound: given enough compute and self-play iterations, the same system could master any well-defined game.

Or could it? The AlphaZero method has a critical dependency that limits its applicability to new games — and this dependency is central to understanding AI's current relationship with BigChess.


Part IV: Modern Engines — Stockfish, Neural Networks, and Variants

The Current State of Chess Engines

Modern Stockfish (the open-source chess engine that has dominated computer chess since the early 2010s) has absorbed the neural network approach pioneered by AlphaZero. Stockfish's current incarnation, NNUE (Efficiently Updatable Neural Network), uses a neural network for evaluation in combination with its traditional alpha-beta search. The result is an engine that combines the speed advantage of classical search algorithms with the positional depth of neural network evaluation.

Stockfish's estimated Elo rating is approximately 3500 — more than 800 points above the strongest human players. The gap between engine and human performance at chess has become so large that direct human-versus-engine competition is essentially meaningless as a contest. Engines are used exclusively as training tools, analysis aids, and benchmarks for evaluating game quality.

The practical consequence for chess players is engine-assisted preparation and post-game analysis. Opening preparation with engine verification is standard at every level above club play. Post-game analysis using engines identifies all significant errors and missed opportunities. The engine is the ultimate arbiter of correctness in chess positions — a role it has occupied so completely that the question "was this move correct?" now means "does the engine agree?"

Engine Cheating Controversies

The dominance of engines has produced its most troubling side effect: engine-assisted cheating. A player using a phone or other device to consult a strong engine mid-game has access to superhuman-level analysis. The difference in playing strength between a strong human and an engine is so large that even infrequent consultation could decide games.

The most prominent cheating controversy in chess history emerged in 2022, when Magnus Carlsen withdrew from the Sinquefield Cup after losing to Hans Niemann, later alleging that Niemann had cheated in that game and "in many more games." Chess.com published a 72-page report citing statistical analysis of Niemann's online game history as evidence of cheating. The controversy generated legal proceedings that were eventually settled out of court, but it brought the engine-cheating problem to mainstream attention in a way that had never occurred before.

The structural problem is that detecting engine assistance is genuinely difficult. Statistical analysis of move quality compared to engine recommendations can identify suspicious patterns, but statistical anomalies occur naturally in real games. No purely statistical method produces certainty, and over-aggressive application produces false positives. The chess world continues to wrestle with detection methods, anti-cheating protocols, and the appropriate consequences for confirmed cheating.

Fairy-Stockfish: Engines for Chess Variants

The standard version of Stockfish evaluates classical chess positions only. It knows nothing about boards of non-standard sizes, pieces with non-standard movements, or rule modifications like the triple pawn advance. It would be of no use analyzing a BigChess position.

Fairy-Stockfish is the answer. An open-source variant engine built on the Stockfish codebase, Fairy-Stockfish supports an enormous range of chess variants by allowing flexible piece definitions, board sizes, and rule parameters. It can be configured to play and analyze Shogi, Xiangqi (Chinese chess), Crazyhouse, Atomic chess, and — crucially — BigChess.

Fairy-Stockfish's approach to variant pieces uses piece mobility tables: for each variant piece type, the engine is given the set of squares the piece can move to from any square, and it incorporates this mobility data into its evaluation. The Clone's diagonal slides and knight jumps can be represented in this framework, and Fairy-Stockfish can evaluate BigChess positions with reasonable competence.

However, there is a critical limitation: Fairy-Stockfish's evaluation is based primarily on piece mobility and material considerations without the deep structural knowledge that the NNUE neural networks in Stockfish provide for classical chess. The NNUE networks were trained on hundreds of millions of strong human and engine games in classical chess. No equivalent dataset exists for BigChess. The engine can play legal BigChess moves and make tactical evaluations, but it lacks the trained intuition that makes modern classical chess engines so powerful in positional assessment.


Part V: Why Training AI for BigChess Is Genuinely Hard

The Data Problem

Modern neural network chess engines — including AlphaZero and the NNUE variant of Stockfish — learn positional evaluation from large datasets of games. The specific learning method varies: AlphaZero learns entirely from self-play, while NNUE was trained on positions extracted from high-quality human and engine games. But in both cases, the quality of the learned evaluation function depends on the quality and quantity of the training data.

For classical chess, the training data problem is solved. There are hundreds of millions of recorded chess games, including millions at high quality from strong human and engine play. This enormous dataset allows neural networks to learn the subtle positional patterns that distinguish strong play from weak play — the evaluation of pawn structures, piece coordination, king safety, long-term initiative — from rich statistical signal.

For BigChess, this dataset does not exist. BigChess is a new game. The total number of recorded BigChess games, as of the game's current stage, is a small fraction of what would be needed to train a high-quality neural network evaluation function. Even a few million BigChess games — which would take years of active play to accumulate — would be modest compared to the datasets used for classical chess engine training.

Self-play training, as AlphaZero demonstrated, can sidestep the human game data requirement. An engine trained purely by playing against itself does not need recorded human games. But self-play training requires enormous computational resources — AlphaZero used specialized hardware at Google scale to achieve its results — and produces an engine that plays the game well according to game-theoretic criteria but may play in ways that are alien to human understanding.

The Clone Requires New Evaluation Knowledge

Even given sufficient training data, the Clone piece presents specific evaluation challenges that have no classical chess equivalent. Classical chess engine evaluation functions have been tuned over decades to accurately assess the value of bishops, knights, rooks, and queens in various position types. The relationship between piece values and position type is well understood: bishops are worth more in open positions, knights are worth more in closed positions, rook pairs are powerful in endgames, and so forth.

The Clone's evaluation is a genuinely open problem. How much is a Clone worth compared to a queen in a specific position type? How does Clone value change as the position opens or closes? When is Clone versus rook better for the Clone, and when for the rook? How much does a Clone outpost contribute to position evaluation compared to a knight outpost?

These questions cannot be answered by importing classical chess evaluation knowledge, because the Clone has no classical analogue. They must be learned from BigChess-specific data or derived theoretically from the Clone's movement properties. A well-trained BigChess engine will have internalized these evaluations; a poorly trained one will make systematic errors in Clone-heavy positions that a human familiar with Clone tactics will exploit.

What Fairy-Stockfish Analysis Reveals About BigChess Depth

Despite its limitations, Fairy-Stockfish's analysis of BigChess positions reveals something important about the game's depth: the 10×10 board and the Clone piece create positions that are extraordinarily complex to evaluate correctly even by strong computer analysis. The expanded game tree — resulting from more legal moves per position on the larger board — is more challenging to search comprehensively, and the Clone's dual-mode movement creates evaluation discontinuities that the engine's position-smoothing evaluation function handles imperfectly.

In practical terms, Fairy-Stockfish analysis of BigChess positions frequently identifies move sequences that surprise experienced chess players — not because they are obviously correct by classical chess standards, but because Clone tactics create combinations that require seeing the dual-mode movement clearly several moves ahead. This suggests that BigChess's combinatorial depth is genuine, not illusory: strong engine analysis consistently finds resource that human calculation misses, in the same way that classical chess engines find tactics in positions human grandmasters evaluate as quiet.


Part VI: The Future of AI in BigChess Development

Self-Play as the Path Forward

The most promising path toward a strong BigChess-specific AI evaluation function is AlphaZero-style self-play: an engine that learns BigChess evaluation purely by playing against itself, starting from only the rules and improving through millions of iterations of reinforcement learning. This method has the advantage of not requiring a historical game database and has been proven to work at superhuman levels for both classical chess and Go.

The practical obstacles are compute requirements and implementation complexity. AlphaZero's training required Google-scale hardware that is not readily accessible to an independent game project. More efficient training methods — using smaller neural networks, less compute, and smarter training curricula — have been developed in the years since AlphaZero's publication, and these may make strong self-play training more accessible for BigChess.

As BigChess's player community grows, the database of human games grows as well. When this database reaches sufficient size and quality, neural network training from human game data (the NNUE approach) becomes viable. The two approaches — self-play training and human game training — are complementary and could be combined to produce a BigChess engine that benefits from both the game-theoretic consistency of self-play and the human-relevant patterns in actual player games.

What a Strong BigChess Engine Would Show

A fully trained, human-quality BigChess engine would provide answers to questions that the BigChess community is currently exploring through practical play:

  • What are the optimal BigChess opening moves? What principles govern the best early Clone development and pawn structures?
  • What is the exact value of the Clone relative to other pieces in different position types?
  • Which endgames are won, drawn, or lost with optimal play? Clone versus rook, Clone versus queen, Clone pair versus various material configurations?
  • How much advantage does White (moving first) have in BigChess? Is it more or less than in classical chess?

These questions define the theoretical frontier of BigChess. Engine analysis will eventually answer them — but in the meantime, human BigChess players are exploring them through practical play, developing intuitions and heuristics that will remain valuable even after strong engine analysis is available.

AI as Training Partner

Beyond competitive use, the most important role of AI in BigChess development is as a training partner. The current Fairy-Stockfish integration in BigChess already serves this function, identifying tactical errors and missed opportunities in post-game analysis. As engine evaluation of BigChess improves, this training partnership will become more valuable.

The history of chess AI's impact on human chess suggests a consistent pattern: engines do not replace human chess but transform how humans learn it. Players who engage seriously with engine analysis of their games improve faster than those who don't. The engine's ability to identify specific errors and demonstrate correct continuations provides a precision of feedback that neither human coaches nor self-analysis alone can match.

This pattern will repeat in BigChess. The players who use Fairy-Stockfish analysis to identify Clone tactical errors, understand Clone evaluation in specific position types, and study the engine's suggestions for improving their play will develop faster than those who rely on intuition alone. AI does not reduce the value of human skill in BigChess — it amplifies the rate at which that skill can develop.

The Cheating-Prevention Challenge for BigChess

As Fairy-Stockfish's BigChess strength grows, the engine-assisted cheating problem that afflicts classical chess will eventually become relevant to BigChess as well. A player consulting a strong BigChess engine mid-game would have a decisive advantage that is both practically difficult to detect and deeply unfair to opponents.

BigChess's online-native design — all games played on a single platform with complete server-side game records — creates detection opportunities that over-the-board chess lacks. Statistical analysis of move quality versus engine agreement can be applied comprehensively to all BigChess games. As the player database grows and statistical baselines become clearer, anomalous performance patterns become more detectable.

The proactive investment in anti-cheating infrastructure, beginning before the cheating problem becomes acute, is one of the important governance challenges that BigChess will face as its community grows. The experience of classical chess — where the cheating problem was largely reactive rather than proactive — provides a clear cautionary example.


Part VII: AI, Chess, and What It All Means

Did Machines Break Chess?

The question that has haunted chess since 1997 — did computers break chess? — has a nuanced answer. Computers broke the illusion that chess was unsolvable by machines. They did not break chess as a human activity. Chess is played by more people today than at any point in its history. Online platforms serve hundreds of millions of games monthly. The game's popularity, measured by player counts, has grown dramatically in the years since Deep Blue's victory.

What changed is the game's relationship with perfection. In classical chess, the ideal of perfect play exists and is known to be achievable by machines. Human players are not playing perfect chess; they are playing human chess, which is beautiful and flawed and the expression of genuine cognitive effort against a similarly flawed opponent. That remains true and remains valuable even in the presence of engines that play better than any human.

BigChess Restores the Balance

In BigChess, the situation is different from classical chess in a specific and valuable way: no engine currently plays BigChess at a level that is clearly and definitively beyond human comprehension. Fairy-Stockfish plays BigChess, but without the trained neural network evaluation that makes classical chess engines incomprehensibly strong. A strong human BigChess player, with deep understanding of Clone tactics, 10×10 positional principles, and BigChess-specific endgame technique, may legitimately be able to identify and exploit evaluation weaknesses in the current generation of BigChess engines.

This is not a permanent situation — it will change as BigChess-specific engine training improves — but it is the current situation. BigChess is a game where human understanding is genuinely frontier knowledge, where no machine has definitively surpassed the best human players, and where the process of exploring the game's depths is a genuine shared discovery between human players and the evolving AI tools they use to study it.

The 1997 Deep Blue match asked what it meant for machines to surpass humans at the world's greatest intellectual game. BigChess asks a different question: what does it mean to play a great intellectual game where human understanding and machine analysis are still, for this brief and valuable moment, genuinely peers?


Conclusion: The Game the Machines Haven't Mastered Yet

The history of chess AI is a history of increasing machine competence against a fixed target: a game whose rules have been stable for five centuries and whose pattern space has been explored for just as long. Given enough compute, enough data, and enough time, machines solved the classical chess benchmark.

BigChess is a new target. The same tools that mastered classical chess — deep search, neural network evaluation, self-play training — can be applied to BigChess. But they must be applied specifically: new evaluation functions for the Clone, new opening theory from self-play or human game data, new endgame tablebases. The work has begun, through Fairy-Stockfish and the game analysis tools BigChess provides to its players. But it is far from complete.

In the meantime, the strongest BigChess players are the ones who understand the Clone most deeply, who have internalized the positional principles of the 10×10 board most thoroughly, and who have played enough BigChess games to build genuine pattern libraries for a game that is still new. That is not a consolation prize while waiting for the machines to catch up. That is the experience of playing a game at the frontier of human knowledge — which is exactly what chess felt like before the machines arrived.


Play where the machines haven't gotten ahead of you yet. Join BigChess at bigchessgame.com — available on iOS, Android, and web browser. The 10×10 board, the Clone, and the Fairy-Stockfish engine are ready for you. The age of exploration starts with your first game.

About the Author

Marat Fatalov

Marat Fatalov

Co-inventor of Big Chess

High School Student, Co-inventor of Big Chess, Second Category chess player.