███╗ ███╗███████╗███╗ ██╗ █████╗ ██████╗███████╗ ████╗ ████║██╔════╝████╗ ██║██╔══██╗██╔════╝██╔════╝ ██╔████╔██║█████╗ ██╔██╗ ██║███████║██║ █████╗ ██║╚██╔╝██║██╔══╝ ██║╚██╗██║██╔══██║██║ ██╔══╝ ██║ ╚═╝ ██║███████╗██║ ╚████║██║ ██║╚██████╗███████╗ ╚═╝ ╚═╝╚══════╝╚═╝ ╚═══╝╚═╝ ╚═╝ ╚═════╝╚══════╝
In 1961, Donald Michie built a tic-tac-toe playing "machine" out of 304 matchboxes and colored beads. No computers. No code. Just physical objects and a simple learning rule.
It learned to play tic-tac-toe through trial and error – and it worked.
I recently implemented MENACE (Machine Educable Noughts And Crosses Engine) in TypeScript, and the experience taught me more about reinforcement learning than any neural network tutorial ever did.
The physical MENACE was beautifully simple:
That's it. No gradients. No backpropagation. No neural networks. Just beads and matchboxes.
And it learned. After about 200 games, MENACE became nearly unbeatable.
The core idea translates beautifully to code. Instead of matchboxes and beads, we use objects and numbers.
The key insight: the number of beads = the probability weight for that move.
If center square has 10 beads and corner has 5, the AI is twice as likely to play center.
One of the coolest parts: the brain persists across sessions. When you close the page and come back, MENACE remembers everything it learned.
This creates a powerful loop: the more you play, the smarter it gets. Leave it running for a few hundred games, and it becomes nearly impossible to beat.
Modern AI has gotten incredibly complex. But at its core, it's still doing what MENACE did: trying things, seeing what works, and adjusting probabilities.
We've added layers of sophistication – neural networks, backpropagation, gradient descent – but the fundamental idea remains: learn from experience.
I've built a live implementation you can play with. Watch MENACE learn in real-time:
Pro tip: Let it lose the first 20-30 games. Watch how it adapts. By game 100, you'll struggle to beat it.
That's reinforcement learning in action – no code changes, just experience.
Building MENACE was one of those rare projects where the implementation was even more educational than I expected.
It's a reminder that AI doesn't have to be black-box neural networks trained on millions of examples. Sometimes the simplest algorithms – the ones you can explain with matchboxes and beads – are the most elegant.
And in 2025, when we're drowning in transformer models and billion-parameter networks, there's something refreshing about an AI you can fully understand in an afternoon.
Donald Michie was onto something in 1961. We'd do well to remember it.