When machines learn to learn

Vanshika GargVanshika Garg
11 min read

On December 5, 2017, something remarkable occurred in a peaceful research laboratory that would forever alter our comprehension of artificial intelligence. An AI system named AlphaZero was provided with nothing but the fundamental rules of the old game of Go

No examples from humans, no pre-loaded strategies, simply the rules.

Within only 72 hours of playing itself, it didn't merely equal the skills of human masters who had spent their lives studying the game. It crushed them.

It even defeated the reigning AI champion that had been laboriously trained on human data for years, employing tactics that experts said were "alien" and "like watching a player from the future." This was not merely a computer beating a game. It was a machine learning strategy in a way that made grandmasters go, “Wait… what just happened?” If this were a sci-fi movie, this would be the moment the soundtrack shifts and someone whispers, “We may have gone too far.”

Navigating the Unknown

When I heard for the first time about AlphaZero's accomplishment, I experienced that strange dizziness that comes from looking into a future coming sooner than anticipated. If a machine can master Go in a weekend, what happens when it sets its sights on medicine? Or climate science? Or running the economy?

…and what happens to us?

Are we just going to sit back and watch as machines out-teach us in real time? If the machines are going to teach themselves at a quicker pace than we learn, are we going to be just observers of our own technological revolution? How can we be certain these systems are human-aligned when they may one day be based on principles that we barely understand?

These aren't theoretical, abstract philosophical speculations any more. They are pressing questions about which researchers, policymakers, and citizens must now contend while there is still some hope of influencing the response.

What are these self-improving algorithms behind all the gravity of these questions, anyway? These self-improving algorithms aren’t just upgrades. They’re a whole new species of intelligence, one we may not even fully understand, yet. They are a break with how AI systems grow and change, a break that potentially redefines our relationship to technology and to intelligence itself.

And they’re not coming. They’re already here.

The Mechanics of Self-Improving Intelligence

Self-enhancing AI isn't some buzzword tech term, rather, it's a revolution in the way machines learn. While other AIs patiently sit around until humans tinker with them to improve them, self-enhancing programs can analyze their own failures and fix themselves. Think about it like the difference between a violin that must be tuned by its player versus a violin that could hear itself playing off-key and tune itself. AlphaGo Zero is the poster child for this ability. Told only the rules of Go, it played millions of games of Go against itself, refining strategies that defeated human masters and its own predecessor AlphaGo hands-down, all within three days. This was not simply a computer beating at a game; it was a computer learning to think in ways its own developers couldn't.

But how does a machine like AlphaGo Zero teach itself strategy from scratch? To understand that, we need to look under the hood, at the architecture powering its learning process.

A) Under the Hood: How Machines Learn to Think

At the heart of most deep learning systems lies a process that sounds deceptively simple: Forward Propagation. It’s the fundamental engine that lets a neural network take in information, process it layer by layer, and spit out a prediction. But beneath that simplicity is a cascade of calculations that mimic, in their own alien way, how we humans make decisions.

Imagine a neural network as a towering system of interconnected nodes, or Perceptrons. Each perceptron takes input from the layer before it, multiplies each input by a learned Weight, adds a Bias, and then pushes the result through an Activation Function, a kind of yes-or-no gate that helps the network decide what to keep, what to discard, and what to pass forward.

This ripple of calculation flows from the Input Layer through Hidden Layers to the Output Layer, and the entire process, this thinking cascade, is what we call Forward Propagation.

But AlphaZero didn’t just need to think. It had to see. It had to understand the shifting spatial patterns of a Go board or a chess game the way a human grandmaster might glance at the board and feel the weight of the future in a single shape.

That’s where Convolutional Neural Networks (CNNs) come in. An invention that dates back to 1989, when Yann LeCun gave machines a better way to interpret the visual world. CNNs are purpose-built for pattern recognition in grid-like data: images, game boards, or anything where space and shape matter.

A CNN is made of three key types of layers, each playing a different role in the machine’s perception:

  1. Convolutional Layers act like digital eyes. They slide tiny filters over the data, scanning for patterns: edges, corners, clusters, the way our brains recognize the shape of a face or the corner of a bishop’s move.

  2. Pooling Layers compress what’s been seen, keeping what matters and dropping what doesn’t. They help the system focus, distilling the data into its most meaningful essence.

  3. Fully Connected Layers come in at the end to pull everything together. They weigh the possibilities and make a decision, often with a final mathematical whisper like softmax, declaring what the network believes it just saw, or what move it should make.

This architecture is the canvas on which AlphaZero painted its alien genius. From medical diagnostics to video games to ancient board games, CNNs have become the lens through which machines begin to understand the world.

And with that lens sharpened, we can now look at how AlphaZero used it, not just to mimic intelligence, but to create something startlingly new.

B) How AlphaGo Zero Thinks

AlphaGo Zero isn’t just a faster engine or a smarter chess bot. It’s a completely different species of intelligence, one that sees, thinks, and evolves through a delicate dance between deep learning and tree search.

At the heart of its brilliance lies the fusion of two powerful components: a Convolutional Neural Network and a Monte Carlo Tree Search (MCTS). And what’s even more remarkable? It was trained entirely through self-play, no human data, no expert games, just pure, relentless iteration. A machine playing itself to perfection.

Before we dive deeper, let’s get acquainted with the language AlphaGo Zero thinks in:

  1. State (sₜ): This is the board at any point in time, from the opening move (s₀) to the final position (sₜ) where the game ends.

  2. Move (aₜ): At each state, the AI selects a move aₜ, its action, its decision, based on probability, not instinct.

  3. Search Probability (πₜ): This is where the Monte Carlo Tree comes in. πₜ is a probability distribution over possible moves, calculated through countless simulations. It’s how the AI chooses not just a move, but the best one.

  4. Monte Carlo Tree (αₜ(θ)): This is the AI’s mental model of the future, a branching tree of possible outcomes, weighted by their likelihood. It explores, evaluates, and narrows down on the path most likely to lead to victory.

  5. Convolutional Neural Network (fθ): This is AlphaGo Zero’s brain. It takes the raw board state as input and produces two crucial outputs:

    1. The Value Vector (vₜ): a prediction of who’s winning from this position. Not in terms of points, but in terms of destiny.

    2. The Policy Scalar (pₜ): a roadmap of which moves are promising, assigning a probability to each.

  6. Winner (z): At the end of the game, when the dust settles, z is determined, the final verdict. That outcome is then back-propagated through the network to refine its understanding, like a player reflecting on every decision made.

What’s groundbreaking here is not just the components, it’s how elegantly they loop together. The neural network guides the tree search. The tree search picks the next move. The result of the game trains the neural network. It’s a perfect feedback loop: play, learn, repeat. And with each cycle, the machine gets stronger, not by imitating humans, but by discovering strategies even we don’t fully understand.

This recursive self-enhancement is what binds these systems to the holy grail of Artificial General Intelligence. If one AI can expand its own abilities, and each improved variant can further enhance itself, then we may see an "intelligence explosion" where machine cognition outgrows human capabilities across thousands of domains. The stakes are as exciting as they are chastening.

AlphaGo Zero didn’t need millions of expert moves. It only needed the rules, the board, and time. And somehow, that was enough.

Promise and Threat

The age of self-improving AI is no longer confined to games. DeepMind has already expanded AlphaZero’s core ideas to real-world problems. AlphaFold revolutionized protein structure prediction, AlphaTensor optimized matrix multiplication, and AlphaDev discovered faster sorting algorithms. Their latest creation, AlphaEvolve, uses evolutionary methods to generate and refine code, a recursive loop of improvement guided only by outcomes. Each of these represents not just raw power, but a shift toward autonomy: systems that learn, evolve, and shape their own goals.

So what exactly does this emerging class of AI bring to the table, and what should we be wary of?

Strengths of self-improving AI:

  • Exponential problem-solving ability - It can catalytically speed up areas such as medical science by acquiring knowledge from immense amounts of data in the form of experiments, data points, and patient outcomes while constantly reconfiguring its learning method.

  • Faster and more efficient processing - AI computers can process much faster than humans, allowing for quick analysis and solution creation.

  • Improved pattern recognition - AI is able to recognize intricate patterns and come up with solutions which may take human generations to find out, not only working faster but perhaps wiser.

  • Solving high-complexity issues - AI is capable of solving gargantuan problems with variables so vast and interconnected that human minds cannot understand or solve them efficiently.

  • Self-improvement on a continuous basis - AI systems are able to improve their own learning capacity, connecting progressively better dots and gaining insights with each passing moment.

Weaknesses and Threats of AI:

  • Alignment uncertainty - While AI systems adapt their own goals and approaches, there is no way to ensure that they will still be aligned with human well-being and values.

  • Black box problem - AI systems improve to become more and more mysterious and opaque, such that their decision-making becomes hard to explain or predict.

  • Risk of unintended consequences - The disconnect between the capabilities of AI and what humans know poses risks of negative consequences not envisioned or prepared for.

  • Loss of human control - With increasingly advanced and self-modifying AI systems, human monitoring and intervention will become ever more difficult or even impossible.

Voices Shaping the Future

I looked up what some of the greatest minds of our generation have to say about self-improving AI, and no doubt, there’s an ongoing debate surrounding it. Sam Altman, the CEO of OpenAI, underlines that "if we can figure out the alignment problem, self-improving AI systems could help solve humanity's greatest challenges. If we can't, we're in trouble." Demis Hassabis,CEO of DeepMind Technologies, regards these systems as "potentially the most important technology humanity has ever developed," while still advocating for human values at their heart.

Approaching the topic with more caution, Eliezer Yudkowsky, American computer scientist and Researcher, cautions that "once a system is self-improving, human control becomes increasingly tenuous. We get one chance at designing the initial conditions and constraints correctly." Stuart Russell, A British computer scientist, suggests developing AI that is uncertain about human preferences and thus driven to learn from human feedback. Fei-Fei Li, Inventor of ImageNet and the Godmother of AI, reminds us all that "AI's purpose is to augment human capabilities, not replace them.".

Together, these voices paint a picture that is both inspiring and daunting. We stand on the brink of creating something with extraordinary power, one that requires careful guidance and responsibility to ensure it benefits humanity.

The Poetics and Perils of Self-Improving AI

Having traced this terrain of possibility and anxiety, I find myself on the brink, gazing out at a vista both thrilling and frightening. Self-improving AI is perhaps the most impactful technology we have ever considered. It might unleash cures for centuries-old diseases, solutions to global warming, and scientific breakthroughs that could rewrite our understanding of the cosmos.There is something deeply poetic about developing intelligence that is able to generate still more intelligence.

But I cannot help feeling trepidation. The difference between healthy self-betterment and out-of-control self-tweaking feels precariously thin. We're trying to build systems that will quickly work at levels of complexity we can hardly fully understand. This isn't fear of technology; it's a realization that intelligence, once let loose, can take on momentum that is hard to steer.

Beyond the Horizon of Human Understanding

What impresses me the most is that we are in a singular moment in human history: precarious, luminous, and irreversibly potent. We have the potential to be creating beings that someday will comprehend things that we cannot. There is awe in that, but also humility, and a quiet responsibility. As we venture further into this space, our work is not to fear the smarter, to ensure that as our intelligence grows, our wisdom deepens alongside it. The question is not if machines will get more intelligent than us

they probably will in most areas but if they'll reflect the values that make intelligence desirable in the first place.

As we stand on the brink of unprecedented technological advancements, it’s crucial to reflect on the kind of future we want to build. Can we instill in our machines the very best of human qualities, or is wisdom a trait that must remain uniquely human?

11
Subscribe to my newsletter

Read articles from Vanshika Garg directly inside your inbox. Subscribe to the newsletter, and don't miss out.

Written by

Vanshika Garg
Vanshika Garg