How does Google know exactly which webpage you’re looking for out of trillions? How does your phone know the next word you’re about to type? And what does all of this have to do with a 100-year-old math feud over a Russian poem?
The answer to all these questions lies in a surprisingly simple and powerful idea: the Markov chain. And understanding this idea is key to understanding not only where AI came from, but where it might be taking us next.
Imagine trying to predict tomorrow's weather. You could look at the weather from a year ago, a month ago, or last week. But your best bet is probably just to look at the weather today. Is there a storm front moving in right now? What's the current temperature and pressure?

This is the core insight of a Markov chain, a concept pioneered by Russian mathematician Andrey Markov around 1906. He theorized that for many complex systems, you can make a pretty good prediction about the next step by only looking at the current state, ignoring almost everything that came before. It’s a "memoryless" way of looking at the world.
Markov proved his point by analyzing Alexander Pushkin's poem, "Eugene Onegin." He showed that the probability of the next letter being a vowel or a consonant depended heavily on the current letter. By creating a simple "chain" of probabilities (if the letter is a vowel, there's an X% chance the next is a consonant), he could model the structure of the Russian language.
This might seem like a simple academic exercise, but this idea—modeling complex dependent systems with simple, memoryless rules—would eventually be worth trillions.
Fast forward to the late 1990s. The internet was exploding, but finding anything was a mess. Early search engines just counted how many times a keyword appeared on a page, which was easy to cheat.
Two Stanford PhD students, Larry Page and Sergey Brin, had a better idea. What if the importance of a webpage wasn't just about its own content, but about how many other important pages linked to it?
They imagined a "random surfer" clicking links on the web. The pages this surfer spent the most time on would, by definition, be the most important. This is a perfect Markov chain:
The surfer's next move depends only on the links on their current page (the current state). They don't need to remember the 50 pages they visited before. By modeling the entire web as a massive Markov chain, Page and Brin created PageRank, the algorithm that powered Google and turned a sea of information into a searchable index. A 100-year-old piece of math from a Russian poem became the foundation of the modern internet.

But Markov's original idea was about predicting the next piece of text. This concept has evolved dramatically, leading directly to the large language models (LLMs) we use today.



So we've gone from the elegant simplicity of memoryless Markov chains to the complex, contextual power of LLMs with "attention." But this leap introduces a strange and powerful new dynamic: the feedback loop.
LLMs are now generating a massive amount of text, code, and images that are being published online. This AI-generated content is becoming part of the training data for the next generation of AI models. We've created a global feedback loop where AI starts learning from itself.
What happens when an AI's main source of information about the world is... content created by other AIs?
If we keep feeding AI its own output, where does this feedback loop lead? The future likely isn't one single outcome, but a splintering of possibilities based on how we manage these loops.

The simple, predictable math of the past gave us a powerful tool. The complex, attention-based systems of today give us a powerful but unpredictable partner. The question is no longer just "what's the next word," but who will have the vision to design the systems that decide which words come next?
It brings the original debate full circle. Markov's work began by questioning the nature of free will in a predictable world. The technology that followed now asks us what human vision and agency truly mean when we have the power to design the very systems that shape our reality.