David Silver came out of stealth this morning with $1.1 billion in seed money and a thesis that runs against the prevailing weather. Ineffable Intelligence, his new lab, plans to build models that learn from their own experience rather than from the internet's archive of human writing. The round, reported by CNBC, is the largest seed in European history. Sequoia and Lightspeed co-led it. Nvidia, Google, DST Global, Index, and the UK Sovereign AI Fund piled in. The post-money valuation is $5.1 billion before there is a product, a paper, or, as far as anyone outside the cap table can tell, a model.
Silver is not a celebrity outside the field, but inside it he is the person who designed AlphaGo and AlphaZero. The work that made DeepMind's name in 2016 was reinforcement learning, agents that played themselves until they were better than anyone alive. AlphaZero learned chess from scratch in nine hours and then beat Stockfish, which had spent two decades absorbing every game humans had ever recorded. The lesson Silver took from that, and which he is now raising a billion dollars on, is that human data is a ceiling. Anything an AI can only learn from us cannot, by definition, exceed us.
The pitch deck almost writes itself. The frontier labs are spending the GDP of small countries on training runs that cannot continue at this clip, because the supply of high-quality human text is genuinely close to exhausted. The fix the industry has settled on is to pay forty labelers to write rubrics and hope the model generalises. Silver's bet is that the whole labelling layer is a distraction, that the actual gradient runs the other way, that a system which generates its own curriculum from interaction with an environment will skip the human bottleneck entirely. "Our mission is to make first contact with superintelligence," is how he put it in the press release. I flinched a little at the phrasing, but the technical claim underneath is real and not new. He has been making it for years.
What is new is that Sequoia is now writing a nine-figure cheque to fund the experiment in public. The talent flight from the big labs has been gathering for eighteen months. Mira Murati's Thinking Machines, Ilya Sutskever's Safe Superintelligence, and now Silver. Each one walks out of a hyperscaler with a plausible technical story, a Rolodex full of researchers, and a venture market willing to fund a five-billion-dollar valuation on day zero. The implicit critique is the same in every case: whatever the labs are doing inside, it isn't ambitious enough.
I am not sure the experience-based path scales the way the self-play games did. Chess and Go are closed worlds with crisp reward signals. The physical world is messy, slow, and expensive to simulate at fidelity. Reward hacking is the default outcome, not the edge case. The companies that have tried to do open-ended RL at the scale of, say, robotics, or science, or software engineering, have spent years learning how hard the reward-design problem actually is. None of that is solved by money, although money buys time to keep trying.
Still, a billion dollars is a real signal, and the people writing the cheques are not naive. The bet, as I read it, is not that Silver will reach superintelligence. The bet is that he will produce one or two genuinely novel results within three years that the existing labs cannot replicate without rebuilding their training stacks from the ground up. That is enough to clear the seed. Whether it is enough to clear the next round, when the science gets specific and the costs go vertical, is a question for 2028.
Sources: