AI Strategy
DeepMind’s David Silver Just Raised $1.1B to Build AI That Learns Without Human Data
9 min read · Published April 28, 2026 · Updated April 28, 2026
By CogLab Editorial Team · Reviewed by Knyckolas Sutherland
A former DeepMind researcher just put a huge number on a very old idea. TechCrunch says David Silver’s new company, Ineffable Intelligence, raised $1.1 billion to build an AI system that learns without human data. The bet is that reinforcement learning can do more than tune a model. It can discover the skills the model needs in the first place.
That sounds abstract until you compare it with the way most AI systems are trained today. The standard playbook leans on large piles of human-written text, human-labeled examples, or both. Silver is pointing at a different path. He wants systems that learn from experience, trial and error, and feedback loops instead of depending on a library of human examples as the main fuel.
TechCrunch says Ineffable’s goal is to build a superlearner that can discover knowledge and skills without relying on human data. That idea matters because it changes the bottleneck. If the best systems learn by interacting with environments, then the scarce resource is no longer just curated text. The scarce resource becomes the right experience, the right feedback, and the right way to let the system improve safely.
Why should everyday professionals care? Because the definition of useful AI could get broader very fast. A chat model is good at answering. A system trained through reinforcement can become better at choosing actions, exploring options, and improving through iteration. That is a different shape of product, and it is the shape most automation workflows actually want.
This also shifts the power of data. Companies have spent years treating clean datasets as the main moat. That still matters. It just stops being the whole story. If experience itself becomes the training fuel, then logs, simulations, workflows, and interactions start looking more valuable than static documents sitting in a drive folder.
Silver is not a random person making that claim. He spent more than a decade at DeepMind and helped build AlphaZero, the system that learned chess and Go by playing against itself rather than copying human playbooks. That history explains why this startup is not selling a softer version of chat. It is chasing the idea that intelligence can be discovered, not only imitated.
The funding round says investors think this path is worth a massive early bet. TechCrunch reports the company raised the money at a $5.1 billion valuation, with participation from names like Sequoia, Lightspeed, Google, Nvidia, and the British Business Bank. That list is a signal on its own. The market is willing to fund a different training philosophy before the product is fully visible.
For operators, the practical takeaway is to start thinking about experience as infrastructure. If your team wants better AI than the generic chat layer, you may need to design the environment the model learns from. That could mean simulations, workflow traces, feedback loops, or controlled tasks where the system can improve from trying, failing, and trying again.
There is also a strategic lesson for vendors. The next wave of AI systems may not be won by whoever has the biggest pile of scraped text. It may be won by whoever can create the richest learning environment and keep the loop running long enough for the model to get better at a real job.
That is why this raise matters beyond the size of the check. It marks a serious attempt to move AI training from imitation to discovery. If it works, the phrase training data will start to sound much narrower than it does now.
The big question is whether reinforcement learning can scale into everyday systems without turning into a science project. If it can, the companies that learn to feed their models experience instead of only examples will have a very different advantage.
Frequently Asked
What did David Silver’s new company raise?
TechCrunch says Ineffable Intelligence raised $1.1 billion at a $5.1 billion valuation to build AI that learns without human data.
Why does reinforcement learning matter here?
Because it lets systems improve through trial and error and feedback, which could reduce dependence on human-labeled examples and static datasets.
What should teams think about now?
Treat simulations, logs, and workflow traces as possible training fuel, not just record-keeping. The learning environment may become part of the product moat.
Sources
Related Articles
Services
Explore AI Coaching Programs
Solutions
Browse AI Systems by Team
Resources
Use Implementation Templates