Why Cohere’s ex-AI research lead is betting against the scaling race


AI labs are racing to build data centers as large as Manhattan, each costing billions of dollars and consuming as much energy as a small city. The effort is driven by a deep belief in “scaling” — the idea that adding more computing power to existing AI training methods will eventually yield superintelligent systems capable of performing all kinds of tasks.

But a growing chorus of AI researchers say the scaling of large language models may be reaching its limits, and that other breakthroughs may be needed to improve AI performance.

That’s the bet Sara Hooker, Cohere’s former VP of AI Research and a Google Brain alumna, is taking with her new startup, Adaption Labs. She co-founded the company with fellow Cohere and Google veteran Sudip Roy, and it’s built on the idea that scaling LLMs has become an inefficient way to squeeze more performance out of AI models. Hooker, who left Cohere in August, quietly announced the startup this month to start recruiting more broadly.

In an interview with TechCrunch, Hooker says Adaption Labs is building AI systems that can continuously adapt and learn from their real-world experiences, and do so extremely efficiently. She declined to share details about the methods behind this approach or whether the company relies on LLMs or another architecture.

“There is a turning point now where it’s very clear that the formula of just scaling these models — scaling-pilled approaches, which are attractive but extremely boring — hasn’t produced intelligence that is able to navigate or interact with the world,” said Hooker.

Adapting is the “heart of learning,” according to Hooker. For example, stub your toe when you walk past your dining room table, and you’ll learn to step more carefully around it next time. AI labs have tried to capture this idea through reinforcement learning (RL), which allows AI models to learn from their mistakes in controlled settings. However, today’s RL methods don’t help AI models in production — meaning systems already being used by customers — to learn from their mistakes in real time. They just keep stubbing their toe.

Some AI labs offer consulting services to help enterprises fine-tune their AI models to their custom needs, but it comes at a price. OpenAI reportedly requires customers to spend upwards of $10 million with the company to offer its consulting services on fine-tuning.

Techcrunch event

San Francisco
|
October 27-29, 2025

“We have a handful of frontier labs that determine this set of AI models that are served the same way to everyone, and they’re very expensive to adapt,” said Hooker. “And actually, I think that doesn’t need to be true anymore, and AI systems can very efficiently learn from an environment. Proving that will completely change the dynamics of who gets to control and shape AI, and really, who these models serve at the end of the day.”

Adaption Labs is the latest sign that the industry’s faith in scaling LLMs is wavering. A recent paper from MIT researchers found that the world’s largest AI models may soon show diminishing returns. The vibes in San Francisco seem to be shifting, too. The AI world’s favorite podcaster, Dwarkesh Patel, recently hosted some unusually skeptical conversations with famous AI researchers.

Richard Sutton, a Turing award winner regarded as “the father of RL,” told Patel in September that LLMs can’t truly scale because they don’t learn from real world experience. This month, early OpenAI employee Andrej Karpathy told Patel he had reservations about the longterm potential of RL to improve AI models.

These types of fears aren’t unprecedented. In late 2024, some AI researchers raised concerns that scaling AI models through pretraining — in which AI models learn patterns from heaps of datasets — was hitting diminishing returns. Until then, pretraining had been the secret sauce for OpenAI and Google to improve their models.

Those pretraining scaling concerns are now showing up in the data, but the AI industry has found other ways to improve models. In 2025, breakthroughs around AI reasoning models, which take additional time and computational resources to work through problems before answering, have pushed the capabilities of AI models even further.

AI labs seem convinced that scaling up RL and AI reasoning models are the new frontier. OpenAI researchers previously told TechCrunch that they developed their first AI reasoning model, o1, because they thought it would scale up well. Meta and Periodic Labs researchers recently released a paper exploring how RL could scale performance further — a study that reportedly cost more than $4 million, underscoring how expensive current approaches remain.

Adaption Labs, by contrast, aims to find the next breakthrough, and prove that learning from experience can be far cheaper. The startup was in talks to raise a $20 million to $40 million seed round earlier this fall, according to three investors who reviewed its pitch decks. They say the round has since closed, though the final amount is unclear. Hooker declined to comment.

“We’re set up to be very ambitious,” said Hooker, when asked about her investors.

Hooker previously led Cohere Labs, where she trained small AI models for enterprise use cases. Compact AI systems now routinely outperform their larger counterparts on coding, math, and reasoning benchmarks — a trend Hooker wants to continue pushing on.

She also built a reputation for broadening access to AI research globally, hiring research talent from underrepresented regions such as Africa. While Adaption Labs will open a San Francisco office soon, Hooker says she plans to hire worldwide.

If Hooker and Adaption Labs are right about the limitations of scaling, the implications could be huge. Billions have already been invested in scaling LLMs, with the assumption that bigger models will lead to general intelligence. But it’s possible that true adaptive learning could prove not only more powerful — but far more efficient.

Marina Temkin contributed reporting.



1,000 artists release ‘silent’ album to protest UK copyright sell-out to AI


The U.K. government is pushing forward with plans to attract more AI companies to the region through changes to copyright law that would allow developers to train AI models on artists’ content on the internet — without permission or payment — unless creators proactively “opt out.” Not everyone is marching to the same beat, though.

On Monday, a group of 1,000 musicians released a “silent album,” protesting the planned changes. The album — titled “Is This What We Want?” — features tracks from Kate Bush, Imogen Heap, and contemporary classical composers Max Richter and Thomas Hewitt Jones, among others. It also features co-writing credits from hundreds more, including big names like Annie Lennox, Damon Albarn, Billy Ocean, The Clash, Mystery Jets, Yusuf / Cat Stevens, Riz Ahmed, Tori Amos, and Hans Zimmer. 

But this is not Band Aid part 2. And it’s not a collection of music. Instead, the artists have put together recordings of empty studios and performance spaces — a symbolic representation of what they believe will be the impact of the planned copyright law changes. 

“You can hear my cats moving around,” is how Hewitt Jones described his contribution to the album. “I have two cats in my studio who bother me all day when I’m working.”

To put an even more blunt point on it, the titles of the 12 tracks that make up the album spell out a message: “The British government must not legalize music theft to benefit AI companies.”

The album is just the latest move in the U.K. to bring attention to the issue of how copyright is being handled in AI training. Similar protests are underway in other markets, like the U.S., highlighting a global concern among artists.

Ed Newton-Rex, who organized the project, has simultaneously been leading a bigger campaign against AI training without licensing. A petition he started has now been signed by more than 47,000 writers, visual artists, actors, and others in the creative industries, with nearly 10,000 of them signing up in just the last five weeks since the U.K. government announced its big AI strategy. 

Newton-Rex said he has also been “running a nonprofit in AI for the last year where we’ve been certifying companies that basically don’t scrape and train on great work without permission.” 

Newton-Rex arrived at advocating for artists after having batted for both sides. Classically trained as a composer, he later built an AI-based music composition platform called Jukedeck that let people bypass using copyrighted works by creating their own. Its catchy pitch, where he rapped and riffed on the virtues of using AI to write music, won the TechCrunch Startup Battlefield competition in 2015. Jukedeck was eventually acquired by TikTok, where he worked for some time on music services. 

After several years at other tech companies like Snap and Stability, Newton-Rex is back to considering how to build the future without burning the past. He’s contemplating that idea from a pretty interesting vantage point: He now lives in the Bay Area with wife Alice Newton-Rex, VP of product at WhatsApp. 

The album release comes just ahead of the planned changes to copyright law in the U.K, which would force artists who do not want their work used for AI training purposes to proactively “opt out.”

Newton-Rex thinks this effectively creates a lose-lose situation for artists since there is no opt-out method in place, or any clear way of being able to track what specific material has been fed into any AI system. 

“We know that opt-out schemes are just not taken up,” he said. “This is just going to give 90% [to] 95% of people’s work to AI companies. That’s without a doubt.”

The solution, say the artists, is to produce work in other markets where there might be better protections for it. Hewitt Jones — who threw a working keyboard into a harbor in Kent at an in-person protest not long ago (he fished it out, broken, afterwards) — said he’s considering markets like Switzerland for distributing his music in the future. 

But the rock and hard place of a harbor in Kent are nothing compared to the Wild West of the internet. 

“We’ve been told for decades to share our work online because it’s good for exposure. But now AI companies and, incredibly, governments are turning around and saying, ‘Well, you put that online for free …” Newton-Rex said. “So now artists are just stopping making and sharing their work. A number of artists have contacted me to say this is what they’re doing.”

The album will be posted widely on music platforms sometime Tuesday, the organizers said, and any donations or proceeds from playing it will go to the charity Help Musicians.