AGI is a pipe dream until we solve one big problem, AI experts say, even as Google celebrates Gemini's success

AI brain coming out of laptop screen
(Image credit: Getty Images / Surasak Suwanmake)

  • AI researchers at NeurIPS 2025 say today’s scaling approach has hit its limit
  • Despite Gemini 3’s strong performance, experts argue that LLMs still can’t reason or understand cause and effect
  • AGI remains far off without a fundamental overhaul in how AI is built and trained

Recent successes by AI models like Gemini 3 don't disguise the more sobering message that emerged this week at the NeurIPS 2025 AI conference: that we might be building AI skyscrapers on intellectual sand.

While Google celebrated its latest model’s performance leap, researchers at the world’s biggest AI conference issued a warning: no matter how impressive the current crop of large language models may look, the dream of artificial general intelligence is slipping further away unless the field rethinks its entire foundation.

All agreed that simply scaling today’s transformer models, giving them more data, more GPUs, and more training time, is no longer delivering meaningful returns. The big leap from GPT‑3 to GPT‑4 is increasingly seen as a one-off; everything since has felt less like breaking glass ceilings than merely polishing the glass.

That’s a problem not just for researchers, but for everyone being sold the idea that AGI is around the corner. The truth, according to this year's scientific attendees, is far less cinematic. What we’ve built are highly articulate pattern-matchers. They’re good at producing answers that sound right. But sounding smart and being smart are two very different things, and NeurIPS made clear that the gap isn’t closing.

The technical term being passed around is the “scaling wall.” This is the idea that the current approach – train ever-larger models on ever-larger datasets – is running up against both physical and cognitive limits. We’re running out of high-quality human data. We’re burning enormous amounts of electricity to extract tiny marginal gains. And perhaps most troubling, the models still make the kind of mistakes that no one wants their doctor, pilot, or science lab to make.

It’s not that Gemini 3 hasn’t wowed people. And Google poured resources into optimizing model architecture and training techniques, rather than simply throwing more hardware at the problem, which makes it perform incredibly well. But Gemini 3’s dominance only underscored the problem. It’s still based on the same architecture that everyone is now quietly admitting isn’t built to scale to general intelligence – it’s just the best version of a fundamentally limited system.

Managing expectations

Among the most discussed alternatives were neurosymbolic architectures. These are hybrid systems that combine the statistical pattern recognition of deep learning with the structured logic of older symbolic AI.

Others advocated for “world models” that mimic how humans internally simulate cause and effect. If you ask one of today’s chatbots what happens if you drop a plate, it might write something poetic. But it has no internal sense of physics and no actual grasp of what happens next.

The proposals aren’t about making chatbots more charming; they’re about making AI systems trustworthy in environments where it matters. The idea of AGI has become a marketing term and a fundraising pitch. But if the smartest people in the room are saying we’re still missing the fundamental ingredients, it may be time to recalibrate expectations.

NeurIPS 2025 might be remembered not for what it showcased, but for admitting that the industry’s current trajectory is impressively profitable but intellectually stuck. To go further, we’ll need to abandon the idea that more is always better.


Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds. Make sure to click the Follow button!

And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form, and get regular updates from us on WhatsApp too.

Purple circle with the words Best business laptops in white
The best business laptops for all budgets
Eric Hal Schwartz
Contributor

Eric Hal Schwartz is a freelance writer for TechRadar with more than 15 years of experience covering the intersection of the world and technology. For the last five years, he served as head writer for Voicebot.ai and was on the leading edge of reporting on generative AI and large language models. He's since become an expert on the products of generative AI models, such as OpenAI’s ChatGPT, Anthropic’s Claude, Google Gemini, and every other synthetic media tool. His experience runs the gamut of media, including print, digital, broadcast, and live events. Now, he's continuing to tell the stories people want and need to hear about the rapidly evolving AI space and its impact on their lives. Eric is based in New York City.

You must confirm your public display name before commenting

Please logout and then login again, you will then be prompted to enter your display name.