While tech enthusiasts celebrate the latest advancements in large language models, AI pioneer Yann LeCun isn’t joining the party.

He’s got some harsh news for the cheerleaders: auto-regressive LLMs are hitting a wall. These models, for all their flashy abilities to spit out Shakespeare-like sonnets or code snippets, fundamentally lack true reasoning capabilities. They’re impressive party tricks, not the path to artificial general intelligence.

The problem? Auto-regressive models can’t plan worth a damn. They generate text one token at a time without any real ability to look back and say, “Wait, that’s wrong.” No self-verification, no error correction. Just words following words following words. Statistics, not thought. Pretty depressing when you think about it.

These systems might fool you into thinking they’re smart. They’re not. They’re just really good at pattern matching from massive training datasets. What looks like brilliance is actually just probability distribution at work. Smoke and mirrors, folks.

LeCun predicts we’ll need entirely different architectures to move forward. Future systems will need built-in mechanisms for reasoning, planning, and fact-checking. The current approach? Dead end. Full stop. Instead, LeCun favors the JEPA architecture for its superior abstract memory capabilities that focus on essential information rather than mere word prediction.

The models’ knowledge representation is another joke. They’re approximating knowledge without truly retaining context. Try having a lengthy conversation with ChatGPT and watch it forget what you discussed ten minutes ago. Not exactly the foundation for true intelligence.

Naturally, not everyone in the AI community agrees with LeCun. Some researchers think auto-regressive models still have untapped potential. The debate rages on in academic circles and Twitter threads alike.

What’s missing is something fundamental: the ability to correct errors. Evolution figured this out billions of years ago with natural selection. Our fancy AI models? Still working on it.

The problem becomes more evident as sequences get longer, with LLMs experiencing accumulated errors when generating text, making mistakes that build upon previous ones as they drift from the original training distribution.

The bottom line is clear: despite the hype and investment pouring into auto-regressive LLMs, they’re ultimately a technological cul-de-sac. If we want AGI, we need to find a different road.