spoonai
TOPAMI LabsJEPAFunding

Yann LeCun's AMI Labs Raises $1.03B Seed – Europe's Largest Ever – Betting Against Autoregressive Models

LeCun spins up AMI Labs to build world models with JEPA architecture, backed by Bezos, Nvidia, and Samsung – a fundamental bet against the current AI paradigm

Yann LeCun AMI Labs
Photo: AMI Labs

The Gamble: A Billion Dollars on a Different Path

Yann LeCun – Meta's AI chief and a Turing Award winner – just made a declaration: everything we've built over the past five years might be a dead end. And he's putting a billion dollars behind that bet.

In March, LeCun launched AMI Labs with $1.03 billion in seed funding. That's the largest seed round in European history. Ever. This isn't just another startup launch. This is a fundamental challenge to how the entire AI industry thinks.

The claim? Current large language models – the GPT line, Claude, all of them – are built on a faulty foundation. They're all trained the same way: predict the next token. And LeCun says that approach has hit a wall. Instead, he's betting everything on a completely different architecture called JEPA (Joint Embedding Predictive Architecture) that learns by building actual world models.

The Context: Why Autoregressive Models Hit a Ceiling

The past five years have been the age of the autoregressive language model. ChatGPT, GPT-4, Claude, Gemini – they all work on the same principle: predict what comes next.

Here's the magic: given billions of parameters and trillions of tokens of training data, this turns into something that can generate coherent text, reason about problems, and sometimes surprise you with its capabilities. It's an engineering miracle.

But LeCun sees a fundamental flaw. To predict the next token accurately, you have to compute very specific probabilities. Every. Single. Time. If I type "Today the weather is," the model has to calculate the probability that "sunny" comes next vs "rainy" vs "cloudy." That's computationally expensive, and it's the only way this architecture learns.

More fundamentally, autoregressive models struggle to learn how the physical world actually works. Drop a ball – what happens next? An autoregressive model treats that like a token prediction game: what's the next pixel value? That's not the same as understanding gravity.

LeCun has been explicit about this. He's said: "The next frontier of AI isn't predicting the next word. It's understanding how the world actually works. Building models that can learn the physics, the causality, the underlying mechanics of reality itself."

Approach Training Method Strength Limitation
Autoregressive (Current LLMs) Next token prediction Unbeatable at text generation Learns pixels, not physics; computationally expensive
JEPA (LeCun's Path) World state embedding Efficient, learns real-world dynamics Unproven at scale, early stage
Multimodal (Google/Meta hybrid) Image + text + audio Diverse input sources Fundamentally still autoregressive underneath

The Payload: What JEPA Actually Does

To understand JEPA, compare it directly to autoregressive prediction.

Autoregressive: "I see tokens A, B, C, D. What's next?" The model calculates: "E has a 90% probability, F has 8%..." Hyperspecific. Expensive.

JEPA: "I'll embed A, B, C, D into a high-dimensional space. In that abstract space, where will the next state probably land?" Instead of discrete tokens, you're working with continuous vectors.

Why is that better?

First, it's far more efficient. Computing probabilities for thousands of possible tokens is expensive. Predicting a vector in continuous space is cheap.

Second, it's better for learning physics. Show the model videos of objects moving, and it naturally learns momentum, gravity, collision. That's easier to capture in a vector space than in "what's the next pixel RGB value."

Third, you need way less data and compute to generalize. Because you're learning the actual structure of reality, not just token sequences.

The Money Behind the Bet

$1.03 billion is absurd for a seed round. Series B companies don't raise that much. But the backers are serious: Bezos, Nvidia, Samsung, Temasek (Singapore's sovereign wealth fund), plus top-tier VCs.

Why?

LeCun's credibility is enormous. He's a legend in computer vision and deep learning. When he says the current path is a dead end, people listen.

The timing matters too. Autoregressive models are hitting diminishing returns. Everyone feels it. GPT doesn't wow people like it used to. More tokens, more parameters – the gains are plateauing.

And the market need is real. Robotics, autonomous driving, manufacturing – these industries absolutely need AI that understands the physics of the world. A language model trained only on text will never cut it.

The Mission: General World Models in 5 Years

AMI's official goal is to build a general world model. That means: feed it video, audio, text, and it learns how reality works. Not just language generation. Not just pattern matching. Actual causal understanding of the physical world.

Concretely, they're targeting a prototype in 3 years, production-grade models in 5. That's insanely ambitious. Trillions of parameters, massive datasets of video/audio/text, and an entirely new architectural foundation.

The Landscape: A Fork in the Road?

This announcement signals something bigger: AI might be splitting into two competing tribes.

One side: OpenAI, Google, Anthropic, Meta (sort of). These companies are pushing autoregressive models to their limits. GPT-5.4 hitting 1 million tokens is proof. They're seeing how far this paradigm can stretch.

Other side: LeCun and AMI Labs. Complete paradigm shift. "Autoregressive is over. Let's build world models."

Who wins? Nobody knows. But the industry is hedging. Meta keeps improving autoregressive models while simultaneously funding LeCun to blow up the whole thing.

Camp Players Strategy 5-Year Goal
Autoregressive Maximalism OpenAI, Google, Anthropic Bigger models, more tokens, scaling Perfect language understanding
World Models AMI Labs, potential competitors Fundamental architecture shift Physical understanding, robot control
Hedging Meta, some others Both paths simultaneously Unknown territory

The Impact: What Changes in Practice

If you're building products on top of autoregressive LLMs right now? Relax. You've got at least three years before anything meaningfully disrupts your stack. AMI's world models won't be production-ready for a while.

But look five years out. If AMI Labs actually ships a working world model? Robotics, autonomous vehicles, industrial automation – these entire sectors could flip upside down. Tasks that are literally impossible with text-based AI become straightforward.

For enterprises, the calculus changes too. Today, you're picking between OpenAI, Google, and Anthropic. In two years, you might face a strategic question: "Do we bet on autoregressive AI or world models?" It's a different game.

Most interesting angle: these camps don't have to be enemies. Imagine a future where GPT-5.4 does high-level planning and reasoning, while AMI's world model handles physical execution and real-world feedback loops. They could be partners.

$1.03 billion isn't just money. It's a public declaration: "The age of predicting the next word is ending. The age of understanding the world is beginning."

What LeCun is saying, essentially, is that the past five years were warming up exercises. We built increasingly sophisticated pattern matchers. Useful? Yes. Impressive? Definitely. But ultimately limited.

The next frontier isn't bigger language models. It's AI that actually understands causality, physics, and how reality works. That shift, if it happens, will be bigger than the LLM revolution itself.

References

관련 기사

무료 뉴스레터

AI 트렌드를 앞서가세요

매일 아침, 엄선된 AI 뉴스를 받아보세요. 스팸 없음. 언제든 구독 취소.