Artificial Intelligence March 14, 2026

AMI Labs raises $1.03B as Yann LeCun backs world models over revenue

Yann LeCun’s new company, AMI Labs, has raised $1.03 billion at a $3.5 billion pre-money valuation to build world models. That's a huge round for a company openly saying it won't chase near-term revenue, and it says a lot about where serious AI money...

AMI Labs raises $1.03B as Yann LeCun backs world models over revenue

AMI Labs lands $1.03B to bet on world models, and that changes the AI conversation

Yann LeCun’s new company, AMI Labs, has raised $1.03 billion at a $3.5 billion pre-money valuation to build world models. That's a huge round for a company openly saying it won't chase near-term revenue, and it says a lot about where serious AI money is heading.

Some investors no longer buy the idea that bigger chatbots alone get you to systems that understand, predict, and act in the physical world. AMI is going after that problem instead.

The company was co-founded by LeCun and is led by CEO Alexandre LeBrun, with backing from Cathay Innovation, Greycroft, Hiro Capital, HV Capital, Bezos Expeditions, and strategic investors including Nvidia, Samsung, Sea, Temasek, and Toyota Ventures. The research bench is strong too: Saining Xie is chief science officer, Pascale Fung is chief research and innovation officer, and Michael Rabbat is VP of world models. Hiring is centered in Paris, New York, Montreal, and Singapore.

That's the financing news. The technical bet matters more.

Why the funding matters

AMI's pitch is straightforward. Today's LLM-centric AI is good with language, weak on grounded understanding, and brittle when decisions depend on how the world changes over time.

Healthcare makes the point quickly. AMI's first disclosed partner is Nabla, LeBrun's former digital health startup. If you're building systems around patient trajectories, test recommendations, or clinical risk, polished prose doesn't help much. The model has to stay tied to observed reality, uncertainty, and time. Hallucinations aren't a UX problem in that setting. They're a liability.

That's the case for world models. The idea predates the current AI boom, but the ingredients are better now: more compute, more multimodal data, and growing frustration with text-only priors as a path to general intelligence.

AMI joins a small, well-funded group pushing in the same direction, including Fei-Fei Li's World Labs, which also raised around $1 billion. Investors are betting that the next serious AI race involves systems that model environments, anticipate outcomes, and support planning.

JEPA is the technical bet

At the center is JEPA, short for Joint Embedding Predictive Architecture, which LeCun proposed in 2022.

The shift is in the training target.

A standard autoregressive language model predicts the next token. Many generative vision models reconstruct pixels or patches. JEPA does something else. It predicts a latent representation of future or missing data instead of reconstructing low-level detail.

That sounds abstract, but the logic is simple. If you want a system to understand a scene, forcing it to reproduce every pixel, texture, and lighting artifact can waste capacity. A predictive latent objective pushes the model toward structure that survives viewpoint changes, noise, and irrelevant detail.

A JEPA-style setup usually has four parts:

  • a context encoder that turns current observations into latent state
  • a target encoder that represents future observations in the same latent space
  • a predictor that maps from context to expected future latent
  • a critic or scoring function that checks whether the predicted latent matches the actual one

The point is what the model can ignore. It doesn't have to generate the full sensory stream token by token or pixel by pixel. That can produce representations that are more useful semantically, and sometimes more sample-efficient too.

Anyone coming from reinforcement learning will recognize the family resemblance. World models, latent dynamics, planning in compressed state spaces, that line runs through PlaNet, Dreamer, MuZero, and TD-MPC. JEPA puts the emphasis on self-supervised representation prediction across messy multimodal data, without making reconstruction the main objective.

It's a serious research gamble. It also makes sense.

Where LLMs struggle

LLMs can be turned into useful systems with retrieval, tool use, agents, and domain-specific pipelines. That works well for software, support, document workflows, and code generation.

Still, the base objective matters.

A model trained on internet text and code learns statistical structure in language. It picks up a lot of implicit world knowledge that way, but it does not learn causality, dynamics, or action-conditioned change the same way a system trained on video, telemetry, and sensor streams can. Synthetic traces and chain-of-thought scaffolding help. They don't cover all of it.

If your system needs to reason about how a patient state evolves, how a robot arm moves after a control input, or how an environment changes under intervention, next-token prediction starts to look like an awkward foundation.

World models probably won't replace LLMs. They'll sit underneath them or beside them. Language is still the easiest interface for humans. But the model tracking state, predicting outcomes, and estimating uncertainty may look nothing like a chatbot with tools bolted on.

The engineering load is real

This is expensive work.

Training a serious world model moves the bottleneck from token throughput to multimodal temporal data. That means:

  • continuous video streams
  • synchronized sensor logs
  • event sequences
  • action traces
  • domain-specific records such as EHR timelines

Storage and I/O get ugly fast. Data pipelines matter as much as model architecture. Teams need timestamp alignment, missing-data handling, chunked storage formats, and preprocessing that preserves temporal structure. Parquet helps for tabular event streams. Zarr or similar formats make sense for large array data. This is not a case of dumping a corpus into object storage and hitting train.

Compute changes too. Long-context temporal models over video and sensor data are brutal. Mixed precision, FP8, fast interconnects like NVLink and InfiniBand, and careful dataloader design are table stakes. Nvidia's spot on the cap table is a practical signal. This category burns hardware.

Evaluation gets harder as well. Perplexity and leaderboard trivia don't tell you much. For world models, you care about rollout accuracy, calibration, control performance, out-of-distribution behavior, and whether downstream planners actually improve with the learned latent state. In healthcare, add intervention safety, error bounds, and clinical outcome proxies. Those metrics are harder to define and much harder to fake.

Better safety in some ways, not all

LeBrun has framed AMI's approach as a way around hallucination-heavy systems. There's some truth in that.

A grounded predictive model trained on constrained multimodal distributions should be less likely to invent free-floating facts than a general-purpose chatbot trained to keep talking. If the task is "predict likely future patient state given this trajectory" instead of "produce a plausible answer in English," you've removed one common failure mode.

Still, nobody should oversell it. Once a world model is wrapped in a language layer, exposed through an assistant, or connected to decision support, familiar problems return: calibration, retrieval quality, uncertainty communication, bad tool calls, and human over-trust. Grounding helps. It doesn't solve safety at the system level.

World models also fail in their own ways. They can overfit narrow environments and break when conditions change. Robotics and autonomous driving have been demonstrating that for years. Distribution drift is still there waiting for you.

Why technical teams should pay attention

Most teams won't build a foundational world model from scratch. AMI may open source meaningful parts of its work, and that would matter more than the funding headline. But the direction is already useful.

A few things stand out.

First, data advantage is shifting. Proprietary text still matters, but high-quality, time-aligned multimodal data could become the stronger moat in some industries. Companies with years of fleet telemetry, industrial video, medical timelines, or rich sensor history just became more interesting.

Second, the stack is changing. Teams moving into this area need stronger temporal data infrastructure, better simulation pipelines, and tighter links between representation learning and downstream control or forecasting. That's a different skill set from prompt engineering and eval harnesses for chat apps.

Third, interfaces will split from cognition. The user-facing layer may still be a language model because English is convenient. Under the hood, the system doing the important work may be a latent dynamics model that never chats at all. Product teams should plan around that split now.

Fourth, healthcare, robotics, and industrial AI are obvious early targets because language-model confabulation is expensive there and temporal prediction has direct value. That doesn't make adoption easy. It means the pain is bad enough to justify the engineering.

A long bet

AMI Labs says it won't chase revenue in the near term. That's both refreshing and risky. The company now has the money to take a real run at foundational research, and LeCun has been making the same argument for years: intelligence needs models of the world, not just bigger autocomplete.

He may be right. The open question is whether JEPA-style systems can scale into something broadly useful outside research demos and narrow verticals. That's still unproven. AI history is full of ideas that were directionally right and early by a decade.

This round lands at a point where the limits of text-first AI are easier to see. For engineers building systems that have to deal with time, uncertainty, and the physical world, that's the part worth paying attention to.

Keep going from here

Useful next reads and implementation paths

If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.

Relevant service
Data science and analytics

Turn data into forecasting, experimentation, dashboards, and decision support.

Related proof
Growth analytics platform

How a growth analytics platform reduced decision lag across teams.

Related article
AMI Labs emerges with Yann LeCun's world model approach to AI

Yann LeCun has spent years arguing that next-token prediction won’t get AI very far in the physical world. Now there’s a startup built around that view. AMI Labs, short for Advanced Machine Intelligence, has surfaced with a plan to build foundational...

Related article
Yann LeCun’s reported Meta exit puts world models at the center of AI

Yann LeCun is reportedly preparing to leave Meta and start a company focused on world models. If that happens, it lands as a management story, a research story, and a product story at the same time. At Meta, LeCun has been the clearest internal criti...

Related article
AI is starting to matter in proofs of hard open math problems

AI-assisted math results used to sound like stunts. That’s getting harder to say. Since Christmas, 15 Erdős-style open problems have reportedly been moved into the solved column, and 11 of those involved AI in some meaningful way. Terence Tao has bee...