News Items

News Items

The Moravec Paradox.

600,000 people have fled Kyiv.

John Ellis, Joanna Thompson, and Tom Smith
Jan 23, 2026
∙ Paid

Give a gift subscription


1. Yann LeCun:

There is a sense in which Large Language Models (LLMs) have not been overhyped, which is that they are extremely useful to a lot of people, particularly if you write text, do research, or write code. LLMs manipulate language really well. But people have had this illusion, or delusion, that it is a matter of time until we can scale them up to having human-level intelligence, and that is simply false.

The truly difficult part is understanding the real world. This is the Moravec Paradox (a phenomenon observed by the computer scientist Hans Moravec in 1988): What’s easy for us, like perception and navigation, is hard for computers, and vice versa. LLMs are limited to the discrete world of text. They can’t truly reason or plan, because they lack a model of the world. They can’t predict the consequences of their actions. This is why we don’t have a domestic robot that is as agile as a house cat, or a truly autonomous car.

We are going to have AI systems that have humanlike and human-level intelligence, but they’re not going to be built on LLMs, and it’s not going to happen next year or two years from now. It’s going to take a while. There are major conceptual breakthroughs that have to happen before we have AI systems that have human-level intelligence. And that is what I’ve been working on. And this company, AMI Labs, is focusing on the next generation. (Sources: amturing.acm.org, technologyreview.com. The MIT Tech Review’s interview with Mr. LeCun is worth reading in full)


2. Google DeepMind:

Anytime we look at the world, we perform an extraordinary feat of memory and prediction. We see and understand things as they are at a given moment in time, as they were a moment ago, and how they are going to be in the moment to follow. Our mental model of the world maintains a persistent representation of reality and we use that model to draw intuitive conclusions about the causal relationship between the past, present and future.

To help machines see the world more like we do, we can equip them with cameras, but that only solves the problem of input. To make sense of this input, computers must solve a complex, inverse problem: taking a video — which is a sequence of flat 2D projections — and recovering or understanding the rich, volumetric 3D world, in motion.

Today, we are introducing D4RT (Dynamic 4D Reconstruction and Tracking), a new AI model that unifies dynamic scene reconstruction into a single, efficient framework, bringing us closer to the next frontier of artificial intelligence: total perception of our dynamic reality. (Sources: deepmind.google, d4rt-paper.github.io)


User's avatar

Continue reading this post for free, courtesy of John Ellis.

Or purchase a paid subscription.
Joanna Thompson's avatar
A guest post by
Joanna Thompson
Science journalist, runner, bookworm, reptile enthusiast. Oxford comma for life.
Subscribe to Joanna
Tom Smith's avatar
A guest post by
Tom Smith
Research director and associate editor for News Items and Political Items
Subscribe to Tom
© 2026 John Ellis · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture