Abstract
How could machines learn as efficiently as humans and animals? How could machines learn to reason and plan? How could machines learn representations of percepts and action plans at multiple levels of abstraction, enabling them to reason, predict, and plan at multiple time horizons? This position paper proposes an architecture and training paradigms with which to construct autonomous intelligent agents. It combines concepts such as configurable predictive world model, behavior driven through intrinsic motivation, and hierarchical joint embedding architectures trained with self-supervised learning.
Meta's Chief AI Scientist Yann Lecun lays out his vision for what an architecture for generally intelligent agents might look like.
I haven’t read the whole paper, but my instinct after reading summaries and the first 12 or so pages is that this will be very difficult to implement with today’s (2025) hardware. So much is going on that is computationally complex. LLMs are all the rage because they are incredibly simple and incredibly scaleable. My hunch is that this will only be practical with a completely different computer paradigm such as neuromorphic computing and their highly efficient spiking neural networks. If we want AGI, money needs to be poured into research, not bigger and bigger LLMs and data centers. The payoff would be long term though, not short term, so I don’t think capital will be directed in a way conducive to this undertaking.