What's Missing Between LLMs and AGI - Vishal Misra & Martin Casado...
a16z PodcastFull Title
What's Missing Between LLMs and AGI - Vishal Misra & Martin Casado
Summary
This episode explores the current limitations of Large Language Models (LLMs) in reaching Artificial General Intelligence (AGI), focusing on the mathematical underpinnings of LLM functionality.
Researchers Vishal Misra and Martin Casado discuss how LLMs currently operate based on correlation rather than causation, and identify key areas for future development to achieve true AGI, such as continued learning and causal reasoning.
Key Points
- Current LLMs function by predicting the next token based on learned correlations, not by understanding cause and effect, limiting their ability to achieve AGI.
- Vishal Misra's research, including the "Bayesian wind tunnel" concept, mathematically demonstrates that transformer architectures can perform Bayesian updating, essentially learning from new evidence within a given context, but this is not equivalent to true understanding or the ability to create new causal models.
- While LLMs excel at tasks involving association and pattern matching (Shannon entropy), they lack the ability for true simulation, intervention, and counterfactual reasoning (Kolmogorov complexity), which are hallmarks of human intelligence and crucial for AGI.
- The distinction between LLM learning and human learning lies in plasticity; LLMs' weights are frozen after training, while human brains remain plastic, continuously learning and adapting throughout life.
- Achieving AGI requires developing AI with continuous learning capabilities (plasticity) and the ability to move from correlation to causation, enabling true simulation and intervention.
Conclusion
Current LLMs are sophisticated correlation engines, not yet capable of true understanding or causal reasoning required for AGI.
Future progress towards AGI depends on developing AI with continuous learning capabilities and the ability to reason causally, moving beyond pattern matching.
The distinction between current LLMs and human intelligence lies in plasticity and the ability to create new conceptual frameworks, not just process existing data.
Discussion Topics
- How can we bridge the gap between LLMs' correlational abilities and the causal reasoning necessary for AGI?
- What are the most promising research directions for developing truly plastic AI systems that can learn continuously like humans?
- Beyond prediction, what new architectures or frameworks are needed to enable AI to generate novel scientific theories or conceptual models?
Key Terms
- AGI
- Artificial General Intelligence, a hypothetical type of intelligent agent that can understand or learn any intellectual task that a human being can.
- LLM
- Large Language Model, a type of artificial intelligence program that uses deep learning techniques and massive amounts of data to generate human-like text.
- In-context learning
- The ability of an LLM to learn a new task from a few examples provided within the prompt, without updating its model weights.
- Bayesian updating
- A statistical method that updates the probability of a hypothesis as more evidence or information becomes available.
- Transformer
- A deep learning model architecture that uses self-attention mechanisms, widely used in natural language processing tasks.
- Shannon entropy
- A measure of the uncertainty or randomness in a set of data, often related to the average amount of information produced by a stochastic source of data.
- Kolmogorov complexity
- A measure of the computational resources needed to specify a mathematical object, such as a string, by a computer program.
- Causation
- The relationship between a cause and its effect, where one event is a direct result of another.
- Correlation
- A statistical measure that describes the extent to which two or more variables fluctuate together.
- Plasticity
- In the context of AI, the ability of a system to adapt and learn continuously after its initial training.
Timeline
Discussion on the limitations of current LLMs and the path to AGI.
Introduction to Vishal Misra's work on modeling LLM functionality, starting with his initial research into in-context learning with GPT-3 and the development of a mathematical model.
Explanation of the matrix abstraction for LLMs and how they generate probability distributions for the next token.
Discussion of in-context learning as the LLM's ability to learn in real-time from provided examples, akin to Bayesian updating.
Explanation of the "Bayesian wind tunnel" concept and its use to mathematically prove that transformer architectures perform Bayesian updating.
The distinction between correlation (Shannon entropy) and causation (Kolmogorov complexity) in AI and human intelligence.
The identified requirements for AGI: continuous learning (plasticity) and causal reasoning.
Discussion on the limitations of current LLMs in creating new causal models and the need for a shift from correlation to causation.
The current limitations of LLMs in producing novel theoretical frameworks or "new manifolds" of understanding, unlike human scientific breakthroughs.
The future direction of research focusing on continuous learning and causal models.
Episode Details
- Podcast
- a16z Podcast
- Episode
- What's Missing Between LLMs and AGI - Vishal Misra & Martin Casado
- Official Link
- https://a16z.com/podcasts/a16z-podcast/
- Published
- March 17, 2026