What's Missing Between LLMs and AGI - Vishal Misra & Martin Casado
What's Missing Between LLMs and AGI - Vishal Misra & Martin Casado
Podcast47 min 35 sec
Listen to Episode
Note: AI-generated summary based on third-party content. Not financial advice. Read more.
Quick Insights

Investors should prioritize companies developing Causal AI and "Post-Transformer" architectures, as current models face a "scaling wall" due to their inability to understand cause-and-effect. While Anthropic (Private) and OpenAI (Private) remain leaders in utility, be wary of "consciousness" hype and focus instead on their efficiency as mathematical engines. Meta (META) and other open-source contributors are critical long-term plays, as their transparent models are essential for the R&D required to reach the next stage of Artificial General Intelligence. Look for opportunities in Retrieval-Augmented Generation (RAG) infrastructure and "front-end" startups that use LLMs to translate complex, domain-specific legacy databases. High-end compute providers remain the primary gatekeepers, making continued exposure to the GPU and AI infrastructure sector a high-conviction necessity for capturing the next architectural breakthrough.

Detailed Analysis

This investment analysis explores the technical evolution of Large Language Models (LLMs) and the path toward Artificial General Intelligence (AGI) based on the discussion between Vishal Misra (Columbia University) and Martin Casado (Andreessen Horowitz).


Artificial Intelligence Sector (LLMs & AGI)

The discussion centers on the mathematical reality of how LLMs function, moving from empirical observation to formal proof that these models operate as Bayesian inference engines.

  • The "Bayesian" Breakthrough: Research confirms that Transformers, the architecture behind ChatGPT and Claude, function by updating their internal probability distributions (posteriors) based on new evidence (prompts).
    • In controlled "Wind Tunnel" experiments, Transformers matched theoretically perfect mathematical answers to an accuracy of 10^-3 bits.
    • Architecture Hierarchy: The study found that Transformers are the most efficient at this mathematical updating, followed by Mamba and LSTMs, while MLPs (Multi-Layer Perceptrons) failed the tasks entirely.
  • The "Wall" for Current LLMs: Despite their power, current models are described as "grains of silicon doing matrix multiplication." They lack:
    • Plasticity: Human brains learn and change constantly; LLM weights are "frozen" after training.
    • Causality: Current AI excels at Correlation (identifying patterns) but fails at Causation (understanding why things happen or simulating "what if" scenarios).
  • AGI Requirements: For the sector to reach true AGI, the experts argue two things must happen:
    1. A move from Shannon Entropy (predicting the next token) to Kolmogorov Complexity (finding the shortest, most logical rule/program to explain data).
    2. The implementation of Continual Learning without "catastrophic forgetting" (the tendency of AI to lose old knowledge when learning new data).

Takeaways

  • Investment Theme: Investors should look for "Post-Transformer" architectures. While current LLMs are economically useful, the next massive valuation leap will likely come from companies solving Causal AI and Plasticity.
  • Scaling Limits: The transcript suggests that "Scale will not solve everything." Simply adding more data or more GPUs to current architectures may result in diminishing returns regarding actual "intelligence."
  • The "Einstein Test": A benchmark for true AGI investment: Can the model be trained on pre-1911 physics and independently derive the Theory of Relativity? Current models cannot, as they are "bound to the manifold" of their training data.

Anthropic (Private)

The podcast briefly discusses the current state of top-tier AI products and the rhetoric coming from industry leaders.

  • Product Quality: Mentioned as making "great products," specifically highlighting Claude and its coding capabilities as "fantastic."
  • Sentiment Check: The speakers expressed skepticism toward Anthropic CEO Dario Amodei’s alleged comments regarding AI consciousness.
    • Insight: The experts firmly categorize these models as mathematical tools, not conscious entities. This suggests a potential "hype bubble" regarding the "sentience" of AI that investors should be wary of.

Takeaways

  • Bullish on Utility: Anthropic remains a leader in functional, "economically useful" AI.
  • Risk Factor: There is a disconnect between the marketing/philosophical claims of AI "consciousness" and the underlying mathematical reality (matrix multiplication). Investors should value these companies based on utility and efficiency, not "consciousness."

OpenAI (Private)

The discussion touches on the shift in the industry from open-source transparency to closed-door development.

  • Data Access: OpenAI previously allowed users to see the probability distributions of tokens (how the AI "thinks"), but they have since removed this feature.
  • Early Innovation: The transcript credits the first implementation of RAG (Retrieval-Augmented Generation) to work done using GPT-3 in 2020 for ESPN.

Takeaways

  • RAG Dominance: RAG (the process of giving an AI a specific database to look at) is confirmed as the primary way LLMs solve domain-specific problems (like the Cricinfo database example). Companies specializing in RAG infrastructure remain highly relevant.

Key Investment Themes & Sectors

1. Domain-Specific Languages (DSL)

  • The podcast highlights how LLMs can be "taught" a brand-new language in real-time via the prompt (In-Context Learning).
  • Insight: There is significant value in companies building "front-ends" for complex legacy databases (like SQL or specialized sports/medical data) using LLMs as translators.

2. Compute & Infrastructure

  • The research mentioned was powered by clusters provided by Andreessen Horowitz (a16z).
  • Insight: Access to high-end GPUs remains the "gatekeeper" for academic and commercial breakthroughs in AI architecture.

3. Open Source vs. Closed Source

  • The researchers used open-source models to "peer inside" and prove the Bayesian nature of AI because frontier models (like those from OpenAI) are increasingly opaque.
  • Insight: Open-source models (like Meta's Llama) are essential for the R&D that will lead to the next generation of AI, making them a critical part of the ecosystem's health.
Ask about this postAnswers are grounded in this post's content.
Episode Description
Vishal Misra returns to explain his latest research on how LLMs actually work under the hood. He walks through experiments showing that transformers update their predictions in a precise, mathematically predictable way as they process new information, explains why this still doesn't mean they're conscious, and describes what's actually required for AGI: the ability to keep learning after training and the move from pattern matching to understanding cause and effect.   Resources: Follow Vishal Misra on X: https://x.com/vishalmisra  https://x.com/martin_casado   Stay Updated: Find a16z on YouTube: YouTube Find a16z on X Find a16z on LinkedIn Listen to the a16z Show on Spotify Listen to the a16z Show on Apple Podcasts Follow our host: https://twitter.com/eriktorenberg   Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures. Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.
About a16z Podcast
a16z Podcast

a16z Podcast

By Andreessen Horowitz

The a16z Podcast discusses tech and culture trends, news, and the future – especially as ‘software eats the world’. It features industry experts, business leaders, and other interesting thinkers and voices from around the world. This podcast is produced by Andreessen Horowitz (aka “a16z”), a Silicon Valley-based venture capital firm. Multiple episodes are released every week; visit a16z.com for more details and to sign up for our newsletters and other content as well!