Dwarkesh and Ilya Sutskever on What Comes After Scaling
Dwarkesh and Ilya Sutskever on What Comes After Scaling
Podcast1 hr 32 min
Listen to Episode
Note: AI-generated summary based on third-party content. Not financial advice. Read more.
Quick Insights

The relentless demand for AI compute presents a powerful, long-term investment trend for the foreseeable future. Consider investing in the key infrastructure providers benefiting from this, such as GPU makers Nvidia and AMD, chip foundry TSMC, and major cloud providers Amazon (AWS) and Microsoft (Azure). A key growth driver is the explosion in inference workloads, which is the cost of running AI models for end-users. Separately, Meta's (META) aggressive acquisition of top AI talent is a strong bullish signal, demonstrating its commitment to becoming a leader in the space. While current AI leaders are dominant, monitor their progress for signs of a technological plateau, which could create opportunities for new disruptors.

Detailed Analysis

SSI (Safe Superintelligence Inc.)

  • Context: A private AI research company co-founded by Ilya Sutskever, a key architect behind major AI breakthroughs like AlexNet and GPT-3.
  • Strategy: SSI is pursuing a "straight shot to superintelligence," focusing purely on foundational research without the immediate pressures of product development and market competition.
  • Core Thesis: The company operates on the belief that the AI industry is moving from an "age of scaling" (where progress came from adding more data and compute) back to an "age of research," where novel ideas are the main bottleneck.
  • Technical Focus: Their primary goal is to solve "reliable generalization," which means creating AI that can learn and adapt as efficiently and robustly as humans—a key weakness in current models.
  • Funding & Resources: SSI has raised $3 billion. Sutskever argues this is sufficient for their research goals, as competitors' larger budgets are often heavily allocated to product support and inference (running models), not just pure research.
  • Timeline: Sutskever gives a rough timeline of 5 to 20 years to develop an AI system that can learn as well as a human.

Takeaways

  • Not Publicly Investable: As a private company, SSI cannot be directly invested in by the public. However, it serves as a crucial barometer for the future direction of the AI industry.
  • A Contrarian Indicator: SSI represents a bet that the current industry approach of simply scaling up existing models will hit a wall. Its progress—or lack thereof—will be a major signal about where the next AI breakthroughs will come from.
  • Potential for Disruption: If SSI's research into generalization is successful, it could leapfrog current market leaders like OpenAI and Google. This would signal a shift in the AI race, where breakthrough ideas become more valuable than just having the largest compute budget. Investors should monitor news from SSI as a gauge of potential disruption.

OpenAI, Google (Gemini), and Anthropic

  • Context: These are described as the "frontier companies" that are currently dominating the "age of scaling."
  • Critique from the Podcast: The discussion suggests their current approach might be "stalling out." While they are positioned to generate "stupendous revenue," they may not achieve the next fundamental breakthrough (true AGI) without a major change in their research paradigm.
  • Problem of Similarity: The models from these different companies are described as being very similar to each other. This is attributed to them all being built on the same pre-training recipe and data, with differentiation only starting to happen at the later fine-tuning stages.
  • Performance vs. Reality Gap: A key criticism is the gap between the models' high performance on benchmarks ("evals") and their often unreliable performance in real-world, practical applications. The podcast suggests this is because the models are over-optimized to pass tests rather than to achieve genuine understanding.
  • Extremely High Costs: These companies operate in a very high-cost environment. It's mentioned that OpenAI is estimated to spend $5-6 billion per year on research experiments alone, highlighting the immense capital required to compete at the top.

Takeaways

  • Incumbent Advantage: These companies have enormous scale, capital, and market penetration. They are set to dominate the current wave of AI applications and generate massive revenues from them.
  • Risk of a Technological Plateau: The primary risk for these giants is technological stagnation. If the scaling-first approach hits a wall of diminishing returns, their growth could slow, making them vulnerable to disruption from companies with more novel, research-driven approaches (like SSI).
  • What to Watch For: Investors should monitor whether these companies can move beyond the current paradigm. True breakthroughs in model capabilities (e.g., reliability, generalization) would be a strong bullish signal. Conversely, if new model releases feel like only incremental improvements, it may validate the "stalling out" thesis.

Meta (META)

  • Context: Meta was mentioned for its attempt to acquire the research lab SSI and for successfully hiring one of SSI's co-founders.
  • Sentiment: This is framed as an aggressive and strategic move, showing Meta's determination to secure top-tier talent and technology in the AI race.

Takeaways

  • Serious AI Ambitions: Meta is clearly positioning itself to be a top-tier AI player. Its willingness to use its vast resources for acquisitions and to poach key talent from leading-edge labs demonstrates a strong commitment to competing at the frontier.
  • Bullish Signal for META: For investors, this is a positive indicator. It shows the company is actively deploying its capital to strengthen its position in what is arguably the most important technological race of our time.

Investment Theme: The "Age of Research"

  • Context: The podcast proposes that the AI industry is shifting from the "age of scaling" (2020-2025), where more compute and data guaranteed better models, to a new "age of research."
  • Core Idea: In this new era, the main barrier to progress is not a lack of computing power, but a lack of fundamental new ideas to solve core problems like generalization and reliability.
  • Historical Precedent: Major breakthroughs like AlexNet (which kicked off the modern AI boom) and the Transformer (the architecture behind GPT) were research ideas that initially required relatively little compute to prove their value.

Takeaways

  • Look Beyond the Giants: This thesis suggests that the next revolutionary AI breakthrough may not come from the biggest, most well-funded companies, but from a smaller, more focused research team with a novel idea.
  • Shift in Competitive Dynamics: If this thesis proves correct, the competitive advantage in AI may shift from being purely capital-based (i.e., who has the most GPUs) to being more talent and idea-based. This could lead to a more dynamic and unpredictable market in the long run, with new winners emerging.

Investment Theme: Compute Demand

  • Context: Even with a potential shift toward research, the demand for massive computing power is not going away. It is described as a critical and ever-growing part of the AI equation.
  • Shifting Workloads: The podcast notes that compute demand is evolving. While pre-training models remains important, an increasing amount of compute is being dedicated to Reinforcement Learning (RL) and inference (the cost of running the models for end-users).
  • Sustained Growth: The discussion implies that overall demand for compute will continue to grow rapidly. The idea of multiple "continent-sized" AI data centers is mentioned as a plausible future scenario.

Takeaways

  • A Strong Secular Trend: The relentless and growing demand for AI compute provides a powerful, long-term tailwind for companies that supply the underlying infrastructure.
  • Implied Beneficiaries: While no specific stocks were named, the clear beneficiaries of this trend are companies in the semiconductor sector that produce GPUs (like Nvidia and AMD), foundries that manufacture advanced chips (like TSMC), and the major cloud providers that rent out this compute at scale (Amazon's AWS, Microsoft's Azure, Google Cloud).
  • Inference is a Key Growth Driver: The specific mention of inference as a major consumer of compute is important. As AI is integrated into more products and services, the demand for chips and cloud resources to run these models will explode, representing a key growth vector for the sector.

Neuralink

  • Context: Mentioned in a highly speculative, long-term discussion about the future of AI alignment.
  • The Idea: The podcast floats the idea that a "Neuralink++" brain-computer interface could allow humans to merge with AI. This is presented as a potential long-term solution to ensure humans remain active participants in a world with superintelligence.
  • Sentiment: This is framed as a far-future, almost sci-fi concept. The speaker notes, "I don't like this solution, but it is a solution," indicating its radical and uncertain nature.

Takeaways

  • High-Risk, Long-Term Venture: Neuralink is a private, high-risk venture operating at the extreme edge of technological possibility. This mention reinforces its potential role in the very distant future of technology and humanity.
  • Not a Near-Term Investment Thesis: For public market investors, this is not an actionable insight for the near future. It serves as a reminder of the transformative and unpredictable nature of the technologies being developed today.
Ask about this postAnswers are grounded in this post's content.
Episode Description
AI models feel smarter than their real-world impact. They ace benchmarks, yet still struggle with reliability, strange bugs, and shallow generalization. Why is there such a gap between what they can do on paper and in practice In this episode from The Dwarkesh Podcast, Dwarkesh talks with Ilya Sutskever, cofounder of SSI and former OpenAI chief scientist, about what is actually blocking progress toward AGI. They explore why RL and pretraining scale so differently, why models outperform on evals but underperform in real use, and why human style generalization remains far ahead. Ilya also discusses value functions, emotions as a built-in reward system, the limits of pretraining, continual learning, superintelligence, and what an AI driven economy could look like.   Resources: Transcript: https://www.dwarkesh.com/p/ilya-sutsk... Apple Podcasts: https://podcasts.apple.com/us/podcast... Spotify: https://open.spotify.com/episode/7naO...   Stay Updated: If you enjoyed this episode, be sure to like, subscribe, and share with your friends! Find a16z on X: https://x.com/a16z Find a16z on LinkedIn: https://www.linkedin.com/company/a16z Listen to the a16z Podcast on Spotify: https://open.spotify.com/show/5bC65RDvs3oxnLyqqvkUYX Listen to the a16z Podcast on Apple Podcasts: https://podcasts.apple.com/us/podcast/a16z-podcast/id842818711 Follow our host: https://x.com/eriktorenberg Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures](http://a16z.com/disclosures.   Stay Updated: Find a16z on X Find a16z on LinkedIn Listen to the a16z Show on Spotify Listen to the a16z Show on Apple Podcasts Follow our host: https://twitter.com/eriktorenberg   Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures. Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.
About a16z Podcast
a16z Podcast

a16z Podcast

By Andreessen Horowitz

The a16z Podcast discusses tech and culture trends, news, and the future – especially as ‘software eats the world’. It features industry experts, business leaders, and other interesting thinkers and voices from around the world. This podcast is produced by Andreessen Horowitz (aka “a16z”), a Silicon Valley-based venture capital firm. Multiple episodes are released every week; visit a16z.com for more details and to sign up for our newsletters and other content as well!