The Fastest Path To Super Intelligence
The Fastest Path To Super Intelligence
Podcast19 min 45 sec
Listen to Episode
Note: AI-generated summary based on third-party content. Not financial advice. Read more.
Quick Insights

Investors should prioritize Alphabet (GOOGL) and Microsoft (MSFT) as foundational "ground layer" plays, as these giants possess the massive capital required to sustain the rapid release cycles of frontier models. Avoid investing in simple AI "wrappers" or startups that rely solely on basic prompt engineering, as they are highly vulnerable to being rendered obsolete by the next model update. Instead, shift focus toward the emerging Agentic AI sector and "Layer 2" companies like Poetic, which build automated reasoning harnesses that sit on top of existing models to boost performance. Monitor the Y Combinator ecosystem for early-stage opportunities in recursive self-improvement technologies that can outperform models like Claude and Gemini at a fraction of the cost. For those with private equity access, watch for Poetic’s move toward a public API or enterprise platform as a high-conviction play in AI efficiency.

Detailed Analysis

Poetic (Private Startup)

Poetic is a specialized AI research company building "recursively self-improving AI reasoning harnesses" for Large Language Models (LLMs). The company aims to provide a layer that sits on top of existing frontier models (like those from OpenAI or Google) to significantly boost their performance and reasoning capabilities.

  • Recursive Self-Improvement: The core technology is a "Metasystem" that uses AI to make the AI smarter. It automatically optimizes prompts, reasoning strategies, and data handling without requiring a human to manually engineer every step.
  • The "Stilts" Concept: Poetic describes its product as "stilts" for AI. It is designed to be model-agnostic; when a new, more powerful model (like a future GPT or Claude) is released, Poetic’s harness can be applied to it immediately to ensure the user stays ahead of the standard "out-of-the-box" performance.
  • Performance Benchmarks:
    • ARC-AGI V2: Poetic reached a 54% score, surpassing Google’s Gemini 3 Deep Think (45%) at roughly half the cost.
    • Humanity’s Last Exam: Achieved 55%, outperforming Anthropic’s Claude 4.6 (53.1%).
  • Cost Efficiency: Unlike traditional fine-tuning, which can cost hundreds of millions of dollars and becomes obsolete when a new model drops, Poetic’s optimization costs are significantly lower (e.g., under $100k for major benchmark runs).

Takeaways

  • Investment Theme: Look for "Layer 2" AI companies. As foundational models become commodities, the value shifts to companies that can extract maximum performance from those models through automated reasoning and "harnesses."
  • Risk Mitigation for Startups: For those building AI products, Poetic offers a way to avoid the "Bitter Lesson"—the risk of spending millions on custom training only to have a foundational model update render your work useless.
  • Actionable Step: While Poetic is currently private and in early access, investors should watch for its potential influence on the Y Combinator ecosystem and its eventual move toward a public API or enterprise platform at poetic.ai.

Foundational Model Providers (OpenAI, Google, Anthropic)

The transcript discusses the major players in the "Frontier Model" space, specifically Google (GOOGL), OpenAI (Microsoft-backed), and Anthropic (Amazon/Google-backed).

  • The "Bitter Lesson": The discussion highlights a major risk for investors in AI startups: foundational model providers frequently "eat the lunch" of smaller companies by incorporating new features or higher intelligence levels in every update.
  • Model Obsolescence: Fine-tuning on older models (like GPT-3.5) is now viewed as a "sunk cost" because newer base models (like GPT-4o or Gemini 1.5) often outperform specialized older models without any extra training.
  • Competitive Landscape: Mention of Claude 4.6 and Gemini 3 Deep Think suggests a rapid release cycle where state-of-the-art (SOTA) leadership changes on a weekly basis.

Takeaways

  • Bullish Sentiment for Infrastructure: The "stilts" analogy confirms that foundational models are the essential "ground layer." Even companies trying to beat them (like Poetic) rely on their existence.
  • Bearish Sentiment for Simple "Wrappers": If a startup's only value is a better prompt or a specific data set, they are highly vulnerable to the next model release from Google or OpenAI.
  • Investment Insight: Focus on foundational companies that have the capital to sustain the "hundreds of millions" required for training runs, as this remains the entry barrier for the base layer of AI.

AI Reasoning & Agentic Systems (Sector)

The podcast identifies a shift from simple "chat" AI to "Agentic Systems"—AI that can use tools, write code, and follow complex reasoning strategies to solve PhD-level problems.

  • Reasoning vs. Knowledge: The discussion emphasizes that "reasoning strategies" (written in code) are more effective than "prompt engineering" (written in natural language). Moving from 5% to 95% performance on certain tasks was achieved through reasoning harnesses, not just better prompts.
  • Automated Engineering: The future of AI development is moving toward "recursive" systems where the human is no longer the bottleneck for optimizing the AI.

Takeaways

  • Sector Trend: The next wave of AI investment is likely in Agentic AI—systems that don't just talk but "do" and "reason."
  • Skill Shift: For the general public/investors, the value is shifting from "knowing how to prompt" to "owning the system that optimizes the prompts."
  • Efficiency Gains: AI is drastically lowering the barrier to software development. The mention of building an iPhone app in a weekend using GPT-5 (likely referring to the latest frontier models) suggests a massive productivity boom for small, lean teams (e.g., Poetic has only 7 employees).

Summary of Key Tickers & Entities

  • Google (GOOGL): Mentioned regarding DeepMind and Gemini models.
  • Anthropic (Private): Mentioned for Claude 4.6 and its performance on "Humanity's Last Exam."
  • OpenAI (Private/MSFT): Mentioned as a primary frontier model competitor.
  • Poetic (Private): The featured startup focusing on recursive self-improvement.
  • Y Combinator (Private): The accelerator hosting the discussion, currently accepting applications for AI-focused startups.
Ask about this postAnswers are grounded in this post's content.
Episode Description
Poetiq is a new startup founded by former DeepMind researchers that recently achieved a major jump on the ARC-AGI and Humanity's Last Exam benchmark by layering a recursive self-improvement system on top of existing models. In this episode of Lightcone, Poetiq's Founder & CEO Ian Fischer joined us to discuss how small teams can build “reasoning harnesses” that outperform base models, what that means for startups and why automating prompt engineering may be one of the most powerful levers in AI today.Chapters:00:00 – Intro00:40 – What Is Poetiq?01:07 – Recursive Self-Improvement Explained02:07 – The Fine-Tuning Trap02:59 – “Stilts” for LLMs03:14 – Recursive Self-Improvement vs. Fine-Tuning05:05 – Taking the Top Spot on ARC-AGI06:37 – Beating Claude on Humanity’s Last Exam08:40 – How the Meta-System Works10:26 – Beyond RL: A New S-Curve11:32 – Automating Prompt Engineering13:37 – From 5% to 95% Performance14:50 – Early Access & Putting Your Agent on Stilts16:17 – From YC Founder to DeepMind Researcher18:29 – Advice for Engineers in the AI EraApply to Y Combinator: https://www.ycombinator.com/applyWork at a startup: https://www.ycombinator.com/jobs
About Y Combinator Startup Podcast
Y Combinator Startup Podcast

Y Combinator Startup Podcast

By Y Combinator

We help founders make something people want.