Beyond Bigger Models: Recursion As The Next Scaling Law In AI
Beyond Bigger Models: Recursion As The Next Scaling Law In AI
Podcast37 min 52 sec
Listen to Episode
Note: AI-generated summary based on third-party content. Not financial advice. Read more.
Quick Insights

Investors should shift focus from massive, parameter-heavy models toward companies specializing in Recursive AI Architectures and Inference-Time Compute, as smaller models like TRM are now outperforming giants in logic-heavy tasks. Prioritize startups that benchmark their technology against the ARC Prize (Abstraction and Reasoning Corpus) rather than standard language tests, as this is the new gold standard for measuring true artificial general intelligence. Look for "alpha" in Small Language Models (SLMs) that utilize Latent Space Reasoning, which allows AI to solve complex problems internally without the high cost and speed bottlenecks of "thinking out loud" via text. This shift toward Recursive Models is particularly actionable for the Biotech, Engineering, and Cryptography sectors, where AI must invent new logic rather than just parrot human data. Monitor the 2025 rollout of Hierarchical Reasoning Models (HRM) as a signal to pivot away from "one-shot" feed-forward architectures toward more efficient, loop-based reasoning systems.

Detailed Analysis

This analysis explores the shift from simply increasing model size (scaling laws) to using Recursion as a method for improving AI reasoning, based on the Y Combinator discussion regarding two landmark 2025 papers: Hierarchical Reasoning Models (HRM) and Tiny Recursive Models (TRM).


Recursive AI Architectures (HRM & TRM)

The discussion highlights a move away from "one-shot" feed-forward models (like standard LLMs) toward models that reuse the same weights repeatedly to "think" through a problem.

  • The Limitation of LLMs: Standard models like GPT-4 or Claude are limited by their architecture; they have a fixed number of layers. If a problem (like a complex Sudoku or a massive sorting task) requires more steps than the model has layers, the model physically cannot solve it in one pass.
  • The "Recursion" Solution: Instead of making a model with 1,000 layers, researchers are using smaller models (e.g., 7M to 27M parameters) that loop back on themselves.
    • HRM (Hierarchical Reasoning Models): Uses different "frequencies" of reasoning (low-level and high-level) to solve puzzles.
    • TRM (Tiny Recursive Models): A simplified version that uses a single transformer layer looped multiple times, outperforming models 1,000x its size on specific reasoning tasks.

Takeaways

  • Efficiency over Size: Investors should look for AI companies focusing on Inference-Time Compute and Recursive Architectures. The era of "just add more GPUs and parameters" may be hitting diminishing returns for logic-heavy tasks.
  • Niche Performance: These models are currently task-specific (e.g., solving the ARC Prize or Sudoku). The next investment frontier is the "General Purpose Recursive Model" that combines LLM knowledge with TRM reasoning.

The ARC Prize & Reasoning Benchmarks

The podcast emphasizes the ARC Prize (Abstraction and Reasoning Corpus) as the gold standard for measuring true AI intelligence versus mere pattern matching.

  • Performance Breakthrough: While massive models like O3 (referenced as a benchmark) struggled with certain logic puzzles, the HRM (only 27M parameters) achieved 70% on ARC Prize 1.
  • TRM Superiority: The TRM model, even smaller at 7M parameters, pushed performance to 87%.
  • Incompressibility: Problems like Sudoku, mazes, and sorting are "incompressible," meaning you can't skip steps. Recursive models handle these by using a "latent memory tape" (hidden states) to store intermediate thoughts.

Takeaways

  • Benchmark Shift: When evaluating AI startups, look for those testing against the ARC Prize rather than just standard language benchmarks (like MMLU). Success on ARC is a stronger indicator of AGI (Artificial General Intelligence) potential.
  • Small Model Value: There is significant "alpha" in companies developing high-performing Small Language Models (SLMs). They are cheaper to run, faster to deploy, and, as shown here, can be more "intelligent" in logic than trillion-parameter giants.

Investment Theme: Latent Space Reasoning

A technical but critical distinction made is the difference between Chain of Thought (CoT) and Latent Reasoning.

  • Chain of Thought (Current Tech): The model "thinks out loud" by printing text. This is slow, expensive, and limited to human language.
  • Latent Reasoning (The Future): Models like TRM reason in "latent space" (mathematical vectors) without printing text. This is faster and allows for more complex "internal" calculations that don't have to be translated into words.
  • The "Einstein Test": The goal is for AI to discover new physics or math from first principles, which requires internal recursion rather than just repeating human-labeled data.

Takeaways

  • Compute Efficiency: Companies that master Latent Space Reasoning will have a massive cost advantage. They require fewer "tokens" to reach a correct answer, reducing the "wall clock time" and energy costs of AI.
  • Sector Impact: This is particularly relevant for Biotech, Cryptography, and Engineering, where the "trace" (the step-by-step solution) isn't always known by humans, and the AI must "invent" the logic.

Risk Factors

  • Training Difficulty: Recursive models suffer from "Backprop Through Time" (BPTT) issues, where errors accumulate as the model loops. While HRM/TRM use tricks (like Fixed Point Iteration) to bypass this, it remains a complex engineering hurdle.
  • Hardware Alignment: Current GPUs are optimized for massive parallel "one-shot" passes. Highly recursive models may require different hardware optimizations or "trading memory for compute time."
  • Generalization Gap: Currently, these recursive breakthroughs are mostly seen in specialized reasoning tasks. It is not yet proven that this architecture can maintain the "creative" and "conversational" fluidity of standard LLMs.
Ask about this postAnswers are grounded in this post's content.
Episode Description
A 7-million parameter model outperforming models a thousand times its size on tasks like ARC Prize. That's what recursive reasoning unlocks.In this episode of Decoded, YC's Ankit Gupta and Francois Chaubard break down two recent papers on recursive AI models, HRMs and TRMs, that are achieving state-of-the-art results with a fraction of the parameters of today's largest models.They explain why standard LLMs hit a fundamental ceiling on certain reasoning tasks, how recursion at inference time gives small models the compute depth to break through it, and what happens when you combine these ideas with the power of large-scale foundation models.
About Y Combinator Startup Podcast
Y Combinator Startup Podcast

Y Combinator Startup Podcast

By Y Combinator

We help founders make something people want.