Terence Tao – Kepler, Newton, and the true nature of mathematical discovery
Terence Tao – Kepler, Newton, and the true nature of mathematical discovery
Podcast1 hr 23 min
Listen to Episode
Note: AI-generated summary based on third-party content. Not financial advice. Read more.
Quick Insights

As AI drives the cost of idea generation toward zero, the highest conviction investment theme is in formal verification and evaluation layers that can validate AI outputs. Focus on companies integrating with Lean or building systems that bridge the gap between natural language AI and formal logic to eliminate "hallucinations" in mission-critical fields. Investors should prioritize firms with proprietary, high-precision "clean" datasets, as these serve as defensive moats against generic models. Look for Quant Hedge Funds and data extraction specialists that employ physicists to find signals in the current "deductive overhang" of unanalyzed big data. Finally, maintain exposure to productivity tools like LabelBox or Wolfram Alpha that automate research drudgery, while remaining cautious of over-optimized companies that have eliminated the "human slack" necessary for breakthrough innovation.

Detailed Analysis

This analysis extracts investment insights from the Dwarkesh Podcast featuring mathematician Terence Tao. The discussion centers on the evolution of scientific discovery, the transition from "eureka moments" to "big data," and the current "plateau" and future potential of AI in mathematics and specialized research.


Artificial Intelligence & Large Language Models (LLMs)

The discussion frames AI as a tool that has fundamentally shifted the economics of scientific discovery by driving the cost of "idea generation" toward zero. However, this has created a new bottleneck: verification and validation.

  • The "Idea Generation" Surplus: AI can generate thousands of hypotheses (e.g., solving 50+ Erdős problems), but it currently lacks the "depth" to verify them or build upon partial progress.
  • The Current Plateau: Tao notes a recent "pause" in AI-driven math breakthroughs. After picking "low-hanging fruit" (problems solvable via one-shot reasoning or obscure literature synthesis), AI is struggling with deeper, multi-step problems that require a "trustworthy co-author" level of interaction.
  • Breadth vs. Depth: AI excels at breadth (applying one technique to a million problems), while humans excel at depth (creating entirely new mathematical frameworks).

Takeaways

  • Investment Theme: Look for companies building verification and evaluation layers for AI. As the cost of generating "slop" or unverified theories drops, the value shifts to systems that can "grade" or formally verify AI outputs (e.g., Lean-based formal verification).
  • Sector Impact: Software engineering and data science are seeing immediate 2x-5x productivity gains in "secondary tasks" (formatting, literature search, boilerplate code), but the "core" creative engine remains human-led for now.
  • Future Catalyst: A major valuation inflection point for AI labs will be the transition from "Artificial Cleverness" (trial and error) to "Artificial Intelligence" (cumulative, adaptive reasoning that learns from partial failures).

Data Science & "Big Data" Paradigms

Tao highlights a reversal in the scientific method: traditionally, one formed a hypothesis and then collected data. Today, we collect "Big Data" first and then use AI/statistics to deduce laws.

  • The "Kepler" Model: Modern data science mirrors Johannes Kepler’s work—using massive, precise datasets (like Tycho Brahe’s) to find empirical regularities that theorists (like Newton) explain much later.
  • Deductive Overhang: There is a massive "overhang" of information in existing datasets that hasn't been extracted because we haven't applied the right "clever tricks" or algorithms yet.

Takeaways

  • Actionable Insight: Investment opportunities lie in "Data Extraction" specialists—companies or funds (like Quant Hedge Funds) that employ astronomers or physicists to find signals in "random bits of data."
  • Asset Class: Companies with proprietary, high-precision "clean" datasets (the modern equivalent of Tycho Brahe’s observatory) hold the most defensive moats against generic AI models.

Specialized Tools & Infrastructure

The transcript mentions specific technical tools and companies that are becoming essential to the "frontier" of research and development.

  • Lean (Formal Proof Language): A critical infrastructure for the future of math. It allows AI-generated proofs to be "atomically" verified, removing the risk of "hallucinations" in mission-critical calculations.
  • Wolfram Alpha / Mathematica: Cited as the precursors to AI automation in math, having already turned 19th-century "frontier math" into a commodity.
  • LabelBox: Mentioned in the context of training models to "think" via rubrics rather than just providing correct answers.
  • Mercury: Mentioned as a fintech platform providing "Insights" for business cash flow management.

Takeaways

  • Infrastructure Play: As AI moves into "hard sciences," the demand for Formal Verification (like the Lean community) will grow. Any platform that can bridge the gap between "natural language AI" and "formal logic" is a high-value target.
  • Productivity Tools: Tools that automate "drudgery" (reformatting, literature search, code maintenance) are providing the most immediate ROI for high-value human capital (PhDs, researchers).

Risks & "The Human Element"

Tao warns that total optimization and the removal of "inefficiency" may carry hidden long-term risks for innovation.

  • Loss of Serendipity: Moving to purely "scheduled" or "AI-searched" environments removes the "accidental" discoveries found in physical libraries or hallway conversations.
  • The "Incomprehensible Proof" Risk: There is a risk that AI solves major problems (like the Riemann Hypothesis) with "assembly code gobbledygook" that provides the answer but no human-usable "insight" or "intuition."
  • Inertia of Systems: Even superior systems (like different computer architectures or base-12 math) often fail to gain traction due to the "inertia" of existing standards (Base-10, QWERTY, etc.).

Takeaways

  • Risk Factor: Be cautious of "over-optimized" companies that have eliminated all "slack" or "randomness." Tao suggests that a certain level of "distraction" and "high-temperature" randomness is essential for breakthrough innovation.
  • Long-term Value: The "Human Side of Science"—persuasion, narrative building, and communication—remains the most "AI-proof" skill set. Investments in "Science Communicators" or platforms that facilitate human collaboration may prove more resilient than pure automation plays.
Ask about this postAnswers are grounded in this post's content.
Episode Description
We begin the episode with the absolutely ingenious and surprising way in which Kepler discovered the laws of planetary motion. People sometimes say that AI will make especially fast progress at scientific discovery because of tight verification loops. But the story of how we discovered the shape of our solar system shows how the verification loop for correct ideas can be decades (or even millennia) long. During this time, what we know today as the better theory can actually make worse predictions. And the reasons it survives this epistemic hell is some mixture of judgment and heuristics that we don’t even understand well enough to actually articulate, much less codify into an RL loop. Hope you enjoy! Watch on YouTube; read the transcript. Sponsors - Jane Street loves challenging my audience with different creative puzzles. One of my listeners, Shawn, solved Jane Street’s ResNet challenge and posted a great walk-through on X. If you want to try one of these puzzles yourself, there’s one live now at janestreet.com/dwarkesh. - Labelbox can get you rubric-based evals, no matter your domain. These rubrics allow you to give your model feedback on all the dimensions you care about, so you can train how it thinks, not just what it thinks. Whatever you’re focused on—math, physics, finance, psychology or something else—Labelbox can help. Learn more at labelbox.com/dwarkesh. - Mercury just released a new feature called Insights. Insights summarizes your money in and out, showing you your biggest transactions and calling out anything worth paying attention to. It’s a super low-friction way to stay on top of your business. Learn more at mercury.com/insights. Timestamps (00:00:00) – Kepler was a high temperature LLM (00:11:44) – How would we know if there’s a new unifying concept within heaps of AI slop? (00:26:10) – The deductive overhang (00:30:31) – Selection bias in reported AI discoveries (00:46:43) – AI makes papers richer and broader, but not deeper (00:53:00) – If AI solves a problem, can humans get understanding out of it? (00:59:20) – We need a semi-formal language for the way that scientists actually talk to each other (01:09:48) – How Terry uses his time (01:17:05) – Human-AI hybrids will dominate math for a lot longer Get full access to Dwarkesh Podcast at www.dwarkesh.com/subscribe
About Dwarkesh Podcast
Dwarkesh Podcast

Dwarkesh Podcast

By Dwarkesh Patel

Deeply researched interviews <br/><br/><a href="https://www.dwarkesh.com?utm_medium=podcast">www.dwarkesh.com</a>