Understanding the Most Viral Chart in Artificial Intelligence
Understanding the Most Viral Chart in Artificial Intelligence
14 days agoOdd LotsBloomberg
Podcast56 min 54 sec
Listen to Episode
Note: AI-generated summary based on third-party content. Not financial advice. Read more.
Quick Insights

The AI sector is currently experiencing an exponential acceleration, with capabilities doubling every four months, making Data Center Infrastructure and hardware providers essential plays as massive capital expenditures for 2025–2027 are already locked in. Anthropic is emerging as a leader in "agentic" AI, with its latest models doubling their autonomous task capacity to 12 hours, signaling a major shift toward Software Engineering Automation. Investors should prioritize US-based frontier labs like OpenAI, Anthropic, and Google, as they maintain a decisive 9-to-12-month lead over Chinese competitors like DeepSeek and Qwen. While revenue growth in the sector is vertical, focus on companies transitioning from simple chatbots to autonomous agents capable of managing complex, multi-hour projects without human intervention. Exercise caution regarding mission-critical automation, as even top-tier models currently maintain only a 50% success rate on long-form complex tasks.

Detailed Analysis

METER (Nonprofit Research Org)

METER is a research nonprofit focused on measuring AI capabilities, specifically regarding autonomy and the potential for catastrophic risks. • They are the creators of the "Time Horizon" chart, which has become an industry-standard benchmark for tracking AI progress. • The organization focuses on agency: the ability of an AI to complete long, complex tasks without human intervention.

Takeaways

Benchmark Influence: Investors and developers are increasingly using METER’s data to make funding and development decisions, as it provides a more "grounded" metric than traditional percentage-based benchmarks. • Talent Gap: Despite the high stakes, METER (and similar nonprofits) struggles to compete with the massive equity packages offered by major labs, creating a bottleneck in independent AI safety research.


Artificial Intelligence Sector (General)

• The "Most Viral Chart" in AI shows an exponential increase in the difficulty of tasks AI can complete. • Doubling Rate: AI capabilities (measured by time horizons) are currently doubling approximately every 4 months. This is an acceleration from previous estimates of 6–7 months. • Compute Correlation: R&D spending on compute (hardware/data centers) is rising exponentially at the same rate as AI capability progress.

Takeaways

Near-Term Momentum: Because massive investments in data centers for 2025–2027 are already "baked in," a slowdown in AI progress is unlikely in the immediate future. • Investment Sentiment: There is a notable "Baptist and Bootlegger" dynamic where lab CEOs warn of "doomsday" risks while simultaneously seeking billions in investment to build the very technology they warn against. • Productivity vs. Capability: While AI "Time Horizons" are expanding, real-world productivity may lag due to "messy" human factors, such as the need for humans to verify AI work and the lack of AI self-awareness in complex project management.


Anthropic (CLAUDE)

Claude Opus 4.6 (projected/current models) has shown a massive leap in performance. • The chart highlights a jump from a ~6-hour task capacity to a 12-hour task capacity (at a 50% success rate). • Anthropic’s revenue chart was described as a "straight line up," indicating hyper-growth.

Takeaways

Market Leadership: Anthropic’s latest models are currently setting the pace for "agentic" AI—AI that can act as an autonomous engineer rather than just a chatbot. • Reliability Gap: While the 50% success rate for 12-hour tasks is impressive, the success rate for 80% reliability is much lower, suggesting businesses should be cautious about replacing humans for mission-critical long-form tasks just yet.


OpenAI (GPT)

• Mentioned in the context of the "Manhattan Project" mentality—building powerful technology while attempting to figure out how to "not build the bomb" (safety). • GPT-5.3 Codex was previously a high-water mark on the time horizon charts before being surpassed by newer models.

Takeaways

Competitive Pressure: The intense rivalry between OpenAI, Anthropic, and Google creates a "race to the finish" that may prioritize speed over safety, despite public statements to the contrary.


Chinese AI Models (e.g., Qwen, DeepSeek)

• Chinese models like Qwen and DeepSeek are generally observed to be 9 to 12 months behind US frontier models in terms of raw capability. • There is a suggestion that some international models may be "gaming" standard benchmarks, performing better on tests than they do on unique, "held-out" real-world problems.

Takeaways

US Dominance: For investors looking for the absolute "frontier" of AI autonomy and engineering capability, US-based labs (OpenAI, Anthropic, Google) currently maintain a significant lead over Chinese competitors.


Investment Themes & Sectors

Software Engineering Automation: This is the primary "task distribution" where AI is seeing the fastest growth. AI is moving from "writing a poem" to "managing a software repository." • The "Singularity" Risk: The point where AI can automate its own R&D (AI building better AI) is the key milestone METER is watching. If this is reached, progress could move beyond human control. • Data Center Infrastructure: A massive amount of capital is being tied up in debt to build the physical infrastructure required to sustain the current exponential growth curve.

Takeaways

Risk Factor: A major risk for the sector is "financial obligation." Companies building massive data centers with debt may be forced to keep "the pedal to the metal" even if safety risks become apparent, simply to service their financial commitments. • Future Jobs: A niche but growing field is "Human Baselining"—talented humans being paid to compete against AI to set the benchmarks for how long tasks should take.

Ask about this postAnswers are grounded in this post's content.
Episode Description
We live in an era of charts that are going up and to the right. This image obviously describes the stock market, particularly any company whose business is adjacent to artificial intelligence. But beyond stocks, another sort of chart we keep seeing is of AI capabilities also going up and to the right. The most famous and viral of these comes from an organization called METR, which stands for Model Evaluation and Threat Research. The organization is focused on understanding the degree to which AI models can engage in autonomous, complex tasks. METR see this is as a particularly important benchmark, given the risk that AI could one day be engaged in recursive self improvement, taking humans out of the loop. But how do you really gauge a model's ability to do complex problems. And what is being measured for exactly? On this episode, we speak with METR's President Chris Painter as well as Joel Becker, a member of the technical staff who works on evaluation methods for the organization. We discuss both the mechanics and the philosophy of METR's work, and what it means when we see a a chart showing that Clause Opus 4.6 can do a task that would take a human nearly 12 hours. Read more: DeepSeek Unveils Flagship AI Model a Year After Breakthrough Meta Inks Deal to Use Amazon’s Graviton Processors for AI Only http://Bloomberg.com subscribers can get the Odd Lots newsletter in their inbox each week, plus unlimited access to the site and app. Subscribe at  bloomberg.com/subscriptions/oddlots Subscribe to the Odd Lots Newsletter Join the conversation: discord.gg/oddlots See omnystudio.com/listener for privacy information.
About Odd Lots
Odd Lots

Odd Lots

By Bloomberg

<p>Bloomberg's Joe Weisenthal and Tracy Alloway explore the most interesting topics in finance, markets and economics. Join the conversation every Monday and Thursday.</p>