Libido Sciendi
Libido Sciendi
The desire to understand the world deeply. Investigating AI & complex adaptive systems.
Home Journal Deep Dives Library About
Contents
  1. 1. Intelligence as a Commodity. When Thinking Gets Cheap
  2. 2. Verifier’s Law: The Measure Becomes the Limit
  3. 3. The Jagged Edge of Intelligence: Why Progress Looks Uneven
  4. Synthesis: The Three Gradients
November 4, 2025 · 9 min ESSAY

Intelligence Isn't the Moat Anymore: The New Frontier Is Verification

Intelligence Isn’t the Moat Anymore: The New Frontier Is Verification


TL;DR: AI isn’t progressing evenly. Jason Wei argues the real question isn’t when it changes everything, but which kinds of tasks it conquers first, and why verification, not intelligence, defines the next frontier.


A quant trader says ChatGPT can’t do his job. An AI researcher thinks she has 2-3 years before being replaced.

Who’s right?

Jason Wei (co-creator of o1 and Deep Research, 90k+ citations) thinks both are asking the wrong question.

The point isn’t when AI changes everything, but what, and in what order.

In a recent talk at Stanford, Wei presents three foundational principles for understanding AI’s trajectory. These principles move beyond the polarised debate between skeptics and believers to offer a predictive framework based on task properties we can identify today.

  1. Intelligence as a Commodity: Once AI achieves a capability, its cost drops exponentially toward zero through adaptive compute and model efficiency improvements.

  2. Verifier’s Law: AI’s ability to solve a task is directly proportional to how easily that task can be verified. Tasks with objective, fast, scalable, low-noise, and continuous verification will be conquered first.

  3. The Jagged Edge of Intelligence: AI capabilities and improvement rates vary dramatically across tasks based on three heuristics: digital vs. physical, human difficulty level, and data abundance.

Key Implication: We won’t see a “fast takeoff” to superintelligence. Instead, expect highly uneven progress where some domains (software, research) accelerate dramatically while others (physical services, rare skills) remain largely untouched.

1. Intelligence as a Commodity. When Thinking Gets Cheap

Wei’s first principle tackles the price of thinking itself. Cognitive work (reasoning, retrieval, and problem solving) is becoming a commodity. Once an ability is unlocked by AI, its cost rapidly approaches zero.

Stage 1: Pushing the Frontier AI gradually unlocks new capabilities. Performance on benchmarks like MMLU improves incrementally as models scale and techniques advance.

Stage 2: Commoditisation Once a capability is achieved, the cost to access that level of intelligence decreases exponentially. Wei demonstrates this with MMLU benchmark data showing year-over-year cost reductions for equivalent performance. Each year, the cost of getting the same performance drops by orders of magnitude.

1.1. Information retrieval time collapse

Wei asks a simple question to make the point: “How many couples got married in Busan, South Korea in 1983?”.

Even more complex queries like “Of the 30 most populated Asian cities in 1983, sort them by number of marriages that year” previously took weeks. Now: minutes.

The shift isn’t about smarter models, it’s about how efficiently they spend their compute.

1.2. Technical inflection: Adaptive Compute

Adaptive compute is the game changer explaining why commoditisation will accelerate indefinitely.

Before, every problem used the same compute. “What’s the capital of California?” consumed the same resources as solving a competition math problem. This was wasteful and expensive.

Now, we can allocate 1 token to simple queries, 100,000 tokens to hard problems. OpenAI o1 demonstrated that increasing test-time compute on difficult problems improves performance proportionally.

Consequence: The cost of easy reasoning collapses without limit. We don’t need to scale model sizes indefinitely. Small, efficient models handle routine tasks while compute scales only where needed.

This is the first time in deep learning history that adaptive compute actually works at scale. It fundamentally changes the economics of intelligence.

1.3. Real-World Validation: BrowseComp Benchmark

OpenAI’s browsing competition benchmark tests questions that are easy to check but take humans hours to solve.

Eg: finding marriage statistics in Korean government database (KOSIS) requires navigating complex interfaces. OpenAI O3 couldn’t do this; OpenAI Operator can through autonomous browsing.

Deep Research now cracks about 50% of these problems. Many tasks that took most humans at least two hours were solved in minutes through autonomous agents.

1.4. Strategic Implications

  1. Violent Democratisation of Knowledge-Gated Fields A talented high schooler with Cursor can now build what a 5-engineer team did in 2020. A biology undergrad with ChatGPT can access medical knowledge equivalent to a competent physician for biohacking experiments. Gatekeeping through knowledge is dead. What matters: taste, distribution, relationships, private data. Fields being democratised: Coding (Cursor, Windsurf, Replit), Research (literature review, synthesis), Personal health (biohacking, self-optimisation), Design (Figma + AI, v0), Writing and content creation.

  2. Increased Relative Value of Private Information

    When public information costs approach zero, private/insider information becomes disproportionately valuable, like:

    • Off-market real estate deals
    • Proprietary customer data
    • Internal company metrics
    • Network effects and relationships
    • Exclusive access to people or places

    The moat shifts from “what you know” to “what only you can know.”

  3. Personalised Internet Experience

    The shift from universal search results to AI-generated, personalised information interfaces. Instead of browsing a public internet, users will interact with dynamically generated, context-aware information environments tailored to their specific query and context.

2. Verifier’s Law: The Measure Becomes the Limit

If the first principle explains why reasoning gets cheap, this one explains why some reasoning improves faster than others.

The Verifier’s Law is a cousin of Goodhart’s Law: when a measure becomes a target, it reshapes the system itself (spoiler alert: it often ceases to be a good measure).

The Verifier’s law states that trainability of AI on a given task is proportional to how easy it is to verify success.

By corollary, any well-defined solvable, verifiable task, with unambiguous success criteria, will eventually be conquered by AI.

2.1. Understanding Asymmetry of Verification

Some problems are much easier to verify than to generate, and vice versa. This asymmetry is the key to predicting AI progress.

Verification asymmetry diagram Tasks can be plotted by ease of generation (x-axis) vs. ease of verification (y-axis). Adding privileged information (e.g. test cases, answer keys) can shift a task upward, making it easier to verify and therefore easier for AI to learn.

You can shift tasks upward by adding privileged information. Providing answer keys for math problems or test cases for code (as in SWE-bench) transforms moderately verifiable tasks into highly verifiable ones.

2.2. The Five Dimensions of Verifiability

Tasks are verifiable to the extent they possess these properties:

  1. Objective Truth: Clear distinction between good and bad responses

  2. Speed: Fast verification time

  3. Scalability: Can verify millions of proposed solutions in parallel

  4. Low Noise: Consistent results across verification attempts

  5. Continuous Reward: Granular feedback beyond binary pass/fail

The more dimensions a task satisfies, the faster AI will master it.

Strategic implications

  • Automation Frontier is Verifiability Frontier: Tasks with trivial or automated verification are first to be fully solved.

  • Measurement as leverage: Designing measurable tasks and metrics becomes a central driver of AI progress and business defensibility. The company that can make an unverifiable task verifiable unlocks AI’s ability to solve it.

  • Recognising Unverifiable Domains: Some tasks will resist AI automation indefinitely because verification is fundamentally hard (long term outcome, high noise, subjective results, etc.)

3. The Jagged Edge of Intelligence: Why Progress Looks Uneven

If the first principle prices intelligence and the second measures it, the third describes its pattern of diffusion.

Technologies don’t spread evenly either; diffusion follows the same jagged logic: fast through digital networks, slow through physical supply chains.

AI progress is uneven and domain-specific, not a smooth march toward general superintelligence.

3.1. Rejecting Fast Takeoff Scenarios

Nick Bostrom in Superintelligence stated (and it’s now a popular belief) that once AI crosses a capability threshold, recursive self-improvement (RSI) will lead to rapid superintelligence.

Year 0: can’t train GPT n+1. Year 1: can train GPT n+1. Year 2: superintelligence.

Wei argues that progress will be gradual and task-specific, not binary. A more realistic model would look like:

Year 0: Can’t even load the codebase autonomously Year 1: Can train something, results are mediocre compared to humans Year 2: Trains autonomously, but not as good as top 10 humans. Year 3 Mostly autonomous, occasional human intervention needed Year 4: Fully autonomous on some tasks, still needs humans on others etc.

And self-improvement happens at different rates: some tasks show rapid gains (math, code, search), others stagnate (rare languages, embodied action, social nuance). Each has a different “rate of improvement curve”.

3.2. Task-based improvement rates

Different domains evolve at different speeds due to: (i) verification ease, (ii), Data availability, and (iii) Digital / physical nature.

Each task has its own trajectory: peaks (fast progress) and valleys (stagnation).

Heuristics for predicting improvement speed

  1. Digital vs. physical: Digital tasks evolve faster due to higher iteration speed (rapid cycles, parallel testing), compute scalability (and safety concerns for physical tasks).

  2. Human difficulty correlation: Tasks easy for humans are often easy for AI (important exceptions: tasks requiring scale or capabilities beyond human biology (analysing 10m images, maintaining consistency across millions of operations (and without fatigue or attention limits)

  3. Data abundance: Performance correlates with available data volume (e.g. math in high-resource languages > low-resource).

  4. Reward clarity: Tasks with explicit, measurable success signals accelerate via reinforcement learning.

Macro implications:

  • Irregular capability curve: Some domains leap ahead, others remain static for decades.

  • Sectoral impact: Software development, data analysis, and content generation see rapid transformation; embodied and social work lags.

  • Self-improvement gradient: Each new capability expands locally before spreading across tasks. A model that gets better at Python doesn’t automatically get better at hairdressing. Progress is task-clustered, not universal.

Synthesis: The Three Gradients

Wei’s framework turns epistemology into strategy. The act of knowing isn’t scarce anymore; the act of confirming is. In that shift, verification becomes a new form of power.

The future won’t be uniformly transformative or uniformly disappointing. It will be specifically transformative based on task properties we can identify and predict.

Each gradient compounds over a different horizon, the timelines below show where they begin to bite.

Timelines by Gradient Position

Wei’s framework predict three gradients. Each company, field, or individual sits somewhere along them, and knowing where you stand might be the only durable edge left.

  • Short-term (1-3 years):

    • Commoditisation of current capabilities (coding, analysis, writing).
    • Emergence of AI-native companies in verifiable domains.
    • Major productivity gains in digital knowledge work.
    • Winner-take-most dynamics in newly automated fields.
  • Medium-term (3-7 years):

    • Jagged progress creates massive arbitrage opportunities between peaks and valleys.
    • Value accrues to measurement infrastructure and private data.
    • Physical-world AI remains bottlenecked despite digital advances.
    • Verification innovation unlocks new domains sequentially.
  • Long-term (7+ years):

    • Intelligence costs → 0 for routine tasks.
    • Human value shifts to unverifiable, relationship-based, or physical domains.
    • Data moats and verification systems become core infrastructure.
    • Jagged edge persists but shifts to different boundaries.

Wei’s framework doesn’t predict a single future but a moving frontier that depends on where measurement can reach.

The Critical Question

The question isn’t ‘will AI change the world’ but ‘where precisely are you positioned on these three gradients?’

If your business depends on (i) commoditisable intelligence (routine reasoning, knowledge retrieval), (ii) easily verifiable tasks (clear success metrics, automated testing), (iii) digital, data-rich domains (software, analysis, content), then you have probably 18-36 months to find a new moat, or you’ll disappear. The middle market in these domains is vanishing rapidly.

If you’re in the valleys (embodied physical work, rare data domains, unverifiable quality metrics, relationship-dependent services), you probably have several years if not decades before significant AI impact.

Your strategic advantage depends on accurately assessing where you sit on these three gradients and positioning accordingly. In the next decade, the bottleneck will shift from intelligence to instrumentation: whoever builds the better feedback loop wins.

Those who understand the jagged edge will capture asymmetric returns. Those who assume uniform progress will be caught off guard.

Intelligence won’t be the scarce resource of the next decade. Verification will.

“Intelligence and knowledge will be fast and cheap. Measurement will drive AI progress. The edge of intelligence is jagged.” Jason, Meta Superintelligence Labs (2025)

Article originally posted on WeLoveSota.com

Willy Braun signature
Libido Sciendi Digest
Essays, journal entries, book notes and curated readings from Libido Sciendi, delivered to your inbox weekly.
Subscribe to Libido Sciendi Digest →
#AI #strategy
← All journal entries
Substack LinkedIn X Instagram
© 2026
Stay curious
Essays, journal entries, book notes and curated readings — delivered to your inbox.