Jason Wei — 3 Key Ideas in AI in 2025

Original source: Stanford AI Club: Jason Wei on 3 Key Ideas in AI in 2025

TL;DR

  • Jason Wei outlines three mental models for 2025: (1) intelligence becomes a commodity as test-time compute and agents drive the cost of “getting the right answer” toward zero; (2) Verifier’s Law: AI learns fastest on tasks that are easy to verify; (3) the edge of intelligence is jagged: capabilities and improvement rates vary wildly by task—no single “fast takeoff.”

Core ideas

1) Intelligence becomes a commodity

  • Once a capability is unlocked, its cost drops rapidly; same target performance gets cheaper each year.
  • Key driver: adaptive/test-time compute (e.g., o1) — spend more reasoning on hard problems, less on easy ones without growing model size.
  • Agents compress time-to-knowledge: public facts go from hours → minutes → seconds as browsing + tool use mature (e.g., finding niche stats via national databases).
  • Implications: democratization of knowledge-gated fields (coding, personal health guidance), rising relative value of private/insider data, and eventual personalized, frictionless information access.

2) Verifier’s Law (asymmetry of verification)

  • Principle: The easier a task is to verify, the easier it is to train AI to do it; easily verifiable tasks will be conquered first.
  • Verification ease depends on: objective truth, speed of checking, scalability (batchable checks), low noise, and continuous reward (not just pass/fail).
  • Examples along the spectrum:
    • Easy to verify: Sudoku, rendering/click-testing an app, code with unit tests, geometry packing tasks.
    • Hard to verify: factual essays, “best diet” claims (slow, noisy, costly to validate).
  • Tactics: supply answer keys/tests to move tasks into the easy-to-verify quadrant (benchmarks, unit tests, graded objectives).
  • Case study: AlphaEvolve-style loops—sample many candidates, grade automatically, feed the best back to the model—show rapid gains when a crisp metric exists.

3) The jagged edge of intelligence

  • No singular step-function “superintelligence takeoff”; instead, capability improves at different rates per task.
  • Heuristics for where progress is fastest:
    • Digital > physical (iteration speed, scale).
    • Easier for humans → easier for AI (in general).
    • Data-abundant tasks improve faster; exception: if a single clear metric exists, RL/self-play can synthesize data (AlphaZero/AlphaEvolve pattern).
  • Rough landscape: already strong on competition math/coding and translation for high-resource languages; slower on low-resource languages, robotics, hands-on trades (plumbing, hairdressing), artisanal crafts, and socially complex goals.

What this means

  • Short term: automation/acceleration hits trivially verifiable, digital, data-rich workflows first (software dev, debugging with tests, structured research + fact retrieval).
  • Opportunity surface: invent measurements and create verification harnesses—whoever defines the metric unlocks progress.
  • Strategy: expect uneven gains; plan around task-level variance, and value private data/connectors as public knowledge commoditizes.

About the speaker

  • Jason Wei is an AI researcher at Meta Superintelligence Labs.
  • Previously at OpenAI, co-created o1 and Deep Research.
  • Earlier at Google Brain, helped introduce Chain-of-Thought prompting and documented emergent phenomena.