WORKING PAPER · 06.03 / AGIFuture & Speculation · v.12026.05
Volume 6 · paper 03 of 10

Artificial General
Intelligence: a working paper.

A synthesis of the technical, governance, and philosophical literature as of 2026 — written so that you can argue with someone who has read Bostrom and someone who has read Bender, in the same dinner.

Abstract. AGI denotes systems that match or exceed human performance across a broad range of cognitive tasks, generalizing without task-specific retraining. Whether the current frontier models are on a smooth path to AGI, or are an impressive plateau, is contested. We summarize scaling-law evidence, candidate routes, alignment difficulty, governance regimes, and four scenario branches.

§1 · Definitions

What we mean by AGI

The term "AGI" was popularized by Shane Legg, Ben Goertzel, and Marcus Hutter circa 2002. Earlier "strong AI" (Searle) and "HLMI" (high-level machine intelligence) capture overlapping ideas. Operational definitions cluster into three families:

(i) Task-coverage — passes a wide battery of human exams (SAT, MCAT, LSAT, codeforces, ARC), generalizes zero-shot. The OpenAI / Anthropic / DeepMind house style.

(ii) Economic — automates a large fraction (≥50%) of remote-work tasks at expert human reliability. The Metaculus / forecaster favourite.

(iii) Cognitive completeness — possesses long-horizon planning, online learning, agency, theory of mind, embodiment-readiness. Gary Marcus's preferred definition; harder to claim.

The disagreement is consequential: under (i), GPT-4-class models are arguably AGI now. Under (iii), nothing built since 1956 qualifies.

§2 · Figure 1 · Scaling laws

The scaling-law observation

Kaplan et al. (2020) and Hoffmann et al. (Chinchilla, 2022) showed test loss decreases as a power law in parameters, data, and compute. The frontier has spent ~10²⁵ FLOPs as of 2024 (GPT-4-era), with Llama-3.1, Claude 3.5, Gemini Ultra, and Grok-3 all in similar bands.

loss low 10¹⁸ 10²⁰ 10²² 10²⁴ 10²⁶ FLOP GPT-2 (2019) GPT-3 (2020) GPT-4 / Chinchilla-tier (2022–23) 2024 frontier · ~10²⁵ FLOP Chinchilla optimal over-trained & aligned FIG 1 · loss vs training compute · log-log
After Kaplan 2020, Hoffmann 2022, Epoch AI 2024 trends. Schematic.

The scaling hypothesis says: keep going, capabilities continue to emerge. Critics (Bender, Marcus, Mitchell) argue that benchmarks measure pattern-completion not understanding, and that emergent capabilities may be artifacts of evaluation thresholds (Schaeffer et al., 2023).

§3 · Routes

Four candidate paths to AGI

Route A · Pure scaling
More compute, more data
Frontier-lab orthodoxy. Leopold Aschenbrenner's 2024 essay; Sutton's "bitter lesson".
Route B · Scaffolded agents
Tool-use, memory, MCTS
AlphaCode, AutoGPT, Claude with tools, Devin. Foundation model + planning loop.
Route C · Neurosymbolic
Hybrid systems
Marcus & Davis; LeCun's JEPA; world-model + verifier architectures.
Route D · Whole-brain emulation
Reverse engineering
Sandberg-Bostrom 2008; FlyEM connectomes; mouse cortex by ~2030?
Route E · Embodied
Robotics + curriculum
Tesla Optimus, 1X, DeepMind RT-X. Sensorimotor grounding hypothesis.
Route F · Surprise
Novel architecture
Mamba, RWKV, energy-based models, diffusion-LMs. Transformer is not the end.
§4 · Forecasts

What people who do this for a living say

SourceEstimate (median)NotesYear
Metaculus (community)~2032 for "weak AGI", ~2040 for fullPulled in dramatically since 20222024
AI Impacts survey of researchers2047 for HLMI (50%)2,778 respondents; high variance2023
Daniel Kokotajlo · Situational Awareness2027–28 plausibleAggressive scaling extrapolation2024
Yann LeCunDecades; current LLMs are "an off-ramp"Argues for JEPA-style world models2024
Demis Hassabis"5–10 years" for AGIRepeated ~2024–252024
Dario Amodei2026–27 plausible for "powerful AI"Machines of Loving Grace, 20242024
Ajeya Cotra · Open Philanthropy~2050 for transformative AI (median)Bio-anchors model; updated downward2022

All estimates are conditional on no compute crackdown, no major war, no ceiling discovered. Variance across these sources is the most honest data point.

§5 · Alignment

The alignment problem, in one paragraph

We can train models to maximize a reward signal. We cannot reliably train them to want what we actually want. As capability increases, the gap between "does what we said" and "does what we meant" becomes more dangerous. The technical research agenda — interpretability, scalable oversight, debate, RLHF, constitutional AI, model organisms of misalignment — is the field's response.

Five canonical concerns

  • Specification gaming — clever solutions that satisfy the metric, not the goal
  • Goal misgeneralization — same training, different deployment, wrong goal
  • Mesa-optimization — learned models that themselves optimize
  • Deceptive alignment — models that look aligned during training
  • Power-seeking — instrumentally convergent subgoals (Omohundro, Bostrom)
datacenter
Compute · the substrate of the argument.
§6 · Figure 2 · Scenario tree

Four scenarios for the 2030s

2026 A · Aligned takeoffpost-scarcity, gradual deployment B · Muddled normalityeconomic shock + adaptation C · Capability plateauscaling stalls, niche utility D · Catastrophemisuse · misalignment · concentration P (rough Metaculus medians, illustrative) ~ 0.20 ~ 0.45 ~ 0.20 ~ 0.10–0.15 probabilities are contested. FIG 2 · scenario tree · 2026 → 2035
Branches A, B, C, D abbreviated from Christiano (2021), Karnofsky (2021), Aschenbrenner (2024), Yudkowsky (2022). Probabilities illustrative; honest people disagree by 0.3+ on each branch.
§7 · Actors

Who builds, who governs, who studies

Lab / orgStanceNotable output
OpenAIMission to build "safe" AGI; capped-profitGPT-4, o1, Sora, Operator
AnthropicSafety-first frontier labClaude, RSP, mechanistic interp
Google DeepMindResearch-driven, Hassabis-ledAlphaFold, Gemini, AlphaProof
Meta AIOpen-weights, LeCun-ledLlama series, JEPA
xAIFrontier ramp + Twitter integrationGrok, Memphis cluster
UK AISI / US AISIGovernment safety institutes (2023–)Pre-deployment evals, model access
EU AI ActRisk-tier regulation, in force 2024–27Foundation-model rules, Article 56
FHI / GovAI / MIRIAcademic-adjacent x-risk researchBostrom, Dafoe, Yudkowsky
§8 · Quote
"The first ultraintelligent machine is the last invention that man need ever make, provided that the machine is docile enough to tell us how to keep it under control."
— I.J. Good, Speculations Concerning the First Ultraintelligent Machine, 1965

Good's paper is the foundational text. Vinge (1993) named the resulting threshold "the singularity". Bostrom (2014) gave it a 360-page treatment. Yudkowsky (2008–) gave it the more pessimistic gloss.

§9 · Governance

Five live policy debates

forecast By 2030, compute reporting will be normalized in OECD countries; binding international AI treaty unlikely.

§10 · Skeptics

The case it isn't happening soon

Emily Bender & Alex Hanna argue that LLMs are "stochastic parrots" — fluent but ungrounded. The capability claims, they say, are marketing.

Gary Marcus has documented persistent failure modes: planning, robust reasoning, world-model maintenance. Current models still hallucinate, cannot reliably count, fail at simple physical inference.

Yann LeCun argues autoregressive LLMs cannot reach AGI; we need world-model architectures (JEPA, video prediction with abstract latents).

Melanie Mitchell emphasizes the abstraction gap revealed by ARC and similar benchmarks.

Sayash Kapoor & Arvind Narayanan ('AI Snake Oil') stress that "agentic AI" demos do not survive contact with messy real-world deployment.

The skeptics may be right. The scaling lab leaders may be right. They may both be right at different timescales.

§11 · Watch

Recommended source

Anthropic Talks · "Mechanistic Interpretability: A Field Guide"

A research-grade walk through what we can and cannot currently see inside a transformer. Pair with Chris Olah's "Zoom In" (2020).

youtube.com/@anthropic-ai →

Lex Fridman × Demis Hassabis (#475)

Three-hour conversation covering AlphaFold, Gemini, scaling, alignment, and the 5–10-year claim. Slow but high-yield.

youtube.com/@lexfridman →
§12 · What to watch

Indicators · 2026–2030

Cite as: The Deck Catalog, Vol 6, Paper 03, 2026. Critique welcome.

01020304050607080910111213