Edition 2026-04-18 · Model v1.4.0 · AGI · Singularity · ASI

The world has never had a calibrated instrument for the most consequential transition in human history. This one runs in real time.

Three data-driven clocks tracking AGI, the Singularity, and Superintelligence, derived from a public ensemble of capability, compute, economic, and alignment signals. Every movement sourced. Every methodology versioned.

PRIMARYOPEN →

AGI ClockArtificial General Intelligence

12 mo
--h --m --s
EST. APR 2027 · 5 SIGNALS
One system, all the cognitive work, no hand-holding.
A single AI system that autonomously performs the full scope of economically valuable cognitive work of a median human knowledge worker, at or above human quality, without task-specific fine-tuning.
Projection, not prediction
Model v1.4.0 · since Apr 15, 2026
CLOCK 2OPEN →

Singularity ClockRecursive Self-Improvement Inflection

3 yr 8 mo
--h --m --s
EST. DEC 2029 · 2 SIGNALS
AI improving AI, faster than humans can.
The inflection point at which AI-driven improvement to AI systems outpaces human-driven improvement, creating a self-reinforcing acceleration in capability gains. Operationalized via two proxies: AI-authored frontier research share crossing 50%, and METR task horizon reaching 1,000 hours (month-scale autonomous research capacity).
Projection, not prediction
Model v1.4.0 · since Apr 15, 2026
CLOCK 3OPEN →

Superintelligence ClockArtificial Superintelligence

5 yr 8 mo
--h --m --s
EST. DEC 2031 · 3 SIGNALS
Beyond the best human, in essentially every domain.
Cognitive capability qualitatively beyond the best human across essentially every economically relevant domain. Operationalized as HLE 95%+ and FrontierMath saturated, or Singularity + 2-year buffer, whichever is later.
Projection, not prediction
Model v1.4.0 · since Apr 15, 2026
Alignment Deficit OPEN →
Capability-to-Safety Velocity Ratio
Moderate
RATIO 1.71× · 39.3% OF SCALE
Keeping pace Moderate Elevated Critical
How fast capability grows relative to safety.
Composite ratio of capability velocity to safety velocity. Blended from structured R&D/safety metrics (60%), arXiv capability-vs-alignment flux (20%), and frontier-lab release flux (20%). 1.0× = pacing; above 1.5× the deficit starts to accumulate.
Gauge, not prediction
Model v1.4.0 · since Apr 15, 2026

Live Signal Readings

11 signals tracked · open source methodology
METR Task Horizon
11.4 hr
+38% QoQ
SWE-bench Verified
88.2%
+4.1pp QoQ
GPQA Diamond
94.8%
+2.3pp QoQ
ARC-AGI-2
62.4%
+18pp QoQ
FrontierMath
41.7%
+12pp QoQ
Humanity's Last Exam
38.9%
+16pp QoQ
Frontier training FLOPs
8.2e26
2.1× YoY
AI R&D capex (global)
$512B
+64% YoY
AI-authored research
14.2%
+5.3pp YoY
Interpretability index
0.31
-0.02 QoQ
Frontier lab safety headcount
12.4%
-1.1pp YoY

Signals from the Field

995 arXiv papers · 318 lab/org posts · LLM-classified
2026-04-16
CAPABILITY LEAP
Agent-Aided Design for Dynamic CAD Models
First agent-aided design system capable of generating 3D assemblies with moving parts; extends agentic capabilities to novel domain.
arXiv · arXiv
Mag 4
2026-04-16
CAPABILITY LEAP
From Procedural Skills to Strategy Genes: Towards Experience-Driven Test-Time Evolution
Demonstrates effective test-time evolution strategy for code-solving; compact gene representation achieves substantial performance gains (9-18% improvement) on scientific tasks.
arXiv · arXiv
Mag 4
2026-04-16
CAPABILITY LEAP
OpenMobile: Building Open Mobile Agents with Task and Trajectory Synthesis
Open mobile agent framework achieving 64.7% on AndroidWorld with transparent task/trajectory synthesis; competitive with closed SOTA and advances autonomous agent capability.
arXiv · arXiv
Mag 4
2026-04-16
CAPABILITY LEAP
Autonomous Evolution of EDA Tools: Multi-Agent Self-Evolved ABC
Self-evolving logic synthesis system using LLM agents to autonomously improve EDA tools at scale; learns new synthesis strategies beyond human heuristics.
arXiv · arXiv
Mag 4
2026-04-16
CAPABILITY LEAP
Towards Faster Language Model Inference Using Mixture-of-Experts Flow Matching
40×-1000× speedup in language model inference with maintained quality is substantial efficiency advance; 3-step NAR generation is qualitatively new.
arXiv · arXiv
Mag 4
2026-04-16
CAPABILITY LEAP
Discovering Novel LLM Experts via Task-Capability Coevolution
Novel coevolution framework (AC/DC) discovers diverse LLM expertise via model merging and synthetic task generation. Demonstrates broader capability coverage than larger models with less compute.
arXiv · arXiv
Mag 4
2026-04-16
MODEL LAUNCH
Introducing GPT-Rosalind for life sciences research
Frontier reasoning model launch for life sciences; significant new model capability.
Lab · OpenAI
Mag 4
2026-04-02
MODEL LAUNCH
Gemma 4: Byte for byte, the most capable open models
Major open model family release with advanced reasoning and agentic capabilities.
Lab · Google DeepMind
Mag 4
2026-04-02
MODEL LAUNCH
Welcome Gemma 4: Frontier multimodal intelligence on device
Gemma 4 frontier multimodal model launch; major version advancement.
Lab · Hugging Face
Mag 4
MODEL LAUNCH
Introducing Claude Opus 4.7
New Claude Opus version release from frontier AI lab Anthropic.
Lab · Anthropic
Mag 4

Recent Clock Movements

Last 90 days · Why two streams?
Apr 15, 2026
AGI
METHODOLOGY Methodology v1.4.0. AGI projection moved later (2026-12-15 → 2027-04-16) after three changes. (1) Benchmark signals (SWE-bench, HLE, ARC-AGI-2, FrontierMath) now use sigmoid (saturating-exponential) extrapolation instead of linear. Benchmarks are bounded; linear models overstate how fast the last 10pp closes and were pulling the AGI date ~4 months too early. (2) METR task-horizon weight raised 0.35 → 0.45: it is the only direct economic-capability measure and should dominate benchmark proxies. (3) Metaculus weight raised 0.15 → 0.20 as a reserved slot; currently absent because Metaculus gates Community Prediction data to a whitelist that excludes AGI questions (access request sent to api-requests@metaculus.com on 2026-04-15). Benchmark aggregate weight dropped 0.50 → 0.35. Source: lib/projection.py v1.4.0 project_agi · 2026-04-15
+122 days
Apr 15, 2026
SING
METHODOLOGY Methodology v1.4.0. Singularity projection moved earlier (2032-11-26 → 2029-12-16) after reweighting. METR task horizon (→1000hr, month-scale autonomous research) weight raised 0.20 → 0.35 as a leading indicator; AI-authored research share (→50%) weight lowered 0.50 → 0.35 because it is a lagging proxy (a linear fit to a pre-phase-transition variable is structurally conservative once capability crosses research-scale horizons). Metaculus+3yr heuristic unchanged at 0.30. Linear extrapolation retained for AI-authored share since it has no natural ceiling at low values. Net: more weight on the leading indicator, which produces the more compressed date. Source: lib/projection.py v1.4.0 project_singularity · 2026-04-15
-1076 days
Apr 15, 2026
ASI
METHODOLOGY Methodology v1.4.0. ASI projection moved earlier (2033-11-26 → 2031-12-16). Three changes. (1) HLE→95% and FrontierMath→95% now use sigmoid extrapolation with weights raised 0.30 → 0.375 each. (2) Singularity+2yr weight reduced 0.40 → 0.25 so ASI does not inherit Singularity's inertia. (3) Cascade minimum buffer from Singularity to ASI widened from 365d to 730d (MIN_BUFFER_DAYS_SING_TO_ASI), reflecting that ASI (qualitatively beyond best human, all domains) is not a one-year hop past recursive-improvement onset. The new date is pinned by cascade (Singularity 2029-12 + 730d = 2031-12); benchmark signals independently project earlier (2028-08 and 2029-03). Source: lib/projection.py v1.4.0 project_asi · 2026-04-15
-711 days
Apr 15, 2026
ALIGN
METHODOLOGY Methodology v1.4.0. Alignment Deficit formula unchanged in this release; gauge reading holds at 1.7× (Moderate). v1.4.0 is a clocks-only rework: sigmoid benchmark extrapolation, reweighted AGI/Singularity/ASI ensembles, widened ASI cascade buffer, plus editorial additions (first-visit disclaimer modal, Limitations methodology section, api.json and RSS/JSON-Feed disclaimer fields). The Alignment blend stays at 0.6 structured + 0.2 arXiv flux + 0.2 frontier flux. Next alignment methodology work will integrate a direct safety-capex proxy when the data source stabilizes. Source: v1.4.0 release notes · 2026-04-15
gauge change
Apr 14, 2026
ALIGN
METHODOLOGY Gauge reading moved Keeping pace → Moderate (1.19× → 1.61×) with the v1.3.0 methodology in effect. This is expected given the structured rework: the prior flow-vs-stock formula mathematically couldn't see that safety headcount share was shrinking and interpretability was declining. Under growth-vs-growth the structured ratio is 1.98×, which blends with arxiv flux (1.11×, post-symmetrization) and frontier flux (1.00×) at 60/20/20 to 1.61×. No real-world change today, the gauge is reflecting an existing reality that prior methodology obscured. Source: v1.3 post-release reading · 2026-04-14
gauge change
Apr 14, 2026
ALIGN
METHODOLOGY Methodology v1.3.0 also symmetrized the arXiv capability-vs-alignment discount. Prior v1.2 formula weighted capability_incremental at 0.25× while alignment_eval_result got 0.5×, an asymmetric thumb on the scale that compressed the ratio toward 1.0 and systematically understated the deficit. Both 'softer' sub-categories now receive the same 0.5× multiplier; landmark categories (capability_leap, alignment_advance) keep full weight. Observed effect: arXiv ratio moved ~1.11× → ~1.48× in subsequent runs (exact value depends on daily classifier output). Source: lib/ingest/arxiv_classifier.py::_aggregate v1.3 · 2026-04-14
gauge change
Apr 14, 2026
ALIGN
METHODOLOGY Methodology v1.3.0: structured composite reworked from flow-vs-stock to growth-vs-growth. Capability velocity is YoY capex growth (1.64×); safety velocity is the equal-weight average of safety-headcount growth (0.91×, share is shrinking 1.1pp/yr off 12.4% base) and interpretability-index growth (0.74×, index is declining 0.02/quarter off 0.31 base). Structured ratio moved 1.28× → 1.98×. The prior formula compared a flow (capex growth) to a stock (headcount level), which mathematically hid the trend. Floored at 0.25 to prevent near-zero growth blowing up the ratio. New safety_headcount_growth and interpretability_growth fields exposed in api.json for audit. Source: lib/projection.py::project_alignment v1.3 · 2026-04-14
gauge change
Apr 14, 2026
ALIGN
METHODOLOGY Methodology v1.3.0: arXiv classifier discount symmetrized. Previous formula weighted capability_incremental papers at 0.25× while alignment_eval_result got 0.5×, an asymmetric thumb on the scale that compressed the ratio toward 1.0. Both 'softer' categories now get 0.5×; landmark categories (capability_leap, alignment_advance) keep full weight. arXiv-flux ratio will re-aggregate at the next ingest cycle. Source: lib/ingest/arxiv_classifier.py::_aggregate v1.3 · 2026-04-14
gauge change
Apr 14, 2026
ALIGN
METHODOLOGY Alignment Deficit reading moved Keeping pace → Moderate (1.19× → 1.61×) following the v1.3 methodology rework. This is a measurement-bias correction, not a real-world worsening: the prior reading was understating the deficit because the structured formula compared a growth rate (capex) to a level (safety headcount). With both sides now expressed as growth rates, the gauge correctly reflects that capability investment is growing 64% YoY while safety headcount share and interpretability are both declining. The previous 'Keeping pace' reading should be considered retrospectively corrected. Source: Alignment gauge v1.3 transition note · 2026-04-14
gauge change
Apr 14, 2026
ALIGN
METHODOLOGY Alignment Deficit reading moved Elevated → Keeping pace (3.0× → 1.29×). This is a methodology effect from the v1.2 blend introduction, not a real-world improvement: the structured ratio (1.28×) is now diluted with arXiv flux (1.35×) and frontier-release flux (1.28×) at 20% weights each. Note three known biases all pulling the blend low: (a) structured formula compares capability flow to safety stock, hiding that safety headcount is shrinking 1.1pp/yr while capex grows 64% YoY; (b) frontier-flux undercounts the safety side because Anthropic, Apollo, METR, Epoch, Meta, Mistral, and xAI lack discoverable RSS feeds; (c) arXiv flux applies a 4× discount to incremental capability that is asymmetric with the alignment-side discount. A more honest reading is likely Moderate (~1.8–2.4×). Restoring the safety-feed coverage and symmetrizing the arXiv discount are tracked for v1.2.1; structured-formula rework is tracked for v1.3. Source: Methodology doc: alignment blend transparency · 2026-04-14
gauge change
Apr 14, 2026
ALIGN
METHODOLOGY Frontier-feed coverage update: dropped 7 lab feeds that 404 with no discoverable RSS (Anthropic, Apollo, Epoch, Meta, Mistral, xAI, plus broken METR URL). Added 3 working safety-side feeds: METR (correct URL: metr.org/feed.xml), MIRI, and LessWrong curated. Added Google Research blog. Net effect: safety-side feed coverage roughly doubles, partially offsetting the structural undercount flagged in the prior entry. Source: data/frontier_feeds.json · 2026-04-14
gauge change
Apr 14, 2026
ALIGN
METHODOLOGY Methodology v1.2.0: Alignment Deficit blend now includes LLM-classified frontier-lab and safety-org RSS feeds (trailing 90 days). Captures launches, capability claims, and safety/policy posts that don't appear on arXiv. Final blend: 60% structured + 20% arXiv flux + 20% frontier-release flux. New 'Signals from the Field' site section surfaces highlighted papers and lab posts. Source: Methodology doc: frontier release monitor · 2026-04-14
gauge change
Apr 14, 2026
ALIGN
METHODOLOGY Methodology v1.1.0: Alignment Deficit now blends LLM-classified arXiv capability/alignment flux (trailing 30 days, cs.AI/cs.LG/cs.CL) with existing structured inputs at 30% weight. Grounds the gauge in observed research output, not only lab-reported headcount and spend. Clock projections unchanged. Source: Methodology doc: arXiv classifier · 2026-04-14
gauge change
Apr 12, 2026
AGI
SIGNAL AGI Clock moved 11 days closer following METR task-horizon update: autonomous task completion length crossed 11 hours, passing the 10-hour inflection 4 months ahead of model expectations. Source: METR Task Horizon v14 release · 2026-04-12
-11 days
Mar 28, 2026
SING
SIGNAL Singularity Clock moved 6 days closer after Anthropic publication on automated red-teaming showed AI-driven interpretability contributions at frontier lab level. Source: Anthropic research post · 2026-03-28
-6 days
Mar 19, 2026
ALIGN
SIGNAL Alignment Deficit gauge moved from Moderate to Elevated: capability/safety velocity ratio crossed 3.0× for first time since index inception. Source: TSC composite index · 2026-03-19
gauge change
Mar 4, 2026
ASI
SIGNAL ASI Clock moved 3 days closer following DeepMind announcement of novel algorithm discovery in combinatorial optimization: first confirmed superhuman contribution to an active research frontier. Source: DeepMind Nature paper · 2026-03-04
-3 days
Feb 21, 2026
AGI
SIGNAL AGI Clock moved 8 days closer after SWE-bench Verified saturation projections updated: extrapolation model now projects 95% by Q2 2027. Source: Epoch AI forecast update · 2026-02-21
-8 days

Methodology

v1.4.0 · Open methodology · Versioned model

What we count to: AGI

A single AI system that can autonomously perform the full scope of economically valuable cognitive work of a median human knowledge worker, sustained across a standard 40-hour work-week, at or above human quality, without task-specific fine-tuning and without human correction.

Operationally testable. Grounded in economic reality. Falsifiable in both directions. Sidesteps unresolvable debates about consciousness or "real" intelligence.

What we count to: Singularity & ASI

Singularity: The inflection point at which AI-driven improvement to AI systems outpaces human-driven improvement, creating a self-reinforcing acceleration in capability gains. Proxy thresholds: AI-authored frontier research crosses 50%, and METR task horizon reaches 1,000 hours.

ASI: Cognitive capability qualitatively beyond the best human in essentially every economically relevant domain. Proxy: HLE saturation (95%+) and FrontierMath saturation, or Singularity + 2-year compute/recursion buffer, whichever is later.

How we project

Each clock is a weighted median across its signal ensemble. Benchmark signals use linear extrapolation to defined saturation thresholds. Capability signals (METR task horizon) use exponential extrapolation with empirical doubling periods. Crowd-forecast signals (Metaculus) enter directly as median predictions.

Confidence intervals reflect signal disagreement (10th–90th percentile of per-signal projections), not statistical uncertainty in any single signal. Wider band = more disagreement among signals. We're more uncertain when signals diverge, tighter when they converge.

Cascade ordering

The three clocks must respect causal ordering: ASI cannot precede Singularity, which cannot precede AGI. Minimum buffers are enforced: AGI ≤ Singularity − 365 days ≤ ASI − 730 days.

When independent signal ensembles produce an incoherent ordering, the later clocks are pinned to the earlier clock plus the buffer, and cascade_adjusted=True is recorded so the adjustment is transparent rather than hidden.

Alignment Deficit blend (v1.2)

The Alignment Deficit gauge is the ratio of capability velocity to safety velocity. As of model v1.2 it blends three independent inputs:

  • 60%: structured composite (R&D capex YoY, frontier-lab safety headcount %, interpretability index)
  • 20%: arXiv flux (LLM-classified cs.AI/cs.LG/cs.CL papers, trailing 30 days)
  • 20%: frontier-release flux (LLM-classified frontier-lab and safety-org RSS, trailing 90 days)

Each flux input is clamped to a sane band [0.25×, 6.0×] to prevent single-week spikes from dominating, and skipped if the sample size is too small. Weights redistribute when an input is unavailable.

Editorial posture

Sober. Transparent. Non-partisan. Quantitative. Continuously updated. We don't use hype language ("imminent," "superhuman," "god-like"). We don't pick a side in the AI-optimist/doomer debate; we present signals and let readers form their own view. Every claim on the site traces to a source URL. Every methodology change bumps the model version and is recorded in the changelog.

Why we version the model

An opaque clock is an opinion in a lab coat. We publish every revision, every weight, every threshold, and every formula on GitHub so you can argue with the data and the method, and reproduce any past reading from a git hash.

That is also why the Recent Clock Movements panel splits into two streams. Signal entries are the world moving the clock: a benchmark advanced, a paper dropped, a gauge input changed. Methodology entries are us moving the instrument: reweighting signals, replacing a formula, widening a cascade buffer.

A methodology revision can shift a clock by years on the day it ships, while the underlying world did not change at all. Mixing the two in a single stream hides that distinction; separating them makes it legible. The current model is v1.4.0, in effect since Apr 15, 2026.

Limitations

These projections are a research instrument, not a forecast you should act on. Specifically:

  • Benchmark saturation is a proxy for capability, not proof of AGI. Models can saturate a benchmark and still fall short of the underlying work.
  • Log-linear and linear extrapolation breaks near benchmark ceilings and at phase transitions; real capability curves are sigmoidal.
  • Signal availability changes (for example, Metaculus community predictions are currently gated). Missing signals widen the band or hide disagreement.
  • Wide confidence intervals reflect real uncertainty, not rounding. A clock sitting on the CI boundary can shift by months on a single ingest cycle.
  • The model is opinion, not prediction. Methodology and signal code are public so you can disagree with both.
  • The Alignment Deficit is a gauge of research intensity, not a measure of actual alignment; a low reading does not mean systems are safe.

Do not use this site for investment, career, medical, legal, or life decisions.

Not advice. The Singularity Clock is a public-interest measurement instrument, not a forecast for personal use. Do not use it for investment, career, medical, legal, or life decisions. Projections are opinion derived from imperfect signals and carry wide uncertainty. See Methodology & limitations.
© 2026 THE SINGULARITY CLOCK · MODEL v1.4.0 · DATA CC-BY 4.0
LAST UPDATED 18 APR 2026 · 08:43 UTC