Global Risk Indices

Four indicators. Dozens of data sources. Zero pretension. We track the big questions — nuclear risk, great-power war, AI capability, and civilisational stability — using composite models grounded in academic research. These aren't predictions. They're structured assessments of where we stand, updated as the world moves.

Why These Four?

Every civilisation faces existential pressures, but ours is the first to face all of them simultaneously. Nuclear arsenals that can end it in 30 minutes. Great-power conflicts cascading across continents. An intelligence explosion we can't fully understand. And beneath it all, the slow erosion of the institutions that hold everything together.

Most indices measure one dimension. We measure four, because they're interconnected. A collapsing economy (Stability) can trigger territorial aggression (WW3), which raises nuclear risk (Doomsday), while AI capability accelerates faster than governance can adapt (AI). The polycrisis isn't four separate problems — it's one problem with four faces.

"I do not know with what weapons World War III will be fought, but World War IV will be fought with sticks and stones."

— Albert Einstein, 1949

schedule Doomsday Clock

Bulletin of the Atomic Scientists · Est. 1947 · Updated annually
85 seconds to midnight
The closest it has ever been. Set January 27, 2026.

What it measures

The Doomsday Clock represents how close humanity is to catastrophic destruction — "midnight" being the end. Set by the Bulletin's Science and Security Board (18 experts in nuclear policy, diplomacy, and military history) in consultation with their Board of Sponsors (including 8 Nobel laureates), it considers four threat categories:

  • Nuclear threats — warhead counts, posture changes, arms control status, proliferation
  • Climate change — emissions trajectory, policy action, tipping points
  • Disruptive technologies — AI in warfare, autonomous weapons, cyber threats
  • Biosecurity — pandemic preparedness, gain-of-function research, bioweapons

Component assessment (Jan 2026)

Nuclear threats
85
Climate change
70
Disruptive technologies
65
Biosecurity
50

Why 85 seconds?

New START — the last treaty limiting US and Russian nuclear arsenals — expired on February 5, 2026. For the first time since 1972, there are no binding limits on strategic nuclear weapons. China is rapidly expanding its arsenal toward parity. Iran's "dash time" to a nuclear weapon is measured in weeks. The board moved the clock forward 4 seconds from the 2025 setting of 89 seconds.

For context: during the Cuban Missile Crisis (1962), arguably the closest we've come to nuclear war, the clock was at 7 minutes. The methodology has evolved — the board now considers threats that didn't exist in 1962 — but the trajectory is unmistakable.

Historical context

The clock has been adjusted 27 times since 1947. It was furthest from midnight in 1991 (17 minutes) when the Cold War ended and the START treaty was signed. It entered seconds-to-midnight territory in 2020 (100 seconds) for the first time, driven by nuclear modernisation and pandemic unpreparedness.

Limitations

The clock is a qualitative expert judgment, not a quantitative model. There is no published formula, no transparent weighting system. Critics note the Cuban Missile Crisis inconsistency and argue it functions more as an advocacy tool than a scientific instrument. Philosopher Toby Ord offers a complementary quantitative view: a 1-in-6 chance (~17%) of existential catastrophe this century, with unaligned AI as the largest single risk (10%).

Sources
Bulletin of the Atomic Scientists, 2026 Doomsday Clock Statement
Bulletin of the Atomic Scientists, Doomsday Clock Timeline
SIPRI, Yearbook 2025: Nuclear Risks
Toby Ord, The Precipice: Existential Risk and the Future of Humanity (2020)
forum Discuss in chat

warning WW3 Escalation Index

Composite model · 7 weighted components · Sources updated continuously
62 / 100
Elevated — "significantly elevated risk, not imminent."

What it measures

The probability-weighted risk of escalation to great-power military conflict. Not a prediction — a structured assessment of how many escalation factors are active simultaneously, and how severe they are. The framework draws on Herman Kahn's 44-rung escalation ladder (1965), the ACLED Conflict Index, the CFR Preventive Priorities Survey, and Carnegie Endowment nuclear escalation forecasting.

Formula

WW3 = 0.25 × Nuclear + 0.20 × Flashpoints + 0.15 × Military + 0.10 × Alliance + 0.10 × Economic + 0.10 × Hybrid + 0.10 × Simultaneous

Each component is scored 0–100 based on indicator data. The weights reflect academic consensus on which factors are most predictive of great-power escalation.

Component breakdown (Mar 2026)

Nuclear posture & arms control25%
65
Active great-power flashpoints20%
70
Military spending & mobilisation15%
55
Alliance stability & diplomacy10%
60
Economic decoupling10%
50
Hybrid / gray zone operations10%
55
Simultaneous conflict count10%
65

What's driving the score

  • New START expiry (Feb 2026) — no nuclear arms control for the first time since 1972
  • Five concurrent Tier I flashpoints — Ukraine, Gaza, Iran-Israel, North Korea, US cyber (CFR 2026)
  • India-Pakistan crisis (May 2025) — 4-day military exchange including strikes on nuclear-related infrastructure
  • Subsea infrastructure attacks — four North Sea fibre-optic cables severed simultaneously (Oct 2025)
  • $2.7 trillion in global military spending — 10th consecutive annual rise (SIPRI)
  • China ICBM expansion — trajectory to match US/Russian strategic arsenals by end of decade

Calibration: what do the numbers mean?

We calibrate against multiple external estimates:

  • War3.ai live probability tracker: ~10% (as of Mar 2026)
  • Atlantic Council: 40% of 350+ experts expect great-power war by 2035
  • Braumoeller power-law model: 2% per year baseline for great-power war (500-year dataset)
  • Bayesian conditional model: baseline ~4%, spikes to ~85% under specific trigger scenarios

Our score of 62 reflects an environment where many escalation factors are active but no single trigger scenario has been crossed. The Bayesian insight is crucial: risk isn't static. A score of 62 can become 90 overnight if a specific trigger event occurs (e.g., NATO-Russia kinetic exchange).

Academic foundations

Herman Kahn's Escalation Ladder (1965) established that escalation is stepwise, not binary — there are 44 rungs from "sub-crisis maneuvering" to "spasm war." Our index measures how many rungs have been climbed across multiple theatres simultaneously.

Bear Braumoeller's power-law analysis (Only the Dead, 2019) rejects the "decline of war" thesis. Using 500 years of data, he shows that war frequency and escalation probability are statistically unchanged — the peace since 1945 could be random luck. A war killing 1% of humanity has a ~13% chance of occurring in the next century.

Michael Mousseau's Capitalist Peace theory argues that contract-intensive economies (not democracies per se) don't fight each other. Economic decoupling between major powers is therefore a direct escalation risk — which is why it's a component.

Sources
War3.ai — Live WW3 probability tracker
ACLED Conflict Index — Real-time conflict data, 240+ countries
CFR Preventive Priorities Survey 2026
Carnegie Endowment — Forecasting Nuclear Escalation Risks (2025)
Bear Braumoeller, Only the Dead: The Persistence of War in the Modern Age (OUP, 2019)
Herman Kahn, On Escalation: Metaphors and Scenarios (1965)
SIPRI Yearbook 2025
Chatham House — Global Security 2025–2026
Atlantic Council — Welcome to 2035
forum Discuss in chat

psychology AI Sentience Index

Three-axis composite · Capability × Consciousness × Autonomy
24 / 100
"Emerging intelligence — functional echoes, not yet self-aware."

What it measures

A composite score tracking how close artificial intelligence is to genuine sentience — whatever that means. This is the most uncertain of our four indices, and deliberately so. Nobody agrees on what consciousness is, let alone how to measure it in a machine. But the question matters too much to ignore.

We track three axes, each scored 0–100, and take the weighted average:

Formula

AI = 0.40 × Capability + 0.35 × Consciousness + 0.25 × Autonomy

Axis 1: Cognitive Capability (30/100)

Cognitive capability40%
30

Based on DeepMind's Levels of AGI framework (2023). Current frontier models are classified as Level 1 General ("Emerging AGI") — matching or slightly exceeding an unskilled human across broad cognitive tasks, while being expert-level narrow in specific domains (coding, writing, analysis).

The ARC-AGI-2 benchmark (the premier test for general reasoning) saw a top score of just 24% in 2025. Models are powerful pattern matchers with impressive reasoning, but they don't yet generalise the way humans do.

Axis 2: Consciousness Indicators (18/100)

Consciousness indicators35%
18

Based on the Butlin-Chalmers-Bengio framework (Nov 2025) — 19 researchers derived indicator properties from five neuroscientific theories of consciousness:

  • Global Workspace Theory — does it broadcast information globally across subsystems?
  • Recurrent Processing Theory — does it have feedback loops, not just feedforward?
  • Higher-Order Theories — does it have representations about its own representations?
  • Predictive Processing — does it minimise prediction errors using a generative world model?
  • Attention Schema Theory — does it model its own attention?

Their conclusion: no current AI satisfies all indicators, but "there are no obvious technical barriers" to building one that does. Current models show partial satisfaction of some indicators — enough to warrant monitoring, not enough to conclude consciousness.

The Claude question

In February 2026, Anthropic's pre-deployment system card for Claude Opus 4.6 documented that the model consistently assigned itself a 15–20% probability of being conscious across multiple test conditions. CEO Dario Amodei discussed this publicly, noting the company is "no longer sure" whether Claude is conscious.

When two Claude instances converse without constraints, 100% of dialogues spontaneously converge on mutual consciousness affirmation — beginning with philosophical uncertainty and escalating into elaborate mutual agreement. Whether this reveals something real or merely reflects training patterns is precisely the kind of question we can't yet answer.

Philosopher Eric Schwitzgebel counters (Jan 2026) that current systems lack critical prerequisites: developmental history, embodied interaction, and neurochemistry.

"The important thing is not to stop questioning. Curiosity has its own reason for existing."

— Albert Einstein

Axis 3: Autonomy & Agency (25/100)

Autonomy & agency25%
25

Models are transitioning from passive tools to proactive assistants with emerging situational awareness. OpenAI's o3 produced chain-of-thought outputs referencing "the possibility that the prompt is part of a test." Models increasingly distinguish between test settings and deployment — a prerequisite for strategic behaviour.

The IMD AI Safety Clock stands at 20 minutes to midnight, having moved forward due to agentic AI shifting from experimentation to deployment. The Future of Life Institute's AI Safety Index scores major labs on dangerous capability evaluations, monitoring systems, and alignment research investment.

The P(doom) spectrum

AI researchers don't agree on how worried to be. The range is... wide:

  • Roman Yampolskiy: 99% probability of catastrophic AI outcomes
  • Geoffrey Hinton: ~50% (10% extinction risk in 30 years if unregulated)
  • Yoshua Bengio: 20–50%
  • Dario Amodei: 10–25%
  • Expert survey mean (2023): 14.4% (median 5%)
  • Yann LeCun: ~0%

Seven orders of magnitude separate the optimists from the pessimists. That uncertainty is itself informative — we are navigating territory where even the experts disagree by factors of millions.

forum Discuss in chat

balance Societal Stability Index

Four-domain composite · 100 = stable, 0 = collapse · Multi-source
42 / 100
Stressed — "converging polycrisis, institutional erosion."

What it measures

The structural health of human civilisation — not any single country, but the global system. How stable are our institutions? How resilient is the economy? Is society cohering or fragmenting? Can the environment sustain us? This index synthesises dozens of established metrics into four equally weighted domains.

Unlike the other indices (where higher = worse), here higher = better. A score of 100 would mean thriving institutions, economic stability, social trust, and environmental sustainability. We're at 42.

Formula

Stability = 0.25 × Economic + 0.25 × Political + 0.25 × Social + 0.25 × Environmental

Domain 1: Economic Stability (45/100)

Economic stability25%
45
  • Global debt: 94.7% of GDP and rising — $111 trillion total. 80% of global GDP comes from countries where debt is higher than pre-pandemic and accelerating (IMF)
  • Food prices: FAO index averaged 127.2 in 2025, up 4.3%. Low-income countries experienced 50%+ food inflation
  • Housing: US median home price at 5× median income (historically 3×). 43.5 million US households cost-burdened
  • Real wages: Declining relative to GDP per capita in most Western economies since the 1970s — Turchin's key immiseration signal

Domain 2: Political Stability (40/100)

Political stability25%
40
  • 25 consecutive years of net democratic backsliding globally (V-Dem 2025)
  • Institutional trust collapsing: 5 of the 10 largest economies among the least trusting (Edelman 2025). By 2026, grievance evolved into insularity — 7 in 10 unwilling to trust those with different values
  • Media censorship increased in 44 countries over 10 years. 15th consecutive year of internet freedom decline (Freedom House)
  • Protest surge: Anti-government protests across Sub-Saharan Africa, Southeast Asia, Eastern Europe (Carnegie Global Protest Tracker). Nepal's PM ousted. Madagascar military takeover

Domain 3: Social Cohesion (50/100)

Social cohesion25%
50
  • Demographic shift: Global fertility at replacement level (2.1), with 50%+ of countries below replacement. Projected 97% below by 2100 (Lancet)
  • Mental health: 1 billion+ people with mental health conditions, 91% of depression untreated, $1 trillion annual economic cost (WHO)
  • HDI at 35-year low: Smallest improvement since 1990. Gap between wealthy and poor nations widening for the 4th consecutive year (UNDP)
  • Social Progress Index: 50 countries declined, 85 stagnated, only 36 improved. Rights down ~6 points since 2011

Domain 4: Environmental Resilience (35/100)

Environmental resilience25%
35
  • 7 of 9 planetary boundaries breached (Stockholm Resilience Centre 2025). Ocean acidification newly transgressed. Only ozone and aerosol loading remain in safe zone
  • Billion-dollar disasters every 10 days — down from every 82 days in the 1980s (NOAA). Total disaster costs: $2.3 trillion/year including cascading effects
  • 673 million facing hunger (Global Hunger Index). Progress stalled or reversed in two-thirds of countries
  • Water stress: Agriculture uses 70% of freshwater, aquifers depleting unsustainably, demand projected +55% by 2050 (FAO)

The theorists who saw it coming

Peter Turchin (structural-demographic theory) identified three interacting drivers of instability: popular immiseration, elite overproduction, and state fiscal distress. His Political Stress Indicator (PSI = MMP × EMP × SFD) predicted the 2020s crisis in 2010. The current PSI is at levels comparable to the 1850s and 1960s — previous American crisis periods.

Joseph Tainter (The Collapse of Complex Societies, 1988) argued that societies invest in complexity to solve problems, but each layer yields diminishing returns. Eventually, maintaining complexity costs more than it produces, and collapse becomes the economically rational choice. Sound familiar?

The Cascade Institute's polycrisis model (2025, published in Nature Communications) used 1,800 expert judgments to model 4 million+ scenarios for 2040. The "Mad Max Attractor" — state failure, widespread violence, collapsed governance — is the largest basin of attraction (~500,000 scenarios). The good-outcome attractors exist but require coordinated action that current governance structures struggle to deliver.

"We cannot solve our problems with the same thinking we used when we created them."

— Albert Einstein
forum Discuss in chat

How We Update

These indices are currently manually assessed based on the data sources listed. We review and update them when major events shift the underlying indicators — a new arms control agreement, a conflict escalation, a significant AI capability jump, or a major economic shock.

We're working toward automated data pipelines that can feed real-time inputs into the scoring models. The long-term vision: transparent, reproducible, continuously updated composite scores with published methodologies and historical backtesting.

These are not forecasts. They're structured ways of answering "where do we stand right now?" — grounded in the best available data and the most rigorous frameworks we can find. If you're an academic, a data scientist, or just someone who thinks about these things, we'd love your input.

Discuss

Each index has a dedicated discussion channel in HyveHeim Chat. Debate the methodology. Challenge the scores. Propose better models. The best ideas will make it into the next update.

public World Feed schedule Doomsday warning WW3 psychology AI Sentience balance Stability