Government DataResearch

The Data Behind the Score: 12 Institutions, One Index

From the U.S. Department of Labor to Oxford University, Stanford HAI to the IMF — a detailed look at every data source powering our AI risk scores, how we pull the data, and how 12 methodologies become one number.

Escape Research TeamFebruary 24, 202612 min read

Any AI risk score is only as credible as the data behind it. A single study, no matter how well-designed, captures one perspective. That's why we built our index on 12 independent data sources — each approaching the question of AI displacement from a different angle, with different methodologies, different sample sizes, and different assumptions.

When 12 institutions independently point in the same direction, the signal is strong. When they diverge, it reveals nuance worth understanding. This article explains exactly what each source provides, how we use it, and where the data comes from.

12

Institutions

997

Occupations Mapped

80k+

Task Evaluations

Government Data Sources

The foundation of our index is U.S. government labor data — the most comprehensive, regularly updated, and publicly accessible occupational data in the world.

🏛

U.S. Department of Labor O*NET

What It Provides

The O*NET (Occupational Information Network) is the most comprehensive occupational database in the world. It contains detailed profiles for 997+ occupations including task descriptions, required skills, abilities, work activities, work context, and education requirements.

How We Use It

We use O*NET as our foundational layer. Every occupation in our index maps to an O*NET-SOC code. Task descriptions feed our Task Automation dimension — we map each of the 20,000+ individual tasks against AI capability benchmarks. Physical demands ratings feed our Physical Requirement dimension. Skill requirements help calibrate Creativity and Social Intelligence scores.

API / Data Access

O*NET Web Services API (developer.onetcenter.org) — RESTful JSON endpoints for occupation details, task listings, and cross-references. Updated quarterly.

Task AutomationPhysical RequirementCreativity DemandSocial Intelligence
📈

Bureau of Labor Statistics (BLS)

What It Provides

The BLS Occupational Employment and Wage Statistics (OES) program produces employment and wage estimates annually for over 800 occupations. The BLS also publishes the Occupational Outlook Handbook with 10-year employment projections.

How We Use It

BLS data enriches every individual report with real-time labor market context. When you search for a job, we pull current employment counts, median/mean wages, and growth projections. This data doesn't directly feed the composite score — it provides context about labor market size and trajectory.

API / Data Access

BLS Public Data API v2 (api.bls.gov) — We construct series IDs from SOC codes (format: OEUM{area}{industry}{SOC}{datatype}) to pull national employment and wage data. Cached with 24-hour revalidation.

Report enrichment — employment & wage context
📋

BLS Occupational Outlook Handbook

What It Provides

10-year employment projections for 800+ occupations, including projected job openings, growth rates, and industry shifts. Published biennially as part of the Employment Projections program.

How We Use It

Growth projections help contextualize risk — an occupation can score high on AI risk but still show employment growth due to expanding demand. We present these projections alongside AI risk scores to give a more nuanced picture in individual reports.

API / Data Access

Available via BLS series data and Employment Projections data files (bls.gov/emp).

Report enrichment — growth trajectory context

Academic Research

Peer-reviewed research provides the methodological backbone. These studies answer the foundational question: which tasks and cognitive capabilities are AI actually displacing?

🎓

Frey & Osborne (University of Oxford)

What It Provides

Carl Benedikt Frey and Michael Osborne's seminal 2013 study 'The Future of Employment' estimated automation probabilities for 702 U.S. occupations. Their methodology used machine learning expert classifications of 70 occupations to train a model that predicted automation probability for the remaining 632.

How We Use It

Their automation probabilities serve as calibration anchors for our Task Automation dimension. While our scoring system uses 6 dimensions (not just one probability), we validate our composite scores against Frey & Osborne's estimates — occupations they identified as highly automatable should score high in our index too. Significant deviations trigger manual review.

Task Automation Potential — calibration benchmark
🤖

OpenAI "GPTs are GPTs" (2023)

What It Provides

This paper analyzed which occupations are most exposed to GPT-family models by mapping AI capabilities to specific work tasks. Rather than broad automation probability, it assessed LLM exposure at the task level — distinguishing between tasks GPTs can do directly vs. with software built on top of them.

How We Use It

This is our primary input for the Cognitive AI Exposure dimension. The paper's task-level exposure scores directly map to our cognitive assessment. Jobs with high LLM task exposure score high on Cognitive Exposure regardless of their physical or regulatory characteristics.

Cognitive AI Exposure — primary data source
📊

Felten AI Occupational Exposure Index

What It Provides

Ed Felten's index maps 10 categories of AI capabilities (image recognition, language modeling, reading comprehension, etc.) to occupational ability requirements from O*NET. It produces an exposure score based on which AI capabilities are relevant to each job's required abilities.

How We Use It

We use Felten's index as cross-validation for our Task Automation and Cognitive Exposure scores. His capability-to-ability mapping provides an independent methodology to check our scoring. Where his index and our scores significantly diverge, we investigate whether our dimensional analysis missed something.

Cross-validation of Task Automation and Cognitive Exposure

International Organizations

Global institutions provide cross-country validation and macro-economic context. Their large-scale analyses help ensure our U.S.-focused scores aren't outliers.

🌎

World Economic Forum — Future of Jobs Report 2025

What It Provides

Surveys 800+ companies across 46 countries about expected changes in job roles, skills demand, and technology adoption over the next 5 years. Includes employer estimates of job creation, displacement, and skills evolution.

How We Use It

WEF data feeds our understanding of which skills are gaining vs. losing value. Their employer surveys help calibrate our Creativity and Social Intelligence dimensions — if employers report that creative and interpersonal skills are becoming more valuable, that validates our resilience scoring for those dimensions.

Creativity DemandSocial Intelligence — trend validation
💼

McKinsey Global Institute

What It Provides

McKinsey's automation research analyzes 800+ occupations across 46 countries at the activity level (not just the job level). They break each job into component activities and assess which activities can be automated with current or near-term technology.

How We Use It

McKinsey's activity-level analysis provides granular validation for our Task Automation scoring. Their distinction between technical automability and actual adoption speed informs how we weight near-term vs. long-term risk. Their cross-country data helps us understand how U.S.-specific our scoring might be.

Task AutomationPhysical Requirement — activity-level benchmarks
🌏

OECD Employment Outlook

What It Provides

The OECD's annual analysis covers G20+ labor markets with focus on automation risk, skills policy, and employment trends. Their methodology assesses risk at the task level within occupations, often finding lower automation estimates than Frey & Osborne because they account for task heterogeneity within jobs.

How We Use It

OECD data provides important methodological calibration. Their finding that most jobs are partially rather than fully automatable informs how we construct our composite — we don't treat automation as binary. Their regulatory analysis feeds our Regulatory Barriers dimension.

Regulatory BarriersComposite calibration methodology
👥

ILO Global Task Index

What It Provides

The International Labour Organization's task index classifies 30,000 tasks across occupations with 50,000 human evaluations of automation susceptibility, cognitive complexity, and physical requirements.

How We Use It

The ILO's massive task classification provides the deepest granular validation of our scoring. 50,000 human evaluations of task automability give us statistical confidence that our AI-based assessments align with expert human judgment. Where they disagree, human evaluations take precedence.

Task Automation — granular validation layer

AI & Policy Research

These sources track the cutting edge — what AI can actually do now, how fast capabilities are advancing, and which populations face the highest exposure.

🏫

Stanford HAI — AI Index Report

What It Provides

Stanford's Human-Centered AI Institute publishes an annual comprehensive report tracking AI technical performance, investment, adoption, policy, and societal impact. Includes benchmarks showing which cognitive tasks AI has surpassed human performance on.

How We Use It

HAI's capability benchmarks directly inform our Cognitive AI Exposure dimension. When their report shows AI surpassing humans on a new cognitive task (e.g., reading comprehension, legal reasoning), we re-evaluate which occupations use that task and adjust scores upward. This makes our index responsive to actual AI progress.

Cognitive AI Exposure — capability trending
🏛

Brookings Institution

What It Provides

Brookings' analysis maps AI exposure to 36 million U.S. workers with metropolitan-level granularity. They focus on which demographics and regions face the highest exposure, including analysis by education level, race, gender, and metro area.

How We Use It

Brookings provides demographic and geographic context that enriches our reports. While our composite score is occupation-level, Brookings data helps us understand which populations within an occupation face higher or lower risk. Their Social Intelligence analysis also feeds our resilience scoring.

Social IntelligenceRegional exposure context
💰

IMF GenAI Analysis

What It Provides

The IMF's 2024 analysis estimates that 40% of global employment is exposed to AI disruption across 108 countries. Advanced economies face higher exposure (60%) than emerging markets (26%) due to their larger share of cognitive-intensive occupations.

How We Use It

The IMF's global exposure estimates provide macro-level calibration. If 40% of jobs are significantly exposed globally, our scoring should reflect a similar proportion in the moderate-to-high risk range. Their country-level analysis helps contextualize U.S.-specific scores within a global framework.

Cognitive Exposure — global calibration layer

How We Synthesize 12 Methodologies Into One Score

Each institution uses different methods, different sample sizes, and different assumptions. That's a feature, not a bug. Convergence across independent methodologies is the strongest signal in research — and our synthesis process is designed to find it.

The process works in layers: O*NET provides the structural foundation (997 occupations, 20,000+ tasks). Academic research provides the AI capability mapping. International organizations provide macro validation. Policy research provides cutting-edge calibration.

When multiple sources agree that a job is high-risk, the composite score reflects that consensus strongly. When sources disagree, the composite reflects the central tendency — and the individual report flags the uncertainty.

The Data Pipeline

1

Ingest & Standardize

Map all 12 sources to O*NET-SOC occupation codes using crosswalk tables

2

Score 6 Dimensions

Each dimension draws from 2-4 sources, weighted by methodology quality and recency

3

Compute Composite

Vulnerability minus Resilience, calibrated to 1-10 scale against Frey & Osborne benchmarks

4

Enrich at Report Time

Pull live BLS employment/wage data, X/Twitter sentiment, and Reddit discussions via API

5

Generate with AI

Claude AI synthesizes all data into a personalized 20+ page career risk report

Data Pipeline APIs

Transparency means showing the technical plumbing too. Here are the specific APIs and data endpoints powering the system:

BLS Public Data API v2

api.bls.gov/publicAPI/v2/timeseries/data/

Series IDs constructed from SOC codes. Format: OEUM{area}{industry}{SOC}{datatype}. National level (area=0000000), all industries (industry=000000).

O*NET Web Services

services.onetcenter.org/ws/

RESTful JSON endpoints for occupation search, task listings, skill profiles, and technology cross-references. Updated quarterly with O*NET database releases.

X/Twitter API v2 — Recent Search

api.twitter.com/2/tweets/search/recent

Real-time social sentiment for live feed and report enrichment. Query-based search with engagement metrics.

Anthropic Claude API

api.anthropic.com/v1/messages

Claude 3.5 Haiku for report generation. Structured prompts with job data, BLS metrics, sentiment analysis, and user personalization. Generates executive summary, task analysis, career pivot plan, and skills roadmap.

See the Data in Action

Search any occupation and see how all 12 data sources combine into a single risk score — with a full dimensional breakdown, task analysis, and career pivot plan.

Check Your Risk Score →

Free score preview. Full report $29.99.