AM I COOKED?
AI Job Displacement Audit
← Back

Methodology

Am I Cooked? is a research-backed tool that shows workers the real economic math behind AI job displacement. It combines peer-reviewed exposure scores, real-world AI usage data from Anthropic, BLS wage statistics, and O*NET task definitions to give you an honest, data-driven picture of where you stand. Every number is sourced, every formula is transparent, and every limitation is stated. The goal is not to alarm — it’s to inform.

⚠ How to interpret the numbers

The AI exposure percentage presented here should not be interpreted as equivalent to the proportion of money saved per employee or the actual financial impact on a company. Rather, this metric is intended to provide a perspective on the potential influence of AI within an organization. By quantifying exposure in relative terms, the aim is to highlight the areas where AI may play a significant role — making the implications more tangible and relatable to stakeholders, without implying any direct monetary consequence.

Audit Pipeline

When you enter a job title, salary, and industry, the following pipeline executes — each step either looks up empirical data, runs deterministic math, or generates prose via an LLM.

01
Job Title
User input
02
O*NET Match
Local CSV search
word-overlap + stemming
03
β Lookup
Eloundou CSV
SOC code → β score
04
Observed
Anthropic CSV
real-world AI usage
05
BLS Wages
Median, mean, %iles
employment counts
06
Task Match
Confidence-aware
fuzzy matching
07
Gemini Grade
Unmatched tasks
batch AI grading
08
Economics
Augmentation vs
automation costs
09
Timeline
Derived formula
β + observed
10
Narrative
Gemini (prose)
cached per prompt
Empirical data lookup
Deterministic math
LLM generation (cached)
System routing

Green steps use published, peer-reviewed data. Amber steps are deterministic formulas — same inputs always produce the same output. Red steps call Gemini, but results are cached — identical prompts return instantly without an API call.

Exposure Score (β)

The core metric comes from Eloundou, Manning, Mishkin & Rock (2023), published in Science. They classified every O*NET task into three exposure levels:

Exposure Rubric

The β Formula

β = E1 + 0.5 × E2 Ntasks where E1 = directly exposed tasks, E2 = tool-exposed tasks (weighted 0.5), Ntasks = total tasks for occupation

The 0.5 weight on E2 reflects that tool-mediated exposure requires additional investment to realize. β ranges from 0 (no exposure) to 1 (fully exposed). The US occupation average is approximately 0.30.

Key finding: approximately 80% of US workers belong to an occupation with at least 10% of tasks exposed, and 19% have occupations where more than half of tasks are exposed.

Observed Exposure

From Massenkoff & McCrory (2026), Anthropic. They measure what fraction of an occupation’s tasks are actually being performed by AI, not just theoretically possible.

Observed Exposure Formula

t = 𝟙{WorkUsaget ≥ 100} × 𝟙t ≥ 0.5} × αt where αt upweights automated (vs. augmentative) tasks. Full automation → α = 1, pure augmentation → α = 0.5
Roccupation = Σ(wt × t) Σ(wt) where wt = fraction of working time spent on task t

Key finding: AI is far from reaching its theoretical capability. Computer & Math occupations have 94% theoretical exposure but only 33% observed adoption.

Task Penetration Matching

O*NET defines ~19,000 tasks. The Anthropic Economic Index measures AI usage per task. O*NET and Anthropic describe the same task differently, so we use confidence-aware multi-stage matching:

  1. Exact text match → confidence: high
  2. Normalized match (lowercase, strip punctuation) → confidence: high
  3. Keyword Jaccard ≥ 0.35 (significant words overlap) → confidence: medium/high
  4. SequenceMatcher ≥ 0.60 (character-level similarity) → confidence: medium/high
  5. ≥ 4 shared significant keywords → confidence: medium

Tasks with no confident match (< medium confidence) are sent to Gemini for AI penetration estimation on a 0.00–0.90 scale. These appear in the report with a ✦ Gemini badge. Results are cached per-task and per-prompt to minimize API calls.

Economic Model

Every cost is labeled: EMP (empirical), EST (estimated with stated basis), or DER (derived from inputs).

Salary at Risk

Srisk = salary × 1.35 × β 1.35 = employer overhead (benefits, payroll taxes, equipment). This represents the economic value of tasks AI could theoretically perform.

Path A — Augmentation

Caugmentation = Csubscriptions ~$720/yr baseline (1 AI assistant + 1 coding copilot + misc tools). Compute is bundled in subscription pricing.

Path B — Automation

Cautomation = Cplatform + Cinference + Cverification Platform (~$12K/yr EST) + API inference (tokens × β × model rate, DER) + human review (20% of automated task value, EST)

Employer Savings

Savings = Srisk Cautomation Range: [savings × 0.5, savings × 1.3]. Assumes proportional headcount reduction — overstates near-term impact.

Users can switch the AI inference model in the report to see how compute costs change across providers (Claude, GPT, Gemini, Grok, DeepSeek, and others).

Timeline Derivation

The automation timeline is calculated from data, not guessed by an LLM.

eff = max(β, observed)
base = 2 + 11 × (1 eff)
ratio = min(observed / β, 2.0)
accel = max(1 ratio × 0.2, 0.4)
years = clamp(round(base × accel), 1, 18) eff=1.0 → 2 years, eff=0.1 → 13 years. Adoption acceleration reduces timeline when observed >> β.

We use max(β, observed) because Eloundou’s β scores are from early 2023 — before widespread tool-use, code agents, and multimodal capabilities. For some occupations, real-world adoption has already exceeded the theoretical prediction.

Gemini Integration

All Gemini API calls go through a unified client with two layers of caching:

  1. Prompt-level cache: Every call is hashed (system + prompt + temperature). Identical prompts return cached results instantly with zero API calls. Persists to disk between restarts.
  2. Task-level cache: Individual task grading scores are stored separately. A task graded in one audit is never re-graded in another, even if the batch prompt differs.

Gemini is used for: task penetration grading (when task_penetration dataset has no confident match) and personalized narrative generation. It is not used for: β scores, costs, savings, timeline, or any quantitative output.

Datasets

DatasetRecordsSourceUsed For
Occupation Level798Eloundou et al. (2023)β exposure scores
Job Exposure Dataset756Massenkoff & McCrory (2026)Observed AI exposure
Task Penetration Dataset17992Anthropic Economic IndexAI usage per task
BLS Wages Dataset1358BLS OES May 2024Wages, employment
Layoffs Dataset2020–2026Kaggle/swaptrIndustry layoffs
Occupations Data~1,016O*NET CenterJob title search
Occupations Tasks Dataset~19,000O*NET CenterTask definitions
LLM Tier & Tokens Pricing4 providersPricetokenSubscription costs

Live API: O*NET Web Services v2 — job outlook and career detail only (changes with BLS projection updates).

Key Research Findings

Eloundou et al. (2023)

Massenkoff & McCrory (2026)

Limitations