Research Paper Interactive Dashboard Vol. 02 · AI Agents

AI Agent TAM Atlas
Software vs Labor Market
across 201 Tracks

Open Full Dashboard Download PDF
TL;DR

A living atlas of where AI agents win

A research framework + interactive dashboard mapping the AI Agent opportunity across 201 sub-tracks in 10 categories, comparing the Software TAM (what AI products charge for) against the Labor TAM (the human-labor spend AI agents can actually disrupt).

Software TAM
$497B
AI product layer · was $450B in 2025-Q1
Labor TAM
$9.6T
Addressable human-labor spend
Avg Labor/SW Ratio
19.4×
The leverage opportunity
Sub-tracks Covered
201
Across 10 categories
01 — The Core Insight

Why the labor market is the real ceiling

The Labor Market is the ceiling for AI Agents, not the Software Market.

Most AI market sizing focuses on software TAM — "how much SaaS revenue can AI capture?" — currently $497B globally for AI-agent-relevant products. That's the wrong ceiling. AI agents don't compete with SaaS; they compete with human labor.

The total addressable human-labor spend across the same 201 use cases is $9.6T — a 19.4× leverage ratio. This $9T gap is the structural opportunity AI agents are aimed at: turn human-labor spend into software spend (Net Compression), or expand the market by making expensive expert work cheap enough for new users (Net Expansion).

This is why we built the framework: to systematically rank where the leverage is largest, how much of each labor pool AI can actually replace, and which startups are credibly going after each pool.

02 — Why This Matters for Hat-Trick

The framework as an operating tool

Use caseHow the framework helps
Deal sourcingFilter by Labor/SW ratio + Net Effect to find structurally underserved tracks (e.g., 100×+ ratio = giant labor pool, tiny software footprint). Use as a hit list for outbound.
Thesis validationCross-check inbound startup pitches against the addressable labor TAM and displacement realism — catch overstated software-TAM claims.
Portfolio trackingMap each portfolio company to its track and watch ratio / displacement / key-player columns evolve over time.
Competitive intel"Representative companies" + "Funding" columns surface who else is in the track and at what valuation.
YC scoutingYC batch mapping (CORE / ADJACENT / INFRA fit) identifies which YC W26 / S26 companies are credible plays in tracks we already have a thesis on.
Capability watch"Capability" entries flag when a new model release (Claude 4.6, GPT-5.4, Helix 02 VLA) unlocks a step that was previously a gap — this is when entire tracks become investable.

Headline takeaway — out of 201 tracks, the framework identifies 22 high-priority deep-dive tracks (Labor/SW ≥ 50×) where the leverage is structurally largest. These are the first places we look when sourcing.

03 — How to Read the Dashboard

A guided tour of the live atlas

The dashboard below is the same interactive build available at agent-mapping.vercel.app — embedded here so you can read and explore in one place. Hover charts, click bubbles, expand table rows.

agent-mapping.vercel.app
Open ↗

Feature-by-feature walkthrough

"What's Changed" banner (top)

Collapsible panel showing the most recent bimonthly update grouped by type — TAM revisions, displacement updates, capability unlocks, and key players & funding. Click any track name → table auto-filters to that track.

KPI cards (6 metrics, recompute live with filters)

Filters

Category · AI Impact (Displacement / Expansion / Mixed) · Net Effect · Valuation Tier · Status · full-text Search.

The four charts

  1. TAM Totals by Category (log-scale horizontal bar) — gold bars (Labor TAM) tower over purple bars (Software TAM); the gap visualizes the leverage per category.
  2. Software × Labor Scatter (log-log bubble) — each bubble is one track; bubble size = displacement rate; click a bubble to drill into that row.
  3. AI Impact Donut — track count by Displacement / Expansion / Mixed, hover shows aggregate Labor TAM per slice.
  4. Valuation Tier Donut — distribution of lead-company valuation maturity per track.

The table — the core artifact

Every row = one of the 201 sub-tracks. Click any row to expand methodology, representative companies, funding, and investment insight. Columns:

ColumnMeaning
SW TAMAddressable AI-software market size for the track ($B), Tier-1 analyst + bottom-up
Labor TAMHuman labor spend addressable by AI agents ($B), BLS/ILO × wage × task-fraction
AI Impact🔴 Displacement / ✅ Expansion / 🟡 Mixed
Disp. Rate% of human work AI agents can autonomously do today (range, e.g. "55-75%")
Net EffectWhether AI grows or shrinks the total category
Labor/SWLabor TAM ÷ Software TAM = the leverage ratio
Valuation / StatusLead-company valuation tier and stage in the track
04 — Methodology

How each number is calculated

4.1 Software TAM

Definition: addressable revenue for the AI software product layer in the track over the next 3-5 years.

Method:

  1. Anchor to Tier-1 analyst reports — Gartner, Forrester, IDC, Grand View Research, MarketsandMarkets — for the broader market.
  2. Subtract non-AI-replaceable revenue (e.g., billing systems that aren't agent-driven).
  3. Cross-check with bottom-up: estimated user base × ARPU of leading product.
  4. Triangulate with company-revenue-multiple for category-defining companies.

Worked example — Code Generation Agent SW TAM ($28B)

Largest single TAM revision in 2026-04 (+133% from $12B), showing all three methods converging:

MethodCalculationOutput
① Analyst anchorMnM "AI in software development" 2026 projection$7B — too low; only counts tools, not agentic-coding spend
② Bottom-up ARRCursor $2B + Copilot ~$600M + Claude Code + Codex + Replit Agent + Cognition + niche tools~$4-5B current; 3× growth → 18-mo forward $15-20B
③ Multiples-impliedCursor $60B ÷ ~30× = ~$2B ARR; ~10-15% market share → category $15-20B today5-year addressable: $25-35B

Reading lesson: when one company's ARR exceeds 10-15% of your published category TAM, your TAM is wrong, not the company.

4.2 Labor TAM

Definition: total annual human-labor spend addressable by AI agents in this track, globally. Built bottom-up via the McKinsey/WEF task-fraction methodology.

Worked example — Tax Filing Assistant Agent Labor TAM ($128B)

  1. Identify occupations (BLS): Tax Preparers ~90K + Accountants ~1.4M (of whom ~30% do tax) = 420K tax-active accountants in the US.
  2. Wages: Tax Preparers ~$50K avg; tax-active Accountants ~$80K avg.
  3. US wage bill on tax work: 90K × $50K + 420K × $80K × 30% + seasonal ≈ $16-17B.
  4. Task automatability (McKinsey): tax prep scores 71% automatable. → US addressable $11.7B.
  5. Global multiplier: US ≈ 9-10% of global tax-prep → multiplier ~10-11×. → $128B globally.

Cross-validation: ILO global accounting workforce × 25% tax × $30K avg comp × 71% ≈ $75B (lower bound). McKinsey GenAI study: $120-180B annual global productivity unlock from AI in tax/audit by 2030. Our $128B sits at the lower end of McKinsey's range — conservative.

Reading lesson: Labor TAM is bigger than people intuit because it counts the full wage bill the AI agent could replace, not just incremental productivity gains.

4.3 Displacement Rate

The most carefully constructed number in the framework. We never take a single analyst's "AI will replace X% of accountants" claim — those ignore that a job is a bundle of dozens of steps. Instead we build it up step-by-step.

6-step methodology (per 50×+ deep-dive track)

  1. Workflow breakdown — decompose the track into 15-25 discrete steps grouped by phase. Example: Contract Review Agent breaks into ~22 steps across Intake / Substantive Review / Negotiation / Closing.
  2. Per-step capability assessment — classify each step:
    • FULL — AI handles 80%+ autonomously, output production-ready (80-95%)
    • PARTIAL — AI-assisted; human applies judgment for the last mile (30-70%)
    • NO — Human-only today (0-15%)
  3. Weight by hours, not step count — Replacement Rate = Σ(Step% × Hour-share). The high-volume boring steps (clause redlining, cross-references) are usually the most automatable, so weighted rate is much higher than naive count.
  4. Triangulate with three external sources — Goldman Sachs Occupational Automation Study, McKinsey Task Automation Fractions, WEF Future of Jobs. If our bottom-up rate sits inside the Goldman-McKinsey-WEF range, we have triangulation; outside → re-examine classification.
  5. Report as a range, use conservative midpoint — never a point estimate. Lower bound used for Net Effect modeling, because under-investing in over-claimed AI is the more expensive mistake.
  6. Update triggers — upward (new benchmark, model release, production data, YC clustering); downward (regulatory ruling, court decision, published failure case).

Worked example — Code Generation displacement (30-50% → 55-75% in 2026-04)

  1. JetBrains Jan 2026 survey: Copilot 29% / Cursor 18% / Claude Code 18% production adoption (not experimentation).
  2. Claude Opus 4.6 hit OSWorld 72.7%, GPT-5.4 hit GDPval 83% — agent benchmarks crossing thresholds where end-to-end coding tasks become FULL.
  3. Cursor's $2B ARR is empirical proof of value capture, validating the productivity gain.
  4. Triangulation: Goldman SDE-automation 47-62%; McKinsey high-adoption scenario 50-70%; our 55-75% sits in the upper half of both → revised up.

4.4 Labor/SW Ratio

Simple division. Categorized:

RatioInterpretation
> 100×Structural gap — massive unaddressed labor pool, tiny AI software footprint. Highest priority.
50-99×High-priority opportunity for compression or expansion.
20-49×Core opportunity — meaningful labor-to-software arbitrage.
< 20×Moderate — software has already penetrated substantially.

4.5 Net Effect — three canonical case studies

Decided by asking: if AI cuts this service's cost by 10×, what happens to total demand?

🔴 Net Compression — Contract Review Agent (Labor/SW 80×)

Corporate-legal contract review is a saturated B2B service. Companies don't draft more contracts because review got cheaper — they draft the contracts they already need. The buyer (general counsel) cares about output (signed deal), not input (associate hours). When Harvey reviews a contract in 3 minutes vs. an associate's 3 hours, the GC pays Harvey instead of the law firm.

✅ Net Expansion — CBT / Mental Health Agent (Labor/SW 437×)

There is a massive unmet demand pool in mental health. ~75% of people who need therapy don't get it — most because $200/session is unaffordable, not because they don't want it. WHO estimates ~1B people globally have a mental health condition, while only ~150M see a clinician. AI therapy at $20-30/month doesn't compete with the human therapist's existing clients — it serves the 850M who weren't being served at all.

🟡 Mixed — Advertising Agency Super-Agent (Labor/SW 62.5×)

Marketing & growth has both dynamics simultaneously, and which dominates depends on customer segment.

Pattern recognition rule: if you can't crisply describe the price-elasticity story for a track in two sentences, you don't know its Net Effect yet — you have a Mixed track and should treat it that way.

05 — Where to Focus

The watchlist

Tier 1 — 100×+ tracks (structural gaps)

Highest priority for sourcing. Both compression (legal cluster) and expansion (mental health) plays.

TrackRatioNet EffectWatch list
CBT / Mental Health437.5×✅ ExpansionWoebot, Wysa, FDA digital-therapeutics guidance
Recruitment Screening155× (was 233×)🔴 CompressionMercor $2B; YC W26 cluster (Perfectly, Skillsync, Vela)
Elder Care200×✅ ExpansionFigure AI $39B; Helix 02 VLA
Compliance Check106× (was 154×)🔴 CompressionOxus, Fenrock, Veriad (YC W26); EU AI Act
IP / Trademark / Patent110× (was 183×)🔴 CompressionArcline, LegalOS, Vector Legal (YC W26)
Tax Filing72× (was 120×)🔴 CompressionDualEntry $10.87B; Intelmarket; Big-4 AI billing
Contract Review80× (was 140×)🔴 CompressionHarvey $11B, Ironclad, Legora $5.55B
Case Research87.5× (was 130×)🔴 CompressionHarvey + Legora super-agent
Law Firm Super-Agent82× (was 100×)🔴 CompressionHarvey $11B; Thomson Reuters CoCounsel $200M+ ARR

Why ratios fell in legal cluster: the AI software TAM caught up fast (Harvey, Legora, super-agent platforms raised aggressively in 2026-Q1), so the leverage compressed. Still 80×+ — investable.

Tier 2 — 50-99× expansion tracks (largest TAM-growth potential)

Tier 3 — Emerging tracks not yet in framework

Likely additions next cycle (2026-06):

06 — Creation Journey

How this was built

Phase 0 — 2025-Q1: Baseline framework

Established the 201-track taxonomy across 10 categories. Built the Labor-vs-Software TAM lens. Cross-validated displacement rates against Goldman / McKinsey / WEF. Hand-curated representative companies and funding data per track.

Phase 1 — 2025-Q1: Deep-dive sheets

Built per-track 4-section workflow analyses (Workflow Breakdown → AI Capability Assessment → Critical Gap Analysis → Strategic Insight) for the 22 highest-priority tracks (Labor/SW ≥ 50×): 9 tracks at 100×+, 13 tracks at 50-99×.

Phase 2 — 2026-Q1: YC W26 mapping

When YC W26 batch was announced (Demo Day 2026-03-24, 199 companies, 74% AI, 64% B2B), we ran every company through the 22-track filter. Output: YC_W26_TAM_Mapping.xlsx — companies ranked CORE / ADJACENT / INFRA per track.

Phase 3 — 2026-04: First bimonthly update

Established the every-2-months update cadence covering 5 update types (new tracks, TAM revisions, displacement updates, capability unlocks, key players). The 2026-04 cycle produced 30 updates — most material being Code Generation Agent SW TAM $12B → $28B (+133%), legal cluster TAM +45-75%, Drug Discovery clinical milestone (Insilico Phase IIa), humanoid robotics breakthrough (Figure AI $39B + Helix 02 VLA), Q1 2026 macro: $300B record VC quarter, AI = 80%.

Phase 4 — 2026-04: Visualization site

Built agent-mapping.vercel.app to make the framework explorable. Self-contained static dashboard (Chart.js + vanilla JS), auto-rebuilds from the master xlsx via build_data.py. Added the "What's Changed" banner so the bimonthly update is the first thing visitors see.

Phase 5 — Ongoing: Bimonthly updates

Next cycle: 2026-06. Tracked: ratio threshold crossings, Net Effect reclassifications, new players in 50×+ tracks, capability unlocks from frontier-model releases, YC S26 batch mapping when announced.

07 — Limitations

What's not in the framework

These are deliberate scope choices. Opening any of them up is a 1-2 month follow-on project.

Appendix — Data Sources

What we cited

Tier 1 — primary, used directly

SourceWhat we use it for
Goldman Sachs Occupational Automation Study (2025 update)Per-occupation 0-100% automation risk score; one of three displacement-rate inputs
McKinsey Generative AI Economic Potential — Task Automation FractionsTask-fraction methodology for Labor TAM; second displacement-rate input
WEF Future of Jobs Report 2025Displacement probability per occupation; third displacement-rate input
BLS Occupational Employment StatisticsUS workforce size and wages by occupation
ILO World Employment & Social OutlookGlobal workforce extension of BLS
OECD Employment OutlookCross-country wage adjustments

Tier 2 — industry analysts (triangulated)

Funding & company data

Industry / professional bodies

Regulatory & policy

Q1 2026 specific (used in 2026-04 update)

Peer-reviewed