AI Could Automate 94% of Tasks, Uses Only 33%


TL;DR

  • Theory vs. Reality: Anthropic found AI could theoretically speed up 94 percent of computer tasks, but observed Claude usage covers only 33 percent.
  • Who Is Affected: AI exposure falls hardest on higher-paid, college-educated workers like programmers, inverting expectations about low-wage job displacement.
  • No Unemployment Spike: No significant unemployment increase has been detected among AI-exposed occupations since ChatGPT launched in late 2022.
  • Early Warning: Hiring for workers aged 22 to 25 in AI-exposed fields dropped 14 percent since 2024, a possible early indicator of structural change.

Anthropic this week released a study finding that AI bears down hardest on workers who are better educated, better paid, more often female, and more often white, not the low-wage service workers many expected. And even for those workers, the disruption hasn’t fully arrived.

The new labor market study finds that workers highly exposed to AI hold jobs paying 47 percent more on average than their unexposed counterparts, with a share of college graduates nearly four times higher in the exposed group. That demographic profile inverts the conventional narrative of AI threatening manual or service labor. Yet the study’s central finding offers a counterweight to alarm: a wide gap exists between what AI theoretically could automate and what it is doing in practice.

Measuring What AI Actually Does

To quantify that gap, Anthropic developed a metric it calls “observed exposure.” The approach combines three data sources: the US occupational database O*NET, theoretical exposure scores from the “GPTs are GPTs” paper by Eloundou et al. (2023, and real-world usage data from the Anthropic Economic Index, a dataset drawn from actual Claude conversations. The combination lets Anthropic compare what AI theoretically could do against what users are actually asking it to do.

Moreover, its scoring scale assigns tasks a value of 1 if a language model alone can complete them at a doubling of speed over unassisted work; 0.5 if additional tools are required; and 0 if there is no meaningful AI speed advantage. The methodology weights fully automated API use more heavily than human-assisted use, and work-related contexts more heavily than personal ones, designed to measure displacement risk rather than casual adoption.

In practice, the design choices produce a deliberately conservative measure. By requiring a doubling of speed and weighting work contexts over personal use, the metric functions as a floor estimate: actual AI influence on work is likely higher, which makes the theory-practice gap more striking when it still appears as wide as the data demonstrate.