C. Nagesh Bhushan
Artificial
intelligence and the future of work
AI is not coming for the factory floor. It is coming for the corner office — and the social contract will need to catch up
Hyderabad Apr 14th 2026
The most exposed
workers today are not on the factory floor — they are at the desk with a
graduate degree
Something strange is
happening to the conventional story about automation. For most of the past
century, the received wisdom ran in one direction: machines replace muscles,
not minds. The assembly-line worker was the perennial casualty; the knowledge
professional was the safe harbour. A university degree, the argument went, was
the best insurance policy against redundancy. That argument is now running in
reverse.
New data from
Anthropic's Economic Index, synthesising observed real-world usage of large
language models across professional settings, reveals that the occupations
facing the highest actual automation today are not the blue-collar roles that
feature in most political rhetoric about job loss. They are the digital-first,
credentialled, white-collar professions — the computer programmers, financial
analysts and customer service managers whose work consists largely of text,
data and structured reasoning. Workers with graduate degrees are now nearly
four times more likely to fall into the highest-exposure category than workers
without them. The shield has become the target.
AI task coverage,
selected occupations — 2026
The flywheel and
its discontents
OpenAI and
Anthropic, publishing new assessments in 2026, describe an "intelligence
flywheel" — a self-reinforcing cycle in which AI accelerates scientific
discovery, which in turn produces better AI. What once took months of
engineering effort can now be accomplished in hours by systems capable of
reasoning across entire research programmes. The transition from narrow tools
to something approaching general problem-solving capability, long forecast for
the 2030s, appears to have arrived rather earlier than expected.
There is, however, a
significant gap between what these systems can theoretically do and what they
are actually doing. Anthropic's researchers estimate that actual automated
coverage in most professions remains well below the technical ceiling — held
back by legal constraints (a licensed pharmacist's signature is still legally
required for a prescription), software interoperability failures, and the
irreducible need for human verification in high-stakes decisions. This gap is
not a sign of the technology's limitations. It is a narrow window — a grace
period during which policy has the opportunity to build safeguards before
theoretical potential becomes wholesale displacement.
"The gap
between what AI can do and what it is doing is not a sign of failure. It is a
narrow window for policy intervention."
The canary in the
coal mine
Headline
unemployment figures are a poor instrument for detecting the early stages of
this transition. Aggregate joblessness in exposed sectors has not yet spiked —
which is why most political attention remains elsewhere. But beneath the
surface, a more telling signal is already audible. The job-finding rate for
workers aged 22 to 25 in highly exposed fields has dropped by 14% since late
2022. The door for entry-level professionals is being quietly locked.
The mechanism is
straightforward. Companies are retaining their experienced senior staff — whose
institutional knowledge and client relationships remain difficult to replicate
— while automating the routine tasks that previously constituted a junior hire's
workload. The harm manifests not as redundancy but as the non-appearance of a
job that would otherwise have existed. A new graduate who finds no opening in
her chosen field does not show up in unemployment data. She shows up, if at
all, in the participation rate, the enrolment figures at graduate schools, or
the headcount of the care economy.
Pathways for
displaced young workers
Where workers aged
22–25 in AI-exposed sectors are going
Source: Anthropic
Economic Index; Current Population Survey, 2026
The efficiency
dividend
If AI systems are
generating unprecedented productivity gains, the question of who captures those
gains is not merely an economic one — it is a political choice. OpenAI's 2026
proposals introduce the concept of an "efficiency dividend": a mechanism
to convert productivity surpluses into time returned to workers rather than
value extracted solely by shareholders. The model being tested in pilot
programmes is a 32-hour, four-day workweek in which output is held constant
while eight hours are returned to the employee. The comparison being made —
advisedly — is to the New Deal's introduction of the five-day workweek in the
1930s, itself a political decision that reshaped the American social contract
for a generation.
The analogy is
instructive, though imperfect. The 1930s reform was driven by organised labour
bargaining with concentrated industrial employers. Today's AI productivity
gains are accruing in a much more fragmented landscape — distributed across
software platforms, gig arrangements and remote-first firms where traditional
union structures have little foothold. Capturing the efficiency dividend will
require new institutional mechanisms, not simply the extension of old ones.
"In an
age of superintelligence, human time is the ultimate luxury. The question is
who gets to keep it."
The public wealth
fund
The efficiency
dividend addresses time. The wealth question is harder. As the intelligence
flywheel generates value at scale, the risk of extreme concentration is real. A
small number of firms — OpenAI, Anthropic, Google DeepMind, a handful of others
— sit at the nexus of the most consequential economic infrastructure in human
history. The market capitalisation implications are staggering; the
distributional implications are more so.
The proposal that
has moved from academic fringe to mainstream policy debate is a Public Wealth
Fund: a publicly held equity stake in the AI economy, whose returns would be
distributed directly to all citizens. The Fund would function as a sovereign
wealth vehicle, investing in the companies driving automation and distributing
dividends broadly — ensuring that the gains of the intelligence age do not flow
exclusively to those who already hold capital. As OpenAI's own framing puts it,
if AI winds up controlled by and benefiting only a handful of actors while most
people lack access to AI-driven opportunity, the technology will have failed
its central promise.
Portable safety
nets
The Public Wealth
Fund provides the upside floor. Containing the downside requires a different
instrument. The labour market toward which AI is pushing us — fluid,
project-based, entrepreneurial — is structurally incompatible with a benefits
architecture built for mid-century employment norms. Healthcare, retirement
savings and retraining accounts that are tied to a specific employer are not
merely inconvenient in a world of frequent job transitions; they are a positive
barrier to the labour market flexibility that would allow workers to move
toward less-exposed roles.
The solution being
advanced — portable benefits that follow the individual across employers, gig
arrangements and entrepreneurial ventures — is conceptually straightforward but
politically demanding. Funding these portable safety nets will require a rebalancing
of the tax base: reducing the burden on labour income while increasing taxes on
capital gains and, potentially, on the automated labour that has displaced
human workers. Paired with wage-linked incentives that reward firms for
retaining and retraining staff rather than replacing them, this would
constitute the most significant revision of the employment social contract
since the post-war welfare state.
Observed AI
exposure by education level
Likelihood of high
AI exposure relative to unexposed workers
Source: Anthropic
Economic Index; Current Population Survey, 2026
A choice, not a
destiny
The tone of both the
OpenAI and Anthropic assessments is notably undefensive about the scale of what
is coming. The question these researchers are asking is not whether
superintelligence will reshape the economy — they regard that as settled — but
whether the democratic process will move quickly enough to shape the terms of
the reshaping. There is a meaningful difference between an economy that deploys
AI to generate broad prosperity and shared agency, and one that uses it
principally to compress costs and concentrate returns. Both are technically
achievable. Only one requires policy effort.
The Inverted
Revolution — automation targeting precisely those workers whom previous
generations of technology had elevated — is already under way. Its early stages
are visible in the hiring data for young graduates, in the career trajectories
of mid-career professionals, in the quiet displacement of tasks that no one has
yet publicly mourned. What is not yet determined is how the surplus it
generates will be distributed, who will bear the transition costs, and whether
the institutions built in earlier centuries are capable of adapting to a change
arriving at this speed.
None of those
questions answer themselves. They are, in the end, political. And the window
for answering them well is, by all the available evidence, shorter than most
politicians have yet appreciated.
This article
draws on published research from the Anthropic Economic Index (2026), OpenAI's
economic impact assessments (2026), Eloundou et al. (2023), and Bureau of Labor
Statistics occupational projections





Comments
Post a Comment