What if AI is both really good and not that disruptive?
The moderate case on AI
There’s a strange dynamic in AI discourse where you’re only allowed to hold one of two positions: either large language models will automate all knowledge work, collapse employment, and fundamentally restructure civilisation within a decade, or they’re stochastic parrots that can’t really do anything useful and the whole thing is a bubble. The measured take, that LLMs are a significant productivity tool comparable to previous technological shifts but not a rupture in the basic economic fabric, doesn’t generate much engagement. It’s boring.
I want to make the case for boring.
The Python Analogy
Consider how we talk about LLMs as a new abstraction layer for programming. You write intent in English, the model translates it to code, you debug at the level of English when things go wrong. This is framed as revolutionary, but there’s another way to see it: it’s the same transition we’ve made repeatedly throughout computing history. Assembly programmers became C programmers became Python programmers. Each transition increased individual productivity dramatically, yet programming employment grew rather than shrank. Higher abstraction didn’t mean fewer programmers; it meant more software got built, which resulted in more demand for people to build it.
If English-to-code is just another abstraction layer, maybe the equilibrium looks like “same number of software engineers, each individually more productive, much more total software in the world.” That’s a big deal, but it’s not mass unemployment. It’s not the end of software engineering as a career. It’s what happens every time we get better tools.
The counterargument is that previous transitions still required learning a formal language with precise syntax, whereas English is natural and everyone speaks it already. This should dramatically lower barriers to entry. Perhaps. Though I suspect the binding constraint was never syntax but the underlying skill of thinking precisely about systems, edge cases, state management, and failure modes. The compiler was pedagogical in that it forced you to confront ambiguity. If the LLM just does something plausible when your specification is vague, you may never develop that precision, and the systems you build may be worse for it. Whether that matters depends on the use case.
The Specification Problem
Here’s a more useful way to think about which work is vulnerable to automation. Consider knowledge work as sitting on a spectrum from well-specified to ambiguous.
At the well-specified end, you have tasks where the inputs are clean, the desired output is clear, and success criteria are obvious: processing a standard form, writing code to a precise spec, translating a document, summarising a report. LLMs are excellent at this, and there’s strong evidence they can automate significant portions of it.
At the ambiguous end, you have tasks where the context is messy, the right approach isn’t obvious, and success depends on knowledge that isn’t written down anywhere. “Put together something on the competitive landscape” where you need to know what the CEO actually cares about, what was discussed in a meeting last month, which competitors are politically sensitive to mention, what level of detail is appropriate for this audience. The specification itself is the work, and it requires drawing on institutional history, unstated preferences, and accumulated judgement that no system has access to.
We’re roughly three years into widespread LLM deployment, and employment in ambiguous knowledge work (strategy, research, engineering, complex negotiation, anything where the right answer isn’t obvious ex ante) hasn’t collapsed. What we’ve seen is productivity tools that let the same people do more, or need fewer junior people supporting them. That’s meaningful, but it’s not replacement.
The response is usually “give it time, organisations are slow to restructure.” Fine, but that’s unfalsifiable in the short term, which makes it a weak argument. At some point the restructuring either happens or it doesn’t, and so far it hasn’t.
The Internal Tension in the Pessimist Narrative
There’s a contradiction at the heart of the most common AI pessimist story that I don’t see addressed often enough.
The pessimist typically argues two things simultaneously: first, that AI will displace workers and suppress wages across large portions of the economy; second, that cost disease sectors like healthcare, education, and childcare will remain very expensive relative to incomes. These two claims are in tension. Healthcare, education, and childcare are expensive primarily because they’re labour-intensive. If labour is getting cheaper or displaced economy-wide, those sectors should eventually feel that impact too.
The most common escape route is to argue that AI specifically displaces knowledge work while physical presence and human trust requirements insulate care sectors. A copywriter loses their job to Claude, but a nurse or nursery worker remains essential because you can’t automate holding someone’s hand or watching a toddler. This might be true, but notice what it concedes: the disruption is contained to specific sectors rather than being the economy-wide transformation usually advertised. That’s a much more modest claim than “AI changes everything”.
The cleanest resolution is that the disruption is more contained than the hype suggests. We’re looking at a productivity tool that makes certain categories of work more efficient, shifts relative wages somewhat, and eventually equilibrates through normal labour market mechanisms, rather than a transformation that breaks existing economic relationships.
Labour Market Reallocation Actually Works
The historical track record on “this technology will end work” predictions is remarkably consistent: they’re always wrong. Agricultural employment went from 90% of the workforce to under 2% in developed economies. Manufacturing employment has declined dramatically since the 1950s. Each time, the prediction was mass unemployment, and each time, labour markets reallocated.
The frictions are real but finite. Retraining takes time. Credentialing is a barrier. A 45-year-old copywriter isn’t going to become a surgeon. Yet at the margin, some people retrain, and the next generation makes different career choices based on the changed landscape. Over a 10-15 year horizon, labour supply shifts toward the sectors that still need humans.
If LLMs displace some knowledge workers, the plausible story is that some of them move into care work, skilled trades, and other labour-intensive sectors that haven’t been automated. Supply in those sectors increases, wages there moderate, and prices eventually stabilise. The displaced workers find new roles, possibly at lower wages than their previous careers but not zero. This is more benign than either extreme narrative, neither “everyone loses their jobs and starves” nor “nothing changes,” just the normal process of technological adjustment that has happened repeatedly throughout economic history.
This story requires that there remain sectors where humans are needed. If AI eventually does everything including physical care and embodied presence, then reallocation has nowhere to go. Yet that’s a much stronger claim than “LLMs are good at generating text,” and we’re nowhere near demonstrating it.
Living Standards vs Income
There’s a confusion that runs through most AI economic doom scenarios: treating income as synonymous with living standards.
Income only matters insofar as it buys things. The relevant question isn’t “what’s on your paycheck” but “what can you actually do with your life?” These come apart when products improve dramatically or entirely new categories emerge.
The median person today has access to the sum of human knowledge through a device in their pocket, can video call anyone in the world for free, and has entertainment, navigation, and communication tools that didn’t exist at any price in 1995. Real wage statistics try to adjust for this through CPI calculations, but quality improvements and new product categories are notoriously difficult to capture. The smartphone in your pocket is not a more expensive Nokia; it’s a categorically different thing.
LLMs themselves are the latest example. A tool that can draft prose, explain concepts, write code, and answer questions on virtually any subject is now available for roughly zero marginal cost. What would this have been worth in 2015? It didn’t exist at any price.
The usual counterargument is that the things that really matter, housing, healthcare, education, childcare, have gotten more expensive relative to wages. This is true, but it circles back to the earlier tension: these sectors are expensive precisely because they’re labour-intensive. If AI really does suppress labour costs broadly, those sectors should eventually feel that pressure too. You can’t simultaneously have AI destroying jobs and human-intensive services remaining expensive forever. One of those has to give.
What Would Change My Mind
I’m not arguing that LLMs are unimportant or that nothing will change. LLMs are clearly a significant technology that will restructure how certain work gets done. I’m arguing that the most likely outcome is something like “computers” or “the internet” rather than “the end of employment as we know it.”
What would falsify this moderate view? A few things.
If ambiguous knowledge work employment actually starts shrinking meaningfully (not just growth slowing, but absolute declines in headcount for strategists, researchers, engineers, people doing work that requires contextual judgement), that would be strong evidence that this viewpoint is wrong. If agentic systems start successfully navigating the illegible organisational context that currently requires human judgement, things like understanding unstated preferences, political sensitivities, and implicit standards, that would be significant.
So far I don’t see compelling evidence of any of these. What I see is a very impressive technology that’s being discussed in maximally dramatic terms because moderate takes don’t go viral.
The boring middle ground is usually closer to the truth. That’s why it’s boring.


This resonated a lot, especially the framing around specification being the work rather than execution.
The Python analogy feels exactly right to me. Each abstraction jump made individual contributors more powerful, but also raised the premium on people who could think clearly about systems, constraints, and failure modes. Syntax was never the real bottleneck. Precision of thought was. If anything, higher abstraction tends to expose that gap more brutally.
I also appreciate the spectrum from well-specified to ambiguous work. That maps cleanly to what I see in practice: LLMs shine when the success criteria are already implicit or externally enforced, and struggle when the task requires knowing what should be asked in the first place. In those cases, the model can only mirror plausibility, not judgement.
The tension you point out in the pessimist narrative is one I rarely see addressed clearly. You can’t simultaneously argue for broad labour displacement and permanently expensive human-intensive services without assuming a very selective, almost magical, form of automation. Either labour becomes cheaper across the board, or the disruption is more contained. Both can’t hold indefinitely.
One thing I keep coming back to is your point about the compiler being pedagogical. If ambiguity is silently resolved by a model rather than surfaced as an error, we may end up with more output but weaker mental models underneath. That doesn’t matter for all use cases, but it matters a lot for the ones that fail expensively.
Overall, this feels like a strong case that the most likely outcome is not collapse or stasis, but redistribution of effort: fewer people doing pure execution, more value placed on framing, judgement, and responsibility for consequences. Which is less dramatic, less viral, and probably closer to how technological change usually unfolds.
Curious how others see this playing out in roles where the spec has traditionally been learned implicitly, rather than written down.
This is a good perspective to have. I think the levels of abstraction concept is interesting. Makes me wonder if that logic still holds as there are still many C programmers out there. I think there is always going to be a space for hand writing code. I don't know for sure, but I find it unlikely that a bunch of Cobol banking software is going to be vibe-coded any time soon.