Twenty years ago, I was a computer science graduate student, building predictive models for natural language. I dabbled with the neural networks that underpin today’s GenAI agents, but at the time they were neither fashionable nor terribly powerful, so my models were built with supervised discriminative machine learning techniques. I ended up with two peer-reviewed publications and a profound sense of ennui from doing work that didn’t feel like it had much practical relevance.
So, you can view this month’s essay as coming from someone with higher-than-average understanding of the field, or you can view it as coming from someone so short-sighted and foolish that she didn’t feel like AI and natural language had enough potential. Both are probably true.
As with all my writing, this post was generated without any assistance from GenAI, and indeed started as old-fashioned, handwritten notes. Because of the subject matter, I took special care to keep track of all the links I read while working on this newsletter. Bye and bye, I will feed them into a GenAI and see what kind of essay comes out. Perhaps it will undo all my convictions and I will wish I had skipped the struggle of writing altogether.
Where I did have help was from the Workomics team. They have a singular ability to question the status quo, surface new perspectives, and insist on better. They pushed me to think more deeply on this topic, and I’m grateful for it.
In comradeship,
S.
GenAI, capitalism, and human flourishing
It’s hard to write about Generative AI and Large Language Models without writing about capitalism. That’s the point of GenAI, isn’t it? To further the capitalist ideal of maximizing profit?
Certainly, that is what the LLM purveyors are selling us: lower your costs while creating many more outputs many times faster than mere human beings. The output is not especially good, but there is so much of it and it is so instant! Who cares about excellence when you can have seemingly infinite volume with fewer of those expensive human beings?
This is the moment where capitalism has lost the plot.
Market capitalism is a system that incentivizes people to maximize efficiency in pursuit of material wealth. It has been very effective in that aim. Economist Brad Delong estimates that an unskilled laborer in 1870 London would have produced roughly the equivalent of 5,000 calories in economic value each day. In other words, not enough to feed his family. Today, an unskilled worker in London produces the equivalent of 2.4 million calories a day — enough to feed her own family and several hundred others besides.
That tremendous leap forward in material well-being is in large measure the result of market capitalism and how it incentivizes corporations and individuals to apply technological progress in pursuit of profit. To be sure, we struggle mightily to distribute the resulting wealth as equitably as we should, but after many thousands of years of humanity, we no longer struggle with not having enough. Capitalism must surely be credited with an assist.
Today, at the societal level, we don’t need more of anything. Indeed, most of us struggle under the weight of too much. Too much carbon. Too much plastic. Too many ultra-processed calories. Too much endless, unfulfilling content in endless, unfulfilling social media feeds. (Capitalism gets an assist here too.)
Nevertheless, here we are, still blithely organizing our society around the pursuit of efficiency and material wealth. Except now we stand on the cusp of a new paradigm that will allow us to create nearly infinitely, with ever less human intervention.
What is a human being for?
In capitalism, corporations are pure vehicles for maximizing profit: more output for less input, without limit. The economists like to pretend that human beings are built along similar lines, but it’s more complicated than that. Yes, we often eschew effort when we can. But left to our own devices, people are forever volunteering for unnecessarily-challenging undertakings: running marathons, baking bread, writing email newsletters, climbing mountains, raising children, playing Bracket City on hard mode.
Psychology professor Michael Inzlicht has coined the term “effort paradox” to capture this notion that effort is sometimes avoided and other times valued. He posits that effort is a mechanism for creating meaning and purpose. Only by expending effort can you feel a sense of mastery. Too, it is a form of social currency — a way to demonstrate your commitment to a group or a cause. Effort is also valuable as an antidote to boredom. People don’t actually like to do nothing for very long, and will seek out ways to spend effort once boredom sets in.
GenAI and LLMs undercut all of these positive attributes of effort. Knowledge workers who are required to integrate AI into their workflows bemoan the loss of meaning and purpose in their work. Human-human relationships are displaced by human-AI transactions. People are consigned to babysit the AI rather than drive the work themselves, representing a loss of agency and autonomy in favour of reaction and monotony.
From a capitalist perspective, none of that matters. The system is working as intended: seek profit, minimize the effort and cost, let the invisible hand take care of the rest. If that means people no longer write their own words, it is no different than when people stopped being blacksmiths. Mustn’t stand in the way of progress.
The struggle is real; the AI is not
Except that language (and creating more generally) is not just some mechanism for generating revenue. It is fundamental to what it means to be human. Writing is not easy, but it is how we get clarity of thought, how we invent new ideas, how we disseminate them to others who might use their own words to build on our ideas and make them better.
If the LLMs do all the writing, what do we lose?
The written word on the page has value as output1. But the messy, frustrating, difficult process of writing is not a problem to be solved. The inefficiency is a feature, not a bug. Setting aside questions of quality, relying on the output of an LLM robs you of the benefits that come through expense of your own effort. Your understanding will be surface-level. Any objections you run into will render you inert — unable to counter because you haven’t done the heavy lifting yourself. You won’t be motivated to see your ideas through because they are not really yours. The output will be forgotten and discarded as easily as it was acquired.
Language, writing, and creating are also mechanisms for discovery and decision-making. When you set out to write about, say, AI and capitalism, the ideas do not emerge fully-formed like Athena from the head of Zeus. Rather, the ideas take shape through thousands of small decisions about words and structure, each one clarifying, refining, defining. When you let GenAI do the work, you delegate most of those decisions2.
Delegation of decision-making is a radical departure from all the automation that has come before. Whatever technical wizardry we deployed, the decision-making still resided with the human beings who designed, built, and ran the machinery. Many decisions were taken away and hidden from front-line workers, but some human somewhere still made the decision before it was programmed into the machine. The GenAI models are different, because they are making all of those small decisions on our behalf, without us ever knowing what was considered and ultimately discarded.
We lose something real when that happens. Recent research surveyed workers who use GenAI at work at least once a week. The researchers found:
“While GenAI can improve worker efficiency, it can inhibit critical engagement with work and can potentially lead to long-term overreliance on the tool and diminished skill for independent problem-solving. ”
The kinds of work that workers did also shifted, with seeking, generating, and doing, displaced by verifying, combining, and supervising. The issue is that the cognitive skills are not switches you can turn on and off at will. We see that when people rely on GPS directions rather than building their own mental maps, the region of the brain associated with mapping literally shrinks. Cognitive skills that aren’t practiced and maintained will atrophy.
We need a certain amount of struggle — grappling with ideas, making trade-offs, doing hard things. It’s how we maintain and expand our capacity to think, and also how we figure out what we think. Written outputs are valuable and useful, but the process of creating them is what is essential to us as humans. We need effortful pursuits to flourish.
Human flourishing is the point
When capitalism was first emerging in the 18th century, there wasn’t enough food or shelter to go around. In that context, it seems preferable to have productivity gains from division of labour, even though it replaced fulfilling work for craftspeople with repetitive drudgery on an assembly line. “Not starving” is pretty foundational to human flourishing.
The world we live in now is very different. Mostly, we are applying GenAI to create “more” where we already have too much. If we prize efficiency and profit above all else, the LLMs seem inevitable. But the capitalist paradigm of efficiency and profit sprung from the trade-offs of a bygone era. The world today is already abundant in material wealth, and approximately no one thinks AI will help distribute it more equitably. LLMs maximize efficiency at a ludicrous scale, but in exchange, we forsake fundamentally human things like discovery, learning, autonomy, relationships, purpose, fulfillment.
We should be very sure the payoff is worth it.
Assessing the payoff will require corporations to be something other than profit-maximizing machines in the capitalist ideal — to grapple with their own version of the effort paradox. Sometimes, effort should be minimized; other times, not.
To wield AI effectively, organizations will need to have a real clarity of purpose — a reason to exist that goes beyond producing a particular output as profitably as possible. They will need to think deeply about when and how AI helps or hinders. Is near-term efficiency worth it, if the result is a long-term decline in the organization’s capacity to create, innovate, differentiate? If there is no one who can supervise the LLMs in 10 years time, because the AIs took the on-the-job learning opportunities away from junior employees? If the LLMs are commoditized and all your competitors can produce the same blandly average offerings?
Ultimately, the decision to leverage GenAI is not unlike the decision to outsource. There’s nothing wrong with it, when it’s a supporting role, in the areas where “adequate” is all you need. But if you outsource the core of your offerings, sooner or later your customers will realize you are just a middleman and cut you out.
Equally, there are some problems out there where we really should privilege efficiency above all else. If you’re using AI to develop new cancer drugs, go with God. I hope market capitalism makes you really rich because you invented new treatments and reduced human misery.
But there are all kinds of other problems where prioritizing efficiency and applying AI is to deeply misunderstand what it is to be human, what makes for a happy life. If you are using AI to call your parents for you, or write your guide to summer, or make your video game, or teach your course, then you have been led astray. When an AI makes these things, the best-case is a mid output that adds to the noise. It’s pointless. Even if there is—improbably—financial profit to be had, you have inflicted a net cost on society and robbed yourself of something meaningful.
But when a real live person makes these things, that’s a very different premise. Human beings are in the arena, connecting with each other, expanding their understanding of the world, making compromises, striving for better. It will be inefficient and effortful. The result might be mediocre or it might be profound. Either way, humans and humanity will have benefitted from the pursuit.
Workomics Staff Picks
Some links from the team this month:
We thought this article on VR in patient experience was super cool: Research recently published in the Clinical Journal of Oncology Nursing revealed that VR tours of Venice, Agra, and the Ecuadorian Amazon can positively impact patients' side effects during longer infusion sessions (e.g., chemotherapy and other cancer treatments).
This article explores how focused attention is one of our most important tools, and that it requires us to have actual systems for preserving it, so that we can do our best work, and so that we can help preserve our teammates’ focus as well. Distractions will happen; we can be prepared with systems for how to deal with them, and certain cases, even welcome them.
Our own Michelle Thompson-McCune was one of four featured voices on Tom Lietz's Storylinking podcast, talking about ways to build engagement with groups.
Anne Helen Petersen digs into the question of whether people are bad at their jobs, or if the jobs are just bad. As work gets faster and more demanding, burnout, poor training, and unclear expectations are making it harder to do good work. Maybe it’s time to stop blaming individuals and start redesigning the way work actually works.
At least, I hope so.
Ted Chiang explores this idea in his New Yorker article “Why AI isn’t going to make art”
Such an excellent and thought provoking article. Thank you Susan.