š¤ AI isnāt āhallucinating.ā We are.
Why 'hallucination' is a problematic term (hint - it's not just because it anthropomorphizes the technology!) and what to do about it.
Hi there, and welcome back to Untangled, a newsletter and podcast about technology, people, and power. May has been a fun month:
In the latest special issue of Untangled, I asked the uncomfortable question: uh, what even is ātechnologyā? Then I spent 3,000 words trying to answer it. š
I offered a behind-the-scenes look at Untangled HQ. I shared what Iāve learned, how Untangled has grown, and where I hope it will go in the future. Plus I extended a personal offer, from me to you.
For paid subscribers, I posed the question, āWhat even is āinformationā?ā and dug into the problems of metaphors. Then I analyzed an alternative metaphor for AI offered by Ted Chiang in The New Yorker: management consulting firms.
This month I decided to answer the important question: is AI āhallucinatingā or are we? If you enjoy the essay or found the pieces above valuable, sign up for the paid edition. Your contributions make a real difference!
Now, on to the show.
If youāve read an article about ChatGPT of late, you might have noticed something odd: the word āhallucinateā is everywhere. The origin of the word is (h)allucinari, to wander in mind, and Dictionary.com defines it this way: āa sensory experience of something that does not exist outside the mind.ā Now, ChatGPT doesnāt have a mind, so to say it āhallucinatesā is anthropomorphizing the technology, which as Iāve written before, is a big problem.
āHallucinateā is the wrong word for another important reason: it implies an aberration; a mistake of some kind, as if it isnāt supposed to make things up. But thatās actually exactly what generative models do ā given a bunch of words, the model probabilistically makes up the next word in that sequence. Presuming that AI models are making a mistake when theyāre actually doing what theyāre supposed to do has profound implications for how we think about accountability for harm in this context. Letās dig in.
In March, tech journalist Casey Newton asked Googleās Bard to give him some fun facts about the gay rights movement. Bard responded in part by saying that the first openly gay person elected to the presidency of the United States was Pete Buttigieg in 2020. Congratulations, Pete! This response was referred to by many as a āhallucinationā ā as if the response wasnāt justified by its training data. But since Bard was largely trained on data from the internet, it likely includes a lot of sequences where the words āgay,ā āpresident,ā āUnited States,ā ā2020,ā and āPete Buttigiegā are close to one another. So on some level, claiming that Buttigieg was the first openly gay president isnāt all that surprising ā it is a plausible response from a probabilistic model.
Now, this example didnāt lead to real-world harm, but who or what should be held accountable when it does? Helen Nissenbaum, a professor of information sciences, explains that weāre quick to āblame the computerā because we anthropomorphize it in ways we wouldnāt with other inanimate objects. Nissenbaum was writing in 1995 about clunky computers, and this problem has become much much worse in the intervening years. As Nissenbaum wrote then, āHere, the computer serves as a stopgap for something elusive, the one who is, or should be, accountable.ā Today, the notion that AI is hallucinating serves as such a stopgap.
Paradoxically, users or operators of the technology often absorb a disproportionate amount of blame. This is what Madeleine Clare Elish, a cultural anthropologist, calls a āmoral crumple zoneā wherein āresponsibility for an action may be misattributed to a human actor who had limited control over the behavior of an automated or autonomous system.ā Traditionally, a ācrumple zoneā is the part of the car designed to absorb the brunt of a crash in order to protect the driver. Elish argues that historically āa moral crumple zoneā has protected the technological system at the expense of the human user or operator. Remember when New York Times journalist Kevin Roose went viral for a very weird and unsettling back-and-forth with Sydney, Bingās new chatbot? The back-and-forth ended with the chatbot proclaiming its love for Roose. In the aftermath, many commentators argued that Roose pushed Sydney too far; that he was to blame for how the chatbot responded.
The group often conspicuously left out of this discussion of potential blame are those building and making key decisions about the model. Engineers, AI researchers, developers, corporate officers, etc. have historically avoided blame for a few reasons. The first reason is the weird idea that weāve come to accept errors or bugs in code as normal. Nissenbaum, calls this āthe problem of bugsā and shows how it leads to an obvious problem: if imperfections are perceived as inevitable, then we canāt hold those designing the system accountable. But that doesnāt hold up ā in a number of industries like car manufacturing and planes, where the cost of an error are very high, weāve proven this idea mostly wrong.
Then there is āthe problem of many handsā that Nissenbaum describes, which is the notion that in modern organizational arrangements, rarely does blame for a decision lie with one person. There are lots of cooks in the proverbial decision-making kitchen, each with varying degrees of authority and power. Also, the ākitchenā has become more complex since Nissenbaumās writing, as machine learning processes introduce more dynamic steps and different stakeholders. In any case, this effect is compounded by the assumption that engineers canāt actually explain whatās happening inside the model. If they canāt explain why specific inputs combined with model decisions contribute to certain outputs, then how can we blame them? Thatās not quite the right question, though. One, itās kind of insane not to assign blame because they donāt understand whatās happening ā if anything, that sounds like all the more reason to dole out a liāl accountability, or at least preemptive thresholds. I argued for this in āA critique of tech-criticism,ā writing:
āThe government has a long history of requiring companies and industries to meet a certain standard before launching a product. I wouldnāt drive a car if federal standards didnāt prevent serious injuries. Nor would I hop on a plane so often if we didnāt render crashes nearly obsolete [ā¦] So what to do? Well, the government could require that AI companies be able to explain how their model produced a result before releasing it. Itās not clear that interpretability is possible, but right now, weāre not even asking that companies try.ā
Furthermore, ācauseā is an especially high bar for blame. Joel Feinberg, a noted moral, social, and political philosopher described a set of conditions under which one would be considered āmorally blameworthyā for a given harm even if they didnāt intend to cause the harm. Nissenbaum summarizes Feinbergās clauses this way:
āWe judge an action reckless if a person engages in it even though he foresees harm as its likely consequence but does nothing to prevent it; we judge it negligent, if he carelessly does not consider probable harmful consequences.ā
In other words, an engineer might deserve blame and accountability ā even if they didnāt mean to cause harm ā if they were reckless or negligent in how they built the technology. For example, researchers have been documenting the harms of AI and LLMs for years. Itās simply no longer reasonable to say āah, we didnāt see that harm coming, the machine must have āhallucinated.āā That sounds pretty negligent to me. Moreover, in a totally insane 2022 survey, AI researchers were asked the question, āWhat probability do you put on human inability to control future advanced A.I. systems causing human extinction or similarly permanent and severe disempowerment of the human species?ā The median reply was 10 percent. I personally find the question itself hyperbolic but yeah, itās fair to say that AI researchers āforesee harm.ā So the question of recklessness comes down to what theyāre doing to proactively prevent it.
Finally, and most importantly, the pursuit of explainability looks to the model itself for answers, when the model is entangled with both its users and creators. Purely from a technical perspective, we can only half explain why a modelās outputs are the way they are. As we saw with Bardās alleged first gay president, probabilistic models will result in weird outputs that donāt technically exist in the training data, because they are just making up sentences based on what words are likely to follow the ones that came before it.
The other half of the explanation exists in how people, culture, norms, race, and gender, inform how the training data is constructed, and then more obviously, how the prompt is created. In some small way, Roose was to blame for Sydneyās response ā his prompts were inputs to what the chatbot generated. So weāre left with a complex system with multiple inputs ā engineers, users, and the technology ā dynamically interacting, each adapting, updating, and changing their actions, decisions, and outputs in response to one another.
What can be done about all this?
We need to accept that the complexity of the engineer-user-tech interaction does not absolve everyone from responsibility for error and harm. As Nissenbaum writes, āInstead of identifying a single individual whose faulty actions caused the injuries, we find we must systematically unravel a messy web of interrelated causes and decisions.ā Right, we can start to disentangle these systems and assign partial blame and accountability to the appropriate stakeholder. This will require a lot (!) more information about the models themselves and the decisions engineers are making. Iāve long had a complicated relationship with ātransparency initiativesā but much more of it will be required if weāre to move closer to accountability. It will also require banishing from our brains the cultural assumption that bugs are inevitable. But the first step along this path is to close the gap between our own expectations of what LLMs can do, and what theyāre actually doing. They arenāt hallucinating ā we are.
As always, thank you toĀ Georgia Iacovou, my editor.
I starte3d not to comment because my expertise is not relevant. But then I realized, no ones expertise is relevant. I have been programming in 13 languages since 1960 but retired in 2000. So my knowledge base stopped in 2000....similar to what ChatGBT says when reminding us that its knowledge stops at 2021. My first response to its making stuff up was when I asked if I could identify poisonous mushrooms by their "white gills" and was told that all mushrooms with white gills were edible, a complete reversal of truth and a dangerous answer if someone took that advice to the field.
My first response was that you need to have an understanding of your subject before you rely on Ai's answers. Then my next response was, 'They cannot drive the human race to extinction because we can un plug them."
Thinking about that I realized that if the commanders of the DEW Line in its early days has relied on Ai, there is a good chance that Russia would not exist today except as a Radioactive no mans land. In those early days there was an incident where the rising moon told the DEW line RADAR operators that a huge swarm of missiles was headed toward the US. Luckily the cooler heads of the HUMAN commanders determined that Russia didnt have that many rockets and in the nick of time, shut down the alert before the button was pushed. (In those days we didnt know that they were totally incapable of a response and believed them to be a threat which was THEIR fault because of their blustering, so we would have destroyed the entire country because of faulty and nonexistent intelligence. that generated an unreasonable fear on our part) OK so that is not total extinction but a similar scenario could cause total extinction.
So unplugging Ai is not the solution. Perhaps human supervision, but then who selects the supervisors? If almost any government is in charge of selecting the supervisors, and I include the UN, there will be biases.
So maybe we should just stamp out all Ai like cockroaches???
Thank you for the article Charley as the issue of transparency in AI and "ignorant bliss" amongst its users is more relevant than ever. You noted complexity in the article and it made me think of its derivative within systems theory, where all parts of the system are interrelated and the actions of one affect several others. This is a very complex problem with many unknown - unknowns. I would offer that the algorithms driving AI/ML are one of the most underrated issues of the time. Thanks again for the wonderful read.