đ¤ Will ChatGPT replace search?
How the limits of chatbots can point us toward a better search system.
Hi there, and welcome back to Untangled, a newsletter and podcast about technology, people, and power. This is the free issue of Untangled, which arrives in your inbox once a month. If you want more sweet, sweet content that helps you â as product managers, technologists, grant-makers, civil society leaders, and concerned citizens alike â analyze the big sociotechnical problems of the day, make better decisions at work, and take daily actions that align with your values, sign up for the paid edition.
Before we get into it, two pieces of exciting news:
Nick Martin, CEO of TechChange, named Untangled his favorite newsletter.
Untangled just eclipsed 2.2K subscribers. Thatâs a lot more than the 200 subscribers I had just one year ago. Thatâs really exciting - thanks for being part of this!
Now, on to the show.
Iâm a real curious person. I enjoy seeking out information, analyzing it, and synthesizing it. Search feels like a fundamental human pursuit to me. Naturally, then, I donât get the appeal of using ChatGPT as a search system. So when Microsoft recently announced that they are working to integrate ChatGPT into its Bing search engine, a lilâ part of me died inside. I kid, but the point is, I donât want a direct answer to my query, I want to search for it.
But also, I get the desire for an alternative to Google. Itâs a private company that monetizes our clicks without our consent and has a market cap that is 60% of the total GDP of all of Sub-Saharan Africa. Googleâs search system is dedicated to its own financial interests, and the experience of using it is getting worse and worse in part because of its business model. Not great. What even constitutes a good search system? In this essay, I start to answer this question by evaluating the limitations of ChatGPT and Google. Letâs dig in.
There was a time when âSearchâ meant having a conversation with a librarian, which veered in various directions and then, eventually, pointed you to a book or two. With the rise of the web, âsearchâ morphed into a process whereby we input a query and get myriad web pages to choose from. Social media has led to more social searching: sites like Quora and Reddit turn the query and response process into a community Q&A of sorts.
The point is, how we understand search is intertwined â some might say âentangledâ â with the technologies of the day and the social behaviors they encourage. One big issue with chatbots like ChatGPT is that it encourages the acceptance of a direct answer. Whereas Google might offer a mix of high-quality and clickbaity information sources from which to choose, ChatGPT generates a single plausible response with an air of authority. As I wrote before:
âWhatâs striking to me about all of the ChatGPT3 examples going around the internet isnât that many are wrong, but how confidently wrong they are. The answers sound authoritative, even if the output is only ever probabilistic. In short, with ChatGPT3, entertainment and confidence substitute for truth and meaning. If thatâs not Trumpy, I donât know what is.â
Indeed, the authoritativeness endowed in a direct answer poses an unavoidable dilemma. In âThe Dilemma of the Direct Answerâ Martin Potthast, Matthias Hagen, and Benno Stein explain that âThe dilemma of the direct answer is a userâs choice between convenience and diligence when using an information retrieval system.â I donât know about you, but Iâm concerned that âmore convenienceâ wins in a fight against âmore researchâ most of the time.
This is a problem because the answers to most questions and queries arenât obvious. If you were talking to a librarian, you might go back and forth. As they probe, you might have an âuh, I donât knowâ moment or think âit totally dependsâ or get a lilâ testy and say âhold up, I disagree with the presupposition of that question.â It takes work and iteration to cut through the ambiguity implicit in a given query. But chatbots just remove that uncertainty and equivocation altogether, leaving no trace that it was ever there in the first place. As the authors put it, âWe donât know what a good answer is because the world is complex, but we stop thinking that when we see these direct answers.â
This directness is also at odds with flexibility and exploration. Search systems need to support a diverse range of search activities because weâre complex lilâ humans with lots of different needs! In âSituating Search,â University of Washington professors Emily Bender and Chirag Shah put it this way,
âInformation sources as well as peopleâs information seeking behavior have become more diverse, which in turn increases the need for flexible tools that can support diverse modes of usage.â
Gary Marchionini, professor at the University of North Carolina at Chapel Hill offers a nice framework to simplify what amount to twenty different âinformation searching strategies:â
Lookup: search as a fact-finding mission. You know what youâre looking for, and your query yields precise results.
Learn: search as knowledge acquisition. Your search involves multiple iterations, followed by interpretation and analysis. Youâre trying to learn or interpret something!
Investigate: search as analysis and evaluation. Youâre searching with a specific intention â e.g. to support âplanning, forecasting, or to transform existing data into new data or knowledge,â as Marchionini writes. He also includes âserendipitous browsing that is done to stimulate analogical thinkingâ in this category.
Chatbots like ChatGPT are useful tools for âlookupâ queries. They can help you find a known item. But they struggle to support search strategies in the latter two buckets. Bender and Shah use the example of someone concerned about being evicted from their home. They imagine a user who enters the query âWho can help me avoid being evicted?â The user isnât on a fact-finding mission, they want to explore different resources available to people at risk of losing their homes. But in an attempt to offer a direct answer, ChatGPT limits the process of sense-making. As Bender and Shah write, in this scenario, a chatbot âprevents the user from being able to build their own model of the space of possibilities available.â
Or take another example from Bender and Shah: âWhat is the number of a 24-hour advice nurse?â As they explain, a human might know important background context, like the fact that the service one can call depends on the userâs healthcare provider and/or insurance plan, which in turn depends on where they live. The user might be able to find this relevant information in the meta-data of a typical set of search results, but the chatbot? Hereâs Bender and Shah:
âA language-model-based dialogue agent, on the other hand, would likely synthesize a string with a phone-number shaped string of digits (possibly not even an actual phone number from a relevant source document) and might link to one or another of the web pages with text about advice nurses (not necessarily the same one with the phone number), but is unlikely to know to foreground the information about which patients the number is available to, nor to provide multiple options differentiated by healthcare provider/insurance plan.â
For a number of search types, context is king, but with chatbots, that context is often erased as it synthesizes strings of text.
Last, we know Google and search engines like it have exacerbated social biases. In Algorithms of Oppression, Safiya Noble reminds us that search results arenât a reflection of their truth or popularity, rather âsearch results reflect the values and norms of the search companyâs commercial partners and advertisers and often reflect our lowest and most demeaning beliefs.â Indeed, Noble goes on to show that search algorithms âprivilege whiteness and discriminate against people of color, specifically women of color.â But ChatGPT might make this even worse. With Google or Bing, the searcher sees racist or sexist results next to others; so theyâre nudged to ask, âum, where do these come from?â Now imagine the racist result coming from an authoritative seeming voice that we anthropomorphize as âintelligent.â As Bender and Shah write,
âWhere are the toe-holds that would allow a user to start to understand where the results are coming from, what biases the source data might contain, how those data were collected, and how modelling decisions might have amplified biases?â
This isnât a hypothetical concern. Microsoft once released a chatbot called Tay that was then quickly removed from the internet for responding to prompts with racist, xenophobic, and otherwise hateful language. Just months ago, Meta stopped developing a chatbot for similar reasons. This is, at least in part, why Google hasnât yet released its competitor into the wild.
đ The other reason Google might be hesitant to double down on chat bots? Itâs business model depends on you clicking ads next to its search results. Amr Awadallah, who worked for Yahoo and Google and now runs Vectara put it this way: âGoogle has a business model issue. If Google gives you the perfect answer to each query, you wonât click on any ads.â This is a spot-on example of the innovatorâs dilemma.
OpenAI is trying to solve this problem on the cheap â they outsourced the job of labeling violent, sexist, and racist text to Kenyan laborers â revealing yet another way in which emerging technologies are entangled with social systems: often hidden global labor!
So what to do about the future of search?
The very first thing OpenAI could do is be more transparent â it could share the corpus of data its using to train ChatGPT. But loyal Untangled readers will know thatâs not enough. Applying the framework of transparency scholar Jonathan Fox, we might also demand information about OpenAI and how internal decisions and practices shaped the development of ChatGPT. As Iâve written before, âalgorithms are shaped by people, practices, and beliefsâ so therefore âalgorithmic assessments should be ongoing, interrogate the decisions and assumptions that led to its design, and be able to demand changes to that design.â Another way to minimize harm might be annotating responses with a disclaimer. Potthast, Hagen, and Stein proposed systems like ChatGPT include the statement âThis answer is not necessarily true. It just fits well to your question.â
However, we should really take this as an opportunity to go beyond simply minimizing harm, and begin imagining alternative search systems. The limits of Google and ChatGPT â of which there are many â should offer strong starting points: we need systems that support exploration and our varied search intentions; systems that preserve context rather than flattening it, and systems that are transparent enough that we can understand how it produced the results that it did.
But above all, we need systems that are public; for which search results arenât unduly influenced by financial incentives. At the moment, we seem to be stuck in a place where we prioritize convenience over the goals of learning, exploration, truth, and equity.
What ChatGPT has done for us is made it clear that search systems can indeed be different, which is why itâs vitally important that we donât go down the same road as before; now is the time when we can strive for search systems that work for the public as a whole; ones that deliver the highest quality of information, to the most people.
Remarkable post, Charley. Thoughtful and critical, but in all the positively necessary ways.
Please Dearsupportt me and i support you.