Can algorithms speak? And should their opinions be protected?

by Jennifer Petersen

Photo of three women working on laptops in a café with festive lights and decorations hanging from the ceiling.

Algorithmic expression is not the classic, liberal model of speech. Instead, it is speech unmoored from personhood

Recommendation systems permeate our online lives. They structure what we buy, read, watch, listen to, even what we think. Though the inner workings of these algorithmic systems remain hidden from us, we encounter them across almost all digital platforms. Google, Amazon, Netflix, Facebook, Alibaba, Twitter, TikTok, Instagram, Spotify and myriad others use recommendation systems to direct us to certain types of content and products. We see the results of these systems any time we try to search for something or are presented with an ordered list of options, often with a ranking and hierarchy. But what exactly are these lists?

You might think of them as mere data. You might suspect they are the preferred choices of a company, brand or search engine. You might think of them as reflections of the archived search histories of countless other users. You may even suspect that they’re based on records of your own purchases, or particular words and syntax from your previous search queries. But would you ever think of these lists of recommendations as opinions – the utterances of an algorithm, or the company that created it?

Recommendation systems pose difficult questions about what it means to speak, and whether speaking is something that only a person does. How do we draw a line between expressions and actions? And who (or what) can be considered a ‘speaker’? In my work, I examine how such questions have been debated and decided in law. What I’ve found has surprised me: machines are reshaping how speech is regulated and defined, and they’re determining who has a voice in the public sphere. This is the case with search results and other algorithmic expressions. But can these recommendations really be considered ‘opinions’ or ‘expressions’? Untangling such problems raises broader questions about the changing nature of speech itself: in a public sphere populated by human, corporate and algorithmic utterances, what exactly does it mean to speak?

Companies like Google have argued that their search results are not data, nor a commercial product, nor an action. Instead, they’re a form of speech or opinion, like that of a newspaper editor. Google has done this, in part, for pragmatic reasons. Within the US, opinions are a form of expression constitutionally protected from regulation and other lawsuits, which is why so much hinges on whether a particular activity is understood as an example of ‘speech’ or ‘expression’. If search results are ‘speech’, then the companies that produce them are insulated from governmental regulation and protected from a range of legal threats, including accusations of anti-competitive practices, allegations of libel, or other complaints from competitors, content creators or users. This is why Google prefers that you – or rather, the US legal system – think of its search results as opinions.

From this perspective, characterising algorithmic outputs and decisions as opinions may seem like spin, an attempt by companies to game the system and find the most advantageous set of precedents. But there is also something much deeper going on in these debates. Google and other tech companies have made these characterisations because their search results really do resemble opinions – but there are important differences, too. That is, these ‘opinions’ overlap with human judgments and statements but are not equal to them. We do not yet have a good way of talking and thinking about this type of ‘speech’ or ‘expression’.

We may be tempted to use words like ‘opinion’ to describe the ‘speech’ of search results because algorithmic outputs share some similarities with human opinions and expression. Algorithms are constructed by humans and involve ideas, beliefs and value judgments. Search algorithms, for example, involve human decisions about how to define relevance and quality in search results, and how to operationalise these definitions. This happens, for example, when an algorithm’s creators must decide whether quality is best defined by many links to a website or by links from highly recognised and reputable sites. In this way, creators are engaging in high-level decision-making that resembles editorial decisions about newsworthiness. Even so, there are important differences between algorithmic outputs and how ‘speech’ is understood.

Speech is often defined in ways that connote human consciousness, intentionality, personality and agency. When we talk about speech, we use terms such as ‘opinion’, ‘judgment’, ‘utterance’, ‘advocacy’, ‘persuasion’, ‘conscience’, ‘beliefs’ and ‘feelings’. These terms have a rich history in humanist thought, as well as in legal reasoning about freedom of speech. Traditionally, speaking has involved an individual externalising an idea or feeling through spoken or written words, images, music or even dress, gesture, dance and other actions. This is reflected in law: legal protections for freedom of speech in the US include actions such as saluting (or burning) a flag, wearing armbands, sitting or standing in silent protest, naked dancing, watching movies, playing phonographs, and more. In all these examples, the act is an expression because it conveys to others some form of mental activity or belief held by the person doing the ‘expressing’. This is the traditional, liberal conception of what it means to speak. In this tradition, speech or speaking are closely tied to what it means to be a person and a participant in a political community. We express this idea when we refer to ‘speaking subjects’ or an individual or group ‘having a voice’.

Algorithmic outputs do not quite fit this definition. They may be shaped by the values, intentions and beliefs of their creators, but they are not equal to them. They display what I call a distinct structure of speech. They do not involve a speaker who wishes to express a particular idea, opinion or belief (as in a typical conversation, publication or expression of opinion). In the complex algorithmic systems that characterise much of our current media landscape, the chains of causality and intent between creators of the algorithm and its outputs are often attenuated or indirect. Problems appear when we try to think about these outputs as expressions of intentionality, agency or judgments – that is, like the liberal conception of speech.

The outputs of algorithmic systems are shaped by, but not reducible to, the ideas, judgments, beliefs or values of their designers. Search results do not reflect or express what either Google or its engineers think about your search query. Instead, these results reflect what they think about how to define relevance and quality in search queries in general. As such, we can’t simply describe these outputs as the expressions of their human creators. In fact, in many such systems, designers are not the only sources of judgment and values. Complex algorithmic systems are dynamic, and the outputs of these systems depend on multiple and often changing streams of data drawn from crowdsourced online activity. As the political geographer (and algorithm expert) Louise Amoore puts it in her book Cloud Ethics (2020), ‘each and every one of us’ is conscripted into writing these systems. We do so without intent, knowledge or coordination.

Another example from Google helps to illustrate this dynamic, in which the company claimed that its search results were not its opinion but a mere reflection of popular belief. Circa 2004, the top result in searches for the term ‘Jew’ was an antisemitic website. While the content of the website was clearly the expression of its creator’s ideas, the source or cause of the recommendation itself was (and remains) less clear. In this instance, Google argued that its recommendations did not reflect the company’s values or beliefs – it was not Google’s opinion, it was a pure reflection of what was popular online. The results were the sum of judgments from a crowd of anonymous others who had linked to or interacted with the antisemitic website. These crowdsourced judgments or decisions are not produced by a set of people deliberately working together towards an end, but rather by a variety of online interactions with varying and often opaque causal relations to the algorithmic output. Again, the company was positioning itself in the most advantageous legal position regarding its search results: in this case, distance, and a lack of responsibility. Yet, at the same time, Google’s arguments demonstrate how algorithmic outputs are both like and unlike opinions. Because recommendations are shaped by the values and priorities of their designers, creators do bear some responsibility for them. But they are also not equal to the expression of a singular set of designers. In other words, recommendations are not equal to the classic, liberal model of speech. Algorithmic expression is something different: speech without speakers, or speech unmoored from personhood.

Attempting to think through the implications of algorithmic systems via the classic models of speech and expression isn’t helpful, but that hasn’t stopped Google and other companies from trying. The presumption of a direct link between expression and intentionality, and of an agent (like an individual human person) behind the speech, has been central to the corporate opportunism outlined above. Google has used these arguments to position itself as a classic speaker, or person, who can claim legal rights – and a voice in the public sphere. These arguments would have us ask: who is really speaking? But that’s the wrong question. Searching for a particular agent or speaker in cases of algorithmic expression isn’t likely to help us think through the complex ethical and political questions surrounding these systems because algorithmic outputs are the products of dynamic systems and many different (often unwitting) actors.

Instead, algorithmic expression might present us with the opportunity to rethink the limits of our current conceptions of what it means to speak as a person. As algorithms increasingly define the ways we know the world in the 21st century, we need more ways to trace lines of causality and intent in expression. We need language and frameworks that will allow us to recognise new forms of agency and responsibility – attenuated forms that fall below the threshold we imagine for human speakers.

Rather than thinking about online recommendations as the opinions of a company or a set of engineers who designed the system, we might understand them as speech with a different structure, exhibiting a different sort of expression and agency. Algorithmic utterances are more than just an amalgamation of the decisions and actions of designers and others at the companies who build them. When you look at your search results and other forms of recommendations, you’re also seeing, all at once: a weave of mere data; the values or goals of a company, brand or set of designers; the opinions of others who have read, watched or linked to the results you are looking at (and to results you do not see); reflections of the archived search histories of myriad users – including you. This kind of distributed expression is not proof that algorithms are persons, or that their ‘opinions’ really are like those of humans, or even companies. Instead, it suggests something we’re perhaps even less prepared for: a way of thinking and talking about speech without the need to find a fully intentional speaker lurking somewhere behind it.

Syndicate this idea

Explore more

A young person lying on a bed in a cluttered room with a suitcase, laundry basket, chest of drawers and desk, illuminated by window.

For young people, AI is now a second brain – should we worry?

As a resident tutor, I’ve seen how students are using AI as more than a tool. It’s a psychological shift we’ll soon all make

by Rhea Tibrewala

A photo showing a person holding a bouquet of mixed flowers wrapped in brown paper, viewed from above and behind.

The words that make an apology convincing

Research on the psychology of apologising finds that certain word choices are more convincing because of their mental cost

by Alice Gregory

A person in the air, mid-flip, on a grassy hill with distant hills in the background.

How to check if an argument is valid

In logic, validity is prime. If you want to make valid arguments, or sniff out invalid ones, here’s what you need to do

by Robert Trueman

A switched-off smartphone on a bright yellow surface with fingerprints and smudges on the black screen.

A brief escape from social media

After leaving my phone behind for a week and coming back to it, I saw my social media use in a stark new light

by Tamur Qutab

A person using a smartphone, with focus on their finger scrolling the screen. Face partially visible.
ADHD

In an era of split attention, there is more than one type of ADHD

ADHD is typically thought to be wired into the brain early. But many cases may be better seen as products of digital life

by Paul Kudlow, Karline Treurnicht Naylor & Elia Abi-Jaoude

Aerial photo of Würzburg, a European cityscape with historic buildings, church spires and a clock tower under a partly cloudy sky.
TRAVEL

Why ‘false authenticity’ is so unsettling

The Old Town of Würzburg seemed to offer a real connection to the past. But none of it was authentic

by Sam Dresser

Collage of animal photos with progressively simplified line drawings of a flamingo, giraffe and cat.

From cave art to subatomic sketches – how drawing has accelerated human progress

Video by MIT Quest for Intelligence

People sitting by a small stream in a wooded area during daytime, with sunlight filtering through bare trees.

Whispers in the wind

How I learned to hear the voice of my ancestors and speak Lenape, a language silenced for generations

Krista Nelson, as told to Dawn Fallik