Sociotechnical Harms of Algorithmic Systems ACM AIES 2023, August 9-11th 2023, Montreal, QC
2.2 Taxonomies of Harm, Risk, and Failure
Structured frameworks can aid practitioners’ anticipation of harms
throughout the product lifecycle [
135
,
239
]. They encourage more
rigorous analysis of social and ethical considerations [
135
], espe-
cially when operationalizing principles seems opaque [
147
]. Taxon-
omizing harms is, however, an exercise in classication, which has
potential limitations: taxonomies can draw action to some issues
over others, shaping how people navigate and act on informa-
tion [
39
]. As such, the epistemological choices made in developing
harm taxonomies focus attention on certain areas over others [
40
].
Many existing harm taxonomies address particular domains of
use (e.g., [
193
,
217
]) and how they are complex assemblages of ac-
tors, norms, practices, and technical systems, which can foster indi-
vidual and collective harm [
174
]. Taxonomies have been developed
related to online content [
19
,
193
,
227
], social media [
157
,
217
,
218
],
users “at-risk" for experiencing online abuse [
216
,
234
], and mali-
cious uses of algorithmic systems [
45
], including cyber attacks [
3
]
and cyberbullying [
12
]. Relatedly, they can focus on particular types
of harms, such as misinformation [
218
] or representational harms
co-produced through image tagging system inputs and outputs
that reinforce social hierarchies [
121
,
228
]. While domain-specic
taxonomies draw attention to how context informs the emergent
nature of harm, they are not easily applicable to a wide range of
algorithmic systems. Many systems deploy across contexts, includ-
ing, for example, ranking and recommendation systems (e.g., search
engines or content sorting algorithms on social media platforms)
and object detection models (e.g., video surveillance systems, self-
driving cars, and accessibility technology).
Another common approach is to orient harm taxonomies around
specic algorithmic or model functions (e.g., [
83
,
235
]). Model-
focused taxonomies have been developed for large language mod-
els [
235
], image captioning systems [
121
,
228
], and so-called “foun-
dational models,” such as GPT-3 and BERT, which are applied in a
wide range of downstream tasks [
37
]. Organizing harm by model
function is highly useful when practitioners’ focus is on a singular
model because it draws attention to relevant computational con-
cerns. It does, however, pose limitations to practitioners working
downstream on products and features, where multiple model types
operate simultaneously, such as in social media, search engines,
and content moderation, and where contextual use signicantly
shapes potential harms.
Scholars have developed harm taxonomies related to system
misbehaviors and failures [
18
,
198
], particularly to aid algorithmic
auditing [
177
]. These taxonomies focus on how algorithmic sys-
tems are sources of harm (e.g., faulty input/outputs, limited testing,
proxy discrimination, and surveillance capitalism) [
203
]. Bandy [
18
]
summarizes four problematic behaviors of algorithmic systems —
discrimination, distortion, exploitation, and misjudgement. Using
such taxonomies focus attention to how specic aordances, train-
ing data, and design choices can co-produce harm [
22
,
40
]. Failure-
based taxonomies are helpful when practitioners examine potential
failure modes of a specic technology, but are often limited in terms
of helping to anticipate who or what is harmed.
In sum, taxonomies can be helpful tools for appreciating and
assessing how harms from algorithmic systems are sociotechnical.
As they retain social and technical elements, they cannot be reme-
died by technical xes [
94
] alone. They require social and cultural
change [
171
]. The proposed taxonomy provides a holistic and sys-
tematic view of the current discourse on types of sociotechnical
harms from algorithmic systems. As the scope of the taxonomy
we propose is broad, and many topic- or application- specic tax-
onomies exist already, we refer to and build on these existing works
when appropriate.
3 METHODOLOGY
In alignment with prior calls to anticipate computational and con-
textual harms [
40
,
151
], we synthesize insights on harms from
computing research to aid anticipation of sociotechnical harms
from algorithmic systems. Our ndings draw on a scoping review
for data collection and a reexive thematic analysis of computing
research on harms.
3.1 Overview of Methodology
Our approach followed prior scoping reviews in HCI literature [
74
,
223
], in alignment with the extension of the PRISMA checklist [
133
],
the PRISMA-ScR (Preferred Reporting Items for Systematic reviews
and Meta-Analyses extension for Scoping Reviews) [
220
]. Scoping
studies, as a knowledge synthesis methodology, map existing litera-
ture, and “clarify working denitions and conceptual boundaries of
a topic or eld” [
168
, p. 141]. They are especially appropriate when
distilling and sharing research ndings with practitioners [
221
],
and are suited to identifying evidence on a topic and presenting it vi-
sually. Compared to systematic reviews, scoping reviews address a
broader range of topics and incorporate dierent study designs [
11
].
A scoping review is an eective method for surfacing current “pri-
orities for research, clarifying concepts and denitions, providing
research frameworks or providing background, or contextual infor-
mation on phenomena or concepts" [
65
, p. 2104]. We implemented
a ve-stage scoping review framework [
11
,
128
]: (1) Identifying
research questions; (2) Identifying relevant studies; (3) Study se-
lection; (4) Charting the data; and (5) Collating, summarizing, and
reporting results.
3.1.1 Identify research questions. To identify the types and range
of sociotechnical harms, we developed the three aforementioned
research questions (see: Section 1).
3.1.2 Identify and select relevant studies. We then employed mul-
tiple strategies to identify relevant resources through dierent
sources: electronic scholarly databases, a citations-based review,
and targeted keyword searches in relevant organizations, and con-
ferences. Using the ACM Guide to Computing Literature as the
primary search system – which reects key computing research
databases – we developed the following initial set of key concepts to
search full text and metadata: “algorithmic harm”, “sociotechnical
harm”, “AI harm”, “ML harm”, “data harm”, “harm(s) taxonomy, “al-
locative harm”, and “representational harm.” Within scoping review
methodology, keyword search strategies should be devised to sur-
face relevant literature [
11
] and include terms common to the eld
[
101
]. We included allocative and representational harm as search
terms because of their conceptual dominance in machine learning
literature since 2017, popularized by responsible ML scholars (e.g.,