Algorithmic Accountability Primer Research
Algorithmic Accountability Primer Research
A Primer
Data & Society is a 501(c)(3) research institute focused on the social and
cultural issues arising from data-centric technological development.
datasociety.net
Algorithmic Accountability: A Primer 1
Table of Contents
What Is an Algorithm? ………………………………………………………………….……………………. 2
How Are Algorithms Used to Make Decisions? ……………………………….…………….………. 2
Example: Racial Bias in Algorithms of Incarceration ……………………….……………….…… 4
Complications with Algorithmic Systems ………………………………………….…………….……. 6
Fairness and Bias ………………………………………………………………….………….……… 6
Opacity and Transparency ……………………………………………………….…….…………. 7
Repurposing Data and Repurposing Algorithms ……………………………….……..…. 7
Lack of Standards for Auditing ……………………………………………………….………. 8
Power and Control ……………………………………………………………………….…………. 8
Trust and Expertise ……………………………………………………………….…………………. 9
What is Algorithmic Accountability? ………………………………………………….……………….. 10
Auditing by Journalists ………………………………………………………….……………….. 10
Enforcement and Regulation ………………………………………………….……………….. 10
What Is an Algorithm?
An algorithm is a set of instructions for how a computer should accomplish a par-
ticular task. Algorithms are used by many organizations to make decisions and allocate re-
sources based on large datasets. Algorithms are most often compared to recipes, which take a
specific set of ingredients and transform them through a series of explainable steps into a pre-
dictable output. Combining calculation, processing, and reasoning, algorithms can be excep-
tionally complex, encoding for thousands of variables across millions of data points. Critically,
there are few consumer or civil rights protections that limit the types of data used to build data
profiles or that require the auditing of algorithmic decision-making. Standards and enforce-
ment for fairness, accountability, and transparency are long overdue for algorithms that allocate
housing, healthcare, hiring, banking, social services, as well as goods and service delivery.1 Al-
gorithmic accountability is the process of assigning responsibility for harm when algorithmic
decision-making results in discriminatory and inequitable outcomes.
While algorithmic decision making can offer benefits in terms of speed, efficiency, and even
fairness, there is a common misconception that algorithms automatically result in unbiased
decisions. While it may appear like algorithms are unbiased calculations because they take in
objective points of reference and provide a standard outcome, there remain many problems with
those inputs and the outputs. As Frank Pasquale, law professor at the University of Maryland,
points out, algorithmic decision-making is “black boxed,” which means that while we may
1Eubanks, Virginia. 2018. Automating Inequality: How High-Tech Tools Profile, Police, and Punish
the Poor. New York, NY: St. Martin’s Press.
know what goes into the computer for processing and what the outcome is, there are currently
no external auditing systems or regulations for assessing what happens to the data during pro-
cessing.2
Algorithms are attractive because they promise neutrality in decision making—they take
in data and deliver results. But algorithms are not “neutral.” In the words of mathematician
Cathy O’Neil, an algorithm is an “opinion embedded in mathematics.”3 And like opin-
ions, all algorithms are different. Some algorithms privilege a certain group of people over an-
other. O’Neil argues that across a range of occupations, human decision makers are being en-
couraged to defer to software systems even when there is evidence that a system is making in-
correct, unjust, or harmful decisions.
When an algorithm’s output results in unfairness, we refer to it as bias. Bias can find
its way into an algorithm in many ways. It can be created through the social context where an
algorithm is created, as a result of technical constraints, or by the way the algorithm is used in
practice.4 When an algorithm is being created, it is structured by the values of its designer, which
might not be neutral. And after an algorithm is created, it must be trained—fed large amounts
of data on past decisions—to teach it how to make future decisions. If that training data is itself
biased, the algorithm can inherit that bias. For these reasons and others, decisions made by
computer are not fundamentally more logical and unbiased than decisions made by people.
Black-boxed algorithms can unfairly limit opportunities, restrict services, and even pro-
duce “technological redlining.” As Safiya Noble, professor of communication at University
of Southern California, writes, technological redlining occurs when algorithms produce inequi-
table outcomes and replicate known inequalities, leading to the systematic exclusion of Blacks,
Latinos, and Native Americans.5 Technological redlining occurs because we have no control over
how data is used to profile us. If bias exists in the data, it is replicated in the outcome. Without
enforceable mechanisms of transparency, auditing, and accountability, little can be known
about how algorithmic decision-making limits or impedes civil rights.
Noble writes, “technological redlining is a form of digital data discrimination, which uses
our digital identities and activities to bolster inequality and oppression. It is often enacted with-
out our knowledge, through our digital engagements, which become part of algorithmic, auto-
mated, and artificially intelligent sorting mechanisms that can either target or exclude us. It is
a fundamental dimension of generating, sustaining, or deepening racial, ethnic, and gender dis-
crimination, and it is centrally tied to the distribution of goods and services in society, like ed-
ucation, housing, and other human and civil rights. Technological redlining is closely tied to
2 Pasquale, Frank. 2015. The Black Box Society: The Secret Algorithms That Control Money and Infor-
mation. Harvard University Press.
3 O’Neil, Cathy. 2016. Weapons of Math Destruction. Crown.
4 Batya Friedman and Helen Nissenbaum, “Bias in Computer Systems,” ACM Transactions on Infor-
longstanding practices of ‘redlining,’ which have been consistently defined as illegal by the
United States Congress, but which are increasingly elusive because of their digital deployments
through online, internet-based software and platforms, including exclusion from, and control
over, individual participation and representation in digital systems.”6 Important examples of
technological redlining were uncovered by ProPublica, who showed how Facebook’s targeted
advertising system allowed for discrimination by race and age.7 These decisions embedded in
design have significant ramifications for those who are already marginalized.
In this memo, we begin by showcasing one example to illustrate how racial bias mani-
fests in an algorithmic system. We then address the trade-offs between and debates about algo-
rithms and accountability across several key ethical dimensions: fairness and bias; opacity and
transparency; the repurposing of data and algorithms; lack of standards for auditing; power and
control; as well as trust and expertise. From there, we provide an overview of algorithmic ac-
countability by highlighting how news coverage and self-governance have further exacerbated
problems related to unfair, unethical, and possibly illegal applications of algorithmic systems.
6 Noble wrote this definition of “technological redlining” specifically for this publication.
7 Julia Angwin, Ariana Tobin. 2017. “Facebook (Still) Letting Housing Advertisers Exclude….” ProPub-
lica. November 21, 2017. https://www.propublica.org/article/facebook-advertising-discrimination-
housing-race-sex-national-origin.
Angwin, Julia, Noam Scheiber, and Ariana Tobin. 2017. “Facebook Job Ads Raise Concerns About Age
Discrimination.” The New York Times, December 20, 2017, sec. Business Day. https://www.ny-
times.com/2017/12/20/business/facebook-job-ads.html.
8 Christin, Angele, Alex Rosenblat, and danah boyd. “Courts and Predictive Algorithms.” CRIMINAL
9 Tim Brennan, Bill Dieterich, Beate Ehret, “Research Synthesis: Reliability and validity of COMPAS,”
Northpointe Inc., September, 2007.
10 Brennan, Tim, William Dieterich, and Beate Ehret. “Evaluating the Predictive Validity of the Compas
Risk and Needs Assessment System.” Criminal Justice and Behavior 36, no. 1 (January 2009): 21–40.
https://doi.org/10.1177/0093854808326545.
11 Blomberg, Thomas, William Bales, Karen Mann, Ryan Meldrum, and Joe Nedelec. “Validation of the
COMPAS Risk Assessment Classification Instrument.” College of Criminology and Criminal Justice,
Florida State University, Tallahassee, FL, 2010. http://criminology.fsu.edu/wp-content/uploads/Vali-
dation-of-the-COMPAS-Risk-Assessment-Classification-Instrument.pdf.
12 Chouldechova, Alexandra. “Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Pre-
Risk and Needs Assessment System.” Criminal Justice and Behavior 36, no. 1 (January 1, 2009): 21–
40. https://doi.org/10.1177/0093854808326545.
14 Blomberg, Thomas, William Bales, Karen Mann, Ryan Meldrum, and Joe Nedelec. “Validation of the
cle/machine-bias-risk-assessments-in-criminal-sentencing.
16 Julia Angwin et. al., “Machine Bias.”
directly affect people’s lives.17 The bill highlights a need for assessment of disproportionate im-
pacts across protected categories as well as a procedure for redress if harms are found.
embedded in the training data. Because black people have historically been arrested at a higher
rate than white people, COMPAS learned to predict that a black person is more at risk of being
re-arrested than a white person. When implemented, this system reflects this learning back into
the criminal justice system at a large scale, injecting a source of racial bias into steps of the
judicial process that come after arrest.
By transferring values from one particular political and cultural moment to a different
context, algorithms create a certain moral rigidity. Unless algorithms are consistently moni-
tored and adjusted as time passes, they reinforce the values they were created with and can
become rapidly outdated. For example, in terms of apportionment of healthcare, service deliv-
ery by insurance companies and hospitals depends on algorithmic decision-making, yet some
doctors and caregivers do not agree with the standardized treatment models because these data
are not robust enough to assess variables unavailable to the computer model, such as the un-
steady living conditions of those in poverty.
17 “The New York City Council - File #: Int 1696-2017.” Accessed April 15, 2018. http://legistar.coun-
cil.nyc.gov/LegislationDetail.aspx?ID=3137815&GUID=437A6A6D-62E1-47E2-9C42-461253F9C6D0.
18 Suresh Venkatsburamanian. “When an algorithm isn’t,” Medium, October 1, 2015, https://me-
dium.com/@geomblog/when-an-algorithm-isn-t-2b9fe01b9bb5
19 Batya Friedman, Peter H. Kahn Jr, and Alan Borning, “Value Sensitive Design and Information Sys-
20 Huet, Ellen. “Server and Protect: Predictive Policing Firm PredPol Promises to Map Crime Before It
Happens.” Forbes. Accessed April 10, 2018. https://www.forbes.com/sites/ellenhuet/2015/02/11/pred-
pol-predictive-policing/.
21 Lartey, Jamiles. “Predictive Policing Practices Labeled as ‘flawed’ by Civil Rights Coalition.” The
may need new, more cost-effective systems for creating algorithms or more standards in place
for evaluating when an algorithm can be successfully adapted from one application to another.
22 “Statement on Algorithmic Transparency and Accountability,” ACM US Public Policy Council, Janu-
ary 12, 2017, https://www.acm.org/binaries/content/assets/public-policy/2017_usacm_statement_al-
gorithms.pdf.
23 Fairness, Accountability, and Transparency in Machine Learning. n.d. “Principles for Accountable
Algorithms and a Social Impact Statement for Algorithms :: FAT ML.” Accessed April 11, 2018.
https://www.fatml.org/resources/principles-for-accountable-algorithms.
24 Simpson, Scott. “Muslim Advocates and Color Of Change Demand Independent Civil Rights Audit of
However, their decisions about relevance are choices shaped by a political agenda, whether that
agenda is implicit or explicit to even the algorithm’s own designers.27 This is especially im-
portant for algorithms that perform a gatekeeping role. Algorithms replicate social values but
also embed them into systems, creating new standards and expectations for what is important
in a given context. While there are laws prohibiting the sharing or sale of health and financial
data by hospitals and banks, discrimination occurs because there are few protections in place
for consumer data brokering, where discrete data points act as proxies for protected categories
that are then assembled into profiles that are sold. This can lead to technological redlining.
Below we outline how journalists, in consultation with academics and whistleblowers, have
taken up the role of auditing algorithms, while also showing how the lack of enforceable regu-
lation led to a deficit in consumer protections.
Auditing by Journalists
Currently, journalists are an important watchdog for algorithmic bias. Data journalism blends
investigative methods from journalism with technical know-how to provide clear and accurate
reporting on computational topics. While many algorithms are proprietary information, skilled
journalists can use techniques of “reverse-engineering” to probe what’s inside the black box by
pairing inputs with outputs. A second approach facilitated by journalists is that of collaborative
research with academics and whistleblowers. Particularly for personalization algorithms, which
can be difficult or impossible to parse from the perspective of an individual user’s account, peer-
sourced research can reveal patterns that give clues about how the underlying algorithms work.
However, for accountability to be meaningful, it needs to come with the appropriate gov-
ernance structures. According to Florian Saurwein, Natascha Just, and Michael Latzer, govern-
ance is necessary because algorithms impose certain risks, such as the violation of privacy rights
and social discrimination.29 These risks need to be dealt with by the appropriate governance
structure, which currently involves little oversight by states. Governance can occur by market
and design solutions, such as product innovation that mitigates risk or consumers’ ability to
substitute risky products for ones they deem safer. Governance can also come from industry
self-regulation, where company principles and collective decision-making favor public interest
concerns. Last is traditional state intervention through mechanisms such as taxes and subsidies
for certain kinds of algorithmic behavior. The appropriate structure must be matched with the
context at hand to ensure the accountability mechanisms are effective.
Because of the ad hoc nature of self-governance by corporations, few protections are in
place for those most affected by algorithmic decision-making. Much of the processes for obtain-
ing data, aggregating it, making it into digital profiles, and applying it to individuals are corpo-
rate trade secrets. This means they are out of the control of citizens and regulators. As a result,
there is no agency or body currently in place that develops standards, audits, or enforces neces-
sary policies.
While law has always lagged behind technology, in this instance technology has become
de facto law affecting the lives of millions—a context that demands lawmakers create policies
for algorithmic accountability to ensure these powerful tools serve the public good.
29Saurwein, Florian, Natascha Just, and Michael Latzer. “Governance of Algorithms: Options and Limi-
tations.” Info 17, no. 6 (September 14, 2015): 35–49. https://doi.org/10.1108/info-05-2015-0025.