Data Analytics and Algorithmic Bias in Policing: Briefing Paper
Data Analytics and Algorithmic Bias in Policing: Briefing Paper
BRIEFING PAPER
This paper summarises the use of analytics and algorithms for policing
within England and Wales, and explores different types of bias that can
arise during the product lifecycle. The findings are based on in-depth
consultation with police forces, civil society organisations, academics and
legal experts. The primary purpose of the project is to inform the Centre
for Data Ethics and Innovation's ongoing review into algorithmic bias in
the policing sector. RUSI’s second and final report for this project will be
published in early 2020, to include specific recommendations for the final
Code of Practice, and incorporating the cumulative feedback received
during the consultation process.
The authors made an equal contribution to this paper, and are very grateful
to Jamie Grace, Pete Fussey, Malcolm Chalmers, Cathy Haenlein and the
RUSI Publications team for providing helpful feedback on an earlier draft
of this report.
SUMMARY
• The use of data analytics and algorithms for policing has numerous
potential benefits, but also carries significant risks, including those
relating to bias. This could include unfair discrimination on the
grounds of protected characteristics, real or apparent skewing of
the decision-making process, or outcomes and processes which are
systematically less fair to individuals within a particular group. These
risks could arise at various stages in the project lifecycle.
• Algorithmic fairness cannot be understood solely as a matter
of data bias, but requires careful consideration of the wider
operational, organisational and legal context, as well as the overall
decision-making process informed by the analytics.
• While various legal frameworks and codes of practice are relevant to
the police’s use of analytics, the underlying legal basis for use must
be considered in parallel to the development of policy and regulation.
Moreover, there remains a lack of organisational guidelines or clear
processes for scrutiny, regulation and enforcement. This should be
addressed as part of a new draft code of practice, which should specify
clear responsibilities for policing bodies regarding scrutiny, regulation
and enforcement of these new standards.
BABUTA AND OSWALD 3
INTRODUCTION
RUSI was commissioned by the Centre for Data Ethics and Innovation (CDEI)
to conduct an independent research study on the use of data analytics by
police forces in England and Wales, with a focus on algorithmic bias. The
primary purpose of the project is to inform CDEI’s review into algorithmic
bias in the policing sector.
CDEI’s review will work towards a draft Code of Practice for the development,
trialling and implementation of data analytics in policing, which will mitigate
risk of bias and address wider legal and ethical concerns.
This briefing paper is the first of two papers to be published as part of this
project. Its purpose is to outline the main issues identified by the research
so far, and to offer contextual information for stakeholders providing
feedback on the draft Code of Practice, due to be published by CDEI. The
second RUSI paper will be published in early 2020 and will include specific
recommendations for the final CDEI Code of Practice, incorporating feedback
received during the consultation process.
This briefing paper summarises the use of analytics and algorithms for
policing in England and Wales, before discussing different types of bias that
can arise during the product lifecycle. The purpose of this paper is not to
offer solutions or recommendations as to how these risks can be addressed;
this will be discussed in detail in the second paper. Instead, this briefing
paper highlights the most significant gaps in the existing policy framework
where future regulatory attention should be focused.
of more sophisticated systems, which are now used for a wider range of
functions. The use of algorithms to make predictions about future crime
and offending raises considerable legal and ethical questions, particularly
concerning the risk of bias and discrimination.8
DOES IT WORK?
Before discussing the risks of bias arising from predictive policing technology,
it is important to address the fundamental question – ‘does it work?’ It is
beyond the scope of this paper to critically assess the (dis)advantages of a
risk assessment-focused approach to resource allocation, or to discuss the
semantic nuances associated with defining ‘risk’. However, on the basis that
police forces must target limited resources to places and people identified as
posing the greatest ‘risk’ (of offending or victimisation), this paper’s starting
point is to question whether algorithmic tools are effective in assisting the
police to identify and understand this risk. Effectiveness and accuracy are
intrinsically linked to ethics and legality: if it cannot be demonstrated that
a particular tool or method is operating effectively and with a reasonable
degree of accuracy, it may not be possible to justify the use of such a tool as
necessary to fulfil a particular policing function.9
the deployment location but also in surrounding areas.12 In the UK, field
trials have found predictive mapping software to be around twice as likely to
predict the location of future crime as traditional intelligence-led techniques
(whereby analysts manually identify future hotspots).13
12. Lawrence W Sherman and David Weisburd, ‘General Deterrent Effects of Police
Patrol in Crime “Hot Spots”: A Randomized, Controlled Trial’, Justice Quarterly
(Vol. 12, No. 4, 1995), pp. 625–48; Rob T Guerette and Kate J Bowers, ‘Assessing
the Extent of Crime Displacement and Diffusion of Benefits: A Review of
Situational Crime Prevention Evaluations’, Criminology (Vol. 47, No. 4, November
2009), pp. 1331–68; College of Policing, ‘The Effects of Hot-Spot Policing on
Crime: What Works Briefing’, September 2013.
13. George Mohler et al., ‘Randomized Controlled Field Trials of Predictive Policing’,
Journal of the American Statistical Association (Vol. 110, No. 512, 2015),
pp. 1399–411; Kent Police Corporate Services Analysis Department, ‘PredPol
Operational Review [Restricted and Heavily Redacted]’, <http://www.statewatch.
org/docbin/uk-2014-kent-police-predpol-op-review.pdf>, accessed 14 August
2019.
14. Conchman, ‘Policing by Machine’, p. 45.
15. Conchman, ‘Policing by Machine’, pp. 45–61; Babuta, ’Big Data and Policing’.
16. For further discussion, see Robyn M Dawes, David Faust and Paul E Meehl,
‘Clinical Versus Actuarial Judgment’, Science (Vol. 243, No. 4899, 1989),
pp. 1668–74; William M Grove and Paul E Meehl, ‘Comparative Efficiency of
Informal (Subjective, Impressionistic) and Formal (Mechanical, Algorithmic)
Prediction Procedures: The Clinical–Statistical Controversy’, Psychology, Public
Policy, and Law (Vol. 2, No. 2, 1996), pp. 293–323.
17. Paul E Meehl, Clinical Vs. Statistical Prediction: A Theoretical Analysis and
a Review of the Evidence (Minneapolis, MN: University of Minnesota Press,
1954); William M Grove et al., ‘Clinical Versus Mechanical Prediction: A
Meta-Analysis’, Psychological Assessment (Vol. 12, No. 1, 2000), p. 19.
18. See, for example, Stephen D Hart and David J Cooke, ‘Another Look at the
(Im‐)Precision of Individual Risk Estimates Made Using Actuarial Risk Assessment
BABUTA AND OSWALD 7
Instruments’, Behavioral Sciences and the Law (Vol. 31, No. 1, January/February
2013), pp. 81–102.
19. Mia Debidin (ed.), A Compendium of Research and Analysis on the Offender
Assessment System (OASys) 2006–2009 (London: Ministry of Justice, 2009),
p. 78.
20. Alan A Sutherland et al., ‘Sexual Violence Risk Assessment: An Investigation
of the Interrater Reliability of Professional Judgments Made Using the Risk for
Sexual Violence Protocol’, International Journal of Forensic Mental Health
(Vol. 11, No. 2, 2012), p. 120.
21. Academic expert during Interdisciplinary Roundtable event hosted by RUSI and
the CDEI, London, 25 July 2019.
BRIEFING PAPER 8
28. ‘Convention for the Protection of Human Rights and Fundamental Freedoms’,
1950.
29. For an assessment of the legal basis of live facial recognition, see Pete Fussey
and Daragh Murray, ‘Independent Report on the London Metropolitan’s Police
Service’s Trial of Live Facial Recognition Technology’, The Human Rights, Big Data
and Technology Project, July 2019.
30. Nicol Turner Lee, Paul Resnick and Genle Barton, ‘Algorithmic Bias Detection and
Mitigation: Best Practices and Policies to Reduce Consumer Harms’, Brookings,
22 May 2019.
31. ‘Equality Act 2010 (UK)’, c.15.
32. This reflects, in particular, administrative law natural justice (rules of fair
administrative procedure applicable to public bodies) and Article 6(1) of the
ECHR.
33. Julia Powles, ‘The Seductive Diversion of “Solving” Bias in Artificial Intelligence’,
Medium, 7 December 2018.
34. Author’s telephone interview with L1, representative of UK law enforcement
agency, 1 July 2019.
BRIEFING PAPER 10
Several legal frameworks and codes of practice are relevant to the development
and deployment of predictive algorithms in England and Wales, including:
The project of which this paper is part assesses the risk of bias in algorithmic
decision-making in relation to these frameworks.
Despite general cuts in police funding since 2010, specific funds for ‘digital
transformation’ have been made available, such as the Police Transformation
Fund, ‘creating strong incentives for forces to frame the development
around digital technology to receive further central support’.48 This may
create a bias in favour of digital solutions, and the risk of modelling complex
social issues in an overly simplistic way,49 without equal consideration of
non-technological measures. This raises questions regarding necessity and
proportionality and data minimisation issues if the use of digital solutions
results in increased data collection to build, test and monitor a model. An
academic expert suggested that, where adoption of predictive policing
occurs rapidly, ‘what is often missing … is a testing and assessment of the
risks and long-term benefits that these systems may provide’.50 Another
police respondent pointed to ‘a risk that we procure capabilities based on
opinion-based decisions rather than evidence-based decisions’.51 With this
Algorithms that are trained on police data ‘may replicate (and in some
cases amplify) the existing biases inherent in the dataset’,53 such as over-
or under-policing of certain communities, or data that reflects flawed or
illegal practices54 (raising further issues regarding the requirement of
accuracy pursuant to the Data Protection Act).55 A police officer interviewed
commented that ‘young black men are more likely to be stop and searched
than young white men, and that’s purely down to human bias. That human
bias is then introduced into the datasets, and bias is then generated in the
outcomes of the application of those datasets’.56 The effects of a biased
sample could be amplified by algorithmic predictions via a feedback loop,57
whereby future policing is predicted, not future crime.58 Another officer
commented that ‘we pile loads of resources into a certain area and it becomes
a self-fulfilling prophecy, purely because there’s more policing going into
that area, not necessarily because of discrimination on the part of officers’.59
52. Solon Barocas and Andrew D Selbst, ‘Big Data’s Disparate Impact’, California Law
Review (Vol. 104, 2016), p. 671.
53. Babuta, Oswald and Rinik, ‘Machine Learning Algorithms and Police
Decision-Making’.
54. Rashida Richardson et al., ‘Dirty Data, Bad Predictions: How Civil Rights
Violations Impact Police Data, Predictive Policing Systems, and Justice’, New York
University Law Review Online (forthcoming).
55. ‘Data Protection Act 2018 (UK)’, c. 12, ss. 37, 38.
56. Author’s telephone interview with L1, representative of UK law enforcement
agency, 1 July 2019.
57. Frederik Zuiderveen Borgesius, ‘Discrimination, Artificial Intelligence and
Algorithmic Decision-Making’, Council of Europe, 2018.
58. Danielle Ensign et al., ‘Runaway Feedback Loops in Predictive Policing’,
Proceedings of Machine Learning Research (Vol. 81, 2018), pp. 1–12.
59. Author’s telephone interview with L2 and L3, representatives of UK law
enforcement agency, 4 July 2019.
BABUTA AND OSWALD 13
60. Robin Allen and Dee Masters, ‘Algorithms, Apps & Artificial Intelligence: The
Next Frontier in Discrimination Law’, updated for the Public Law Project session
entitled ‘AI Justice: Artificial Intelligence Decision-Making and the Law’ on 16
October 2018, Cloisters, 2018.
61. See Jennifer Schwartz et al., ‘Trends in the Gender Gap in Violence: Reevaluating
NCVS and Other Evidence’, Criminology (Vol. 47, No. 2, June 2009), pp. 401–25.
62. David P Farrington, ‘Age and Crime’, Crime and Justice (Vol. 7, 1986), pp. 189–250.
For further discussion, see, for example, Stephen D Hart, Christine Michie and
David J Cooke, ‘Precision of Actuarial Risk Assessment Instruments: Evaluating
the “Margins of Error” of Group v. Individual Predictions of Violence’, British
Journal of Psychiatry (Vol. 190, No. S49, 2007), s63; Nancy R Cook and Nina P
Paynter, ‘Performance of Reclassification Statistics in Comparing Risk Prediction
Models’, Biometrical Journal (Vol. 53, No. 2, March 2011), p. 237.
63. Chris Webster, Quazi Haque and Stephen J Hucker, Violence Risk: Assessment and
Management: Advances Through Structured Professional Judgement and Sequential
Redirections, 2nd Edition (Chichester: John Wiley & Sons, 2013), p. xiii; Kevin S
Douglas, Melissa Yeomans and Douglas P Boer, ‘Comparative Validity Analysis of
Multiple Measures of Violence Risk in a Sample of Criminal Offenders’, Criminal
Justice and Behavior (Vol. 32, No. 5, October 2005), pp. 501–02.
64. Laurens Naudts, ‘How Machine Learning Generates Unfair Inequalities and How
Data Protection Instruments May Help in Mitigating Them’, in Ronald Leenes et
al. (eds), Data Protection and Privacy: The Internet of Bodies (Hart, 2018).
BRIEFING PAPER 14
Deployment
65. Julia Angwin et al., ‘Machine Bias’, ProPublica, 23 May 2016, <https://www.
propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing>,
accessed 14 August 2019.
66. Alex Albright, ‘If You Give a Judge a Risk Score: Evidence from Kentucky
Bail Decisions’, The Little Dataset, 15 July 2019, <https://thelittledataset.
com/2019/07/15/if-you-give-a-judge-a-risk-score/>, accessed 14 August 2019.
67. Michael Rovatsos, Brent Mittelstadt and Ansgar Koene, ‘Landscape Summary: Bias
in Algorithmic Decision-Making’, CDEI, 2019.
68. Author’s telephone interview with A1, academic expert in data protection,
privacy and surveillance policy, 28 June 2019.
69. Author’s interview with A2, academic expert in computer science and data
protection, London, 28 June 2019.
70. See Orla Lynskey, ‘Criminal Justice Profiling and EU Data Protection Law:
Precarious Protection from Predictive Policing’, International Journal of Law in
Context (Vol. 15, Special Issue 2, June 2019), pp. 162–76.
71. Author’s telephone interview with L4, representative of UK law enforcement
agency, 9 July 2019.
72. Dencik et al., ‘Data Scores as Governance’.
BABUTA AND OSWALD 15
Bias may also occur in the way that the human decision-maker adheres to
or deviates from the algorithm’s prediction or insight.73 A factor which may
arise during deployment (thus necessitating attention during design) is the
risk of automation bias, the tendency to over-rely on automated outputs
and discount other correct and relevant information.74 Inappropriate
fettering of discretion through use of an algorithm may make the resulting
decision unfair,75 or de facto an automated one.76 As explained by one police
officer interviewed: ‘Officers often disagree with the algorithm. I’d expect and
welcome that challenge. The point where you don’t get that challenge, that’s
when people are putting that professional judgement aside’.77 However,
another officer noted that ‘professional judgement might just be another
word for bias’, explaining that ‘whenever we have to decide an outcome
there’s always an opportunity for bias’.78 It is essential that the correct
balance is struck to ensure due regard is paid to the insights derived from
analytics, without making officers over-reliant on the tool and causing them
to disregard other relevant factors. Adequate training focused on cognitive
bias and fair decision-making would appear essential to ensure officers are
able to consistently achieve the correct balance.
Furthermore, while the risk of cognitive bias is often used to argue in favour
of the statistical approach, a ‘risk score’ is potentially highly prejudicial to the
decision-maker.79 One police respondent referred to the output getting ‘into the
head’ of the officer.80 Quantifications of risk are also liable to misinterpretation
by those not directly involved in the assessment.81 For example, a ‘low-risk’
label could be interpreted to mean that an individual requires no further
monitoring or intervention. Yet such ‘low-risk’ individuals may have specific
needs that should be addressed as part of a bespoke risk-management plan.
Such individuals may then fail to receive the support to prevent them returning
to problematic behaviour.
73. Albright, ‘If You Give a Judge a Risk Score’; Bo Cowgill, ‘The Impact of Algorithms on
Judicial Discretion: Evidence from Regression Discontinuities’, working paper, 2018,
<http://www.columbia.edu/~bc2656/workingpapers.html>, accessed 14 August 2019.
74. Danielle Keats Citron, ‘Technological Due Process’, Washington University Law
Review (Vol. 85, No. 6, 2008).
75. Oswald, ‘Algorithmic-Assisted Decision-Making in the Public Sector’.
76. ‘Data Protection Act 2018 (UK)’, Part 3.
77. Author’s telephone interview with L4, representative of UK law enforcement
agency, 9 July 2019.
78. Author’s telephone interview with L1, representative of UK law enforcement
agency, 1 July 2019.
79. David J Cooke, ‘More Prejudicial than Probative’, Journal of the Law Society of
Scotland (Vol. 55, No. 1, 2010), pp. 20–23; David J Cooke and Christine Michie,
‘Violence Risk Assessment: From Prediction to Understanding – Or From What?
To Why?’, in Caroline Logan and Lorraine Johnstone (eds), Managing Clinical
Risk: A Guide to Effective Practice (Abingdon: Routledge, 2013), p. 10.
80. Authors’ telephone interview with L6, representative of UK law enforcement
agency, 10 July 2019.
81. Cooke, ‘More Prejudicial than Probative’.
BRIEFING PAPER 16
OVERSIGHT
IMPLEMENTATION
Design
Testing
Incomplete access to Lack of influence
• Unclear evaluation criteria
data and information • Misleading accuracy rates
• Model shrinkage
• High false positives/negatives
• Lack of access to training data
• Testing limited to ‘predictive accuracy’
Deployment
Limited to data
science oversight
EMERGING FINDINGS
Interviews conducted to date evidence a desire for clearer national guidance
and leadership in the area of data analytics, and widespread recognition
and appreciation of the need for legality, consistency, scientific validity and
oversight. It is also apparent that systematic investigation of claimed benefits
and drawbacks is required before moving ahead with full-scale deployment
of new technology.82 As one law enforcement practitioner commented,
‘there’s as much value in understanding what doesn’t work, as what does’,83
but to achieve this, controlled space for experimentation is required,
recognising that ‘policing is about dealing with complexity, ambiguity and
inconsistency’.84
Lessons can be learned from recent trials of live facial recognition, particularly
concerning the need to demonstrate an explicit legal basis for the use of new
technology, the need for clearer guidance relating to trials and evaluation, and
the importance of meaningful public engagement during the development
and testing phase. The development of a draft Code of Practice provides
an opportunity, not only to consider bias, but to improve understanding of
the application of data analytics in different contexts, and of methods of
assessing potential benefits and intrusions. It will be incumbent on users to
evidence such assessments when determining whether use of a particular
tool can be deemed ‘necessary’, in order to decide whether there are less
intrusive means of achieving the same policing aim.85
Any new code of practice for algorithmic tools in policing should establish a
standard process for model design, development, trialling, and deployment,
along with ongoing monitoring and evaluation. It should provide clear
operationally relevant guidelines and complement existing authorised
professional practice and other guidance in a tech-agnostic way.86 Existing
surveillance codes and related inspections were suggested by a number of
interviewees as a potential model. The new code should ensure sufficient
attention is paid to meeting legal and ethical requirements throughout all
stages of the product lifecycle, from project inception through to model
procurement, development and testing, including ongoing tracking and
mitigation of discrimination risk when the tool is deployed operationally,
82. See Albert Meijer and Martijn Wessels, ‘Predictive Policing: Review of Benefits
and Drawbacks’, International Journal of Public Administration (Vol. 42, No. 12,
2019), pp. 1–9.
83. Author’s telephone interview with L7, representative of UK law enforcement
agency, 18 July 2019.
84. Authors’ telephone interview with L8, representative of UK law enforcement
agency, 23 July 2019.
85. Author’s telephone interview with A4, academic expert in human rights and
technology, 11 July 2019.
86. Author’s telephone interview with L1, representative of UK law enforcement
agency, 1 July 2019.
BRIEFING PAPER 18
The Royal United Services Institute (RUSI) is the world’s oldest and the UK’s leading defence and security
think tank. Its mission is to inform, influence and enhance public debate on a safer and more stable world.
RUSI is a research-led institute, producing independent, practical and innovative analysis to address today’s
complex challenges.
Since its foundation in 1831, RUSI has relied on its members to support its activities. Together with revenue
from research, publications and conferences, RUSI has sustained its political independence for 188 years.
The views expressed in this publication are those of the authors, and do not necessarily reflect the views
of RUSI or any other institution.
Published in 2019 by the Royal United Services Institute for Defence and Security Studies. RUSI is a
registered charity (No. 210639).
This work is licensed under a Creative Commons Attribution – Non-Commercial – No-Derivatives 4.0
International Licence. For more information, see <http://creativecommons.org/licenses/by-nc-nd/4.0/>.