Articles 8
Articles 8
Abstract
AI ethics is a field that has emerged as a response to the growing concern regarding the impact of
artificial intelligence (AI). It can be read as a subset of the wider field of digital ethics, which addresses
concerns raised by the development and deployment of new digital technologies, such as AI, big data
analytics and blockchain technologies. The principle aim of this article is to provide a high-level
overview of the field as it stands by introducing basic concepts and sketching approaches and central
themes in AI ethics. The first part introduces basic definitions of the terms i.e. defining ‘AI’ and ‘ethics’,
etc.; the second part explores some predecessors to AI ethics, namely engineering ethics, philosophy
of technology and science and technology studies; the third part discusses three current approaches
to AI ethics namely, principles, processes and ethical consciousness; and finally, the fourth part
discusses central themes in translating ethics in to engineering practice. We conclude by summary and
noting future directions and debates.
1. Introduction
AI ethics is a field that has emerged as a response to the growing concern regarding the impact of
artificial intelligence (AI). It can be read as a subset of the wider field of digital ethics, which addresses
concerns raised by the development and deployment of new digital technologies, such as AI, big data
analytics and blockchain technologies. There is a growing literature that has significantly increased in
the past number of years (2017 - ) and this literature continues to evolve. The principle aim of this
article is to provide a high-level overview of the field as it stands. We do this by introducing basic
concepts and sketching approaches and central themes in AI ethics.
The overview is structured into four parts; the first part introduces basic definitions of the terms i.e.
what is meant by key terms such as ‘AI’ and ‘ethics’; the second part explores some predecessors to
AI ethics, namely engineering ethics, philosophy of technology and science and technology studies;
the third part discusses three current approaches to AI ethics namely, principles, processes and ethical
consciousness; and finally, the fourth part discusses central themes in translating ethics in to
engineering practice. We conclude by summary of the overview and by noting future directions and
debates.
2. Definitions
In this section we will offer and explicate definitions of key terms. As a point of departure, we begin
by defining first ‘digital ethics’ and then ‘AI ethics’. This ordering follows the fact that AI ethics is a
subdiscipline of the broader umbrella of digital ethics. Following this we then move on to defining how
the term ‘digital’ is being used in digital ethics and then expand on how the term ‘AI’ is being used
(similar to how AI ethics falls under digital ethics, AI is shown to fall under ‘digital’). Further, we
explicate how the term ‘ethics’ is being used in this context and expand upon the dominant ethical
philosophies that AI ethics draws upon. This section closes with an exploration of ‘Human Centric AI’,
which we take to be the overarching value framework of AI ethics (Lukowicz 2019).
Turing to AI ethics:
This flows from the definition of digital ethics presented above but is specified for AI. In the case of AI
the psychological refers to the likes of mental autonomy, protection from undue manipulation and
the right to know when one is interacting with a non-human agent; the social refers to the likes of
issues of justice and fairness (both procedural and substantive and; the political refers to impacts on
democratic processes and the economy. Like digital ethics, AI ethics is thus highly interdisciplinary
(Kriebitz and Lutge 2020).
Below we further flesh out this working definition by exploring the key constitutive terms, namely
‘digital’ and ‘ethics’.
2.2 Digital
In the context of digital ethics, ‘digital’ is a reference to emerging technologies that are based on
developments in computer science over the past decade.
Importantly, these technologies are to be thought of together and as interrelated. For example, AI/ML
has been developed and discussed since the mid-twentieth century, however it has grown in
importance and application through advances in computational power and the emergence of large
datasets. As such, big data analytics is a product of AI and big data. Other examples are Federated
Learning, which utilises AI and blockchain (Bonawitz 2019), and the evolution towards Smart Cities ,
which utilse AI, IoT and big data analytics (Burange and Misalkar 2015).
The current epoch is being referred to as the ‘4th industrial revolution’ (Floridi 2014). Where the 1st,
2nd and 3rd industrial revolutions were characterised, respectively, by the use of water and steam
power in the mechanisation of production (circa 1750-1820) , the use of electricity to power mass
production (circa 1870-1920), and the use of electronics and information technologies in the mass
automation of production and processing (circa 1950 -). The 4th is a development upon the third (circa
1990 -), and it is characterized by a fusion of technologies that blur the digital, physical, and biological
spheres (ex. cyberspace, virtual and augmented reality, body-machine interface and robotics)
(Andelka and Mian 2019).
Indeed, two additional themes are i. ubiquitous adoption of these technologies, and ii. futurism.
Where the former is a reference to the increasing use and normalisation of such technologies in
everyday life, government service provision and industry. The latter is a reference to the
In the digital realm algorithms can be expressed in a computer through programming language. In
order to understand this, we must first note that broadly there are two classes of AI algorithms, which
might be termed: static algorithms – traditional programs that perform a fixed sequence of actions,
usually classified as knowledge-based systems (Schreiber et al. 1993; Giarratano and Riley 1998); and
dynamic algorithms – that learn and evolve by interacting with the environment, usually classified as
ML algorithms (Hastie et al. 2009; Sutton and Barto 2018).
We can think of this as an ‘AI continuum’ of epistemological models (Russell and Norvig 2016), with
the current most successful of which is
ML algorithms - a type of program with the ability to learn without explicit programming, and
can change when exposed to a new environment or information
Supervised learning: a program is trained on available and processed data, where specified
inputs are used to predict outputs
Unsupervised learning: the goal of a program is uncover a hidden structure in the data,
thereby ‘discovering’ previously unknown patterns
Reinforcement learning: the goal of the program is to make decisions, achieved through an
iterative trial and error process that is mediated by a reward/penalisation mechanism for
decisions chosen in the development
Machine learning applications in financial services can provide examples of these: Suppose a database
of financial reports is available, if some of them have been historically labelled as positive and
negative, we can leverage this to automatically tag future documents. This can be accomplished by
training an algorithm in a Supervised fashion. If these documents were unstructured, and spotting
relations or topics is the goal (political events, economic data, etc.), an algorithm trained in an
Unsupervised manner can help uncover these hidden structures. Also, these documents can
characterise the current state of the capital markets. Using that, an algorithm can decide which actions
should be taken in order to maximize profits, hedge against certain risks, etc. By interacting and
gaining feedback from the environment (Markets), an algorithm can Reinforce some behaviours so as
to avoid future losses or inaccurate decisions.
In addition to the above mentioned subdivision of ML there are further and disruptive forms of more
advanced ML systems that are making the resolution of previous problems cheaper, faster and more
scalable, like Deep Learning (Goodfellow et al. 2016), Adversarial Learning (Huang et al. 2011),
Transfer and Meta Learning (Andrychowicz et al. 2016; Devlin et al. 2018).
Ethics: the rational and systematic study of the standards of what is right and wrong
Morality: the commonly used term for notions of good and bad
Morality is closely associated with notions of virtue, which concerns the internal state of a person (we
can think of a moral/virtuous act as one that is rooted in the internal state of the said person).
Importantly the law is enforceable i.e. there is a coercive core (usually by the executive branch of the
government). In philosophical ethics we think of this in terms of external constraints, which compares
to the internal states described by notions of virtue.
These concepts are highly related and in natural language use the terms are often interchangeable
and synonymous. Indeed, a law may be considered ethical or unethical; breaking a law may be
considered (im)moral or simply procedural: for example, a person may cross an empty road in a non-
designated place i.e. ‘jaywalk’, however this would not be considered immoral, whereas,
contrastingly, stealing would be a clear example of an immoral breaking of the law. Additionally,
something may be considered immoral, and unethical, but not illegal (ex. infidelity, lying in casual
opinion, etc.). As such, the definitions offered should be understood as notions that allow AI ethics
discussions to be structured rather than stable and inflexible concepts.
There are numerous interpretations of utilitarianism (ex. ‘act’ and ‘rule’ utilitarianism) as well as
questions regarding how the terms ‘pleasure’ and ‘displeasure’ are to be understood and quantified.
Putting these concerns aside, the operative concept is that ethics is about weighing the consequences
of actions. One domain where this approach to ethics dominates is that of the justification of
government policy and decision making (ex. policy regarding health care is justified by appeal to
maximization of health outcomes for citizens, economic policy is justified through maximisation of
Gross Domestic Product, etc.).
Here, the ethical framework is such that a series of ‘rights’ i.e. that which is referred to as entitlements
in the definition above, are conferred to a person simply by belonging to a class. Where class is
understood as a generic category of identity. Two central examples of rights are human and civil rights.
Human rights are rights entitled to anyone in the class ‘human’, and civil rights are rights
entitled/conferred upon any citizen of a particular jurisdiction. Whereas human rights are considered
inviolable and fundamental simply by virtue of being a human, civil rights are conferred to members
of the political community.
Virtue: development of the character of an individual and actions that result as a consequence
of good character
Virtue ethics (also known as natural ethics) is a classical position that is rooted in pre-enlightenment
Aristotelianism. It is an approach to ethics that emphasises character development – it is closely
associated with ‘perfectionism’, where a person develops over time towards an idealised notion of
the perfect Self (often described in the religious terms of becoming god-like or being in union with the
divine). Good character is understood in terms of values (read, virtues) such as honesty, self-control,
integrity, courage, generosity and fairness.
2.6 Human-Centric AI
In all three ethical approaches, the central subject of concern is the human being i.e. ‘persons’. As
such, it is necessary to offer a working definition of human, which we take to be a rational animal.
There are numerous ways in which this definition can be understood, however, for the present
purpose it is sufficient for us to emphasise that the class ‘human’ is principally defined in terms of
possession of the rational faculty. Thus, humans share all other characteristics with animals
(movement, reproduction, etc.) but are differentiated into a separate class by reason. Reason itself
requires fleshing out and can be thought of in broader terms that include, ‘freedom’, ‘volition’,
‘intentionality’ and ‘agency’. These terms themselves are hotly debated within the philosophical
literature, however for our purposes, we can read them all as referring to reason as an ability to
meaningfully make choices i.e. agency and autonomy.
As such, the human is defined as an ‘agent’ and ‘meaningful choice’ is understood as self-conscious
decision-making. Following from this we can define ‘dignity’ as respect for the moral status of human
beings as rational agents making meaningful choices i.e. existing autonomously. In the context of AI
ethics, autonomy can be subdivided into mental and physical autonomy, where mental autonomy
concerns respect for a person’s deliberative faculties and processes (for example, the right not to be
manipulated consciousness or subconsciously), and, where physical autonomy concerns respect for a
person’s body and choices over their own body.
Definition: the development and deployment of AI systems that respect human dignity and
autonomy
Indeed, Human Centric AI can be thought of in positive terms i.e. that automated systems should be
developed and deployed for the betterment of humankind, to advance well-being, human dignity and
human flourishing. When a system reflects this overarching value framework then it can be thought
of as trustworthy i.e. Trustworthy AI – as discussed by the European Commission’s ‘Ethics Guidelines
for Trustworthy AI’ (2019). In section 5.6 below we expand upon themes that fall under
trustworthiness in the context of AI systems.
3. Predecessors to AI Ethics
AI ethics is an emergent field that is still in its nascent phase. However, there are a number of
disciplines that have long traditions and literature from which AI ethics draws and can be seen as, in
various ways, a continuation of. The three bodies of literature most relevant are, i. engineering ethics,
ii. philosophy of technology, and iii. science and technology studies. Below we discuss each in turn.
Engineering is a term that refers to structuring, design, and building. It is perhaps the most inherently
practical discipline within the broadly construed sciences (often referred to as an applied science). As
such, it is natural that there already exists a body of literature discussing and debating the social and
environmental impact of engineering. Although engineering encompasses numerous subdisciplines
(such as civil, mechanical, computer and chemical engineering), all of which have specific societal
impacts, the field has matured to the point where non-subdiscipline specific, general community-
based engineering codes have emerged.
For example, the UK based Royal Academy of Engineers, which is a membership by nomination and
selection community of fellows, has a published ‘Engineering Ethics’ guideline ‘Code of Practice’
(2020). The guideline is divided into two parts: first, a ‘Statement of Ethical Principles’ (namely, i.
honesty and integrity, ii. respect for life, law, the environment and public good, iii. accuracy and rigor,
and, iv. leadership and communication); and second, ‘Engineering Ethics in Practice’, which is the
fleshing out of the statement of ethical principles through granular real-world case studies. There are
other similar examples of this (see, IEEE ‘Code of Ethics’ (2020)).
The philosophy of technology emerges circa 1920s and can be read as continuing till today (major
figures are Martin Heidegger (d. 1976), Herbert Marcuse (d. 1979), and Jurgen Habermas). It is
differentiated from the philosophy of science, which has a longer legacy in the history of philosophy
and is concerned with method and knowledge. Contrastingly, philosophy of technology emerges as a
result of technological innovation (where technology ‘tekhnē’ means art or craft); indeed, it concerns
the applications and uses of discoveries in science. The principal focus is in appraisal of how technology
affects the human condition and whether the technology is neutral or value laden (ex. exploration of
whether nuclear technology is inherently good or bad, or whether it is dependent on the deployment
of the technology in particular contexts) (Balabanian 2006). From a historical perspective, the
philosophy of technology can be read as a response to the overtly optimistic attitudes of the
enlightenment and ‘positivism’, as well as the Post-WWII world which had an ever-present nuclear
treat and a 1960s-counter culture that was typified by ‘social conscious’. Indeed, it challenges the idea
that there is a necessary connection between scientific discovery and scientific progress, and that this
progress includes and is extended to society. The literature is typically negative, highlighting the
dangers, risks and loss of meaning through adoption of new technologies and increased
technocratisation with respect to the ordering principles of society. Key themes are automation,
alienation, destruction and loss of connection to nature, uniformity, shallow consumption, and
excessive rationalisation.
In addition to noting the general negative critique of technology in the philosophy of technology, it is
also noteworthy that the perspective of the engineers and scientists themselves i.e. the practitioners,
are missing. This negativity also ignores empirical evidence that those who have a less negative view
could cite, namely that technologies such as vaccinations and other medical equipment (ex.
pacemakers etc.) have tangibly and demonstrably decreased mortality rates. This ‘turn to evidence’ is
discussed in the next subsection.
The philosophy of technology can be criticised from several points. Two of these are that it is i. typified
by moral panic, and ii. that it is non-empirical. Considering moral panic, namely the phenomena of an
acute reaction to a shocking/dramatic event, typified by knee-jerk reactionaryism - the negative
critique and commentary is read as excessive and blind to the benefits that such technologies have
conferred to humanity. Indeed, it is read as highly politicised, with the philosophy of technology being
instrumentalised as a polemical force of political intervention rather than as a considered investigation
into the nature and impact of technological innovation.
The accusation that the philosophy of technology is non-empirical is motivated by the evolution and
development of the social ‘sciences’, where sciences are placed in quotation marks due to the
increased incorporation and methodological approach of the empirical (natural) sciences in the
humanities (such as sociology and anthropology). Increasingly sophisticated mechanisms of empirical
investigation - that survey, test hypothesis and observe through data analysis (c.f. natural sciences) –
were brought to questions that the philosophers of technology were commentating on.
This empirical turn gave birth to science and technologies studies (circa 1980s -) (Franssen 2016).
Major figures include Bruno Latour and Andrew Light, Donna Haraway. Science and technology studies
problematised the overwhelmingly negative critique of technology by pointing out that technology
7
3.4 Conclusion
In the previous section we presented a number of definitions, in this section we have presented
predecessor disciples. These definitions and predecessor disciplines will inform what is meant and
what is being drawn upon, in the burgeoning field of AI ethics. In the section below we sketch the
current landscape of AI ethics through a high-level analysis of the main approaches. These are i. the
principles approach, ii. ethical-by-design/processes approach, and iii. ethical consciousness approach.
4.1 Principles
The most vocal response to the harms and risks of new digital technologies, in particular AI, has been
to call for guidelines that inform and direct the use and development of this technologies. Attempts
to delimit these principles fall into three main categories, i. abstract first-principles, ii. development
and application of legislative standards and norms, and iii. to make analogy with bio/medical ethics.
Below we sketch these.
4.1.2 Legislation
The most direct way in which to approach AI ethics is to ensure that the technologies are developed
and deployed in a lawful manner. Indeed, legal compliance is a clear and objective standard by which
to judge and evaluate ethics (where lawfulness can be read as a necessary but insufficient condition
of ethics). However, straightforward recourse to the law is not possible. This is because there are a
number of nuanced concerns. First, the question is posed as to whether it is necessary to create new
laws or to update and ensure application of existing one European Commission 2020). Indeed, one
approach that can be taken is to allow a body of case law to emerge and derive standards, and if need
be legislate accordingly based on this i.e. a ‘bottom-up’ approach (see Committee on Standards in
Public Life 2020). Second, there is the question of whether legislation is appropriate at all, with other
options possible, such as self-regulation and/or a standards body (Simon 2019; Lauterbach 2019).
Third, there is the issue of jurisdiction, where technological development and deployment are
obscured through internationalism, which does not respect or easily lend to jurisdictional oversight
(ex. nation states and international unions (European Union, African Union, etc.)). This also allays itself
the problem of enforcing the law. Fourth, in the common law tradition (typically found in the
anglophone), based on case and precedent, the statutes accommodate ambiguity and contradiction
(relying on judicial discernment). This is a problem with respect to automated systems, where
increasingly there is a call to automate regulatory compliance via expressing laws in codes/protocols
(the ambiguity and contradiction accommodated by common law does not translate in this context).
Regarding this specific concern, continental law (exemplified by EU’s legislative agenda) typified
through a top-down (first principles) legal philosophy is less likely to have this problem. Finally, and
more specifically, in the context of AI there is the question regarding the legal status of an algorithm
(ex. will algorithms follow how companies have rights and obligations, and will AI systems will have
artificial personhood status) (Treleaven et al. 2019), and how the questions of legal culpability, which
rely on judgements of intent, can be formulated in the context of AI systems (Rajakishore and Sahu
2020; Vetrò et al. 2019).
4.2 Processes
A second approach to AI ethics is to address risk and harm that can result because of design issues and
lack of appropriate governance.
4.2.1 Ethical-by-Design
An ethical by design approach is a commitment to building systems ethically and, in the hope, that
harm can be prevented. There are several approaches to ethical-by-design. First, via co-design, which
is a reference to interdisciplinarity in the design processes. The idea here is that AI engineers may not
be best placed to understand and discern the ethical dimension and potential impact of the
technology and as such experts from anthropology, sociology, philosophy, psychology, law, etc. i.e.
‘ethicists’, can be integrated into the team at the development stage. Second, by having clear
principles, laws, standards and guidelines, with which to structure and judge design. As noted above,
there is considerable ambiguity and lack of consensus in the field of digital regulation (perhaps with
the exception of GDPR) and standards, thus making it difficult to establish best practices in the domain
of translating principles into engineering practice. Finally, and as a corollary to the previous point,
from a design perspective, implementation of ethical principles will need to be balanced and ‘traded-
off’. For example, in the design, emphasis on transparency and openness may come at the cost of
privacy. These judgements need to be justified, articulated and expressed in different forms in various
contexts. With this practical dimension and necessity of trade-offs in mind, several practical manuals
have been published, all of which note these trade-offs and discuss ethical evaluation (including
interdisciplinarity) in the design, development and deployment phases, respectively (IEEE 2017;
Whittlestone et al.; 2019; Leslie 2019).
4.2.2 Governance
Within the processes approach to AI ethics questions of governance are emerging. More generally,
with respect to novel digital technologies governance can be divided into two broad streams, namely
technical and non-technical (with ethical-by-design falling into the former category). Technical
governance concerns systems and processes that render the activity of the technology itself
accountable and transparent – this includes justifying what design choices are made and ensuring the
system is accessible. Non-technical governance concerns systems and processes that focus on
allocating decision makers, providing appropriate training and education (in the context of new digital
technologies such as AI, education and training will require continuous updating), and keeping the
human-in-the-loop with respect to how automated decisions are used while respecting human rights
(often referred to in the context of ‘human centric AI’ (Lukowicz 2019).
Falling under governance is the growing literature on auditing and impact assessments. Auditing and
impact assessments involve the creation of metrics for tracing and tracking decisions, making the
technologies accessible for verification and accountability. The most well-established form of this is
Data Protection Impact Assessments (DPIA) (Information Commissioner’s Office 2019). To date more
general ‘data ethics’ canvases and process frameworks have emerged (DCMS 2018; Open Data
Institute 2020), and impact assessments specific to AI are being called for and developed (European
Commission 2020; Jordan et al. 2019; Babuta and Oswald 2020).
10
4.4 Conclusion
The above approaches can be thought of in terms of i. the theoretical and abstract, ii. the practical
and process, and iii. culture and society, and in turn can be thought of as all necessary in the
development of a mature AI ethics i.e. one that is truly reflective of the nuances and the inherent
complexity found in all forms of applied ethics. This also challenges any attempt to silo questions and
responsibilities; AI ethics is not exclusively in the purview of philosophers or lawyers or sociologists or
engineers, etc. rather it is inherently interdisciplinary. As such, the call for interdisciplinary must be
followed with development of methods and structures by which they can be fulfilled. It is likely that
this aligns with the call for training and education (noted above in the context of discussions regarding
governance); as such an increase in ‘literacy’ is crucial in informing all relevant stakeholders and
society at large and will be facilitated through a holistic education and training agenda (touching upon
and integrating ethics, policy and engineering).
● Impact on human agency: This touches on the impact on individuals and, in particular, mental
autonomy. For example, consider whether a system directly or indirectly diminishes the
deliberative/rational capacity of humans (ex. cognitive shifts in attention spans). Another
issue is consent, where respect for human agency would entail meaningful and informed
11
The above may be read as a re-statement of the ethical imperative that grounds Human Centric AI,
i.e. that automated systems should be developed and deployed for the betterment of humankind, to
advance well-being (or at least not adversely affect it), human dignity and human flourishing. Indeed,
although the discussion is presented in terms of mitigating risks and potential harms, it is important
to bear in mind the considerable benefits of AI for people and society (Taddeo and Floridi 2018).
5.3 Safety
This theme is based on the ethical principle of preventing harm, where harm is defined in terms of
adverse effects on human well being i.e. the psychological, social and environmental human
wellbeing.
Here the approach is one of identifying risks and then mitigating for them: crucially the approach is
preventative. Key themes here are:
● Robustness: Systems should be robust against adversarial attacks i.e. hacking. Here resilience
is important and that there are measures to stop/resist exploitation of a system (ex. data
poisoning, model leakage).
● Malicious use: A system may have been developed for one use and then be appropriated
and/or modified for another, malicious, use i.e. dual use (ex. the weaponization of delivery
drones).
● Reliability and Reproducibility: Reliability concerns the system working within the framework
of why it was developed and deployed, whereas reproducibility concerns consistent behaviour
when given the same set of inputs and under the same conditions. In the context of robustness
this is important because a system that is unreliable and does not reproduce results will lead
to untrustworthiness in the system.
● Fallback Plans and Unknown risks: A concern for robustness is to address known risks (such as
those cited above, i.e. security, malicious use, reliability and reproducibility) and unknown
risks. With respect to the former, safeguards can be put in that specifically monitor and track
usage and/or metrics of known risks and put in place stops or other mechanisms that would
mitigate this risk. With respect to the latter, it is not possible to fully anticipate risk and as
such, mechanisms can be put into place to mitigate this (ex. fallback mechanisms, automatic
stops – statistical or rule-based, metrics, periodic request for human operators to continue
operating, etc).
5.4 Privacy
This theme is based on the public and political demand to respect a human’s personal information.
This relies on a distinction between the private-personal and the public-political/communal sphere,
where the former is seen as demanding a higher level of respect for privacy than the latter (Mittelstadt
2017). Informed consent is crucial here, where people are informed and updated regarding the
storage and use of their data. Further, there are debates concerning the value derived from personal
data and the distribution of financial benefits derived thereof (ex. mass data driven business models).
Additionally, privacy has emerged as a political concern, with mass surveillance and personal data
being used to target and engage in recommendation and manipulation (for both political and
economic ends).
12
5.5 Transparency
This theme is based on the principle of openness, which is crucial to establishing trust and
accountability. Transparency can be thought of with respect to what decisions are being made
regarding how the AI system is used and with respect to how the system comes to its decisions. The
former touches on governance (which is also expanded upon in the accountability section below),
whereas the latter concerns explainability of automated decision makers. As such key themes are:
● Explainability: Being able to explain how a system has come to a decision (c.f. black-box
problem) and making that decision explicable to various stakeholders i.e. explicability will
depend on the technical knowledge of the person, what role they play in the development
and deployment of the system and what kind of end-user they are. Furthermore, there are a
host of technical requirements and tools that may be grouped under explainability, namely,
accuracy, traceability, tracking, general (global/model) and specific (local/data point)
explanations.
● Communication: In addition to explicability i.e. communication, of automated decisions, there
is also the concern for communicating the capabilities and purposes of the system to those
both directly and indirectly impacted (Binns 2018). One crucial dimension is that in cases
where a system may be mimicking human subjectivity (ex. a chatbot) it should be
communicated clearly that the user is interacting with an AI system.
5.6 Fairness
This theme is based on the ethical principle of human equality. Fairness falls under debates about
justice and is hotly contended. A central question is what definition(s) of fairness/justice to commit to
i.e. there are mutually exclusive theories of fairness such as corrective, distributive, procedural,
substantive, comparative, … etc. The question is also raised as to the scope or remit within which
notions of fairness/justice are being discussed i.e. fairness in the context of political communities
(citizenship rights) and/or universal human concerns, and, if appropriate, how to define demographics
i.e. gender, nationality, race, socio-economic background, … etc. Key themes here are:
13
5.7 Accountability
Ethical AI is a branch of applied ethics and, as such, is inherently concerned with how AI systems
impact human beings. How the systems are developed, the processes, logic of decision making, the
allocation of duties with respect to who makes decisions, and how and to what extent, where impacts,
risks and harms gauged and measured. All of this falls under the remit of accountability, and, as the
previous sentence indicates accountability relates to knowing who had made decisions, how those
decisions were made and what systems or tools were put in place to measure and track i.e.
governance. Finally, accountability is central for the possibility of redress and assigning legal liability.
Key themes here are:
● Keeping it Human
Crucial to accountability is ensuring that there are robust human oversight mechanisms, this is based
on the principle, and current legal standing (see section 4.1), that humans are ultimately the
accountable and thereby responsible for harms that may result from AI systems. Within the literature,
there is a growing discourse regarding keeping the ‘human-in-the-loop’ (Wang 2019), which is
discussed in two ways, firstly as ensuring that there is human intervention in the decision processes
of automated systems and secondly in terms of human oversight regarding automated decisions i.e.
in the context of ‘decision support systems’. With respect to the latter, a ‘semi-automated decision’
scheme can be thought of, where a system generates results, directions and recommendations (ex.
whether to hire someone, or reject a loan application) and is followed by human review in order to
affirm or reject the recommendation (Information Commissioner’s Office 2020). In these cases, checks
should be in place to assure that the human review does not become a rubber stamp exercise
rendering the decisions effectively solely automated. Ensuring human responsibility also entails
mapping of duties and risks to responsibilities and roles within an institution.
● Algorithmic Impact Assessments: This relates to direct mechanisms by which to assess and
thereby put in measures to mitigate potential harms of AI systems. We can divide this into
two, namely, i. impact assessments and ii. auditing of technology (Ada-Lovelace Institute and
DataKind UK 2020).
i. Impact Assessments can range from assessments of fundamental rights,
psychological and social well being (ex. social cohesion), citizen rights,
democracy, economic and environmental impacts.
ii. Auditing can be directed to the technology itself, and focus on fairness (ex.
tracking bias metrics), explainability (ex. providing global and local
explanations of models and individual decisions) and robustness (ex. testing
how resilient a system is to hacking).
It is important that these are conducted in such a way as to facilitate inspection (perhaps even
independently conducted). Moreover, clear documentation is necessary. This includes
documenting any trade-offs and the methodology and logic behind trade-off choices.
6. Conclusion
In this high-level overview we have introduced basic definitions of terms such as ‘AI’ and ‘ethics’.
Following this we explored some predecessors to AI ethics, namely engineering ethics, philosophy of
technology and science and technology studies/ We then discussed three current approaches to AI
ethics namely, principles, processes and ethical consciousness. Turning to translating AI ethics into
engineering practice we surveyed the themes of human centric AI, safety, transparency, fairness and
privacy.
Themes and developments in AI ethics that we anticipate:
14
7. References
1. Allen C, Varner G, & Zinser J. (2000). Prolegomena to any future artificial moral agent.
Journal of Experimental & Theoretical Artificial Intelligence, 12(3), 251-261.
2. Andrychowicz M, Denil M, Gomez S., Hoffman MW, Pfau D, Schaul T, & De Freitas N. (2016).
Learning to learn by gradient descent by gradient descent. In Advances in neural information
processing systems (pp. 3981-3989).
3. Artificial Intelligence (AI) in healthcare and Research. (2018). Nuffield Council on BioEthics.
4. Artificial Intelligence and Public Standards: report. (2020). Committee on Standards in Public
Life, Chair, Lord Evans of Weardale KCB DL.
5. Arvan M. (2018). Mental time-travel, semantic flexibility, and AI ethics. AI & SOCIETY, 1-20.
6. Awad E, Dsouza S, Kim R, Schulz K, Henrich J, Shariff A, & Rahwan I. (2018). The moral
machine experiment. Nature, 563(7729), 59-64.
7. Babuta A, Oswald M. (2020). Data Analytics and Algorithms in Policing in England and Wales:
Towards A New Policy Framework. Royal United Services Institute.
8. Balabanian N. (2006). On the presumed neutrality of technology. IEEE Technology and
Society Magazine, 25(4), 15-25.
9. Bellamy RK, Dey K, Hind M, Hoffman SC., Houde S, Kannan K, ... & Nagar S. (2018). AI
Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted
algorithmic bias. arXiv preprint arXiv:1810.01943.
10. Binns R. (2018). Algorithmic accountability and public reason. Philosophy & Technology,
31(4), 543-556.
11. Bonawitz K, Eichner H, Grieskamp W, Huba D, Ingerman A, Ivanov V, & Van Overveldt T.
(2019). Towards federated learning at scale: System design. arXiv preprint arXiv:1902.01046.
12. Burange AW, & Misalkar HD. (2015, March). Review of Internet of Things in development of
smart cities with data management & privacy. In 2015 International Conference on Advances
in Computer Engineering and Applications (pp. 189-195). IEEE.
13. Char DS, Shah NH, & Magnus D. (2018). Implementing machine learning in health care—
addressing ethical challenges. The New England journal of medicine, 378(11), 981.
15
16
17
18