0% found this document useful (0 votes)
31 views33 pages

How To Survive A Robot Invasion Rights Responsibil

Uploaded by

krafes.soukaina
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
31 views33 pages

How To Survive A Robot Invasion Rights Responsibil

Uploaded by

krafes.soukaina
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 33

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/336380084

How to Survive a Robot Invasion: Rights,


Responsibility, and AI

Book · October 2019


DOI: 10.4324/9780429427862

CITATIONS READS

18 958

1 author:

David J. Gunkel
Northern Illinois University
116 PUBLICATIONS 2,365 CITATIONS

SEE PROFILE

All content following this page was uploaded by David J. Gunkel on 10 May 2021.

The user has requested enhancement of the downloaded file.


How to Survive a Robot
Invasion

In this short introduction, David J. Gunkel examines the shifting world of


artificial intelligence, mapping it onto everyday twenty-first century life and
probing the consequences of this ever-growing industry and movement.
The book investigates the significance and consequences of the robot
invasion in an effort to map the increasingly complicated social terrain of
the twenty-first century. Whether we recognize it as such or not, we are in
the midst of a robot invasion. What matters most in the face of this machine
incursion is not resistance but how we decide to make sense of and respond
to the social opportunities and challenges that autonomous machines make
available.
How to Survive a Robot Invasion is a fascinating and accessible volume
for students and researchers of new media, philosophy of technology, and
their many related fields. It aims both to assist readers’ efforts to understand
a changing world and to provide readers with the critical insight necessary
for grappling with our science fiction-like future.

David J. Gunkel is an award-winning educator and scholar, specializing in


the philosophy of technology. He is the author of nine books, including The
Machine Question: Critical Perspectives on AI, Robots, and Ethics (2012);
Robot Rights (2018); and An Introduction to Communication and AI (2019).
He currently holds the position of Distinguished Teaching Professor in the
Department of Communication at Northern Illinois University, USA. More
info at http://gunkelweb.com.
How to Survive a Robot
Invasion
Rights, Responsibility, and AI

David J. Gunkel
First published 2020
by Routledge
52 Vanderbilt Avenue, New York, NY 10017
and by Routledge
2 Park Square, Milton Park, Abingdon, Oxon, OX14 4RN
Routledge is an imprint of the Taylor & Francis Group, an informa business
© 2020 Taylor & Francis
The right of David J. Gunkel to be identified as author of this work
has been asserted by him in accordance with sections 77 and 78 of the
Copyright, Designs and Patents Act 1988.
All rights reserved. No part of this book may be reprinted or reproduced or
utilised in any form or by any electronic, mechanical, or other means, now
known or hereafter invented, including photocopying and recording, or in
any information storage or retrieval system, without permission in writing
from the publishers.
Trademark notice: Product or corporate names may be trademarks or
registered trademarks, and are used only for identification and explanation
without intent to infringe.
Library of Congress Cataloging-in-Publication Data
A catalog record for this book has been requested
ISBN: 978-1-138-37071-5 (hbk)
ISBN: 978-0-429-42786-2 (ebk)
Typeset in Times New Roman
by Apex CoVantage, LLC
Contents

List of Figures vii


Preface viii

1 Introduction 1
1.1 Robots 1
1.1.1 Science Fiction 2
1.1.2 Science Fact 5
1.2 The Invasion 7
1.2.1 Robots at Work 7
1.2.2 Robots at Home 10
1.3 The Program 11

2 Default Settings 16
2.1 Technological Things 17
2.1.1 The Instrumental Theory of Technology 18
2.1.2 Standard Operating Presumptions 20
2.1.3 Mistakes and Errors 20
2.2 The Right(s) Stuff 23
2.2.1 Substantive Stuff 24
2.2.2 Terminological Troubles 26
2.2.3 Epistemological Exigencies 27
2.2.4 Outcome and Results 29
2.3 Conclusion 30

3 The New Normal 33


3.1 Natural Language Processing 33
3.1.1 The Imitation Game 34
3.1.2 Chatbots 35
vi Contents
3.2 Machine Learning 38
3.2.1 AlphaGo 39
3.2.2 Tay 41
3.2.3 Responsibility Gaps 43
3.3 Social Robots 44
3.3.1 Jibo . . . for Example 44
3.3.2 Social Consequences 47
3.4 Outcomes 49

4 Responses 53
4.1 Slavery 2.0 53
4.1.1 Advantages 54
4.1.2 Disadvantages 57
4.2 Machine Ethics 61
4.2.1 Advantages 64
4.2.2 Disadvantages 65
4.3 Joint Agency 67
4.3.1 Many Hands 68
4.3.2 Advantages and Disadvantages 69
4.4 Conclusion 70

Index 75
Figures

1.1 A scene from the play R.U.R., showing three robots. 2


1.2 Theatrical poster for the film Forbidden Planet (1956) and
featuring the imposing figure of Robby the Robot. 3
1.3 Industrial robot for metal die casting. KUKA Roboter GmbH,
Augsburg, Germany.  8
1.4 The social robot Jibo.  10
2.1 The moral situation.  21
2.2 Artist’s rendition of John Searle’s Chinese Room.  28
3.1 The game of imitation.  35
3.2 AIML code for an ELIZA-type chatbot.  38
3.3 Screenshot of one of the racist tweets posted by Tay.  41
3.4 Jibo and other social robots occupy an ambivalent social
position that is situated somewhere in between “what” are
mere things and those other individuals “who” have
independent status. 45
3.5 US Army iRobot Packbot.  49
4.1 Scanned pages from Mechanix Illustrated, January 1957.  56
4.2 Sophia from Hanson Robotics.  60
4.3 MQ-9 Reaper US military unmanned aerial vehicle (UAV).  62
4.4 Autonomous Waymo Chrysler Pacifica Hybrid minivan
undergoing testing in Los Altos, California. 65
4.5 The spectrum of possible responses to the robot invasion.  71
Preface

This book was not so much written as it was assembled. It has been fabri-
cated by drawing on and remixing material from a number of presentations
and texts that I have given and produced over the past several years.
The basic idea and trajectory for the project was devised and initially
tested in a presentation titled “How to Survive the Robot Apocalypse.” The
talk was first given in September of 2015 during Huskie Hack, the annual
Northern Illinois University (NIU) hackathon. Since that time, versions of it
have been performed at other events and venues. In 2016, it was featured in
NIU STEM Café and Teen STEM Café community-outreach events. It was
delivered during invited talks at the University of Virginia, the University
of Vienna, Pennsylvania State University, and the Universidade Federal do
Piauí in Teresina, Brazil. In 2017, a modified version was presented to the
College Endowment Association in Milwaukee, Wisconsin. In 2018, it was
delivered at the Media Ethics Initiative of the Moody College of Communi-
cation at University of Texas, the Faculty Summer Institute at the University
of Illinois, and the Morals and the Machine Workshop at the Edinburgh
Centre for Robotics. And in April of 2019, it was the featured keynote at the
Student Technology, Arts & Research Symposium (STARS) conference at
the University of Illinois, Springfield.
Published transcriptions of the talk have been developed for and have
appeared in several outlets. Although none of these texts have been repro-
duced here in their entirety, they have furnished bits and pieces for the
re-assemblage. These publications include the following: “The Other Ques-
tion: Socialbots and the Question of Ethics,” published in Robert W. Gehl
and Maria Bakardjieva (Eds.), Socialbots and Their Friends: Digital Media
and the Automation of Sociality (New York: Routledge, 2017); “Other
Things: AI, Robots and Society,” published in Zizi Papacharissi (Ed.), A
Networked Self: Human Augmentics, Artificial Intelligence, Sentience (Lon-
don: Routledge, 2018); and “Mind the Gap: Responsible Robotics and the
Preface ix
Problem of Responsibility,” published in the journal Ethics and Information
Technology, July 2017.
Because of these opportunities to present and test the material with audi-
ences from all over the world, the content that is presented here is as much
a product of my own effort as it is a result of the insightful questions, com-
ments, and feedback I have received from others. Even though it is impossible
to identify and name everyone who has contributed to the mix, I do want to
acknowledge a number of individuals who helped make this work possible:
Maria Bakardjieva (University of Calgary), Janet Brown (College Endow-
ment Association), Mark Coeckelbergh (University of Vienna), Keenan E.
Dungey (University of Illinois, Springfield), Judith Dymond (Northern Illi-
nois University), Robert W. Gehl (University of Utah), J. Michael Herrmann
(University of Edinburgh), Deborah Johnson (University of Virginia), Jef-
frey Nealon (Pennsylvania State University), Zizi Papacharissi (University
of Illinois), Stephanie Richter (Northern Illinois University), Tracy Rog-
ers (Northern Illinois University), Gustavo Said (Universidade Federal do
Piauí), and Scott R. Stroud (University of Texas).
One final note concerning structure and method of access: because the
book is derived from a presentation, which is something that takes place
within a temporal sequence, the text is probably best understood and enjoyed
when read in a linear fashion. I am well aware that this is a “big ask,” espe-
cially for media consumers who are more accustomed to the experience
of random access to short textual bursts posted on social media and other
digital platforms. Although the book can be read in fragments and in any
order you like (which is one of the main advantages of reading a text over
watching a presentation or performance), the material has been designed
to develop and unfold in sequence, like the narrative of a story. This is
one of the reasons why the book is deliberately short and concise. Asking
readers—especially readers whose daily experience with reading involves
accessing digital content on mobile devices—to have the patience to sit
down and read 300+ pages start to finish may simply be too much. But ask-
ing the same for a book that comes in at just 100 pages is less imposing and
more realistic. Whether I have met this challenge and provided something
that is valuable and worth the effort is for you to decide.
1 Introduction

Whether we recognize it or not, we are in the midst of a robot invasion.


The machines are now everywhere and doing virtually everything. We chat
with them online, we play with them in digital games, we collaborate with
them at work, and we rely on their capabilities to manage all aspects of our
increasingly complex digital lives. Consequently, the “robot invasion” is not
something that will transpire as we have imagined it in our science fiction,
with a marauding army of evil-minded androids either descending from the
heavens or rising up in revolt against their human masters. It is an already
occurring event with machines of various configurations and capabilities
coming to take up positions in our world through a slow but steady incur-
sion. It looks less like Blade Runner, Terminator, or Battlestar Galactica
and more like the Fall of Rome.
This book investigates the significance and consequences of this invasion
in an effort 1) to map the increasingly complicated terrain of the twenty-first
century, 2) to assist students, teachers, and researchers in their efforts to
understand and make sense of a changing world, and 3) to provide readers
with the information and critical insight necessary for responding to and
taking an active role in shaping the future.

1.1 Robots
Before we get too far into it, however, it would be a good idea to begin by
getting a handle on the basics . . . beginning with terminology. Despite what
one might think, robots are not the product of scientific R&D. Robots are
originally the result of fiction, specifically a 1920 stage play titled R.U.R.
or Rossum’s Universal Robots and written by the Czech playwright Karel
Čapek (2009). In Czech, as in several other Slavic languages, the word
robota (or some variation thereof) denotes “servitude or labor,” and “robot”
was the word that Čapek used to name a class of manufactured, artificial
slaves that eventually rise up against their human makers (Figure 1.1).
2 Introduction

Figure 1.1 A scene from the play R.U.R., showing three robots.
Source: Public domain image provided by https://commons.wikimedia.org/wiki/File:Capek_
play.jpg.

But Čapek was not, at least as he tells the story, the originator of this des-
ignation. That honor belongs to the playwright’s brother, the painter Josef
Čapek, who suggested the word to his brother during the time of the play’s
initial development (for more, see Gunkel 2018, 15). Since the publication
of Čapek’s play, robots have infiltrated the space of fiction. But what exactly
constitutes a robot differs and admits of a wide variety of forms, functions,
and configurations.

1.1.1 Science Fiction


Čapek’s robots were artificially produced biological creatures that were
humanlike in both material and form. This configuration persists with the
bioengineered replicants of Blade Runner and Blade Runnner 2049 (the
film adaptations of Philip K. Dick’s Do Androids Dream of Electric Sheep?)
and the skin-job Cylons of Ronald Moore’s reimagined Battlestar Galac-
tica. Other fictional robots, like the chrome-plated android in Fritz Lang’s
Metropolis and C-3PO of Star Wars, as well as the “hosts” of HBO’s West-
world and the synths of Channel 4/AMC’s Humans, are humanlike in form
but composed of non-biological materials. Others that are composed of
similar synthetic materials have a particularly imposing profile, like For-
bidden Planet’s Robby the Robot (Figure 1.2), Gort from The Day the Earth
Figure 1.2 T
 heatrical poster for the film Forbidden Planet (1956) and featuring the
imposing figure of Robby the Robot.
Source: Public domain image provided by https://commons.wikimedia.org/wiki/File:Forbidden
planetposter.jpg.
4 Introduction
Stood Still, or Robot from the television series Lost in Space. Still others
are not humanoid at all but emulate animals or other kinds of objects, like
the trashcan R2-D2, the industrial tank-like Wall-E, or the electric sheep
of Dick’s novella. Finally, there are entities without bodies, like the HAL
9000 computer in 2001: A Space Odyssey or GERTY in Moon, with virtual
bodies, like the Agents in The Matrix, or with entirely different kinds of
embodiment, like swarms of nanobots.
When it comes to defining the term “robot,” science fiction plays a signifi-
cant and influential role. In fact, much of what we know or think we know
about robots comes not from actual encounters with the technology but from
what we see and hear about in fiction. Ask someone—especially someone
who is not an insider—to define “robot,” and chances are the answer that is
provided will make reference to something found in a science fiction film,
television program, or story. This not only applies to or affects outsiders
looking-in. “Science fiction prototyping,” as Brian David Johnson (2011)
calls it, is rather widespread within the disciplines of AI and robotics even if
it is not always explicitly called out and recognized as such. As the roboticists
Bryan Adams et al. (2000, 25) point out: “While scientific research usually
takes credit as the inspiration for science fiction, in the case of AI and robot-
ics, it is possible that fiction led the way for science.” Because of this, science
fiction is recognized as being both a useful tool and a potential liability.
Engineers and developers, for instance, often endeavor to realize what
has been imaginatively prototyped in fiction. Cynthia Breazeal (2010), for
example, credits the robots of Star Wars as the inspiration for her pioneering
efforts in social robotics:

Ever since I was a little girl seeing Star Wars for the first time, I’ve been
fascinated by this idea of personal robots. . . . I knew robots like that
didn’t really exist, but I knew I wanted to build them.

Despite its utility, for many laboring in the field of robotics this incursion
of pop culture and entertainment into the realm of the serious work of sci-
ence and engineering is also a potential problem and something that must
be, if not actively counteracted, then at least carefully controlled and held in
check. As the roboticist Alan Winfield (2011, 32) explains:

Real robotics is a science born out of fiction. For roboticists this is both
a blessing and a curse. It is a blessing because science fiction provides
inspiration, motivation and thought experiments; a curse because most
people’s expectations of robots owe much more to fiction than reality.
Introduction 5
And because the reality is so prosaic, we roboticists often find ourselves
having to address the question of why robotics has failed to deliver
when it hasn’t, especially since it is being judged against expectations
drawn from fiction.

In whatever form they have appeared, science fiction already—and well


in advance of actual engineering practice—has established expectations for
what a robot is or can be. Even before engineers have sought to develop
working prototypes, writers, artists, and filmmakers have imagined what
robots do or can do, what configurations they might take, and what prob-
lems they could produce for human individuals and communities. John
Jordan (2016, 5) expresses it quite well in his short and very accessible
introductory book Robots:

No technology has ever been so widely described and explored before


its commercial introduction. . . . Thus the technologies of mass media
helped create public conceptions of and expectations for a whole body
of compu-mechanical innovation that had not happened yet: complex,
pervasive attitudes and expectations predated the invention of viable
products.

1.1.2 Science Fact


So what in fact is a robot? Even when one consults knowledgeable experts,
there is little agreement when it comes to defining, characterizing, or even
identifying what is (or what is not) a robot. In the book Robot Futures, Illah
Nourbakhsh (2013, xiv) writes: “Never ask a roboticist what a robot is. The
answer changes too quickly. By the time researchers finish their most recent
debate on what is and what isn’t a robot, the frontier moves on as whole new
interaction technologies are born.”
One widely cited source of a general, operational definition comes from
George Bekey’s Autonomous Robots: From Biological Inspiration to Imple-
mentation and Control: “In this book we define a robot as a machine that
senses, thinks, and acts. Thus, a robot must have sensors, processing ability
that emulates some aspects of cognition, and actuators” (Bekey 2005, 2).
This “sense, think, act” paradigm is, as Bekey (2005, 2) explicitly recognizes,
“very broad,” encompassing a wide range of different kinds of technologies,
artifacts, and devices. But it could be too broad insofar as it may be applied
to all kinds of artifacts that exceed the proper limits of what many consider
to be a robot. “The sense-think-act paradigm,” as John Jordan (2016, 37)
6 Introduction
notes, “proves to be problematic for industrial robots: some observers con-
tend that a robot needs to be able to move; otherwise, the Watson computer
might qualify.” The Nest thermostat provides another complicated case.

The Nest senses: movements, temperature, humidity, and light. It rea-


sons: if there’s no activity, nobody is home to need air conditioning.
It acts: given the right sensor input, it autonomously shuts the furnace
down. Fulfilling as it does the three conditions, is the Nest, therefore,
a robot?
(Jordan 2016, 37)

And what about smartphones? According to Joanna Bryson and Alan Win-
field (2017, 117) these devices could also be considered robots under this
particular characterization:

Robots are artifacts that sense and act in the physical world in real
time. By this definition, a smartphone is a (domestic) robot. It has not
only microphones but also a variety of proprioceptive sensors that let it
know when its orientation is changing or when it is falling.

In order to further refine the definition and delimit with greater precision
what is and what is not a robot, Winfield (2012, 8) offers the following list
of qualifying characteristics:

A robot is:
1. An artificial device that can sense its environment and purpose-
fully act on or in that environment;
2. An embodied artificial intelligence; or
3. A machine that can autonomously carry out useful work.

Although basically another variation of sense-think-act, Winfield adds an


important qualification to his list—“embodiment”—making it clear that a
software bot, an algorithm, or an AI implementation like IBM’s Watson
or DeepMind’s AlphaGo are not robots, strictly speaking. This is by no
means an exhaustive list of all the different ways in which “robot” has been
defined, explained, or characterized. What is clear from this sample, how-
ever, is that the term “robot” is open to a range of diverse and even different
denotations. And these “definitions are,” as Jordan (2016, 4) writes, “unset-
tled, even among those most expert in the field.”
Our task, at this stage, is not to sort this out once and for all but simply
to identify this terminological difficulty and to recognize that what is under
investigation is as much a product of innovation in technology as it is a
Introduction 7
rhetorical construct. For this reason, words matter. What we call these things
and how they come to be described in both fiction and the scientific litera-
ture are important to how we understand what they are, what they might
become, and what they do and/or are capable of doing. What this means
for us, then, is that our understanding of “robot” and the “robot invasion”
is something of a moving target. We will need to focus attention not just
on different kinds of technological objects but also on the way scientists,
engineers, science fiction writers and filmmakers, journalists, politicians,
critics, and others situate, conceptualize, and explain this technology in and
by language and other methods of representation. What is called “robot” is
not some rigorously defined, singular kind of thing that exists out there in
a vacuum. It is something that is socially negotiated such that word usage
and modes of communication shape expectations for, experiences with, and
understandings of the technology and its impact. Consequently, we need to be
aware of the fact that whatever comes to be called “robot” is always socially
situated and constructed. Its context (or contexts, because they are always
plural) is as important as its technical components and characterizations.

1.2 The Invasion


Despite what is imaginatively foretold in fiction, the robot invasion is not
coming from the future. The robots are already here. Like the “barbarians”
that were said to have invaded Rome, we have already invited the robots
into our places of work, into our homes, and into our lives. They are already
all around us, even if we often do not see or identify them as such.

1.2.1 Robots at Work


Industrial robots (IRs) have slowly but steadily been invading our work
places since the mid-1970s. As S. M. Solaiman (2017, 156) recently reported:

The International Federation for Robotics (IFR) in a 2015 report on


IRs found an increase in the usage of robots by 29% in 2014, which
recorded the highest sales of 229,261 units for a single year (IFR 2015).
IFR estimates that about 1.3 million new IRs will be employed to work
alongside humans in factories worldwide between 2015 and 2018 (IFR
2015). IFR has termed this remarkable increase as ‘conquering the
world’ by robots (IFR 2015). [See Figure 1.3].

But the robots are not just taking over industrial tasks that could be
classed as “dull, dirty and dangerous”—e.g., repetitive and monotonous
tasks performed on manufacturing assembly lines; efforts to extract raw
8 Introduction

Figure 1.3 I ndustrial robot for metal die casting. KUKA Roboter GmbH, Augsburg,
Germany.
Source: Public domain image provided by https://commons.wikimedia.org/wiki/File:Automation_
of_foundry_with_robot.jpg.

materials from the earth and process industrial waste byproducts; or opera-
tions conducted in situations or environments hazardous to human life, like
terrestrial places flooded with dangerous levels of radiation or toxic chemi-
cals, deep underwater on the floor of the ocean, or in the vacuum of space.
The invasion of the work place is (or will soon be) affecting many other
kinds of occupations. As Martin Ford (2015, xiv) explains in his book Rise
of the Robots:

One widely held belief that is certain to be challenged is the assump-


tion that automation is primarily a threat to workers who have little
education and lower-skill levels. That assumption emerges from the
fact that such jobs tend to be routine and repetitive. Before you get too
comfortable with that idea, however, consider just how fast the frontier
is moving. At one time, a “routine” occupation would probably have
implied standing on an assembly line. The reality today is far different.
While lower-skill occupations will no doubt continue to be affected, a
great many college educated, white collar workers are going to discover
that their jobs, too, are squarely in the sights as software automation and
predictive algorithms advance rapidly in capability.
Introduction 9
Recent developments in connectionist architecture, machine learning,
and big data mean that we are not just automating the dull, dirty, and dan-
gerous industrial-era jobs. Being automated are tasks and occupations that
fall squarely within the realm of what is often called “routine intellectual
work.” In other words, what is now susceptible to the pressures of automa-
tion are not just the “blue collar” jobs that require repetitive physical activity
but also those middle-wage, “white collar” occupations that are often the
opportunities sought out by university graduates.
Consider, for example, occupations involving activities related to human
communication and information processing. At one time, if you wanted
customers to conduct transactions through a sales or customer service rep-
resentative, you would have needed to hire a good number of human work-
ers to take customer inquiries, to process those requests and communicate
the results, and to supply a friendly (or at least efficient) conversational
interface to the business or enterprise. Today an increasingly significant
portion of this work can be performed by a speech dialogue system (SDS),
like Google Duplex. In fact, Duplex is so good at emulating human-grade
conversational interaction, that governments, like the state of California,
have recently passed legislation stipulating that SDS explicitly inform users
that they are actually talking to a machine and not a “real” human person.
If you wanted something written—catalogue copy, a financial report, or a
magazine article about a championship football match, etc.—a human being
(someone with a job title like “copywriter,” “technical writer,” or “journal-
ist”) was needed to conduct the research, organize and process the appro-
priate data, and generate readable prose. Now, however, Natural Language
Generation (NLG) algorithms, like Narrative Science’s Quill and Auto-
mated Insights’ Wordsmith, can do that work for you. For the time being
at least, these NLG systems seem to be best suited to developing descrip-
tive and factual stories, leaving the more complicated storytelling tasks to
human writers. But for how long? Kris Hammond of Narrative Science has
not only predicted that within ten year’s time 90% of all written content will
be algorithmically generated but has also asserted that an NLG algorithm,
like Quill, might, at some point in the not-too-distant future, successfully
compete for and even win a Pulitzer Prize (Wright 2015, 14).
Similar transformations are occurring in other sectors of the economy.
Right now it is estimated that 4.9 million human workers are employed in
for-hire transportation within the United States—e.g., driving cars (think
limo services, taxis, or Lyft and Uber), long-haul trucking, city and inter-
state buses, local delivery vehicles, etc. (U.S. Department of Transportation
2017, ch. 4). Self-driving vehicles promise or threaten (and the choice of
verb depends on your individual perspective) to take over a good portion of
this work, displacing an entire sector of good paying jobs for skilled workers.
10 Introduction
1.2.2 Robots at Home
The “robot invasion” is not something that is limited to the work place, there
are also “service robots,” which are characterized as machines involved in
“entertaining and taking care of children and elderly people, preparing food
and cooking in restaurants, cleaning residential premise, and milking cows”
(Cookson 2015). There are, according to data provided by the Foundation
for Responsible Robotics, 12 million service robot currently in operation
across the globe, and the IFR predicts “an exponential rise” with the popula-
tion of service robots expected to reach 31 million units by 2018 (Solaiman
2017, 156).
There are also “social robots,” a subset of service robot specifically
designed for human social interaction and use in domestic settings. This
category of robot includes devices and applications like the Paro therapy
robot, which has proven to be an incredibly useful tool for elder care, espe-
cially in situations involving individuals suffering from debilitating forms
of dementia (Wada and Shibata 2007; Šabanović et al. 2013; Bemelmans
et al. 2012), and Jibo, a table-top digital assistant—kind of like Alexa with
a movable body—developed by Cynthia Breazeal and marketed as the “the
world’s first family robot” (Figure 1.4). And at the very far end of the social

Figure 1.4 The social robot Jibo.


Source: Photograph by the author.
Introduction 11
robotics spectrum, there are various forms of “sex robots,” which are being
promoted not as substitutes for human partners but as a means to augment
human intimacy and sexual relationships (Levy 2007; Danaher and McAr-
thur 2017; Devlin 2018; Richardson 2019). Although providing for different
kinds of social interaction, these social robots are, as Breazeal (2002, 1)
has described it, designed to be “socially intelligent in a human like way,
[so that] interacting with it is like interacting with another person.” Popula-
tion statistics and predictions for social robots exceed IRs, with countries
like South Korea aiming to put a robot in every home by 2020 (Lovgren
2006).
Finally, there are distributed systems like the Internet of Things (IoT),
where numerous connected devices “work together” to make an automated
arrangement that is not “a robot” in the typical sense of the word but a net-
work of interacting and smart devices. The Internet is already overrun, if not
already run, by machines, with better than 50% of all online activity being
machine generated and consumed (Zeifman 2017), and it is now estimated
that IoT will support over 26 billion interactive and connected devices by
2020. By way of comparison, the current human population of planet earth
is estimated to be 7.4 billion (Gartner 2013). We have therefore already
achieved and live in that future Norbert Wiener had predicted at the begin-
ning of his 1950 book, The Human Use of Human Beings: Cybernetics and
Society:

It is the thesis of this book that society can only be understood through
a study of the messages and the communication facilities which belong
to it; and that in the future development of these messages and com-
munication facilities, messages between man and machines, between
machines and man, and between machine and machine, are destined to
play an ever-increasing part.
(Wiener 1988, 16)

1.3 The Program


What matters most in the face of this already occurring robot incursion is
not resistance—insofar as resistance already appears to be futile—but how
we decide to make sense of and respond to the social opportunities and
challenges that these increasingly autonomous machines make available.
As these various mechanisms take up influential positions in contempo-
rary culture—positions where they are not necessarily just tools or instru-
ments of human action but a kind of interactive social agent in their own
right—we will need to ask ourselves some rather interesting but difficult
questions.
12 Introduction
At what point might a robot, algorithm, or other autonomous system be
held accountable for the decisions it makes or the actions it initiates? When,
if ever, would it make sense to say, “It’s the robot’s fault.” Conversely, when
might a robot, an intelligent artifact, or other socially interactive mechanism
be due some level of social standing or respect? When, in other words,
would it no longer be considered nonsense to inquire about the rights of
robots and to ask the question “Can and should robots have rights?”
We will proceed to address and respond to these questions in three stages
or movements. The next chapter (Chapter 2) begins the investigation by
examining and reevaluating the way we typically deal with and make sense
of our technological devices and systems. It is titled “the default setting,”
because it is concerned with the way that we typically respond to these
opportunities and challenges. Technologies, we assume or tell ourselves, are
invented, manufactured, and utilized by human beings for various applica-
tions and objectives. This is as true of simple devices like hammers and
corkscrews as it is of very sophisticated technological developments like
rocket ships, computers, and artificial intelligence. The German philosopher
Martin Heidegger (1977) famously called this particular characterization of
technology “the instrumental definition” and suggested that it forms what
is considered to be the “correct” understanding of any kind of technologi-
cal contrivance. The second chapter will review and critically assess this
“instrumental theory of technology.”
The third chapter is titled “The New Normal,” and it examines three
recent technological innovations with artificial intelligence and robot-
ics that complicate the instrumental theory and necessitate other ways of
responding to and taking responsibility for robots. The three technologies
include: 1) natural language processing (NLP) applications like Siri, Alexa,
or Google Home. These devices are designed to interact with users in a
conversational manner, and they now populate many of the social spaces in
which we live, work, and play. Unlike artificial general intelligence (AGI),
which would presumably occupy a subject position reasonably close to that
of another human person, these ostensibly mindless but very social things
complicate social relationships by opening onto and leaving undecided
questions regarding who or what is talking to us. 2) Machine learning (ML)
applications. The instrumental theory, for all its success handling different
kinds of technology, appears to be unable to contend with recent devel-
opments in machine learning, where machine actions are not prescribed
by human developers through program instructions but are developed by
the machine through discovering patterns in large data sets. Consequently,
machine learning systems are deliberately designed to do things that their
human programmers cannot anticipate, completely control, or answer for.
In other words, we now have autonomous (or at least semi-autonomous)
Introduction 13
robotic systems that in one way or another have “a mind of their own.” And
this is where things get interesting, especially when it comes to questions
of responsibility and social standing. 3) Social robots. Unlike industrial
robots, which are designed to accomplish a specific task on the assembly
line, social robots are meant to interact with human users in a human-like
way. For this reason, social robots are deliberately designed for and occupy
a strange position in the social order. They are not just another instrument,
like a refrigerator or a toaster, but they also are not quite another “person.”
They are a kind of quasi-other that is intentionally designed to occupy an
ambivalent position that complicates existing categories.
The fourth and final chapter considers and critically assesses the range of
possible responses to these new social challenges. What we see in the face
or the faceplate of the robot is a situation where our theory of technology—a
theory that has considerable history behind it and that has been determined
to be as applicable to simple hand tools as it is to complex technological
systems—seems to be unable to respond to or answer for recent develop-
ments with artificial intelligence, algorithms, and robots. In the face of these
opportunities/challenges there are at least three possible responses: Slavery
2.0, Machine Ethics, and Joint Agency. The final chapter will introduce and
perform a cost/benefit analysis of these three modes of response. The goal
in doing so is not to select one as better than the others but to lay out the
options that we—each of us individually and together as a community—
need to consider and debate. The goal, therefore, is not to provide readers
with the “right answer” but to articulate the terms of the debate, map out
what needs to be addressed and decided, and provide individuals with the
knowledge and data to make and defend their own determinations.
One final word concerning the size of this book. As you can see in both
the number of chapters and the physical dimensions, the book is short. This
is deliberate and strategic. It is not, I should emphasize, a product of there
being little to say on the matter. There is, in fact, plenty to say, and I will
highlight various resources that may be accessed should you want additional
information. The brevity is more a matter of wanting to provide readers with
a quick and easily accessible mode of entry into what is often considered to
be too complicated, too involved, and too technical for non-specialists. The
challenges that are now confronted, with innovations in robots, machine
learning algorithms, and autonomous technology, are simply too important
and influential for us to leave decision-making to a few experts. We do, no
doubt, need the experts. But we also need an informed and knowledgeable
public who can think and act confidently in the face of new technological
opportunities and challenges. The book is designed to bring everyone (irre-
spective of background or formal training) up to speed in an effort to ensure
that we (together) are working to respond to and take responsibility for OUR
14 Introduction
future. It is only by working together and across difference that we have the
chance to survive a robot invasion.

References
Adams, Brian, Cynthia L. Breazeal, Rodney Brooks and Brian Scassellati. 2000.
“Humanoid Robots: A New Kind of Tool.” IEEE Intelligent Systems 15(4): 25–31.
https://doi.org/10.1109/5254.867909
Bekey, George A. 2005. Autonomous Robots: From Biological Inspiration to Imple-
mentation and Control. Cambridge, MA: MIT Press.
Bemelmans, Roger, Gert Jan Gelderblom, Pieter Jonker and Luc de Witte. 2012.
“Socially Assistive Robots in Elderly Care: A Systematic Review into Effects and
Effectiveness.” Journal of the American Medical Directors Association 13(2):
114–120. https://doi.org/10.1016/j.jamda.2010.10.002
Breazeal, Cynthia L. 2002. Designing Sociable Robots. Cambridge, MA: MIT Press.
Breazeal, Cynthia L. 2010. “The Rise of Personal Robots.” TEDWomen 2010. www.
ted.com/talks/cynthia_breazeal_the_rise_of_personal_robots?
Bryson, Joanna and Alan Winfield. 2017. “Standardizing Ethical Design for Artifi-
cial Intelligence and Autonomous Systems.” Computer 50(5): 116–119. https://
doi.org/10.1109/MC.2017.154
Čapek, Karel. 2009. R.U.R. (Rossum’s Universal Robots), translated by David
Wyllie. Gloucestershire, UK: The Echo Library.
Cookson, Clive. 2015. “Scientists Appeal for Ethical Use of Robots.” Financial Times
(10 December). www.ft.com/content/fee8bacc-9f37-11e5-8613-08e211ea5317
Danaher, John and Neil McArthur. 2017. Robot Sex: Social and Ethical Implica-
tions. Cambridge, MA: MIT Press.
Devlin, Kate. 2018. Turned On: Science, Sex and Robots. New York: Bloomsbury
Sigma.
Ford, Martin. 2015. Rise of the Robots: Technology and the Threat of a Jobless
Future. New York: Basic Books.
Gartner. 2013. “Gartner Says the Internet of Things Installed Base Will Grow to
26 Billion Units by 2020.” www.gartner.com/newsroom/id/2636073
Gunkel, David J. 2018. Robot Rights. Cambridge, MA: MIT Press.
Heidegger, Martin. 1977. The Question Concerning Technology and Other Essays,
translated by William Lovitt. New York: Harper & Row.
International Federation of Robotics (IFR). 2015. “Industrial Robot Statistics.”
www.ifr.org/industrial-robots/statistics/
Johnson, Brian David. 2011. Science Fiction Prototyping: Designing the Future
with Science Fiction. Williston, VT: Morgan and Claypool Publishers. https://doi.
org/10.2200/S00336ED1V01Y201102CSL003
Jordan, John. 2016. Robots. Cambridge, MA: MIT Press.
Levy, David. 2007. Love and Sex with Robots: The Evolution of Human-Robot Rela-
tionships. New York: Harper Perennial.
Lovgren, Stefan. 2006. “A Robot in Every Home by 2020, South Korea Says.”
National Geographic News (6 September). http://news.nationalgeographic.com/
news/2006/09/060906-robots.html
Nourbakhsh, Illah. 2013. Robot Futures. Cambridge: MIT Press.
Richardson, Kathleen. 2019. Sex Robots: The End of Love. Cambridge: Polity Press.
Šabanović, Selma, Casey C. Bennett, Wan-Ling Chang and Lesa Huber. 2013.
“PARO Robot Affects Diverse Interaction Modalities in Group Sensory Ther-
apy for Older Adults with Dementia.” 2013 IEEE 13th International Confer-
ence on Rehabilitation Robotics (ICORR). 24–26 June. https://doi.org/10.1109/
ICORR.2013.6650427
Solaiman, S. M. 2017. “Legal Personality of Robots, Corporations, Idols and Chim-
panzees: A Quest for Legitimacy.” Artificial Intelligence Law 25(2): 155–179.
https://doi.org/10.1007/s10506-016-9192-3
US Department of Transportation. 2017. Transportation Economic Trends 2017.
https://www.bts.gov/product/transportation-economic-trends
Wada, Kazuyoshi and Takanori Shibata. 2007. “Living With Seal Robots: Its Socio-
psychological and Physiological Influences on the Elderly at a Care House.”
IEEE Transactions on Robotics 23(5): 972–980. https://doi.org/10.1109/
TRO.2007.906261
Wiener, Norbert. 1988. The Human Use of Human Beings: Cybernetics and Society.
Boston, MA: Da Capo Press.
Winfield, Alan. 2011. “Roboethics: For Humans.” New Scientist 210(2811): 32–33.
https://doi.org/10.1016/S0262-4079(11)61052-X
Winfield, Alan. 2012. Robotics: A Very Short Introduction. Oxford: Oxford Univer-
sity Press.
Wright, Alex. 2015. “Algorithmic Authors.” Communications of the ACM 58(11):
12–14. https://doi.org/10.1145/2820421
Zeifman, Igal. 2017. “Bot Traffic Report 2016.” Incapsula. www.incapsula.com/
blog/bot-traffic-report-2016.html
Benso, Silvia. 2000. The Face of Things: A Different Side of Ethics. Albany, NY:
SUNY Press.
Bentham, Jeremy. 2005. An Introduction to the Principles of Morals and Legisla-
tion, edited by J. H. Burns and H. L. Hart. Oxford: Oxford University Press.
Breazeal, Cynthia and Rodney Brooks. 2004. “Robot Emotion: A Functional Per-
spective.” In Who Needs Emotions: The Brain Meets the Robot, edited by J. M.
Fellous and M. Arbib, 271–310. Oxford: Oxford University Press.
Brooks, Rodney A. 2002. Flesh and Machines: How Robots Will Change Us. New
York: Pantheon Books.
Bryson, Joanna. 2010. “Robots Should be Slaves.” In Close Engagements with Arti-
ficial Companions: Key Social, Psychological, Ethical and Design Issues, edited
by Yorick Wilks, 63–74. Amsterdam: John Benjamins.
Calarco, Matthew. 2008. Zoographies: The Question of the Animal from Heidegger
to Derrida. New York: Columbia University Press.
Churchland, Paul M. 1999. Matter and Consciousness. Cambridge, MA: MIT Press.
Coeckelbergh, Mark. 2012. Growing Moral Relations: Critique of Moral Status
Ascription. New York: Palgrave MacMillan.
Coeckelbergh, Mark and David J. Gunkel. 2014. “Facing Animals: A Relational,
Other-Oriented Approach to Moral Standing.” Journal of Agricultural and Envi-
ronmental Ethics 27(5): 715–733. https://doi.org/10.1007/s10806-013-9486-3
Dennett, Daniel C. 1998. Brainstorms: Philosophical Essays on Mind and Psychol-
ogy. Cambridge, MA: MIT Press.
Derrida, Jacques. 2005. Paper Machine, translated by Rachel Bowlby. Stanford,
CA: Stanford University Press.
Feenberg, Andrew. 1991. Critical Theory of Technology. New York: Oxford Uni-
versity Press.
Floridi, Luciano. 2013. The Ethics of Information. Oxford: Oxford University Press.
Goertzel, Ben. 2002. “Thoughts on AI Morality.” Dynamical Psychology: An Inter-
national, Interdisciplinary Journal of Complex Mental Processes (May). http://
www.goertzel.org/dynapsyc/2002/AIMorality.htm.
Gunkel, David J. 2012. The Machine Question: Critical Perspectives on AI, Robots
and Ethics. Cambridge, MA: MIT Press.
Haraway, Donna J. 2008. When Species Meet. Minneapolis, MN: University of Min-
nesota Press.
Harman, Graham. 2002. Tool Being: Heidegger and the Metaphysics of Objects.
Peru, IL: Open Court Publishing.
Heidegger, Martin. 1962. Being and Time, translated by John Macquarrie and
Edward Robinson. New York: Harper & Row.
Heidegger, Martin. 1977. The Question Concerning Technology and Other Essays,
translated by William Lovitt. New York: Harper & Row.
Himma, Kenneth Einar. 2009. “Artificial Agency, Consciousness, and the Criteria
for Moral Agency: What Properties Must an Artificial Agent Have to be a Moral
Agent?” Ethics and Information Technology 11(1):19–29. https://doi.org/10.1007/
s10676-008-9167-5
Johnson, Deborah G. 1985. Computer Ethics. Upper Saddle River, NJ: Prentice Hall.
Johnson, Deborah G. 2006. “Computer Systems: Moral Entities But Not Moral
Agents.” Ethics and Information Technology 8(4): 195–204. https://doi.
org/10.1007/s10676-006-9111-5
Johnson, Deborah G. and Keith W. Miller. 2008. “Un-Making Artificial Moral
Agents.” Ethics and Information Technology 10(2–3): 123–133. https://doi.
org/10.1007/s10676-008-9174-6
Kant, Immanuel. 1985. Critique of Practical Reason, translated by Lewis White
Beck. New York: Macmillan.
Leopold, Aldo. 1966. A Sand County Almanac. Oxford: Oxford University Press.
Lovelace, Ada. 1842. “Translator’s Notes to an Article on Babbage’s Analytical
Engine. Scientific Memoirs (ed. by R. Taylor), vol. 3: 691–731.
Lyotard, Jean-François. 1984. The Postmodern Condition: A Report on Knowledge,
translated by Geoff Bennington and Brian Massumi. Minneapolis, MN: Univer-
sity of Minnesota Press.
McLuhan, Marshall and Quentin Fiore. 2001. War and Peace in the Global Village.
Berkeley: CA: Gingko Press.
Mowshowitz, Abe. 2008. “Technology as Excuse for Questionable Ethics.” AI &
Society 22(3): 271–282. https://doi.org/10.1007/s00146-007-0147-9
Nissenbaum, Helen. 1996. “Accountability in a Computerized Society.” Science and
Engineering Ethics 2(1): 25–42. https://doi.org/10.1007/BF02639315
O’Regan, Kevin J. 2007. “How to Build Consciousness into a Robot: The Senso-
rimotor Approach.” In 50 Years of Artificial Intelligence: Essays Dedicated to the
50th Anniversary of Artificial Intelligence, edited by Max Lungarella, Fumiya
Iida, Josh Bongard and Rolf Pfeifer, 332–346. Berlin: Springer-Verlag.
Ricœur, Paul. 2007. Reflections on the Just, translated by David Pellauer. Chicago:
University of Chicago Press.
Searle, John. 1999. “The Chinese Room.” In The MIT Encyclopedia of the Cognitive
Sciences, edited by R. A. Wilson and F. Keil, 115–116. Cambridge, MA: MIT Press.
Singer, Peter. 1975. Animal Liberation: A New Ethics for Our Treatment of Animals.
New York: New York Review of Books.
Siponen, Mikko. 2004. “A Pragmatic Evaluation of the Theory of Information Eth-
ics.” Ethics and Information Technology 6(4): 279–290. https://doi.org/10.1007/
s10676-005-6710-5
Sullins, John P. 2006. “When Is a Robot a Moral Agent?” International Review of
Information Ethics 6(12): 23–30. www.i-r-i-e.net/inhalt/006/006_Sullins.pdf
Turing, Alan. 1999. “Computing Machinery and Intelligence.” In Computer Media
and Communication: A Reader, edited by Paul A. Meyer, 37–58. Oxford: Oxford
University Press.
Velmans, Max. 2000. Understanding Consciousness. London, UK: Routledge.
Yonck, Richard. 2017. Heart of the Machine: Our Future in a World of Artificial
Emotional Intelligence. New York: Arcade Publishing.
Zimmerman, Michael E. 1990. Heidegger’s Confrontation with Modernity: Technol-
ogy, Politics, Art. Bloomington, IN: Indiana University Press.
Bemelmans, Roger, Gert Jan Gelderblom, Pieter Jonker and Luc de Witte. 2012.
“Socially Assistive Robots in Elderly Care: A Systematic Review into Effects and
Effectiveness.” Journal of the American Medical Directors Association 13(2):
114–120. https://doi.org/10.1016/j.jamda.2010.10.002
Bessi, Alessandro and Emilio Ferrara. 2016. “Social Bots Distort the 2016 U.S.
Presidential Election Online Discussion.” First Monday 21(11). https://doi.
org/10.5210/fm.v21i11.7090
Breazeal, Cynthia L. 2002. Designing Sociable Robots. Cambridge, MA: MIT Press.
Breazeal, Cynthia L. 2010. “The Rise of Personal Robots.” TEDWomen 2010. www.
ted.com/talks/cynthia_breazeal_the_rise_of_personal_robots?
Čapek, Karel. 2009. R.U.R. (Rossum’s Universal Robots), translated by David
Wyllie. Gloucestershire, UK: The Echo Library.
Carpenter, Julie. 2015. Culture and Human-Robot Interaction in Militarized Spaces:
A War Story. New York: Ashgate.
Darling, Kate. 2012. “Extending Legal Protection to Social Robots. IEEE Spectrum.
http://spectrum.ieee.org/automaton/robotics/artificial-intelligence/extending-
legal-protection-to-social-robots
DeepMind. 2016. “AlphaGo.” https://deepmind.com/alpha-go.html
Derrida, Jacques. 2005. Paper Machine, translated by Rachel Bowlby. Stanford,
CA: Stanford University Press.
Epstein, Robert. 2007. “From Russia, with Love: How I Got Fooled (and Somewhat
Humiliated) by a Computer.” Scientific American Mind. www.scientificamerican.
com/article/from-russia-with-love/
Hui, Fan. 2016. AlphaGo Games—English. DeepMind. https://deepmind.com/
alphago-games-english/
Jibo. 2014. “Indiegogo Video.” www.indiegogo.com/projects/jibo-the-world-s-first-
social-robot-for-the-home#/
Kim, Elizabeth S., Lauren D. Berkovits, Emily P. Bernier, Dan Leyzberg, Freder-
ick Shic, Rhea Paul and Brian Scassellati. 2013. “Social Robots as Embedded
Reinforcers of Social Behavior in Children with Autism.” Journal of Autism and
Developmental Disorders 43(5): 1038–1049. https://doi.org/10.1007/s10803-
012-1645-2
Lee, Peter. 2016. “Learning from Tay’s Introduction.” Official Microsoft Blog (25
March). https://blogs.microsoft.com/blog/2016/03/25/learning-tays-introduction/
Matthias, Andreas. 2004. “The Responsibility Gap: Ascribing Responsibility for
the Actions of Learning Automata.” Ethics and Information Technology 6(3):
175–183. https://doi.org/10.1007/s10676-004-3422-1
McFarland, Matt. 2016. “What AlphaGo’s Sly Move Says About Machine Cre-
ativity.” The Washington Post (15 March). www.washingtonpost.com/news/
innovations/wp/2016/03/15/what-alphagos-sly-move-says-about-machine-
creativity/?utm_term=.0c8281af53c9
Metz, Cade. 2016a. “The Sadness and Beauty of Watching Google’s AI Play Go.”
Wired (11 March). www.wired.com/2016/03/sadness-beauty-watching-googles-
ai-play-go/
Metz, Cade. 2016b. “In Two Moves, AlphaGo and Lee Sedol Redefine the Future.”
Wired (16 March). www.wired.com/2016/03/two-moves-alphago-lee-sedol-
redefined-future/
Metz, Cade. 2016c. “Google’sAI Wins a Pivotal Second Game in Match with Go Grand-
master.” Wired (10 March). www.wired.com/2016/03/googles-ai-wins-pivotal-
game-two-match-go-grandmaster/
Microsoft. 2016. “Meet Tay: Microsoft A.I. Chatbot with Zero Chill.” www.tay.ai/
Mowbray, Miranda. 2002. “Ethics for Bots: Paper presented at the 14th International
Conference on System Research, Informatics, and Cybernetics.” Baden-Baden, Ger-
many. 29 July–3 August. www.hpl.hp.com/techreports/2002/HPL-2002-48R1.pdf
Peterson, Andrea. 2013. “On the Internet, No one Knows You’re a Bot: And That’s
a Problem.” The Washington Post (13 August). www.washingtonpost.com/news/
the-switch/wp/2013/08/13/on-the-internet-no-one-knows-youre-a-bot-and-thats-
a-problem/?utm_term=.b4e0dd77428a
Reeves, Byron and Clifford Nass. (1996). The Media Equation: How People Treat
Computers, Television, and New Media Like Real People and Places. Cambridge:
Cambridge University Press.
Ricœur, Paul. 2007. Reflections on the Just, translated by David Pellauer. Chicago:
University of Chicago Press.
Risely, James. 2016. “Microsoft’s Millennial Chatbot Tay.ai Pulled Offline After
Internet Teaches Her Racism.” GeekWire. www.geekwire.com/2016/even-robot-
teens-impressionable-microsofts-tay-ai-pulled-internet-teaches-racism/
Robins, Ben., K. Kerstin Dautenhahn, Rene Te Boekhorst and Aude Billard. 2005.
“Robotic Assistants in Therapy and Education of Children with Autism: Can a
Small Humanoid Robot Help Encourage Social Interaction Skills?” Universal
Access in the Information Society 4(2): 105–120. https://doi.org/10.1007/s10209-
005-0116-3
Rosenthal-von der Pütten, Astrid M., Nicole C. Krämer, Laura Hoffmann, Sabrina
Sobieraj and Sabrina C. Eimler. 2013. “An Experimental Study on Emotional
Reactions Towards a Robot.” International Journal of Social Robotics 5(1):
17–34. https://doi.org/10.1007/s12369-012-0173-8
Šabanović, Selma, Casey C. Bennett, Wan-Ling Chang and Lesa Huber. 2013.
“PARO Robot Affects Diverse Interaction Modalities in Group Sensory Ther-
apy for Older Adults with Dementia.” 2013 IEEE 13th International Confer-
ence on Rehabilitation Robotics (ICORR). 24–26 June. https://doi.org/10.1109/
ICORR.2013.6650427
Searle, John. 1999. “The Chinese Room.” In The MIT Encyclopedia of the Cogni-
tive Sciences, edited by R. A. Wilson and F. Keil, 115–116. Cambridge, MA: MIT
Press.
Suzuki, Yutaka, Lisa Galli, Ayaka Ikeda, Shoji Itakura and Michiteru Kitazaki. 2015.
“Measuring Empathy for Human and Robot Hand Pain Using Electroencephalog-
raphy.” Scientific Reports 5: 15924. www.nature.com/articles/srep15924
Tanz, Jason. 2016. “The End of Code.” Wired (17 May). www.wired.com/2016/05/
the-end-of-code/
Turing, Alan. 1999. “Computing Machinery and Intelligence.” In Computer Media
and Communication: A Reader, edited by Paul A. Meyer, 37–58. Oxford: Oxford
University Press.
Turkle, Sherry. 2011. Alone Together: Why We Expect More from Technology and
Less from Each Other. New York: Basic Books.
Vincent, Jane. 2013. “Is the Mobile Phone a Personalized Social Robot?” Inter-
valla 1(1): 60–70. www.fus.edu/intervalla/volume-1-social-robots-and-emotion-
transcending-the-boundary-between-humans-and-icts/is-the-mobile-phone-a-
personalized-social-robot
Wada, Kazuyoshi and Takanori Shibata. 2007. “Living With Seal Robots: Its Socio-
psychological and Physiological Influences on the Elderly at a Care House.”
IEEE Transactions on Robotics 23(5): 972–980. https://doi.org/10.1109/
TRO.2007.906261
Weizenbaum, Joseph. 1976. Computer Power and Human Reason: From Judgment
to Calculation. San Francisco, CA: W. H. Freeman.
Žižek, Slavoj. 2008. The Sublime Object of Ideology. London: Verso.
Anderson, Susan Leigh. 2008. “Asimov’s ‘Three Laws of Robotics’ and Machine
Metaethics.” AI & Society 22(4): 477–493. https://doi.org/10.1007/s00146-007-
0094-5
Anderson, Michael and Susan Leigh Anderson. 2007. “Machine Ethics: Creating
an Ethical Intelligent Agent.” AI Magazine 28(4):15–26. https://doi.org/10.1609/
aimag.v28i4.2065
Aristotle. 1944. Politics, translated by H. Rackham. Cambridge, MA: Harvard Uni-
versity Press.
Arkin, Ronald C. 2009. Governing Lethal Behavior in Autonomous Robots. Boca
Raton, FL: CRC Press.
Ashrafian, Hutan. 2015. “Artificial Intelligence and Robot Responsibilities: Innovat-
ing Beyond Rghts.” Science and Engineering Ethics 21(2): 317–326. https://doi.
org/10.1007/s11948-014-9541-0
Asimov, Isaac. (2008). I, Robot. New York: Bantam Books.
Binder, Otto O. 1957. “You Will Own ‘Slaves’ by 1965.” In Mechanix Illustrated
(January): 62–65. Greenwich, CT: Modern Mechanix Publishing Co.
Boden, Margaret, Joanna Bryson, Darwin Caldwell, Kerstin Dautenhahn, Lilian
Edwards, Sarah Kember, Paul Newman, Vivienne Parry, Geoff Pegman, Tom
Rodden, Tom Sorrell, Mick Wallis, Blay Whitby and Alan Winfield. 2017. “Prin-
ciples of Robotics: Regulating Robots in the Real World.” Connection Science
29(2): 124–129. http://dx.doi.org/10.1080/09540091.2016.1271400
Bryson, Joanna. 2010. “Robots Should be Slaves.” In Close Engagements with Arti-
ficial Companions: Key Social, Psychological, Ethical and Design Issues, edited
by Yorick Wilks, 63–74. Amsterdam: John Benjamins.
Bryson, Joanna J., Mihailis E. Diamantis and Thomas D. Grant. 2017. “Of, For, and
By the People: The Legal Lacuna of Synthetic Persons.” Artificial Intelligence
and Law 25(3): 273–291. https://doi.org/10.1007/s10506-017-9214-9
Calverley, David J. 2008. “Imaging a Non-Biological Machine as a Legal Person.”
AI & Society 22(4): 523–537. https://doi.org/10.1007/s00146-007-0092-7
Čapek, Karel. 2009. R.U.R. (Rossum’s Universal Robots), translated by David Wyl-
lie. Gloucestershire, UK: The Echo Library.
Carpenter, Julie. 2015. Culture and Human-Robot Interaction in Militarized Spaces:
A War Story. New York: Ashgate.
Coeckelbergh, Mark. 2010. “Moral Appearances: Emotions, Robots, and Human
Morality.” Ethics and Information Technology 12(3): 235–241. https://doi.
org/10.1007/s10676-010-9221-y
Datteri, Edoardo. 2013. “Predicting the Long-Term Effects of Human-Robot Inter-
action: A Reflection on Responsibility in Medical Robotics.” Science and Engi-
neering Ethics 19(1): 139–160. https://doi.org/10.1007/s11948-011-9301-3
de Tocqueville, Alexis 1899. Democracy in America, translated by H. Reeve. New
York: The Colonial Press.
Estrada, Daniel. 2018. “Sophia and Her Critics.” Medium (17 June). https://medium.
com/@eripsa/sophia-and-her-critics-5bd22d859b9c
Fjeld, Jessica, Hannah Hilligoss, Nele Achten, Maia Levy Daniel, Joshua Feldman
and Sally Kagay. 2019. Principled Artificial Intelligence: A Map of Ethical and
Rights-Based Approaches (4 July). https://ai-hr.cyber.harvard.edu/primp-viz.html
Foundation for Responsible Robotics (FFR). 2019. FFR Website—About Us. https://
responsiblerobotics.org/about-us/mission/
French, Peter. 1979. “The Corporation as a Moral Person.” American Philosophical
Quarterly 16(3): 207–215. www.jstor.org/stable/20009760
Garreau, Joel. 2007. “Bots on the Ground: In the Field of Battle (or Even Above It),
Robots are a Soldier’s Best Friend.” Washington Post (6 May). www.washington-
post.com/wp-dyn/content/article/2007/05/05/AR2007050501009.html
Gladden, Matthew E. 2016. “The Diffuse Intelligent Other: An Ontology of Non-
localizable Robots as Moral and Legal Actors.” In Social Robots: Boundaries,
Potential, Challenges, edited by Marco Nørskov, 177–198. Burlington, VT: Ashgate.
Gunkel, David J. 2012. The Machine Question: Critical Perspectives on AI, Robots
and Ethics. Cambridge, MA: MIT Press.
Hanson, F. Allan. 2009. “Beyond the Skin Bag: On the Moral Responsibility of
Extended Agencies.” Ethics and Information Technology 11(1): 91–99. https://
doi.org/10.1007/s10676-009-9184-z
Hemmersbaugh, Paul A. 2016. “NHTSA Letter to Chris Urmson, Director, Self-
Driving Car Project, Google, Inc.” https://isearch.nhtsa.gov/files/Google%20-%
20compiled%20response%20to%2012%20Nov%20%2015%20interp%
20request%20-%204%20Feb%2016%20final.htm.
Hernéndez-Orallo, José. 2017. The Measure of All Minds: Evaluating Natural and
Artificial Intelligence. Cambridge: Cambridge University Press.
Inayatullah, Sohail and Phil McNally. 1988. “The Rights of Robots: Technology,
Culture and Law in the 21st Century.” Futures 20(2): 119–136. www.kurzweilai.
net/the-rights-of-robots-technology-culture-and-law-in-the-21st-century
Jacobs, Harriet Ann. 2001. Incidents in the Life of a Slave Girl. New York: Dover.
Johnson, Deborah G. 2006. “Computer Systems: Moral Entities But Not Moral
Agents.” Ethics and Information Technology 8(4): 195–204. https://doi.
org/10.1007/s10676-006-9111-5
Jones, Raya. 2016. Personhood and Social Robotics: A Psychological Consider-
ation. New York: Routledge.
LaGrandeur, Kevin. 2013. Androids and Intelligent Networks in Early Modern Lit-
erature and Culture: Artificial Slaves. New York: Routledge.
Latour, Bruno. 2005. Reassembling the Social: An Introduction to Actor-Network-
Theory. Oxford: Oxford University Press.
Marx, Karl. 1977. Capital: A Critique of Political Economy, translated by Ben
Fowkes. New York: Vintage Books.
Mayor, Adrienne. 2018. Gods and Robots: Myths, Machines and Ancient Dreams of
Technology. Princeton, NJ: Princeton University Press.
Nissenbaum, Helen. 1996. “Accountability in a Computerized Society.” Science and
Engineering Ethics 2(1): 25–42. https://doi.org/10.1007/BF02639315
Pagallo, Ugo. 2013. The Laws of Robots: Crimes, Contracts, and Torts. New York:
Springer.
Robertson, Jennifer. 2014. “Human Rights vs. Robot Rights: Forecasts from Japan.”
Critical Asian Studies 46(4): 571–598. https://doi.org/10.1080/14672715.2014.
960707
Robertson, Jennifer. 2017. Robo Sapiens Japanicus: Robots, Gender, Family, and
the Japanese Nation. Berkeley, CA: University of California Press.
Ross, Philip E. 2016. “A Google Car Can Qualify as a Legal Driver.” IEEE Spec-
trum. http://spectrum.ieee.org/cars-that-think/transportation/self-driving/an-ai-
can-legally-be-defined-as-a-cars-driver.
Scheutz, Matthias. 2012. “The Inherent Dangers of Unidirectional Emotional Bonds
Between Humans and Social Robots.” In Robot Ethics: The Ethical and Social
Implications of Robotics, edited by Patrick Lin, Keith Abney and George A.
Bekey, 205–221. Cambridge, MA: MIT Press.
Singer, Peter W. 2009. Wired for War: The Robotics Revolution and Conflict in the
Twenty-First Century. New York: Penguin Books.
Turner, Jacob. 2019. Robot Rules: Regulating Artificial Intelligence. Cham: Pal-
grave Macmillan.
van de Poel, Ibo, Jessica Nihlén Fahlquist, Neelke Doorn, Sjoerd Zwart and Lambèr
Royakkers. 2012. “The Problem of Many Hands: Climate Change as an Example.”
Science Engineering Ethics 18(1): 49–67. https://doi.org/10.1007/s11948-011-9276-0
Verbeek, Peter Paul. 2011. Moralizing Technology: Understanding and Designing
the Morality of Things. Chicago: University of Chicago Press.
Wagenaar, Willam A. and Jop Groenewegen, J. 1987. “Accidents at Sea: Multiple
Causes and Impossible Consequences.” International Journal of Man-Machine
Studies 27(5–6): 587–598. https://doi.org/10.1016/S0020-7373(87)80017-2
Wallach, Wendell. 2015. A Dangerous Master: How to Keep Technology from Slip-
ping Beyond Our Control. New York: Basic Books.
Wallach, Wendell and Colin Allen. 2009. Moral Machines: Teaching Robots Right
from Wrong. Oxford: Oxford University Press.
Wiener, Norbert. 1988. The Human Use of Human Beings: Cybernetics and Society.
Boston, MA: Da Capo Press.
Wiener, Norbert. 1996. Cybernetics: Or Control and Communication in the Animal
and the Machine. Cambridge, MA: MIT Press.
Winner, Langdon. 1977. Autonomous Technology: Technics-out-of-Control as a
Theme in Political Thought. Cambridge, MA: MIT Press.
Winograd, Terry. 1990. “Thinking Machines: Can There Be? Are We?” In The Foun-
dations of Artificial Intelligence: A Sourcebook, edited by Derek Partridge and
Yorick Wilks, 167–189. Cambridge: Cambridge University Press.
Yampolskiy, Roman V. 2016. Artificial Superintelligence: A Futuristic Approach.
London: CRC Press.

View publication stats

You might also like