0% found this document useful (0 votes)
32 views21 pages

Bauer Et Al 2023 Expl(Ai)Ned the Impact of Explainable Artificial Intelligence on Users Information Processing

Uploaded by

Catur Huda
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
32 views21 pages

Bauer Et Al 2023 Expl(Ai)Ned the Impact of Explainable Artificial Intelligence on Users Information Processing

Uploaded by

Catur Huda
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 21

INFORMATION SYSTEMS RESEARCH

Vol. 34, No. 4, December 2023, pp. 1582–1602


https://pubsonline.informs.org/journal/isre ISSN 1047-7047 (print), ISSN 1526-5536 (online)

Expl(AI)ned: The Impact of Explainable Artificial Intelligence on


Users’ Information Processing
Kevin Bauer,a,* Moritz von Zahn,b Oliver Hinzb
a
Information Systems Department, University of Mannheim, 68161 Mannheim, Germany; b Information Systems Department, Goethe
University, 60323 Frankfurt am Main, Germany
*Corresponding author
Contact: [email protected], https://orcid.org/0000-0001-8172-1261 (KB); [email protected],
https://orcid.org/0000-0003-1160-1007 (MvZ); [email protected], https://orcid.org/0000-0003-4757-0599 (OH)

Received: June 11, 2021 Abstract. Because of a growing number of initiatives and regulations, predictions of mod­
Revised: June 2, 2022; October 28, 2022 ern artificial intelligence (AI) systems increasingly come with explanations about why they
Accepted: December 17, 2022 behave the way they do. In this paper, we explore the impact of feature-based explanations
Published Online in Articles in Advance: on users’ information processing. We designed two complementary empirical studies
March 3, 2023 where participants either made incentivized decisions on their own, with the aid of opaque
predictions, or with explained predictions. In Study 1, laypeople engaged in the deliberately
https://doi.org/10.1287/isre.2023.1199 abstract investment game task. In Study 2, experts from the real estate industry estimated
listing prices for real German apartments. Our results indicate that the provision of feature-
Copyright: © 2023 The Author(s)
based explanations paves the way for AI systems to reshape users’ sense making of infor­
mation and understanding of the world around them. Specifically, explanations change
users’ situational weighting of available information and evoke mental model adjustments.
Crucially, mental model adjustments are subject to the confirmation bias so that misconcep­
tions can persist and even accumulate, possibly leading to suboptimal or biased decisions.
Additionally, mental model adjustments create spillover effects that alter user behavior in
related yet disparate domains. Overall, this paper provides important insights into potential
downstream consequences of the broad employment of modern explainable AI methods. In
particular, side effects of mental model adjustments present a potential risk of manipulating
user behavior, promoting discriminatory inclinations, and increasing noise in decision mak­
ing. Our findings may inform the refinement of current efforts of companies building AI
systems and regulators that aim to mitigate problems associated with the black-box nature
of many modern AI systems.

History: Alessandro Acquisti, senior editor; Jason Chan, associate editor.


Open Access Statement: This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License. You are free to download this work and share with others,
but cannot change in any way or use commercially without permission, and you must attribute this
work as “Information Systems Research. Copyright © 2023 The Author(s). https://doi.org/10.1287/
isre.2023.1199, used under a Creative Commons Attribution License: https://creativecommons.org/
licenses/by-nc-nd/4.0/.”
Funding: This work was supported by the Deutsche Forschungsgemeinschaft (DFG) (Projek 449023539),
Volkswagen Foundation (ML2MT), and LeibnizInstitute for Financial Research SAFE.
Supplemental Material: The online appendix is available at https://doi.org/10.1287/isre.2023.1199.

Keywords: explainable artificial intelligence • user behavior • information processing • mental models

1. Introduction restricted contestability, and limited accountability (see


Contemporary artificial intelligence (AI) systems’ high Rosenfeld and Richardson 2019 for a review). Having rec­
predictive performance frequently comes at the expense ognized these problems, organizations developing AI
of users’ understanding of why systems produce a certain and governments increasingly adopt principles and regu­
output (Gunning et al. 2019, Meske et al. 2022). For AI sys­ lations (EU 2016, 2021; Google AI 2019; Meta AI 2021)
tems that provide predictions to augment highly conse­ effectively stipulating that AI systems need to provide
quential processes such as hiring decisions (Hoffman meaningful explanations about why they make certain
et al. 2018), investment decisions (Ban et al. 2018), or med­ predictions (Goodman and Flaxman 2017, Cabral 2021).
ical diagnosing (Jussupow et al. 2021), this “black box” In light of these developments, the implementation and
nature can create considerable downsides. These issues use of explainable AI (XAI) methods are becoming more
include impaired user trust, reduced error safeguarding, widespread and mandated by law.

1582
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1583

The purpose of XAI methods is to make AI systems’ that the explanation also, and maybe more importantly,
hidden logic intelligible to humans by answering the affects his processing of currently available information
question: Why does an AI system make the predictions it and his underlying mental models of the determinants of
does? Thereby, XAI methods aim to achieve high predic­ creditworthiness. By changing mental models, explana­
tive performance and interpretability at the same time. tions may even reshape the loan officer’s behaviors in
Many state-of-the-art XAI techniques convey insights related domains beyond the loan approval decision, for
into AI systems’ logic after training and explain beha­ example, assessing the faithfulness of his daughter’s new
viors by depicting the contribution of individual input boyfriend based on the smartphone charging behavior.2
features to the outputted prediction (Doshi-Velez and Considerable challenges arise when trying to answer
Kim 2017). Although there is reason to believe that XAI our research questions. First, measuring how XAI meth­
can mitigate black-box problems (Bauer et al. 2021), the ods affect users’ situational processing of information
pivotal question is how users respond to modern expla­ and mental models is extremely difficult because these
nations, given that the human factor frequently creates cognitive processes are typically unobserved. Second,
unanticipated, unintended consequences even in well- we need to control for possible external cues, unin­
designed information systems (Willison and Warkentin tended stimuli, additionally attainable information, and
2013, Chatterjee et al. 2015). preferences that may affect these cognitive processes in
Nascent research on human-XAI interaction examines any given situation. Third, whether people interact with
how explainability affects humans’ perceptions, attitu­ an (X)AI system, let alone rely on it, is highly endoge­
des, and use of the system, for example, trust (Erlei et al. nous and depends on factors such as culture, technolog­
2020), detection of malfunctioning (Poursabzi-Sangdeh ical literacy, and the socio-technological environment.
et al. 2021), (over)reliance (Bussone et al. 2015), and task Thus, isolating effects associated with the provision of
performance (Senoner et al. 2021). Prior research, how­ explanations in addition to predictions is particularly
ever, does not consider the potential consequences of demanding, if not outright impracticable, in a natural
providing explanations for users’ situational information (organizational) setting. To address these challenges,
processing (the use of currently available information in we rely on two complementary, incentivized experi­
the given situation) and mental models (cognitive repre­ mental studies.
sentations that encode beliefs, facts, and knowledge). By In Study 1 (n � 607), laypeople played a series of
depicting the contribution of individual features to speci­ investment games (Berg et al. 1995), making sequential
fic predictions, feature-based XAI enables users to recog­ economic transaction decisions in an intentionally ab­
nize previously unknown relationships between features stract setting. In Study 2 (n � 153), experts from the
and ground truth labels that the AI system autono­ real-estate industry predicted listing prices for real
mously learned from complex data structures. In that apartments located in Germany. Study 2 extends Study
sense, XAI may constitute the channel through which AI 1 by testing the generalizability of our findings and
systems impact humans’ conceptualization and under­ elaborating on mechanisms driving the results. In both
standing of their environment. This effect could reinforce studies, conditional on the treatment, participants either
the already considerable influence contemporary AI sys­ received no decision support, support from an AI system
tems have on human societies (Rahwan et al. 2019) by, in the form of opaque predictions or an XAI system
for better or worse, allowing human users to adopt sys­ with predictions plus feature-based explanations. We
tems’ inner logic and problem-solving strategies. Despite answer our research questions by eliciting and compar­
the increasing (legally required) implementation of XAI ing changes in both participants’ decision-making pat­
methods, a systematic study of these effects is yet miss­ terns and their beliefs about feature-label relationships.
ing. The paper at hand aims to fill this important gap. The two studies strongly complement each other for
We ask three research questions. Does the additional three reasons. First, laypeople (Study 1) and experts
provision of feature-based explanations affect AI system (Study 2) are the two diametrical archetypes of AI sys­
users’ situational processing of observed information? tem users affected by growing explainability require­
Does it affect users’ underlying mental models? What are ments. Studying both types’ responses to XAI methods
important moderating factors? Consider, for instance, a enables us to identify possibly differential effects and
loan officer who works with an AI system to predict an make inferences about the generalizability of our find­
applicant’s risk parameters and determine the credit ings. Second, we consider two fundamental types of
approval. Because of legal requirements (e.g., Artificial prediction problems where AI systems are frequently in
Intelligence Act; EU 2021), the AI system recently started use: transaction outcome predictions (Study 1) and price
to provide feature-based explanations, showing that it predictions (Study 2) (Ban et al. 2018, Rico-Juan and de
strongly relies on people’s smartphone charging be­ La Paz 2021). Examining the two settings allows us to
havior to predict creditworthiness.1 Although previous understand better whether the interplay between XAI and
research examines how this explanation may affect the cognitive processes is task specific. Third, using local inter­
loan officer’s perceptions of the system, we conjecture pretable model-agnostic explanations (LIME) (Study 1)
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1584 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

and SHapley Additive exPlanations (SHAP) explanations 2021, Molnar 2020). LIME (Ribeiro et al. 2016) and SHAP
(Study 2), the two most popular feature-based XAI meth­ (Lundberg and Lee 2017) provide explanations through
ods (Gramegna and Giudici 2021), allows us to draw additive feature attributions, that is, linear models that
more general conclusions about the interplay between depict the numeric contribution of each feature value to
feature-based explainability and cognitive processes. the overall black box model prediction. Both approaches
Our findings paint a consistent picture: Providing learn these interpretable “surrogate models” on input-
explanations is the critical factor that enables AI systems prediction pairs of the black box model and are applica­
to influence the way people make sense of and leverage ble to virtually all classes of ML models, that is, are model
information, both situationally and more permanently. agnostic. On the individual level, SHAP and LIME
Crucially, we find an asymmetric enduring effect that can provide contrastive explanations that inform users why
foster preconceptions and spill over to other decisions, predictions for a specific instance diverge from the pre­
thereby promoting certain (possibly biased) behaviors. diction for an average instance (Molnar 2020). For
Our paper proceeds as follows. Section 2 presents the­ example, if the SHAP value for the feature Balcony
oretical foundations, whereas Section 3 explains our equals +500 (�200), it indicates that having a balcony
experimental studies and results. Section 4 concludes by marginally increases (decreases) the current apartment’s
discussing our results, the limitations of our work, and listing price prediction by $500 ($200). The big difference
directions for future research. between LIME and SHAP is the way of estimating the
additive feature attributions. LIME creates synthetic,
2. Theory perturbed data points in the local neighborhood of the
observation of interest and fits a weighted linear model
In this section, we first discuss modern XAI methods (Sec­
to explain the relationship between the synthetic data
tion 2.1). Subsequently, we outline the relation between
and the relevant black box predictions. Importantly,
providing explanations and cognitive processes (Section
LIME weights synthetic instances based on their proxim­
2.2) and discuss our work’s contribution to the literature
ity to the original data point. By contrast, SHAP is
(Section 2.3).
inspired by coalitional game theory and treats input fea­
tures as a team of players that cooperate to generate a
2.1. Explainable AI
payoff (the prediction). The method essentially estimates
Following Doshi-Velez and Kim (2017), we conceptual­
the marginal contribution of each player to the overall
ize XAI as methods that possess the ability to present in payoff, Shapley values (Shapley 1953), using a linear model
understandable terms to a human why an AI system that weights instances based on characteristics of coalitions.
makes certain predictions. Over the last couple of years, Given these mathematical differences, the two methods
researchers developed ample XAI methods that help can produce (slightly) different feature attributions for the
elucidate the opaque logic of machine learning (ML)- same instance. However, from the perspective of a user
based AI systems (Ribeiro et al. 2016, Lundberg and Lee who is not familiar with these details, the intuition and
2017, Koh and Liang 2017, Lakkaraju et al. 2019). Very gen­ interpretation of the two methods’ explanations are rea­
erally, XAI methods aim to alleviate problems associated sonably similar (Molnar 2020). Notably, LIME and SHAP
with the black-box nature (e.g., distrust, lack of accountabil­ closely relate to the seminal description of Gregor and Ben­
ity, and error safeguarding) while maintaining a high level basat (1999) of “why and why not explanations” in the con­
of prediction accuracy (Bauer et al. 2021). text of knowledge-based expert systems.
Our study focuses on feature-based XAI methods, With the development of modern explainability meth­
hereafter XAI methods, that can explain the behavior of ods, research on the impact of contemporary XAI on user
any ML-based AI system by showing the contribution behavior has become increasingly essential (Vilone and
of individual features to the prediction. We do so for Longo 2021). Nascent research in this domain typically
several reasons. First, these explanations are the most focuses on how explanations affect user attitudes and reli­
widespread in practice (Bhatt et al. 2020, Senoner et al. ance on the AI system (Lu and Yin 2021). These studies
2021, Gramegna and Giudici 2021). Second, they are produce mixed evidence on the consequences of XAI on
highly intuitive and straightforward to interpret as they decision performance, user trust, perception, and decision-
satisfy most requirements for human-friendly explana­ making performance. Several studies depict that explana­
tions (Molnar 2020). Third, they are typically applicable tions can enhance trust in and positive perceptions of the
to systems using structured and unstructured data (Gar­ system (Rader et al. 2018, Dodge et al. 2019, Yang et al.
reau and Luxburg 2020). Fourth, these methods can ex­ 2020), whereas others provide reversed evidence (Erlei
plain individual predictions, local explainability, which et al. 2020, Poursabzi-Sangdeh et al. 2021). Although
might be the only method legally compliant with (upcom­ prior studies produce important insights regarding the
ing) regulations (Goodman and Flaxman 2017). interplay between XAI and user perceptions, none of
Many researchers recognize two related XAI methods them considers that the additional provision of explana­
as state-of-the-art: LIME and SHAP (Gramegna and Giudici tions may also reshape users’ information processing,
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1585

both situationally and more permanently. For instance, 2017). This mental process might entice people to revise
using SHAP to show the contribution of input features to their expectations and thus make different decisions
a creditworthiness prediction may not only affect a loan because the machine prediction effectively substitutes
officer’s perception of the AI system in use. Instead, she for people’s own mental model driven formation of
may process currently available information about the expectations (Agrawal et al. 2019). However, the black
applicant differently and develop a novel understanding box nature does not allow users to directly compare
of the determinants of creditworthiness, that is, adjust her their underlying beliefs and logic with that of the AI sys­
mental model. With the increasing adoption of explain­ tem. This comparison can only occur when they learn
ability principles by organizations (Google AI 2019, Meta AI how the system combines available information to arrive
2021) and the growing number of regulatory transparency at a prediction. In the previous example, the real estate
requirements (EU 2016, 2021), it is pivotal to understand agent may have access to an XAI system that provides a
how contemporary XAI methods influence cognitive pro­ listing price prediction together with an explanation of
cesses that lie at the heart of people’s knowledge, how specific apartment attributes contribute to it. The
behavior, and problem-solving capabilities. agent can compare the explanation to her own initial per­
ception of the individual attribute contributions to the
2.2. Cognitive Perspective on XAI Employment listing price. As a result, the agent may detect inconsis­
Through feature-based explanations about an AI sys­ tencies that prompt her to revise her logic by putting
tem’s prediction, human users can observe possibly more or less emphasis on specific information currently
unknown feature-label relationships that the system available to evaluate the apartment. This explanation-
learned from complex data structures by itself (Agarwal enabled situational process (Schön 2017) can reconcile
and Dhar 2014, Berente et al. 2021). Although providing the distinct logic that humans and machines apply to
explanations, in general, can have a variety of cognitive arrive at a certain assessment. From this perspective, pro­
effects, researchers across disciplines generally agree viding explanations on top of predictions may constitute
that they primarily enhance people’s understanding of a pivotal factor in allowing users to reflect on how they
someone or something, improve reasoning, and facili­ leverage information to solve a problem and adapt it
tate learning (Gregor 2006, Malle 2006). From a cogni­ according to the AI system’s logic for the given task.
tive perspective, obtaining explanations can entail two Apart from situationally changing cognitive processes
effects: First, it may change people’s situational proces­ that shape the current decision, the interaction between
sing of available information: their use of available mental models and explanations may also yield lasting
information while observing explanations. Second, it effects because mental models possess the dynamic
can lead to an adjustment of their beliefs about feature- capacity to change (Jones et al. 2011). Repeatedly ob­
label relationships the AI system inherently models: serving explanations about how feature X contributes
their mental representation of real-world processes. In to prediction Ŷ and engaging in reflection processes may
this paper, we follow previous work in information sys­ evoke adjustments of the underlying mental model in use.
tems and rely on the “Mental Models Framework” to con­ Following Vandenbosch and Higgins (1996), exposure to
ceptualize relevant cognitive processes (Vandenbosch and external stimuli, here explanations, can lead to two mental
Higgins 1996, Lim et al. 1997, Alavi et al. 2002). model adjustment processes: maintenance and building.
Mental models are “all forms of mental representa­ Under mental model maintenance, people feel encour­
tion, general or specific, from any domain, causal, inten­ aged to maintain or reinforce current beliefs and decision-
tional or spatial” (Brewer 1987, p. 193), encoding beliefs, making rules. This process occurs when they perceive or
facts, and knowledge (Jones et al. 2011). Through imagi­ select new information to fit into their current beliefs and
nary manipulations of model components, people can routines. Under mental model building, individuals
reason and make inferences about how to solve pro­ profoundly restructure or build new mental models in
blems (Rouse and Morris 1986). Much of the people’s response to handling novel, disconfirming information.
decision making is based on these simulations that figu­ As a result of these processes, individuals may adopt dif­
ratively create informal algorithms for carrying out spe­ ferent beliefs about how X contributes to the real label Y,
cific tasks (Johnson-Laird et al. 2017). For instance, real enticing them to process information differently even
estate agents can mentally simulate how listing prices when explanations are no longer present. Put differently:
might change if an apartment for sale had a balcony. users may not merely combine situationally observed
When people perform tasks, they draw on relevant explanations with their own logic to solve a given task.
mental models that guide their processing of incoming Instead, observing the system’s logic may more funda­
information to form expectations and make (expectedly) mentally reshape users’ way of solving problems in gen­
optimal decisions. Working with an AI system that pro­ eral, that is, evoke learning. Therefore, users may exhibit
vides black box predictions, that is, information relevant different problem-solving strategies whenever they draw
to the task, allows people to reflect on their own expecta­ on the explanation-adjusted mental model, even in situa­
tions and compare it to the machine prediction (Schön tions where they do not observe explanations anymore.
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1586 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

In sum, cognitive theories give reason to believe that behaviors and malfunctions (Kaur et al. 2020, Lakkar­
providing explanations in addition to predictions can aju and Bastani 2020). Hence, explanations may be a
influence users’ processing of information about feature security concern if adversaries use perturbations of
X, both situationally and more fundamentally. Because inputs and model attributes to produce intentionally
of the latter effect, modern XAI methods may constitute misleading explanations that manipulate users’ trust
a cornerstone of effective knowledge transfers from and behaviors (Ghorbani et al. 2019). We complement
ML-based AI systems to human users, helping them to this pivotal and insightful work by examining the
learn from the AI how X relates to Y. Hence, explana­ impact of contemporary XAI on users’ situational infor­
tions could facilitate learning machine knowledge: new mation processing and mental models. Understanding
knowledge AI systems autonomously learned from how the provision of explanations about the workings
Big Data and previously missed by domain experts of ML-based AI systems may reshape these cognitive
(Teodorescu et al. 2021, van den Broek et al. 2021). processes is pivotal for anticipating the downstream
consequences of this technology on human societies
2.3. Contribution to the Literature and designing effective transparency and explainability
Our study complements three different streams of liter­ regulations.
ature. The first and most closely related line of work The second set of literature we complement explores
studies the interplay between XAI techniques and user the mechanisms of learning in socio-technological envir­
behavior (see Rosenfeld and Richardson (2019) and onments. A common theoretical foundation builds on
Vilone and Longo (2021) for an overview). About two Bayes rule as a rational benchmark of how humans
decades ago, several studies found that suitably de­ accommodate new information (Holt and Smith 2009).
signed explanations about the functioning and purpose However, research has shown systematic deviations
of legacy knowledge-based expert systems can increase from Bayes’ rule. Reasons include over- or underweight­
users’ trust in the systems, improve users’ perceptions ing of new information (Rabin and Schrag 1999) and a
of the system, and enhance decision-making perfor­ general tendency to asymmetrically discount information
mance (Dhaliwal and Benbasat 1996, Gregor and Benba­ conflicting with prior beliefs while readily internalizing
sat 1999, Ji-Ye Mao 2000, Wang and Benbasat 2007). confirming information (Yin et al. 2016). We complement
However, these expert systems codify knowledge from this research stream by showing how human users devi­
human experts as explicit procedures, instructions, rules, ate from Bayes rule in the context of learning from mod­
and constraints in a digital format. They do not represent ern AI systems. Notably, there exists a limited number of
machine knowledge that modern ML-based AI systems prior research examining how black box predictions
learn independently of domain experts by training on change users’ decision-making habits (Abdel-Karim et al.
large data sets (van den Broek et al. 2021). Given the 2020, 2022; Fügener et al. 2021a, b; Jussupow et al. 2021).
inherent distinctions between expert systems and ML- Relatedly, in a formal model, Agrawal et al. (2019) show
based AI systems in terms of encoded knowledge, con­ that the predictions of black box AI systems can alter
temporary explainability methods present an entirely users’ abilities by providing them with incentives to learn
different form of reasoning to users, namely that of to assess the (negative) consequences of their actions for
machines (Vilone and Longo 2021, Meske et al. 2022). the task supported by the AI.3 None of these studies,
More recent research on the impact of explainability on however, examines the role of feature-based explanations
user behavior mainly focuses on how contemporary in learning, which could pave the way for more funda­
XAI methods impact users’ perceptions of the AI sys­ mental changes in the way users understand real-world
tem. This nascent literature shows that explainability processes. Our paper intends to fill this gap. We study
often improves reliance on and trust in the system (Bus­ how the provision of explanations about how an AI sys­
sone et al. 2015), fairness perceptions (Dodge et al. 2019), tem solves prediction tasks allows users to integrate the
human-AI collaboration (Yang et al. 2020), task efficiency presented machine knowledge into their mental models,
(Senoner et al. 2021), and users’ understanding of the sys­ that is, learn from XAI. A better understanding of how
tem’s malfunctions (Rader et al. 2018). However, there explainability may contribute to machine teaching, the
is also evidence of disadvantages relating to infor­ notion that AI systems first learn novel knowledge that
mational overload (Poursabzi-Sangdeh et al. 2021), experts neither conceive nor anticipate from data and
reduced user trust (Erlei et al. 2020), and overreliance then transfer this knowledge to human users (Abdel-
(Bussone et al. 2015). Moreover, explanations that are Karim et al. 2020), is particularly significant given the
unstable and sensitive even to small perturbations to growing requirements to implement explainability meth­
inputs have the potential to mislead human users into ods when using AI systems.
trusting a problematic black box, for example, by selec­ The third stream of literature we add to studies how
tively providing explanations that conceal biased humans collaborate with computerized systems to solve
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1587

Figure 1. Structure of Empirical Studies

Notes. We provide an overview of the main sequence of our two empirical studies.

problems. Previous research in this area dates back 3. Empirical Studies


decades. Several studies document that humans resist We now present the design and results of Studies 1 and
using computerized decision aids, despite possible per­ 2. In both studies, participants made decisions under
formance benefits (Kleinmuntz 1990), whereas others uncertainty (providing loans and predicting apartment
find that humans possess a strong preference for using listing prices) either with the aid of an opaque AI, an
them (Dijkstra 1999). With the growing employment of explainable AI, or without any support. We paid partici­
modern AI systems in a broad range of domains, the pants according to their decision-making performance
examination of human-machine collaboration has seen to reveal actual preferences and beliefs.4 We implemen­
a considerable resurgence, for example, in the domain ted both studies using oTree, Python, and HTML and
of finance (Ge et al. 2021), medicine (Jussupow et al. ran them online. In Study 1, we recruited 607 partici­
2021), customer service (Schanke et al. 2021), and on- pants on Prolific and let them engage in deliberately
demand tasks (Fügener et al. 2021a). Research on abstract investment games (Berg et al. 1995). Results
“centaur” systems (Goldstein et al. 2017, Case 2018) allow us to observe how the provision of explanations
documents how hybrid human-AI systems (i.e., cen­ on top of predictions shapes information processing
taur systems) achieve superior results in comparison and mental models for laypeople in a very general
with the entities operating independently (Dellermann sequential transaction domain. Study 2 extends the first
et al. 2019, Tschandl et al. 2020), promising consider­ study by testing the generalizability of mental model
able benefits from successful human-AI collaboration.
adjustments regarding the task domain (listing price
Several factors moderate the interaction of humans
predictions), decision-maker expertise, and the explana­
and AI systems including the perceived subjectivity of
tion presentation, and elaborates on important asym­
the task (Castelo et al. 2019, Logg et al. 2019), seeing the
metric effects. With the help of our industry partner, the
system err (Dietvorst et al. 2015), being able to modify
Real Estate Association Germany (IVD), we recruited 153
predictions (Dietvorst et al. 2018), the divergence bet­
experts from the real estate industry to participate in
ween actual and expected predictive performance (Jus­
Study 2. We report the designs and results of the two
supow et al. 2020), and, most importantly for our
studies consecutively. Figure 1 portrays an overview of
research, understanding the system’s internal logic
the experimental designs.
(Gregor and Benbasat 1999, Hemmer et al. 2021). Fol­
lowing our conjecture that explanations pave the way
for AI systems to affect people’s cognitive processes, 3.1. Study 1
contemporary XAI methods introduce another layer of 3.1.1. Design. In Study 1, participants repeatedly en­
complexity in human-AI interaction and its success: gaged in one-shot investment games (Berg et al. 1995)
an interaction between machine and human problem- that possess the following structure. An investor receives
solving strategies. Our work provides novel insights into 10 monetary units (MU). The investor initially observes
whether and under what circumstances people prefer to 10 deliberately abstract borrower characteristics and deci­
rely on their own way of leveraging information or will­ des whether to invest her 10 MU with the borrower. If
ingly adjust it according to machine explanations. In this she does not invest, the game ends without the borrower
sense, our work contributes to the literature on (hybrid) making a decision, and both the investor and borrower
human-AI collaboration by analyzing the underlying earn a payoff of 10 MU. If she invests, the borrower pos­
cognitive processes that may facilitate or hinder the reali­ sesses 30 MU and can keep the whole amount without
zation of the promise of this technology. repercussions. Crucially, the borrower can repay the
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1588 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

investor 10 MU, thereby reciprocating the investor’s ini­ Stage II introduced our treatment variation. Partici­
tial trust. In case of repayment, the investor receives 20 pants made 20 decisions for new random borrowers
MU (we double the amount); otherwise, the investor observing all 10 borrower traits. Additionally, base­
earns 0 MU while the borrower gets 30 MU. The bor­ line participants saw an AI system’s prediction about
rower, in the absence of sufficiently strong social motives, whether borrowers will repay an initial investment.
for example, altruism, egalitarian concerns, or moral pre­ Again, we did not provide intermediary feedback. We
ferences (Miettinen et al. 2020), will not make a repay­ trained the AI system on 1,054 distinct data points col­
ment and maximize his personal income. As a result, the lected in a previous field study, the same data set that
payoff structure of the investment game is of an adver­ the borrowers that participants encounter in the exper­
sarial nature from the investor’s perspective because her iment stem from (see the online appendix for details).7
material well-being is at the mercy of the borrower if The system did not continue to learn during the experi­
she invests. The investor loses her initial investment of ment. Treatment participants, on top of predictions,
10 MU whenever the borrower pursues pure income- observed LIME explanations (Ribeiro et al. 2016) for
maximizing or adversarial motives like wanting to each borrower characteristic, informing them of its con­
minimize the investors’ payoffs. Given this payoff tribution to the repayment prediction. Revealing LIME
structure, an income-maximizing investor in the exper­ values on top of identical predictions constituted the
iment will only invest if (i) her belief that the bor­ treatment variation. As is often the case, we depicted
rower’s motive leads him to repay her is sufficiently LIME values graphically using colored bars of different
strong, and (ii) she ultimately judges that the prospect lengths. Participants received detailed information about
of doubling her income is worth risking the loss of her the model, input features, performance on a representa­
investment.5 Study 1 participants always played as inves­ tive test set, and how to interpret LIME explanations.
tors. Borrowers are subjects from a previous incentivized Stage III perfectly mirrored Stage I. Importantly, par­
field study who had to decide on repayment assuming ticipants engaged with the same borrowers from Stage I
an initial investment; that is, they have already commit­ in random order. We did not draw participants’ atten­
ted to a repayment decision and cannot strategically tion to this fact to alleviate concerns about the experi­
change this choice ex post. We did not provide inter­ menter’s demand effect. The study concluded with a
mediary feedback to prevent the development of idio­ brief questionnaire on socio-economic control variables.
syncratic expertise, experience, or investment strategies
that may confound our results. We randomly matched 3.1.2. Results. Throughout our analyses of Study 1, we
investor and borrower decisions to determine game out­ mainly rely on the following regression model:
comes at the end of the study and pay both according to
the earned MU. Yijs � β1 · Xj + β2 · (Xj × Is ) + β3 · (Xj × Expli )
Study 1 comprised a baseline (AI) and a treatment + β4 · (Xj × Expli × Is ) + γijs + ɛ: (1)
(XAI) condition, each with three stages.6 In Stage I, each
participant made 10 investment decisions for distinct, Yijs is a dummy indicating whether participant i in­
randomly drawn borrowers without intermediary feed­ vested with borrower j in Stage s. Hence, β coefficients
back. They always observed the 10 characteristics of a measure variation in the probability to invest with a bor­
borrower and did not obtain any aid. The idea is that the rower, and Xj is a vector reflecting the 10 observed bor­
10 borrower characteristics allow investors to get an rower traits, the overall prediction, and LIME values.8
idea of the likelihood that an individual borrower will Most relevant to our analyses, Is and Expli are dummy
make a repayment, for whatever motives, and to assess variables, respectively, indicating whether a decision
whether it is worth taking the risk of losing their invest­ takes place in Stage s compared with Stage I (i.e., Stage I
ment. We deliberately chose 10 unintuitive traits corre­ serves as the reference category) and whether partici­
lated with a person’s repayment inclination so that pant i is in the XAI treatment (observes explanations
participants did not possess strong prior beliefs about on top of predictions in Stage II), and γis represents
the informativeness of characteristics for someone’s individual-state fixed effects. We report standardized
repayment behavior (see Table 4 in the online appen­ regression coefficients with robust standard errors. Our
dix). The main reason for choosing just these character­ main interest lies in the interaction terms β3 and β4,
istics is that previous empirical tests have shown that respectively, capturing the isolated effects of observing
they are appropriate features for developing an AI sys­ the prediction and additionally observing LIME explana­
tem that accurately predicts repayment with which par­ tions. As β4 constitutes a difference-in-difference (DiD)
ticipants interact in Stage II. Importantly, participants estimator, it is pivotal to check that before the interven­
learned that the AI system makes predictions based on tion, there are no treatment differences (parallel trends
the same 10 borrower characteristics they also observe, assumption). Regression analyses reveal that baseline
mitigating concerns that they believed the AI system to and treatment participants in Stage I did not place signifi­
have access to more information. cantly different weight on any trait; hence, the use of a
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1589

DiD identification strategy appears generally valid. Nev­ provision of explanations on top of predictions entailed
ertheless, because participants placed significant weight significant weight changes that mirror the relationship
on Gender, Conscientiousness, Neuroticism, and Younger between borrower traits and repayment behavior as
Siblings in only one of the two conditions, there is still depicted by the LIME values. Here, the average magni­
some concern about the appropriate interpretation of tude of absolute weight changes equals 73.9%. Figure
DiD estimates for these traits.9 To avoid drawing incor­ 2(a) shows that the predicted repayment probability
rect conclusions, we conservatively refrain from inter­ markedly decreases (increases) with a borrower’s level
preting these traits’ estimates while still including them of Competitiveness (Patience). Figure 2(b) reveals that
as controls in the model. these are the two traits whose weighting the provision
of explanations significantly fostered: observing expla­
3.1.2.1. Situational Information Processing. We start nations rendered the relationship between a borrower’s
analyzing how participants’ weighting of borrower char­ Competitiveness (Patience) and a participant’s investment
acteristics changed from Stage I to II, that is, changes in likelihood significantly more negative (positive). Nota­
participants’ situational information processing. Figure 2 bly, explanations as such increased the absolute magni­
illustrates our results. Figure 2(a) depicts the average LIME tude of the coefficient for Competitiveness (Patience) by
values (color saturation) participants observed for different 240.0% (94.6%). LIME values reveal that Agreeableness,
feature values (y and x axis). Higher positive (negative) the trait participants initially weighted the most, has
LIME values depict a higher positive (negative) contribu­ almost no impact on the repayment prediction. Accord­
tion of a given feature value to the predicted probability ingly, we find that the provision of explanations led to a
that a borrower makes a repayment. Figure 2(b) portrays significant decrease in the magnitude of the weight parti­
how the provision of predictions and explanations affected cipants placed on this trait (�44.7%). Additional analyses
the weighting of a given borrower trait. The diamond confirm that LIME values for these three characteristics
marker represents the original weighting in Stage I (β1). had a significantly positive influence on participants’
The dashed and solid arrows, respectively, illustrate investment decisions, corroborating the notion that parti­
the isolated effects of observing predictions (β3) and cipants paid attention to and adjusted their weighting of
additional explanations (β4). Depicted results stem from traits according to observed explanations (see Table 11 in
regressions reported in Table 9 in the online appendix. the online appendix). Taken together, participants signifi­
There are two main insights. First, prediction effects cantly adjusted their weighting of information in the
in Figure 2(b) suggest that the provision of opaque pre­ direction of observed explanations for (i) the trait they ini­
dictions generally decreased the weight participants tially perceived as most important and (ii) the traits LIME
placed on observed borrower traits. On average, the highlighted as most important.10 Finally, although not
absolute magnitude of coefficients changed by 63.6%. shown in the Figure 2 for ease of interpretation, regres­
Although only the estimates for Agreeableness, Patience, sion analyses further reveal that explanations signifi­
and Older Siblings are significant, predictions reduced cantly reduced the weight participants placed on the
the absolute magnitude of all variables. Second, the prediction as such (magnitude of coefficient decreased

Figure 2. (Color online) Prediction and Explanation Effects on Situational Information Processing

Notes. We illustrate how the provision of opaque predictions and LIME explanations on top of predictions affect participants situational informa­
tion processing. (a) LIME values (z axis) for different feature values (x axis) participants observed in the study. For the binary feature Older sib­
lings, we show the LIME values for No and Yes at the outer limits of the continuous feature scale. (b) Estimated prediction and explanation
effects, respectively, of β3 and β4 in Model (1) with s � 2. Initial values represent β1. We denote significance levels by *p < 0.1, **p < 0.05, and ***p
< 0.01.
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1590 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

by 26.8%); that is, they were less likely to follow a predic­ solid arrows, respectively, show how having observed
tion that a borrower makes a repayment.11 predictions (β3) and explanations on top of predictions
(β4) did fundamentally alter participants’ information
Result 1.1. Observing explanations changed participants’
processing, that is, mental models.
situational processing of the overall prediction and borrower
Observing opaque predictions did not result in a sig­
traits that explanations or they themselves consider most
nificant change in participants’ weighting of borrower
important. The direction of adjustments mirrors explanations.
traits. By contrast, depicted results suggest that providing
Result 1.1 agrees with our theoretical elaborations: explanations did entail an adjustment of mental models
People adjust their situational information processing in with the absolute magnitude of coefficients changing by
response and according to explanations they currently 61.8% on average. Importantly, this adjustment was
observe. Notably, elicited expectations about the predic­ asymmetric. Observing explanations led participants to
tion accuracy did not differ significantly for predictions place significantly more weight on borrowers’ Competi­
with or without explanations (71.8% and 70.6%, respec­ tiveness (+148.6%) and Patience (+59.4%) in Stage III than
tively; p � 0.751, Wilcoxon rank-sum test). Therefore, in Stage I. The weight changes again mirror the observed
changes in the weighting of predictions do not seem to LIME explanations. After observing explanations that the
result from lower performance expectations. Next, we AI system places the most weight on borrowers’ Competi­
test the conjecture that explanations affect beliefs about tiveness and Patience, participants increased their weight­
the relationship between borrower characteristics and ing of these attributes even for investment decisions
repayment behavior, that is, mental models. where they no longer observed explanations. Intrigu­
ingly, we do not find that explanations about the low rel­
3.1.2.2. Mental Model Adjustments. We compare par­ evance of Agreeableness led participants to adjust their
ticipants’ information weighting across Stages I and III marked weighting of this trait significantly. Although par­
to test the conjecture that explanations affect mental mod­ ticipants weighted Agreeableness significantly less while
els about the relationship between borrower traits and observing explanations, they returned to their original
repayment behavior. We rely on the regression model weighting of it once they lost access to the XAI system.
(1), setting s � 3 and excluding controls for the prediction Naturally, one may wonder about this asymmetry’s ori­
and LIME values. Figure 3 illustrates regression results gins. One plausible interpretation is that explanations are
that we report in Table 12 in the online appendix. less likely to evoke pronounced mental model adjust­
Figure 3 portrays how the provision of predictions ments when they conflict with strong preconceptions. Put
and explanations lastingly changed the weighting of a differently, people are more inclined to engage in mental
given borrower trait across Stages I and III, where parti­ model maintenance rather than building because it is less
cipants had no (X)AI aid. The diamond marker depicts cognitively demanding and creates less psychological dis­
the original weighting in Stage I (β1). The dashed and tress (Vandenbosch and Higgins 1996). In Stage I, partici­
pants put by far the most emphasis on a borrower’s
Agreeableness to decide on investing. LIME values, how­
Figure 3. Mental Model Adjustments
ever, suggested that this conception is incorrect because it
is among the least relevant predictors for borrowers’
repayment inclination. Although one would expect that
participants engaged in mental model building to reshape
their beliefs about the relationship between Agreeableness
and repayment behavior, we do not find significant
adjustments. For Competitiveness (Patience), explanations
depicted an important negative (positive) influence,
which, given their initial weighting of it, confirmed par­
ticipants’ prior beliefs. Following the Mental Models
framework, confirming explanations should evoke the
maintenance or reinforcement of prior beliefs. Given
the significant explanation effects, it seems that partici­
pants willingly engaged in this process. This inclination
to engage in mental model maintenance rather than
building more generally concurs with the frequently
Notes. We depict participants’ mental model adjustments as mea­ documented confirmation bias (Yin et al. 2016), that is,
sured by their change in the weighting of borrower traits across the tendency to selectively process information in a
Stages I and III. The estimated prediction and explanation effects
respectively represent β3 and β4 in Model (1) with s � 3. Initial values
way that allows for the continuation or strengthening
represent β1. We denote significance levels by *p < 0.1, **p < 0.05, and of beliefs. We elaborate on this issue in Study 2 and the
***p < 0.01. discussion.12
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1591

Result 1.2. Machine explanations entailed asymmetric men­ of the cases. As a result, treatment and baseline partici­
tal model adjustments. Participants reinforced priors that pants respectively achieved a decision accuracy of 51.7%
explanations confirmed but did not abandon priors that and 57.2% (�9.6%, p < 0.01, F test) for most competitive
explanations markedly contradicted. borrowers and 59.5% and 62.8% (�5.3%, p < 0.05, F test)
for other borrowers. Notably, participants already associ­
3.1.2.3. Investment Performance. Thus far, it remains ated very high competitiveness with a low repayment
open how providing explanations on top of predictions likelihood in Stage I: Most competitive borrowers re­
affected participants’ decision-making performance in ceived an investment in 56.3% of the cases, whereas all
our setting. Table 1 summarizes participants’ perfor­ others did so in 69.5% of the cases (there do not exist
mance measured by the accuracy (share of payoff maxi­ treatment differences). Against this background, expla­
mizing decisions) and recall (share of investments with nations seem to have exacerbated this inaccurate pat­
repaying borrowers). We also report p values of F tests tern16 to an extent that treatment participants made
to illustrate significant treatment differences.13 significantly worse decisions than before. Put differently,
Although there are no differences in Stage I, treatment confirming explanations inappropriately reinforced pre­
participants performed significantly worse than base­ conceptions about most competitive borrowers not re­
line ones in Stage II (�8.9% and �11.0% for accuracy paying an investment in our setting.
and recall, respectively).14 Treatment participants’ rela­
tively lower performance in Stage II stems from not Result 1.3. Participants excessively increased the isolated
investing with the most competitive borrowers (with weighting of a trait they already believe to be evidence
most negative LIME values), whereas the overall predic­ against repayment. This reaction inefficiently decreased
tion implies doing so, that is, from overruling positive participants’ likelihood to invest with repaying borrowers
predictions.15 that were highly competitive.
They overruled positive predictions and refrained In sum, the results for Study 1 are highly consistent
from investing in 46.5% of these cases, resulting in a with the notion that the provision of explanations cre­
decision accuracy of merely 53.5%. Baseline partici­ ates a novel channel through which AI systems may
pants, for most competitive borrowers, overruled posi­ reshape users’ way of processing information, both situ­
tive predictions only in 21.2% of the cases and achieved ationally and more permanently. For the latter effect,
a decision accuracy of 78.9%; that is, they are 47.5% we observe an asymmetry that is reminiscent of a confir­
more likely to make an income maximizing decision mation bias and, in our setting, decreased participants’
than treatment participants. For all other borrowers, decision-making performance by excessively reinfor­
treatment (baseline) participants overruled positive pre­ cing inaccurate preconceptions.
dictions and made optimal decisions in 23% (19.4%) and
69.6% (71.1%) of the cases, respectively. Hence, treat­
ment participants seem to have placed too much weight 3.2. Study 2
on very high competitiveness, leading them to overrule The goal of Study 2 is twofold. First, we extend Study 1
the overall prediction inefficiently often. results by testing the generalizability of mental model
Examining Stage III, we find that this overweighting of adjustment findings regarding the task domain, user
the highest competitiveness level persisted even when expertise, and explanation presentation and examining
participants did not observe explanations anymore (see whether the asymmetry we found for explanation-driven
Table 13 in the online appendix). In Stage III, treatment mental model adjustments in Study 1 is indeed a manifes­
(baseline) participants invested with most competitive tation of the confirmation bias. Second, we explore if men­
borrowers in 44.7% (54.7%, p < 0.01, F test) of the cases tal model adjustments spill over to related but disparate
and with other borrowers in 68.2% (67.6%, p < 0.7, F test) domains.

Table 1. Investment Performance Across Stages

Stage I (no aid) Stage II (with aid) Stage III (no aid)

Accuracy Recall Accuracy Recall Accuracy Recall

Baseline (AI) (%) 60.3 64.9 63.1 64.6 62.7 65.1


Treatment (XAI) (%) 60.7 67.4 57.5 57.5 56.5 60.2
F test: Baseline versus treatment p � 0.79 p � 0.31 p < 0.01*** p < 0.01*** p < 0.02** p < 0.04**
Notes. We depict participants’ investment performance as measured by their accuracy (share of payoff maximizing decisions) and recall (share
of investments with repaying borrowers) in Stages I, II, and III. We report results separately for baseline (AI) and treatment (XAI) participants. F
tests reveal the significance of treatment differences per measure and stage.
*p < 0.1; **p < 0.05; and ***p < 0.01.
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1592 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

3.2.1. Design. Study 2 comprises four consecutive stages, importantly, the apartment was in a midsize city in east­
where recruited real estate experts estimated the list­ ern Germany (Chemnitz). For historical, demographic,
ing price per square meter in Euros of apartments that and socioeconomic reasons, Chemnitz is very different
we previously collected from a large online plat­ from “A-cities” such as Frankfurt and Cologne, so the
form.17 Participants saw 10 apartment characteristics housing market is also very different. Germans in gen­
to make an informed guess and did not receive inter­ eral and real estate agents in particular are usually
mediate feedback. To reduce the task complexity and aware of this East-West disparity.20 The study con­
avoid informational overload, we fixed seven apart­ cluded with a questionnaire on participants’ socio-
ment characteristics across all stages, that is, apartments demographics.
only differed regarding the same three characteristics:
Location (Frankfurt/Cologne), Balcony (Yes/No), and 3.2.2. Results. We report our results in three steps.
Green voter share in the district (Below city average/City First, we outline the experts’ belief adjustments from
average/Above city average).18 We provide screenshots Stage I to Stage III. Second, we examine the occurrence
of the interfaces from each stage in the online appendix. of confirmation bias in these adjustment processes.
In Stage I, we elicited participants’ initial beliefs about Finally, we analyze experts’ listing price estimates in
the relationship between the three variable apartment Stage IV.
characteristics and listing prices. Participants estimated
the listing price of four random apartments with differ­ 3.2.2.1. Mental Model Adjustments. Figure 4 shows
ent combinations of the variable attributes by entering the distribution of absolute differences between experts’
their marginal contributions to the price using a slider. beliefs about the marginal contribution of the three vari­
Sliders ranged from minus to plus 2.500e in steps of 50e. able attributes before and after the treatment interven­
We initially set the marginal contributions and overall tion. We show results for the NoAid, AI, and XAI
price estimation to 0e and the average listing price conditions. The distributions for the NoAid and AI con­
(9,600e), respectively. Participants additionally stated ditions are remarkably similar and skewed toward zero,
their confidence in the entered marginal contributions indicating that experts frequently did not adjust beliefs.
and the resulting price estimation on a five-point scale. The distribution for XAI participants is considerably less
Stage II introduced our treatment variations. In all right-skewed; that is, they adjusted their beliefs across
variations, participants estimated listing prices for eight Stages I and III more. On average, NoAid, AI, and XAI
random apartments with different combinations of vari­ participants’ absolute belief adjustments equaled 166.4e,
able attributes they did not encounter in Stage I. In con­ 165.4e, and 299.1e, respectively. Only the differences
trast to Stage I, participants directly entered the estimated between NoAid versus XAI (+79.7%, p < 0.01, F test), and
listing price. As a reference point, they again observed the AI versus XAI (+80.8%, p < 0.01, F test) conditions are sta­
average listing price for an apartment. Participants stated tistically significant (see Table 24 in the online appendix),
their confidence on a five-point scale. In our baseline con­ that is, observing explanations led to remarkably stron­
dition (NoAid), participants estimated the price without ger adjustments of beliefs. Our notion is that real estate
any aid. Participants in the AI condition observed opaque experts updated initially held mental models about the
listing price predictions of a steady, that is, nonlearning, relationship between apartment attributes and listing
AI system trained on 4,975 collected observations.19 In prices as they encountered SHAP explanations. Contrast­
our XAI condition, in addition to observing these predic­ ing our first study, we directly measure participants’
tions, participants also saw numerically presented SHAP prior and posterior beliefs about the contribution of dis­
values for the three variable apartment characteristics, tinct apartment characteristics to listing prices in Study 2.
that is, marginal contributions to the prediction in Euros. This design facet enables us to estimate mental model
After they entered all eight listing price estimates, parti­ adjustments directly, leveraging the accepted framework
cipants in treatments with decision support filled out a by DeGroot (1974). Specifically, we assume that agent i’s
survey containing items on their trust, degree of reli­ posterior belief about the relationship of characteristic j
ance, and perceived transparency of the AI system (and and the listing price Posti,j � ai,j · Priori,j + (1 � ai,j ) · Expli,j
explanations). is a weighted combination of the corresponding prior
Stage III replicated Stage I to measure posterior beliefs. belief Priori,j and the personally observed explanation
Independent of the condition, participants again made Expli,j ; 1 � ai,j represents the extent of belief adaptation in
decisions without any aid for the same apartments. the direction of the explanation, whereas ai,j describes the
Finally, in Stage IV, participants estimated the listing anchoring of the previous belief. For instance, in the
price for one last apartment without any decision aid. extreme case of 1 � ai,j � 1, individual i completely aban­
Across participants, we varied the balcony and green dons her prior mental model and adopts the observed
voter attribute of the apartment, whereas the seven fixed explanation as her new one. We estimate the weights
attributes were identical to the previous listings. Most (1 � ai ) and ai for our three study conditions using a
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1593

regression model comprising treatment interactions that Table 2. Posterior Belief Formation
has the following form:
Dependent variable: Posterior belief (1) (2)
Posijk � β1 · Priijk + β2 · (AIi × Priijk ) + β3 · (Expli × Priijk ) Prior belief (β1) 0.634*** 0.782***
+ β4 · SVij + β5 · (AIi × SVij ) + β6 · (Expli × SVij ) (0.060) (0.063)
Prior belief × AI (β2) 0.070 �0.027
+ γi + δk + ɛ: (2) (0.104) (0.084)
Prior belief × Expl. (β3) �0.276*** �0.240***
The variables Posijk and Priijk, respectively, represent (0.084) (0.075)
expert i’s posterior and prior beliefs about attribute j’s Avg. SHAP (β4) 0.025 0.033
contribution to apartment k’s listing price in Euros. (0.040) (0.039)
Avg. SHAP × AI (β5) 0.078 0.083
Most importantly, AIi is a dummy variable indicating (0.053) (0.050)
that expert i observed a prediction, whereas the dummy Avg. SHAP × Expl. (β6) 0.265*** 0.249***
Expli equals one if a participant additionally observed (0.053) (0.052)
explanations; SVij represents the average SHAP value Fixed effects No Yes
for apartment attribute j of the apartments participant i N 1,836 1,836
R2 0.740 0.787
encountered in Stage II; and γi and δk are expert and
apartment controls, respectively. Notes. We depict results from OLS regression models with robust
standard errors reported in parentheses. The dependent variable
On an individual level, Model (2) estimates how equals participants’ posterior belief about the marginal contribution
observed SHAP values affected participants’ adjustments of apartment attributes to the listing price in euros. The main
of beliefs about the relationship between a given charac­ independent variables of interest are participants’ prior beliefs, the
teristic and the listing price. It enables us to quantify the average SHAP values for apartment attributes in Stage II, a dummy
indicating that participants observed a prediction in Stage II (AI), a
“stickiness” of prior beliefs (β1 � β3 ) and “gravitational dummy indicating that participants observed explanations in Stage II
pull” of explanations (β4 � β6 ) and directly test the occur­ (XAI), and interaction terms. We further control for the overall
rence of confirmation bias. Importantly, this estimation is posterior listing price participants entered for the apartment and its
interaction with treatment dummies, and the average prediction they
only possible for Study 2, where we elicited prior and observed in Stage II. In column (2), we additionally include
posterior beliefs about distinct feature-label relationships. individual and apartment fixed effects.
In Study 1, we measured the ultimate investment deci­ *p < 0.1; **p < 0.05; and ***p < 0.01.
sions only and observed belief changes indirectly through
changes in those decisions. As a result, we cannot individ­
ually quantify the impact of observed explanations on (unsurprisingly) have no significant explanatory power
regarding posterior beliefs (see β4 and β5).21 When parti­
specific beliefs nor can we analyze confirmation bias: a
cipants did not obtain machine aid or only observed
key contribution of our second study.
predictions, their prior and posterior beliefs were more
Table 2 depicts regression results for Model (2). Re­
than 60% positively correlated (β1 and β2); that is, parti­
sults show that in our NoAid and AI conditions where
cipants barely adjusted their beliefs. Only when partici­
participants did not observe explanations, SHAP values
pants observed explanations in addition to predictions
did the displayed SHAP values have positive, statisti­
Figure 4. (Color online) Distribution of Absolute Belief
cally significant effects. β6 reveals that XAI participants
Changes significantly adjusted their beliefs in the direction of
observed explanations. According to the estimate, pos­
terior beliefs resembled SHAP values more closely
in the XAI treatment condition compared with the
NoAid and AI conditions (approximately +25 percent­
age points). Observing explanations also caused XAI
participants’ posterior beliefs to resemble their prior
significantly less (β3), that is, prior beliefs became less
“sticky” compared with the NoAid and AI conditions
(approximately �25 percentage points). In sum, these
results suggest that observing SHAP explanations led
participants to adjust their beliefs in the direction of
explanations and abandon their priors. This insight
corroborates our Result 1.2 in Study 1 on an individual
level, revealing that explanation-driven mental model
Notes. We depict the distribution of experts’ absolute belief adjust­
ments across Stages I and III. We aggregate the belief adjustments
adjustments also occur for experienced experts, who
over all apartment attributes. Different distributions show results sep­ are arguably familiar with apartment traits and listing
arately for NoAid, AI, and XAI participants. price predictions.22
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1594 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

3.2.2.2. Confirmation Bias. In Study 1, we observed Table 3. Confirmation Bias and Posterior Belief Formation
asymmetric mental model adjustments that are reminis­
Dependent (1) (2) (3)
cent of the confirmation bias. The design of Study 2
variable: Low High
allows us to test for confirmation bias in mental model Posterior belief Overall confidencebeliefs confidencebeliefs
adjustment processes more directly by examining
whether XAI participants’ adjustments depended on Prior belief 0.492*** 0.483*** 0.496***
(0.091) (0.105) (0.136)
the alignment of explanations and prior beliefs. Avg. SHAP 0.303*** 0.344*** 0.145**
We define that explanations confirmed an expert’s (0.043) (0.055) (0.067)
preconception about the price contribution of a specific Confirm 12.039 �10.838 115.724
apartment attribute if the prior and the observed aver­ (27.949) (39.552) (73.702)
age SHAP value for the corresponding attribute have Avg. SHAP 0.166*** 0.107 0.301***
× Confirm (0.059) (0.077) (0.094)
the same sign. With this definition, observed explana­ N 708 481 222
tions confirm prior beliefs in 49.6% of the cases.23 We R2 0.746 0.725 0.843
analyze differences in belief adjustments with respect to
Notes. We depict results from OLS regression models with individual
confirming and conflicting explanations using a modi­ and apartment fixed effects. We report robust standard errors reported
fied version of Model (2). Specifically, we are interested in parentheses. The dependent variable equals XAI participants’
in whether the convergence of XAI participants’ poste­ posterior belief about the marginal contribution of apartment attributes
to the listing price in euros. The main independent variables of interest
rior beliefs toward observed SHAP values only occurred
are participants’ prior beliefs, the average SHAP values for apartment
when explanations confirmed prior beliefs. Therefore, attributes in Stage II, a dummy indicating that observed SHAP values
we focus on the subsample of XAI participants allowing in Stage II confirmed participants’ priors, measured by an equal sign of
us to omit treatment dummies and interaction terms prior beliefs and average SHAP values for a given attribute, and
interaction terms. We further control for the overall posterior listing
which facilitates the interpretation of results. Along the price participants entered for the apartment and the average prediction
lines of Model (2), we regress XAI participants’ posterior they observed in Stage II. Column (1) presents results for all decisions.
beliefs about the relationship between apartment charac­ Columns (2) and (3) respectively depict results for the shares of
teristics and the listing price on their prior beliefs and decisions where XAI participants report low and high confidence in
their prior.
observed SHAP values. Most importantly, we now add *p < 0.1; **p < 0.05; and ***p < 0.01.
a dummy variable (Confirm) indicating whether explana­
tions confirmed prior beliefs and its interaction with
average SHAP values and prior beliefs as independent 2009). To test the existence of such heterogeneity, we
variables. The interaction Avg. SHAP × Confirm will pro­ consider experts’ reported confidence in prior beliefs
vide insights into whether the influence of observed and define that an expert possessed low (high) confi­
SHAP values on belief adjustments depended on the dence in a prior, if, on a five-point scale, they reported a
alignment of explanations and prior beliefs, which are confidence level of less than 4 (at least 4). In columns (2)
insights we cannot obtain from Study 1 using Model (1). and (3) of Table 3, we, respectively, repeat the regression
Corroborating our interpretation of Result 1.2 from analysis reported in column (1) for the subsamples of
Study 1, we find that explanation-driven belief adjust­ low- and high-confidence prior beliefs.
ment processes depended on whether explanations con­ Reported estimates provide further evidence that
firmed or conflicted with prior beliefs. The estimate for explanation-enabled mental model adjustments were
the interaction term Avg. SHAP × Confirm is positive subject to confirmation bias. According to the estimated
and statistically significant (see column (1) in Table 3). coefficient of Avg. SHAP × Confirm, for low-confidence
Following the estimate, posterior beliefs resembled ob­ priors, the influence of observed SHAP values on poste­
served SHAP values significantly more closely (about rior beliefs did not depend on whether explanations
50% more) if they confirmed their prior beliefs. Hence,
confirmed prior beliefs (see column (2)). Considering
consistent with confirmation bias, the belief adjustment
the positive and significant estimate of Avg. SHAP, the
was asymmetric regarding the confirmatory nature of
belief updating was in line with Bayes rule. By con­
explanations. If participants had updated beliefs ratio­
trast, for high-confidence priors, belief adjustments
nally according to Bayes rule, the interaction term should
were highly sensitive to whether SHAP values confirmed
be insignificant as Bayesian observers would not weight
priors (see column (3)). The estimate for Avg. SHAP ×
explanations conditional on their alignment with prior
Confirm suggests that the magnitude of the adjustment of
beliefs (Rabin and Schrag 1999).
high-confidence priors was about two times larger when
To elaborate on the notion that these asymmetric
observed explanations were in line with them.
belief adjustments are a manifestation of confirmation
bias, we further consider the role of experts’ confidence Result 2.1. Study 1 findings extend to expert users, SHAP
in their prior beliefs. Prior research shows that confirma­ explanations, and the domain of apartment price predic­
tion bias is strongest for entrenched beliefs (Pyszczynski tions: SHAP explanations led real estate experts to adjust
and Greenberg 1987, Knobloch-Westerwick and Meng prior beliefs about the relation between apartment attributes
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1595

and listing prices. Adjustment processes were subject to the Specifically, compared with observing no predictions
confirmation bias. (opaque predictions), observing explained predictions
decreased Chemnitz price estimates by 33.9% (38.9%) if
3.2.2.3. Spillover Effects. Although we observe that the share of green voters was low and increased price
real estate experts (asymmetrically) adjusted prior be­ estimates by 16.5% (32.4%) if the share of green voters
liefs, all previously reported results pertain to the same was high. As one might expect, the direction of the differ­
market: Participants observed SHAP explanations for ence in experts’ evaluation of the green voter share attri­
the same two A-cities in Western Germany, for which bute is in line with explanations observed in Stage II:
we elicited prior and posterior beliefs. What remains SHAP values indicated that in Frankfurt and Cologne, a
open is whether explanation-driven belief adjustments high (low) share of green voters marginally contributes
spilled over to the listing price estimation for apart­ to listing prices by about +652e (�613e). We do not find
ments in different markets. We put this idea to the test any effect for experts who only observed opaque predic­
by examining the distribution of participants’ final price tions in Stage II.
predictions for an apartment in a medium-sized eastern To elaborate on these findings, we also perform a
German city that is not an “A city”: Chemnitz.24 median split and analyze the subsamples of experts
Figure 5 shows the distribution of listing price esti­ whose average absolute belief adjustment for the attri­
mates conditional on the share of green voters in the dis­ bute “Green voter” is below and above the median.
trict for NoAid, AI, and XAI participants. The results Consistent with the idea that belief spillover effects
indicate that observing explanations impacted partici­ drive differences in listing price estimates in Chemnitz,
pants’ price estimates for Chemnitz apartments in neigh­ experts who strongly adjusted their beliefs about the rel­
borhoods with high and low proportions of green voters. evance of “Green voters” from Stage I to III drive our
Figure 5(a) shows that the distribution of listing prices for aggregate-level results. We do not find significant treat­
an apartment in a district with a low green voter share is ment differences in the accuracy of participants’ listing
considerably more right-skewed for XAI than NoAid or price estimates as measured by the absolute deviation
AI participants; that is, they estimate relatively low prices from actual prices. Nevertheless, our results show that
more frequently. NoAid, AI, and XAI participants on using XAI as a decision support tool in one market can
average estimated a listing price of 4,752e, 5,141e, and affect aggregate listing prices in another market in an eco­
3,140e, respectively. Only the differences between NoAid nomically considerably way (average absolute change:
versus XAI and AI versus XAI are statistically significant approximately 20%), which is not the case for opaque
in regression analyses (p < 0.05, F test, for both). The dis­ systems. This result demonstrates that XAI methods can
tribution of price estimates in districts with high shares of link disparate decision-making tasks.
green voters has a stronger left-skew for XAI participants
Result 2.2. Pronounced explanation-driven belief adjust­
than their NoAid and AI counterparts (Figure 5(b)). On
ments spill over to experts’ listing price estimation in a fun­
average, NoAid, AI, and XAI participants estimated a
damentally different market.
listing price of 5,231e, 4,600e, and 6,092e, respectively, for
an apartment in a district with a high percentage of green In summary, our results from Study 2 (i) demonstrate
voters. Again, we only find significant explanation effects the robustness of our results from Study 1 on mental
(p < 0.1, F test, for both). These results reveal the eco­ model adjustments in terms of system user expertise,
nomic significance in the changes of price distributions. explanation representation, and decision domain; (ii)

Figure 5. (Color online) Price Distributions in Chemnitz

Notes. We depict the distribution of experts’ listing price estimates in Chemnitz. (a) and (b) Price distribution for apartments in a district with a
low and high share of green voters, respectively. Different distributions show results separately for NoAid, AI, and XAI participants.
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1596 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

provide strong evidence that explanation-driven mental From a theoretical perspective, our results contribute
model adjustments are subject to confirmation bias; and to our understanding of the role of popular XAI meth­
(iii) show that explanation-driven mental model adjust­ ods in effective knowledge transfers from ML-based AI
ments generate significant spillover effects. systems to human users. A key promise of modern AI
systems is that the application of ML techniques will
4. Discussion and Conclusion discover new knowledge from Big Data that has previ­
We report results from two empirical studies that pro­ ously eluded even experienced experts (Berente et al.
vide novel insights into the interplay between the use of 2021, van den Broek et al. 2021). This “machine knowl­
feature-based XAI methods and users’ cognitive pro­ edge” is typically codified in the form of a complex pre­
cesses. Our main contribution is the identification of dictive model that outperforms humans. We show that
considerable side effects of providing feature-based providing predictions alone is insufficient to achieve
explanations, the most popular form of XAI methods, systematic knowledge transfers from AI systems to
on users’ situational information processing and mental human users. In both our studies, neither laymen nor
models. We find that the latter effect (i) is subject to the experts adapted their understanding of the relation­
confirmation bias so that misconceptions can persist ships between features X and label Y according to
and even accumulate, possibly leading to suboptimal “machine knowledge” when observing only opaque
decisions, and (ii) can create spillover effects into other predictions. Merely in treatments where users also had
access to explanations, they began to adapt their app­
decision domains. These overarching results suggest that
roach to solving the task so that it more closely matched
the growing, partially legally required, use of feature-
the strategy of the AI system. Therefore, XAI methods
based XAI methods opens a new channel through which
appear to be a pivotal factor contributing to an effective
AI systems may fundamentally reshape the way humans
channel through which AI systems can pass on their
understand real-world relationships between features X
self-learned knowledge to human users. Crucially,
and target variables Y. In the following, we discuss our
feature-based XAI methods seem to induce an asymme­
results, present implications for organizations and soci­
try in mental model adjustments: users adjust their
ety, and, based on the limitations of our studies, provide
beliefs more in the direction of observed explanations if
directions for future research.
they confirm rather than disconfirm their priors. This
asymmetry contradicts with the updating behavior of a
4.1. Discussion of Results
Bayesian observer who would neither over- nor under­
Study 1 demonstrates that the provision of explanations
weight explanations conditional on them confirming or
can situationally lead lay users to adjust their weighing disconfirming prior beliefs. This asymmetry occurred
of features accordingly, the average absolute change in regardless of whether we provide graphically visual­
estimates equals 73.9%, and to put less emphasis on the ized LIME or numerically represented SHAP explana­
overall prediction (�26.8%). Explanations also evoked tions. It therefore seems as if additive feature-based
asymmetric changes in lay users’ conceptions about the explanations more generally evoke cognitive processes
relationship between borrower traits and repayment leading users to learn from the machine selectively.
inclinations that influence behaviors even when they do Researchers across disciplines commonly refer to such an
not observe explanations anymore, the average absolute asymmetry as confirmation bias (Yin et al. 2016). Study 2
change in estimated coefficients equals 61.8%; that is, provides consistent evidence that explanation-driven
explanations affect mental models. Explanation-driven knowledge transfers from an AI to a human similarly suf­
effects decreased lay users’ decision-making perfor­ fer from confirmation bias as knowledge transfers in the
mance in our setting. Compared with opaque predic­ human-to-human domain. For example, confidence in
tions, explanations decreased participants investment prior conceptions and their difference from the new
performance by 8.9% while observing them and by 9.8% information moderate confirmation bias (Pyszczynski
even when not observing explanations anymore. Study and Greenberg 1987). Similar to learning from other
2 extended these results in three ways. First, we find humans, users seem unwilling to internalize potentially
that even expert users in a considerably more applied helpful, XAI-channeled machine knowledge if it is in­
domain adjusted mental models by about 25 percentage consistent with what they already, perhaps incorrectly,
points. Second, results indicate that asymmetric mental believe to be true. From the perspective of the Mental
model adjustments were a manifestation of the confirma­ Models framework, individuals more frequently engage
tion bias because posterior beliefs resembled observed in maintaining rather than in building mental models of
explanations about 50% more closely if explanations con­ the relationships between features and labels. One reason
firmed prior beliefs. Third, Study 2 reveals that mental for this effect could be the need to attain or maintain a
model adjustments created spillover effects leading to an high level of self-esteem (Klayman 1995), leading users
average absolute change in apartment price estimates for to focus inappropriately on explanations that make
a different market by approximately 20%. them feel competent. In other words, they may derive
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1597

a positive intrinsic benefit from being in the right from those estimated by experts who either had no deci­
(Gilad et al. 1987). From this perspective, people may sion aid or only observed opaque predictions. This spill­
misuse the XAI as a tool to enhance their self-esteem. If over effect seems to occur because of the adjustment of
left unaddressed, the asymmetric adaptation of mental mental models that experts draw on in both situations.
models by humans may prevent modern (X)AI appli­ Therefore, as an unintended side effect, increasing public
cations from fulfilling their promise of making humans and private efforts to promote the use of XAI methods
smarter, which (ironically) may also hinder the further may extend the already significant influence of AI sys­
development of AI applications by humans. tems from areas where we interact with them (Rahwan
Interpreting our results in the light of the model by et al. 2019) to areas where such systems are not in use.
Agrawal et al. (2019) yields another theoretical insight Feature-based XAI methods’ potential to link different
regarding the ramifications of XAI. Our results indicate domains is particularly concerning given recent evidence
that users’ willingness to follow XAI predictions de­ on their susceptibility to intentional manipulation and
pends on whether the explanations conform with their adversarial attacks (Lipton 2018). Many modern XAI
mental models. One way to rationalize this behavior is methods, including LIME and SHAP, optimize fidelity,
that their objective function includes a component that that is, ensure that explanations accurately mimic the pre­
accounts for experiencing some positive (negative) in­ dictions of the black box model. However, even small
trinsic utility when obtaining a signal that their mental perturbations of the input data (e.g., deliberate manipula­
model may (not) be accurate (Festinger 1962, Gilad et al. tion and measurement errors) can lead to considerably
1987, Harmon-Jones 2019). In the model by Agrawal different explanations for identical predictions, that is,
et al. (2019), AI systems make predictions about uncer­ depict different feature-label relations (Ghorbani et al.
tain states of the world that relate to the profitability of 2019, Lakkaraju and Bastani 2020). The potential instabil­
taking specific actions. Human users, in turn, assess the ity of explanations allows manipulating user behaviors.
expected payoffs associated with specific actions, that is, Following our results, the creation of misleading explana­
make judgments. Our results suggest that human judg­ tions may not only affect users’ trust in the AI system
ment in this model encompasses not only the material (Lakkaraju and Bastani 2020) but also lead to an (asym­
consequences of an action but also the psychological metric) adjustment of mental models that affect users’
impact of receiving a signal that implicitly shows decision making beyond the XAI augmented decision at
whether current mental models are correct. If expla­ hand. Specifically, the depiction of certain feature-label
nations reveal that the AI system arrived at a prediction relationships that are not present can evoke inappropri­
in a way that contradicts their held mental models, tak­ ate mental model adjustments that, given the documen­
ing an action that follows this prediction effectively con­ ted asymmetry, will cause users who already believe
stitutes a signal to oneself that the current mental model these patterns to be true, to feel vindicated and reinforce
is incorrect, creating psychological distress, for example, these beliefs. In general, the documented spillover effects
in the form of a cognitive dissonance (Harmon-Jones may magnify the reach and impact of intentional manip­
2019). This mental toll may lead users not to follow the ulations of explanations, increasing deceiving parties’
prediction in the first place. Conversely, users may fol­ incentive to do so.
low unreliable predictions more often if the explanations
are consistent with their current mental models because 4.2. Implications
doing so provides a psychologically valuable self-signal Reported results have important practical implications
that they are in the right (Gilad et al. 1987). Against this for organizations and policymakers. Our finding that
background, users’ inclination to follow predictions of XAI can change human thinking points to potential pit­
an XAI system, and thus their ultimate decisions and falls for companies that want, or have to, use XAI. Con­
gains, is subject to greater variance than with a black-box sider a company that plans to implement XAI methods
AI. That is because users’ propensity to follow predic­ to explain to its employees why an AI system makes
tions depends on the consistency of the explanations certain predictions. As Study 1 shows, providing exp­
with their mental models. lanations in addition to predictions may draw users’
Another theoretical contribution of our work is to attention excessively to the explanations, to the detri­
show the potential of feature-based XAI to link different ment of the prediction itself. Users may place too much
decision domains by influencing users’ beliefs about the emphasis on individual explanations that confirm their
feature-label relationship. Study 2 results show that prior beliefs, rather than adhering to the overall predic­
observing explanations for listing price predictions for tion. As a result, employees’ decision-making perfor­
apartments in Market A influenced the price estimation mance for the task at hand may deteriorate, which is in
of experts in a different Market B, where the learned pat­ line with evidence from related research (Poursabzi-
tern does not exist, and they did not have access to XAI Sangdeh et al. 2021). In domains where explanations are
decision support. We find that listing prices estimated by becoming a regulatory standard, managers need to take
experts who observed explanations differed significantly such potential downsides into account and contemplate
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1598 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

the ramifications of implementing explainability mea­ From a societal perspective, our results indicate that
sures. Following our results, managers who, in the broad, indiscriminate implementation of XAI methods
future, are obliged to put XAI methods in place, should may create unintended downstream ramifications. Our
not take these steps too lightly. From a business perspec­ finding that XAI can lead users to adjust mental models
tive, our documented downsides of explainability could in a confirmatory way and carry over learned patterns
render the continued use of AI-based decision support to other domains may, in an extreme case, foster dis­
systems unattractive. Considering that AI systems are crimination and social divisions. Assume all recruiters
often deeply interwoven with business processes, this start to collaborate with an XAI system to support hiring
XAI-driven discontinuance may entail considerable orga­ decisions. For example, a subgroup of recruiters may dis­
nizational change. As a result, managers may be well criminate against women because they believe female
advised to assess potential inconsistencies between the applicants to be less productive on the job. If the XAI
AI system’s internal logic and employees’ understanding (occasionally) provides local explanations that depict
of the task it supports before rolling out explainability being female as negative evidence for high future per­
measures. This puts managers in a position to evaluate formance, the subgroup that statistically discriminates
the magnitude of the potential downside of explainability based on gender will readily reinforce its prior belief,
and use countermeasures. For example, managers may that is, engage in mental model maintenance. As a
obviate confirmation bias by openly discussing explana­ result, these recruiters may become more biased and
tions that conflict with employees’ mental models and less noisy in their behavior as they hire female appli­
showcasing arguments in support of the explanation. cants consistently less. Given the spillover effects we
Another pitfall for companies concerns the transfer of find, they may even carry over their strengthened con­
knowledge from AI systems to human users. As Study 2 ceptions about women’s productivity to other jobs, fur­
shows, even experts can overgeneralize learned feature- ther reinforcing discriminatory patterns. Additionally,
label relationships that are only applicable in the context because nondiscriminating recruiters will most likely
in which they interact with the system. With the confir­ refrain from adjusting their mental model, that is, not
matory learning from explanations, existing differences engage in mental model building, social divisions among
in employees’ initial conceptions may lead to differ­ recruiters may develop and accumulate along the lines of
ences in how they collaborate with and what they learn gender biases. Hence, without any malicious intent, the
from the XAI, for example, fostering the biased weight­ broad use of XAI may ironically foster human discrimi­
ing of certain information. From this perspective, pro­ natory tendencies and divide social groups. Notably,
viding explanations might decrease individual level with the possibility to manipulate explanations, deceiv­
noise in the decision-making process (Kahneman et al. ing third parties could also intentionally cause explana­
2021) because individuals’ decisions become more con­ tions to exhibit specific prediction contributions for
sistent. This is in line with Fügener et al. (2021b), who sensitive attributes such as race, gender, or age. This
find decisions to be increasingly consistent among users effect could lead human users who already hold preju­
engaging with opaque predictions. On a more aggregate dices, stereotypes, or discriminatory tendencies to rein­
level, however, our results suggest that explained pre­ force their views, which could promote certain political
dictions may additionally foster differences in the agendas.
decision-making process across subgroups of users
that possess heterogeneous priors. As a consequence, 4.3. Limitations and Future Research
the variation of decisions on a group level can grow. As with any other research study, ours is not without
As pointed out by Kahneman et al. (2021), variation in limitations. In light of increasing regulatory require­
decisions can substantially contribute to errors and ments and private initiatives, we believe that these lim­
ultimately harm business performance. Consider our itations open up fruitful avenues for future research.
previous example of loan officers. XAI may cause loan One limitation of our work concerns the lack of feed­
approval decisions to increasingly depend on the par­ back on the decision outcomes and thus the perfor­
ticular employee, with idiosyncratic mental models, mance of the AI system. In both our studies, we did not
assessing the applicant’s creditworthiness. This in­ provide feedback for two reasons. First, it adds a consid­
crease in loan approval variation may create consider­ erable layer of complexity that impedes the measure­
able business, legal, and reputational risks. Against ment and interpretation of isolated explanation-driven
this background, managers should closely monitor the effects on users’ cognitive processes. Second, in practice,
introduction of XAI to identify a possible increase in many AI-supported decisions do not yield immediate
decision variance. For instance, managers could com­ feedback, or only yield feedback for some of the predic­
plement XAI with “noise audits” and the development tions. Hence, users have to interact with the system
of “reasoned rules” (as proposed by Kahneman et al. without learning its prediction accuracy, at least for a
2021) to overcome the hidden costs of XAI-driven certain period. Examples include hiring decisions sup­
increases in inconsistent decision making. ported by an on-the-job performance predicting AI
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1599

system, investment decisions supported by a return Additionally, the global representation may facilitate
predicting AI system, and drug treatment decisions comparison and reflection processes that ultimately
supported by an effectiveness predicting AI system. Con­ improves the transfer of knowledge from the AI sys­
sequently, explanations may alter users’ situational infor­ tem to the user.
mation processing and mental models before feedback
on system performance arrives. Nonetheless, we strongly 4.4. Conclusion
encourage future research to examine the role of feed­ A concluding remark is worth making. Of course, our
back as it may introduce unexpected dynamics in the work is not meant to be an argument, let alone a plea,
cognitive effects we document. For instance, the (selec­ against making “black box” AI systems more explain­
tive) reinforcement of their mental models through exp­ able or transparent. Instead, we comprehend our find­
lanations, may lead users to be more forgiving and ings as a warning that the indiscriminate use of modern
maintain trust in the AI system, even if they eventually XAI methods as an isolated measure may lead to unin­
see it making mistakes. In this way, the interaction tended, unforeseen problems because it creates a new
between feedback and explanations might constitute a channel through which AI systems can affect human
factor contributing to unwarranted algorithm apprecia­ behaviors across domains. The pervasive human incli­
tion (Logg et al. 2019), leading users to rely on incorrect nation to process information in a way that confirms
outcomes blindly. Additionally, people’s adjustments of their preconceptions while ignoring potentially helpful
the situational information processing and existing men­ yet conflicting information needs addressing if exp­
tal models possibly depend on the extent to which the lainability is to become an effective means to combat
XAI system’s predictions outperform their own. If users accountability, transparency, and fairness issues with­
learn that an XAI system’s predictions perform consider­ out creating adverse second-order effects. For instance,
ably better than their subjective ones, the magnitude of one might restrict the provision of explanations of sen­
reported confirmation biases may vary. Conversely, sitive features for end users of the system and only use
when users’ predictions are better than the XAI, their them to ensure the proper and unbiased functioning of
confirmation bias might be even stronger. Future re­ the AI system during the development process. Addi­
search could examine to what extent our reported tionally, it might be important to provide developers
effects, at the intensive margin, depend on users’ per­ and data scientists with cognitive awareness trainings
ceptions about differences in their own and the XAI to make them more sensitive to their own biased men­
system’s predictive performance. tal processes.
Another limitation of our work originates from letting
participants interact with local, feature-based XAI meth­ Endnotes
ods. We opted to use these explanations because they 1
For anecdotal evidence of such nontraditional data use, see Lend
are already widely in use in practice and because there doEFL.com or https://money.cnn.com/2016/08/24/technology/
are arguments that feature-based explanations on an lenddo-smartphone-battery-loan/index.html.
individual level are necessary to comply with (upcom­ 2
On a high level, both decisions effectively constitute sequential
ing) regulatory requirements (Goodman and Flaxman economic transactions under uncertainty that strongly depend on
2017). Yet, there exist other forms of explanations, for trust.
3
example, global feature-based explanations or even Explainability may enter the model of Agrawal et al. by changing
example-based explanations. Although an investiga­ the prediction reliability. Following Proposition 2, the necessity for
providing explanations decreases with the users’ judgment. How­
tion and comparison of the interplay between different ever, the model does not consider the idea presented in our paper
forms of explanations and cognitive processes are that explainability may also affect users’ understanding of the pro­
beyond the scope of this paper, it is worthwhile for cess that determines the uncertain state of the world the AI tries to
future research to explore whether, and if so why, the predict. One could integrate this notion into the framework by
effects we document would change if users (addition­ modeling that explanations affect users’ judgment capabilities by
influencing beliefs about underlying processes. Extending the
ally) obtain other forms of explanations. Consider, for
model of Agrawal et al. in this direction may be a fruitful endeavor
instance, global explanations. Although local explana­ to better understand whether explainability modulates the relation­
tions help understand why an AI system produces a ship between prediction and judgment. However, an extension of
prediction on a case-by-case basis, global explanations the formal model is beyond the scope of this paper and left for
reveal important high-level patterns and nonlinearities future research.
4
in the system’s logic. Such global explanations effec­ See the online appendix for details on the experimental proce­
tively aggregate individual-level information for the dures including payments, instructions, and screenshots.
5
user and help to understand the system’s overall logic. When a risk-neutral, purely self-interested investor expects that the
borrower repays her with a probability of p > 0.5, for example, because
By taking over this information aggregation task, global
she believes the borrower to possess altruistic, efficiency, or fairness
explainability could mitigate concerns about the selec­ preferences, they have a strict incentive to invest because they maxi­
tive processing of isolated local explanations that argu­ mize their expected earnings. Importantly, holding such expectations
ably contribute to the occurrence of confirmation bias. about the borrower’s preferences is justified and frequently observed
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1600 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

in sequential games: A considerable share of people does respond (+64.3%). As a result, participants could have benefited from relying
reciprocally in sequential exchanges if they are trusted (see Miettinen on the predictions, which baseline participants did at least partially.
et al. 2020 for an overview). 15
Across Stages I and II, baseline participants’ access to the AI system
6
To reduce the complexity for the reader, we only report the three significantly increased the accuracy by 4.6% (p < 0.01, F test), whereas
main stages of the experiment. Right before and after Stage II, we the recall effectively remained constant (p < 0.82, F test). XAI partici­
additionally measured participants’ prior and posterior preferences pants performance significantly decreased regarding both the accuracy
to observe three borrower characteristics. We use these measures as (�5.3%; p < 0.01, F test) and recall score (�14.6%; p < 0.01, F test).
robustness and consistency checks. We provide a detailed descrip­ 16
A purely linear distinction between most competitive and other
tion of these measurements in the online appendix. borrowers does not allow to draw conclusions about their repay­
7
The questionnaire items included in the field study were selected ment likelihood: they, respectively, made a repayment in 77.4% and
partly for exploratory reasons and partly motivated by previous 79.8% of the cases (p � 0.85, F test).
research documenting their association with individuals’ repayment 17
We scraped data from a large online platform in February 2022.
behavior in investment games (Ben-Ner and Halldorsson 2010). We collected observations for all apartments listed for sale in the
8
For most traits, values and LIME values are almost perfectly corre­ seven major cities of Germany (“A-Cities”) and a medium-sized
lated producing severe problems of multicollinearity (see Table 7 in eastern German city (Chemnitz). We constructed a data set consist­
the online appendix). Therefore, in our regression analyses, we only ing of eight apartment attributes and the listing price directly
include LIME explanations for which there exists a tolerable correla­ obtained from the platform and two additionally collected features
tion between the trait and LIME values: Openness, Agreeableness, from public statistics. We provide summary statistics in the online
and Conscientiousness. appendix (Table 6).
9 18
See Table 8 in the online appendix. We selected these three characteristics for technical reasons regard­
10 ing the ML model and based on the input from our industry partner.
These results do not allow us to isolate how explanations affect
The notion is that these characteristics together are (i) sufficiently rele­
what investors consider to be a borrower’s motivation to repay
vant to the prediction and (ii) familiar/accessible to experts.
them or not. The change in the weighting of competitiveness could
19
stem from a reinforced perception that competitiveness predicts a The AI system is a random forest that achieves a performance of
low repayment likelihood because it proxies for antisocial, income- R2 � 0:72 on unseen test data. See the online appendix for additional
maximizing, or relative income-maximizing motives. Although we information.
cannot isolate investors’ latent belief(s) about borrowers’ motives, 20
For instance, A-cities exhibit considerably higher average wages,
our results effectively show that the provision of explanations does more liberal political attitudes, and faster population growth (Cajias
entail a change in at least one of these perceived latent motives, that et al. 2020).
is, that XAI can change the processing of information. A similar 21
The positive coefficient for β5 may be related to the fact that
argument applies regarding mental model adjustments outlined SHAP values and overall predictions are inextricably linked. Merely
later. observing high (low) predictions may lead to adjustments of
11
Reported results are robust to excluding participants who always reported beliefs upward (downward), creating a positive, however,
or never invested in our analyses, respectively, alleviating concerns insignificant correlation with underlying SHAP values in the data.
that our results are driven by pure altruists or players who always 22
Participants, on average, have worked in the real estate industry
choose the game-theoretically dominant strategy (see the subsection for 13.8 years and, on a scale from 1 to 10, report that their experi­
on additional robustness checks in the online appendix). Instead, ence level in rating apartment listing prices is 5.7.
our results stem from those participants whose behavior suggests 23
that they try to invest with borrowers whom they believe will make Our main insights are robust to defining more restrictively that
a repayment, that is, individuals who, from a conceptual point of explanations confirm priors if the absolute distance between the
view, should be most inclined to learn to recognize repaying bor­ prior and the observed average SHAP value is smaller than the
rowers. Results 1.2 and 1.3 are equally robust to excluding these absolute distance between the prior and 0e and, at the same time,
“extreme” types, warranting a similar interpretation. smaller than the absolute distance between the prior and the closest
12
extreme, that is, 62,500e (see Table 25 in the online appendix).
The significant explanation effect for Openness and Extraversion 24
We did not include Chemnitz observations in the data to train
may be a consequence of participants’ significantly stronger weight­
the AI model. We conducted several analyses showing that the
ing of borrowers’ Competitiveness and Patience and a limited capacity
most important predictors for listing prices in Frankfurt and
to process information. Specifically, XAI participants in Stage III
Cologne (cities in Stages I to III) differ considerably from listing
place similarly low weight on all borrower traits but Competitiveness,
price predictors in Chemnitz. Real estate experts are arguably aware
Agreeableness, and Patience. This pattern may suggest that partici­
of the structural differences in apartment markets.
pants heuristically focus on the three characteristics that they them­
selves and the AI system deemed most relevant to the decision. As
a result, they place less weight on all other traits, which for Open­
ness led to a statistically significant effect. References
13
Abdel-Karim BM, Pfeuffer N, Carl V, Hinz O (2022) How AI-based
We show ROC curves in Figures 17 to 19 in the online appendix. systems can induce reflections: The case of AI-augmented diag­
14
Participants neither knew their own nor the AI system’s perfor­ nostic work. Management Inform. Systems Quart. Forthcoming.
mance because we did not provide intermediate feedback. There­ Abdel-Karim BM, Pfeuffer N, Rohde G, Hinz O (2020) How and
fore, they could not see how much better or worse the system what can humans learn from being in the loop? German J. Artifi­
performs compared with themselves. Although unknown to partici­ cial Intelligence 34(2):199–207.
pants, predictions are accurate in about 69.3% of the cases. This per­ Agarwal R, Dhar V (2014) Big data, data science, and analytics: The
formance holds equally for both repaying (69.7%) and nonrepaying opportunity and challenge for IS research. Inform. Systems Res.
borrowers (67.7%). Participants in Stage I correctly invested with (non- 25(3):443–448.
)repaying borrowers in 66.1% (41.2%) of the cases and overall in 60.5% Agrawal A, Gans JS, Goldfarb A (2019) Exploring the impact of arti­
of the cases. Put differently, the AI system outperforms them overall ficial intelligence: Prediction vs. judgment. Inform. Econom. Pol­
(+14.5%) and especially for the identification of nonrepaying ones icy 47:1–6.
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s) 1601

Alavi M, Marakas GM, Yoo Y (2002) A comparative study of dis­ EU (2021) Proposal for a regulation EU of the European Parliament
tributed learning environments on learning outcomes. Inform. and of the Council of April 21, 2021, laying down harmonised
Systems Res. 13(4):404–415. rules on artificial intelligence (Artificial Intelligence Act) and
Ban GY, El Karoui N, Lim AE (2018) Machine learning and portfolio amending certain Union legislative acts. Official J. Eur. Union
optimization. Management Sci. 64(3):1136–1154. Law 119. https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=
Bauer K, Hinz O, van der Aalst W, Weinhardt C (2021) Expl(AI)n it celex%3A52021PC0206.
to me: Explainable AI and information systems research. Bus. Festinger L (1962) Cognitive dissonance. Sci. Amer. 207(4):93–106.
Inform. Systems Engrg. 63(2):79–82. Fügener A, Grahl J, Gupta A, Ketter W (2021a) Cognitive challenges
Ben-Ner A, Halldorsson F (2010) Trusting and trustworthiness: in human–artificial intelligence collaboration: Investigating the path
What are they, how to measure them, and what affects them. J. toward productive delegation. Inform. Systems Res. 33(2): 678–696.
Econom. Psych. 31(1):64–79. Fügener A, Grahl J, Gupta A, Ketter W (2021b) Will humans-in-the-
Berente N, Gu B, Recker J, Santhanam R (2021) Managing artificial loop become borgs? Merits and pitfalls of working with AI. Man­
intelligence. Management Inform. Systems Quart. 45(3):1433–1450. agement Inform. Systems Quart. 45(3b):1527–1556.
Berg J, Dickhaut J, McCabe K (1995) Trust, reciprocity, and social Garreau D, Luxburg U (2020) Explaining the explainer: A first theo­
history. Games Econom. Behav. 10(1):122–142. retical analysis of LIME. Proc. Internat. Conf. on Artificial Intelli­
Bhatt U, Xiang A, Sharma S, Weller A, Taly A, Jia Y, Ghosh J, et al. gence and Statist.
(2020) Explainable machine learning in deployment. Proc. Conf. Ge R, Zheng Z, Tian X, Liao L (2021) Human–robot interaction:
on Fairness, Accountability, and Transparency (Association for When investors adjust the usage of robo-advisors in peer-to-peer
Computing Machinery, New York). lending. Inform. Systems Res. 32(3):774–785.
Brewer WF (1987) Schemas vs. mental models in human memory. Ghorbani A, Abid A, Zou J (2019) Interpretation of neural networks
Morris P, ed. Modelling Cognition (John Wiley & Sons, Oxford, is fragile. Proc. AAAI Conf. on Artificial Intelligence. 33(1):
UK), 187–197. 3681–3688.
Bussone A, Stumpf S, O’Sullivan D (2015) The role of explanations Gilad B, Kaish S, Loeb PD (1987) Cognitive dissonance and utility maxi­
on trust and reliance in clinical decision support systems. Proc. mization: A general framework. J. Econom. Behav. Organ. 8(1):61–73.
Internat. Conf. on Healthcare Informatics (Institute of Electrical Goldstein IM, Lawrence J, Miner AS (2017) Human-machine collab­
and Electronics Engineers (IEEE), New York). oration in cancer and beyond: The centaur care model. JAMA
Cabral TS (2021) AI and the right to explanation: Three legal bases Oncology 3(10):1303–1304.
under the GDPR. Data Protection Artificial Intelligence 13:29–56. Goodman B, Flaxman S (2017) European Union regulations on algo­
Cajias M, Freudenreich P, Freudenreich A, Schäfers W (2020) rithmic decision-making and a “right to explanation”. AI Maga­
Liquidity and prices: A cluster analysis of the German residen­ zine 38(3):50–57.
tial real estate market. J. Bus. Econom. 90(7):1021–1056. Google AI (2019) Responsible AI practices: Interpretability. Accessed
Case N (2018) How to become a centaur. J. Design Sci. https://jods. March 8, 2022, https://ai.google/responsibilities/responsible-ai-
mitpress.mit.edu/pub/issue3-case/release/6?version=53b19e72- practices/?category=interpretability.
d43a-4eda-8c48-6ed3cdc03218. Gramegna A, Giudici P (2021) SHAP and LIME: An evaluation of
Castelo N, Bos MW, Lehmann DR (2019) Task-dependent algorithm discriminative power in credit risk. Frontiers Artificial Intelligence
aversion. J. Marketing Res. 56(5):809–825. 4:752558.
Chatterjee S, Sarker S, Valacich JS (2015) The behavioral roots of Gregor S (2006) The nature of theory in information systems. Man­
information systems security: Exploring key factors related to agement Inform. Systems Quart. 30(3):611–642.
unethical IT use. J. Management Inform. Systems 31(4):49–87. Gregor S, Benbasat I (1999) Explanations from intelligent systems:
DeGroot MH (1974) Reaching a consensus. J. Amer. Statist. Assoc. Theoretical foundations and implications for practice. Manage­
69(345):118–121. ment Inform. Systems Quart. 23(4):497–530.
Dellermann D, Ebel P, Söllner M, Leimeister JM (2019) Hybrid intel­ Gunning D, Stefik M, Choi J, Miller T, Stumpf S, Yang GZ (2019)
ligence. Bus. Inform. Systems Engrg. 61(5):637–643. XAI—explainable artificial intelligence. Sci. Robot 4(37):eaay7120.
Dhaliwal JS, Benbasat I (1996) The use and effects of knowledge- Harmon-Jones EE (2019) Cognitive Dissonance: Reexamining a Pivotal
based system explanations: Theoretical foundations and a frame­ Theory in Psychology (American Psychological Association).
work for empirical evaluation. Inform. Systems Res. 7(3):342–362. Hemmer P, Schemmer M, Vössing M, Kühl N (2021) Human-AI
Dietvorst BJ, Simmons JP, Massey C (2015) Algorithm aversion: Peo­ complementarity in hybrid intelligence systems: A structured
ple erroneously avoid algorithms after seeing them err. J. Exper­ literature review. Proc. 28th Pacific Asia Conf. on Inform. Systems.
iment. Psych. General 144(1):114–126. Hoffman M, Kahn LB, Li D (2018) Discretion in hiring. Quart. J.
Dietvorst BJ, Simmons JP, Massey C (2018) Overcoming algorithm Econom. 133(2):765–800.
aversion: People will use imperfect algorithms if they can (even Holt CA, Smith AM (2009) An update on Bayesian updating. J.
slightly) modify them. Management Sci. 64(3):1155–1170. Econom. Behav. Organ. 69(2):125–134.
Dijkstra JJ (1999) User agreement with incorrect expert system Ji-Ye Mao IB (2000) The use of explanations in knowledge-based
advice. Behav. Informs. Tech. 18(6):399–411. systems: Cognitive perspectives and a process-tracing analysis.
Dodge J, Liao QV, Zhang Y, Bellamy RK, Dugan C (2019) Explaining J. Management Inform. Systems 17(2):153–179.
models: An empirical study of how explanations impact fairness Johnson-Laird PN, Goodwin GP, Khemlani SS (2017) Mental models
judgment. Proc. Internat. Conf. on Intelligent User Interfaces. and reasoning. The Routledge International Handbook of Thinking
Doshi-Velez F, Kim B (2017) Toward a rigorous science of interpret­ and Reasoning (Routledge, Abingdon-on-Thames, UK), 346–365.
able machine learning. Preprint, submitted March 2, https:// Jones NA, Ross H, Lynam T, Perez P, Leitch A (2011) Mental mod­
arxiv.org/abs/1702.08608. els: An interdisciplinary synthesis of theory and methods. Eco­
Erlei A, Nekdem F, Meub L, Anand A, Gadiraju U (2020) Impact of logical Soc. 16(1). https://www.jstor.org/stable/26268859#metadata_
algorithmic decision making on human behavior: Evidence info_tab_contents.
from ultimatum bargaining. Proc. AAAI Conf. on Human Com­ Jussupow E, Benbasat I, Heinzl A (2020) Why are we averse toward
put. and Crowdsourcing. algorithms? A comprehensive literature review on algorithm
EU (2016) Regulation EU 2016/679 of the European Parliament and aversion. Proc. Eur. Conf. on Inform. Systems.
of the Council of 27 April 2016, article 22. Official J. Eur. Union Jussupow E, Spohrer K, Heinzl A, Gawlitza J (2021) Augmenting
Law 119:59. medical diagnosis decisions? An investigation into physicians’
Bauer, von Zahn, and Hinz: Explainable AI and Users' Information Processing
1602 Information Systems Research, 2023, vol. 34, no. 4, pp. 1582–1602, © 2023 The Author(s)

decision-making process with artificial intelligence. Inform. Sys­ Rader E, Cotter K, Cho J (2018) Explanations as mechanisms for
tems Res. 32(3):713–735. supporting algorithmic transparency. Proc. CHI Conf. on Human
Kahneman D, Sibony O, Sunstein CR (2021) Noise: A Flaw in Human Factors in Comput. Systems.
Judgment (Little, Brown). Rahwan I, Cebrian M, Obradovich N, Bongard J, Bonnefon JF, Brea­
Kaur H, Nori H, Jenkins S, Caruana R, Wallach H, Wortman zeal C, Crandall JW, et al. (2019) Machine behaviour. Nature
Vaughan J (2020) Interpreting interpretability: Understanding 568(7753):477–486.
data scientists’ use of interpretability tools for machine learn­ Ribeiro MT, Singh S, Guestrin C (2016) “Why should I trust you?”
ing. Proc. CHI Conf. on Human Factors in Comput. Systems. Explaining the predictions of any classifier. Proc. ACM SIGKDD
Klayman J (1995) Varieties of confirmation bias. Psych. Learning Internat. Conf. on Knowledge Discovery and Data Mining.
Motives 32:385–418. Rico-Juan JR, de La Paz PT (2021) Machine learning with explain­
Kleinmuntz B (1990) Why we still use our heads instead of formu­ ability or spatial hedonics tools? An analysis of the asking
las: Toward an integrative approach. Psych. Bull. 107(3):296. prices in the housing market in Alicante, Spain. Expert Systems
Knobloch-Westerwick S, Meng J (2009) Looking the other way: Appl. 171:114590.
Selective exposure to attitude-consistent and counterattitudinal Rosenfeld A, Richardson A (2019) Explainability in human–agent
political information. Comm. Res. 36(3):426–448. systems. Autonomic Agent Multi Agent Systems 33(6):673–705.
Koh PW, Liang P (2017) Understanding black-box predictions via Rouse WB, Morris NM (1986) On looking into the black box: Pro­
influence functions. Proc. Internat. Conf. on Machine Learn. spects and limits in the search for mental models. Psych. Bull.
Lakkaraju H, Bastani O (2020) “How do I fool you?” Manipulating 100(3):349.
user trust via misleading black box explanations. Proc. AAAI/ Schanke S, Burtch G, Ray G (2021) Estimating the impact of
ACM Conf. on AI, Ethics, and Society. “humanizing” customer service chatbots. Inform. Systems Res.
Lakkaraju H, Kamar E, Caruana R, Leskovec J (2019) Faithful and cus­ 32(3):736–751.
tomizable explanations of black box models. Proc. AAAI/ACM Schön DA (2017) The Reflective Practitioner: How Professionals Think in
Conf. on AI, Ethics, and Society. Action (Routledge, Abingdon-on-Thames, UK).
Lim KH, Ward LM, Benbasat I (1997) An empirical study of com­ Senoner J, Netland T, Feuerriegel S (2021) Using explainable artifi­
puter system learning: Comparison of co-discovery and self- cial intelligence to improve process quality: Evidence from
discovery methods. Inform. Systems Res. 8(3):254–272. semiconductor manufacturing. Management Sci. 68(8):5704–5723.
Lipton ZC (2018) The mythos of model interpretability: In machine Shapley LS (1953) A value for n-person games. Contributions to the
learning, the concept of interpretability is both important and Theory of Games (AM-28), vol. II (Princeton University Press,
slippery. Queue 16(3):31–57. Princeton, NJ).
Logg JM, Minson JA, Moore DA (2019) Algorithm appreciation: Teodorescu MH, Morse L, Awwad Y, Kane GC (2021) Failures of
People prefer algorithmic to human judgment. Organ. Behav. fairness in automation require a deeper understanding of human-
Human Decision Processes 151:90–103. ML augmentation. Management Inform. Systems Quart. 45(3b):
Lu Z, Yin M (2021) Human reliance on machine learning models 1483–1499.
when performance feedback is limited: Heuristics and risks. Tschandl P, Rinner C, Apalla Z, Argenziano G, Codella N, Halpern
Proc. CHI Conf. on Human Factors in Comput. Systems. A, Janda M, et al. (2020) Human–computer collaboration for
Lundberg SM, Lee SI (2017) A unified approach to interpreting model skin cancer recognition. Nature Medicine 26(8):1229–1234.
predictions. Proc. Conf. on Neural Inform. Processing Systems. van den Broek E, Sergeeva A, Huysman M (2021) When the
Malle BF (2006) How the Mind Explains Behavior: Folk Explanations, machine meets the expert: An ethnography of developing
Meaning, and Social Interaction (MIT Press, Cambridge, MA). AI for hiring. Management Inform. Systems Quart. 45(3):
Meske C, Bunde E, Schneider J, Gersch M (2022) Explainable artifi­ 1557–1580.
cial intelligence: Objectives, stakeholders, and future research Vandenbosch B, Higgins C (1996) Information acquisition and men­
opportunities. Inform. Systems Management 39(1):53–63. tal models: An investigation into the relationship between
Meta AI (2021) Facebook’s five pillars of responsible AI. Accessed behaviour and learning. Inform. Systems Res. 7(2):198–214.
March 8, 2022, https://ai.facebook.com/blog/facebooks-five- Vilone G, Longo L (2021) Notions of explainability and evaluation
pillars-of-responsible-ai/. approaches for explainable artificial intelligence. Inform. Fusion
Miettinen T, Kosfeld M, Fehr E, Weibull J (2020) Revealed prefer­ 76:89–106.
ences in a sequential prisoners’ dilemma: A horse-race between Wang W, Benbasat I (2007) Recommendation agents for electronic
six utility functions. J. Econom. Behav. Organ. 173:1–25. commerce: Effects of explanation facilities on trusting beliefs. J.
Molnar C (2020) Interpretable Machine Learning: A Guide for Making Management Inform. Systems 23(4):217–246.
Black Box Models Explainable. Accessed January 14, 2022, https:// Willison R, Warkentin M (2013) Beyond deterrence: An expanded
christophm.github.io/interpretable-ml-book. view of employee computer abuse. Management Inform. Systems
Poursabzi-Sangdeh F, Goldstein DG, Hofman JM, Wortman Vaughan Quart. 37(1):1–20.
JW, Wallach H (2021) Manipulating and measuring model inter­ Yang F, Huang Z, Scholtz J, Arendt DL (2020) How do visual expla­
pretability. Proc. CHI Conf. on Human Factors in Comput. Systems. nations foster end users’ appropriate trust in machine learning?
Pyszczynski T, Greenberg J (1987) Toward an integration of cognitive Proc. Internat. Conf. on Intelligent User Interfaces.
and motivational perspectives on social inference: A biased Yin D, Mitra S, Zhang H (2016) Research note—When do consumers
hypothesis-testing model. Adv. Experiment. Soc. Psych. 20:297–340. value positive vs. negative reviews? An empirical investigation
Rabin M, Schrag JL (1999) First impressions matter: A model of con­ of confirmation bias in online word of mouth. Inform. Systems
firmatory bias. Quart. J. Econom. 114(1):37–82. Res. 27(1):131–144.

You might also like