Decoding AIs Nudge A Unified Framework To Predict
Decoding AIs Nudge A Unified Framework To Predict
net/publication/379284314
CITATION READS
1 13
3 authors, including:
Zhuoyan Li Zhuoran Lu
Purdue University Purdue University
9 PUBLICATIONS 22 CITATIONS 14 PUBLICATIONS 119 CITATIONS
All content following this page was uploaded by Zhuoyan Li on 04 April 2024.
10083
The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)
this gap by proposing such a computational framework1 . tion (Green and Chen 2019; Guo et al. 2019; Zhang,
Specifically, inspired by Callaway, Hardy, and Grif- Liao, and Bellamy 2020; Levy et al. 2021), These indi-
fiths (2022) that explores the designs of optimal nudges for cators may help decision makers gauge the credibility of
cognitively bounded agents, we conceptualize the AI assis- AI recommendation and calibrate their trust in AI. Since
tance as a “nudge” to the human decision making process, various indicators of the AI recommendation serve simi-
which would modify how humans weigh different informa- lar purposes, aligning with prior research (Tejeda et al.
tion in making their decisions. Therefore, in our framework, 2022; Wang, Lu, and Yin 2022), we focus on model-
we first establish an independent decision model that reflects ing how immediate assistance influences human decision
how humans form their independent decisions without any makers when the model’s prediction confidence is used
AI assistance. We then model the nudge of AI assistance to as the indicator in this study.
humans as the alterations to their decision models. To eval- 2. Delayed recommendation (Park et al. 2019; Grgic-Hlaca,
uate the performance of the proposed framework, we collect Engel, and Gummadi 2019; Lu and Yin 2021; Buçinca,
data on real human subjects’ decisions in AI-assisted dia- Malaya, and Gajos 2021; Fogliato et al. 2022; Ma et al.
betes prediction tasks with the aids of three common types 2023): Humans need to first make an initial decision on
of AI assistance through a randomized experiment. By fit- the task before the AI model’s decision recommendation
ting various computational models to the behavior dataset is revealed to them; this type of AI assistance forces hu-
collected, we find that our proposed framework consistently mans to engage more thoughtfully with the AI recom-
outperforms other baseline models in accurately predicting mendation.
the human decision behavior under different forms of AI 3. Explanation only (Lucic, Haned, and de Rijke 2019;
assistance. Furthermore, the proposed framework demon- Alqaraawi et al. 2020; Rader, Cotter, and Cho 2018;
strates robust performance in accurately predicting human Schuff et al. 2022; van Berkel et al. 2021): Only the AI
behavior in AI-assisted decision making even with limited model’s decision explanation but not its decision recom-
training data. Lastly, through a detailed analysis of the nudg- mendation is provided to decision makers. The explana-
ing effects of AI assistance identified by our framework, we tion often points out important features of the task that
offer quantitative insights into how individuals with different contribute the most to the AI model’s unrevealed decision
cognitive styles are nudged by AI assistance differently. For recommendation, aiming to highlight information that AI
instance, we observed that AI explanations appear to show a believes as highly relevant for decision making.
larger effect in redirecting the attention of intuitive decision
For more details of the literature review, please see the sup-
makers than reflective decision makers.
plementary material. In this study, we focus on building
a computational framework to characterize how different
Related Work forms of AI assistance, such as the three types identified
Empirical Studies in AI-Assisted Decision Making. The above, impact humans in AI-assisted decision making.
increased usage of decision aids driven by AI models has Modeling Human Behavior in AI-assisted Decision Mak-
inspired a line of experimental studies that identify different ing. Most recently, there has been a surge of interest among
forms of AI assistance to enhance human-AI collaboration researchers in computationally modeling human behavior
in decision making (Lai et al. 2023). By surveying the lit- in AI-assisted decision making (Bansal et al. 2021; Ku-
erature related to AI-assisted decision making in the ACM mar et al. 2021; Tejeda et al. 2022; Pynadath, Wang, and
Conference on Human Factors in Computing Systems, ACM Kamireddy 2019; Li, Lu, and Yin 2023; Lu et al. 2023).
Conference on Computer-supported Cooperative Work and Many of these studies build their models on economics
Social Computing, ACM Conference on Fairness, Account- frameworks (Wang, Lu, and Yin 2022), which explain hu-
ability, and Transparency, and ACM Conference on Intelli- man decision making behavior under uncertainty, or on psy-
gent User Interfaces from 2018 to 2021, we identify three chological frameworks that describe the relationship be-
common types of AI assistance: tween human trust and reliance on automated systems (Aje-
1. Immediate assistance: The AI model’s decision recom- naghughrure et al. 2019; Li, Lu, and Yin 2023). However,
mendation on the decision making task and other in- most existing works are either tailored to one specific form
dicators of the recommendation are provided to deci- of AI assistance or lack interpretations of how AI assistance
sion makers upfront. Typical indicators of the AI rec- influences human decision making processes. Inspired by
ommendation include the AI model’s accuracy (Lai, Liu, the recent research in computationally modeling the effects
and Tan 2020), explanations of the AI recommenda- of nudges (Callaway, Hardy, and Griffiths 2022), we take a
tion (Poursabzi-Sangdeh et al. 2018; Cheng et al. 2019; different approach in this paper and build a framework to
Smith-Renner et al. 2020; Liu, Lai, and Tan 2021; Tsai characterize diverse forms of AI assistance as nudges in the
et al. 2021; Bansal et al. 2020; Zhang, Liao, and Bel- human decision making process.
lamy 2020), and confidence levels of the recommenda-
1
Methods
In this study, we narrow down the scope of our framework to
model the influence of AI assistance on human decision makers on
Problem Formulation
each individual decision making task. That is, we do not consider We now formally describe the AI-assisted decision making
the sequential or temporal influence of AI assistance on humans in scenario in this study. Suppose a decision task can be char-
a sequence of decision making tasks. acterized by an n-dimensional feature x ∈ Rn , and y is
10084
The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)
the correct decision to make in this task. In this study, we given a training dataset of the DM’s independent decisions
focus on decision making tasks with binary choices of de- D = {xi , yih }N
i=1 , we adopt a Bayesian approach and set out
cisions, i.e., y ∈ {0, 1}, and each feature xi of the task x to learn from the training dataset the posterior distribution
is normalized to fall within the interval of [0, 1]. We use of model parameters for a population of diverse DMs, i.e.,
M(x; wm ) to denote the AI model’s output on the deci- P(wh |D), instead of learning a point estimate. As directly
sion task (wm are model parameters), and it is within the computing this posterior P(wh |D) is intractable, we lever-
range of [0, 1]. Given M(x; wm ), the AI model can pro- age variational inference to approximate it using the param-
eterized distribution qϕ (wh ) = N (wh ; µϕ , Σϕ ) and mini-
vide a binary decision recommendation to the human deci-
sion maker (DM), i.e., y m = 1(M(x; wm ) > 0.5). The AI mize the KL divergence between qϕ (wh ) and P(wh |D):
model’s confidence in this recommended decision is cm = Z
qϕ (wh )
max{M(x; wm ), 1 − M(x; wm )}. Following explainable KL(qϕ (wh )∥P(wh |D)) = qϕ (wh ) log dwh
wh P(wh |D)
AI methods like LIME (Ribeiro, Singh, and Guestrin 2016) Z
and SHAP (Lundberg and Lee 2017), the AI model could qϕ (wh )
= qϕ (wh )(log − log P(D|wh ) + log P(D))dwh
also provide some “explanations” of its decision recommen- wh P(wh )
dation, e = E(M(x; wm )), e ∈ Rn , by highlighting the = KL(qϕ (wh )∥P(wh )) − Eqϕ (wh ) [log P(D|wh ) − log P(D)]
“important” features that contribute the most to the decision (1)
recommendation. Here, ei ∈ {0, 1}, where ei = 1 means the
feature xi is highlighted as important, while ei = 0 means where P(wh ) is the prior distribution of wh and P(D) is a
the feature xi is not highlighted. In addition, we assume that constant2 . Given the learned qϕ (wh ), and without additional
the human DM also independently forms their own judg- knowledge of a human DM’s unique independent decision
ment of the decision task, which is characterized by the func- making model, we can only postulate that the DM follows
tion H(x; wh ) whose output is in the range of [0, 1]. Thus, an average model to make their decision:
y h = 1(H(x; wh ) > 0.5) represents the human DM’s inde-
pendent binary decision. y h,t = 1(Eqϕ (wh ) [H(xt ; wh )] > 0.5) (2)
We consider the setting where the human DM is asked to
Moreover, after we possess additional observations of the
complete a set of T decision tasks with the help of the AI
human DM’s decision making behavior (e.g., the initial de-
model. For each task t (1 ≤ t ≤ T ), the human DM is given
cision y h,t that they make), we can update our belief of the
the feature vector xt and the AI assistance. As discussed
DM’s independent decision making model from the general
previously, we focus on studying the following three forms
parameter distribution qϕ (wh ) in order to align with the ob-
of AI assistance:
served human behavior:
• Immediate assistance: The AI model’s binary decision
recommendation y m,t and its confidence cm,t are imme- q̂ϕ (wh ) ∝ qϕ (wh ) · 1(1(H(xt ; wh ) ≥ 0.5) = y h,t ) (3)
diately provided to the DM along with the task xt .
• Delayed recommendation: The DM is required to first Without loss of generality, in this study, we assumed that hu-
make an initial independent decision y h,t on the task. Af- mans’ decision making model H(xt ; wh ) follows the form
ter that, the AI model’s binary decision recommendation of logistic model:
y m,t will be revealed to the DM. H(xt ; wh ) = sigmoid(wh · xt ) (4)
• Explanation only: The DM is only provided with the AI
model’s explanation et , which highlights the important Quantify the Nudging Effects of AI Assistance
features of the task that contributes the most to the AI
model’s unrevealed decision recommendation. Inspired by a recent computational framework for under-
standing and predicting the effects of nudges (Callaway,
The DM’s independent judgement on the task is y h,t —this
Hardy, and Griffiths 2022), in this study, we introduce a
is observed as the DM’s initial decision when AI assistance
computational framework to provide an interpretable and
comes in the form of delayed recommendation, but is unob-
quantitative characterization of the influence of diverse
served (thus requires inference) when AI assistance comes
forms of AI assistance on human decision makers, which
in the other two forms. Given both their own judgement and
enables us to predict human behavior in AI-assisted decision
the AI assistance, the DM then makes a final decision ŷ t
making. The core idea of this framework is to conceptualize
on the task. The goal of our study is to quantitatively char-
the AI assistance as a “nudge” to the human decision making
acterize how the DM is “nudged” by different forms of AI
process, such that it can modify how the human DM weighs
assistance in making their final decision on each task.
different information in their decision making and alter their
Model Decision Makers’ Independent Judgement independent decision model accordingly. Depending on the
type of AI assistance used, this alternation could be oper-
To characterize how AI assistance nudges human DMs in
AI-assisted decision making, it is necessary to first under- ationalized as the human DM changing their belief in the
stand how human DMs form their independent judgement relevance of certain task feature to their decisions, or as the
without being nudged by AI. That is, we need to quan- human DM redirecting their attention to certain task feature
tify human DMs’ independent decision model H(x; wh ). when making their decisions.
Since each DM may have their own unique independent de-
2
cision making model with different model parameters wh , In this study, P(wh ) is set to be N (wh ; 0, In ).
10085
The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)
10086
The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)
10087
The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)
Table 1: Comparing the performance of proposed method with baseline methods on three forms of AI assistance, in terms of
NLL, Accuracy, and F1-score. “↓” denotes the lower the better, “↑” denotes the higher the better. Best result in each column is
highlighted in bold. All results are averaged over 5 runs. “-” means the method can not be applied in this scenario.
10088
The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)
(a) Immediate assistance: (b) Delayed recommendation (c) Delayed recommendation (d) Explanation only: δexp
||δdirect || (y h = y m ): ||δaffirm || (y h ̸= y m ): ||δcontra ||
Figure 5: Comparing the nudging effects of AI assistance on decision makers with different cognitive styles across three forms
of AI assistance.
rize the cognitive style of each human DM (i.e., each sub- reflective DMs are significantly more likely to align their fi-
ject), we utilized their scores in the 3-item Cognitive Re- nal decisions with the AI recommendation when AI affirms
flection Test (CRT) in the experiment—Following previous rather than contradicting their initial judgement (p < 0.05).
research (Frederick 2005), subjects scoring 3 were classi- In contrast, the intuitive and moderately reflective DMs do
fied as having a reflective thinking style, those with a score not appear to be impacted by AI significantly differently un-
of 0 were categorized as having an intuitive thinking style, der these two conditions.
and those scoring 1 or 2 were categorized as the moderate Finally, under the Explanation only scenario, we also ob-
reflective style. serve that intuitive DMs tend to place more emphasis on
Figure 5 shows the comparison results of the nudging ef- the features highlighted by the AI explanations. The nudg-
fects of AI assistance on DMs with different cognitive styles. ing effect of AI explanations on intuitive DMs is found to
To examine whether the nudging effects of AI assistance are be significantly greater than that on moderately reflective
different across DMs with different cognitive styles, we first DMs (p < 0.05). While the nudging effect also appears to
used a one-way ANOVA test 3 to determine if there are sta- be slightly larger for intuitive DMs compared to reflective
tistically significant differences in the values of sign(δ)||δ|| DMs, the difference is not statistically significant.
across different groups of DMs. If significant differences are
detected, we proceed with post-hoc pairwise comparisons Conclusion
using Tukey’s HSD test 4 . Overall, our findings suggest that In this paper, we propose a computational framework to
under the Immediate assistance and the Delayed recommen- characterize how various AI assistance nudges humans in
dation scenarios (when AI affirms human decision), AI as- AI-assisted decision making. We evaluate the proposed
sistance exerts a larger influence on DMs with a reflective model’s performance in fitting the real human behavior data
thinking style than intuitive DMs (p < 0.05) and DMs with collected from a randomized experiment. Our results show
moderate reflective styles (p < 0.05). One potential expla- that the proposed model consistently outperforms other
nation is that reflective DMs are inclined to deliberate more baselines in accurately predicting humans decisions under
extensively on tasks and the AI model’s recommendations. diverse forms of AI assistance. Additionally, further analy-
Thus, through their interactions with the AI model, reflec- ses based on the proposed framework provided insights into
tive DMs may have sensed the high performance of the AI how individuals with varying cognitive styles are impacted
model (its accuracy is 87% for the task), making them more by AI assistance differently.
willing to align their decisions with the AI recommendation, There are a few limitations of this study. For example, the
especially when the AI recommendation affirms their own behavior data is collected from laypeople on the diabetes
independent judgement. However, when the human DM’s prediction task, which contains a relatively small number
initial decision differs from the AI recommendation in the of features. It remains to be investigated whether the pro-
Delayed recommendation scenario, there isn’t a statistical posed model can perform well on tasks that involve many
difference in the AI’s nudging effects across the three types more features and thus more complex. Additionally, the AI-
of decision makers. In fact, by comparing the AI’s nudging assisted decision scenario we examined in this study lacks
effects on different groups of DMs under the two conditions sequential or temporal feedback regarding AI performance.
of the Delayed recommendation scenario—AI affirms hu- Further exploration is required to generalize the propose
man decisions or contradicts human decisions—we find that framework to the sequential settings. Lastly, we assumed
3
Analysis of Variance (ANOVA) is a statistical test for identi- that the independent human decision model follows the form
fying significant differences between group means. of logistic regression. Additional research is needed to ex-
4
Tukey’s HSD (Honestly Significant Difference) is a post-hoc plore how to adapt the current ways of altering humans’ de-
test used to determine specific differences between pairs of group cision models for reflecting the nudging effect of AI assis-
means after a one-way ANOVA test has found significance. tance on human DMs to other forms of decision models.
10089
The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)
10090
The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)
Park, J. S.; Berlin, R. B.; Kirlik, A.; and Karahalios, K. 2019. Wang, X.; Lu, Z.; and Yin, M. 2022. Will you accept the ai
A Slow Algorithm Improves Users’ Assessments of the Al- recommendation? predicting human behavior in ai-assisted
gorithm’s Accuracy. Proceedings of the ACM on Human- decision making. In Proceedings of the ACM Web Confer-
Computer Interaction, 3: 1 – 15. ence 2022, 1697–1708.
Passi, S.; and Vorvoreanu, M. 2022. Overreliance on AI Lit- Zhang, Y.; Liao, Q. V.; and Bellamy, R. K. E. 2020. Effect of
erature Review. Microsoft Research. confidence and explanation on accuracy and trust calibration
Poursabzi-Sangdeh, F.; Goldstein, D. G.; Hofman, J. M.; in AI-assisted decision making. Proceedings of the 2020
Vaughan, J. W.; and Wallach, H. M. 2018. Manipulating and Conference on Fairness, Accountability, and Transparency.
Measuring Model Interpretability. Proceedings of the 2021
CHI Conference on Human Factors in Computing Systems.
Pynadath, D. V.; Wang, N.; and Kamireddy, S. 2019. A
Markovian Method for Predicting Trust Behavior in Human-
Agent Interaction. Proceedings of the 7th International Con-
ference on Human-Agent Interaction.
Rader, E. J.; Cotter, K.; and Cho, J. 2018. Explanations as
Mechanisms for Supporting Algorithmic Transparency. Pro-
ceedings of the 2018 CHI Conference on Human Factors in
Computing Systems.
Ribeiro, M. T.; Singh, S.; and Guestrin, C. 2016. ” Why
should i trust you?” Explaining the predictions of any clas-
sifier. In Proceedings of the 22nd ACM SIGKDD interna-
tional conference on knowledge discovery and data mining,
1135–1144.
Schuff, H.; Jacovi, A.; Adel, H.; Goldberg, Y.; and Vu, N. T.
2022. Human Interpretation of Saliency-based Explanation
Over Text. Proceedings of the 2022 ACM Conference on
Fairness, Accountability, and Transparency.
Smith-Renner, A.; Fan, R.; Birchfield, M. K.; Wu, T. S.;
Boyd-Graber, J. L.; Weld, D. S.; and Findlater, L. 2020. No
Explainability without Accountability: An Empirical Study
of Explanations and Feedback in Interactive ML. Proceed-
ings of the 2020 CHI Conference on Human Factors in Com-
puting Systems.
Subrahmanian, V.; and Kumar, S. 2017. Predicting human
behavior: The next frontiers. Science, 355(6324): 489–489.
Tejeda, H.; Kumar, A.; Smyth, P.; and Steyvers, M. 2022.
AI-Assisted Decision-making: a Cognitive Modeling Ap-
proach to Infer Latent Reliance Strategies. Computational
Brain & Behavior, 5: 491 – 508.
Tsai, C.-H.; You, Y.; Gui, X.; Kou, Y.; and Carroll, J. M.
2021. Exploring and Promoting Diagnostic Transparency
and Explainability in Online Symptom Checkers. Proceed-
ings of the 2021 CHI Conference on Human Factors in Com-
puting Systems.
van Berkel, N.; Gonçalves, J.; Russo, D.; Hosio, S. J.; and
Skov, M. B. 2021. Effect of Information Presentation on
Fairness Perceptions of Machine Learning Predictors. Pro-
ceedings of the 2021 CHI Conference on Human Factors in
Computing Systems.
Wang, X.; Liang, C.; and Yin, M. 2023. The effects of AI
biases and explanations on human decision fairness: a case
study of bidding in rental housing markets. In Proceedings
of the Thirty-Second International Joint Conference on Ar-
tificial Intelligence, IJCAI-23, Edith Elkind (Ed.). Interna-
tional Joint Conferences on Artificial Intelligence Organiza-
tion, 3076–3084.
10091