Clinical Decision Support System
Clinical Decision Support System
White Paper ■
IDA SIM, MD, PHD, PAUL GORMAN, MD, ROBERT A. GREENES, MD, PHD,
R. BRIAN HAYNES, MD, PHD, BONNIE KAPLAN, PHD,
HAROLD LEHMANN, MD, PHD, PAUL C. TANG, MD
A b s t r a c t Background: The use of clinical decision support systems to facilitate the practice
of evidence-based medicine promises to substantially improve health care quality.
Objective: To describe, on the basis of the proceedings of the Evidence and Decision Support
track at the 2000 AMIA Spring Symposium, the research and policy challenges for capturing
research and practice-based evidence in machine-interpretable repositories, and to present
recommendations for accelerating the development and adoption of clinical decision support
systems for evidence-based medicine.
Results: The recommendations fall into five broad areas—capture literature-based and practice-
based evidence in machine-interpretable knowledge bases; develop maintainable technical and
methodological foundations for computer-based decision support; evaluate the clinical effects and
costs of clinical decision support systems and the ways clinical decision support systems affect and
are affected by professional and organizational practices; identify and disseminate best practices
for work flow–sensitive implementations of clinical decision support systems; and establish public
policies that provide incentives for implementing clinical decision support systems to improve
health care quality.
Conclusions: Although the promise of clinical decision support system–facilitated evidence-based
medicine is strong, substantial work remains to be done to realize the potential benefits.
■ J Am Med Inform Assoc. 2001;8:527–534.
Affiliations of the authors: University of California–San Francisco, Clinical decision support systems (CDSSs) have been
California (IS); Oregon Health and Science University, Portland, hailed for their potential to reduce medical errors1
Oregon (PG); Harvard Medical School (RAG); McMaster
University, Hamilton, Ontario, Canada (RBH); Yale University
and increase health care quality and efficiency.2 At
School of Medicine, Hamden, Connecticut (BK); Johns Hopkins the same time, evidence-based medicine has been
University School of Medicine, Baltimore, Maryland (HL); Palo widely promoted as a means of improving clinical
Alto Medical Foundation, Palo Alto, California (PCT). outcomes, where evidence-based medicine refers to
This work was supported in part by a United States Presidential the practice of medicine based on the best available
Early Career Award for Scientists and Engineers awarded to Dr. scientific evidence. The use of CDSSs to facilitate evi-
Sim and administered through grant LM-06780 of the National
Library of Medicine. dence-based medicine therefore promises to substan-
Correspondence and reprints: Ida Sim, MD, PhD, Department of
tially improve health care quality.
Medicine and Program in Biological and Medical Informatics, The Evidence and Decision Support track of the 2000
University of California–San Francisco, 400 Parnassus Avenue, AMIA Spring Symposium examined the challenges
Room A-405, San Francisco, CA 94143-0320; e-mail: <sim@medi-
cine.ucsf.edu>. in realizing the promise of CDSS-facilitated evidence-
Received for publication: 2/19/01; accepted for publication:
based medicine. This paper describes the activities of
7/11/01. this track and summarizes discussions in specific
528 SIM ET AL., Decision Support Systems for Evidence-based Medicine
In contrast, if the research literature were available as The Internet and other sources of research evidence
shared, machine-interpretable knowledge bases, then have provided patients with many more options for
CDSSs would have direct access to the newest obtaining health information but have also increased
530 SIM ET AL., Decision Support Systems for Evidence-based Medicine
the potential for patients to misinterpret or become ■ Continue to develop better methods for synthesiz-
misinformed about research results.24,25 As a result, ing results from a wide variety of study designs,
patients are now less dependent on clinicians for from randomized trials to observational studies .
information, but still trust clinicians the most for help
■ Develop shareable, machine-interpretable reposi-
with selecting, appraising, and applying a profusion
tories of up-to-date evidence of multiple types
of information to health decisions.26 Clinical decision
(e.g., from clinical trials, systematic reviews, deci-
support systems can support this growing involve-
sion models).
ment of patients in clinical decision making through
interactive tools that allow patients to explore rele- ■ Develop shareable, machine-readable repositories
vant information that can foster shared decision mak- of executable guidelines that are linked to up-to-
ing.27,28 Systems that provide both patients and clini- date evidence repositories.
cians with valid, applicable, and useful information
may result in care decisions that are more concordant ■ Define and build standard interfaces among these
with current recommendations, are better tailored to repositories, to allow evidence to be linked auto-
individual patients, and ultimately are associated matically among systems for systematic review-
with improved clinical outcomes. The actual effects ing, decision modeling, and guideline creation
of these CDSSs on care decisions and outcomes and maintenance.
should be evaluated. ■ Develop an informatics infrastructure for prac-
tice-based research networks to collect practice-
Recommendations based evidence.
The gap between the current state of CDSSs and the Establishment of a Technical and
full promise of CDSSs for evidence-based medicine Methodological Foundation
suggests a research and development agenda. On the
basis of the expert panels and discussion sessions at Figure 1 depicts the informatics architecture that we
the Congress, we recommend the following steps for suggest is needed for CDSSs to facilitate evidence-
researchers, developers, and implementers to take in based practice. In this architecture, CDSSs are situat-
the five areas of activity essential to increasing adop- ed in a distributed environment that comprises mul-
tion of evidence-adaptive CDSSs. tiple knowledge repositories as well as the electronic
medical record. Vocabulary and interface standards
Capture of Literature-based and will be crucial for interoperation among these sys-
Practice-based Evidence tems. To provide patient-specific decision support at
the point of care, CDSSs need to interface with the
If clinical research is to improve clinical care, it must electronic medical record to retrieve patient-specific
be relevant, of high quality, and accessible. The data and, increasingly, also to effect recommended
research should provide evidence of efficacy, effec- actions through computerized order entry. Evidence-
tiveness, and cost-effectiveness for typical inpatient adaptive CDSSs also need to interface with up-to-
and outpatient practice settings.29 If CDSSs are to date repositories of clinical research knowledge. No
help translate this research into practice, CDSSs must longer should CDSSs be thought of as stand-alone
have direct machine-interpretable access to the expert systems.
research literature, so that automated methods can be
brought to bear on the myriad tasks involved in In addition to establishing standardized communica-
“keeping up with the literature.” Thus, the establish- tion among CDSSs, electronic medical records, and
ment of shared, machine-interpretable knowledge knowledge repositories, we also need better models
bases of research and practice-based evidence is a of individualized patient decision making in real-
critical priority. On the basis of discussions at the world settings. Formal models of decision making
conference, we identify six specific recommendations such as decision analysis are not commonly used;
for action: much methodological work remains to be done on
mapping real-world decision-making challenges to
Recommendations for Clinical and Informatics tractable computational approaches.
Researchers
We identify several additional priorities for evi-
■ Conduct better quality clinical research on the effi- dence-adaptive CDSSs in particular. These priorities
cacy, effectiveness, and efficiency of clinical inter- include the development of methods for adjusting for
ventions, particularly in primary care settings. the quality of the evidence base, and efficient, sus-
Journal of the American Medical Informatics Association Volume 8 Number 6 Nov / Dec 2001 531
able findings on the clinical and organizational evaluated to establish the presence or absence of
aspects of CDSS use. A wide variety of evaluation clinical benefits. Any randomized clinical trials
methods are available,35–37 and both quantitative and that are conducted should have sufficient sample
qualitative methods should be used to provide com- sizes to detect clinically meaningful outcomes,
plementary insight into the use and effects of CDSSs. should randomize physicians or clinical units
All types of evaluation studies, not just randomized rather than patients, and should be analyzed using
trials, deserve increased attention and funding.38,39 methods appropriate for cluster randomization
studies.
In light of the current focus on errors in medicine, a
special class of evaluation study deserves particular ■ Establish partnerships between academic groups
mention. These studies are ongoing, iterative reeval- and community practices to conduct evaluations.
uations and redesigns of CDSSs that identify and
amplify system benefits while identifying and miti- Promotion of the Implementation of CDSSs
gating unanticipated system errors or dangers. The
rationale for these types of studies is that automation Relatively few examples of CDSSs can be found in
in other industries has not always been beneficial, practice. In part, this limited adoption may be
and indeed, automation can interfere with and because CDSSs are as much an organizational as a
degrade overall organizational performance.40 technical intervention, and organizational, profes-
Woods and Patterson 41 offer a cautionary note from sional, and other challenges to implementing CDSSs
the transportation industry: may be as daunting as the technical challenges.
Despite the fact that these systems are often justified Recommendations for CDSS Implementers
on the grounds that they would help offload work
from harried practitioners, we find that they in fact ■ Establish a CDSS implementation team composed
create new additional tasks, force the user to adopt of clinicians, information technologists, managers,
new cognitive strategies, require more knowledge and evaluators to work together to customize and
and more communication, at the very times when the
implement the CDSS.
practitioners are most in need of true assistance .
■ Develop a process for securing clinician agree-
Clinicians and health care managers must be contin-
ment regarding the science underlying the recom-
uously vigilant against unforeseen adverse effects of
mendations of a CDSS. For evidence-adaptive
CDSS use.
CDSSs, a process is also needed for maintaining
Recommendations for Evaluators clinician awareness of and agreement with any
changes in CDSS recommendations that may
■ Evaluate CDSSs using an iterative approach that result from new evidence.
identifies both benefits and unanticipated prob-
lems related to CDSS implementation and use: all ■ Plan explicitly for work flow re-engineering and
CDSSs can benefit from multiple stages and types other people, organizational, and social issues and
of testing, at all points of the CDSS life cycle. incorporate change management techniques into
system development and implementation. For
■ Conduct more CDSS evaluations in actual practice example, a CDSS that recommends immediate
settings, including ambulatory settings. angioplasty instead of thrombolysis as a new treat-
■ Use both quantitative and qualitative evaluation ment option for acute coronary syndromes will
methodologies to assess multiple dimensions of necessitate a major restructuring of the hospital’s
CDSS use and design (e.g., the correctness, relia- resource use and work practices.
bility, and validity of the CDSS knowledge base;
the congruence of system-driven processes with Establishment of Public Policies That Provide
clinical roles and work routines in actual practice; Incentives for Implementing CDSSs
and the return-on-investment of system imple- Significant financial and organizational resources are
mentation). Qualitative studies should incorporate often needed to implement CDSSs, especially if the
the expertise of ethnographers, sociologists, orga- CDSS requires integration with the electronic med-
nizational behaviorists, or other field researchers ical record or other practice systems. In a competitive
from within and without the medical informatics health care marketplace, financial and reimburse-
community, as applicable. ment policies can therefore be important drivers both
n If preliminary testing suggests that a CDSS could for and against the adoption of effective CDSSs. As
improve health outcomes, the CDSS should be more evaluation studies become available, policy
Journal of the American Medical Informatics Association Volume 8 Number 6 Nov / Dec 2001 533
makers will be better able to tailor these policies to ACP J Club. 1993;Nov–Dec:A16.
promote only those CDSSs that are likely to improve 8. Schor S, Karten I. Statistical evaluation of medical journal
manuscripts. JAMA. 1966;195(13):1123–8.
health care quality. 9. Fletcher RH, Fletcher SW. Clinical research in general medical
journals: a 30-year perspective. N Engl J Med. 1979;301(4):180–3.
Recommendations for Policy Makers 10. Moher D, Jadad AR, Nichol G, Penman M, Tugwell P, Walsh
S. Assessing the quality of randomized controlled trials: an
■ Develop financial and reimbursement policies that
annotated bibliography of scales and checklists. Control Clin
provide incentives for health-care providers to Trials. 1995;16(1):62–73.
implement and use CDSSs of proven worth. 11. Sacks HS, Reitman D, Pagano D, Kupelnick B. Meta-analysis:
an update. Mt Sinai J Med. 1996;63(3–4):216–24.
■ Develop and implement financial and reimburse- 12. Moher D, Cook DJ, Eastwood S, Olkin I, Rennie D, Stroup DF.
ment policies that reward the attainment of meas- Improving the quality of reports of meta-analyses of ran-
urable quality goals, as might be achieved by domised controlled trials: the QUOROM statement—Quality of
CDSSs. Reporting of Meta-analyses. Lancet. 1999;354(9193):1896–900.
13. Shaneyfelt TM, Mayo-Smith MF, Rothwangl J. Are guidelines
■ Promote coordination and leadership across the following guidelines? The methodological quality of clinical
health care and clinical research sectors to leverage practice guidelines in the peer-reviewed medical literature.
JAMA. 1999;281(20):1900–5.
informatics promotion and development efforts 14. Williamson JW, German PS, Weiss R, Skinner EA, Bowes FD.
by government, industry, AMIA, and others. Health science information management and continuing edu-
cation of physicians: a survey of U.S. primary care practitioners
Conclusions and their opinion leaders. Ann Intern Med. 1989;110(2):151–60.
15. McAlister FA, Graham I, Karr GW, Laupacis A. Evidence-
based medicine and the practicing clinician. J Gen Intern Med.
The coupling of CDSS technology with evidence- 1999;14(4):236–42.
based medicine brings together two potentially pow- 16. Greer AL. The state of the art vs. the state of the science: the
erful methods for improving health care quality. To diffusion of new medical technologies into practice. Int J
realize the potential of this synergy, literature-based Technol Assess Health Care. 1988;4(1):5–26.
17. Sim I, Owens DK, Lavori PW, Rennels GD. Electronic trial
and practice-based evidence must be captured into banks: a complementary method for reporting randomized tri-
computable knowledge bases, technical and method- als. Med Decis Making. 2000;20(4):440–50.
ological foundations for evidence-adaptive CDSSs 18. Cina CS, Clase CM, Haynes BR. Refining the indications for
must be developed and maintained, and public poli- carotid endarterectomy in patients with symptomatic carotid
stenosis: a systematic review. J Vasc Surg. 1999;30(4):606–17.
cies must be established to finance the implementa-
19. Chassin MR. Appropriate use of carotid endarterectomy [edi-
tion of electronic medical records and CDSSs and to torial]. N Engl J Med. 1998;339(20):1468–71.
reward health care quality improvement. 20. Goldstein LB, Bonito AJ, Matchar DB, et al. U.S national sur-
vey of physician practices for the secondary and tertiary pre-
The authors thank the many discussion participants whose anony- vention of ischemic stroke: design, service availability, and
mous comments were included in this paper. They also thank common practices. Stroke. 1995;26(9):1607–15.
Patricia Flatley Brennan for her helpful comments on an earlier 21. Nutting PA. Practice-based research networks: building the
draft of this manuscript, and Amy Berlin for her assistance in infrastructure of primary care research. J Fam Pract. 1996;
preparing the manuscript. 42(2):199–203.
22. Nutting PA, Baier M, Werner JJ, Cutter G, Reed FM, Orzano
References ■ AJ. Practice patterns of family physicians in practice-based
research networks: a report from ASPN. Ambulatory Sentinel
1. Bates DW, Cohen M, Leape LL, et al. Reducing the frequency Pratice Network. J Am Board Fam Pract. 1999;12(4):278–84.
of errors in medicine using information technology. J Am Med 23. van Weel C, Smith H, Beasley JW. Family practice research
Inform Assoc. 2001;8:299–308. networks: experiences from 3 countries. J Fam Pract. 2000;
2. Teich JM, Wrinn MM. Clinical decision support systems come 49(10):938–43.
of age. MD Comput. 2000;17(1):43–6. 24. Kaplan B, Brennan PF. Consumer informatics supporting
3. Sackett DL, Rosenberg WM, Gray JA, Haynes RB, Richardson patients as co-producers of quality. J Am Med Inform Assoc.
WS. Evidence-based medicine: what it is and what it isn’t [edi- 2001;8:309–16.
torial]. BMJ. 1996;312(7023):71–2. 25. Jadad AR, Haynes RB, Hunt D, Browman GP. The Internet
4. Nordin-Johansson A, Asplund K. Randomized controlled tri- and evidence-based decision-making: a needed synergy for
als and consensus as a basis for interventions in internal med- efficient knowledge management in health care. CMAJ. 2000;
icine. J Intern Med. 2000;247(1):94–104. 162(3):362–5.
5. Suarez-Varela MM, Llopis-Gonzalez A, Bell J, Tallon-Guerola 26. Harris Interactive. Ethics and the Internet: Consumers vs.
M, Perez-Benajas A, Carrion-Carrion C. Evidence-based gen- Webmasters, Internet Healthcare Coalition, and National
eral practice. Eur J Epidemiol. 1999;15(9):815–9. Mental Health Association. Oct 5, 2000.
6. Ellis J, Mulligan I, Rowe J, Sackett DL. Inpatient general med- 27. Morgan MW, Deber RB, Llewellyn-Thomas HA, Gladstone P,
icine is evidence based. A-Team, Nuffield Department of Clin- O’Rourke K, et al. Randomized, controlled trial of an interac-
ical Medicine. Lancet. 1995;346(8972):407–10. tive videodisc decision aid for patients with ischemic heart
7. Haynes RB. Where’s the meat in clinical journals [editorial]? disease. J Gen Intern Med. 2000;15(10):685–93.
534 SIM ET AL., Decision Support Systems for Evidence-based Medicine
28. Health Dialog Videos. Available at: http://www.healthdia- 35. Friedman CP, Wyatt JC. Evaluation methods in medical infor-
log.com/video.htm. Accessed Feb 20, 2001. matics. New York: Springer-Verlag, 1997.
29. Haynes B. Can it work? Does it work? Is it worth it? The test- 36. Anderson JG, Aydin CE. Evaluating the impact of health care
ing of healthcare interventions is evolving [editorial]. BMJ. information systems. Int J Technol Assess Health Care. 1997;
1999;319(7211):652–3. 13(2):380–93.
30. United States Preventive Services Task Force. Guide to Clinical 37. Kaplan B. Evaluating informatics applications: social interac-
Preventive Services. 2nd ed. Appendix A: Task Force Ratings. tionism and call for methodological pluralism. Int J Med Inf.
1996. Available at: http://text.nlm.nih.gov/ftrs/directBrowse. 2001;84:39–56.
pl?dbName=cps&href=APPA. Accessed Jun 18, 2001. 38. Tierney WM, Overhage JM, McDonald CJ. A plea for con-
31. Kaplan B. Evaluating informatics applications: review of the trolled trials in medical informatics [editorial]. J Am Med
clinical decision support systems evaluation literature. Int J Inform Assoc. 1994;1(4):353–5.
Med Inf. 2001;64:15–37. 39. Stead WW, Haynes RB, Fuller S, Friedman CP, Travis LE, Beck
32. Hunt DL, Haynes RB, Hanna SE, Smith K. Effects of comput- JR, et al. Designing medical informatics research and library:
er-based clinical decision support systems on physician per- resource projects to increase what is learned. J Am Med
formance and patient outcomes: a systematic review. JAMA. Inform Assoc. 1994;1(1):28–33.
1998;280(15):1339–46. 40. Woods D. Testimony of David Woods, Past President, Human
33. Balas E, Austin S, Mitchell J, Ewigman B, Bopp K, Brown G. Factors and Ergonomics Society. In: National Summit on
The clinical value of computerized information services. Arch Medical Errors and Patient Safety Research. Washington, DC:
Fam Med. 1996;5:271–8. Quality Interagency Coordination Task Force, Sep 11, 2000.
34. Shea S, DuMouchel W, Bahamonde L. A meta-analysis of 16 41. Woods DD, Patterson ES. How unexpected events produce an
randomized controlled trials to evaluate computer-based clin- escalation of cognitive and coordinative demands. In:
ical reminder systems for preventive care in the ambulatory Hancock PA, Desmond PA (eds). Stress Workload and
setting. J Am Med Inform Assoc. 1996;3:399–409. Fatigue. Hillsdale, NJ: Erlbaum, 2001.
Appendix
PANELISTS AND GROUP LEADERS