Research Methods A Framework For Evidence-Based Clinical Practice
Research Methods A Framework For Evidence-Based Clinical Practice
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 382
LibraryPirate
97689_FM 12/07/10 5:41 PM Page i
RESEARCH METHODS
A Framework for Evidence-Based
Clinical Practice
All rights reserved. This book is protected by copyright. No part of this book may be reproduced or transmitted in
any form or by any means, including as photocopies or scanned-in or other electronic copies, or utilized by any
information storage and retrieval system without written permission from the copyright owner, except for brief
quotations embodied in critical articles and reviews. Materials appearing in this book prepared by individuals as
part of their official duties as U.S. government employees are not covered by the above-mentioned copyright. To
request permission, please contact Lippincott Williams & Wilkins at Two Commerce Square, 2001 Market Street,
Philadelphia, PA 19103, via email at [email protected], or via website at lww.com (products and services).
Printed in China
DISCLAIMER
Care has been taken to confirm the accuracy of the information present and to describe generally accepted
practices. However, the authors, editors, and publisher are not responsible for errors or omissions or for any
consequences from application of the information in this book and make no warranty, expressed or implied, with
respect to the currency, completeness, or accuracy of the contents of the publication. Application of this information
in a particular situation remains the professional responsibility of the practitioner; the clinical treatments described
and recommended may not be considered absolute and universal recommendations.
The authors, editors, and publisher have exerted every effort to ensure that drug selection and dosage set
forth in this text are in accordance with the current recommendations and practice at the time of publication.
However, in view of ongoing research, changes in government regulations, and the constant flow of information
relating to drug therapy and drug reactions, the reader is urged to check the package insert for each drug for any
change in indications and dosage and for added warnings and precautions. This is particularly important when the
recommended agent is a new or infrequently employed drug.
Some drugs and medical devices presented in this publication have Food and Drug Administration (FDA)
clearance for limited use in restricted research settings. It is the responsibility of the health care provider to
ascertain the FDA status of each drug or device planned for use in their clinical practice.
To purchase additional copies of this book, call our customer service department at (800) 638-3030 or fax orders to
(301) 223-2320. International customers should call (301) 223-2300.
Visit Lippincott Williams & Wilkins on the Internet: http://www.lww.com. Lippincott Williams &
Wilkins customer service representatives are available from 8:30 am to 6:00 pm, EST.
9 8 7 6 5 4 3 2 1
LibraryPirate
97689_FM 12/07/10 5:41 PM Page iii
FOREWORD
This research methods textbook distinguishes itself from other textbooks by provid-
ing a unique framework and perspective for users/students to establish the rele-
vancy of research in their clinical practice. Many, if not most, students in profes-
sional preparation allied health care programs view the research methods/statistics
course requirement of the curriculum as an obstacle to be overcome, or at best, as a
necessary evil. Most research methods textbooks promote these notions because of
the way they are presented. Of course, most times they are written by researchers or
statisticians and are absolutely correct in presenting the theoretical underpinnings
and mechanistic applications of the scientific method. They correctly present expla-
nations as to why one type of methodology requires a certain type of statistical
analysis based on the characteristics of the study population, the type of data col-
lected, or the underlying assumptions pertinent to a specific statistical model. So,
while technically beyond reproach, their failure is in establishing how and why re-
search activity and understanding is integral to a professional practice.
In this textbook, the authors appreciate professional realities that have rele-
vancy to professional preparation and the role of research within that preparation.
First, they promote that understanding research is a required competency in allied
health care preparation. That is, most allied health care professionals are going to
be consumers of research literature for the duration of their professional lives.
This is often enforced by continuing education requirements that reflect the un-
derstanding that medical and allied medical bodies of knowledge are ever ex-
panding and/or being refined and must be communicated to clinicians. Thus, at a
minimum, professional preparation programs must provide the opportunity for
the student to establish competencies in reading, critically evaluating, and synthe-
sizing research into their clinical practice.
The reality is that most medical and allied medical professions have accepted
that their disciplines need to follow an evidence-based best practice approach to
clinical practice and professional preparation in order to be effective today. Most
of their professional organizations actively promote through funding and dissem-
ination vehicles this research paradigm. This is not at the exclusion of other valu-
able research paradigms, but rather as an emphasis commensurate with current
disciplinary interest.
The authors have taken these realities and developed the research elements of
this textbook to be professionally relevant to clinical movement practitioners (e.g.,
physical therapists and certified athletic trainers). They use a commonsense-style
presentation of the conceptual and theoretical bases for the clinical research enter-
prise and provide clear examples of application, evaluation, and integration into
clinical practice.
iv
LibraryPirate
97689_FM 12/07/10 5:41 PM Page v
Foreword v
Part I introduces the reader to the basic tenets of the research enterprise as it
relates to clinical practice. A wonderful chapter is presented on how to read and
evaluate research articles. This is a very pragmatic chapter that highlights the ele-
ments of an article so the reader can critically appraise the quality of the article.
The history of evidence-based clinical practice is presented along with a chapter
about how research can be used to establish best practices. The last chapter in this
section provides an overlay of ethical principles that need to be operative in re-
search and clinical practice enterprises.
Part II has seven chapters that effectively make the case that the conceptual,
elemental parts of research have a role in the readers effort to establish an
evidence-based practice. This is accomplished by describing how a reader could
have a clinical question and that the question is what dictates the research
methodology and analysis. One chapter provides the reader with a way to find re-
sources on clinical practice topics and evaluate the quality of the resources.
Another establishes that there is a hierarchy of evidence and the reader needs
to know what level or quality of evidence is important for them to use in their
clinical practice. Several chapters are nicely presented that introduce the most
common types of scientific inquiry and statistical analysis related to evidence-
based practice.
Part III carries this utilitarian presentation approach to the research enterprise
even further. The chapters are organized into clinical categories rather than re-
search method categories. That is, Chapter 13 presents what clinical research
methodology and analysis is appropriate when the reader is trying to find out
about the evaluation and diagnosis of orthopaedic conditions. Chapter 14 is about
screening and prevention of illnesses and injuries and what kinds of clinical re-
search and analysis are appropriate. Chapter 15 presents the research options that
clinical movement practitioners can use to develop evidence about the efficacy of
treatment/intervention options they may want to use. Chapter 16 further devel-
ops the concepts of treatment outcomes and broadens the focus to present com-
mon methods and relevant data analysis techniques. The last two chapters in this
part focus on how to extend research results to a clinical practice and how to or-
ganize many research results into a usable body of clinical evidence.
Parts IV and V provide sections that are unique to this text. But, again, the util-
itarian approach is taken. These sections are how-tos for a clinical movement
practitioner to appreciate how clinical research results are disseminated. As a re-
search consumer, this is important for them to know. Also, the authors provide
some guidance on how to effectively write a funding proposal for a research proj-
ect. Again, many students in professional preparation programs begin to favor the
research element of their discipline and want to become actively involved in devel-
oping and answering their own clinical questions. Finally, the authors provide in-
sight into how clinical evidence can be used in clinical learning and teaching.
The genius of this text is that it is written for clinical professionals to under-
stand and appreciate research elements that are going to positively affect their
LibraryPirate
97689_FM 12/07/10 5:41 PM Page vi
vi Foreword
clinical practice. All the research methodology and analysis contained herein is
within the context of improving clinical outcomes. In other words, the authors il-
lustrate how clinical practice should drive the research enterprise rather than the
converse. In this way, the information becomes relevant to something clinical pro-
fessionals value and are therefore willing to accept and incorporate research com-
petencies in their clinical practice.
PREFACE
viii Preface
APPROACH
The focal point of our approach to research methods is to provide guidance and
direction for students, instructors, and practitioners on how to acquire, read, in-
terpret, assess, and apply research as evidence in clinical practice has not been
provided in traditional research methods textbooks. We tend to view statistical
analysis as the flip side of research methodology, or two sides of the same coin.
We approach the subject matter conceptually and practically. Using a common-
sense style, conceptual and theoretical frameworks are introduced and discussed
LibraryPirate
97689_FM 12/07/10 5:41 PM Page ix
Preface ix
CHAPTER FEATURES
Each chapter of Research Methods provides the following elements to enhance the
usability of the text and offers a fresh approach to research methods for clinical
movement practitioners and students interested in learning about evidence-based
clinical practice:
Key terms and concepts are bolded throughout the chapter to help the
reader focus their attention on scientific nomenclature and vocabulary es-
sential to a basic understanding of the chapter content and context.
Chapter objectives detail what the reader will learn in the chapter and
highlight important pedagogical outcomes while also serving as a three-
fold self-assessment for readers (What are some important reasons for me
to read this chapter? How does this knowledge help me prepare to read it?
And, after reading this chapter, can I satisfy this list of learning outcomes?).
Readers are encouraged to make use of the chapter objectives to help guide
their reading and assess their level of reading comprehension.
Concept checks reinforce important chapter content and purposefully reit-
erate noteworthy theory and viewpoints.
Examples throughout chapters make concepts easier to grasp and apply to
real-life research and clinical decision-making situations.
A chapter summary at the end of each chapter provides a comprehensive
review of the chapter and provides a take-home message for the readers.
A list of key points further elucidates concepts, theories, and viewpoints
presented and elaborated throughout the chapter as both foundation to the
current chapter and groundwork to subsequent chapters, thus presenting a
full-circle approach by reinforcing the importance of key terms, chapter ob-
jectives, concept checks, examples, and the chapter summary within the
textbook.
Chapter references and suggested readings are provided to aid the reader
with supplemental materials for breadth and depth of knowledge, and
demonstrate appropriate use and formal citation of original sources in em-
pirical research.
Figures and tables offer illustrations to provide the reader with visual ex-
amples that help support important information detailed in the text.
LibraryPirate
97689_FM 12/07/10 5:41 PM Page x
x Preface
ORGANIZATION
The book is organized into five parts. Part I (Chapters 1 to 5) is structured as an
opening presentation of underlying conceptual frameworks and theoretical un-
derpinnings in clinical research and evidence-based practice, and an introduction
to the skills for critiquing and analyzing research is introduced. Chapter 1 covers
the concept of empirical research and the basic tenants of research methodology
as a collective paradigm. Chapter 2 provides a guide for how to read research and
offers a framework for evaluating research articles. Chapter 3 introduces the no-
tion of evidence-based clinical practice and explains how to distinguish best prac-
tices. Chapter 4 addresses the historical perspective of evidence-based medicine to
provide lead into the necessity of ethics in research and practice (covered in
Chapter 5).
Part II (Chapters 6 to 12) is organized around the research process in terms of sta-
tistical analyses and the idea of research as evidence. More specifically, this section
addresses how the question drives the methods when seeking answers to clinical
questions, because we are of the pedagogical and theoretical perspectives that it is
difficult to teach research methods without simultaneously addressing statistics.
Chapter 6 covers informational sources, search strategies, and critical appraisal of re-
search as evidence. Chapter 7 addresses the issue of hierarchy of evidence. Chapter 8
deals with qualitative inquiry, while Chapter 9 begins to sort out quantitative inquiry.
Chapters 10 and 11 describe research designs and data analysis while introducing the
statistical perspective of research methods. Chapter 10 discusses the fundamentals of
statistical analysis, focusing on validity and reliability of measures. Chapter 11 cov-
ers tests of comparison. Chapter 12 highlights measures of association.
The concepts that are introduced in Parts I and II and then developed and elab-
orated in Part III focuse on clinical research diagnosis, prevention, and treatment. In
this way, Part III (Chapters 13 to 18) is ordered to describe research designs and data
analysis for each type of study, and then provide some examples to illustrate its
LibraryPirate
97689_FM 12/07/10 5:41 PM Page xi
Preface xi
application. Our goal is to address and apply concepts of research methods into pre-
vention, diagnostics, and intervention outcomes. We are of the perspective that
these are the issues at the heart of evidence-based medicinewhy disease and in-
jury occur (epidemiology) and what can I, as a clinician, do to prevent the condition
in the patient in my office (clinical epidemiology), etc. Chapter 13 addresses evalu-
ation and diagnosis from the perspective of research methods and data analysis.
Chapter 14 looks at screening and prevention of illness and injuries, again from the
perspective of research methods and data analysis. Chapter 15 explores the notion
of treatment outcomes across the disablement spectrum. Chapter 16 expands on this
and looks at treatment outcomes from the perspective of research methods and data
analysis. Chapter 17 covers the topics of clinical prediction rules and practice guide-
lines. And, Chapter 18 explores the application and usefulness of systematic review
and meta-analysis in evidence-based clinical practice.
Parts IV and V are unique to our book on several levels. The topics covered in
these chapters are uncommon in most traditional textbooks and are even rarely
mentioned in a book on research methods. Part IV (Chapters 19 and 20) is concen-
trated on the dissemination of research. Specifically, Chapter 19 offers guidelines
and suggestions for presenting research findings; and, Chapter 20 offers a detailed
guide with examples for writing the funding proposal. Part V (Chapters 21 and 22)
covers the integration of evidence-based medicine into the education experience.
Chapter 21 explores the notion of evidence in learning and teaching. Chapter 22
wraps up the discussion and comes full circle to revisit the topic of evidence in the
context of the clinical experience.
ADDITIONAL RESOURCES
Research Methods includes additional resources for both instructors and students
that are available on the books companion website at http://thePoint.lww.com/
Hurley.
Instructor Resources
Approved adopting instructors will be given access to the following additional
resources:
xii Preface
Student Resources
All purchasers of the textbook can access the searchable Full Text On-line by going
to the Research Methods website at http://thePoint.lww.com/Hurley. See the in-
side front cover of this textbook for more details, including the passcode you will
need to gain access to the website.
ACKNOWLEDGMENTS
No book is the result of any one persons efforts, yet we accept any mistakes as
ours alone. We would like to thank the following people whose efforts were essen-
tial to this project:
Our product manager, Meredith Brittain, and her editorial staff for their tal-
ents, continued hard work, patience, and assistance in bringing this vision
to fruition.
Our reviewers, for their time, feedback, and suggestions for improvement.
Our contributing authors for their time, expertise, and generous participa-
tion. These include the coauthors of Chapter 20, Amy Henderson-Harr
(Assistant Vice President, Research and Sponsored Programs, State
University of New York College at Cortland) and Allan Shang (Assistant
Professor of Anesthesiology, Duke University School of Medicine, and
Senior Research Scientist, The Fitzpatrick Institute for Photonics, Duke
University Pratt School of Engineering).
We would also like to acknowledge the following individuals who con-
tributed suggestions for the text: Timothy J. Bryant (Lecturer III, Kinesiology
Department, State University of New York College at Cortland), for his
help with Chapters 5 and 8; Brent Thomas Wilson (Assistant Professor,
Communication Disorders and Sciences Department, State University of
New York College at Cortland), for his help with Chapters 5 and 8; and,
Amy Henderson-Harr, for her help with Chapter 5.
Our students and patients, past and present. We offer special acknowledg-
ment to the following students for their assistance in this project: Caitlin
Latham, Lauren Lenney, Patrick Sullivan.
Our contemporaries in education, research, and medicine.
The many past generations of teachers, clinical researchers, and medical
practitioners who came before us. We have benefited from your countless,
sometimes unrecognized, sacrifices, and your successes and failures paved
the way for evidence-based practice.
We would also like to thank our families and friends for their understanding
and support throughout the many hours during which our work keeps us away
from those most precious parts of our lives: the persons, the times, and the occa-
sions that we cannot recover after they are missed or gone.
LibraryPirate
97689_FM 12/07/10 5:41 PM Page xiii
CONTENTS
Foreword . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iv
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
PART I
INTRODUCTION TO CLINICAL RESEARCH 1
1 Research: An Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
Introduction 4
What Research is and What Research is Not 4
The Role of Theory in Research 7
The Research Process 7
Research and the Clinical Practitioner 14
Chapter Summary 15
Key Points 15
Critical Thinking Questions 16
Applying Concepts 16
REFERENCES 16
SUGGESTED READING 17
xiv Contents
Key Points 40
Critical Thinking Questions 41
Applying Concepts 41
REFERENCES 41
SUGGESTED READING 42
PART II
SEEKING ANSWERS: HOW THE QUESTION
DRIVES THE METHODS 75
6 Finding the Evidence: Informational Sources,
Search Strategies, and Critical Appraisal . . . . . . . . . . . . . . . . . . 77
Introduction 78
Informational Sources and the Reference Librarian 78
Electronic Sources and Databases 79
Search Strategies 80
LibraryPirate
97689_FM 12/07/10 5:41 PM Page xv
Contents xv
xvi Contents
Contents xvii
PART III
CLINICAL RESEARCH: DIAGNOSIS, PREVENTION, AND TREATMENT 213
13 Evaluation and Diagnosis: Research Methods
and Data Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 215
Introduction 216
Physical Examination Procedures 217
How Can Clinical Research Improve Patient Evaluation? 218
Design of Studies of Diagnostic Testing 219
Assessing Research of Diagnostic Instruments 222
Statistics and Interpretations 224
The BasicsSensitivity and Specificity 226
Likelihood Ratios 229
Chapter Summary 236
Key Points 236
Critical Thinking Questions 237
Applying Concepts 237
REFERENCES 237
SUGGESTED READING 238
xviii Contents
Contents xix
PART IV
DISSEMINATION OF RESEARCH 327
19 Presentation of Findings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 329
Introduction 330
Types of Presentations 330
Chapter Summary 344
Key Points 344
Critical Thinking Questions 344
Applying Concepts 345
REFERENCES 345
PART V
INTEGRATING EVIDENCE-BASED MEDICINE INTO
THE EDUCATIONAL EXPERIENCE 381
21 Evidence in Learning and Teaching . . . . . . . . . . . . . . . . . . . . . 383
Introduction 384
Learning Critical Appraisal 385
Learning and Teaching from a Body of Evidence 387
LibraryPirate
97689_FM 12/07/10 5:41 PM Page xx
xx Contents
Glossary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 405
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 415
LibraryPirate
97689_FM 12/07/10 5:41 PM Page xxi
Reviewers xxi
Reviewers
Diane P. Brown, PhD, OTR
Associate Clinical Professor
School of Occupational Therapy
Texas Womans University
Denton, TX
xxii Reviewers
PART I
INTRODUCTION TO
CLINICAL RESEARCH
LibraryPirate
97689_c01 01/07/10 3:39 PM Page 2
LibraryPirate
97689_c01 01/07/10 3:39 PM Page 3
CHAPTER 1
RESEARCH: AN OVERVIEW
CHAPTER OBJECTIVES
After reading this chapter, you will:
Know what research is and what research is not.
Understand the role of theory in research.
Know the appropriate steps to follow during the research process.
Learn how to identify a topic and be aware of the ways in which to search and review
the literature.
Be able to explain how to define a topic.
Understand the importance of an operational definition when defining a topic and how
to correctly phrase an operationally defined hypothesis.
Know how to plan methods and to gather and interpret data in an attempt to test the
hypothesis.
Understand the role that research methods play in evidence-based clinical practice.
KEY TERMS
anecdotal evidence-based practice theoretical research hypothesis
data independent variable theory
dependent variable paradigm validity
empirical research reliability variables
3
LibraryPirate
97689_c01 01/07/10 3:39 PM Page 4
INTRODUCTION
For many of us, the notion of research can seem daunting, ambiguous, or even
conceptually intangible. Some of the more imaginative among us may envision
the scientific research process as brainiac, dangerous, secretive, or even science
fiction-like. For others, the results of research may have an unrecognized yet as-
sumed impact on multiple levels of everyday living. We may presume the deci-
sions and decision processes inherent in medicine and healthcare are not only
necessary but also valid and reliable. But what about the false assumptions that
we make regarding research and/or how it is used to protect us?
Many of us might be as truly unaware of the extent of research in our lives as
we are uninformed about the pitfalls of quackery that expose us to worthless or
even harmful products that are legally marketed. Sometimes it might seem hard
to keep up with the latest reports of clinical studies and medical guidelines. At
times it might even seem as though one news report contradicts another, and it
could be a challenge to determine which information is trustworthy. How do we
differentiate credible research and determine what is the right and most appropri-
ate answer to the questions we ask? If we seek to apply research, how do we
weigh research with instinct-honed professional expertise and personal prefer-
ence? These are among the issues that are central to clinicians who seek to become
practitioners of evidence-based medicine. For example, how can clinicians use the
most directly relevant research (evidence) to help guide them when facing diag-
nosis and treatment decisions for individual patients (evidence-based practice
[EBP])? Similarly, how can clinicians learn to access, evaluate, and interpret re-
search literature as they seek to apply the principles of clinical epidemiology to
day-to-day clinical practice (Bulpitt, 1987; Godfrey, 1985; Haynes et al., 1986;
Sackett et al., 1991)?
Research can at times feel overwhelming, especially to those not formally
trained in research methodology. The goal of this chapter is to provide common-
sense style guidance and direction in the methods of research. The intent is to help
bridge the gap between clinical practice and research. This chapter will provide
a theoretical framework that will enable students and practitioners to become
informed consumers of research as they learn to evaluate, interpret, and apply re-
search into an evidence-based best practice model.
Hollywood depictions, research is more than the mission of a scientist gone mad
with an all-consuming passion for truth.
What Is Research?
Research is a careful, logical, and systematic process of investigation. Empirical
research is a methodological approach to problem solving in which decisions are
based on findings and conclusions established as a result of analyzing collected
data about variables and the relationships that connect them. Data are collected
through observation and/or experimentation, and later scrutinized through a
series of statistical analyses to determine results. The results of research often lead
to more questions and more areas of investigation.
Further investigation requires future research. Thus, like an expanding puz-
zle, knowledge is pieced together as research begets research. This complex, se-
quential method of inquiry is usually based on tentative explanation and
discussion of facts, findings, and theory. For itself, theory is central to the research
process. While the goal of research may be to originate theory, an existing theory
can be replaced only by a new theory that has been empirically tested and sup-
ported by data. This is the cyclic nature of research.
In the continuum of research, applied research offers direct clinical applicabil-
ity while basic research may have little direct clinical application. In between these
two types of research lies the debate of ecologic validity. It has been suggested
that levels of relevance for finding solutions to practical problems might incorpo-
rate some degree of basic and applied research, depending on the setting and pri-
mary objective of the research (Christina, 1989).
Basic research tends to take place in carefully monitored and controlled labo-
ratory settings. The goal of basic research is to addresses theoretical issues or un-
derlying explanations to questions in basic science (e.g., biology). Applied
research can take place in laboratory and nonlaboratory settings. In either circum-
stance, the settings are carefully designed to approximate authentic and function-
ing environments (e.g., clinical or sport environments). Applied research can
provide and contribute to theory-based knowledge, but also aims to provide di-
rect solutions to practical problems.
of research is limited by (and dependent on) the accuracy, validity, and reliability
of each step of the process.
Research is not a method of proof. In other words, research does not
prove anything. Rather, it lends supportive evidence for or against the existence
or nature of relationships among or between variables of interest. It is a process of
investigation that provides perspective on how (or if) one variable or group of
variables affects or influences another variable or group of variables.
Research is not haphazard. It is planned and procedural. Research is a time-
consuming and, at times, tedious process of examination and investigation. It is a
careful, unhurried, deliberate series of steps that are completed in an exact order.
The steps are designed to confirm precise measurement.
The outcome of research is not predetermined. Research is not conducted to
justify results; it is done to determine results. In the same respect, research reports
results that were found. It does not report results that were expected or hoped to
be found. Sometimes, finding nothing at all is as important and profound as find-
ing something as a result of the research process.
Research is not unrepeatable. By its very nature, research is subject to and
intended for replication. The results of research must be repeatable in order to be
reliable and valid.
CONCEPT CHECK
Research lends supportive evidence for or against the existence or nature of relation-
ships among or between variables of interest. It is a process of investigation that pro-
vides perspective on how (or if) one variable or group of variables affects or
influences another variable or group of variables. Research is not a method of proof.
CONCEPT CHECK
The role of theory in research is that of a frame of reference, a school of thought,
or a paradigm of science. Theory can be a goal or a guide to research.
now, and it provides us with a widening and deepening knowledge base that we
can use to find better answers, as we learn more. Research produces new data, and
adds to what is already known.
The focus of this text is on the stages of the scientific research process;
nonetheless, it is noteworthy at this point to mention chapter sequence in the for-
mal written proposal or thesis.
your topic in a more specific context. For example, a general topic of knee injury
would be an appropriate starting point. The topic is broad, but until youve ex-
plored and reviewed the literature, you may not appreciate the various ap-
proaches and avenues of research on the topic. Perhaps, after beginning to search
the literature on this topic, you discover that your interest lies more within the
context of anterior cruciate ligament (ACL) injury. However, this is still a broad
category, and it will require more searching and reviewing of the literature.
A common mistake is to become infatuated with with a topic before it has
been thoroughly searched and reviewed in the literature. Stated differently, at the
beginning stage of the research process, the topic is still evolving as a work in
progress and it is important to stay open-minded to possibilities perhaps not yet
considered. Most likely, the topic will change or become modified to some extent. A
broad, general topic can become clarified, focused, and refined; however, an overly
specific, narrow topic can be so limited that it leaves little room for exploration.
CONCEPT CHECK
When choosing a research topic, we suggest you begin by selecting a general subject
or issue that is of interest to you either personally or professionally (or both).
problem can often be addressed from more than one perspective, using more than
one type of methodology. Searching and reviewing the existing literature can pro-
vide perspective on how similar questions or problems have been examined in the
past. Additionally, reviewing the discussions, limitations, and suggestions for fu-
ture research directions from relevant earlier studies can be helpful in planning
the methods to test the hypothesis of the current study.
When planning the methods and procedures, it is important that what is
being measured in the study is clearly connected to the hypothesis and the state-
ment of the problem. The methods must focus on the variables of interest and how
these variables will be measured or evaluated so as to best answer the research
question. The methods are like a detailed recipe or formula. They explain the step-
by-step directions describing what variables will be measured or tested, and how.
The methods must be accurately followed and must be able to be replicated.
Simply stated, if someone else were to follow the methods of a study exactly, then
they should be able to come to the same results as the original researchers. It is im-
portant to keep in mind that the research questions drive the research methods,
and that methods need to be flexible, broad, and available for clinical use
(e.g., practical).
CONCEPT CHECK
The methods must focus on the variables of interest, and how these variables will be
measured or evaluated so as to best answer the research question.
Students should recognize that statistics are not simply a tool that is used
after the data has been collected. In evidence-based approaches, it is particu-
larly important to simultaneously merge research methods and data analysis at
every level of the research process in order to help generate more meaningful
research questions that are both clinically meaningful and methodologically
sound. This notion will be further discussed and elaborated in Part III of this
book. Part III will take a big picture look at the application and integration of
data analysis and statistical methods in diagnosis, prevention, and treatment in
clinical research. Research designs and data analysis will be discussed for
each type of study introduced, and examples will be provided to illustrate
application.
While you might not need to know the details of statistical analysis, research
methodology will be presented in detail, because you might become researchers
or professors in academic and clinical settings in the future. Introduction to basic
statistics, or common statistical tools traditionally covered in an entry-level sta-
tistics course, usually precedes a course in research methods. Whether or not you
may have had formal training in statistics, statistical concepts must be applied in
a research methods course as part of the research processparticularly when
approached from an EBP perspective. Rather than separating the concepts of re-
search methods and statistics, coverage of statistics is thematically linked and
woven throughout chapters in Parts II and III of this book, including clinical in-
terpretation of statistical results and application to real-world examples.
CHAPTER SUMMARY
Research is a formal process of questioning and exploring particular relationships
among or between variables of interest. The research process can be explained in
nine steps: (i) identifying a topic, (ii) searching and reviewing the literature, (iii)
defining a topic, (iv) stating a general question or a problem, (v) phrasing an
operationally defined hypothesis, (vi) planning the methods to test the hypothe-
sis, (vii) collecting data, (viii) analyzing data and interpreting results, and (ix)
writing about the findings. Although the various steps may be explained with
minor dissimilarity, together they are commonly recognized as research methods.
During the research process, ideas are tested but not proven. The information
yielded from research can be used to generate a more complex series of ideas or
theories. Research findings can be translated into clinical practice so that treat-
ment protocols and evaluative techniques are the result of careful scientific study.
Consequently, evidence-based clinical practices are developed from a systemati-
cally organized body of knowledge about a specific subject or problem. As such,
evidence-based clinical practices are carried out according to methods that are
grounded in theory and science, and then balanced with clinical expertise and
practical experience. By considering research methods as a type of decision-
making framework, clinicians can interpret and apply scientific findings to an
evidence-based clinical practice model.
KEY POINTS
Empirical research is a problem-solving method for decision making.
Evidence-based clinical practices are based on tested relationships and
logical, deductive reasoning.
Research lends supportive evidence for already existing relationships
between variables.
Research does not prove anything.
Research results must be repeatable in order for the research to be reliable.
Problem statements should formally announce the intention of the research
question.
LibraryPirate
97689_c01 01/07/10 3:39 PM Page 16
Applying Concepts
For the following problem-solving exercises, visit the MEDLINE and
PubMed websites and search a variety of topics that you think are inter-
esting to you. Then attempt to narrow down your search and select a spe-
cific topic. Bring three articles pertaining to your specific topic to class
with you.
1. Consider the clinical implications of a research article in which the re-
sults do not match the conclusion(s) and an article that manipulates the
findings of the study to justify a desired clinical effect or outcome.
2. Explain and describe why the topic you selected is important to you ei-
ther personally or professionally, and consider why it might be a poten-
tially interesting research topic.
REFERENCES
Baird D. A Thousand Paths to Wisdom. Naperville, IL: Sourcebooks, Inc.; 2000.
Bulpitt CJ. Confidence intervals. Lancet. 1987;1(8531):494497.
Christina RW. Whatever happened to applied research in motor learning? In: Skinner
JS, Corbin CB, Landers DM, et al., eds. Future Directions in Exercise and Sport Science
Research. Champaign, IL: Human Kinetics; 1989:411422.
Godfrey K. Simple linear regression in medical research. N Engl J Med. 1985;313(26):
16291636.
LibraryPirate
97689_c01 05/07/10 5:40 PM Page 17
Haynes RB, McKibbon KA, Fitzgerald D, et al. How to keep up with the medical literature:
V. Access by personal computer to the medical literature. Ann Intern Med.
1986;105(5):810816.
Knight KL, Ingersoll CD. Developing scholarship in athletic training. J Athl Train.
1998;33:271274.
Pyrczak F, ed. Completing Your Thesis or Dissertation: Professors Share Their Techniques and
Strategies. Los Angeles, CA: Pyrczak Publishing; 2000.
Sackett DL, Haynes RB, Guyatt GH, et al. Clinical Epidemiology, a Basic Science for Clinical
Medicine. 2nd ed. Boston, MA: Little Brown & Co Inc; 1991:218.
The Pennsylvania State University Graduate School. Thesis Guide. Available at:
http://www.gradsch.psu.edu/current/thesis.html. Accessed November 2006.
Thomas JR, Nelson JK. Research Methods in Physical Activity. 4th ed. Champaign, IL: Human
Kinetics; 2001.
SUGGESTED READING
1. Kuhn TS. The Structure of Scientific Revolutions. Chicago, IL: University of Chicago
Press; 1970.
2. Popper KR. The Logic of Scientific Discovery. London: Hutchinson; 1968.
LibraryPirate
97689_c02 01/07/10 3:44 PM Page 18
CHAPTER 2
HOW TO READ RESEARCH:
EVALUATING RESEARCH
ARTICLES
CHAPTER OBJECTIVES
After reading this chapter, you will:
Know some basic tips for reading research.
Understand why most research is read.
Know the appropriate steps to follow when learning to read research.
Learn how to identify various sections of a research article.
Be able to explain what to expect in each section of a scientific article.
Understand common difficulties in reading research.
Know how to plan an effective and efficient reading strategy.
Understand the role and benefit of self-assessment in determining your level of reading
comprehension.
KEY TERMS
appraisal plan scientific writing
critical analysis preparation self-assessment
interpretation reading comprehension strategy
18
LibraryPirate
97689_c02 01/07/10 3:44 PM Page 19
INTRODUCTION
When first learning to read research, many students and practitioners alike
complain of feeling overwhelmed by variations in formal layout, writing styles,
and technical terminology from journal to journal. A common complaint is a
resounding expression of frustration at not being able to see the forest for
the trees when attempting to read formal, scientific reports of research. The
purpose of this chapter is to provide tips for reading research. Our goal is to
provide a mini guidebook to help you keep away from the pitfalls of what is
sometimes considered a shadowy and mazelike journey through the minutiae
of scientific writing.
As a brief overview of the general ideas of this chapter, you will learn the fol-
lowing points about reading and evaluating research: The purpose should be
clearly stated in the introductionthat the methods permit achieving the pur-
pose, that the data are tied to the central question or questions, that the results
make sense, given the data, and justify the conclusion, and finally authors are
open regarding the limitations of the study and thus the strength of the conclu-
sions (Figure 2-1). Our goal is to weave this theme as a thread throughout this
chapter and guide you through the elements of scientific writing.
For explanation purposes, we use the example of reading a map to illustrate
likenesses to reading research. Similarly, we submit that research should be
viewed as a respectful and purposeful process of appreciation and discovery,
Point 4: The results make sense given the data and justify the conclusion.
CONCEPT CHECK
Some common difficulties when reading research include feeling overwhelmed by
variations in formal layout, writing styles, and technical terminology from journal to
journal.
WHEREABOUTS
Before you begin your travels, it could be argued that it would be helpful to have
in mind either a destination or an objective. In other words, what is the reason or
purpose for your intended travel? For example, are you traveling to get some-
where? Or, are you traveling to leave somewhere? We suggest that it is practical
to take into account four initial aspects before you set out:
Consider the following question: Would you prepare differently if you were
planning to trek along a public walkway versus in a national park? How might you
prepare if you were planning a multistate hike of the 2175-mile footpath known as
the Appalachian Trail that follows the crest of the Appalachian Mountain ridgelines
from Georgia to Maine? By comparison, what might you plan differently if you in-
tended to stroll leisurely on a jaunt through the local city park during your 30-minute
lunch break? Your purpose will largely influence and determine your strategy.
Just as there are obvious advantages to planning ahead and preparing a travel
strategy before embarking on a demanding outdoor adventure trip, it can be to
your advantage to plan ahead and prepare a reading strategy before you delve
into evaluation and appraisal of a challenging scientific article. The essential crux
of your reading strategy requires that you clearly establish your intention or rea-
son for reading the specific article. For example, are you reading for pleasure? Or,
are you reading for evidence discovery because you seek accurate and reliable an-
swers to a specific clinical question? Here again, your purpose will largely influ-
ence and determine your strategy.
LibraryPirate
97689_c02 01/07/10 3:44 PM Page 21
CONCEPT CHECK
We suggest that our students consider the following questions in preparation for read-
ing scientific writing: who, what, where, when, how, and why?
LibraryPirate
97689_c02 01/07/10 3:44 PM Page 22
more of these questions is not clear, then we recommend rereading the relevant
section(s) to find out the answer(s). The redundancy of this rudimentary system
is deliberate so as to help students and practitioners become skillful at planning
ahead and preparing to read scientific writing, as well as to learn to self-assess
their level of reading comprehension when attempting critical analysis of a re-
search article.
CONCEPT CHECK
The headings of a scientific article indicate the type of information to expect in the
various sections of a research article.
Where to Start
Although there are certainly many different approaches and anecdotal tips for
how to read research, we offer the following steps as a soft, flexible framework
rather than a rigid, uncompromising progression. As you become more adept at
reading scientific writing, you will undoubtedly discover your own preferred
steps. We suggest the steps suggested in Table 2-1.
3. Consider the key terms: Are you familiar with the terminology?
5. Read the body of the article Do you get a sense of how one section
methods, results, and discussion: transitions to the next?
7. Consider the references: Are they current, have you read any of them?
LibraryPirate
97689_c02 01/07/10 3:44 PM Page 24
If you can answer NO to any one of the questions 1 to 6 from Table 2.1, then
we urge you to reevaluate the appropriateness of your article selection, as you
may not yet be ready for the level of complexity or difficulty of the article at this
point. If the article was assigned to you rather than selected by you, then we
strongly encourage you to carefully repeat the suggested steps while keeping a
notepad nearby. List and look up any unfamiliar words. Use the notepad to list
any remaining questions you have after reading the article section by section at
least two times in its entirety. The list of questions will be helpful when you later
discuss your confusions with a professor or colleague. Lastly, we suggest that you
consider reading some of the supportive articles listed as references. The refer-
ences can often provide important background information and prerequisite
knowledge about the topic or relevant points of discussion made by the authors
of the current article.
CONCEPT CHECK
We suggest that you consider reading some of the supportive articles listed as references.
The Title
The title should be highly descriptive and provide direct insight into the topic of the
article; however, the title is merely a starting point and cannot fully represent the
scope and content of the article. Regardless of the length of the title, it is important
to remember that the title alone usually lacks the necessary details upon which you
will need to determine the relevancy and readability of the article. In most cases
when conducting a search of the literature, it is necessary to read the abstract as well
as the title before determining if an article is germane to your investigation.
The Abstract
Abstracts are typically limited in length and many journals have abstract word
limits as low as 150. The abstract provides a snapshot view of the article and high-
lights the author(s) perception of the rudimentary yet fundamental information
detailed in the body of the article.
LibraryPirate
97689_c02 01/07/10 3:44 PM Page 25
The Body
Just as nonmilitary GPS receivers are accurate to within 10 feet, there may be some
variation in the structure of a research article; however, you can comfortably ex-
pect for published research in health science to organize the body of the article
using Introduction, Methods, Results, and Discussion sections.
CONCEPT CHECK
The purpose should be clearly stated in the introduction.
The Methods Section The Methods section of an article should permit the reader to
replicate the study completely. The reader should expect the Methods section to
accurately describe how, when, and where data were acquired. Likewise, the meth-
ods should provide sufficient detail to replicate the data analysis. It is not uncom-
mon to see the use of subheadings in a Methods section.
CONCEPT CHECK
The methods permit achieving the purpose; and, the data are tied to the central ques-
tion or questions.
The Results Section The Results section should provide answers to the question or
questions posed in the introduction. The reader should expect the reports of statis-
tics, as well as tables and figures to complete and enhance the presentation of the
findings. The reader should expect the most important results to be reported first.
CONCEPT CHECK
The results make sense, given the data, and justify the conclusion.
The Discussion Section The Discussion section should put the results in scientific
and, where appropriate, clinical perspective and offer one or more conclusions.
The reader should expect the discussion of the results to logically connect the
LibraryPirate
97689_c02 01/07/10 3:44 PM Page 26
results of the current study with previous literature as well as formulate recom-
mendations and ideas for future research. The limitations of the study should be
addressed by the authors.
CONCEPT CHECK
Authors of scientific writing are open regarding the limitations of the study and thus
the strength of the conclusions.
CHAPTER SUMMARY
The objective of this chapter was to provide basic guidance and direction for read-
ing research. We offered suggestions and insights to help the reader progress from
section to section of a scientific article. Although we noted the benefits of planning
and preparation, we feel it is necessary to stress the significance of patience and
diligence when reading scientific writing. Checklists and helpful tips are only as
useful as the time you commit to using them; becoming skillful at reading research
takes time and repetition. The more familiar you become with the research litera-
ture, the more skillful you will become at reading research, and vice versa.
Eventually, students of research methods realize that they must immerse them-
selves in the research literature if they are to fully understand the research process.
A hasty review of the literature or a superficial grasp of the research is difficult to
hide from others more experienced and/or knowledgeable. There are no shortcuts
to reading or conducting research, and accurate interpretation of research is key
to a successful evidence-based approach.
KEY POINTS
The essential crux of your reading strategy requires that you clearly establish
your intention or reason for reading the specific article.
Most research is read for the like purpose of finding something out (i.e., ob-
serving, finding, or revealing certain information).
Sections and section headings provide order and enhance the transitions
through the path of the manuscript.
The title is merely a starting point and cannot fully represent the scope and
content of the article.
Abstracts are typically limited in length with a 150 word limit being common.
The Introduction section should identify a problem or a question in need of study.
The Methods section of an article should permit the reader to replicate the
study completely.
LibraryPirate
97689_c02 05/07/10 5:40 PM Page 27
Applying Concepts
For the following problem-solving exercises, visit the MEDLINE and
PubMed Web sites and search a variety of topics that you think are inter-
esting to you. Then attempt to narrow down your search and select a spe-
cific topic. Bring three articles pertaining to your specific topic to class
with you. You may use the same topic as the three articles previously at-
tained for the problem-solving exercises in Chapter 1, but you may not
use the same three articles.
1. Complete the questions listed in Box 2.1 for each of the three articles you
selected. Be prepared to lead discussion and verbally summarize these
articles using the answers to the questions in Box 2.1 as your guide. Keep
in mind that you are required to recap the articles using your own words
to explain and describe the answers to questions in Box 2.1. Your sum-
mary cannot be a replica of the authors abstract. Explain and describe;
do not plagiarize (i.e., do not copy illegally, lift, bootleg, or reproduce
published words or creative work to which you do not own copyright).
2. Explain and describe why the results of the study you selected are im-
portant to you either personally or professionally, and consider how
this study might be potentially expanded or further explored as a re-
search topic.
SUGGESTED READING
1. For more information on Leave No Trace outdoor ethics, visit http://www.lnt.org.
LibraryPirate
97689_c03 01/07/10 3:47 PM Page 28
CHAPTER 3
EVIDENCE-BASED CLINICAL
PRACTICE: DISTINGUISHING
BEST PRACTICES
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand what is meant by evidence-based clinical practice.
Be able to explain and discuss the issues at the heart of evidence-based medicine.
Recognize the roles of theory and models in research.
Recognize the roles of theory and models in evidence-based clinical practice.
Learn how to follow an evidence-based practice approach.
Understand the difference between a systematic review and a traditional review.
Be able to describe and explain the role of evidence in defining and advancing clinical
practice.
Recognize the role and limits of research in evidence-based practice.
Understand the steps in the systematic inquiry process.
Be able to describe and explain the cyclic nature of accessing the best information avail-
able for evidence-based practice.
28
LibraryPirate
97689_c03 01/07/10 3:47 PM Page 29
KEY TERMS
clinical epidemiology epidemiology paradigm shift
critical appraisal evidence prevention
diagnostics exhaustive search theoretical model
disablement model outcomes
INTRODUCTION
Perhaps the greatest challenge for many practitioners who wish to engage
in evidence-based clinical practice is to distinguish how to address research
into prevention, diagnostics, and intervention outcomes. The issues at the
heart of evidence-based medicine (EBM) reflect this quandary of how to apply
the clinical research literature and attend to the important underlying
matters of concern: why disease and injury occur, understanding how wide-
spread the conditions are (epidemiology); and, what can we do to prevent the
condition in the patients in our offices (clinical epidemiology). Clinical epi-
demiology should encompass diagnosis, treatment, and prognosis of condi-
tions in patients in addition to prevention. Diagnostics is an obvious issue
because of the fact that without identification the next piece (treatment) is
meaningless. The other issue is cost containment.
The research methods and thus the data analysis are markedly different for
prevention and diagnostic studies thus requiring separate approaches. The same
can be said for treatment outcomes, particularly with a disablement model.
Students and practitioners alike may fumble with these issues along with the ques-
tions of what constitutes evidence and how the strength of evidence can be
weighed, interpreted, and applied. These issues will be covered in detail in
Chapter 15, which explores the concept of evidence as it applies to the advance-
ment of health care practice and patient care, but before we get ahead of ourselves
let us first consider the role of evidence in defining and advancing clinical practice.
In the era of evidence-based practice (EBP), there is growing expectation for
students and practitioners to be able to access the research literature efficiently, an-
alyze the strength of evidence, conduct article appraisal, and understand basic re-
search designs. While it might not be necessary to know the details of statistical
analysis, undergraduate and postprofessional doctoral students will need to
know the research methodology in detail because they will be researchers or pro-
fessors in academic and clinical settings in the future. Acquiring skills in research
and critical appraisal is fundamental in learning to assess and regard the signifi-
cance of efficiently accessing and effectively determining the applicability, reliabil-
ity, and validity of published research in order to distinguish best clinical
practices. The goal of this chapter is to guide you through academic theory and
LibraryPirate
97689_c03 01/07/10 3:47 PM Page 30
CONCEPT CHECK
Acquiring skills in research and critical appraisal is fundamental in learning to assess
and regard the significance of efficiently accessing and effectively determining the ap-
plicability, reliability, and validity of published research in order to distinguish best clin-
ical practices.
ASSESS
clinical or policy problems and identify key issues
APPLY
evidence to clinical or policy
problems
ASK
well-built questions that can
be answered using
evidence-based resources
APPRAISE
the validity, importance, and
applicability of evidence that
has been retrieved
ACQUIRE
evidence using selected, pre-appraised resources
CONCEPT CHECK
Fundamentally, EBP can be viewed as a systematic inquiry process through which stu-
dents and/or practitioners (i) assess, (ii) ask, (iii) acquire, (iv) appraise, and (v) apply
evidence to answer clinical problems.
and implementation has conceivably added to the debate and confusion over the
development of evidence-based clinical practice.
While the notion of applying evidence to clinical practice is not novel, the more
historical customs of following traditional dogma and anecdotal accounts as evi-
dence of alternative treatment approaches have generated misapprehension about
todays judicious and systematic attempts to establish clinical practices that are con-
sistent with the best research evidence. Recognizing that valid evidence and sound
data from high quality, ethical clinical research does not immediately translate into
EBP, students and practitioners must be trained to independently and skillfully ap-
praise the research evidence (Claridge & Fabian, 2005; Guyatt et al., 2000). Those
same skills necessary to provide an evidence-based solution to a clinical dilemma
can be recognized as a series of actions and principles that symbolize the evolution
of the evidence-based clinical practice paradigm itself: identify and define the prob-
lem, plan and carry out a well-organized search to find the best evidence, critically
appraise the evidence/literature, and carefully weigh the evidence and its implica-
tions with regard to patient values and belief system (Guyatt et al., 2000; Haynes
et al., 2002; Sackett et al., 2000). With regard to the evolution of the (evidence-based
clinical practice) paradigm, the distinguishing characteristic of EBP is the hierarchy
it assigns to specific categories of evidence (Sehon & Stanley, 2003).
CONCEPT CHECK
The role of theory in evidence-based clinical practice is a type of conceptual frame-
work for a set of ideas, principles, or guidelines that can provide a contextual outline
for students and practitioners in the process of coming to a conclusion or determina-
tion about a clinical situation.
Theory in Research
As discussed in Chapter 1, theory provides both a goal and a guide to empirical
research. The term theory can refer to a set of hypothetical circumstances or
principles used to systematically examine complex concepts. It can also be used to
provide the underlying explanation for particular areas of speculative knowledge
concentrated on causality or phenomenological existence in the basic sciences.
Stated more simply, theory represents a frame of reference, school of thought, or
paradigm of science.
LibraryPirate
97689_c03 01/07/10 3:47 PM Page 34
CONCEPT CHECK
In evidence-based clinical practice, the information made known through research is
interpreted and applied by clinicians to help guide and support practice decisions.
concerned about educational practices may wish to implement EBM into their
clinical practice simply to update themselves and stay current with the latest
medical literature.
The clinical practitioner may seek to incorporate holistic perspectives and in-
clude EBM principles in their practice by asking clear and answerable questions
about specific clinical problems, systematically searching and obtaining relevant
research literature, appraising the research evidence for quality and appropriate-
ness, carefully interpreting and weighing the evidence with their own clinical ex-
pertise and experiences, and then deciding if or how best to apply the evidence
with regard to the patients personal values and belief systems (Gronseth, 2004;
Khan & Coomarasamy, 2006; Sackett et al., 1996). Although students are taught
the process of EBM, the process may seem overwhelming and intimidating to the
clinical practitioner faced with day-to-day practice issues of economics and time
constraints. The goal of the following sections is to empower clinicians with EBM
techniques for finding, identifying, and using the best evidence help bridge the
gap between research and clinical practice.
clear, answerable question about it. In EBP, it is imperative that clinicians identify
the clinical problem and formulate a good, clear, answerable clinical question be-
cause the question actually guides the search strategies for finding the evidence,
determining the relevance of the evidence, and evaluating the quality of the evi-
dence. The answers will only be as good as the questions asked. Assessment
yields information; and, informed questions yield informed answers. Good infor-
mation generates more good information.
individual points in a blueprint for a detailed inquiry and examination into some
particular topic of interest. Each stage serves as a necessary step in the process;
and, collectively, the steps frame a method of problem solving. Systematic review
is a method of investigative problem solving used in EBP. Accordingly, it is impor-
tant to keep in mind that EBP is a framework for applying research evidence to
clinical practice.
CONCEPT CHECK
Essentially, EBP is about recognizing clinical problems, asking good clinical questions,
finding, critically evaluating, analyzing, and synthesizing the evidence, and applying the
best, most relevant evidence to clinical decisions and patient care recommendations.
CHAPTER SUMMARY
In evidence-based clinical practice, research findings are applied to clinical practice
in the form of clinical recommendations. Clinical recommendations are suggested
or approved by the practitioner as the best course of action based on conclusions
drawn from the evidence that was summarized and interpreted during a careful
systematic review of the existing research literature. The research evidence is
weighed with clinical insights and expertise to determine optimal patient care.
The systematic review process is guided and directed based on the clinical
question. The more skilled the clinician or student is at precisely defining a patient
problem, and what information is required to resolve the problem (assessing), the
better the clinical question will be (ask). A better clinical question leads to a more
efficient search of the literature (acquire); better selection of relevant studies and
better application of the rules of evidence leads to determination of the validity of
the evidence (appraise); better evaluation of the evidence leads to better applica-
tion of evidence to the patient problem (apply). This process of literature review
and evaluation is known as the critical appraisal of evidence; and, critical ap-
praisal is the crux of evidence-based clinical practice and deciding best practices.
KEY POINTS
The distinguishing characteristic of EBP is the hierarchy it assigns to specific
categories of evidence.
Evidence-based practice is not an attempt to replace clinician judgment with
scientific research.
Theory is based on general principles rather than specific instances.
In scientific research, theory is introduced and based on a set of testable hy-
potheses or a conjectured relation between independent and dependent
variables.
Rigor refers to methodological soundness and addresses the appropriate-
ness of the research methods in the study.
Credibility refers to the integrity of the research findings and addresses the
presentation and meaningfulness of the results.
LibraryPirate
97689_c03 01/07/10 3:47 PM Page 41
Relevance refers to the clinical importance of the study and addresses the
usefulness and applicability of the results to the clinician and patient.
There is no substitute for clinical experience; thus, the role of research in
EBM is that of a complement to clinical experience and patient preference.
The stages of the scientific research process are the same, regardless of the
problem or question being addressed.
The systematic review process is guided and directed based on the clinical
question.
Applying Concepts
1. Form small discussion groups of three to five people. Consider the pro-
gression of evidence-based clinical practice. Talk about why it is im-
portant that clinicians be able to access, evaluate, interpret, and apply
the medical literature by following methodological criteria used to sys-
tematically determine the validity of the clinical evidence. Provide an
example based on a clinical scenario and identify applicable preven-
tion, diagnostics, and intervention outcomes.
2. Provide examples of theory and law as they connect to clinical practice
search topic.
REFERENCES
CASP. The CASP elearning resource: An Introduction to Evidence-Based Practice.
(http://www.caspinternational.org/ebp) This website was developed by the Critical
Appraisal Skills Programme (CASP) and the Department of Knowledge and
Information Science, part of the Public Health Resource Unit, Oxford, UK; 2006.
Claridge JA, Fabian TC. History and development of evidence-based medicine. World
J Surg. 2005;29(5):547553.
LibraryPirate
97689_c03 05/07/10 5:41 PM Page 42
SUGGESTED READING
1. Couto JS. Evidence-based medicine: a Kuhnian perspective of a transvestite non-the-
ory. J Eval Clin Pract. 1998;4:267275.
2. Guyatt GH, Rennie D. Users' Guide to the Medical Literature: A Manual for Evidence-Based
Clinical Practice. Chicago: AMA Press; 2002.
3. Miles A, Bentey P, Polychronis A, et al. Evidence-based medicine: why all the fuss?
This is why. J Eval Clin Pract. 1997;3:8386.
4. Russell Keeps Post by Vote of 11 to 7; Court Fight Begun: Leaders in the Russell
Controversy, Times Wide World Photos. New York Times (1857current file). New York:
March 19, 1940:p. 1 (2pp.).
LibraryPirate
97689_c04 01/07/10 3:54 PM Page 43
CHAPTER 4
EBM: A HISTORICAL
PERSPECTIVE
CHAPTER OBJECTIVES
After reading this chapter, you will:
Have an appreciation for the role of clinical epidemiology in evidence-based medicine.
Understand why the practice of evidence-based health care requires the continued
pursuit of clinical research and the dissemination of that research to the provider.
Know that the research-consuming clinician must be able to appraise the research literature
in order to apply the best available evidence to their clinical decision making.
Learn how and why evidence-based medicine is central to the advancement of medicine.
Be able to explain the differences between basic research, field research, and translational
research.
Know that clinical research asks questions about the usefulness of diagnostic tools and
the effectiveness of prevention and treatment efforts.
Understand the general goals of clinical research and evidence-based practice.
Appreciate how the study of cost analysis will bring additional evidence to the decision-
making process once a level of proficiency is developed in consuming and critically ap-
praising research related to screening, diagnosis, prevention, and treatment.
43
LibraryPirate
97689_c04 01/07/10 3:54 PM Page 44
KEY TERMS
art of medicine clinical judgment paradigm
bench research Cochrane collaboration randomized clinical trials
biomedical research information technology translational research
clinical epidemiology interventions
INTRODUCTION
Perhaps the most remarkable aspect of the history of evidence-based medicine
(EBM) is that it is not very long. In 2005, Dr. David Eddy acknowledged that since
1990 EBM has become increasingly more accepted and established in the field of
medicine. He continued by stating that the term evidence-based medicine first
appeared in print in papers authored by Dr. Gordon Guyatt in 1991 and 1992
(Evidence-Based Medicine Working Group, 1992). Given the long history of ef-
forts to treat conditions of the body and mind, EBM is in its relative infancy.
As with many new concepts, the development of what has been described as
the dominant new paradigm in medicine has origins dating back before the label
was attached. Many have labored and written to develop the central concepts of
EBM, but most new, large endeavors can be traced to a few pioneers. The work
of Chalmers I. Archie Cochrane (1909 to 1988), known as Archie and after whom
the Cochrane collaboration is named, has been identified as the springboard
for the development of EBM (2006). While even a short biography on Archie
Cochrane (Chalmers, 2006) is fascinating, his efforts to promote the use of
randomized clinical trials to collect data to inform clinical practice certainly were
important to the development of the paradigm of EBM.
Drs. David Sackett, Gordon Guyatt, and colleagues have also had a vast im-
pact on the promotion and teaching of EBM as well as the methods of identifying
best evidence. These authors also coined the term clinical epidemiology, which is
really the vehicle through which evidence-based health care is practiced.
While medicine was the original focus of the EBM, the paradigm has spread
throughout health care. The concepts and principles apply to nonphysician providers
including physical therapists, occupational therapists, and certified athletic trainers.
In todays complex health system, an understanding of EBM and clinical epidemiol-
ogy is important to all providers striving to improve patient care by employing the
best diagnostic strategies, and most effective prevention and treatment intervention.
CONCEPT CHECK
Clinical epidemiology can be described as the vehicle through which evidence-based
health care is practiced.
LibraryPirate
97689_c04 01/07/10 3:54 PM Page 45
practice patterns, when based on the best evidence, plans of care will likely
become more similar.
The universal practice of evidence-based health care is a lofty goal. Over
time, advances in information technology and clinical research are changing
health care practices, improving the outcomes of care, and in some cases, re-
ducing the costs of care by eliminating unnecessary diagnostic procedures and
ineffective interventions. Much of this book is devoted to research methods
and data analysis, not because we strive to prepare clinical researchers but be-
cause the research-consuming clinician must be able to appraise the research
literature in order to apply the best available evidence to their clinical decision
making.
CONCEPT CHECK
The paradigm of evidence-based practice has emerged so that the best clinical re-
search is applied to the treatment decisions made on similar patients across disciplines,
facilities, and regions. The best evidence is now more available to care providers every-
where because of the advances in information technology practice patterns.
RESEARCH IN MEDICINE
Research is without a doubt central to the advancement of medicine. Biomedical
research is a vast and diverse field where questions related to the functions of the
body, disease, responses to medications, injury mechanisms, and disease and in-
jury patterns are but a few that are addressed. Biomedical research is much like a
spiderweb, with strands representing areas of study that are often intricately con-
nected to address complex problems.
Take, for example, the challenge of preventing anterior cruciate ligament
(ACL) tears. In the last 50 years, much has been learned regarding the anatomy
and histology of the ligament. Biomechanists have explored normal and patho-
logic loading. Epidemiologists have identified those most at risk of injury, and fac-
tors contributing to risk have been identified. Each research effort has expanded
the understanding of ligament structure, injury, and treatment. With greater un-
derstanding, efforts at prevention of ACL injuries have included rule changes,
bracing, and exercise programs. These efforts have been based on existing knowl-
edge and are well intended. Sound reasoning and good intentions, however, do
not necessarily lead to effective interventions.
Archie Cochrane championed randomized-controlled clinical trials as the
best means of determining which interventions are truly effective in preventing
and treating illnesses and injuries. The need for and methods of such research
LibraryPirate
97689_c04 01/07/10 3:54 PM Page 47
CONCEPT CHECK
Biomedical research is a vast and diverse field where questions related to the functions
of the body, disease, responses to medications, injury mechanisms, and disease and in-
jury patterns are but a few that are addressed. Biomedical research is much like a spi-
derweb, with strands representing areas of study that are often intricately connected
to address complex problems.
CLINICAL RESEARCH
Medical research, like research in other fields, is not of singular design or purpose.
The diversity in how and where research is conducted has lead to components
of the large research picture being labeled. Basic science or bench research is often
thought of as being conducted in a laboratory environment under tightly con-
trolled conditions. Field research is conducted away from the laboratory, often
in a natural setting. Translational research is a more recent term used to describe
investigations that apply the results from basic science to the care of patients.
Sometimes referred to as bench-to-bedside, translational research seeks to speed
the development of more effective patient care strategies. The completion of the
translation of research findings to patient care requires what is perhaps best la-
beled clinical research.
CONCEPT CHECK
Basic science or bench research is often thought of as being conducted in a labora-
tory environment under tightly controlled conditions. Field research is conducted away
from the laboratory, often in a natural setting. Translational research is a more recent
term used to describe investigations that apply the results from basic science to the
care of patients.
LibraryPirate
97689_c04 01/07/10 3:54 PM Page 48
Clinical research asks questions about the usefulness of diagnostic tools and the
effectiveness of prevention and treatment efforts by enrolling patients and at-risk
individuals. Consider the discussion of ACL injuries in the previous sections. It
makes sense that exercise regimens that train athletes to land in ways that reduce
loading of the ACL would decrease the incidence of injury, especially in popula-
tions shown to be at higher risk. However, this assumption warrants testing be-
fore investing resources in such training initiative and perhaps failing to pursue
effective alternatives. Such clinical research completes the translation from basic
science to patient applications and lies at the heart of EBM.
CONCEPT CHECK
Clinical research asks questions about the usefulness of diagnostic tools and the effec-
tiveness of prevention and treatment efforts by enrolling patients and at-risk individuals.
CONCEPT CHECK
Certainly a general goal of clinical research and EBM is to seek tests and procedures
that will identify problems when they exist (specificity) but rarely lead to false-positive
findings (sensitivity). Clinicians also want to recommend prevention efforts that are
generally effective and pose a low risk of adverse events.
LibraryPirate
97689_c04 01/07/10 3:54 PM Page 49
CONCEPT CHECK
Once a level of proficiency is developed in consuming and critically appraising re-
search related to screening, diagnosis, prevention, and treatment, the study of cost
analysis will bring additional evidence to the decision-making process.
CHAPTER SUMMARY
The paradigm of evidence-based health care is neither old nor fully established.
The past few decades have witnessed tremendous changes in health care. The
health care system continues to change, and new discoveries and technologies
offer hope of better treatment of a host of diseases and medical conditions. Greater
emphasis on translational research seeks to speed the use of new information and
technology to patient care. Unfortunately, not all new procedures and treatments
truly improve the outcomes of health care, and sometimes there are adverse unan-
ticipated events. Thus, clinical research is essential to informing the practices of in-
dividual providers. The volume of clinical research continues to increase rapidly.
Advances in information technology also permit clinicians greater and more rapid
access to the medical literature than at any time in history. It is essential that
practicing clinicians and, more importantly, tomorrows clinicians are prepared to
access, read, critically appraise, and lastly apply research findings to the clinical
decisions and recommendation made in the care of each patient.
According to David Katz (2001):
Evidence has securely claimed its place among the dominant concepts in modern
medical practice. To the extent possible, clinicians are expected to base their deci-
sions (or recommendations) on the best available evidence. Physicians may see
this as one possible silver lining in the dark cloud of managed care.
LibraryPirate
97689_c04 01/07/10 3:54 PM Page 50
Despite the challenges of working within the health care system, this is an ex-
citing time in the history of health care. The practice of health care has changed,
and while the decisions of providers are often questioned, more tools are available
to develop best practice than ever. Ultimately, learning how to practice evidence-
based health care is not about the provider but about the improved care delivered
to the patient, and we are all patients many times in our lives.
KEY POINTS
The term evidence-based medicine first appeared in print in papers
authored by Dr. Gordon Guyatt in 1991 and 1992.
The efforts of Archie Cochrane to promote the use of randomized clinical
trials to collect data to inform clinical practice were important to the devel-
opment of the paradigm of EBM.
The paradigm of evidence-based health care is neither old nor fully
established.
Greater emphasis on translational research seeks to speed the use of new
information and technology to patient care.
Clinical research is essential to informing the practices of individual providers.
Advances in information technology also permit clinicians greater and more
rapid access to the medical literature than at any time in history.
It is essential that practicing clinicians and, more importantly, tomorrows
clinicians are prepared to access, read, critically appraise, and lastly apply
research findings to the clinical decisions and recommendation made in the
care of each patient.
Ultimately, learning how to practice evidence-based health care is not about
the provider but about the improved care delivered to the patient, and we
are all patients many times in our lives.
Applying Concepts
1. Consider the role of clinical epidemiology in EBM. Would it have been
plausible for EBM to have developed without the study of clinical epi-
demiology? Is it possible to practice EBM without considering, inter-
preting, or applying clinical epidemiology? Explain and provide
rationale as to why or why not.
2. Consider the roles and influences of randomized clinical trials (RTC)
and clinical epidemiology in the practice of EMB.
REFERENCES
1. Chalmers I. Archie Cochrane (19091988). In: The Lind Library; 2006. Available at:
www.jameslindlibrary.org. Accessed August 19, 2008.
2. Eddy DM. Evidence-based medicine: a unified approach. Health Aff. 2005;24:917.
3. Evidence-Based Medicine Working Group. Evidence-based medicine: a new approach
to teaching the practice of medicine. JAMA. 1992;268(17):24202425.
4. Guyatt GH. Evidence-based medicine. ACP J Club. 1991;114(2):A-16.
5. Katz D. Clinical Epidemiology and Evidence-based Medicine. Thousand Oaks, CA: Sage
Publications; 2001.
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 52
CHAPTER 5
ETHICS AND RESPONSIBLE
CONDUCT IN RESEARCH
AND CLINICAL PRACTICE
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand what role culture takes on in the advancement of medical care.
Know how the Belmont Report supports the clients/patients right to know the procedures
that they will be exposed to or the extent to which they will be involved in as participants.
Be aware of how ethical procedures transcend the boundaries of the United States via in-
ternational agreements on health, medicine, and medical research.
Realize how the federal government can intercede to make ethical guidelines clear and
aid in the advancement of medical practices.
KEY TERMS
Belmont Report confidentiality human participants
beneficence disclosure
clinical practice ethical issues
52
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 53
INTRODUCTION
The question of ethics goes beyond the notion of shaping right from
wrong. It could be reasoned that the majority of ethical considerations and
moral standards in research and clinical practice have evolved from two
basic tenets. Arguably, these tenets are both ideological and philosophical in
nature.
The first tenet can be summarized as follows: It is essential to think about all
probable risks and/or potential harm regarding the use of humans as participants.
This first guideline is reassuring and honorable; however, not necessarily plausi-
ble as we cannot logically foresee and prevent all possible harms. Nevertheless,
the prevailing theme (intended to reflect our ethical values in the research com-
munity) is that our first obligation is to protect the participant from physical and
psychological injury. This principle is similar to the modern translation of the
Hippocratic oath, First do no harm. (Hippocratic oath, OrkoV, attributed to
Hippocrates of Cos [460370 BC] taken from the Latin phrase, Primum non nocere,
translated by Francis Adams as quoted in the Internet Classic Archive:
http://classics.mit.edu/Hippocrates/hippooath.html)
The second tenet is an associated line of reasoning: the benefit (to the partici-
pant) of participation in the research must outweigh any potential risk (to the par-
ticipant) associated with participation. Yet again, this second guideline is not
always attainable. There may be occasions when participation in research poses
only negligible risk (to the participant), yet participants derive no obvious benefit
from their participation.
The goal of this chapter is to provide an overview of historical context
and justification for the theoretical and practical application of ethics, responsi-
ble conduct, and ethical decision-making in medicine, research, and clinical
practice. Determination of rationale and intention will be suggested as keys to
help recognize high-standard, transparent ethics and responsible conduct,
and as tools to help students and clinicians draw more parallels to clinical prac-
tice and identify some definite ethics breaches. A big-picture objective
will be to reinforce the notion of evidence-based practice (EBP) and its intercon-
nectedness with ethical codes of conduct in both research and clinical
practice.
It is important to note that this chapter will promote the integration of ethical
and professional conduct at the highest standards across culturally diverse patient
populations, as well as translational and interdisciplinary research necessary for
advancement in the medical community. Ethical guidelines, however, are not ex-
empt from the changes and updates develop from research that identifies best
practices or conflicting ethical values. We take the approach that ethical and re-
sponsible conduct in research and clinical practice is crucial for both establishing
and promoting excellence, as well as public trust, in medicine. As this chapter will
provide a relatively brief synopsis on this vast subject, we strongly encourage
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 54
GENERAL BACKGROUND
As stated in Chapter 1, research is a process of scientific, methodological investi-
gation designed to acquire knowledge about particular clinical issues or support
or revise theory in clinical practice based on the knowledge discovered. Ethics and
responsible conduct guide the research process. In establishing ethical guidelines,
as in evidence-based clinical practice, the information made known through re-
search is interpreted and applied to help guide and support practice decisions and
advance ethical and responsible conduct.
In other words, research results can offer evidence that provides empirical
support and validation for clinical decisions (i.e., causation, prognosis, diagnos-
tic tests, and treatment strategy), as well as provide a stimulus for federal regu-
latory requirements and institutional policies made in an attempt to provide
ethical guidelines for responsible and professional conduct and optimal patient
care. For example, beginning January 4, 2010, all institutions that apply for fi-
nancial assistance from the National Science Foundation (NSF) were required
to provide, as part of their grant proposal, certification of training and oversight
in the responsible and ethical conduct of research for all researchers who
participate in NSF funded projects (NSF-10, January 2010; AAG Section IV B).
Similarly, all researchers who apply for financial assistance from the National
Institutes of Health (NIH) must complete online training education for ethical
and responsible conduct in research with human participants (NIH Human
Participant Protections Education for Research Teams at http://cme.cancer.gov/
c01/nih_intro_02.htm).
Ethical codes do not exist in a vacuum or in isolation from mainstream soci-
ety, as a whole. Ethical codes in the healthcare profession are developed and es-
tablished to protect the patient or client legally and provide a procedural
conduct approach for the practitioner. The degree of importance of having such
codes in the healthcare profession has resulted in many ethical standards of
practice that have been legislated into state and federal laws. While such codes
of conduct are passed into law, the origin, practice, and interpretation of them
are left to those in the specific profession where they are applied at a variety of
institutional levels.
The issues of ethics and responsible behavior in research and healthcare
have not always been reflective of clear-cut, contemporary convention. Ethics in
medical research and clinical practice have evolved over centuries of unscripted
human interaction, religious dogma, war, and government report. Today, ethics
have developed into a three-pronged Buddhist-like approach to determining
and demonstrating right thought, right word, and right action. Yet, the distinction
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 55
between right and wrong is not always evident or easily predictable. Often,
what is more easily recognizable and agreed upon is the apparent discrepancy
between that which is deemed wrong versus right. More than judgments of
wrong versus right actions or bad versus good decisions, ethical issues
contrast appropriate and inappropriate behaviors, and are based on personal,
moral, professional (i.e., cultural), and societal views of accepted, principled
criteria and guidelines for responsible conduct (Bourdieu, 1995). These levels
influence one another in a nonlinear and give-and-take fashion, yet do have
hierarchical order with societal norms serving as the supervisory rank
(Baudrillard, 1990).
Currently, many institutionalized professions have adopted an explicit
code of ethical procedures or conduct to reduce misinterpretations and en-
courage proper behavior in the workplace while reducing confusion on how to
act properly. Since the medical profession maintains a strong and stable stance
on acceptable practices, the only way the standard or paradigm of clinical
practice is changed is through the advances of legitimized research (Kuhn,
1970). When well-respected medical professionals conduct research on new
clinical procedures, diagnostic or treatment techniques (and find benefit from
their use), then these new practices become more legitimized. This is impor-
tant to understand for the practitioner when dealing with the general public
because if there is doubt in the treatment they are to receive, patients will want
to confirm that the recommended treatment is a normal and highly successful
procedure.
CONCEPT CHECK
Often a code of ethics is learned through experience and implicitly understood by all
members of the group, and these norms are often unwritten.
three parts needs to be examined with respect to their role in this model of
practice.
CONCEPT CHECK
It is important to remember that EBP takes the perspective that the consideration and
experience of the clinician, the most current research/literature available, and the pa-
tient themselves must all be weighed in determining the best or most suitable clin-
ical practice.
The Clinician
When clinicians are considered as to their roles or status it may vary based on
the designated role, professional experiences, and the period when they re-
ceived their education. For instance, the type of clinician may vary from certi-
fied athletic trainers to physical therapists to doctors (general practitioner/GP).
These three allied professions and their knowledge of injury diagnosis and treat-
ment can either vary greatly or be on an equal level. The reasons for this simi-
larity or disparity have to do with the level and frequency of experience that a
clinician has with the problem the patient is experiencing, the quality and extent
of their own education (formal, learned experiences, and self-taught), as well as
their knowledge and understanding of relevant, current research on the prob-
lem. It is important for the patient to understand and be aware of the parity of
knowledge that exists within the diagnosis and treatment of injuries in the med-
ical community.
EXAMPLE
The Patient
The patient is the third part of the EBP model and must be viewed as an indi-
vidual. Similarly, it must be understood that (based on the medical problem
they may have) patients also bring with them a variety of emotions, skills, ed-
ucation, religious beliefs (culture), and personal history when they seek treat-
ment. To direct a patient to the most productive and successful treatment, each
aspect of their background should be considered. The decision to guide an in-
dividual to a proper treatment should entail finding out what other treatments
they have received in the past and how successful were those treatments. It also
helps to find out the religious background of a patient. Without that knowl-
edge, a clinician can run into a number of problems such as recommending
someone for a treatment that violates their belief system. An example of this
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 58
consistently occurs with those practicing under the tenets of the Christian
Science Church that does not traditionally accept medical practices but instead
promotes healing of physical and mental illnesses and disorders through
prayer.
Another aspect to consider with the patient is educating the patient on the
various types of treatments available. The role of the clinician becomes twofold,
educator and advocator at the same time. A part of the education process of fol-
lowing proper ethical codes is to allow the patient to be involved in the process of
their own treatment.
HISTORICAL CONTEXT
A physical therapist friend of mine once offered the following words of
advice to me in terms of patient treatment and learning to gauge how far to
push someone as they regain muscular strength, range of motion, and so on:
sometimes you have to cross the line in order to establish it. She was not sug-
gesting blatant disregard for patient perceptions of pain, nor was she suggest-
ing cruelty or mistreatment under the excuse of a Darwinian mindset of
species adaptation and survival. She was simply noting that sometimes we
learn through trial and error, even when we are acting deliberately and inten-
tionally exercising appropriate precautions. Perhaps, the same approach could
be recognized in the development and advancement of ethical and profes-
sional standards for conducting research. However, the lessons of history
have shown bold denial in the disturbing convictions with which our prede-
cessors repeatedly chose to ignore prudence and benevolence in the name of
science.
Source: McGuire Dunn C., Chadwick G. Protecting Study Volunteers in Research: A Manual for Investigative
Sites. CenterWatch, Inc., University of Rochester Medical Center; 1999:2 4; Henderson-Harr, A. (2004). Historical
Perspectives and their Significant Impact on the Development of Ethical Standards Governing Research Using
Human Participants. Cortland, NY: State University of New York College at Cortland, Research and Sponsored
Programs Office.
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 60
EXAMPLE
German military (in particular, the Air Force), eliminating the Jewish race, and
curing homosexual prisoners of their homosexuality.
These so-called medical experiments were performed without consent on
detained prisoners of war, as well as unfortunate civilians of occupied countries.
Some of the torturous experimentation included poison experiments; freezing ex-
periments; sterilization experiments; twins experiments; incendiary bomb experi-
ments; malaria experiments; high altitude experiments; sea water experiments;
sulfonamide experiments; battlefield medicine experiments including treatment
of gunshot wounds, burns, traumatic amputations, and chemical and biological
agent exposures; and mustard gas experiments. These experiments were later
deemed crimes and the doctors who conducted the experiments were put on
trial at what became known as the Doctors Trial. Widespread news broadcasts of
these beyond-abusive crimes eventually led to trial judgments and a set of stan-
dards that we now know as the Nuremberg Code of medical ethics.
The Nuremberg Code (http://ohsr.od.nih.gov/guidelines/nuremberg.html)
was actually a list of ethical standards that collectively established a new para-
digm shift in morality by which the defendants were judged and their guilt deter-
mined. Thus began the Modern era of human research protections that
originated from the Nuremberg Code.
Medical ethics commonly include the following six values that developed out
of the Code: (i) autonomy; (ii) beneficence; (iii) nonmaleficence; (iv) justice; (v) dig-
nity; and (vi) truthfulness and honesty (informed consent has stemmed from these).
obligations, and any potential risks they may incur by participating in the research
project being presented to them. Consent forms have to be extremely explicit in
their language, and must clearly state the roles of the researcher/practitioner and
participant/client in the research study. Depending on the nature of the research,
the participant must be informed of such matters of confidentiality. If partici-
pants are going to be referred to by name they need to agree to this within the in-
formed consent. Likewise, due to the nature of most medical research it is clearly
laid out in the informed consent that the participant will remain anonymous in the
study. However, as mentioned above, there are special cases where approval for
identification might be needed as in compiling social histories or in case studies.
Either way, it is critical that the language of the consent form makes this point
clear to the participant.
For the research practitioner informed consent forms may take on the mantle
of just another procedure to go through prior to beginning their research.
Nevertheless, the informed consent form is in fact a contract between the partici-
pant and the researcher. Another way of viewing this is while the research practi-
tioner should already have the knowledge of ethical procedures such as Helsinki
Agreement and the Belmont Report, the participant often is not informed of such
detailed ethical protocols and procedures. This is where the research practitioner
has to take on the role of educator and advocator for their participants/clients in
order to protect not only the participants but also themselves. Consent forms are
the foundational bridge between researcher and research participant, and part of
the necessary procedure to maintain ethical standards while conducting research
with human participants.
Boundaries between Practice and Research In a medical setting these two terms are
often blended together and thought of as the same. However, they need to be
separated and defined as the separate entities they are. The Belmont report
makes it clear about how these two are different. Practice is often when the prac-
titioner adjusts or slightly deviates from the standard protocol of treatment to
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 63
Basic Ethical Principles The attempt here is to outline the justifications for apply-
ing codes of conduct while proceeding forward with research on human subjects.
Much of this is framed in the cultural and societal codes we value. The most rele-
vant ones here relate to respect for persons, beneficence, and justice (NCPHS,
1979).
Beneficence
Beneficence refers to the ability of the practitioner to secure and stabilize the con-
dition or well-being of the client while they are receiving treatment or are involved
in a research project (NCPHS, 1979). Beneficence can also be understood as acts of
kindness, charity, or comfort to the individual client that go beyond their normal
obligation to the client. The guiding statute here for the practitioner is that the
client should be receiving maximum benefit from participation in the study while
minimizing their exposure to threat or harm (NCPHS, 1979). Another considera-
tion for beneficence is being able to see when there is a depreciable return of bene-
fits to the client in the face of constant or increased risks to their health over time.
In these cases beneficence requires that a client be removed from such involvement.
Justice
This aspect of ethical codes of conduct arises over who should benefit from the
findings of the research produced. It is often thought that the benefactors of med-
ical research are those coming from a better social class, ideal health, age or
weight, and financial standing. When this does occur an injustice has occurred.
The goal is to equalize the playing field by applying the idea of equally distribut-
ing knowledge and treatment to correct the problem of injustice.
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 64
Applications and Informed Consent under the Belmont Report Under the Belmont
Report guidelines of informed consent consist of disclosure to the client of
what the research entails. The potential client has the choice to accept or reject
participation once all procedures and requirements of participation have been
disclosed to them. This also includes understanding of the study by the subject
and what will happen to the data collected on them once it has concluded
(NCPHS, 1979).
The disclosure of information to clients includes but is not limited to research
procedures, the goal or purpose of the study, any potential risks or possible ben-
efits, adaptations to the stated procedures, explicit statements regarding the
opportunity for the client to ask questions about the study, and the option to
drop out of the study at any time for any reason (NCPHS, 1979). Another con-
sideration is to disclose to the client their individual results and the overall re-
sults once the study has been concluded. In the case of researchers conducting
interviews, case studies, or other forms of qualitative research, the client should
be given a copy of their responses and the opportunity to edit or add to their
responses and sign off on a second and final informed consent before conclud-
ing the study. Here the overall guiding theme throughout this process is that
the standard of care is to be given to the client who is volunteering for the
study.
In most cases full disclosure is the norm; however, there are some cases where
incomplete disclosure is needed in order to maintain the validity of the study.
Such cases arise where if the clients were to have full disclosure of all intentions
of the study the results would be manipulated by their own actions (NCPHS,
1979). The ethical rule to follow in such cases is to go back to the well-being of the
client/subjects participating in the study, whereby not fully informing them no
harm will come to them (NCPHS, 1979). And again the client would be informed
of the results and the reason for incomplete disclosure at the completion of the
study. What is not acceptable here is when a researcher chooses to give only in-
complete disclosure of the study because it is an inconvenience to them or in some
cases the researcher fears that they will lose some of their subject pool if they give
full disclosure.
With any study the client needs to be fully aware of what is being asked of
them. An adequate amount of time must transpire before any client is allowed to
participate in a study. This gives the client the opportunity to fully understand
their role in the study and ask for clarification of just what is being asked of them.
The degree of understanding of what the client is being asked to do is ensured by
giving the client a questionnaire regarding the procedure(s) or putting them
through a simulation or practice run prior to commencing the studying with the
subject (NCPHS, 1979). Another consideration for comprehending the role of the
client in a study is when the subject or subject pool have limited cognitive capa-
bilities (e.g., infants, toddlers, and coma patients) and require a third party (often
the closest family member) acting in the clients best interest.
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 65
CONCEPT CHECK
Under the Belmont Report guidelines of informed consent consist of disclosure to the
client of what the research entails. The potential client has the choice to accept or re-
ject participation once all procedures and requirements of participation have been dis-
closed to them.
Subject Selection
In the selection of subjects for research studies all of the guidelines already stated
apply. The selection process of subjects needs to follow the guidelines of what has
been approved by what is considered an expert but yet objective institutional
body that is not directly conducting the research. In some cases the idea of ran-
domness satisfies the need of where the pool of subjects can be recruited to partic-
ipate. However, subject selection must be taken with care so that certain
populations are not exploited or repeatedly used for studies (World Medical
Association, 2008; NCPHS, 1979). This is what is known as a form of injustice in
the selection of subjects for research work. In the past there have been abuses that
have warranted the protection of groups or entire social classes of people from
being overused and even abused in the name of research. Groups that are the most
vulnerable to overselection in human research work include institutionalized
groups/individuals, economically disadvantaged groups, minority groups, and
the very sick (NCPHS, 1979).
Purpose of WMA
The overall purpose of the WMA is twofold. The first being, The health of my pa-
tient will be my first consideration (World Medical Association, 2008). The sec-
ond tenet relates to an ethical code, A physician shall act only in the patients
interest when providing medical care which might have the effect of weakening
the physical and mental condition of the patient (World Medical Association,
2008). Much of the general practice and procedures in the treatment of
subjects/clients by the WMA mirror the standards set by the Belmont Report,
which has already been outlined.
provide full disclosure aims, methods, sources of funding, any possible conflict of
interest (cultural, religious, or otherwise), the relationship of the researcher(s) to
the institution, and the benefits and the risks of the proposed research to the
client/subject (World Medical Association, 2008). Another obligation to be
stressed here that falls within the guidelines of the WMA is that when the data go
to publication the results are reported accurately, all facets of the research are re-
ported, sources of funding, institutional affiliation, and any possible conflict of in-
terest (World Medical Association, 2008).
CONCEPT CHECK
Also in agreement with the Belmont Report the WMA is clear on any inherent risks
to human subjects, provided the subjects are healthy. Here as in the Belmont
Report, medical research should be conducted only if the importance of the med-
ical objectives outweighs the risks and hazards to the subject (World Medical
Association, 2008).
protocols that have the potential to save the life of the patient, restore health, or
eliminate suffering. The applications of these protocols still can only be adminis-
tered with the patients informed consent but then also knowing the method to be
used on them is untried (World Medical Association, 2008).
of what was developed in the research study. Prior to 1980, universities and other
nonprofit organizations were faced with the problem over who had the rights to
inventions or medicines when they had been developed with the assistance of fed-
eral taxpayers money and who would receive the windfall from these new inven-
tions. Often the process for public distribution of new inventions and medicines was
arduous and time consuming when it came to distributing them to the public be-
cause of legal and ethical snarls over ownership of the patents and trademarks due
to federal funding. In 1980, Senators Robert Dole and Birch Bayh put forth legisla-
tion to allow private research firms and universities to retain title to inventions with
federal funding which in turn allows these agencies to promote commercial con-
cerns for these inventions (Council on Governmental Relations, 2008). Retaining
title to inventions by the institution did not help, but ensured the practice of private
institutions or universities being able to accept funding from interested research
partners in the future (Council on Governmental Relations, 2008). Further additions
to this act were made in 1984 and 1986 to ensure that the products developed from
federal funding in private institutions continued to find quicker access to the pub-
lic for commercial use (Council on Governmental Relations, 2008). Additionally,
universities are encouraged by the government to file for patents on inventions they
develop with federal grants (Council on Governmental Relations, 2008). The trade-
off for the government under the Bayh-Dole Act occurs because they retain the license
to practice the patent throughout the world as well as retaining march-rights to the
patent (Council on Governmental Relations, 2008).
The Bayh-Dole Act has several provisions that address a number of ethical is-
sues related to the titling of inventions. For instance, universities have to properly
manage the invention in the public market when it is distributed for commercial
use (Council on Governmental Relations, 2008). Proper managing of these inven-
tions is set and enforced by the related or appropriate federal agency that granted
the funding to the research firm or university (Council on Governmental
Relations, 2008). While this may appear to be the government interfering with re-
search and development, it is in reality quite the opposite. Agencies such as the
National Institutes of Health (NIH) have electronic reporting systems that allow
institutions such as universities to enter reports and data directly into database
systems that allow the NIH to manage and review the progress of the agency re-
ceiving the grant (National Institutes of Health, 2008). Additionally, such meas-
ures have not inhibited research but have actually increased the development and
distribution of new inventions and medicines tenfold since the passing of the
Bayh-Dole Act (Council on Governmental Relations, 2008). Because of the
Bay-Dole Act some of the following discoveries have been made possible: devel-
oping artificial lung surfactant for use with newborn infants, process for inserting
DNA into eucaryotic cells and for producing proteinaceous materials, recombi-
nant DNA technology, central to the biotechnology industry, and TRUSOPT (dor-
zolamide) ophthalmic drop used for glaucoma (Council on Governmental
Relations, 2008).
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 70
CONCEPT CHECK
The Bayh-Dole Act put in place a series of ethical checks and balances for proper dis-
bursement of said inventions, along with a greater attention to the details to monitor
proper protocols and procedures are being followed by agencies receiving grants.
CHAPTER SUMMARY
The role of ethics and responsible conduct in treatment of human participants is a
product of protection for the research participant. In short, the researchers pledge
two primary assurances to their participants: (i) that they will not be placed in
harms way and (ii) that they will not be denied access to information concerning
their well-being. Such ethical guidelines for participant treatment go through a se-
ries of checks and balances to ensure proper treatment of the client, whereby the
client will be provided with practitioners within the medical field to advocate for
them. It is important to understand that the institutions (medical, academic, or
otherwise) must establish and uphold guidelines of ethical practice for uniform
treatment of research participants.
While all professionals should adhere to written and formally agreed upon
ethical standards sometimes their ability to deliver services within those parame-
ters might be different due to their educational background, professional position,
world-view, and age. In other words, although style and technique of application
may vary, all professionals within a given institution have ethical guidelines to
keep them within a common code or practice for how they treat and interact with
clients. Ethical treatment includes, but is not limited to, the methods of selection
of research participants and patients who agree to participate in experimental
studies or treatment protocols.
While the interaction with the ethical treatment of patients for medical re-
search or returning them to health is the goal of healthcare professionals, there are
other matters of importance that require ethical consideration within the health-
care field. As noted previously, legislation has helped in reducing the problems
of who has the rights to medical developments and other products that would
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 71
benefit clients. The solution of the Bayh-Dole Act helped to bridge that gap. The
institution funding the development of a new product and the institution produc-
ing the product from outside funding can have equitable common ground. This
shared role in the delivery of the product without getting entangled in territorial
snares and differences of ethical choices allows both parties to adhere to the same
practice and to speed the delivery of the product to clients.
KEY POINTS
The role of ethical treatment of clients is a product of protection for the client.
Ethical guidelines for client treatment go through a series of check and bal-
ances to ensure proper treatment of the client.
Institutions, medical or otherwise, must establish and uphold guidelines of
ethical practice for uniform treatment of clients.
Ethical treatment includes but is not limited to the methods of selection of
clients for research or aiding them in regaining their health status.
Legislation has helped in reducing the problems of who has the rights to
medical developments and other products that would benefit clients.
Applying Concepts
1. Part of maintaining ethical standards is making the right decisions when
there seems to be a degree of uncertainty. Consider the position that ath-
letic trainers are often put in with regards to injured athletes at the inter-
scholastic or intercollegiate levels. Often an athletic trainer is asked to
(continued)
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 72
compromise their ethical stance of care of the athlete due to the influence
of the athlete, coach, or in some cases the parents. For example, when an
athlete at either of these two levels (regardless of the sport) receives a dis-
located shoulder the discussion quickly moves to the severity of the dis-
location and how soon can the athlete return to competition. The issue
can be further complicated by when the injury occurred and the pressure
to return the athlete to competition especially when the injury happens
anywhere between the midpoint of the season to the end of the season.
This is where pressure to rush the athlete back to action often comes from
the coaching staff. Likewise, pressure also comes from the athlete as well
if they sense that their immediate future to continue competing in the
sport is threatened. For the high school athlete this means school losing
interest in offering a scholarship to the athlete. At the intercollegiate
level, the athlete could be facing a financial decision, whereby their draft
status could be compromised by their sitting out during critical compe-
titions in the second half of the season. Often the athletic trainer is put in
a situation where they are asked to compromise their values system of
care for an athlete on several levels. What are the inherent problems in
this situation with athletes returning too soon to competition?
2. The idea of giving options that are viable to patients is an ethical deci-
sion that will need to be made by clinicians. Take for instance the choice
of recommending a treatment procedure for a patient with allergies.
Traditionally, the choice would be to recommend antihistamines for a
patient. However, there are other therapies available and many pa-
tients are choosing to see acupuncturist for permanent allergy relief.
Explain why both options should be made to patients and the prob-
lems and successes each have for this problem.
3. Real world problems often exist for researchers attempting to further
their line of research. One of them is often funding of the research itself.
For some practitioners they are faced with problems that comprise their
ability to continue their research and ultimately their jobs. Take for ex-
ample a research team that is devoted to the preservation and the exten-
sion of life and is doing research on the stresses on the body to extreme
environmental conditions. The group has submitted several grant pro-
posals to a variety of agencies including the Federal Government to fund
their research. The only grant awarding agency at this time that will ap-
prove their research and give them funding is the United States Army.
While the nature of the research is designed to enhance life, the factor of
military interest always carries with it the potential of destruction of
human life. Discuss the ethical decisions that need to be made here in-
cluding the subject selection process if the grant is accepted.
LibraryPirate
97689_c05 05/07/10 5:42 PM Page 73
REFERENCES
Baudrillard J. Seduction. New York: St. Martins Press; 1990.
Bourdieu P. Outline of a Theory of Practice. Cambridge, UK: Cambridge University Press; 1995.
Kuhn T. The Structure of Scientific Revolutions. 2nd ed. Chicago, IL: University of Chicago
Press; 1970.
McGuire Dunn C, Chadwick G. Protecting Study Volunteers in Research: A Manual for
Investigative Sites. CenterWatch, Inc., University of Rochester Medical Center; 1999:24.
McPherson JM. Battle Cry of Freedom: The Civil War era. New York: Ballantine Books; 1988.
Council on Governmental Relations. (2008). The Bayh-Dole Act: a guide to the law and im-
plementing regulations. University of California technology transfer. Available at:
http://www.ucop.edu./ott/faculty/bayh.html. Accessed October 24, 2008.
National Institutes of Health. period (2008). Developing sponsored research agreements:
considerations for recipients of NIH research grants and contracts. NIH Guide. Available
at: http://www.grants.nih.gov/grants/guide/notice-files/not94-213.html. Accessed
October 24, 2008.
World Medical Association. (2008). World Medical Association declaration of Helsinki: eth-
ical principles for medical research involving human subjects. Available at: http://
www.wma.net/e/policy/b3.htm. Accessed October 24, 2008.
NCPHS. (1979). The Belmont Report: ethical principles and guidelines for the protection of
human subjects of biomedical and behavioral research. Department of Health,
Education, Welfare. Available at: http://www.bop.gov/news/BelmontReport.jsp.
Accessed October 24, 2008.
SUGGESTED READING
1. National Bioethics Advisory Commission. Ethical and Policy Issues in Research Involving
Human Participants. Vol 1: Report and recommendations of the National Bioethics
Advisory Commission. Bethesda, MD: National Bioethics Advisory Commission; 2001.
2. Jones J. Bad Blood: The Tuskegee Syphilis Experiment: A Tragedy of Race and Medicine. New
York: The Free Press; 1981.
3. National Science Foundation.Proposal and Award Policies and Procedures Guide (PAPPG).
NSF 10-1 January 2010; 2010.
4. National Institutes of Health. NIH Online Course: Human Participant Protections
Education for Research Teams. Avialable at: http://cme.cancer.gov/c01/nih_intro_02.htm.
5. (NA). Nuremberg Doctors Trial. BMJ. 1996;313(7070):14451475.
6. The Nuremberg Code. Available at: http://ohsr.od.hih.gov/nuremberg.php3.
7. The Journal of Research Administration. Commemorative Anniversary Edition. Volume
XXXVII, 2007.
8. Online Ethics Center. Available at: http://www.onlineethics.org.
9. Nazi Medical Experimentation. US Holocaust Memorial Museum. Available at:
http://www.ushmm.org/wlc/article.php?lang=en&ModuleId=10005168. Accessed
LibraryPirate
97689_c05 01/07/10 3:55 PM Page 74
FURTHER READING
Bulger JW. An approach towards applying principlism. Ethics Med. 2009;25(2):121125.
Fortin S, Alvarez F, Bibeau G, et al. Contemporary medicine: applied human science or
technological enterprise? Ethics & Med. 2008;24(1):4150.
Fuchs S. Relativism and reflexivity in the sociology of scientific knowledge. In: Ritzer G, ed.
Metatheorizing (Key Issues in Sociological Theory). Newbury Park, CA: Sage; 1992.
Goldblatt D. Ask the ethicist: must physicians respect an incompetent patients refusal of
treatment? Med Ethics. 2006;13(2):3.
Hoffmann D. The legal column: choosing paternalism. Med Ethics. 2006;13(2):4, 12.
LibraryPirate
97689_c06 01/07/10 3:56 PM Page 75
PART II
SEEKING ANSWERS:
HOW THE QUESTION
DRIVES THE METHODS
LibraryPirate
97689_c06 01/07/10 3:56 PM Page 76
LibraryPirate
97689_c06 01/07/10 3:56 PM Page 77
CHAPTER 6
FINDING THE EVIDENCE:
INFORMATIONAL SOURCES,
SEARCH STRATEGIES, AND
CRITICAL APPRAISAL
CHAPTER OBJECTIVES
After reading this chapter, you will:
Be able to describe the role of the reference librarian.
Understand the concept of the Invisible Web.
Be able to use Boolean logic as part of your search strategy.
Appreciate the notion search strategies for purposeful Web-based navigation for deliber-
ate inquiry.
Learn how to critically review and appraise scientific literature as evidence.
Understand the benefits of peer-review in journal article publication.
Recognize that various factors influence and determine the quality of informational
sources.
Appreciate the benefits and limitations of the journal impact factor.
77
LibraryPirate
97689_c06 01/07/10 3:56 PM Page 78
KEY TERMS
boolean search Invisible Web pubMed
database peer-review search engine
evidence proof URL
hierarchy prop
INTRODUCTION
Before attempting to effectively interpret, critically appraise, apply, or conduct re-
search, it is prudent to be familiar with how and where to find the plausible
evidence to be had in various informational sources. Evidence consists of infor-
mational particulars of the theoretical, hypothetical, anecdotal, or empirical na-
ture that can be corroborated, substantiated, or confirmed. In a research paradigm,
data are gathered as evidence to either support or refute scientific viewpoints. In
this manner, evidence can be used to corroborate or contradict previous research
findings. Consequently, evidence is sometimes erroneously inferred as proof
rather than prop for theoretical perspectives.
Evidence is evaluated based on a hierarchy of its derivational and presenta-
tional formats. Chapter 7 discusses the hierarchy of evidence in terms of its deri-
vation (i.e., research design and research methodology). This chapter discusses the
hierarchy of evidence in terms of its presentation/publication (i.e., prestige or im-
portance of a refereed scientific journal in its field).
CHAPTER 6 Finding the Evidence: Informational Sources, Search Strategies, and Critical Appraisal 79
categories and academic disciplines. They can provide instruction and guidance in
navigating print periodicals, electronic and digital reference systems, commercial
search engines, new standards for reference metrics, as well as more traditional
topics such as organizing print and electronic resources, proper citation style for-
mats and preventing plagiarism. The role of the reference librarian is important
and while they can offer a vast amount of guidance and direction in conducting
your search for evidence, they do not represent the culmination of your search.
CONCEPT CHECK
The goals of the reference librarian are to assist you in finding suitable informational
sources and to provide guidance and instruction for using effectual search strategies.
that have undergone peer-review prior to acceptance for publication. The peer-
review process is one by which a panel of experts judge the content and correct-
ness of ones work before accepting a research document for journal publication.
Peer-review is a time-consuming and arduous practice thought to be necessary
and meritorious to test and safeguard the quality of scholarly work.
CONCEPT CHECK
Those areas of the Internet that are inaccessible to search engines are collectively
known as the Invisible Web.
SEARCH STRATEGIES
Regardless of your preferred search strategy, it is important to remember that all
searches are time-consuming to some extent. The effectiveness of your search
strategies and the narrowness of your topic(s) will largely influence the amount of
time it takes to generate good, usable results. Notwithstanding the chance bene-
fits of luck and guessing, search results are only as good as the search strategies
employed. It is prudent to understand the limits of hit and miss tangential style
searches, and the value of planned investigation. Planned investigation begins
with the understanding that searching and reviewing the literature is a purpose-
ful and, sometimes, tiresome process. Guidelines and suggestions for search and
review strategies are provided in Chapter 1.
There are online tutorials for search strategies, as well as Internet searching
classes and books to aid in purposeful Web-based navigation for deliberate in-
quiry. While these instructional materials and classes can offer valuable recom-
mendations and instructions, it may helpful to consider a few basic guidelines
before beginning your search. Simple, self-explanatory steps include:
You may also wish to consider using Boolean logic as part of your search strat-
egy. Boolean searching is based on the notion of logical relationships among
search terms. Specifically, from a computer programming perspective, the opera-
tor terms of OR, AND, and NOT effect strict logically collated outcome or
LibraryPirate
97689_c06 01/07/10 3:56 PM Page 81
CHAPTER 6 Finding the Evidence: Informational Sources, Search Strategies, and Critical Appraisal 81
search results. Stated differently, each of the operator terms (i.e., OR, AND, NOT)
used to combine search terms instructs a different operation or a set of search di-
rections for the computer to follow.
Let us consider this concept in the following example using the search terms
of ankle and knee. The terms could be combined as such: ankle OR knee,
ankle AND knee, ankle NOT knee. Search results based on these combinations
would vary logically in this manner:
ankle OR knee: would generate results in which at least one of the search
terms was found
ankle AND knee: would generate results in which both of the search terms
were found
ankle NOT knee: would generate results in which only the term ankle
was found, but not if the term knee was found, even if the term ankle
was found with it.
CONCEPT CHECK
The effectiveness of your search strategies and the narrowness of your topic(s) will
largely influence the amount of time it takes to generate good, usable results.
by stating the purpose, methods, results, and a list of key terms. Depending on
the journal format and specific author submission guidelines, more or less infor-
mation may be included in the abstract. In addition to reading the abstract, the
paper is scanned for appropriateness of content, comparable context, date of pub-
lication, highly regarded authors, reputation of journal, and references. Articles
of uncertain yet potential relevance are kept until they can be eliminated without
question.
Subsequent scrutiny and critical appraisal occur through a more serious and
careful review. Critical review involves structured reading to answer questions,
identify key points, and recognize significance. Reading critically and creatively is
a skill that requires one to examine the content and context of the paper so as to
extract important information about scientific contributions from writings using
vocabulary and journal formats specific to a field or profession. Rather than
merely annihilating the work, your goal is to identify areas of strength as well as
areas for improvement. This process takes time. Seasoned professionals often read
a paper more than a few times in order to discern the scientific contributions and
nuances of an article. With this in mind, most people should plan to devote suffi-
cient time to each paper they read for critical review.
Following critical review and appraisal, it may be helpful to skim the major
parts of the paper before attempting to analyze or critique the paper. First, read
the entire paper to get a sense of the big picture and identify any unfamiliar ter-
minology or phrases. Next, read one section at a time for comprehension. Read to
figure out answers to the following questions to help you locate information in the
paper:
1. Topic: In your own words, clearly state the main topic and focus of the
study. What was the article about? What was the issue or research prob-
lem in the article?
2. Purpose: In your own words, clearly state the point of the article. Briefly
put in plain words why the study or experiment was done. Hint: The ra-
tionale for conducting a research study is typically to determine or inves-
tigate or explore the connection between or among variables of interest.
3. Methods and Procedures: In your own words, briefly describe what was
being measured in the study and how it was tested. What was the general
idea behind the nitty-gritty nuts and bolts of the study? What thing or
list of things were being questioned or measured (variables of interest)?
How were these variables measured or evaluated? Hint: identify the inde-
pendent and dependent variables of interest; state how the variables were
measures; summarize what was done and how it was done step by step
to collect and/or test the variables of interest; provide appropriate, objec-
tive detail.
4. Results: What were the statistical findings from the study or experiment?
If some thing of interest was measured, what was the outcome? How did
LibraryPirate
97689_c06 01/07/10 3:56 PM Page 83
CHAPTER 6 Finding the Evidence: Informational Sources, Search Strategies, and Critical Appraisal 83
the existence of any rules for informational sources and/or expectations for
quality of evidence in those informational sources. For example, perhaps the
use of Web sites as informational sources is frowned upon or even disallowed. Or,
perhaps, only peer-reviewed journal articles published within the last 5 years are
approved for inclusion in your search results.
University libraries sometimes provide Web sites and informational
brochures with guidelines for evaluating and critically analyzing the quality of
informational sources, including Web sites. Likewise, employers, academic de-
partments, graduate schools, or instructors may provide detailed directions
and guidelines that clarify requirements for use of various informational
sources. It is beneficial to know expectations regarding the quality of informa-
tional sources prior to initiating search strategies; be that as it may, it is imper-
ative to confirm quality requirements before finalizing critical review and
appraisal. As a rule, scholarly journals are usually preferred over nonscholarly
periodicals and/or popular magazines when conducting a search for credible
evidence.
The quality of informational sources and the notion of hierarchies of evidence
as they relate to the strength of evidence in making decisions about patient care
will be presented in detail in Chapter 7. For now, it suffices to note that differences
in research methods (e.g., sampling) and data analysis will influence the quality or
strength of evidence. Chapter 7 explores the concept of evidence, what consti-
tutes evidence, and how the strength of evidence can be weighed, interpreted, and
applied as it applies to the advancement of healthcare practice and patient care.
CONCEPT CHECK
University libraries sometimes provide Web sites and informational brochures with
guidelines for evaluating and critically analyzing the quality of informational sources,
including Web sites.
CHAPTER 6 Finding the Evidence: Informational Sources, Search Strategies, and Critical Appraisal 85
Each year, scientific journals are ranked according to impact factor in an at-
tempt to establish their relative order of importance or influence within a
professional field or scientific discipline (i.e., physical therapy). The use of JIF
has allowed for journal comparison in professional fields. Consequently, JIF has
become associated with prestige of publication status. Of recent years, debate
has ensued over whether JIF is merely a reflection of reputation as determined
by citation frequency. The crux of the issue is the distinction between quality
(based on rigor of peer-review) and perceived status (based on JIF). Simply
stated, the quality of the peer-review process is arguably not an extension of JIF
(Wu et al., 2008; Kurmis, 2003; Benitez-Bribiesca, 2002). So, what is JIF and what
does it tell us? Journals with high impact factors are regularly assumed to be
more prestigious, but are those assumptions accurate?
The impact factor of a journal for any specific year is mathematically calcu-
lated by dividing the number of times a journal was cited in the previous 2 years
by the number of articles it published in the same 2 years (Garfield, 2006). For ex-
ample, a journal with a current impact factor of 1 would have a 1:1 citation-to-
publication ratio, or an average of 1 citation for each article published in the
previous 2 years. Arguably, this metric comparison does not tell us anything
about the reputation of the peer-review process for scientific literature and a
great deal of challenge and dispute has erupted around questions of validity and
misuse.
JIF rankings are limited to the extent that frequency in citation and publica-
tion do not necessarily equate to scientific expertise, quality of scholarship, or an
active research agenda. It has been argued that the originally intended value and
purpose of JIF (Garfield, 2006) have been bastardized to encompass author im-
pact. Professionals in scientific and academic research communities are often re-
quired to achieve a tally of scholarship requirements. In other words, there is a
requirement to produce a certain number of publications within a certain time
frame. This expectation for an active and productive research agenda is regularly
referred to as the publish or perish peril of the academic tenure process.
Additionally, it has become increasingly more common for institutions to require
that a certain number of the publications happen in journals with rankings at or
above a certain JIF.
Critics argue that depending on the mass of the scientific discipline and the ex-
tensiveness of the circulation of the journal, the JIF ranking could be skewed
(Seglen, 1997; Hoeffel, 1998; Opthof, 1999; Yu et al., 2005; Garfield, 2006;
Monastersky, 2005). For example, it is conceivable that a quarterly published jour-
nal with a wide circulation in a major, well-established professional organization
would have more opportunity for citation than a semiannually published journal
with a fairly small circulation in an emergent, yet obscure and highly specialized
scientific discipline. More frequent publication equates to a larger number of arti-
cles published each year. Wider circulation reaches a bigger readership, and name
recognition happens with bigger readership. Thus, it is plausible that a high impact
LibraryPirate
97689_c06 01/07/10 3:56 PM Page 86
factor would be more likely for the quarterly published journal with extensive
readership and assumption(s) of high status in a major, well-established profes-
sional organization than for the semiannually published journal with a fairly small
circulation in an emergent, yet obscure and highly specialized scientific discipline.
Human judgment is arguably the Achilles heel of JIF, and this quandary fuels
the publication debate over journal popularity versus journal influence and impor-
tance. In other words, how many (e.g., quantity, impact) versus whom your work
reaches and the process by which it gets to that point (e.g., quality, influence).
CONCEPT CHECK
Impact factor is a mathematical rating system based on the numbers of journal cita-
tions and article publications within a 2-year time frame. This is a controversial, yet in-
fluential measure of journal evaluation, not to be confused with the hierarchy of
evidence that applies to quality of evidence as determined by the pecking-order level
of the informational source.
CHAPTER SUMMARY
Resourcefully and capably locating credible informational sources can prove to be a
tedious endeavor. Organized search strategies along with a practiced system for crit-
ical review and appraisal can greatly improve the overall outcome when attempting
to find, read, and interpret scientific literature. Strategies such as Boolean searching
and carefully comparing the reference lists of peer-reviewed research articles can
help as you learn to piece together the literature as evidence. Learning to decipher
how the evidence fits together requires patience, prudence, intellectual curiosity, and
a willingness to look beyond the obvious. There are no shortcuts to finding the evi-
dence; and, the results of your search will only be as good as the strategies you em-
ploy as you learn to better locate, critically review, and appraise the literature.
KEY POINTS
In a research paradigm, data are gathered as evidence to either support or re-
fute scientific viewpoints.
Evidence is evaluated based on a hierarchy of its derivational and presenta-
tional formats.
The role of the reference librarian is important and while they can offer a
vast amount of guidance and direction in conducting your search for evi-
dence, they do not represent the culmination of your search.
LibraryPirate
97689_c06 01/07/10 3:56 PM Page 87
CHAPTER 6 Finding the Evidence: Informational Sources, Search Strategies, and Critical Appraisal 87
Often the information provided in general Web sites and nonrefereed docu-
ments is questionable in terms of the accuracy and reliability of information
content.
The peer-review process is one by which a panel of experts judge the content
and correctness of ones work before accepting a research document for jour-
nal publication.
There are online tutorials for search strategies, as well as Internet searching
classes and books to aid in purposeful Web-based navigation for deliberate
inquiry.
The abstract is an abbreviated summation of the main points of importance
from the paper, and provides a synopsis of the article by stating the purpose,
methods, results, and a list of key terms.
JIF is widely considered a determinant of journal status associated with
Journal Citation Reports.
Applying Concepts
1. Discuss the steps you might follow to gather evidence if you were in-
terested in learning more about whether or not muscle atrophy due to
nerve damage can be prevented.
2. Review the Web pages for your university library or for your alma
mater library. Locate and compare library Web pages guidelines for
evaluating and critically analyzing the quality of informational
sources, including Web sites.
3. Select a peer-reviewed journal article of interest to you. Exchange arti-
cles with a classmate, and read them one at a time for the purpose of
critical review. Prepare an annotated bibliography. Compare your bibli-
ographic citations and summaries for accuracy and comprehensiveness.
LibraryPirate
97689_c06 01/07/10 3:56 PM Page 88
REFERENCES
Benitez-Bribiesca L. The ups and downs of the impact factor: the case of Archives of
Medical Research. Arch Med Res. 2002;33(2):9194.
Garfield E. The history and meaning of the journal impact factor. JAMA. 2006;295(1):9093.
Hoeffel C. Journal impact factors. Allergy. 1998;53:1225.
Kurmis AP. Understanding the limitations of the journal impact factor. J Bone Joint Surg Am.
2003;85:24492954.
Monastersky R. The number thats devouring science. Chron High Educ. 2005. October 14.
Available at: http://chronicle.com/free/v52/i08/08a01201.htm.
Opthof T, Submission, acceptance rate, rapid review system and impact factor. Cardiovasc
Res. 1999;41:14.
Seglen PO. Why the impact factor of journals should not be sued for evaluating research.
Br Med J. 1997;314(7079):498502, Available at:
http://www.bmj.com/cgi/content/full/314/7079/497.
Wu X-F, Fu Q, Rousseau R. On indexing in the Web of Science and predicting journal im-
pact factor. J Zhejiang Univ Sci B. 2008;9(7):582590. [Electronic version].
Yu G, Wang X-Y, Yu D-R. The influence of publication delays on impact factors.
Scientometrics. 2005;64:235246.
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 89
CHAPTER 7
THE HIERARCHY
OF EVIDENCE
First learn the meaning of what you say, and then speak.
Epictetus (AD 55135), as quoted in The Classic Quotes
Collection on The Quotations Page
(www.quotationspage.com)
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand when clinical experience should drive patient care decisions.
Be able to describe the hierarchies of evidence as they relate to the strength of evidence
in making decisions about patient care.
Be able to explain how differences in research methods and data analysis across the spec-
trum of patient care can result in some differences as to how the hierarchy of evidence
is described.
Learn why larger samples are more likely to estimate true population values and result in
narrower confidence intervals than small samples.
Understand the principal difference between an RCT and a prospective cohort study.
Understand why the RCT is a superior research design.
Describe and explain the acronyms SpPIN and SnNOUT as they relate to diagnostic studies.
Learn that there is more than one model of a research method hierarchy.
KEY TERMS
diagnosis patient values treatment plan
hierarchy prevention treatment response
89
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 90
INTRODUCTION
The paradigm of evidence-based medicine or evidence-based clinical practice has
been introduced in earlier chapters. The purpose of this chapter is to explore the
concept of evidence as it applies to the advancement of healthcare practice and
patient care. Before delving into what constitutes evidence and how the strength
of evidence can be weighed, interpreted, and applied, lets consider the alterna-
tive to evidence in defining and advancing clinical practice.
Isaacs and Fitzgerald (1999) tackled the subject of alternatives to evidence-
based medicine in their short and light-hearted paper entitled Seven alternatives
to evidence based medicine. Although entertaining, this paper confirms the ne-
cessity of evidence in the pursuit of improved patient care.
The first alternative, Eminence based medicine, suggests that experience,
seniority, and recognition are sufficient to train and advance the practices of
junior colleagues. While senior clinicians and faculty often are well positioned
for this task because of their efforts to continue to critically review the best cur-
rent literature many of us have encountered those who continue to make the
same mistakes with increasing confidence over an impressive number of years.
Thus, while clinical experience is an important consideration, only when
combined with appraisal of the best available literature should it drive patient
care decisions and the advancement of health care. Isaacs and Fitzgerald contin-
ued by identifying Vehemence based medicine, Eloquence based medicine, and
Providence based medicine as alternatives to evidence-based medicine. Certainly,
proclaiming your opinion more loudly or smoothly may work in politics or the
sale of used vehicles but has not been shown to benefit the patient or advance
the practice.
These authors conclude by identifying Diffidence based medicine, Nervousness
based medicine, and Confidence based medicine. In the first case nothing is done due
to a lack of plausible solutions to the problem and, in second, too much is done
since one does not want to miss a diagnosis or fail to use every opportunity or tool
to foster recovery. It is rare indeed that a condition exists for which the medical lit-
erature fails to provide evidence of some merit to offer guidance. Equally true is
that the literature provides guidance on diagnostic procedures and interventions
that provide no benefit to the patient. More is not necessarily better. Lastly, we re-
turn to confidence. Certainly, confidence gained from experience and continued
study of the literature helps the clinician cope with the stresses of providing
optimal care and eases the anxiety of the patients we treat. However, as alluded to
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 91
CONCEPT CHECK
While clinical experience is an important consideration, only when combined with ap-
praisal of the best available literature should it drive patient care decisions and the ad-
vancement of health care. Thus, medical decisions become evidence based.
analysis across the spectrum of patient care also result in some differences as to
how the hierarchy of evidence is described. For the purposes of this chapter the
hierarchy of evidence related to treatment responses and prevention is presented
first and developed to the greatest extent. The hierarchy of evidence related to di-
agnostic procedures and prognosis will be presented in a manner that compares
and contrasts the levels of evidence with those of treatment outcomes. The
evaluation of screening effectiveness often involves cost-benefit assessment and
lies beyond the scope of this chapter. An excellent overview of these issues was
provided by Katz (2001).
CONCEPT CHECK
It is important to keep in mind that differences in research methods and data analysis
across the spectrum of patient care can result in differences as to how the hierarchy
of evidence is described.
CONCEPT CHECK
Data from a large clinical trial with solid research methods will provide more com-
pelling evidence than a small study of similar methodological quality.
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 93
Level of Evidence
Level 2 - Cohort studies
2-a: systematic reviews of cohort studies
2-b: individual cohort studies
2-c: outcomes research
Level 4
Case series
Poorly designed cohort studies
Poorly designed case-control studies
Level 5
Anecdotal evidence
Animal research
Bench research
Unpublished clinical observations
FIGURE 7-1 The Oxford Center for Evidence-Based Medicine hierarchy. (Adapted with
permission from http://www.cebm.net/index.aspx?o=1025. Oxford Center for Evidence-Based
MedicineLevels of Evidence [March 2009]. Accessed May 10, 2010.)
Level 1
1-a: Systematic review of level 1 diagnostic studies and clinical
decision rules with level 1b studies from multiple centers
1-b: Cohort studies of good methodological quality and clinical
decision rules tested within a single center
1-c: Absolute sensitivity or specificity
Level 3
3-a: Systematic reviews including 3b, 2b, and 1b studies
3-b: Studies not enrolling consecutive patients or studies with
inconsistently applied reference standards
Level 4
Case-control studies, studies with poorer methodological quality,
studies without independent review of reference standard
Level 5
Expert opinion, animal and bench
FIGURE 7-2 Hierarchy of evidence for diagnosis. (Adapted with permission from
http://www.cebm.net/index.aspx?o=1025. Oxford Center for Evidence-Based MedicineLevels
of Evidence [March 2009]. Accessed May 10, 2010.)
trials are studies in which patients receive all interventions under consideration in
random order over the duration of the investigation. Since patients essentially serve
as their own control, within patient, randomized treatment order trials studies min-
imize the influence of differences between patients in terms of responsiveness to in-
tervention. While potentially providing strong evidence, these studies require that
the condition being treated be relatively stable and that the effect of an intervention
does not influence the response to the other interventions being investigated.
Yelland et al. (2007) provides an example of such a study. These investigators
studied the effects of celecoxib, SR paracetamol, and placebo on pain, stiffness and
functional limitation scores, medication preference and adverse effects in
patients suffering from osteoarthritis of the knee. Each patient completed three cy-
cles of paired treatment periods (2 weeks for each treatment over the course of 12
weeks) in a randomized order. The value in within patient, randomized treatment
order trials studies is readily apparent when one considers that patients will differ
considerably in their reports of pain, stiffness and functional limitation despite
having similar radiographic changes. Moreover, while symptoms may worsen
over time, large changes in a 3-month period are unlikely. Furthermore, the effects
of the medications studied are not likely to result in long-term changes. Thus the
within patient, randomized treatment order trials design provides strong evidence
and minimizes the number of patients needed as there are no comparison groups.
Within patient, randomized treatment order trials studies are not common,
however, simply because few conditions treated by physical therapists, occupa-
tional therapists, and athletic trainers are stable across long periods of time and
because the natural history and interventions directed at change are not re-
versible. Consider, for example, the management of an acute lateral ankle sprain.
First, natural history projects improvement in pain, loss of motion, and function
over time without intervention. Secondly, once impaired motion is improved it re-
mains improved. Thus, investigations into the treatment of conditions such as the
lateral ankle sprain require other research designs such as a randomized-
controlled clinical trial to be discussed shortly.
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 96
trials from designs that fall lower on the hierarchies. First, as noted previously
randomized-controlled clinical trials are prospective that is in contrast to case se-
ries or case studies as examples where patients are identified after a course of care
for analysis. Prospective designs allow for greater control over factors that con-
found the identification of causeeffect relationships. Take, for example, an effort
to determine if a protocol of three times per week treatment with low-power laser
reduces the need for surgery in the treatment of carpal tunnel syndrome. If one
were to address the question prospectively, uniform instructions could be pro-
vided to all of the patients receiving real and sham laser treatment regarding med-
ication use, right splint use, and exercise. Thus, if the intervention was found to
be successful, the differences between treatment and sham could not be attributed
to differences in medications, brace use, or exercise. However, if one addressed the
question retrospectively by reviewing the charts of patients treated for carpal
tunnel syndrome to determine whether those receiving laser were less likely to
undergo surgery, several potentially confounding factors may exist, clouding the
ability to attribute differences solely to laser intervention.
In a similar vein, RCTs assign patients to groups without bias; all patients
have the same likelihood of assignment to a particular group. When patients are
studied in groups (often referred to as cohorts) the influence of factors such as en-
vironment is not distributed equally across all groups. For example, the results of
laser treatment might be studied by comparing outcomes of care at a specialized
hand clinic that uses laser in a treatment protocol to the outcomes of patients
treated in a busy physical therapy clinic in a large hospital where laser is not used.
Patients treated in the hand clinic may be biased when assessing their response to
treatment by the individualized attention and clinic environment thus leading to
misleading conclusions regarding the efficacy of laser. Thus, while cohort, case se-
ries, and case studies can provide important information, randomized-controlled
clinical trials represent the strongest single study research design when assessing
responses to treatments.
All-or-None Studies
The Oxford Center for Evidence-Based Medicine hierarchy identifies all or none
studies (level of evidence 1c), which is missing from the listing in Box 7-1. Such
studies are often linked to mortality where prior to an intervention all patients
died while now some survive or where prior to intervention some patients died
and now all survive. Certainly, death or survival is not the only measure that lends
to the completion of all-or-none reports but absolute outcomes are uncommon.
Box 7-1 categorizes all studies that are not RCTs collectively while The Oxford
Center for Evidence-Based Medicine classifies such studies based on whether the
study was prospective or retrospective and cohort based or case based. Systematic
reviews of nonrandom assignment of treatment studies on the hierarchy generally
provide stronger evidence than individual reports, if for no other reason than the
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 98
fact that more patients are studied. As with systematic reviews of RCTs, a thor-
ough systematic review of nonrandom assignment of treatment studies regardless
of the specific methods of the study is more likely to provide data to influence
clinical decisions than a single report.
Cohort Studies
Cohort studies, which may be prospective or retrospective in design, involve the
study of groups based on exposure or intervention and assessed for differences in
outcomes. The principal difference between an RCT and a prospective cohort is
that in the RCT individuals or, in some case, groups are randomly assigned to in-
terventions, while cohort studies lack random assignment. In some cases the ran-
dom assignment of a group to intervention is possible. Retrospective cohort studies
identify groups based on exposure or intervention at some point in the past and
then follow the groups forward. Thus, random assignment is not possible.
At this juncture one may ask, If the RCT is a superior research design then
why consider prospective cohort studies? To answer this question to a reasonable
degree lets consider circumstances where random assignment of individuals to
exposure or intervention would be difficult or impossible. Consider, for example,
the investigation of a pre-practice exercise regimen purported to reduce lower ex-
tremity injuries in a team sport (Myklebust et al., 2007). Since teams generally
warm-up together assigning teams to receive or not receive the intervention poses
significant challenges. Moreover, the culture of sport may preclude participation
of organizations and teams if random assignment to control is perceived nega-
tively. Thus, cohorts might be identified for study rather than seeking to ran-
domly assign individual or groups to treatment.
If in this example a decrease in injuries was observed at the conclusion of the
study, one could conclude that the intervention was effective. The challenge for
the investigator and research consumer is confidence that the conclusion drawn
reflects the true effect. In this scenario differences in coaching, practice and game
facilities, equipment, climate, or a number of other factors including simply being
studied (Hawthorne effect) could be responsible for the observed effect. Certainly,
one would have greater confidence in the conclusions if efforts were taken before
the study to maximize the similarities across the cohorts on all other factors except
for the intervention of interest. When done well prospective cohort designs can
maximize efficiency, reduce research costs, and yield important information. The
appraisal of such studies, however, can be more difficult for the research con-
sumer since variables with effects that are likely to be normally distributed in an
RCT can influence the behavior and response to intervention of a group.
The results of retrospective cohort studies can also be affected by numerous
factors outside the investigators control. Such studies, however, may offer efficien-
cies in time and costs that lead to more rapid advances in science. Moreover, in
some circumstances it is not permissible to randomly assign research participants
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 99
CONCEPT CHECK
The principal difference between an RCT and a prospective cohort is that in the RCT
individuals are randomly assigned to interventions while in a cohort study groups are
investigated. Random assignment in retrospective cohort studies is not possible be-
cause researchers identify groups based on exposure or intervention at some point in
the past and then follow the groups forward.
Outcomes Studies
Population-based outcomes research (level 2c in Figure 7-1) is similar to cohort
studies but has been defined as research that:
seeks to understand the end results of particular health care practices and inter-
ventions. End results include effects that people experience and care about, such
as change in the ability to function. In particular, for individuals with chronic con-
ditionswhere cure is not always possibleend results include quality of life as
well as mortality. By linking the care people get to the outcomes they experience,
outcomes research has become the key to developing better ways to monitor
and improve the quality of care. (Agency for Healthcare Research and Quality,
U.S. Department of Health and Human Services, http://www.ahrq.gov/clinic/
outfact.htm, accessed October 21, 2009)
Case-Control Studies
Case-control studies are similar to retrospective cohort studies or imbedded
within a prospective cohort study. In case-control studies, however, comparisons
are made between groups of subjects based on an outcome rather than an expo-
sure or intervention. Take, for example, the finding that drinking moderate amounts
of alcohol, particularly red wine, reduces the incidence of heart attack (Yusuf et al.,
2004; Carevic et al., 2007). This finding resulted from retrospectively studying
groups of people who did or did not suffer a cardiac event and attempting to
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 100
100 PART II Seeking Answers: How The Question Drives The Methods
identify factors that increased or decreased risk. Several factors including smok-
ing, obesity, diet, and exercise habits along with alcohol consumption alter risk.
The data, however, suggest that alcohol consumption has a cardioprotective effect.
Our purpose here is not to debate the benefits and potential harms of alcohol con-
sumption but to consider the means by which this finding was originated.
Consider addressing the question of whether red wine reduces the incidence of
heart attack in an RCT. First, subjects would have to agree to drink or not drink as
part of the study. Certainly, many individuals make this personal choice and
would be adverse to changing their behavior for a long time in order to partici-
pate in the research. Next consider what a long time meansthe following of sub-
jects until they experienced a cardiac event or died of other causes; truly an
impossible study. Thus, while shorter-term RCTs related to the mechanisms be-
hind the observed benefits have been conducted (Jensen et al., 2006; Tsang et al.,
2005) much has been learned about cardiovascular risk factors from case-control
studies. Cohort and case-control studies have an important place in research. The
research consumer, however, must again be aware of factors out of the investiga-
tors control that might bias the results.
series provide clinicians with important evidence of safety and an alternative ap-
proach to treating patients with loss of motion at the ankle.
102 PART II Seeking Answers: How The Question Drives The Methods
CONCEPT CHECK
The acronym SpPIN refers to tests with near perfect specificity (high specificity rules in)
while SnNOUT refers to tests with near perfect sensitivity (high sensitivity rules out).
Studies revealing absolute SpPIN or SnNOUT constitute level 1c evidence.
Level 1
1-a: Systematic review of cohort studies from inception of care
and clinical decision rules validated in different populations
1-b: Individual cohort studies from inception of care and clinical
decision rules validated in a single population
1-c: All or none case series
Level 2
2-a: Systematic reviews of retrospective cohort studies or
Level 3
No criteria
Level 4
Case series studies, cohort studies with poorer
methodological quality
Level 5
Expert opinion-based research or physiological principles
FIGURE 7-3 Hierarchy of evidence for prognosis. (Adapted with permission from
http://www.cebm.net/index.aspx?o1025. Oxford Center for Evidence-Based Medicine
Levels of Evidence [March 2009]. Accessed May 10, 2010.)
104 PART II Seeking Answers: How The Question Drives The Methods
Consider these questions as you read the remaining chapters in Part II of this
book, develop skills in the retrieval and appraisal of the clinical literature, and
apply your reading to the care of individual patients. Certainly, caution is war-
ranted if the validity of the data reported is of concern. Moreover, one must con-
sider the whole of the patient they are treating when applying the results of
students. Individuals may differ in age, gender, race and culture, the presence of
co-morbidity, and a host of other factors from the subjects in a research report.
Rarely are there clear recommendations to be made as to whether the results of a
clinical trial apply to the individual patient. Recall that the practice of evidence-
based medicine requires the integration of the best available evidence with clini-
cal experience and patient values. Consideration of the first two questions then
leads to the most important: Will (and perhaps how) the results help me? When
the evidence is very strong and the patient presentation is clearly consistent with
those studied in clinical trials the evidence will most certainly be of help in devel-
oping and explaining a plan of care. The real dilemma comes when the evidence
is less strong or conflicting and when the patient differs from those described in
the research. The patient is seeking care and decisions must be made. We may
elect to cast aside evidence that poses more questions than answers, yet we must
continue to pursue the research that will ultimately best inform our decisions.
CHAPTER SUMMARY
This chapter has provided an overview of how the sources of evidence used to
make clinical decisions are generated and how the importance of the evidence
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 105
based on the methods used to gather data is viewed. Two commonly cited hierar-
chies (with that of the Oxford Center for Evidence-Based Medicine being divided
by category in Figures 7-2 to 7-3) are presented, primarily to highlight the similar-
ities in presentation, but also to address subtle differences. The practice of evi-
dence-based health care requires the consideration of the best available evidence
along with clinical experience and patient values in recommending a course of
care. Within patient, randomized treatment order trials and level 1 evidence
should be weighted more heavily in clinical decisions than sources lower on the
hierarchy. The clinician, however, must appreciate that this level of evidence is not
always available and carefully consider how strongly the available research will
influence their decisions. Not all evidence is created equal, nor can be applied
equally for all patients. All evidence, however, is worthy of consideration for the
patients seeking our care.
KEY POINTS
Confidence gained from experience and continued study of the literature
helps the clinician cope with the stresses of providing optimal care and eases
the anxiety of the patients we treat.
Within patient, randomized treatment order trials and level 1 evidence
should be weighted more heavily in clinical decisions than sources lower on
the hierarchy.
Within patient, randomized treatment order trials are studies in which pa-
tients receive all interventions under consideration in random order over the
duration of the investigation.
Hierarchies of evidence relate to the strength of evidence in making deci-
sions about patient care, not the sophistication of the research methods and
data.
Differences in the strength of evidence come from two sources, sampling and
research methods.
Case series and case studies provide detailed descriptions of a series or sin-
gle case.
Poor methods increase the likelihood of biased data threatening internal as
well as external validity.
Clinicians must carefully consider how strongly the available research will
influence their decisions.
All evidence is worthy of consideration for the patients seeking our care.
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 106
106 PART II Seeking Answers: How The Question Drives The Methods
Applying Concepts
1. Form groups and discuss how the differences in research methods and
data analysis across the spectrum of patient care also result in some dif-
ferences as to how the hierarchy of evidence is described. Provide clin-
ical examples along with examples from the literature to support your
viewpoint.
2. Form groups and discuss the question of, If the RCT is a superior re-
search design then why consider cohort studies? Remember to consider
circumstances where random assignment of individuals to exposure or
intervention would be difficult or impossible, and provide clinical exam-
ples along with examples from the literature to support your viewpoint.
REFERENCES
Carevic V, Rumboldt M, Rumboldt Z. Coronary heart disease risk factors in Croatia and
worldwide: results of the Interheart study (English abstract). Acta Med Croatica.
2007;61(3):299306.
DeBerard MS, Lacaille RA, Spielmans G, et al. Outcomes and presurgery correlates of lum-
bar discectomy in Utah Workers Compensation patients. Spine J. 2009;9:193-203.
Fiala KA, Hoffmann SJ, Ritenour DM. Traumatic hemarthrosis of the knee secondary to he-
mophilia A in a collegiate soccer player: a case report. J Athl Train. 2002;37(3):315319.
Guyatt GH, Haynes B, Jaeschke RZ, et al; for the Evidence Based Medicine Working Group.
EBM: Principles of Applying User's Guide to Patient Care. Center for Health Evidence,
2001. (Available at: http://www.cche.net/text/usersguides/applying.asp. Accessed
May 10, 2010.)
LibraryPirate
97689_c07 01/07/10 3:57 PM Page 107
CHAPTER 8
QUALITATIVE INQUIRY
There are in fact two things, science and opinion; the former
begets knowledge, the latter ignorance.
Hippocrates, Law Greek physician (460 BC377 BC) from
Classic Quotes as quoted in The Quotations Page
(www.quotationspage.com)
CHAPTER OBJECTIVES
After reading this chapter, you will:
Be able to describe the nature of qualitative inquiry.
Be able to contrast the difference between inductive and deductive vantage points of where
to begin research projects.
Understand the various perspectives and theoretical traditions that use qualitative research
methods.
Understand that the interpretation of the world should happen in natural settings when prac-
ticable and practical.
Understand the various types of qualitative data collection.
Develop respect for the rigors of qualitative inquiry.
Appreciate the time-intensive makeup of qualitative inquiry.
Explain the role of observation in qualitative inquiry.
Understand the role of theory development in qualitative inquiry.
Understand the various rigors of qualitative research.
Understand that qualitative research requires multiple views, research methods, and realities
to bring a better understanding to the social world.
108
LibraryPirate
97689_c08 01/07/10 4:47 PM Page 109
KEY TERMS
credibility ethnography interpretivism
dependability grounded theory observation
deductive hypotheses paradigm
epistemology inductive phenomenology
INTRODUCTION
This chapter introduces you to the notion of a qualitative research paradigm.
Qualitative inquiry allows for the exploration of naturally occurring social phenom-
ena in actual contexts (Wilson, 2007). Within scientific endeavors, the qualitative
research paradigm has long served the needs of social scientists who were interested
in the rigors of systematic and disciplined data collection and analysis but who did
not want to sacrifice the complexity of the data under investigation for the purposes
of scientism (Mills, 1998; Ross, 1992). We will introduce you to a number of perspec-
tives and theoretical traditions, and methods for conducting qualitative research
and qualitative data collection. Considerations of reliability and validity will also
be discussed.
110 PART II Seeking Answers: How the Question Drives the Methods
The purpose of qualitative inquiry can be to explain, predict, describe, or explore the
why or the nature of the connections among unstructured information. The focus
of qualitative inquiry is the inductive process rather than the deductive outcome.
Nominal data from qualitative inquiry provides unencumbered information; thus,
the researcher must search for narrative, explanatory patterns among and between
variables of interest, as well as the interpretation and descriptions of those patterns.
Rather than starting with hypotheses, theories, or precise notions to test, qualitative
inquiry begins with preliminary observations and culminates with explanatory
hypotheses and grounded theory.
The process of qualitative work is one where the hypothesis(es) form or
develop as the research progresses during the study (Lincoln & Guba, 1985). In
other words, the overriding or guiding theory may be augmented or changed
entirely as the study runs its course and data are collected (Wilson, 2007). One of
the unique features of qualitative research designs is that it is data-driven versus
hypothesis-driven. Subject(s) are often studied in their natural environment and
the data responses come unfiltered from the subject(s) (Patton, 1990). This means
the data are collected in natural settings where it is to be found: workplace, street,
gym, or home. The most familiar characteristic of qualitative inquiry is that it pro-
vides detailed data and rich interpretation of that data, often without the need for
statistical procedures (Wilson, 2007).
The task of the researcher in qualitative inquiry is to observe to the point of sat-
uration. The idea is not about large numbers; it is about observing a phenomenon
until all observations have been repeated and main themes fit into emerging cate-
gories. The well-conducted progression of qualitative inquiry is extremely time-
intensive due to the fact that the researcher is constantly observing, documenting,
and analyzing.
CONCEPT CHECK
The well-conducted progression of qualitative inquiry is extremely time-intensive
because the qualitative researcher is constantly observing, documenting, and analyzing.
112 PART II Seeking Answers: How the Question Drives the Methods
inquiry remains largely disregarded in the fields of medicine and human movement
sciences in favor of the more widely recognized epistemological category of quan-
titative inquiry. Quantitative inquiry, which is largely based on causal associations
and statistical models, is discussed in Chapter 9.
As described by some of the overall themes that follow in this paragraph,
Wilson (2007) explained that the underlying premise of intrepretivism is that to
understand the meaning of an experience for particular participants, one must
understand the phenomena being studied in the context of the setting for the par-
ticipants of interest. During qualitative inquiry, the researcher attempts to find
meanings attached to and/or associated with the institution, phenomena, or group
culture under investigation. Thus, the process of qualitative inquiry is a process by
which inferences and generalizations are extrapolated to explain underlying pat-
terns and intricacies of observable situations or occurrences. Stated differently,
qualitative research is an interpretive progression of investigation and examina-
tion. The qualitative researcher reports extensive detail to interpret or understand
phenomenon directly observed or experienced. Results of qualitative inquiry are
expressed with word descriptions rather than numerical statistics to capture and
communicate the essence of that which is being studied.
Qualitative fieldwork often does not follow the same procedures of investiga-
tion that would be used by a chemist or an exercise physiologist; however, obser-
vations and analyses are purposeful and take place within philosophic paradigms
that provide theoretical frameworks to guide the inquiry process for each type of
qualitative research. The main types of qualitative inquiry include grounded the-
ory, case study, interviews, observation, participant observation, and artifact
analysis. Qualitative research provides an extensive narrative interpretation of
results unique in context and dependent on the manner in which the data were
collected.
It is important to re-emphasize the point that qualitative inquiry leads to new
theory, whereas quantitative data provides support to answer questions and sup-
ports or refutes a pre-existing theory. As stated earlier, the goal of the qualitative
researcher is to observe phenomenon, occurrences, or experiences to the point of
saturation and repetition. The starting point for qualitative inquiry is inductive in
nature. By comparison, quantitative inquiry is deductive and causal in nature.
Various types of qualitative perspectives and data collection methods are presented
in the next sections.
CONCEPT CHECK
Results of qualitative inquiry are expressed using word descriptions rather than numer-
ical statistics to capture and communicate the essence of that which is being studied.
LibraryPirate
97689_c08 01/07/10 4:47 PM Page 113
Grounded Theory
Grounded theory research is a method of qualitative inquiry that results in the gen-
eration of a theory. The theory can be thought of as a by-product of the data
because it develops and perhaps even changes during the qualitative inquiry
process. Major and minor themes emerge from the data as the researcher uses
informed judgments to analyze it and offer expounding theory to explain or
describe the observations. The resultant theory evolves and is derived from the
data; thus the theory is considered anchored or grounded in the research. Hence
the term grounded theory research. One widely held example is that of
Aristotles (fourth century BC) careful and detailed observations of the elements
and heavenly bodies in motion that led to what we now refer to as gravitational
theory.
Case Study
Case study research involves a thorough study of a single institution, situation
(cultural group), or one individual (Yin, 1994). A case study provides detailed
information about unique characteristics. The researcher collects data about the
individual or situation by conducting firsthand observations, interviews, or other
forms of data collection (qualitative or quantitative). Clinical case reports are
examples of case study research in medical and human movement science disci-
plines. The purpose of a case study is to observe, collect, document, and analyze
detailed information and knowledge about one individual. If this process is con-
ducted over time (i.e., several months or years), then it is termed longitudinal
research.
Interviews
The use of interviews in qualitative research is central to the data collection
process. Interviews are often recognized as the most common instrument for
qualitative data collection. They allow the persons under study to contribute
their knowledge and experience to the research project. Interviewing takes on
many aspects where the researcher may do repeated interviews with the same
person or people for the purpose of validating the data given during the inter-
viewing process and as attempt to gain greater breadth and depth of the data
given by key research informants. The format of the interview content may also
vary. The researchers, in many cases, often start with a list of prepared closed and
open-ended questions to ask informants but they will not limit themselves to the
script if the informant goes into greater detail on subject related to the research
project.
LibraryPirate
97689_c08 01/07/10 4:47 PM Page 114
114 PART II Seeking Answers: How the Question Drives the Methods
Observation
Qualitative researchers often observe the cultural landscape to view how the
action of everyday life unfolds in the organization or group under investigation.
This goes back to the idea within qualitative methods of studying groups in their
natural setting and not manipulating the research environment.
Participant Observation
In some cases the researcher may have to become directly involved in the study
itself by taking part in or at least being present for ordinary or important practices
and rituals performed by the group under investigation. It is often acceptable
for researchers to immerse themselves in group practices to better understand
what the group does that makes the group unique, interesting, or insightful as com-
pared with existing data or other groups. The obvious concern with immersion
is that the researcher might risk becoming overly involved with the subjects(s)
(i.e., group under investigation) and lose objectivity; however, this technique
often gives a rare vantage point that is useful for understanding the group and
its world.
Artifact Analysis
This is the evaluation of written or electronic primary and secondary sources
for the topic being researched. Often the form of investigation is archival
retrieval of data sources that aid in giving interpretation and meaning to the
subject being studied because the gathering and evaluation of written docu-
ments may be relevant in providing background/historical information on a
particular person or group, change of policy when examining previous inter-
ventions for a group, or strengthening and validating other types of collected
qualitative data within the study.
Phenomenology
Phenomenological research focuses on the study of a question (Patton, 1990).
Phenomenology refers to how people describe things and how they experience
them via their senses and the role of the researcher(s) is to then study and describe
these experienced situations. Phenomenology is used as an approach to enhance
meaning and details about the subject of inquiry. In essence, phenomenology is at
the heart of qualitative inquiry because the qualitative researcher seeks to find
meaning in the data and develop explanations, trends, and patterns of behavior
from the particular data set so as to understand the phenomena observed. This
means the researcher often has to take part as a participant in the study, and in some
way experience what others in the group are describing as their reality or life expe-
rience. Within this perspective the researcher develops an understanding for the
group experiences by immersing themselves in the project or situation.
LibraryPirate
97689_c08 01/07/10 4:47 PM Page 116
116 PART II Seeking Answers: How the Question Drives the Methods
EXAMPLE
Ethnography
Ethnographic research focuses on the study of the group and their culture.
Ethnography is closely associated with the fields of anthropology and sociology.
Ethnographers utilize firsthand observation and participant field research meth-
ods as a means of investigation. Ethnographic research is often associated with
studying cultural or cultural groups, but it is not limited to this inquiry and
includes topics of investigation such as office settings/corporate culture, youth
sport groups, cults, community-based groups, feminism/sexism in communities
or institutions, policy application and delivery, and racism (Agar, 1986a,b).
Ethnographers start with a premise about group culture. The tradition here is to
immerse oneself in the culture for usually a year, conduct interviews, be a partic-
ipant observer, and conduct other fieldwork connected to the group. Examples of
fieldwork would include tasks such as reviewing other primary and secondary
sources of written and archival data. The starting point for studying any cultural
group is when people form a group they will generate their own values, traditions,
policies/laws, and physical artifacts. For example, anthropologist Alan Klein
studied the subculture of male body builders in Southern California in the mid-
1980s. From his research, Klein (1993) presents a world where in public displays
LibraryPirate
97689_c08 01/07/10 4:47 PM Page 117
(competitions) these men are presented as the idealized male physique but the
reality is that the athletes involved in this sport suffer from narcissism, question-
able or confused sexual identities, as well as needs and rituals for human perform-
ance-enhancing aids to achieve the highest levels of athletic and commercial
achievement within the sport (Klein, 1993).
Symbolic Interactionism
This perspective investigates symbols in relation to each other, and how people
give meaning to their interactions with others. This perspective comes from the
tradition of the social-psychologist. The push of this perspective is to derive mean-
ing and interpretation from what others consider meaningful and is a shared
experience. These shared meanings and experiences become reality, and develop
the shared reality of the group. Erving Goffmans works on stigmas (Goffman,
1986) and dramography (Goffman, 1999) are key to understanding this perspec-
tive. One of Goffmans (1999) insights was that we live in a front stage (how we
present ourselves and what we do)back stage (who we are) world that we create.
Front stage is the world in which we create and perform roles that the general
public takes part in and views, while the back stage is the unseen or protected
realm whereby people present their real selves to the group(s) in which they work
and share their lives. In other words, the back-stage world real self is unknown
to the public masses, while the front-stage world Social Face is shown and
known in the general public. An example of this type of constructed social world
study in the health professions can be found in Howard Beckers study (Becker et
al., 1976), Boys In White, which examines how medical students learn the role
of how to become a medical doctor and the symbolic importance attached to
wearing a white lab coat as a doctor or health professional when interacting with
clients or other nonhealth professionals.
118 PART II Seeking Answers: How the Question Drives the Methods
with one another; how professionals define their roles; how satisfied clients feel
about their treatments at this particular facility. For researchers to observe and eval-
uate an EBA program they would need to include experts from a number of health
fields to assist in the data collection and interpretation of the data. Other considera-
tions within a systems approach would be to work on site for the data collection,
describe how doctors, athletic trainers, physician assistants, physical therapists, reg-
istered nurses, and orderlies work together to meet the needs of the clients. Under
this perspective the evaluative process is all-inclusive in order to reveal revenue
streams, credentials, education levels, and roles and responsibilities on site.
Questions that might be asked under a systems approach entering health provider
facility that is attempting to follow an EBA system include: how is treatment and
care administered by those within the system, types of care offered (consideration
given to context-local, budgets, etc.); changes in the facility if the EBA system has
been recently implemented; how responsive and adaptive is the staff to problems
and situations that occur on site.
Orientational Qualitative
The orientational perspective examines how a particular ideology or phenomena
manifests itself in a social situation or group (Patton, 1990). What differs from this
perspective as compared to most other qualitative traditions is that researchers go
into the project with an explicit theoretical perspective that determines what vari-
ables and concepts are most important and how the findings will be interpreted.
A predetermined and defined framework determines this type of qualitative
inquiry. For instance, in order to investigate if a medical caregiver is in line with
the EBA framework, it needs to be investigated by the framework or mandates of
what an ideal EBA system looks like and functions as a medical caregiver.
Historical
Historical research shares much in common with qualitative research designs.
The development of social histories is often a feature of many of the qualitative
perspectives used where the form of investigation is non-experimental and often
analytical. The data resources used for investigation by historians include but are
not limited to the following: archival records, interviews (past recordings of
interviews and/or those recently conducted), oral histories, books, journals,
newsprint, and film (Yow, 1994). Historical research deals with events that have
already occurred. In historical research, the researcher provides a detailed narra-
tion in an attempt to relate meaning to past events. A study that uses data col-
lected from deceased patients hospital records to better understand how to treat
a medical problem that has been dormant for a period is an example of historical
research in medical science and human movement studies.
LibraryPirate
97689_c08 01/07/10 4:47 PM Page 119
CONCEPT CHECK
The basic idea of triangulation is that if similar findings are produced as a result of mul-
tiple methods, or multiple sources, or multiple researchers then such findings would
be indisputably more credible than findings obtained from one method, one source,
or one researcher.
120 PART II Seeking Answers: How the Question Drives the Methods
taking more than enough diligent and detailed notes. Detailed notes must then be
logged, chronicled, and transcribed. Detailed notes are essential in the systematic
approach for observing a phenomenon until nothing new appears. The researcher
observes and records the observations via notes until all observations have been
repeated, and main themes fit into emerging categories that represent the social
world that was researched or investigated.
experiences with different clients with the same problems, and the numbers of
years the two have worked together in the same facility. The community con-
structed in a health care facility is not built around segregated roles of each pro-
fession but the integration of them. Additionally, each health care facility is
interconnected by networks that each individual has with other people in their
own field and those they know outside of their domain of health care. For exam-
ple, the physician's assistant who has a network of contacts at a number of other
health care facilities where three people in their network are long-time friends
who are now medical doctors and certified athletic trainers working with sport
medicine specialist. These network connections help build and strengthen the
existing community within a particular facility.
CHAPTER SUMMARY
This chapter provided an overview of qualitative inquiry. Common types of quali-
tative research as well as common types of qualitative data collection were pre-
sented. It is our intention that you will now have a clear understanding of what
qualitative research is and what is involved to turn out credible and dependable
results through a careful and detailed qualitative inquiry.
LibraryPirate
97689_c08 01/07/10 4:47 PM Page 122
122 PART II Seeking Answers: How the Question Drives the Methods
KEY POINTS
Qualitative research is an interpretive progression of investigation and exam-
ination.
The task of the researcher in qualitative inquiry is observation to the point of
saturation.
Qualitative inquiry is largely established through the collection of observable
data.
Qualitative data yields a research question or a new theory, whereas quantita-
tive data answers a question and supports or refutes a pre-existing theory.
Qualitative inquiry is also referred to as interpretivism.
Validity in qualitative research is referred to as credibility.
In qualitative research, the notion of reliability is referred to as dependability.
Applying Concepts
1. Select a qualitative research study of interest to you. In your own
words, prepare a 1-page summary of the study including relevant
details about the research question and the qualitative methods. Be
sure to include critical analysis of strengths and areas for improvement
in the study.
2. Consider an answerable research question of interest to you. Suggest a
do-able qualitative research design to explore answers to this question.
Be prepared to discuss plausible strengths, areas for improvement, and
clinical applications of the design.
ACKNOWLEDGMENT
We wish to thank Timothy J. Bryant and Brent Thomas Wilson for their input and
contributions to this chapter.
LibraryPirate
97689_c08 01/07/10 4:47 PM Page 123
REFERENCES
Agar M. Speaking of Ethnography. Newbury Park, CA: Sage Publications; 1986a.
Agar MH. Speaking of Ethnography. Vol. 2, Qualitative Research Methods Series. Newbury Park,
CA: Sage Publications; 1986b.
Becker HS, Geer B, Hughes EC, et al. Boys in White: Student Culture in Medical School. Edison,
NJ: Transaction Publishers; 1976.
Berger PL, Luckmann T. The Social Construction of Reality. A Treatise in the Sociology of
Knowledge. New York: Anchor Books; 1967.
Blumer H. Symbolic Interaction. Englewood Cliffs, NJ: Prentice-Hall; 1969.
Creswell JH. Research Design: Qualitative, Quantitative, and Mixed Methods Approaches. 2nd ed.
Thousand Oaks, CA: Sage Publications, Inc.; 2003.
Creswell JW. Qualitative Inquiry and Research Design. Choosing Among Five Traditions. Thousand
Oaks, CA: Sage Publications; 1998.
Damico JS, Oelschlager M, Simmons-Mackie NN. Qualitative methods in aphasia research:
conversation analysis. Aphasiology. 1999a;13:667680.
Damico JS, Simmons-Mackie NN, Oelschlager M, et al. Qualitative methods in aphasia
research: basic issues. Aphasiology. 1999b;13:651665.
Fosnot CT, ed. Constructivism. Theory, Perspectives, and Practice. 2nd ed. New York: Teachers
College Press; 2005.
Garfinkel H. Studies in Ethnomethodology. New York: Prentice-Hall; 1967.
Geertz C. Local Knowledge. New York: Basic Books; 1983.
Goffman E. Strategic Interaction. Philadelphia, PA: University of Pennsylvania Press; 1969.
Goffman E. Stigma: Notes on the Management of Spoiled Identity. New York: Simon & Schuster;
1986.
Goffman E. Presentation of Self in Everyday Life. Gloucester, MA: Smith Peter Publisher; 1999.
Holtgraves TM. Language as Social Action. Social Psychology and Language Use. Mahwah, NJ:
Lawrence Erlbaum Associates, Inc.; 2002.
Klein AM. Little Big Men: Bodybuilding Subculture and Gender Construction. Albany, NY: State
University of New York Press; 1993.
Lincoln YS, Guba EG. Naturalistic Inquiry. Newbury Park, CA: Sage Publications; 1985.
Mills JA. Control: A History of Behavioral Psychology. New York: New York University Press; 1998.
Patton MQ. Qualitative Evaluation and Research Methods. Newbury Park, CA: Sage
Publications; 1990.
Plimpton G. Out of my League: The Classic Hilarious Account of an Amateurs Ordeal in
Professional Baseball. Guilford, CT: Lyons Press; 2003.
Plimpton G. Paper Lion: Confessions of a Last String Quarterback. Guilford, CT: Lyons Press;
2006.
Rempusheski VF. Qualitative research and Alzheimers disease. Alzheimer Dis Assoc Disord.
1999;13(1):S45S49.
Ross D. Origins of American Social Science. Cambridge, UK: Cambridge University Press; 1992.
Schultz SJ. Family Systems Therapy: An Integration. Northvale, NJ: Aronson; 1984.
Wilson BT. A functional exploration of discourse markers by an individual with dementia
of the Alzheimer's type: A conversation analytic perspective. Unpublished doctoral dis-
sertation, The University of Louisiana at Lafayette; 2007.
Yin RK. Case Study Research: Design and Methods. 2nd ed. Thousand Oaks, CA: Sage
Publications; 1994.
Yow VR. Recording Oral History: A Practical Guide for Social Scientists. Thousand Oaks, CA:
Sage Publications; 1994.
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 124
CHAPTER 9
QUANTITATIVE INQUIRY
You can use all the quantitative data you can get, but you still
have to distrust it and use your own intelligence and judgment.
Alvin Toffler from Coles Quotables as quoted in The
Quotations Page (www.quotationspage.com)
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand the characteristics of quantitative inquiry.
Appreciate the significance behind collecting data in a systematic manner.
Comprehend how the concepts of central tendency are used in statistical analysis to test
hypotheses.
Be able to describe the basic steps of the scientific method.
Be able to explain the difference between the types of measurement data.
Understand the different research experimental designs.
Understand the difference between Type I and II errors and the relevance to data
analysis.
KEY TERMS
a priori power analysis external validity scientific method
confounding variables independent variable
dependent variable research hypothesis
124
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 125
INTRODUCTION
Quantitative inquiry forms the basis for the scientific method. Through the per-
formance of controlled investigation, researchers may objectively assess clinical
and natural phenomena and develop new knowledge. By collecting data in a sys-
tematic manner and using statistical analysis to test hypotheses, the understand-
ing of human physiology and behavior may be advanced. Quantitative inquiry is
central to advancing the health sciences. The aim of this chapter is to identify the
characteristics of quantitative research. This chapter is presented based on classi-
cal test theory (pretestposttest) rather than from the perspective of clinical epi-
demiology (case control, cohort, randomized-controlled clinical trial). The
principles of clinical epidemiology are essential for applying research in clinical
practice and are addressed in subsequent chapters specifically related to preven-
tion, diagnosis, and treatment outcomes. The material presented here is, however,
an essential foundation for laboratory and clinical research.
5. Answer the research question on whether the experiment confirms or refutes the hypothesis.
126 PART II Seeking Answers: How the Question Drives the Methods
CONCEPT CHECK
The formulation of a good research question cannot be based simply on the clinical
observation alone, but instead must be developed by framing the observation within
the existing knowledge base of a specific discipline.
EXAMPLE
The mean (x) represents the arithmetic average of scores across all sampled
subjects and is computed by summing the scores of all subjects and dividing by
the total number of subjects. This is expressed mathematically as: x xi> n
where xi represents the score of each individual subject and n is the total number
of subjects.
To calculate the mean,
xi
x
n
(118 127 123 104 116 116)
x
6
704
x
6
x 117.33
Sometimes one or even a few individual scores are much higher or lower
than the other scores. In this case, the outlying scores can skew the data in such
(continued)
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 128
128 PART II Seeking Answers: How the Question Drives the Methods
a way that the mean is not a good representative measure of the entire data set.
The median (Md) represents the individual score that separates the higher half of
scores from the lower half of scores. It is the score that is in the middle of score of
all observations. To determine the median, all scores are first ranked in ascend-
ing order from the lowest to the highest. The median score is determined with the
formula: Md (n 1)/2. If there are an odd number of samples, the score that
falls in the rank-ordered slot equal to Md represents the median. If a data set has
an even number of samples, the average of the scores from the rank-ordered slots
immediately above and below Md is computed that represents the median score.
To calculate the median, the scores are first rank ordered from the lowest to
the highest:
104, 116, 116, 118, 123, 127
Because there are six samples in the data set, the average of rank-ordered
scores #3 (116) and #4 (118) is calculated, so the median value of this data set is 117.
Half of the scores fall above and half fall below this value.
The mode (Mo) is the score that occurs most frequently out of all included
observations. To calculate the mode, we observe that two subjects have a score of
116, while all other subjects have unique scores. Thus, the mode is 116.
We can thus see that in this data set, the mean (117), median (117), and mode
(116) are all very similar in their estimates of the central tendency of knee flexion
range of motion scores. Now, let us look at a data set that produces markedly dif-
ferent estimates of central tendency. This data set represents the total billable
health care costs of five patients who had an acute onset of low back pain:
18161
x
5
x 3632.20
We can see that a mean of $3632.20 is not a good estimate of central tendency
for this data set because four out of the five samples are less than half of the
mean.
To calculate the median, the scores are first rank ordered from the lowest to
the highest:
Estimates of Dispersion
Besides the estimates of the magnitude of subject scores, another important char-
acteristic of a data set is the dispersion, or variability, of observed values. The
range of scores represents the arithmetic difference between the highest and the
lowest scores in a data set. A limitation of the range is that outliers, individual
scores at the low and/or high extremes of the data set, can skew this estimate of
dispersion.
A more robust estimate of dispersion is the standard deviation. The standard devi-
ation (s) estimates how much the scores of individual subjects tend to deviate from the
mean. The formula to calculate standard deviation is s 2((xi x) 2 )> (n 1).
The numerator of this equation is termed the sum of squares, and is used in a wide
variety of statistical analyses.
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 130
130 PART II Seeking Answers: How the Question Drives the Methods
EXAMPLE
Standard Deviation
Using the knee range of motion data from earlier in this chapter the standard devi-
ation is calculated as follows:
Patient #1 118 Patient #2 127 Patient #3 123
Patient #4 104 Patient #5 116 Patient #6 116
x 117.33
(xi x) 2
s
B n1
(118 117.33) 2 (127 117.33) 2 (123 117.33) 2 (104 117.33) 2 (116 117.33) 2 (116 117.33) 2
s
B 61
means with these statistical tests is based on the probability of overlap in the nor-
mal distribution of the two data sets being compared.
Hypothesis Testing
Quantitative inquiry usually involves statistical analysis of the collected data. The
statistical analysis is typically performed to test the hypothesis, or hypotheses, that
the investigator has established. When establishing a research hypothesis (also called
(continued)
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 131
+
1 SD
68%
+
2 SD
95%
+
3 SD 99%
the alternative hypothesis), the investigator generally has an idea that the independent
variable (the intervention being manipulated) is going to cause a change in the
dependent variable (the characteristic being measured). For example, an investigator
may form a research hypothesis that the application of a newly designed wrist brace
will limit wrist flexion range of motion in patients with carpal tunnel syndrome. In
contrast to the research hypothesis, the null hypothesis is that the independent vari-
able will not cause a change in the dependent variable. Using the previous example,
the null hypothesis would be that the newly designed wrist brace would not cause a
change in wrist flexion range of motion in carpal tunnel syndrome patients.
Despite the fact that the use of the scientific method normally results in
investigators generating a directional research hypothesis, inferential statistical
analysis, the most common type of analysis, actually is a test of the null hypoth-
esis. In other words, most statistical analyses are performed to determine if there
is not a difference between two measures. Hypothesis testing will yield a yes or
no answer as to whether or not there is a statistically significant difference
between measures in the study sample. This result, however, may not be repre-
sentative of the population at large. When this occurs, either a Type I or Type II
error has occurred (see Table 9-2).
Sampling
When a clinician makes an observation on an individual patient they know that
they cannot immediately generalize that observation to all of their patients.
Likewise, when a scientist performs an experiment, he or she must perform that
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 132
132 PART II Seeking Answers: How the Question Drives the Methods
If the results of the tested sample do not match with what is true in the population at large, either a Type I or a
Type II error has occurred.
CONCEPT CHECK
The most common reason for a Type II error is an inadequate sample size.
CONCEPT CHECK
Type I error occurs when a difference is found in the study sample but there is, in fact,
no difference present in the population at large.
134 PART II Seeking Answers: How the Question Drives the Methods
sampling frame has an equal chance of being selected for study participation.
Computerized random number generators are typically utilized to select potential
subjects from the sampling frame when random sampling is employed. Systematic
random sampling is a method of sampling in which every xth individual out of the
entire list of potential subjects is selected for participation. For example, out of a
list of 200 potential subjects in a sampling frame, every 10th individual on the list
might be selected to produce a random sample of 20 subjects.
Stratified random sampling provides a method for dividing the individual mem-
bers of the sampling frame into groups, or strata, based on specific subject charac-
teristics. For example, members of the sampling frame may be initially grouped
into strata by sex (males, females), health status (diagnosed with osteoporosis, not
diagnosed with osteoporosis), or some other important characteristic, and then a
random sample of potential participants is selected from each stratum. The use of
stratified random sampling can be a useful means of preventing potential con-
founding variables from contaminating the results of a study.
Cluster random sampling is a process of dividing the sampling frame into clus-
ters (or groups) based on some common characteristics and then randomly select-
ing specific clusters to participate in the study out of all possible clusters. All
individuals in a selected cluster thus have the opportunity to participate in the
study. Performing a study of high-school athletes provides a good example of
cluster random sampling. Imagine that a university health system in a large met-
ropolitan area provides outreach athletic training services to 24 different high
schools. Researchers at the university want to assess the effectiveness of core sta-
bility training on the prevention of lower extremity injuries in high-school basket-
ball players. The researchers could choose to randomly select individual players
from each school as potential subjects but this could create several logistical chal-
lenges. Alternatively, they could randomly select specific schools (each school is a
cluster) to participate. All individual athletes from each selected school are then
given the opportunity to enroll in the study.
Random sampling is considered superior to nonrandom sampling because
the results of the study are more likely to be representative of the population at
large. However, when random sampling is not feasible, nonrandom sampling
may be employed. With convenience sampling, potential subjects are selected based
on the ease of subject recruitment. Studies which seek volunteer participants from
the general population is a specific geographic area utilize convenience sampling.
Another type of nonrandom sampling is purposive sampling. Purposive sampling
entails potential subjects from a predetermined group to be sought out and sam-
pled. For example, researchers performing a study of patients with a specific
pathology, such as multiple sclerosis, will specifically seek out individuals with
this pathology and recruit them to participate in their study. Because of logistical
reasons, these two nonrandom sampling techniques are frequently utilized in
health care research.
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 135
Pre-Experimental Designs
The one shot posttest design consists of a single measurement completed after a
group of subjects has already received the treatment of interest. In clinical
research, this is also referred to as a case series. This type of study design lacks
both a baseline measurement (pretest) and a control group for comparison. It is a
very primitive study design that has considerable threats to internal validity.
Group 1: X O X intervention, O observation
The one group pretestposttest design adds a pretest to the previous design. The
inclusion of baseline data and repeated measures is helpful, but the continued lack
of control group does not allow the investigator to be confident that any change
in measures is strictly due to the intervention. Considerable threats to internal
validity remain.
Group 1: O X O
The static group posttest design has two groups, one that receives the intervention
and one that does not. Both groups are assessed only once. There is no pretest for
either group so the investigator cannot be confident that group differences are due
exclusively to the interventions. Considerable threats to internal validity remain.
Group 1: X O
Group 2: O
Quasi-Experimental Designs
The nonrandomized pretestposttest design compares two groups before and after
intervention. The two groups receive different interventions; however, the assign-
ment of subjects to groups is based on convenience rather than randomization.
The lack of randomization limits the internal validity of the study as there may be
considerable bias of subject selection into specific groups.
Group 1: O X O
Group 2: O X O
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 136
136 PART II Seeking Answers: How the Question Drives the Methods
The time series design compares a single group at multiple, but regular, time inter-
vals both before and after intervention. This design allows for an understanding of
how the intervention affects the course of progression of the dependent variable over
time; however, the lack of control group is a threat to the internal validity of the study.
Group 1: O O X O O
R Group 1: X O R Randomization
R Group 2: X 0
The randomized pretestposttest design is the gold standard for most experi-
ments. Subjects are pretested, then randomized to assigned groups, and
posttested after they receive their assigned intervention. By randomizing subjects
after they are baseline tested, it prevents any potential bias on the part of the
research team member who is conducting the baseline measures.
Group 1: O R X O
Group 2: O R X O
CONCEPT CHECK
In general, there are three categories of study designs: pre-experimental, quasi-
experimental, and true experimental.
CHAPTER SUMMARY
By following the steps of the scientific method, clinician observations can be
translated into hypotheses that can be evaluated via experimental study. The con-
cepts of central tendency, most commonly represented by the mean and standard
deviation, are the foundation for inferential statistics. Hypothesis testing allows
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 137
KEY POINTS
Clinician observations can be translated into hypotheses that can be evalu-
ated via experimental study by following the steps of the scientific method.
The concepts of central tendency are the foundation for inferential statistics.
Hypothesis testing allows for the research hypothesis to be confirmed or
refuted.
Sampling of potential study volunteers is important to the generalizability of
the study results.
Experimental design of a study forms the infrastructure for the project.
Quantitative inquiry is central to advancing the health sciences.
Applying Concepts
1. Using classical test theory:
a. come up with an idea for a mini-study.
b. provide an example of a plausible research design for the mini-study
you planned.
2. Discuss how and why the material presented in this chapter provides an
essential foundation for laboratory and clinical research.
a. Provide current and historical examples from medical research to
support your position.
LibraryPirate
97689_c09 01/07/10 3:59 PM Page 138
138 PART II Seeking Answers: How the Question Drives the Methods
SUGGESTED READING
Greenfield ML, Kuhn JE, Wojtys EM. A statistic primer. Descriptive measures for continu-
ous data. Am J Sports Med. 1997;25(5):720723.
Hopkins WG, Marshall SW, Batterham AM, et al. Progressive statistics for studies in sports
medicine and exercise science. Med Sci Sports Exerc. 2009; 41(1):313.
Tate DG, Findley T Jr, Dijkers M, et al. Randomized clinical trials in medical rehabilitation
research. Am J Phys Med Rehabil. 1999;78(5):486499.
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 139
CHAPTER 10
VALIDITY AND RELIABILITY
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand the difference between internal and external validity.
Be able to discuss participant blinding and explain how it is utilized in experimental
design.
Understand the differences between validity and reliability.
Understand the different research experimental designs.
Understand the concept of objective measurement.
Understand the different types of quantitative data and be able to explain how they are
obtained.
KEY TERMS
blinding limits of agreement (LOA) standard error of measurement
delimitations objective measurement validity
dependent variables precision
internal validity selection bias
139
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 140
140 PART II Seeking Answers: How the Question Drives the Methods
INTRODUCTION
A key to quantitative inquiry is unbiased and objective measurement of the
dependent variables. Measures may be placed on different types of measurement
scales and each type of measure is subjected to distinctive forms of statistical
analysis. Validity is an inherent principle in research design and this important
concept has many components. Reliability and agreement of measures are also
key components of validity.
CONCEPT CHECK
Measurement data can usually be classified into one of three types: categorical, ordi-
nal, or continuous.
respond to how they feel about a certain statement and give five choices: strongly
agree, agree, neutral, disagree, and strongly disagree. For statistical analysis, a
numeric value will be assigned to each possible response, but in this case the order
of the numeric assignment is of consequence. The Likert scale data are likely to be
classified as 1strongly agree, 2agree, 3neutral, 4disagree, and 5strongly
disagree (the order of the numbers could, in theory, be reversed as well). The num-
bers assigned to each response are of consequence because there is an inherent
order to the responses that has a clear meaning. Ordinal data are not strictly lim-
ited to Likert scale measures.
Continuous data are measured on a scale that can continuously be broken down
into smaller and smaller increments. For example, when measuring the length of a
certain object, it could be measured in meters or centimeters or millimeters and so
on with each successive measurement unit becoming more and more precise. This
property is in stark contrast to categorical or ordinal data. If 1male and 2female
for categorical data, it is not possible to have a score of 1.6. Likewise with ordinal
data on a Likert scale, it is not possible to have a score in between any of the adja-
cent responses. With continuous data, however, it is always possible to make a
measure more precise provided the appropriate measurement tool is available.
VALIDITY
Validity is a concept that is often discussed in experimental design. It is critical to
understand that validity has multiple contexts. Validity may be discussed in rela-
tion to the structure of the overall design of an experiment, the intervention
assessed in an experiment, or the measurements performed in an experiment.
Internal Validity
Internal validity refers to the validity of a studys experimental design. Most
experiments are designed to show a causal relationship between an independent
variable and a dependent variable. Investigators manipulate an independent vari-
able (e.g., comparing two treatment regimens) to assess its effects on a dependent
variable (an outcome measure). If an experiment can conclusively demonstrate
that the independent variable has a definite effect on the dependent variable, the
study is internally valid. If, however, other factors may influence the dependent
variable and these factors are not controlled for in the experimental design, the
studys internal validity may be questioned. It is therefore essential for investiga-
tors to identify potential threats to internal validity and appropriately control for
these threats in the design of the experiment.
Internal validity should be thought of along a continuum rather than as a
dichotomous property. In laboratory experiments it is often easier to control for
confounding factors, and thus enhance internal validity, than it is in clinical trials.
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 142
142 PART II Seeking Answers: How the Question Drives the Methods
If all pretest and posttest measures and interventions are performed in a single
laboratory session, it is much easier to control extraneous factors that may influ-
ence the measures. In contrast, if a clinical trial has measurements taken at time
points weeks or months apart, it is much more difficult to control extraneous fac-
tors that may influence the follow-up measures. Such extraneous factors are
referred to as confounding variables and may result in spurious relationships.
Brainstorming for potential confounding variables must be done when the exper-
iment is being designed. If extraneous factors may influence measures of the
dependent variable over the course of a study, these factors must either be con-
trolled or quantified.
EXAMPLE
Internal Validity
If a study was investigating two physical therapy treatment regimens for hip
osteoarthritis and one of the dependent variable was pain, the investigators
should realize factors such as analgesic drug use (over the counter and prescrip-
tion) by the subjects may influence pain measures. The investigators could
attempt to control for this confounding factor by instructing subjects to not take
any analgesic medications. This stipulation may raise concerns during institu-
tional review of the research protocol and limit the number of patients willing to
participate. Alternatively, the investigators could ask subjects to maintain a daily
log of analgesic medication use and then use this information as a covariate in their
statistical analysis. Either approach is better than not controlling for analgesic drug
use at all and thus would improve the internal validity of the study.
Potential threats to internal validity often involve some sort of bias. Bias may
be inherent to either the subjects in the study or the experimenters themselves. In
terms of the experimental subjects, selection bias is an important consideration.
The characteristics that subjects have before they enroll in a study may ultimately
influence the results of the study. These may include things like age, maturation,
sex, medical history, injury or illness severity, and motivation, among many oth-
ers. Investigators must take such factors into account when establishing their
inclusion and exclusion criteria for study participation in an effort to ensure inter-
nal validity. Such decisions that investigators make to improve the internal valid-
ity of their studies are referred to as delimitations.
Likewise, once subjects are enrolled in a study, the formation of study groups
(e.g., an intervention group and a control group), care must be taken to make sure
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 143
that the subjects in different groups have similar characteristics. This is most
often accomplished by either randomly assigning subjects to treatment groups
or employing some type of matching procedure to make sure that groups are
equal in important characteristics at baseline.
The issue of blinding is also important to the internal validity of a subject.
There are three entities that may be blinded in a study. The subjects may be blinded
to whether they are receiving an experimental treatment or a control treatment.
This is often accomplished by having a placebo, or sham, treatment that prevents
subjects from knowing whether or not they are receiving the active treatment.
Blinding of subjects is easier with some interventions such as medication than oth-
ers such as receiving therapeutic rehabilitation or wearing an external brace.
Members of the experimental team who are performing outcome measures should
also be blinded to the group assignments of individual subjects and values of pre-
vious measurements for individual subjects (this same blinding principle holds
true for subjects who are providing self-report information). Lastly, in some
instances it is possible to blind clinicians who are treating patients in clinical trials
to the group assignments of individual subjects. Again this is easier with some
experimental interventions such as medications than with other interventions such
as rehabilitation or external appliances. If one of these entities is blinded, a study is
referred to as being single-blinded; if two entities are blinded, a study is double-
blinded; and if all three are blinded, a study is triple-blinded.
External Validity
While internal validity is related to the design of an experiment, external validity
relates to how generalizable the results of a study are to the real world. There is a
definite trade off between internal validity and external validity. The more tightly
controlled a study is in terms of subject selection, administration of interventions,
and control of confounding factors, the less generalizable the study results are to
the general population. In health care research this is an important issue in terms
of translating treatments from controlled laboratory studies to typical clinical
practice settings. This concept is also referred to as ecological validity. Ultimately,
investigators must make decisions that provide the appropriate balance for influ-
encing both the internal and external validity of a study.
CONCEPT CHECK
While internal validity refers to the validity of a studys experimental design, external
validity relates to how generalizable the results of a study are to the real world.
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 144
144 PART II Seeking Answers: How the Question Drives the Methods
Validity of Measures
There are multiple ways of assessing the validity of data and, thus, the measures
used in research. Each of the validity characteristics should be considered when
assessing the measures utilized in a research study. Face validity refers to the prop-
erty of whether a specific measure actually assesses what it is designed to meas-
ure. This is an important issue in the development of functional tests for
patients in the rehabilitation sciences. For example, it is very important that clini-
cal tests that are part of a functional capacity evaluation to determine if a workers
compensation patient is prepared to return to work actually assess an individuals
ability to return to their specific job. Face validity is determined subjectively and
most often by expert opinion.
Content validity refers to the amount that a particular measure represents all
facets of the construct it is supposed to measure. Content validity is similar to face
validity but is more scientifically rigorous because it requires statistical analysis of
the opinion of multiple content experts rather than only intuitive judgment.
Accuracy is defined as the closeness of a measured value to the true value of
what is being assessed. For example, it is well accepted that a goniometric meas-
ure of knee flexion range of motion based on external bony landmarks provides
an accurate assessment of the actual motion occurring between the femur and the
tibia. Accuracy should not be confused with precision of measurement (see dis-
cussion in section Reliability of this chapter).
Concurrent validity refers to how well one measure is correlated with an exist-
ing gold standard measure. This is an important property to be established for
new measures aiming to assess the same properties as an existing test. New meas-
ures are often developed because existing measures are expensive, or perhaps
cumbersome to perform. Measurement of leg length discrepancy is a good exam-
ple. The gold standard for measuring lower extremity length is via radiographs.
A less expensive procedure would be the use of standard tape measures with well-
defined procedures. Part of the validation of the tape measure technique was to
determine concurrent validity with the existing gold standard of radiographic meas-
ures. A challenge does exist, however, when there is no gold standard. In this case,
the other types of validity must be adequately demonstrated to validate a new test.
Construct validity refers to how well a specific measure or scale captures a
defined entity. Like many measurement properties, this concept stems from psy-
chology but is applicable to other areas of study such as the health sciences. For
example, balance may be considered an important construct of musculoskeletal
function. There are, however, many tests to assess the multidimensional con-
structs of balance. In an elderly population, clinical tests such as the Berg Balance
Scale, the Timed Get Up and Go Test, and the Functional Reach Test may be used
as part of a comprehensive balance screening protocol. Using specific statistical
analysis techniques such as factor analysis, it can be determined whether these
three tests all assess the same construct, or whether in fact they each assess dif-
ferent constructs of balance.
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 145
CONCEPT CHECK
When discussing the validity of a measurement, it is critical to understand in what
context validity is being discussed.
146 PART II Seeking Answers: How the Question Drives the Methods
CONCEPT CHECK
Reliability refers to the consistency of a specific measurement.
Any single observed measure can be thought of as being the true score
error. For a group of measurements, the total variance can be due to true variance
(the variability between subjects) and error variance (the difference between the
true score and the observed score). Sources of error may include error or biolog-
ical variability by the subject, or error by the tester or instrumentation used to
take the measure. Recall that, mathematically, variance (s2) is simply the standard
deviation (s) squared. Essentially, reliability may be thought of as the proportion
of total variability (sT2 ) that stems from between subject variability (st2 ) . The
remainder of the variability is thus attributable to error (se2 ) . This may be
expressed with the following formula:
st2
Reliability
st2 se2
2
The error term can be further divided into systematic error (sse ) and random
2
error (sre ) . Systematic error may include constant error and bias. Constant error
affects all measures in the same manner, whereas bias affects certain types of
scores in specific ways (e.g., scores of higher magnitudes may consistently be
overestimated while scores of lower magnitudes are underestimated). The relia-
bility formula can thus be expressed more robustly as:
st2
Reliability
st2 ss2e sre
2
The calculation of the ICC stems from the parsing of the variability contribu-
tions from an analysis of variance (ANOVA). ANOVAs are discussed in detail in
Chapter 11. There are several types of ANOVAs but ICC estimates are calculated
specifically from the single within factor ANOVA model (also called a repeated
measures ANOVA). This analysis is performed to determine if two (or more) sets
of measurements are significantly different from each other.
Tables 10-1 through 10-5 illustrate how the results of a repeated measures
ANOVA can be used to calculate ICCs from a testretest study (Weir, 2005). Shrout
and Fleiss (1979) presented formulas for six different ICCs. The formula nomen-
clature includes two terms. The first term may be expressed as either 1, 2, or 3 and
the second term as 1 or k. For the first term, in model 1 each subject is assessed by
a different set of raters than the other subjects and these raters are randomly
selected from all possible raters. In model 2, each subject is assessed by the same
group of raters who are randomly selected from all possible raters. In model 3,
each subject is assessed by the same group of raters but these raters are the only
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 147
90 99 +9 150 159 +9
Reprinted with permission from Weir JP. Quantifying test-retest reliability using the intraclass correlation coefficient
and the SEM. J Strength Cond Res. 2005;19(1):231240.
raters of interest (and results thus do not generalize to any other raters). For the
second term, 1 indicates that a single observation is being assessed for each trial
performed by an individual subject, while k indicates that a mean of multiple tri-
als is being assessed for each subject. The six different ICC formulas are shown in
Table 10-5. Essentially, each formula estimates the sources of error in the data set.
TABLE 10-2 Two-way analysis of variance summary table for data set Aa
SOURCE DF SS MEAN SQUARE F P VALUE
Total 15 15,119.8
a
MSB, between-subjects mean square; MSE, error mean square; MSS, subjects mean square; MST, trials mean
square; MSw, within-subjects mean square; SS, sum of squares; DF, degrees of freedom; F is calculated value
based on MS explained/MS unexplained (error).
Reprinted with permission from Weir JP. Quantifying test-retest reliability using the intraclass correlation coeffi-
cient and the SEM. J Strength Cond Res. 2005;19(1):231240.
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 148
148 PART II Seeking Answers: How the Question Drives the Methods
Total 15 1760
a
MSB, between-subjects mean square; MSE, error mean square; MSS, subjects mean square; MST, trials mean
square; MSw, within-subjects mean square; SS, sum of squares.
Reprinted with permission from Weir JP. Quantifying test-retest reliability using the intraclass correlation coeffi-
cient and the SEM. J Strength Cond Res. 2005;19(1):231240.
CONCEPT CHECK
Precision of measurement may be thought of in terms of how confident one is in the
reproducibility of a measure. Precision is reported as the SEM in the unit of measure
and takes into account the ICC of the measure as well as the standard deviation.
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 150
150 PART II Seeking Answers: How the Question Drives the Methods
Limits of Agreement
Bland and Altman (1986) have recommended that the limits of agreement (LOA)
be calculated when two measurement techniques (or two raters) are being com-
pared to each other. This technique compares the absolute differences between
two measurement techniques and specifically looks for systematic error. To per-
form this analysis, data must be on a continuous scale and both techniques must
produce the same units of measurement. The arithmetic difference in measures
between the two techniques is calculated for each subject. If the difference is zero,
the two techniques are identical. The LOA for the entire data set is computed as
LOA (x1 x2 )>n 1.96(sdiff ), where x1 is measurement 1, x2 is measurement
2, n is the number of subjects, and sdiff is the standard deviation of the difference.
The LOA represents a 95% confidence interval of the difference between the two
measures.
In addition to calculating the LOA, a BlandAltman plot should also be con-
structed and analyzed when comparing two measurement techniques. To do this,
for each subject the mean of the two measures is expressed on the x-axis and the
arithmetic difference between the measures is expressed on the y-axis. Once the
data from all subjects have been plotted, the investigator can observe for system-
atic error. Figure 10-1 illustrates a BlandAltman plot using data from data set A
in Table 10-6.
Mean difference of measure 1 and measure 2 2.75
Standard deviation of difference 10.0
Limits of agreement 10.0 (1.96 10.0) 16.8, 22.3
TABLE 10-6 Fictional data of limb girth measures taken two times by the same
clinician
MEASURE 1 MEASURE 2 MEASURE 1 MEASURE 2 MEAN OF 1 AND 2
90 99 9 94.5
The arithmetic difference between measures and the mean are calculated for use in the BlandAltman plot
shown in Figure 10-1.
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 151
30
20
Measure 1 Measure 2
10
10
20
30
0 50 100 150 200 250
Mean of Measures 1 and 2
FIGURE 10-1 BlandAltman plot.
Agreement
Estimates of the consistency or reproducibility of categorical data are called
agreement. Intrarater and interrater agreement are defined the same as with reliabil-
ity measures. Estimates of agreement are reported with the kappa (k) statistic which
also ranges from 1 to 1 with 1 indicating perfect agreement, 0 is what would be
expected by chance, and negative values indicate agreement occurring less than by
chance alone.
Calculation of k is as follows:
po pe
k
1 pe
where po is the observed agreement and pe is the expected agreement. po and pe are
both derived from a 2 2 contingency table of rater agreement.
EXAMPLE
Agreement
Table 10-7 shows data from two clinicians who independently observed the same
chronic regional pain syndrome patients for swelling of their involved upper
extremities.
(continued)
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 152
152 PART II Seeking Answers: How the Question Drives the Methods
Calculation of po is as follows:
ab
po
n
20 15 35
po 0.70
50 50
This indicates that the two clinicians both agreed that swelling was present or
absent in 70% of the observed patients.
Calculation of pe is as follows:
pe ([n1>n] [m1>n] ) ([n0>n] [m0>n] )
pe ([25>50] [30>50] ) ([25>50] [20>50] )
pe ([0.5 0.6] [0.5 0.4] ) (0.3) (0.2) 0.50
This indicates that in this data set the expected agreement between the two
clinicians due to chance was 50%.
Thus, k is calculated as:
Po Pe
k
1 pe
0.70 0.50 0.20
k 0.40
1 0.50 0.50
The most common interpretation for k is as follows:
<0 Agreement is less than by chance
0.010.20 Slight agreement
0.210.40 Fair agreement
(continued)
LibraryPirate
97689_c10 01/07/10 4:03 PM Page 153
CHAPTER SUMMARY
Whether a measure is nominal, ordinal, or continuous will dictate how that meas-
ure should be analyzed statistically. Validity is an inherent principle in research
design and this important concept has many components. Reliability and agree-
ment of measures are also key components of validity.
KEY POINTS
A key to quantitative inquiry is unbiased and objective measurement of the
dependent variables.
Validity is an inherent principle in research design and this important con-
cept has many components.
Reliability and agreement of measures are key components of validity.
To ensure objective results, the design of an experiment to test the hypothe-
sis must be done in an unbiased way.
Blinding affects and is pertinent to the internal validity of a study.
Agreement is reported with the kappa statistic that ranges from 0 to 1.
The randomized pretestposttest design is the gold standard for most exper-
iments.
Estimates of the consistency or reproducibility of categorical data are called
agreement.
154 PART II Seeking Answers: How the Question Drives the Methods
Applying Concepts
1. Consider whether a study is valid if it is not reliable or vice versa?
Argue your point with examples using contemporary issues from peer-
reviewed studies.
2. Design a quantitative study. Include research design, research hypothe-
sis/hypotheses, type of data that will be collected, describe any poten-
tial threats to internal validity, external validity of the study, how the
study will be determined reliable.
REFERENCES
Bland JM, Altman DG. Statistical methods for assessing agreement between two methods
of clinical measurement. Lancet. 1986;1:307310.
Shrout PE, Fleiss JL. Intraclass correlations: uses in assessing rater reliability. Psychol Bull.
1979;36:420428.
Weir JP. Quantifying test-retest reliability using the intraclass correlation coefficient and the
SEM. J Strength Cond Res. 2005;19(1):231240.
SUGGESTED READING
Hopkins WG. Measures of reliability in sports medicine and science. Sports Med.
2000;30:375381.
Viera AJ, Garrett JM. Understanding interobserver agreement: the kappa statistic. Fam Med.
2005;37:360363.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 155
CHAPTER 11
TESTS OF COMPARISON
CHAPTER OBJECTIVES
After reading this chapter, you will:
Know how estimates of variance are used in testing hypotheses and comparing group
values.
Understand types of data and the differences between parametric and nonparametric
statistics.
Understand the differences between interval and ratio data.
Learn how to interpret the clinical meaningfulness of data rather than simply focusing on
the statistical result.
Be able to explain issues of statistical significance, clinical meaningfulness, confidence
interval analysis, and effect size.
Understand the importance of context for the critical appraisal of clinical research.
Know how to consider tests of comparison and the clinical meaningfulness of data rather
than simply focusing on the statistical result.
Understand the need to compile evidence that will assist in making sound decisions about
the care of patients and clients.
155
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 156
156 PART II Seeking Answers: How the Question Drives the Methods
KEY TERMS
analysis of covariance parametric and nonparametric statistics statistical power
null hypothesis post-hoc comparisons type I and II errors
INTRODUCTION
This chapter introduces statistical procedures used to test hypotheses and investi-
gate differences between groups or within groups across time. The chapter is an
extension of the previous chapter devoted to sampling, probability, and measure-
ment reliability and validity, which are central to the critical appraisal and appli-
cation of the results of statistical analyses.
The chapter first focuses on types of data and the differences between paramet-
ric and nonparametric statistics. The primary focus of the discussion of parametric
statistics is on analysis of variance (ANOVA) procedures. This section leads the
reader through a working example to illustrate how estimates of variance are used
in testing hypotheses and comparing group values. This section also introduces con-
cepts of Type I and II errors, statistical power, and interaction between independ-
ent variables and concludes with an introduction to planned and post-hoc
comparisons and analysis of covariance.
The use of t-tests and the link between t-tests and ANOVA follow the section
devoted to ANOVA. In addition to introducing statistical procedures emphasis is
placed on the clinical meaningfulness of data rather than simply focusing on the
statistical result. An introduction to issues of statistical significance, clinical mean-
ingfulness, confidence interval analysis, and effect size provides a context for the
critical appraisal of clinical research.
The chapter concludes with an overview of nonparametric test of comparison
and provides a working example of one such statistical procedure, a Mann
Whitney U test. Entire texts have been devoted to tests of comparison. It is not the
intention here to prepare the reader to perform complex analyses but rather gain
an understanding of the basic principles so that the clinical research can be
appraised and the results of clinical research applied in a thoughtful process
rather than accepted or refuted.
These words offer some insight into the approach to this chapter. First, we do
not prove anything with statistics. Proof emerges over time through the accumu-
lation of convincing evidence. Lesson 1 is: do not read to accept the conclusions of
a research report as an absolute or final answer. Read to consider and compile evi-
dence that will assist in making a sound decision about the care of patients and
clients. Lesson 2: numbers can lie and the misinterpretation of data and statistical
analyses can mislead. Intentional misrepresentation and research fraud, while a
serious concern and potentially dangerous activity, is thankfully rare. However,
the adage dont believe everything you read must guide consumption of the
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 157
clinical literature. This chapter expands on the issue of sampling from populations
introduced in the previous chapters. The statistics discussed in this chapter are
exercises in probability and there is always the possibility that data drawn from a
sample doesnt represent the whole population. Moreover, an inappropriate selec-
tion of statistical techniques may also lead to conclusions that are not supported
by the data. Careful consideration and critical appraisal informs quality clinical
practice. This chapter provides the foundation for critical appraisal of research
comparing responses of groups to interventions or treatments used in the care of
patients and the prevention of injury and illness.
Lastly, this book in general and this chapter especially was written with the
full appreciation that most students preparing for careers in health care are not
fond of statistics. Our intent is not to change your mind but to make the medicine
as palatable as possible. The practice of evidence-based care depends on con-
sumption (and critical appraisal) of the available evidence (research) so in fact all
practitioners claiming to practice from an evidence base must also understand the
principles of statistics.
CONCEPT CHECK
It is important to gain an understanding of the basic principles so that the clinical
research can be appraised and the results of clinical research applied in a thoughtful
process rather than accepted or refuted.
CONCEPT CHECK
The first consideration in selecting an analysis is to determine the type or level of the
data to be analyzed.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 158
158 PART II Seeking Answers: How the Question Drives the Methods
Nominal simply means to name. The assignment of numeric values for analy-
sis of nominal (categorical) data is arbitrary. For example, if one were investigating
if there is a difference in the prevalence of Carpal tunnel syndrome between left-
and right-handed office workers, it would make no difference if left-hand workers
were coded as 1s and right-handed workers 2s or vice versa as long as the investi-
gator was clear as to what the 1s and 2s meant while interpreting the results.
Ordinal data, unlike nominal data, are ordered in a particular and meaning-
ful manner. Numeric pain scales represent a common form of ordinal data. We
acknowledge that a pain rating of 8 is worse than 5 and 5 is worse than 2. Thus,
coding of data is not an arbitrary assignment of numbers. Nominal and ordinal
data are analyzed through procedures that differ from those usually appropriate
for interval and ratio data.
CONCEPT CHECK
The assignment of numeric values for analysis of nominal data is arbitrary. Ordinal
data, unlike nominal data, are ordered in a particular and meaningful manner.
CONCEPT CHECK
Nonparametric statistical methods of comparison are often used to analyze nominal
data. Parametric statistics are appropriate to analyze interval and ratio data under most
circumstances.
Ordinal measures are a little more fickle. Some ordinal scales have interval
characteristics, an issue discussed later in this chapter. However, lets consider a
simple ordinal scale of poor, fair, good, and very good, applied to the rating of
videotaped performance of a set of therapeutic exercises incorporated in a home
program for patients recovering from rotator cuff tendinopathy. If the values are
coded as 0 poor, 1 fair, 2 good, 3 very good, then, how would one inter-
pret a mean of 1.5? It is not possible for an individual to score 1.5 and we are really
not sure if 1.5 is really halfway between fair and good. In this case the appropri-
ate measure of central tendency is the median (score that lies in the middle from
high to low when all scores are ranked) while range could be provided as a meas-
ure of dispersion.
What are the differences between interval and ratio data? First, lets consider
interval data. Interval implies that the differences between points of measure
are consistent and meaningful. The measurement of elbow flexion range of
motion provides an example of interval data. We know that the motion is meas-
ured in degrees and the difference between 20 and 30 degrees is the same as the
difference between 80 and 90 degrees. A gain or loss of 10 degrees is the same
throughout the spectrum of measurement. Contrast range of motion measures to
pain scale measures. Is an increase in pain from 1 to 2 the same as an increase from
8 to 9? There is really no way to be sure and this uncertainty affects the ability to
interpret average values and variance estimates.
Ratio data are similar to interval data but as the label suggests this level of
data can yield meaningful ratio values. Consider measures of force production
in comparison to measures of elbow range of motion. If a patient recovering
from knee surgery increased their knee extensor force production from 10 to 20
N over a period of time through strength training, we could conclude that their
strength has doubled. However, if a patient increased their elbow extension
range of motion from 55 to 110 degrees we should not conclude their motion
doubled. Why? In the case of the strength measures, which are ratio data, there
is an absolute 0 value. It is not possible to have negative strength. However, we
know that the extension of the elbow commonly exceeds the 0 value when
measured with a standard goniometer. The absence of an absolute 0 precludes
the calculation of meaningful ratios. In all other respects interval and ratio data
are similar and both types of data are analyzed with the same statistical
procedures.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 160
160 PART II Seeking Answers: How the Question Drives the Methods
CONCEPT CHECK
The absence of an absolute 0 precludes the calculation of meaningful ratios.
CONCEPT CHECK
Variance and standard deviation are measures of dispersion for interval and ratio data.
Variance is the difference between a score or value and a mean; and, a standard devi-
ation is the square root of variance. Parametric statistics analyze the (distribution of)
variance.
ANALYSIS OF VARIANCE
ANOVA is at once simplistic and highly complex. The simplicity lies in the calcu-
lation of variance. Complexity grows when variance is classified as explained and
unexplained and the ratio of explained and unexplained variance is linked to
probability. Complexity grows further when more complex research designs are
introduced. This section seeks to provide an understanding of variance and
ANOVA procedures and interpretation. Issues related to the analysis of data from
complex designs are introduced but the reader is referred to texts such as Keppel
and Wickens (2004) for further explanation.
To help in understanding variance and therefore ANOVA consider the fol-
lowing scenario (see Box 11-1 to review the steps necessary to complete an analy-
sis): A clinician wants to know if patients recovering from total knee arthroplasty
(knee replacement surgery) treated with neuromuscular electrical stimulation
(NES) and resistance exercise gain more quadriceps strength in the first 4 weeks
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 161
than patients who only complete resistance exercise. Sixteen patients consent to
participate in the study and are randomly assigned to one of the two groups.
Quadriceps strength is measured using a dynamometer on the first day of treat-
ment and after the 4th week of treatment consisting of resistance training on a pre-
scribed schedule. The dependent variable entered into the analysis was the
difference between strength measures at the 4th week minus the initial values.
2. Write the research question in a null form. The addition of neuromuscular electrical stim-
ulation in conjunction with a resistance exercise regimen does not result in greater improve-
ments in isometric quadriceps strength 6 weeks following knee replacement surgery than an
exercise program alone.
a. Abbreviated, the null is written NES = no NES
3. Collect data.
162 PART II Seeking Answers: How the Question Drives the Methods
With the data collected it is time to proceed to the analysis, but how does one
complete the task? From the outset the question was if the addition of NES-
enhanced strength gain in the first 4 weeks of recovery. One look at the data above
would lead to the conclusion that indeed it does. However, the process is far more
complicated than eyeballing the descriptive statistics (mean and standard devia-
tion). First, it is important to remember the purpose of most research involving
comparisons is to infer the results to the population from which the samples
were drawn. Thus, the analysis we will undertake with these data will address
the question as to whether, if we could study the entire population of knee
replacement patients, we would find that NES results in more rapid strength
gains in the quadriceps. Since these fictitious data were acquired from small
samples of a larger populations (knee replacement patients treated with and
without NES) it is possible that subjects in the NES group were inclined to make
greater gains for reasons other than NES. The analysis we will perform will esti-
mate the probability that the differences found in our study reflect real population
differences.
CONCEPT CHECK
The purpose of most research involving comparisons is to infer the results to the pop-
ulation from which the samples were drawn. The analysis we perform will estimate the
probability that the differences found in our study reflect real population differences.
CONCEPT CHECK
While our research question asks if one group differs from another the hypothesis
tested in statistics, called the null hypothesis, is that there is no difference between
groups. In rejecting the null the only viable conclusion is that the groups are different.
Note that it is not possible to accept a null since two groups will not truly be equal.
If we fail to reject a null, we must suspend judgment as to whether groups differ.
Now lets return to the data and test the null hypothesis that gains in strength
through the exercise program alone are equal to those when exercise is supple-
mented with NES. To do this we will perform ANOVA. Certainly, the use of com-
puter software is recommended for analyzing larger data sets typical of research.
However, we will go through the process the old fashion way to illustrate what
information obtained through a computerized analysis means. The result of
ANOVA is an F value which is a point on an F distribution that permits estimates
of probability. However, it is the calculation of an F that really allows for an under-
standing of the process of ANOVA. The formula for an F is a ratio of variance
estimates thus the term analysis of variance. F mean square explained/mean
square unexplained (also sometimes referred to as ms error). What the heck is a
mean square? A mean square is essentially the sum of the squared differences from
each score and a mean divided by the number of scores minus 1. To calculate a
mean square the first step is to calculate the differences between individual scores
and a mean score. These values are then squared since the sum of these values
always equals zero. The sum of the squared values is then divided by the degrees
of freedom to produce a mean square. A more thorough discussion of degrees of
freedom (df) is reserved for later but in most cases df n 1, where n the num-
ber of scores forming the mean. This process is illustrated with our data below.
164 PART II Seeking Answers: How the Question Drives the Methods
If we sum the squared values from both groups, the value is 224 and forms the
numerator for calculating a mean square. The df in each group equals the number
of subjects 1 or n 1 and since the numerator was derived from both groups
the denominator will be equal to (8 1) (8 1) 14. The mean square value
thus 176 48/n 2 224/14 16. This mean square represents unexplained
variance and will form the denominator of our F ratio.
CONCEPT CHECK
The result of ANOVA is an F value that is a point on an F distribution that permits esti-
mates of probability. F = mean square explained/mean square unexplained. A mean
square is essentially the sum of the squared differences from each score and a mean
divided by the number of scores minus 1.
CONCEPT CHECK
Unexplained variance is described as variation from the mean that is attributed to factors
beyond the scope of the research design; plausible yet undefined, superfluous factors.
Now we look at the group means. If we were to characterize the average per-
formance of subjects within a group, we could assign each subject the mean value.
The value for the group receiving NES differs from the group that did not. These
scores differ from the average of all scores or the grand mean. Why are the scores
in the group treated with NES different from the grand mean? This can be
explained by the treatment received and the sum of the difference between each
subjects score, when assigned the group mean score, and the grand mean allows
for calculation of explained variance. Now lets complete the process of calculating
ms explained. The grand mean (mean of all scores) 26. Thus if each subject in the
group receiving NES were given the group mean score, the variance from the
grand mean of each subject is 4 and the squared variance is 16. There are 8 subjects
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 165
in the group and the sum of the squared deviations is 8 16 128. If we move to
the group that received exercise alone and assign each subject the group mean of
22, the process repeats and the total ms explained 256. The ms explained 256/k
1, where k the number of group means. In this case there are two, so ms
explained 256. When a large portion of variance in scores can be explained we
are much more likely to conclude that differences observed in a sample reflect real
population differences. In this case it appears that a large portion of variance is
explained by the treatment rendered since the F 256/16 16.0.
INTERPRETING F
F, like the normal distribution (see Figures 9-1 and 11-1) is a distribution of values
along a continuum. There are, however, an infinite number of F distributions that
are reflective of the number of df in the numerator and denominator which is deter-
mined by the number of variables, the number of levels of those variables, and the
number of subjects in a study. In our example there is one variable with two levels
(NES vs. no NES). If we added a third group that received a subsensory stimulation,
a new level would be added, while if we divided the subjects by gender and then
assigned them to treatment, a new variable (gender) would be added. The appropri-
ate F distribution is identified by finding the degrees of freedom associated with
explained variance and unexplained variance. Fortunately, computer software takes
care of selecting the correct reference F distribution for an analysis but our simple
example allows for a more thorough explanation of degrees of freedom. In this
process we assume that the mean value of a group of scores is fixed or not free to vary.
Taking the scores in the NES group as an example, the mean 30 is fixed. Given there
are 8 scores contributing to that mean 7 could be any value. However, regardless of
those 7 scores the 8th would have to result in a group mean of 30 and thus not free to
vary or be any value. Thus, 7 scores are free to vary resulting in 7 degrees of freedom.
166 PART II Seeking Answers: How the Question Drives the Methods
The same case exists for the second group in our analysis thus the degrees of freedom
associated with the ms unexplained 14. Now if we examine the ms explained, we
find two group means forming the grand mean. Thus, using the same logic we find
that only one of those scores is free to vary since we again assume the grand mean to
be a fixed value. Thus the degrees of freedom for explained variance is 1. The F distri-
bution that reflects the probability of the F value 16 in this analysis is the one asso-
ciated with 1 df in the numerator and 14 in the denominator.
CONCEPT CHECK
The appropriate F distribution is identified by finding the degrees of freedom associ-
ated with explained variance and unexplained variance.
The larger the F (ratio of explained/unexplained variance), the less likely that
the differences observed were chance occurrences. In other words, the likelihood of
a large F resulting from chance is quite small. By consulting an F distribution table
found in many statistics texts (see Table 11-1) we learn that the chances of an
F 16 occurring by chance is less than 1 in 100. By convention, researchers are gen-
erally willing to accept less than a 5 in 100 risk that an F value obtained is a chance
occurrence. The level of risk is known as an alpha value and usually established
before the study is begun. When the F value is larger and the probability therefore
great that the results reflect real population difference, we reject the null hypothe-
sis and thus conclude that differences observed are due to the effects of our inter-
vention (in this case the introduction of NMS into the treatment). It is important to
note that an alpha 0.05 (acceptance of a 5 in 100 risk of incorrectly concluding
that observed differences do not reflect true population differences) is the most
common level of risk of being wrong used in research, but it is not an absolute stan-
dard. In some cases one might want to be more conservative and accept, for exam-
ple, a 1 in 100 risk (alpha 0.01) of being wrong while in exploratory work a more
liberal approach might be to accept a 10 in 100 risk (alpha 0.10).
CONCEPT CHECK
When the F value is larger and the probability therefore great that the results reflect
real population difference we reject the null hypothesis and thus conclude that differ-
ences observed are due to the effects of our intervention. The alpha value specifies
the level of accepted risk of incorrectly concluding that observed differences do not
reflect true differences in a population of 100.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 167
6 5.987 5.143 4.757 4.533 4.387 4.060 3.938 3.874 3.808 3.774 3.740 3.705 3.669
7 5.591 4.737 4.347 4.120 3.972 3.637 3.511 3.445 3.376 3.340 3.304 3.267 3.230
8 5.318 4.459 4.066 3.838 3.688 3.347 3.218 3.150 3.079 3.043 3.005 2.967 2.928
9 5.117 4.257 3.863 3.633 3.482 3.137 3.006 2.937 2.864 2.826 2.787 2.748 2.707
10 4.965 4.103 3.708 3.478 3.326 2.978 2.845 2.774 2.700 2.661 2.621 2.580 2.538
11 4.844 3.982 3.587 3.357 3.204 2.854 2.719 2.646 2.571 2.531 2.490 2.448 2.405
12 4.747 3.885 3.490 3.259 3.106 2.753 2.617 2.544 2.466 2.426 2.384 2.341 2.296
13 4.667 3.806 3.411 3.179 3.025 2.671 2.533 2.459 2.380 2.339 2.297 2.252 2.206
14 4.600 3.739 3.344 3.112 2.958 2.602 2.463 2.388 2.308 2.266 2.223 2.178 2.131
15 4.543 3.682 3.287 3.056 2.901 2.544 2.403 2.328 2.247 2.204 2.160 2.114 2.066
16 4.494 3.634 3.239 3.007 2.852 2.494 2.352 2.276 2.194 2.151 2.106 2.059 2.010
17 4.451 3.592 3.197 2.965 2.810 2.500 2.308 2.230 2.148 2.104 2.058 2.011 1.960
A segment of the full F distribution illustrating the critical values with 1 and 14 degrees of freedom (df). The critical
value of F at an alpha value = 0.05 is 4.600.
Complete F distribution tables are available in many statistics books and websites including
http://www.statsoft.com/textbook.
168 PART II Seeking Answers: How the Question Drives the Methods
CONCEPT CHECK
Type I errors occur when a null is rejected when in fact population differences do not
exist. The alpha value is really the level of risk of Type I error. Type II error occurs when
a null is not rejected yet a study of the population would reveal differences between
groups.
Again turning to the example above we find the mean differences between
groups to be fairly large. Suppose, however, the group means were 24 for the NES
group and 22 for the exercise group.
In reality, the only factor investigators can control is sample size. Research is
costly and time consuming and in some cases poses a risk to participants. Thus,
increasing power is not always an easy matter of just adding some subjects. In
many cases the variance in a measure can be estimated from previous research
and a magnitude of difference between groups that would be of clinical value can
be entered into equations to estimate the number of subjects needed to produce an
F that would reject a null hypothesis. Such power calculations can be useful but
the amount of variance in samples may differ a good bit, especially if the samples
are small, and in some cases it is not reasonable to predetermine the size of mean
differences that could be deemed important. Moreover, if large samples are stud-
ied, it is possible that statistically significant differences (null is rejected) are
reported that are of little clinical significance. The issue of magnitude of differ-
ences is explored further with the introduction of confidence intervals shortly.
CONCEPT CHECK
Researchers guard against Type I error by selecting the alpha level. Statistical power is
required to decrease the risk of Type II error.
170 PART II Seeking Answers: How the Question Drives the Methods
CONCEPT CHECK
When a between-subjects variable and a within-subjects variable exist in a research
design the design may be referred to as a mixed model. Examples of such research
designs are very common in health care research.
INTERACTION
Greater complexity in research designs and thus the data analysis is not necessar-
ily an indicator of better research. In some cases complexity seems to confuse
rather than clarify. However, many questions addressed in health carerelated
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 171
research must take into consideration more factors or variables to derive maxi-
mum benefit from the work. Consider the example above where gender was
added as a second independent, between-subjects variable. The purpose for
adding gender to the design is not to determine whether men and women differ
in the recovery of strength following knee replacement but rather to investigate
whether they differ in response to the intervention being studied. This is a subtle
difference in wording but large difference in purpose because the investigator is
really seeking to understand the interaction between treatment and gender. The
concept of studying the interaction between independent variables is important
for investigators and research consumers alike.
CONCEPT CHECK
The concept of studying the interaction between independent variables is important
for investigators and research consumers alike.
Computer software has made it far easier to analyze data from research where
multiple variables are considered in a single analysis and to study the interactions
between variables. Unfortunately, the software does not generate results that
readily reveal precisely how variables interact. Lets suppose we added gender as
a between-subjects variable to our study and the analysis revealed a significant
interaction between the treatment regimen and gender. What does significant
interaction really mean? First, recalling our previous discussion significant sug-
gests that the finding is unlikely a chance occurrence but rather a reflection of a
population phenomenon. Unfortunately, the statistical result tells us little else. To
better understand how variables interact we can turn to figures (see Figure 11-2)
that include cell means and standard deviations as appears below.
CONCEPT CHECK
It is important to remember that significant suggests that the finding is unlikely a chance
occurrence but rather a reflection of a population phenomenon. Unfortunately, the sta-
tistical result tells us little else.
172 PART II Seeking Answers: How the Question Drives the Methods
Pre Post
< 30
> 50
6 RM
Pre Post
FIGURE 11-2 The significant age by time interaction indicates that the two age groups
responded differently over time. There was a slight increase from pre to post 6 RM perform-
ance in the 30 age group, but a much larger increase was seen in the 50 age group.
LEVELS OF VARIABLES
In the previous section we explored adding more than one variable into an analy-
sis. It is also possible that there may be multiple levels within a variable. This con-
cept was introduced in the discussion of collecting data at multiple time points
with time becoming a within-subjects variable. Many variables may include
more than two levels. It might be that an investigator wants to study two different
forms of NES and thus has three levels of treatment (NES 1, NES 2, and exercise
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 173
without NES). As with the addition of variables into a research design the addition
of levels within a variable adds to the complexity of interpretation. However, the
addition of levels of independent variables can maximize efficiencies by allowing
completion of one larger study rather than multiple smaller studies and may yield
greater insights into the interactions between the variables of interest.
CONCEPT CHECK
For our purpose, it is simply necessary to understand that when one encounters ref-
erence to procedures of post-hoc testing the investigator/author is conveying that addi-
tional analyses were performed to isolate the sources of significant (likely not due to
chance) differences between sets of scores.
At this point one may ask why, in the scenario described, an investigator
would not simply use the procedures previously described and simply perform
three comparisons, A to B, A to C, and B to C, to identify where differences
between groups might be. The problem with this approach, and the necessity
of planned comparison and post-hoc analyses is that the risk of Type I error
exists with each analysis performed. Thus, if three analyses are performed with
an alpha 0.05, the potential for Type I error occurring somewhere is 15%. Thus,
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 174
174 PART II Seeking Answers: How the Question Drives the Methods
the conventional approach is to see if any difference exists between means (or in
other words a portion of a null hypothesis can be rejected) with a known and rea-
sonable risk of Type I error and finding the sources of the differences once it is
known that at least a portion of the null can be rejected.
CONCEPT CHECK
The risk of Type I error exists with each analysis performed. Stated differently, with
each additional analysis there is an added possible risk for Type I error.
ANALYSIS OF COVARIANCE
Before moving on to other topics two matters related to ANOVA warrant a brief
mention. The research consumer will occasionally encounter ANCOVA that signi-
fies analysis of covariance. ANCOVA is a special case of ANOVA where a variable
is introduced for the purpose of accounting for unexplained variance. When used
appropriately, which is not always the case, ANCOVA increases statistical power
(chance of rejecting the null hypothesis). The introduction of a covariant, however,
must be planned and readily justified. From the perspective of the research con-
sumer ANCOVA has little impact on the process of critical appraisal and interpre-
tation of data assuming the introduction of a covariate is based on sound
rationale. If the rationale does not appear to justify the analysis, move on, the data
may not support the conclusions drawn by the investigator.
CONCEPT CHECK
ANCOVA is a special case of ANOVA where a variable is introduced for the purpose
of accounting for unexplained variance.
observed between function of patients undergoing the specific exercise and a gen-
eral recommendation for progression to walking and running, follow-up analyses
would be needed to determine on which, if any, functional tests patients differed. It
is possible that differences only emerge when the combination of dependent meas-
ures is analyzed. This issue is addressed in greater detail by Pedhazur (1997).
CONCEPT CHECK
MANOVA refers to multivariate analysis of variance or case where more than one depend-
ent measure is analyzed simultaneously. It is best applied when the investigator is inter-
ested in the affect of the independent variable(s) on the collection of dependent variables.
T-TESTS
Before exploring confidence intervals a brief discussion of t-tests is warranted. Some
readers might have expected this discussion earlier since t-tests are often introduced
before ANOVA. t-tests, however, are really a special case of ANOVA where there are
only two sets of data in the comparison. However, t2 F so it is equally permissible
to perform ANOVA as was done above. Like F, t values are points on a curve and
there are an infinite number of t distributions, each corresponding to the degrees of
freedom associated with unexplained variance. The degrees of freedom associated
with explained variance is always 1 since there are only two sets of data.
t- tests and ANOVA are also similar in that between and within subject data
can be analyzed. A dependent t-test is used to analyze within subject data. For
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 176
176 PART II Seeking Answers: How the Question Drives the Methods
example, if a cohort were tested for gains on a timed test of stair climbing before
and after a 3 week lower extremity exercise program a dependent t-test would be
used to test the null hypothesis the stair climbing speed before and after the exer-
cise program were equal.
An independent t-test is appropriate for between subjects analyses. The formu-
lae for dependent and independent t-tests differ. Statistical software used in these
analyses will guide the selection of the appropriate calculations. To illustrate the
process, the formula for an independent t-test is presented below and the process of
performing the analysis is described using the data from the study of NES described
previously. To illustrate the process, the formula for an independent t-test is pre-
sented below and the process of performing the analysis is described using the data
from the study of NES described previously. The independent t-test formula is:
meanA meanB
t
1 1
Spool
B nA nB
where:
SA SB
Spool
B 2
Note that in this example the group sizes are equal and an assumption of
equal within group variance is made. Large differences in group size and variance
affect the outcome of t-test analyses and must be accounted for through the use of
modified formulas. This work is typically done with commercially available soft-
ware but the investigator must recognize the potential problems as well as the
available solutions under special circumstances.
Recall that SD is the square root of variance and thus the link between the sim-
ple formula for t and ANOVA. If the values from our example are entered into the
equation we find, with a bit of compensation for rounding error that
t 30 22/2 4 or t square root of F (16). When we compare the calculated
value for t (4.0) with the critical value (see Table 11-2) we again find (as we did
when performing ANOVA on these data) the calculated value exceeds the critical
value and thus the null hypothesis is rejected.
A quick glance at an F and a t distribution reveals that all F values are posi-
tive while t values may be positive or negative (see Figure 11.3). This leads to one
other issue associated with t, directional hypotheses. The null associated with
ANOVA is that two groups are equal (A B) while with t we have a choice of a
null of A B or vice versa A B.
Again lets return to our example of the study of NES. If we assign A to the NES
group and B to the exercise group, we could write a null hypothesis A B. If the null
is rejected, then A must be greater than B or NES more effective than exercise alone.
We really do not care if exercise alone is better than NES or equal to NES because if
NES is not greater then we would not use the treatment. In this case the point
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 177
A segment of the full t distribution illustrating the critical values with 14 degrees of freedom (df). Note that the
numerator df of t is always = 1. The critical value of a one-tailed test = 1.761, while that of the two-tailed test is 2.145.
Complete t distribution tables are available in many statistics books and websites including
http://www.statsoft.com/textbook.
associated with a 5 in 100 error would be found on the right side of the distribution
and the t required to reject the null a bit smaller (creating greater statistical power)
than if we had to account for error on both sides of the distribution. When we con-
sider only one side of the distribution we are performing a one-tailed t-test. If both
AB and BA would be of interest then a two-tailed t-test is performed. In fact, it is
unusual to be interested in both alternatives and thus one-tailed tests are most com-
monly performed. Thus, when performing a one-tailed t-test there is a greater chance
of rejecting the null than when performing a two-tailed test or ANOVA.
CONCEPT CHECK
t-tests are really a special case of ANOVA where there are only two sets of data in the
comparison. Like F, t values are points on a curve and there are an infinite number of t
distributions, each corresponding to the degrees of freedom associated with unexplained
variance.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 178
178 PART II Seeking Answers: How the Question Drives the Methods
CONCEPT CHECK
It is possible to reject a null hypothesis and conclude little clinical consequence or
conversely fail to reject a null when the possibility of clinically meaningful differences
exists.
This result could be obtained by hand using the following formulas. The
first step in calculating CI is to calculate a pooled standard deviation using the
formula:
SE(d) Spool # a b
1 1
n1 n2
CI d t SE(d) to d t SE (d)
How does this result compare to that obtained by ANOVA or t-test analysis? In
all cases the null that improvements in quadriceps force generation following NES
and exercise is equal the improvements with exercise alone is rejected. The reporting
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 180
180 PART II Seeking Answers: How the Question Drives the Methods
of this result with a CI requires more explanation. The 95% CI in this case does not
include 0 as the lower limit is 3.7. Thus, we can conclude with 95% certainty that pop-
ulation differences exist. Moreover, we now have a sense of the magnitude of the dif-
ferences that would be observed if the whole population could be studied. In other
words, we are 95% certain that the true population difference would lie between 3.7
and 12.3 Nm. Now we can ask if these differences are clinically meaningful.
It is likely that most clinicians would agree that a 12-Nm greater improvement
in force production would warrant the use of NES. However, a 4-Nm difference is a
little less convincing and thus one may hesitate to apply these results in clinical prac-
tice. Is the extra time and effort warranted for this level of improvement? Of course,
this is a judgment call but now a decision that can be better informed than if we were
to rely solely on the mean differences observed in our study of samples and the
knowledge that there is 95% certainty that some real population difference exists.
In our case the 95% CI is wide given the nature of the data. A statistically
meaningful difference may not reflect clinically meaningful differences and the
clinician may await more information before including NES into a plan of care.
How would more information affect the 95% CI and the clinical decision process?
Similar to our discussion regarding Type II error and statistical power, a larger
sample will increase power and narrow the 95% CI. Sometimes data from multi-
ple studies can be pooled in a process known as meta-analysis to provide more
refined estimates of CIs. Meta-analysis is discussed in greater detail in Chapter 18.
It is also important to note the relationship between an alpha value and a CI. A
95% CI implies an alpha value 0.05. If a more conservative alpha of, for exam-
ple, P 0.01 were selected then the 99% CI would be calculated.
CONCEPT CHECK
A statistically meaningful difference may not reflect clinically meaningful differences;
thus, additional information may be needed before deciding on a plan of care.
Now consider the case where a null is not rejected or, in other words, statisti-
cally significant differences were not found. Recall the scenario in this chapter
where the mean of the participants receiving NES was reduced to 24. Entering
these data into the CIA software we obtain the following 95% CI: 2.36.3. Since
the 95% CI includes 0 we are not certain that, in fact, the population difference is
not 0. However, we might believe that a difference of 4 or more units would be
clinically meaningful. Thus, rather than rejecting the notion the NES improves
strength gains we again await more information and a narrower confidence inter-
val before including NES in the routine care of knee replacement patients, but we
do not fully reject the possibility that the treatment might yield benefit as the true
population mean may lie within the range we believe clinically important.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 181
CONCEPT CHECK
The fundamental problem of relying on the probability that differences observed in a
sample reflect population differences is that the magnitude of a difference is not con-
sidered. The reliance on probability estimates alone in the consideration of research
findings may result in dismissing differences that require further investigation or
accepting as important or significant differences that are of such magnitude that they
are of little clinical importance.
To review, consider the result of our ANOVA (F 16, P 0.05) and the 95%
CI 3.712.3 reported previously. The CI does not include 0 and thus we are
95% certain that a true population difference exists. This conclusion is the same
as rejecting a null hypothesis when the selected alpha level 0.05. In other
words, if the 95% CI does not include 0, the null hypothesis can be rejected
with the same level of certainty as was done through ANOVA or completion of
a t-test. Moreover, the research consumer now has a clearer understanding of
the magnitude of the effect. If only the statistical result and mean differences
and variance estimates are reported, as was once common, the reader is left
either to conclude that a difference exists and the point estimate of the differ-
ence is the mean value, calculate the CI on their own or to calculate and inter-
pret effect size. The latter two options are possible but require additional work
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 182
182 PART II Seeking Answers: How the Question Drives the Methods
Small, Large
clinically Clinically clinically
unimportant meaningful important
difference difference difference
and understanding on the part of the consumer. It is better for all when inves-
tigators simply provide the CI associated with the preselected alpha value
(P 0.05 corresponds to a 95% CI, P 0.01 to a 99% CI, etc.) because the
consumer has more work to do.
Take, for example, the four bars on the graph representing CIs in Figure
11-4. The first represents a situation where the null is rejected but the CI approaches
0 and is very large. The consumer is left with a great degree of uncertainty regard-
ing the true magnitude of treatment effect. In this case more data are needed to nar-
row the CI, an issue of power to be discussed shortly, before being convinced that
the treatment will generate, on average, a clinically beneficial effect.
In the second bar the CI is narrow and regardless of where the true popula-
tion effect lies within the CI the intervention is of clinical importance. This is the
best-case scenario in data reporting. The third bar includes 0 so we cannot be
certain that a difference between treatments truly exists. However, the interval
is also large and if the true population difference falls toward the right side of
the bar, the intervention might help a lot of patients. Lastly, we have an interval
that includes 0 and does not include differences that are of clinical interest.
Thus, even if this result represents a Type II error the small magnitude of the dif-
ference would lead us to conclude that the treatment is of little value. These
examples hopefully convey the value to the consumer when CIs are reported. In
some cases the CI can confuse rather than clarify when limits approach 0 and
intervals are very large but it is likely better to be a bit confused and reserve
judgment than accepting an outcome as important simply because a null
hypothesis was rejected.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 183
EFFECT SIZES
Another useful approach to understanding what the observed differences
between groups mean in terms of magnitude of effect, or in other words the typ-
ical response to intervention, is through the calculation of effect size. Like other
aspects of statistics, the subject of effect size is broader than our focus here. There
are multiple approaches to effect size depending on the purpose of the research
and the nature of the data. However, when we consider differences between
groups, effect size calculations place the magnitude of differences between groups
in the context of group variance.
CONCEPT CHECK
When we consider differences between groups, effect size calculations place the mag-
nitude of differences between groups in the context of group variance.
Jacob Cohen (1988) contributed greatly to the use and interpretation of effects
size and Cohens d is one of the most commonly referenced methods of calculat-
ing effect size as follows:
meana meanb
d
spool
Hedges (1981) offers a similar formula; however, the denominator was based on
the degrees of freedom as follows:
(7 2.62) 53.35
s 2(7 5.01) 3.33
88 16
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 184
184 PART II Seeking Answers: How the Question Drives the Methods
CONCEPT CHECK
There is no substitute for knowledge of the nature of data and clinical populations when
considering the magnitude of effect. The research consumer should use effect size esti-
mates as one element of critical appraisal of research reporting group comparisons.
NONPARAMETRIC STATISTICS
In the introduction of this chapter the type of data or level of measurement was
linked to the selection of the appropriate analysis. The notion of population
parameters of mean and standard deviation was introduced in preparation of our
discussion of parametric statistics in general and ANOVA in particular. It is time
to turn attention to nonparametric statistics but before doing so a couple of points
of clarification are warranted.
First, the terms nonparametric and distribution-free are sometimes used
interchangeably. It is important to note that when parametric analyses are com-
pleted it is assumed that the data are based on observations of a normally dis-
tributed population with similar variance and samples are drawn at random.
These assumptions can be tested (histogram or box plot allows for assessment of
distribution while Levenes test tests a null that population variances are equal).
If these assumptions are not met, nonparametric procedures may be the appro-
priate analytical methods. Such circumstances may occur with small samples
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 185
and one may see a nonparametric analysis of interval or ratio data from a small
sample. These issues, however, rarely impact the analysis of data in the clinical
research literature and the use of nonparametric procedures is generally
reserved for ordinal and nominal data. Moreover, violation of the assumptions
is unlikely to have a substantial impact on the statistical outcome as procedures
such as ANOVA are robust and not highly sensitive to departures from these
assumptions nor the conclusions drawn from research if one considers the mag-
nitude of effect in conjunction with the probability that differences observed
represent a true population difference. Keppel and Wickens (2004) offer a more
complete discussion of the issues related to violating the assumptions underly-
ing ANOVA.
CONCEPT CHECK
It is important to note that when parametric analyses are completed it is assumed that
the data are based on observations of a normally distributed population with similar
variance and samples are drawn at random.
CONCEPT CHECK
Nonparametric statistics test hypotheses about medians or in the case of nominal data,
distribution. The most appropriate nonparametric test in a particular circumstance
depends on the nature of the data and research method of the study.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 186
186 PART II Seeking Answers: How the Question Drives the Methods
MannWhitney U Test
The MannWhitney U test is analogous to the paired t-test discussed and the
ANOVA performed for illustration earlier in this chapter. The analysis tests the null
hypothesis that the median score in one group (A) is or to the median score of
a second group (B) (A
B). If the analysis reveals the median of B A, we might
reject the null hypothesis. However, we would want to know the probability that
the result was a chance finding. As with parametric tests reference tables allow us
to know the probability of a test result occurring by chance (see Table 11-3). Recall
that the parametric statistics discussed previously result in F and t values. The
MannWhitney U result is designated as a T (capital vs. t used to designate a para-
metric test). As with parametric tests the null hypothesis (A
B) is only rejected if
the probability of obtaining a T-value is sufficiently small (e.g., less than 5%). The
process of performing a MannWhitney U is described by Daniel (see section
Suggested Reading at the end of this chapter).
N2
6 31 36 40 44 49
7 36 41 46 51 56
8 40 46 51 57 63
9 44 51 57 64 70
10 49 56 63 70 77
A segment of the full MannWhitney distribution illustrating the critical value with eight participants (N1 and N2)
in each of the groups at an alpha value = 0.05.
Complete MannWhitney U contingency tables are available in many statistics books and websites including
http://www.zoology.ubc.ca/~bio300/StatTables.pdf.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 187
suffering from chronic ankle instability might be asked to rank four commercial
ankle stabilizers for comfort and effectiveness. Freidman labeled the test statistic
for this test an X2r and the probability of a result can be determined by referring
to a reference table specific to the Friedman analysis. The reference tables for these
nonparametric procedures can be found in statistics texts devoted to, at least in
part, nonparametric procedures. These sources are noted because these analyses
are relatively easily completed by hand with small data sets. The formulas for
these analyses are also found in these texts. The formulas differ from one another
and might appear a little strange as illustrated in the MannWhitney U analysis
but all are similarly easy to negotiate.
Note that none of these nonparametric tests allow for the analysis of
repeated measures from multiple groups known as a mixed model design.
This represents one of the major limitations of these statistical tests in clinical
research since change across time with different interventions is often the pur-
pose of an investigation. This reality has led to the analysis of ordinal data
such as pain ratings with parametric statistics. In these cases the data are
assumed to possess interval properties. In many cases this may be a reasonable
assumption but the research consumer should be aware of the assumptions
made and be provided with reasonable justification for the analyses per-
formed.
188 PART II Seeking Answers: How the Question Drives the Methods
Formula:
n(n 1)
T S
2
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 189
where S is the sum of the ranks in group X and n the number of participants in
group X. The assignment of X is arbitrary but in this case we are interested in X Y
and/or discomfort with NES exercise without NES so NES was assigned as X.
Continuing:
8(9)
T 76.5 or 76.5 36 40.5
2
190 PART II Seeking Answers: How the Question Drives the Methods
the work. There is no black and white answer but like many compromises in
research has shades of gray.
CHAPTER SUMMARY
This chapter is devoted to examining differences between groups. The selection of
the appropriate test requires a determination as to the nature of the data. Ratio
and interval data are associated with parametric procedures while ordinal and
nominal data are managed using nonparametric methods. ANOVA is the most
common parametric procedure used in analyzing data related to group compar-
isons and the chapter devotes considerable attention to the basic components of
such an analysis.
Attention is briefly turned to t-tests before examining the issues of magnitude
of effect and clinical meaningfulness. Advances in statistics and the evolution of
evidence-based practice as well as the field of clinical epidemiology have altered
the perspective as to how research data are best interpreted and applied in clini-
cal practice. We are in fact living in an era of change, which, while a step forward
in health care, adds another layer of uncertainty as to what the research we read
really means in the context of caring for our patients. Many investigators and cli-
nicians received instruction in statistics at a time where the risk of Type I error was
the dominant concern and nonsignificant differences (in which the calculated F or
t values could occur by chance 5 in 100 times) were rarely reported in the clini-
cal literature because such reports were rejected by editors.
The fundamental problem of relying on the probability that differences
observed in a sample reflect population differences is that the magnitude of a dif-
ference is not considered. The reliance on probability estimates alone in the con-
sideration of research findings may result in dismissing differences that require
further investigation or accepting as important or significant differences that
are of such magnitude that they are of little clinical importance. The reporting of
confidence intervals addresses the issues of probability while providing an inter-
val in which true population differences lie. Effect size reporting provides the
consumer with a sense of the magnitude of the difference between group mean
values in relation to the variance among subjects. Each has the potential to assist
the research consumer appraise clinical research that is the overarching mission
of this book.
Nonparametric statistics is a broad topic well beyond the scope of this chap-
ter. The limited coverage here is not to imply that these analyses are of less impor-
tance. There are, however, multiple nonparametric techniques. We chose to
perform one such technique to illustrate the process and discuss the interpretation
of findings. It is hoped that this offers a foundation for exploration of research
related to diagnostic procedures, prevention, prognosis, and intervention found in
the next section.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 191
See Table 11-4 for a summary of the tests of comparison covered in this chapter.
KEY POINTS
Statistics do not prove anything.
Do not read to accept the conclusions of a research report as an absolute or
final answer.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 192
192 PART II Seeking Answers: How the Question Drives the Methods
Numbers can lie and the misinterpretation of data and statistical analyses
can mislead.
Most students preparing for careers in health care are not fond of statistics.
Careful consideration and critical appraisal informs quality clinical practice;
thus, it is necessary to understand the principles of statistics.
Applying Concepts
1. Discuss the various factors that influence statistical power.
a. Consider the role and responsibility of a research in controlling these
factors.
2. What are the parameters that differentiate the type of analysis to be
performed?
a. Discuss examples of clinical relevance.
b. Consider methodical and practical implications of exploring the
same clinic question from more than one research perspective, using
different types of analyses to test the same or different hypotheses in
order to compare findings.
c. Consider the possible clinical implications that might result from
such an approach.
3. Consider the implications of the following scenarios, and discuss plau-
sible clinical examples:
a. A research study where misinterpretation of data and statistical
analyses can mislead.
b. A clinical study in which the results do not match the conclusion(s).
c. Manipulating the findings of a study to justify a desired clinical
effect or outcome.
d. Critical appraisal of research comparing responses of groups to
interventions or treatments used in the care of patients and the pre-
vention of injury and illness.
LibraryPirate
97689_c11 01/07/10 4:03 PM Page 193
REFERENCES
Altman DG, Machin D, Bryant TN, et al. Statistics with Confidence. 2nd ed. London, UK: BMJ
Books; 2000.
Chottiner S. Statistics: towards a kinder, gentler subject. Journal of Irreproducible Results.
1990;35(6):1315.
Cohen J. Statistical Power Analysis for the Behavioral Sciences. 2nd ed. Mahwah, NJ: Lawrence
Erlbaum Associates; 1988.
Hedges LV. Distribution theory for Glasss estimator of effect size and related estimators.
J Educ Stat. 1981;6:107128.
Keppel G, Wickens TD. Design and Analysis: A Researchers Handbook. 4th ed. Upper Saddle
River, NJ Prentice-Hall; 2004.
Pedhazur EJ. Multiple Regression in Behavior Research. New York: Holt, Rinehart and
Winston; 1997.
SUGGESTED READING
Daniel WW. Biostatistics: A Foundation for Analysis in the Health Sciences. 9th ed. New York:
John Wiley & Sons; 2008.
Glass GV, Peckham PD, Sanders JR. Consequences of failure to meet assumptions underly-
ing fixed effects analysis of variance and covariance. Rev Educ Res. 1972;42:237288.
Available at: http://www.basic.northwestern.edu/statguidefiles/levene.html.
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 194
CHAPTER 12
MEASURES OF ASSOCIATION
Do not put your faith in what statistics say until you have carefully
considered what they do not say.
William W. Watt (as quoted in
www.quotegarden.com/statistics.html)
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand that the purpose of research can be to investigate relationships between vari-
ables rather than compare group differences.
Learn how to introduce the methods and statistical analysis of research into relationships
or correlations between variables.
Understand that the type of data dictates the appropriate statistical approach and the sta-
tistical approach affects how results are interpreted.
Learn to recognize and understand the relationships between variables.
Understand that there are limitations to the conclusions that can be appropriately drawn
from statistical results.
Learn that probability estimates can be, and often are, provided in association with point
estimates of association.
Know that the interpretation of the P-values requires some elaboration.
KEY TERMS
conclusions interpretations predictor variables
correlations between variables limitations probability estimates
criterion variable P-values results
data point estimates of association variables
194
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 195
INTRODUCTION
In many circumstances the purpose of research is to investigate relationships be-
tween variables rather than compare group differences. This chapter introduces
the methods and statistical analysis of research into the relationships or correla-
tions between variables. As with research into group comparisons, the type of
data dictates the appropriate statistical approach and the statistical approach af-
fects how results are interpreted. While understanding the relationships between
variables can be highly informative, there are limitations to the conclusions that
can be appropriately drawn. Moreover, while probability estimates can be, and
often are, provided in association with point estimates of association, the inter-
pretation of the P-values requires some elaboration. These issues are discussed in
more detail at the end of the chapter.
CONCEPT CHECK
In many circumstances the purpose of research is to investigate relationships between
variables rather than compare group differences.
196 PART II Seeking Answers: How the Question Drives the Methods
variable and any other variable unless the direction and magnitude of error is sim-
ilar for each of the variables being studied. Thus, it is possible that a strong rela-
tionship exists between two variables but the strength of the relationship is
underestimated due to measurement imprecision. Validity of data may or may not
affect the strength of the relationship between two measurements. If a measure-
ment is reliable and fairly precise, a strong association may be found; however, the
interpretation of the association may be misleading since the data from one or
more measurements may not reflect the underlying construct that the investigator
was intending to measure.
CONCEPT CHECK
When investigating association (between variables), data are collected that allow the
estimation of the strength of the relationships between variables.
CONCEPT CHECK
The researcher and research consumer must consider the timing of measurements and
the sample being studied.
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 197
In addition to, and perhaps more important, the critical appraisal of research
into the associations between variables is a complete description and careful con-
sideration of the sample tested. The more homogeneous a sample, the more likely
the strength of the association between variables will be underestimated. Take, for
example, the association between undergraduate grade point average (GPA) and
performance (graduate GPA) in a highly competitive graduate program. Since the
only students accepted into the program had a very high GPA as an undergradu-
ate (lets say 3.5 or better) there is very little variance between them (homogene-
ity). Moreover, since acceptable grades in the graduate school are A and B there is
limited variance between students in graduate school grades. In other words, the
sample has resulted in a restricted range of values from a larger scale (e.g., 0 to 4.0
GPA). An analysis of these data is likely to lead to the conclusion that there is lit-
tle association between undergraduate achievement and graduate school per-
formance. If, however, all students across the spectrum of undergraduate GPA
were enrolled in the graduate program and a full spectrum of grades (A to F)
awarded, a much stronger association between undergraduate and graduate
achievement would likely emerge. The research consumer should consider
whether the sample studied truly represents the range of the population of inter-
est and, from a clinical perspective, whether the sample is similar to patients in
their practice. Generalization of results into clinical practice is a judgment of the
clinician but the investigator/author must provide sufficient description to allow
a thoughtful decision on the part of the clinician.
CONCEPT CHECK
The more homogeneous a sample, the more likely the strength of the association be-
tween variables will be underestimated.
DATA ANALYSIS
As with the statistical methods used for comparisons between groups described
in the previous chapter, the first consideration when analyzing data to estimate
the strength of an association between variables is to determine the type of data at
hand. Recall that interval and ratio data can be analyzed with analysis of variance
(ANOVA) because variance estimates can be calculated. Variance estimates are
also central to estimating the strength of a relationship between variables that are
interval or ratio data. Ordinal and nominal data are analyzed through nonpara-
metric procedures. We will begin with the analysis and interpretation of measures
of association involving interval and ratio data and then address nonparametric
procedures.
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 198
198 PART II Seeking Answers: How the Question Drives the Methods
CONCEPT CHECK
The first consideration when analyzing data to estimate the strength of an association
between variables is to determine the type of data at hand.
REGRESSION
The fundamental statistic for the analysis of interval and ratio data to estimate the
strength of the association between variables is regression analysis, which is also
the foundation for ANOVA. In fact, ANOVA is really a special case of regression.
As with ANOVA, the more the variables under consideration, the more complex
the analysis. Similar to the analyses for comparisons between groups where t-tests
were identified as the simplest form of ANOVA and used when only pairs of data
were being considered, simple linear regression is the simplest form of regression.
Simple linear regression, often identified as a Pearson product moment correla-
tion (PPMC) and designated as a Pearson or simply an value, is used to meas-
ure the strength of the association between pairs of data that are interval or
ordinal.
CONCEPT CHECK
Simple linear regression is often identified as a Pearson product moment correlation
(PPMC) and is designated as a Pearson or simply an value.
EXAMPLE
Regression
Zhou et al. (2009) explored the relationships between obesity indices and blood
pressure in a large sample of Chinese adults. Each of the measures of obesity (e.g.,
BMI, waist circumference) and blood pressure are interval data. Thus, PPMC co-
efficients were calculated to estimate the strength of the relationships of interest.
Of course, computer software was used as there were data from more than 29,000
people. However, a small subset of the data might have generated the table below
that will allow illustration of the calculation of a PPMC.
r N XY (X )(Y )/ 2 N X2 (X )2 2 N Y 2 (Y )2
r 8(30,154) 23,8206/( 2 8(6,288) 2222) ( 2 (8)14,5959 1,0732)
r 3,026/(31.9) (127.8) 0.74
In this very small sample a strong relationship was observed between BMIs.
Zhou reported the association between BMI and systolic blood pressure to be 0.43
for men and 0.51 for women. Although fictitious our data illustrate two points.
First, small samples, even in real settings, may yield estimates well above true
population values. Second, even with a very small data set calculations are cum-
bersome and time consuming. Computer software is really essential for manag-
ing data that is informative to clinical practice; however, the interpretation of the
results continues to require an understanding of the analytical procedures used
for analysis.
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 200
200 PART II Seeking Answers: How the Question Drives the Methods
Y Y
X1 X2 X1 X2
CONCEPT CHECK
Multiple regression yields an R 2 value that can be defined as the variance in the criterion
variable (Y ) explained by a predictor variable (X ) or predictor variables (X1, X 2, etc.).
Before discussing the interpretation of and R2 lets return to our example and
consider what happen when X2 was added to the analysis. Note that the R2 when
X1 and X2 were entered is not the sum of the variance in Y explained by X1 plus the
variance in Y explained by X2, but rather the variance explained in Y by X1 plus any
additional variance in Y explained by the addition of X2 as illustrated below. The
only exception to this rule is when there is no association between X1 and X2 in
which case both variables explain completely unique portions of variance in Y.
It is possible to explore the relationships between variables while controlling
for the influence of other variables through the calculation of partial and semipar-
tial correlations. While beyond the scope of this discussion you may refer to
Pedhazur for a detailed explanation.
EXAMPLE
1 10.8 18 72
2 9.5 27 64
3 11.4 23 85
(continued)
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 202
202 PART II Seeking Answers: How the Question Drives the Methods
4 9.8 34 70
5 7.8 38 50
6 10.2 20 88
7 8.0 25 75
8 12.0 15 90
9 8.6 17 77
10 13.0 12 95
11 8.8 14 88
12 9.0 17 80
13 8.2 28 60
14 10.5 17 75
Data were analyzed through regression analysis and the calculation of PPMCs between individual variables
with SPSS Version 16.0 software.
Model Summary
Analysis of Varianceb
Total 2071.214 13
a
Predictors: (Constant), Stair climbing (sec), Strength (N/kg).
b
Dependent variable: Community mobility assessment.
Coefficientsa
Multiple regression analysis revealed a statistically significant (very low proba-
bility that a relationship 0) and strong association between strength, stair
climbing time, and self-reported community mobility with approximately 75% of
variance in community mobility explained by these two predictor variables. This
(continued)
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 203
finding does not identify the strength of the individual contributions to commu-
nity ambulation. Examining the correlation between the individual factors yields
further clarity.
N 14 14
N 14 14
a
Correlation is significant at the 0.01 level (two-tailed).
The analysis reveals a strong correlation between stair climbing ability and
self-reported community ambulation. The correlation of 0.826 reveals that as
time to climb stairs decreases, self-reported community mobility increases (hence
the negative value), and the stair climbing ability explains 68% (0.8262) of the
variance in community ambulation. The results of the regression analysis and this
finding mean that strength only added an additional 7% of explained variance in
community mobility scores when combined with stair climbing ability and there is
not 95% certainty that strength added any explained variance. (Note t 1.796, P
0.100.) This means that the addition of strength data to the model did not signifi-
cantly (null 0) improve the explanation of variance in community mobility.
Coefficientsa
UNSTANDARDIZED STANDARDIZED
MODEL COEFFICIENTS COEFFICIENTS T SIG.
Std. error
204 PART II Seeking Answers: How the Question Drives the Methods
COMMUNITY MOBILITY
STRENGTH (N/KG) ASSESSMENT
N 14 14
N 14 14
a
Correlation is significant at the 0.01 level (two-tailed).
The answer lies in the strong correlation between the predictor variables. The
analysis reveals a fairly strong link ( 0.533) between leg press strength and
stair climbing time, and examination of the scores confirms greater strength is as-
sociated with faster climbing.
STAIR CLIMBING
STRENGTH (N/KG) (SEC)
N 14 14
N 14 14
a
Correlation is significant at the 0.05 level (two-tailed).
(continued)
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 205
The assessment of these associations does not imply cause and effect.
However, a logical argument could be made for exercises to promote strength
and stair climbing repetitions in an effort to improve reported community mobil-
ity. This would provide the foundation for clinical research into these questions.
INTERPRETING AND R2
The first consideration when interpreting estimates of association has yet to be in-
troduced. P-values or probability values were discussed in the preceding chapter
in the context of hypothesis testing. When exploring differences between groups the
P-value is an indication that the differences observed when studying a sample are
reflective of true population differences. P-values relate to statistical differences but
as noted in the preceding chapter offer little to understanding clinical meaningful-
ness. The P-values found in association with and R2 also relate to hypothesis test-
ing and also generally are of little value when considering the meaningfulness of an
association between variables. While it is possible to test the hypothesis that the as-
sociation between variables exceeds any value (e.g., 0.50), by far the most com-
mon use tests the hypothesis that the association between X and Y does not equal 0
(rXY 0). Thus, the fact that a significant relationship has been identified is
nearly useless when considering the meaningfulness of the analysis.
CONCEPT CHECK
When exploring differences between groups the P-value is an indication that the differ-
ences observed when studying a sample are reflective of true population differences.
If a P-value is not informative, then what statistic should the research consumer
be most concerned with? The answer is the R2 value (with a confidence interval) or
the portion of variance in Y explained by the predictor variable(s). This is also why
it is important to recognize the relationship between PPMC and multiple regression.
The usefulness of the R2 value is revealed when one considers interpretation of
values although interpretation of R2 is far from straightforward. Investigators and
statisticians have provided some guidelines for interpreting values (Note: table
values are for interpreting , not -square) such as:
(continued)
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 206
206 PART II Seeking Answers: How the Question Drives the Methods
Domholdt E. Rehabilitation Research: Principles and Applications. 3rd ed. St. Louis, MO: Elsevier Saunders; 2005.
However, the application of these guidelines may prove misleading. Lets as-
sume, for example, that investigators found that the correlation between BMI and
serum cholesterol in men over 40 years of age was 0.49. Such a finding could
be interpreted as a low correlation and therefore unimportant. However, many
factors may influence serum cholesterol and an R2 0.24 (0.492) indicates that 24%
of the variance in serum cholesterol is explained by BMI. Such a finding might
suggest a strategy for reducing serum cholesterol and thus quite important.
CONCEPT CHECK
The usefulness of the R 2 value is revealed when one considers interpretation of values.
There are two important points to be made here. The first is that measures of
association do not imply cause and effect. It is possible that reducing BMI would
have no effect on serum cholesterol. Additional clinical research would be needed
to determine if lowering BMI would affect serum cholesterol. If the issue of cause
and effect is still puzzling since it seems reasonable that lowering BMI would be
associated with reduced serum cholesterol across a sample of older men, consider
the converse. It would be unlikely that lowering serum cholesterol through med-
ication would impact BMI. The issue of cause and effect will be discussed further
near the end of the chapter but the introduction here leads to a second key point;
investigators and research consumers really need to understand the nature of the
data at hand to interpret measures of association and comparison. As noted in the
previous chapter statistically significant differences may be of little clinical mean-
ing and thus misinterpreted. Interpreting the strength and importance of meas-
ures of association and teasing out cause and effect relationships requires
knowledge of the subject matter and careful critical appraisal rather than reference
to a published scale that, while well intended and applicable in some circum-
stances, can potentially mislead.
CONCEPT CHECK
Measures of association do not imply cause and effect.
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 207
The second point to be made is that while measures of association cannot con-
firm cause and effect, these analyses can build useful predictive models. Recall that
in the discussion of multiple regression the terms predictor and criterion variables
were introduced. These terms imply that the values of the predictor variables, in
fact, predict the value of the criterion variable. Take, for example, the use of skin-
fold measures to estimate body composition. These formulas are derived from re-
gression analyses where the variance in body composition as measured from a
standard (e.g., underwater weighing) explained by skin-fold measures was identi-
fied. Since a large portion of the variance in the criterion variable body composi-
tion (as measured through underwater weighing) was explained by the predictor
variables (skin-folds) the use of skin-fold measures is an acceptable (although not
perfect) and more easily performed means of estimating body composition. Note
that the prediction model does not imply cause and effect, but rather a means of
predicting an outcome in one measure, based on the values of predictor measures.
CONCEPT CHECK
While measures of association cannot confirm cause and effect, these analyses can
build useful predictive models. Note that the prediction model does not imply cause
and effect, but rather a means of predicting an outcome in one measure, based on
the values of predictor measures.
EXAMPLE
Calculating a Spearman
A clinician wants to learn about the association between perceived pain and
changes in systolic blood pressure in patients suffering from tension headaches. To
investigate the question a small pilot study is conducted where blood pressure is
recorded on three occasions when patients are free of pain. Patients are then in-
structed to contact the investigators when a headache develops and blood pressure
(continued)
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 208
208 PART II Seeking Answers: How the Question Drives the Methods
is again recorded and perceived pain is recorded on a 10-cm visual analog scale.
Since the pain measures are deemed to be ordinal data the investigator selects a
nonparameter Spearman rank order correlation for data analysis.
The formula for a Spearman rank order correlation is as follows:
Spearman ( ) 1 ((6d2)/n(n2 1))
where d is the difference in ranks between pairs of data and n is the number
of subjects.
The pilot study yielded the following data set:
CHANGE IN SYSTOLIC
SUBJECT BLOOD PRESSURE RANK PAIN SCORE RANK d d2
1 10 47 5 1 1
2 5 1 5 1 0 0
3 15 6 5.5 2 4 16
4 0 2 9 7 5 25
5 5 3 6 3 0 0
6 20 7 8 6 1 1
7 12 5 6.5 4 1 1
Sum 44
The pilot study suggests there is a small relationship between changes in sys-
tolic blood pressure and perceived pain with a tension headache. Such a result
may cause investigators to question whether the expense of conducting a larger
clinical trial can be justified.
The result is similar to that of a PPMC, with values ranging from 1.0 to 1.0,
negative and positive values reflecting the direction of the relationship.
The analysis of the association between ordinal values is more complicated
than those for ordinal, interval, or ratio data. The most common approach is a
Cramers V correlation that is based on chi square (x2) values. One application of
an x2 analysis is to test the hypothesis that two nominal variables are independ-
ent. The text by Daniel (see section Suggested Reading at the end of this chap-
ter) provides a detailed discussion of x2 including tests of independence. If the
null hypothesis that two variables are independent is rejected, a Cramers V can
be calculated to measure the strength of the relationship.
The following example illustrates this process.
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 209
EXAMPLE
Calculating a Cramers V
An investigator sought to determine if there is a relationship between living envi-
ronment and activity level. Participants in the study were classified as living in a
rural, suburban, or urban environment based on population density and classified
as being sedentary, moderately active, or active based on scores on an
International Physical Activity Questionnaire Short Form (IPAQ-SF). Note that the
data related to the community in which individuals lived is nominal while the ac-
tivity level data is ordinal. As with other analyses the appropriate statistic is based
on the lowest level of data. Thus a x2 analysis was performed on the data set
found below to determine if activity level is independent of where people live.
The sample consisted of 800 urban dwellers, 845 suburban dwellers, and 680
rural dwellers with activity profiles described below.
Thus the hypothesis that living environment and activity level are independ-
ent is rejected. What is the strength of the association between living environment
and activity level observed in this sample? The Cramers V is calculated as follows:
x2
V where n is the sample size and t is the smallest number of rows or
B nt
columns minus 1. In this case there are 2325 participants and 3 rows and columns,
so t 2.
The result is Cramers V 19.4. One might conclude that there is a rather
modest association between living environment as defined for the study and ac-
tivity level measured in the study with urban and rural dwelling people less likely
to be sedentary.
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 210
210 PART II Seeking Answers: How the Question Drives the Methods
CONCEPT CHECK
In some cases two variables may appear highly associated due to correlation with a
third variable.
CONCEPT CHECK
It requires more than measures of association to establish cause and effect.
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 211
CHAPTER SUMMARY
This chapter provides an introduction to the research and statistical methods
used to investigate the association between variables. Consideration for critically
appraising and generalizing research that provides estimates of association as
well as the most common analyses are reviewed. Although the calculation of es-
timates of association is fairly easy, interpretation of estimates of association for
clinical applications is often less straightforward. Some investigators have relied
on P-values to make a case for the clinical meaningfulness of their findings210
ather than considering the magnitude of an observed association in the context
of multifactorial relationships. In some instances inferences of cause and effect
have evolved from studies of association. The knowledge of the associations be-
tween variables can inform clinical decisions and generate new research ques-
tions. The consumer-clinician, however, must consider the magnitude and
clinical meaningfulness of estimates of association in the context of the complex-
ity of interrelated variables rather than accepting, at face value, that reported re-
lationships are of importance.
KEY POINTS
Knowledge of the association between variables can inform health care
practice.
The less precise a measurement is, the lower the association between that
variable and any other variable unless the direction and magnitude of error
is similar for each of the variables being studied.
Validity of data may or may not affect the strength of the relationship be-
tween two measurements.
Generalization of results into clinical practice is a judgment of the clinician
and requires thoughtful decision on the part of the clinician.
The fundamental statistic for the analysis of interval and ratio data to esti-
mate the strength of the association between variables is regression analysis.
LibraryPirate
97689_c12 01/07/10 4:07 PM Page 212
212 PART II Seeking Answers: How the Question Drives the Methods
Applying Concepts
1. Discuss considerations related to the research methods investigating
association.
2. Consider why the answer to question #2 above is important to meas-
ures of association; and, provide a clinical example to illustrate and
support your opinion.
REFERENCES
Cohen P, Cohen J, West SG, et al. Applied Multiple Regression/Correlation Analysis for the
Behavioral Sciences. 3rd ed. Marwah, NJ: Lawrence Erlbaum; 2002.
Domholdt E. Rehabilitation Research: Principles and Applications. 3rd ed. St. Louis, MO:
Elsevier Saunders; 2005.
Pedhazur EJ. Multiple Regression in Behavior Research. New York: Holt, Rinehart and Winston;
1997.
Zhou Z, Hu D, Chen J. Association between obesity indices and blood pressure or hyper-
tension: which index is the best? Public Health Nutr. 2009;12(8):10611071.
SUGGESTED READING
1. Daniel WW. Biostatistics: A Foundation for Analysis in the Health Sciences. 9th ed.
New York: John Wiley & Sons; 2008.
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 213
PART III
CLINICAL RESEARCH:
DIAGNOSIS,
PREVENTION, AND
TREATMENT
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 214
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 215
CHAPTER 13
EVALUATION AND
DIAGNOSIS: RESEARCH
METHODS AND DATA
ANALYSIS
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand and explain the purpose of patient interview and observation.
Be able to explain and describe the importance of physical examination procedures.
Be able to explain why observations are a key part in the evaluation process.
Be able to explain the purpose and process of clinical research.
Be able to define and decrease investigational bias.
Understand the concepts of sensitivity and specificity.
Be able to calculate sensitivity and specificity values.
Understand and calculate likelihood ratios.
Understand positive and negative prediction values.
Understand and interpret receiver operator characteristic (ROC) curves.
215
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 216
KEY TERMS
blinding false positive positive and negative
prediction values
clinical trials generalizability of research findings special tests
control biases methodologic flaws therapeutic intervention
diagnostic continuum methodologic quality in research validity
estimates of sensitivity nonclinical research
and specificity
false negative results positive and negative likelihood
ratios or receiver operator
characteristic curves
INTRODUCTION
The ability to identify tissue damage or disease that is responsible for a patient's
symptoms is the foundation of therapeutic intervention. The clinician needs to
know what is wrong before it can be made right. Health care providers use a vari-
ety of diagnostic procedures including laboratory testing, diagnostic imaging, phys-
ical examination, interviewing, and observation. Decisions regarding return to work
or sports participation, referral for emergent and nonemergent care, as well as the
development of rehabilitation plans of care, stem from the evaluation process.
Patient interviews and the performance of physical examination procedures,
often referred to as special tests, are the primary tools available to the physical
therapists, occupational therapists, and athletic trainers evaluating an individual
seeking care. Unfortunately, not everyone with a particular condition will report
the same symptoms. Moreover, many of the physical examination procedures
used by these clinicians may fail to detect the condition that the procedure is in-
tended to identify or render positive findings in the absence of the condition.
The limitations related to diagnostic testing are not exclusive to the patient in-
terview and physical examination procedures. Diagnostic imaging and laboratory
procedures may also lead to incorrect conclusions. For example, magnetic reso-
nance imaging (MRI) without contrast may fail to detect glenoid labrum pathol-
ogy (Volpi et al., 2003) or identify damage to the intervertebral discs unrelated to
the pain generator in a patient complaining of low back pain (Ernst et al., 2005;
Grane, 1998).
The limitations associated with diagnostic testing should not dissuade the cli-
nician from performing these assessments. Despite the inherent flaws most diag-
nostic tests do provide the clinician with useful information. What is needed is an
understanding of how well the diagnostic procedures employed perform in rou-
tine practice so that the results can be interpreted in the context of other informa-
tion and examination results.
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 217
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 217
CONCEPT CHECK
The purpose of clinical observation is to guide the opening questions in an interview
with a patient. Through observation and interview the list of diagnostic possibilities
shrinks and one or two conditions emerge as the most likely culprits of the patients
complaints.
The clinician can then select physical examination procedures that will help
confirm or refute the existence of the suspect conditions. This process then guides
decisions related to returning to work or sport participation making a referral to a
physician or, if indicated, an other care provider. In many cases the ultimate diag-
nosis has been established prior to referral only to be confirmed through re-
evaluation and additional diagnostic testing.
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 218
Advances in health care have greatly expanded the diagnostic tools available
to physicians and other providers. MRI, for example, was not available 40 years
ago. Such technology offers greater diagnostic accuracy in many situations but at
a greater cost. MRI has not replaced observation, interview, and physical exami-
nation. In fact these assessment techniques are necessary to identify patients likely
to benefit from MRI. The Ottawa Ankle Rules (Stiell et al., 1992) provide an excel-
lent example of observation and physical examination guiding the decision as to
whether a radiographic assessment of the injured ankle is warranted. Research re-
veals that when the Ottawa Ankle Rules are applied and radiographs are not war-
ranted the risk of not identifying a fracture is very low (Bachmann et al., 2003).
Through the application of the Ottawa Rules unwarranted exposure to radiation
and the costs associated with radiographic studies are reduced.
Unfortunately, not all assessment techniques perform as well as the Ottawa
Ankle Rules nor have most been as extensively studied. There are, however, a lot
of reports related to diagnostic testing that are of importance to the clinician.
Research into the performance of diagnostic tests differs from research into, for ex-
ample, prevention efforts and treatment outcomes. The results of research into the
outcomes of treatments and prevention efforts are far more convincing when the
responses to particular intervention are compared with the outcome following an-
other treatment, a placebo treatment, or a no treatment control. The key point is
that the outcomes in a group of subjects receiving a treatment of interest need to
be compared to some other group. The details of methods of controlling investi-
gational bias in these types of studies are presented in preceding chapters. In stud-
ies of diagnostic procedures comparison to another group of patients or subjects
is unnecessary. As we shall see these studies require that the results of the diag-
nostic test of interest are compared to the results of an established standard.
Investigational methods, and therefore sources of investigational bias, differ sub-
stantially and are the subject of this chapter.
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 219
The process of searching for clinical studies and an overview of clinical re-
search was provided in earlier chapters. It will take only a few attempts at re-
viewing the literature on specific topics to find disagreement in the conclusions
drawn by researchers based upon the data they have collected and analyzed.
Sorting through clinical research to find truth is a real challenge. Differing re-
sults can be due to differences in sample sizes between studies. Large samples
are more likely to reflect the true state of a population than smaller samples.
Differences may result from subtle variations in the populations from which
samples were drawn. Differences may even result from chance where, through
random selection, an investigator enrolls individuals who are more or less
likely to benefit from a particular procedure. Lastly, differences in study out-
comes may be due to a failure to control biases that threaten the validity of
data.
The clinicianconsumer of clinical research must understand the relation-
ship between study methods and threats to the validity of data. The methods of
studies of prevention, treatment, and diagnostic procedures, and therefore the
steps that minimize investigational bias, differ. Issues related to diagnostic stud-
ies are discussed here, while those related to prevention and treatment are found
in Chapters 1416 of Part III.
EXAMPLE
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 221
Two issues, at a minimum, emerge from this example. First, how much confi-
dence should the examiner place in the results of the assessment of the ligamen-
tous integrity of the knee and how should one proceed if multiple diagnoses
remain highly plausible? The first issue leads to the central issue in this chapter.
Studies of diagnostic procedures are needed to guide clinicians as they ponder the
certainty of their diagnosis. The less certain one is of a particular diagnosis the
greater the likelihood that something else is wrong. When this is the case the use
of advanced medical technologies may be warranted. The use of advanced diag-
nostic procedures (e.g., MRI) comes at a cost and in some cases (e.g., spinal tap to
work up a patient who may have meningitis) with a risk. The more that is known
about how well a structured interview, observation, and when indicated, physical
examination procedures perform in establishing specific diagnoses, the more se-
lective the use of higher cost and riskier procedures can become. Thus, studies of
the diagnostic accuracy of responses to specific questions and the finding of stud-
ies of diagnostic procedures are important to the advancement of health care.
CONCEPT CHECK
An effectively structured physical examination will substantially increase or decrease
the clinicians suspicion regarding the presence of one or more specific condition.
EXAMPLE
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 223
Reprinted with permission from Whiting P, et al. BMC Med Res Meth. 2003;3:25.
It is also important that all subjects of a research study into diagnostic tests re-
ceive all testing (items 5 to 11). If only those believed to have rotator cuff pathol-
ogy, perhaps based on a drop arm test, are evaluated by MRI. The true number of
false negative results will be reduced due to the study methods employed. Thus,
the data are biased toward underreporting of false negative results. This form of
bias is referred to as work-up bias. It is also important that the diagnostic test is
studied and the reference test is administered in a manner that it is unlikely the
condition changes between examinations (item 4).
Lastly, it is important to note that some tests may not yield useful data and
not all subjects can complete all testing (items 11 and 12). Those that have
struggled with integrating data acquisition systems such as electromyography,
motion analysis, and force plates know that some trials do not yield useful
data. Furthermore, those who suffer from claustrophobia may not tolerate
being confined in an MRI unit. While generally less critical to making deci-
sions about the use of a diagnostic procedure in ones practice, these issues be-
come important when repeated testing may be necessary and to help identify
those patients for whom a diagnostic test may be contraindicated or poorly
tolerated.
CONCEPT CHECK
To decrease investigational bias, blinding of investigators performing, or interpreting,
the results of the diagnostic test being investigated as well as those responsible for the
results of gold standard assessment to each others finding is essential.
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 225
foundation and require nothing more than simple multiplication and division.
The second is to emphasize that the values derived are estimates of test perform-
ance. It is not possible to study the administration of a diagnostic test in all pa-
tients for whom such testing would be appropriate. Thus, researchers study
samples from a population and the diagnostic testing is typically completed by
one or a few clinicians. Thus, one can search the literature and find varying esti-
mates of test performance. Larger samples are more likely to offer more accurate
estimates and in some cases there are a sufficient number of studies to permit a
meta-analysis (described in Chapter 18) that will yield even more precise esti-
mates of test performance.
The foundation of the above noted statistical values was actually presented in
the previous chapter. When one compares the results from a diagnostic test with
a dichotomous result (positive or negative) of interest to the results of a gold
standard diagnostic test, four subgroups or cells are formed. Table 13-2 illustrates
the four cells. The values included will be used in an upcoming example so that
the calculation process can be followed. From such a table sensitivity and speci-
ficity, positive and negative prediction values, and positive and negative likeli-
hood ratios can be calculated. An expansion of the table vertically is needed to
generate receiver operator characteristic curves; a topic reserved for the end of this
chapter.
POSITIVE NEGATIVE
(CONDITION PRESENT) (CONDITION ABSENT) TOTAL
Total 23 15 35
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 227
condition of interest. Conversely, tests with high specificity will have few false
positives; thus, a positive result using a diagnostic test with high specificity has
identified the target disorder.
The calculation and interpretation of sensitivity and specificity are predi-
cated on the effort to identify individuals with and without a target disorder of
interest. These values are calculated vertically using cells A and C or B and
D, respectively. Positive and negative prediction values offer a different ap-
proach to assessing the usefulness of diagnostic tests. A positive prediction
value addresses the question: if a diagnostic test is positive, what is the proba-
bility the target condition is present? Conversely, a negative prediction value
provides an estimate that the target condition is not present when the diagnos-
tic test is negative.
Since the positive and negative prediction values address the meaning of a
positive or negative exam finding rather than the presence or absence of the tar-
get condition the cells used to calculate these values differ. Positive predictive val-
ues are calculated by:
A
PPV
AB
Using the data from Table 13-2, PPV 18/18 3 0.86. Thus, when an an-
terior draw test is positive there is an 86% chance that the ACL is torn.
Negative prediction values are calculated by:
D
NPV
DC
Again using the data in Table 13-2 and referring to the previous example,
NPV 12/12 5 0.70. Thus, a negative anterior drawer test may occur in 30%
of individuals with torn ACL.
A few questions and issues are starting to arise. First, it is important to point
out that the data used in this example is fictitious. The sensitivity and specificity
of the anterior drawer test have been estimated to be 0.62 and 0.88, respectively.
Second, and more importantly, sensitivity and specificity are calculated differ-
ently than PPV and NPV. Which values are most useful to the clinician?
The answers to this question are found when one considers the prevalence of
the target disorder and the application of the values into the examination of indi-
vidual patients. Sackett et al. (1992) illustrated the relationship between preva-
lence and PPV and NPV. The PPV associated with a condition with lower
prevalence will be less than the PPV of a condition with a higher prevalence. The
converse is true regarding NPV: a condition with a lower prevalence will yield a
higher NPV than the one with a higher prevalence. In summary as prevalence
falls, PPV also fall and NPV rise (Sackett et al., 1992).
In the example above 23 of 35 patients entering the study had torn their ACL.
This represents 66% of the enrolled patients, a rather high prevalence. However, if
the selection criteria were athletic individuals who experienced an acute, disabling
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 228
POSITIVE NEGATIVE
(CONDITION PRESENT) (CONDITION ABSENT) TOTAL
knee injury while playing contact sports and presented with hemarthrosis, such a
prevalence rate might not be surprising. To examine the effect of prevalence on the
values we have calculated lets assume a prevalence rate of 6.6% instead of 66%.
The values in cells A and C can remain the same, 23 ACL injured patients are in-
cluded in the study. The low prevalence rate translated into many more (10 times
as many in this example) patients without ACL injury. If the rate of false positive
findings remains the same, a reasonable assumption, then the table would appear
as Table 13-3.
The calculation of sensitivity and specificity yield the same values as
before because the values are calculated vertically. (Sensitivity 18/23 0.78,
Specificity 120/150 0.80.) However, the PPV 18/(18 30) 0.47 has fallen
dramatically; NPV 120/(120 5) 0.96 has risen. It is difficult to interpret PPV
and NPV unless the prevalence of the target condition is known or to compare val-
ues from studies with differing prevalence rates. Sensitivity and specificity values
are relatively stable across a spectrum of prevalence values, one reason that they
are of more use to the clinician.
While more stable estimates than PPV and NPV, values related to sensitivity
and specificity are still difficult to apply in clinical practice. What does a sensitiv-
ity estimate really mean when a clinician examines a patient with an injured knee?
The answer unfortunately is, not too much. Now you are left wondering why you
have spent the time to read this far. PPV and NPV estimates are unstable across
varying prevalence rate and sensitivity and specificity are difficult to apply. Be
patient, help is on the way in the form of likelihood ratios.
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 229
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 229
CONCEPT CHECK
Sensitivity and specificity are related to the ability of a test to identify those with and
without a condition and are needed to calculate likelihood ratios.
LIKELIHOOD RATIOS
Likelihood ratios (LRs) are values derived from estimates of sensitivity and speci-
ficity. Like sensitivity and specificity LRs offer insight into those diagnostic proce-
dures that are generally more and less useful. However, unlike sensitivity and
specificity LRs can be applied by the clinician to the examination of individual pa-
tients. Knowledge of the LRs associated with the diagnostic procedures a clinician
employs will influence the level of certainty that a condition does or doesnt exist
at the end of the examination.
A positive likelihood (LR) ratio is indicative of the impact of a positive ex-
amination finding on the probability that the target condition exists. For tests with
dichotomous results a LR is calculated as follows:
Sensitivity
()LR
1 Specificity
Using the examples above the LR would equal 0.90/(1 0.8) 3.9 (95%
CI 1.4 11.0). This means that, based upon our hypothetical numbers, a posi-
tive anterior drawer is 3.9 times more likely to occur in a patient with a torn ante-
rior cruciate than the one with an intact ligament. Further applications of LRs in
the context of clinical practice will be developed later.
A negative LR addresses the impact of a negative examination on the proba-
bility that the condition in question is present. A negative result of a diagnostic test
with a small likelihood ratio suggests that the chance that the target condition ex-
ists is very low. Negative likelihood is calculated as follows:
1 Sensitivity
()LR
Specificity
Again using the values from the preceding tables the negative likelihood
ratio would equal (1 0.90)/0.8 0.28 (95% CI 0.12 0.64). In this test the ex-
aminer would find a positive anterior drawer 28/100th as often in uninjured
knees as compared to injured knees. Jaeschke et al. (1994) summarized likelihood
ratios (positive and negative) into broader categories of clinical value as follows:
CONCEPT CHECK
A positive likelihood (LR) ratio is indicative of the impact of a positive examination
finding on the probability that the target condition exists. A negative likelihood ratio
addresses the impact of a negative examination on the probability that the condition
in question is present.
Applications
The diagnostic test used in the examples above, like all other diagnostic procedures,
is not conducted in isolation. Diagnosis is a process during which information is
gathered to narrow the range of diagnostic possibilities. A carefully conducted inter-
view and observation form the foundation of a physical examination. As the exami-
nation proceeds, a level of suspicion that a narrowing list of diagnostic possibilities
is developed. Certainly one is more reporting an acute injury and presenting on
crutches with a very swollen knee has torn the ACL than the one who reports a grad-
ual onset of anterior knee pain that is made worse by prolonged sitting. Once the cli-
nician has narrowed the diagnostic possibilities consideration is turned to the
physical examination and other diagnostic assessment that may confirm or rule out
specific diagnoses. Thus, the clinician has some level of concern that a specific con-
dition exists before proceeding with additional testing.
This pretest level of suspicion can be quantified as pretest probability. Pretest
probability values will vary between clinicians and the circumstances of the indi-
vidual patient. The key is to recognize that a level of suspicion regarding diagnos-
tic possibilities exists before the examination procedure of interest (e.g., anterior
test) is performed. The results of diagnostic tests will change the level of suspicion
of a diagnosis, but how much?
Fritz and Wainner (2001) described the relationship between pretest probabil-
ity and likelihood ratios. Understanding this link is prerequisite to understanding
the impact of diagnostic test results on posttest probability or the degree of cer-
tainty a condition does, or does not exist, after a clinical examination is completed.
Before demonstrating the impact of diagnostic tests with higher and lower
LRs, stop and consider probability. Probability implies uncertainty. Consider how
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 231
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 231
often you are absolutely, positively, certain of a diagnosis at the end of a physical
examination. These events happen, of course, but not as often as most of us would
like. Uncertainty is inherent in the clinical practice of athletic training. However,
decisions regarding referral, plans of treatment, and a physicians use of addi-
tional diagnostic studies revolve around the level of certainty (probability) that a
condition does or does not exist. The value of specific examination procedures
may be best viewed in the context of their impact upon patient care decisions
based on a positive or negative result.
EXAMPLE
alternative answers. The second option is to pursue additional testing. Lastly, the
clinician can accept the diagnosis of the target disorder and address treatment op-
tions with the patient. The first option is selected when the likelihood of the target
disorder is very low. Take for example an athlete complaining of ankle pain follow-
ing an inversion mechanism injury. The athlete is able to weight bear into the clinic
although there is a noticeable limp. Upon palpation there is no tenderness in the
distal 6 cm of the lateral malleolus, nor tenderness over the tarsal navicular or base
of the fifth metatarsal. The literature suggests that the prevalence of fractures fol-
lowing inversion ankle injuries is about 15%. The LR of the Ottawa Ankle Rules
has been estimated at 0.08. If the 15% prevalence is used as the pretest odds
(0.015/1 0.015 0.176) the odds following application of the Ottawa Ankle
Rules 0.176 0.08 0.014. Thus, there is only a 1.4% (posttest probability
0.014/1.014) chance of a fracture of clinical significance. In this case the clinician
abandons the target disorder of fracture as the problem and seeks alternative diagnoses.
At the other end of the spectrum lets consider a truck driver who landed
awkwardly when he slipped while exiting the truck cab. He states his knee
twisted and he heard a loud pop. On initial evaluation a Lachman test was posi-
tive and a large hemarthrosis developed within a few hours. The history and the
development of the hemarthrosis causes the clinician to estimate a 70% probabil-
ity (pretest odds 0.7/1 0.7 2.33) of a tear of the anterior cruciate. Using a LR
of 9.6 for the Lachman Test, posttest odds 2.33 9.6 22.4. Thus, the posttest
probability 22.4/23.4 96%. While further diagnostic testing may be warranted
to identify collateral damage to the menisci and other structures, a plan of care
should be developed for this ACL deficient knee.
The preceding examples lie at the comfortable end of the diagnostic spectrum,
first because of the high degree of certainty that exists and second because the con-
sequences of the rare missed diagnoses (a treatable fracture of the ankle and an in-
tact ACL) are likely fairly minor. A treatable ankle fracture will likely be detected
upon follow-up and the intact ACL will likely be confirmed before beginning a
surgical procedure. How does the clinician manage the patient in the middle
where after a diagnostic work-up a fair degree of uncertainty remains?
This is a difficult decision. To begin with what is a fair degree of uncer-
tainty? In general the more serious the consequences of being wrong the broader
our definition of uncertainty. At this point the clinician really has only two op-
tions, perform additional testing or refer to a provider better prepared to evaluate
the patient. Additional testing (e.g., MRI) is indicated when a fair degree of
uncertainty exists. Such additional assessment should be taken in the context of
the diagnostic process. For example, the posttest probability of 75%, following a
Lachman test, is really the pretest probability for MRI. The LR for MRI assessment
of anterior cruciate tears has been estimated at 21.5 (Vaz et al., 2005). These values
(pretest odds 0.75/1 0.75 3) yield posttest odds of 64.5/1 or a probability
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 233
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 233
of 98%. Once again the probability of an ACL tear leaves the clinician in a position
to accept the diagnosis and move on to discussion of treatment option. Before we
move forward, however, please appreciate even with advanced technology such
as MRI complete certainty is not achieved.
CONCEPT CHECK
Pretest probability, LR and test result will yield a posttest probability. Posttets probability
can be quickly estimated with a nomogram.
NOTE: Refer to suggested readings for example illustrations of the nomogram.
EXAMPLE
ROCcs
To illustrate the process a fictitious data set (Table 13-4) has been developed to
consider. Preseason evaluation of 200 college pitchers reveals the following about
the differences in internal rotation between the dominant and nondominant arms.
(continued)
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 234
45 5 degrees 11% 5
The ROCc (Table 13-5) is developed from the sensitivity and specificity esti-
mates at each level by creating five A, B, C and, D plots. Plotting sensitivity and
specificity reveals the associated ROCc Figure 13-1. In general the point nearest
the upper left corner northwest is best identifies the best balance between sen-
sitivity and specificity. In this case intervening to improve internal shoulder rota-
tion in players with greater than 10 degree side-to-side differences offers the best
opportunity to reduce risk (LR 2.3 for 10 degrees or more loss). Those with
less than a 10 degree side-to-side difference are lower risk (LR 1.4) which may
allow for attention to be directed at those with high risk.
I O
< 5
510
10.115
Sensitivity
15.120
> 20
O
Specificity
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 235
PRESENT ABSENT
15.120 degrees 27 13
10.115 degrees 27 28
510 degrees 7 33
5 degrees 5 40
20 degrees 18 a 2c
66 b 114 d
Although less commonly cited in the clinical literature than measures of speci-
ficity, sensitivity, or likelihood ratios, ROCcs allow for the identification of critical
points along continuous measures to guide clinical practice. ROCcs can also be ap-
plied to the analysis of clusters of diagnostic or prognostic criteria. These applica-
tions were discussed in detail in previous chapters in Part II.
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 236
CHAPTER SUMMARY
The principal purpose of this chapter was to provide the reader with an under-
standing of the research methods and statistical techniques used to assess how
well diagnostic tests perform in clinical practice. Investigators need to strive to im-
plement research methods that minimize the risk of biasing their data. Consumers
of the research literature must be able to assess the quality of the research meth-
ods employed in the studies they read to permit critical analysis prior to applying
research findings to their clinical practices.
There are multiple measures of diagnostic performance including sensitivity
and specificity, positive and negative prediction values, positive and negative like-
lihood ratios, and receiver operator characteristic curves. For the reasons discussed
in the chapter likelihood ratios provide the most stable and clinically useful esti-
mates of a performance of tests with dichotomous outcomes. A receiver operator
characteristic curve provides clinicians information about tests that generate meas-
ures on a continuous scale such as blood pressure, range of motion, and serum en-
zyme levels. There are multiple applications of receiver operator characteristic
curves beyond investigation of diagnostic techniques. However, the basic princi-
ples of curve generation always apply and thus can be generalized across research
disciplines.
KEY POINTS
Diagnostic testing should be used as an aid instead of a crutch for a
clinician.
History and observation should help greatly in narrowing the scope of pos-
sible injuries during the evaluation process.
Diagnostic tests have limitations.
An effectively structured physical examination will substantially increase or
decrease the clinicians suspicion regarding the presences of one or more
specific conditions.
The only way to find answers to clinical questions is through clinical trials.
Special tests will not always deliver a clear-cut decision.
The amount of confidence a clinician has in assessment results determines
the likelihood of a correct diagnosis.
Clinical research must be sorted through because of all the varying results in
the data.
Comparing the clinical study with the current standard is important in
research.
Statistics are crucial to research methods because they provide numerical ev-
idence that can be ranked to form valid conclusions.
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 237
CHAPTER 13 Evaluation and Diagnosis: Research Methods and Data Analysis 237
Applying Concepts
1. Is it possible to have a positive test in someone without the target
condition?
2. How can this information be applied in my practice and teaching?
REFERENCES
Bachmann LM, Kolb E, Koller MT, et al. Accuracy of Ottawa Ankle rules to exclude frac-
tures of the ankle and mid-foot: systematic review. BMJ 2003;326:417.
Cooperman JM, Riddle DL, Rothstein JM. Reliability and validity of judgments of the in-
tegrity of the anterior cruciate ligament of the knee using the Lachman test. Phys Ther.
1990;70:225233,
Ernst CW, Stadnik TW, Peeters E, et al. Prevalence of annular tears and disc herniations on
MR images of the cervical spine in symptom free volunteers. Eur J Radiol.
2005;55:409414. Epub 2005 Jan 1.
Fritz JM, Wainner RS. Examining diagnostic tests: an evidence based perspective. Phys Ther.
2001;81:15461564.
Grane P. The postoperative lumbar spine. A radiological investigation of the lumbar spine
after discectomy using MR imaging and CT. Acta Radiol Suppl. 1998;414:123.
Jaeschke R, Guyatt JH, Sackett DL. Users guide to the medical literature, III: how to use
an article about a diagnostic test. B: what are the results and will they help me in car-
ing for my patients? The Evidence-Based Medicine Working Group. JAMA.
1994;271:703707.
Katz DL. Clinical Epidemiology and Evidence-Based Medicine. Thousand Oaks, CA: Sage
Publications; 2001:41.
Magee DJ. Orthopedic Physical Assessment. 3rd ed. Philadelphia, PA: W.B Saunders;
1997:560572.
Sackett DL, Haynes RB, Guyatt GH, et al. Clinical Epidemiology: A Basic Science for Clinical
Medicine. 2nd ed. Boston MA: Little, Brown & Co; 1992.
LibraryPirate
97689_c13 01/07/10 4:08 PM Page 238
Stiell IG, Greenberg GH, McKnight RD, et al. A study to develop clinical decision rules for
the use of radiography in acute ankle injuries. Ann Emerg Med. 1992;21:384390.
Torg JS, Conrad W, Kalen V. Clinical diagnosis of anterior cruciate ligament instability in
the athlete. Am J Sports Med. 1976;4:8493.
Vaz CE, Camargo OP, Santiago PJ, et al. Accuracy of magnetic resonance in identifying
traumatic intraarticular knee lesions. Clinics. 2005;60:445450.
Volpi D, Olivetti L, Budassi P, et al. Capsulo-labro-ligamentous lesions of the shoulder:
evaluation with MR arthrography. Radiol Med (Torino). 2003;105:162170.
Whiting P, Rutjes AW, Reitsma JB, et al. The development of QUADAS: A tool for the qual-
ity assessment of studies of diagnositc accuracy included in systematic review. BMC
Med Res Meth. 2003;3:25.
SUGGESTED READING
Myers JB, Laudner KG, Pasquale MR, et al. Glenohumeral range of motion deficits and pos-
terior shoulder tightness in throwers with pathologic internal impingement. Am J
Sports Med. 2006;34:385391.
Ruotolo C, Price E, Panchal A. Loss of total arc of motion in collegiate baseball players. J
Shoulder Elbow Surg. 2006;15:6771.
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 239
CHAPTER 14
SCREENING AND
PREVENTION OF ILLNESS
AND INJURIES: RESEARCH
METHODS AND DATA
ANALYSIS
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand the concept of injury prevention.
Be able to discuss the principles of injury risk identification and prevention research
design based on the tenets of clinical epidemiology.
Be able to explain why risk identification and injury prevention are concepts often linked
to each other.
Understand why prospective study designs yield much more robust information about
injury risk factors than retrospective study designs.
Understand that randomized-controlled trials are the hallmark of injury prevention research
design.
Explain how randomized-controlled trials provide higher levels of evidence than retro-
spective or nonrandomized designs.
239
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 240
Understand that the limitations of case control designs must be emphasized when dis-
cussing the results of case control studies.
Understand that risk factor studies and screening studies share similar characteristics.
Explain why the issue of injury definition can be more problematic in retrospective stud-
ies as opposed to prospective studies.
KEY TERMS
etiology prevalence prospective
incidence prevalence ratio relative risk
injury rate prevention retrospective
INTRODUCTION
Injury prevention is often touted as one of the hallmarks of the clinical practice of
sports medicine and occupational medicine; however, there is a remarkable dearth
of research literature in this area. Risk identification and injury prevention are con-
cepts that are often linked to each other. It is logical to assume that before injury
prevention initiatives are implemented, the factors most likely to predict specific
injuries in particular populations should first be well understood. Unfortunately,
this is not often the case as injury prevention initiatives are often initiated before
proven risks and predispositions are identified in a scientific fashion.
The principles of injury risk identification and prevention research design are
based on the tenets of clinical epidemiology. Prospective study designs yield
much more robust information about injury risk factors than retrospective study
designs. Similarly, randomized-controlled trials are the hallmark of injury preven-
tion research design providing higher levels of evidence than retrospective or
nonrandomized designs.
CHAPTER 14 Screening and Prevention of Illness and Injuries: Resarch Methods and Data Analysis 241
an injury is ideally done through prospective injury surveillance. The best exam-
ple of this is the Injury Surveillance System utilized by the National Collegiate
Athletic Association (Agel et al., 2007). In this system, certified athletic trainers
from representative schools that offer each sport report the details of all sports-
related injuries as well as the number of athletes participating in all practices in
games. The tracking of injuries as well as exposures (participation in practices or
games) allows for the estimation of not only the number of injuries incurred by
athletes participating in a particular sport but also the rate of injuries per unit of
exposure and the overall risk of participating athletes of being injured.
The second step in this paradigm is to establish the etiology and mechanisms
of the sports-related injuries. The cause of a sports injury may be traumatic or
atraumatic. In the case of traumatic injuries documenting the source of trauma is
important. For example, determining whether a particular traumatic injury is
typically caused by an athlete being struck with a ball, with an implement such
as a bat, or from contact (legal or illegal) with an opposing player will provide
considerably different potential injury prevention solutions. Identifying the
mechanism of injury is normally done through purely observational research. In
contrast, identifying risk factors that may be involved in the etiology of a specific
injury may require the collection of baseline data as an assessment of potential
risk factors followed by a prolonged period of injury surveillance. An example of
this is that high-school soccer and basketball players who have poor balance are
at an increased risk of suffering acute ankle sprains (McGuine & Keene, 2006).
The third step in the injury prevention paradigm is to introduce a preventative
measure. Care must be taken when selecting the preventative measures to be im-
plemented. The intervention should be based on the information gained in the first
two steps of the paradigm. For example, if a large number of field hockey players
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 242
were observed to be suffering eye injuries due to being struck in the face with
the ball, an intervention of mandatory eye protection may be an appropriate injury
prevention initiative to attempt. Conversely, if numerous soccer players were
suffering leg and knee injuries due to hard tackles from behind by their opponents,
an intervention of more strict enforcement of penalties for illegal tackles may be ap-
propriate. Lastly, if specific biomechanical or physiological factors have been shown
to increase injury risk (i.e., poor balance), injury prevention that addresses such
deficits (i.e., balance training) may be warranted. One of the biggest challenges in
injury prevention research is to avoid the use of the shotgun approach of injury
prevention. For example, if an experimental injury prevention program is imple-
mented that includes numerous interventions (i.e., dynamic warm-up, flexibility ex-
ercises, balance training, strength training, and plyometrics) and that program is
shown to prevent injuries, researchers and clinicians are left not knowing which as-
pect(s) of the intervention actually caused the reduction in injuries. For this reason,
the implementation of injury prevention programs that contain single intervention
strategies (i.e., flexibility exercises) is strongly encouraged.
The fourth and final step of the injury prevention paradigm is to reassess the
extent of the injury problem after the implementation of the injury prevention
intervention. If the incidence and severity of injuries has been substantially re-
duced, the permanent implementation of the intervention is likely warranted. In
contrast, if the intervention is unsuccessful, then its use should not be retained.
CONCEPT CHECK
The steps in the Bahr et al. (2002) paradigm are:
1. To establish the extent of the injury problem;
2. To establish the etiology and mechanisms of the sports-related injuries;
3. To introduce a preventative measure; and
4. To reassess the extent of the injury problem after the implementation of the injury
prevention intervention.
By repeating the steps of the paradigm, it may become apparent that particu-
lar athletes (either certain position players or those who have specific physiologi-
cal or biomechanical profiles) are, in fact, at greater risk of certain injuries and that
they would benefit most from targeted injury prevention programs.
CHAPTER 14 Screening and Prevention of Illness and Injuries: Resarch Methods and Data Analysis 243
Participants who
suffer injury of
interest
Statistical
Large number of Participants followed comparison
participants baseline for the occurrence of of baseline
tested on the measures injury of interest over a measures
of interest defined time period between
groups
Participants who
do not suffer injury
of interest
FIGURE 14-2 Flow chart of the design of a prospective study of injury risk factors.
to, adequate sample size, clear definition of injury with specific diagnostic crite-
ria, and the determination of a representative comparison group.
Prospective Designs
A prospective cohort design is the gold standard for studies of injury risk factors
(see Figure 14-2). A large group of participants who are potentially at risk for
suffering the injury of interest are baseline tested on a number of measures that
represent potential risk factors. These participants are subsequently followed over
time to determine whether or not they go on to suffer the injury of interest.
Statistical comparisons of the potential risk factors are then made between partic-
ipants who were injured and those who did not suffer injury. These results allow
for the most robust determination of injury risk factors.
Determining an appropriate sample size is a critical step in designing a
prospective study of injury risk factors. The use of injury surveillance data is very
useful in determining sample size. Please refer to the following examples.
EXAMPLE
ACL Injuries
While considerable attention is paid to anterior cruciate ligament (ACL) injuries
in female athletes participating in sports such as basketball, the incidence rate
of these injuries has been reported as 0.23 per 1000 athlete-exposures (one athlete-
(continued)
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 244
exposure equals participation in one practice or game) (Agel et al., 2007). Thus, we
could expect an ACL injury to occur in a female collegiate basketball player for
every 4000 athlete-exposures. To put this into perspective, consider that the typi-
cal NCAA division I womens basketball player participates in an average of 27
games and 89 practices per year for a total of 116 athlete-exposures per year.
Assuming that each team has 12 players in each game or practice, a team can ex-
pect to generate 1392 athlete-exposures per year. Accordingly, at least three teams
would have to be followed during a single season to expect one ACL injury.
Alternatively, the same team could be followed for 3 years with the expectation of
a single ACL injury occurring.
It becomes obvious that in order to generate a substantial number of womens
basketball players who sustain ACL injuries, a very large number of participants
would need to be baseline tested and followed for subsequent injuries. For exam-
ple, if an a priori power analysis indicated that 30 injured subjects were required,
an estimated total of 130,435 athlete-exposures would be required to obtain 30
ACL injuries at the reported rate of 0.23 injuries per 1000 athlete-exposures. With
each team having an average of 1392 athlete-exposures per season, in order to ob-
tain a total of 130,435 athlete-exposures, researchers would need to baseline test
and follow 93.7 teams in a single season if they hoped to have 30 ACL-injured par-
ticipants. Alternatively, they could test and follow 46.9 teams over 2 years, 31.2
teams over 3 years, or 18.7 teams over 5 years. These calculations are designed to
put the large scale of prospective risk factor studies into focus.
Ankle Sprains
As another example of sample size estimation for prospective risk factor studies,
lets use ankle sprains, a more commonly occurring injury in womens basketball.
The incidence rate for ankle sprains has been reported to be 1.89 per 1000 athlete-
exposures (Agel et al., 2007). Again, assume that an a priori power analysis esti-
mates that 30 participants suffering ankle sprains are necessary to adequately
power the study. An estimated total of 15,873 athlete-exposures would be required
to obtain 30 ankle sprains at the reported rate of 1.89 injuries per 1000 athlete-
exposures. With each team having an average of 1392 athlete-exposures per sea-
son, in order to obtain a total of 15,873 athlete-exposures, researchers would need
to baseline test and follow 11.4 teams in a single season if they hoped to have
30 ankle spraininjured participants. Alternatively, they could test and follow
5.7 teams over 2 years, 3.8 teams over 3 years, or 2.3 teams over 5 years.
As these examples show, the incidence rate of the injury being studied greatly
influences the estimation of the number of participants that need to be baseline
tested in prospective studies of injury risk.
With a prospective study of injury risk, the definition of the injury of interest
can be tightly controlled. For example, confirmation of injuries such as fractures
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 245
CHAPTER 14 Screening and Prevention of Illness and Injuries: Resarch Methods and Data Analysis 245
CONCEPT CHECK
The incidence rate of the injury being studied greatly influences the estimation of the
number of participants that need to be baseline tested in prospective studies of injury risk.
Screening Studies
Prospective studies for risk factors may also be described as screening studies.
In sports medicine and occupational medicine it is very common for athletes and
workers to receive a physical exam before they are allowed to participate in organ-
ized athletics or before they are hired for specific jobs. Essentially, clinicians are
trying to identify risk factors or screen these individuals for particular injuries
or diseases. Risk factor studies and screening studies share similar characteristics.
EXAMPLE
CONCEPT CHECK
Prospective studies for risk factors may also be described as screening studies.
Retrospective Designs
Studies of injury risk factors are sometimes performed using a case control design
(see Figure 14-3). Case control studies are by definition retrospective in nature and
are sometimes called ex post facto (literally, after the fact) designs. In this case,
Statistical
comparison
of measures
between
groups
Controls: Participants
Measures of interest
who have not previously
are performed
suffered injury of interest
CHAPTER 14 Screening and Prevention of Illness and Injuries: Resarch Methods and Data Analysis 247
individuals who already have suffered the injury of interest are compared to a
control group that has not suffered the injury of interest. Each case subject who
has suffered the injury of interest is assigned a matched control subject who
has not suffered the injury of interest, hence the use of the term case control
study design.
CONCEPT CHECK
Studies of injury risk factors are sometimes performed using a case control design.
Case control studies are by definition retrospective in nature and are sometimes called
ex post facto (literally, after the fact) designs.
The most obvious advantage of the case control design over the prospective
cohort design is that the number of subjects required for testing is dramatically
lower. If we again use our example of risk factors to ACL injuries among womens
collegiate basketball players and assume that our power analysis provides a rec-
ommended sample size estimate of 30 subjects per group, we find that in order to
obtain our 30 ACL-injured players we simply have to recruit 30 female collegiate
basketball players who have previously injured their ACL and 30 matched con-
trols who have not injured their ACL; this is in comparison to the prospect of hav-
ing to baseline test and follow all of the participants of over 90 womens basketball
players over a single season in order to generate 30 ACL-injured participants
using the prospective design. The retrospective design does, however, have a se-
rious limitation in comparison to the prospective design. Because the measures of
risk factors are being taken after the injury of interest has occurred, there is no
way to know if these measures were present before the injury occurred (truly
being a risk factor) or if the current measure is a result of being an adaptation to
the injury that developed after the onset of injury. The importance of this limita-
tion has led some to recommend that the term injury risk factor should never be
used when describing the results of a study that utilized a case control design.
This limitation, however, does not mean that case control designs should
never be used. First, the results of a case control study of potential risk factors to
the injury of interest can provide pilot data that is very helpful in identifying what
potential risk factors should be measured when designing a prospective study.
Second, some injuries are very rare (i.e., they have a very low incidence rate) and
performing a prospective study of risk factors to this injury may not be feasible.
Even in such situations, however, the limitations of case control designs must be
emphasized when discussing the results of case control studies; and, thus, the re-
sults must be put into their proper context.
The issue of injury definition can be more problematic in retrospective studies
as opposed to prospective studies. Ideally, injury confirmation can be obtained
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 248
using imaging methods or strict diagnostic criteria that can be derived from
participants medical records. The reality is, however, that many case control
studies rely on patient self-report for their injury history. This may work for
some major injuries, such as an ACL tear, where individuals are likely to
truly know whether or not they have been diagnosed with this injury; but may
be more problematic for an injury such as an ankle sprain that may be quite
difficult to operationally define for a lay person. Care must be taken in the
design of case control studies to assure tight control over inclusion criteria for
the case group.
CONCEPT CHECK
The issue of injury definition can be more problematic in retrospective studies as
opposed to prospective studies.
Prevalence a b 100
No. of injured subjects
No. of total subjects
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 249
CHAPTER 14 Screening and Prevention of Illness and Injuries: Resarch Methods and Data Analysis 249
For example, if there are 100 athletes on a middle school girls track team and
15 of these athletes are currently injured, the prevalence of injury is 15%. Note
that prevalence provides estimates of injury or illness occurrence at one snap-
shot in time. It does not provide any indication to how many injuries have pre-
viously occurred or how many injuries will occur in the future. The prevalence
of an injury or illness may fluctuate at different time points. With some occupa-
tions or athletes, more individuals may suffer injuries during different seasons
of the year or points in an athletic season. For example, in northern climates
work-related injuries may increase for laborers who work outdoors during
the winter because of the increased likelihood of injury causing falls due to ice-
covered ground or sidewalks. Likewise in competitive athletes, overuse injuries
may be more common early in competitive seasons when the greatest change in
training volume is likely to have occurred. Prevalence is an estimate of the num-
ber of injured or ill individuals in comparison to the entire sample at risk at a
given time point.
The incidence of an injury or illness refers to the number of new cases of the
pathology in a given period of time. To establish the incidence of pathology, a sur-
veillance system must be in place to record the number of new cases. This requires
a prospective study design. For example in sports medicine, the number of ath-
letes who suffer ankle sprains during a full season of competitive basketball may
be of interest. To calculate the incidence of ankle sprains during a season, partici-
pating athletes must be monitored for the occurrence of new ankle sprains for the
entire length of the season. If in one collegiate conference there are 10 basketball
teams each with 15 players, and 8 of them suffer new ankle sprains during the
course of a season, the incidence may be calculated as 15 newly injured players
out of a population at risk of 150 players, or 10%.
CONCEPT CHECK
Prevalence is an estimate of the number of injured or ill individuals in comparison to
the entire sample at risk at a given time point. Incidence refers to the number of new
cases of the pathology in a given period of time.
Incidence Proportion a b
No. of newly injured subjects
No. of total subjects participating
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 250
The number of athletes suffering ankle sprains during a basketball season de-
scribed in the previous paragraph is an example of an incidence proportion. With
incidence proportion, an individual either sustains an injury or illness or they do
not. There is no way of accounting for multiple injuries in the same individual
over the course of time when using incidence proportion.
Incidence rate is defined as the number of new cases that occur per unit of
person-time at risk. In order to calculate incidence rate, the amount of time that
each individual is exposed to injury or illness risk must be calculated. For exam-
ple in sports, this may be determined by quantifying the number of hours (or
games or practices) that an individual athlete participates during a defined period
of time. In occupational medicine, the total number of days (or hours) worked
during a defined period of time may be quantified. Importantly, the use of inci-
dence rate also allows for the tabulation of multiple new injuries or illnesses
among the same individual over the defined time period. For example, a basket-
ball player may suffer an ankle sprain to their right limb in the first week of the
season and a sprain to their left ankle in the last week of the season. When using
incidence proportion, this individual would only be counted once; however, with
incidence rate, both new injuries could be expressed. Incidence rate can be ex-
pressed with the formula:
Incidence Rate a b
No. of new injuries
Total exposure time
While the terms injury risk and injury rate may seem interchangeable,
understanding the difference between injury risk and rate is a very important con-
cept. Injury risk refers specifically to the probability of new injury per individual.
The numerator of injury risk is the number of individuals suffering a new injury
in a given period of time, while the denominator is the total number of individual
exposed to risk of injury over that given period of time. The incidence proportion
described above is an example of injury risk.
Alternatively, injury rate specifically refers to the number of new injuries per
unit of exposure time. The numerator for injury rate is the number of new injuries
and the denominator is the total number of person-time at risk of all individuals
at risk. The incidence rate described above is an example of injury rate.
The estimation of the amount of exposure to risk of injury or illness that each
individual has is very important. For risk of chronic diseases such as pancreatic
cancer, the amount of exposure would typically be expressed as person-years.
Because each individual is under constant, albeit relatively low, risk for develop-
ing this disease over their lifespan every year they live is considered a unit of ex-
posure. In occupational medicine, every day (or hour) that a worker performs
their job is considered a unit of exposure. In athletics, the unit of exposure may be
defined in ways such as player-seasons, hours (or minutes) of sport participation,
or most commonly as an athlete-exposure. An athlete-exposure refers to one
athlete participating in one practice or game.
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 251
CHAPTER 14 Screening and Prevention of Illness and Injuries: Resarch Methods and Data Analysis 251
CONCEPT CHECK
With incidence proportion, an individual either sustains an injury or illness or they do
not. There is no way of accounting for multiple injuries in the same individual over the
course of time when using incidence proportion.
STATISTICAL ANALYSIS
There are several statistical analyses that are important to understand when
designing, executing, and interpreting screening and prevention studies. These
include concepts such as risk ratios, rate ratios, relative risk reduction, absolute
risk reduction, numbers needed to treat, and odds ratios.
The simplest comparison between two measures of injury incidence is to cal-
culate the ratio of the injury incidence between two groups. If the ratio of the in-
jury prevalence estimates between two groups is taken, this is referred to as the
prevalence ratio. If the ratio of injury risk estimates between two groups is calcu-
lated, this is referred to as the risk ratio. Lastly, if the ratio of the injury rate esti-
mates between two groups is determined, this is termed the rate ratio. As with
most epidemiologic statistics, the point estimate must be evaluated along with its
95% confidence interval. The key cutline for determining the efficacy of treatment
(or increased risk) for ratio estimates is whether or not the confidence interval
crosses the 1.0 line on the log scale (see Figure 14-4).
When assessing injury risks or rates between two groups, another simple
method of comparison is the calculation of the relative risk of injury between the two
groups. Relative risk (RR) is simply calculated by dividing the injury rate in the inter-
vention group by the injury rate in the control group. (It must be noted that either in-
jury rates or injury risks may be used to calculate relative risk and the other statistics
in this chapter.) Relative risk simply provides a proportion of injury incidence be-
tween two groups and is identical to the calculation of risk ratio (or rate ratio).
Unfortunately, comprehension of this calculation is not necessarily intuitive.
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 252
Graphing ratios
Harmful treatment Effective treatment
Ineffective treatments
A statistic that is more easily understood is the relative risk reduction (RRR).
This is calculated by simply taking 1 minus the relative risk and multiplying by
100. Relative risk reduction represents the percentage that the experimental con-
dition reduces injury risk compared to the control condition. If the experimental
condition is found to lead to heightened risk of injury, rather than expressing the
relative risk reduction as a negative number, the sign is changed to positive and
this value is termed the relative risk increase (RRI). As with most epidemiologic sta-
tistics, the point estimate must be evaluated along with its 95% confidence inter-
val. The key cutline for determining the efficacy of treatment for RRR is whether
or not the confidence interval crosses the zero line (see Figure 14-5).
Ineffective treatments
100 75 50 25 0 25 50 75 100
Increased risk % Decreased risk
Relative risk increase Relative risk reduction
CHAPTER 14 Screening and Prevention of Illness and Injuries: Resarch Methods and Data Analysis 253
CONCEPT CHECK
The absolute risk reduction is dependent on the magnitude of the injury incidence.
A statistic that builds on the absolute risk reduction is the numbers needed to
treat (NNT). The NNT is calculated by simply taking the inverse of the absolute
risk reduction (1/ARR) comparing two groups. The NNT represents the number
of patients that need to be treated with the experimental treatment to prevent one
injury in comparison with receiving the control condition. The ideal NNT is 1,
meaning that for every patient treated with the experimental treatment, an injury
is prevented. Conversely, the worst NNT is infinity ( q ), meaning that an infinite
number of patients would need to be treated to prevent one injury.
When NNT refers to a treatment that benefits patients, it is often termed
the numbers needed to treat to benefit (NNTB), but when it refers to a treatment
that is deleterious to patients (i.e., causing adverse events), it is referred to
as the numbers needed to harm (NNTH). As with other epidemiologic statistics,
the point estimate for NNT must be evaluated along with its 95% confidence
interval. The key cutline for determining the efficacy of treatment for
NNT is whether or not the confidence interval crosses the infinity line (see
Figure 14-6).
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 254
Ineffective treatments
1 15 35 75 100 100 75 35 15 1
Numbers needed to treat Numbers needed to treat
to benefit to harm
FIGURE 14-6 Interpretation of numbers needed to treat
point estimates and confidence intervals. Note that the
unique scale of this graph with 1 NNTB and 1 NNTH on the
ends of the x-axis and infinity in the center.
CONCEPT CHECK
The NNT represents the number of patients that need to be treated with the experimen-
tal treatment to prevent one injury in comparison with receiving the control condition.
The last statistic we will discuss in this chapter involves the odds of being in-
jured. The odds of any given population becoming injured are calculated by di-
viding the injury risk by 1 minus the injury risk. Remember that odds must be in
comparison to another number (i.e., the odds are 2:1 that a certain event will
occur).
EXAMPLE
CHAPTER 14 Screening and Prevention of Illness and Injuries: Resarch Methods and Data Analysis 255
basketball players not spraining their ankle are approximately 9:1. Assume that
none of these basketball players used any type of prophylactic ankle supports
such as bracing or taping. The following season, all of these athletes are taped for
all practices and games and only 5 of 150 players suffer a sprain. The odds of suf-
fering an ankle sprain with this intervention are 0.034:1. We could take the inverse
of this and determine that the odds of these players not suffering an ankle sprain
while being taped are approximately 29:1.
By themselves, the raw odds among the no intervention (control) group and
the taping intervention group are difficult to interpret. However, by calculating an
odds ratio between the two groups a more clear estimation of the treatment effect
of the intervention becomes evident. By dividing the odds of athletes suffering an
ankle sprain without tape (0.11:1) by the odds of athletes suffering a sprain while
taped (0.034:1), we calculate an odds ratio of 3.2. This indicates that the odds of
suffering an ankle sprain without tape compared to that with tape are 3.2:1 in this
population of basketball players.
CONCEPT CHECK
The odds of any given population becoming injured are calculated by dividing the in-
jury risk by 1 minus the injury risk. Remember that odds must be in comparison to an-
other number (i.e., the odds are 2:1 that a certain event will occur).
As with other epidemiologic statistics, the point estimate for odds ratios must
be evaluated along with its 95% confidence interval. The key cutline for determin-
ing the efficacy of treatment (or increased risk) for odds ratio is whether or not the
confidence interval crosses the 1.0 line on the log scale (see Figure 14-4).
Figure 14-7 provides an example of the calculation of incidence risk, incidence
rate, odds ratio, relative risk reduction, and numbers needed to treat from an in-
jury prevention study (McGuine & Keene, 2006).
CHAPTER SUMMARY
Studies of injury risk factors and injury prevention are extremely important to
the advancement of the clinical practice of sports and occupational medicine.
Injury surveillance lies at the heart of the identification of risk factors and the
determination of the efficacy of injury prevention initiatives. When designing
injury prevention studies, it is important that a shotgun approach to injury
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 256
Incidence
Ankle Incidence
Subjects Exposures Rate
Sprains Risk
(Per 1000 AE)
(39/392) = (39/20,828) *
Control 392 20,828 39
0.099 1000 = 1.87
AE = Athlete Exposure
Relative Risk Reduction RRR = (1RR)*100 (10.63)*100 = 37% (2% RRI 62% RRR)
Absolute Risk Reduction ARR = CI Risk EI Risk 0.099 0.062 = 0.037 (0.001 ARI 0.07 ARR)
(13 NNTB to to
Numbers Needed to Treat NNT = 1/ARR 1/0.037 = 27
707 NNTH)
prevention not be used so that the factors that may lead to injury prevention
may be clearly determined. The gold standard for the performance of studies of
injury risk is the prospective cohort design. This type of study requires the base-
line testing of a large number of participants and the following up of these indi-
viduals for injury occurrence over a defined period of time. An alternative is the
case control design that compares measures between participants who have al-
ready suffered the injury of interest and matched controls who have not suffered
the injury of interest. While easier to perform, there are inherent limitations to
the case control study design that must be carefully weighed when developing
a study of injury risk factors. These are important nuances in reporting and in-
terpreting injury or illness risks and rates and the statistics utilized to assess
these values.
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 257
CHAPTER 14 Screening and Prevention of Illness and Injuries: Resarch Methods and Data Analysis 257
KEY POINTS
Determining an appropriate sample size is a critical step in designing a
prospective study of injury risk factors.
Due to the limitations of case control designs, results of case control studies
must be put into their proper context.
Care must be taken to avoid any bias when determining subject selection.
The estimation of injury rates and risks is accomplished through injury sur-
veillance systems.
Prevalence may be presented as either a proportion or a percentage.
The simplest comparison between two measures of injury incidence is to cal-
culate the ratio of the injury incidence between two groups.
Understanding the difference between injury risk and rate is a very impor-
tant concept.
Many case control studies rely on patient self-report for their injury history.
In sports medicine and occupational medicine, the definition of time loss
and nontime loss injuries must be considered.
An athlete-exposure refers to one athlete participating in one practice or game.
The ideal NNT is 1, meaning that for every patient treated with the experi-
mental treatment an injury is prevented.
The worst NNT is infinity ( q ), meaning that an infinite number of patients
would need to be treated to prevent one injury.
Applying Concepts
1. Consider and discuss implications of the term injury risk factor when
describing the results of a study that utilized a case control design.
2. Consider that retrospective design does have a serious limitation in
comparison to the prospective design.
(continued)
LibraryPirate
97689_c14 01/07/10 4:09 PM Page 258
REFERENCES
Agel J, Olson DE, Dick R, et al. Descriptive epidemiology of womens collegiate basketball
injuries: National Collegiate Athletics Association Injury Surveillance System, 19881989
through 20032004. J Athl Train. 2007;42(2):202210.
Bahr R, Kannus P, van Mechelen W. Epidemiology and sports injury prevention. In: Kjaer M,
Krogsgaard M, Magnusson P, et al., eds. Textbook of Sports Medicine: Basic Science and
Clinical Aspects of Sports Injury and Physical Activity. Malden, MA: Blackwell Science Ltd;
2003.
Basavarajaiah S, Wilson M, Whyte G, et al. Prevalence of hypertrophic cardiomyopathy in
highly trained athletes: relevance to pre-participation screening. J Am Coll Cardiol.
2008;51:10331039.
Barratt A, Peter CW, Hatala R, et al. For the evidence-based medicine teaching tips work-
ing group. Tips for learners of evidence-based medicine: 1. Relative risk reduction, ab-
solute risk reduction and number needed to treat. CMAJ. 2004;171(4). doi:10.1503/
cmaj.1021197.
Knowles SB, Marshall SW, Guskiewicz KM. Issues in estimating risks and rates in sports in-
jury research. J Athl Train. 2006;41:207215.
McGuine TA, Keene JS. The effect of a balance training program on the risk of ankle sprains
in high school athletes. Am J Sports Med. 2006;34(7):11031111.
LibraryPirate
97689_c15 01/07/10 4:33 PM Page 259
CHAPTER 15
TREATMENT OUTCOMES
ACROSS THE DISABLEMENT
SPECTRUM
If you do not expect the unexpected you will not find it, for it
is not to be reached by search or trail.
Heraclitus from Heraclitus Quotes as quoted in BrainyQuote
(www.brainyquote.com/quotes/authors/h/heraclitus.html)
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand why performing research on injured or ill patients is important to advancing
the health sciences.
Appreciate the challenges in conducting clinical research.
Be able to explain the importance of selecting appropriate outcomes measures to assess
in patients as their pathology progresses over time.
Understand a conceptual framework for the measurement of treatment outcomes.
Be able to delineate between disease-oriented and patient-oriented measures.
Be able to describe global and region-specific measures.
Understand how to choose the appropriate outcomes instruments.
KEY TERMS
conceptual framework measurement patient-oriented measures
disease-oriented measures outcomes instruments treatment outcomes
global and region-specific measures pathology
259
LibraryPirate
97689_c15 01/07/10 4:33 PM Page 260
INTRODUCTION
Performing research on injured or ill patients is of utmost importance to advanc-
ing the health sciences. One of the primary challenges in conducting clinical re-
search is selecting the appropriate outcomes measures to assess in patients as their
pathology progresses over time. The purpose of this chapter is to provide the con-
ceptual framework for the measurement of treatment outcomes, delineate be-
tween disease-oriented and patient-oriented measures, describe global and
region-specific measures, and provide guidance in choosing the appropriate out-
comes instruments.
CONCEPT CHECK
The central difference between DOE and POE is that POE outcomes measures assess
issues that are of the utmost interest to the patient, as opposed to information that is
important to clinicians.
DISABLEMENT MODELS
Any valid measurement system must be based on a sound theoretic rationale. The
measurement of treatment outcomes in clinical research must be built upon dis-
ablement models. Several contemporary, but related, disablement models have
been put forth by organizations such as the Institute of Medicine, World Health
Organization, and National Center for Medical Rehabilitation Research. Each dis-
ablement model provides a framework using standard terminology for the de-
scription of health status. A common foundation for all of these models is the
emphasis being placed on the psychosocial functioning of the individual patient
(POE) as opposed to the physiologic or structural functioning of the patient
(DOE). The concept of disability as a sociologic construct was initially champi-
oned by a sociologist, Saad Nagi, in the 1960s. The contemporary definition of dis-
ability is described as the inability or limitation in performing socially defined
roles and tasks expected as an individual within a sociocultural and physical en-
vironment (Nagi, 1991). This is in contrast to the common perception of disability
as a permanent physical or mental handicap.
CONCEPT CHECK
A common foundation for disablement models is the emphasis being placed on the
psychosocial functioning of the individual patient (POE) as opposed to the physiologic
or structural functioning of the patient (DOE).
Nagis Model
Nagis model (1991) is illustrated in Figure 15-1. In this model, the active pathology
refers to the patients injury or illness. Impairment refers to abnormality in physio-
logic function at the site of injury or illness. Functional limitation refers to limita-
tions in actions due to the associated impairments. Disability is operationally
defined in this model as an inability to perform normal socially expected activities
due to functional limitations. Quality of life is defined as an individuals vitality
and level of satisfaction with their current state of existence. Impairments, functional
LibraryPirate
97689_c15 01/07/10 4:33 PM Page 262
Active Functional
Impairments Disability
pathology limitations
Quality of life
limitations, and disability can all influence quality of life. In the context of health
care outcomes, this concept is often referred to as health-related quality of life
(HRQOL) to distinguish it from socioeconomic or interpersonal issues that may
also influence an individuals overall quality of life.
Measures of impairments are typically considered DOE. Measures of func-
tional limitations may represent DOE or POE depending on the context of the
measures. Measures of both disability and quality of life are almost always con-
sidered POE. Examples of four patients using Nagis disablement model are pro-
vided (see Table 15-1) to illustrate these distinctions.
Nagis disablement model has been updated in the past decade in an effort to
get beyond its linear approach to disability. Nagis core concepts do, however, re-
main the foundation for the newer disablement models.
The WHO model has the same foundations as the Nagi model but also adds to
it in unique ways. The WHO model uses a standardized documentation system
that allows clinicians and health systems to classify and quantify specific descrip-
tors of a patients disability. At the top of the WHO disability model is a patients
health condition; this represents the patients pathology that may be in the form of
a disorder, disease, or injury. The cause of the health condition can typically be de-
scribed in terms of abnormalities in body functions and structures. These represent
altered physiologic functions and anatomic structures, respectively. At the center
of the WHO Disablement Model is activities and these are operationally defined
similar to functional limitations in the Nagi and NCMRR models. Participation in-
volves the performance of activities in societal contexts.
The remaining portions of the WHO model are referred to as contextual fac-
tors. These include personal factors and environmental factors that influence a pa-
tients ability to function. Personal factors include, but are not limited to, issues
Health condition
(Disorder or disease)
Body functions
Activity Participation
and structures
(Impairments) (Limitations) (Restrictions)
Contextual factors
such as a patients age, sex, socioeconomic status, and previous life experiences.
Environmental factors refer to the physical, social, and attitudinal environments in
which patients live and include support and relationships, attitudes, and services
and policies; these can be considered at the level of the individual or society. The
addition of these contextual factors allows for an individualization of the factors
influencing a particular patients disability. Without also considering personal
and environmental factors of a patient, it is not possible to truly understand the
barriers that exist to positively affect their disability.
CONCEPT CHECK
The biggest mistake an investigator can make is developing or adopting a new out-
comes measurement tool that is not based on a sound disablement model and vali-
dated appropriately.
CLINICIAN-DERIVED MEASURES
Most health care providers routinely take measures on their patients that could be
considered clinical outcomes. These vary widely by specialty but could include
measures such as core body temperature or body mass index. Almost always,
these measures represent DOE. As state previously, measures of DOE are impor-
tant but they also have limitations when they are used in clinical trials without ac-
companying POE.
Some clinical outcomes instruments exist that ask clinicians to rate the level of
function or disability of their specific patient. Such items that involve a compo-
nent of the physical exam (i.e., swelling, ecchymosis, etc.) that can be directly ob-
served or perhaps even measured (these are examples of DOE), there is little
LibraryPirate
97689_c15 01/07/10 4:33 PM Page 266
concern over clinician bias in reporting. However, when clinicians are asked to
subjectively assess their patients outcome or level of disability, clinicians typically
rate their patients success substantially higher than the patients themselves do.
Therefore, clinician subjective reports of patient outcomes should be considered
DOE rather than POE.
CONCEPT CHECK
When clinicians are asked to subjectively assess their patients outcome or level of dis-
ability, clinicians typically rate their patients success substantially higher than the pa-
tients themselves do.
PATIENT SELF-REPORT
The most common type of POE comes from those outcomes that patients self-
report on their current health status. There are a huge number of patient self-
report instruments that range from having patients rate a single question or item
to having scores of specific questions and items. These instruments may be used
to assess a subjects general or global health status or may be region- or condition-
specific. It is essential that researchers and clinicians utilize outcomes instru-
ments that are appropriate (i.e., valid) for the patients to whom they are being
administered.
CONCEPT CHECK
The most common type of POE comes for outcomes those patients subjectively self-
report on their current health status.
groups (especially pediatrics) and also translated into other languages. In either
case, it is essential that the revised instruments be validated before they are used
extensively.
Challenges exist in assessing global health outcomes in patients who have
pathologies (i.e., relatively minor musculoskeletal injuries) that are not chronic in
nature and are associated with transient disability. In such cases, region-specific
scales may be used, but it must be noted that efforts are underway to develop and
validate outcomes instruments that focus on transient disability. Examples of
common global health outcomes instruments are provided in Box 15.1.
CONCEPT CHECK
Dimension-specific outcomes instruments are often designed to assess a specific phys-
ical or emotional phenomenon such as pain, anxiety, or depression.
LibraryPirate
97689_c15 01/07/10 4:33 PM Page 268
How would you rate your current level of function during your usual activities of daily
living from 0 to 100 with 100 being your level of function prior to your foot or ankle
problem and 0 being the inability to perform any of your usual daily activities?
.0%
FIGURE 15-4 Example of a Single Assessment Numeric Evaluation
(SANE score) from the Foot and Ankle Ability Measure.
Mark an X on the spot on the line that represents the worst pain that you have
been in over the past 24 hours. (Scored as a percentage of the length of the line.)
Since your last clinic visit, has there been any change in activity limitation,
symptoms, emotions, and overall quality of life, related to your asthma?
with knee meniscus injuries would be inappropriate because the instrument was
not designed to be used in this population, nor have its clinimetric properties
been validated in this population. Care must be taken not to overgeneralize the
utility of an outcomes instrument to populations that it was not designed to
measure.
In Chapter 10, the importance of establishing measurement properties of out-
comes measures was discussed in detail. Survey instruments must be developed
with the same rigor as laboratory or clinical measures employed in research.
Clinimetric properties such as reliability, validity, sensitivity to change, and inter-
nal consistency must all be assessed during the development of an outcomes in-
strument. The process of creating a new outcomes scale is neither easy, nor brief,
nor simple.
CONCEPT CHECK
Clinimetric properties such as reliability, validity, sensitivity to change, and internal
consistency must all be assessed during the development of an outcomes
instrument.
LibraryPirate
97689_c15 01/07/10 4:33 PM Page 271
CHAPTER SUMMARY
The importance of assessing patient-oriented evidence (POE) in addition to
disease-oriented evidence (DOE) is critical to understanding the true effectiveness
of clinical intervention. The most common measures of POE are instruments that
require patients to self-report their health status. Outcomes instruments should be
based on an accepted disablement framework. When selecting specific outcomes
instruments, clinicians and practitioners must ensure that the instruments they
select are appropriate for the population to be tested and that the instruments
have had their clinimetric properties rigorously established.
KEY POINTS
Disease-oriented evidence (DOE) is critical to understanding the pathophysio-
logy of specific conditions.
Patient-oriented evidence (POE) outcomes measures provide critical informa-
tion regarding the impact that a persons health care status has on their abil-
ity to function in society and their quality of life.
Each disablement model provides a framework using standard terminology
for the description of health status.
The WHO model uses a standardized documentation system that allows cli-
nicians and health systems to classify and quantify specific descriptors of a
patients disability.
Clinician subjective reports of patient outcomes should be considered DOE
rather than POE.
Eliminating redundancy in items is an important step in developing an out-
comes instrument.
Matching the purpose of using a specific outcomes instrument with the in-
tended utility of the instrument is essential.
Care must be taken not to overgeneralize the utility of an outcomes instru-
ment to populations that it was not designed to measure.
Applying Concepts
1. Consider the three disablement models presented in this chapter.
Present arguments for or against any of the three disablement models
presented here, then suggest and rationalize specific improvements
that you expect to see in new models proposed in the coming years.
2. Discuss why it is imperative that clinical outcomes measurement is
done within the context of an accepted framework.
REFERENCES
Bellamy N, Buchanan WW, Goldsmith CH, et al. Validation study of WOMAC: a health sta-
tus instrument for measuring clinically important patient relevant outcomes to an-
tirheumatic drug therapy in patients with osteoarthritis of the hip or knee. J Rheumatol.
1988;15:18331840.
Binkley JM, Stratford PW, Lott SA, et al. The Lower Extremity Functional Scale (LEFS): scale de-
velopment, measurement properties, and clinical application. Phys Ther. 1999;79:371383.
Denegar CR, Vela LI, Evans TA. Evidence-based sports medicine: outcomes instruments for
active populations. Clin Sports Med. 2008;27(3):339351.
Nagi S. Disability concepts revisited: implications for prevention. In: Pope A, Tarlov A, eds.
Disability in America: Toward a National Agenda for Prevention. Washington, DC: National
Academy Press; 1991:309327.
Snyder AR, Parsons JT, Valovich McLeod TC, et al. Using disablement models and clinical
outcomes assessment to enable evidence-based athletic training practice, part I: disable-
ment models. J Athl Train. 2008;43(4):428436.
Valovich McLeod TC, Snyder AR, Parsons JT, et al. Using disablement models and clinical
outcomes assessment to enable evidence-based athletic training practice, part II: clinical
outcomes assessment. J Athl Train. 2008;43(4):437445.
LibraryPirate
97689_c16_Hurley_Templet6.qxd 7/1/10 4:48 PM Page 273
CHAPTER 16
TREATMENT OUTCOMES:
RESEARCH METHODS
AND DATA ANALYSIS
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand why assessment of the effectiveness of therapeutic interventions on patient
outcomes is the centerpiece of clinical research.
Be able to describe the tenets of evidence-based practice.
Appreciate the value of evidence-based practice in making clinical decisions.
Understand that the best research evidence ideally consists of patient-oriented evidence
from well-conducted clinical trials.
Appreciate the issues involving study design, statistical analysis, and interpretation of the
results of clinical trials that assess treatment outcomes.
KEY TERMS
a priori clinical trial limitations
assessment evidence patient values
basic and translational research infrastructure translational research
clinician experience levels of evidence
273
LibraryPirate
97689_c16_Hurley_Templet6.qxd 7/1/10 4:48 PM Page 274
INTRODUCTION
The assessment of the effectiveness of therapeutic interventions on patient out-
comes is the centerpiece of clinical research. The tenets of evidence-based prac-
tice call for the use of the best research evidence in conjunction with clinician
experience and patient values to make clinical decisions. The best research ev-
idence ideally consists of patient-oriented evidence from well-conducted clinical
trials. This chapter will focus on issues involving the study design, statistical
analysis, and interpretation of the results of clinical trials that assess treatment
outcomes.
CONCEPT CHECK
The performance of clinical trials to assess treatment outcomes must be done as part
of a planned process with foresight. An infrastructure should be put in place that al-
lows for the regular collection of treatment outcomes, including both disease-oriented
and patient-oriented measures.
STUDY DESIGNS
The design of a clinical trial is the greatest factor that influences the level of evi-
dence stemming from that studys results. The two most common schema for
classifying levels of evidence are from the Centre for Evidence-Based Medicine
(CEBM) and the Strength of Recommendation Taxonomy (SORT). For the purpose
of this chapter we will utilize the CEBM classification.
The CEBM has five general categories of levels of evidence. Level 5 evidence
is the lowest level and consists of expert opinion and disease-oriented evidence
derived from the basic and translational research. Level 4 evidence is derived
from case series. Level 3 evidence stems from case control studies, which are al-
ways retrospective in design. Level 2 evidence is derived from prospective cohort
studies that lack randomization. The highest level of evidence is Level 1 evidence
that comes from randomized controlled trials (RCTs), which are the gold standard
for clinical trials methodology.
CONCEPT CHECK
Following the CEBM classification of levels of evidence, the highest level of evidence
is Level 1 evidence that comes from randomized controlled trials (RCTs), which are the
gold standard for clinical trials methodology.
CONCEPT CHECK
A case series does not utilize an experimental design, as there is no comparison group
treated with a different intervention, thus limiting the internal validity of the study de-
sign. Despite the lack of an experimental design, case series must still be performed
in a structured and organized manner.
controlled for in the experimental design. Another limitation is that often there is
no baseline data with which to compare the follow-up measures. While the results
of case control studies must be interpreted in the context of the study design
limitations, they may be used to justify the performance of a more rigorous
prospective investigation of treatment interventions.
CONCEPT CHECK
In case control studies, patients are treated with the respective treatments based on
clinician judgment and not because of experimental allocation. Numerous confound-
ing factors may influence the decision to treat patients with respective interventions;
the inherent weakness of this approach is that none of these confounding factors can
be controlled for in the experimental design.
CONCEPT CHECK
Prospective cohort studies involve baseline measurement before patients receive the
prescribed treatment.
Introduction
Methods
Randomization
Results
(continued)
LibraryPirate
97689_c16_Hurley_Templet6.qxd 7/1/10 4:48 PM Page 280
Discussion
The study provides both point measures and meas- No Yes Where:
ures of variability for at least one key outcome
CONCEPT CHECK
By assigning interventions randomly, the potential of confounding biases between
groups is limited.
LibraryPirate
97689_c16_Hurley_Templet6.qxd 7/1/10 4:48 PM Page 282
Enrollment
Excluded (n = ...)
Not meeting inclusion
criteria (n = ...)
Refused to participate
(n = ...)
Randomized (n = ...) Other reasons (n = ...)
Allocation
(n = ...) (n = ...)
Did not receive allocated Did not receive allocated
intervention (give reasons) intervention (give reasons)
(n = ...) (n = ...)
Follow up
significance of the involved comparisons, but does not necessarily assess the clin-
ical meaningfulness of the results. Is it possible to have statistically significant re-
sults that are not clinically meaningful? Is it possible for results to not be
statistically significant but to have clinical meaningfulness? The answer to both
questions is an emphatic, yes.
CONCEPT CHECK
Traditional statistical analysis assesses the statistical significance of the involved com-
parisons, but does not necessarily assess the clinical meaningfulness of the results.
LibraryPirate
97689_c16_Hurley_Templet6.qxd 7/1/10 4:48 PM Page 284
The aim of this section is not to discount the role of statistical hypothesis testing
in the analysis of clinical trials data. There is clearly a role for hypothesis testing and
it should be viewed as an integral part of data analysis, but not the only means of
analysis. Hypothesis testing does not provide information about the magnitude of
mean differences between comparisons and this is the limitation of this technique.
EXAMPLE
CONCEPT CHECK
There is clearly a role for hypothesis testing and it should be viewed as an integral part
of data analysis, but not the only means of analysis. Hypothesis testing does not pro-
vide information about the magnitude of mean differences between comparisons and
this is the limitation of this technique.
In the above example, the unit of measurement (degrees) is quite easy for clini-
cians to interpret and thus the meaningfulness of the results can be easily ascertained.
When measures are not so easily understood on face value, other analysis con-
cepts such as confidence intervals, effect sizes, and minimal clinically important
differences may be employed.
Confidence Intervals
Confidence intervals are an estimate of dispersion, or variability, around a point es-
timate. In the case of treatment outcomes data, the point estimate is typically a mean,
or average, value for a given sample. (It is important to recognize that confidence
intervals can be calculated around almost any statistical measure that gives a point
estimate.) If a different sample were measured out of the entire relevant population,
the point estimate would likely vary somewhat from the original point estimate. If
different samples were taken continually and the means graphed, we would eventu-
ally be likely to have a normal distribution of point estimates (see Figure 16-2).
LibraryPirate
97689_c16_Hurley_Templet6.qxd 7/1/10 4:48 PM Page 285
67%
95%
0 20 40 60 80 100
Percent function score
FIGURE 16-2 Point estimate (represented by the black rectangle) and 95%
confidence interval (represented by the horizontal black lines extending from
the rectangle). We are 95% confident that the true population mean lies
within the confidence interval.
CONCEPT CHECK
Confidence intervals are an estimate of dispersion, or variability, around a point estimate.
Posttest
Pretest
0 20 40 60 80 100
Posttest
Pretest
0 20 40 60 80 100
CONCEPT CHECK
The width of the confidence interval is influenced by two factors: the variance in the
data (indicated by the SD) and the sample size.
Effect Sizes
Another way to interpret data is the use of effect sizes. Effect sizes provide an es-
timate of the strength of a treatment effect and, thus, an indication of the mean-
ingfulness of results. Where confidence intervals involved the unit of
LibraryPirate
97689_c16_Hurley_Templet6.qxd 7/1/10 4:48 PM Page 287
measurement, effect sizes are on a standardized, unitless scale. There are several
ways to compute effect sizes, but the most straightforward method is Cohens d.
Quite simply, d (Mean 2 Mean 1)/SD 1, where Mean 2 represents the experi-
mental condition, and Mean 1 and SD 1 represent the control condition.
Essentially, this effect size estimate treats the mean difference between conditions
as a proportion of the standard deviation.
CONCEPT CHECK
Effect sizes provide an estimate of the strength of a treatment effect and, thus, an
indication of the meaningfulness of results.
In cases where there is not a clear control condition, effect size can be calcu-
lated as d (Mean 2 Mean 1)/pooled SD. In this case the denominator repre-
sents the pooled standard deviation of both conditions.
The most common interpretation of effect size estimates is derived from the
social sciences (Cohen, 1988). Effect sizes greater than 0.8 are considered
strong, those between 0.5 and 0.8 moderate, between 0.2 and 0.5 small,
and less than 0.2 weak. There is some debate about whether this same scale can
be simply applied to health outcomes measures, but at this time these values are
widely accepted (Rhea, 2004).
Whether an effect size estimate is positive or negative simply reflects which
condition is placed first in the numerator of the effect size equation. In general, ef-
fect sizes are reported as positive values unless there is a specific reason to report
them as negative values. It must be noted that confidence intervals can also be cre-
ated around effect size point estimates as a method of providing an assessment of
the precision, or certainty, of the treatment effect. The interpretation of these con-
fidence intervals is very similar to those previously described in this chapter.
CONCEPT CHECK
In general, effect sizes are reported as positive values unless there is a specific reason
to report them as negative values.
CONCEPT CHECK
ROC analysis provides one a means of identifying the change in a health status meas-
ure that is associated with improvements that are meaningful to patients. This value is
referred to as the minimally clinically important difference (MCID) or minimally impor-
tant difference (MID).
The structure of the ROC for these applications is very similar to that previ-
ously described in Chapter 13. Recall that the use of an ROC analysis in diagnos-
tic studies is needed when the measure in question is continuous rather than
dichotomous. The point on that continuous measure that best differentiates those
with and without the condition or disease of interest based on a dichotomous
gold standard is identified on the ROC. In the case of estimating an MCID, the
continuous measure is the change reported on a patient self-report measure or a
clinician-derived measure such as range of motion while the dichotomous meas-
ure may be a report of being at least much improved on a Global Rating of
Change (GROC) scale or the ability to return to work or sport. As with ROC analy-
ses applied in diagnostic studies a point (remember northwest corner) can be
identified on the continuous measure that best differentiates those that are im-
proved or return to an activity from those patients that do not.
In the case of clinical prediction rules the continuous measure is the number
of factors present following patient evaluation that are related to the outcome of
treatment. Along this scale there will be a point representing the number of fac-
tors present that best differentiates patients who responded well (e.g., much im-
proved based on GROC or 50% reduction in pain) from those who did not (e.g.,
no change in condition reported on GROC or 50% reduction in pain). Examining
examples from the clinical literature best reveals the value of these analyses in es-
timating MCID and deriving clinical prediction rules.
CONCEPT CHECK
In the case of clinical prediction rules the continuous measure is the number of fac-
tors present following patient evaluation that are related to the outcome of treatment.
LibraryPirate
97689_c16_Hurley_Templet6.qxd 7/1/10 4:48 PM Page 289
80 80
60 60
40 40
20 20
0 0
0 20 40 60 80 100 0 20 40 60 80 100
80 80
60 60
40 40
20 20
0 0
0 20 40 60 80 100 0 20 40 60 80 100
The papers by Stucki et al. (1995) and Flynn et al. (2002) are examples of the
utility of ROCs across varying research purposes. The analyses essentially identify
cut-offs or points along continuous data that best correspond to dichotomous sub-
groups of a sample drawn from a population. Although the concept of ROCs may
appear foreign and perhaps intimidating, once an understanding of sensitivity
and specificity (see Chapter 9 for a detailed explanation) is gained the use of ROC
analysis in research and interpretation of these analyses in the clinical literature
we read become much more user-friendly.
CONCEPT CHECK
Traditional hypothesis testing is the most common way to analyze data statistically, but
its limitations in assessing clinical meaningfulness must be acknowledged.
CHAPTER SUMMARY
Clinical outcomes studies form the basis of much of evidence-based practice.
Performing successful clinical trials requires a solid infrastructure for data collec-
tion and management, the use of a sound study design, and a comprehensive data
analysis plan. Careful planning is essential for successful execution of clinical out-
comes research.
KEY POINTS
The importance of building a strong research infrastructure for clinical out-
comes research cannot be overemphasized.
The design of a clinical trial is the greatest factor that influences the level of
evidence stemming from that studys results.
LibraryPirate
97689_c16_Hurley_Templet6.qxd 7/1/10 4:48 PM Page 292
Applying Concepts
1. Consider and discuss what is meant by the phrase best research evi-
dence. Provide historical and current examples from clinical and
health-related research.
2. Discuss and debate the advantages and disadvantages of prospective
cohort studies. Keep in mind that the choice of treatment is not ran-
domly assigned but is instead left to the discretion of the treating clini-
cian. What are some of the confounding factors that may influence the
decision to treat patients with the respective interventions? Provide
specific examples and details to support your position.
3. Is it possible to have statistically significant results that are not clini-
cally meaningful? Is it possible for results to not be statistically signifi-
cant but to have clinical meaningfulness? The answer to both questions
is an emphatic, yes. Discuss and explain why the answer to both of
these questions is yes.
LibraryPirate
97689_c16_Hurley_Templet6.qxd 05/07/10 5:43 PM Page 293
REFERENCES
Chodron P. When Things Fall Apart: Heartfelt Advice for Difficult Times. Boston, MA:
Shambala Publication, Inc.; 1997.
Cohen J. Statistical Power Analysis for the Behavioral Sciences. 2nd ed. Hillsdale, NJ: Lawrence
Erlbaum Associates; 1988.
Copay AG, Glassman SD, Subach BR, et al. The minimum clinically important difference in
lumbar spine surgery: a choice of methods using the Oswestry Disability Index, Medical
Outcomes Study Questionnaire Short Form 36, and pain scales. Spine J. 2008;8:968974.
Flynn T, Fritz J, Whitman J, et al. A clinical prediction rule for classifying patients who
demonstrate short-term improvement with spinal manipulation. Spine. 2002;27:
28352843.
McBrier NM, Neuberger T, Denegar CR, et al. MR imaging of acute injury in rats and ef-
fects of Buprenex on limb volume. J Am Assn Lab Animal Sci. 2009;48:15.
Rhea MR. Determining the magnitude of treatment effects in strength training research
through the use of the effect size. J Strength Cond Res. 2004;18:918920.
Stucki G, Liang MH, Fossel AH, et al. Relative responsiveness of condition-specific health
status measures in degenerative lumbar spinal stenosis. J Clin Epidemiol. 1995;48:
13691378.
SUGGESTED READING
Hopkins WG, Marshall SW, Batterham AM, et al. Progressive statistics for studies in sports
medicine and exercise science. Med Sci Sports Exerc. 2009;41(1):313.
LibraryPirate
97689_c17 01/07/10 4:36 PM Page 294
CHAPTER 17
CLINICAL PRACTICE
GUIDELINES AND CLINICAL
PREDICTION RULES
The men of experiment are like the ant, they only collect
and use; the reasoners resemble spiders, who make cobwebs
out of their own substance.
Francis Bacon, from Aphorism 95 from Bacons 1620 work
The New Organon, or True Directions Concerning the
Interpretation of Nature, as quoted on goodreads
(http://www.goodreads.com/quotes/)
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand the difference between clinical prediction guides (also known as clinical pre-
diction rules) and clinical practice guidelines.
Understand the grading system for clinical practice guidelines.
Understand the concept and application of clinical prediction guides.
Be able to explain and describe why and how the Ottawa Ankle Rules are used as
prediction guides.
Be able to discuss how practice guides are connected to diagnostic studies.
Discuss how patient care is improved through the use of clinical practice guidelines and
clinical prediction guides.
Explain and describe the concept of strength of evidence in clinical research.
Identify outside influences when implementing a strategy to take care of a patient.
294
LibraryPirate
97689_c17 01/07/10 4:36 PM Page 295
KEY TERMS
clinical practice guidelines likelihood ratios receiver operator characteristic curve
clinical prediction guides population values regression analysis
clinical prediction rules position statements strength of the evidence
intervention predictor variables
INTRODUCTION
This section of the book has been devoted to research methods and data analyses
across a spectrum of clinical research. We have addressed diagnostic procedures,
screening and treatment outcomes, and explored systematic review and meta-analy-
sis. Before closing out this section two additional topics warrant attention: clinical
prediction guides (also known as clinical prediction rules) and clinical practice
guidelines. These topics could be addressed in separate short chapters. However,
while quite different in scope and developmental process each helps answer the
question of how to proceed with specific patients based on presentation and exami-
nation findings. Thus, we have chosen to include these topics in a single chapter.
EXAMPLE
radiographic evaluation is not indicated. If the patient is unable to bear weight for
4 steps or has pain with palpation at distal 6 cm of the medial or lateral malleolus,
the base of the fifth metatarsal or the navicular radiographic evaluation is needed
to determine whether a fracture has occurred. This clinical prediction guide has
been studied in a number of settings using a variety of assessors (e.g., orthopaedic
surgeons [Springer et al., 2000], family physicians [McBride, 1997], physical ther-
apists [Springer et al., 2000], nurse practitioners [Mann et al., 1998], and nurses
[Derksen et al., 2005; Fiesseler et al., 2004; Karpas et al. 2002]) and settings (e.g.,
emergency department [Broomhead & Stuart, 2003; Karpas et al., 2002; Papacostas
et al., 2001; Yazdani et al., 2006] and clinics [Papacostas et al., 2001; Wynn-Thomas
et al., 2002]) has performed well. As noted previously, this level of development
has resulted in the Ottawa Ankle Rules emerging as practice guidelines, ad-
dressed at the end of this chapter, in some settings.
CONCEPT CHECK
The Ottawa Ankle Rules are an example of a clinical prediction guide based on the
observation of weight-bearing status and areas of tenderness upon palpation.
CONCEPT CHECK
Clinical prediction guides are developed from a cluster of exam findings or character-
istics and may assist in the evaluative or treatment phase of patient care.
LibraryPirate
97689_c17 01/07/10 4:36 PM Page 297
DATA ANALYSIS
Data analysis in studies conducted to develop and validate clinical predication
guides may compare outcomes measures between groups as described in
Chapters 15 and 16 devoted to outcomes measures. Childs et al. (2004) compared
patients receiving or not receiving care based on the clinical predication guide de-
veloped by Flynn et al. (2002). As noted above, those that were treated with spinal
manipulation based on the criteria proposed by Flynn et al. (2002) had superior
outcomes at 6 months in comparison to patients identified as candidates for ma-
nipulation but assigned to treatment with exercise alone. The analyses conducted
by Flynn et al. (2002) used data from a single cohort to generate likelihood ratios
that translated into the probability of a successful outcome based on a dichoto-
mous measure when a patient presents with a set of predictive characteristics.
This process is further illustrated in the following example.
EXAMPLE
Table 17-1 The six variables forming the clinical prediction rule and
the number of subjects in each group at each levela
Symptoms 30 days
FABQPA score 12
Diminished upper thoracic spine kyphosis
6 2 0
5 3 0
4 9 1
3 18 4
2 7 11
1 3 14
0 0 6
a
FABQPA, Fear-Avoidance Beliefs Questionnaire Physical Activity subscale; ROM, range of motion. (Reprinted
with permission from Cleland JA, Childs JD, Fritz JM, et al. Development of a clinical prediction rule for guiding
treatment of a subgroup of patients with neck pain: use of thoracic spine manipulation, exercise, and patient
education. Phys Ther. 2007;87:923.)
LibraryPirate
97689_c17 01/07/10 4:36 PM Page 299
CONCEPT CHECK
Sackett et al. (2000, pp. 173, 177) presented a system of grading recommendations
with letters A, B, C, and D based on the strength of the research evidence available for
diagnostic, intervention (treatment and prevention), prognostic, and economic analy-
sis clinical practice guidelines. Table 17-2 is adapted from Sackett et al. to summarize
grades and levels of evidence from intervention (prevention and treatment) and diag-
nostic studies.
LibraryPirate
97689_c17 01/07/10 4:36 PM Page 301
D 5 Expert opinion without ex- Expert opinion without explicit critical ap-
plicit critical appraisal (system- praisal (systematic review) or expert
atic review) or expert opinion opinion based on nonclinical research
based on nonclinical research
a
All or none is generally associated with death and survival but might be applied to functional recovery (e.g., be-
fore the development of a surgical technique no athletes returned to professional level sports and now some do).
b
See page 226 for discussion of SpPIn and SnNOut. (Adapted with permission from Sackett DL, Straus SE, Richardson
WS, et al. Evidence-based medicine: how to practice and teach EBM. Edinburgh: Churchill Livingston; 2000.)
LibraryPirate
97689_c17 01/07/10 4:36 PM Page 302
LEVELS OF EVIDENCE
Level of Evidence and Grade of Recommendation
Grade A recommendations are the result of the availability of level 1 evidence
which in the case of interventions consists of systematic reviews of randomized
controlled trials with homogeneous results (1a) (a - is added to identify rec-
ommendations where there is concern resulting from heterogeneity of findings
or wide confidence intervals), individual randomized controlled trials with
narrow confidence intervals (1b), or all or none results (1c) where either all pa-
tients died prior to the intervention in question and some now survive or
where some patients died prior to the intervention and now some survive.
Grade B recommendations stem from systematic reviews based on level 2 evi-
dence consisting of systematic reviews of cohort studies with homogeneous re-
sults (2a), individual cohort studies or RTC with less than 80% follow-up (2b),
outcomes research (2c), systematic reviews of case control studies with homo-
geneous results (3a), or individual case control studies (3b). Grade C recom-
mendations are based on case series or lower-quality case control or cohort
studies (evidence level 4) while grade D recommendations stem from expert
opinions, logical applications of physiologic principles, or bench science (evi-
dence level 5).
ADDITIONAL CONSIDERATIONS
Once the consumer has ascertained that the clinical practice guidelines have re-
sulted from a comprehensive (all languages, all journals) review and is up to date
and considered the strength of the recommendations, there are additional consid-
erations before proceeding with the recommended patient care. In the case of pre-
vention efforts one must decide if the event rate and/or seriousness of the
problem warrant the expenditure and effort required to implement the interven-
tion. In all cases where a patient or the patients guardians are able to provide
input one must also consider whether a proposed plan for diagnosis or treatment
is in conflict with their beliefs and values. Lastly, one must consider whether there
are barriers to implementation. For example, while the work of Childs et al. (2004)
cited previously might result in a clinical practice guideline calling for thoracic
spine manipulation in a 40-year-old patient presenting with thoracic kyphosis,
limited cervical extension, and an absence of radicular symptoms, the treatment
may not fall within the practice domain based on the state in which a clinician is
credentialed to practice. Organizational rules, entrenched local traditions, and re-
source availability may also pose barriers that the clinician cannot overcome in an
effort to implement a clinical practice guideline.
CONCEPT CHECK
In all cases where a patient or the patients guardians are able to provide input one
must also consider whether a proposed plan for diagnosis or treatment is in conflict
with their beliefs and values.
CHAPTER SUMMARY
In summary, while attractive and frequently very helpful, the clinician must be
prepared to critically review clinical practice guidelines as with other forms of
clinical literature. The development of clinical prediction guides and clinical
practice guidelines, combined with the evolution of the Internet however, greatly
expands access to information likely to improve patient care. For example, Fritz
et al. (2007) investigated the impact of following guidelines recommending active
intervention in the treatment of patients with low back pain. Those treatments
that adhered to the clinical prediction guide experienced greater pain relief,
greater functional recovery, and were 77% more likely to be deemed to have a
successful outcome in physical therapy despite requiring less care rendered at a
lower cost. Thus, when appropriately assessed and implemented, prediction
guides and practice guidelines enhance care and may lower health care costs.
LibraryPirate
97689_c17 01/07/10 4:36 PM Page 304
KEY POINTS
Clinical prediction guides are formed from a cluster of exam findings and
may help with forming a diagnosis or the treatment of a patient.
Position statements are often clinical practice guidelines developed to im-
prove health care practice.
Implementation of clinical practice guidelines can be hindered by factors in-
cluding organizational rules, availability of proper resources, and tradition.
The goal of clinical practice guidelines is to combine the best evidence and
provide the most useful recommendations for the patient care.
Clinical practice guidelines should be regularly updated in order to reflect
the best evidence to date.
Clinical prediction guides can help with identifying a problem when a sin-
gle examination does not provide results.
Clinical prediction guides can be used to help determine the need for refer-
ral for radiographic examination and other diagnostic work-up.
Clinical practice guidelines are professional edicts developed by a team of
clinicianscholars.
While clinical practice guidelines are valuable resources, the practice of evi-
dence-based care calls for the integration of patients values and clinician ex-
perience into clinical decision-making. Thus, a patients values and
preferences must be taken into consideration when deciding the appropriate
course of action.
Applying Concepts
1. Use Sackett's grading recommendations with letters (A, B, C, and D) to
base the strength of the research evidence available for diagnostic, inter-
vention (treatment and prevention), prognostic, and economic analysis
clinical practice guidelines for a current clinical study of your choice.
2. Discuss how you would use the Ottawa Ankle Rules as prediction
guidelines to help determine the need for radiographic evaluation to
judge the chances of an ankle fracture following a lateral sprain.
LibraryPirate
97689_c17 01/07/10 4:36 PM Page 305
REFERENCES
Broomhead A, Stuart P. Validation of the Ottawa Ankle Rules in Australia. Emerg Med
(Fremantle). 2003;15:126132.
Browder DA, Childs JD, Cleland JA, et al. Effectiveness of an extension-oriented treatment
approach in a subgroup of subjects with low back pain: a randomized clinical trial. Phys
Ther. 2007;87:16081618.
Childs JD, Fritz JM, Flynn TW, et al. A clinical prediction rule to identify patients with low
back pain most likely to benefit from spinal manipulation: a validation study. Ann Intern
Med. 2004;141:920928.
Cleland JA, Childs JD, Fritz JM, et al. Development of a clinical prediction rule for guiding
treatment of a subgroup of patients with neck pain: use of thoracic spine manipulation,
exercise, and patient education. Phys Ther. 2007;87:923.
Derksen RJ, Bakker FC, Geervliet PC, et al. Diagnostic accuracy and reproducibility in the
interpretation of Ottawa ankle and foot rules by specialized emergency nurses. Am J
Emerg Med. 2005;23:725729.
Fiesseler F, Szucs P, Kec R, et al. Can nurses appropriately interpret the Ottawa ankle rule?
Am J Emerg Med. 2004;22:145148.
Flynn T, Fritz J, Whitman J, et al. A clinical prediction rule for classifying patients with low
back pain who demonstrate short-term improvement with spinal manipulation. Spine.
2002;27:28352843.
Fritz JM, Cleland JA, Brennan GP. Does adherence to the guideline recommendation for ac-
tive treatments improve the quality of care for patients with acute low back pain deliv-
ered by physical therapists? Med Care. 2007;45:973980.
Hicks GE, Fritz JM, Delitto A, et al. Preliminary development of a clinical prediction rule
for determining which patients with low back pain will respond to a stabilization exer-
cise program. Arch Phys Med Rehabil. 2005;86:17531762. Available at
http://www.ohri.ca/emerg/cdr.html, accessed January 5, 2010.
Institute of Medicine. Clinical Practice Guidelines: Directions for a New Program. Washington,
DC: National Academy Press; 1990.
Karpas A, Hennes H, Walsh-Kelly CM. Utilization of the Ottawa ankle rules by nurses in a
pediatric emergency department. Acad Emerg Med. 2002;9:130133.
Laslett M, Aprill CN, McDonald B, et al. Diagnosis of sacroiliac joint pain validity of indi-
vidual provocation tests and composites of tests. Man Ther. 2005;10:207218.
Mann CJ, Grant I, Guly H, et al. Use of the Ottawa ankle rules by nurse practitioners. J Accid
Emerg Med. 1998;15:315316.
McBride KL. Validation of the Ottawa ankle rules. Experience at a community hospital. Can
Fam Physician. 1997;43:459465.
Papacostas E, Malliaropoulos N, Papadopoulos A, et al. Validation of Ottawa ankle rules
protocol in Greek athletes: study in the emergency departments of a district general hos-
pital and a sports injuries clinic. Br J Sports Med. 2001;35:445447.
Sackett DL, Straus SE, Richardson WS, et al. Evidence-based Medicine: How to Practice and
Teach EBM. Edinburgh: Churchill Livingston; 2000.
Springer BA, Arciero RA, Tenuta JJ, et al. A prospective study of modified Ottawa ankle
rules in a military population. Am J Sports Med. 2000;28:864868.
Wynn-Thomas S, Love T, McLeod D, et al. The Ottawa ankle rules for the use of diagnos-
tic X-ray in after hours medical centres in New Zealand. NZ Med J. 2002;115(1162):U184.
Yazdani S, Jahandideh H, Ghofrani H. Validation of the Ottawa ankle rules in Iran: a
prospective survey. BMC Emerg Med. 2006;6:3.
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 306
CHAPTER 18
SYSTEMATIC REVIEW
AND META-ANALYSIS
CHAPTER OBJECTIVES
After reading this chapter, you will:
Learn how to complete a systematic review.
Understand the hierarchy of evidence.
Be able to identify the domains and elements of a systematic review.
Understand the difference between narrative and systematic reviews.
Identify obstacles clinicians must overcome to practice evidence-based medicine.
Be able to describe and explain the differences between systematic review and meta-analysis.
Understand how to conduct a meta-analysis.
Be able to discuss the concept of validity in systematic review.
Explain the role of the Cochrane Collaboration in evidence-based health care.
Understand the concept and implications of publication bias in systematic review.
KEY TERMS
case reports interpretation randomized treatment
clinical research intervention order trials
Cochrane Collaboration intervention outcomes strength of evidence
evidence-based medicine (EBM) patient-important outcomes systematic review
306
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 307
INTRODUCTION
Throughout the preceding chapters attention has been directed toward the de-
sign of studies and analysis of data related to the prevention, diagnosis, and
treatment of illness and injury. In these chapters attention has been directed to
the fact that data derived from samples drawn from populations provide only
estimates of true population values. For instance, the positive likelihood ratio
(LR) 3.9 and negative likelihood ratio (LR) 0.28 generated from our mock
data set (see pages 228230) provide only an estimate of the diagnostic usefulness
of the diagnostic test. The calculation of confidence intervals provides a range
(LR 1.4 to 11.0 and LR 0.12 to 0.64 in this case) within which we can be 95%
confident that a true population value lies. The larger a sample, the more likely
the values derived are reflective of the true population values and thus truth.
For example, if one maintains the same proportions of true positives, true nega-
tives, and false findings as described in Table 13-3 with a 10-fold increase in
sample size (see Table 18-1) the LR and LR values are unchanged but the
confidence intervals narrow substantially LR 3.9 (95% CI 2.8 to 5.4), LR
0.27 (95% CI 0.21 to 0.35). Thus, it would seem that what is needed to answer
the multitude of clinical questions facing health care providers are large, well-
designed studies that generate narrow confidence intervals and thus confidence
that the findings reflect the usefulness of diagnostic tests in a population or re-
sponses to interventions.
This simple solution crumbles when we are faced with the realities of the time,
effort, and costs associated with clinical research, challenges of recruiting patients
into studies, losing patients to drop-out and balancing research with clinical prac-
n 180 n 30
n 50 n 120
Sensitivity 18/23 0.78; Specificity 12/15 0.80; LR 3.9 (95% CI 2.85.4); LR 0.27 (95%
CI 0.210.35).
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 308
CONCEPT CHECK
The clinician seeking to practice evidence-based medicine faces two significant
obstacles.
SYSTEMATIC REVIEW
What For? And, How To?
Before we move forward with the how to of systematic review one additional
point needs to be made regarding the what for. Systematic reviews are clinician-
friendly and summarize the evidence into a single paper. But how are systematic
reviews viewed in comparison to other forms of evidence? Several authors have
described hierarchies of the strength of the evidence provided in various forms of
clinical research (see Boxes 7.1, 18.1, and Figure 18-1)
In each of the hierarchies described in Box 18.1 and Figure 18-1, systematic re-
view takes a rightful place as the most compelling evidence in most circumstances.
Note that within patient, randomized treatment order trials, while excellent, re-
quire that each patient enrolled in a study receives all interventions for prescribed
periods of time in randomized order. While possible in studying interventions in
Systematic
reviews
FILTERED
INFORMATION
Quality of evidence
articles (Article synopses)
UNFILTERED
INFORMATION
Cohort studies
Case-controlled studies
Case series/reports
Data sources and Which databases were searched Comprehensive search of many
search strategy and search strategy are not databases as well as the so-called
typically provided. gray literature. Explicit search
strategy is provided.
Study quality If assessed, may not use formal Some assessment of quality is
quality assessment. almost always included as part
of the data extraction process.
research the subsequent description of the research methods should appear ap-
propriate to address the question. In the case of systemic review the description of
the research methods begins with a clear description as to how the investigators
conducted their search. Most will begin by describing the search terms entered
and the databases searched.
CONCEPT CHECK
Systematic review is a research process where the investigators identify previous stud-
ies that address a particular question, summarize findings, and when possible collapse
data for meta-analysis, a process where statistical analysis is performed on data com-
bined from multiple studies.
thus eliminate cohort studies, case reports, and literature reviews from further con-
sideration. Papers may be limited by language such as including only those pub-
lished in English. Regardless of the criteria, a reader reproducing this process
should obtain the same results. Flow diagrams (see Figure 18-2) can help the reader
follow the process from search to a final pool of papers included in the review.
Once a description of the search process that identified papers for consideration
is provided a description of how final papers were included for data analysis should
be provided. The incorporation of multiple reviewers using criteria determined a pri-
ori and blinded to the assessment of other investigators reduces the potential for bias.
Similarly, methods by which data are extracted from individual papers for
analysis should be described, including a description of investigators blinded to
the assessment of others. Additional concerns may include the grading of method-
ologic quality of included papers. There are multiple grading scales of method-
ologic quality of diagnostic and intervention studies, since studies of differing
methodologic quality can lead to differences in the results of individual trials.
Thus the assessment of methodologic quality can become an important element in
the interpretation of data analysis.
CONCEPT CHECK
Literature reviews provide a discussion around multiple references related to an issue
but often limit the included literature to that which supports a position taken by the
author or authors a priori. By comparison, a systematic review begins with an answer-
able question and works through a planned process of investigation.
DATA SYNTHESIS
Data synthesis is the process through which data from multiple studies are com-
bined. Data from studies using similar research methods and measurements can
be combined through meta-analysis. This process can substantially increase the
size of the sample contributing data for analysis thus increasing statistical power
and narrowing confidence intervals. Several factors, however, warrant considera-
tion before combining data for meta-analysis. One must consider whether the pa-
tients included in individual studies are sufficiently similar to warrant combining
data. For example, Verhagen et al. (2007) did not pool data in their systematic re-
view of the treatment of whiplash pain. One consideration in this decision was the
fact that, while all of the patients in the trials included for review-fulfilled criteria
for having suffered whiplash mechanism injuries, the timing of interventions var-
ied from acute to chronic symptoms. It is certainly reasonable to suspect patients
with chronic symptoms may respond differently to treatment than those entering
a clinical trial more acutely.
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 315
Inclusion criteria:
Random allocation of treatments
Inclusion of skeletally mature patients of either sex with a current fracture
Blinding of both the patient and assessors to treatment group
Administration of either LIPUS or PEMF to one of the treatment groups
Assessment of time to fracture healing as determined clinically and/or
radiographically
Limits added
Human
Clinical trials
English
FIGURE 18-2 Flow diagram depicting the process from establishment of criteria for inclu-
sion through search process to a final pool of papers included for review of low-intensity
pulsed ultrasound and pulsed electromagnetic field treatments of tibial fractures. (Reprinted
with permission from Walker NA, Denegar CR, Preische J. Low-intensity pulsed ultrasound
and pulsed electromagnetic field in the treatment of tibial fractures: a systematic review. J
Athl Train. 2007;42:530535.)
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 316
META-ANALYSIS
Raw data or mean difference data from multiple studies can be pooled for meta-
analysis. The reader of meta-analyses will find reference in the methods section of
these papers to fixed or random effects models. Katz (2001, p. 162) summarized
the differences by describing fixed effects models as asking the question, is there
evidence here (within the data available) of an outcome effect while random ef-
fects models address whether the available data indicate that the larger popula-
tion of data from which they were drawn provides evidence of an outcome effect.
Thus, random effects analysis projects greater generalizability of the results of an
investigation.
When mean difference values are pooled the values should be weighted based
on the sample size of the study. Mean values calculated in studies with larger sam-
ples are more likely to reflect population values than mean scores reported in
smaller studies. The details of these calculations are beyond the scope of this
text. However, there are several software packages (e.g., RevMan, see http://
www.cc-ims.net/RevMan/) available to assist in the completion of meta-analysis.
Furthermore, weighing can be used to account for other factors that may affect the
results of an analysis such as controlling for publication bias by placing more
weight on studies failing to show a treatment effect or funding bias by giving
more weight to studies not funded through industry sources. The reader is re-
ferred to the writings of Petitti (2000) for more detailed discussions of these issues.
The numbers in the body of the tables are the NNTs for the corresponding odds ratio at that
particular patients expected event rate (PEER).
OR 1
0.30 46 22 14 10 8
0.40 40 19 12 9 7
0.50 38 18 11 8 6
0.70 44 20 13 9 6
OR 1
0.30 49 25 17 13 11
0.40 43 23 16 12 10
0.50 42 22 15 12 10
0.70 51 27 19 15 13
0.90 121 66 47 38 32
Can you apply this valid, important evidence from a systematic review in caring for your
patient?
What are your patients potential benefits and harms from the therapy?
(continued)
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 320
Should you believe apparent qualitative differences in the efficacy of therapy in some subgroups
of patients?Only if you can say yes to all of the following:
FIGURE 18-3 The Systematic Review (of therapy) Worksheet from the University of
Torontos University Health Network. (Reprinted with permission from University of Torontos
University Health Network, available at http://www.cebm.utoronto.ca/teach/materials/sr.htm.)
make decisions about the extent to which the results of a systematic review should
influence a plan of care.
The components of the Are the results of this systematic review valid? section
provide a check sheet regarding issues of validity. The worksheet then leads the
clinician to patient-specific concerns that must be considered before allowing the
evidence to direct recommendations for care. The first question raised under Can
you apply this valid, important evidence from a systematic review in caring for your pa-
tient? is Is your patient so different from those in the study that its results cannot
apply? This is a rather rhetorical question in that it suggests the need for an ab-
solute yes or no response. First, each patient is an individual who possesses a
unique history and physical state. Second, the clinician treating a patient that is
very different from those included in a systematic review will rarely find evidence
from the study of a group representative of the patient to guide their decisions.
This question, however, forces a weighing of the evidence from which a rationale
for treatment recommendation can be developed and presented to the patient.
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 321
The next question Is the treatment feasible in your setting? leads to further de-
liberation. Obviously, if a treatment has been shown to be highly effective in sim-
ilar patients and can be administered, the clinician will likely proceed in
recommending the treatment. However, if the treatment is not feasible in the cur-
rent setting, the clinician must decide whether treatment in that setting is likely to
be of similar benefit and poses a similar risk or whether referral for the therapy in
question is warranted. Referral is sometimes an easy process but in some cases the
patients circumstances and preferences weigh heavily in the decision.
Once these questions are answered and the clinician is prepared to apply the
results from a systematic review in recommending a course of treatment it is time
to consider the likelihood that the patient will benefit from the intervention and
the potential for adverse responses to treatment. There are a number of ways to
convey the response to an intervention, either preventative or therapeutic. These
measures have been introduced in previous chapters but are summarized here.
The response to an intervention can be conveyed as the magnitude of change on
one or more measures or as a probability of a favorable or adverse outcome. For
example, Medina et al. (2006) completed a systematic review of the response to
hyaluronic acid injections on pain, stiffness, and function in patients with os-
teoarthritis of the knee. The paper identified modest benefits based on analysis
of confidence intervals derived from the data. While useful, these data do not
help the clinician or the patient in knowing the probability of an intervention re-
sulting in improvement or harm in an individual case. Some patients, in fact, re-
port feeling much better after a series of hyaluronic injections. The potential for
such outcomes is estimated with odds ratios, risk reduction measures, and num-
bers-needed-to-treat (NNT) measures. Since NNT is the inverse of absolute risk
reduction (ARR) the same information is provided in differing units. It is also
possible to calculate NNT from odds ratio data if the patients expected event
rate (PEER) is known. The tables found in Figure 18-3 assist in this conversion
and will be applied shortly.
Consider the works of Grindstaff et al. (2006) and Myklebust et al. (2003). The
first is a meta-analysis developed from studies where specific exercise programs
were designed and implemented in an effort to prevent anterior cruciate ligament
injuries. These authors completed an NNT analysis estimating that 89 athletes
need to be enrolled in an intervention program to prevent one ACL injury per sea-
son. Mykleburst et al. (2003) completed an investigation of the effects of an ACL
injury prevention program in team handball players. These authors reported a re-
duction in ACL injuries, especially in elite level players. Although the injury re-
duction was not statistically significant, the point odds ratio estimate for ACL
injury in elite players completing the exercise program was 0.51 in the first year
and 0.37 during the second year. During the year prior to initiating the prevention
program 29 ACL injuries were recorded in 924 players (PEER 0.031). If we round
up to a PEER 0.05, an odds ratio of 0.50 yields an NNT 41. Thus, a clinician could
be fairly confident that enrolling female team handball players in a prevention
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 322
program will reduce ACL injuries at least as effectively as has been reported in
studies of a wider array of athletes.
NNT is a measure familiar to clinicians but a concept that can be difficult to
convey to patients. ARR may be less confusing since it is a percentage. ARR is
simply the event rate in a control group minus an event rate in a treatment group.
In the Mykleburst et al. (2003) paper the event rate of ACL injuries in female ath-
letes was reduced to 0.02. Thus the risk of injury was reduced from 3 in 100 athletes
per season to 2 in 100 athletes per season. A 1% reduction in injury doesnt sound
like much. Relative risk reduction (RRR) calculated by (event rate in a control
group event rate in treatment group)/(event rate in control group) may yield a
different perspective. Using the values above, RRR (0.031 0.02)/0.031 or 0.35.
Thus the RRR reflects a 35% reduction in the risk of ACL injury with participation
in an exercise regimen that sounds a lot different from a reduction of 1 injury per
100 athletes per season.
Why are these values so different and which is most useful? The differences
in ARR and RRR lie in the incidence rate of the problem. In this example, the in-
cidence rate is low and thus the ARR is small. When the incidence rate is low,
however, small differences in incidence rate can yield a large RRR. While this ex-
plains the differences in ARR and RRR, it does not answer the latter question of
which value is most useful. Unfortunately, all of the measures can provide help-
ful information and thus no single measure is most useful.
In this case we are concerned with a serious musculoskeletal injury that is
costly in terms of lost productivity, diminished quality of life, and health care ex-
penditure. The injury may preclude future high-level sport participation and will
lead to early degenerative changes in the affected knee. While neither of the stud-
ies noted above addressed adverse events during the exercise intervention, it
would be reasonable to assume that such events (serious injury occurring during
the exercise training) are very rare. The intervention can also be offered at rela-
tively low cost. Stop and consider these last statements. Notice that the ARR, RRR,
and NNT are being discussed in a new context rather than as isolated values.
Many clinicians would interpret these data as a reason to recommend an ACL in-
jury prevention program. Interventions that pose a greater risk, require more time
or money, or conflict with a patients values or willingness to participate might
yield recommendations that differ markedly despite being associated with identi-
cal NNT, ARR, and RRR measures.
The worksheet (Figure 18-3) leads the clinician to two final considerations,
patient preferences and decisions regarding patient subgroups. The role patient
preferences and choices play in the clinical decision-making process was dis-
cussed by Haynes et al. (2002). Figure 18-4A depicts a model where patient pref-
erence, research evidence, and clinical expertise impact upon clinical decisions.
Figure 18-4B presents a more contemporary model that leaves the clinician at the
center of the decision-making process and introduces the clinical state of the pa-
tient and the surrounding circumstances into the process. This model returns the
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 323
Clinical Research
expertise evidence
Clinical
decision
Patients preferences
A
Clinical decision
Patients preference
and actions
B
FIGURE 18-4 Schematic representations of clinical
decision-making. A: Early model. B: Contemporary model
with the clinician as the actor making and being responsi-
ble for clinical decisions. (Adapted with permission from
Haynes RB, Devereux PJ, Guyatt GH. Clinical expertise in
the era or evidence-based medicine and patient choice.
Evid Based Med. 2002;7(2):3638.)
patient is a unique story and that the individual, not the research, is the focus of
our attention. The last section of the worksheet addresses situations where there
is evidence to suggest that a subset of patients may respond differently from a
larger sample. In the desire to help the patient clinicians may consider an inter-
vention despite a lack of evidence suggesting benefit because their individual
patient differs in some respect to the average patient studied. This may be
quite reasonable in some cases but by completing the worksheet the clinician is
once again forced into a critical analysis of the evidence supporting the effective-
ness of the intervention being considered. Once again the values of the patient
and the clinician expertise will influence the ultimate plan of care but the best
available evidence will have been critically assessed.
CHAPTER SUMMARY
Systematic review with or without meta-analysis is a research strategy that often
provides clinicians with the best current evidence to integrate into patient manage-
ment. Systematic review can also identify where further investigation is needed to
guide effective, cost-efficient health care. As with all clinical research the methods
employed in an investigation can threaten data validity. The investigator should
strive to control threats validity by employing sound research methods and convey
the conduct of the research to the consumer. The consumer must appraise system-
atic reviews critically before applying the results of an investigation in their practice.
This chapter was written from the perspective of the research consumer. In
identifying the components of the systematic review, however, the investigator is
also provided a sound foundation from which to pursue an investigation. The
busy clinician will continue to seek well-conducted systematic reviews to provide
summaries of the clinical literature on selected topics in a time-efficient manner.
Since larger samples are more likely to represent true population values a system-
atic review also is at the top of the evidence value hierarchy. The clinician will con-
tinue to be challenged by conflicting results and a dearth of research in some
areas. Systematic reviews however have become common features across much of
the health care research literature and often a click away on the computer.
KEY POINTS
Systematic review provides the clinician with results compiled from multi-
ple clinical trials in a single document, sometimes with the effect of defining
the best current practice.
A literature review is often limited because it supports a position taken by
the author, whereas a systematic review begins with an answerable question
and works through a planned process.
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 325
Applying Concepts
1. Discuss the use of systematic review with or without meta-analysis in
a research strategy and consider how/if it provides clinicians with the
best current evidence to integrate into patient management.
2. Following the notion of EBP, consider and discuss the extent to which
that patient values serve as a reminder that we should not get carried
away interpreting the numbers and forget that the patient is the focus of
our attention. Provide examples from historical and current research.
LibraryPirate
97689_c18 01/07/10 4:37 PM Page 326
REFERENCES
Bjordal JM, Couppe C, Ljunggren AE. Low level laser therapy for tendinopathy. Evidence
for a dose-response pattern. Phys Ther Rev. 2001;6:9199.
Grindstaff TL, Hammill RR, Tuzson AE, et al. Neuromuscular control training programs
and noncontact anterior cruciate ligament injury rates in female athletes: a numbers-
needed-to-treat analysis. J Athl Train. 2006;41:450456.
Haynes RB, Devereaux PJ, Guyatt GH. Clinical expertise in the era of evidence-based med-
icine and patient choice. Evid Based Med. 2002;7:3638. Available at http://www.cc-ims.
net/RevMan; accessed March 12, 2008.
Jewell D. Guide to Evidence-Based Physical Therapy Practice. Boston, MA: Jones & Bartlett
Publishers; 2008:370.
Katz DL. Clinical Epidemiology and Evidence-Based Medicine. Thousand Oaks, CA: Sage
Publications; 2001:162.
Kerkhoffs GM, Handoll HH, de Bie R, et al. Surgical versus conservative treatment for
acute injuries of the lateral ligament complex of the ankle in adults. Cochrane Database
Syst Rev. 2007;2:CD000380.
Medina JM, Thomas A, Denegar CR. Effects of hyaluronic acid on pain, stiffness and dis-
ability: a meta analysis. J Fam Pract. 2006;55:669675.
Morgan O. Approaches to increase physical activity: reviewing the evidence for exercise-
referral schemes. Public Health. 2005;119:361370.
Myklebust G, Engebretsen L, Braekken IH, et al. Prevention of anterior cruciate ligament
injuries in female team handball players: a prospective intervention study over three
seasons. Clin J Sport Med. 2003;13:7178.
Petitti DB. Meta-analysis, Decision Analysis and Cost Effectiveness Analysis. Methods for
Quantitative Synthesis in Medicine. 2nd ed. New York: Oxford University Press; 2000.
Stasinopoulos DI, Johnson MI. Effectiveness of low-level laser therapy for lateral elbow
tendinopathy. Photomed Laser Surg. 2005;23:425430.
Verhagen AP, Scholten-Peeters GGGM, van Wijngaarden S, et al. Conservative treatments
for whiplash. Cochrane Database Syst Rev. 2007; 2. Art. No.: CD003338. DOI: 10.1002/
14651858.CD003338.pub3.
Walker NA, Denegar CR, Preische J. Low-intensity pulsed ultrasound and pulsed electro-
magnetic field in the treatment of tibial fractures: a systematic review. J Athl Train.
2007;42:530535.
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 327
PART IV
DISSEMINATION
OF RESEARCH
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 328
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 329
CHAPTER 19
PRESENTATION OF FINDINGS
CHAPTER OBJECTIVES
After reading this chapter, you will:
Develop an appreciation for the dissemination of research findings as a part of the
research process.
Understand the differences between presentation and publication of findings.
See the point of poster and platform presentations at professional meetings.
Recognize the advantages and unique features of each presentation forum.
Understand why publication of peer-reviewed papers is held in higher regard than
presentations at professional meetings.
Appreciate the advantages of publishing research findings.
Recognize the workings of a well-prepared presentation.
Understand the components of a well-written research paper.
KEY TERMS
call for abstracts peer reviewed relevant literature
CINAHL publication guidelines scientific paper
limitations PubMed validity
329
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 330
INTRODUCTION
The preceding 18 chapters are intended to provide a foundation for conducting re-
search that generates data that guide clinical decision-making and to prepare prac-
titioners in evidence-based clinical care and students in professional preparation
allied health care programs to use research data in their daily practice. The ad-
vancement of evidence-based practice requires investigators and clinicians to work
together in collecting data and disseminating the results of their research. This
chapter is devoted to the dissemination of findings and therefore targeted more to
the investigator than the research consumer. We believe, however, that the more in-
sight consumers have into the research process the better prepared they will be to
critically appraise what they read. The dissemination of research findings is as
much a part of the process as is the development of research methods and for some
a more daunting task. This chapter provides an overview of research presentation
and manuscript preparation with an emphasis on the latter subject.
TYPES OF PRESENTATIONS
The dissemination of research findings occurs almost exclusively through presen-
tations at professional meetings and publication of manuscripts. It is common that
research findings are first made public at professional meetings with subsequent
publication in a peer-reviewed scholarly journal. Each forum has advantages and
unique features. For the consumer the opportunity to see data presented for the
first time at a professional meeting, pose questions, and discuss the research as
well as the application of results is often enjoyable and professionally rewarding.
From the perspective of the investigator/presenter professional meetings provide
an opportunity to showcase their work and receive feedback that often refines and
directs future projects.
The publication of peer-reviewed papers is held in higher regard than presenta-
tions at professional meetings. This forum also offers some distinct advantages. The
reasons that publication is viewed as being a greater accomplishment than presenta-
tion include the fact that manuscripts receive the most thorough and critical review
from peers since journals receive far more submissions than they have the capacity
to print. The great advantages of publishing, especially in this electronic age, are the
opportunity to reach a worldwide audience with work that is permanently available.
an abstract for review. Each organization provides guidelines regarding the length
and style of the abstract and publishes submission deadlines. As with all publica-
tion guidelines it is important that the submitted materials conform to the guide-
lines of the organization. Abstracts are typically limited in length with a 400 to
600-word limit being common. Writing a good abstract is challenging since the au-
thor has relatively few words to convey the findings from an extensive research
effort. We suggest following the example format guidelines and tips for abstract
preparation given in Box 19.1.
Once an abstract is submitted, a panel of professionals within the organization
usually reviews the work with those submissions meeting the standards for ac-
ceptance being included in the meeting program. The accepted abstracts are then
presented either in poster or platform (oral) presentations. These formats of pres-
entation, which are usually assigned rather than selected by the presenter, require
different presentation skills. Posters that describe an entire research project must
convey the most important points in an orderly format within a limited space.
Consumers attending a meeting often have large blocks of time to browse through
and read groupings of posters. Authors are typically assigned to be present to dis-
cuss their work at prespecified times during the meeting. The opportunity to discuss
research one-on-one with an author can lead to dynamic discussions and profes-
sional collaborations. The challenge of preparing a really good poster requires
skills in layout and design. Posters that appear poorly organized, cluttered, or
generally unattractive tend to attract few readers regardless of the quality of the
research. We suggest the tips for producing effective posters in Table 19-1. See also
3 Expand abstract and stick to the key points and big picture concepts
3 Overall layout should have balanced organization and appear aesthetically pleasing
3 When viewing the poster from a distance of approximately 3 feet, your eye should
be drawn to the most important graphic, table, or figure
3 Make sure there is a clear, easily identifiable, and understandable 'take-home' message
3 Poster must present professional quality appearance (i.e., use of materials, fonts,
color(s), graphics, etc.)
3 Proof read
CONCEPT CHECK
The latitude in developing a presentation has resulted in many a message being lost
in the production. The graphics used must support the presentation, not become the
presentation.
3 General planning
Know your audience
Know your time limit
Know your material
Know your technology
Practice, practice, and practice your presentation
Scientific Writing
Conveying and discussing the results of an investigation in a manuscript is an es-
sential final step in the research process. Reading the research literature can foster
new research ideas and identify strategies to improve patient care. A well-written
research paper leads the reader through the purpose of the work, carefully de-
scribes the research methods, presents the results, and concludes with a discus-
sion of the meaning and application of the findings in a logical and easily
understood manner. The relative ease of reading a well-prepared paper belies the
work required in the writing process. A few scholars simply write well in a man-
ner that seems effortless. The extensive editing and revision of most published pa-
pers, however, reveals the reality that for most writing is hard work. Nearly all
papers published in scientific journals are considerably improved from the sub-
mitted and earlier draft versions through the input of reviewers and editors. Since
manuscripts receive such scrutiny authors must develop thick skin so as to see
the merits in constructive criticism rather than taking such criticism personally.
While writing well requires effort and practice, there are some recommenda-
tions that can be applied that can make it easier to get your first and subsequent
papers published.
Author Guidelines
As noted previously, following the published guidelines is essential for profes-
sional success and acceptance of submitted work. Professional journals publish
Authors Guides, which explain the format and style requirements for papers
published in that journal. The more closely the guidelines are adhered to, the bet-
ter the chances of success. In some cases a failure to comply with an Authors
Guide may lead to the rejection of a paper without peer review. The guidelines re-
late to all aspects of the manuscript including the title, abstract, body of the paper,
and standards for the format of tables, figures, and references. In some cases lim-
its are placed on the length of an abstract, or the entire manuscript and the num-
ber of tables, figures, and references permitted.
CONCEPT CHECK
Authors need to attend to the details of preparing an abstract that best represents their
work while consumers must be able to scan abstracts to identify the research that is
priority reading.
The Introduction. Over time, scientific writing has evolved resulting in changes to
instructions to authors and generally accepted writing style. The greatest change
to the body of a research paper has occurred in the Introduction section. From an
editors perspective the problem with many an introductory section is length. A
well-written Introduction identifies a problem or a question in need of study, pro-
vides a relatively brief review of the most important relevant literature, and con-
cludes with a clear statement of purpose for the research. In most cases three or
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 337
perhaps four paragraphs are sufficient to introduce the reader to the research and
capture their interest. There appear to be a few problems common to Introduction
sections that miss the mark for quality. The first is the tendency to cite more of the
literature than required to orient the reader and substantiate the need for the re-
search being reported. Perhaps this common problem, which most often is found
in the writing of younger scholars, is attributable to experiences in the writing of
theses and dissertations. While there is more latitude in the format of these docu-
ments today a lengthy and detailed Introduction demonstrating the students
mastery of the subject remains a common expectation. The reader of a scientific
manuscript is not interested in judging the cumulative scholarly accomplishments
of the author(s). Moreover, a lengthy Introduction poses the risk of losing the in-
terested reader when the text drifts from the central purpose of the research. In
summary, save the detail related to previously published literature for the
Discussion section of the paper and cut to the chase in the Introduction.
Perhaps the second greatest problem encountered as editors and reviewers of
scientific papers is the failure to concisely define the purpose of the investigation
and weave the thread through the entire paper to a concluding statement. When
the research methods, reported results, and/or discussion of the Results do not re-
flect the stated purpose something is amiss. Hopefully, such a paper does not sur-
vive the review process and get published without revision; however, when one
does, readers beware for greater than usual effort will be required to sort how the
value of the information in ones research and practice.
The last concern with Introduction sections that bears mentioning here is the
effort on the part of authors to identify the need for the research being reported.
On the one hand, implying that the work isnt of importance but of merely of
some interest will be a death blow to the attention to all but the most dedicated
readers. More commonly, however, the importance of the work is overstated. Big
problems are usually solved over time through multiple research efforts while
smaller problems, often of limited importance to most and greater importance to
a few, are just that, smaller problems. Moreover, how often do the limitations im-
posed on the research and the inability to study a very large sample lead to
statements such as we conclude that , however further investigation is
needed to . The point is that a balance is needed to capture the readers atten-
tion without overstating (or overselling) the importance of ones work.
CONCEPT CHECK
A well-written Introduction identifies a problem or a question in need of study, pro-
vides a relatively brief review of the most important relevant literature, and concludes
with a clear statement of purpose for the research.
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 338
EXAMPLE
sections provide overlapping information or that text will flow best in a single
section than it is to insert subheadings when sections become too lengthy and
cumbersome.
CONCEPT CHECK
The Method section of a paper should permit the reader to replicate the study
completely.
EXAMPLE
Statements of Results
Consider the following two statements of hypothetic results from a study of a
6-week home exercise program involving stationary cycling versus normal daily
activity in patients with arthritis of the knees that included a 20-point self-report
of daily pain (0 no pain, 20 completely disabling pain).
First statement:
Second statement:
In the first statement, the statistical procedure (ANOVA) is noted before the
result of interest, which is distracting to the reader. If the Method section was
properly developed, the fact that the analysis of variance (ANOVA) was per-
formed should already be known to the reader. Furthermore, the inclusion of con-
fidence intervals (CI) in the latter provides estimates of the magnitude of change
in each group in addition to addressing the issue of differences due to chance.
EXAMPLE
Discontinuation of Participation
An author might report:
Five participants (three cycling and two normal daily activity) dropped out of the
study. Two of the cycling participants reported increases in knee pain and
(continued)
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 341
swelling after 5 and 8 days of participation and one moved to another country. Of
the normal activity group one was injured in an accident and the other failed to
respond to reminders about follow-up.
CONCEPT CHECK
As a reader of research, it is important to carefully consider the methods used to
acquire and analyze data so that one can make some judgment regarding the validity
of the data reported.
Tables and Figures. The subject of tables and figures also warrants discussion at
this point. Tables are set off from the main text and used to present numerical as
well as textual information. Each table should have a unique title and be arranged
with clearly labeled column headers and row identifiers or headings. The units of
measure should be clearly identified when appropriate. Footnotes that add clarity
to the table are encouraged. A well-developed table will convey extensive infor-
mation to the reader yet occupy relatively little space in the paper. The format of
tables is prescribed in the authors guide of journals.
Figures include an array of materials that convey information to the reader,
often far more effectively than can be accomplished with words alone. Figures may
include photographs, computer-generated images, algorithms, and graphic dis-
plays of data. Bar graphs and line graphs can convey information regarding trends
and uncover the meaning of interactions between variables in the most effective
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 342
Discussion and Conclusion. The last section of a scientific paper is devoted to a dis-
cussion of the results and a conclusion. The Discussion section may include the
authors conclusions or may be followed by a brief Conclusion paragraph. The
Discussion section likely has a greater impact as to how favorably a paper is
judged once published than any other portion of the work and plays a significant
role in seeing a paper accepted for publication. Certainly, a flawed research design
that poses a significant threat to the validity of data is good reason for rejecting a
paper from consideration for publication. Barring fatal design flaws, however,
the Discussion permits the author(s) the opportunity to suggest applications for
their results, compare their results to works previously published, probe theoret-
ical foundations, and identify topics for future research. A well-written Discussion
brings data to life and leaves the reader filled with new insights and ideas. Poorly
prepared Discussion sections first suffer from the plague of the poor Introduction;
they are simply too long. A lengthy Discussion often drifts from the central focus
of the research and puts the reader at risk of losing the key points. Aside from
being too lengthy, the most common problem we have encountered in editing and
reviewing papers is the excessive speculation that some authors are prone to.
While the Discussion should not simply repeat the Results, the results should be
kept in perspective. The generalization of results beyond the bounds of the study
setting and population must be addressed with caution and the limitations of the
study fully disclosed. In reviewing papers we frequently refer back to the purpose
of the research stated in the Introduction. When the Discussion extends beyond
the purpose it is likely too long and perhaps excessively speculative. Regardless
of whether the conclusion is imbedded in the Discussion or placed under a unique
heading labeled Conclusion, the final words should directly address the main
question or questions posed in the statement of purpose. It is always concerning
when the end does not appear directly tied to the beginnings of a scientific paper.
In those cases where the conclusions do not appear directly related to the purpose
one is likely to find more confusion than clarity in between.
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 343
CONCEPT CHECK
It is always concerning when the end does not appear directly tied to the begin-
nings of a scientific paper. In those cases where the conclusions do not appear
directly related to the purpose one is likely to find more confusion than clarity in
between.
Final Suggestions
Scientific writing is hard, time-consuming work. It is, in our experiences, not dif-
ficult to get lost in the forest that is a scientific paper and lose track of the path
from Introduction through to Conclusion. While there is not a single solution we
suggest, we do offer some final suggestions to keep in mind throughout your
writing process:
1. Writing mud
2. Seek review
3. Be patient
The concept of writing mud comes from recollections of the late Dr. George
Sheehan, the philosopher of the exercise boom in the 1960s and 1970s. His books
and columns are still most enjoyable reading and over the course of his writing ca-
reer which overlapped with his career as a cardiologist he wrote a lot. He con-
fessed in some of his writings and lectures to struggling at times to get ideas onto
paper (in the age of the manual typewriter!) and talked of writing what came to
mind or what he referred to as mud. It is advice we still share particularly with
students and young scholars. He made the point that it is not possible to revise
and edit until something is in writing. Certainly, some of this book began as mud
and some of the mud was discarded. However, you have to begin somewhere and
we concur with Dr. Sheehan, if all of the thoughts are not coming together begin
with mud.
Once the mud has taken form seek constructive criticism. Does the paper
make sense to those knowledgeable of the subject and can the interested reader
not an expert in the subject follow the logic and identify the most important find-
ings? Once the feedback has been incorporated into the paper the last step is to let
the paper sit for a few days.
Be patient rather than rushing to submit the work for review by an editorial
board in consideration of publication. After a few days where other activities dis-
tract from near full focus on the paper, read from a fresh perspective. If all is in
order the paper is ready to submit, but we have found this is the prime time to re-
vise wording, shorten and generally fine-tune the writing.
LibraryPirate
97689_c19 01/07/10 4:38 PM Page 344
CHAPTER SUMMARY
This chapter is devoted to the dissemination of research results through presenta-
tion and publication. Public speaking and scientific writing are skills that require
practice to hone. Unfortunately, it is not possible to write an instructional chapter
that will decrease the work of writing or assure immediate proficiency. However,
the advice shared in this chapter is, to a large extent, passed down from our men-
tors and certainly helped us in our personal development as presenters and writ-
ers. Hopefully, it will help some who read this book through the challenges of
presenting their work to the health care community.
KEY POINTS
Titles should be highly descriptive without being exceedingly long.
Abstracts are typically limited in length with a 400 to 600-word limit being
common.
The body of a scientific paper is divided into sections, which provide order
and enhance the transitions through the course of the manuscript.
The use of subheadings in a Method section is encouraged and often re-
quired as per many authors guidelines.
The Methods section should convey exactly how, when, and where data
were acquired and provide sufficient detail to replicate the data analysis.
The Results section should provide answers to the question or questions
posed in the Introduction.
Authors should report the most important results first.
Information presented in tables and figures should not be repeated in the
text of a Results section.
A well-written Discussion brings data to life and leaves the reader filled with
new insights and ideas.
Applying Concepts
1. Consider the potential problems encountered as editors and reviewers
of scientific papers if the author of the paper fails to concisely define
the purpose of the investigation and weave the thread through the
entire paper to a concluding statement. Discuss the likely implications
if the research methods, reported results, and/or discussion of the re-
sults do not reflect the stated purpose of the study/paper.
2. Locate a Call for Papers or Call for Abstracts for an upcoming or
recent professional meeting. Bring it to class with you for review and
comparison of authors guidelines.
3. Choose three abstracts on a topic of interest to you in three different
scholarly journals. Bring them to class with you for review and com-
parison of style format.
REFERENCES
Day RA. How to Write and Publish a Scientific Paper. 3rd ed. Phoenix, AZ: Oryx Press; 1988:28.
Physical Therapy Information for Authors. Available at http://www.ptjournal.org/misc/
ifora.dtl, accessed January 5, 2009.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 346
CHAPTER 20
WRITING THE FUNDING
PROPOSAL
CHAPTER OBJECTIVES
After reading this chapter you will:
Understand the culture of grants.
Understand how to search for funding sources.
Understand how to write a competitive grant proposal.
KEY TERMS
grants internal culture procurement principles
INTRODUCTION
The preceding chapters provide detailed steps for conducting research with a focus
on problem solving using the research process. In particular, Chapter 1 outlines
nine steps to frame a method of problem solving, including: (1) identifying a topic;
346
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 347
(2) searching and reviewing the literature; (3) defining a topic; (4) stating a general
question or a problem; (5) phrasing an operationally defined hypothesis; (6) plan-
ning the methods to test the hypothesis; (7) data collection; (8) data analysis and in-
terpretation of results; (9) writing about the findings. Generally, these steps mirror
the sequence used to develop a well-conceived and convincing grant proposal.
Although the grant proposal process is anxiety provoking in many people,
such anxiety is unnecessary. With careful preparation and critical review of suc-
cessful proposals, the grant writing neophyte becomes fully capable of submitting
a competitive grant proposal designed to improve evidence-based practice. This
chapter provides a brief overview of the culture of grants, identifies resources
helpful in locating funding opportunities, and draws upon the nine steps of re-
search problem solving to compose and create a competitive grant application.
Services Act (FPASA), The Armed Services Property Act (ARPA), The Contracts
Disputes Act (CDA), and several additional statutes that govern ethics, labor laws,
environmental laws, and small business development. These acts establish the
principles governing the competition for federal funds and the process by which
the awards are issued. Using the acts (also called laws or statutes), Congress cre-
ates regulations through legislation applicable to all federal agencies. Although
oversimplified, these regulations provide the framework for open access to fund-
ing information via public announcements of all funding opportunities for grant
and contract activity supported by US tax dollars.
Today, the federal Office of Management and Budget (OMB) is the central
agency responsible for overseeing procurement regulations for federal grants and
contract activity. The OMB uses well-established management systems and tools
to assist federal agencies with their compliance with Congressional provisions. All
federal agencies announcing calls for proposals require OMB clearance for the use
of certain proposal forms to assure compliance with federal statutes. For example,
OMB approval numbers appear at the top and/or bottom of proposal face pages
and budget and certification forms of federal grant applications. The numbers in-
dicate that agency procurement procedures meet federal mandates and regula-
tions for fair and unbiased competition.
Additionally, the OMB manages the federal budget and communicates re-
search and development allocations to the broader community. The OMB commu-
nicates through announcements using the Federal Register and the Commerce
Business Daily, and also posts information alerts on their website and the central
government announcement portal for federal funding opportunities, Grants.gov
(http://www.grants.gov). Lastly, the OMB staff regularly present workshops and
seminars describing federal research priorities and budgetary allocations at na-
tional conferences targeted at research investigators and administrators.
CONCEPT CHECK
Understanding the historical evolution of grant funding can provide valuable insights
for orienting the grant process and can significantly increase the competitiveness of
the grant application.
1
There are also several other types of procurement applications such as memorandum of agreements,
fixed priced agreements, material transfer agreements, subcontracts, consulting agreements, and in-
tellectual property agreements. These agreements inclusively fall under the header of contract.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 350
are the core regulations that apply to all government agencies issuing contracts for
research and development activity. The FAR is also available on the World Wide
Web at http://www.acqnet.gov/FAR/.
SOURCES OF FUNDING
Currently, over $400 billion dollars is awarded for research annually, through 22
different federal agencies. These monies represent approximately 17% of the fed-
eral budget and 11% of the US Gross National Product. All federal funding oppor-
tunities are announced in the Federal Register, Commerce Business Daily, and on the
governments central information and grants application portal, www.grants.gov.
In addition to the central World Wide Web portal, funding announcements are
listed on each individual federal agencys website. The announcements are easily
accessed by entering the term grants or program announcements or fund-
ing in the websites main search field. For those investigators seeking funding
from a specific federal agency, automated e-mail alerts announcing funding op-
portunities are generally available via an agency-specific listserv.
The management of funding opportunities has itself developed into a chal-
lenging and respected profession, where talented and experienced individuals are
highly sought after by both academia and industry. Inevitably, any institution of
higher education that performs research of some sort will have an individual or
an entire department dedicated to the search and pursuit of funding opportuni-
ties. For example, in 2005, the number of grant-making foundations totaled 71,095
(The Foundation Center, 2007). Given the number of US foundation and corporate
giving programs it is impractical to attempt to search individual foundation web-
sites one at a time. Instead, a new industry has been created to provide database
mining software specifically designed to assist academic institutions with the
search for funding opportunities.
The three leading search engines that colleges and universities use today to
conduct funding opportunity searches are the Sponsored Programs Information
Network (SPIN), the Illinois Research Information System (IRIS), and the
Community of Science (COS). Each of these database systems allows an investi-
gator to input key words to attempt to match their research interest(s) to key
words contained within the description of funding opportunities. Once an inves-
tigator enters key words, the software searches all applicable funding opportuni-
ties and announcements and generates a report of funding announcements with
key word matches. The announcements provide a brief synopsis describing the
funding opportunity, including the name of the sponsor, their location, amount
and average size of awards, deadline, and geographic limitations. The investiga-
tor can conduct advanced searches using Boolean logic (using quotes around
and or or to narrow a search) or take advantage of links within a search that
cast a wider net to other opportunities. For example, at the end of a SPIN search, the
investigator is provided with a list of key words synonymous to those originally
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 351
supplied. The synonymous or related key words provide links to new searches,
effectively expanding the yield of funding opportunities.
State funding opportunity announcements vary depending on state congres-
sional sponsorship, state priorities or procedures, and budgetary constraints. The
state funding opportunity announcements usually lag behind the electronic noti-
fications of other sponsors. To locate state sources, the investigator is advised to
search the website of a specific state agency(s) and contact an agency representa-
tive to discuss pending opportunities proposed for funding in the state legislature.
Investigators should also request enrollment on the agencys e-mail distribution
list or mailing list. Some states, such as New York, fund regional libraries. A re-
gional library receives congressional funding to maintain a collection of periodi-
cals, funding directories and searchable databases for funding opportunities for
the region it serves. A reference librarian can assist users unfamiliar with the in-
dexes and databases.
For those investigators interested in funds dispersed from the more than
71,095 (2005) US foundations or corporate giving programs, the leading search-
able databases are the Foundation Directory and Metasoft Systems BigDatabase
and Foundationsearch program. These are just two of many commercially avail-
able products used to navigate and yield targeted searches for specific research
and other funded projects. Universities, regional libraries, and even some eco-
nomic development agencies like business bureaus and chambers of commerce
may provide access to users depending on community needs and any licensing
arrangements with database vendors.
Locating the optimum funding source requires advanced planning, practice,
and patience. The challenge presented is not the lack of funding opportunities,
rather locating the right funding opportunity. The proper funding opportunity is
one that is targeted to an investigators interest and expertise. The opportunity
must be closely aligned with the investigators research program and the sponsor
should provide sufficient funding to complete the research plan. If the announce-
ment matches the mission of the sponsor and the interest and intent of the investi-
gator, two additional criteria must be met: (1) the sponsor awards funds in the
investigators geographic location; and (2) the investigators host institution or em-
ployer meets the eligibility criteria.
Every announcement or request for proposal (RFP) includes a description of
eligibility criteria to permit interested applicants to determine an announcements
suitability. The previously described databases also include fields that allow
Boolean searches to factor in geographic limitations and eligibility criteria.
Once the investigator has found the right source of funding, further work en-
sues. Take, for example, attempting to procure funding from the National
Institutes of Health (NIH) for research. According to survey data collected from
colleges and universities in 2006, the Department of Health and Human Services
(HHS), including the NIH, continues to provide the majority of the federal gov-
ernments funding to universities and colleges. In FY 2006 HHS contributed 57%
of the total federal funding ($17.1 billion), primarily in support of the medical and
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 352
biological sciences (Britt, 2007). Every year the NIH budget contains the largest
amount of competitive funding in the United States for research in health care and
disease prevention (Britt, 2007). According to Federal Grants and Contracts Weekly
(2007), the total funds administered by the NIH for research, health care, injury
prevention, and treatment for all sectors of the economy including universities
and colleges comprise approximately $30 billion annually with about 10,000 new
research projects receiving awards in the same period.
There are hundreds of NIH program announcements made available each year.
In order to identify the announcements most relevant to an investigators particular
research interests, the search for opportunities may need to be culled by targeting a
specific Institute within the overall NIH. Although the NIH is one sponsor, the NIH
is actually comprised of 26 different Institutes, with multiple programs available in
each Institute and many funding opportunities cross listed among different Institutes
for joint funding. It is up to the investigator to browse the Institutes to match their
expertise to the individual Institutes mission. Once the investigator chooses a spe-
cific Institute, then the individual Web pages of the Institute(s) should be examined
in depth to assure that the correct program announcement has been chosen.
As with other federal agencies, within the NIH there exist many programs tai-
lored to the higher education research community. For example, the NIH offers
program opportunities that are specifically designed for new investigators, prima-
rily undergraduate institutions, small business innovation research programs,
training programs, international programs, and so forth. Finding the right funding
source takes advanced planning, communication with the sponsor and with any
collaborators, and with the investigators organization to ensure that the investiga-
tor has identified a viable opportunity for which he/she can realistically compete.
Investigators are strongly advised to contact their institutional research offi-
cer(s), if available, to help match their expertise and interests with the best-fit
funding opportunities and programs. Institutional research officers receive thou-
sands of funding opportunities from hundreds of sponsors each year. These advi-
sors have valuable experience and knowledge concerning which programs best
match their institutions eligibility, geographic restrictions, and investigators
strengths. In addition, they are also familiar with the particular nuances and intri-
cacies of the agency divisions and programs within their state, the federal govern-
ment, and any particular requirements of specific foundations.
CONCEPT CHECK
All federal funding opportunities are announced in the Federal Register, Commerce
Business Daily, and on the governments central information and grants application
portal. In addition to the central World Wide Web portal, funding announcements are
listed on each individual federal agencys website.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 353
2
As reported by Jean Feldman, Director of NSF Policy Division, at the National Council of University
Research Administrators Conference, Annual Meeting, Washington, DC, November, 2007.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 354
commitment to the subject makes the proposal believable and literally makes the
words come alive through the passion and drive of the author. Those authors not
fully committed to a particular philosophy, hypothesis, or strategy quickly lose
the attention of the reviewer. A proposal should tell a compelling story: a story
that the investigator believes in and sees clearly the tasks at hand.
3
David G. Bauer is author to 15 books on grant writing and getting and is nationally renowned for
his work with training university faculty in increasing the odds of their proposals being funded.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 355
importance to the agency, and the background, experience, and training of the re-
viewers scoring the application. The reviewers background information is essen-
tial for attempting to target the proposal to a receptive audience. For example, a
frequent mistake made by new investigators is assuming that the reviewers of the
proposal are as knowledgeable as they are in their specific discipline. Often that is
not the case. Proposers need to be cognizant of composing in an easily under-
standable style so that all readers can comprehend what is being proposed.
Consider an investigator writing a proposal suggesting the creation of a signifi-
cant new manufacturing procedure, followed by a detailed methodological de-
sign, using acronyms and discipline-specific jargon. Although the proposal might
wow the reviewer with a similar technical background, another reviewer may
have little knowledge of the acronyms and significance of the new method be-
cause the language is written in such a way as to exclude some readers while in-
cluding others. It is the responsibility of every investigator to write in such a
manner that invites the fundamental understanding of the topic, its significance,
and application to specific disciplines.
However, a certain balance of lyric, prose, and technical description is re-
quired. The composition cannot be so general as to dilute or minimize the con-
tent. Rather, difficult concepts should be clearly articulated and presented in a
manner that educates the reviewer without being condescending. This process is
often termed scaffolding. The basic concept of scaffolding is to introduce a gen-
eral concept and build upon the concepts foundation with increasing complex-
ity so that the reviewer expands their understanding as they progress through
the text.
Having acquired information about scoring priorities and the background
and types of reviewers scoring the proposal, the prepared investigator now has a
distinct advantage over other applicants who did not contact the program officer.
According to David G. Bauer, almost all successful applicants have had some
prior contact with an appointed program officer. In his book titled The How To
Grants Manual, Bauer shares that in a study of 10,000 federal proposals, the only
variable that was statistically significant in separating the funded and rejected
proposals was pre-proposal contact with the funding source. Chances for success
increased an estimated threefold when contact with the funding source takes
place before the proposal is written (Bauer, 1984, p. 106).
Early and frequent contact with the program officer may help the investigator
discover other useful elements. Program officers often offer advice in interpreting
guidelines and may share information that is not included in the written RFP. As
an example, perhaps a current award cycle or request for proposals seeks spe-
cific proficiencies in the use of technology to improve scientific methodology.
Those applicants unaware of the technology proficiency criteria would be at a dis-
advantage in composing their proposal. Through constructive dialogue with the
program officer, the applicant may discover hints, nuances, and subtle interpreta-
tions of the evaluation guidelines. These nuances and interpretations help target
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 356
the content and better align the writing style and level of detail provided based on
the intended audience.
Given the responsibility of reviewing hundreds of research proposals each
year, program officers become particularly proficient at recognizing cutting edge
research, and having a general awareness of the current state-of-the-art. Via dia-
log and open communication with this valuable resource, the investigator can
gain perspective about similar projects that have received funding, and which
specific topics are most interesting to the sponsor. The information obtained in
this exchange will place the investigator in an excellent position for effectively
proposing the research topic, assuring alignment within the larger context of the
sponsors interests.
CONCEPT CHECK
An effective proposal writer engages a reviewer in wanting to read more and will
hopefully create in the reviewer some positive emotion, action, or thought.
Literature Review
Before beginning the composition of a grant application, investigators need to
conduct a thorough literature search to understand the history, evolution, and cur-
rent and potential future trends in the research topic being considered. An appro-
priate start to a literature review begins in a library or in the librarys electronic
holdings via remote access using discipline-specific databases, online catalogs,
and periodicals. The sponsors website is also an excellent resource. Most spon-
sors provide a list of prior awardees and some provide a brief synopsis, abstract,
or project summary of previously funded grants.
By using both library resources and sponsor databases and archives, the ap-
plicant can comprehensively search past and current literature on their subject.
Investigators should also learn to scrutinize publications acknowledgment sec-
tions. Almost all sponsors require an acknowledgment statement indicating that
the published research was supported in full, or in part, by the sponsor and that
the ideas presented are not necessarily endorsed by the sponsor. The investigator
can glean important clues about funding sources in the acknowledgment section
and follow up with that source to determine if any current or pending funding op-
portunities would be applicable to their area of interest.
this process the investigator is able to expand their vision in a new direction or re-
configure procedures, interventions, or treatments in comparison to previous
study designs. Establishing and validating a sound methodological procedure is
the foundation for which a grant proposal is conceived and reviewed for funding.
However, the crucial first step to grant writing is having a good idea. Ideas need
to be novel. Novel in this context means having an interesting and potentially use-
ful idea.
Many proposals are effectively written and convincing, but the concept itself
appears insignificant or inconsequential. The letters received by the authors of
nonfunded proposals typically state We are sorry that your project was not
funded. The topic did not match the XYZs highest priority this round. In other
words, sponsors do not want to fund projects that do not appear important, use-
ful, or are of low priority.
Basing an idea on the significance of prior work as demonstrated through a
literature search develops the ideas credibility and usefulness. An effective way
to hone an idea is to select a broad concept that you want to explore, conduct an
exhaustive literature search, and then identify gaps in knowledge or inconsisten-
cies in published results that merit further exploration. Writers must demonstrate
to sponsors that they are familiar with previously published research, and that
the proposed research study builds upon past practices to make improvements
to the future. Once the idea is demonstratively sound, then the writer can begin
partitioning the concept(s) into methodological steps. It is these gaps in knowl-
edge that sponsors are funding and that are considered novel and important
to science.
exciting, and deliverable. The composition style engages the reviewer and con-
vinces the reviewer to support the proposal.
Understanding the psychology of grant writing greatly helps the writer in
choosing their examples, selecting evidence, and articulating a plan of action. A
good reviewer wants to be involved in selecting projects that are worthy and takes
his/her role seriously in selecting the best proposals that will simultaneously ad-
vance the mission of the sponsor or match the program announcements objec-
tives. Therefore, writing to the readers, understanding and using the terms
identified in the program announcement, familiarity with the sponsors history,
prior grants, and priorities, and combining those with your own knowledge and
interest in the project being proposed is the best recipe for successfully engaging
reviewers and being awarded funding.
Elements of a Proposal
Most proposals include the following sections:
Each of these sections is discussed below. There are variations in the number
or content of the sections, but in general the information presented is fairly stan-
dardized. In every case, the investigator should read and reread proposal guide-
lines to assure that the directions are carefully followed. Proposals that do not
adhere to formatting instructions (margins, paginations, fonts, page or character
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 359
limitations, order, appendices, etc.) will not be reviewed. If there are any questions
about the instructions, investigators should contact the sponsor for guidance.
Additionally, investigators should familiarize themselves with the review criteria
and any specialized requirements such as supplemental instructions or electronic
proposal submission details. Unfortunately, many of the electronic proposal pro-
cessing systems are in their infancy and require more time and effort than antici-
pated. As with the case of new programs, multiple data entry fields occur and
some of the bridges/crosswalks between computer systems are not understood or
even retrievable. Therefore, plan well in advance and test the sponsors proposal
processing system to avoid unnecessary stress in order to successfully meet the
proposal deadline.
Title Page
A title page provides useful contact information and a basic description of the pro-
posal. Generally, a title page includes a descriptive, yet succinct title for the proj-
ect. The title of the project will establish an immediate image in the reviewers
mind of what is expected in the narrative. Put some effort into choosing a good
title. The title page also includes the name(s) of the principal investigator, any co-
investigators, office address, phone number and e-mail information, and the name
of the investigators home institution. Additionally, the title page typically con-
tains project start and end dates, and the total funding requested. Note that state
and federal title pages require the name and position of the authorizing official
who assumes legal responsibility for assuring the contents of the proposal adhere
to state and federal statutes governing grants and contracts. The title page is ei-
ther hand or electronically signed by the authorizing official prior to submission
to the sponsor. Authorizing officials generally reside in the research or grants
office, but on occasion may be the chief executive officers of the institution
or agency.
Abstract
The abstract, or proposal summary, is in some ways the most important text of the
entire proposal. It is often the only section of the proposal that all reviewers read.
The abstract serves as a summation of the project. It includes a description of what
problems the researcher is trying to address, why the project is significant, and
what impact the project will have if funded. Abstracts should be no longer than
one page, succinctly written and composed as the last task of the writing phase in
order to capture a full summation of the entire project. Abstracts should include
the research questions and procedures used to address the questions, and should
be written in terms understandable to anyone. If the proposal is funded, the ab-
stract can be published verbatim and many sponsors release the abstract for pub-
lic announcements about the project.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 360
EXAMPLE
Abstract
Below is an example of an abstract.
The purpose of this study is to investigate therapeutic exercises to reduce the
common ailments and foot pain in females diagnosed with plantar fasciitis, an in-
flammation caused by excessive stretching of the fibrous tissue that runs along the
bottom surface of the foot and, when excessively stretched, can lead to plantar
fasciitis that results in heel pain, arch pain, or heel spurs. The study will explore
therapeutic exercises for improvements in gait to increase participants exercise
length and strength. Although several studies conducted by leading podiatrists
(list authors names in parenthesis) show multiple strategies of stretching, the use
of orthotics, low-dose pain medication, and weight control as effective treat-
ment for managing plantar fasciitis, significant numbers of patients continue to
complain of pain and therefore reduced exercise duration and selection. This
study will evaluate the efficacy of diet, massage therapy, and yoga on pain reduc-
tion and increased length and strength for women golfers who regularly walk 18
holes of golf and who have been diagnosed with plantar fasciitis. At least 50 fe-
male participants will be enrolled in the study, each at least 15 lbs overweight, be-
tween 45 and 50 years old, active in the sport, and interested in a lifestyle change
of dieting, foot massage therapy, and yoga. The study will take place in upstate
New York and provide valuable information for womens health, sport and fitness
practitioners, and alternative treatments to enhance the quality of life of women
golfers who suffer from plantar fasciitis.
Table of Contents
Currently, most proposals require electronic submission via the World Wide Web
with many sponsors relying on the use of software programs that automatically
enter the applicants information in the table of contents page. The advantage of
using automated forms is that the investigator receives an effective management
tool for following page limitations and staying on task by section headers.
Similarly, the sponsor conserves resources by streamlining administrative respon-
sibilities such as checking page limitations and completion requirements and can
focus on the intellectual merit and quality of the proposal itself.
Not all sponsors use specifically tailored electronic processing programs for
proposal submissions. In developing a nonelectronic table of contents, authors
should use the same sort of algorithm as encoded into the electronic processors.
The goal is to develop a table of contents that follows sponsor guideline headers
and sections, adheres to the proposal guideline page limitations, and presents
a well-organized chronology that matches the sections of the narrative and
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 361
Project Narrative
The project narrative is the core of the proposal. The narrative section includes all
relevant information used for determining the importance of the proposal, what
particular problem will be addressed, and which procedures will be used to solve
the described problem. The narrative also includes a description of the relevant
prior experience and credentials the investigator has that will enable him/her in
carrying out the proposed project. In total, the narrative section provides the re-
viewer with a description of ambitious, yet achievable goals and objectives and a
sensible management plan to complete the project within the designated time-
frame. Therefore, the goals and objectives must be achievable and not so expan-
sive as to appear overly ambitious in relationship to the timeframe allowed to
accomplish the work. The investigator(s) must also bear in mind their other re-
sponsibilities and not promise more dedicated time to the project than they have
available.
a. Introduction To begin the narrative, the investigator should orient the reader
with one or two introductory paragraphs as space allows to introduce the signifi-
cance of the topic and to establish the credibility of the organization or individ-
ual(s) conducting the research. Since most proposals are submitted by
organizations on behalf of individual investigators, introductory paragraphs can
be used to provide important information about the institution and the context of
the research environment. The introductory paragraphs frame an image of the ex-
isting research environment and the support structure available to assist the inves-
tigator throughout the project. Although not necessarily explicit, in essence the
writer is establishing credibility of the organization. This descriptive portion of
the narrative can be used to suggest that the parent organization is the perfect lo-
cation for the performance of the research. Note the distinction of these para-
graphs from an abstract that is used to summarize the entire proposal.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 362
EXAMPLE
Introductory paragraphs
Here is an example of introductory paragraphs:
The XYZ University was established in 1868 as a teachers college with a
strong emphasis on physical fitness. For over 100 years, XYZ University has pro-
vided national leadership on human performance, especially in the areas of ath-
letic training, kinesiology, and physical education. The faculty at XYZ University
are selectively recruited to advance teaching, research, and service. The faculty to
student ratio is 1 to 12, allowing for individualized instruction and collaborative
research and practicum experiences to prepare each graduate with the skills and
know how of addressing some of todays most plaguing problems confronting
human performance in the twenty-first century. With the escalation and zealous-
ness of human performance lawsuits, XYZ University has been called upon by
National College Athletic Association (NCAA) to provide case law curriculum for
the nations athletic training programs. The NCAA call is intended to reach all ac-
credited athletic training programs to serve as a guide in helping instructors nav-
igate civil and criminal law suits creeping into each of the NCAA divisions. The
curriculum will be designed to guide students in understanding and creating de-
fendable protocols for preliminary diagnostics, treatments, and physician refer-
rals. The protocol documentation and insurance company claims and court
testimonial course sections will include the most recent case law on human per-
formance to guide institutions in best preparing the next generation of athletic
trainers confronting civil and criminal law suits.
specifics should be addressed in the methods section. Lastly, if the research project
addresses regional, state, or federal priorities, refer to those priorities in this section.
In general, the significance section should not exceed two to three pages. The
majority of the proposal narrative should be focused on the method section that is
described later on in this chapter. Rather, the significance section is used to edu-
cate the reader about the identified problem and its importance and further en-
tices the reader to continue to the next section.
c. Rationale Now that the investigator has educated the reader about the significance
of the proposed research, the rationale section should then document the scientific
literature covering the investigators topic. This section typically demonstrates a gap
in the body of knowledge and supports the argument that the proposed investiga-
tion will contribute significant advances to the body of scientific knowledge. Any
criticisms of previous studies or methods when identifying gaps in the literature
should be avoided. Instead, the rationale should be composed in a positive light,
identifying alternatives that may be considered to address unanswered questions.
Given that it is quite possible that the cited references may include previous
publications from the proposals reviewers, the wise investigator will retain a pos-
itive respectful tone in addressing all prior work. Any perceived weaknesses, lim-
itations, or inaccuracies, should be couched as interesting or provocative questions
remaining to be addressed. For example, imagine a reviewer having devoted years
to researching a problem, arduously pursuing alternative solutions and serving on
a proposal panel with peers who have done similar work and made substantive
commitments to research. The proposal being reviewed and currently under dis-
cussion in their review group references their life careers work and provides a con-
vincing summation that their approach is limited, misguided, and has done little
to advance the understanding of the problem. The reviewers discussion group col-
leagues, also considered top experts in the field, have never met the investigator
who purports that an alternative strategy deviating from commonly accepted prac-
tices by the reviewers is methodologically superior and is certain they know the
right questions and research and have a clearer vision of the solution. From this
example, one might surmise that there would be a lack of enthusiasm for the pro-
posal, even if the limitations and criticisms detailed are true.
To navigate the political minefield of proposing improvements to prior work,
the writer should still cite the accepted study(s), but carefully emphasize how the
strengths of the previous study piqued their interest and reinforced their commit-
ment for further investigation. Additional research should hopefully be able to
build upon the accepted work of prior experts and lead the science in a new direc-
tion. The writer can achieve the intended result of identifying the limitations of
prior work but accentuate the positive while doing so. In this manner, all persons
are given credit for their contributions and the review team can discuss the next
phase of the research or an interesting alternative that may yield new findings
rather than dwell upon the criticism of an investigator.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 364
As previously stated, the federal government allocates over $400 billion annually
for US research and development. According to the OMB, approximately 80% of that
federal funding goes directly to states for managing their priorities and coordinating
competitions based on joint priorities. As a result, state and federal government re-
ports are considered credible sources of citation information and help frame the sig-
nificance of the proposed research on a state or national level. Effective and timely
citations may advance the reviewers desire to fund a project to meet parallel pur-
poses: that of their own mission along with the states or nations priorities.
As a general rule, the rationale and literature cited within this section should
not exceed two to three pages. Although important, the rationale section is not the
most important section of the proposal. The emphasis of the proposal should be
directed toward the methods section as outlined below.
d. Methods Section The methods section is the heart and soul of a proposal. This sec-
tion provides readers with a step-by-step process of solving a particularly important
problem. The methods section is a merger of traditional academic writing paired
with business and economic writing to create a feasible plan of action that has a high
probability of achieving important results. The methods section, like a business plan,
is not reflective like the rationale section, but rather an active line of inquiry that ar-
ticulates in great detail the investigators plan. This section is the most lengthy in the
entire proposal and requires a logical order to educate the reader. Once again using
a scaffolding concept is recommended, introducing the basic concepts initially, and
then increasing the complexity as the reader progresses through the text.
The methods section should be introduced by identifying the research ques-
tion(s) or study objectives. The questions or objectives should be accented or high-
lighted using bold font, indentations, numbering, or underlines. This allows the
reviewers to quickly analyze the overall goals of the project, and then permits
rapid scanning of the subheaders to provide an initial impression of the general
cohesiveness, organization, and feasibility of the study.
Clarity and brevity are key when writing research questions or objectives.
Impractical or highly theoretical questions and goals will appear unfocused.
Rather, specificity, practicality, and the use of active tense help the reviewer em-
brace the questions and objectives as important, valid, and timely.
EXAMPLE
1. What impact, if any, has repeated shoulder injuries had on the quality of
life of retired NFL players between the ages of 55 and 60?
2. What pain treatments, if any, are these same retired NFL football players
using to maintain their quality of life after having experienced repeated
shoulder injuries during their NFL career?
3. What impact, if any, has self-prescribed pain treatment had on the health
and physical functionality of these same retired NFL football players?
4. What impact, if any, has physician-prescribed pain treatment had on the
health and physical functionality of these same retired NFL football players?
5. What impact, if any, has repeated shoulder injuries had on the psycho-
logical and social aspects of the lives of these same retired NFL football
players?
Note how each successive question builds upon the previous question logically
and how the overall set of questions appears manageable during a typically funded
1- to 3-year study. Also note that the questions are composed in a manner such that
several answers may result, rather than a simple positive or negative response.
EXAMPLE
Once the research questions are identified, subsections can be created using
the research questions as headers. These subheaders keep both the investigator
and the reader focused. Subheaders may be included in the table of contents.
Regardless, the discovery process and methodology must be detailed and de-
scribed to the reviewer.
Draw upon sound scientific methods that have been confirmed as accurate,
valid, and reliable in similarly structured studies. For example, if the study typi-
cally uses an experimental group and a control group, then the proposed study
should reference prior studies and offer a rationale for the number of participants
that will yield statistically reliable data for analysis and interpretation of the re-
sults. If a study is a longitudinal analysis of a specified group and follows up with
focus group analysis, then the investigator should reference prior work of similar
size and scope and alert the reader that control questions will be embedded in the
overall survey to assure reliability of data. If available, the help of a statistician
should be enlisted to ensure that the study size and design is powerful enough to
ensure statistical significance, and that the analytical methods are properly suited
for appropriate analysis of the specific type of data collected.
With all methodological procedures, be straightforward and honest about any
particular methodological limitations. Reviewers appreciate knowing the limita-
tions and may already have forecasted challenges with particular procedures be-
fore reading the entire methods section. The effort to describe any weaknesses in
advance may actually illustrate to the reviewer that the investigator has antici-
pated other potential problems and issues ahead of time. It is essential to describe
the contingencies in place for dealing with the known limitations and assurances
should be made to the reviewer that these limitations do not affect the overall
study integrity. One method of illustrating this concept uses parallels of prior
work that included similar limitations with little impact. In this manner, the inves-
tigator can reference how other projects confronted the same issue and used alter-
native measures to alleviate the difficulties. Specific citations and references must
be used. If the investigator is proposing a new methodology for simplifying ear-
lier work or developing a new instrument, it is important that some pilot data and
a description of preliminary work, design, or studies is provided. Pilot or prelim-
inary data offers the reviewers some evidence that preliminary investigations are
promising and that further study may yield important information.
It is important to stay focused and committed to the research plan. The meth-
ods section must describe in detail the investigators proposed solutions to the re-
search questions. All pertinent information supporting the described procedures
and methods used must be included. For example, necessary information includes
listing data encoding software packages, a description of the actual encoding pro-
cedures, the detailing of how validity scores are assigned within the method used,
the expected time duration of the various subsections along with the anticipated
deliverables or goals from each subsection, and the assignment of responsibilities
for each subsection to various members of the team. Using a well-organized outline,
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 367
ample spacing, diagrams, charts, and a well-illustrated timeline will not only
greatly assist reviewers in the analysis and interpretation of the text, but will also
create appeal in the attractiveness of a well thought out, well-organized, concise,
and focused proposal.
Lastly, an investigator must be realistic in the self-evaluation of strengths,
skills, weaknesses, and deficiencies. Ones strengths must be emphasized, and
ones weaknesses must be supplemented with team members or consultants pos-
sessing expertise in these weak areas. As described above, statisticians are almost
always useful in designing appropriate data analysis algorithms. Other potential
team members include clinicians to augment the research scientist, engineers to
augment clinicians, and so on.
EXAMPLE
recognized study on ABC funded by the Robert Wood Johnson Foundation. This
long, highly successful career has prepared her well for undertaking the current
proposed study. Her specific responsibilities will include oversight of the research
team that is comprised of one research specialist and two graduate assistants. All
administrative components of the grant (purchasing, payroll, budgeting, etc.) will
be supported by a departmental secretary assigned, but not charged, to the pro-
posed grant.
f. Timeline Timelines are effective tools to chronologically organize, order, and se-
quence the events and activities in a proposal. The investigator must be realistic
about the time required for each activity or experiment. Many timelines include
multiple headers including information such as Task, Description of Task, Date of
Completion, Team Leader. Others use headers such as Goal, Objective, Activity,
Timeframe, Person(s) Responsible. Use of a table format that can be created in
Microsoft Word or Excel is highly desirable to easily mesh the generated timeline
with the other charts or data outlined previously in the prior sections.
part of their study personnel to assure unbiased feedback and analysis of the va-
lidity of the study design and its broader impacts. Evaluations should indicate the
expected outputs and outcomes and/or deliverables so that there can be tracking
measurements from the beginning to the end of a study. Outputs, outcomes, and
deliverables also aid the investigator in keeping focused and on track with the
main purpose and significance of the research.
i. References The references section identifies the sources used, quoted, and re-
ferred to throughout the proposal. If only a few references are used, they can eas-
ily be embedded within the text within parentheses. If there are a number of
references, a separate section using a style format acceptable within the discipline
is appropriate. Formats are readily available on the World Wide Web by using any
search engine and searching the words reference style manuals. A list of style
guides and information distinguishing between citations for monographs/books
and electronic material is also available. References should be listed in chronological
order as they appear throughout the proposal narrative. Software such as Endnote
may be used for ease of reference manipulation and storage for future use.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 370
PREPARING A BUDGET
Federal proposals submitted by colleges and universities must adhere to the cost
principles outlined in OMB Circulars A-110, A-21, and A-133. The Federal Acquisition
Regulations refer to the OMB Circulars when costing and pricing out a contract (FAR
Part 33). The Circulars outline the basic cost principles for (i) the organizational sys-
tems institutions must have in place for the acceptance and management of federal
funding (A-110); (ii) the determination of what constitutes an allowable and unallow-
able cost, and what can be budgeted as a direct cost verses and an indirect cost (A-
21); and (iii) what documentation and costing analysis is required for federal
compliance for audits (A-133). The term costing refers to the act of establishing
pricing by activity or by unit of cost. State and local governments must adhere to
Circular A-87 and nonprofit organizations must follow Circular A-122 for their budg-
eting principles. The different circulars have similarities, but also have their differ-
ences. Investigators should follow cost principles appropriate to their institution.
Given the array of regulations, cost accounting standards, and financial
principles involved in federal budgeting, institutions rarely establish different
procedures for budgeting for federal, state, nonprofit, or corporate proposal appli-
cations. Rather, one universal system applies to all proposals. And, because so
many accounting regulations apply to budgeting within organizational contexts,
sponsoring agencies require authorizing officials approval of the budget prior to
the final proposal submission.
The circulars applicable to budget preparation are A-21, A-87, and A-122. These
circulars communicate three fundamental principles for budget development. First,
the funds requested must be allowable. Budgets may not include requests for
gifts, alcohol, unrelated project supplies, advertising, memberships, and other costs
that are not directly related to the project. Second, costs must be reasonable.
Reasonable is defined as what a prudent business person in the industry would con-
sider a normal cost based on fair market value. This concept necessitates that inves-
tigators obtain a minimum of three quotes for equipment items over $5000 and any
unusually large bulk supply purchases (i.e., laboratory cages, reference materials).
Sponsors desire a good value for their money and it is essential to present costs
that are competitive, attractive, and comparable for similar work or supplies.
The following categories are used when preparing budgets:
a. Personnel
b. Employee Benefits
c. Equipment
d. Travel
e. Supplies and Materials
f. Other Direct Costs
g. Indirect Costs
Each of these sections is discussed in the following paragraphs.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 371
a. Personnel
When presenting costs for salaries, investigators should use their own institu-
tional base salary (current salary) as a basis for computation with an added small
escalation factor for cost of living increases each year. If a position is new, the
salary level should be appropriate to the position and the level of expertise (e.g.,
education level and experience) required for the job. Costs budgeted for person-
nel should be realistic and not overinflated. Some types of federal grants, such as
those administered by the NIH, have salary caps, which may be determined via
NIH guidelines and discussion with the grant administrator. Investigators are re-
minded that they cannot devote more salaried time toward a grant than allowed
within the demands of any other responsibilities. For example, academic investi-
gators may be required to devote 50% of their time toward teaching. Therefore,
time available for grant work is capped at 50%, and only 50% of the investigators
salary may be drawn from grant funds. Within the culture of grant giving and
getting investigators are also encouraged to contribute some of their time/effort
toward the project at no charge to the sponsor.
The lead investigator on a project is called the Principal Investigator (PI) or
Project Director (PD), and he/she either charges or contributes a percentage of
effort reflected in the budget. The amount of time allocated to the project is a func-
tion of the PIs other commitments and the amount of involvement required.
Investigators should propose a reasonable amount of dedicated time. What con-
stitutes reasonable is a judgment call. If the scope of the project requires substan-
tive time and effort, then perhaps 30% to 50% effort should be budgeted.
Conversely, if there are other personnel working on the project, the PI should de-
vote less effort to the project. Levels of effort on a project vary depending on the ex-
pertise, role, credentials, and number of additional staff on the project. Questions
about the management plan and personnel for the research project should be di-
rected to the institutions authorizing official. This individual will be able to assess
the required level of staffing that will then appear competitive to a sponsor.
EXAMPLE
Dr. Jane Doe, PI: $65,000 base salary 15% effort $9,750 requested from the
sponsor in year 1
PI: $66,950 (3% escalation of base salary) 15% $10,042 requested in year 2
PI: $68,959 (3% escalation of base salary) 15% $10,344 requested in year 3
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 372
EXAMPLE
Dr. Jane Doe, PI: $65,000 base salary 5% effort $3,250 requested from the
sponsor in year 1
PI: $66,950 (3% escalation of base salary) 5% $3,348 requested in year 2
PI: $68,959 (3% escalation of base salary) 5% $3,448 requested in year 3
Dr. John Miller, Co-PI: $62,000 base salary 5% effort $3,100 requested from the
sponsor in year 1
Co-PI: $63,860 (3% escalation of base salary) 5% $3,193 requested in year 2
Co-PI: $65,776 (3% escalation of base salary) 5% $3,288 requested in year 3
Senior Research Associate (SRA): $40,000 base salary 100% $45,000 re-
quested in year 1
SRA: $41,200 (3% escalation of base salary) 100% effort $41,200 requested
in year 2
SRA: $42,436 (3% escalation of base salary) 100% effort $42,436 requested
in year 3
In the above illustration, two senior personnel are co-directing the project and
plan to hire a full-time researcher to carry out the daily activities of the project.
Less seasoned investigators often have less staff and devote more time to a proj-
ect than a senior PI or Co-PI. Regardless of the level of experience of the various
members of the research team, salary budgeting needs to be appropriate and real-
istic as compared to the scope of the project.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 373
Many reviewers already have unofficially established a project cost target well
before reviewing the budget. Therefore, budgets should be cost-effective. If the
personnel costs are unrealistic, the proposal may be rejected.
b. Employee Benefits
In this section, investigators should use the percentage rates that are negotiated by
their Human Resource Office and multiply the percentage(s) for the salary listed
under personnel costs. Sometimes different benefit rates apply to different job ti-
tles or employee union classifications. Like personnel costs, the employee benefit
rates typically include a small escalation factor for the cost of inflation for each ad-
ditional year of the project. If an investigator works for a small organization with-
out a Human Resource Officer, the investigator can apply a pro-rated amount for
benefits that is based on the insurance plans offered within the organization for
health coverage, dental, eye, or workmans compensation programs. In this cir-
cumstance, it is best to fully detail the calculation process used to determine the
benefit percentage applied to the project. The description of computation and
types of benefit coverage should be described in the budget narrative and must be
attested by a certified accountant.
EXAMPLE
Employee benefits
In either scenario above, employee benefits consist of a percentage rate that is ap-
plied to a salary in the following manner:
Dr. Jane Doe, PI: $65,000 base salary 15% effort $9,750 35% benefits
$3,413 requested in year 1
PI: $66,950 (3% escalation of base salary) 15% $10,042 35% benefits
$3,515 requested in year 2
PI: $68,959 (3% escalation of base salary) 15% $10,344 35% benefits
$3,620 requested in year 3
c. Equipment
Equipment is defined by the federal government (and most state governments) as
a cost over $5000 and a useful life of 2 years. Any purchase under a $5000 thresh-
old is considered supplies and should be listed within the supply budget line. As
stated earlier, equipment purchases require vendor quotes to assure the sponsor
that they are receiving the best possible rate for equivalent merchandise.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 374
d. Travel
Travel costs include funds needed for airfare, ground transportation, car rental,
taxis, hotel, meals and general incidentals (M&IE). General incidentals are consid-
ered gratuity costs that are incurred for tipping cab drivers, baggage handlers, hotel
room service providers, and dining staff. The U.S. General Services Administration
(GSA) is the official federal agency that publishes the maximum lodging and per
diem rates that can be charged to a federal grant or contract. The GSA website is lo-
cated at http://www.gsa.gov/Portal/gsa/ep/home.do?tabId0. The GSA rates for
lodging and M&IE should be applied to all travel expenses listed in the budget.
To obtain reasonable airfare costs, the investigator should compare airline ticket
prices using travel websites or by contacting a travel agent. Since most awards are
received several months after a proposals submission, investigators should use a
small escalation factor for airline tickets to take into account cost increases over time.
EXAMPLE
Travel costs
An example of how to present travel costs follows:
New York City, NY to Ocean City, MD (1 traveler) total $978 per traveler
Airfare from NYC to Salisbury, MD roundtrip $366
Per diem @ $149/day (per diem includes $85 lodging $64 M&IE allowance)
3 days $447
Car rental @ $55/day for 3 days $165
with partnering institutions, consultants, facility costs such as hotel rent for host-
ing a conference outside of their institution, participant costs for human subjects,
and so on. When budgeting for subcontracts, the total cost of the subcontractor
should be listed in one budget line, accompanied by a note indicating that the in-
dividual budget of the subcontractor is attached. Subcontracts are used to engage
collaborators who are considered essential to the scope of work and without
whom the project could not be completed. Because subcontractors may have their
own budgeting principles (based on whether they are a nonprofit, corporate, edu-
cational, or medical institution), they prepare their own budget for reviewers
analysis. In general, the PI will reach an agreement with a subcontractor during the
initial phase of budget construction, and then allow the subcontractor to determine
the specific financial details as long as they follow the guidelines of the sponsor.
g. Indirect Costs
Once all of the direct costs have been identified in sections a to f above, indirect costs
are applied to the above. Indirect costs are also called facilities and administrative
(F&A) costs. The F&A costs for institutions that conduct significant amounts of re-
search are negotiated between the institution and a federal agency that specializes in
reviewing cost accounting formulas for an organizations general operating costs.
The general operating costs include utilities (e.g., lighting, heat, and air condition-
ing), snow removal, library resources, administrative and budgeting support, space,
secretarial and janitorial services, computer networking, liability insurance coverage,
and other logistical and physical plant services. Every 4 to 5 years F&A agreements
are established between research institutions and the federal government and a per-
centage rate is added to the overall budgetary costs of personnel, fringe benefits, travel,
equipment, supplies and materials, and other costs for contracted services. Typically,
costs exceeding $25,000 of a subcontract, participant support costs that support the
attendance of individuals at conferences or workshops, and the amount equivalent
to tuition and fees are excluded from indirect cost calculations. Once the subcontract-
ing costs, participant support costs, and tuition and fees have been subtracted from
the total direct cost, the remainder of the costs is multiplied by the F&A rate to de-
rive the indirect costs included in the budget. A sample budget is given in Table 20-1.
Investigators should submit a budget narrative that details the costs outlined in
the budget to help reviewers understand how costs are derived and how the project
will be managed to assure alignment of the project objectives with the expenditures.
CHAPTER SUMMARY
Getting funded is an honor and a privilege. Once the process is understood, the
most important part of grant writing is convincing others that the proposed inves-
tigation is relevant, timely, important, and within the investigators capabilities in
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 376
B. EMPLOYEE BENEFITS
C. EQUIPMENT
D. TRAVEL
F. OTHER
H. INDIRECT COSTS
order to advance the human, animal, and global condition. The concept of re-
search through grant award is founded on the principle of doing good for hu-
manity. An investigators ability to collaborate with sponsors through relationship
building and conveying a convincing argument using sound evidence is the
recipe for successfully obtaining funding. In essence, grants provide funding for
leaders and visionaries who seek change. Leadership is an important responsibil-
ity that includes the widespread sharing of advances so that the greatest number
of individuals can learn and benefit from work supported by federal, state, corpo-
rate, and nonprofit sponsors.
KEY POINTS
Grant writing and the concept of research through grant award are founded
on the principle of doing good for humanity.
Collaborating with sponsors through relationship building and conveying a
convincing argument using sound evidence is the recipe for successfully ob-
taining funding.
Grants provide funding for leaders and visionaries who seek change.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 378
Applying Concepts
1. Consider reasons why leadership is an important responsibility that in-
cludes the widespread sharing of advances so that the greatest number
of individuals can learn and benefit from work supported by federal,
state, corporate, and nonprofit sponsors.
2. Offer examples of how the concept emphasized above applies to health
care and clinical practice.
3. Consider important ethical issues associated with grant writing and
seeking funding from sponsors for advancement in health care and clin-
ical practice.
4. Discuss the culture of grants.
5. Offer examples illustrating ways to search for funding sources for a
specific clinical issue of your choice.
6. Consider key points that you would need to include if you were to
write a competitive grant proposal for the issue you identified above in
question #5.
REFERENCES
Bauer D. The How to Grants Manual: Successful Grantseeking Techniques for Obtaining Public
and Private Grants. 2nd ed. New York: Atheneum; 1984.
Britt R. Universities report stalled growth in federal R&D funding in FY 2006. INFOBRIEF,
Science Resources Statistics, NSF 07-336. 2007.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 379
Pallister A. Magna Carta: The Heritage of Liberty. London: Oxford University Press; 1971.
The Foundation Centers Statistical Information Service. [http://foundationcenter.org/findfun-
ders/statistics/pdf/02_found_growth/03_05.pdf]. New York, NY: The Foundation
Center [Producer]; 2007.
The Foundation Centers Statistical Information Service. NIH issues call for perspectives on
peer review. Federal Grants & Contracts Weekly. 2007;3(13).
SUGGESTED READING
1. Bauer D. How to Evaluate and Improve Your Grants Effort. 2nd ed. Westport, CT: The
American Council on Education and The Oryx Press; 2001.
2. Bauer D. The How to Grants Manual: Successful Grantseeking Techniques for Obtaining
Public and Private Grants. 6th ed. Westport, CT: Praeger Publishers; 2007.
3. Bowers L. Physical Educators Guide to Successful Grant Writing. Reston: National
Association for Sport and Physical Education; 2005.
4. Markin K. The mysteries of budgeting for a grant. The Chronicle of Higher Education.
2005;May 27:C2C4.
5. McVay BL. Proposal that Win Federal Contracts. Woodbridge: Panoptic Enterprises; 1989.
6. Molfese V, Cerelin J, Miller P. Demystifying the NIH proposal review process. The
Journal of Research Administration. 2007;38:127134.
7. Orlich D. Designing Successful Grant Proposals. Alexandria, VA: Association for
Supervision and Curriculum Development; 1996.
8. Quinlan Publishing Group. Writing the Winning Grant Proposal, A Quinlan Special
Report.[Report]. Boston, MA: DiMauro J [Managing Editor].
9. U.S. General Services Administration website was reviewed on February 23, 2008,
available at http://www.gsa.gov/Portal/gsa/ep/home.do?tabId0.
LibraryPirate
97689_c20.qxd 06/07/10 11:53 AM Page 380
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 381
PART V
INTEGRATING
EVIDENCE-BASED
MEDICINE INTO THE
EDUCATIONAL
EXPERIENCE
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 382
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 383
CHAPTER 21
EVIDENCE IN LEARNING
AND TEACHING
CHAPTER OBJECTIVES
After reading this chapter, you will:
Recognize the elements necessary for successful learning and teaching in an evidence-
based curriculum.
Understand that clinical research often does not provide the black and white answers
we seek.
Know that skill development is fostered by discussion and constructive feedback.
Learn why making decisions about individual patients is the central focus of clinical
epidemiology.
Be able to explain why the reality of clinical practice is that most cases involve common
diagnoses with patients responding to conventional interventions and natural history.
Understand how learning and teaching from a body of evidence requires active pursuit
of the best available evidence by the student.
Know why the case analysis must not focus on the decisions of the individual care
provider.
Understand that often the evidence is insufficient to fully guide clinical practice.
383
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 384
KEY TERMS
active learning credentialing exam framework
clinical rotations critical appraisal outcomes
cognitive learning theory evidence hierarchy problem-based learning
INTRODUCTION
Learning to practice evidence-based health care requires more than classroom
teaching focused on clinical epidemiology. Such a course and the contents of this
book provide only a framework on which talented professionals-in-training will
build upon in their quest to integrate the best available evidence in their practices.
The next generation of practitioners-in-training must develop the ability to iden-
tify and critically appraise clinical research; and, these learning objectives should
be introduced in texts as we have strived to do, and in entry-level courses. If, how-
ever, such critical appraisal ends with a course rather than becoming a common
theme across a professional curriculum, the graduates are not likely to be well pre-
pared to practice evidence-based health care.
The challenges of learning and teaching in a developing curriculum that em-
braces evidence-based practice are not trivial. To build upon the preparation in
clinical epidemiology in advanced courses, time for learning and teaching clinical
epidemiology must be allotted early in the curriculum. Preparation in inferential
statistics may help students more readily grasp the nuances of probability, risk,
odds, and likelihood that permeate discussions related to diagnostic testing, prog-
nostication, and intervention outcomes. The greater challenges, however, usually
lie in building upon a foundation in clinical epidemiology in clinical courses. One
of the greatest challenges we have faced is the desire of students to have the right
answer and know how to examine, evaluate, diagnose and treat patients. Some
are better than others at coping with the uncertainties uncovered after careful con-
sideration of the myriad of decisions made daily in clinical practice. Testing strate-
gies used to evaluate student performance that ask for the correct answer, while
easing the burden of assessment, only add to the stresses emanating from uncer-
tainty. Katz (2001) wrote that the life span of medical facts is short and shorten-
ing further all of the time, certainly not comforting news to the student working
hard to learn what the current facts are.
Uncertainty is a stressor on those who teach, as well as those who seek to
learn. How does one approach, for example, teaching students the best way to
learn an examination technique that they will likely be expected to perform on
clinical rotations and perhaps be held accountable for on a credentialing exam
when the best available evidence suggests the technique is of little use in effec-
tively treating patients? How does the dynamic of a shared learning environment
unfold when a well-informed student presents evidence that is contradictory (and
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 385
stronger) to that offered as the instructor? How does an instructor assist the stu-
dent caught between the best available evidence and a clinical instructor/precep-
tors order to practice like I have shown you? Lets face it, it is easier for
instructors to be authoritarian, to ask students to identify correct answers and
strive to become replicas of themselves as faculty. And, many students are both ac-
customed to and comfortable with such an approach to learning. In this approach,
however, we drift toward expert opinion, falling toward the bottom of the evi-
dence hierarchy, thus failing ultimately as practitioners.
The success of a curriculum, students, and a faculty embracing evidence-
based health care requires humility, a willingness to address rather than hide from
conflicting conclusions and insufficient data, and the time to keep abreast of the
best available evidence. Students must be prepared to appraise the clinical litera-
ture, empowered to debate the issues, and encouraged to view uncertainty as a
challenge rather than a threat. Idealistic goals perhaps, but so is the goal of the
universal practice of evidence-based health care optimizing the care of each pa-
tient. So the students, the educators, and the clinicians missions merge and it all
begins in a classroom.
CONCEPT CHECK
The success of a curriculum, students, and a faculty embracing evidence-based health
care requires humility, a willingness to address rather than hide from conflicting con-
clusions and insufficient data, and the time to keep abreast of the best available
evidence.
These words of wisdom are perhaps more applicable today than ever before
as the volume of and accessibility to the clinical literature continues to expand.
Contradictory findings are common, as is the need to make decisions as to
whether the results of clinical research can or should be applied to decisions about
the care of individual patients. In other words, the clinical research often does
not provide the black and white answers we seek, thus adding to the sense of
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 386
CONCEPT CHECK
Clinical research often does not provide the black and white answers we seek, thus
adding to the sense of uncertainty experienced by students, faculty, and clinicians.
CONCEPT CHECK
Students working in small groups to address a clinically relevant question often result
in dynamic discussion regarding the strength of the clinical literature. When the groups
report out further discussion ensues allowing a large number of students to learn and
develop their skills. Such a process is often messy and requires faculty to empower stu-
dents and guide rather than direct the activities of the class.
CONCEPT CHECK
Learning and teaching from a body of evidence requires active pursuit of the best
available evidence by the students who, while becoming a labor force, gain the expe-
rience necessary to support a lifetime of independent learning.
PROBLEM-BASED LEARNING
Making decisions about individual patients is the central focus of clinical epidemi-
ology and is really the vehicle that permits the integration of the best available re-
search with clinician experience and patient values in the hospital and clinic
environment. Thus, once students have developed the basic skills of information re-
trieval and critical appraisal, learning through case analysis provides a powerful
means of integrating evidence-based practice into a curriculum. Case- or problem-
based learning (PBL), sometimes referred to as situation-based learning, also effec-
tively integrates the continuum of the patient care process into the learning process.
EXAMPLE
CONCEPT CHECK
Making decisions about individual patients is the central focus of clinical epidemiology
and is really the vehicle that permits the integration of the best available research with
clinician experience and patient values in the hospital and clinic environment. Thus,
once students have developed the basic skills of information retrieval and critical ap-
praisal, learning through case analysis provides a powerful means of integrating evi-
dence-based practice into a curriculum.
Case analysis also offers the opportunity to integrate classroom and clinical
experiences. The use of information regarding patients receiving care or those
recently discharged raises concerns regarding confidentiality as well as opening
the door for criticism of care providers known to students and faculty. Obtaining
permission to use personal information in case analyses is essential when it is
possible that the information may identify the individual patient. Potentially
identifying information can take many forms. Unique diagnoses and injuries sus-
tained in particular incidents may identify a patient as can their profession, posi-
tion on a team, or role in the community. The point is that it is best to secure
permission to discuss a case rather than assuming the steps taken to secure confi-
dentiality will be sufficient. This is especially true when the care was provided in
the local community.
Cases are analyzed to help students learn to make clinical decisions based on
the best available evidence. The reality is that often the evidence is insufficient to
fully guide clinical practice and the case analysis is completed outside of the in-
fluence of clinical experience, patient preferences, and other factors influencing
care. Thus the case analysis must not focus on the decisions of the individual care
provider but rather focus on how the research literature can inform practice when
similar cases are encountered. Moreover, the process is best not confined behind
the closed doors of a classroom open only to faculty and students but rather be in-
clusive of those participating in the case when possible.
One final consideration regarding problem and case-based teaching and
learning that seems to perplex students relates to the nature of the case or prob-
lem of interest. Since most cases presented in the literature are rather unique, often
in the form of a fairly rare diagnosis, students often assume that the cases and
problems they investigate must also be unique. The reality of clinical practice is
that most cases involve common diagnoses with patients responding to conven-
tional interventions and natural history. Common cases offer as many or more
opportunities for critical appraisal of the related clinical research than those
cases that warrant presentation in peer-reviewed settings. While every clinician
strives to identify those patients who warrant additional evaluations to rule out
potentially serious but uncommon conditions, it is the critical review of routine
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 390
practices that will have the greatest impact on the broadest spectrum of patients.
Consider the once common use of superficial heat, ultrasound, and massage to
treat nonspecific low back pain. Certainly, the diagnosis and the interventions are
very common and not worthy of a published case report. However, the review of
a case involving a patient with nonspecific low back pain is ripe for analysis from
perspectives including risk factors and prevention, diagnostic considerations,
and intervention effectiveness. Students exploring such a case might gain a wealth
of understanding of the challenges faced by clinicians in some settings, very
frequently.
CONCEPT CHECK
While every clinician strives to identify those patients who warrant additional evaluations
to rule out potentially serious but uncommon conditions, it is the critical review of rou-
tine practices that will have the greatest impact on the broadest spectrum of patients.
CONCEPT CHECK
The reality is that often the evidence is insufficient to fully guide clinical practice and
the case analysis is completed outside of the influence of clinical experience, patient
preferences, and other factors influencing care. Thus the case analysis must not focus
on the decisions of the individual care provider but rather focus on how the research
literature can inform practice when similar cases are encountered.
assessment must occur along a continuum that begins with a general understand-
ing of clinical epidemiology and the paradigm of evidence-based practice, contin-
ues into information retrieval and critical appraisal, and ends when students
demonstrate the integration of the best available research into their patient man-
agement decisions.
Assessment related to the mechanical issues of data retrieval, methodologic
quality appraisal, and data analysis is relatively straightforward. Students and cli-
nicians must also value and strive to practice evidence-based care, which while re-
fining skills also promotes expansion of the paradigm. Successful integration of
clinical epidemiology and evidence-based care can only be assessed once students
are completing clinical experiences or have graduated and begin practice. Thus,
assessment of the student regarding application of the principles developed in the
classroom must continue into their clinical experiences. Such assessment requires
a planned communication between faculty, clinical instructors, coordinators of
clinical education, and students. Such additional assessment may be viewed as
adding to the burden on clinical instructors already stretched to provide patient
care and quality instruction.
In the big picture evidence-based practice is a new paradigm that many cur-
rent clinical instructors were not exposed to during their training. Over time more
clinical instructors will enter this role with expectations that students will inte-
grate clinical research into their patient care recommendations. This demographic
shift will permit greater on-site assessment of how effectively students function in
an evidence-based practice model. In the current environment, however, some
creativity is required to assess students attempting to adapt to treating patients
in the real world and integrating research evidence into their developing practice
patterns. We have found that case reports and postclinical affiliation debriefings
permit a level of assessment and remediation when necessary to develop skills
within the context of the overall academic mission. The truest test of success
comes from the assessment of practicing alumni. Assessment of alumni can be a
tall challenge due to geographic dispersion and often low return rates on alumni
surveys. Including items related to the value for and practice of evidence-based
health care on such surveys, however, provides data related to how effectively an
academic curriculum impacted upon practice. Recurring assessment from the
classroom to clinical experiences to postgraduation practice also conveys the
measure of importance a faculty and an academic unit place on evidence-based
health care.
CHAPTER SUMMARY
The goal of this chapter was to encourage the use of evidence in the classroom as
we strive to prepare our students with the research skills that are necessary for cli-
nicians in this new millennium of evidence-based medicine (EBM). As American
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 392
medical schools and professional curriculum are moving from providing instruc-
tion toward promoting learning education is shifting towards modes of learning
that require students to apply concepts, follow procedures, solve problems, and
make predictions (Amato, Konin, & Brader, 2002; Clark & Harrelson, 2002). These
methods of learning have evolved from recent developments in the science of
learning and from constructivist theory (Brooks & Brooks, 2001; National
Research Council, 2001/2002). In contrast to traditional approaches to learning
based on rote memory and recall, these contemporary theories of learning empha-
size the way knowledge is represented, organized, and processed. Thus, we are
faced with the challenge of introducing research methods as a framework for ev-
idence-based clinical practice. Part of this challenge is to prepare our students
with the necessary skills for accessing, synthesizing, interpreting, and applying
empirical research findings to determine the best clinical practices when they have
not yet become practicing clinicians.
Research skills must be learned, and to do so, they must be taught and prac-
ticed (Wingspread Group, 1993). These skills include written and oral communi-
cation, information gathering, critical analysis, interpersonal competence, and the
ability to use data and make informed judgments following an evidence-based
best practice model. Spence (2001) states that we will not meet the needs for a
more and better higher education until professors shift to designers of learning ex-
periences and not teachers. The focus in the classroom must shift from how will
I teach this information toward how will my students learn this information.
As instructors of preservice students of physical therapy and athletic training,
we have realized the need to model the use of high-quality learning experiences.
Under this premise, we have delved into PBL, situation-based and case-based
learning to help foster the skills of critical analysis and integration and application
of research findings following hierarchies of evidence (in EBM) among our stu-
dents. We have found that PBL, if done well, can improve learning by promoting
enduring understanding (Wiggins & McTighe, 1998) and help connect the class-
room to authentic learning experiences that increase the transfer of key concepts
to other contexts outside of the classroom.
What has become clearer to us over the past few years is that process is con-
tent. Stated more specifically, our teaching philosophies have matured and de-
veloped around the belief that the learning process is as equally important as the
subject discipline content. For that reason, when we teach research methods
courses there is a strong focus on process skills in addition to traditional research
method content. As part of the evolution from professional lecturer to facilita-
tor of learning, we try to overtly model the following process skills throughout
the course: teamwork, assessment, written and oral communication, technology
integration, critical thinking, and problem solving (SCANS, 1993; NASPE/
NCATE, 2001). We also try to have our students think and write about their per-
formance of these skills in the context of the physical therapy and athletic training
disciplines. These skills are more obviously important to the next generation of
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 393
educators; however, it is important that future certified athletic trainers and phys-
ical therapists be able to peer teach, communicate effectively with the health care
system and parents, locate available community-based resources for patients in-
volved with substance abuse or other psychosocial issues, demonstrate technol-
ogy use, interpret professional literature, and make effective presentations
(Knight, 2001).
Cognitive learning theory and its constructivist approach to knowledge
suggest that we should not look for what students can repeat or mimic, but for
what they can generate, demonstrate, and exhibit (Brooks & Brooks, 1999). As con-
structivists and instructors of future clinical practitioners (e.g., physical therapists
and certified athletic trainers), weve found PBL to be a logical and refreshing in-
structional practice for us to follow when teaching research methods as a frame-
work for evidence-based clinical practice.
PBL is an active learning strategy. Active learning suggests that students
demonstrate what they know and are able to do. Rather than emphasizing dis-
crete, isolated skills, PBL emphasizes the application and use of knowledge in re-
search methods and EBM. In order for PBL and EBM to be effective, we believe
both PBL and EBM must be done well, and that comes with practice by the in-
structor and students alike.
KEY POINTS
One of the greatest classroom challenges is the desire of students to have
the right answer and know how to examine, evaluate, diagnose and treat
patients.
Contradictory findings are common.
Critical appraisal is a skill that is developed with practice over time.
Skill development is fostered by discussion and constructive feedback.
Staying current requires a commitment of time to identify, read, appraise,
and integrate new research.
Students often assume that the cases and problems they investigate must
also be unique.
Obtaining permission to use personal information in case analyses is essen-
tial when it is possible that the information may identify the individual
patient.
The connections between clinical epidemiology and evidence-based practice
must be appreciated when developing instructional strategies and assess-
ment instruments.
Instruction and assessment must occur along a continuum.
Assessment requires a planned communication between faculty, clinical in-
structors, coordinators of clinical education, and students.
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 394
Applying Concepts
1. Consider cognitive learning theory and its constructivist approach to
knowledge. Choose a course (i.e., epidemiology, therapeutic modali-
ties, etc.) and suggest ways that students can generate, demonstrate,
and apply their understanding of course content through case studies
and other problem-based learning strategies.
2. Offer examples of how to efficiently and effectively assess students
basic skills of information retrieval and critical appraisal learning
through case analysis.
REFERENCES
Amato H, Konin, JG, Brader H. A model for learning over time: the big picture. J Athl Train.
2002; 37(4 Suppl): S236S240.
Brooks JG, Brooks MG. In Search of Understanding: The Case for Constructivist Classrooms.
Alexandria, VA: Association for Supervision and Curriculum Development; 1999.
Clark R, Harrelson G. Designing instruction that supports cognitive learning processes.
J Athl Train. 2002;37(4 Suppl): S152S159.
Katz DL. Clinical Epidemiology & Evidence-based Practice. Thousand Oaks, CA: Sage
Publications; 2001:xvii.
Knight K. Assessing Clinical Proficiencies in Athletic Training: A Modular Approach.
Champaign, IL: Human Kinetics; 2001.
National Association for Sport and Physical Education/National Council for the
Accreditation of Teacher Education (NASPE/NCATE). Guidelines for Initial Physical
Education Program Reports: NASPE/NCATE 2001 Initial Physical Education Standards. 3rd
ed. Reston, VA: NASPE; 2001.
LibraryPirate
97689_c21 01/07/10 4:39 PM Page 395
National Research Council. How people learn: brain, mind, experience, and school.
Committee on developments in the science of learning and committee on learning re-
search and educational practice. In: Bransford JD, Brown AL, and Cocking RR, (eds).
Commission on Behavioral and Social Sciences and Education. Washington, DC: National
Academy Press; 2002.
National Research Council. Knowing what students know: the science and design of edu-
cational assessment. Committee on the foundations of assessment. In: Pelligrino J,
Chudowsky N, and Glaser R, (eds). Board on Testing and Assessment, Center for Education.
Division of Behavioral and Social Sciences and Education. Washington, DC: National
Academy Press; 2001.
Secretarys Commission on Achieving Necessary Skills (SCANS). What Work Requires of
Schools: A SCANS Report for America 2000. Washington, DC: US Department of Labor;
1991.
Spence L. The case against teaching. Change Magazine. 2001; 33 (6):1019.
Wiggins G, McTighe J. Understanding by Design. Alexandria, VA: Association for
Supervision and Curriculum Development; 1998.
Wingspread Group on Higher Education. An American Imperative: Higher Expectations for
Higher Education. Johnson Foundation; 1993, pp. 1, 13.
LibraryPirate
97689_c22 01/07/10 4:40 PM Page 396
CHAPTER 22
THE CLINICAL EXPERIENCE
CHAPTER OBJECTIVES
After reading this chapter, you will:
Understand the importance of evidence-based medicine (EBM) in the classroom.
Be able to discuss the role of clinical education in student learning.
Understand the similarities and differences between didactic education and clinical
education.
Be able to describe the concept and practice of grand rounds in clinical education.
Understand the differences between case reports and case studies.
Recognize the three parts of a case study.
Be able to discuss the benefits of case-based learning in clinical education.
Understand strategies to integrate evidence-based medicine into classroom experiences.
Appreciate the continued preparation of clinicians in the practice of evidence-based medicine.
Understand how patients, students, and the public can be educated through grand rounds.
KEY TERMS
case reports evaluation interventions
case-based learning evidence-based medicine outcomes
critical review grand rounds
didactic lecture informed consent
396
LibraryPirate
97689_c22 01/07/10 4:40 PM Page 397
INTRODUCTION
Chapter 21 discussed strategies to help students and instructors integrate evidence-
based medicine into classroom experiences. These experiences build an important
foundation for students. It is through clinical education, however, that students
apply what has been learned, demonstrate professional behaviors, and often come
to appreciate that the real world of health care practice differs from the comfort
of the classroom where greater certainty often exists and mistakes dont affect real
patients. Classroom or didactic education and clinical education are not two un-
related parts of an educational experience. In fact, it is likely that the more these
experiences can be integrated, the richer the educational program will be. Thus,
this chapter serves as extension of Chapter 21 and focuses on the continued prepa-
ration of clinicians in the practice of evidence-based medicine once they have pro-
gressed into clinical education. The timing of clinical experiences for students in
the health care professions varies widely between disciplines and academic pro-
gram. Thus, there is not a single strategy. We have chosen to focus on two com-
mon assignments required of studentclinicians, case reports and grand rounds
but also identify other opportunities to apply many of the principles found in this
text during clinical education.
CASE REPORTS
Case reports, also referred to as case studies, have appeared in many clinical jour-
nals and are frequently required of students in the health care professions. There
is continuing debate as to the values of case reports in todays clinical literature,
although some are valuable additions to the knowledge base. Our purpose is not
to debate the extent to which case reports should be included in the literature but
rather to focus on those student-generated reports that rarely warrant publication.
In this context the case report really becomes a form of case-based learning. In
many instances students are seeing what they have learned being applied to the
care of a real patient.
Case reports may be retrospective or prospective and when prepared by
students either a description of what was observed or what the student did. As the
clinical education experience progresses, the student assumes more responsibility
and thus can prepare a case report prospectively based on their actions. This type
of case report generally requires the student to acquire and review a detailed
medical history, complete and describe a thorough examination, review findings
of additional testing and imaging, and consider the diagnostic possibilities. The
student is then faced with the challenge of developing a plan of care that is agree-
able to the patient. Lastly, the outcome of the case must be described. Regardless
of the students role in the case these reports are usually presented in written
form, sometimes followed by oral presentation.
LibraryPirate
97689_c22 01/07/10 4:40 PM Page 398
CONCEPT CHECK
Case reports may be retrospective or prospective and when prepared by students
either a description of what was observed or what the student did.
EXAMPLE
In posing the questions in the above example, the student and their learning,
rather than an external audience, becomes the focus of the case study. Ending the
case study at the evaluation stage paints an incomplete picture, but is it necessary
for the student to go further? The answer will depend on where the student is
in their training and the context in which the case study is completed. A case
study that serves as a culminating experience likely needs to be further developed
to address interventions and outcomes. For the student early in their academic
preparation focusing on the evaluation of patients the purpose of the case study
has been served and the process ended. We suggest that purpose-specific case
studies are valuable in teaching student to practice evidence-based health care. By
breaking cases into evaluation, intervention, and outcome assessment across an
academic program permits sufficient attention to the issues related to diagnostic
accuracy, efficacy, and effectiveness of interventions and the properties of the
outcomes measurement instruments to truly identify and apply the best available
evidence.
Before leaving the subject of case studies we would like to share additional
observations. The first is that cases that are sufficiently unique to be of interest to
an external audience are relatively uncommon. Those in academia that have been
subjected to culminating case reports can likely attest to the state of boredom that
sets in somewhere in the grading process and the jubilation that accompanies
completion. However, when one asks students to examine the components of a
case and address diagnosis, intervention, or outcomes, the common cases we en-
counter daily provide sufficient fodder because it is not the case but the appraisal
and application of the literature that draws the readers attention. The second ob-
servation is that detailed, well-prepared case studies require a good bit of effort.
Our collective observation is that when a student prepares a case report they often
lose the forest through the trees, therefore losing the opportunity to critically eval-
uate the management of the case and thus clinical practice in general. There is not
a single or simple solution to this dilemma; however, failure to appreciate the time
required to acquire and carefully appraise the salient literature will sentence the
instructor to the continued reading of case reports that are neither interesting nor
generally informative.
CONCEPT CHECK
The practice of evidence-based medicine requires consideration of the best available
evidence rather than the identification of evidence that supports or refutes a particu-
lar position.
LibraryPirate
97689_c22 01/07/10 4:40 PM Page 400
GRAND ROUNDS
In using the term grand rounds we are borrowing from the long-standing prac-
tice of hospitals and other health care facilities of scheduling regular plenary ses-
sions for providers (or subgroups of providers) in the organization. In their
original form grand rounds involved live patient evaluations and interactive dis-
cussion of the diagnostic process and plan of care. The format of grand rounds
has changed markedly over the past few decades and now the presence of
patients is uncommon (Herbert & Wright, 2003). Often grand rounds that were
intended to spark debate and discussion resemble a typical classroom lecture.
The most common format is now a didactic lecture (Mueller et al., 2006). A
single speaker completes a presentation and provides, hopefully, time for ques-
tions and, more importantly, discussion. Moreover, the didactic format tends to
lead the audience through a case rather than engaging in an effort to identify the
best course.
We chose to use the term grand rounds in the context of integrating evidence-
based health care into the educational experience because the critical analysis of
cases often does not occur in typical courses. Moreover, active student participa-
tion in such classes is often limited. We also wish to distinguish what we have
labeled grand rounds from presentations made by students in a classroom as a
component of an academic course. It is not our intention to judge the values of
such presentations but simply point out that most are delivered in front of one or
a few faculty rather than a community of clinicians. In fact, it is the community of
clinicians sharing their knowledge and debating the issues that best distinguishes
events as grand rounds. While the presence of patients is uncommon at grand
rounds today, returning to this format has some attraction especially for students
early in their training with limited patient contact. We also do not believe that
grand rounds be limited to being led by a recognized expert. In fact, we would en-
courage some sessions be led by students. Siri et al. (2007) described the experi-
ences of surgical residents charged with reviewing components of a case
involving the death of a patient. While we hope that our readers will not face this
challenge in the context of the preparation as health care providers, cases with un-
satisfactory outcomes often provide the best opportunities to learn. Regardless of
the outcome of the case, grand rounds create a format where the case defines the
direction of the learning rather than a syllabus or an instructor. Often a clinician
attending grand rounds is quite familiar with the various diagnostic and treat-
ment procedures discussed. However, the pearls of wisdom that enhance our abil-
ities are frequently found at grand rounds. These may be in the form of the
consideration of a diagnosis that had not crossed our mind, a better technique for
treatment, and a better means of assessing treatment outcomes. Filling in the gaps
in what we know is at the core of lifelong learning and grand rounds can be a
wonderful forum for our professional development.
LibraryPirate
97689_c22 01/07/10 4:40 PM Page 401
CONCEPT CHECK
We chose to use the term grand rounds in the context of integrating evidence-based
health care into the educational experience because the critical analysis of cases often
does not occur in typical courses.
Requirements
Clinician Support
The first piece to developing a system of regularly scheduled grand rounds that
will impact student learning is the support of clinicians associated with the aca-
demic program and/or the community. Certainly, faculty members are welcome
but unless the faculty member sees patients on a regular basis, they will not bring
fresh cases to the table.
The Cases
The next pieces to grand rounds are the cases. Circumstances will influence whether
patients are present for all or a portion of the session or whether someone will be
charged with presenting the case. In all circumstances steps must be taken to pro-
tect the confidentiality of medical information. Patients should be informed of the
process if they are to appear and provide informed consent for their information to
be discussed. The participants in the grand rounds should hold the information
they have about the patient with the same rigor as that of a patient seen during rou-
tine clinical care. If a case is presented by a clinician, permission to discuss the case
should also be sought. While it is possible to present a case without revealing the
identity of the patient, the more unique the case and the smaller the practice com-
munity, the more likely confidentiality will be breached. It is also likely that efforts
to maintain anonymity will lead to some failure to present information germane to
the case. Thus, we recommend obtaining informed consent from the patient that is
the subject of the case regardless of whether they will be present.
LibraryPirate
97689_c22 01/07/10 4:40 PM Page 402
CONCEPT CHECK
The participants in the grand rounds should hold the information they have about the
patient with the same rigor as that of a patient seen during routine clinical care.
Other Considerations
Professionalism, mutual respect, and the quest for knowledge that defines prac-
tice are more critical to the success of grand rounds than rules and structure. The
focus is on learning rather than teaching and while senior clinicians often guide
those with lesser experience and expertise, every effort should be made to avoid
top-down presentation of information characteristic of many classrooms found in
health care education programs.
Case-based learning is not a new concept. The quest for the best available ev-
idence, however, is not always evident especially when cases are recycled in
course materials. Case reports and grand rounds foster the pursuit of the best
available evidence and help students identify what they dont know as opposed
to a structured course where efforts are made to help them learn what the faculty
believes they should know. As educators strive to prepare students to practice ev-
idence-based health care and apply the contents of this text, these educational
strategies can prove quite useful.
CONCEPT CHECK
Case reports and grand rounds foster the pursuit of the best available evidence and
help students identify what they dont know as opposed to a structured course where
efforts are made to help them learn what the faculty believes they should know.
LibraryPirate
97689_c22 01/07/10 4:40 PM Page 403
CHAPTER SUMMARY
Grand rounds are a ritual of medical education, consisting of presenting the med-
ical problems and treatment of a particular patient to an audience consisting of
doctors, residents, and medical students. The patient is usually present for the
presentation and may answer questions. Grand rounds have evolved consider-
ably over the years, with most current sessions rarely having a patient present and
being more akin to lectures.
KEY POINTS
The practice of evidence-based medicine requires consideration of the best
available evidence rather than the identification of evidence that supports or
refutes a particular position.
Case reports may be retrospective or prospective and when prepared by
students either a description of what was observed or what the student
did.
The literature review component of case reports is a critical juncture in the
integration, or lack thereof, of evidence-based medicine in the academic
experience.
The format of grand rounds has changed markedly over the past few
decades and now the presence of patients is uncommon.
Grand rounds create a format where the case defines the direction of the
learning rather than a syllabus or an instructor.
Grand rounds can foster the development of professional behaviors partially
because participants are taking responsibility for their continuing education
rather than being compelled to participate in order to receive a grade or meet
a requirement imposed by an employer.
In all circumstances steps must be taken to protect the confidentiality of
medical information.
Each case should provide or stimulate a quest for the best available
evidence.
Much like a good research paper, the value of grand rounds often is found
in the pursuit of what is left unanswered rather than in the answers
provided through the event.
Case reports and grand rounds foster the pursuit of the best available
evidence and help students identify what they dont know as opposed to a
structured course where efforts are made to help them learn what the faculty
believes they should know.
LibraryPirate
97689_c22 01/07/10 4:40 PM Page 404
Applying Concepts
1. Consider various clinical and health care situations in which steps
must be taken to protect the confidentiality of medical information.
Discuss potential challenges and ethical considerations for the respon-
sible conduct in clinical practice (refer to Chapter 5, if necessary).
2. Discuss the educational and medical significance of grand rounds in
teaching and learning from an evidence-based practice perspective.
REFERENCES
Herbert RS, Wright SM. Re-examining the value of medical grand rounds. Acad. Med.
2003;78:12481252.
Mueller PS, Segovis CM, Litin SC, et al. Current status of medical grand rounds in depart-
ments of medicine at US medical schools. Mayo Clin Proc. 2006;81:313321.
Siri J, Reed AI, Flynn TC, et al. A multidisciplinary systems-based practice learning experi-
ence and its impact on surgical residency education. J Surg Educ. 2007;64:328332.
LibraryPirate
97689_Gloss 6/30/10 12:45 PM Page 405
GLOSSARY
Note: The more basic key terms found in the chapters are not defined in this
glossary. For these definitions, please consult Merriam-Websters online diction-
ary (http://www.merriam-webster.com/dictionary/).
A
active learning suggests that students demonstrate what they know and are able to do
analysis of covariance (ANCOVA) a special case of ANOVA where a variable is
introduced for the purpose of accounting for unexplained variance. ANCOVA
increases statistical power (chance of rejecting the null hypothesis)
a priori power analysis estimation of an appropriate sample size for a study is done
before the study takes place
B
Belmont Report is a written report that originated from the Department of Health,
Education and Welfare (HEW) which concentrates on and lays down the guide-
lines for the protection of human subjects used in research
bench research is often thought of as being conducted in a laboratory environ-
ment under tightly controlled conditions. Also known as basic science
beneficence is the ability of the practitioner to secure and stabilize the condition
or well being of the client while they are receiving treatments or involved in a
research project. It can also be understood as acts of kindness, charity, or com-
fort to the individual client that go beyond their normal obligation to the client
benefit versus risk a written document that states the benefits and the risks of a po-
tential activity. The benefits must outweigh the risks to be useful
biomedical research a vast and diverse field where questions related to the func-
tions of the body, disease, responses to medications, injury mechanisms, and
disease and injury patterns are addressed
blinding is a technique where the subjects, members of the experimental team, or
clinicians are not fully aware of a certain component of a study. An example is
if an individual is a control or part of experiment
Boolean search is a search based on the notion of logical relationships among
search terms. Specifically, from a computer programming perspective, the op-
erator terms of OR, AND, and NOT effect strict logically collated out-
come or search results. Stated differently, each of the operator terms (i.e., or,
and, not) used to combine search terms instructs a different operation or set of
search directions for the computer to follow
405
LibraryPirate
97689_Gloss 6/30/10 12:45 PM Page 406
406 Glossary
C
call for abstracts in order to present research findings at a professional meeting
the investigators must respond to a call for abstracts from the organization
sponsoring the meeting and submit an abstract for review
case reports a detailed report of the diagnosis, treatment, and follow-up of an
individual patient. Case reports also contain some demographic information
about the patient (for example, age, gender, ethnic origin) (1)
case-based learning case reports and grand rounds foster the pursuit of the best
available evidence and help students identify what they dont know as opposed
to a structured course where efforts are made to help them learn what the fac-
ulty believes they should know
CINAHL electronic database
clinical epidemiology what can be done to help prevent or treat a patients condi-
tion (disease/injury) in a clinic or office
clinical practice guidelines systematically developed statements to assist practi-
tioner and patient decisions about appropriate health care for specific clinical
circumstances
clinical prediction guides also known as clinical prediction rules. Prefer the term
clinical practice guide, as opposed to rule, since the true purpose of these re-
ports is to guide, rather than dictate, clinical decisions
clinical prediction rules developed from a cluster of exam findings or characteris-
tics and may assist in the evaluative or treatment phase of patient care
clinical research research that either directly involves a particular person or group
of people or uses materials from humans, such as their behavior or samples of
their tissue, that can be linked to a particular living person (2)
clinical rotations a period in which a medical student in the clinical part of
his/her education passes through various working services (3)
Cochrane collaboration an international not-for-profit and independent organiza-
tion, dedicated to making up-to-date, accurate information about the effects of
health care readily available worldwide. It produces and disseminates system-
atic reviews of health care interventions and promotes the search for evidence in
the form of clinical trials and other studies of interventions (4)
cognitive learning theory suggests that we not look for what students can repeat
or mimic, but for what they can generate, demonstrate, and exhibit
conceptual framework a group of concepts that are broadly defined and system-
atically organized to provide a focus, a rationale, and a tool for the integration
and interpretation of information (5)
control biases methods of studies of prevention, treatment and diagnostic proce-
dures, and steps that minimize investigational bias
correlations between variables the methods and statistical analysis of research into
relationship
LibraryPirate
97689_Gloss 6/30/10 12:45 PM Page 407
Glossary 407
D
data information collected through observation and/or experimentation, and
later scrutinized through a series of statistical analyses to determine results.
database electronic libraries of indexed journals, books, and nonjournal biblio-
graphic literature and documents that are overseen, managed, and updated on
a regular basis
dependability in research is associated with a judgment of the repeatability of re-
search findings. If findings are repeatable, then they are considered to be reliable.
In qualitative research, the notion of reliability is referred to as dependability.
Also associated with consistency
dependent variable something that is measured by the researcher
diagnosis the determination of the nature of a disease, injury, or congenital
defect (7)
diagnostic continuum the differences between research into diagnostic testing and
prevention or treatment strategies
didactic lecture a single speaker completes a presentation and provides hopefully,
time for questions and more importantly, discussion
disablement model an evaluation and treatment model based on specific impair-
ment, functional loss, and attainable quality of life rather than a medical diag-
nosis (8)
disease-oriented measures measures that provide insight into the physiology of
illness or injury. DOE measures provide information about a patients pathol-
ogy and are of most interest to health care providers, as opposed to being im-
portant to patients
LibraryPirate
97689_Gloss 6/30/10 12:45 PM Page 408
408 Glossary
E
empirical research a methodological approach to problem solving in which deci-
sions are based upon findings and conclusions established as a result of analyz-
ing collected data about variables and the relationships that connect them. A
process or set of ideas used for asking and answering questions
epidemiology the study of why disease and injury occur in a population
estimates of sensitivity and specificity the ability of a test to identify those with and
without a condition and are needed to calculate likelihood ratios
ethical issues the contrast of appropriate and inappropriate behaviors, and are
based on personal, moral, professional (i.e., cultural), and societal views of ac-
cepted, principled criteria, and guidelines for responsible conduct
ethnography focuses on the study of the group and their culture. Closely associ-
ated with the fields of anthropology and sociology. The studying of cultural or
cultural groups, but it is not limited to this inquiry and includes topics of inves-
tigation such as office settings/corporate culture, youth sport groups, cults,
community-based groups, feminism/sexism in communities or institutions,
policy application and delivery, and racism
evidence hierarchy rank reviews based on the strength of the argument made and
the evidence given
evidence-based medicine (EBM) the process of applying relevant information
derived from peer-reviewed medical literature to address a specific clinical
problem; the application of simple rules of science and common sense to deter-
mine the validity of the information; and the application of the information to
the clinical problem (7)
evidence-based practice the process by which decisions about clinical practice are
guided from evidence in research based on scientific models and theoretical
paradigms
external validity relates to how generalizable the results of a study are to the real
world
F
false negative results test results may fail to detect pathology when it is present
false positive results test results may suggest pathology is present when it is not
G
generalizability of research findings the reader must assess whether the investiga-
tors used research methods that minimized potential bias and maximized data
validity
global and region-specific measures survey instruments used to assess global
health status typically focus on quality of life and disability. These often are
LibraryPirate
97689_Gloss 6/30/10 12:45 PM Page 409
Glossary 409
H
human participants a individual who partakes in a activity
I
independent variable something that is manipulated by the researcher
Injury Rate refers to the number of new injuries per unit of exposure time
internal culture standards, values, philosophy and professional guidelines shared
by a societal ethos
internal validity refers to the validity of a studys experimental design. For ex-
ample, most experiments are designed to show a causal relationship between
an independent variable and a dependent variable. If an experiment can con-
clusively demonstrate that the independent variable has a definite effect on
the dependent variable, the study is internally valid. If, however, other fac-
tors may influence the dependent variable and these factors are not con-
trolled for in the experimental design, the studys internal validity may be
questioned
interpretivism also known as qualitative inquiry. To understand the meaning of an
experience for particular participants, one must understand the phenomena
being studied in the context of the setting for the participants of interest
intervention the magnitude of change on one or more measures or as a probabil-
ity of a favorable or adverse outcome
intervention outcomes results attributed to treatment
invisible web areas of the Internet that are inaccessible to search engines
L
levels of evidence following the CEBM classification of levels of evidence, the
highest level of evidence is Level 1 evidence that comes from randomized con-
trolled trials (RCT), which are the gold standard for clinical trials methodology
likelihood ratios the likelihood ratio, often denoted by (the capital GREEK LETTER
LAMBDA), is the ratio of the maximum PROBABILITY of a RESULT under two differ-
ent hypotheses (8)
LibraryPirate
97689_Gloss 6/30/10 12:45 PM Page 410
410 Glossary
M
methodologic flaws threaten the validity of data reported in studies
methodologic quality in research Whiting et al. developed a 14-item assessment
tool to assist consumers in evaluating the methodological quality of research
into diagnostic tests. Scores of 10 or more are considered to reflect sound re-
search methods
N
nonclinical research research and development of new diagnostic technologies.
However, not until a diagnostic test is studied in the population it is intended
to benefit can the true magnitude of benefit and risk be elucidated
O
outcome measures the data analysis conducted in preparation of a guide can also
generate likelihood ratios that translate the probability of a successful outcome
with an intervention when a patient presents with a group of characteristics de-
rived from investigations where outcomes are dichotomous
outcomes instruments instruments designed to address functional limitations and
disabilities that are often associated with patients who have the specific injury
or illness
P
paradigm shift an adjustment in thinking that comes about as the result of new
discoveries, inventions, or real-world experiences (8)
parametric and nonparametric statistics nonparametric statistical methods of com-
parison are used to analyze nominal data. Parametric statistics analyze the (dis-
tribution of) variance and are appropriate to analyze interval and ratio data
under most circumstances
pathology the medical science, and specialty practice, concerned with all aspects
of disease but with special reference to the essential nature, causes, and devel-
opment of abnormal conditions, as well as the structural and functional
changes that result from the disease processes (7)
patient-important outcomes consideration of patient values serves as a reminder
that clinicians should not get carried away interpreting numbers and forget
that the patient is the focus of attention
LibraryPirate
97689_Gloss 6/30/10 12:45 PM Page 411
Glossary 411
412 Glossary
R
randomized clinical trials clinical trials that involve at least one test treatment and
one control treatment, concurrent enrollment and follow-up of the test- and
control-treated groups, and in which the treatments to be administered are se-
lected by a random process (11)
randomized trials require that each patient enrolled in a study receive all interven-
tion for prescribed periods of time in randomized order
receiver operator characteristic curves provides a clinicians information about
tests that generate measures on a continuous scale such as blood pressure,
range of motion, and serum enzyme levels
regression analysis a technique used for the modeling and analysis of numerical
data consisting of values of a dependent variable (response variable) and of one
or more independent variables (explanatory variables) (8)
relative risk provides a proportion of injury incidence between two groups and is
identical to the calculation of risk ratio
relevant literature information that is related to the topic of discussion, it is most
important in the introduction of the paper
research hypothesis predicts the answer to the question (also see Chapter 1)
S
scientific method a systematic process where the creation of a research hypothesis
that is based on existing knowledge and unconfirmed observations. Experiments
are then designed and conducted to the test these hypotheses in an unbiased
manner. The scientific method requires that data be collected via observation or,
in the health sciences, often via instrumented devices in a controlled manner
scientific paper conveying and discussing the results of an investigation
scientific writing writings exhibiting the methods and principles of science
selection bias characteristics that subjects have before they enroll in a study may
ultimately influence the results of the study. These may include things like age,
maturation, sex, medical history, injury or illness severity, and motivation,
among many others
special tests patient interviews and the performance of physical examination pro-
cedures
standard error of measurement (SEM) a test based on error with regard to reliability.
The difference between the obtained test result and the hypothetical true result (7)
statistical power chance of rejecting the null hypothesis. Statistical power increases
the likelihood of finding statistically significant differences or in other words
rejecting the null hypothesis
strength of evidence can vary between and even within a clinical practice guide-
line depending upon the quality and quantity of research available
LibraryPirate
97689_Gloss 6/30/10 12:45 PM Page 413
Glossary 413
T
theoretical model a strict statement of relations that can be put into equation (e.g.,
Theory of Relativity). Used to represent a type of physical analogy (e.g., the
brain is like a computer)
theoretical research hypothesis describes how expected changes in the dependent
variable depend on the independent variable as a result of the study
theory a tentative explanation for the facts and findings that evolve from the re-
search process
therapeutic intervention the clinician needs to know what is wrong before it can
be made right
translational research a more recent term that is used to describe investigations
that apply the results from basic science to the care of patients. Sometimes re-
ferred to as bench-to-bedside, translational research seeks to speed the devel-
opment of more effective patient care strategies
translational research consists of expert opinion and disease-oriented evidence
treatment outcomes evaluation undertaken to assess the results or consequences
of management and procedures used in combating disease in order to deter-
mine the efficacy, effectiveness, safety, practicability, etc., of these interventions
in individual cases or series (11)
Type I and II errors Type I errors occur when a null is rejected and in fact popula-
tion differences do not exist. Type II error occurs when a null is not rejected yet
a study of the population would reveal differences between groups
V
variables the things or stuff being questioned or measured in the study. Must have
two or more categories of distinguishing qualities or distinctive characteristics,
a range of values, parameters, or quantifiers
REFERENCES
1. National Cancer Institute. Retrieved March 08, 2010, from http://www.cancer.gov/
dictionary/?CdrID44007
2. Clinical Research & Clinical Trials. (2009, July 23). Retrieved March 8, 2010, from
http://www.nichd.nih.gov/health/clinicalresearch/
3. McGraw-Hill Concise Dictionary of Modern Medicine. (2002). Retrieved March 4, 2010,
from http://medical-dictionary.thefreedictionary.com/
LibraryPirate
97689_Gloss 6/30/10 12:45 PM Page 414
414 Glossary
INDEX
Page numbers followed by f, t and b indicate figures, tables, and boxes, respectively.
415
LibraryPirate
97689_Index 6/30/10 12:15 PM Page 416
416 Index
Index 417
418 Index
Index 419
420 Index
Index 421
M N
Magna Carta, 347 Nagi, Saad, 261
Magnetic resonance imaging (MRI), 216, 218 Nagis disablement model, 261262, 262f, 263t
Management plan section. See Personnel, section visual representation of, 262f
Mann-Whitney U test, 186, 186t, 187189 Narrative review
example, 187189 vs. systematic review, 311t
MANOVA. See Multivariate analysis of variance Narrative section
(MANOVA) in funding proposal writing, 361369.
MCID. See Minimally clinically important See also Proposal writing, funding
difference (MCID) National Center for Medical Rehabilitation
Meals and general incidentals (M&IE), 374 Research (NCMRR) disablement model,
Mean, 127128, 158 262, 264f
mathematical expression, 127 National Collegiate Athletic Association, 241
pairs of National Institutes of Health (NIH), 54, 69
comparisons between, 173 funding and, 351352
Measurement data National Science Foundation (NSF), 54, 353
classification of, 140141 NCMRR. See National Center for Medical
categorical data, 140 Rehabilitation Research (NCMRR)
continuous data, 141 disablement model
ordinal data, 140141 Nervousness based medicine, 90
Likert scale and, 140, 141 NES. See Neuromuscular electrical stimulation
Median (Md), 128, 129 (NES)
formula, 128 Neuromuscular electrical stimulation (NES), 160
Medical Newtons Law of Gravity, 34
care NIH. See National Institutes of Health (NIH)
medical research combined with, 6768 NNT. See Numbers-needed-to-treat (NNT)
decisions, 45 NNTB. See Numbers needed to treat to benefit
research (NNTB)
combined with medical care, 6768 NNTH. See Numbers needed to harm (NNTH)
principles for, 6667 Nominal data, 158. See also Categorical data
Medicine. See also Evidence-based medicine nonparametric statistical analysis of, 158
(EBM) Nonclinical research, 218
research in, 4647 Nonparametric statistics, 158, 184185
Meta-analysis, 180, 317 nominal data and, 158
Metasoft Systems BigDatabase, 351 Nonprobability sampling, 133
Methods section vs. probability sampling, 133
in funding proposal writing, 364367 Nonrandomized pretestposttest design, 135
in scientific paper, 338339 Non-time loss injury
MID. See Minimally clinically important defined, 251
difference (MCID) Normal distribution, 130
M&IE. See Meals and general incidentals (M&IE) for range of motion, knee, 131f
Milgram Study, 59b NSF. See National Science Foundation (NSF)
Military research, 6870 Null hypothesis, 131, 162
background, 69 vs. research hypothesis, 131
Minimally clinically important difference Numbers needed to harm (NNTH), 253, 254f
(MCID), 288289 Numbers-needed-to-treat (NNT), 321
estimation, 289 calculation of, 253
ROC analysis in, 289, 290f definition of, 253
Minimally important difference (MID). Numbers needed to treat to benefit (NNTB),
See Minimally clinically important 253, 254f
difference (MCID) Nuremberg Code, 59b, 61
Mixed model, 170 Nuremberg Doctors Trial, the, 59b, 6061
ANOVA and, 170
Mode (Mo), 128, 129
MRI. See Magnetic resonance imaging (MRI) O
Multiple regression, 200205 Objective measurement, 140
PPMC and, 200 Observation
Multivariate analysis of variance (MANOVA), in qualitative inquiry, 114
174175 Office of Management and Budget (OMB), 348
LibraryPirate
97689_Index 6/30/10 12:15 PM Page 422
422 Index
OMB. See Office of Management and Budget Pearsons r correlation coefficient, 145
(OMB) vs. intraclass correlation coefficients (ICC), 145
OMB Circulars PEDro scale, 281t
in budget preparation, 370 PEER. See Patients expected event rate (PEER)
One group pretestposttest design, 135 Peer-reviewed papers, 330
One shot posttest design, 135 Peer-review process, 80
Oral presentations. See Platform presentations Personal factors in WHO model, 264265
Ordinal data, 140141, 158, 159 Personnel
parametric analysis of, 189190 budgeting for, 371373. See also Budget
Spearman rank order correlation and, 207 preparation
Orientational qualitative research, 118 section
Ottawa Ankle Rules, 218 in funding proposal writing, 367368
clinical prediction guide and, 295296 Phenomenology, 115
Ottawa Health Research Institute, 296 PHS. See U.S. Public Health Service (PHS)
Outcome, case studies, 399 Physical examination procedures, 217218
Outcomes measures, 297 PI. See Principal investigator (PI)
types of Planning
DOE, 260 in grant writing, 353354
POE, 260 Platform presentations, 333335
Outliers, 129 challenges in, 333
Oxford Center for Evidence-Based Medicine tips for producing effective, 334t335t
hierarchy, 93f POE. See Patient oriented evidence (POE)
all or none studies, 95b, 9798 POEM. See Patient-oriented evidence that
matters (POEM)
P Point estimate, 284, 285f
Pairs of means of association, 195
comparisons between, 173. See also Mean Population-based outcomes research, 93f, 99
Paradigm, 44 Population values
of EBM, 90 estimates
qualitative research, 109 clinical prediction guides and, 299
of science, 7 Position statements, 299
Parametric statistics, 158, 160 Poster presentation, 331333
interval data and, 158 tips for producing effective, 332t
ordinal data and, 189190 Post-hoc analysis, 173174
ratio data and, 158 Post-hoc comparisons, 156
Participant observation Posttest probability
in qualitative inquiry, 114 application of LRs on, 231232
Participation Power analysis, a priori, 133
of students PPMC. See Pearson product moment
grand rounds and, 401 correlation (PPMC)
in WHO model, 264 Precision
Patient of measurement, 144, 149
care, 9192 Predictor variables, 200, 299
systematic review in, 317324 Pre-experimental designs, 135
evaluation, 218219 Presentation
information, confidentiality of, 401 professional meetings, 330335
values, 104, 274 scientific writing, 335343
Patient oriented evidence (POE), 260 Pretest probability
self-report, patient, 266 and LR, relationship between, 230231
vs. DOE, 260 Prevalence
Patient-oriented evidence that matters of injury, 248249
(POEM), 260 ratio, 251
Patients expected event rate (PEER), 321 Prevention, 29, 91. See also Injury(ies), prevention
PBL. See Problem-based learning (PBL) Principal investigator (PI)
Pearson product moment correlation budgeting for, 371
(PPMC), 198 Probability
multiple regression and, 200 estimates, 195
Pearson , 198 posttest, application of LRs on, 231232
LibraryPirate
97689_Index 6/30/10 12:15 PM Page 423
Index 423
424 Index
Index 425
426 Index