Get Applied Univariate, Bivariate, and Multivariate Statistics Using Python: A Beginner's Guide to Advanced Data Analysis 1st Edition Daniel J. Denis PDF ebook with Full Chapters Now
Get Applied Univariate, Bivariate, and Multivariate Statistics Using Python: A Beginner's Guide to Advanced Data Analysis 1st Edition Daniel J. Denis PDF ebook with Full Chapters Now
com
https://textbookfull.com/product/applied-univariate-
bivariate-and-multivariate-statistics-using-python-a-
beginners-guide-to-advanced-data-analysis-1st-edition-
daniel-j-denis/
OR CLICK BUTTON
DOWNLOAD NOW
https://textbookfull.com/product/applied-univariate-bivariate-and-
multivariate-statistics-1st-edition-daniel-j-denis/
textboxfull.com
https://textbookfull.com/product/biota-grow-2c-gather-2c-cook-loucas/
textboxfull.com
https://textbookfull.com/product/advanced-data-analysis-modelling-in-
chemical-engineering-1st-edition-denis-constales/
textboxfull.com
https://textbookfull.com/product/multivariate-data-analysis-joseph-f-
hair/
textboxfull.com
https://textbookfull.com/product/practical-machine-learning-for-data-
analysis-using-python-1st-edition-abdulhamit-subasi/
textboxfull.com
Applied Univariate, Bivariate, and Multivariate Statistics Using Python
Daniel J. Denis
All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or
transmitted, in any form or by any means, electronic, mechanical, photocopying, recording or otherwise,
except as permitted by law. Advice on how to obtain permission to reuse material from this title is available
at http://www.wiley.com/go/permissions.
The right of Daniel J. Denis to be identified as the author of this work has been asserted in accordance with
law.
Registered Office
John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, USA
Editorial Office
111 River Street, Hoboken, NJ 07030, USA
For details of our global editorial offices, customer services, and more information about Wiley products
visit us at www.wiley.com.
Wiley also publishes its books in a variety of electronic formats and by print-on-demand. Some content that
appears in standard print versions of this book may not be available in other formats.
10 9 8 7 6 5 4 3 2 1
Contents
Prefacexii
References 273
Index 276
Preface
Applied Univariate, Bivariate, and Multivariate Statistics Using Python: A Beginner’s Guide to
Advanced Data Analysis, First Edition. Daniel J. Denis.
© 2021 John Wiley & Sons, Inc. Published 2021 by John Wiley & Sons, Inc.
(e.g. see Guttag, 2013). For those who want some introductory exposure to Python on
generating data-analytic results and wish to understand what the software is
producing, it is hoped that the current book will be of great use.
In a book such as this, limited by a fixed number of pages, it is an exceedingly dif-
ficult and challenging endeavor to both instruct on statistics and software simultane-
ously. Attempting to cover univariate, bivariate, and multivariate techniques in a book
of this size in any kind of respectable depth or completeness in coverage is, well, an
impossibility. Combine this with including software options and the impossibility fac-
tor increases! However, such is the nature of books that attempt to survey a wide vari-
ety of techniques such as this one – one has to include only the most essential of
information to get the reader “going” on the techniques and advise him or her to
consult other sources for further details. Targeting the right mix of theory and software
in a book like this is the most challenging part, but so long as the reader (and instruc-
tor) recognizes that this book is but a foot-in-the-door to get students “started,” then I
hope it will fall in the confidence band of a reasonable expectation. The reader wishing
to better understand a given technique or principle will naturally find many narratives
incomplete, while the reader hoping to find more details on Python will likewise find
the book incomplete. On average, however, it is hoped that the current “mix” is of
introductory use for the newcomer. It can be exceedingly difficult to enter the world of
statistics and computing. This book will get you started. In many places, references are
provided on where to go next.
Unfortunately, many available books on the market for Python are nothing more
than slaps in the face to statistical theory while presenting a bunch of computer code
that otherwise masks a true understanding of what the code actually accomplishes.
Though data science is a welcome addition to the mathematical and applied scien-
tific disciplines, and software advancements have made leaps and bounds in the area
of quantitative analysis, it is also an unfortunate trend that understanding statistical
theory and an actual understanding of statistical methods is sometimes taking a back
seat to what we will otherwise call “generating output.” The goal of research and
science is not to generate software output. The goal is, or at least should be, to
understand in a deeper way whatever output that is generated. Code can be
looked up far easier than can statistical understanding. Hence, the goal of the book is
to understand what the code represents (at least the important code on which tech-
niques are run) and, to some extent at least, the underlying mathematical and philo-
sophical mechanisms of one’s analysis. We comment on this important distinction a
bit later in this preface as it is very important. Each chapter of this book could easily be
expanded and developed into a deeper book spanning more than 3–4 times the size of
the book in entirety.
The book is the fourth in a series of books published by the author, all with Wiley.
Readers wishing a deeper discussion of the topics treated in this book are encouraged
to consult the author’s first book, now in its second (and better) edition titled Applied
■■ “Don’t Forget!”
Brief “don’t forget” summaries serve to emphasize and reinforce that which is most
pertinent to the discussion and to aid in learning these concepts. They also serve to
highlight material that can be easily misunderstood or misapplied if care is not prac-
ticed. Scattered throughout the book, these boxes help the reader review and empha-
size essential material discussed in the chapters.
■■ Each chapter concludes with a brief set of exercises. These include both conceptu-
ally-based problems that are targeted to help in mastering concepts introduced in
the chapter, as well as computational problems using Python.
■■ Most concepts are implicitly defined throughout the book by introducing them in
the context of how they are used in scientific and statistical practice. This is most
appropriate for a short book such as this where time and space to unpack definitions
in entirety is lacking. “Dictionary definitions” are usually grossly incomplete any-
way and one could even argue that most definitions in even good textbooks often fail
to capture the “essence” of the concept. It is only in seeing the term used in its proper
context does one better appreciate how it is employed, and, in this sense, the reader
is able to unpack the deeper intended meaning of the term. For example, defining a
population as the set of objects of ultimate interest to the researcher is not enlighten-
ing. Using the word in the context of a scientific example is much more meaning-
ful. Every effort in the book is made to accurately convey deeper conceptual
understanding rather than rely on superficial definitions.
■■ Most of the book was written at the beginning of the COVID-19 pandemic of 2020
and hence it seemed appropriate to feature examples of COVID-19 in places through-
out the book where possible, not so much in terms of data analysis, but rather in
examples of how hypothesis-testing works and the like. In this way, it is hoped
examples and analogies “hit home” a bit more for readers and students, making the
issues “come alive” somewhat rather than featuring abstract examples.
■■ Python code is “unpacked” and explained in many, though not all, places. Many
existing books on the market contain explanations of statistical concepts (to varying
degrees of precision) and then plop down a bunch of code the reader is expected to
simply implement and understand. While we do not avoid this entirely, for the most
part we guide the reader step-by-step through both concepts and Python code used.
The goal of the book is in understanding how statistical methods work, not
arming you with a bunch of code for which you do not understand what is behind it.
Principal components code, for instance, is meaningless if you do not first under-
stand and appreciate to some extent what components analysis is about.
Having now taught at both the undergraduate and graduate levels for the better part of
fifteen years to applied students in the social and sometimes natural sciences, to the
delight of my students (sarcasm), I have opened each course with a lecture of sorts on
the differences between statistical vs. software knowledge. Very little of the warn-
ing is grasped I imagine, though the real-life experience of the warning usually sur-
faces later in their graduate careers (such as at thesis or dissertation defenses where
they may fail to understand their own software output). I will repeat some of that ser-
mon here. While this distinction, historically, has always been important, it is perhaps
no more important than in the present day given the influx of computing power avail-
able to virtually every student in the sciences and related areas, and the relative ease
with which such computing power can be implemented. Allowing a new teen driver to
drive a Dodge Hellcat with upward of 700 horsepower would be unwise, yet newcom-
ers to statistics and science, from their first day, have such access to the equivalent in
computing power. The statistician is shaking his or her head in disapproval, for good
reason. We live in an age where data analysis is available to virtually anybody with a
laptop and a few lines of code. The code can often easily be dug up in a matter of sec-
onds online, even with very little software knowledge. And of course, with many soft-
ware programs coding is not even a requirement, as windows and GUIs (graphical
user interfaces) have become very easy to use such that one can obtain an analysis in
virtually seconds or even milliseconds. Though this has its advantages, it is not always
and necessarily a good thing.
On the one hand, it does allow the student of applied science to “attempt” to conduct
his or her data analyses. Yet on the other, as the adage goes, a little knowledge can
be a dangerous thing. Being a student of the history of statistics, I can tell you that
before computers were widely available, conducting statistical analyses were available
only to those who could drudge through computations by hand in generating their
“output” (which of course took the form of paper-and-pencil summaries, not the soft-
ware output we have today). These computations took hours upon hours to perform,
and hence, if one were going to do a statistical analysis, one did not embark on such
an endeavor lightly. That does not mean the final solution would be valid necessar-
ily, but rather folks may have been more likely to give serious thought to their analy-
ses before conducting them. Today, a student can run a MANOVA in literally 5 minutes
using software, but, unfortunately, this does not imply the student will understand
what they have done or why they have done it. Random assignment to conditions
may have never even been performed, yet in the haste to implement the software rou-
tine, the student failed to understand or appreciate how limiting their output would
be. Concepts of experimental design get lost in the haste to produce computer out-
put. However, the student of the “modern age” of computing somehow “missed” this
step in his or her quickness to, as it were, perform “advanced statistics.” Further, the
result is “statistically significant,” yet the student has no idea what Wilks’s lambda is
or how it is computed, nor is the difference between statistical significance and
effect size understood. The limitations of what the student has produced are not
appreciated and faulty substantive (and often philosophically illogical) conclusions
follow. I kid you not, I have been told by a new student before that the only problem
with the world is a lack of computing power. Once computing power increases, experi-
mental design will be a thing of the past, or so the student believed. Some incoming
students enter my class with such perceptions, failing to realize that discovering a cure
for COVID-19, for instance, is not a computer issue. It is a scientific one. Computers
help, but they do not on their own resolve scientific issues. Instructors faced with these
initial misconceptions from their students have a tough road to hoe ahead, especially
when forcing on their students fundamental linear algebra in the first two weeks of the
course rather than computer code and statistical recipes.
The problem, succinctly put, is that in many sciences, and contrary to the opinion
you might expect from someone writing a data analysis text, students learn too
much on how to obtain output at the expense of understanding what the out-
put means or the process that is important in drawing proper scientific con-
clusions from said output. Sadly, in many disciplines, a course in “Statistics” would
be more appropriately, and unfortunately, called “How to Obtain Software Output,”
because that is pretty much all the course teaches students to do. How did statistics
education in applied fields become so watered down? Since when did cultivating
the art of analytical or quantitative thinking not matter? Faculty who teach such
courses in such a superficial style should know better and instead teach courses with
a lot more “statistical thinking” rather than simply generating software output. Among
students (who should not necessarily know better – that is what makes them students),
there often exists the illusion that simply because one can obtain output for a multiple
regression, this somehow implies a multiple regression was performed correctly in
line with the researcher’s scientific aims. Do you know how to conduct a multiple
regression? “Yes, I know how to do it in software.” This answer is not a correct
answer to knowing how to conduct a multiple regression! One need not even
understand what multiple regression is to “compute one” in software. As a consultant,
I have also had a client or two from very prestigious universities email me a bunch of
software output and ask me “Did I do this right?” assuming I could evaluate their code
and output without first knowledge of their scientific goals and aims. “Were the statis-
tics done correctly?” Of course, without an understanding of what they intended to do
or the goals of their research, such a question is not only figuratively, but also literally
impossible to answer aside from ensuring them that the software has a strong repu-
tation for accuracy in number-crunching.
This overemphasis on computation, software or otherwise, is not right, and
is a real problem, and is responsible for many misuses and abuses of applied statistics
in virtually every field of endeavor. However, it is especially poignant in fields in the
social sciences because the objects on which the statistics are computed are often sta-
tistical or psychometric entities themselves, which makes understanding how sta-
tistical modeling works even more vital to understanding what can vs. what cannot be
concluded from a given statistical analysis. Though these problems are also present in
fields such as biology and others, they are less poignant, since the reality of the objects
in these fields is usually more agreed upon. To be blunt, a t-test on whether a COVID-
19 vaccine works or not is not too philosophically challenging. Finding the vaccine is
difficult science to be sure, but analyzing the results statistically usually does not
require advanced statistics. However, a regression analysis on whether social distanc-
ing is a contributing factor to depression rates during the COVID-19 pandemic is not
quite as easy on a methodological level. One is so-called “hard science” on real objects,
the other might just end up being a statistical artifact. This is why social science
students, especially those conducting non-experimental research, need rather
deep philosophical and methodological training so they do not read “too much” into a
statistical result, things the physical scientist may never have had to confront due to
the nature of his or her objects of study. Establishing scientific evidence and support-
ing a scientific claim in many social (and even natural) sciences is exceedingly diffi-
cult, despite the myriad of journals accepting for publication a wide variety of incorrect
scientific claims presumably supported by bloated statistical analyses. Just look at the
methodological debates that surrounded COVID-19, which is on an object that is rela-
tively “easy” philosophically! Step away from concrete science, throw in advanced sta-
tistical technology and complexity, and you enter a world where establishing evidence
is philosophical quicksand. Many students who use statistical methods fall into these
pits without even knowing it and it is the instructor’s responsibility to keep them
grounded in what the statistical method can vs. cannot do. I have told students count-
less times, “No, the statistical method cannot tell you that; it can only tell you this.”
Hence, for the student of empirical sciences, they need to be acutely aware and
appreciative of the deeper issues of conducting their own science. This implies a heav-
ier emphasis on not how to conduct a billion different statistical analyses, but on
understanding the issues with conducting the “basic” analyses they are performing. It
is a matter of fact that many students who fill their theses or dissertations with applied
statistics may nonetheless fail to appreciate that very little of scientific usefulness has
been achieved. What has too often been achieved is a blatant abuse of statistics
masquerading as scientific advancement. The student “bootstrapped standard
errors” (Wow! Impressive!), but in the midst of a dissertation that is scientifically
unsound or at a minimum very weak on a methodological level.
A perfect example to illustrate how statistical analyses can be abused is when per-
forming a so-called “mediation” analysis (you might infer by the quotation marks
that I am generally not a fan, and for a very good reason I may add). In lightning speed,
a student or researcher can regress Y on X, introduce Z as a mediator, and if statisti-
cally significant, draw the conclusion that “Z mediates the relationship between Y and X.”
That’s fine, so long as it is clearly understood that what has been established is statis-
tical mediation (Baron and Kenny, 1986), and not necessarily anything more. To say
that Z mediates Y and X, in a real substantive sense, requires, of course, much more
knowledge of the variables and/or of the research context or design. It first and fore-
most requires defining what one means by “mediation” in the first place. Simply
because one computes statistical mediation does not, in any way whatsoever, justify
somehow drawing the conclusion that “X goes through Z on its way to Y,” or any-
thing even remotely similar. Crazy talk! Of course, understanding this limitation
should be obvious, right? Not so for many who conduct such analyses. What would
such a conclusion even mean? In most cases, with most variables, it simply does not
even make sense, regardless of how much statistical mediation is established. Again,
this should be blatantly obvious, however many students (and researchers) are una-
ware of this, failing to realize or appreciate that a statistical model cannot, by itself,
impart a “process” onto variables. All a statistical model can typically do, by
itself, is partition variability and estimate parameters. Fiedler et al. (2011)
recently summarized the rather obvious fact that without the validity of prior assump-
tions, statistical mediation is simply, and merely, variance partitioning. Fisher,
inventor of ANOVA (analysis of variance), already warned us of this when he said of
his own novel (at the time) method that ANOVA was merely a way of “arranging the
arithmetic.” Whether or not that arrangement is meaningful or not has to come from
the scientist and a deep consideration of the objects on which that arrangement is
being performed. This idea, that the science matters more than the statistics on which
it is applied, is at risk of being lost, especially in the social sciences where statistical
models regularly “run the show” (at least in some fields) due to the difficulty in many
cases of operationalizing or controlling the objects of study.
Returning to our mediation example, if the context of the research problem lends
itself to a physical or substantive definition of mediation or any other physical process,
such that there is good reason to believe Z is truly, substantively, “mediating,” then the
statistical model can be used as establishing support for this already-presumed rela-
tion, in the same way a statistical model can be used to quantify the generational trans-
mission of physical qualities from parent to child in regression. The process itself,
however, is not due to the fitting of a statistical model. Never in the history of science
or statistics has a statistical model ever generated a process. It merely, and poten-
tially, has only described one. Many students, however, excited to have bootstrapped
those standard errors in their model and all the rest of it, are apt to draw substantive
conclusions based on a statistical model that simply do not hold water. In such cases,
one is better off not running a statistical model at all rather than using it to draw inane
philosophically egregious conclusions that can usually be easily corrected in any intro-
duction to a philosophy of science or research methodology course. Abusing and
overusing statistics does little to advance science. It simply provides a cloak of
complexity.
So, what is the conclusion and recommendation from what might appear to be a
very cynical discussion in introducing this book? Understanding the science and
statistics must come first. Understanding what can vs. cannot be concluded from a
statistical result is the “hard part,” not computing something in Python, at least not at
our level of computation (at more advanced levels, of course, computing can be excep-
tionally difficult, as evidenced by the necessity of advanced computer science degrees).
Python code can always be looked up for applied sciences purposes, but “statistical
understanding” cannot. At least not so easily. Before embarking on either a statistics
course or a computation course, students are strongly encouraged to take a rigorous
research design course, as well as a philosophy of science course, so they might
better appreciate the limitations of their “claims to evidence” in their projects.
Otherwise, statistics, and the computers that compute them, can be just as easily
misused and abused as used correctly, and sadly, often are. Instructors and supervisors
need to also better educate students on the reckless fitting of statistical models and
computing inordinate amounts of statistics without careful guidance on what can vs.
cannot be interpreted from such numerical measures. Design first, statistics
second.
One important aspect of learning and understanding any craft is to know where and
why making distinctions is important, and on the opposite end of the spectrum,
where divisions simply blur what is really there. One area where this is especially true
is in learning, or at least “using,” a technical discipline such as mathematics and statis-
tics to better understand another subject. Many instructors of applied statistics strive
to teach statistics at a “conceptual” level, which, to them at least, means making the
discipline less “mathematical.” This is done presumably to attract students who may
otherwise be fearful of mathematics with all of its formulas and symbolism. However,
this distinction, I argue, does more harm than good, and completely misses the point.
The truth of the matter is that mathematics are concepts. Statistics are likewise
concepts. Attempting to draw a distinction between two things that are the same does
little good and only provides more confusion for the student.
A linear function, for example, is a concept, just as a standard error is a concept.
That they are symbolized does not take away the fact that there is a softer, more mal-
leable “idea” underneath them, to which the symbolic definition has merely attempted
to define. The sooner the student of applied statistics recognizes this, the sooner he or
she will stop psychologically associating mathematics with “mathematics,” and instead
associate with it what it really is, a form of conceptual development and refine-
ment of intellectual ideas. The mathematics is usually in many cases the “packaged
form” of that conceptual development. Computing a t-test, for instance, is not mathe-
matics. It is arithmetic. Understanding what occurs in the t-test as the mean differ-
ence in the numerator goes toward zero (for example) is not “conceptual understanding.”
Rather, it is mathematics, and the fact that the concepts of mathematics can be
unpacked into a more verbal or descriptive discussion only serves to delineate the
concept that already exists underneath the description. Many instructors of applied
statistics are not aware of this and continually foster the idea to students that mathe-
matics is somehow separate from the conceptual development they are trying to
impart onto their students. Instructors who teach statistics as a series of recipes
and formulas without any conceptual development at all do a serious (almost
The book can be used at either the advanced undergraduate or graduate levels, or for
self-study. The book is ideal for a 16-week course, for instance one in a Fall or Spring
semester, and may prove especially useful for programs that only have space or desire
to feature a single data-analytic course for students. Instructors can use the book as a
primary text or as a supplement to a more theoretical book that unpacks the con-
cepts featured in this book. Exercises at the end of each chapter can be assigned weekly
and can be discussed in class or reviewed by a teaching assistant in lab. The goal of the
exercises should be to get students thinking critically and creatively, not simply
getting the “right answer.”
It is hoped that you enjoy this book as a gentle introduction to the world of applied
statistics using Python. Please feel free to contact me at [email protected]
or [email protected] should you have any comments or corrections. For data files
and errata, please visit www.datapsyc.com.
Daniel J. Denis
March, 2021
CHAPTER OBJECTIVES
■■ How probability is the basis of statistical and scientific thinking.
■■ Examples of statistical inference and thinking in the COVID-19 pandemic.
■■ Overview of how null hypothesis significance testing (NHST) works.
■■ The relationship between statistical inference and decision-making.
■■ Error rates in statistical thinking and how to minimize them.
■■ The difference between a point estimator and an interval estimator.
■■ The difference between a continuous vs. discrete variable.
■■ Appreciating a few of the more salient philosophical underpinnings of applied statistics
and science.
■■ Understanding scales of measurement, nominal, ordinal, interval, and ratio.
■■ Data analysis, data science, and “big data” distinctions.
The goal of this first chapter is to provide a global overview of the logic behind statisti-
cal inference and how it is the basis for analyzing data and addressing scientific prob-
lems. Statistical inference, in one form or another, has existed at least going back to the
Greeks, even if it was only relatively recently formalized into a complete system. What
unifies virtually all of statistical inference is that of probability. Without probability,
statistical inference could not exist, and thus much of modern day statistics would not
exist either (Stigler, 1986).
When we speak of the probability of an event occurring, we are seeking to know the
likelihood of that event. Of course, that explanation is not useful, since all we have
done is replace probability with the word likelihood. What we need is a more precise
definition. Kolmogorov (1903–1987) established basic axioms of probability and was
thus influential in the mathematics of modern-day probability theory. An axiom in
mathematics is basically a statement that is assumed to be true without requiring any
proof or justification. This is unlike a theorem in mathematics, which is only consid-
ered true if it can be rigorously justified, usually by other allied parallel mathematical
results. Though the axioms help establish the mathematics of probability, they surpris-
ingly do not help us define exactly what probability actually is. Some statisticians,
Applied Univariate, Bivariate, and Multivariate Statistics Using Python: A Beginner’s Guide to
Advanced Data Analysis, First Edition. Daniel J. Denis.
© 2021 John Wiley & Sons, Inc. Published 2021 by John Wiley & Sons, Inc.
scientists and philosophers hold that probability is a relative frequency, while others
find it more useful to consider probability as a degree of belief. An example of a rela-
tive frequency would be flipping a coin 100 times and observing the number of heads
that result. If that number is 40, then we might estimate the probability of heads on the
coin to be 0.40, that is, 40/100. However, this number can also reflect our degree of
belief in the probability of heads, by which we based our belief on a relative frequency.
There are cases, however, in which relative frequencies are not so easily obtained or
virtually impossible to estimate, such as the probability that COVID-19 will become a
seasonal disease. Often, experts in the area have to provide good guesstimates based on
prior knowledge and their clinical opinion. These probabilities are best considered
subjective probabilities as they reflect a degree of belief or disbelief in a theory
rather than a strict relative frequency. Historically, scholars who espouse that proba-
bility can be nothing more than a relative frequency are often called frequentists,
while those who believe it is a degree of belief are usually called Bayesians, due to
Bayesian statistics regularly employing subjective probabilities in its development and
operations. A discussion of Bayesian statistics is well beyond the scope of this chapter
and book. For an excellent introduction, as well as a general introduction to the rudi-
ments of statistical theory, see Savage (1972).
When you think about it for a moment, virtually all things in the world are probabil-
istic. As a recent example, consider the COVID-19 pandemic of 2020. Since the start of
the outbreak, questions involving probability were front and center in virtually all
media discussions. That is, the undertones of probability, science, and statistical infer-
ence were virtually everywhere where discussions of the pandemic were to be had.
Concepts of probability could not be avoided. The following are just a few of the ques-
tions asked during the pandemic:
■■ What is the probability of contracting the virus, and does this probability vary as
a function of factors such as pre-existing conditions or age? In this latter case, we
might be interested in the conditional probability of contracting COVID-19 given
a pre-existing condition or advanced age. For example, if someone suffers from heart
disease, is that person at greatest risk of acquiring the infection? That is, what is the
probability of COVID-19 infection being conditional on someone already suffering
from heart disease or other ailments?
■■ What proportion of the general population has the virus? Ideally, researchers wanted
to know how many people world-wide had contracted the virus. This constituted a
case of parameter estimation, where the parameter of interest was the proportion
of cases world-wide having the virus. Since this number was unknown, it was typi-
cally estimated based on sample data by computing a statistic (i.e. in this case, a
proportion) and using that number to infer the true population proportion. It is
important to understand that the statistic in this case was a proportion, but it could
have also been a different function of the data. For example, a percentage increase
or decrease in COVID-19 cases was also a parameter of interest to be estimated via
sample data across a particular period of time. In all such cases, we wish to estimate
a parameter based on a statistic.
■■ What proportion of those who contracted the virus will die of it? That is, what is the
estimated total death count from the pandemic, from beginning to end? Statistics
such as these involved projections of death counts over a specific period of time
and relied on already established model curves from similar pandemics. Scientists
who study infectious diseases have historically documented the likely (i.e. read:
“probabilistic”) trajectories of death rates over a period of time, which incorporates
estimates of how quickly and easily the virus spreads from one individual to the
next. These estimates were all statistical in nature. Estimates often included confi-
dence limits and bands around projected trajectories as a means of estimating the
degree of uncertainty in the prediction. Hence, projected estimates were in the
opinion of many media types “wrong,” but this was usually due to not understand-
ing or appreciating the limits of uncertainty provided in the original estimates. Of
course, uncertainty limits were sometimes quite wide, because predicting death
rates was very difficult to begin with. When one models relatively wide margins
of error, one is protected, in a sense, from getting the projection truly wrong.
But of course, one needs to understand what these limits represent, otherwise they
can be easily misunderstood. Were the point estimates wrong? Of course they were!
We knew far before the data came in that the point projections would be off.
Virtually all point predictions will always be wrong. The issue is whether the
data fell in line with the prediction bands that were modeled (e.g. see Figure 1.1). If
a modeler sets them too wide, then the model is essentially quite useless. For
instance, had we said the projected number of deaths would be between 1,000 and
5,000,000 in the USA, that does not really tell us much more than we could have
guessed by our own estimates not using data at all! Be wary of “sophisticated mod-
els” that tell you about the same thing (or even less!) than you could have guessed
on your own (e.g. a weather model that predicts cold temperatures in Montana in
December, how insightful!).
Figure 1.1 Sample death predictions in the United States during the COVID-19 pandemic in
2020. The connected dots toward the right of the plot (beyond the break in the line) represent
a point prediction for the given period (the dots toward the left are actual deaths based on
prior time periods), while the shaded area represents a band of uncertainty. From the current
date in the period of October 2020 forward (the time in which the image was published), the
shaded area increases in order to reflect greater uncertainty in the estimate. Source: CDC
(Centers for Disease Control and Prevention); Materials Developed by CDC. Used with Permission.
Available at CDC (www.cdc.gov) free of charge.
■■ Measurement issues were also at the heart of the pandemic (though rarely addressed
by the media). What exactly constituted a COVID-19 case? Differentiating between
individuals who died “of” COVID-19 vs. died “with” COVID-19 was paramount, yet
was often ignored in early reports. However, the question was central to everything!
“Another individual died of COVID-19” does not mean anything if we do not know the
mechanism or etiology of the death. Quite possibly, COVID-19 was a correlate to death
in many cases, not a cause. That is, within a typical COVID-19 death could lie a virtual
infinite number of possibilities that “contributed” in a sense, to the death. Perhaps one
person died primarily from the virus, whereas another person died because they already
suffered from severe heart disease, and the addition of the virus simply complicated the
overall health issue and overwhelmed them, which essentially caused the death.
To elaborate on the above point somewhat, measurement issues abound in scientific
research and are extremely important, even when what is being measured is seem-
ingly, at least at first glance, relatively simple and direct. If there are issues with how
best to measure something like “COVID death,” just imagine where they surface else-
where. In psychological research, for instance, measurement is even more challeng-
ing, and in many cases adequate measurement is simply not possible. This is why
some natural scientists do not give much psychological research its due (at least in
particular subdivisions of psychology), because they are doubtful that the measure-
ment of such characteristics as anxiety, intelligence, and many other things is even
possible. Self-reports are also usually fraught with difficulty as well. Hence, assessing
the degree of depression present may seem trivial to someone who believes that a self-
report of such symptoms is meaningless. “But I did a complex statistical analysis using
my self-report data.” It doesn’t matter if you haven’t sold to the reader what you’re
analyzing was successfully measured. The most important component to a house is its
foundation. Some scientists would require a more definite “marker” such as a bio-
logical gene or other more physical characteristic or behavioral observation before
they take your ensuing statistical analysis seriously. Statistical complexity usually does
not advance a science on its own. Resolution of measurement issues is more often the
paramount problem to be solved.
The key point from the above discussion is that with any research, with any scien-
tific investigation, scientists are typically interested in estimating population parame-
ters based on information in samples. This occurs by way of probability, and hence one
can say that virtually the entire edifice of statistical and scientific inference is based on
the theory of probability. Even when probability is not explicitly invoked, for instance
in the case of the easy result in an experiment (e.g. 100 rats live who received COVID-
19 treatment and 100 control rats die who did not receive treatment), the elements of
probability are still present, as we will now discuss in surveying at a very intuitive level
how classical hypothesis testing works in the sciences.
Armed with some examples of the COVID-19 pandemic, we can quite easily illustrate
the process of statistical inference on a very practical level. The traditional and classi-
cal workhorse of statistical inference in most sciences is that of null hypothesis
significance testing (NHST), which originated with R.A. Fisher in the early 1920s.
Fisher is largely regarded as the “father of modern statistics.” Most of the classical
techniques used today are due to the mathematical statistics developed in the early
1900s (and late 1800s). Fisher “packaged” the technique of NHST for research workers
in agriculture, biology, and other fields, as a way to grapple with uncertainty in evalu-
ating hypotheses and data. Fisher’s contributions revolutionized how statistics are
used to answer scientific questions (Denis, 2004).
Though NHST can be used in several different contexts, how it works is remarkably
the same in each. A simple example will exemplify its logic. Suppose a treatment is
discovered that purports to cure the COVID-19 virus and an experiment is set up to
evaluate whether it does or not. Two groups of COVID-19 sufferers are recruited who
agree to participate in the experiment. One group will be the control group, while the
other group will receive the novel treatment. Of the subjects recruited, half will be
randomly assigned to the control group, while the other half to the experimental
group. This is an experimental design and constitutes the most rigorous means
known to humankind for establishing the effectiveness of a treatment in science.
Physicists, biologists, psychologists, and many others regularly use experimental
designs in their work to evaluate potential treatment effects. You should too!
Carrying on with our example, we set up what is known as a null hypothesis,
which in our case will state that the number of individuals surviving in the control
group will be the same as that in the experimental group after 30 days from the start of
the experiment. Key to this is understanding that the null hypothesis is about popula-
tion parameters, not sample statistics. If the drug is not working, we would expect,
under the most ideal of conditions, the same survival rates in each condition in the
population under the null hypothesis. The null hypothesis in this case happens to
specify a difference of zero; however, it should be noted that the null hypothesis does
not always need to be about zero effect. The “null” in “null hypothesis” means it is the
hypothesis to be nullified by the statistical test. Having set up our null, we then
hypothesize a statement contrary to the null, known as the alternative hypothesis.
The alternative hypothesis is generally of two types. The first is the statistical alter-
native hypothesis, which is essentially and quite simply a statement of the comple-
ment to the null hypothesis. That is, it is a statement of “not the null.” Hence, if the
null hypothesis is rejected, the statistical alternative hypothesis is automatically
inferred. For our data, suppose after 30 days, the number of people surviving in the
experimental group is equal to 50, while the number of people surviving in the control
group is 20. Under the null hypothesis, we would have expected these survival rates to
be equal. However, we have observed a difference in our sample. Since it is merely
sample data, we are not really interested in this particular result specifically. Rather,
we are interested in answering the following question:
relatively high under the null, then we have no reason to reject the null hypothesis in
favor of the statistical alternative. However, if the probability of the sample result is
low under the null, then we take this as evidence that the null hypothesis may be false.
We do not know if it is false, but we reject it because of the implausibility of the data in
light of it. A rejection of the null hypothesis does not necessarily mean the null
is false. What it does mean is that we will act as though it is false or potentially
make scientific decisions based on its presumed falsity. Whether it is actually false or
not usually remains an unknown in many cases.
For our example, if the number of people surviving in each group in our sample were
equal to 50 spot on, then we definitely would not have evidence to reject the null
hypothesis. Why not? Because a sample result of 50 and 50 lines up exactly with what
we would expect under the null hypothesis. That is, it lines up perfectly with expecta-
tion under the null model. However, if the numbers turned up as they did earlier, 50
vs. 20, and we found the probability of this result to be rather small under the null,
then it could be taken as evidence to possibly reject the null hypothesis and infer the
alternative that the survival rates in each group are not the same. This is where the
substantive or research alternative hypothesis comes in. Why were the survival
rates found to be different? For our example, this is an easy one. If we did our experi-
ment properly, it is hopefully due to the treatment. However, had we not performed a
rigorous experimental design, then concluding the substantive or research hypoth-
esis becomes much more difficult. That is, simply because you are able to reject a null
hypothesis does not in itself lend credit to the substantive alternative hypothesis of
your wishes and dreams. The substantive alternative hypothesis should naturally drop
out or be a natural consequence of the rigorous approach and controls implemented
for the experiment. If it does not, then drawing a substantive conclusion becomes very
much more difficult if not impossible. This is one reason why drawing conclu-
sions from correlational research can be exceedingly difficult, if not impossi-
ble. If you do not have a bullet-proof experimental design, then logically it becomes
nearly impossible to know why the null was rejected. Even if you have a strong experi-
mental design such conclusions are difficult under the best of circumstances, so if you
do not have this level of rigor, you are in hot water when it comes to drawing strong
conclusions. Many published research papers feature very little scientific support for
purported scientific claims simply based on a rejection of a null hypothesis. This is due
to many researchers not understanding or appreciating what a rejection of the null
means (and what it does not mean). As we will discuss later in the book, rejecting a
null hypothesis is, usually, and by itself, no big deal at all.
The goal of scientific research on a statistical level is generally to learn about population
parameters. Since populations are usually quite large, scientists typically study statistics
based on samples and make inferences toward the population based on these samples.
Null hypothesis significance testing (NHST) involves putting forth a null hypothesis and then
evaluating the probability of obtained sample evidence in light of that null. If the probability of
such data occurring is relatively low under the null hypothesis, this provides evidence against
the null and an inference toward the statistical alternative hypothesis. The substantive
alternative hypothesis is the research reason for why the null was rejected and typically is
known or hypothesized beforehand by the nature of the research design. If the research design
is poor, it can prove exceedingly difficult or impossible to infer the correct research alternative.
Experimental designs are usually preferred for this (and many other) reasons.
We have discussed thus far that a null hypothesis is typically rejected when the prob-
ability of observed data in the sample is relatively small under the posited null. For
instance, with a simple example of 100 flips of a presumably fair coin, we would for
certain reject the null hypothesis of fairness if we observed, for example, 98 heads.
That is, the probability of observing 98 heads on 100 flips of a fair coin is very small.
However, when we reject the null, we could be wrong. That is, rejecting fairness
could be a mistake. Now, there is a very important distinction to make here. Rejecting
the null hypothesis itself in this situation is likely to be a good decision. We have
every reason to reject it based on the number of heads out of 100 flips. Obtaining 98
heads is more than enough statistical evidence in the sample to reject the null.
However, as mentioned, a rejection of the null hypothesis does not necessarily mean
the null hypothesis is false. All we have done is reject it. In other words, it is entirely
possible that the coin is fair, but we simply observed an unlikely result. This is the
problem with statistical inference, and that is, there is always a chance of being
wrong in our decision to reject a null hypothesis and infer an alternative.
That does not mean the rejection itself was wrong. It means simply that our decision
may not turn out to be in our favor. In other words, we may not get a “lucky out-
come.” We have to live with that risk of being wrong if we are to make virtually any
decisions (such as leaving the house and crossing the street or going shopping during
a pandemic).
The above is an extremely important distinction and cannot be emphasized enough.
Many times, researchers (and others, especially media) evaluate decisions based not
on the logic that went into them, but rather on outcomes. This is a philosophically
faulty way of assessing the goodness of a decision, however. The goodness of the deci-
sion should be based on whether it was made based on solid and efficient decision-
making principles that a rational agent would make under similar circumstances,
not whether the outcome happened to accord with what we hoped to see. Again,
sometimes we experience lucky outcomes, sometimes we do not, even when our deci-
sion-making criteria is “spot on” in both cases. This is what the art of decision-making
is all about. The following are some examples of popular decision-making events and
the actual outcome of the given decision:
■■ The Iraq war beginning in 2003. Politics aside, a motivator for invasion was presum-
ably whether or not Saddam Hussein possessed weapons of mass destruction. We
know now that he apparently did not, and hence many have argued that the inva-
sion and subsequent war was a mistake. However, without a proper decision anal-
ysis of the risks and probabilities beforehand, the quality of the decision should not
be based on the lucky or unlucky outcome. For instance, if as assessed by experts in
the area the probability of finding weapons of mass destruction (and that they would
be used) were equal to 0.99, then the logic of the decision to go to war may have been
a good one. The outcome of not finding such weapons, in the sense we are discuss-
ing, was simply an “unlucky” outcome. The decision, however, may have been cor-
rect. However, if the decision analysis revealed a low probability of having such
weapons or whether they would be used, then regardless of the outcome, the actual
decision would have been a poor one.
■■ The decision in 2020 to essentially shut down the US economy in the month of
March due to the spread of COVID-19. Was it a good decision? The decision should
not be evaluated based on the outcome of the spread or the degree to which it
affected people’s lives. The decision should be evaluated on the principles and logic
that went into the decision beforehand. Whether a lucky outcome or not was
achieved is a different process to the actual decision that was made. Likewise, the
decision to purchase a stock then lose all of one’s investment cannot be based on the
outcome of oil dropping to negative numbers during the pandemic. It must be
instead evaluated on the decision-making criteria that went into the decision. You
may have purchased a great stock prior to the pandemic, but got an extremely
unlucky and improbable outcome when the oil crash hit.
■■ The launch of SpaceX in May of 2020, returning Americans to space. On the day of
the launch, there was a slight chance of lightning in the area, but the risk was low
enough to go ahead with the launch. Had lightning occurred and it adversely
affected the mission, it would not have somehow meant a poor decision was made.
What it would have indicated above all else is that an unlucky outcome occurred.
There is always a measure of risk tolerance in any event such as this. The goal in
decision-making is generally to calibrate such risk and minimize it to an acceptable
and sufficient degree.
As discussed thus far, decision-making is risky business. Virtually all decisions are
made with at least some degree of risk of being wrong. How that risk is distributed and
calibrated, and the costs of making the wrong decision, are the components that must
be considered before making the decision. For example, again with the coin, if we start
out assuming the coin is fair (null hypothesis), then reject that hypothesis after obtain-
ing a large number of heads out of 100 flips, though the decision is logical, reality itself
may not agree with our decision. That is, the coin may, in reality, be fair. We simply
observed a string of heads that may simply be due to chance fluctuation. Now, how are
we ever to know if the coin is fair or not? That’s a difficult question, since according to
frequentist probabilists, we would literally need to flip the coin forever to get the true
probability of heads. Since we cannot study an infinite population of coin flips, we are
always restricted on betting based on the sample, and hoping our bet gets us a lucky
outcome.
What may be most surprising to those unfamiliar with statistical inference, is that
quite remarkably, statistical inference in science operates on the same philosophical
principles as games of chance in Vegas! Science is a gamble and all decisions have
error rates. Again, consider the idea of a potential treatment being advanced for
COVID-19 in 2020, the year of the pandemic. Does the treatment work? We hope so,
but if it does not, what are the risks of it not working? With every decision, there are
error rates, and error rates also imply potential opportunity costs. Good decisions are
made with an awareness of the benefits of being correct or the costs of being wrong.
Beyond that, we roll the proverbial dice and see what happens.
TESTIMONY II.
TESTIMONY IV.
Testimony of Capt. James Millar.
August 7. Mr. Blaisdel came to my house and desired me to go to
his own, where I might hear and see for myself. He also went to
Capt. Samuel Simson’s with the same request. Capt. Simson and his
wife, S—— B——, and N—— G——, who were there, came with him
to my house, and we all went to Mr. Blaisdel’s. When we had been
there some minutes, Capt. Simson, by desire, prayed. His prayer was
immediately followed by a knocking, and we all went into the cellar.
Mr. Blaisdel asked what was wanted, and who it was. It answered, “I
was once N. H.” I asked, “How was man made?” “Out of the dust,”
said the voice, “Dust thou art, and unto dust thou shalt return. You
have the Bible, and that is God’s truth, and do you abide by it. Love
God and keep his commandments.” After some conversation with
Mrs. Simson and others, she said, “I must go,” and we heard no
more. It was now broad day-light, the outer cellar door being open,
and utterly impossible that any living person should be there, but
those whom we could see and know. The voice was about six feet
from me.
August 9. I went to that house with many people, among whom I
observed much disorderly behaviour. The Spirit spake but little, and I
returned with a resolution to go no more to that house on such an
errand.
August 14. Just before day-light, I heard singing, as I lay in bed,
approaching to my house. Presently, by my leave, my house was
filled with people, and I heard knockings on the floor. By the desire
of certain persons, I went into the cellar with Capt. P—— B——.
After some discourse of the voice with him, which I understood not,
I heard sounds of knocking near me. I asked, what do you want of
me. It answered, “I have come to let you know that I can speak in
this cellar as well as in the other.[29] Are you convinced?” I
answered, “I am.” “Now,” said the voice, “the company must be
solemn and stand in order before your door, I am going to appear.
Now do you remember that I was once N. H.” We went up and
complied with her direction, and I saw a personal shape coming
toward us, white as the light. By the Spectre’s order, as I was
informed, Mrs. Butler went toward her. “Lydia,” said the Spectre,
“you are scared. You must sing.” Then she sung an hymn. The Spirit
came almost to us; then turned, and Mrs. Butler with her, and went
several rods towards Capt. Simson’s and appeared to take her by the
hand to urge her on further; and disappeared in our sight.
Mrs. Butler returned and informed the company, as I was told,
that if they would walk to Mr. Blaisdel’s solemnly as to a funeral, the
Spirit would walk with Mrs. Butler behind them. The company did so.
But I being far forward, saw nothing. Mrs. Butler had expressed her
unwillingness to go to Capt. Simson’s, and was excused, as she
afterwards told us. JAMES MILLAR.
TESTIMONY V.
Testimony of Mrs. M. G.
On the 4th of August, 1800, about two hours before day-light,
while I slept in Mr. Blaisdel’s house, I was waked by the sound of
knocking. I got up, and with about twenty others went into the
cellar. There I heard such a voice speaking to us as I never heard
before nor since. It was shrill, but very mild and pleasant.
Mr. Blaisdel, in addressing the voice, said that several persons, (of
whom I was one) had come from a distance to obtain satisfaction,
and desired that she would tell us who she was, and the design of
her coming. She answered, “I was once N. H. and after I was
married, I was N. B.” After much conversation upon the design of
her coming, she appeared to us. At first the apparition was a mere
mass of light: then grew into personal form, about as tall as myself.
We stood in two ranks about four or five feet apart. Between these
ranks she slowly passed and repassed, so that any of us could have
handled her. When she passed by me, her nearness was that of
contact; so that if there had been a substance, I should have
certainly felt it. The glow of the apparition had a constant tremulous
motion. At last the personal form became shapeless—expanded
every way, and then vanished in a moment.
Then I examined my own white gown and handkerchief, but I
could no more see them than if they had been black.
Nothing more being now seen or heard, we were moving to go
up, when the voice spake again and desired us to tarry longer. We
did so, and the Spirit talked with us another hour, even till broad
day-light. She mentioned to us the ill treatment which Mr. Blaisdel’s
family had suffered by reproach and false accusation, and told us
that they would, on her account, be more despised and ridiculed in
time to come, than they had been already.
Her discourse concluded by a solemn exhortation to the old, the
middle aged and the young. The present life, she said, was but a
moment, in which we must be renewed or be miserable forever.
In her address to the youth she observed that it was now the
Lord’s day, and that we must retire to our homes, read the Bible,
pray and keep the day holy. It was then she uttered these lines of
Dr. Watts.
“This is the day when Christ arose
So early from the dead;
Why should I keep my eyelids closed,
And waste my hours in bed?”
After speaking much more which I cannot remember, she sang
praises and left us.
Her notes were very pleasant. Her words were in no higher style
than common, yet were they exceedingly impressive.
MARY GORDON.
TESTIMONY VI.
Testimony of Mrs. Sally Wentworth.
On the 2d of January 1800, Hannah Blaisdel came to Mr. Butler’s
house and informed me that the extraordinary voice which they had
heard, had declared itself to be that of my sister, and that I must go
to her father’s house.
I told her to her face that I did not believe it.
The next day I received the same message by three other
persons, two of which belonged to two other families, and returned
the same answer. Nevertheless, to give satisfaction, Capt. Butler, Mr.
Wentworth and I went with them to that house. Capt. Butler and I
examined the cellar with a candle, and in a few minutes after, Lydia
and I went down there. Capt. S——n and some others, went with
us, but none of them stood before us. While I held Lydia by the arm,
we heard the sound of knocking. Lydia spoke, and a voice answered,
the sound of which brought fresh to my mind that of my sister’s
voice, in an instant; but I could not understand it at all; though it
was within the compass of my embrace, and, had it been a creature
which breathed, it would have breathed in my face, and I had no
impediment of hearing. But Lydia told me that it said, “We must live
in peace and be united.” Then we came up. But Capt. S——n with
Lydia and others, went down again. I passed through the room
which led to the cellar into another room, and there I was much
surprised when I plainly understood by the same kind of voice, still
speaking in the cellar, these words, “I am the voice of one crying in
the wilderness;” and other sentences, which I cannot remember.
This is testified by several others who were with me.[30]
From this time I cleared Lydia as to the voice, and accused the
devil.
August 8, I was there again with about thirty others, and heard
much conversation. Her voice was still hoarse and thick, like that of
my sister on her death bed,[31] but more hollow. Sometimes it was
clear, and always pleasant. A certain person did, in my opinion very
unwisely, ask her whether I was a true Christian. The reply was,
“She thinks she is, she thinks she is. She is my sister.”
August 13-14, I heard the same voice in the same place, and did
then believe it was that of my sister. She talked much with Capt. S
——n, and exhorted the people. Mr. Sp——r asked her if I believed
that she was my sister. The answer was, “She believes now.” By the
direction of the Spirit we went to Capt. M——r’s, but I never saw her.
Sometime after this, Mr. Butler brought to me from the Spectre,
the private conversation which I know I had with my sister in her
lifetime, at a certain hour, when we were alone together, and which
he declares he never knew before; as a token that I was her sister.
It is true I had never revealed it to any person, and I do not believe
that my sister ever did; but could not some evil spirit hear that
conversation, and afterwards personate my sister, and reveal it to
Capt. Butler? For what purpose should my sister become visible to
us?—There was certainly no dispute nor difficulty in my father’s
family or that of Mr. Butler’s, which could be any reason for her
coming. SALLY WENTWORTH.
Mrs. Wentworth had now an opportunity to hear the voice of Lydia and the
voice of the Spectre in the same time and place, that she might have the best
advantage to judge whether or not there was the least agreement between them.
And that Lydia had never learned to utter two voices in the same minute, the one
her natural voice, the other the dying voice of this woman’s sister, appears from
the certainty that through all the time of the Spectre’s last sickness and death,
Lydia was two hundred miles distant from her.
When Mrs. Wentworth heard in the east room that sentence of the ghost, “I am
the voice of one crying in the wilderness;” this was the only time in which the
ghost uttered these words for that day, as several witnesses (nem. con.) declare.
Hence it follows that this was the exact minute when Captain Simson, in the cellar,
within eight feet of the voice, and free from deafness, heard only a sound, while
they who stood by him understood the words plainly. (Compare this with a
sentence in Test. 4, 2d part.)
The reality of the token appears from the undoubted veracity of Mrs. W., her
inflexible opposition and the oath of Capt. Butler, the reputed dupe of the whole
business.
TESTIMONY VII.
TESTIMONY I.
TESTIMONY II.
Testimony of Miss Mary Card.
I am not only a witness to many things in the preceding relation,
but I further declare that, on August 13, about two o’clock in the
day, while Mr. Blaisdel was gone for evidence, the Spirit knocked,
and M. M. and I went near to the place of the sound and asked what
was wanted.
The Spirit answered “I have come, I have come. Make room for I
am coming among you.”[34] I pleaded that she would not. I must, I
must, said she, don’t be scared. I answered we were poor sinful
creatures, and could not help it, and again earnestly entreated her
not to come.
I now plainly saw her appear in shining white; and she asked me
if any person in the house wanted to see her. I then called on Miss P.
C. who in the same hour had said that “though she had heard the
voice speak, she would not believe it was that of a ghost.” I asked
her now to come and see her, and she did. Now, said the ghost,
satisfy yourselves. Here I am, here I am, satisfy yourselves. Miss P.
C. answered that she was satisfied.[35] The ghost then spake several
other things which I cannot remember. M. CARD.
TESTIMONY III.
TESTIMONY IV.
Testimony of Capt. Paul Simson.
January 3, 1800, I was at the house of Mr. Blaisdel. His son P.
desired me to go with him into the cellar. I went down with him and
his two sisters, and Mrs. C. M.
I heard a rapping, and asked in the name of Christ what it
wanted. I heard a voice considerably loud, but could not understand
it. But some who were present, told me that it said, “I am the voice
of one crying in the wilderness. Prepare ye the way of the Lord.
Make his paths strait. Seek ye the Lord, while he may be found; call
upon him while he is near.”
After some silence it rapped again. I spake to it in the name of
Christ, and said, if there was any thing it could utter for peace, to
utter it. It answered, “I am not to be trifled with. I am not to be
trifled with. I am not to be trifled with. Peace, peace, peace.”[36]
Then they all went up, except Lydia and myself, and I held her by
the hand. She was much terrified, and said, “I feel so I cannot stay:
I must go up.” Stay awhile, said I, perhaps it will speak again. “I
cannot,” said she, and began to urge me away. I consented, and
when we had come up, she told me what I had myself perceived,
that the cellar began to grow light, where the voice was uttered and
that she heard a rushing noise.
Sometime after I was in the same cellar with a number of people
and heard a plain voice; clearly understood by others, but not at all
by myself, though as near to it as others and free from deafness.
The voice appeared to me inimitable.
August 8, I was there again with thirty others and heard the
conversation of the Spectre with several persons. Mr. N. H.
mentioned his desire to handle her. “Handle me and see,” said she,
“for Christ tells you that a spirit is not flesh and bones.”
Mr. T. U. said, “if you are a happy soul, intercede for me.” The
reply was, “None but Christ intercedes.” “There are among us,” said
Mr. U., “several denominations of Christians: Presbyterians,
Congregationalists, Baptists, and Methodists. Which of all these are
right?” The voice answered, “There are good and bad of all these
sorts for the elect’s sake.”[37]
In August 13-14, while I was at the same house, the Spirit
informed us that she could not speak freely in the night of August 9,
because the behaviour of the people had been so rude, but
expressed her joy in discoursing with christians. You know, said I, a
thousand times more than any of us. Yes I do, said she. Mr. Blaisdel
said to me, you stand too near her. Then I asked, do I stand too
near you? No, said she, stand as near as you please. I felt surprised,
and said, it was a wonderful event. Yes, the spirit replied, it is a
wonderful event indeed—do you not remember what you told me,
just after I was married that if I was married to Christ, how happy I
should live? Yes, I answered, I remember it very well. She exhorted
the young people and told them that without a change they would
be miserable.
After this, the spirit expressed her resolution to convince us that
Mr. Blaisdel and his family were clear of the evils alleged against
them.
They say I am a witch and a devil, said she; and they said that
Christ was a devil. The spirit said other things, which I remember
not.
At last she told them she was about to appear in order to
convince them. Then by her direction we went up, and having
prayed together, and heard the token, we walked on to Capt. M’s,
singing the 84th Psalm. I was one of the foremost of the company
and did not see it then; but the greatest number of those who were
behind me, said they saw it plainly.
When we were at Capt. M’s house, we stood in the field, while
Mrs. Butler, in great fear, walked with the Spirit before us,[38] a few
rods toward Capt. Simson’s; and then returned and told us that we
must walk back, two and two, to her father’s house singing, and the
spirit would follow us back. We did so. Mr. D. and I walked behind
all, except Mrs. Butler, in order if possible to see the apparition.
When we had walked about fifteen rods, I saw a white appearance
forward of us to the left hand. As we passed by it, it fell in after us
and walked with Mrs. Butler.
Mr. D. and I turned and looked upon them, and heard them talk.
We walked a little way further, stopped and looked upon them, and
heard them talk again; but they spoke with so low a voice we could
not understand them. The spirit appeared in personal form, white as
snow, about as tall as Mrs. Butler.
It was now day-break. I turned my eyes from the object, and in
half a minute looked toward it again, but it was gone. Mr. D. then
told me he saw it disappear. PAUL SIMSON.
TESTIMONY V.
TESTIMONY VI.
TESTIMONY VIII.
TESTIMONY IX.
TESTIMONY X.
TESTIMONY XI.
TESTIMONY XII.
TESTIMONY XIV.
TESTIMONY XV.
TESTIMONY XVI.
TESTIMONY XVII.
Our website is not just a platform for buying books, but a bridge
connecting readers to the timeless values of culture and wisdom. With
an elegant, user-friendly interface and an intelligent search system,
we are committed to providing a quick and convenient shopping
experience. Additionally, our special promotions and home delivery
services ensure that you save time and fully enjoy the joy of reading.
textbookfull.com