Download Complete (Ebook) Sparse Estimation with Math and Python: 100 Exercises for Building Logic by Joe Suzuki ISBN 9789811614378, 9811614377 PDF for All Chapters
Download Complete (Ebook) Sparse Estimation with Math and Python: 100 Exercises for Building Logic by Joe Suzuki ISBN 9789811614378, 9811614377 PDF for All Chapters
com
https://ebooknice.com/product/sparse-estimation-with-math-
and-python-100-exercises-for-building-logic-35997448
OR CLICK HERE
DOWLOAD EBOOK
ebooknice.com
ebooknice.com
ebooknice.com
https://ebooknice.com/product/sparse-estimation-with-math-
and-r-33795076
ebooknice.com
(Ebook) doing math with python doing math with python by
kan
https://ebooknice.com/product/doing-math-with-python-doing-math-with-
python-50196050
ebooknice.com
ebooknice.com
ebooknice.com
ebooknice.com
ebooknice.com
Joe Suzuki
This work is subject to copyright. All rights are solely and exclusively
licensed by the Publisher, whether the whole or part of the material is
concerned, specifically the rights of reprinting, reuse of illustrations,
recitation, broadcasting, reproduction on microfilms or in any other
physical way, and transmission or information storage and retrieval,
electronic adaptation, computer software, or by similar or dissimilar
methodology now known or hereafter developed.
The publisher, the authors and the editors are safe to assume that the
advice and information in this book are believed to be true and accurate
at the date of publication. Neither the publisher nor the authors or the
editors give a warranty, expressed or implied, with respect to the
material contained herein or for any errors or omissions that may have
been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer
Nature Singapore Pte Ltd.
The registered company address is: 152 Beach Road, #21-01/04
Gateway East, Singapore 189721, Singapore
Preface
I started considering the sparse estimation problems around 2017
when I moved from the mathematics department to statistics in Osaka
University, Japan. I have been studying information theory and
graphical models for over thirty years.
The first book I found is “Statistical Learning with Sparsity” by T.
Hastie, R. Tibshirani, and M. Wainwright. I thought it was a monograph
rather than a textbook and that it would be tough for a non-expert to
read it through. However, I downloaded more than fifty papers that
were cited in the book and read them all. In fact, the book does not
instruct anything but only suggests how to study sparsity. The contrast
between statistics and convex optimization gradually attracted me as I
understand the material.
On the other hand, it seems that the core results on sparsity have
come out around 2010–2015 for research. However, I still think further
possibilities and expansions are there. This book contains all the
mathematical derivations and source programs, so graduate students
can construct any procedure from scratch by getting help from this
book.
Recently, I published books “Statistical Learning with Math and R”
(SLMR), “Statistical Learning with Math and Python” (SLMP), and
“Sparse Estimation with Math and R” (SEMR). The common idea is
behind the books (XXMR/XXMP). They not only give knowledge on
statistical learning and sparse estimation but also help build logic in
your brain by following each step of the derivations and each line of the
source programs. I often meet data scientists engaged in machine
learning and statistical analyses for research collaborations and
introduce my students to them. I recently found out that almost all of
them think that (mathematical) logic rather than knowledge and
experience is the most crucial ability for grasping the essence in their
jobs. Our necessary knowledge is changing every day and can be
obtained when needed. However, logic allows us to examine whether
each item on the Internet is correct and follow any changes; we might
miss even chances without it.
What makes SEMP unique?
I have summarized the features of this book as follows.
1. Developing logic
This book explains how to use the package and provides examples
of executions for those who are not familiar with them. Still,
because only the inputs and outputs are visible, we can only see the
procedure as a black box. In this sense, the reader will have limited
satisfaction because they will not be able to obtain the essence of
the subject. SEMP intends to show the reader the heart of ML and is
more of a full-fledged academic book.
The reader can ask any question on the book via https://bayesnet.
org/books.
Exercises 1–20
2 Generalized Linear Regression
2.1 Generalization of Lasso in Linear Regression
2.2 Logistic Regression for Binary Values
2.3 Logistic Regression for Multiple Values
2.4 Poisson Regression
2.5 Survival Analysis
Appendix Proof of Proposition
Exercises 21–33
3 Group Lasso
3.1 When One Group Exists
3.2 Proxy Gradient Method
3.3 Group Lasso
3.4 Sparse Group Lasso
3.5 Overlap Lasso
3.6 Group Lasso with Multiple Responses
3.7 Group Lasso Via Logistic Regression
3.8 Group Lasso for the Generalized Additive Models
Appendix Proof of Proposition
Exercises 34–46
4 Fused Lasso
4.1 Applications of Fused Lasso
4.2 Solving Fused Lasso Via Dynamic Programming
4.3 LARS
4.4 Dual Lasso Problem and Generalized Lasso
4.5 ADMM
Appendix Proof of Proposition
Exercises 47–61
5 Graphical Models
5.1 Graphical Models
5.2 Graphical Lasso
5.3 Estimation of the Graphical Model Based on the Quasi-
Likelihood
5.4 Joint Graphical Lasso
Appendix Proof of Propositions
Exercises 62–75
6 Matrix Decomposition
6.1 Singular Decomposition
6.2 Eckart-Young’s Theorem
6.3 Norm
6.4 Sparse Estimation for Low-Rank Estimations
Appendix Proof of Propositions
Exercises 76–87
7 Multivariate Analysis
7.1 Principal Component Analysis (1):SCoTLASS
7.2 Principle Component Analysis (2):SPCA
7.3 -Means Clustering
7.4 Convex Clustering
Appendix Proof of Proposition
Exercises 88–100
References
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021
J. Suzuki, Sparse Estimation with Math and Python
https://doi.org/10.1007/978-981-16-1438-5_1
1. Linear Regression
Joe Suzuki1
(1) Graduate School of Engineering Science, Osaka University, Toyonaka, Osaka, Japan
Joe Suzuki
Email: [email protected]
In general statistics, we often assume that the number of samples N is greater than the
number of variables p. If this is not the case, it may not be possible to solve for the best-
fitting regression coefficients using the least squares method, or it is too
computationally costly to compare a total of models using some information
criterion.
When p is greater than N (also known as the sparse situation), even for linear
regression, it is more common to minimize, instead of the usual squared error, the
modified objective function to which a term is added to prevent the coefficients from
being too large (the so-called regularization term). If the regularization term is a
constant times the L1-norm (resp. L2-norm) of the coefficient vector, it is called Lasso
(resp. Ridge). In the case of Lasso, if the value of increases, there will be more
coefficients that go to 0, and when reaches a certain value, all the coefficients will
eventually become 0. In that sense, we can say that Lasso also plays a role in model
selection.
In this chapter, we examine the properties of Lasso in comparison to those of Ridge.
After that, we investigate the elastic net, a regularized regression method that combines
the advantages of both Ridge and Lasso. Finally, we consider how to select an
appropriate value of .
is denoted by .
First, for the sake of simplicity, we assume that the jth column of X,
and y have already been centered. That is, for each , define
, and assume that has already been subtracted from each so that is
satisfied. Similarly, defining , we assume that was subtracted in advance
from each so that holds. Under this condition, one of the parameters
holds. Thus, from now, without loss of generality, we may assume that the intercept
is zero and use this in our further calculations.
We begin by first observing the following equality:
(1.1)
Thus, when we set the right-hand side of (1.1) equal to 0, and if is invertible, then
becomes
(1.2)
For the case where , write each column of X as ; we see that
(1.3)
If we had not performed the data centering, we would still obtain the same slope ,
(1.4)
In this book, we focus more on the sparse case, i.e., when p is larger than N. In this
case, a problem arises. When , the matrix does not have an inverse. In fact,
since
(that is, whether we choose each of the variables or not). This means we have to
compare a total of models. Then, to extract the proper model combination by using
an information criterion or cross-validation, the computational resources required will
grow exponentially with the number of variables p.
To deal with this kind of problem, let us consider the following. Let be a
constant. We add a term to that penalizes for being too large in size.
Specifically, we define
(1.5)
or
(1.6)
Our work now is to solve for the minimizer to one of the above quantities. Here,
for , is the L1-norm and is the L2-
either (1.5) or (1.6), then minimize it with respect to the slope (this is called
1.2 Subderivative
To address the minimization problem for Lasso, we need a method for optimizing
functions that are not differentiable. When we want to find the points x of the maxima
or minima of a single-variable polynomial, say, , we can differentiate
(1.7)
Therefore, it is not convex (Fig. 1.1, right). If functions f, g are convex, then for any
, the function has to be convex since the following holds:
Fig. 1.1 Left: is convex. However, at the origin, the derivatives from each side differ; thus, it
is not differentiable. Right: We cannot simply judge from its shape, but this function is not convex
Next, for any convex function , fix arbitrarily. For all , we
say that the set of all that satisfies
(1.8)
is a subderivative of f at .
If f is differentiable at , then the subderivative will be a set that contains only 1
element, say, .2 We prove this as follows.
First, the convex function f is differentiable at ; thus, it satisfies
. To see this, since f is convex,
This can be rewritten as
left and, at the same time, be less than or equal to the derivative on the right at . Since
f is differentiable at , those 2 derivatives are equal; this completes the proof.
The main interest of this book is specifically the case where and .
Hence, by (1.8), its subderivative is the set of z such that for any , . These
values of z lie in the interval greater than or equal to and less than or equal to 1, and
is true. Let us confirm this. If for any x, holds, then for and ,
and , respectively, need to be true. Conversely, if , then
is true for any arbitrary .
Example 1 By dividing into 3 cases , , and , find the values x that
attain the minimum of and . Note that for , we can find
. Neither is differentiable at . Despite not being differentiable, the point is a minimum for
1.3 Lasso
As stated in Sect. 1.1, the method considered for the minimization of
(1.9)
holds, and let . With this assumption, the calculations are made much
simpler.
Solving for the subderivative of L with respect to gives
(1.10)
(1.11)
At a certain , all the coefficients will be 0. The at which each coefficient becomes 0 varies
As we can see in Fig. 1.4, as increases, the absolute value of each coefficient
decreases. When reaches a certain value, all coefficients go to 0. In other words, for
each value of , the set of nonzero coefficients differs. The larger becomes, the
smaller the set of selected variables.
When performing coordinate descent, it is quite common to begin with a large
enough that every coefficient is zero and then make smaller gradually. This method is
called a warm start, which utilizes the fact that when we want to calculate the
coefficient for all values of , we can improve the calculation performance by setting
the initial value of to the estimated from a previous . For example, we can write
the program as follows:
Example 3 We use the warm start method to reproduce the coefficient for each
in Example 2.
First, we make the value of large enough so that all are 0 and
then gradually decrease the size of while performing coordinate descent. Here, for
simplicity, we assume that for each , we have ; moreover, the
values of are all different. In this case, the smallest that makes
larger than this formula, it will be satisfied that for all , . Then, we
have that and
hold. Thus, when we decrease the size of , one of the values of j will satisfy
. Again, since ( ), if we continue to make it
Fig. 1.5 The execution result of Example 4. The numbers at the top represent the number of estimated
coefficients that are not zero
The glmnet package is often used [11].
Example 4 (Boston) Using the Boston dataset and setting the variable of the 14th
column as the target variable and the other 13 variables as predictors, we plot a graph
similar to that of the previous one (Fig. 1.5).
So far, we have seen how Lasso can be useful in the process of variable selection.
However, we have not explained the reason why we seek to minimize (1.5) for .
Why do we not instead consider minimizing the following usual information criterion
(1.12)
1.4 Ridge
In Sect. 1.1, we made an assumption about that the matrix is
invertible, and, based on this, we showed that the that minimizes the squared error
is given by .
First, when , the possibility that is singular is not that high, though we
may have another problem instead: the confidence interval becomes significant when
the determinant is small. To cope with this, we let be a constant and add to the
squared error the norm of times . That is, the method considering the minimization
of
($1.6$)
is commonly used. This method is called Ridge. Differentiating L with respect to gives
positive, which is the same as saying that is nonsingular. Note that this
less than or equal to N, and hence, the matrix is singular. Therefore, for this case, the
following conditions are equivalent:
Example 5 We use the same U.S. crime data as that of Example 2 and perform the
following analysis. To control the size of the coefficient of each predictor, we call the
function ridge and then execute.
In Fig. 1.6, we plot how each coefficient changes with the value of .
Fig. 1.6 The execution result of Example 5. The changes in the coefficient with respect to based
give the same value of (1.13). The rhombus in the left figure is the L1 regularization constraint
, while the circle in the right figure is the L2 regularization constraint
the least squares process, we solve for the values that minimize
. For now, let us denote them by , respectively. Here,
if we let , we have
(1.13)
and, of course, if we let here, we obtain the minimum ( RSS).
Thus, we can view the problems of Lasso and Ridge in the following way: the
minimization of quantities (1.5) and (1.6) is equivalent to finding the values of
that satisfy the constraints and , respectively, that also
minimize the quantity of (1.13) (here, the case where is large is equivalent to the case
where are small).
The case of Lasso is the same as in the left panel of Fig. 1.7. The ellipses are centered
at and represent contours on which the values of (1.13) are the same. We
expand the size of the ellipse (the contour), and once we make contact with the
rhombus, the corresponding values of are the solution to Lasso. If the rhombus
is small ( is large), it is more likely to touch only one of the four rhombus vertices. In
this case, one of the values will become 0. However, in the Ridge case, as in the
right panel of Fig. 1.7, a circle replaces the Lasso rhombus; hence, it is less likely that
will occur.
In this case, if the least squares solution lies in the green zone of Fig. 1.8,
larger, which is the reason why Lasso performs well in variable selection.
Fig. 1.8 The green zone represents the area in which the optimal solution would satisfy either
the VIF (variance inflation factor). The larger this value is, the better the jth column
variable is explained by the other variables. Here, denotes the coefficient of
determination squared in which is the target variable, and the other variables are
predictors.
Example 6 We compute the VIF for the Boston dataset. It shows that the 9th and 10th
variables (RAD and TAX) have strong collinearity.
In usual linear regression, if the VIF is large, the estimated coefficient will be
unstable. Particularly, if two columns are precisely the same, the coefficient is
unsolvable. Moreover, for Lasso, if two columns are highly related, then generally one of
them will be estimated as 0 and the other as nonzero. However, in the Ridge case, for
, even when the columns j, k of X are the same, the estimation is solvable, and both
of them will obtain the same value.
In particular, we find the partial derivative of
the coefficients change relative to the value of in Fig. 1.9. Naturally, one might expect
that similar coefficient values should be given to each of the related variables, though it
turns out that Lasso does not behave in this way.
Fig. 1.9 The execution result of Example 7. The Lasso case is different from that of Ridge: related
variables are not given similar estimated coefficients. When we use the glmnet package, its default
horizontal axis is the L1-norm [11]. This is the value , which becomes smaller as increases.
Therefore, the figure here is left-/right-reversed compared to that where or is set as the
horizontal axis
(1.14)
This is called the elastic net. For each , if we find the subderivative of (1.14)
with respect to , we obtain
fact that
more it is able to handle collinearity, which is in contrast to the case of (Lasso), where the
Example 9 We apply the function cv.glmnet to the U.S. crime dataset from
Examples 2 and 5, obtain the optimal , use it for the usual Lasso, and then obtain the
coefficients of . For each , the function also provides the value of the least squares of
the test data and the confidence interval (Fig. 1.11). Each number above the figure
represents the number of nonzero coefficients for that .
For the elastic net, we have to perform double loop cross-validation for both .
The function cv_glmnet provides us the output cvm, which contains the evaluation
values from the cross-validation.
Example 10 We generate random numbers as our data and try conducting the double
loop cross-validation for .
For the problem, we changed the values and compared the evaluation for the best
to find that the difference is within 1% through .
Exercises 1–20
In the following exercises, we estimate the intercept and the coefficients
from N groups of p explanatory variables and a target variable data
. We subtract from each the
and from each the such that each of
them has mean 0 and then estimate (the estimated value is denoted by ). Finally, we
let . Again, we let and
for a function called linear with Python. This function accepts a matrix
and the estimated slope as outputs. Please complete blanks (1), (2) below.
3.
(a)
(b)
(a) If the functions g(x), h(x) are convex, show that for any , the function
is convex.
(c) Decide whether the following functions of are convex or not. In addition,
provide the proofs.
(i)
(ii)
(iii)
At this point comes the tug of war. If cows have pox, how do they
contract the malady? Speaking at the London Conference on Animal
Vaccination in December, 1879, Professor J. B. Simonds, Principal of
the Royal Veterinary College, said—
My contention is, that the existence of Cowpox has to be proved.
Jenner’s account of the disease was an illusion. In my experience among
animals for forty years, I have never seen a case of Cowpox, and I do not
believe that any form of variola belongs to the bovine race. Sheep are
afflicted with pox, but not cattle. We hear of Cowpox, but who ever heard
of Bullpox? And is it credible that a disease should be confined to cows
and never attack bulls and steers? Let any one point out an affection of
females that does not extend to the males of the same species.
JOHN BIRCH.
Birch wished to know what cowpox was. Jenner had said it was
derived from horsegrease, but “that origin is proved to be erroneous,
and is now given up, even by his best friends. On all hands it is
admitted,” Birch continued, “that it is not a disease of the cow, but
communicated to the cow by the milker. No cow that is allowed to
suckle her own calf ever has the complaint.” What, then, is the
disease in the milker? asked Birch. Is it smallpox? Is it lues venerea?
Is it itch? A man came to St. Thomas’s Hospital from an adjacent
dairy with a hand and arm covered with ulcerations. He said several
of the milkers and the cows’ teats were affected in the same way,
and he was told they had got cowpox. Birch called one of his country
pupils and asked him what was wrong with the man. “It is itch—rank
itch,” was his reply. A box of Jackson’s ointment for the itch was
given to him, and at the end of a week he reappeared at the
Hospital cured. If cowpox be itch, argued Birch—
Then if a patient be inoculated with the disorder, though it may suspend
the capacity for Smallpox for a season in the constitution, it will ultimately
prove no security—
Which was to say, that it was not probable that smallpox and itch
could occur together, and that a person inoculated with itch would
pass through the variolous test successfully. In this connection we
may recall the fact that Jenner found it impossible to vaccinate a
regiment at Colchester, the men with their women and children all
being afflicted with itch.
Still farther to complicate the mystery of cowpox, Jenner began to
describe it as genuine and spurious, but which was the one and
which was the other he left in bewildering uncertainty. Said Birch—
Though Dr. Jenner could not tell us what Cowpox was, he soon came
forward to inform us that it was of two sorts—the one genuine and
harmless, the other spurious and hurtful.
Spurious Cowpox is a term I do not admit of. I know of no such thing as
spurious Smallpox, spurious Measles, spurious Lues Venerea, spurious
Scrofula.
From out the muddle as to the origin of cowpox and its genuine
and spurious varieties, Birch demanded, What had Jenner
discovered? It is not that cowpox prevents smallpox; for that has
been asserted by dairy-folk for generations, and has been
disregarded by physicians because proved to be untrue. What then
is it? Let him define his discovery that we may know how to respect
it. Let him explain why it is forbidden to inoculate direct from the
cow. Is genuine cowpox invisible and to be taken on trust? Or is the
disease so virulent on its first communication that it has to be
meliorated in the body of some victim ere it is fit for public use?
Birch asked these questions as we continue to ask What is cowpox?
Is it a disease of the cow? Or is it communicated to the cow by man
or by horse? However definite the answers, the contradictions are
equally definite, and the authorities equally trustworthy.
Practical men answered for Jenner, as they presume to answer at
this day, “Whatever may be the origin of cowpox, we know that
vaccination is harmless, and that it prevents smallpox; and more we
neither demand nor care to inquire.” The credulity and conceit of
such practical men is that stupidity against which, says Goethe, even
the gods are powerless. It was practical men who “on the mere
show of reason” accepted vaccination before it could be tested, and
on most superficial evidence, said Birch—
Recommended Dr. Jenner to the munificence of Parliament for a
discovery in practice which was never to prove fatal; which was to excite
no new humours or disorders in the constitution; which was to be, not
only a perfect security against Smallpox, but would, if universally adopted,
prevent its recurrence for ever.
Nor did Brown rest satisfied with proving that vaccination did not
prevent smallpox. He also showed the fallacy of the variolous test.
He adduced twelve cases in which vaccinated persons had been
variolated as if they had never been vaccinated. Also four cases in
which vaccination and variolation were effected simultaneously, the
diseases running their courses concurrently, proving there was no
antagonism between them; and since they could occur together,
what reason was left for supposing that one might not succeed the
other?
Having found liberty in the truth, he reverted to Jenner’s writings,
and reading them with opened eyes, he was not slow to detect and
to demonstrate the laxity of statement, the contradictions, and
absurdities with which they were pervaded. No reply was attempted:
no reply, indeed, was possible. The surgeons of the Edinburgh
Vaccine Institution issued An Examination of Mr. Brown’s Opinions
and Statements,[167] but they merely carped over non-essential
details, and left the main issues wholly unaffected. What they had to
show was that Brown’s patients were either unvaccinated, or had not
had smallpox; and unable to do this, they were unable to do
anything.
Brown remained victor. He did not overthrow vaccination, nor
restore variolation, but he did make an end in Scotland of confidence
in vaccination as an omnipotent safeguard against smallpox. The rite
continued to be practised on humbler terms: “it did no harm”: even
Mr. Brown allowed that it might keep off smallpox for a time: and
“there was reason to believe that it tended to make the disease
milder when it did occur.” Thirty years after his first publication, in
1842, Brown reaffirmed his position in a series of letters[168] to Dr.
George Gregory, a sympathetic friend, and advised a return to
variolation in view of “the acknowledged defects of the Jennerian
practice”—a dismal alternative. But it is in vain to expect any man to
be much in advance of his time: it suffices for honourable distinction
that he be in advance. When Brown commenced practice, smallpox
and other fevers were regarded as inevitable as storms and
earthquakes, and the knowledge with which we are now so familiar,
that they are engendered in foul habits and habitations, was for
practical purposes unknown. Our reproach is, that knowing so much
better, we surrender ourselves to a superstitious observance
conceived in days of darkness.
FOOTNOTES:
[164] Cases of Smallpox subsequent to Vaccination, with Facts
and Observations read before the Medical Society at Portsmouth,
29th March, 1804: addressed to the Directors of the Vaccine
Institution. By William Goldson. Portsea, 1884.
[165] Baron’s Life of Jenner, vol. ii. pp. 338, 346, 348.
[166] An Inquiry into the Anti-Variolous Power of Vaccination; in
which from the state of the Phenomena and the Occurrence of a
great variety of Cases, the most Serious Doubts are suggested of
the Efficacy of the Whole Practice, and its Powers at best proved
to be only temporary. From which also will appear the Necessity
of and the proper period for again submitting to Inoculation with
Variolous Virus. By Thomas Brown, Surgeon, Musselburgh.
Edinburgh, 1809. Pp. 307.
[167] Report of the Surgeons of the Edinburgh Vaccine
Institution, containing an Examination of the Opinions and
Statements of Mr. Brown of Musselburgh on Vaccination.
Edinburgh, 1809.
[168] An Investigation of the Present Unsatisfactory and
Defective State of Vaccination, and the Several Expedients
proposed for removing the now Acknowledged Defects of the
Jennerian practice. In a Series of Letters addressed to Dr. George
Gregory, Physician to the Smallpox and Vaccination Hospital,
London; which also are intended as an Answer to the Queries of
the Academy of Science in Paris, proposed as the subject of a
Prize Essay. By Thomas Brown, formerly Medical Practitioner in
Musselburgh. Edinburgh, 1842, pp. 139.
CHAPTER XXI.
Furious was the controversy, but why was it furious? There are
often great fights over little matters, but the reason is that the little
matters are vitally related to the self-love of the combatants; and
thus it was with the Cowpoxers and the Smallpoxers. The Cowpoxers
set out with the absolute assertion that whoever submitted to their
prescription would be secure from smallpox for life. Without proof, or
with powerful sham proof, the assertion was endorsed by the mass
of the medical profession, and there followed the conversion of the
community in that sort of faith-panic which is described by Carlyle as
Swarmery—
All the world assenting, and continually repeating and reverberating,
there soon comes that singular phenomenon called Swarmery, or the
gathering of men in swarms; and what prodigies they are in the habit of
doing and believing when thrown into that miraculous condition! Singular,
in the case of human swarms, with what perfection of unanimity and
quasi-religious conviction the stupidest absurdities can be received as
axioms of Euclid, nay, as articles of faith, which you are not only to
believe, unless malignantly insane, but are (if you have any honour or
morality) to push into practice, and, without delay see done , if your soul
would live. [169]
Hill and Jenner were great friends. Hill visited Jenner at Berkeley,
and Jenner heard Hill when he preached at Cheltenham. Introducing
Jenner to a nobleman, Hill remarked—
“Allow me to present to your Lordship my friend Dr. Jenner, who
has been the means of saving more lives than any other man.”
To which Jenner, being of a pious turn, sighed with meek effusion
—
“Ah! would I, like you, could say souls.”[172]
So committed and so possessed, Hill naturally resented the
growing distrust of vaccination. It cut him deeply to be supposed a
quack; and in 1806 he issued a pamphlet[173] relating his
experiences as a Jennerite, defending his practice, and denouncing
those who treated it despitefully. Moseley especially was subjected
to severe and contemptuous condemnation. Hill’s sanctimony and
virulence, his vigour and venom compose a piquant mixture, and if
we could tarry for amusement we might produce it abundantly from
a variety of elegant extracts. Consider, for instance, this his
adjuration, and its pitiful object—
Oh, the blessing of the Jennerian inoculation! Did ever man stand as
Jenner so much like an Angel of God, an instrument in the hands of Divine
Providence between the living and the dead till the plague was stayed!
Talk evidently sincere, and widely repeated, but with how little
consideration for truth!
To return to Moseley. He was not the man to endure Hill’s
aggression submissively, and in a pamphlet entitled An Oliver for a
Rowland,[174] he made a terrific reprisal. The public were delighted
with it, and in the course of a few months it ran through ten
editions. Hill was generally regarded as a clerical mountebank, with
more impudence than piety, and to see him knocked over, kicked,
and rolled in the mire, was sport that carried many sympathisers.
Moseley’s opening address gives the key to the whole performance—
R owland ,—I bought your pamphlet, entitled Cowpox Inoculation
Vindicated, dated the 25th of March, 1806.
I paid a shilling for it. Rowland,—it is not dear. The same quantity of
folly, falsehood, and impudence, could not have been bought for twice the
money of any other Cowpoxer from the Ganges to the Mississippi.
But let me ask you, Rowland, what could induce you to take up your
pen to attack me on the subject of Physic, who never attacked you on the
subject of Religion? Would it not have been more prudent in you to have
continued to expose yourself in your own trade in your own shop?
As to my learned friend Dr. Lettsom (who is never out of the way when
there is good to be done) being moved to instigate you, a Methodist
parson, to enter into a medical controversy—that can only be accounted
for by supposing he owes you a grudge, and put you into my hands for
payment.
The fury had subsided in 1805, and Rowley held that many
medical men were deeply ashamed of the extravagance into which
they had been committed, but lacked courage to make frank
confession.
Rowley died in 1806, and the regard in which he was held was
manifest in the crowds who flocked to his funeral. In the Roll of
Physicians, Dr. Munk observes—
Dr. Rowley was a determined opponent of Vaccination, and obtained an
unenviable notoriety by his association with Dr. Moseley in opposing every
conceivable obstacle to the reception and progress of that invaluable
discovery.
Our website is not just a platform for buying books, but a bridge
connecting readers to the timeless values of culture and wisdom. With
an elegant, user-friendly interface and an intelligent search system,
we are committed to providing a quick and convenient shopping
experience. Additionally, our special promotions and home delivery
services ensure that you save time and fully enjoy the joy of reading.
ebooknice.com