Categorical and Nonparametric Data Analysis Choosing The Best Statistical Technique 1st Edition E Michael Nussbaum Download
Categorical and Nonparametric Data Analysis Choosing The Best Statistical Technique 1st Edition E Michael Nussbaum Download
https://ebookbell.com/product/categorical-and-nonparametric-data-
analysis-choosing-the-best-statistical-technique-1st-edition-e-
michael-nussbaum-34121254
https://ebookbell.com/product/categorical-and-nonparametric-data-
analysis-2nd-e-michael-nussbaum-56329242
Applied Categorical And Count Data Analysis Second Edition 2nd Edition
Wan Tang Hua He Xin M Tu
https://ebookbell.com/product/applied-categorical-and-count-data-
analysis-second-edition-2nd-edition-wan-tang-hua-he-xin-m-tu-48249068
Applied Categorical And Count Data Analysis 2nd Edition Wan Tang
https://ebookbell.com/product/applied-categorical-and-count-data-
analysis-2nd-edition-wan-tang-49420098
https://ebookbell.com/product/statistics-categorical-and-time-
dependent-data-analysis-mathematics-research-developments-
suzuki-32354852
Applied Categorical And Count Data Analysis 1st Edition Tang
https://ebookbell.com/product/applied-categorical-and-count-data-
analysis-1st-edition-tang-33727342
https://ebookbell.com/product/applied-categorical-and-count-data-
analysis-tang-wanhe-huatu-232343782
https://ebookbell.com/product/regression-models-for-categorical-and-
count-data-peter-martin-46541038
https://ebookbell.com/product/generalized-linear-models-for-
categorical-and-continuous-limited-dependent-variables-michael-
smithson-52688452
https://ebookbell.com/product/regression-models-for-categorical-and-
count-data-martin-peter-232949664
CATEGORICAL AND NONPARAMETRIC
DATA ANALYSIS
Featuring in-depth coverage of categorical and nonparametric statistics, this book pro-
vides a conceptual framework for choosing the most appropriate type of test in various
research scenarios. Class tested at the University of Nevada, the book’s clear explanations
of the underlying assumptions, computer simulations, and Exploring the Concept boxes
help reduce reader anxiety. Problems inspired by actual studies provide meaningful illus-
trations of the techniques. The underlying assumptions of each test and the factors that
impact validity and statistical power are reviewed so readers can explain their assump-
tions and how tests work in future publications. Numerous examples from psychology,
education, and other social sciences demonstrate varied applications of the material. Basic
statistics and probability are reviewed for those who need a refresher. Mathematical deri-
vations are placed in optional appendices for those interested in this detailed coverage.
Highlights include the following:
• Unique coverage of categorical and nonparametric statistics better prepares readers to se-
lect the best technique for their particular research project; however, some chapters
can be omitted entirely if preferred.
• Step-by-step examples of each test help readers see how the material is applied in a
variety of disciplines.
• Although the book can be used with any program, examples of how to use the tests in SPSS
and Excel foster conceptual understanding.
• Exploring the Concept boxes integrated throughout prompt students to review key
material and draw links between the concepts to deepen understanding.
• Problems in each chapter help readers test their understanding of the material.
• Emphasis on selecting tests that maximize power helps readers avoid “marginally”
significant results.
• Website (www.routledge.com/9781138787827) features datasets for the book’s exam-
ples and problems, and for the instructor, PowerPoint slides, sample syllabi, answers
to the even-numbered problems, and Excel data sets for lecture purposes.
E. Michael Nussbaum
First published 2015
by Routledge
711 Third Avenue, New York, NY 10017
and by Routledge
27 Church Road, Hove, East Sussex BN3 2FA
Preface xv
About the Author xix
CHAPTER 8 Advanced Rank Tests (for Interactions and Robust ANOVA) 215
Preface xv
About the Author xix
Problems 73
Technical Notes 78
Appendix 3.1: Linear Combination of Two Random Variables 79
Problems 276
Technical Notes 280
Appendix 9.1: Proofs of Selected OLS Theorems 282
When scientists perform research, they often collect data that do not come in a
form that is suitable for traditional methods, such as analysis of variance (ANOVA)
or ordinary linear regression. Categorical and nonparametric data analysis are use-
ful for data that are nominal or ordinal, or when assumptions of traditional metric
tests have been violated, for example, when the sample is small. Finally, even when
all the assumptions of traditional tests have been met, alternative tests are some-
times more statistically powerful, meaning that those tests are more likely to find a
statistically significant result when there is a real effect to be found. There is noth-
ing more frustrating than to spend a year or two on a research project, only to have
your results come in as not quite significant (e.g., p = .07).
This book was developed both out of my efforts to advise students on disserta-
tions as well as from my own research in the field of argumentation. I found that
while much of what I was taught about statistics in graduate school was useful at
times, it was also somewhat limited. It was as if I was only taught half the story.
In fact, one of my statistics teachers at Stanford University was upset that cover-
age of categorical data analysis (CDA) was no longer required, because there was
too much material already crammed into the statistics core. It was only later, after
several years conducting research with actual messy data sets, that I discovered the
immense value of categorical and nonparametric data analysis.
It is important to know when to use these techniques and when to use tradi-
tional parametric analysis. One major goal of the book is to provide a conceptual
framework for choosing the most appropriate type of test in a given situation. One
has to consider the underlying assumptions of each test and the factors that impact
each test’s statistical power. One also has to be able to explain these assumptions
(and how the test works) conceptually to both oneself and others (including the
audience of a journal article or dissertation). This allows one to make statistically
based arguments that can serve as a rationale for using that test. Therefore, another
major goal of this book is to provide readers with a conceptual framework that
underlies the statistical methods examined and, to some extent, traditional para-
metric methods as well.
xvi Preface
Intended Audience
The primary intended audience is researchers and students in the social sciences
(particularly psychology, sociology, and political science) and in related profes-
sional domains (such as education). Readers should have some prior knowledge of t
tests and ANOVA. It is preferable for readers to also have some prior knowledge of
linear regression, although bright students can also pick up the basics from reading
Chapter 9.
This book is fairly unique in covering both nonparametric statistics and CDA
in one volume. With the exception of contingency tables, most textbooks address
either one or the other. In the one-semester course that I teach on this topic, I suc-
cessfully cover both and—as noted above—provide a framework for choosing the
best statistical technique. Using the framework requires knowledge of both CDA
and nonparametrics. Although it is a challenging course, given students’ financial
and time constraints, it is often not practical for graduate students who are not spe-
cializing in statistics to take separate courses in nonparametric statistics and CDA.
Instructors who can provide a two-semester course can of course cover more topics
and provide students with more extensive practice; however, a one-semester course
can still provide students with some familiarity with the topics and a framework for
choosing a statistical methodology for particular research projects. (Students can
then get more practice and mastery of that particular technique, as the best way to
learn a technique is to use it in the context of applied research.) The book could also
be used for a course in only CDA or nonparametrics by just focusing on the appli-
cable chapters, leaving the rest of the material for optional, supplementary reading.
Finally, the book is also suitable for researchers and graduate students who are not
necessarily enrolled in a course but who desire some knowledge of these alternative
techniques and approaches for enhancing statistical power.
The book is closely tied to those techniques currently available in SPSS and provides
direction on using SPSS. (Incorporation into SPSS was in fact a major criterion for
choosing what topics would be covered.) Although the book can be used with other
statistical packages, SPSS is widely used and there is a need for a text that addresses
that audience. Some of the examples and problems in the book also use Excel as a
pedagogical tool for building conceptual understanding. Also available is a website
for the book at www.Routledge.com/9781138787827 that contains selected data
sets for the book’s examples and problems and, for the instructor, PowerPoint slides
for each chapter, answers to even-numbered problems, and the author’s course
syllabus.
Preface xvii
An important feature of the text is its conceptual focus. Simple computer simu-
lations and the inclusion of Exploring the Concept boxes are used to help attach
meaning to statistical formulas. Most homework problems were inspired by actual
research studies; these problems therefore provide authentic and meaningful illus-
trations of the techniques presented. Mathematical derivations have been kept to
a minimum in the main text but are available in appendices at the end of each
chapter.
Content
Acknowledgments
I am truly thankful to all those who assisted me in reviewing chapter drafts. Special
thanks to Lori Griswold for reviewing the entire book for clarity and typographical
mistakes. Sections of the book were also reviewed by my graduate assistants: Ivan
Ivanovich, Jason Boggs, and Marissa Owens. I would also like to thank my editor,
Debra Riegert, for her assistance, and Mark Beasley and Hossein Mansouri, both
of whom gave valuable input on Chapter 8. I would also like to acknowledge the
xviii Preface
following reviewers who were solicited by the publisher: Randall H. Rieger, West
Chester University; Margaret Ross, Auburn University; Sara Tomek, University of
Alabama; Haiyan Wang, Kansas State University; and three anonymous reviewers.
Finally, the book would not be possible without all the encouragement and feed-
back provided by the graduate students who completed my course in categorical/
nonparametric data analysis.
ABOUT THE AUTHOR
Categorical and nonparametric data analysis is designed for use with nominal or
ordinal data, or for metric data in some situations. This chapter therefore reviews
these different levels of measurement before turning to the topic of probability.
Levels of Measurement
In statistics, there are four basic types of variables: (a) nominal, (b) ordinal, (c) inter-
val, and (d) ratio.
A nominal variable relates to the presence or absence of some characteristic.
For example, an individual will be either male or female. Gender is a dichotomous
nominal variable. In contrast, with a multinomial nominal variable, cases are classi-
fied in one of several categories. For example, ethnicity is multinomial: Individuals
can be classified as Caucasian, African American, Asian/Pacific Islander, Latino, or
Other. There is not any particular ordering to these categories.
With an ordinal variable, there is an ordering. For example, an art teacher might
look at student drawings and rank them from the most to the least creative. These
rankings comprise an ordinal variable. Ordinal variables sometimes take the form
of ordered categories, for example, highly creative, somewhat creative, and uncreative.
A number of individuals may fall into these categories, so that all the drawings clas-
sified as highly creative would technically be tied with one another (the same for
the moderately creative and uncreative categories). With ranks, on the other hand,
there are typically few if any ties.
With an interval or ratio variable, a characteristic is measured on a scale with
equal intervals. A good example is height. The scale is provided by a ruler, which
may be marked off in inches. Each inch on the ruler represents the same distance;
as a result, the difference between 8 and 10 inches is the same as between 1 and
3 inches. This is not the case with an ordinal variable. If Drawing A is ranked as
2 Measurement, Probability, and Binomial Formula
more creative than Drawing B, we do not know if Drawing A is just slightly more
creative or significantly more creative; in fact, the distances are technically unde-
fined. As a result, we need to use different statistics and mathematical manipula-
tions for ordinal variables than we do for interval/ratio variables. (Much of this
book will be devoted to this topic.)
As for the difference between an interval variable and a ratio variable, the
defining difference is that in a ratio variable a score of zero indicates the complete
absence of something. Height is a ratio variable because zero height indicates that
an object has no height and is completely flat (existing in only two dimensions).
Counts of objects are also ratio variables. The number of people in a classroom can
range from zero on up, but there cannot be a negative number of people. With an
interval variable, on the other hand, there can be negative values. Temperature
is a good example of something measured by an interval scale, since 0o Celsius is
just the freezing point of water, and negative temperatures are possible. However,
for the tests discussed in this book, it will usually not be necessary to differentiate
between ratio and interval variables, so we will lump them together into one level
of measurement. The distinction will only become important when we consider the
analysis of count data with Poisson regression. For ease of exposition, in this book
I will use the term metric variable to refer to those at the interval or ratio levels of
measurement.
Metric variables are often also referred to as continuous, but this usage fails to
recognize that some metric variables are discrete. For example, a count cannot have
fractional values; for example, it would be incorrect to say that there are 30.5 people
enrolled in a class.
Figure 1.1 shows the three levels of measurement. The metric level is shown
on top because it is the most informative. Metric data can always be reduced to
ordinal data by using the numerical values to rank the data (e.g., ranking people
from the tallest to the shortest based on their heights). Likewise, ordinal data can
be reduced to nominal data by performing a median split and classifying cases as
above or below the median. Transforming data from a higher level to a lower level
is known as data reduction. Data reduction throws away information; for example,
knowing that Marie is taller than Jennifer does not tell one how much taller Marie
is. Nevertheless, data reduction is sometimes performed if the assumptions of a
statistical test designed for metric data are not met. One might then reduce the
data to ordinal and perform a statistical test that is designed for ordinal data. One
can also reduce ordinal (or metric) data to nominal. One cannot move from a
lower level to a higher level in the figure because that requires information that
is missing.
Categorical and nonparametric statistics is concerned with statistical methods
designed for ordinal- and nominal-level data. However, these methods are often
used with metric data when sample sizes are small (and therefore some of the
assumptions of t tests, ANOVA, and linear regression are not met) or when the data
Measurement, Probability, and Binomial Formula 3
Ratio/Interval (Metric)
Ordinal
Nominal
Figure 1.1 Three levels of measurement. The figure shows that metric data can be
reduced to ordinal data, which can in turn be reduced to nominal data. Metric data are
the most informative because they carry information on how different the cases are on
a variable in quantitative terms. Nominal data are the least informative because they
contain no information regarding order or ranks.
are skewed or otherwise highly abnormal. In the latter cases, standard methods may
not be as statistically powerful as categorical and nonparametric ones. In reading
this book, it is very important to remember the definition of statistical power:
Statistical power refers to the ability to reject the null hypothesis and find a
“result.” (To be more technically precise, it is the probability that one will reject
the null hypothesis when the alternative hypothesis is true.)
Because conducting a study requires a large amount of work, one usually wants
to use the most powerful statistical methods. (Obtaining a p value of .06 or .07 is
not sufficient to reject the null hypothesis and therefore can be very disappointing
to researchers.) That is why, in planning a study, one wants to use the most valid
and statistically powerful methods one can.
Probability
assuming that all possibilities are equally likely and mutually exclusive.
So a probability of ½ means that there is one favorable possibility (heads) out
of two (heads or tails). However, this example assumes that the coin is fair and not
biased, meaning that the possibilities are equally likely. A biased coin might have a
little more metal on one side so that heads results 60% of the time. Such coins have
been created to cheat at games of chance.
Eq. 1.1 is often useful but because of the restrictive equal-probability assump-
tion, a more general definition of probability is needed. The probability of some
event A occurring is the proportion of time that A will occur (as opposed to not-A)
in the limit as n approaches infinity, that is:
f ( A)
Pr ob ( A) = lim ( n ∞) , (Eq. 1.2)
n
where f(A) is the frequency of A. Thus, the meaning of the statement “The probabil-
ity that the coin will come up heads is one-half” is that over a large number of flips,
about half the time the coin will come up heads. The amount of error decreases as
n (the number of flips) increases, so that the proportion will approach Prob(A) as n
approaches infinity. Now to assess the probability, one might flip the coin 1,000 times
and gauge the relative proportion that the coin comes up heads as opposed to tails.
This procedure will only give one an estimate of the true probability, but it will give
one a pretty good idea as to whether the coin is fair or biased. Basically, what we are
doing is taking a random sample of all the possible flips that could occur.
The definition in Eq. 1.2 reflects a frequentist view of probability. Technically,
Eq. 1.2 assigns probabilities only to general statements, not to particular facts or
events. For example, the statement that 80% of Swedes are Lutherans is a meaning-
ful probability statement because it is a generalization; but the statement that “the
probability that John is Lutheran, given that he is Swedish, is 80%” is not mean-
ingful under this definition of probability, because probability applies to relative
frequencies, not to unique events.
This position, however, is extreme, given that we apply probability to unique
events all the time in ordinary discourse (maybe not about Lutherans, but cer-
tainly about the weather, or horse racing, etc.). In my view, probability statements
can be meaningfully applied to particular cases if one makes the appropriate back-
ground assumptions. For example, if one randomly selects one individual out of the
Measurement, Probability, and Binomial Formula 5
population (of Swedes), then one can meaningfully say that there is an 80% chance
that she or he is Lutheran. The background assumption here is that the selection
process is truly random (it may or may not be). The existence of the background
assumptions means that probability values cannot be truly objective because they
depend on whether one believes the background assumptions. Nevertheless, these
assumptions are often rational to make in many situations, so in this book I will
assign probability values to unique events. (For further discussion of subjectivist,
Bayesian notions of probability, which view probability statements as measures of
certainty in beliefs, see Nussbaum, 2011).
Probability Rules
Probability of Joint Events
What is the probability that if one flips two coins, both will come up heads? Using
Eq. 1.1, it is ¼, because there is one favorable possibility out of four, as shown below.
H H
H T
T H
T T
Another way of calculating the joint probability would be to use the following
formula:
Here, A represents the first coin coming up heads and B represents the second coin
coming up heads; the joint probability is ½ * ½ = ¼. The formula works because
A represents one-half of all possibilities and of these, one half represent favorable
possibilities, where B also comes up heads. One-half of one-half is, mathematically,
the same as ½ * ½.
An important background assumption, and one that we shall return to repeat-
edly in this book, is that A and B are statistically independent. What this means
is that A occurring in no way influences the probability that B will occur. This
assumption is typically a reasonable one, but we could imagine a scenario where it
is violated. For example, suppose someone designs a coin with an electrical trans-
mitter, so that if the first coin comes up heads, this information will be transmitted
to the second coin. There is a device in the second coin that will tilt it so that it
will always comes up heads if the first coin does. In other words: Prob(B | A) = 1.
This statement means that the probability of B occurring, if A occurs, is certain. We
6 Measurement, Probability, and Binomial Formula
will also assume the converse: Prob(not-B | not-A) = 1. There are therefore only two
total possibilities:
H H
T T
The joint probability of two heads is therefore one-half. The more general rule for
joint probabilities (regardless of whether or not two events are statistically inde-
pendent) is:
Note that in the rigged coin example, the joint probability has increased from
¼ (under statistical independence) to ½ (under complete dependence). This out-
come reflects a more general principle that the probability of more extreme events
increases if cases are not statistically independent (and positively correlated). For
example, if I throw 10 coins up in the air, the probability that they will all come
up heads is
½ * ½ * ½ * ½ * ½ * ½ * ½ * ½ * ½ * ½ = (½)10 = .001,
or 1 in 1,000. But if the coins are programmed to all come up heads if the first
coin comes up heads, then the joint probability is again just one-half. If the first
coin coming up heads only creates a general tendency for the other coins to come
up heads [say Prob(B | A) = 0.8], then the joint probability of two coins coming up
heads would be 0.5 * 0.8 = 0.4 (which is greater than ¼ th, assuming independence).
If 10 coins are flipped, the probability is 0.5 * (0.8)9 = .067. This probability is still
far greater than the one under statistical independence.
A violation of statistical independence is a serious problem, as it is a violation
of the first axiom of statistical theory. For example, my own area of research is on
how students construct and critique arguments during small-group discussions. In
small-group settings, students influence one another, so, for example, if one stu-
dent makes a counterargument, it becomes more likely that other students will do
so as well, due to modeling effects and other factors. The probability that Student
Measurement, Probability, and Binomial Formula 7
Applying Eq. 1.5 to our question, we find that the Prob(spades or clubs) =
Prob(spades) + Prob(clubs) = 14 + 14 = 21 . This result makes sense because half the
cards will be black. The reason Eq. 1.5 works is because Event A (card being a spade)
represents ¼ of the favorable possibilities, and Event B (card being a club) repre-
sents another ¼ of the favorable possibilities—so together, the probability that the
selected card will be black is the sum of all the favorable possibilities (which is why
we add). The background assumption is that the card cannot be both spades and
8 Measurement, Probability, and Binomial Formula
clubs, that is, A and B cannot both occur—the events are mutually exclusive. If the
events are not mutually exclusive, one should use Eq. 1.6:
For example, if one flips two coins, the probability that one or the other (or
both) will come up heads is 21 + 21 − 14 = 43 , which is the correct answer. If one were
to erroneously use Eq. 1.5 to calculate the probability, one would double count the
possibility where both coins come up heads.
Summary
In summary, remember that if joint events are involved (involving ands), one should
multiply (assuming statistical independence), and where alternative events are
involved (involving ors), one should add (assuming the events are mutually exclusive).
Conditional Probabilities
The last type of probability to be considered is conditional probability. These take an
IF . . . THEN form; for example, if a student is male, then what is the probability that
the student will complete high school? Condition probabilities will be discussed in
Chapter 4.
Each of the four outcomes is called a permutation. If one counts the number of
hits (i.e., heads), outcomes (2) and (3) both reflect one head. Although these two
outcomes reflect different permutations, they reflect the same combination, that is, a
Measurement, Probability, and Binomial Formula 9
0.6
0.5
0.4
Probablity
0.3
0.2
0.1
0
1 2 3
Hits
combination of one head and one tail. Combinations reflect the number of heads,
but whether the head comes up on the first trial or the second does not matter. The
order does matter with permutations.
Let us now make a histogram (frequency graph) of the different combinations
(see Figure 1.2).
This figure is shape like an inverted U, except that it is not smooth. Many prob-
ability distributions—like the normal curve—have this upward concave characteris-
tic. This is because there are more permutations corresponding to the combinations
in the middle than at the extremes, making the more central possibilities “more
likely.”
.000000000000000000000000000788861.
On the other hand, the probability of obtaining 50 heads out of 100 is much larger
(8%).
10 Measurement, Probability, and Binomial Formula
0.09
0.08
0.07
0.06
Probability
0.05
0.04
0.03
0.02
0.01
0
0
5
10
15
20
25
30
35
40
45
50
55
60
65
70
75
80
85
90
95
100
Hits
This is because there are a lot of ways of getting a combination of 50 coins out of
100 (i.e., there are a lot of corresponding permutations). The first 50 coins could come
up heads and the second 50 could come up tails; the first 50 coins could come up tails
and the second 50 could come up heads; every other coin could come up heads, and
so forth. There are also a lot more “random-looking” permutations, such as
HHHTTTHTTH...,
which is a random sequence that I generated in EXCEL. In fact, the number of permu-
tations comes out to be 1.00891 * 1029. How this number is derived will be explained
later in the chapter. Notice that in comparing Figures 1.2 and 1.3, the one with 100
coins is skinnier, if one thinks of the extremes in percentage terms (100 heads is 100%
hits, tails is 0% hits). This result occurs because the more extreme outcomes become
less likely as the number of trials increases. For example, the probability of 100% hits
in the two-coin case is 25%, not .5100 (as calculated above). Later we will see that this
is the reason why larger samples provide more precise estimates.
The binomial formula is used to measure the probability of a certain number of hits in
a series of yes/no trials. It is a complicated formula, so I will introduce it bit by bit. In
my example, I will use a biased rather than a fair coin to make the mathematics easier
to follow. The Greek letter pi (π) denotes the probability of a hit. (Notice that the
words pi and probability both start with the letter p. Statisticians often choose Greek
letter for concepts that, in English, start with the same letter.) Let us suppose that the
probability of this particular biased coin coming up heads is 70% (π = 70%), and ask:
What is the probability of tossing 10 coins (with π = .7) and obtaining 8 heads?
Measurement, Probability, and Binomial Formula 11
For the permutation considered previously, the probability is .78 * .32 = 0.5%.
Note that all the permutations associated with 8 coins out of 10 coming up heads
are equally likely. For example, the permutation “H H H T H H H H H T” has the
probability 70% * 70% * 70% * 30% * 70% * 70% * 70% * 70% * 30% = .78 * .32 =
0.5%. The only thing that has changed is the order in which the coins come up
heads or tails.
Remember that our overall goal is to find the probability of a combination of
8 hits out of 10. We shall see that this combination is associated with 45 different
permutations, including the two shown above. The probability of one or another
of these 45 different permutations occurring can be calculated by adding the indi-
vidual probabilities together. Because the permutations are mutually exclusive, we
can use the addition formula (Eq. 1.5). Because the permutations are equally likely,
the calculation reduces to
n n!
= . (Eq. 1.9)
k k !(n − k )!
12 Measurement, Probability, and Binomial Formula
Here, ( nk ) is the notation for k hits out of n trials. On the right-hand side of the equa-
tion, the exclamation mark signifies a factorial product (e.g., 4! = 4 * 3 * 2 * 1).
Substituting the values from our example yields
10 ! 3, 628, 000
= = 45.
8 !(10 − 8)! 40, 320 ∗ 2
Putting Eqs. 1.9 and 1.7 together yields the binomial formula, which gives the
probability of a certain number of k hits out of n trials:
n!
Prob (k hitsoutof n trials) = ∗ π k ∗ (1 − π )n−k . (Eq. 1.10)
k !( n − k ) !
In our example, the terms are 45 * 0.5% = 23%. The first term represents the
number of permutations associated with a combination of 8 heads out of 10 coin
tosses, and the second term reflects the probability of each permutation.
The calculation can also be performed in EXCEL with the following command:
= BINOMDIST(k, n, π, 0),
the expression to one. (A one in the last place of the EXCEL command tells EXCEL
to use the cumulative probabilities.) So we have
Now I am going to change the problem a little. The question is, “What is the
probability of obtaining 9 heads or more?” (again, out of a toss of 10 coins, with
π = 70%). This is calculated by subtracting the 85% from 100% (= 15%). Note that
the probabilities of obtaining (a) 8 coins or less and (b) 9 coins or more must sum
to 100%. This is because the two possibilities are exhaustive of all possibilities and
mutually exclusive; one or the other must happen.
Now, instead of asking, “What is the probability of obtaining 9 hits or more?”
we ask, “What is the probability of obtaining 8 hits or more?” This value is 100%
minus the probability of obtaining 7 hits or less. The EXCEL syntax is
Some students err by putting an 8 rather than a 7 in the formula, because the
question asks about the probability of getting 8 heads or more, yielding an incor-
rect answer of 15%. This approach is incorrect because it is tantamount to calculat-
ing the probability of 8 heads or more plus 8 heads or less. These are not mutually
exclusive possibilities, because if one obtains exactly 8 heads, both scenarios occur.
According to the formula for alternative probabilities, one has to subtract out the
probability of getting exactly 8 heads so as to not double count it (see Eq. 1.6). We
previously calculated the probability of obtaining eight heads or less at 85% and for
eight heads or more at 38%. The two possibilities sum to 123%, which is not a legal
probability. We have double counted the probability of obtaining exactly 8 heads
(which we calculated above at 23%).
This fact is so important that one should spend a minute now making a men-
tal rule to remember that when you are asked cumulative probability problems
that involve X or more (rather than X or less), you need to use X - 1 in the EXCEL
formula. EXCEL will only give you cumulative probabilities for X or less problems.
Heads (X = 1) 70%
Tails (X = 0) 30%
Sum 100%
0.8
0.7
0.6
0.5
Probability
0.4
0.3
0.2
0.1
0
0 1
Hits
(A)
0.3
0.25
0.2
Probability
0.15
0.1
0.05
0
0 1 2 3 4 5 6 7 8 9 10
Heads
(B)
0.3
0.25
0.2
Probability
0.15
0.1
0.05
0
0 1 2 3 4 5 6 7 8 9 10
Heads
Figure 1.5 Binomial distribution (n = 10) for different probabilities of a hit. (A) π =
50%. (B) π = 70%.
(A)
0.25
0.2
Probability
0.15
0.1
0.05
0
0 2 4 6 8 10 12 14 16 18 20
Heads
(B)
0.1
0.09
0.08
0.07
Probability
0.06
0.05
0.04
0.03
0.02
0.01
0
0
5
10
15
20
25
30
35
40
45
50
55
60
65
70
75
80
85
90
95
100
Heads
Figure 1.6 Binomial distribution for π = 70% for different values of n. (A) n = 20.
(B) n = 100.
= 40%]. The probability of 40% hits is becoming less likely with the larger sample,
making the distribution more symmetric.
X −mean
z = . (Eq. 1.11)
SD
X Probability
1 15%
2 33%
3 7%
4 12%
5 33%
Total 100%
0.35
0.3
0.25
Probability
0.2
0.15
0.1
0.05
0
1 2 3 4 5
X
X only takes on the discrete values of 1, 2, 3, 4, and 5. On the other hand, if X could
take on values such as 1.3, 2.67, or 4.3332, it would be continuous.
The probability distribution for the above discrete variable is displayed graphi-
cally in Figure 1.7. Ignoring the spaces between the bars, note that the area under
the curve sums to 1.0. The area for the value X = 2.0 is given by the width of the bar
(1 unit) times the height (which is 0.33), so the area is 0.33. The individual subareas
sum to 1 because the probabilities sum to 1.
If X were continuous, the area for a value (such as 2.67, a number I arbitrarily
chose) is given by the height (2.67) times the width, except that the width would
be 0. This produces a paradox that the probability of any specific value of a continu-
ous variable is 0. Obviously, this situation cannot occur, as the variable will take
on specific values in particular cases. To resolve this paradox, we stipulate that for
continuous variables, one can only meaningfully talk about the values of cumula-
tive probabilities, for example, the probability of obtaining a value of 2.67 or less (or
the probability of obtaining a value of 2.67 or more). The probability of obtaining
exactly 2.67 is undefined.
Measurement, Probability, and Binomial Formula 19
Table 1.3 Means and Variances of the Bernoulli and Binomial Distributions
Parameter Distribution
Bernoulli Binomial
Mean π nπ
Variance π(1 - π) nπ(1 - π)
that we might get 49 or 53 coins coming up heads rather than exactly 50, but the
number of heads is still likely to be around 50. The mean is also known as the
expected value.
In our example, the variance is nπ(1 - π). This is sometimes written as nPQ,
with P representing the probability of a hit and Q representing the probability of a
miss. In statistics, by convention, Greek letters refer to population parameters and
English letters to sample estimates of these parameters. Because we are not dealing
with samples here, I use the Greek letters in writing the formulas.
To continue with the example, the variance is nπ(1 - π) = 100 * 50% * 50% = 25.
The standard deviation is the square root of the variance, or 5. The z score is then
z = X −SDmean = 60 − 50 = 2.00,
5 Prob(z ≤ 2.00) = 97.7%. This result does not jive
with the calculation using the binomial formula (98.2%) because we have not yet
applied the continuity correction. When using z scores, it is assumed that the val-
ues are continuous. But we are here using discrete raw score values, for example,
62, 61, 60, 59, 58, and so forth. We want to find the probability of getting a value
of X ≤ 60. The area of the probability distribution corresponding to 60 is assumed
with the normal-curve method to have a width of zero (because continuous vari-
ables are assumed), when in fact the bar has a width of 1. The area of the bar to the
right of the midpoint corresponds to the probability excluded from the calculation
with the normal-curve method. This method leaves out “half a bar” associated
with the binomial formula method, which gives the exact value. The point of the
continuity correction when using the normal approximation is basically to “add
half a bar” to the probability estimate. We can do this if we use the value of X =
60.5 rather than 60 in the z-score calculation:
X − mean 60.5 − 50
z= = = 2.10, Prob( z ≤ 2.10 ) = 98.2%.
SD 5
This result is equal to the result from using the binomial method.
The normal approximation is an important procedure for the material discussed
in this book. It may seem unnecessary here to go through all the gyrations when we
can calculate the exact probability with the binomial formula, but in many other
situations we will not have recourse to the binomial formula or some other “exact”
formula and therefore will need to use the normal approximation.
Measurement, Probability, and Binomial Formula 21
X − mean 59.5 − 50
z= = = 1.90, Prob( z ≤ 1.90 ) = 97.1%.
SD 5
Problems
6. Three cards are drawn at random (with replacement) from a card deck with
52 cards. What is the probability that all three will be spades?
7. Three cards are drawn at random (without replacement) from a card deck with
52 cards. What is the probability that all three will be clubs? [Hint: When
two events are not statistically independent, and A occurs first, Prob(A & B) =
Prob(A) * Prob(B | A).]
4
8. Evaluate the following expression ( 1 ). [Hint: Remember that ( nk ) = n ! .]
k !( n−k !)
9. Evaluate the following expression: ( 34 ) .
10. You toss four coins. Using the binomial coefficient formula, compute how many
permutations are associated with a combination of two heads and two tails.
11. You toss four coins. Using the binomial formula, compute the probability of
obtaining exactly two heads if the coin is fair. [Hint: The probability is given
n k ( n−k )
by ( k ) ∗ π (1 − π ) .]
12. In a toss of four coins, use the binomial formula to compute the probability of
obtaining exactly two heads if the coin is biased and the probability of a head
is 75%.
13. Repeat the previous problem using EXCEL.
14. Using EXCEL, find the probability of obtaining two heads or less (again, with
n = 4 and π = 75%).
15. Using EXCEL, find the probability of obtaining two heads or more (again, with
n = 4 and π = 75%).
16. Chart the frequency distribution when n = 4 and π = 75%.
17. You construct a science achievement test consisting of 100 multiple-choice
questions. Each question has four alternatives, so the probability of obtaining a
correct answer, based on guessing alone, is 25%. If a student randomly guesses
on each item, what is the probability of obtaining a score of 30 or more correct?
a) Use the binomial function in EXCEL. (Hint: Find the probability of 29 or
less, and then subtract from 1.)
b) Use the normal approximation method, without a continuity correction.
You will need to calculate a z score, which means you will need to calculate
the mean expected correct (nπ) and standard deviation of this nπ(1- π ) .
(Hint: Find the probability of 29 or less, and then subtract from 1.)
c) Use the normal approximation with a continuity correction. (Hint: Find
the probability of 29.5 or less, then subtract from 1.)
18. If Y is a binomial random variable with parameters n = 60 and p = 0.5, esti-
mate the probability that Y will equal or exceed 45, using the normal approxi-
mation with a continuity correction.
19. Let X be the number of people who respond to an online survey of attitudes
toward social media. Assume the survey is sent to 500 people, and each has
a probability of 0.40 of responding. You would like at least 250 people to
respond to the survey. Estimate Prob(X ≥ 250), using the normal approxima-
tion with a continuity correction.
Measurement, Probability, and Binomial Formula 23
APPENDIX 1.1
HDT
DHT
HTD
DTH
THD
TDH
There are six permutations. Any one of the three people can fill the first seat.
Once the first seat is filled, there are two people left, so either one could fill the
second seat. Once the second seat is filled, there is just one person left, and so there
is just one possibility for filling the third seat. The total number of possibilities is
3 * 2 * 1 = 6. This is the factorial of 3 (3!). More generally, there are n! distinct ways
of arranging n objects.
Suppose now that Harry and Dick are identical twins. Some of the six permu-
tations will no longer appear distinct. If H = D and substituting H for D, the six
permutations become
HDT HHT
DHT HHT
HTD HTH
DTH HTH
THD THH
TDH THH
The number of distinct arrangements has been cut in half. We could just modify
the formula by dividing 6 by 2, or more specifically, 2!. In general, when k of the
objects (or people) are identical, the number of distinct permutations is given by nk!! .
We need to divide by k!, and not just k, because there are k! ways of arranging
the k identical objects, and each of these correspond to a possibility that is duplica-
tive of another possibility. There are therefore only three possible outcomes: HHT,
HTH, and THH. If instead of people, we let H represent a coin toss coming up heads
and T represent tails, we can see that there are three permutations corresponding
24 Measurement, Probability, and Binomial Formula
to a combination of two heads and one tail, and that ( 32 ) = 2 !( 33−! 2 )! = 3 . More gener-
ally, ( nk ) = k !( nn−! k )! , where (n - k!) is the number of misses. Although in the example,
(n - k)! = 1, in more complex examples (n - k)! may reflect duplicative possibilities,
so we need to divide by this term as well as k!.
Reference
Estimation
likely value of X in the sample. This number also corresponds to an expected value
of P of 12
20
=60%. This example shows why it is important to use random samples:
The laws of probability make it likely that the characteristics of the sample will be
representative (i.e., similar) to the population. In this case, if the sample contained
12 students who had been married, the P will be 60%, just like the population,
where π = 60%.
However, recall that this expected value is only the most likely value of our
sample estimate P; it is possible that we could draw a sample with 11 or 14 married
students. We should expect that we will obtain a sample of about 12 students who
have been married but that our estimate will be off a little. The “off a little” reflects
sampling error, the fact that even with a random sampling the sample might not
represent the population perfectly. Because of sampling error, we always represent
statistical estimates in the form of confidence intervals, for example .58 ± .02 (with
95% confidence).
It can be shown that the mean of the sampling distribution is π, the popu-
lation proportion. (This claim will be demonstrated in the next chapter and in
Appendix 2.1.)
The population distribution is a Bernoulli distribution (see Figure 1.4 in the pre-
vious chapter), and so the graph of the population distribution will have two bars:
one for X = 1 [with Prob(X = 1) = π] and one for X = 0 [with Prob(X = 0) = 1 - π].
Recall that the variance of a Bernoulli distribution is π(1 - π ). Because we are using
P to estimate π, we can use PQ to estimate the population variance, where Q is the
probability of a miss (Q = 1 - P).
However, our interest is in finding the standard deviation of the sampling dis-
tribution, not of the population distribution. (The former is known as the standard
error, or SE). The applicable formula is
PQ
Estimated SE of a Proportion = . (Eq. 2.1)
n
Appendix 2.1 presents the proof (see also Chapter 3).
Note from Eq. 2.1 that as n increases, the standard error decreases, meaning
that the sampling distribution becomes “thinner.” Eq. 2.1 implies that the standard
error approaches zero as n approaches infinity. In fact, the standard error would
approach zero as n approaches N¸ which is the size of the population. Technically,
therefore, we should include this additional constraint in the formula by writing
PQ PQ
Estimated SE of a Proportion = − . (Eq. 2.2)
n N
28 Estimation and Hypothesis Testing
When the population size is large, the second term is negligible, so it is typi-
cally left out of the formula for the standard error. One should use Eq. 2.2 when the
population size is small.
50%
45%
40%
35%
Probability
Figure 2.1 Bernoulli and binomial sampling distributions for varying sample sizes. For
n = 1, the distribution is Bernoulli (and equal to the population distribution). For n = 2
and n = 3, the distributions are binomial.
Estimation and Hypothesis Testing 29
1 1 100% π2 25%
1 0 50% π(1 - π) 25%
0 1 50% (1 - π) π 25%
0 0 0% (1 - π)2 25%
2 100% 25%
1 50% 50%
0 0% 25%
normality as the sample size approaches infinity (see Technical Note 2.1). So if the
sample size is large enough, the sampling distribution will be approximately nor-
mal. When we assume normality because the sampling distribution is so close to
being normal, this is known as using a normal approximation. How large the sample
needs to be for a normal approximation to be reasonable is a complex question that
will be addressed later in this chapter.
For now, I wish to build the reader’s conceptual intuitions about the CLT with
another thought experiment. Suppose I take a sample of students at my university
(N = 30,000), but my interest is now in a metric rather than a categorical variable,
specifically, estimating the average height of males at the university. If we were to
take a sample of n = 1 and drew an individual with a height of 6 feet, then for that
sample, X =6′ . If we drew someone of 5 feet, then X =5′ . The distribution of pos-
sible X ′s would be identical to the population distribution. Because it is well known
that the distribution of heights by gender in a large population is normal, it would
follow that the sampling distribution is normal. This leads to the following impor-
tant result, which holds even for larger samples (n > 1): If the population distribution
is normal, then the sampling distribution will be exactly normal.
However, many if not most population distributions are not normal. Many are
skewed—for example, the distribution of household income or the number of high
school students in different schools who become professional actors. In addition,
we usually have little information about the shape of the population distribution;
although the sample can provide some information, we will nevertheless still be
uncertain about the shape of the sampling distribution.
Fortunately, when we are uncertain about the shape of the population distribu-
tion, or when we know it is not normal, we can invoke the CLT if the sample is large
enough. Suppose we want to estimate the average annual income of students at
the university and (going to the other extreme) we sampled the entire population,
so n = 30,000. There is only one possible sample we could draw, so the sampling
distribution will have no variance, as indicated by Eq. 2.1. Graphically, it would
be a spike, as shown in Figure 2.3 (light gray line in the middle). This spike is the
central limit.
Now let us take the next step and give the sampling distribution a little variance.
Suppose we sample all individuals in the population but one, so that n = 29,999.
Because there are 30,000 people in the population, there are 30,000 different
Discovering Diverse Content Through
Random Scribd Documents
The Project Gutenberg eBook of Gardens of the
Caribbees, v. 2/2
This ebook is for the use of anyone anywhere in the United States
and most other parts of the world at no cost and with almost no
restrictions whatsoever. You may copy it, give it away or re-use it
under the terms of the Project Gutenberg License included with this
ebook or online at www.gutenberg.org. If you are not located in the
United States, you will have to check the laws of the country where
you are located before using this eBook.
Language: English
Volume II.
GARDENS OF
THE CARIBBEES
Sketches of a C r u i s e to the W e s t
Indies and the Spanish Main
By
Ida M. H. Starr
IN TWO VOLUMES
VOL. II.
ILLUSTRATED
Boston
L. C. Page & Company
MDCCCCIV
Copyright, 1903
By L. C. Page & Company
(INCORPORATED)
———
All rights reserved
Colonial Press
Electrotyped and Printed by C. H. Simonds & Co
Boston Mass., U. S. A.
CONTENTS
CHAPTER PAGE
I. Island of Trinidad. “Iere” 11
II. Island of Trinidad. La Brea 35
III. The Spanish Main 64
IV. In Venezuela. Caracas 101
V. In Venezuela. Caracas to Puerto Cabello 125
VI. Curaçao. City of Willemstad 153
VII. The Southern Cross 189
VIII. Kingston, Jamaica 198
IX. “Cuando Salide La Habana” 239
X. A Memory of Martinique 247
INDEX
LIST OF ILLUSTRATIONS
Volume II.
PAGE
From Our Balcony, Caracas, Venezuela Frontispiece
The Barracks, through Live-oaks and
Mahogany-Trees, Trinidad 17
Governor’s Palace and Public Gardens,
Port of Spain, Trinidad 21
On the Way to the Savannah, Port of
Spain, Trinidad 31
The Beach of la Brea, Trinidad 39
Asphalt for Northern Pavements, Pitch
Lake, Trinidad 47
Loading Cars, Pitch Lake, Trinidad 53
A Native Washerwoman on the Pitch
Lake, Trinidad 57
Where the Mountains Meet the Sea, La
Guayra, Venezuela 65
Caracas and the Mountains, Venezuela 75
Equestrian Statue of Bolivar, the
Liberator, Caracas, Venezuela 85
An Interior Court, Caracas, Venezuela 91
Cathedral and Plaza, Caracas,
Venezuela 111
A House Beside the Sea, Puerto
Cabello, Venezuela 127
A South American Street, Puerto
Cabello, Venezuela 149
Across Ste. Anne Bay, Harbour of
Willemstad, Curaçao 157
Some of Our Friends at Willemstad.—
Where the Basket-Women Waited,
Willemstad, Curaçao 161
The Landing, Willemstad, Curaçao 165
A Jolly Dutch Port, Willemstad,
Curaçao 173
A Snug Harbour, Willemstad, Curaçao 185
Kingston, Jamaica, from the Bay 199
Rio Cobre, near Spanish Town, Jamaica 203
A Native Hut, Jamaica 209
The Bog Walk Road, near Spanish
Town, Jamaica 213
Where We Landed, Kingston, Jamaica 223
El Morro, Entrance to Harbour,
Santiago de Cuba 229
The Plaza, Cienfuegos, Cuba 233
The Grave of Cervera’s Fleet, West of
Santiago de Cuba 237
The Wreck of the Maine, Havana
Harbour, Cuba 241
Cabañas, La Punta, and Harbour
Entrance, Havana, Cuba 245
St. Pierre and Mt. Pelée before the
Eruption, Martinique 249
St. Pierre and Mt. Pelée after the
Eruption, Martinique 253
Rue Victor Hugo Before the Eruption,
St. Pierre, Martinique 257
Rue Victor Hugo after the Eruption,
St. Pierre, Martinique 261
Gardens of the Caribbees
CHAPTER I.
I.
The garden once entered, we pass a great, squarely built mansion, the
governor’s residence, and are in the midst of a wonderful vegetation from
the first step. At the very entrance, we are greeted with, perhaps, the most
unique tree in these latitudes.
After all, there is something stupefying in the effort to describe tropical
wonders. When they are passing before one’s eyes, each has a feature
distinct to itself, which, in a way, is its own manner of description. Each has
its peculiar wonder, its own glory,—no two alike—and yet, when one sits
down to think it over, there is the same old alphabet from which to draw
new pictures, new miracles; and how to make each different with the same
letters is a question indeed.
If I could only tell you the name of this particular tree which stands at
the entrance to the garden, you might some day hunt it up yourself, but as I
know neither its family nor home, we will let that all go, and just tell you
how it is dressed.
It is a heavily, glossily leafed, symmetrical, low tree, just about the size
of those dear old cherry-trees we used to climb, oh, so long,—so long ago!
From the tip of every branch there drops a cord-like fibre about a foot and a
half long, and at the end of this little brown string there hangs a cluster of
delicate pink flowers. These are suspended in almost exact length in rows
from the lowest to the highest branch, and it really seems as if Nature were
experimenting to see what wonderful living garlands she could create for a
canopy above our heads.
III.
The character of the garden is defined at once upon entrance. It is a
botanical garden, pure and simple, a place for strange plants from far away,
a sort of orphan asylum for everybody’s vegetable baby. It is not, like
Martinique, an enchanted forest with cascades and glens fit for nymphs and
dryads; it is matter-of-fact, orderly, prim, and businesslike. Aside from its
unique trees, there is little to attract one, so we decide for once it would be
wise to engage a guide who can tell us something about the inhabitants of
the place, which otherwise promises to be rather dull.
IV.
Quite as one strokes the head of a favourite animal, Manuel leads us to
an insignificant-looking tree, takes a branch caressingly in his hand, brings
out his clumsy knife, selects just the right spot, cuts off a bit, and hands us a
piece of camphor wood.
Into the dear St. Thomas basket it goes, with the leaves of coffee, the
pink and white clove blossoms, and a long spray of araucaria from the
Norfolk Islands,—a strange company, indeed!
Yonder long yellow avenues are cinnamon and spice groves with
reddish-yellow bark, smooth as wax, casting slender shadows in the golden
light. Here is the shaddock, entirely weaned from its Malayan home, and
farther on a clump of low bushes, in among the nutmeg trees and coffee,
with small satin-like leaf, brings us to the herb that “cheers but does not
inebriate,”—the tea.
Just see those glorious great lemons, glowing in the ever-splendid
sunlight, which transmits to every living object a radiance, a dazzling
brilliancy, in which life progresses and finally dissolves out of sheer
exhaustion from the exuberance of vitality.
Oh, to our starved eyes of the North; to our senses benumbed by dreary
days of darkened sky, hearts chilled by bitterness of wind and gray,
unyielding frost, this never-ending, unspeakable sunlight, filtering through
the yellow vistas of clove and cinnamon, comes like the actual presence of
Apollo, the Shining One! We may, in unguarded moments, in ungrateful
moments, maybe, consider his embrace too positive, and we may raise the
white umbrella, but we never quite lose our rejuvenated love for his golden
glory.
Manuel, but half-clad, looks as if he would dismember at any moment.
His trousers are hitched by a couple of old leathers, and his shirt looks as if
it wished it “didn’t have to,” and his old hat is only there on sufferance, and
his shoes—old flippety-flops—have dragged their ill-shaped existence
through many a weary mile. But Manuel doesn’t care; he loves his garden,
and the sunshine and the luscious fruit, all his children so well behaved and
so obedient to his voice. He takes a bamboo pole and gives one of the big,
juicy lemons a rap, and down it falls on Wee One’s head with such a thump!
Then Manuel is very sorry, and he apologises for his child’s misdemeanour
in his funny, mixed-up Portuguese-English-Spanish and the rest, and we
understand and don’t mind a bit; in fact, we wouldn’t care if more would
fall in the same way.
Once upon a time, in the far-off golden days, when the Divine in
Creation had not been quite forgot, there came to this shore a band of men,
—not faultless, no, not faultless—but great men “for a’ that,” who, with
glittering cross aloft, christened this fair land after the blessed Trinity. But
this was not her first sacrament. Deep in the eternal silence of the forest, the
dweller in the High Wood had sought expression of the divine through
beauty, and chose a name from out the radiant wilderness which would tell
for ever of its wonderment: “Iere,” the land of the humming-bird, they
called her—those dusky children of the High Wood—and to this day she
clings lovingly to her maiden name.
We look about us. Where are the birds once peopling these forests, like
myriads of rainbows? Oh, sisters! members of Humane Societies! Hunt up
your old bonnets and see the poor little stuffed carcasses ornamenting your
cast-off finery! So Trinidad has been bereft of her wonderful birds, and now
there is but a name, a sad-sounding, meaningless name—Iere—to tell of
days which knew not the pride and cruelty of women.
Think of it!—at one time, there averaged twenty thousand humming-
birds a year exported from Trinidad to England alone!
And now, well—there are none left to export. We must find new islands
to denude, to ravage, to desolate, for our adornment. But it’s too unpleasant,
—this seeing things as they are; we’ll hide the poor little innocent card
which the black woman gave us at the hotel; we’ll cover up the word “Iere”
with these coffee leaves. There, now the spray of araucaria, now the stick
of camphor, and I think the lemon will fit right in among the nutmegs.
Come along, Manuel, we are ready; and we follow through the birdless
paths, down where the Nux Vomica grows, and the pepper, and the lime and
the calabash, and the orange and breadfruit, and tamarind, and pineapple;
and we go on and wake up the comical lizards who scurry away like brown
flashes of whip-cord. What ridiculous creatures they are, and how
desperately frightened! Why, surely they must be fifteen inches long, and
fully four inches high, and what funny, nimble legs! They start off in the
same spasm-like way as do the toy lizards we buy for the youngsters.
Manuel brings us to the plant house where the great forest wonders of
the Far East are babied and loved into strength, and I could not but think of
Daudet’s dear old Tartarin of Tarascon, dreaming by the homesick little
baobab-tree, which grew in his window-garden; and of the long nights
under the mellow moon of sunny France; and how he fought great beasts
and achieved great fame in the land of sweet illusion.
Dream on, Tartarin, wherever you may be! The time will come when it
will all be true, and you, too, will rest under the yellow splendour of the
golden trees; and the earth, the great Mother Earth, will open her heart to
you and breathe upon you the spirit of limitless possibilities!
Good-bye, Manuel! The basket is heavy to carry with its spoils of fruit
and flowers; and we take “turn about” across the savannah.
The races are on, and horses are dashing around the grassy turf, and the
Trinidadians are yelling, the cricket games are going, and the picnic parties
are gathering up their baskets for home; and the Hindoo girls clamour to
carry our basket, and we gladly give over the load to a tough little head; and
the merry-go-round wheezes out its squeaking tunes, and we pass through
the black crowd, and narrowly escape taking a cab, for the way to the quay
looks long, and we waver and weaken, and are just about to give in, when
up comes a tinkling tram, and we jump in, with a penny to the Hindoo girl,
and rumble away.
The man with the two monkeys, and the man with the green and blue
parrot, and the boy with the shells, are still waiting.
Alackaday! Where is the woman with the baskets?
CHAPTER II.
I.
ebookbell.com