Error Analysis
Error Analysis
• Introduction
• Average error, r.m.s error, probable error and error
propagation
• Significant digit and figures
• Uncertainties in Measurements: Measuring Errors,
Uncertainties, Parent and Sample Distributions, Mean and
Standard Deviation of Distributions,
• Numericals
What is an error?
• Error is a measure of the lack of certainty in a value.
• Error means the inevitable uncertainty that attends all
measurements.
HOW ERRORS ARISE IN MEASUREMENT (SYSTEMS)?
A measurement under ideal conditions has no errors.
Failure to account for a factor (usually systematic) - The most challenging part of
designing an experiment is trying to control or account for all possible factors except
the one independent variable that is being analyzed. For instance, you may
inadvertently ignore air resistance when measuring free-fall acceleration, or you may
fail to account for the effect of the Earth's magnetic field when measuring the field of a
small magnet. The best way to account for these sources of error is to brainstorm with
your peers about all the factors that could possibly affect your result. This brainstorm
should be done before beginning the experiment so that arrangements can be made to
account for the confounding factors before taking data. Sometimes a correction can be
applied to a result after taking data, but this is inefficient and not always possible.
Parallax (systematic or random) - This error can occur whenever there is some distance
between the measuring scale and the indicator used to obtain a measurement. If the
observer's eye is not squarely aligned with the pointer and scale, the reading may be too
high or low (some analog meters have mirrors to help with this alignment).
Instrument drift (systematic) - Most electronic instruments have readings that drift over
time. The amount of drift is generally not a concern, but occasionally this source of error
can be significant and should be considered.
Lag time and hysteresis (systematic) - Some measuring devices require time to reach
equilibrium, and taking a measurement before the instrument is stable will result in a
measurement that is generally too low. The most common example is taking temperature
readings with a thermometer that has not reached thermal equilibrium with its
environment. A similar effect is hysteresis where the instrument readings lag behind and
appear to have a "memory" effect as data are taken sequentially moving up or down
through a range of values. Hysteresis is most commonly associated with materials that
become magnetized when a changing magnetic field is applied.
Estimation of Error
• When attempting to estimate the error of a measurement, it is often
important to determine whether the sources of error are systematic or
random. A single measurement may have multiple error sources, and these
may be mixed systematic and random errors.
Recognize that a discrepancy may or may not be significant. The two measurements
just discussed are illustrated in Figure (a), which shows clearly that the discrepancy
of 10 ohms is significant because no single value of the resistance is compatible with
both measurements. Obviously, at least one measurement is incorrect, and some
careful checking is needed to find out what went wrong.
Figure (a): Two measurements of the
same resistance. Each measurement
includes a best estimate, shown by a black
dot, and a range of probable values, shown
by a vertical error bar. The discrepancy
(difference between the two best
estimates) is 10 ohms and is significant
because it is much larger than the
combined uncertainty in the two
measurements. Almost certainly, at least
one of the experimenters made a mistake.
Suppose, on the other hand, two other students had reported these results:
Error bars only need to be used when the uncertainty in one or both of the plotted
quantities are significant. Error bars are not required for trigonometric and
logarithmic functions.
To add error bars to a point on a graph, we simply take the uncertainty range
(expressed as "± value" in the data) and draw lines of a corresponding size above
and below or on each side of the point depending on the axis the value
corresponds to.
An error bar tells you how closely your measured result should be matched by true
result.
An error bar estimates how confident you are in your own measurement or result.
When plotting points (x, y) with known uncertainties on a graph, we plot the
average, or mean, value of each point and indicate its uncertainty by means of
“error bars.” Figure 1 shows how the upper error bar at ymax and the lower error
bar at ymin are plotted. If the quantity x also has significant uncertainty, one adds
horizontal error bars (a vertical error bar rotated 90°) with the rightmost error bar
at position xmax and the leftmost error bar at position xmin.
Figure 1 Diagram of error bars showing uncertainties in the value of the x- and y-
coordinates at point (xavg , yavg).
Significant Figures (digits):
Significant figures are the digits in a value that are known with some degree of
confidence. As the number of significant figures increases, the more certain the
measurement. As precision of a measurement increases, so does the number of
significant figures.
1 All non zero digits are significant. 549 has three significant figures
1.892 has four significant figures
2 Zeros between non zero digits are 4023 has four significant figures
significant. 50014 has five significant figures
3 Zeros to the left of the first non zero 0.000034 has only two significant figures.
digit are not significant. (This is more easily seen if it is written as
3.4x10-5) 0.001111 has four significant
figures.
4 Trailing zeros (the right most zeros) 400. has three significant figures
are significant when there is a 2.00 has three significant figures
decimal point in the number. For 0.050 has two significant figures
this reason it is important to give
consideration to when a decimal
point is used and to keep the
trailing zeros to indicate the actual
number of significant figures.
5 Trailing zeros are not significant in 470,000 has two significant figures
numbers without decimal points. 400 or 4x102 indicates only one significant
figure. (To indicate that the trailing zeros
are significant a decimal point must be
added. 400. has three significant digits and
is written as 4.00x102 in scientific
notation.)
6 Exact numbers have an infinite If you count 2 pencils, then the number of
number of significant digits but pencils is 2.000...
they are generally not reported. The number of centimeters per inch (2.54)
Defined numbers also have an has an infinite number of significant digits,
infinite number of significant digits. as does the speed of light (299792458
m/s).
Fractional Uncertainties
The uncertainty in a measurement,
However, the uncertainty by itself does not tell the whole story. An uncertainty
of one inch in a distance of one mile would indicate an unusually precise
measurement, whereas an uncertainty of one inch in a distance of three inches
would indicate a rather crude estimate.
Obviously, the quality of a measurement is indicated not just by the uncertainty
but also by the ratio of dx , which leads us to consider the fractional
uncertainty,
fractional uncertainty =
|
|
(The fractional uncertainty is also called the relative uncertainty.) In this definition,
the symbol | | denotes the absolute value .
Random Error and Systematic Error
All experimental uncertainty is due to either random errors or systematic
errors.
Random errors are statistical fluctuations (in either direction) in the measured
data due to the precision limitations of the measurement device.
Random errors usually result from the experimenter's inability to take the
same measurement in exactly the same way to get exact the same number.
Or
Random errors in experimental measurements are caused by unknown and
unpredictable changes in the experiment.
Or
Random errors cause a measurement to be as often larger than the true value,
as it is
smaller.
Systematic errors are often due to a problem which persists throughout the
entire experiment.
The main differences between these two error types are:
Random errors are (like the name suggests) completely random. They are
unpredictable and can’t be replicated by repeating the experiment again.
Systematic Errors produce consistent errors, either a fixed amount (like 1 lb) or a
proportion (like 105% of the true value). If you repeat the experiment, you’ll get
the same error.
Preventing Errors
Once again, the best estimate for x is usually the average of , , … … , . That
is,
= ̅ ,
where
+ + ⋯ + ∑
̅ = =
In the last line, I have introduced the useful sigma notation, according to which
= = + + ⋯ + ;
Standard deviation (SD)
The standard deviation of the measurements is an estimate of the average
uncertainty of the individual measurements.
The mean ̅ is our best estimate of the quantity x, it is natural to consider the
difference − ̅ = " .
This difference, often called the deviation (or residual) of from x, tells us how
much the ith measurement differs from the average ̅ . If the deviations
" = − ̅ are all very small, our measurements are all close together and
presumably very precise. If some of the deviations are large, our measurements are
obviously not so precise.
Notice that the deviations are not (of course) all the same size; di is small if the ith
measurement xi happens to be close to ̅ , but di is large if xi is far from ̅ . Notice
also that some of the di are positive and some negative because some of the xi are
bound to be higher than the average ̅ , and some are bound to be lower.
The average of the deviations is not a useful way to characterize the reliability of
the measurements.
The best way to avoid this annoyance is to square all the deviations, which will
create a set of positive numbers, and then average these numbers.
If we then take the square root of the result, we obtain a quantity with the same
units as x itself. This number is called the standard deviation .
The Standard Deviation of the Mean (SDOM) / Standard Error
The uncertainty in the final answer xbest = ̅ is given by the standard deviation
divided by . This quantity is called the standard deviation of the mean (SDOM),
or standard error, and is denoted ̅ :
̅ = .
Thus, based on the N measured values , , … … , , we can state our final
answer for the value of x as
$%&'( )* = xbest ± ,
When you multiply two numbers, you more or less multiply the uncertainties. Thus it
is the percentage by which you are uncertain that is important -- the uncertainty in
the number divided by the number itself. This is given roughly by the number of
digits, regardless of their placement in terms of powers of ten. Hence the number of
digits is what is important.