E. Theory
E. Theory
A large part of geophysics concerns understanding how material deforms when it is squeezed, stretched,
or sheared. Elasticity theory is the mathematical framework which describes such deformation. By elastic,
we mean that the material rebounds to its original shape after the forces on it are removed; a rubber
eraser
is a good example of an elastic material. Play-dough, on the other hand, is a good example of a material
which is not elastic- if you squeeze it and then let go, the material stays in the shape it was when you
squeezed it. Materials of this type are described as being plastic, rather than elastic, and elasticity theory
is not suitable for understanding how these plastic materials deform. In reality, most materials are both
elastic and plastic, but on different time scales. Silly-putty is one such example- it stretches like hot
bubble gum, which is plastic behavior, but you can also roll it up and bounce it like a ball, which means
it’s also elastic. The key is the time duration of the external force- when you stretch it like gum, the force
is acting over several seconds, while when you bounce it off the floor, the impact on the floor lasts only
a fraction of a second. So silly putty is plastic for forces acting over time scales of several seconds, by
is elastic for very short –duration forces. Most of the materials that make up the earth are in fact elastic
over time scales of years, and plastic over time scales of thousands to millions of years.
Stress
Before we describe how a medium deforms under an external force, we need to be able to describe
precisely these external forces in terms of how they are oriented. Stress is defined as the force per
unit area. When a deforming force is applied to a body, the stress is the ratio of the force to the area over
which it is applied. If a force of 1 Newton(N) is applied uniformly to an area of 1 square meter, the stress
is 1 N/M2 = 1 (Kg x M/S2) / M2 = 1 Kg/(M S2) = 1 Pascal. Your intuition should tell you that it is stress,
rather than just force, which determines how a material deforms- think of an elephant inside a fancy
ballroom, in one case standing on its feet, and in another case balanced on the tip of a needle. Clearly,
the needle is going to puncture the floor much more effectively than the elephant’s feet. The reason
is because although the force is the same, ie , the mass of the elephant times gravity, the area over
which it is distributed is much smaller in the case of the needle, which, since stress is force/area, means
There are two basic types of stress. If the force is perpendicular (normal) to the surface over which
it is acting, then the stress is termed a normal stress; if it is tangential to the surface, it is called
a shear stress. Usually, the foce is neither enitrely normal nor tangential, but is at some arbitrary
intermediate
angle, in which case it can be resolved into components which are both normal and tangential to the
surface; so the stress is composed of both normal and shearing components. The sign convention is that
tensional stresses are positive and compressional stresses negative (in geophysics).
Directionality
Stress also has directionality to it- clearly a force acting in one direction will deform an object differently
than a force which is oriented purpendicular to it. So, we need to develop how we incorporate
directionality into stress. To do this, we use a hypothetical cube of material, it can be a free standing cube,
or one imbedded deep in a larger piece of material. Consider a small parallelipiped with sides dx,dy, and
Figure 1 The parallelipiped is in static equilbrium. Stresses acting on the shaded back face are (-sxx, -sxy, and
-sxz). Stress acting on the front face are (sxx, sxy, and sxz).
the stresses can be resolved into components in the x,y, and z directions. The stresses acting on the shaded
face are -sxx, -sxy, and -sxz. The notation is that sxy refers to the stress s acting in the y direction on the face
which is perpendicular to the x axis. The normal stress is thus -sxx and the shearing stresses are -sxy, and -sxz.
If the parallelipiped is to be in static equilibrium (not moving), then the stresses on opposite faces must
balance, and there must be no net couple which would rotate the parallelipiped. This requires that the
stresses on opposite faces be equal in magnitude and opposite in direction. The shearing stresses on
opposite faces of the parallelipiped (eg, -sxy and sxy on the back and front faces as shown in Figure 1)
provide a couple of forces, or couple, which will roate the parallelipiped. Since the parallelipiped must not
rotate, this couple must be balanced by the couple provided by the shearing stresses -syx, and syx
acting on the two side faces. This means that sxy must equal syx. . The same conditions apply to the other
shearing stresses: sxy = syx, sxz = szx, syz = szy (ie, the stress tensor must be symmetric).
Stress and Strain are both tensor quantities, which means that they are more than just simple scalars
or vectors. Everyday scalar numbers are just that- numbers, the dollars in my wallet, the pounds that I weigh, etc. However, there are other types of quantites
that require more information than just the amount-
for example, the velocity of my Subaru headed eastbound on I-90. While I might be driving 65 mph, that
doesn’t actually describe my motion- we need the direction also, since 60mph east is very different from 60mph west if I’m analyzing rear-ending a semi going
east bound 70mph. So velocity is a vector, meaning
it has both a value and a direction associated with it. Tensors are just an extension of this, it’s a number which has any number of directions associated with it. In
the case of stress, it is a 2-direction tensor, containing the direction of the force and the direction of the perpendicular to the face on which the force is acting.
Strain is like wise a tensor with two directions, or a rank 2 tensor. A physical quantity which had
Numerical Mathematics: Theory, Methods and Applications. Numerical Mathematics: Theory, Methods
and Applications (NM-TMA) publishes high-quality original research papers on the construction, analysis
and application of numerical methods for solving scientific and engineering problems.
Numerical analysis
From Wikipedia, the free encyclopedia
Jump to navigationJump to search
This article includes a list of references, but its sources remain unclear because it
has insufficient inline citations. Please help to improve this article
by introducing more precise citations. (November 2013) (Learn how and when to
remove this template message)
Babylonian clay tablet YBC 7289 (c. 1800–1600 BC) with annotations. The approximation of the square root
of 2 is four sexagesimal figures, which is about six decimal figures. 1 + 24/60 + 51/602 + 10/603=
1.41421296...[1]
Numerical analysis is the study of algorithms that use numerical approximation (as opposed to
general symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete
mathematics). Numerical analysis naturally finds application in all fields of engineering and the physical
sciences, but in the 21st century also the life sciences, social sciences, medicine, business and even the arts
have adopted elements of scientific computations. As an aspect of mathematics and computer science that
generates, analyzes, and implements algorithms, the growth in power and the revolution in computing has
raised the use of realistic mathematical models in science and engineering, and complex numerical analysis
is required to provide solutions to these more involved models of the world. Ordinary differential
equations appear in celestial mechanics (planets, stars and galaxies); numerical linear algebra is important
for data analysis; stochastic differential equations and Markov chains are essential in simulating living cells
for medicine and biology.
Before the advent of modern computers, numerical methods often depended on hand interpolation in
large printed tables. Since the mid 20th century, computers calculate the required functions instead. These
same interpolation formulas nevertheless continue to be used as part of the software algorithms for
solving differential equations.
One of the earliest mathematical writings is a Babylonian tablet from the Yale Babylonian Collection (YBC
7289), which gives a sexagesimal numerical approximation of the square root of 2, the length of
the diagonal in a unit square. Being able to compute the sides of a triangle (and hence, being able to
compute square roots) is extremely important, for instance, in astronomy, carpentry and construction.[2]
Numerical analysis continues this long tradition of practical mathematical calculations. Much like
the Babylonian approximation of the square root of 2, modern numerical analysis does not seek exact
answers, because exact answers are often impossible to obtain in practice. Instead, much of numerical
analysis is concerned with obtaining approximate solutions while maintaining reasonable bounds on
errors.
General introduction[edit]
The overall goal of the field of numerical analysis is the design and analysis of techniques to give
approximate but accurate solutions to hard problems, the variety of which is suggested by the following:
Advanced numerical methods are essential in making numerical weather prediction feasible.
Computing the trajectory of a spacecraft requires the accurate numerical solution of a system
of ordinary differential equations.
Car companies can improve the crash safety of their vehicles by using computer simulations of car
crashes. Such simulations essentially consist of solving partial differential equations numerically.
Hedge funds (private investment funds) use tools from all fields of numerical analysis to attempt to
calculate the value of stocks and derivatives more precisely than other market participants.
Airlines use sophisticated optimization algorithms to decide ticket prices, airplane and crew
assignments and fuel needs. Historically, such algorithms were developed within the overlapping field
of operations research.
Insurance companies use numerical programs for actuarial analysis.
The rest of this section outlines several important themes of numerical analysis.
History[edit]
The field of numerical analysis predates the invention of modern computers by many centuries. Linear
interpolation was already in use more than 2000 years ago. Many great mathematicians of the past were
preoccupied by numerical analysis, as is obvious from the names of important algorithms like Newton's
method, Lagrange interpolation polynomial, Gaussian elimination, or Euler's method.
To facilitate computations by hand, large books were produced with formulas and tables of data such as
interpolation points and function coefficients. Using these tables, often calculated out to 16 decimal places
or more for some functions, one could look up values to plug into the formulas given and achieve very
good numerical estimates of some functions. The canonical work in the field is the NIST publication edited
by Abramowitz and Stegun, a 1000-plus page book of a very large number of commonly used formulas and
functions and their values at many points. The function values are no longer very useful when a computer
is available, but the large listing of formulas can still be very handy.
The mechanical calculator was also developed as a tool for hand computation. These calculators evolved
into electronic computers in the 1940s, and it was then found that these computers were also useful for
administrative purposes. But the invention of the computer also influenced the field of numerical analysis,
since now longer and more complicated calculations could be done.
Direct and iterative methods[edit]
Direct vs iterative methods
Consider the problem of solving
3x3 + 4 = 28
for the unknown quantity x.
Direct method
3x3 + 4 = 28.
Subtract 4 3x3 = 24.
Divide by 3 x3 = 8.
Take cube roots x = 2.
For the iterative method, apply
3
the bisection method to f(x) = 3x − 24. The
initial values are a = 0, b = 3, f(a) = −24, f(b)
= 57.
Iterative method
a b mid f(mid)
0 3 1.5 −13.875
1.5 3 2.25 10.17...
1.5 2.25 1.875 −4.22...
1.875 2.25 2.0625 2.32...
We conclude from this table that the
solution is between 1.875 and 2.0625. The
algorithm might return any number in that
range with an error less than 0.2.
Discretization and numerical
integration[edit]
Direct methods compute the solution to a problem in a finite number of steps. These methods would give
the precise answer if they were performed in infinite precision arithmetic. Examples include Gaussian
elimination, the QR factorization method for solving systems of linear equations, and the simplex
method of linear programming. In practice, finite precision is used and the result is an approximation of
the true solution (assuming stability).
In contrast to direct methods, iterative methods are not expected to terminate in a finite number of steps.
Starting from an initial guess, iterative methods form successive approximations that converge to the exact
solution only in the limit. A convergence test, often involving the residual, is specified in order to decide
when a sufficiently accurate solution has (hopefully) been found. Even using infinite precision arithmetic
these methods would not reach the solution within a finite number of steps (in general). Examples
include Newton's method, the bisection method, and Jacobi iteration. In computational matrix algebra,
iterative methods are generally needed for large problems.
Iterative methods are more common than direct methods in numerical analysis. Some methods are direct
in principle but are usually used as though they were not, e.g. GMRES and the conjugate gradient method.
For these methods the number of steps needed to obtain the exact solution is so large that an
approximation is accepted in the same manner as for an iterative method.
Discretization[edit]
Furthermore, continuous problems must sometimes be replaced by a discrete problem whose solution is
known to approximate that of the continuous problem; this process is called discretization. For example,
the solution of a differential equation is a function. This function must be represented by a finite amount
of data, for instance by its value at a finite number of points at its domain, even though this domain is
a continuum.
of , after 10 or so iterations, we conclude that the root is roughly 1.99 (for example). We therefore
have a truncation error of 0.01.
Once an error is generated, it will generally propagate through the calculation. For instance, we have
already noted that the operation + on a calculator (or a computer) is inexact. It follows that a calculation of
Many algorithms solve this problem by starting with an initial approximation x0 to , for instance x0 =
1.4, and then computing improved guesses x1, x2, etc. One such method is the famous Babylonian method,
which is given by xk+1 = xk/2 + 1/xk. Another method, which we will call Method X, is given by xk+1 = (xk2 −
2)2 + xk.[3] We have calculated a few iterations of each scheme in table form below, with initial guesses x0 =
1.4 and x0 = 1.42.
... ...
Observe that the Babylonian method converges quickly regardless of the initial guess, whereas Method X
converges extremely slowly with initial guess x0 = 1.4 and diverges for initial guess x0 = 1.42. Hence, the
Babylonian method is numerically stable, while Method X is numerically unstable.
Numerical stability is affected by the number of the significant digits the machine keeps on, if we
use a machine that keeps only the four most significant decimal digits, a good example on loss of
significance is given by these two equivalent functions
and
by looking to the two results above, we realize that loss of significance (caused here
by catastrophic cancelation) has a huge effect on the results, even though both functions are
equivalent, as shown below
Areas of study[edit]
The field of numerical analysis includes many sub-disciplines. Some of the
major ones are:
Computing values of functions[edit]
Interpolation: We have observed
the temperature to vary from 20
degrees Celsius at 1:00 to 14
degrees at 3:00. A linear
interpolation of this data would
conclude that it was 17 degrees at
2:00 and 18.5 degrees at 1:30pm.
Extrapolation: If the gross domestic
product of a country has been
growing an average of 5% per year
and was 100 billion dollars last year,
we might extrapolate that it will be
105 billion dollars this year.
Software[edit]
Main articles: List of numerical analysis software and Comparison of numerical
analysis software
Since the late twentieth century, most algorithms are implemented in a variety
of programming languages. The Netlib repository contains various collections
of software routines for numerical problems, mostly in Fortran and C.
Commercial products implementing many different numerical algorithms
include the IMSL and NAG libraries; a free-software alternative is the GNU
Scientific Library.
There are several popular numerical computing applications such
as MATLAB, TK Solver, S-PLUS, and IDL as well as free and open source
alternatives such as FreeMat, Scilab, GNU Octave (similar to Matlab),
and IT++ (a C++ library). There are also programming languages such
as R (similar to S-PLUS) and Python with libraries such
as NumPy, SciPy and SymPy. Performance varies widely: while vector and
matrix operations are usually fast, scalar loops may vary in speed by more than
an order of magnitude.[5][6]
Many computer algebra systems such as Mathematica also benefit from the
availability of arbitrary precision arithmetic which can provide more accurate
results.
Also, any spreadsheet software can be used to solve simple problems relating
to numerical analysis.
An analytical solution involves framing the problem in a well-understood form and calculating the exact
solution.
A numerical solution means making guesses at the solution and testing whether the problem is solved well
enough to stop.
An example is the square root that can be solved both ways.
We prefer the analytical method in general because it is faster and because the solution is exact.
Nevertheless, sometimes we must resort to a numerical method due to limitations of time or hardware
capacity.
A good example is in finding the coefficients in a linear regression equation that can be calculated
analytically (e.g. using linear algebra), but can be solved numerically when we cannot fit all the data into
the memory of a single computer in order to perform the analytical calculation (e.g. via gradient descent).
Sometimes, the analytical solution is unknown and all we have to work with is the numerical approach.
Analytical Solutions
Many problems have well-defined solutions that are obvious once the problem has been defined.
For example, you know what operation to use given a specific arithmetic task such as addition or
subtraction.
In linear algebra, there are a suite of methods that you can use to factorize a matrix, depending on if the
properties of your matrix are square, rectangular, contain real or imaginary values, and so on.
We can stretch this more broadly to software engineering, where there are problems that turn up again
and again that can be solved with a pattern of design that is known to work well, regardless of the specifics
of your application. Such as the visitor pattern for performing an operation on each item in a list.
Some problems in applied machine learning are well defined and have an analytical solution.
For example, the method for transforming a categorical variable into a one hot encoding is simple,
repeatable and (practically) always the same methodology regardless of the number of integer values in
the set.
Unfortunately, most of the problems that we care about solving in machine learning do not have analytical
solutions.
Numerical Solutions
There are many problems that we are interested in that do not have exact solutions.
We have to make guesses at solutions and test them to see how good the solution is. This involves framing
the problem and using trial and error across a set of candidate solutions.
The numerical solutions to each sub-problem along the way influences the space of possible solutions for
subsequent sub-problems.
The core of a given machine learning model is an optimization problem, which is really a search for a set of
terms with unknown values needed to fill an equation. Each algorithm has a different “equation” and
“terms“, using this terminology loosely.
The equation is easy to calculate in order to make a prediction for a given set of terms, but we don’t know
the terms to use in order to get a “good” or even “best” set of predictions on a given set of data.
This is the numerical optimization problem that we always seek to solve.
It’s numerical, because we are trying to solve the optimization problem with noisy, incomplete, and error-
prone limited samples of observations from our domain. The model is trying hard to interpret the data and
create a map between the inputs and the outputs of these observations
Analytical is exact, numerical is approximate. For example, some differential equations cannot be solved
exactly (analytic or closed form solution) and we must rely on numerical techniques to solve them.Dec 16,
2015