Basic Engineering For Medics and Biologists PDF
Basic Engineering For Medics and Biologists PDF
and Biologists
An ESEM Primer
Edited by
T. Clive Lee
Royal College of Surgeons in Ireland
and Trinity Centre for Bioengineering, Dublin, Ireland
and
Peter F. Niederer
Institute for Biomedical Engineering
Swiss Federal Institute of Technology (ETH)
and University of Zurich, Zurich, Switzerland
All rights reserved. No part of this book may be reproduced, stored in a retrieval system,
or transmitted, in any form or by any means, without prior written permission from the publisher.
Publisher
IOS Press BV
Nieuwe Hemweg 6B
1013 BG Amsterdam
Netherlands
fax: +31 20 687 0019
e-mail: [email protected]
LEGAL NOTICE
The publisher is not responsible for the use which might be made of the following information.
Preface
Making Engineering Accessible to
Medics and Biologists
Research and development in bioengineering and medical technology, conducted dur-
ing recent decades, have led to spectacular progress in clinical medicine. These
achievements have triggered an enormous increase in the number of courses offered in
the areas of bioengineering, clinical technology and medical informatics; nowadays,
most major universities offer curricula oriented towards these fields. The majority of
participants however come from engineering backgrounds and so modules dealing with
basic biological and medical sciences have been included. These have been facilitated
by the ready availability of textbooks in this area, many of which were specifically
written for nursing A & P (Anatomy & Physiology) programmes.
In contrast, relatively few participants from medicine and biology have taken up
courses in biomedical engineering, to the detriment of scientific exchanges between
engineers and medical doctors. The reasons for this imbalance are many and may vary
from country to country, but a consistent finding is the difficulty (or courage) in taking
the first step. ‘Introductory’ texts in bioengineering tend to involve vector algebra and
calculus early in Chapter 1. While in most countries entry to medical school is very
competitive and requires, among others, high grades in mathematics, little more than
arithmetic is required during the course itself, so numeracy undergoes disuse atrophy.
Furthermore, to paraphrase George Bernard Shaw, medics and engineers are sepa-
rated by a common language. To the medic, stress and strain are both symptoms of
anxiety, while to the engineer they are defined by equations involving symbols such as
σ and ε and are related to forces and deformations. To the average medic this is, liter-
ally, all Greek.
ESEM is a bridge between medicine and engineering. It promotes cultural and sci-
entific exchanges between engineers and medics and training courses in biomedical
engineering. What better way to achieve these objectives than to address this problem
and help medics and biologists to take this first step.
To this end, we herewith present a series of First Step chapters in book form aimed
at medics and biologists to help bring them to the level where they can begin an MSc in
biomedical engineering, or at medics who “simply” wish to understand a particular
medical technology. Written by engineers who are leaders in their field, with input
from medical colleagues, they will cover the basic engineering principles underpinning
biomechanics, bioelectronics, medical informatics, biomaterials, tissue engineering,
bioimaging and rehabilitation engineering, and will include clinically relevant exam-
ples.
As Albert Einstein may have said ‘Everything should be made as simple as possi-
ble, but not simpler’.
Dublin & Zurich, December 2009 Clive Lee & Peter Niederer
Cover Illustration
Stress trajectories in a curved Culmann crane (left) compared with a schematic repre-
sentation of the trabecular pattern in the proximal human femur (right). From: Wolff J.
Ueber die innere Architectur der Knochen und ihre Bedeutung für die Frage vom
Knochenwachsthum (On the intrinsic architecture of bone and its significance with
respect to the question of bone growth). Virchow’s Arch. 1870; 50: 389–450.
vii
Contents
Preface: Making Engineering Accessible to Medics and Biologists v
Clive Lee and Peter Niederer
I.1. Statics 3
Gijsbertus J. Verkerke and T. Clive Lee
I.2. Mechanics of Materials 13
Prashant K. Sharma
I.3. Dynamics of Human Movement 27
Bart H.F.J.M. Koopman
I.4. Biofluid Mechanics & the Circulatory System 45
Pascal Verdonck and Kris Dumont
I.5. Biomechanics of Implants 58
Jan G. Hazenberg, Johannes Schmid, T. Clive Lee and
Gijsbertus J. Verkerke
Introduction to
Chapter I: Biomechanics
Gijsbertus J. VERKERKE, Pascal VERDONCK and T. Clive LEE (eds.)
Biomechanics is the science that examines forces acting upon and within a biological
structure and effects produced by such forces (B.M. Nigg, Biomechanics of the
musculoskeletal system, 1995). As such, biomechanics is part of the world of physics
and concentrates on forces that act on biological structures. These forces either
originate from inside the human body, like muscle forces, or forces, created externally,
like an external load or impact. The structures they act upon can be classified as solids,
like skeletal parts or organs and by fluids, like blood and air.
The study on the effect of these forces can be classified as deformation and
displacement.
Biomechanics concentrates on various situations. To apply biomechanics four
subspecialisations are created, each focusing on a specific situation and simplifying the
real world in a specific way to allow a practical model that can be analysed
theoretically.
Statics deals with solid structures that will not move. In this way the forces that act
upon a structure can be calculated. All structures are assumed to be undeformable.
It is based on Newton’s third law, F12 = - F21: when a structure exerts a force on
another structure (for instance by its weight) that last body will exert an equal force in
opposite direction to the first body. It is also based on Newton’s first law:
F = 0 ⇔ v = constant: If the resulting force F on a structure is zero, then that
structure will not change its velocity, so non-moving structures will remain motionless.
Mechanics of materials deals with solid structures that will not move, but are
deformable.
Dynamics deals with solid structures that can move, but cannot deform. It is based
on the second law of Newton, F = m a: when a force F acts on a body with mass m, it
will undergo an acceleration a.
And fluid mechanics deals with fluids that can move and that can deform.
These four different biomechanical subspecialisations and their applications will
be discussed in this chapter.
This page intentionally left blank
Basic Engineering for Medics and Biologists 3
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-3
I.1. Statics
Gijsbertus J. VERKERKEa,b and T. Clive LEEc
a
University Medical Center Groningen, University of Groningen
Dept of Biomedical Engineering, Groningen, The Netherlands
b
University of Twente, Dept of Biomechanical Engineering, Enschede,
The Netherlands
c
Royal College of Surgeons in Ireland, Department of Anatomy, Dublin, Ireland
Abstract. The forces that act on an object determine its dynamic behaviour and
defromation. Analysis of all forces and moments is essential. A free-body diagram
summarizes all forces and moments that act on an object. To calculate the
magnitude of the forces we can use the static equilibrium of forces and moments.
Introduction
F=m•a (1)
a = Δv / Δt (2)
4 G.J. Verkerke and T.C. Lee / Statics
v = Δs / Δt (3)
1. Resultant Force
When two forces with different lines of action are acting on the same point of an object
(for instance two muscles that are active with the same insertion point but different line
of action), then we can combine them into a single resultant force. The magnitude and
line of action of this force can be found using the rules for adding vectors (Figure 1).
Figure 1. Two vectors A and B can be replaced by a resultant force R using the parallelogram
construction (left) or putting them head-to-tail (middle and right).
A force can also be resolved into two different forces using the same rule. Often
this is performed to simplify the effect of a force. If the seat of your chair is not
horizontal, but inclined under an angle β, then the force F that is acting on this seat due
to gravity can be replaced by two forces: a force N that is acting perpendicular to the
seat (also called the normal direction) and a force V acting parallel to the seat (also
called the shear direction). The force V lets you slide off the seat, while the force N
determines whether or not the chair will be able to carry you.
You can find the force V by multiplying force F with sin β (sinus β). Force N is
found by multiplying force F with cos β (cosinus β). The value of sin β and cos β can
be found on most pocket calculators and using your Windows calculator (scientific
view). The angle β is expressed in degrees. So:
V = F • sin β
N = F • cos β (4)
G.J. Verkerke and T.C. Lee / Statics 5
1.1. Friction
Four main types of forces exist in and on the human body: gravity forces, muscle
forces, friction forces and forces exerted by others or other objects, like a boxing glove.
A friction force always acts perpendicular to a surface and prevents an object moving
along that surface. So a friction force prevents you from sliding off the inclined seat.
Friction has some strange features:
First of all its magnitude is variable and only as large as is necessary to prevent
movement. If a force V is acting upon you, because you are sitting on an inclined seat,
then the friction force has the same size as V, but acts in the opposite direction.
Secondly there is a limit to the friction force, of course, because you will slide off
a steeply inclined seat. The friction force, acting during motion is slightly smaller than
the static friction force.
Thirdly, the friction force is linearly related to the normal force:
Ff = μs • FN (5)
with Ff = static friction force (N), μs = coefficient of friction and FN = normal force (N).
As you can derive from formula (5), μs is dimensionless. Its value depends on the two
surfaces that are in contact with each other and can be found in handbooks.
Fourthly, the friction force is independent of the area of contact between the two
objects. You can check this by placing a book on a board, start to incline the board and
remember at which angle it starts to move. Then place the book on its side on the board
and repeat the experiment. The angle will be the same. An explanation for this is that
both surfaces are rough and in theory to keep the book stable it rests only at three
contact points on the board. The coefficient of friction is determined by the nature of
these three contact points and thus is not determined by the contact surface.
Apart from the tendency to move an object along its line of action a force also has a
second effect: think of a billiard ball that is hit by your cue right in the middle: it will
move straight forward. Now hit it slightly off-centre. It will again move straight
forward, but also start to rotate around its axis (Figure 2). This rotating effect of a force
is called a moment. The effect is expressed by:
M=F•d (6)
with F the force (N), d the distance between the centre of the billiard ball and the line
of action of the cue (m) and M the moment. The distance d, also called the lever arm of
the force, must be perpendicular to the line of action of the force. Note that a moment
always is related to the point of which you would like to know the rotating effect of the
force, in this case the centre of the billiard ball.
Using formula (6), the unit of M is Nm. (Note that mN is not the correct notation
of this unit, since it means milli-Newton)
A moment allows you to open a door, to unscrew bolts, and to spin a ball. To
increase the effect of a moment you can either increase the force or the distance. Our
body uses this effect: since the muscle force is limited, enlarging d is an additional
strategy to increase the rotating effect, that is why you have a patella; it increases the
6 G.J. Verkerke and T.C. Lee / Statics
effect of your quadriceps muscle and makes it possible to extent your knee more
powerfully.
Figure 2. When a force is acting eccentrically over a distance d, a moment appears, resulting in a
rotation (left). Two parallel forces in opposite direction are called a couple (middle), also denoted by a
curved arrow (right).
A specific situation appears when two parallel and equal forces with an opposite
line of action are acting on an object (Figure 2, middle). The net action of the forces in
horizontal movement is zero, they only have a rotating action caused by their moment,
which is twice the size of the moment when one force is acting. Such a pair of forces is
called a couple and denoted by a curved arrow (Fig. 2, right), since its line of action
(the rotation line) is pointing out of the plane. A couple has a magnitude and
orientation and therefore is a vector.
In statics, moving a force along its line of action - translation - does not have any
influence. It does not matter whether you push an object or pull it. However, moving a
force perpendicular to its line of action does have an influence. Rotation is affected by
it, since its arm is changing. Moving a couple over an object is allowed. Although the
arm of one force is increasing, the arm of the other force is decreasing by the same
amount, leaving no net change. Therefore a couple is called a free vector.
In practice the terms couple and moment are used inaccurately and often
exchanged, which is not logical, since a couple is a free vector, but a moment is related
to a point on an object. The term couple moment is often used instead of couple, which
is also confusing.
2. Static Equilibrium
Now we know the effects of forces, either translation (movement along the line of a
force) or rotation, we can apply Newton’s First Law. According to this law, if an object
is not subjected to a force, this object will not start to move or change its velocity.
Since, in statics, we consider all objects to be at rest, no net forces should act on them.
Also the sum of their moments should also be zero to avoid rotation. So in statics the
sum of all forces should be zero and the sum of all moments in relation to a certain
point on the object is also zero. This means that the object has no degree of freedom to
move.
G.J. Verkerke and T.C. Lee / Statics 7
Σ Fx = 0
Σ Fy = 0 (7)
Σ Mp = 0
where Σ is capital sigma and means ‘the sum of’.
Note that when the sum of all moments is zero in relation to a certain point on the
object, it will also be zero to another point on that object. So the reference point is a
free choice, but must be used subsequently for all forces.
In the three-dimensional world, which we will not consider in this paper, six
equations of equilibrium exist:
Σ Fx = 0
Σ Fy = 0
Σ Fz = 0 (8)
Σ Mx = 0
Σ My = 0
Σ Mz = 0
3. Free-Body Diagram
In 2D, in order to apply the three equations of equilibrium, we have to know which
forces are acting upon an object. We already found out, that not only obvious external
forces are active, like gravity, but also hidden ones, like the force of a seat acting upon
your bottom. To make this last type of force visible, we have to free the object from its
surroundings. After that we have to draw all forces of the surroundings that act upon
this object. These forces are unknown, but by applying the three equations of
equilibrium we can calculate them and so complete our static analysis. Thus, all forces
that are acting upon an object are known.
Then we can proceed to dynamics or mechanics of materials to find out the
movement of this object or its deformation.
The main problem in making an object free from its surroundings is to determine
the reaction forces from the surroundings. In Figue 3 a list of connections between the
object and its surroundings (the supports) is shown, along with the replacing reaction
forces and couples. As a general rule, if a support prevents translation of an object in a
given direction, then a force is present on the object in the opposite direction. Similarly,
if a rotation is prevented, a couple is exerted on the object.
8 G.J. Verkerke and T.C. Lee / Statics
F2
F1
M
F2
F1
Figure 3. Overview of possible supports and their replacing forces and couples.
Making an object free from its surroundings is a very delicate action, since we
have to decide which object we select. If we select a whole human body as our object,
and we are interested in the forces acting in our shoulder joint when lifting a bag, we
will not be able to calculate these forces, because they are hidden inside the object. So
in that case we have to select the upper limb as our object and free it from the rest of
the body. We again have to determine the reaction forces of the rest of the body. Most
supports consist of joints, which in general cannot be characterised as regular hinge-
type connections. We can distinguish 6 types of joints (Figure 4):
1. Ball and socket joints, like the hip joint, allow three rotations and no translations.
2. Condyloid (ellipsoid) joints, like the knee, allow in general rotation around one
axis and some rotation around another axis. When the knee is extended there is no
axial rotation possible, when it is flexed some axial rotation is possible.
3. Saddle joints, like the thumb (between the 1st metacarpal and trapezium), allow
rotation around two axes.
4. Hinge joints, like the elbow (between the humerus and the ulna), act like a door
hinge, allowing only flexion and extension around one axis
G.J. Verkerke and T.C. Lee / Statics 9
5. Pivot joints, like the elbow (between radius and ulna). This is where one bone
rotates about another.
6. Gliding joints, such as in the carpals of the wrist (not shown in Figure 4). These
joints allow a wide variety of movement, but not over a large distance.
A moment allows you to open a door, to unscrew bolts, and to spin a ball. To
increase the effect of a moment you can either increase the force or the distance. Our
body uses this effect: since the muscle force is limited, enlarging d is an additional
strategy to increase the rotating effect, that is why you have a patella; it increases the
effect of your quadriceps muscle and makes it possible to extent your knee more
powerfully.
A moment allows you to open a door, to unscrew bolts, and to spin a ball. To
increase the effect of a moment you can either increase the force or the distance. Our
body uses this effect: since the muscle force is limited, enlarging d is an additional
strategy to increase the rotating effect, that is why you have a patella; it increases the
effect of your quadriceps muscle and makes it possible to extend your knee more
powerfully.
Figure 4. Schematic drawing of 5 joint types, ball and socket (1), condyloid (2), saddle (3), hinge (4)
and pivot joints (5).
Apart from separating the joint surfaces we also have to cut through muscles to
free our upper limb from the rest of our body. A muscle can be categorised as a cable
and must be replaced by a single force with an orientation from origin to insertion.
Let us now summarize the procedure for determining the shoulder joint and muscle
forces in case we think we need to lift some weights (of in total 100 kg, see Figure 5,
left). First we must define a positive x-axis, y-axis and direction of moments (Figure 5,
right). Then we need to free the object of interest, in this case the upper limb.
10 G.J. Verkerke and T.C. Lee / Statics
The upper limb is connected to the rest of the body by the shoulder joint, which
can characterised as a ball and socket joint in 3D, and in 2D as a hinge. So the
shoulder joint can be replaced by two forces, F1 and F2. The upper limb is also
connected with muscles to the rest of the body. In this situation the deltoid muscle will
be (very) active. A muscle can be replaced by a single force, pointing along its line of
action, Fs. Now we can make the free-body diagram (Fig. 5, middle), including all
forces acting on the upper limb: half of the weight of the dumb-bell (50 kg x 10 m/s 2 =
500 N), the weight of the upper limb (assumed to be 2.5 kg, resulting in 25 N), the
unknown muscle force Fs, and the unknown joint reaction forces F1 and F2. To be able
to use our equations of motion, we first have to split F1, F2 and Fs in a force, acting in
x-direction and a force, acting in y-direction using formula (4).
600
M+
Fg Fw
Fs F1 x
F2
0.05 m
0.2 m 0.25 m
Figure 5. How to calculate shoulder joint and deltoid forces when lifting a weight (left): a free-body
diagram is made of the arm (middle), including the orientation of the x- and y-axis and the positive direction
of moments (right)
Finally we can calculate the unknown forces using the three equations of
equilibrium:
Fs = 4600 N (9)
F1 = √3 F2 – Fs (10)
Obviously we have chosen both F1 and F2 in the wrong direction, because the
outcome is negative for both forces. So the direction of the unknown forces can be
chosen arbitrarily, after the calculation will become clear, if the assumed direction is
the proper one.
The magnitude of Fs and F1 are considerable, corresponding with a mass of 460
and 505 kg, respectively! The main reason for this is the small arm of Fs compared to
the arm of Fw; muscle arms are small due to geometric considerations. So muscles must
be very strong to compensate for this short arm.
4. Internal Loads
Now we are able to determine the forces, acting on an object and to calculate the
magnitude of these forces, next we can analyse two situations:
Firstly we can find out what will happen if an extra force or couple is applied and
thus the equilibrium of forces and moments is disrupted. In the ‘Dynamics’ chapter we
will calculate the resulting accelerations.
Secondly we can consider our object to be not rigid, so deformable and calculate
the deformations that result from the forces and determine, if our object can resist the
forces, and thus if it is strong enough. For this last action we need to do an extra step in
statics, we have to calculate the internal loads and determine what location is the
weakest point.
d
y
600
M+
M
Fg Fw
F2 x
F1
0.25 m
Figure 6. For the body builder on the left the internal forces of his arm are calculated (middle), using a
co-ordinate system (right).
The first step we have to do is to isolate our object (Figure 6, left), define a
positive x- and y-axis and direction of moments (Fig. 6, right) and calculate the external
12 G.J. Verkerke and T.C. Lee / Statics
reaction forces (in our case there are none). Then we have to cut our object in two parts
and replace one part of our object (in our case the lower left part) by its reaction forces
F1, F2 and M at the cutting edge (Fig. 6, middle).
The resulting free-body diagram (Fig. 6, middle) contains three unknowns, F1, F2
and M that can be calculated using the three equations of equilibrium. Note that the
position of the cut is defined by the distance d between the cut and the upper right end
of our object. By varying this distance we can scan the entire arm and thus calculate the
internal forces at any spot. It then is possible to determine the site at which the highest
loads will occur and thus the site where failure will occur first:
→ F1 = √3 F2 (15)
Again, both F1 and F2 are acting in the opposite direction as was presumed.
Now we can find out, where the most critical spot in the upper arm is. Both F1 and
F2 are constant, M is maximal for d = maximal. So the highest load will act where d is
maximal, which is the lower left end of the upper arm.
References
[1] G.L. Lucas, F.W. Cooke, E.A. Friis, A Primer of Biomechanics, Springer-Verlag,
1998
[2] R.C. Hibbeler, Statics and Mechanics of Materials (Second edition), Pearson –
Prentice Hall, 2004
[3] N. Özkaya and M. Nordin, Fundamentals of Biomechanics; Equilibrium, Motion
and Deformation (Second edition), Springer Science+Business Media, 1999
Basic Engineering for Medics and Biologists 13
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-13
Introduction
This field of science deals with the relation of external loads acting on a body to the
induced internal forces experienced by the material of that body. Internal forces either
result in deformation i.e. gross changes in the external shape or failure.
Assume a bucket of mass 1 kg hanging on a rope, then from intuition we know that the
bucket is pulling the rope downwards i.e. exerting a force due to gravity. The
magnitude of the force is equal to the mass of the bucket times the acceleration due to
gravity, g, with a value of 9.8 m s-2. Thus the force experienced by the rope (F = m x g)
is 9.8 kg m s-2, or 9.8 N (≈ 10 N). In other words, one Newton of force will act if 102
(≈100) g of mass is hanging from the rope.
Let us now analyze a simple situation where mass m is hanging on a cylinder with
radius of r (Figure 1, left) and a mass of 2m on a cylinder with radius 2r (Figure 1,
right). Can we say which cylinder will fail first? A force of mg is acting on the left
cylinder and twice the force 2mg on the right cylinder. Cross-sectional area of the
cylinder on which this force is active is also different, the left cylinder has an area of
πr2 (pi, a constant = 3.142 x radius squared) whereas the right cylinder has 4 times the
area i.e. 4πr2. We have to invoke the concept of Stress, denoted by σ, (sigma) in order
to answer this question.
14 P.K. Sharma / Mechanics of Materials
F
σ= (1)
A
where F is the force and A is the cross-sectional area on which the force is active. Unit
of σ is N/m2, also called Pascal (Pa)
r 2r
m 2m
The left cylinder is experiencing a stress of mg/πr2 whereas the right cylinder
experiences half that much (mg/2πr2), therefore we can now say that the left cylinder
will fail first.
The situation we analyzed is called uniaxial state of stress because the cylinders
are experiencing stress along one axis and the stress is called normal stress because
the force is acting perpendicular to the cross-sectional area (Figure 2A). Normal stress
always tries to either pull, called tension, or push, called compression, a body (Figure
2B).
A B
90o
F Tension Compression
Material inside the cylinders as shown in Fig 1 responds to this tensile stress and
deforms by stretching itself resulting in increase in the length of the cylinder at the cost
of a decrease in the diameter. This deformation is expressed in terms of normal strain,
P.K. Sharma / Mechanics of Materials 15
denoted by ε (epsilon). Strain is a concept where the deformation due to the applied
stress is normalized by the initial dimensions of the body.
ΔL
ε= (2)
L
where L is the original length and ΔL is the increase in length (Figure 3A). Strain
is a dimension-less quantity since both ΔL and L are expressed in the same unit.
L L
ΔL
A. Tension
ΔL
L
B. Compression
In case of compression (Figure 3B) the formula to calculate the strain remains the
same but the meaning of ΔL changes into a decrease in height thus takes a negative
sign.
Why can’t we simply measure the applied force, in Newtons and deformation in
centimeters instead of using stress and strain? The concept of stress and strain helps us
to compare the mechanical properties of say stainless steel and aluminum cylinders
without having to make them exactly of the same size and shape and applying exactly
the same force. It helps us to compare our results with laboratories from the other side
of the world and most importantly by measuring the mechanical properties of a small
piece of metal in the laboratory we can predict the performance of a long beam of the
same metal placed in a bridge.
In day to day life we talk about the strength of different materials and compare them by
saying that one is stronger, tougher or harder than the other.
16 P.K. Sharma / Mechanics of Materials
Elastic Plastic
I II III IV
Strain to
failure
Stress, σ (Pa)
Elastic
limit
Tensile Failure
strength stress
Yield
stress
Strain ε
What do we mean by all these words- strength, toughness and hardness? Strength,
toughness and hardness depend on the ability of a material to sustain load without
undue deformation. If we keep on increasing the force on the left cylinder in Figure 1
then we know that the cylinder will deform, but do we know at what applied force it
will fail and how long it will become? The answer is that it depends on the strength of
material the cylinder is made of. Strength of a material can be determined by
performing either a tensile or compression test and record the results in terms of stress
and strain. During a tensile test, an increasing amount of stress is applied to a standard
specimen and the resulting strain is measured, these stress and strain data are plotted on
a so called stress-strain diagram (Figure 4). Starting from zero, as the applied stress is
increased the strain increases linearly up to a limit called the elastic limit. In this region
I the material shows elastic behavior and follows Hooke’s Law (Eq. 3), where the strain
is linearly proportional to the applied stress and the constant of proportionality E is the
Young’s modulus having the same units as stress i.e. Pascal
σ = Eε (3)
If the applied stress is removed within the elastic limit then the material regains its
original size meaning that the strain reduces to zero and there is no permanent
deformation in the specimen (Figure 5A). At the end of the elastic region, the plastic
region starts where a slight increase in stress above the elastic limit i.e. yield stress
results in deformation almost without an increase in stress, this process is called
yielding (region II). In region III the applied stress again increases, resulting in an
increase in strain. The point where the maximum amount of stress is supported by the
specimen is called the tensile strength. In region IV the supported stress continuously
decreases leading finally to failure. If the applied stress is removed in the plastic region,
the strain does not reduce to zero but some residual strain remains in the form of
permanent deformation (Figure 5B). The start of the plastic region is difficult to
determine, since the linear increase of the stress is gradually changing into the yielding
phase. For practical reasons the start of the plastic region is defined by the point where
P.K. Sharma / Mechanics of Materials 17
if the applied stress is removed the specimen keeps a small amount of permanent
deformation (ε = 0.002 = 0.2 %); this stress is therefore called the yield stress σ0.2.
Figure 6 shows two different type of materials; material A (Figure 6A) is strong
because it can support very high stresses but this material is also brittle, meaning that it
cannot be given any permanent (plastic) deformation, just after reaching the maximum
stress the material fails giving rise to a flat fracture surface (Figure 6C). Material B
18 P.K. Sharma / Mechanics of Materials
(Figure 6B) on the other hand is less strong, but tough and ductile because it deforms a
lot before failure meaning that it can absorb a lot of energy by deforming permanently
before failure. This amount of energy is represented by the area under the curve,
Figure 6B inset.
For a tough material the cross sectional area reduces at a localized area and causes
a large plastic deformation, called necking, before the material fails (Figure 6D).
While constructing a bridge, engineers choose a tougher material because then they
can see their bridge change its shape before it finally fails. If they use a strong and
brittle material then the bridge will fail catastrophically without any prior warning.
Hardness tells us the resistance of a material to scratching by another material, this
property is important if we do not want our material to wear out during rubbing with
other surfaces. This is the reason for use of hard materials as artificial joint surfaces and
drill bits. Similarly cortical bone is stronger and harder to avoid wear and tear whereas
trabecular bone is tougher to absorb energy during jumping and falling.
Till now we have talked about the tensile strength of a material, but the normal
stress can also be compressive in nature. Materials like rubber can deform to up to 10
times their original size under tension where as concrete cannot (Figure 7). Concrete
has a very low tensile strength but have very high compressive strength; that is why
steel rods are embedded in concrete to bear the tensile part of the stress in a building
column. In the human body most of the bone structure and teeth are good at bearing
compressive stresses whereas ligaments are specialized structures meant to bear tensile
stresses.
Rubber
ε
1 10
σ
Cement
ε
-0.06 0.01
A B
F
Figure 8. Shear Stress
We rarely feel the presence of yet another type of stress, which is called shear stress.
This occurs when the force is active parallel to the cross-sectional area (Figure 8A) and
tries to slide one part of the body in relation to the other part (Figure 8B). Shear stress
is denoted by τ (tau) and also equal to the force acting per unit cross-sectional area
with the unit Newton (N) (Eq.4).
F
τ= (4)
A
A B
L
d
Figure 9. Shear Strain
The material reacts to this shear stress as shown in Figure 9A and the shear strain,
denoted by γ (gamma), is calculated using equation 5 (Figure 9B) and again it is a
dimension-less number
d
γ= (5)
L
A material experiencing shear stress also demonstrates a shear stress-strain curve
very similar to a normal stress strain curve, with an elastic region where it follows
Hooke’s law and shear strain is linearly proportional to shear stress and having a
modulus of rigidity, G, (Eq.6). Above the elastic limit the material deforms
permanently
20 P.K. Sharma / Mechanics of Materials
τ = Gγ (6)
Normal stresses (both tensile and compressive) and shear stress are the only two
types of stresses which can exist inside the material and hence the material manifests
only normal and shear strain. All the different types of external forces acting on the
body induce normal or shear stresses, although active in more than one direction.
There are two more ways in which external forces can be active on a body; one in
which the cylinder is twisted called torsion (Figure 10) and second where the cylinder
is bent called bending (Figure 14A). We shall now analyze how these external forces
translate into internal stresses.
T T
τ τ
Cylinder Tube
If the cylinder surface is marked with a black square (Figure 11A) and external
torsional moment, T, is applied then the cylinder deforms in such a way that the square
is transformed to a rhombus shape (Figure 11B&C). We have seen that this
transformation takes place only when the forces are active parallel to the cross-
sectional area i.e. shear stress (Figure 8, 9). Thus externally applied torsional forces
induce shear stresses in the cylinder hence the cylinder shows shear strain (γ=d/L,
Figure11B,C,D).
If we know the magnitude of the externally applied torsional moment, also called
torque, on a cylinder of radius r then we can calculate the induced maximum shear
stress at the cylinder surface using Eq.7.
2T
τmax = (7)
πr 3
If we look at the cross section of the cylinder then we will see that the induced
shear stresses vary linearly from zero in the center to a maximum value at the surface,
whereas in a hollow tube the stress is a finite value at the inside diameter with
maximum at the outer diameter (Figure 12). The maximum shear stress at the hollow
tube surface with outside radius, ro, and inside radius, ri, is calculated using Eq. 8. For
the same applied torque and outside tube diameter, as the inside tube diameter increases
the maximum induces shear stress at the outside surface increases.
P.K. Sharma / Mechanics of Materials 21
2Tro
τmax = (8)
π( ro4 − ri4 )
A A B
h
b
F
L
B C D D
C
d
L E σ
+ τ
- 0 0
Figure 11. Internal shear stress Figure 14. Normal and shear stresses induced by the external
induced due to external torsional bending moment
forces.
If we know the material properties then we can calculate the shear strain at each
point, this shear strain cumulatively results in twisting of the cylinder. The angle of
twist, θ (theta), (Figure 13) for a cylinder made of material with modulus of rigidity G
and length L experiencing an external torque of, T, can be calculated using Eq. 9. The
longer and thinner a cylinder greater is the angle of twist.
2TL
θ= (9)
πGr 4
If a square cross-section beam is fixed at one end and a force is applied at the
other end perpendicular to its axis then there is a induced bending moment M in the
bar. A bending moment is the applied force multiplied by the distance of the force
from the fixed end (M=F.L) and is thus reported in N.m (Figure 14A).
22 P.K. Sharma / Mechanics of Materials
T
Figure 13. Angle of twist due to torsion
This bending moment induces both normal and shear stresses in the beam; Figure
14A shows how a rectangle drawn on the side of a beam changes to a trapezoid due to
stretching of the beam above the central axis and compression below it. This means that
tensile stresses are induces on the beam above and compressive stresses below the
central axis (Figure 14C&E). The maximum tensile or compressive stress at the top or
bottom surface can be calculated by Eq. 10.
6M
σmax = (10)
Tensile / Compressive
Top / bottom bh 2
If we imagine the beam to be constructed with a pile of much thinner beams
(Figure 14B) then by intuition, the same as bending a thick book along its binding axis,
we know that these thin beams will slide and move relative to each other when a
bending moment is applied (Figure 14D). This can only happen if there is a force active
along the interface between the thin beams i.e. shear stress is being active. Even when
the beam is not composed of thinner beams we can see that these shear stresses will be
induced, the proof for this can be seen from the longitudinal fracture through the center
visible in old wooden beams used to support the roof. Induced shear stresses are
maximum at the central axis and drop to zero at the top and bottom surfaces (Figure
14E). Maximum shear stress at the central axis can be calculated by equation 11 where
V is the internal shear force resulting from the applied bending moment.
3V
τmax
centralaxis = (11)
2bh
The deformation caused by a bending moment is deflection “v” and slope of the
beam. Both parameters can be calculated at any point on the beam defined by distance
x from the fixed end by solving a differential equation (Eq. 12), where the Young’s
modulus (E), moment of inertia (I) and bending moment at the point (M(x)) is known.
d 2v
EI = M ( x) (12)
dx 2
P.K. Sharma / Mechanics of Materials 23
Some typical solutions to the differential equation for a beam loaded with force, F
(N), bending moment, M (Nm) and distributed load, Q (N/m) is presented in Figure 15.
Figure 15. Deflection and slope of a beam under different loading situations
4. Failure
Failure can now be decided in terms of the maximum stress above which we consider
our material to be rendered useless. While constructing a bridge or making a load
bearing implant, we would never like that our structure deforms permanently thus we
can say that the upper limit of any stress induced in the material should be the yield
stress. In building and bridge construction a margin is given to be always on the safe
side i.e. the calculated induced stresses in the material should be always, say, less than
70% of the yield stress and this is called the maximum allowed stress (σallowed , τallowed).
In simple external loading situation like tension, compression, torsion or bending
we can calculate the maximum internal stresses using equations 1, 4, 7, 8, 10 and 11
and if these stresses are higher than the allowed stresses then we know that the
structure will fail. In real situation two or more of these external loading takes place
simultaneously, what happens then? One such simple example of combined loading is
a spherical liquid storage tank, for example the heart. The tank is pressurized and if we
look at a small square piece of material on the tank surface we can see that the square
experiences tensile stresses on both sides as shown in Figure 16. To handle these
situations with combined loading we use Mohr’s circle. Most of the materials have to
be analysed 3-dimensionally but since they are geometric in nature, using the
symmetry we can convert them to 2 dimensions like we did for the spherical storage
tank (Figure 16).
24 P.K. Sharma / Mechanics of Materials
−
σy
d c b σ
σx − σ x +σ y +
2θ p
σ avg =
2 2θ s τ xy
a
e
σx
τ +
Mohr’s circle is a plot between the normal stress (σ) on the horizontal or x-axis
and the shear stress (τ) on the vertical or y-axis (Figure 17). Tensile normal stresses are
positive and compressive are negative, for the shear stress we follow a sign convention
shown in Figure 18 where a very small volume element inside the bulk of the material
is shown. Coming back to Figure 17, the position of the center of the circle c is the
average of the two normal stresses (σx and σy and reference point a can be plotted with
τxy on the y-axis and σx on the x-axis. Now we can join points c and a to get the radius
and then the circle can be drawn with a radius given by Eq. 13. From this Mohr’s
circle for a given situation we can find out the maximum normal stress, point b, and
shear stress, point e, active in the material.
2
⎛ σx − σ y ⎞
R = ⎜⎜ ⎟ + τ 2xy (13)
2 ⎟
⎝ ⎠
If we take a simple situation where a tensile stress of 100 Pa is active on a square
cross section beam then the state of stress at a very small volume element will
+σy -σy
+τxy -τxy
+σx -σx
be as shown in Figure 19A. Surprisingly, the Mohr’s circle (Figure 19B) of this volume
element shows that there exists an orientation (45o clockwise, Fig19C) where
P.K. Sharma / Mechanics of Materials 25
A B σ x −σ y
= 50
2
100 Pa
a σ
+σx σ avg = 50 90O τ xy = 0
σx = 100, σy = 0 e
τ = 50 Pa
τxy = 0
τ
C D
100 Pa 50 Pa
100 Pa
e(-50,-70.7)
100 Pa
a(-100,-50)
T
σy=0 σ c(-50,0) b(20.7,0)
d(-120.7,0)
σx=-100 Pa
τxy = -50 Pa
τ
Figure 20. Mohr’s circle for combined compressive and torsional loading
26 P.K. Sharma / Mechanics of Materials
References
[1] R.C. Hibbeler, Mechanics of Materials. Printice Hall International, New Jersey, 2003
[2] N. Özkaya and M. Nordin, Fundamentals of Biomechanics: Equilibrium, Motion, and Deformation.
Springer, 1999
Basic Engineering for Medics and Biologists 27
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-27
Abstract. The part of (bio)mechanics that studies the interaction of forces on the
human skeletal system and its effect on the resulting movement is called rigid
body dynamics. Some basic concepts are presented: A mathematical formulation to
describe human movement and how this relates on the mechanical loads acting on
the skeletal system. These equations of motion depend on the mechanical
properties of the skeletal system, such as dimensions and mass distribution. It is
applied to describe and analyze human gait.
Introduction
Human movement dynamics relates the applied forces or moments of force (i.e. the
kinetics) to the movement (i.e. kinematics) of the skeletal system. In the control loop to
displace body parts (Figure 1), the Central Nervous System (CNS) is the controller that
activates the motors (muscles). The movement dynamics are represented by the skeletal
system, and the resulting movement is measured and fed back into the CNS.
In this chapter we focus on the skeletal system only. The part of (bio)mechanics
that studies this is called rigid body dynamics. Some basic concepts are presented: A
mathematical description of human movement; how this depends on the mechanical
loads acting on the skeletal system and how this depends on the mechanical properties
of the skeletal system, such as dimensions and mass distribution. It will be applied to
describe human walking.
The human body consists of a more or less rigid structure (the bones of the skeleton) to
which soft tissues (muscles, fat, organs etc.) attach. A structure is considered rigid
when under practical loading situations, the structure does not noticeably deform, i.e.
the mechanical stiffness seems to be infinite. Since the soft tissues are not rigid and do
deform in practical loading situations, an accurate mechanical description of human
28 H.F.J.M. Koopman / Dynamics of Human Movement
movement involves the description of both soft tissues and skeleton. The description of
the dynamics of the human body would involve very complex calculations, since not
only movement, but also deformations have to be considered. This approach is only
feasible in some very specific applications, for example to study the effect of car crash
impacts on brain damage. As it is not possible to measure all soft tissue deformations,
some assumptions have to be made to be able to study human movement.
sensory physiological
feedback sensors
The most important assumption in rigid body dynamics is that movement only
occurs in the joints. As a consequence of this assumption, the structures in-between the
joints, the so-called segments, are assumed to be rigid, so they have an infinite large
stiffness. All passive soft tissue structures are considered to be rigid as well. The
muscles are replaced by the forces they produce. So it is assumed that the human body
behaves like a segments model or a linked system of rigid bodies. Apart from the
enormous reduction of complexity, this viewpoint has other advantages as well: The
rigid body equations of motion, when applied to each segment, are sufficient to
describe the dynamics of the entire system.
Rigid body dynamics describes the kinematics (movement) and kinetics (forces) of
a given segments model. The equations of motion are the link between these two,
which is a set of two first-order nonlinear differential equations for each movement
component (or, alternatively, a single second order equation that includes a dependency
on the second time derivative, the acceleration). For a given segments model the forces
and movements provide interchangeable information: The forces can be calculated
when the movements are known (i.e. inverse dynamics approach) and the movements
can be calculated when the forces are known (i.e. direct or forward dynamics approach).
2. Kinematics
Two bones can move with respect to each other by virtue of the joint in between. In the
human body, only the synovial joints allow for large movements. In principle, a bone
has six Degrees-of-Freedom (DOF) of motion with respect to the other bone: three
rotations and three translations. The motions of the joint are limited by passive
structures like the articular surfaces and the ligaments. These passive structures pose
H.F.J.M. Koopman / Dynamics of Human Movement 29
restraints to the joint motions: Though motions in the direction of the restraint are still
possible, these motions will be very small.
For example, cartilage can be compressed a few millimeters. Most of the time
these small motions are neglected, the motion is said to be constrained. For each
constraint the number of DOF diminishes by one. As soon as constraints come into the
picture, one has already started modeling the joint. In Figure 2 all combinations of
(constrained) rotations and translations are shown. Many of them are merely
hypothetical, and will not be found in the human body.
Figure 3. Schematic drawing of 5 joint types, ball and socket (1), condyloid (2),
saddle (3), hinge (4) and pivot joints (5).
Traditionally, joints have been studied (e.g. by Fick [4]) by comparing them with
standard revolute joints: Hinges, spherical joints, ellipsoidal or condyloidal joints,
saddle joints (see Figure 3). The shoulder joint and the hip joint behave approximately
as spherical joints only permitting three rotational DOF. The elbow joint and finger
joints can be regarded as hinge joints with one rotational DOF. The first metacarpo-
phalangeal (thumb) joint and the ankle joint resemble a saddle joint, having two non-
intersecting rotational axes. But often one cannot derive the potential motions from the
shape of the articular surfaces.
Since anatomists were the first to study joint motions, medical definitions still
dominate the way joint motions are described. The goal of such medical definitions is
to distinguish between pathological and normal motion, and to evaluate the outcome of
treatment: Is there improvement in the range of motion or not.
As a starting position the anatomical position (Figure 4) is used. From this position
the motion is defined for each single rotation. For spherical joints, the rotation axes are
defined along the axis of the global coordinate system: A vertical axis and two
horizontal axes, pointing backward-forward (antero-posteriorly) and side-side (medio-
laterally, from the centre of the body to the left or right side). Commonly used terms
are flexion-extension, abduction-adduction, and medial and lateral rotation. A problem
occurs if the motion is not about just one of the standardized axes, but is a combination
of rotations. The order of rotations is not defined, and it is not clear whether the
rotation axes move with the bone or not. This causes much confusion about rotation
angles, and makes comparison between studies often impossible.
H.F.J.M. Koopman / Dynamics of Human Movement 31
All models to analyze or simulate human movement are based on segments. These are
coupled systems of rigid bodies, with dimensions and inertial properties that are
deduced from the dimensions and inertial properties of the human body. Initially, these
segments models were used to dimension the dummies used in car crash experiments
[3] and were therefore based on averaged data of human bodies. With the introduction
of the segments models to computer simulations, a wider variety became possible. The
segments models became scalable, with inertial properties usually depending on local
segment dimensions and total body weight. In this way, the segments model could be
matched to fit each individual [2].
The choice of the number of segments should be large enough to simulate the
movement adequately. Too large a number of segments, however, would lead to
unnecessary complexity and larger computational efforts. To simulate walking,
segments models varying from 3 segments [10] up to 17 segments [6] have been
proposed. To further reduce complexity, symmetry between right and left leg is often
assumed [1] and the movement is often restricted to the sagittal plane only.
In a model for normal walking, there are segments for the thighs, shanks and feet.
The head, arms and trunk (HAT) and the pelvis are modeled as two separate segments.
The segments are connected to each other at the joints (Figure 5). Although it is
possible to have more than two joints in a segment (e.g. the pelvis), each joint is
32 H.F.J.M. Koopman / Dynamics of Human Movement
connecting just two segments. To define the position of the segments in space, an
absolute or reference frame is attached to the floor, with the x-axis pointing in the
walking direction, the y-axis pointing upward and the z-axis perpendicular to the xy-
plane in the lateral direction. Figure 5 shows an 8-segmental model with numbering of
the segments, the joints and definition of the reference frame. It should be noted that
the shape of the segments is of no importance as long as the positions of the joints and
the mass properties are well defined. In each segment, a local frame is defined,
standardized with the help of some bony landmarks on the segment.
Figure 5. The segments model for the body at rest. From left to right: Numbering of segments and joints,
definition of the reference frame, definition of local frames and positions of the centers of mass.
The segments model must include the following mechanical properties, which are
dependent on individual characteristics: The dimensions and positions of the segments
in space, the positions of the joints, and the mass properties of the segments (mass,
position of the center of mass and moment of inertia tensor). Other properties, such as
the ranges of motion of the joints, the maximal moments of force that can be exerted on
the joints, and muscle models, may also be included [8].
The segments model is the basis for a sound technical description of human movement.
In general, a movement of a rigid body is defined as the displacement and orientation
of the local frame relative to the reference frame or another local frame. Therefore, any
description of a (body or joint) rotation implicitly assumes that some kind of segments
model is defined, otherwise the rotation has no unique meaning. The anatomical
position is usually taken as the offset position for the segments model where the
relative rotations are defined as zero.
H.F.J.M. Koopman / Dynamics of Human Movement 33
Bold capitals are used to denote matrices; underlined lowercase characters are used
to denote vectors. Gp is the position vector of point P in the global coordinate system G,
A
p is the position vector of point P in the local coordinate system A. Gt is the translation
vector from the origin of A expressed in coordinate system G. Obviously At, the origin
of A expressed in coordinates of A equals the zero vector 0. GAR is the rotation matrix
that describes the rotation from the local coordinate system A to the coordinate system
G (the global coordinate system).
Element-wise, equation (1) looks like
3. Dynamics
The next step after defining the movement of rigid bodies is to derive the equations of
motion that define the relations between the acting forces and torques and the
movement. A straightforward way to do this is to apply the Newton-Euler approach.
Although alternative formulations exist that have specific applications, such as the
Lagrange approach, the Newton-Euler approach is simplest because:
The method is identical for each segment.
It is based on the Free Body Diagram of the segment. With this Free Body
Diagram there is a simple recipe to obtain the equations of motion.
The equations of motion for the entire body is the sum of the equations for each
segment.
Newton formulated the equations of motion for systems of mass particles. Euler
recognized that a rigid body is a special case for such a system: The positions of the
particles are constrained with respect to each other. This leads to the notion that the
internal forces (the forces acting between the particles) do not perform work and do not
contribute to the equations of motion for the entire system, the rigid body. Since a rigid
body has six degrees of freedom (DOF), there must be six equations describing the
relation between forces and motion. This leads to the formulation of the Newton-Euler
equations of motion for each segment:
H.F.J.M. Koopman / Dynamics of Human Movement 35
d 2p
CM
F CM = m
dt 2 (4)
d ( J CM ω )
M=
dt
Where FCM is the resulting external force acting on the center of mass (CM) of the
rigid body with mass m and pCM is the position of CM. The first derivative with respect
to time of pCM reflects the change of position in time, or the velocity of CM. Likewise,
the second derivative of pCM (as in equation 4) reflects the change of velocity in time,
or the acceleration of the center of mass. Any force, acting on the rigid body, can be
divided in a force, acting on CM, and a moment of force M. The second vector
equation in (4) resembles the first, as the resulting torque equals the product of rotation
inertia tensor JCM and angular acceleration (or the first time derivative of the angular
velocity vector ω (omega). However, unlike the mass m, JCM is in general not constant
in each coordinate system. This leads to considerable complications in 3-D; in 2-D on
the other hand JCM reduces to a single constant component.
MP FP
P rAP
A θ
rA
G
y
G mg
G
x
Consider now the 2-D situation of a (foot) segment A with mass m and moment of
inertia JA during the swing phase of walking. A local frame is attached to the center of
mass with an angle θ to the global (inertial) frame G. The segment connects to another
segment in joint P, the ankle joint (Figure 7). Of course, the lower leg segment applies
some connection forces on the foot in the ankle joint, expressed by the joint force FP
and joint moment of force MP. Finally, a gravity force applies to the foot, thus
completing the free body diagram. When all vectors, which have both magnitude and
direction, are expressed in global coordinates, equation (4) reduces to 3 scalar
equations – scalars have magnitude only:
36 H.F.J.M. Koopman / Dynamics of Human Movement
G d 2 G rAx
FPx = m ⋅
dt 2
G
d 2 G rAy
FPy − m ⋅ g = m ⋅ (5)
dt 2
dω
M P +( G rAPx ⋅G FPy − G rAPy ⋅G FPx ) = J A ⋅
dt
further processed in muscle models to estimate the distribution of muscle forces, which
allows for a validation with measured EMG patterns.
In the direct dynamics approach, the movements of the segments are calculated by
integrating the equations of motion, usually based on a Lagrangian formulation. This is
only possible when the joint moments of force are known or assumed to be zero. The
latter is the case in ballistic walking [10]. The joint moments of force can be found by
trial and error such that a normal walking pattern results [12], from estimations of the
muscle forces [11], as the result of an inverse dynamics model or by optimization
techniques. This direct dynamics modeling will not be discussed further.
4.1. Step-Parameters
The step-parameters are used for a general characterisation of the walking pattern.
They are determined by the points of heel contact (HC) and toe-off (TO) of the left and
right feet. For a walking cycle beginning with left heel contact (HCL), this is followed
by right toe-off (TOR), right heel contact (HCR) and left toe-off (TOL). A cycle is
completed with HCL again (Figure 8).
These points divide the walking cycle in four different phases. For the left leg, the
stance phase is from HCL to TOL and the swing phase (or the single stance phase of
the right leg) is from TOL to HCL. There are two double stance phases where both feet
are on the floor. The left double stance phase is from HCL to TOR. One stride consists
of two steps: The left step is usually defined from HCR to HCL.
For the step-parameters, a distinction is made between the time-parameters and the
distance-parameters. With the cycle beginning at HCL (tHCL=0), four other time-
parameters will suffice to define the points of HC and TO. These may be normalized
with the stride or cycle time T to make these time parameters dimensionless for
comparison purposes. Likewise, the distance parameters can be made dimensionless
with the stride length S. For symmetrical walking, the step-parameters are not
independent any more.
38 H.F.J.M. Koopman / Dynamics of Human Movement
Two step-parameters are derived from S and T, which are the average forward
velocity v and the step ratio r. These are calculated from:
v=S (6)
T
r= 1
4S ⋅T (7)
The step ratio is usually defined as the step length divided by the step frequency.
In contrast to this, definition (7) does not depend on differences between right and left
leg for asymmetrical walking. The reason for introducing the step ratio is that this
parameter is shown to be reasonably constant for a wide range of walking velocities
and for different subjects [13]. For normal walking, r ranges from 0.39 to 0.44 m·s for
men and from 0.34 to 0.40 m·s for women [7, 15]. For a constant r, S and T are
determined by the forward velocity only:
T =2 r (8)
v
S = 2 r ⋅v (9)
Figure 9. Joint rotations in the sagittal plane for a cycle beginning with heel contact (in degrees). Hip flexion,
knee flexion and ankle dorsiflexion; average values for 19 subjects with measuring deviations. From (Winter,
1991).
4.2. Kinematics
A more detailed description of the walking movement is obtained when the joint
rotations are measured at each point of time of the walking cycle. The hip, knee and
ankle flexion are well documented in the literature, so measured data can be compared
H.F.J.M. Koopman / Dynamics of Human Movement 39
with "standard" data. As an example, the rotations that are shown in Figure 9 are
average values for normal walking, measured by Winter [16].
The other rotations that also contribute to the walking movement (e.g. hip
adduction, pelvic rotations) are less well documented; accepted average values have yet
to be established. The kinematics of the walking movement could also be described
with the displacements of the joints as time functions or with a combination of
displacements and rotations. The choice of one of these possibilities mostly depends on
the measuring system that is available.
Note that a joint angle (e.g. hip angle) describes the orientation of one bone
relative to another bone, or more precisely, the orientation of one local frame with
respect to the other local frame of two connecting rigid bodies. A segment angle, on the
other hand (e.g. the foot angle) describes the orientation of a local frame with respect to
the global frame.
Figure 10. Forward (- - -) and vertical (─) ground reaction forces, normalized with body weight, ensemble
average from 19 subjects. From Winter (1991).
4.3. Kinetics
An alternative description of the walking movement is provided with the internal and
external forces and moments of force, acting on the body segments. Of these forces
only the ground reaction forces can be measured directly. The average ground reaction
forces, as measured by Winter [16], are shown in Figure 10.
Note that the average vertical ground reaction force should equal body weight.
Variations from this average must result from accelerations of the total body centre of
mass, according to equation (4).
The joint moments of force are determined by applying the Newton-Euler
equations of motion (4). This always implies that some modeling assumptions,
concerning the mechanical properties of the human body, have to be made. The joint
moments of force are computed from the measured ground reaction forces and
40 H.F.J.M. Koopman / Dynamics of Human Movement
accelerations. Where the variation in the ground reaction forces between different
subjects is reasonably small, due to the assumptions there may be a larger variation in
the joint torques. Figure 11 shows some of the joint moments of force, as determined
for four different subjects [7]. The joint moments of force can be viewed as the net
result of all muscular, ligament and frictional forces acting on the joint. For example,
the push-off with the calf musculature is clearly visible in the ankle torque of Figure 11.
For some applications, the muscular forces can be related to electromyographic (EMG)
signals, which may validate the modeling assumptions.
Figure 11. Joint moments of force for four different subjects (in Nm).
From Inman et al. (1981).
E w = a + bv 2 (10)
This relation is confirmed by various investigators [7]. When Ew is expressed in
watt per kg body mass and v is in m/s, the experimental values for the constants a and b
are a = 2.24 W/kg and b = 1.26 Hz. Em is defined by
Ew a
Em = = + b⋅v (11)
v v
Ew and Em are shown in Figure 12.
The optimal velocity vopt is defined as the velocity where Em is minimal.
Differentiating equation (11) with respect to v and equating to zero yields vopt = 1.33
H.F.J.M. Koopman / Dynamics of Human Movement 41
m/s. The comfortable walking velocity vcomf is the velocity a person tends to adopt in a
natural walk. vcomf is found to have an average value of 1.39 m/s [7], which differs from
vopt by only 4 %. Inman et al. called this an example of a fundamental feature of human
motor behavior: "In freely chosen rate of activity, a rate is chosen that represents
minimal energy expenditure per unit task". In the case of walking a speed is adopted,
with a specific stride length S and cycle period T, such that each meter is covered as
cheaply as possible. The energy expenditure Em can in this sense be interpreted as
inversely proportional to the efficiency of walking.
In Figure 12 is seen that Em varies only little for a wide range of velocities: The
sensitivity of the efficiency to the velocity is small in this range. When the velocity is
enforced, for example in a treadmill, the stride length, cycle period and joint rotations
can still be chosen freely. The choice is such that the efficiency of the movement is
maximized.
Figure 13. Joint powers (in W/kg). Average values for 19 subjects.
Obtained from Winter (1991).
To be able to tell something about the work done at the joints, usually a distinction
is made between the positive and the negative work (Wj+ and Wj-) at joint j with power
Pj:
T
⎧ P j ( P j > 0)
W j+ = ∫ Pj+ dt; Pj+ = ⎨
0 ⎩ 0 ( P j ≤ 0)
(12)
T 0 ( ≥ 0 )
⎧ P j
W j− = ∫ Pj− dt; Pj− = ⎨
0 ⎩ Pj ( P j < 0)
This definition results in a work balance as shown in table 1. From the table it is
obvious which joints in general generate energy and which joints dissipate energy. In
most cases, the hip joints generate as much as they dissipate, the knee joints are mainly
dissipaters (in the beginning of the stance phase to accept body weight) and the ankle
joints are mainly generators (at the end of the stance phase to push off, also visible in
Figure 12). The total work for all joints during one cycle should equal zero. In table 1,
the total Wj++Wj- equals –2 J/cycle. This value reflects some inaccuracies of the
calculation, especially the fact that the properties of the segments model do not
perfectly match the properties of the human body.
The last column shows the absolute amount of mechanical work done. When it is
assumed that mono-articular muscles do all the work, no energy is stored in passive
structures, and the efficiency equals 1 for concentric contractions and –1 for eccentric
contractions, this number can be related to the metabolic energy consumption. The
metabolic energy consumption for a person of 80 kg with a stride length of 1.5 m/cycle
is about 3.4×80×1.5=408 J/cycle (see Figure 11).
H.F.J.M. Koopman / Dynamics of Human Movement 43
R hip 18 -14 -4 32
L hip 18 -14 -4 32
R ankle 13 -3 10 16
L ankle 13 -3 10 16
About half this amount is needed for basal metabolism, the energy consumption at
zero velocity. This leaves about 200 J/cycle spendet on walking alone. This value is
close to the estimated mechanical work (table 1).
However, if more realistic values for the efficiencies are assumed (i.e. 0.3 and -1.0
for concentric and eccentric contractions respectively, see McMahon [10], the total
mechanical work predicts a metabolic energy consumption of 418 J/cycle. This implies
that energy transfer between joints through biarticular muscles and energy storage in
passive structure are important mechanisms to reduce the over-all metabolic energy
consumption. The work balances are especially useful when analyzing situations that
deviate from normal.
References
[1] R.A. Brand, R.D. Crowninshield, C.E. Wittstock, D.R. Pedersen, C.R. Clark and F.M. van Krieken, A
model of lower extremity muscular anatomy. J. Biomech. Eng. 104 (1982), 304-310.
[2] R.F. Chandler, C.E. Clauser, J.T. McConville, H.M. Reynolds and J.W. Young, Investigation of the
inertial properties of the human body. Report DOT HS-801430, National Technical Information Service,
Springfield Virginia 22151, U.S.A, 1975.
[3] C.E. Clauser, J.T. McConville and J.W. Young , Weight, volume, and center of mass of segments of the
human body. Aerospace Medical Research Laboratory TR-69-70 (AD 710 622), Wright-Patterson Air
Force base, Ohio, 1969.
[4] R. Fick, Handbuch der Anatomie und Mechanik der Gelenke (Handbook of joint anatomy and
mechanics), Gustav Fischer, Jena. 1911
[5] H. Goldstein, Classical mechanics (second edition). Addison Wesley Publishing company, Reading,
Massachusetts. ISBN 0-201-02969-3, 1980.
[6] H. Hatze, Quantitative analysis, synthesis and optimization of human motion. Hum. Movem. Sc. 3
(1981), 5-25.
[7] V.T. Inman, Ralston, H.J., Todd, F., Human walking, Baltimore, Williams & Wilkins, 1981.
[8] B. Koopman, Grootenboer H.J., Jongh H.J. de, An inverse dynamic model for the analysis
reconstruction and prediction of bipedal walking. J. Biomech 28 (1995), 1369-1376.
[9] H.F.J.M. Koopman, The three-dimensional analysis and prediction of human walking, Ph.D.
Dissertation, University of Twente, Enschede, 1989.
[10] T.A. McMahon, Muscles, reflexes, and locomotion. Princeton University Press, Princeton, New Jersey,
1984.
44 H.F.J.M. Koopman / Dynamics of Human Movement
[11] S.J. Olney and D.A. Winter, Predictions of knee and ankle moments of force in walking from EMG and
kinematic data. J. Biomech. 18 (1985), 9-20.
[12] M.G. Pandy and N. Berme, A numerical method for simulating the dynamics of human walking. J
Biomech 21 (1988), 1043-51.
[13] R.H. Rozendal, P.A.J.B.M. Huijing, Y.F. Heerkens, R.D. Woittiez, Inleiding in de kinesiologie van de
mens. (Introduction in the human kinesiology) Culemborg: Educaboek, 1990.
[14] H.J. Ralston, Energy-speed relation and optimal speed during level walking. Int. Zeitschrift für
angewandte Pysiologie 17 (1958), 277.
[15] R.L. Waters, B.R. Lunsford, J. Perry, R. Byrd, Energy-speed relationship of walking: standard tables. J
Orthop Res 6 (1988), 215-22
[16] D.A. Winter, The Biomechanics and motor control of human gait, Waterloo (Canada), University of
Waterloo Press, 1991.
Basic Engineering for Medics and Biologists 45
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-45
Introduction
The performance of a heart valve depends on the flow of blood passing through it.
Cardiac output is the amount of blood pumped by each ventricle in one minute. It is a
function of heart rate (HR), the number of heart beats per minute, and stroke volume
(SV), the amount of blood pumped out by each ventricle per minute. The stroke
volume (SV [ml]), times the number of beats per minute (heart rate, HR), equals the
cardiac output (CO [l/min]; equation 1). Cardiac output in humans varies between 4-6
l/min at rest to 20-35 l/min during exercise, with heart rate variations from 40-70 bpm
at rest to 170-200 bmp during exercise and SV from 60-125 ml during rest to 100-200
ml during exercise. A well trained person develops a larger SV and so CO can be
achieved with a lower HR. [13].
CO = SV ⋅ HR (1)
Changes in either stroke volume or heart rate will alter cardiac output. The SV can
be calculated from velocity measurements over the valve (equation 2):
SV = VTI ⋅A (2)
with VTI, the velocity time integral (VTI [cm]) and A the valve area in [cm2]. A
velocity-time curve is obtained using Doppler ultrasound and the VTI (= ∫ v ⋅ dt ) is the
area under the velocity curve.
46 P. Verdonck and K. Dumont / Biofluid Mechanics & the Circulatory System
1. Pressure Gradient
The resistance of the valve against blood flow can be quantified by pressure/energy
losses over the valve.
A dynamic systolic gradient is found between left ventricular and aortic pressures
(Figure 1a). Assuming proper alignment of the pressure tracings, a series of
instantaneous pressure differences can be measured, and the mean systolic gradient
ΔPmean can be calculated. The gradient ΔPpeak between the peak systolic left
ventricular pressure and the systolic aortic pressure can also be determined, even
though these two peaks are non-simultaneous. Note that the maximum instantaneous
gradient ΔPmax is always larger than the peak-to-peak gradient. Similar gradients can
be determined for the mitral valve during ventricular filling or diastole [16].
The pressure gradient can be calculated using the simplified Bernoulli equation. This
equation is derived from the Bernoulli equation which is based on the assumptions that
flow through the stenosis is laminar and inviscid (viscous forces negligible).
Bernoulli’s hydraulic formula states the conservation of mechanical energy between
two points along a streamline (Figure 1b) [22]; equation 3:
2
1 dv
P1 − P2 = ρ(v 22 − v12 ) + ρ g ( z 2 − z1 ) + ρ∫ ds (3, Bernoulli’s equation)
2 1
dt
where P1 [Pa], v1 [m/s], z1 [m] are the pressure, velocity, and height from a reference
level at the upstream location, and P2 [Pa], v2 [m/s], z2 [m] are the pressure, velocity,
and height at the downstream location. The integral term accounts for the flow
acceleration or deceleration between the two locations. If the equation is applied at
peak systolic velocity, this term becomes zero, because the rate of change of velocity,
the derivative, becomes zero. When further applied on two points along a stream line
located at the same height, one finds that [22]; equation 4:
1
P1 − P2 = ρ (v22 − v12 ) (4)
2
When the downstream velocity is much higher than the upstream velocity
(v2 >> v1), v1 can be neglected. Filling in the value for the density of blood at 37°C
(1060 kg/m3) ρ (rho), and converting pressure from Pa to mmHg (133 Pa = 1
mmHg), results in a factor of about 4, in the so-called simplified Bernoulli’s
equation [22]; equation 5:
b c
where ΔP is in [mmHg] and v2 is in [m/s] [22]. This equation has found wide
application clinically in determining pressure drops across severe stenoses from
noninvasive velocity measurements using Doppler ultrasound. Bernoulli’s equation
is not valid in cases in which viscous forces are significant, such as in long
constrictions with a small diameter, or in flow with separation regions. In addition,
it does not account for turbulent energy losses in cases of severe stenosis. Such
losses should be taken into account, because they reduce the energy content of the
48 P. Verdonck and K. Dumont / Biofluid Mechanics & the Circulatory System
fluid [22]. In normal individuals, there is a very slight (1-2 mmHg) pressure
difference between the left ventricle and aorta during ejection that helps drive the
blood across the aortic valve. In contrast, very high pressure gradients are observed
in patients with a stenotic valve. Patients with a pressure gradient ΔP of 100
mmHg, in the case of severe aortic stenosis, and of 16 mmHg in the case of mitral
stenosis, are referred for clinical assessment..
Results from in vitro experiments by Feng et al [5] are displayed in Figure 1a. The
SJM HP (St. Jude Medical Hemodynamic Plus) and On-X valves produced the lowest
mean positive pressure drop, and the CM (CarboMedics) valve produced the highest
under every condition. The differences in these pressure drops were related to the
geometric orifice diameters and the degrees of valve opening. That the On-X valve
produced the smallest pressure drop is mainly due to its larger internal orifice diameter
and the parallel opening of its leaflets. The SJM HP valve also benefits from a larger
internal orifice diameter and a large opening angle. The ATS valve has a lower mean
positive pressure drop than the CM valve in pulsatile flow, despite the findings that its
maximal opening angle is less than that of the CM valve. Furthermore, it can be
observed from Figure 1a, that the pressure drop over the valves increases with the
cardiac output. In the clinical setting, this maximum pressure gradient cannot be
measured invasively because it is not possible to position a catheter across or between
mechanical prosthetic valve leaflets in patients [18]. Doppler gradients across the
central orifices are significantly higher than the transvalvular and net catheter pressure
gradients measured across the valve (Figure 1b). These differences are due to
downstream pressure recovery (Figure 1c) [6, 18], which gives the pressure at a certain
time along the axis of the valve. The pressure shows a drop, but recovers at a certain
distance downstream the valve, due to conversion of kinetic energy into potential
energy upon deceleration of the flow through the valve.
When blood is flowing through an orifice with cross section area A [m 2], convergence
of streamlines make that only part of the cross section is effectively used for flow
passage. This is particularly important in valve dynamics where the effective valve area
should be maximized to avoid high velocities, turbulence, shear, and associated high
pressure drops. The effective orifice area (EOA) is the standard parameter for the
clinical assessment of the severity of the valve stenosis. It is also a parameter that can
be used to compare performance of different mechanical and biological valve
prostheses.
As demonstrated in Figure 2, the flow through the valve is narrowing, using only
the EOA, i.e. a fraction of the total geometric orifice area (GOA). The effective orifice
area (EOA [cm2]) is calculated - based on the continuity equation - as the ratio of
forward stroke volume and VTI; equation 6:
Besides the continuity equation, the Gorlin equation provides an alternative way to
calculate the effective orifice area (EOAGorlin). With Q the flow rate [m3/s] and v the
maximal velocity [m/s] in the vena contracta (narrowed vessel), the EOA is given by
EOA = Q / [ C1 v ] (7)
where C1 is a coefficient to adjust for the velocity profile in the vena contracta.
Through the orifice, there is a conversion of potential energy to kinetic energy, and this
conversion is described by [22]:
v = C2 2 ⋅ g ⋅ Δh (8)
with the pressure head across the orifice Δh = Δz + ΔP/(ρg) and where C2 is a
coefficient to adjust for loss in the conversion of energy from potential to kinetic
energy. With Δz = 0 and assumption of C1·C2 = 1, the formula can be written as:
Q Q Q
EOA = = = (9)
v ΔP ρ Hg ⋅ g ⋅ ΔhHg
2 2
ρ ρ
The commonly used SI units [m2], [m3/s], [Pa] are converted into clinical used
units [cm2], [ml/s], [mmHg] for the area, the flow and pressure respectively.
Q ⋅ 10 −6
EOA ⋅ 10 − 4 = (12)
ρ Hg ⋅ g ⋅ 10 −3 ⋅ ΔP
2
ρ
Q
EOA Gorlin = (13, Gorlin equation)
51,6 ΔP
with EOA in [cm2], ΔP in [mmHg] and the mean forward flow Qfwd,mean in [ml/s],
calculated as the ratio of the “positive” area under the flow curve (forward stroke
volume SV) divided by the duration of forward flow. EOAGorlin is a fraction θ (theta), of
the available geometric orifice area GOA, so that EOAGorlin = θ ·GOA. The
performance index (PI [no dimension]) of the valve is calculated as the ratio of the
effective orifice area and the geometric orifice area (GOA):
Many clinical and experimental studies are done in order to estimate the EOA and
calculate the performance index PI which enables clinicians and engineers to compare
different types of heart valves. The higher the PI, the better the hydrodynamic
performance is of the valve. This means the contraction of flow through the valve is
minimized and the flow utilizes a high percentage of the available geometric orifice
area. Bech-Hanssen et al. showed similar effective orifice areas in vitro for St Jude
Medical and Omnicarbon valves (Figure 3) [1].
4. Regurgitation
Regurgitation is reversed flow through the ‘one way’ valve. The percentage of
regurgitation (% reg) in relation to the total SV is calculated as
Vreg [ml] is the total volume of regurgitation and can be split into V close and Vleak.
Vclose is the volume of regurgitation due to the closing of the valve. V leak is the volume
of regurgitation due to leakage of the closed valve (Figure 4).
P. Verdonck and K. Dumont / Biofluid Mechanics & the Circulatory System 51
Figure 3. Effective orifice area of two different mechanical valve types [1]. Results are shown for the in vivo
(black o) and in vitro (open o) studies. Bars indicate mean ± SD. Boxes show the geometric orifice area.
(a) Caged ball valve. (b) Tilting disc valve. (c) Bileaflet valve. (d) Bioprosthesis.
Figure 4. Velocity profiles in cm/s based on in vitro measurements on 27 mm aortic valve designs.
Measurements were done downstream the valve on a centerline at peak systole,
at a cardiac output of 6 l/min and a heart rate of 70 beats/min [21].
4.1. Regurgitation
Figure 5 shows regurgitation data from different types of bileaflet valves [5]. As
reported by Wu et al. [20], the closing volume, expressed as a percentage of the
forward flow volume, increased with decreasing cardiac output. The closing volumes
52 P. Verdonck and K. Dumont / Biofluid Mechanics & the Circulatory System
of the valves, shown in Figure 5, are within the acceptable range (<8%) [5]. The SJM
HP valve under low cardiac output condition at 100 bpm showed a somewhat higher
closing volume of 11.4%. Closing volume is believed to be proportional to opening
angle of the mechanical heart valve [5].
Figure 5. Closing volumes of different bileaflet heart valves (ATS: Advancing The Standard, CM:
Carbomedics, SJM HP: St. Jude Medical Hemodynamic Plus, On-X) studied in mitral position in an
experimental setup at 70 beats/min and at 100 beats/min [5].
The Reynolds number (Re) is a quantity that engineers use to assess whether fluid flow
is laminar or turbulent. This is important, because increased mixing and shearing occur
in turbulent flow. This results in increased energy losses which affects the efficiency of
heart valves. A good example of laminar and turbulent flow is the rising smoke from a
cigarette. The smoke initially travels in smooth, straight lines (laminar flow) then starts
to ”wave” back and forth (transition flow) and finally seems to randomly mix
(turbulent flow). The dimensionless Reynolds number Re [-] is defined as:
U ⋅D
Re = ρ (16)
μ
with ρ (rho) the density of the fluid (kg/m3), U the velocity of the flow (m/s), D the
diameter of the vessel (m) and µ (mu) the dynamic viscosity, or internal friction, of the
fluid (mPa·s). Figure 6 demonstrates the different flow regimes in function of Re for
the flow in a straight tube. Generally, a fluid flow is laminar in a stiff tube from Re = 0
P. Verdonck and K. Dumont / Biofluid Mechanics & the Circulatory System 53
to some critical value (Re < 2000) at which transition flow begins. Transition flow is
fluctuating.
Figure 6. Diagram of flow regimes in pipe flow. between laminar and turbulent flow (2000 < Re < 4000).
Fluid flow becomes unstable for higher Reynolds numbers (Re > 4000). In
turbulent flow, there is increased mixing that results in viscous losses, which are
generally much higher than those in laminar flow [19]. The Reynolds number can reach
up to 4500 at peak flow in the normal aortic valve. The Strouhal number gives an
indication of a steady or transient regime of the flow. The Strouhal number is defined
as
D
Sr = (17)
Tp ⋅U
with Tp the period of time of the observed flow phenomenon. A very small Strouhal
number (Sr << 1) is considered quasi-steady. For a Strouhal number close to one, the
flow is considered transient.
As blood flows through a vessel or valve, it exerts a physical force on the vessel wall or
the valve leaflet. This force can be resolved into two principal vectors (Figure 7). Shear
stress, being tangential to the wall, represents the frictional force exerted by the flowing
blood on the endothelial surface of the wall. The shear stress on the vessel wall or on
the valve leaflet is normally called wall shear stress or WSS. Normal stress, or pressure,
is perpendicular to the wall. In the case of laminar flow shear stress is calculated with
equation 18.
r
r ∂U
τ lamin ar = μ = μ ⋅ γ& [Pa] or [N/m2] (18)
∂y
with τ (tau), the shear stress, μ (mu) the dynamic viscosity of the fluid, U the velocity
vector and γ& the shear rate. Turbulence in the blood system increases resistance to flow,
resulting in higher pressure gradients. Turbulent shear stresses are calculated in a
slightly different manner, as shown in equation 19 [3]:
r
r ∂U
τ turbulent = μ − ρu ′v ′ (19)
∂y
with u’ and v’ the turbulent fluctuations of the velocities u and v respectively. The
vascular and valvular endothelial cells are subjected at all times to shear forces that act
on the vessel surface as a result of the flow of viscous blood. Fluid shear stress
regulates endothelial phenotype by altering its gene expression profile, including
growth factors [11,12]. Fluid shear stress transforms polygonal, cobblestoneshaped
endothelial cells of random orientation (Figure 8a) into fusiform endothelial cells
aligned in the direction of flow (Figure 8b). High shear stress (> 40 N/m 2) can cause
direct endothelial injury [11]. Low and oscillating shear stress regions (< 0.4 N/m 2) can
lead to atherosclerotic lesions or plaques, mainly at arterial bifurcations and in the
coronary arteries [11].
Figure 8. Transformation of endothelial cell morphology by fluid shear stress: aortic endothelial cells
exposed to low shear stress are randomly oriented (a), while those exposed to physiological shear stress (1.5
N/m2, right panel) for 24 hours align in the direction of blood flow [11,12]
P. Verdonck and K. Dumont / Biofluid Mechanics & the Circulatory System 55
High shear stresses in the blood may create platelet activation [10,15] leading to
thrombosis [23], and the subsequent risk of embolism. Blood platelet damage starts to
occur at shear stress values of 10 N/m2 [8]. Furthermore, the magnitude, exposure time,
and spatial distribution of the shear stresses coincide with damage to red blood cells.
Shear stresses above 200 N/m2 will cause hemolysis [2,4,14]. Hemolysis can also
occur at lower shear stress values if there is a long exposure time [3,7]. Therefore
recirculation zones in prosthetic heart valves are to be avoided, unless they are washed
out by the next heart beat [9], so that the residence time of particles remains limited. In
vitro studies of aortic prosthetic heart valve designs have revealed shear stress values
sufficiently high to cause lethal or sublethal damage to blood cells [17,24]. Shear stress
cannot be measured directly in vitro nor in vivo on a moving heart valve leaflet. For
that reason computational fluid dynamics is used to estimate the shear stress.
Natural heart valves allow blood to flow straight through the center of the valve. This
property, known as central flow, minimises the work done by the heart. With non-
central flow, the heart works harder to compensate for momentum lost to change in
direction of the fluid. Caged ball valves completely block central flow (Figure 9a);
blood therefore requires more energy to flow around the occluder. In addition,
collisions with the occluder ball caused damage to blood cells. Caged-ball valves are
also notorious for stimulating thrombosis, requiring patients to take lifelong
prescriptions of anticoagulants.
High turbulent shear stresses are observed at the edges of the jet (Fig 9a). The
intensity of turbulence during peak systole does not decay very rapidly downstream of
the valve. Elevated shear stresses occurred during most of systole [21]. This type of
valve is no longer used clinically.
This valve design allows more central flow while still preventing backflow. However,
flow is not completely central (Figure 9b), because blood has to make a significant
directional change to flow around the occluder disc. Newer models of monoleaflet
valves improve this shortcoming. In the major orifice region (Figure 9b), high turbulent
shear stresses are confined to narrow regions at the edges of the major orifice jet.
During the acceleration and deceleration phases, the turbulent shear stresses are
relatively low [21]. High turbulent shear stresses are more dispersed in the minor
orifice than those in the major orifice regions (Figure 9b)
These valves provide the closest approximation to central flow yet achieved in a
mechanical heart valve (Figure 9c). The two leaflets block flow to some extent, leading
to a three-part flow pattern. High turbulent shear stresses occur at locations
immediately distal to the valve leaflets (Figure 9c).
56 P. Verdonck and K. Dumont / Biofluid Mechanics & the Circulatory System
(a) Caged ball valve. (b) Tilting disc valve. (c) Bileaflet valve. (d) Bioprosthesis.
Figure 9. Turbulent shear stress profiles in 0.1·N/m2 based on in vitro measurements on 27 mm aortic valve
designs. Measurements were taken downstream of the valve on a centerline at peak systole, at a cardiac
output of 6 l/min and a heart rate of 70 beats/min [21].
Biological valves, optimally mimic native heart valves and provide central flow (Figure
9d). Turbulent shear stress measurements during the deceleration phase show low shear
stresses spread out over a wide region [21]. During peak systole, the high turbulent
shear stresses were confined in a narrow region (Figure 9d).
7. Conclusions
References
[1] O. Bech-Hanssen, K. Caidahl, I. Wallentin, P. Ask and B. Wranne, Assessment of effective orifice area
of prosthetic aortic valves with doppler echocardiography: An in vivo and in vitro study, The Journal of
Thoracic and Cardiovascular Surgery 122(2), (2001) 287–295.
[2] P. Blackshear, F. Dorman and E. Steinbach, Shear, wall interaction and hemolysis, Transactions of the
American Society of Artificial Internal Organs 12, (1966) 113–120.
[3] D. Bluestein, Y. Li and I. Krukenkamp, Free emboli formation in the wake of bi-leaflet mechanical
heart valves and the effects of implantation techniques, Journal of Biomechanics 35(12), (2002), 1533–
1540.
[4] J.T. Ellis, T.M. Wick and A.P. Yoganathan, Prosthesis-induced hemolysis: mechanisms and
quantification of shear stress, Journal of Heart Valve Disease 7(4), (1998) 376–386.
P. Verdonck and K. Dumont / Biofluid Mechanics & the Circulatory System 57
[5] Z. Feng, T. Nakamura, T. Fujimoto and M. Umezu, In vitro investigation of opening behavior and
hydrodynamics of bileaflet valves in the mitral position, Artificial Organs 26 (2002), 32–39.
[6] D. Garcia, J. Dumesnil, L.G. Durand, L. Kadem and P. Pibarot, Discrepancies between catheter and
doppler estimates of valve effective orifice area can be predicted from the pressure recovery
phenomenon: Practical implications with regard to quantification of aortic stenosis severity, Journal of
the American College of Cardiology (JACC) 41 (2003), 435–442.
[7] L. Goubergrits and K. Affeld, Numerical estimation of blood damage in artificial organs, Artificial
Organs 28(5), (2004) 449–507.
[8] J. Hellums and R. Hardwick, The Rheology of Blood, Blood Vessels and Associated Tissues, chapter
Response of Platelets to Shear Stress - a Review, 160–183, Sijthoff & Noordhoff, Alphen aan den Rijn,
1981.
[9] S. Kelly, P. Verdonck, J. Vierendeels, K. Riemslagh, E. Dick and G.V. Nooten, A threedimensional
analysis of flow in the pivot regions of an ATS bileaflet valve, International Journal of Artificial
Organs 22(11), (1999) 754–763.
[10] M. Kroll, J. Hellums, L. McIntire, A. Schafer and J. Moake, (1996) Platelets and shear stress, Blood
88(5), (1996) 1525–1541.
[11] A.M. Malek, S.L. Alper and S. Izumo, Hemodynamic shear stress and its role in atherosclerosis,
Journal of the American Medical Association (JAMA) 282(21), (1999) 2035–2042.
[12] A. Malek, and S. Izumo, Mechanism of endothelial cell shape change and cytoskeletal remodeling in
response to fluid shear stress, Journal of Cell Science 109, (1996) 713–726.
[13] K. Matthys, Assessment of Vascular Hemodynamics: Investigation of non-invasive and minimally
invasive methods for assessment of vascular function at rest and during cardiovascular challenge, Ph.D.
Dissertation, Ghent University, 2004.
[14] A. Nevaril, E. Lynch, C. Alfrey and J. Hellums, Erythrocyte damage and destruction induced by
shearing stress, Journal of Laboratory and Clinical Medicine 71(5), (1968) 781–790.
[15] Z. Ruggeri, Mechanisms of shear-induced platelet adhesion and aggregation., Thrombosis and
Haemostasis 70 (1), (1993) 119-123.
[16] K. Schmailz and O. Ormerod, ed., Ultrasound in Cardiology, Blackwell Science, 1998.
[17] W. Tillmann, H. Reul, M. Herold, K. Bruss and J. van Gilse, In vitro wall shear measurements in aortic
valve prostheses, Journal of Biomechanics 17(4), (1984) 263–279.
[18] P.M. Vandervoort, N.L. Greenberg, M. Pu, K.A. Powell, D.M. Cosgrove and J.D. Thomas, Pressure
recovery in bileaflet heart valve prostheses: localized high velocities and gradients in central and side
orifices with implications for doppler-catheter gradient relation in aortic and mitral position,
Circulation 92 (1995), 3464–3472.
[19] J. Welty, C. Wicks, R. Wilson and G. Rorrer, Fundamentals of momentum, heat, and mass transfer,
Wiley Text Books; 4 edition, ISBN 0471381497, 2000.
[20] Z. Wu, B. Gao, J. Slonin. and N. Huang, N., Bileaflet mechanical heart valves at low cardiac output,
ASAIO Journal 42(5), (1996) 747–749.
[21] A.P. Yoganathan, The Biomedical Engineering Handbook, Volume I: Second Edition., chapter Cardiac
Valve Prostheses, CRC Press LLC, ISBN 0-849-38594-6, 2000.
[22] A.P. Yoganathan and G. Chatzimavroudis, PanVascular Medicine: Integrated Clinical Management,
Chapter 7: Hemodynamics, Springer, 2002.
[23] A.P. Yoganathan, T. Wick and H. Reul, H., Current issues in heart valve disease, chapter Thrombosis,
Embolism and Bleeding, London: ICR, 1992.
[24] A.P. Yoganathan, Y.R. Woo and H.W. Sung, Turbulent shear stress measurements in aortic valve
prostheses, Journal of Biomechanics 19(6), (1986) 433–442.
58 Basic Engineering for Medics and Biologists
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-58
Introduction
Human life expectancy is increasing and so too, therefore, is the need for joint
replacements to maintain mobility. However, the functional life expectancy of the
implants is limited to 10-20 years, thus in many patients the original implant must be
removed and replaced. One of the reasons for the limited lifetime of these devices is
the fact that bone adapts its shape to a change in load distribution as the geometry of an
artificial joint differs from the original, particularly in its stem.. As the bone adapst its
shape, the fixation of implants is affected resulting in movement of the artificial joint,
malfunction and pain, making revision surgery necessary. Such revision surgery is
difficult and dangerous, so an increase in the lifetime of joint prostheses is preferable.
To improve those implants, an analysis of the loading situation and response of the
implant is required However, the geometry of such an implant and the surrounding
bone is complicated, and bone is a complex material, with non-homogenous properties.
The Finite Element Analysis (FEA), which originates from the aerospace and nuclear
industries, offers a solution, Finite element packages avail of advanced imaging
J.G. Hazenberg et al. / Biomechanics of Implants 59
techniques and are capable of solving the most sophisticated problems, not just in
structural analysis, but for a wide range of phenomena such as static and dynamic
loading of orthopaedic implants and fluid flow problems in cardiovascular applications.
Its popularity is due to a number of factors, which include:
• A variety of loading conditions and environmental parameters can be analysed
• A variety of materials can be analysed and tested under in vivo circumstances
• Geometrical modifications are easily made
• It reduces the number of animal and pre-clinical tests required
• It is cost and time efficient
1.1. Geometry
-20° 0° 20°
A B C
Figure 1. From the MRI or μCT scan a model is created of a mandible (A). The data points are
connected through lines and a solid model is created (B). Finally the solid model is meshed (C).
To carry out a finite element analysis, the model must be divided into a number of
small pieces known as finite elements. In simple terms, a mathematical net or "mesh" is
required to carry out a finite element analysis. If the system is one dimensional, it can
be represented by line elements, if it is two dimensional, then a 2D mesh is required.
60 J.G. Hazenberg et al. / Biomechanics of Implants
The next step is to give the finite elements mechanical properties such as Young’s
modulus, which determines its elasticity. The final step is to place a load (or different
loads) at one or more nodes, or intersection points of the mesh, and to fix other nodes
to the environment.
Now the load on all elements will be computed and can be compared to the
maximal allowable load before failure. If this load is exceeded in some element, we
know that the structure will collapse.
If the maximal load is not exceeded, the resulting deformation on every element
can be calculated and, by adding all deformations, the total deformation of the
complete structure is found..
In more detail, the finite element is a mathematical method for solving ordinary
and partial differential equations. As these types of equations occur naturally in
virtually all fields of the physical sciences, the applications of the finite element
method are limitless as regards the solution of practical design problems. Within each
element, the variation of displacement is assumed to be determined by simple
olynomial shape functions and nodal displacements. Equations for the strains and
stresses are developed in terms of the unknown nodal displacements. From this, the
equations of equilibrium are assembled in a matrix form, which can easily be
programmed and solved on a computer. After applying the appropriate boundary
conditions, the nodal displacements are found by solving the matrix stiffness equations.
Once the nodal displacements are known, element stresses and strains can be calculated.
From the elements and their nodes the 'stiffness' matrix is formulated. This square
matrix contains details of the material properties, the model geometry and any
assumptions of the stress-strain field.
Figure 2. Two these nodes we can apply forces, fi and fj respectively, resulting in a
displacement ui and uj. The displacement of these two nodes will be dependant on the
stiffness of the spring ‘k’. In any case where linear elastic material properties can be
assumed, the displacement u increases proportionally with the force f (if k is constant).
In order for this spring to be in a state of equilibrium, the forces fi and fj have to be
of equal magnitude in opposite directions. Mathematically we can say that:
⎧ fi ⎫ ⎡ k − k ⎤ ⎧ ui ⎫
⎨ ⎬=⎢ ⎥•⎨ ⎬
⎩ f j ⎭ ⎣− k k ⎦ ⎩u j ⎭ (3)
This is similar to F = K ⋅ u in this case. Here F is the matrix of the load vectors,
derived from the loads of every cases, u is the matrix of the displacement of nodes and
K is the coefficient matrix of the system, the so called stiffness matrix. Since FEA does
not use one single element but several, we can do this in the same way for a spring
assembly.
1 k1 2 k2 3
In Figure 3 we have two elements; one is connected by nodes 1 and 2, while the
second element is connected through the elements 2 and 3. This can be expressed in
matrix form as follows:
⎧⎪ f 11 ⎫⎪ ⎡ k 1 − k 1 ⎤ ⎧ u1 ⎫
⎨ 1⎬ = ⎢ ⎥ • ⎨ ⎬ for element 1 and (4)
⎪⎩ f 2 ⎪⎭ ⎣− k1 k1 ⎦ ⎩u 2 ⎭
⎧ f 12 ⎫ ⎡ k 2 − k 2 ⎤ ⎧u 2 ⎫
⎨ 2⎬ = ⎢ • ⎨ ⎬ for element 2. (5)
⎩ f 2 ⎭ ⎣− k 2 k 2 ⎥⎦ ⎩u 3 ⎭
62 J.G. Hazenberg et al. / Biomechanics of Implants
⎧ F1 ⎫ ⎡ k1 − k1 0 ⎤ ⎧ u1 ⎫
⎪ ⎪ ⎢ ⎥ ⎪ ⎪
⎨F2 ⎬ = ⎢− k1 k1 + k 2 − k 2 ⎥ • ⎨u 2 ⎬ (6)
⎪F ⎪ ⎢ 0 − k2 k 2 ⎥⎦ ⎪⎩u 3 ⎪⎭
⎩ 3⎭ ⎣
Figure 4. On the left-hand side (A) an image of a free-body-diagram in which P is the posterior part of
temporalis, M1 is the superficial part of masseter, M2 is the deep part of masseter, Pm is the medial
pterygoid and J is the lateral pterygoid muscle. On the right-hand (B) side is an image of the forces applied to
the meshed model of the mandible.
J.G. Hazenberg et al. / Biomechanics of Implants 63
but with large displacements (in the analysis of muscles). Muscles contractions are also
a good example where nonlinear forces are applicable. This is where the magnitude or
direction of the applied forces change with application to the structure. Displacement
boundary condition nonlinearities, such as occurs in modelling of friction in knee joints.
This is where the displacement boundary conditions depend on the deformation of the
structure. The most important and obvious application is in contact problems, the
displacement is highly dependent on the relationship between two contact surfaces (e.g.
a fracture site in bones). It is important to note that the bodies in contact could be in a
state of linear elastic stress; the nonlinearities all come from the contact definition in
this case.
All non-linearities are solved by applying the load slowly (dividing it into a
number of small load increments). The model is assumed to behave linearly for each
load increment, and the change in model shape is calculated at each increment. Stresses
are updated from increment to increment, until the full applied load is reached. In a
nonlinear analysis, an initial condition at the start of each increment is the state of the
model at the end of the previous one. This dependency provides a convenient method
for following complex loading histories. At each increment, the solver iterates for
equilibrium. Due to the iterative nature of the calculations, non-linear FEA is
computationally expensive, but reflects the real life conditions more accurately than
linear analyses.
The previous section has provided some of the background regarding FEA. But why is
it commonly used in the analysis of implant biomechanics and implant design? These
days it is very easy to obtain accurate models of various body parts of interest using
MRI and CT-scanning. This means that irregular shaped objects are easy to analyse.
Anatomical studies can be used as a great source of information regarding muscle
attachments, insertions and directions. Analysis can be done for static as well as for
dynamic loading conditions. All these features contribute to more sophisticated models
that mimic in vivo situations in a more realistic way. But the true power of FEA lies in
the ability to change parameters relatively easily once the model has been created.
In the case of the analysis of a fractured mandible, we can choose a specific
material for the plate that stabilises the fracture site. The analysis will indicate what the
local stress or strains are in the tissue. Perhaps one wants to investigate what happens if
a softer and more flexible material is used. This would only require altering the
material properties and rerunning the analysis. The same applies if the location and size
of the plate or the fracture angle is to be investigated. This can provide valuable
information, even at tissue level. The phenomenon of stress shielding is a common
problem in implant design, which can result in tissue loss or altered structure. Typically
a tissue is stimulated by the daily loads to which it is subjected. However, following the
insertion of an implant, this load is partly reduced since it is shared by the implant.
Factors that are of importance here are the geometry and the fit of the implant with the
natural structure. Interface conditions between the implant and its surrounding tissue
might be designed to promote bounding between them and therefore realise a better
stress distribution from the implant to the surrounding tissue and so increase the
lifetime of the implant.
J.G. Hazenberg et al. / Biomechanics of Implants 65
References
[1] T.M. Wright and S.B. Goodman, Implant Wear in Total Joint Replacement: Clinical and Biological
Issues, Material and Design Considerations, American Academy of Orthopaedic Surgeons, 2002
[2] P.J. Prendergast, Finite element models in tissue mechanics and orthopaedic implant design,. Clin
Biomech (Bristol, Avon) 12 (1997),.343-366.
This page intentionally left blank
Basic Engineering for Medics and Biologists 67
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-67
Introduction to
Chapter II. Bioelectronics
Richard B. REILLY and T. Clive LEE (eds.)
Abstract. The paper describes the physical phenomena involved in the conduction
of electricity, with particular reference to living tissue. The conduction of
electricity depends on the flow of charge carriers in the material, while the
dielectric properties are due to the rotation of dipoles that can align along an
applied electric field. The relation between the electric variables in a conducting
medium and their physical meanings are explained. The phenomena responsible
for the electric and dielectric properties of living tissues are described. The
presence of cells limits the flow of charge carriers, in particular at low frequency,
and the membranes are responsible for dielectric relaxation. The passive response
of cell membranes to weak applied signals enables bioelectrical tissue
characterisation. Practical tools for electrical impedance spectroscopy are given
with an overview of the most recent applications.
1. Definitions
The electric field is a property of space in the presence of an electric charge. A point
charge q0 produces an electric field which is a vector, denoted E , passing by the point
charge and verifying the following equation
70 J. Jossinet / Elementary Electrodynamics
(
E = q 0 / 4π ε 0 d 2 ) (1)
The quantity ε 0 is the dielectric permittivity of free space and d is the distance
to the source. The major property of the electric field is the force, f = q E , it exerts on
a charge. Two charges exert opposite forces on each other. The electric force is
attractive when the charges have opposite signs and repulsive when the charges have
identical signs.
Voltage or electric potential, u, at a given point is the electric potential energy per unit
charge (in joules per coulomb = volts). The definition of potential is generally
presented under the form: E = − grad ( v ) , where grad is the vector operator gradient.
This vector represents the magnitude and direction of the greatest rate of change of the
potential. This definition indicates that the existence of electric field in a medium
implies a non uniform potential distribution. Conversely, if the potential is uniform, the
field is equal to zero at any point (e.g. in a conductor at electric equilibrium).
i = ∂q / ∂t (2)
It is often useful to consider the flow of electric charge per unit surface. This quantity is
a vector termed current density, j (A/m2). The current density through a surface
element of area dS, is given by:
j = n.di / dS (3)
where n is the unity vector perpendicular to the surface. The dot product shows that
only the orthogonal component of the flow gives rise to a current density through the
surface.
The current density determines the effects of electricity in a medium. A practical
example is the heating of a wire by the Joule effect. For a given current, the heating is
greater in a conductor of small cross section than in a conductor of larger cross section.
In body tissues, the desired or adverse effects of electrical current depend on how the
current concentrates in certain pathways or spreads in a larger section, resulting in
different current densities.
J. Jossinet / Elementary Electrodynamics 71
σ 0 = F ∑ γ i α i C i ni μ i (4)
i
The activity coefficient, γ i , ( γ ≤ 1) accounts for the deviation from linear behaviour
for increasing ionic concentrations. For a given species, Ci is the concentration of this
species in the solution, ni the number of charges of the ion, and μ i is the mobility and
Σ denotes summation. The above equation deals with the static conductivity for direct
current. However, the viscous forces delay the movement of charges carriers and the
establishment of the steady state is not instantaneous. This effect however becomes
appreciable for rapidly varying fields. Hence, in biomedical applications the
conductivity of ionic solutions can be considered independent of the applied field
frequency.
Ohm's law relates to the proportionality between the DC current, denoted I, and the DC
voltage difference, denoted V, across an ideal conductor, circuit or current pathway.
The general expression of Ohm's law is
V = R× I (5)
E = J / σ = ρJ (6)
dW = J2 / σ (7)
(where σ is the electric conductivity and J the current density). This equation shows
that in a heterogeneous medium, the local Joule effect depends on the conductivity and
current density at each point.
In a circuit passed by a sinusoidal signal, the voltage and the current vary in synchrony,
but are shifted in time, according to the properties of the circuit. The voltage-to-current
ratio is termed impedance.
A sinusoidal signal (voltage or current) is characterised by its magnitude and phase
angle and is represented by a complex number formed by a real part and an imaginary
part.
It is expressed as a complex number denoted Z. The real part and the imaginary
part of Z are denoted R and X respectively.
The magnitude |Z| of this complex number is the ratio of the magnitudes (V/I, V in
volt and I in ampere).The argument, θ , is the phase difference (in radians) between
voltage and current. Simple mathematical considerations lead to the following classic
equations.
( )
Z = R 2 + X 2 , R = R Z cos θ and X = Z sin θ (8a)
The real part, R, of the impedance is termed resistance and its imaginary part, X, is
termed its reactance. Impedance is the generalisation to sinusoidal signals for the
notion of resistance for DC current. The reciprocal of impedance is also a complex
number denoted Y ( Y , θ' ), termed admittance, and verifying:
M = q × Q − Q+ (9)
The dipoles in a dielectric medium cannot move but can rotate in the presence of an
electric field. The two charges experience opposite forces. It can be shown that these
forces create a torque tending to align the dipole along the direction of the electric
field.
The rotation of dipoles creates the dielectric polarisation of the material. Near the
boundary surface, the orientation of dipoles makes charges of a given polarity come
closer to the surface and the charges of the opposite polarity move away from it. This
results in a distribution of charges at the outer surface of the material. This distribution
depends on the concentration of dipoles and their ability to rotate. A dielectric material
is characterised either by its dielectric permittivity ε (farad/metre) or by its "dielectric
constant" ε r = ε ε 0 (dimensionless).
S
C = ε0ε (10)
d
S is the area of charged surfaces, d is the distance between them and ε is the dielectric
constant of the material. The main feature of a capacitor is to block DC signals and be
permeable to varying signals.
74 J. Jossinet / Elementary Electrodynamics
Δε
ε(ω) = ε ∞ + (11)
1 + jωτ
The conduction of electricity in the human body presents a particular set of features
that is not met is other materials, except in animal tissues and, to some extent in plants.
The main feature is the simultaneous existence of electric conduction and dielectric
behaviour. The presence of cells is responsible for the remarkable electric and
dielectric properties of tissue.
The cell is the basic building block of the human body. The living cell is delimited by
the plasma membrane providing rigidity and protection. The cell membrane also binds
different cells together. The membrane consists of an asymmetric molecular lipid bi-
layer which is basically impermeable to ions. However, protein molecules inserted in
the bi-layer ensure the transport of materials, including ions, across the membrane. The
interior of the cell contains the nucleus and the cytoplasm which is a jelly-like material
consisting of the cytosol and the organelles. The cytosol is made up of water, salts, and
organic molecules and is therefore conductive. The content and structure of the interior
of a cell depend of each cell type.
the flow produced by the ionic concentration gradient. This potential difference is
called the Nernst potential. The resting value for nerve cells is about 70 mV.
The overall conduction of electric signals through a tissue depends on the conductivity
of the interstitial space, that of the intracellular medium and the capacitive behaviour of
the cell membranes. The membrane is practically impermeable to DC current, except
for the active transport of materials through it by the trans-membrane proteins. For
applied signals, the membrane acts like a passive insulator separating two conducting
media and causes capacitive coupling between the interstitial space and the interior of
the cell. The capacitance of the cell membrane is about 1 µF/cm2 [7, 12]. Finally, DC
and low frequency signals do not penetrate into cells. For increasing applied signal
frequency, the interior of cells is progressively involved in the overall conduction due
to capacitive coupling through the membrane.
Observation shows that body tissues are dispersive media. In a dispersive dielectric
medium, the permittivity and dielectric constant vary with the applied field frequency.
In general, the dielectric relaxation in body tissue does not follow in the Debye
equation. Experiments show that, in the frequency range from a few hertz to many
gigahertz (1 GHz = 109 Hz), the dielectric permittivity of tissues is a decreasing
function of frequency and shows several dispersion domains. Several types of
phenomena, preponderant in different frequency domains, are responsible for the
observed dispersion [6, 14, 15].
2.5.1. α -Relaxation
This relaxation occurs at low frequency roughly below 1 Hz. This effect is due to the
counter-ion atmosphere surrounding the membrane surface. In the presence of an
electric field, the ions in the cell vicinity accumulate on diametrically opposed
locations and then form a cell-size dipole. Such a dipole cannot follow rapidly varying
76 J. Jossinet / Elementary Electrodynamics
signals. This low frequency relaxation does not give information on either cell
membrane or cellular content.
2.5.2. β -Relaxation
This relaxation is due to the orientation of dipoles. Several types of dipoles contribute
to the polarisation of the medium. It takes place typically in the microwave frequency
domain (f < 1 GHz). The permanent dipoles include dipolar molecules (mainly H2O),
hydrogen ionic or covalent bonding. Induced dipoles result from the temporary
separation from the barycentre (the point between two objects where they balance each
other) of positive and negative charges in non-polar molecules by either neighbouring
molecules or the applied field. Transient dipoles (also called London forces or van der
Waals forces) result from the random mutual perturbation of electron clouds of two
atoms or molecules. When an uneven distribution occurs, a temporary dipole is created.
In polar molecules, transient dipoles are usually smaller than the permanent dipolar
moment.
The above relaxation phenomena give tissues high dielectric constants compared
to usual materials (Table 1).
An equivalent circuit model (ECM) is a circuit yielding the same impedance as a tissue
in a given frequency domain. There are lumped circuits and discrete element circuit
models. The most commonly used discrete element ECM are composed of two resistors
and one capacitive element. The response of such circuits is characterised by a single
central frequency; denoted f0 according to the general equation [10, 11]:
R0 − R∞
Z = R∞ + (12)
1 + j ( f / f 0 )β
Z denotes the overall impedance of the circuit, R0 the low frequency limit resistance, R∞
the high frequency limit resistance and j the base of the complex numbers (j2 = -1).
Exponent β is termed the "fractional power" ( β ≤ 1). At frequency f0 the magnitude
of the imaginary part passes by a maximum. The plot in the complex plane of the
imaginary part against the real part (termed an Argand diagram) is an arc of circle the
centre of which is shifted from the horizontal axis (Fig. 1).
J. Jossinet / Elementary Electrodynamics 77
Table 1. Example values of electric conductivity (S.m-1) and dielectric constant in tissues and some usual
materials at 50 Hz or "below 100 Hz" [2]. The above figures are supplied for illustration and different values
can be found according to measurement frequencies and experimental conditions [12].
Figure 1. Sketch of Eq. (13) in the complex plane: Horizontal axis is the resistance and vertical axis is the
reactance. The reactance of body tissues is negative. The white dots represent particular measurement
frequency points. The abscissa of the intercepts with the horizontal axis are the so-called “limit resistances”
at low frequency ( R0 ) and at high frequency ( R∞ ). The bottom of the arc corresponds to f 0 in Eq. (13).
The shift of the centre from the horizontal axis is a function of the fractional power β.
78 J. Jossinet / Elementary Electrodynamics
There are two types of 3-elements circuits giving rise to the above response in Fig.
2. In these circuits, Zcpa is a "pseudo-capacitance" (subscript "cpa" stands for "constant
phase angle"). Its frequency response is given by Zcpa = K ( j ω )−β . The exponent β is
the fractional power in (12). A constant phase element is a hypothetical element that
does not exist in practice, but is useful to describe the frequency response of body
tissues. If the particular case where β is equal to unity, Zcpa is equivalent to a capacitor
of capacity C=1/K. The relation between ECMs' elements and the variables in (12) are
given in Table 2.
Figure 2. Equivalent circuit models comprising 3-element and yielding identical frequency responses. Type
(a) circuit is generally used for the deep tissue, with Re representing the resistance of the interstitial space
and Ri the intracellular resistance. Type (b) circuit is generally used to represent the impedance of layered
media such as skin.
In practice, fitting an arc of circle to the experimental data points enables the
calculation of R0, R∞, and β (Table 2). The calculation of these parameters makes it
possible to compare data sets obtained at different measurement frequencies. Table 3
gives the equations enabling the calculation of the elements of ECM using the four
parameters resulting from the fit of a circular arc to the experimental data. The table
confirms that, with appropriate values of their elements, both circuits can model a
given response (same R0, R∞, K and β ).
Table 2: Equations for R0, R∞ in equation (12) for the equivalent circuit models of Fig. 2.
ECM R0 R∞ FT
1β
R0 = Re R∞ = Re Ri (Re+ Ri ) 1 ⎛⎜ R0 − R∞ ⎞
⎟
(a) FT= K
2π ⎜⎝ R02 ⎟
⎠
1β
R0 = Rp + Rs R∞ = Rs 1 ⎛ K ⎞
(b) FT= ⎜ ⎟⎟
2π ⎜⎝ R0 − R∞ ⎠
J. Jossinet / Elementary Electrodynamics 79
Table 3: Equations to calculate the elements of the equivalent circuit models of Fig. 2a and Fig. 2b from the
variable in Equ. (12) Quantity Xc denotes the ordinate of the centre of the circular arc. Centre's abscissa is
equal to (R0+R∞)/2.
ECM β K
Ri = 2 ⎛ R − R∞ ⎞
(a) Re = R0 atan⎜⎜ 0 ⎟ (R0 − R∞ )(2πFT )β
R0 R∞ (R0 − R∞ ) π ⎝ 2X C
⎟
⎠
Rp = 2 ⎛ R − R∞ ⎞
(b) Rs = R∞ atan⎜⎜ 0 ⎟
⎟ (R0 − R∞ )(2πFT )β
R0 − R∞ π ⎝ 2X C ⎠
3. Applications
The measurement of the electric and dielectric properties of tissues has given rise to
range of biomedical applications [1, 8]. The most widespread are EIS, EIT and BIA.
Electrical Impedance spectroscopy (EIS) is the study of the frequency response of a
tissue or an organ to characterise its physiological and/or pathological state (e.g.
detection of oedema, ischaemia, inflammation, and tumour growth) [7]. Electrical
Impedance Tomography is an imaging method that appeared in the 1980s. The purpose
is the reconstruction of the distribution of conductivity inside the body from multiple
surface measurements. The main features of this method are the relatively small
number of measurements, limited spatial resolution, high time resolution, real-time
imaging capability and true3-D image reconstruction. This method can be associated
with EIS [3, 4, 5, 9].
Impedance measurements at the body level or across limb segments are termed
Bio-Impedance Analysis (BIA) [13]. The purpose is to monitor the changes in fat free
mass and/or total body water. This method has given rise to devices for various
applications including fitness, sports and nutrition. One popular implementation is the
incorporation of BIA system in personal weighing scales to monitor the effects of
exercise and/or nutritional regimes.
The above applications involve measurements using low intensity currents (of the
order of milliamps or less) avoiding any significant interaction with the explored tissue
at the used measurement frequencies. Therapeutic devices are based on the interaction
of relatively strong signals with the body. Typical applications include electrosurgery,
transcutaneous electric stimulation of muscles or nerve (TENS), cardiac pacing and
defibrillation.
80 J. Jossinet / Elementary Electrodynamics
References
[1] J. Ackman and M. Seitz, Methods of complex impedance measurements in biologic tissue, CRC Crit.
Rev. in Biomed. Eng. 11(1984), 281-311.
[2] D. Andreuccetti, R. Fossi and C. Petrucci, Internet resource for the calculation of Dielectric Properties
of Body Tissues, Italian National Research Council, Institute for Applied Physics "Nello Carrara",
Florence.
[3] K. Boone, D. Barber and B.H. Brown, Imaging with electricity: report of the European concerted
action on impedance tomography, J. Med .Eng. Technol. 21(1997), 201-232.
[4] S.G. Dawids, Evaluation of applied potential tomography: a clinician's view, Clin. Phys. Physiol.
Meas., 8A(1987), 175-180.
[5] A.M. Dijkstra, B.H. Brown et al., Clinical applications of electrical impedance tomography, J. Med.
Eng. Technol. 17(1993), 89-98.
[6] K.R. Foster and H.P. Schwan1, Dielectric properties of tissues and biological materials: a critical
review, CRC Critical Reviews in Biomedical Engineering, 17(1989), 25-104.
[7] J.P. Grant and N.M. Spyrou, Complex permittivity differences between normal and pathological
tissues: mechanisms and medical significance, J. Bioelectricity, 4(1985), 419-458.
[8] P. Heroux and M. Bourdages Monitoring living tissues by electrical impedance spectroscopy, Ann.
Biomed. Eng., 22(1994), 328-337.
[9] H.C.N. Jongschapp, R. Wytch, J.M.S. Hutchinson andV. Kulkarni, Electrical impedance tomography:
A review of current literature, Europ. J. of Radiology, 18(1994):165-174.
[10] J.R. MacDonald, Impedance spectroscopy, Ann. Biomed. Eng., 20(1992), 289-305.
[11] J.P Morucci, M.E. Valentinuzzi, B. Rigaud, C.J. Felice, N. Chauveau and P.M. Marsili, Bioelectrical
Impedance techniques in Medicine, Critical Reviews in Biomedical Engineering, J.R. Bourne Ed.,
24(1996), ISSN 0278-940X.
[12] R. Pethig, Dielectric properties of biological materials, Clin. Phys. Physiol. Meas, 8A(1984), 5-12.
[13] P.J. Riu, J. Rosell, R. Bragosand and O. Casas, Electrical Bioimpedance Methods, Ann. NY Acd. Sci.,
873(1999), ISBN 1-57331-191-X.
[14] H.P. Schwan, Electrical properties of tissue and cell suspensions, Advan. Biol. Med. Phys., 5(1957),
147-208.
[15] R.D. Stoy, K.R. Foster and H.P. Schwan, Dielectric properties of mammalian tissues from 0.1 to 100
MHz:A summary of recent data, Phys. Med. Biol., 27(1982), 501-513.
Basic Engineering for Medics and Biologists 81
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-81
Introduction
The contact with either a man-made or natural source of electrical energy can create
injury. Such a contact is termed "electrification". The term "electrocution" should be
reserved for cases resulting in a victim's death. The detrimental effects of electrification
depend on several parameters including the source of electrical energy and current
conduction through the body.
For electrification to occur, the victim must complete the circuit: at least two
contact points are necessary with sufficient voltage difference to drive the current
through the circuit. Electrification can have direct and indirect effects. For instance, the
explosive force produced by lightning can throw the victim a distance and the sudden
muscular spasm caused by the current can cause the person to fall from a height.
However, most lesions and trauma due to electrification are the direct effects of
current. It must be pointed out that it is the current that produces lesions. This is
confirmed, for instance, by the harmlessness of the static discharge, where the voltage
is high, but the current is limited in intensity and duration. This feature is reminded by
the familiar slogan:
"The volts jolt, the mills kill", mills signifying milliamps, the unit of current
For a given applied voltage, the current is controlled by the series combination of
the impedance of the sources and the impedance of the body through the contact points.
In lightning and electrostatic discharge (ESD), the resistance of the body is small
compared to that of the total circuit and has very little influence on current intensity. In
contrast, body impedance is the most important variable when considering
electrocution by power lines.
Note: In the literature on electrical safety, tissues are often characterised by their
resistances. This is incorrect use of the terminology since for AC current a tissue is
characterised by its impedance. However, at the frequencies used for power distribution
82 J. Jossinet / Electrical Safety
1.2. Lightning
Lightning is the natural and violent discharge between an electrically loaded cloud
(cumulonimbus) and the ground. Once the "stepper" has established the path, the
current flows along the "arc", i.e. a tube of plasma (high temperature, highly
conductive ionised matter). Lightning voltages are typically greater than 20 kV and can
reach millions of volts. The average current during the discharge is about 20-30 kA.
The lightning generally consists of several strokes and lasts from about 0.2 to 1 sec.
Thunder is due to the shock wave resulting from the rapid temperature change of the air
in the close vicinity of the arc.
Direct stroke is not the only danger of lightning. The lightning current spreads
through the ground and gives rise to gradients in ground potential. The "step voltage" is
defined as the voltage difference in electrical potential between a person’s feet. As a
body is a far better conductor of electricity than the earth, a current can flow through
the legs and cause lesions, trauma and even death, especially if excessive current passes
through the heart [8].
By construction, power lines (and the mains) maintain a constant voltage even if large
current is current is drawn either in the circuit connected to it or through the body of a
person. The contact with power lines can therefore result in severe electrical injury, as
a large current can continue to pass through the body as long as the contact is
J. Jossinet / Electrical Safety 83
maintained. This is particularly critical when the victim is not able to release the
contact, above the so-called "let go threshold".
High-voltage power lines can also give rise to arc-flash between the line and
conducting objects connected to the ground. In air, this may occur for distances shorter
than 1 cm for voltage levels of approximately 4000 volts. The arc is made of plasma
and gases at high temperature produce second and third degree burns by direct contact
or by ignition clothing. Arcs can also produce pressure wave caused by the rapid
thermal expansion of gases, giving rise to the explosion of circuit components.
2. Electrical injury
The detrimental effects of electricity through the body can range from a barely
perceptible tingle to immediate cardiac arrest [2, 12]. The undesirable effects of electric
current consist of burns and electric shocks. Burns are due to the excessive electrical
energy dissipated in the tissue. An electrical current flowing through the body can be
hazardous or fatal if it causes local current densities in vital organs that are sufficient to
interfere with their functions.
2.1. Burns
Burns are produced by the heating of body tissues (Joule effect). The appreciable
conductivity differences between organs result in an inhomogeneous distribution of
current density. Consequently, the thermal effects on electrification (burns) depend on
the pathway of the current through the body. Electrical burns are usually most severe at
the contact points, where current concentration results in high current density. This
effect is exploited in electrocautery devices for the cauterization of tissue, where the
tissue is burned by an electrical current. Electrosurgery uses alternating current to
directly heat the tissue itself. This is known as diathermy.
The severity of electrical injury depends on the type of source, the intensity of the
current, the pathway through the body and the duration of the contact. Other factors are
the applied current frequency, the phase of the heart cycle when the shock occurs and
the general health status of the person.
The effect of electric shock decreases with applied signal frequency [7]. High
frequencies currents do not excite muscles and do not cause cardiac arrhythmias.
84 J. Jossinet / Electrical Safety
Unless otherwise noted, the values quoted here are implicitly for the frequency of
power distribution.
The current and pathway of the current through the body depend on the location
and impedance of contact points and the resistance of the body between these points.
This latter issue depends on each subject's morphology and body composition. Finally,
it must be kept in mind that there are no absolute thresholds or known values
quantifying the exact injury from a given current or voltage.
3.1. Voltage
Sources producing electrical injury are generally classified into "low voltage" (less than
500 volts) and "high voltage" (greater than 1000 volts). Both high and low voltage can
be fatal. Low voltage injuries occur at home or in a residential environment.
Electrocutions in bathtubs and by electric dryers are the most common causes of low-
voltage deaths.
Low voltage does not imply low hazard. In most cases, electrification by the mains
does not cause skin damage unless the contact point is small or the victim has delicate
skin. A low-voltage source can, however, produce major cardiopulmonary
complications and death if a sufficient current passes across the chest (depending on
the resistance) to induce ventricular fibrillation (depending on the pathway).
High-voltage injuries generally occur in an outdoor environment near power
sources and power lines, either by direct contact or arc flashing. High-voltage has a
greater potential for tissue destruction and can be responsible for severe injuries
leading to amputations and tissue loss. However, remembering that the adverse effect
of electrification is due to current a shock of 100 volts is not less dangerous than a
shock of 10000 volts. Individuals have been electrocuted by appliances connected to
the mains.
3.2. Amperage
The term "amperage" refers to the intensity of the current passing through the body.
The "let-go threshold" is the limit at which a person becomes unable to let go off the
current source because of muscular tetany. Typical values of let-go currents for men,
women and children are about 9, 7 and 4 mA, respectively. Electricians familiar with
this effect often refer to an immobilized victim of electric shock as being "frozen on the
circuit." Typical effects of electrification of the human body by 50 or 60 Hz AC
currents are summarized in Table 1.
The type of circuit involved, either direct current (DC) or alternating current (AC) is
one of the factors affecting the nature and severity of electrical injury. Direct current
(ESD, lightning, batteries) tends to throw the victim from the source after. Alternating
current (power lines) is more dangerous for it causes muscle contraction that can
maintain the contact with the power source. High-voltage AC injuries are more severe
than DC ones mainly because of increased risk of muscular tetany and prolonged
contact. AC exposures are said to be three times more dangerous than DC exposure at
the same voltage.
J. Jossinet / Electrical Safety 85
Table 1. Archetypal effects of electric current on the human body against the applied currents for frequency
of 50/60 Hz
1 - 10 Muscular contraction, slight shock, not painful but disturbing, “no let go”
danger
The severity of injury is increases with the time of current flow through the body. The
effects of short duration current are more shock than burns. For instance, the electric
ray produce discharges of several hundred volts. This fish, however, does not produce
lethal shocks since the discharge duration, of the order of tens of microseconds, is too
short.
However, even brief exposures to very high amperage can produce important tissue
damage. Lightning current is seldom long enough to cause severe burns, but it produces
electric shock than can be fatal. Thoracic tetany can occur at levels just above the let-
go current and result in respiratory arrest.
The pathway of the current between the contact points determines the tissues at risk,
the type of injury regardless of whether high, low, or lightning voltages are being
86 J. Jossinet / Electrical Safety
considered.
The hand is the most common site of source contact point. Bone, tendon, and fat
have a very high resistance and tend to heat up and coagulate rather than transmit
current. Similarly high resistance skin resistance dissipates much of the energy,
producing surface burns and limiting internal damage. If the current were delivered
through low resistance contact points, it can produce cardiac arrest but no surface
burns, such as in a bathtub injury.
The impedance at the contact point depends on the piece of material in contact with the
skin and the properties of the skin. In particular, the contact impedance depends on the
contact surface area: the wider the area, the lower the contact resistance. Hence, the
contact impedance can be different if the victim is grasping a handle or coming
accidentally into contact with an electrified object. In practice, the contribution of skin
resistance to the contact impedance is preponderant.
For usual contacts, practical values of skin resistance vary from 1000 Ω for wet
skin to over 500 KΩ for dry skin. When a victim comes into contact with a high-
voltage power source, the epidermis may be destroyed by heat within milliseconds.
Moisture, sweating and immersion in water can dramatically reduce skin resistance.
Pure water is a poor conductor, but tap water is almost as conductive as body tissues. In
practice, water always increases the risk of electrical injury by decreasing the resistance
of contact points.
Body tissues other than skin are relatively good conductors; large currents can then
pass through them and produce tissue damage [1,6]. The resistances of the organs
determine the distribution of the current through the body for the current tends to pass
through low resistance regions. Nerves, muscles, and blood are better electrical
conductors than bone, tendon, and fat. Other body tissues have intermediate
resistances.
The actual resistance of the body varies depending upon the points of contact and
the skin condition (moist or dry). For instance, skin resistance excluded, the internal
resistance between the ears is about 100 ohms, while the resistance from hand to foot is
about 500 ohms.
The outcome of an electric shock depends on the speed and adequacy of the treatment
[3,9]. The current's path through the body determines for prognosis and therapy. It is
impossible to know the exact amperage because of the variability of the resistance.
However, an estimate can be calculated knowing the voltage of the source. The
pathway can be estimated knowing the contact points.
Pain is the least significant result of electric shock. The outside of the victim’s
body may appear to have only minor lesions, but the internal injuries may still be
significant. More generally, the symptoms include:
J. Jossinet / Electrical Safety 87
Skin burns,
tingling, weakness, unconsciousness,
Muscle contraction, muscular pain,
Bone fractures
Headache, hearing impairment, seizures
Respiratory failure, arrhythmias, cardiac arrest.
Simple rules improve electrical safety at home and at the laboratory. The use of
electricity at the working place must be in agreement with the general safety
regulations and the specific ones applying to every particular type of activity (e.g. use
of chemical, flammable gases, protection against ESD). However, the application of
simple basic rules minimizes electrical hazards at home and in the laboratory. Some of
them are given below.
Bibliography
[1] C. Biegelmeier, New knowledge of the impedance of the human body, in Electric shock safety criteria,
J. Bridges, L. Ford, L. Sherman, M. Vainberg, eds, Pergamon Press, 1985.
[2] C.F. Dalziel, The threshold of perception currents, in: IEEE Trans. Power Apparatus and Systems .
73(1954), 990-996.
[3] A.R. Dimick, Electrical Injuries, in Harrison's Principles of Internal Medicine, Anthony S. Fauci et al.
Eds, McGraw-Hill, New York, 1997.
[4] R. Fish, Electric shock- Part I: Physics and pathophysiology, J Emerg Med., 11(1993), 309-312.
[5] R. Fish, Electric shock, Part II: Nature and mechanisms of injury, J Emerg Med., 11(1993), 457-462.
[6] IEC Technical Specification, Electrical impedance of the human body, effects of sinusoidal alternating
current in the range of 15 Hz to 100 Hz, effects of direct current, IEC publication IEC/TS 60479, Part
1, (1994-09).
[7] IEC Technical Specification, Effects of alternating current with frequencies above 100 Hz, Effects of
special waveforms of current, Effects of unidirectional single impulse currents of short duration", IEC
publication IEC/TR 60479, Part 2 (1987-03).
[8] IEC Technical Specification, Effects of lightning strokes on human beings and livestock, IEC
publication IEC/TR 60479, Part 4 (2004-07).
[9] A.C. Koumbourlis, Electrical injuries, Crit Care Med., 30(2002), 424-430.
J. Jossinet / Electrical Safety 89
[10] R.C. Lee, D. Zhang and J. Hannig, Biophysical injury mechanisms in electrical shock trauma, Ann. Rev
Biomed Eng., 2(2000), 477-509.
[11] D. Leibovici, J. Shemer, S.C. Shapira, Electrical injuries: current concepts, Injury, 26(1995), 623-627.
[12] J.P.Reilly, Scales of reaction to electric shock, in Electrical Injury, Annals of N.Y. Acad. Sci.,
720(1994), 21-37.
90 Basic Engineering for Medics and Biologists
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-90
1. Biopotentials
The term biopotential refers to the electric potential that is measured between points in
living cells, tissues, and organisms and which accompanies all biochemical processes.
Biopotentials from the organs of the body provide rich physiological and clinical
information, playing often significant role in diagnosis.
Bioelectric potentials are produced as a result of electrochemical activity in a
certain class of excitable cells. These cells are components of nervous, muscular, and
glandular tissues. Electrically, these cells exhibit a resting potential and, when
appropriately stimulated, an action potential. Their activity is necessary for information
transfer (e.g. sensory information in the nervous system or coordination of blood
pumping in the heart).
various molecules (i.e. K+, Na+, Cl-, and Ca2+). The cell membrane is a very thin (7-15
nm) lipoprotein complex that is essentially impermeable to intracellular protein and
other organic anions (A-). The membrane in the resting state is only slightly permeable
to sodium Na+ and rather freely permeable to potassium K+ and chlorine Cl-. The
permeability of the resting membrane to potassium ions is approximately 50~100 times
larger than its permeability to sodium. The diffusion and electrical forces acting across
the membrane are opposed to each other, and a steady state is ultimately achieved. This
steady membrane potential is called the equilibrium potential for potassium on account
of the main ion potassium K+ involved in the resting state. As a result an approximation
of steady membrane potential can be calculated using the Nernst equation , Eq. (1). The
Nernst equation provides a quantitative value for potential generated, EK given the
intra- and extracellular concentrations of potassium.
RT [K ] o [K ] o (1)
EK = ln = 0.0615 log10
nF [K ] i [K ] i
where n is the valence of potassium, [K]i and [K]o are the intra- and extracellular
concentrations, R is the universal gas constant, T is the absolute temperature in Kelvin,
F is the Faraday constant [1].
A more accurate expression for the membrane equilibrium potential can be
calculated from the Goldman- Hodgkin-Katz equation, Equations 2 and 3, which
provide a more accurate estimate of the potential E, by taking into consideration the
intra- and extracellular concentrations of sodium and chlorine ions as well as
potassium.
⎡ ⎤ (2)
RT ⎢ PK [K ]o + PNA [Na ]o + PCl [Cl ]i ⎥
E= ln
F ⎢ PK [K ]i + PNA [Na ]i + PCl [Cl ] ⎥
⎣⎢ o ⎦⎥
⎡ ⎤ (3)
P [K ]o + PNA [Na ]o + PCl [Cl ]i ⎥
E = 0.0581 log10 ⎢ K
⎢ PK [K ]i + PNA [Na ]i + PCl [Cl ] ⎥
⎢⎣ o ⎥⎦
ion channels into the cell open, and sodium ions rush in, reversing the electrical charge
at the point of stimulus (In the l millisecond that a channel is open, about 20,000
sodium ions flow through). Thus, at the point of stimulus, the inside of the membrane
becomes positively charged relative to the outside, a condition known as
depolarization. When a stimulus is strong enough to cause depolarization, the neuron is
said to fire.
Once a small area on the neuron is depolarized, it stimulates the adjacent area, and
an action potential, or nerve impulse, is initiated and conducted along the plasma
membrane.
Shortly after depolarization, tbe original balance of sodium and potassium ions is
restored by the action of the membrane pumps; the relative electrical charges inside and
outside the membrane are also restored. The membrane is then said to be repolarized.
The transmission of a nerve impulse along the plasma membrane may be visualized as
a wave of depolarization and repolarization.
After each firing, there is an interval of from 0.5 to 1 millisecond before it is
possible for an adequate stimulus to generate another action potential. This period is
called the refractory period. Most nerve fibers are capable of generating about 300
impulses per second
In the resting state there exists a steady electrical potential difference between
internal and external environments—typically between -70 to -90mV, relative to the
external medium.
In the active state an electrical response is instigated by adequate stimulation. It
consists of an “all-or-none” action potential after the cell threshold potential has been
reached.
The body is an inhomogeneous volume conductor and these ion fluxes create
measurable potentials on the body surface. By recording these biopotentials, a process
known as electrography, an analysis of the underlying physiological can be carried out
providing information for clinical diagnosis.
Biopotentials have extremely low voltages values, typically in the order of micro
or millivolts. Amplification of these biopotentials is required prior to their acquisition
by computer and subsequent analysis.
2. Bioamplifers
3. Electrocardiogram (ECG)
The heart consists of four chambers: the left and right atria and also the left and right
ventricles. Blood enters the right atrium and passes through the right ventricle, which in
turn pumps the blood to the lungs where it becomes oxygenated. The oxygenated blood
is brought back to the heart by the pulmonary veins which enter the left atrium, flowing
then into the left ventricle. The left ventricle pumps the blood to the aorta which will
distribute the oxygenated blood to all parts of the body.
There are three groups of cells in the heart. Pacemaker cells, Electrical Conducting
Cells and Myocardial Cells [2]. In the resting state these cells are electrically polarised,
but it is their depolarisation that is the fundamental electrical event of the heart.
Depolarisation is propagated from cell to cell, producing a wave of depolarisation that
is transmitted across the heart. This wave sets up a biopotenital across the heart,
resulting in the flow of electrical current [3]. This depolarisation can be detected by a
number of electrodes attached at specific points on the body. Once the depolarisation
has occured, repolarisation commences to restore the cardiac cells to their resting state.
This repolarisation phase can also be recorded with the same electrodes. The ECG is a
recording of the depolarisation and repolarisation of the myocardial cells of heart,
which makeup the majority of heart cells and represent the heart’s contractile
machinery. The ECG indicates the overall rhythm of the heart and weaknesses in
different parts of the heart muscle [4].
Since the human body can be regarded as a volume conductor, changes in potential
are transmitted throughout the body. Therefore to be able to record myocardial activity,
one needs to be able to detect small changes in potential on the body surface. These
waves can be measured at electrodes attached to the skin. An ECG displays the voltage
between pairs of these electrodes and the muscle activity in different locations.
In order to record the ECG, each lead pair is connected to a bioamplifier. Bipolar
leads measure the difference in electrical potential between any two points on the body,
unipolar leads measure the potential at a single point on the body and a virtual
reference point, with zero electrical potential, located in the centre of the heart.
A lead records the electrical signals of the heart from a particular combination of
recording electrodes that are placed at specific points on the patient's body. In various
cases, the detected signals have the following polarities (see Figure 1):
94 R.B. Reilly and T.C. Lee / Electrograms (ECG, EEG, EMG, EOG)
Figure 1. Graphic showing the relationship between positive electrodes, depolarization wavefronts (or mean
electrical vectors), and complexes displayed on the ECG.
The Einthoven limb or Standard leads are defined in the following way:
Lead I: VI = Φ L - Φ R
Lead II: VII = Φ F - Φ R
Lead III: VIII = Φ F - Φ L
where
VI = the voltage of Lead I, VII = the voltage of Lead II,
VIII = the voltage of Lead III
Φ L = potential at the left arm (LA)
Φ R = potential at the right arm (RA)
Φ F = potential at the left leg (LL)
According to Kirchhoff's law, these lead voltages have the following relationship:
V I + V III = V II (4)
For measuring the potentials close to the heart, Wilson introduced the precordial leads
(chest leads) in 1944. These leads (V1 to V6) are located over the left chest as shown in
Figure 3.
The most commonly used clinical ECG-system, the 12-lead ECG system, consists of
the following 12 leads:
• I, II, III
• aVR, aVL, aVF
• V1, V2, V3, V4, V5, V6
R.B. Reilly and T.C. Lee / Electrograms (ECG, EEG, EMG, EOG) 97
Of these 12 leads, the first six are derived from the same three measurement points.
Therefore, any two of these six leads include exactly the same information as the other
four.
In order to interpret the ECG, one needs to understand the electrical events that
underpin the cardiac cycle of contraction (systole) and relaxation (diastole) [2]. The
dominant pacemaker cells in the heart, are known as the Sinoatrial Node. These are
located at the top right hand side of the heart, in the right atrium. These cells fire
spontaneously at a rate of 60 to 100 times per minute, and a wave of depolarization
begins to spread outward into the atrial myocardium cells. It is the depolarization of
atrial myocardium cells that results in atrial contraction. Electrodes on the surface of
the body will record this contraction as a burst of electrical activity lasting a fraction of
a second. This is called the P wave. The atria begin contracting around 100 ms after
the start of the P wave. Once the atrial depolarisation is complete the ECG returns to
baseline.
Following this the ventriclar myocardial cells depolarise, causing a ventriclar
contraction. Electrodes on the surface of the body will record this contraction as a large
increase in amplitude. This is known as the QRS complex. The amplitude of the QRS
complex is larger than the P wave as the mass of the ventricular muscle is much larger
than that of the atria. The ventricles begin contracting shortly after the R wave. The
smaller T wave indicates ventricular repolarization. The atria repolarize during the
QRS complex and therefore this repolarization cannot be observed separately.
When the heart depolarizes and repolarizes, it is convenient to represent the electrical
activity as an electric dipole or mathematically as a vector. A vector has both
magnitude and an angle of orientation. The magnitude of the heart vector represents a
maximum amplitude of the recorded electrical activity, while the angle of orientation
represents the average direction of the current flow in the heart. The magnitude and
direction of this heart vector can be beneficial in diagnosis.
The augmented and standard leads allow analysis of the heart on what is known as
the frontal plane. Therefore, the heart vector refers to the direction of the electrical
depolarization obtained from the sum of all different vectors in the frontal plane. It is
usually oriented in a right shoulder to left leg direction. The normal angle of
orientation of the QRS vector generally lies between 0 and 90 degrees. A change in the
angle of orientation provides diagnostic information. For example, a left deviation (-30 o
to -90o) may indicate left anterior fascicular block or Q waves from inferior myocardial
infarction (MI). Right axis deviation (+90o to +180o) may indicate left posterior
fascicular block, Q waves from high lateral myocardial infarction, or a right ventricular
strain pattern.
ECG is the standard method for the diagnosis of cardiac arrhythmias (Figure 5).
Pathology that can be identified and studied using ECGs includes rhythm disturbances,
ischaemia and infarction, chamber enlargements, electrolyte abnormalities, and drug
toxicities.
4. Electroencephalography
Hans Berger was the first to standardize recordings of the electrical activity in the
brain. In a series of experiments in 1920s, he pioneered electroencephalography (EEG),
by placing electrodes on the scalp and measuring electrical activity during different
tasks. The name is derived from the Greek encephalo meaning brain.
The conventional EEG is recorded from scalp electrodes, and shows cortical
electrical activity. This includes cortical manifestations of the sub cortical regions
(projection pathways, thalamus, reticular formation, mesencephalon or midbrain) [5].
As the EEG reflects the functional status of the brain, it can be used to monitor its
functional integrity and thus assist clinicians in the diagnosis of a variety of
neurological problems. It is very important to relate scalp potentials to the underlying
neurophysiology. The pathological states most commonly diagnosed using EEG
include, common headaches and dizziness, seizure disorders, stroke, brain tumours,
epilepsy, multiple sclerosis, sleep disorders and movement disorders [5].
A method similar to the EEG is intracranial EEG (icEEG), also referred to as
subdural EEG (sdEEG) and electrocorticography (ECoG). The electrodes are placed
directly on the exposed surface of the brain to record electrical activity from the
cerebral cortex. ECoG is an invasive procedure in which electrodes are placed over a
specific brain area. In a very small selected group of patients being considered for
epilepsy surgery, ECoG may be recorded over several days to map the distribution and
spread of seizure activity.
EEG records the spontaneous activity in the brain. The amplitude the signal record
on the scalp is typically 100 µV, while that of ECoG from the surface of the cortex it is
in the order of 1-2 mV.
There is a standard system for electrode placement, known as the 10-20 international
system which includes 64 electrodes. In this system the electrodes are located on the
surface of the scalp, as shown in Figure 6. The electrode positions are determined as
follows: Reference points are nasion, which is the delve at the top of the nose, level
with the eyes; and inion, which is the bony lump at the base of the skull on the midline
at the back of the head, the external occipital protrusion. The first mark is placed 10%
of the distance along this line and others are arranged at 20% intervals. With this
electrode placement system, activity at specific brain locations can be recorded with
specific electrodes and thus can easily be compared across individuals [5].
When carrying out EEG experimentation, it is typical to record specific sensory
processing areas of the cortex: visual, auditory or somatosensory. For example, visual
processing is carried out in the occipital regions at the back of the brain. Activity at
electrodes O1 and O2 would be more active during visual processing and these
electrodes would be termed active electrodes while the brain is engaged in processing
visual information.
64-channel EEG systems are typical, but some high-density systems exist, with
128 and 256 electrodes. Placement is based again on the 10-20 system, with the new
electrodes assigned to locations in between the 32 electrodes of the 10-20 system.
In order to obtain good signal quality, impedance between the scalp and each
electrode should ideally be kept below 5kOhms. If the electrode impedance increases,
background noise and movement artifacts may obscure the EEG signal. As a
R.B. Reilly and T.C. Lee / Electrograms (ECG, EEG, EMG, EOG) 101
consequence, the signal-to-noise ratio decreases and the ability to extract diagnostic
information may be limited.
Figure 6. The international 10-20 system seen from (A) left and (B) above the head. A = Ear lobe, C =
central, Pg = nasopharyngeal, P = parietal, F = frontal, Fp = frontal polar, O = occipital.
The EEG signal recorded on the scalp or on the cortex is generated by the summation
of the synchronous activity of thousands of neurons that have similar spatial orientation
and placement radial to the scalp. An action potential in a pre-synaptic axon causes the
release of neurotransmitter into the synapse. The neurotransmitter diffuses across the
synaptic cleft and binds to receptors in a post-synaptic dendrite. The activity of many
types of receptors results in a flow of ions into or out of the dendrite. This results in
currents in the extracellular space. It is these extracellular currents which are
responsible for the generation of EEG potentials.
The character of the EEG signal is highly dependent on the degree of the activity of the
cerebral cortex. EEG observed during states of wakefulness and sleep are remarkably
102 R.B. Reilly and T.C. Lee / Electrograms (ECG, EEG, EMG, EOG)
The EEG reflects the functional status of the brain. The strict definition is continuous
EEG, as it represents ongoing brain activity. Another useful electrical signal that can be
recorded is the evoked or event related potential. Evoked potentials arise in response to
a stimulus (auditory, visual or somatosensory, etc.). They provide great diagnostic
information on sensory processing. They are much smaller in amplitude than
continuous EEG and are not readily distinguished. Therefore, a train of stimuli is often
applied to the individual and signal averaging method used to improve the signal-to-
noise ratio of the recorded signals to allow interpretation [5],[6].
R.B. Reilly and T.C. Lee / Electrograms (ECG, EEG, EMG, EOG) 103
Measurements are made of the latencies and amplitudes of particular peaks. In the
normal VEP, a very distinguishable peak is found approximately 100ms after
stimulation, known as the P100. There are also two clear troughs called the N75 and
N135, around 75ms and 135ms after stimulation. The latencies and amplitudes of these
peaks provide objective diagnostic information on the brain’s ability to process visual
information. Individual data can be with compared to normative data.
4.4.2. Auditory Evoked Potentials
Audio evoked potentials (AEP) reflect the function of the auditory pathway. AEP can
be evoked by repeated clicks of short duration (100-500 ms) given into an ear piece.
Trigger-synchronized averaging of a defined number of EEG-segments is used to
extract the AEP by reduction of the underlying EEG signal (background noise). The
reduction in background noise is proportional to the square root of the number of
averaged segments; the more averaged segments, the better the quality of the AEP. The
extracted AEP signal consists of a number of peaks and troughs. As with VEP analysis
measurements are made of latencies and amplitudes of particular peaks. Three main
groups of peaks can be distinguished and they can be correlated to the anatomical
structures:
104 R.B. Reilly and T.C. Lee / Electrograms (ECG, EEG, EMG, EOG)
Changes in EEG are not necessarily specific to underlying mechanisms [5]. For
example, slowing of EEG may reflect either changes in anesthetic concentrations or
cerebral ischemia. EEG has poor spatial resolution, meaning that it not possible to
know exactly the source of electrical activity from the recordings made on the scalp.
Mathematically one can consider the electrical activity being generated by an electric
dipole. Considerable research attention is focused on providing estimates of the
location of electric dipoles that would generate the potential similar to those recorded.
The magnitude and orientation of these dipoles can be used as a basis for diagnosis in
neurological conditions.
Despite these disadvantages, EEG has considerable advantages over neuroimaging.
Its poor spatial resolution is offset by excellent temporal resolution, which is in
milliseconds. The ease of recording EEG makes it of great clinical importance.
5. Electromyography (EMG)
Muscle consists of motor units, which is areas consisting of a motor neuron and the
muscle fibres it innervates. Each motor unit is controlled by the motor neurons. The
motor end plate is the synaptic junction between the motor neuron and the controlled
motor unit. Depolarization of the post synaptic membrane arises in case of activation of
the motor unit. End plate potential (EPP) is the potential that is recorded. The
depolarization wave moves along the direction of the muscle fibers. The signal between
the EMG electrodes corresponds to the depolarization wave front and to the subsequent
repolarization wave.
R.B. Reilly and T.C. Lee / Electrograms (ECG, EEG, EMG, EOG) 105
The electrical source is the muscle membrane potential of about -70mV. Due to
movement of the muscle, the measured potentials range between less than 50 μV and
20 to 30 mV. EMG signals are made up of superimposed motor unit action potentials
(MUAPs) from several motor units. For a thorough analysis, the measured EMG
signals can be decomposed into their constituent MUAPs.
The electrode should be placed between a motor point and the tendon insertion or
between two motor points, and along the longitudinal midline of the muscle. The
longitudinal axis of the electrode should be aligned parallel to the length of the muscle
fibers. The reference electrode also called the ground electrode should be placed as far
away as possible and on electrically neutral tissue [7].
The voltage waveform that is recorded is the difference in potential between the
two electrodes. The electrical signal generated in the muscle fibers is called a muscle
action potential (MAP). As mentioned above, the surface electrodes or indwelling
electrodes record the algebraic sum of all MAPs that are being transmitted along the
muscle fibers.
6. Electrooculography (EOG)
Figure 8. An illustration of the electrooculogram (EOG) signal generated by horizontal movement of the
eyes. Two electrodes are employed and each electrode connected to the input of a bioamplider. The polarity
of the signal is positive at the electrode to which the eye is moving
7. Conclusion
Electrical activity associated with the brain, heart, muscle and eye can be recorded and
analysed. Comparisons against normative data provide important clinical information.
It is important to relate the electrical activity to the underlying physiology so that this
information can be used for diagnosis.
References
[1] Deutsch S. and Deutsch A. (1993), Understanding the Nervous System-An Engineering Perspective,
IEEE Press, New York.
[2] Katz A.M. (1986), Physiology of the Heart, Raven Press, New York
[3] DiFrancesco D. and Noble D. (1985) “A model of cardiac electrical activity incorporating ionic
pumps and concentration changes”, Phil Trans R. Soc, London [B] 307, 307-353
[4] Geselowitz D.B. (1989) “On the theory of the electrocardiogram”, Proc IEEE 77, 857.
[5] Nunez P.L., Srinivasan R. (2005) Electric Fields of the Brain: The Neurophysics of EEG, Oxford
University Press, USA.
[6] Buzsaki G. (2006), Rhythms of the Brain, Oxford University Press, USA.
[7] Katirji B. (2007) Electromyography in Clinical Practice: A Case Study Approach, 2nd Edition,
Mosby.
R.B. Reilly and T.C. Lee / Electrograms (ECG, EEG, EMG, EOG) 107
Glossary of Terms
II.4. Biosensors
Richard B. REILLY and T. Clive LEE*
Trinity Centre for BioEngineering, Trinity College, Dublin 2, Ireland
*Department of Anatomy, Royal College of Surgeons in Ireland, Dublin 2, Ireland
Introduction
The ability to accurately measure physiological and chemical changes in the body is
fundamental to the development of new diagnostic and therapeutic methods. In order to
carry out such measurements, specific tools are necessary to detect physiological and
chemical changes and to transform these changes into a form that can easily be
analyzed and processed. A biosensor is one such tool and can be defined as an
analytical device which converts a biological response into an electrical signal. The
name signifies that the device is a combination of two components, a biological
element and a sensor or transducer element.
The biological element may be tissue, living cells, an enzyme, antibody or antigen.
The sensor/transducer element includes electric current, electric potential, intensity and
phase of electromagnetic radiation, mass, conductance, impedance, temperature and
viscosity.
This paper will outline the components and characteristics of a biosensor, describe
some of the typical transducers found in biomedical engineering and biomedical
instrumentation amplifiers. A focus is placed on sensing within electrophysiology.
How transducers are used in medicine to measure key biophysical characteristics in the
clinic will form the concluding section.
1. Components of a Biosensor
An ideal biosensor must fulfill a number of criteria to measure a physical quantity [1].
The biosensor must be able to provide a signal that is directly proportional to the
quantity under investigation, providing all the useful information about the physical
event. The biosensor must also meet requirements connected with measurement such as
repeatability, sensitivity, reproducibility, selectivity, specificity, good response times
and linearity of response. These are defined as follow:
Specificity/selectivity: Ability to recognize a single event/condition among other
events/conditions in the same sample/signal. The selectivity of the biosensor is
determined by both the sensor and transduction method.
Sensitivity: The sensitivity of a sensor is measured by the change in its response as
a function of the analogous change in the quantity examined. The sensitivity s of a
biosensor is defined by the ratio
Δoutput
s=
Δinput
(1)
where
Δoutput = change in the amplitude of output and
Δinput = change in the amplitude of input.
The sensitivity determines the aptness of the sensor for a particular application.
Repeatability: The results are deemed to have high repeatability if two sets of
results, obtained using the same sensor and the same sample, are numerically close to
each other.
Reproducibility: The method is reproducible if similar results are obtained by other
scientists at different laboratories.
Linear response: The linear region of a biosensor is obtained from a calibration
curve of its response to different inputs. A good calibration curve is also indicative of
stability of the biosensor’s response, which should neither drift nor oscillate with time.
R.B. Reilly and T.C. Lee / Biosensors 111
There are several types of transducers and they can be classified based on their
mechanism of action. Some of the types are listed below with examples.
Displacement Transducers: A displacement transducer measures distance or angle
traversed by the sensor. A device that produces movement in a straight line is a linear
displacement transducer and if it measures movement through an angle it is angular
displacement transducer. Displacement transducers are widely used in kinesiology, the
study of human movement.
Resistive Transducers: A resistive transducer contains a translational or angular
displacement sensing shaft. This shaft drives a wiper in the transduction element that
slides along the resistive layer. The resistance measured between the wiper and one of
the ends of the layer is a function of the position, motion, or displacement. Resistive
transducers are often used in kinesiological studies.
Inductive Transducers: Inductive transducers are used to sense small
displacements of movements and are based on the inductance L of a coil. This type of
transducer measures movement by changing the value of L, in proportion to the
movement of a ferrite or iron core in the bore of the coil assembly. L is given by the
following equation L = n2 x G μ, where n is the number of turns in the coil, x the
distance of the ferrite core within the coil assembly, G is the Geometric form constant,
and μ is the permeability of magnetically susceptible medium inside the coil. Inductive
transducers are often used to measure respiration, where the sensor is connected to a
flexible woven belt worn around the chest.
Capacitive Transducers: The sensing shaft in a capacitive transducer changes the
position of the dielectric between the capacitor's plates in the transduction element, or it
changes the distance and area between the plates. A change in these three parameters
leads to a change in capacitance, which is then measured.
Piezoelectric Transducers: Piezoelectric crystals (e.g. quartz) vibrate under the
influence of an electric field. The frequency of this oscillation depends on their
thickness and cut, each crystal having a characteristic resonance frequency.
Piezoelectric crystals can generate an electric field when subjected to physical pressure.
As a result, piezoelectric crystals can sense light pressures, such as touch. They are the
112 R.B. Reilly and T.C. Lee / Biosensors
basis of tactile sensors. They can also be used to provide vibrotactile stimulation in the
study of movement disorders and in force plates to measure ground reaction forces.
Transducers in temperature Measurement: The most popular method of
temperature measurement is by using a mercury thermometer, but this is slow, difficult
to read, not reliable and prone to contamination. In many cases, continuous monitoring
of temperature is required, as in an operating theatre or intensive care unit. Electronic
thermometers are convenient, reliable and hence used for continuous monitoring. They
use probes incorporating a thermistor or thermocouple sensor, which is an electronic
component that has rapid response characteristics of voltage with temperature.
Thermocouples measure the temperature difference between two points and not
absolute temperature.
Electrochemical Transducers: The underlying principle for this class of biosensors
is that many chemical reactions produce or consume ions or electrons which in turn
cause some change in the electrical properties of the solution. These electrical
properties can be measured.
Sensing electrical activity within the body is an important task in clinical diagnosis.
This section focuses on the physics underlying the acquisition of this electrical activity.
2.1. Electrodes
The amplifier that is often used in conjunction with transducers is the operational
amplifier or Op-Amp [2]. An op-amp is a differential amplifier that amplifies the
difference between two inputs. One input has a positive effect on the output signal; the
other input has a negative effect on the output. The op-amp is powered by a dual
polarity power supply in the range of +/- 5 volts to +/- 15 volts. The theoretically
R.B. Reilly and T.C. Lee / Biosensors 113
perfect op-Amp has an infinite voltage gain, theoretically being able to amplify a signal
to any level. It also theoretically has infinite bandwidth, being able to amplify any
frequency without distortion. It also theoretically has infinite input impedance, being
able to sense an input voltage level without distorting that voltage in any way. The
perfect Op-Amp also has zero-Ohm output impedance, allowing it to be connected to
any other electronic device without distortion. In reality such characteristics are not
met, but real op-amps do achieve great performance when correctly configured.
The inverting amplifier is one such configuration [2]. The op-amp is connected using
two resistors RA and RB such that the input signal is applied in series with RA and the
output is connected back to the inverting input through RB. The non-inverting input is
connected to a reference voltage, typically electrical ground or zero volts. This can be
achieved by connecting to the centre tap of the dual polarity power supply (Fig. 2a).
During operation, as the input signal becomes positive, the output will become
negative. The opposite is true as the input signal becomes negative. The amount of
voltage change at the output relative to the input depends on the ratio of the two
resistors RA and RB. As the input changes either positively or negatively, the output
will change in the opposite direction, so that the voltage at the inverting input remains
constant or zero volts in this case. If RA is 1K and RB is 10K and the input is +1 volt
then there will be 1 mA of current flowing through RA and the output will have to
move to -10 volts to supply the same current through RB and keep the voltage at the
inverting input at zero. The voltage gain in this case would be R B/RA or 10K/1K = 10.
Since the voltage at the inverting input is always zero, the input signal will see input
impedance equal to RA, or 1K in this case. For higher input impedances, both resistor
values can be increased.
The voltage follower, also called an electrical buffer, provides high input impedance, a
low output impedance, and unity gain. As the input voltage changes, the output and
inverting input will change by an equal amount (Fig. 2c). The voltage follower is often
used to electrical isolate the transducer from a computer, to allow subsequent analysis
and signal processing to take place.
114 R.B. Reilly and T.C. Lee / Biosensors
2.8. Filtering
Filtering is a necessary processing step with any biosensor, in order to reduce noise to a
minimum and accentuate frequencies of interest. Filtering may be carried out in
software using digital signal processing, but it can also be accomplished in hardware
[3].
When designing a filter, the range of frequencies we wish to pass through the filter
without attenuation is known as the passband. The range of frequencies we wish to
attenuate is known as the stopband. The steepness of the filter’s transition from the
passband to the stopband, is related to the filter order. The higher the filter order, the
sharper the transition and more precise is the filter in pass and attenuating frequencies.
Three 2nd order filters are described here. A low pass, high pass, and bandpass
filter. Each of these filters will attenuate frequencies outside their passband at a rate of
12dB per octave equivalent to 25% of the voltage amplitude for each octave of
frequency increase or decrease outside the passband (Fig. 3(a) and Fig. 3(b) ).
First order low or high pass cutoff frequency is given by the choice of a resistor
and capacitance value.
1
FrequencyCutoff =
2πRC
(2)
2nd order low or high pass cutoff frequency (-3dB point) is also chosen by resistor
and capacitance values.
1
FrequencyCutoff =
2π R1R2C1C 2
(3)
from the centre to the upper, but not the absolute amount. The filter bandwidth (BW) is
the difference between the upper and lower passband frequencies. A formula relating
the upper, lower, and center frequencies of the passband is
R1 C1
R1 R2
C1 C2
C2
R2
Figure 3. (a) High Pass Filter Figure 3. (b) Low Pass Filter
R1
C1
R1 R2
C1 C2
R2 C2
This section provides an example of the use biosensors in clinical engineering, a branch
of biomedical engineering which is concerned with equipment routinely used in
hospitals and clinics [4].
Direct blood pressure measurements are made by introducing a cannula (needle) into an
artery and then coupling this to a pressure transducer. The advantage of this approach is
that pressure is constantly monitored beat-by-beat and a clinically important waveform
(a graph of pressure against time) can be displayed. Patients with invasive arterial
monitoring require very close supervision, as there is a danger of severe bleeding if the
cannula line becomes disconnected. It is generally reserved for critically ill patients
where rapid variations in blood pressure are anticipated.
Spirometry is the measurement of lung function and is the most common pulmonary
function test. It is an important tool used in assessing conditions such as asthma,
pulmonary fibrosis, cystic fibrosis, and chronic obstructive pulmonary disease (COPD).
There are many designs of spirometers, but most modern devices are electronic based.
They all involve the measurement of the volume and flow of breath during inhalation
and exhalation. The most common electronic based devices consist of a small handheld
118 R.B. Reilly and T.C. Lee / Biosensors
tube into which the patient exhales as hard as possible and for as long as possible,
having been just requested to take the deepest breath possible. A flow sensor in the
tube measures the direction and volume of the breath flow. Following exhalation, the
patient rapidly inhales (inspiration). Inspiration is important when assessing possible
upper airway obstruction and is sensed in the tube by the directional sensor.
Sometimes, the test will be preceded by a period of quiet breathing in and out from the
tube, so that sensor can measure tidal volume. During the test soft nose clips may be
used to prevent air escaping through the nose.
4. Conclusions
As can be appreciated, the majority of biosensors are essentially the same as sensors
used in other commercial or industrial applications. The distinctive aspect about
biosensors is their application. However, there are problems that are encountered by
biosensors that are unique to them. These problems are associated with the interface
between the biosensor and the biological system being measured. The presence of
foreign materials, especially implanted materials, can affect the biological environment
surrounding the sensor. Therefore one must be aware of potential physical or chemical
reactions as a result of the sensing process. The presence of the sensor may also hinder
the subject’s physical and cognitive ability. Therefore the placement of the sensor on
the body must be given due care and attention.
Despite these issues, biosensors have an enormous clinical impact both in
diagnosis and therapy, as well for being critical for data collection for biomedical
research.
References
[1] Geddes L.A. and Baker L.E., (1989), Principles of Applied Biomedical Instrumentation, Wiley-
Interscience, New York.
[2] Horowitz P. and Hill W., (1989), The Art of Electronics, Cambridge University Press.
[3] Wise D.L. (1991), Bioinstrumentation and Biosensor, Marcel Dekker, New York
[4] Harsanyi G. (2000), Sensors in Biomedical Applications: Fundamental Technology and Applications,
CRC, Florida.
Basic Engineering for Medics and Biologists 119
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-119
Introduction to
Chapter III: Medical
Informatics for Biomedical
Engineering
Paul McCULLAGH and T. Clive LEE (eds.)
Introduction
The computer is ubiquitous in everyday-life and is an essential tool for the biomedical
engineer. It is fundamental to the acquisition, storage, processing and analysis of data,
122 P.J. McCullagh et al. / Medical Informatics and eHealth
and can be used to summarize and display information via the human computer
interface (HCI) so that it is comprehensible to users. This interface is normally
graphical nowadays and can be very sophisticated, e.g. 3-dimensional image
construction of the fetus in the womb, by processing reflections of externally applied
ultrasound, providing the means for clinical assessment. Indeed processing capacity is
such that images can be constructed and displayed in real-time to provide a non-
invasive ‘movie’ of the fetus [4]. Alternatively, a more straightforward example is the
use of automated speech recognition to allow a specialist to record clinical observations
directly into the computer. An example of this is the reporting of diagnostic imaging
results in radiology [5].
The ability to be connected to private networks and the Internet is essential for
information sharing and for information retrieval from large authoritative information
repositories. The Electronic Patient Record (EPR), which is now commonplace in
Primary Care, is an example of the former, with many additional, often incompatible
systems existing in Secondary Care departments of large hospitals. In the United
Kingdom, a key aim of the government's Connecting for Health program [6] was to
establish a central database of 50 million patients' medical records, accessible over
NHSnet1. This is the world’s largest Information Technology development program,
with an estimated cost from six to twenty billion pounds. The United States National
Library of Medicine’s publicly accessible database PUBMED [7, 8] with over 16
million citations from MEDLINE is an example of an Internet based resource. Others
include large data banks such as the human genome project.
Van Bemmel has provided an important conceptual model for structuring
computer applications in health [9]. It comprises six levels, with the lowest level being
largely automated and the higher levels requiring progressively more human
interaction. This model provides a useful context in which to discuss the interaction of
computers, networks and eHealth. We will use the recording, storage and interpretation
of the electrocardiogram (ECG), an electrical recording used to assess the physiology
of the heart, as an example.
Level 1: Communication and telematics deals with data acquisition and transfer,
e.g. the recording of the ECG potential waveform with electrodes and physiological
amplifier and its subsequent transfer, in a standardized format, to a computer screen for
human interpretation. The computer can be thousands of miles distant, a capability
which has promoted the emergence of telemedicine as a discipline.
Level 2: Storage and retrieval deals with the storage of information in a data
repository, e.g. multimedia EPR which capitalizes upon advances in memory storage
capacity and database technology to form digital libraries comprising labeled text,
physiological recordings and images.
Level 3: Processing requires a computer program to extract information from the
captured data. For an ECG, this could be important diagnostic parameters such as heart
rate, or pattern abnormalities such as ST elevation2, which may have some diagnostic
significance. Similarly for an image (e.g. an echocardiogram), the computer can be
used to segment boundaries and highlight potential abnormalities, for human
verification.
Level 4: Diagnosis combines features extracted from the processing of acquired
data with existing medical knowledge to assist with decision making, e.g. the ‘ECG is
1
NHSnet is a private network connecting hospitals and general practitioner sites in the UK
2
The well know ECG complex is designated by QRST markers
P.J. McCullagh et al. / Medical Informatics and eHealth 123
Logically, the computer comprises three elements: (i) a processor (control unit and
arithmetic and logic unit (ALU)), (ii) random access memory, some form of persistent
storage, currently a hard disk (memory) and (iii) input/output devices. This is the
classical von Neumann architecture [10] or "stored-program computer", which has
been in use for decades and forms the design of most of today’s computers. Parallel
architectures are also in use, particularly for high performance applications. At the
heart of a computer is an algorithm, which is a set of programmed instructions to carry
out some task. The design uses a single store to hold both instructions and data, and the
control unit instructs the ALU to execute instructions sequentially.
Any hardware specification quickly becomes outdated due to the pace of
technological progress. This can be summed up by Moore's Law (proposed by Gordon
E. Moore a co-founder of Intel, the major silicon chip manufacturer) which observes
that the transistor density of integrated circuits, with respect to minimum component
cost, doubles every 24 months [11]. A similar law has held for hard disk storage cost
per unit of information, and indeed random access memory (RAM) storage capacity
has increased at the same rate as processing power. Kurzweil’s expansion of Moore's
Law [12] shows that the underlying trend has held true from integrated circuits to
earlier transistors, vacuum tubes, relays and electromechanical computers. He projects
continued progress until 2019 with transistors just a few atoms in width, and predicts
that it is possible that Moore's Law may be extended by a paradigm shift into new
technology, e.g. quantum computers.
The computer has many different incarnations: supercomputer, grid computer,
mainframe, server, desktop, laptop, personal data assistant, smart phone and embedded
device. A summary of characteristics is provided in Table 1.
124 P.J. McCullagh et al. / Medical Informatics and eHealth
The supercomputer and grid perform specialized tasks, typically in the research
world, where enormous processing is required. For example, supercomputers have
been used to model avian flu progression by scientists from Los Alamos National
Laboratory [13], and to model HIV's vulnerability to a class of drugs known as protease
inhibitors [14] by researchers at Stony Brook University. The former application was
implemented on a platform developed for the nuclear weapons programme comprising
2048 processors, the largest cluster in the world. The latter simulation took 20,000
hours of processing (about three months) on the USA National Center for
P.J. McCullagh et al. / Medical Informatics and eHealth 125
3
Terms for storage: Terabytes (1012 bytes ), Gigabytes (109 bytes), Megabytes (106 bytes), Kilobytes (103
bytes),
126 P.J. McCullagh et al. / Medical Informatics and eHealth
1000 Mbps (106 or million bits per second). While this provides adequate quality of
service for medium sized networks (up to say 30 users), the growing demand for
multimedia applications (including the transfer of sound, images and movies, all of
which are required in the health domain) continues to challenge the performance of the
network.
Mini desktop PCs are beginning to emerge. These have smaller footprints, and
significantly lower energy requirements than the conventional tower or desktop.
Average Power Consumption of only 29Watts, is approximately 30% less energy than
a standard PC. This is important in the reduction of the carbon footprint of computer
technology. Virtualisation lets one larger computer do the job of multiple computers,
by sharing the resources of a single computer across a network. A small energy
efficient ‘thin client’ (no hard disk and minimal peripherals) is required on the desktop.
In addition to energy savings and lower cost due to more efficient use of hardware,
virtualisation provides better desktop management, increased mobility and security,
and improved disaster recovery.
The home computer sacrifices little functionality by comparison to the desktop.
The network connection was initially via a telephone dial-up (Public Switched
Telephone Network) to an Internet Service Provider (ISP) at 56 kbps (10 3 or thousand
bits per second) but increasing is via broadband (Asymmetric Digital Subscriber Line
or cable television) at speeds typically of 1-10 Mbps for download, and 512 kbps for
upload. The speeds have increased as sophisticated signal processing yields more bits
per second from the available bandwidth.
For mobility, a number of options are available. The laptop comprises practically
all the functionality of the desktop, but with a slightly inferior specification-price ratio.
Wireless Bluetooth using radio waves or infrared light (The Infrared Data Association,
IrDA) provide unfettered desktop connectivity. Both wired Ethernet and wireless
Ethernet (known as wireless fidelity or WiFi) are normally supported. Wireless offers
nominal speeds of the order of 11-54 Mbps, depending on the characteristics of the
network access point, and the wireless interface built into the laptop. This is inferior to
the wired network speed of typically 100 Mbps. Reliability of connection is probably a
more important issue for the user, with the wired network providing better service. The
tablet PC is a variant of the laptop, with automatic handwriting, a standard input
feature. A major issue with portable computing is the capacity of the battery, typically
supporting a couple of hours of unfettered computing and hence a major design issue is
the conservation of power. This has spawned research into harnessing power from the
environment, e.g. from kinetic energy or solar power.
The personal data assistant (PDA) provides computing power which can be stored
in the pocket. This is very appealing in the area of medicine, e.g. a ward round in a
hospital where even a lightweight laptop can be obtrusive. The PDA comprises similar
connectivity interfaces to the laptop and supports similar applications to desktop such
as word processing, spreadsheet, database, calendar, Internet browsing, and email,
normally with limited functionality. However, due to the small screen size and small
keypad, usability is the major difficulty. Some models have virtual keypads, others
have small keyboards or incorporate speech recognition. Many medical applications
have been developed for the PDA, including dictionaries, calculators, guidelines and
databases, see [18].
In many respects, from technology and user perspectives, the PDA and the smart
phone have converged. The smart phone primarily provides telephony within a Global
System for Mobile communication (GSM) cell (IS-95 in USA). However the use of
P.J. McCullagh et al. / Medical Informatics and eHealth 127
data services such as Short Message Service (SMS), and Internet access using Wireless
Access Protocol (WAP) or I-Mode4 using General Packet Radio Service (GPRS) are in
widespread use. The telephone companies have introduced third generation (3G)
cellular systems based on Universal Mobile Telecommunications System (UMTS),
with data rates of 384 kbps and beyond, which can support video conferencing and
hence basic telemedicine. Further technological enhancements offering performance
approaching LAN speeds, e.g. High Speed Packet Download Access (HSPDA) are
under development (download speed of 3.6 Mbps with 14.4 Mbps planned for the
future). The smart phone also contains a processor and can be programmed for bespoke
applications. By enhancing a PDA with GPRS connectivity, the PDA essentially
becomes a telephone. Similar convergence from ICT means that a computer, laptop or
PDA connected to a ‘hot-spot’ can utilize voice over internet protocol (VOIP) to
provide telephony or ‘chat’ services.
Of course, embedded devices with microprocessor controllers are also ubiquitous
in biomedical engineering. These can utilize advances in miniaturization to perform
dedicated functions, under the control of a program. An example is an ECG monitor,
which uses purposely designed hardware and software to acquire and display the
electrocardiogram. If a new clinical measure is required, then an updated program can
be downloaded from a web site to the device firmware. Semi-automated ECG
defibrillation systems rely more heavily on embedded software for decision support.
An operating system (OS), is a computer program that manages the hardware and
software resources of a computer. The OS performs basic tasks such as controlling and
allocating memory, prioritizing system requests, controlling input and output devices,
facilitating networking, and managing files. It may provide a graphical user interface.
Microsoft Windows is the dominant OS on the PC and Apple Macintosh platforms,
with 90% market share. The main competition comes from Open Source5 protagonists
who support Unix-variants such as Linux. On larger mainframe computers Unix and
proprietary systems such as OpenVMS dominate. The latest Windows OS, Windows 7,
includes added functionality in security and network administration.
The most common operating systems used by smart phones are Symbian, Linux,
Windows Mobile, and Palm OS. Windows Compact Edition (CE) is an OS for minimal
computers, such as PDAs. It is optimized for devices that have minimal storage and
may run in under a megabyte of memory. Embedded systems normally use dedicated
operating systems and versions of Linux.
Programming environments for the smart phone and PDA are Microsoft.NET
Compact Framework and Java6. For the former, applications may be authored in
languages such as C# or Visual Basic.NET. For the latter, Java programs (know as
MIDlets) may be authored for embedded devices.
4
NTT DoCoMo's I-Mode is a wireless Internet service made popular in Japan, providing access to a number
of consumer sites
5
Open principles and practices are applied to the development of source code that is made available for
public collaboration, and released as open-source software
6
Java is an open source programming language from Sun Microsystems
128 P.J. McCullagh et al. / Medical Informatics and eHealth
Networks may be categorized by scale as: personal area networks, local area networks
and wide area networks.
A personal area network (PAN) is a computer network used for communication
among computer devices (including telephones and personal digital assistants) a few
meters (normally 10m but up to 100m is possible) apart. PANs can be used for
communication among the personal devices themselves, or for connecting to a higher
P.J. McCullagh et al. / Medical Informatics and eHealth 129
level network and the Internet. Personal area networks may be wired (USB and
FireWire) or wireless (WPAN) using technologies such as IrDA and Bluetooth. ZigBee
is the name of a specification for a suite of high level communication protocols using
small, low-power digital radios based on the IEEE 802.15.4 standard for wireless
personal area networks. ZigBee operates in the Industrial, Scientific and Medical (ISM)
radio bands; 868 MHz in Europe, 915 MHz in the USA and 2.4 GHz worldwide. The
technology is simpler and cheaper than Bluetooth.
Request for:
http://www.esem.org/index.html
Client Server
PC PC
Document index.html
returned to client
leased line, a router connects to the LAN on one side and a hub within the WAN on the
other. Network protocols including TCP/IP deliver transport and addressing functions.
The Internet is the worldwide, publicly accessible network of interconnected
computer networks that transmit data by packet switching using the standard Internet
Protocol (IP). It is a "network of networks" that consists of home, academic, business,
and government networks, which together carry various information and services, such
as electronic mail, online chat, file transfer, and the interlinked Web pages.
1.5. Security
1.6. eHealth
We will use the following case study to emphasize the importance of computers and
networking to the eHealth paradigm and relate this to the van Bemmel conceptual layer
model [2], introduced in section 2.
In the United Kingdom, stroke is the most significant cause of severe disability
affecting a quarter of a million people [21]. At six months post-stroke, around half of
patients need help with bathing, a third need help with dressing and a third need help
with feeding [22]. Rehabilitation is primarily concerned with maximising the functional
and cognitive ability of the patient combating social exclusion and increasing patient
independence [22]. The National Framework for Older People recommends that
rehabilitation should continue until maximum recovery has been achieved [24].
Research suggests that intensive, task specific and repetitive training may be necessary
to modify neural organisation [25]. However, due to cost factors, in-patient length of
stay is decreasing and outpatient rehabilitation facilities are limited. Furthermore,
current rehabilitation techniques used in hospitals and rehabilitation centres require
access to specialised equipment and a dedicated laboratory set-up [26]. Therefore there
is a strong argument for the need to develop a low-cost, accessible system that can
132 P.J. McCullagh et al. / Medical Informatics and eHealth
augment existing rehabilitation services for post-stroke patients. Currently, two main
approaches have been explored, namely tele-rehabilitation and home-based
rehabilitation.
Tele-rehabilitation, was firstly proposed in 1997 by the National Institute on
Disability and Rehabilitation Research (United States Department of Education), it is
defined as:
the use of communication and information technologies to deliver rehabilitation
services and exchange information over geographical distances [27].
SMART rehabilitation: technological applications for use in the home with stroke
patients, examined the scope, effectiveness and appropriateness of eHealth to support
home-based rehabilitation for older people and their carers [29]. It utilized Information
and Communication Technology and motion sensors to provide tele-rehabilitation
alongside home-based rehabilitation.
The design and development phases were guided by an expert panel of national
and international academics. In addition an expert group of stroke patients and their
carers contributed to developments by way of focus groups and consultations [30].
Throughout the development, neurological physiotherapists engaged with decisions
about the interface and the feedback given to the user. A biomedical engineering design
team was involved in the development of the sensor attachment to clothing ensuring
that the clothing was acceptable and usable by the stroke patients.
PC
(Level 3). The decision support platform (Level 4) operating at the base station provides
various types of feedback to user, such as movement display, exercise history and
variable measurement results. A personal computer and touch screen monitor make up
the base station which is connected to a central server providing web services to the
user. Patients and their carers can view their activities in a three dimensional (3D)
environment on the ICT platform, and can communicate with their therapists using the
web services (Level 1).
Various technologies have been applied in the area of motion tracking and
rehabilitation:
• Switches: pedometer and functional electrical stimulation (FES);
• Video systems;
• Gyroscopes (angle / rate of turn);
• Accelerometers (velocity);
• Robot arms.
In the SMART project, inertial sensors (Xsens Dynamics Technologies,
Netherlands, known as MT9) were used to track patients’ upper limb movements. The
MT9 sensor consists of a three axis accelerometer, a three axis gyroscope, and a three
axis magnetic field sensor, which can be used to measure rotation rates and linear
134 P.J. McCullagh et al. / Medical Informatics and eHealth
Figure 3: Accelerometer based motion tracking unit in Smart rehabilitation system with embedded
processing
The ICT platform consists primarily of five modules: database, interface, decision
support, communication and a feedback module.
The database module (Level 2) maintains information such as system information,
personal details, rehabilitation movement parameters; comments and instructions from
healthcare professionals and movement analysis. The system information stores a series
of questions designed to ascertain the patient’s ability to safely complete the daily
exercises. A calendar service stores a patient’s entire history of rehabilitation. The
database module also provides access to relevant personal information on the patient.
The design of the interface module was largely informed by the fact that users
would interact with the software via a touch screen display. Its aim is to provide easily
accessible and understandable menus. One such menu is illustrated in Figure 4
allowing the user to start an exercise, view their calendar and diary and obtain some
feedback on their rehabilitation progress.
The decision support module (Level 4) analyses the movement data and provides
key outcome variables relating to physical performance such as length of reach, elbow
angle. The communication module manages the transfer of information between the
base station and the central server. The feedback module provides different types of
P.J. McCullagh et al. / Medical Informatics and eHealth 135
Figure 4: Welcome screen after user login. Menus are designed using large icons to accommodate for touch
screen technology, and promote usability.
All the SMART home-based systems are linked to a central server via the public
Internet, using a virtual private network. Patients’ movement data and comments or
queries can be sent to the central server. A therapist can log into the central server to
view his/her patients’ movement data and their comments, provide instructions after
analysing their exercises and provide feedback on their queries.
There are three processes that run on the server to manage the system; MySQL
database server to control the storage of data, Apache TomCat web server (see section
2.3) to deal with Internet traffic and Matlab to provide a tool for analysing the
movement data (Level 3).
136 P.J. McCullagh et al. / Medical Informatics and eHealth
Figure 5: Software screen shot of 3D rendering providing personalized feedback to promote rehabilitation
The user groups involved in the rehabilitation system are post-stroke patients, their
carers, and therapists. The system must be as simple as possible to use, and adaptable
to individual needs. Special care must be taken with people who have had a stroke as
they have complex impairments often incorporating cognitive difficulties such as
problems with perception, attention, information processing, language and memory. In
order to provide a user friendly rehabilitation system, the user-centered design strategy
was applied throughout the projects development.
In the early stages of the project, focus groups were held with patients and
healthcare professionals to ensure that proposed technical solutions, methodology and
outputs were acceptable. In the later stages, a group of expert users provided feedback
on key aspects of the system such as user interface, type of feedback and computer
interface. These data were collected by qualitative researchers, summarised and fed
back to the engineering teams.
3. Summary
In this review, we have examined computers and communication which are drivers for
medical informatics and the eHealth paradigm. A conceptual model for medical
informatics, proposed by van Bemmel, is applicable to eHealth. Having provided the
P.J. McCullagh et al. / Medical Informatics and eHealth 137
Acknowledgement
The SMART consortium consists of Royal National Hospital for Rheumatic Diseases,
University of Bath, Sheffield Hallam University, University of Essex, University of
Ulster and The Stroke Association. We would like to thank the patients and carers, who
have given their time and support to the project. We would also like to thank Design
Futures from Sheffield, Xsens, Charnwood Dynamics Ltd. and the Stroke Association
for their support. SMART was funded under the EQUAL 4 (extend quality of life)
initiative from the UK Engineering and Physical Sciences Research Council (EPSRC).
Glossary
References
[1] D.L. Sackett, W.M.C. Rosenberg, J.A. Muir Gray, R.B. Haynes, W. Scott Richardson, Evidence based
medicine: what it is and what it isn't. BMJ 312, (1996), 71-72.
[2] J.H. van Bemmel, The structure of medical informatics. Med Informics 9, (1984), 175-180.
[3] P. Wilson, C. Leitner, A. Moussalli, Mapping the Potential of eHealth: Empowering the citizen through
eHealth tools and services. European Institute of Public Administration 2004.
[4] 4D Ultrasound, http://www.4dfetalimaging.com/, accessed Oct 2007.
[5] R. Morin, S. Langer. Speech Recognition System Evaluation. Journal of the American College of
Radiology 2(5), 449-451, May 2005.
[6] Connecting for health, http://www.connectingforhealth.nhs.uk/ accessed Dec 2007.
[7] S. Kotzin, Medline and PubMed will be able to synthesise clinical data BMJ. 324 (2002), 791.
[8] PUBMED, National Institute of Health, http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?DB=pubmed,
accessed Oct 2007.
[9] J.H. van Bemmel and M.A. Musen (eds.), Handbook of Medical Informatics, Springer 1997.
[10] W. Aspray, John Von Neumann and the Origins of Modern Computing. Cambridge, Massachusetts:
The MIT Press, 1990.
[11] G.E. Moore, Cramming more components onto integrated circuits. Electronics Magazine 19 April
1965.
[12] R. Kurzweil, The Law of Accelerating Returns,
http://www.kurzweilai.net/articles/art0134.html?printable=1, March 2001.
[13] Los Alamos National Laboratory, http://www.nigms.nih.gov/News/Results/FluModel040306, accessed
Oct 2007.
[14] M.L. Baker Supercomputer Serves as Weapon in AIDS Fight,
http://www.eweek.com/article2/0,1759,1972257,00.asp, accessed Oct 2007.
[15] Health Grids, http://www.healthgrid.org/, accessed Oct 2007.
[16] Biomedical Informatics Research Network, http://www.nbirn.net/, accessed Oct 2007.
P.J. McCullagh et al. / Medical Informatics and eHealth 139
Introduction
Medical data are at the centre of many healthcare-related activities, ranging from
providing medical care for each individual patient to supporting scientific research and
education. This review offers an introduction to medical data structures and
management. Examples of coding and classification systems are also presented.
1. Medical Data
Medical data include a broad range of data types, including text-based alphanumeric
data, recorded physiological signals and medical picture/images. In addition to
traditional clinical data, recent progress in medical science and technology has led to
the accumulation of a tremendous amount of genomic data in support of individualised
healthcare and personalised therapies. In this section, the characteristics of typical
clinical data are introduced, followed by a brief introduction to genomic data.
In general, there are three types of clinical data in the practice of medical and health
science 0, [2].
• Alphanumeric data are typically recorded as unstructured, free text in the
medical record. They account for a large portion of the information that is
gathered verbally in the care of patients, such as patient demographics, a
description of patients’ illnesses and responses to physician’s questions.
Results of physical examination such as vital signs and laboratory tests, and
other narrative descriptions, such as treatment plans and surgical procedures,
also belong to this category.
• Physiological signals, captured from the human body by using appropriate
biomedical sensors, are usually represented by a series of values along two
dimensions with the x-axis representing time and the y-axis displaying voltage
values associated with the underlying physiological activity. Advances in
physiological signal acquisition techniques have generated a vast quantity of
H. Zheng et al. / Data Structures, Coding and Classification 141
Figure 1. Illustration of an ECG signal: (a) a typical ECG wave, (b) ECG intervals. RR, QT and PR
represent the intervals between two respective wave peaks.
Advances in biological science are fostering a new clinical practice, where clinical
diagnosis and treatment will be supported by the information encoded in relevant
genomic data, such as DNA (deoxyribonucleic acid) sequences, protein structures and
microarray gene expression data. Examples of genomic applications in medical
research can be found in the Programs for Genomic Applications (PGAs) funded by the
National Heart, Lung and Blood Institute (NHLBI) of the NIH (National Institutes of
Health) (http://www.nhlbi.nih.gov/resources/pga/).
Each type of genomic data has a unique data structure. For instance, DNA may be
represented by a linear sequence from the alphabet {A, G, T, C}. Each of these
characters refers to a nucleotide. A typical microarray experiment dataset includes
expression levels of thousands of genes in a number of experimental samples
(conditions). These samples may correspond to different conditions or serial time
points taken during a biological process. An expression dataset can be summarised by a
matrix, in which the horizontal rows represent genes, one gene in each row, and the
vertical columns contain the various samples corresponding either to the time points or
to various conditions, with one sample in each column. Thus, various models of
electronic medical records have been developed to integrate traditional clinical data and
diverse genomic data 0.
The strong interest in integrating genomic data into clinical research is driven by
the hope that it will result in the prevention and diagnosis of complex genetic disease
and the design of highly targeted therapies [4]. For example, with DNA microarray
technology, scientists are able to measure gene expression of thousands of genes under
selected physiological conditions at a given time. Thus, by comparing expression
profiles of an individual gene under various conditions or by comparing expression
profiles of samples to understand their similarities and differences and to find the
relationships among genes and samples, many important physiological and medical
phenomena can be studied at the molecular level.
Apart from privacy and confidentiality requirements, medical data have important and
unique features. For example, a patient record is a mixture of opinion and factual data
generated from a wide range of clinical professionals. Some of these data, such as
observations and interpretations, are difficult to record in a structured form. Recorded
physiological signals and medical images are contaminated by noise generated from the
machine recording the data or from the patient. For instance, an ECG record could be
contaminated by physiological variability, baseline wander noise, 50 or 60 Hz power
line interference, and muscle noise 0. Due to the quantum noise inherent in photo
detection, images generated by computed tomography-based techniques such as
positron emission tomography (PET) and single photo emission tomography (SPECT)
have a high noise content 0.
The characteristics of medical data pose a great challenge to efficient data storage and
management. In this section, electronic patient records are described. The examples of
H. Zheng et al. / Data Structures, Coding and Classification 143
management and storage of medical images and physiological signals are introduced
respectively.
A patient record contains all the clinical data relevant to the care of one patient. Given
the increasing volume and scope of medical data available nowadays, traditional paper-
based patient record systems are not sufficient to address many issues related to
efficient medical data management. The Electronic Patient Record (EPR) is
revolutionising storage, management and communication of medical information. With
EPR comes all related medical data, including patient history, physical examination,
drug prescriptions, and recorded biological signals, linked into one record that can be
easily accessed for different purposes. The ability to store all patient-related medical
information in one record, which may be accessed at any time or from any place,
ultimately contributes to improvements in healthcare delivery.
It has been shown that an EPR-based system offers several significant advantages
in the process of data management and analysis 0]. One of the most obvious benefits is
that the speed of data retrieval is significantly improved. A physician should be able to
search through thousands of records for specific attributes such as medication or
diagnosis with little effort. With structured data entry, medical data collected in an EPR
are well structured and organised, allowing the electronic sharing of patient recording
among different clinical professionals. Moreover, an EPR-based record system has the
inherent flexibility to provide task-specific views on all the available patient data to
meet the needs of different clinical tasks, as illustrated in Figure 3.
Figure 3. An EPR provides an integrated view of patient data and has the flexibility to generate task-specific
views on all the available data to meet the needs of different tasks.
oriented, and object-oriented data models. Among these, the relational data model is
most widely used.
A relational data model, developed by E.F. Codd [8], is one in which the data and
relations between them are organised into a series of tables. Each table is made up of
columns and rows. All values in a given columns have the same data type, which could
be numeric or textual. All values in a row conceptually belong together. Typically, each
row, known as a record, is uniquely identified by a primary key which may contain a
single column or multiple columns in combination. Figure 4 shows an example of a
relation database with two tables: PatientDemographics and PhysicalExamination
tables.
the resolution of imaging systems, therefore we can expect to see the computer network
and storage requirements also increase.
The main DICOM functions are to transmit and store medical images, to enable
searching and retrieval, to facilitate printing of hard copy, to support workflow
management and to ensure quality and consistency of image presentation 0. The
DICOM file format is well defined and contains two specific sections. The first part of
the file, called the header, holds patient demographics, hospital information and details
of the referring physician. Also, the header contains detailed data on the actual specific
imaging modality parameters for each patient scan. For example, in a magnetic
resonance (MR) image data file, the slice thickness, field of view, pulse sequence
details, table position, and scan date are stored. The header also contains hidden fields
used by the equipment vendor to store proprietary information on image reconstruction
algorithms. Another feature of the DICOM header is that it contains a unique
identification number that identifies the scanner manufacturer, the individual scanner,
the modality types and the patient study number. This is called the unique identifier
(UID). DICOM data files written to storage media like DVD or to another computer
will be named using the UID as part or the entire file name.
Figure 5 shows a screen shot of part of a DICOM image header from OSIRIS
developed by The University Hospital of Geneva, Switzerland. Osiris is a free DICOM
viewer available at http://www.sim.hcuge.ch/osiris/01_Osiris_Presentation_EN.htm.
The second part of the DICOM data file is the image pixel data where pixel values are
stored using between 4 and 32 bits, depending on the imaging modality. A typical file
size for MR is 0.5 MBytes (one patient examination may contain up to 1000 images),
whilst a Computed Radiography chest X-ray may be 32 MB. The DICOM standard
includes the functionality to compress medical images using lossless or “lossy”
techniques 0. JPEG 2000 is the common standard achieving a lossless compression of
approximately 3:1. Further compression is possible, however, there is some reluctance
to reduce file size further as diagnostic integrity of the original image may be lost after
significant compression.
The benefit of DICOM is the ability to connect a number of different vendor
medical imaging modalities to a Picture Archive and Communication System (PACS).
This is the main image database and transmission system in the health environment.
This system enables multiple copies of images to be available at any time and at any
place that has the appropriate network connections and permissions. PACS also enables
the creation of large digital libraries for teaching and research.
2.3. SCP-ECG
Due to its ability and flexibility to define markups for specific type of data, XML
(eXtensible Markup Language, http://www.w3.org/xml/) and its related techniques
have gained great attention in medical data representation and management. Currently,
various organizations within healthcare such as HL7 (health level 7) and ASTM
(American Society for Testing and Materials) are working on recommendations of
XML-based e-healthcare technologies within the medical domain. For example, HL7
has published the Clinical Document Architecture (CDA), which uses XML as the
representation format for medical and administrative data
(http://xml.coverpages.org/ni2004-08-20-a.html). Along with the CDA, HL7 version 3
messaging standard has also been developed using XML as the standard exchange
format. The use of XML syntax for the exchange of electronic patient records was
illustrated in the EU project Synapses (https://www.cs.tcd.ie/synapses/public/) and its
implementations [12]. The application of XML in DICOM can be found in several
related projects [13]. The U.S Food and Drug Administration (FDA) Centre for Drug
Evaluation and Research has proposed recommendations for the XML-based
techniques for the exchange of time-series data. In December 2003, the FDA XML
format was finalized and the Annotated ECG (aECG) format is now part of the HL7
family of standards
(http://www.hl7.org/v3annecg/foundationdocuments/welcome/index.htm).
Each chapter is divided into blocks and categories. Unlike ICD-9, in which
diseases are coded with three digits, category codes in ICD-10 start with an
alphabetical character followed by two numeric characters, ranging from A00 to Z99.
For instance, code I01 represents rheumatic fever with heart involvement. This greatly
increases the number of available codes, which can be used to code common diseases
in more detail.
Most 3-character categories in ICD-10 are further subdivided into up to 10
subcategories by adding a fourth digit after a decimal point. This can be used, for
instance, to classify varieties of a disease, e.g. as illustrated in Figure 7, acute
rheumatic pericarditis is coded with I01.0, while acute reheumatic endocarditis is
represented by I01.1.
Read Codes were initially introduced into the UK in the middle 1980s as a 4-byte set
terminology to summarise patient treatment, including clinical process and
administrative data, in primary care. This version only consists of about 40000 codes.
Its subsequent Version 2 released in 1990 was based on the content of ICD-9 and
Classification of Surgical Operations and Procedures, Fourth Revision (OPCS-4). Read
Codes in this version were extended to 5-digit alphanumeric characters, each
representing one level of hierarchy. While being easy to understand and implement,
such a rigid hierarchical classification structure poses some difficulties in the
H. Zheng et al. / Data Structures, Coding and Classification 149
representation of clinical terms. For example, without giving two Read Codes, a
clinical term can only be located within one hierarchy.
Due to increasing demands from clinical professionals, Read Codes were further
expanded by the Clinical Terms Project, a joint project established by the UK’s
Conference of Medical Royal Colleges and the government’s National Health Service
(NHS). As a result, the Clinical Term Version 3 (CTV3) was developed and first
released in 1994.
In CTV3, a concept refers to any clinical disorder, procedure or observation and
can be described by more than one clinical term. Like its previous version, CTV3
adopts a 5-digit alphanumeric code, giving more than 650 million possible codes to
represent clinical concepts. However, CTV3 implements a new, more flexible structure
to address the problems caused by its early version. For example, the coded-based
hierarchical structure has been abandoned. A Read Code is only used to uniquely
identify a concept and no longer represents its hierarchy position. Instead, hierarchical
relationships between concepts are defined by a set of parent-child links described in a
Hierarchy file.
Core release files included in CTV3 are Concept file, Term file, Descriptions file,
Hierarchy file, key file, Specialty files, Cross-mapping files, Template file, Redundant
codes mapping file and Description Change file. The structures and examples of
Concept file (describing concepts), Term file (describing terms), Descriptions file
(linking terms to Read Codes) and Hierarchy file (describing a hierarchy) are illustrated
in Figure 8. The reader is referred to [14] for a more detailed description of CTV3.
Figure 8. The structures and examples of (a) Concept file; (b) Term file; (c) Descriptions file; and (d)
Hierarchy file.
The current version of SNOMED CT includes more than 300,000 clinical concepts,
770,000 descriptions and 900,000 semantic relationships. SNOMED CT can also map
to other medical classification systems such as ICD-10.
The ecgML model, a markup language for supporting representation and coding of
ECG data, was first published in 2003 [15]. Based on XML, ecgML offers several
advantages over existing ECG coding systems. For example, it includes all the
components required for ECG representation. The terms and structure used can be
further expanded or reviewed at any stage. This protocol can be used as a common
coding platform between different ECG acquisition devices and visualization programs.
152 H. Zheng et al. / Data Structures, Coding and Classification
ECG waveform data are a key component of an ECG record. Given that a wide range
of ECG recoding devices are available today, the size of waveform data can vary
dramatically. Moreover, they are usually expressed in binary format. In order to
accommodate the full spectrum of ECG data, ecgML provides the following three
options to handle ECG waveform summarized in Figure 12.
• For large data files, typical of Holter recording, ECG data are maintained in an
external file, which can be referenced in ecgML using a FileLink element;
• Using RealValue elements to directly include actual waveform values in
ecgML as the content of the element;
154 H. Zheng et al. / Data Structures, Coding and Classification
5. Summary
The above coding systems are developed for a specific purpose - they are goal-oriented.
For example, ICD takes a pathophysiological and aetiological view of medicine and
classifies diseases primarily according to the organ system involved with some
important exceptions such as infectious diseases [18] It is unable to fulfill the needs of
all users.
Another problem is that the terms are changing over time, thus the meaning of a
same term can be different over time, new terms are created and old terms may be
replaced. When the meaning of a standard code is changed, the data will be interpreted
incorrectly. When a standard code is removed from the coding system, the data are no
longer interpretable. Additionally there is a lack of comprehensive, standardized
medical terminology, and highly skilled experts are required to interpret the systems.
Despite these limitations, coding and classification systems have been successfully
applied in health care. A widely accepted coding system is essential for storing and
exchanging patient records efficiently.
Glossary
References
[1] E. Shortliffe, L. Perreault, G. Wiederhold, L. Fagan, eds., Medical Informatics: Computer Applications
in Health Care and Biomedicine, New York: Springer-Verlag;Second Edition 2001.
[2] W. Tompkins, Biomedical Digital Signal Processing: C lauguage Examples and Laboratory
Experiments for the IBM PC. Englewood Cliffs, N.J.; London : Prentice Hall, 1993.
[3] B. Robson, R. Mushlin, Genomic Messaging System and DNA Mark-Up Language for Information-
Based Personalized Medicine with Clinical and Proteome Research Applications. J. Proteome Res. 3(5),
(2004) 930 – 948.
[4] E. Coiera, Guide to Health Informatics. Second Edition. London: Hodder Arnold, 2003.
[5] Y. Pu, R. Patterson, Comparison of R-wave detection errors of four wireless heart rate belts in the
presence of noise. Physiol. Meas. 24, 913-924, 2003.
[6] P. Razifar, M. Lubberink, H. Schneider, B. Langstrom, E. Bengtsson, and M. Bergstrom. Non-isotropic
noise correlation in PET data reconstructed by FBP but not by OSEM demonstrated using auto-
correlation function. BMC Medical Imaging, 5:3, (2005).
[7] J. H. van Bemmel, M.A. Musen, Handbook of medical informatics. Heidelberg, Germany ; AW
Houten, Netherlands : Springer Verlag, 1997.
[8] E.F. Codd, A relational model of data for large shared data banks. Communications of the ACM, 13(6),
(1970), 377–387.
[9] A.E. Flanders and J.A. Carrino, Understanding DICOM and IHE. Seminars In Roentgenology, 38(3),
(2003) 270-281.
H. Zheng et al. / Data Structures, Coding and Classification 157
[10] Graham RNJ, Perriss RW, Scarsbrook AF. DICOM demystified: A review of digital file formats and
their use in radiological practice. Clinical Radiology 60, (2005), 1133 – 40.
[11] ENV 1064 standard communications protocol for computer-assisted electrocardiography. European
Committee for Standardisation(CEN), Brussels, Belgium, 1996.
[12] B. Jung, E.P. Andersen, J. Grimson, Using XML for Seamless Integration of Distributed Electronic
Patient Records. In Proceedings of XML Scandinavia 2000 conference, Gothenburg, Sweden, May
2000.
[13] A. Tirado-Ramos, J. Hu, K.P. Lee. Information object definition-based unified modeling language
representation of DICOM structured reporting: a case study of transcoding DICOM to XML. J Am Med
Inform Assoc, 9(1), (2002), 63-71.
[14] http://www.connectingforhealth.nhs.uk/terminology/readcodes/, accessed Mar 2008.
[15] H.Y. Wang, F. Azuaje, B. Jung, N.D. Black, A markup language for electrocardiogram data acquisition
and analysis (ecgML). BMC Medical Informatics and Decision Support, 3 (4), 2003.
[16] H.Y. Wang, B. Jung, F. Azuaje, N.D. Black. ecgML: Tools and technologies for multimedia ECG
presentation. In the Proc. of XML Europe Conference, London, United Kingdom, May 2003.
[17] H.Y. Wang, F. Azuaje, G. Clifford, B. Jung, N.D. Black, Methods and tools for generating and
managing ecgML-based information. In the Proc. of 2004 Computers in Cardiology, IEEE Press,
Chicago, September 2004.
[18] http://www.opengalen.org/, accessed Mar 2008.
158 Basic Engineering for Medics and Biologists
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-158
Modern commercial clinical DSSs have evolved from the work of early investigators
who set about developing systems that could flag the presence and genre of disease
using the computational tools available at that time. An example of such a system was
that developed by DeDombal in 1972 [1]. This system, designed to diagnose acute
abdominal pain, used statistical techniques to suggest to which of 13 diagnostic
categories a patient was most likely to belong. Although most early systems utilised
statistical techniques, investigators soon began applying other methods. An example of
this was in the implementation of MYCIN [2], a rule-based system designed to
diagnose and recommend treatment for certain blood infections and other infectious
diseases. In this system clinical knowledge was modelled as a set of IF-THEN rules
with certainty factors attached to diagnoses. More recent systems have also been
developed with the ability to provide support over a comprehensive range of
conditions. The INTERNIST system [3] was capable of providing hundreds of
D.D. Finlay et al. / Mining, Knowledge and Decision Support 159
diagnoses in internal medicine, based on thousands of symptoms. This has since been
superseded by Quick Medical Reference (QMR) which contains information on almost
700 diseases and more than 5,000 symptoms. Iliad, DXplain and DiagnosisPro are
similar interactive consultation systems, which use clinical findings to produce a
ranked list of diagnoses which might explain the clinical manifestations.
In Table 1, adapted from [4], likely examples of DSS utility are presented. From this
table it is evident that DSSs can be designed to provide a wide range of functionality
and hence clinical support. We want, however, to give further consideration to two of
the most common and hence important genres of functionality, namely “classification”
and “prediction”. These forms of support are analogous to the processes of diagnosis
and prognosis respectively.
1.2.1. Classification
Classification is the term used to describe the process of analysing data to form
decisions regarding the class membership of individual instances. An instance can be
thought of as the individual case or record that is observed while a class relates to the
possible categories to which the instance can belong. For example, consider a patient
who has arrived at the emergency department with chest pain. This chest pain could
result from some form of cardiac pathology or perhaps be attributed to less serious
causes such as indigestion or a panic attack. The information collected through the
medical assessment of the patient is used to describe the instance. Using this
information, a diagnosis is made which places the instance into some predefined class,
such as those outcomes outlined above. The actual diagnosis is formed by comparing
and locating similarities from the details of the medical assessment with previously
documented assessments for which there exists a known diagnosis.
1.2.2. Prediction
Similar to classification, prediction involves the analysis of measured information in
order to form a decision. While classification is synonymous with diagnosis, prediction
can be compared to prognosis. In this sense the analysis of information recorded to date
is used to produce a forecast of the expected outcomes in the future or over time.
When dealing with chronic diseases, for example, medical staff often refer to the
records (instances) of patients who have, in the past, suffered from the same disease.
These cases can then be used to predict, for example, the life expectancy or treatment
outcomes for current patients.
Regardless of functionality rendered, all decision support systems are based upon the
same underlying processes. These processes are summarised in the block diagram
presented in Figure 1, which consists of three stages. The first stage involves the
measured data and its treatment and should result in useful ‘features’. The second stage
involves the interpretation of these features to provide some outcome. The last stage
considers the relevance of the outcomes with some quantification of performance.
160 D.D. Finlay et al. / Mining, Knowledge and Decision Support
Table 1: Examples categories of clinical decision support systems, after Coeira [1]
2. 1. Features
Measured/
recorded Decision Outcome
data
The interpretation algorithm can be considered the core component of a DSS as it must
take the presented features and make a decision based upon their values. Thinking
again of the simple example of analysing blood glucose levels for the presence of
diabetes, a very simple algorithm could be implemented that would flag the presence of
diabetes if a certain threshold has been reached. The threshold which dictates the
presence of this disease can easily be specified by anyone with a prior knowledge of
the domain. In more complex diseases, where there are perhaps many features and a
less comprehensive knowledge of the disease process, development of the decision
algorithm is more difficult. Specifically, in cases where the meaning of the features and
the relationships between them are not obvious, it is necessary to develop an algorithm
that will learn how to make decisions based on a pool of exemplar data. In this case,
we say that the knowledge that is contained within the decision algorithm is ‘learned’
from the existing data. The process of ‘learning’ knowledge involves exposing an
algorithm to a set of ‘example cases’ that represents the classification task at hand.
Subsequently, when unseen data in the form of ‘test cases’ are presented, the algorithm
will suggest which group or category each case belongs to, based on the previous
learning. There are many ways in which interpretation algorithms can learn, these are
broadly categories as either ‘supervised’ or ‘unsupervised’.
2.2.1. Supervised Learning
In the supervised approach, the algorithm is exposed to a number of example cases.
Each will consist of a set of features along with the actual outcome that was observed
when those features were measured. After application of a suitable number of training
examples, the algorithm is considered ‘trained’. Subsequent to this, when the algorithm
is exposed to a new example case which will consist of an unseen set of features, it
should be able to indicate what the output would be. The number of training examples
required to adequately train the algorithm relies on many factors. These include the
type of algorithm used, the number of features and the overall complexity of the
problem. The final performance of the algorithm is also dependent on many such
factors. A diagram illustrating supervised learning is displayed in Figure 2.
Actual Desired
Respons Respons
Input Algorithm e e Teacher
Input and
Error desired
output
2.3.1. Accuracy
When trying to understand how well a given interpretation algorithm performs it is
desirable to quantify how well it predicts the presence or absence of disease. This
quantification is based on a comparison with the actual or true outcomes. The easiest
way to achieve this is to calculate the percentage of cases that have been correctly
classified. Refereed to as ‘accuracy’ (ACC), this can be defined as [9]:
x
ACC(%) = 100 × (1)
N
where x is the number of subjects that have been correctly classified and N is the total
number of subjects in the dataset. Although this measure gives an indication of overall
interpretation performance, no information can be attained from this metric regarding
the breakdown of the performance of a particular group within the dataset, e.g. how
many ‘normals’ (i.e. people without the disease) have been correctly assigned, or how
many ‘abnormals’ (i.e. people with the disease) have been correctly assigned. To
achieve this further measures of ‘sensitivity’ and ‘specificity’ are required.
2.3.2 Sensitivity and Specificity
Sensitivity (SEN) is a measure of how well a given interpretation algorithm can predict
the presence of disease. Specificity (SPE), on the other hand, is how well a given
interpretation algorithm can predict the absence of disease (SPE).
negative by the given algorithm. In general a high level of SEN will be exhibited by an
algorithm that performs well in the detection of presence of disease, a high level of
SPE shall be exhibited by a classifier that performs well in the detection of absence of
disease.
3. Interpretation Algorithms
3.1.1. Statistical
The main objective of the statistical approach is the allocation of a test case to one of a
number of possible diagnostic categories, with minimum probability of miscalculation
[11]. When using such a classification technique, the background knowledge which is
known is used to manipulate the data into a number of groups within the data set as
opposed to using the knowledge to design the ‘rules’ for the algorithm itself. Once the
data has been arranged, it is then possible to calculate a number of statistical measures
relating to how likely a specific feature is in belonging to a specific category. These
types of ‘probability’ measurements form the underlying basis of the statistical
interpretation process.
3.1.2. Induction Algorithm
An induction algorithm builds a set of rules from a set of training examples which can
then be used to assign unseen examples. After development, these rules are often
represented as a ‘decision tree’ structure. In this type of system, a hierarchy of decision
nodes connected by branches forms a decision tree. Each node represents a binary test,
the result of each test determines the path through the tree network. After the rule set
has been established, this type of algorithm is easy to implement, although the
induction algorithms themselves are complex. The main disadvantage of the resulting
rule set is their rigid nature i.e. only a very strict set of conditions can be tested for with
sharp boundaries between normal and abnormal.
3.1.3. Neural Networks
Neural Networks (NNs) are vast interconnections of elements called neurons. The
biological neuron itself is a building block of the brain and in NNs an attempt is made
to replicate the behavior of these neurons. An artificial neuron consists of a number of
inputs which are weighted before reaching a main body or processing unit, the signal is
processed and passed to an output. The operation of the NN depends greatly on how
the neurons are interconnected. After design, a NN must be trained. During this process
the NN adjusts its own internal parameters causing the network to converge to the
desired response. Although in this section we are focusing on supervised techniques,
NNs can also be configured to work in an unsupervised manner.
D.D. Finlay et al. / Mining, Knowledge and Decision Support 165
some cases, it could be difficult to predict what the number of k should be. Fixed
number of clusters can make it difficult to predict what k should be.
Figure 3. A typical SOM architecture consisting of two layers: an input layer with i input samples and a 2D
layer which contains a grid of neurons. The rectangle in dotted lines represents a topological neighborhood,
Nc(t), of winning neuron, Wj, which is monotonically decreased as training progresses.
In the introductory paragraphs of this article we have discussed DSSs mainly in the
context of clinical care. There is also scope for utilizing DSSs in everyday life to
support independent living [19]. We now live in an age where we can take more control
of our own health by using readily available devices which support personal vital sign
monitoring. The availability of such devices provides the opportunity for the
deployment of decision support systems outside the boundaries of primary or
secondary healthcare establishments. If we consider very basic home based measuring
168 D.D. Finlay et al. / Mining, Knowledge and Decision Support
devices such as the blood pressure monitor or blood glucose monitor, it is possible to
add a simple algorithm which would sound an alarm if an excessive measurement were
recorded. Obviously this intermediate form of assessment is merely a pre-cursor to
assessment by a healthcare professional; however, it is a step forward in terms of the
automated analysis of vital signs outside of the conventional clinical setting. In the
following paragraphs we discuss three, albeit closely related, areas which we believe
define new trends and challenges in decision support.
With all of the aforementioned scenarios, the decision support systems will be required
to analyse information from a number of different sources. Given complementary
advances within the telecommunications domain it is possible for all of the information
to be analysed on the device or system within the home environment or relayed to some
central healthcare facility. The advantage of the latter is the remote support which may
be offered to a person within the comfort of their own home. For example, a decision
support system based in a hospital may receive information regarding a possible alarm
for a person’s high blood pressure. In addition, the system may receive information
D.D. Finlay et al. / Mining, Knowledge and Decision Support 169
suggesting the person has not taken their anti-hypertensive medication. Based on this
information the decision support system could raise an alarm resulting in a healthcare
professional calling the person at home to check on their general status. Although this
is a simplistic scenario, it conveys the end-to-end element of healthcare monitoring and
delivery which is directing the future development and deployment of medical decision
support systems.
DSSs have the ability to assist the decision making process in many applications. These
range from classical clinical decision making to more contemporary applications
associated with independent living. The developments in DSSs have also spurred the
development of elaborate computational techniques and algorithms which in turn have
created new opportunities evident in the emergence of new domains such as knowledge
discovery and data mining. These domains and the tools used have also opened up new
possibilities for gaining an understanding of and streamlining the process of healthcare
delivery. In this case study we look at how some of the techniques that form part of the
DSS process can be utilized for solving less conventional problems. We approach this
discussion not by focusing on the technical rigors of this application, but on the
application and need itself.
A medical discipline that has gained a lot from these developments is cardiology,
particularly electrocardiology. This particular area saw one of the first applications of
computers in medicine [20], and computerized electro cardiology is now a domain in
its own right [21]. The application of computers in this area is widespread with
computerized enhancement of the acquisition process right through to computerized
interpretation and diagnosis being commonplace. In more recent years computers and
particular data mining techniques have been applied in an attempt to gain more of an
understanding of the principles of electrocardiology and of the underlying
electrophysiology. To provide a typical example of how these techniques can be
applied we look at the application of data mining techniques in the selection of
electrocardiographic recording sites. The most effective way to capture
electrocardiographic information is to record information from recording sites which
are positioned all over the torso. Such an approach is referred to as body surface
potential mapping and can involve recording information from in excess of 200
recording sites. This technique is deemed to provide the most diagnostic information,
as effectively all information as projected on to the body surface is recorded. Although
comprehensive, this technique is seldom used in clinical practice due to the excessive
number of recording sites required and the associated complexity of the hardware. A
more favored technique is the 12-lead ECG which as the name suggests renders just 12
channels of information for interpretation. Although widely used, the 12-lead ECG is
also met with some skepticism as it is well appreciated that the locations of the
recording sites are known to be sub optimal. There therefore is a very valid research
question of where should electrocardiographic information be recorded from a patient’s
torso to increase diagnostic yield. Assuming that something around the number of
recording sites that are required to record the 12-lead ECG is suitable, many
investigators have set about analyzing body surface potential maps to find which of the
200+ recording sites available yield the most information. Historically this analysis has
170 D.D. Finlay et al. / Mining, Knowledge and Decision Support
been conducted using statistical methods [22]. However, recently more contemporary
techniques have been endorsed.
The problem of recording site selection can be described as trying to locate the
leads in a body surface potential map that yield the most diagnostic information. If we
take this a step further and think of this problem in the context of automated
classification, we can think of the information from each electrocardiographic lead as a
feature and we want to reduce the number of these features prior to classification. We
have therefore defined a ‘feature selection’ problem. As described earlier, feature
selection deals with the removal of redundant variables, and the retention of those
which are deemed more useful. This is in contrast to feature extraction, which
transforms the features and hence will not tell us which electrocardiographic leads we
should measure.
Regardless of the problem domain, feature selection would involve exhaustively
evaluating all possible combinations of input features and choosing the best subset. In
reality, the computational cost will be prohibitive if there are many features to be
considered. This is likely to be the case in the ECG lead selection problem where, for
this example, we have in excess of 200 features and hence leads. For this reason, there
has been an interest in developing algorithms and strategies that locate optimal features
at low computational cost [23].
6. Summary
Glossary
ACC Accuracy
BSPM Body surface potential map
DSS Decision support system
EPR Electronic patient records
MRI Magnetic resonance imaging
NN Neural Network
QMR Quick medical reference
SEN Sensitivity
SOM Self organizing map
SPE Specificity
D.D. Finlay et al. / Mining, Knowledge and Decision Support 171
References
[1] F.T. de Dombal, D.J. Leaper, J.R. Staniland, A.P. McCann and J.C. Horrocks, Computer aided
diagnosis of acute abdominal pain, British Medical Journal 2 (1972), 9-13.
[2] E.H. Shortliffe, Computer Based Medical Consultations: MYCIN, Elsevier, North Holland, New York,
1976.
[3] J.D. Myers, The Background of INTERNIST-I and QMR, in A History of Medical Informatics, eds.
Bruce I. Blum and Karen Duncan (New York: ACM Press, 1990), 427-433.
[4] E. Coiera, Guide to Medical Informatics, The Internet and Telemedicine, Chapman & Hall Medical,
1992.
[5] N. Wyse, R. Dubes, and A. K. Jain, A critical evaluation of intrinsic dimensionality algorithms, in
Pattern Recognition in Practice, E. S. Gelsema, and A. K. Jain, Eds. Morgan Kaufmann Publishers,
Inc, 1980, 415-425.
[6] L. Huan, and H. Motoda, Feature transformation and subset selection. IEEE Intelligent Systems, 13
(1998), 26-28.
[7] A. Jain, and D. Zongker, Feature selection: Evaluation, application, and small sample performance,
IEEE Transactions on Pattern Analysis and Machine Intelligence 19 (1998), 153-158.
[8] J. Reunanen, Overfitting in making comparison between variable selection methods, Journal of
Machine Learning Research, 3 (2003), 1371-1382.
[9] C. D. Nugent, PhD. Dissertation, University of Ulster, 1998.
[10] P. M. Rautaharju, H. W. Blackburn, and J. W. Warren, Review: The concepts of sensitivity, specificity
and accuracy in evaluation of electrocardiographic, vectorcardiographic and polarcardiographic criteria,
Journal of Electrocardiology, 9 (1976), 275-281.
[11] J. A. Kors, J. H. van Bemmel, Classification Methods for Computerised Interpretation of the
Electrocardiogram, Methods of Information in Medicine, 29 (1990), 330-336.
[12] J.C. Sharkley, Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems. London:
Springer, 1999.
[13] T. Kohonen, Self-Organising Maps. Heidelberg, Germany: Springer-Verlag, 1995.
[14] G. Bortolan and W. Pedrycz, An interactive framework for an analysis of ECG signals, Artificial
Intelligence in Medicine, 24 (2002), 109-132.
[15] T. S. Douglas, S. E. Solomonidis, V. S. P. Lee, W. D. Spence, W. A. Sandham and D. M. Hadley,
Automatic segmentation of magnetic resonance images of the trans-femoral residual limb, Medical
Engineering & Physics 20 (1999), 756-763.
[16] K. Simelius, M. Stenroos, L. Reinhardt, J. Nenonen, I. Tierala, M. Mäkijärvi, L. Toivonen, and T.
Katila, Spatiotemporal characterization of paced cardiac activation with body surface potential mapping
and self-organizing maps, Physiological Measurement 24 (2003), 805-816.
[17] M. B. Eisen, P. T. Spellman, P. O. Brown, and D. Botstein, Cluster analysis and display of genome-
wide expression patterns, Proc Natl Acad Sci USA, 95 (1998) 14863-14868.
[18] T. R. Golub, D. K. Slonim, P. Tamayo, C. Huard, M. Gassenbeck, J. P. Mesirov, H. Coller, M. L Loh,
J. R. Downing, M. A. Caligiuri, C. D. Bloomfield and E. S. Lander, Molecular classification of cancer:
class discovery and class prediction by gene expression monitoring, Science 286 (1999), 531-537.
[19] C. D. Nugnet, J Augusto, Smart Homes and Beyond. Amsterdam, Netherlands: IOS Press, 2006.
[20] R.J.A. Schijvenaars, Intraindividual variability of electrocardiogram. Assessment and exploiuting in
computerized ECG analysis, PhD Dissertation, University Rotterdam, 2000.
[21] www.isce.org, accessed Mar 2008.
[22] F. Kornreich, P. M. Rautaharju, J. Warren, T. J. Montague, and B. M. Horacek, Identification of best
electrocardiographic leads for diagnosing myocardial infarction by statistical analysis of body surface
potential maps, American Journal of Cardiology, 56 (1985), 852-856.
[23] D. D. Finlay, C. D. Nugent, P. J. McCullagh, and N. D. Black, Mining for diagnostic information in
body surface potential maps: A comparison of feature selection techniques, Biomedical Engineering
Online, 4 (2005) 51.
172 Basic Engineering for Medics and Biologists
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-172
now larger and hence places a larger demand on healthcare services. Secondly, there is
an increased prevalence of healthcare problems and instances of disease within the
elderly and thirdly there are now fewer younger people to care for the elderly [2].
Taking all of this into account has resulted in healthcare providers, governments
and members of society, searching for new paradigms of healthcare delivery.
Population Catergories
85+ 2030
70-80
2000
40-50
1950's
0-10
Population (Males) Population (Females)
Figure 1 Estimated profile of age categories in 1950, 2000 and predictions for 2030.
We have previously introduced the concept that the increase in the size of the
population has a direct impact on the prevalence of instances of healthcare problems
and of long term chronic diseases. In addition to placing a resource burden on primary
and secondary healthcare organisations, this increase in population size has additional
societal and economic implications. There are benefits from both perspectives which
would support the introduction of remote healthcare monitoring.
From a societal perspective, people are wishing to play a more active role in their
own healthcare. This can be facilitated, at least notionally, via the realisation of remote
healthcare monitoring where the patient is given a higher degree of responsibility to
become more proactive with the management of their own healthcare and its
assessment. Involvement in such a manner can also lead to the patient having a greater
understanding of not only their condition, but also its means of assessment. In addition,
there is the general appreciation that people recover more quickly in their own homes
and would prefer to remain in their own homes, even if they are at greater risk [3].
Remaining at home for an extended period of time can also be linked to improvements
in the patient’s perceived quality of life. Taking this into account, the use of remote
healthcare monitoring has the potential to avoid instances of institutionalisation when
the patient requires high levels of healthcare assessment. It can also avoid lengthy and
174 C.D. Nugent et al. / Remote Healthcare Monitoring and Assessment
inconvenient trips being made for the purposes of routine clinical assessments. Patients
can also benefit from remote expert clinical diagnosis. For example, it is possible to
send a patient’s clinical details from one institution to another and receive expert
clinical assessment via remote means.
In addition to the societal benefits, deployment of technology to support healthcare
assessment via remote means has a number of economic benefits. If a person can have
their healthcare assessed within their own homes, removing the requirement of
temporary or permanent institutionalisation, then significant healthcare savings can be
gained. Additionally, those patients receiving clinical care within a hospital setting can
be discharged in a much shorter space of time and continue to have routine assessments
conducted via remote means. Finally, consultants can offer remote consultations to
patients all over the world without having to leave their own office. This offers benefits
to the patient in terms of them receiving treatment from 'world class' specialists who,
because of geographical distances, would otherwise not be able to provide such
treatment.
Traditionally, the means by which those involved in healthcare provision interact with
their patients is based on a bi-lateral communication model (Figure 2 (a)). In this
respect it is normal for the patient to have direct contact with a number of healthcare
providers, however, commonly there is a potential lack of a unified infrastructure to
support communication between healthcare providers. Deployment of technology and
the establishment of care models facilitates communication channels to be established
between both patients and healthcare providers as well as intercommunication between
the healthcare providers (Figure 2 (b)). This offers an improved means of healthcare
delivery from the patient’s perspective as all those involved in the delivery of their care
can interact.
Physician
Physician
Professional
Healthcare
Carer
Patient
Healthcare
Patient
Professional
Pharmacist
Although the aforementioned care models can be used to facilitate the deployment
of technology, their conceptualisation can at times be faced with resistance. The
deployment of technology within the realms of healthcare provision can offer benefits
to the patients, however, the changes of practice which may be required from the
healthcare providers’ perspective, in addition to the problem of new service integration,
creates major challenges concerning the widespread deployment of the technology. On
the one hand, drivers such as the benefits of improved treatment results and alleviation
of the problems associated with the shortage of physicians and nurses encourage such
developments. Nevertheless, such innovations are faced with the constraints such as
technology costs, reimbursement issues and legal issues, to name but a few. This
inevitably results in a ‘push-pull’ scenario (Figure 3) from a technological perspective.
The development of new technology results in new paradigms being pushed into the
healthcare market. These can be considered as high risk solutions. If, on the other hand,
a user driven approach is adopted, a pull effect can be witnessed. This may be viewed
by some as the desired effect, where the end solution will be based on mature
technology, which, from a technological perspective may not offer high levels of
innovation, however, from the healthcare perspective can be seen as a low risk solution
meeting the needs of the end users. In order to provide the patient with the best solution
to their needs, a user centered design process offers many benefits. Adopting this
approach allows the patient to be involved during the early stages of the design process.
s
iont
ca
Push Effect
pli
Ap
new solutions
Pull Effect
gy
mature solutions
n
ch
Te
w
Ne
The Internet is one of the most widely used and accepted forms of modern day
communication, as well as a rich source of information. Patients can now seek
healthcare advice, research their symptoms and determine possible diagnosis, in
addition to researching available courses of treatments. Additionally, the Internet can
be used as a forum to support communities offering discussion groups dealing with
health and healthcare topics. Furthermore, the Internet offers great potential for
distributed healthcare, which is important especially for sparsely populated rural areas,
and for people with limited mobility.
The Internet makes it possible for user-centric healthcare, where the user’s needs
are in focus. More demands are also placed on being able to locate the necessary
information and help in user understanding. This has created more demands on
information and communication resources, as well as on available healthcare
professionals. Nevertheless, this also means that for many health-related issues, the
user can receive help from centralized healthcare resources. For example, a parent
concerned about a child's skin rash could be taught how to acquire high-resolution
pictures of the affected area and subsequently forward such pictures to a centralized
healthcare service. This approach would offer a fast and efficient means of receiving
direct feedback regarding the condition of the skin, without the necessity of the parent
and child having to travel to the hospital for a physical examination.
2. Telemedicine
This Section aims to introduce the notion of Telemedicine and how technology can be
deployed to link healthcare professionals together in an effort to improve the quality of
the healthcare delivery process.
Telemedicine has received many definitions over the years. Most notably it can be
described as the remote use of medical expertise at the point of need [4]. This results in
the use of telecommunications to facilitate the sharing of medical knowledge and the
delivery of healthcare over a distance [5]. With this approach medical expertise can be
delivered to remote sites, rural areas or understaffed regions. lthough Telecare may also
be considered as a form of Telemedicine, for the purposes of this article we will treat
these two topics individually and focus within this Section on the element of
Telemedicine and the associated technology which fosters a collaborative working
environment between medical experts.
Telephony: A rudimentary means of providing telemedicine is via telephone
communications. This can be in the form of two specialists discussing a patient's
diagnosis or prognosis over large distances or similarly can relate to a consultation
between a patient and a specialist. Other examples include automated services which
patients access, via the telephone, to provide details regarding their current state of
health. An example of this type of system is an automated diabetes service which
permits patients to verbally enter information regarding their condition via an
automated telephone system [6]. Such a system facilitates the routine assessment of a
C.D. Nugent et al. / Remote Healthcare Monitoring and Assessment 177
patient’s condition without the need for direct consultation thus alleviating pressure on
healthcare services.
Video Conferencing Systems: Video Conferencing Systems have become of one the
most popular platforms used to facilitate telemedicine and subsequently improve
communications between healthcare professionals [7]. Such systems typically include
the ability to send and receive video, audio, and text between one or more users.
Additionally, some systems provide tools, such as shared whiteboards, which facilitate
the communication of ideas and thoughts via informal means. The main purpose of
using video conferencing systems in telemedicine is to provide a sense of presence
through the provision of visual instructions, or demonstrations to patients, and to
visually present or describe problems and symptoms to healthcare professionals.
Video conferencing provides a wide range of services that healthcare providers can
avail of for example nurses can use video conferencing to make house-calls while
keeping contact with the doctor, whilst patients can use the system to set up meetings
with healthcare professionals from their home, instead of having to travel to the
hospital.
2.1.1. Point-to-Point Ssystems
Similar to video conferencing, point-to-point systems offer commutations between
stakeholders, however, they only provide communications between two points. At
present, several such systems are freely available. While these systems are often
sufficient in providing simple communication between a patient and the doctor, they do
not offer the support group functionality that video conferencing systems do.
2.1.2. Remote Surgery
At the other end of the spectrum from telephony systems exists the concept of remote
surgery. This technology provides the means to support the delivery of remote care
through robotics [8]. An obvious advantage of this approach is that it permits world
renowned surgeons from across the globe to provide patient care. The first remote
surgery was conducted in 2001 and was referred to as "Operation Lindbergh". Through
the use of three robotic arms setup in an operating theatre in France, surgeons located
in New York where able to remotely perform a gallbladder operation over a fibre optic
connection. Since this date, several other remote surgical procedures have taken place.
Store and Forward Services: In addition to real-time telemedicine systems such as
those already described, offline telemedicine is also possible. Such care delivery does
not require both parties to be present at the same time of communication. For example,
technology within one consultant’s office can be used to record some patient
information which can be subsequently stored and forwarded at a set time to be
reviewed offline by another consultant.
3. Telecare
This Section aims to introduce the notion of Telecare and how technology can be
deployed to link remote patient information with healthcare providers in an effort to
improve the quality and the healthcare delivery process.
178 C.D. Nugent et al. / Remote Healthcare Monitoring and Assessment
Telecare combines the usage of sensing devices and telecommunications to support the
remote monitoring of a patient within their home environment with the goal of
providing a means of care support. Therefore, via remote patient monitoring it is
possible for a Telecare service to react to abnormal situations which are a cause for
concern. In response, a Telecare service can issue an alert to a caregiver or a family
member. It is also possible for the person themselves to raise an alarm in instances
when they may require support, for example, following a fall.
3.1.1. Technology Platforms Available for Home Based Monitoring
To provide a home based monitoring system to patients who require additional health
care services requires exploiting current technology platforms that are already in place.
These should be widely available and practical in terms of cost, performance and
speed.
The Public Switched Telephone Network (PSTN) is a commonly used means of
supporting the transmission of information from the patient’s home to a healthcare
organisation. The PSTN is available throughout the world and although originally
based on analogue lines, is now almost entirely digital at its core. The primary use of
the PSTN is the ability to connect people together to provide a phone based voice
service. However, in more recent times, the PSTN has been used as a digital
communication channel allowing information to be transferred to and from patients’
homes. The PSTN itself has a number of drawbacks in comparison with more recent
technologies based on the same infrastructure. It can be more difficult to set up and is
expensive to run if the application is resource intensive and operates at relatively slow
speeds making it suitable for only a small percentage of applications.
Broadband technology is another possibility. As its name suggests, broadband
technology offers a wider band of information. One example of broadband is a Digital
Subscriber Line (DSL) which encodes digital information at a higher frequency on one
channel and sends voice on a lower frequency forming another voice based channel.
DSL technology offers high availability, as it is based upon the already existing PSTN
system which has worldwide coverage. There are a number of advantages with
broadband technology. Firstly, it is easy to set up and easy to use and can even support
wireless connectivity. Although dependent on line conditions such as the distance from
the exchange, the speed and reliability are greatly improved. Broadband technology,
and DSL in particular, are essential to providing the underlying communication
infrastructure to allow home based services to become successful.
3.1.2. Telecare Services
One of the most common forms of Telecare services exist in the form of alarm based
pendants. Alarm based pendants are essentially a device worn by a patient which can
be used in either a passive or active manner. Their primary use is within the home
environment and can be used to trigger various alarms. A passive pendant is one which
involves no interaction between the patient and the device; this could take the form of a
fall detector which passively monitors the orientation of the patient during home based
activities. Falls are common place among elderly patients living alone and are the
leading cause for such people having to enter into permanent institutional care. Such a
device could alert a relative or health care personnel to a serious situation. A more
obvious and active approach of an alarm pendant is one that is activated by the patient
C.D. Nugent et al. / Remote Healthcare Monitoring and Assessment 179
whenever a security situation arises. Once the pendant has been activated, an
emergency message is relayed to a healthcare professional or family member so that
the necessary intervention can take place.
In addition to pendant alarms, Telecare services can be based upon a number of
devices which may exist within the home environment, for example medication
management devices, cognitive prosthetics, fire/smoke/water alarms and various door
and device interaction sensors. The exchange of physiological data recorded from the
patient in their home and transferred to a remote site can be referred to as Telehealth.
Although the patient and the communication infrastructure for both Telecare and
Telehealth are the same, it has been usual in the past to keep these two terms separate.
Typical parameters which can be monitored within the auspices of Telehealth include
blood pressure, electrocardiogram, weight and blood glucose levels.
The way in which healthcare can be delivered is changing rapidly. The convergence of
mobile communications, decision support systems and Internet computing are offering
a wealth of new healthcare paradigms. In this Section, we focus on a number of new
and emerging technologies and demonstrate how they can be deployed within the
realms of home based healthcare.
Perhaps the most notable advance in care delivery has been through the introduction of
the Intelligent Home environment [9]. This environment aims to promote the use of
embedded technology to support a person within their own living environment. The
desired goal is to offer a means of independent living and extend the period of time a
person can remain in their own home prior to institutionalisation {2}. Within such an
environment, it is common to find two types of devices; sensors and actuators. Sensors
are the devices which can record information about the person in the environment.
These may be motion sensors detecting which room the person is in, or they may be
pressure sensors detecting whether the person is in, for example, a chair or their bed.
Other types of sensors include temperature sensors, door sensors, water sensors,
appliance sensors etc. Such sensors can provide sufficient information so that the
current status of the person can be inferred and subsequently be reused to recommend
how the environment should be modified or controlled. The environment itself is
managed through the use of actuators. These can, for example, control the ambient
temperature if the environment becomes too warm, or raise an alarm if a hazardous
situation arises. For example, a person turns on the cooker and then turns on the taps in
the bath. The challenge at present is to find the correct balance between the information
gathered by the sensors and the ensuing processing to support the dynamic change of
the environment through the use of the actuators to support the changing needs of the
person.
180 C.D. Nugent et al. / Remote Healthcare Monitoring and Assessment
The ability to monitor a number of vital signs from a patient in a pervasive manner has
been realised through the introduction of Wearable Computing or Smart Clothes [10].
Smart clothing is a result of the consolidation of sensor technologies, advanced textiles
and information and communication technologies [11]. Sensing devices can be
embedded into clothing to offer a means of recording information such as heart rate,
perspiration levels, body motion / movement and breathing rates. Smart clothing offers
the ability to record from a larger surface area on the body, for example the torso,
hence the limitations of taking recordings from the finger or the wrist can be avoided
[12]. It is possible, during the manufacture of the garment, to fabricate the sensors by
either coating a textile with a form of sensing material or to form a sensing material
which can then be knitted or woven into the garment. Although many challenges now
exist regarding which are the optimal places to locate the recording electrodes [13] and
what information should be recorded, smart clothing provides a realistic solution to
continuously monitor the person in an unconstrained manner and either provide a
means of local processing or relay the information to a central location for assessment
by healthcare staff. Smart clothing has huge potential in conjunction with the
aforementioned intelligent environments as an augmented means, not only to assess the
activities of the person, but also, their health status.
We will now show how some of the aforementioned topics have been addressed in a
real project scenario – Home Based Medication Management. Within this context we
present an exemplar of how the problem domain has been identified, how a technical
C.D. Nugent et al. / Remote Healthcare Monitoring and Assessment 181
solution has been deployed and how emerging technologies have been integrated into
the second generation of the system.
It is a well recognised problem that patients do not fully adhere to their medication
regimen having received a prescribed course of medication. This results in a
significant burden being placed on the healthcare service due to extended patient
welfare and healthcare costs. According to the World Health Organisation (WHO)
adherence is defined as “the extent to which a person’s behaviour – taking medication,
following a diet, and/or executing lifestyle changes, corresponds with agreed
recommendations from a healthcare provider” [15]. There have been numerous
attempts to quantify the impact of non-adherence to medication, including:
- one third of people take all their medication, one third take some and one third do
not take any at all.
- as many as 125,000 deaths may occur each year in the US as a direct result of
non-adherence to medication.
- up to 23% of admissions to nursing homes can be attributed to non-adherence.
Adherence to medication is a complex issue and stems from problems such as the
patient not understanding their medication regimen, perceived side effects and financial
issues [16]. These can also be further complicated by any confusion which may arise
through the context in which the medication is prescribed, delivered and taken by the
patient.
Given the huge impact associated with non-adherence, many efforts have been made
to deploy technology as a potential means to alleviate the effects of the problem [17].
Technology has been deployed in three different manners. In the first instance, pill
holders can be offered to the patient which store the medication in labeled
compartments for various times throughout the day. The second type of device is an
extension of the first, however, electronic modules support the inclusion of pre-
programmable alarms. The third is a monitoring device which provides a remote means
to assess if the patient has taken their medication. This last type of device will be the
focus of this Case Study.
The aim of the MEDICATE system was to develop an Internet based care model and
associated peripherals to support the needs of all stakeholders within the prescribe to
intake chain of medication. The anticipated benefits of the system were the ability to
improve and support the patient with the management of their medication in addition to
providing a means to support communications between all stakeholders. To achieve
this vision it was necessary to develop a suite of interfaces which could be used by
each of the stakeholders. The interfaces took the form of both custom electronic
devices and software interfaces all connected via an Internet based care model as
shown in Figure 4.
182 C.D. Nugent et al. / Remote Healthcare Monitoring and Assessment
Figure 4 Overview of MEDICATE care model to support home based medication management.
This care model supported communication between the patient within their home
and their remote formal or informal carer, in addition to a communication channel
between healthcare professionals. Figure 5 below shows the various stakeholder
interfaces. Briefly these include:
Patient: a mobile medication management system (Figure 5 (a)) and a base station
acting as a reservoir of medication and a means to connect the patient to the Internet
portal. These devices had the ability to store the medication and would remind the
patient at the appropriate time to take their medication. Any instances of non-adherence
would be recorded by these devices
Doctor: a web based interface (Figure 5 (b)) to support the prescribing of
medication and assessment of patient adherence. This interface provides a means for
the medication regime of the patient to be entered onto the system.
Pharmacist: a software interface (Figure 5 (c)) to support the filling of medication
containers to be used by the patient’s medication device according to their prescribed
medication regimen. This system has the ability to retrieve the information entered onto
the system by the Doctor.
Caregiver: this software interface (Figure 5 (d)) provides a means to allow the
patient’s adherence to their medication regimen to be monitored in real time and, in
instances of non-adherence, will raise the attention of the care giving staff.
Within this scenario we have demonstrated how the basic concepts of utilising
modern day communications have supported the integration of a number of
stakeholders in the supply-to-intake chain of medication. Evaluation of this system has
demonstrated its usefulness in terms of supporting a patient’s adherence to medication.
C.D. Nugent et al. / Remote Healthcare Monitoring and Assessment 183
Nevertheless, it has also highlighted the importance which should be given to the
integration of such systems into existing practice.
The natural evolution of this system has involved the deployment of the concepts
of medication management onto mobile phone based platforms [17]. These facilitate a
similar means of providing management and of monitoring of a patient’s intake of
medication, in addition to maintaining the link with healthcare professionals. An
additional extension to the system would involve integration with context aware
services, where the means by which the reminder was offered to the patient would be
dependent upon their current activity.
5.4. Conclusions
Overall, within this review we have shown how technology can be used to establish a
number of care paradigms, all of which aim to improve the delivery of care provided to
the patient and to overcome geographical boundaries. In some cases existing
technological infrastructures may be used and in others new forms of technology are
required. Prior to widespread deployment, a number of challenges still remain. These
challenges are only partly related to the technology. Other challenges are more related
to organizational issues and issues relating to who is actually going to pay for the
184 C.D. Nugent et al. / Remote Healthcare Monitoring and Assessment
technology and its associated services. Nevertheless, it is widely accepted that there are
benefits to be accrued, hence it is anticipated that further effort will be directed towards
these areas in future years.
Glossary
References
[1] Ageing and life course, World Health Organisation, http://www.who.int/ageing/en/ (accessed May
2008).
[2] M.E. Pollack, Intelligent technology for an aging population: The use of ai to assist elders with
cognitive impairment, AI Magazine, (2005), 9-24.
[3] D.J. Cook and S.K. Das, How smart are our environments? An updated look at the state of the art,
Journal of Pervasive and Mobile Computing, 3 (2007), 53-73.
[4] L. Belochi, Telemedicine Glossary: Glossary of Concepts, Technologies, Standards and Users, 5th
ed., European Commission, Brussels, 2003.
[5] H.R. Fischer, S. Reichlin, J.P. Gutzwiller, A. Dyson, C. Beglinger, Telemedicine as a new
possibility to improve health care delivery, in M-Health: Emerging Mobile Health Systems, eds.
RSH Istepanian, S Laxminarayan, CS Pattichis, (Springer, 2006), 203-218.
[6] L.A. Black, C. McMeel, M. McTear, N. Black, R. Harper, M. Lemon, Journal of Telemedicine and
Telecare, 11 (2005), 6-8.
[7] M. Drugge, J. Hallberg, P. Parnes, K. Synnes, Wearable systems in nursing home care: prototyping
experience, IEEE Pervasive computing, (2006), 86-91.
[8] J. Rosen and B. Hannaford, Doc at a distance, IEEE Spectrum, 43 (2006), 34-39.
[9] J.C. Augusto and C.D. Nugent, Designing Smart Homes: The Role of Artificial Intelligence, LNAI
4008, Springer, 2006.
[10] F. Axisa, P.M. Schmitt, C. Gehin, G. Delhomme, E. McAdams and A. Dittmar, Flexible
technologies and smart clothing for citizen medicine, home healthcare, and disease prevention, IEEE
Transactions on Information Technology in Biomedicine, 9 (2005), 325-336.
[11] C.D. Nugent, P.J. McCullagh, E.T. McAdams and A. Lymberis, Personalised Health Management
Systems: The Integration of Innovative Sesning, Textile, Information and Communication
Technologies, (Amsterdam: IOS Press, 2005).
[12] S. Brady, L. Dunne, A. Lynch, B. Smyth and D. Diamond, Wearable Sensors? What is there to
sense?, in Personalised Health Management Systems: The Integration of Innovative Sensing, Textile,
Information and Communication Technologies, (Amsterdam: IOS Press, 2005), 80-88.
[13] M. Donnelly, C. Nugent, D. Finlay, P. McCullagh and N. Black, Making smart shirts smarter:
Optimal electrode placement for cardiac assessment, International Journal of Assistive Robotics and
Mechatronics, 8 (2007), 53-60.
[14] S.S. Intille, Designing a home of the future, IEEE Pervasive Computing, 1 (2002), 76-82.
[15] World Health Organisation (WHO), Adherence to Long-Term Therapies, Evidence for Action, 2003.
[16] L. Blonde, Removing polytherapy as a barrier to adherence, Managed Care, Compliance and
Persistence with Medication Therapy, 9 (2000), 1.
[17] C.D. Nugent, D. Finlay, R. Davies, M. Mulvenna, J. Wallace, C. Paggetti, E. Tamburini, N. Black,
The next generation of mobile medication management solutions, International Journal of Electronic
Healthcare, 3 (2007), 7-31.
Basic Engineering for Medics and Biologists 185
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-185
Introduction to
Chapter IV: Biomaterials and
Tissue Engineering
Fergal J. O’BRIEN and Brian O’CONNELL (eds.)
The goal of tissue engineering is to develop cell, construct, and living system
technologies to restore the structure and functional mechanical properties of damaged
or degenerated tissue. The term “tissue engineering”, which is interchangeably used
with the more recent term of regenerative medicine, was officially coined at a National
Science Foundation workshop in 1988 to mean “the application of principles and
methods of engineering and life sciences toward fundamental understanding of
structure-function relationships in normal and pathological mammalian tissues and the
development of biological substitutes to restore, maintain or improve tissue function.”
While the field of tissue engineering may be relatively new, the idea of replacing tissue
with another goes as far back as the 16th century when an Italian, Gasparo Tagliacozzi
(1546-99), Professor of Surgery and Anatomy at the Bologna University described a
nose replacement that he had constructed from a forearm flap in his work ‘De
Custorum Chirurigia per Insitionem” (The Surgery of Defects by Implantation) which
was published in 1597. In modern times, the techniques of transplanting tissue from
one site to another in the same patient (an autograft) or from one individual to another
(transplant or allograft) have been revolutionary and lifesaving. However major
problems exist with both techniques. Harvesting autografts is expensive, painful,
constrained by anatomical limitations and associated with donor-site morbidity due to
infection and haemorrhage. Transplants have serious constraints. The major problem is
accessing enough tissue and organs for all of the patients who require them.
Transplants are strongly associated with rejection by the patient’s immune system and
they are also limited by the potential risks of introducing infection or disease.
The field of tissue engineering is highly multidisciplinary and draws on experts
from mechanical engineering, materials science, surgery, genetics, and related
disciplines from engineering and the life sciences. Tissue engineering technologies are
based on a biological triad and involve the successful interaction between three
components: (1) the scaffold that holds the cells together to create the tissue’s physical
form i.e. acts as a template for new tissue formation, (2) the cells that synthesise the
tissue and (3) signalling mechanisms (i.e. mechanical and/or chemical signals) that
direct the cells to express the desired tissue phenotype. The specific properties of each
of these three components is critically important in determining the quality of the
186 F.J. O’Brien and B. O’Connell / Biomaterials and Tissue Engineering
engineered tissue and therefore the potential for clinical success when the engineered
tissue has been implanted into an injured site in vivo. For example, the choice of
biomaterial and the fabrication process used to produce the porous scaffold which will
act as the template for in vitro tissue formation should be selected based on the specific
tissue type and anatomical site in which it will be implanted. Similarly, consideration
must be given to the choice of cell type used, for example, whether the cells should be
autologous or allogeneic and whether stem cells or primary mature cells should be used.
Finally, the process of cellular differentiation into a specific phenotype followed by the
production of extracellular matrix is stimulated by either the provision of specific
growth factors and cytokines and/or by subjecting a cell-seeded construct to
biophysical stimulation in the form of a bioreactor. The types of growth factors,
choice of bioreactor and duration of exposure all play a key role in determining the
success of the engineered tissue. The mechanical properties of the scaffold and
engineered tissue will govern the ability of the construct to function in vivo and to
allow infiltration of host cells and in most tissues (cartilage being an exception),
vascularisation. It is therefore important to develop techniques to test biomaterial
scaffold and tissue properties prior to implantation and to be able to predict how the
constructs will behave when implanted in vivo.
This chapter focuses on all three areas of the tissue engineering triad: Scaffolds &
Surfaces, Cellular & Molecular Biomechanics and Bioreactors in Tissue Engineering.
In addition, the fourth part of the chapter is devoted to Characterisation and Testing of
Biomaterials. It is pitched at a level that should allow either an engineer or a clinician
to understand the basic principles of tissue engineering and the fundamental interaction
between cells & scaffolds and their response to biophysical stimuli and the different
types of bioreactor that exist. The chapter finishes with a section which discusses the
fundamentals of materials testing which will be of interest to clinicians and biologists
involved not only in tissue engineering, but in any area of biomechanical analysis. The
contributors are all experts in their fields and are involved in research in tissue
engineering and a number of cognate disciplines and come from all across Europe.
They include an anatomist/bioengineer (O’Brien), dentist (O’Connell), materials
scientist (Partap), orthopaedic surgeon (Lyons), physiologist (Campbell), engineers
(Plunkett, Dendorfer and Hammer) and trauma surgeon (Lenich).
Basic Engineering for Medics and Biologists 187
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-187
Introduction
Every day thousands of clinical procedures are performed to replace or repair tissues in
the human body that have been damaged through disease or trauma. Current therapies
are focused on the replacement of the damaged tissue by using donor graft tissues
(autografts, allografts or xenografts). Problems associated with this approach include
shortage of donors or donor sites, the volume of donor tissue that can be safely
harvested, donor site pain and morbidity, the possibility of harmful immune responses,
transmission of disease and rejection of grafts [1]. Alternatively, the field of tissue
engineering (a phrase that is interchangeably used with regenerative medicine) aims to
regenerate damaged tissues instead of replacing them (with grafts) by developing
biological substitutes that restore, maintain or improve tissue function [2, 3].
In native tissues, cells are held within an extracellular matrix (ECM) which guides
development and directs regeneration of the tissue, serves to organise cells in space and
provides them with environmental signals to direct cellular behaviour. The goal of
tissue engineering is to synthesise substitutes that mimic the natural ECM to help guide
the growth of new functional tissue in vitro or in vivo. At a simplistic level, biological
tissues consist of cells, signalling mechanisms and extracellular matrix. Tissue
engineering technologies are based on this biological triad and involve the successful
interaction between three components: (1) the scaffold that holds the cells together to
create the tissues physical form, (2) the cells that create the tissue and, (3) the
biological signalling molecules (such as growth factors) that direct the cells to express
the desired tissue phenotype (Figure 1). Tissue engineering is a multidisciplinary field
that harnesses expertise and knowledge from a variety of fields, including those of the
medical profession, materials scientists, engineers, chemists and biologists.
There are differences in cell behavior in three dimensional (3-D) vs. two dimensional
(2-D) environments. In vitro 3-D cell culture conditions more accurately model in vivo
biological responses, as the conditions more closely resemble the natural structure and
function of tissues in vivo [4]. These conditions can be created by using a 3-D scaffold
that acts as a template, allowing cells to produce and deposit extracellular matrix
(ECM) that would not be possible in 2-D environments. 2-D cell culture does not allow
cells to move or assemble with the freedom they have in vivo, and thus cannot replicate
188 S. Partap et al. / Scaffolds & Surfaces
Figure 1. The tissue engineering triad; factors that need to be considered when designing a suitable structure
for tissue engineering applications.
All scaffolds for tissue engineering applications are designed to perform the following
functions: (1) to encourage cell-material interactions i.e. cell attachment, differentiation
and proliferation, eventually leading to the deposition of extracellular matrix, (2) to
permit the transport of nutrients, wastes and biological signalling factors to allow for
cell survival, (3) to biodegrade at a controllable rate which approximates the rate of
natural tissue regeneration, and (4) to provoke a minimal immune and/or inflammatory
response in vivo. The following parameters must be considered when designing a
scaffold for tissue engineering.
S. Partap et al. / Scaffolds & Surfaces 189
1.1. Biocompatibility
The implantation of a scaffold may elicit different tissue responses depending on the
composition of the scaffold. If the scaffold is non-toxic and degradable, new tissue will
eventually replace it; if it is non-toxic and biologically active then the scaffold will
integrate with the surrounding tissue. However, if the scaffold is biologically inactive,
it may be encapsulated by a fibrous capsule, and in the worst case scenario if the
scaffold is toxic, rejection of the scaffold and localised death of the surrounding tissue
can occur [6]. Biocompatibility is the ability of the scaffold to perform in a specific
application without eliciting a harmful immune or inflammatory reaction. For a
scaffold to positively interact with cells and with minimal disruption to the surrounding
tissue, it should have an appropriate surface chemistry to allow for cellular attachment,
differentiation and proliferation. Cells primarily interact with scaffolds via chemical
groups on the material surface or topographical features. Topographical features
include surface roughness and pores where cell attachment is favoured. Alternatively,
cells may recognise and subsequently bind to the arginine-glycine-aspartic acid (RGD)
cell adhesion ligand. Scaffolds synthesised from natural extracellular materials (e.g.
collagen) already possess this specific ligands, whereas scaffolds made from synthetic
materials may be designed to deliberately incorporate them.
1.2. Biodegradability
The scaffold provides structural integrity to the engineered tissue in the short term.
Furthermore, it provides a framework for the three dimensional (3-D) organisation of
the developing tissue as well as providing mechanical stability to support the growing
tissue during in vitro and/or in vivo growth phases [9]. The mechanical properties of the
scaffold should be designed to meet the specific requirements of the tissue to be
regenerated at the defect site. Furthermore, at the time of implantation, the scaffold
should have sufficient mechanical integrity to allow for handling by the clinician, be
able to withstand the mechanical forces imposed on it during the implantation
procedure and survive under physiological conditions. Immediately after implantation,
the scaffold should provide a minimal level of biomechanical function that should
progressively improve until normal tissue function has been restored, at which point the
construct should have fully integrated with the surrounding host tissue.
190 S. Partap et al. / Scaffolds & Surfaces
Porous structures allow for optimal interaction of the scaffold with cells. The pore
architecture is characterised by pore size and shape, pore interconnectivity/tortuosity,
degree of porosity and surface area. The microstructure determines cell interactions
with the scaffold, as well as molecular transport (movement of nutrients, wastes and
biological chemicals e.g. growth factors) within the scaffold. Specifically, pore size
determines the cell seeding efficiency into the scaffold [10]; very small pores prevent
the cells from penetrating the scaffold, whilst very large pores prevent cell attachment
due to a reduced area and therefore, available ligand density. Subsequently, cell
migration within a scaffold is determined by degree of porosity and pore
interconnectivity/tortuosity. A scaffold with an open and interconnected pore network,
and a high degree of porosity (>90 %) is ideal for the scaffold to interact and integrate
with the host tissue [11].
phase solid, and the solvent is then removed yielding a porous polymer, this is known
as thermally induced phase separation (TIPS) [14]. In the second, a polymer is
dispersed in a solvent which is then frozen to induce crystallisation of the solvent to
form solvent crystals that act as templates for the pores. These crystals are then
removed by freeze drying to yield a porous foam. Manipulation of the processing
conditions enables the creation of different pore sizes and distributions [15].
1.5.3. Foaming
Foaming techniques use gaseous porogens that are produced by chemical reactions
during polymerisation, or are generated by the escape of gases during a temperature
increase or drop in pressure. Nam et al. 2000 [16] synthesised poly (lactic acid) [PLA]
scaffolds using ammonium bicarbonate which acted as both a gas foaming agent and as
a solid salt porogen, an increase in temperature caused the formation of carbon dioxide
and ammonia to create a highly porous foam. Also, high pressure carbon dioxide can
be used to foam polymers by saturating a prefabricated polymer monolith. A
subsequent reduction in pressure causes a decrease in solubility of the carbon dioxide
within the polymer, and as the carbon dioxide gas tries to escape it causes the
nucleation and growth of bubbles resulting in a porous microstructure [17].
1.5.4. Emulsion Templating
Porous structures can also be obtained by using emulsion templating techniques. The
internal phase of the emulsion acts as a template for the pores whilst polymerisation
occurs in the continuous phase (in which the monomer is dissolved). After
polymerisation, the internal phase is removed to give a templated porous material. The
resulting porous microstructures are replicas of the internal phase droplets around
which polymerisations were performed. The size of the emulsion droplets is preserved,
producing polymer foams with approximately the same size and size distributions as
that of the emulsion at the point of polymerisation [18].
1.5.5. Solid Free Form (SFF) Fabrication
Solid free form (SFF) fabrication or rapid prototyping (RP) technologies uses layer
manufacturing techniques to create three dimensional scaffolds directly from computer
generated files. There are a few techniques that come under this group including
stereolithography, selective laser sintering, fused depositional modeling and three
dimensional printing. However, all the techniques share the same principle where
powders or liquids are solidified one layer at a time to gradually build a three-
dimensional scaffold. The layering is controlled by computer assisted design (CAD)
programs where the scaffold architecture is designed and modelled. Data collected
from computed tomography (CT) or magnetic resonance imaging (MRI) scans may
also be used to create CAD models that are specific to the tissue to be regenerated [19].
1.5.6. Combination of Techniques
The techniques discussed above can also be combined with each other depending on
the exact requirements of the scaffold, e.g. phase separation (freeze drying) techniques
can be combined with emulsion templating processes. Whang et al. 1995 created an
emulsion that was quenched using liquid nitrogen, which was then freeze dried to
produce porous PLGA polymeric monoliths [20].
192 S. Partap et al. / Scaffolds & Surfaces
Figure 2 Scanning electron microscopy images of porous (a) collagen-GAG scaffolds made by freeze
drying15, (b) poly-L-lactide (PLLA) foams made by solvent casting-particulate leaching12, (c) alginate
scaffolds made by emulsion templating18 and (d) polycaprolactone–calcium phosphate composites made by
solid free form fabrication methods51.
2.1. Ceramics
Ceramics (inorganic, non metallic materials) used within the biomedical field are
classified as being either bioinert or bioactive. The bioinert ceramics include materials
such as alumina and zirconia that are typically used as implants for musculoskeletal,
oral and maxillofacial applications whilst the bioactive group include the calcium
phosphates, the bioglasses and glass-ceramics [6]. All bioceramics are also further
defined as being osteoconductive (supporting bone growth) or osteoinductive
(stimulating bone growth); all types of bioceramics are osteoconductive as all support
the formation of bone, but not all are osteoinductive. The calcium phosphate based
bioceramics, bioglasses and glass-ceramics are commonly used as scaffolds for bone
tissue engineering as they have a compositional similarity to the mineral phase of bone
S. Partap et al. / Scaffolds & Surfaces 193
[21]. Hydroxyapatite (HA) and tri-calcium phosphate (TCP) are two of the most
commonly used calcium phosphate bioceramics in tissue engineering applications. TCP
is used as a degradable scaffold, whilst HA, which is non-resorbable and has the added
advantage of being osteoinductive, is typically used for coating biomedical implants to
induce bone regeneration, allowing the implant to integrate with the surrounding tissue.
For this reason, HA has shown much popularity for use as a scaffold for tissue
engineering.
Figure 3 Chemical structures of some biodegradable synthetic polymers used as scaffolds in tissue
engineering applications
Structural proteins such as fibrin are also utilised in tissue engineering applications.
Fibrin can be used as a natural wound healing material, and has found applications as a
sealant and adhesive in surgery. It can be produced from the patient’s own blood, to be
used as an autologous scaffold. However, the stability of the material is limited as it
can be easily degraded unless apronitin, a protein inhibitor, is used to control the rate of
degradation. Fibrin hydrogels have been used to engineer tissues with smooth muscle
cells [30] and chondrocytes [31]. Alternatively, gelatin (a derivative of collagen) that is
produced by altering the helical structure of the collagen molecule by breaking it into
single strand molecules) has been investigated for cartilage tissue regeneration. [32].
However, as one of the main disadvantages of gelatin is its poor mechanical strength, it
has also been cross-linked with hyaluronic acid for skin tissue engineering, and with
alginate for wound healing applications [33]. Instead, collagen, the main component
found in the extracellular matrix of mammalian connective tissues has found use in
tissue engineering applications including skin substitutes [34], scaffolds for bone and
cartilage, vascular applications and as drug delivery systems. As is typical of all natural
polymers, collagen gels also display poor mechanical properties. However, these can be
improved by employing both chemical and physical crosslinking methods. Physical
crosslinking methods include UV radiation and dehydrothermal treatments, whilst
cross-linking agents such as glutaraldehyde and carbodiimides (EDAC) can be used to
produce chemically cross-linked collagen hydrogels with improved physical properties
(Figure 4).
Figure 4 Chemical structures of some natural polymers used as scaffolds in tissue engineering applications
196 S. Partap et al. / Scaffolds & Surfaces
2.4. Composites
Due to some of the problems associated with using scaffolds synthesised from a single
phase biomaterial (eg. poor mechanical properties and biocompatibility of natural and
synthetic polymers respectively, and poor degradability of bioceramics), a number of
researchers have developed composite scaffolds comprising of two or more phases to
combine the advantageous properties of each phase. For example, polymer/ceramic
composites of poly(lactic-co-glycolic acid (PLGA) and hydroxyapatite have been
investigated for tissue engineering applications [35], whilst Cui et al. [36] have
produced tri-phasic scaffolds by depositing nano-hydroxyapatite particles onto cross-
linked collagen-chitosan matrices. However, even though composite scaffolds such as
these have shown some promise as grafts for bone and cartilage, each one consists of at
least one phase which is not found naturally in the body and therefore has problems
with either biocompatibility or biodegradability or both. Table 1 summarises the
different types of biomaterials described above and lists the advantages and
disadvantages of each type for use as scaffolds in tissue engineering applications.
Table 1 Properties, advantages and disadvantages of biomaterials used as scaffolds in tissue engineering
applications
198 S. Partap et al. / Scaffolds & Surfaces
There are also reports of triphasic scaffolds made from collagen, a bioceramic and
a synthetic polymer. Scaffolds made from nano-HA, collagen and PLA were placed in
defects of rabbit radius and they integrated with the defect site within 12 weeks [42].
These studies indicate that by finding an adequate balance between pore structure,
mechanical properties and biocompatibility, a collagen-based construct can potentially
support bone growth and may have real potential for bone tissue engineering.
Economic activity within the tissue engineering sector has grown five-fold in the past 5
years. In 2007, approximately 50 companies offered commercially available tissue-
regenerative products or services, with annual sales recorded in excess of $1.3 billion,
whilst 110 development-stage companies with over 55 products in FDA-level clinical
trials and other preclinical stages spent $850 million on development [43]. The tissue
engineering approach was originally conceived to address the gap between patients
waiting for donors and the amount of donors actually available. To date the highest
rates of success have been achieved in the areas of skin regeneration where tissue-
engineered substitutes have been successfully used in patients [44].
However, much research still remains to be performed in all aspects of tissue
engineering [45]. Cellular behaviour is strongly influenced by signals (biochemical and
biomechanical) from the extracellular matrix, the cells are constantly receiving cues
from the extracellular matrix about their environment and are constantly remodelling it
accordingly. Therefore, an appropriate three dimensional structure that is
predominantly thought of as playing a mechanical role is not enough to promote the
growth of new tissue. It is important that the scaffold provides adequate signals (e.g.
through the use of adhesion peptides and growth factors) to the cells, to induce and
maintain them in their desired differentiation stage, for their survival and growth [46].
Thus, equal effort should be made in developing strategies on how to incorporate the
adhesion peptides and growth factors into the scaffolds, as well as in identifying the
chemical identity of adhesion peptides and growth factors that influence cell behaviour,
along with the distributions and concentrations required for successful outcomes. An
example would be to incorporate angiogenic growth factors in scaffolds for different
types of tissue in an attempt to generate vascularised tissues. Tissue vascularisation can
be used to establish blood flow through the engineered tissues and strategies involving
the incorporation of vasculature, as well as innervation will be of great importance [47].
Additionally, the incorporation of drugs (i.e. inflammatory inhibitors and/or antibiotics)
into scaffolds may be used to prevent any possibility of an infection after surgery [48].
The field of biomaterials has played a crucial role in the development of tissue
engineered products. An alternative to using prefabricated scaffolds is to use a polymer
system that is injected directly into the defect site which is polymerised in situ using
either heat [49] (thermoresponsive polymers) or light [50] (photoresponsive polymers).
The advantages for the patient with this approach over current therapies are that
injectable delivery systems fill both regularly and irregularly shaped defects (“get a
custom fit”), they represent a minimally invasive procedure therefore avoiding surgery
and the potential risks associated with it, eliminate the need for donor tissue or a donor
site, and waiting time for treatment is reduced, as it can be used whenever treatment is
required.
S. Partap et al. / Scaffolds & Surfaces 199
References
[1] R. Langer, Biomaterials in drug delivery and tissue engineering: One laboratory’s experience, Acc
Chem Res 33 (2000), 94-101.
[2] A. Atala, Tissue engineering and regenerative medicine: Concepts for clinical application,
Rejuvenation Res 7 (2004), 15-31.
[3] L. J. Bonassar and C. A. Vacanti, Tissue engineering: The first decade and beyond, J Cell Biochem
Suppl 30-31 (1998), 297-303.
[4] M. P. Lutolf and J. A. Hubbell, Synthetic biomaterials as instructive extracellular microenvironments
for morphogenesis in tissue engineering, Nat Biotechnol 23 (2005), 47-55.
[5] L. G. Griffith and M. A. Swartz, Capturing complex 3d tissue physiology in vitro, Nat Rev Mol Cell
Biol 7 (2006), 211-24.
[6] L. L. Hench, Bioceramics, J. Am. Ceram. Soc 81 (1998), 1705-28.
[7] J. E. Babensee, A. G. Mikos, J. M. Anderson and L. V. McIntire, Host response to tissue engineered
devices, Adv. Drug Del. Rev. 33 (1998), 111-139.
[8] W. E. Hennink and C. F. van Nostrum, Novel crosslinking methods to design hydrogels, Adv Drug
Deliv Rev 54 (2002), 13-36.
[9] D. W. Hutmacher, Scaffolds in tissue engineering bone and cartilage, Biomaterials 21 (2000), 2529-
2543.
[10] F. J. O’Brien, B. A. Harley, I. V. Yannas and L. J. Gibson, The effect of pore size on cell adhesion in
collagen-gag scaffolds, Biomaterials 26 (2005), 433-41.
[11] T. M. Freyman, I. V. Yannas and L. J. Gibson, Cellular materials as porous scaffolds for tissue
engineering, Prog. Mater Sci. 46 (2001), 273-282.
[12] L. Lu, S. J. Peter, M. D. Lyman, H. L. Lai, S. M. Leite, J. A. Tamada, S. Uyama, J. P. Vacanti, R.
Langer and A. G. Mikos, in vitro and in vivo degradation of porous poly(dl-lactic-co-glycolic acid)
foams, Biomaterials 21 (2000), 1837-45.
[13] S. H. Oh, S. G. Kang, E. S. Kim, S. H. Cho and J. H. Lee, Fabrication and characterization of
hydrophilic poly(lactic-co-glycolic acid)/poly(vinyl alcohol) blend cell scaffolds by melt-molding
particulate-leaching method, Biomaterials 24 (2003), 4011-21.
[14] A. S. Rowlands, S. A. Lim, D. Martin and J. J. Cooper-White, Polyurethane/poly(lactic-co-glycolic)
acid composite scaffolds fabricated by thermally induced phase separation, Biomaterials 28 (2007),
2109-21.
[15] F. J. O’Brien, B. A. Harley, I. V. Yannas and L. Gibson, Influence of freezing rate on pore structure in
freeze-dried collagen-gag scaffolds, Biomaterials 25 (2004), 1077-86.
[16] Y. S. Nam, J. J. Yoon and T. G. Park, A novel fabrication method of macroporous biodegradable
polymer scaffolds using gas foaming salt as a porogen additive, J Biomed Mater Res 53 (2000), 1-7.
[17] D. J. Mooney, D. F. Baldwin, N. P. Suh, J. P. Vacanti and R. Langer, Novel approach to fabricate
porous sponges of poly(d,l-lactic-co-glycolic acid) without the use of organic solvents, Biomaterials 17
(1996), 1417-22.
[18] S. Partap, J. A. Darr, I. U. Rehman and J. R. Jones, "Supercritical carbon dioxide in water" Emulsion
templated synthesis of porous calcium alginate hydrogels, Adv. Mater. 18 (2006), 501-504.
[19] E. Sachlos and J. T. Czernuszka, Making tissue engineering scaffolds work. Review: The application
of solid freeform fabrication technology to the production of tissue engineering scaffolds, Eur Cell
Mater 5 (2003), 39-40.
[20] K. Whang, C. H. Thomas, K. E. Healy and G. Nuber, A novel method to fabricate bioabsorbable
scaffolds, Polymer 36 (1995), 837-842.
[21] K. A. Hing, Bioceramic bone graft substitutes: Influence of porosity and chemistry, Int. J. Appl.
Ceram. Technol. 2 (2005), 184-199.
200 S. Partap et al. / Scaffolds & Surfaces
[46] C. A. Pangborn and K. A. Athanasiou, Growth factors and fibrochondrocytes in scaffolds, J Orthop
Res 23 (2005), 1184-90.
[47] R. Langer, Tissue engineering: Perspectives, challenges, and future directions, Tissue Eng 13 (2007),
1-2.
[48] M. V. Risbud and M. Sittinger, Tissue engineering: Advances in in vitro cartilage generation, Trends
Biotechnol 20 (2002), 351-6.
[49] L. Klouda and A. G. Mikos, Thermoresponsive hydrogels in biomedical applications, Eur J Pharm
Biopharm 68 (2008), 34-45.
[50] K. T. Nguyen and J. L.West, Photopolymerizable hydrogels for tissue engineering applications,
Biomaterials 23 (2002), 4307-14.
[51] M. J. Mondrinos, R. Dembzynski, L. Lu, V. K. Byrapogu, D. M. Wootton, P. I. Lelkes and J. Zhou,
Porogen-based solid freeform fabrication of polycaprolactone-calcium phosphate scaffolds for tissue
engineering, Biomaterials 27 (2006), 4399-408
202 Basic Engineering for Medics and Biologists
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-202
Introduction
Mechanosensitive (MS) ion channels form pores that span the plasma membrane and
are therefore ideally situated to respond to external forces applied to the cell. These
channels serve as mechano-electrical switches converting a mechanical force into a
change in the cellular electrochemical gradient (see Figure 1), and they play a role in
coordinating diverse physiological events such as touch and hearing in mammals, to
adaptation to turgor in bacteria [1, 2]. The patch clamp technique has facilitated the
electrophysiological analysis of single MS channel currents and has demonstrated that
these channels open in response to a hypo-osmotic environment which causes the
V.A. Campbell and B. O’Connell / Cellular & Molecular Biomechanics 203
membrane to swell, or by the direct application of suction [3]. It has been suggested
that MS channels may be gated directly by changes in the lipid bilayer tension, or in the
‘tethered model’ of gating, by displacement of the MS channel with respect to
components of the cytoskeleton or extracellular matrix [1]. Amongst the
mechanosensitive channels are the transient receptor potential channels (TRPs) [4] and
stretch-activated ion channels (SACs) [5]. However, mechanosensitive channels also
include other ion channels such as voltage-dependent Ca2+ channels [6], the NMDA
receptor [7] and Ca2+-dependent BK channels [8]. Many MS channels are permeable to
cations in general, with highest permeability to Ca2+, which results in an increase in
intracellular Ca2+ concentration following channel opening. It is postulated that the Ca2+
then acts as a second messenger resulting in the appropriate cell response to
mechanical stimulation [9]. Other classes of MS ion channels exhibit selective
permeability for Na2+, K+ or Cl- ions. It is therefore apparent that changes in the
mechanical environment initiate electrical and chemical changes in the cell which
define the cellular response.
MS channels are also modulated by cellular components, such as the contractile
actin cytoskeleton. The cytoskeleton provides support to the plasma membrane and
forms part of the coupling mechanism to the extracellular matrix. It reacts and
rearranges its configuration in response to alterations in the extracellular environment.
The tensegrity model [2] proposes that the cytokeleton is the architectural basis for
cellular mechanotransduction, and application of physical stress promotes changes in
actin polymerization to control cell shape and realignment [10]. A direct coupling may
exist between MS channels and the cytoskeleton since agents that disrupt the
organization of actin filaments, such as colchicine and cytochalasin-D, affect MS
channel activity. Studies suggest that the cytoskeleton exerts a tonic inhibition on MS
channel activity and support the contention that MS ion channels are functionally
linked with the dynamics of the intracellular cytoskeleton [11, 12].
that is generated in the plasma membrane [16]. In the nematode, C. elegans, the
mechanosensitive TRPN channel has recently been found to be involved in stretch
receptor-mediated proprioception to regulate sensory-motor integration during
locomotion [17]. Thus, the TRP family of receptors are mechanically regulated via
distinct mechanisms, and are involved in sensing changes in lipid tension precipitated
by stretch and a hypotonic environment.
Figure 1. Cell pathways associated with mechanotransduction. Cells respond to mechanical stimulation via
activation of membrane-associated mechanoreceptor proteins which undergo a conformation change resulting
in the activation and recruitment of a cascade of intacellular signaling molecules. The signaling molecules
may be released into the extracellular matrix (e.g nitric oxide) to influence neighbouring cells, or may target
organelles such as the nucleus and cytoskeleton to evoke changes in gene transcription or cell motility,
respectively.
Enzymes associated with the intracellular domain of the plasma membrane, such as
phospholipase A2, phospholipase C and tyrosine kinases have also been reported to
have a mechanosensitive role [18]. Membrane stretch increases the production of
prostaglandins via phospholipase A2-mediated release of arachidonic acid [19]. In a
number of cell types, flow and shear have been shown to induce activation of
phospholipase C with subsequent activation of the phosphatidylinositol pathway. In
glial cells exposured to pressure, cell proliferation is increased and this effect is
blocked by genistein, the tyrosine kinase inhibitor, but not by blockers of
mechanosensitive channel blockers, implicating tyrosine kinases as mechanosensitive
elements in some cell types [20].
V.A. Campbell and B. O’Connell / Cellular & Molecular Biomechanics 205
Mechanical force has been shown to activate the mitogen-activated protein kinase
(MAPK) superfamily [21]. The extracellular-regulated protein kinase (ERK) has been
definitively implicated in mechanotransduction, but a role for p38 and c-jun N-terminal
kinase (JNK) has also been reported [22]. ERK activation has been observed in
response to steady and pulsatile stretch in the vasculature [23], renal mesangial cells
[24], osteoblasts [25] and gingival fibroblasts [26] where downstream consequences of
ERK activation include proliferation, immediate early gene expression and
extracellular matrix accumulation. Furthermore, the strain-induced activation of ERK is
thought to play a role in ‘mechanical strain memory’ in airway smooth muscle,
whereby a mechanically-induced deformation of the smooth muscle cell produces a
structural change which modulates mechanical performance in subsequent contractions
[27].
Mechanical strain activates nitric oxide synthase in endothelial cells [28], stromal
cells [29], chondrocytes [30], and osteocytes [31] leading to the formation of nitric
oxide (NO), a gaseous diffusible second messenger that regulates vascular remodeling,
increases expression of bone-related genes, regulates protoeoglycan synthesis, and
alters cytoskeletal fibre alignment, respectively. The strain-induced changes in NO
formation are coordinated by ERK signaling [32], demonstrating substantial crosstalk
between these intracellular pathways.
A number of lines of evidence support the hypothesis that stem cells are responsive to
mechanical stimulation, and that stem cell differentiation may be regulated by strain.
Certainly during development the embryo is exposed to stresses imposed by the
external environment of the uterus, as well as to stresses imposed by neighbouring cells.
These growth-generated strains and pressures can regulate expression of a number of
developmental genes [39]. Administration of a neuromuscular blocker to a chick
embryo causes paralysis, and interferes with the development of the musculoskeletal
system [40], thereby reflecting the importance of the local mechanical environment to
tissue differentiation. At a cellular level, there is substantial evidence demonstrating
that mechanical stresses exert a direct effect on stem cells. For example, mouse
embryonic stem cells generate reactive oxygen species in response to mechanical strain
and this pathway initiates differentiation into cells of the cardiovascular system [41].
Adult mesenchymal stem cells (MSCs) isolated from the marrow are multipotent cells
that have the proclivity to differentiate along osteogenic, chondrogenic and adipogenic
lineages [42]. Human MSCs exposed to cyclic uniaxial strain in the presence of
osteoinductive factors increase the formation of bone-associated extracellular matrix
[43], whilst exposure to low-intensity ultrasound promotes chondrogenesis [44]. MSCs
isolated from adipose tissue respond to mechanical loading applied via pulsating fluid
flow by increasing nitric oxide production and undergoing osteogenesis [45]. As well
as controlling stem cell differentiation potential, mechanical signals are also important
for regulating MSC proliferation via Ca2+ -dependent mechanisms [46]. Thus
mechanical stimulation regimes may have potential to modulate stem cell proliferation
and differentiation profiles for tissue engineering applications (Figure 2).
The walls of blood vessels react to the mechanical stimuli of pressure and shear stress
evoked by the flowing blood. The responses to such mechanical events are necessary to
control blood flow, or to adapt the vessel structure to its requirements. Shear stress is a
critical biophysical phenomenon for determining vascular homeostasis, vascular
remodeling and development of the cardiac system, as well as for governing the
development of atherosclerotic lesions. The endothelial cells which line blood vessels
are responsible for vascular responses to the shear stress that is evoked by fluid flow.
Shear stress is transmitted from the apical surface of the endothelial cell through the
cytoskeleton to points of attachments at cell-cell and cell-matrix junctions. These
junctions, composed of proteins, such as platelet endothelial cell adhesion molecule
(PECAM), transmit mechanical force via interaction with the intracellular signaling
proteins, Src and P13kinase, leading to activation of the transcription factor, NFNB [47].
In laminar shear, this integrin pathway promotes endothelial cell alignment in the
direction of flow, and a transient activation of the NFNB pathway. In contrast, in the
face of disturbed shear, such as may occur at vessel branch points, bifurcations and
regions of high curvature, the PECAM pathway is activated in a sustained manner. The
sustained activation of NFNB leads to prolonged induction of NFNB-regulated genes,
(e.g adhesion molecules, ICAM, VCAM, E-selectin) which are molecular hallmarks of
V.A. Campbell and B. O’Connell / Cellular & Molecular Biomechanics 207
Figure 2. Bone-associated gene expression in stem cells is modulated by constraining the substrate or the
application of uniaxial strain.Effects of one-dimensional clamping (Clamped) and 5% uniaxial cyclic strain
(Clamped & Stretched) (1 Hz) on levels of bone-associated RNAs from mesenchymal stem cell-seeded
collagen glycosaminoglycans scaffold constructs incubated for 7 days in the presence of osteogenic factors,
when compared to unconstrained constructs (Unconstrained). BSP; bone sialoprotein, OCN; osteocalcin.
*p<0.05, **p<0.01. Data courtesy of Dr. Elaine Byrne, RCSI
The skeletal system can regulate bone density and architecture in order to meet the
body’s strucural and metabolic demands. While the metabolic demands of the skeleton
are controlled by the calcitropic hormones, the responsiveness of the skeleton to
mechanical load governs skeletal strucure. The cells of the skeleton – osteoblasts,
osteocytes, osteoclasts, chondrocytes, as well as their progenitor bone marrow-derived
cells – display mechnosensitivity which influences how these cells function, and thus
the skeletal characteristics. At a macro-level it is clear that application of load is
anabolic to bone [48], while un-weighting, such as occurs in the absence of gravity,
induces bone loss [49]. At the micro-level a number of cells involved in the
maintenance of bone may be regulated by mechnical force. The marrow stromal
progenitor cells display alterations in proliferation, gene expression, and nitric oxide
formation in response to mechnical load [32], and RANKL expression, a key factor in
osteoclast formation, is increased in response to fluid shear stress [50].
Osteocytes are the principal cell in cortical bone and by extending cellular
processes through a network of fluid-filled cannaliculi, they are ideally placed to
monitor mechanical strain as it is transmitted to the cell surface. Application of
mechanical load increases the production of bone-related proteins [39]. In relation to
208 V.A. Campbell and B. O’Connell / Cellular & Molecular Biomechanics
Figure 3. Comparison of osteoporotic and osteoarthritic cell responses to shear fluid flow Effect of pulsating
fluid flow (PFF) at low (0.4±0.1 Pa), medium (0.6± 0.3 Pa), or high shear stress (1.2±0.4 Pa) on nitric oxide
(NO) and prostaglandin E2 (PGE2) production by osteoporotic (OP, black squares) and osteoarthritic (OA,
white circles) bone cells. Data are expressed as PFF-treated over static culture ratios calculated per individual
donor. A; NO response, B; PGE2 response. PFF/Stat=1 means no effect. a implies a significant effect of PFF,
b implies significantly different from OA cells, P<0.05. Reproduced from [57].
V.A. Campbell and B. O’Connell / Cellular & Molecular Biomechanics 209
Figure 4. Generation of nitric oxide (NO) in a single MC-3T3 osteoblast cell in response to stimulation by
atomic force microscopy. Time-series images of NO fluorescence intensity from 0 s to 1600 s. Periodic
indentation with peak force of ~22nN was applied from 30 s until 60 s at the location indicated (white arrow).
Reproduced from [65].
(second messenger production, ion flux, protein phosphorylation) is often more readily
and reliably measured from a large number of cells. Similarly, it may be more feasible
to expose a population of cells to a particular stimulus (stretch, shear, pulsating flow)
than a single cell. However, there are drawbacks with this approach, including the need
to ensure a homogeneous cell population and the assumption that all cells in the
experiment will experience precisely the same strain. Nonetheless, the observation of
biomechanical responses in cells has been very useful in undertanding how cells are
likely to behave in a variety of conditions, for example in coronary heart disease, bone
fractures, during skeletal development or orthodontic tooth movement. An extension of
the purely cell-based experiment is to use an entire tissue or organ in culture to study
biomechanical responses. Isolated muscle, bone, cartilage, tendon or blood vessel may
be maintained long enough in vitro to study cellular responses to applied strain over a
period of hours to perhaps a few days. Improved systems for maintaining tissue
viability, ‘bioreactors’, are discussed elsewhere in this text and they have been used, for
example, to study the response of bone to the mechanical stimulation of artificial
implants [66], the apoptosis of vertebral disc cartilage cells caused by compression [67]
and the stimulation of growth at cranial sutures [68].
Biomechanical studies may be extended further, to the whole animal, to observe a
wide variety of responses, particularly those that depend on the function of entire
systems. For instance, rat hindlimb suspension has been used to compare the cellular
(and matrix) response of bone to the loss of mechanical loading as it would be difficult
to model this effect in an in vitro system [69]. Similarly, cellular changes have been
studied resulting from shear stresses in arterial remodelling [70] and gene expression
caused by mechanical stresses generated following lung injury [71]. Such data may
clarify transduction pathways or even inform clinical decisions [72]. It is important to
note that studies of cellular biomechanics do not necessarily follow the paradigm of
application of stress followed by an observation of altered cell function. The advent of
transgenic and knockout animal models means that key molecules in biomechanical
pathways can be modified or ablated and any change in function revealed [73]. In this
way, valuable new insights into cell function and disease may be gained.
4. Cellular & Molecular Biomechanics: State of the Art and Future Directions
The mechanical environment in which cells exist dictates the physiological attributes of
the cell, and this in turn regulates the function and maintenance of the tissue in which
the cell resides. It is clear that a number of cellular elements serve roles as
mechanoreceptors, and that the intracellular signaling pathways that become activated
by mechanical strain are diverse. Understanding exactly how cells respond to strain is
an emerging area of research, and the knowledge gained from investigations in this area
will be relevant to the tissue engineering field as we strive to develop paradigms with
which to control cellular differentiation and the maintenance of tissue phenotype.
References
[1] B. Martinac, Mechanosensitive ion channels: molecules of mechanotransduction, J Cell Sci. 117
(2004), 2449-60.
V.A. Campbell and B. O’Connell / Cellular & Molecular Biomechanics 211
[2] D.E. Ingber, Cellular mechanotransduction: putting all the pieces together again, Faseb J. 20 (2006),
811-27.
[3] B. Martinac and A. Kloda, Evolutionary origins of mechanosensitive ion channels, Prog Biophys Mol
Biol. 82 (2003), 11-24.
[4] S.Y. Lin and D.P. Corey, TRP channels in mechanosensation, Curr Opin Neurobiol. 15 (2005), 350-7.
[5] F.B. Kalapesi, J.C. Tan, and M.T. Coroneo, Stretch-activated channels: a mini-review. Are stretch
activated channels an ocular barometer? Clin Experiment Ophthalmol. 33 (2005), 210-7.
[6] A.J. el Haj, L.M. Walker, M.R. Preston, and S.J. Publicover, Mechanotransduction pathways in bone:
calcium fluxes and the role of voltage-operated calcium channels, Med Biol Eng Comput. 37 (1999),
403-9.
[7] M. Casado and P. Ascher, Opposite modulation of NMDA receptors by lysophospholipids and
arachidonic acid: common features with mechanosensitivity, J Physiol. 513 (1998), 317-30.
[8] T. Kawakubo, K. Naruse, T. Matsubara, N. Hotta, and M. Sokabe, Characterization of a newly found
stretch-activated KCa,ATP channel in cultured chick ventricular myocytes, Am J Physiol. 276 (1999),
H1827-38.
[9] J. Lammerding, R.D. Kamm, and R.T. Lee, Mechanotransduction in cardiac myocytes, Ann N Y Acad
Sci. 1015 (2004), 53-70.
[10] M.J. Cipolla, N.I. Gokina, and G. Osol, Pressure-induced actin polymerization in vascular smooth
muscle as a mechanism underlying myogenic behavior, Faseb J. 16 (2002), 72-6.
[11] F. Guharay and F. Sachs, Stretch-activated single ion channel currents in tissue-cultured embryonic
chick skeletal muscle, J Physiol. 352 (1984), 685-701.
[12] X.Wan, P. Juranka, and C.E. Morris, Activation of mechanosensitive currents in traumatized
membrane, Am J Physiol. 276 (1999), C318-27.
[13] S.F. Pedersen, G. Owsianik, and B. Nilius, TRP channels: an overview, Cell Calcium. 38 (2005), 233-
52.
[14] M.J. Caterina, T.A. Rosen, M. Tominaga, A.J. Brake, and D. Julius, A capsaicin-receptor homologue
with a high threshold for noxious heat, Nature 398 (1999), 436-41.
[15] H. Watanabe, J. Vriens, J. Prenen, G. Droogmans, T. Voets, and B. Nilius, Anandamide and
arachidonic acid use epoxyeicosatrienoic acids to activate TRPV4 channels, Nature 424 (2003), 434-8.
[16] R. Maroto, A. Raso, T.G. Wood, A. Kurosky, B. Martinac, and O.P. Hamill, TRPC1 forms the stretch
activated cation channel in vertebrate cells, Nat Cell Biol. 7 (2005), 179-85.
[17] W. Li, Z. Feng, P.W. Sternberg, and X.Z. Xu, A C. elegans stretch receptor neuron revealed by a
mechanosensitive TRP channel homologue, Nature 440 (2006), 684-7.
[18] G. Apodaca, Modulation of membrane traffic by mechanical stimuli, Am J Physiol Renal Physiol. 282
(2002), F179-90.
[19] K.S. Ko and C.A. McCulloch, Partners in protection: interdependence of cytoskeleton and plasma
membrane in adaptations to applied forces, J Membr Biol. 174 (2000), 85-95.
[20] Y. Oishi, Y. Uezono, N. Yanagihara, F. Izumi, T. Nakamura, and K. Suzuki, Transmural compression
induced proliferation and DNA synthesis through activation of a tyrosine kinase pathway in rat
astrocytoma RCR-1 cells, Brain Res. 781 (1998), 159-66.
[21] L.C. Martineau and P.F. Gardiner, Insight into skeletal muscle mechanotransduction: MAPK activation
is quantitatively related to tension, J Appl Physiol. 91 (2001), 693-702.
[22] C. Yan, M. Takahashi, M. Okuda, J.D. Lee, and B.C. Berk, Fluid shear stress stimulates big mitogen
activated protein kinase 1 (BMK1) activity in endothelial cells. Dependence on tyrosine kinases and
intracellular calcium, J Biol Chem. 274 (1999), 143-50.
[23] S. Lehoux, B. Esposito, R. Merval, and A. Tedgui, Differential regulation of vascular focal adhesion
kinase by steady stretch and pulsatility, Circulation. 111 (2005), 643-9.
[24] J.C. Krepinsky, Y. Li, D. Tang, L. Liu, J. Scholey, and A.J. Ingram, Stretch-induced Raf-1 activation
in mesangial cells requires actin cytoskeletal integrity, Cell Signal. 17 (2005), 311-20.
[25] J.P. Hatton, M. Pooran, C.F. Li, C. Luzzio, and M. Hughes-Fulford, A short pulse of mechanical force
induces gene expression and growth in MC3T3-E1 osteoblasts via an ERK 1/2 pathway, J Bone Miner
Res. 18 (2003), 58-66.
[26] T.E. Danciu, E. Gagari, R.M. Adam, P.D. Damoulis, and M.R. Freeman, Mechanical strain delivers
anti-apoptotic and proliferative signals to gingival fibroblasts, J Dent Res. 83 (2004), 596-601.
[27] H.R. Kim and C.M. Hai, Mechanisms of mechanical strain memory in airway smooth muscle, Can J
Physiol Pharmacol. 8 (2005), 811-5.
[28] R.J. Hendrickson, C. Cappadona, E.N. Yankah, J.V. Sitzmann, P.A. Cahill, and E.M. Redmond,
Sustained pulsatile flow regulates endothelial nitric oxide synthase and cyclooxygenase expression in
cocultured vascular endothelial and smooth muscle cells, J Mol Cell Cardiol. 31 (1999), 619-29.
212 V.A. Campbell and B. O’Connell / Cellular & Molecular Biomechanics
[29] X. Fan, J.A. Rahnert, T.C. Murphy, M.S. Nanes, E.M. Greenfield, and J. Rubin, Response to
mechanical strain in an immortalized pre-osteoblast cell is dependent on ERK1/2, J Cell Physiol. 207
(2006), 454- 60.
[30] M. Matsukawa, K. Fukuda, K. Yamasaki, K. Yoshida, H. Munakata, and C. Hamanishi, Enhancement
of nitric oxide and proteoglycan synthesis due to cyclic tensile strain loaded on chondrocytes attached
to fibronectin, Inflamm Res. 53 (2004), 239-44.
[31] J.G. McGarry, J. Klein-Nulend, and P.J. Prendergast, The effect of cytoskeletal disruption on pulsatile
fluid flow-induced nitric oxide and prostaglandin E2 release in osteocytes and osteoblasts, Biochem
Biophys Res Commun. 330 (2005), 341-8.
[32] J. Rubin, T.C. Murphy, L. Zhu, E. Roy, M.S. Nanes, and X. Fan, Mechanical strain differentially
regulates endothelial nitric-oxide synthase and receptor activator of nuclear kappa B ligand expression
via ERK1/2 MAPK, J Biol Chem. 278 (2003), 34018-25.
[33] G. Jiang, A.H. Huang, Y. Cai, M. Tanase, and M.P. Sheetz, Rigidity sensing at the leading edge
through alphavbeta3 integrins and RPTP_, Biophys J. 90 (2006), 1804-9.
[34] K. Yamada, K.G. Green, A.M. Samarel, and J.E. Saffitz, Distinct pathways regulate expression of
cardiac electrical and mechanical junction proteins in response to stretch, Circ Res. 97 (2005), 346-53.
[35] S.T. Rhee and S.R. Buchman, Colocalization of c-Src (pp60src) and bone morphogenetic protein 2/4
expression during mandibular distraction osteogenesis: in vivo evidence of their role within an
integrinmediated mechanotransduction pathway, Ann Plast Surg. 55 (2005), 207-15.
[36] D. Riveline, E. Zamir, N.Q. Balaban, U.S. Schwarz, T. Ishizaki, S. Narumiya, Z. Kam, B. Geiger, and
A.D. Bershadsky, Focal contacts as mechanosensors: externally applied local mechanical force induces
growth of focal contacts by an mDia1 dependent and ROCK-independent mechanism, J Cell Biol. 153
(2001), 1175-86.
[37] M.E. Chicurel, R.H. Singer, C.J. Meyer, and D.E. Ingber, Integrin binding and mechanical tension
induce movement of mRNA and ribosomes to focal adhesions, Nature 392 (1998), 730-3.
[38] S.M. Norvell, M. Alvarez, J.P. Bidwell, and F.M. Pavalko, Fluid shear stress induces beta-catenin
signaling in osteoblasts, Calcif Tissue Int. 75 (2004), 396-404.
[39] H. Henderson and D.R. Carter, Mechanical induction in limb morphogenesis: the role of growth
generated strains and pressures, Bone. 31 (2002), 645-53.
[40] B.K. Hall and S.W. Herring, Paralysis and growth of the musculoskeletal system in the embryonic
chick, J Morphol. 206 (1990), 45-56.
[41] M. Schmelter, B. Ateghang, S. Helmig, M. Wartenberg, and H. Sauer, Embryonic stem cells utilize
reactive oxygen species as transducers of mechanical strain-induced cardiovascular differentiation,
Faseb J. 20 (2006), 1182-4.
[42] M.F. Pittenger, A.M. Mackay, S.C. Beck, R.K. Jaiswal, R. Douglas, J.D. Mosca, M.A. Moorman,
D.W. Simonetti, S. Craig, and D.R. Marshak, Multilineage potential of adult human mesenchymal
stem cells, Science 284 (1999), 143-7.
[43] A. Wiesmann, H.J. Buhring, C. Mentrup, and H.P. Wiesmann, Decreased CD90 expression in human
mesenchymal stem cells by applying mechanical stimulation, Head Face Med. 2 (2006), 8.
[44] J.H. Cui, K. Park, S.R. Park, and B.H. Min, Effects of low-intensity ultrasound on chondrogenic
differentiation of mesenchymal stem cells embedded in polyglycolic acid: an in vivo study, Tissue Eng.
12 (2006), 75-82.
[45] M. Knippenberg, M.N. Helder, B.Z. Doulabi, C.M. Semeins, P.I. Wuisman, and J. Klein-Nulend,
Adipose tissue-derived mesenchymal stem cells acquire bone cell-like responsiveness to fluid shear
stress on osteogenic stimulation, Tissue Eng. 11 (2005), 1780-8.
[46] R.C. Riddle, A.F. Taylor, D.C. Genetos, and H.J. Donahue, MAP kinase and calcium signaling mediate
fluid flow-induced human mesenchymal stem cell proliferation, Am J Physiol Cell Physiol. 290 (2006),
C776-84.
[47] E. Tzima, M. Irani-Tehrani,W.B. Kiosses, E. Dejana, D.A. Schultz, B. Engelhardt, G. Cao, H.
DeLisser, and M.A. Schwartz, A mechanosensory complex that mediates the endothelial cell response
to fluid shear stress, Nature 437 (2005), 426-31.
[48] C.T. Rubin and L.E. Lanyon, Regulation of bone mass by mechanical strain magnitude, Calcif Tissue
Int. 37 (1985), 411-7.
[49] G. Carmeliet, L. Vico, and R. Bouillon, Space flight: a challenge for normal bone homeostasis, Crit
Rev Eukaryot Gene Expr. 11 (2001), 131-44.
[50] M. Mehrotra, M. Saegusa, S.Wadhwa, O. Voznesensky, D. Peterson, and C. Pilbeam, Fluid flow
induces Rankl expression in primary murine calvarial osteoblasts, J Cell Biochem. 98 (2006), 1271-83.
[51] N. Kizer, X.L. Guo, and K. Hruska, Reconstitution of stretch-activated cation channels by expression
of the alpha-subunit of the epithelial sodium channel cloned from osteoblasts, Proc Natl Acad Sci U S
A. 94 (1997), 1013-8.
V.A. Campbell and B. O’Connell / Cellular & Molecular Biomechanics 213
[52] J. Li, R.L. Duncan, D.B. Burr, and C.H. Turner, L-type calcium channels mediate mechanically
induced bone formation in vivo, J Bone Miner Res. 17 (2002), 1795-800.
[53] J. Rubin, C. Rubin, and C.R. Jacobs, Molecular pathways mediating mechanical signaling in bone,
Gene 367 (2006), 1-16.
[54] S. Kapur, D.J. Baylink, and K.H. Lau, Fluid flow shear stress stimulates human osteoblast proliferation
and differentiation through multiple interacting and competing signal transduction pathways, Bone 32
(2003), 241-51.
[55] F.A. Weyts, Y.S. Li, J. van Leeuwen, H. Weinans, and S. Chien, ERK activation and alpha v beta 3
integrin signaling through Shc recruitment in response to mechanical stimulation in human osteoblasts,
J Cell Biochem. 87 (2002), 85-92.
[56] T. Ishida, T.E. Peterson, N.L. Kovach, and B.C. Berk, MAP kinase activation by flow in endothelial
cells. Role of beta 1 integrins and tyrosine kinases, Circ Res. 79 (1996), 310-6.
[57] A.D. Bakker, J. Klein-Nulend, E. Tanck, I.C. Heyligers, G.H. Albers, P. Lips, and E.H. Burger,
Different responsiveness to mechanical stress of bone cells from osteoporotic versus osteoarthritic
donors, Osteoporos Int. 17 (2006), 827-33.
[58] Y.J. Kim, R.L. Sah, A.J. Grodzinsky, A.H. Plaas, and J.D. Sandy, Mechanical regulation of cartilage
biosynthetic behavior: physical stimuli, Arch Biochem Biophys. 311 (1994), 1-12.
[59] Q.Q. Wu and Q. Chen, Mechanoregulation of chondrocyte proliferation, maturation, and hypertrophy:
ion-channel dependent transduction of matrix deformation signals, Exp Cell Res. 256 (2000), 383-91.
[60] [60] T.T. Chowdhury, R.N. Appleby, D.M. Salter, D.A. Bader, and D.A. Lee, Integrin-mediated
mechanotransduction in IL-1 beta stimulated chondrocytes, Biomech Model Mechanobiol. 5 (2006),
192-201.
[61] C.V. Gemmiti and R.E. Guldberg, Fluid flow increases type II collagen deposition and tensile
mechanical properties in bioreactor-grown tissue-engineered cartilage, Tissue Eng. 12 (2006), 469-79.
[62] T.P. Lele, J.E. Sero, B.D. Matthews, S. Kumar, S. Xia, M. Montoya-Zavala, T. Polte, D. Overby, N.
Wang, and D.E. Ingber, Tools to study cell mechanics and mechanotransduction, Methods Cell Biol. 83
(2007), 443-72.
[63] D. Stamenovic and N. Wang, Invited review: engineering approaches to cytoskeletal mechanics, J Appl
Physiol. 89 (2000), 2085-90.
[64] [64] P. Maguire, J.I. Kilpatrick, G. Kelly, P.J. Prendergast, V.A. Campbell, B.C. O’Connell, and S.P.
Jarvis, Direct mechanical measurement of geodesic structures in rat mesenchymal stem cells, HFSP
Journal Epub ahead of print (2007).
[65] J.G. McGarry, P. Maguire, V.A. Campbell, B.C. O’Connell, P.J. Prendergast, and S.P. Jarvis,
Stimulation of nitric oxide mechanotransduction in single osteoblasts using atomic force microscopy, J
Orthop Res, 26 (2008), 513-21.
[66] V. Mann, C. Huber, G. Kogianni, D. Jones, and B. Noble, The influence of mechanical stimulation on
osteocyte apoptosis and bone viability in human trabecular bone, J Musculoskelet Neuronal Interact. 6
(2006), 408-17.
[67] K. Ariga, K. Yonenobu, T. Nakase, N. Hosono, S. Okuda, W. Meng, Y. Tamura, and H. Yoshikawa,
Mechanical stress-induced apoptosis of endplate chondrocytes in organ-cultured mouse intervertebral
discs: an ex vivo study, Spine 28 (2003), 1528-33.
[68] S.S. Tholpady, T.F. Freyman, D. Chachra, and R.C. Ogle, Tensional forces influence gene expression
and sutural state of rat calvariae in vitro, Plast Reconstr Surg. 120 (2007), 601-11; discussion 612-3.
[69] D.A. Hardiman, F.J. O’Brien, P.J. Prendergast, D.T. Croke, A. Staines, and T.C. Lee, Tracking the
changes in unloaded bone: Morphology and gene expression, Eur J Morphol. 42 (2005), 208-16.
[70] Y.H. Li, C.Y. Hsieh, D.L. Wang, H.C. Chung, S.L. Liu, T.H. Chao, G.Y. Shi, and H.L. Wu,
Remodeling of carotid arteries is associated with increased expression of thrombomodulin in a mouse
transverse aortic constriction model, Thromb Haemost. 97 (2007), 658-64.
[71] B.A. Simon, R.B. Easley, D.N. Grigoryev, S.F. Ma, S.Q. Ye, T. Lavoie, R.M. Tuder, and J.G. Garcia,
Microarray analysis of regional cellular responses to local mechanical stress in acute lung injury, Am J
Physiol Lung Cell Mol Physiol. 291 (2006), L851-61.
[72] M.C. Meikle, Remodeling the dentofacial skeleton: the biological basis of orthodontics and dentofacial
orthopedics, J Dent Res. 86 (2007), 12-24.
[73] J.P. Schmitt, E.P. Debold, F. Ahmad, A. Armstrong, A. Frederico, D.A. Conner, U. Mende, M.J.
Lohse, D. Warshaw, C.E. Seidman, and J.G. Seidman, Cardiac myosin missense mutations cause
dilated cardiomyopathy in mouse models and depress molecular motor function, Proc Natl Acad Sci U
S A. 103 (2006), 14525-30.
214 Basic Engineering for Medics and Biologists
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-214
Introduction
What is a Bioreactor?
Bioreactors have been used for many years in areas other than tissue engineering. They
have been used in diverse areas such as in fermentation, in water treatment, in food
processing and in the production of pharmaceuticals [1]. All of these bioreactors are
devices in which biological or biochemical processes develop under a closely
monitored and tightly controlled environment. Bioreactors have been used in animal
cell culture since the 1980s in order to produce vaccines and other drugs and to culture
large cell populations. Bioreactors for use in tissue engineering have progressed from
such devices.
A tissue engineering bioreactor can be defined as a device that uses mechanical
means to influence biological processes [2]. In tissue engineering, this generally means
that bioreactors are used to stimulate cells and encourage them to produce extra-cellular
matrix (ECM). There are numerous types of bioreactor which can be classified by the
means they use to stimulate cells. A number of these will be discussed below.
Tissue engineering technologies are based on the biological triad of cells, signalling
mechanisms and extracellular matrix. To simulate the development of tissues in vitro,
tissue engineering aims to optimise cell growth, by providing regulatory signals in the
form of growth factors and a regeneration template in the form of scaffold. Bioreactors
may be used as an alternative to or in conjunction with growth factors in the signalling
part of the triad. Defects requiring tissue-engineering solutions are typically many
millimetres in size [3]. Scaffolds in such a size range are easily fabricated. However,
problems arise when culturing cells on these scaffolds. Static culture conditions result
in scaffolds with few cells in the centre of the construct [4]. This heterogeneous cell
distribution is a major obstacle to developing any three-dimensional tissue or organ in
vitro. It has been shown that despite homogeneous cell seeding, after long periods in
culture, more cells are found on the periphery of demineralised trabecular bone
constructs [4]. This is due to cell necrosis and cell chemotaxis. Necrosis occurs at the
centre of the scaffold due to a lack of nutrient delivery to, and waste removal from that
area. The only mechanism by which nutrients and waste can move when a scaffold is in
N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering 215
static culture is by diffusion. As the size of the scaffold increases, diffusion to the
centre of the construct becomes more difficult. In addition, as cells on the periphery
grow and secrete extracellular matrix, movement of fluid to the interior of the scaffold
is further impeded. Chemotaxis of the cells from the interior towards the periphery
occurs because of the concentration gradient in nutrients that has been set up [3].
Nutrient concentration is greater at the periphery so cells move along this gradient
towards the periphery in order to obtain the nutrients they require. It has been found
that mineralised bone matrix reaches a maximum penetration depth of 240 μm on
poly( DL-lactic-co-glycolic acid) scaffolds seeded with stromal osteoblasts, which is
far thinner than an ideal bone graft replacement [5].
Figure 1. Histological sections of tissue engineered vascular tissue cultured using a bioreactor (on left) and
statically (on right). Sections stained with Verhoff’s elastin stain [6].
In general, bioreactors are designed to perform at least one of the following five
functions: 1) providing uniform cell distribution, 2) maintaining the desired
concentration of gases and nutrients in the medium, 3) providing mass transport to the
tissue, 4) exposing tissue to physical stimuli, or 5) providing information about the
formation of 3D tissue [10].
While the detailed requirements for bioreactor design are tissue- and/or
application- specific, there are a few general principles which have to be adhered to
when developing a bioreactor. The material selection is very important as it is vital to
ensure that the materials used to create the bioreactor do not elicit any adverse reaction
from the cultured tissue. Any material which is in contact with media must be
biocompatible or bioinert. This eliminates the use of most metals, although stainless
steel can be used if it is treated so that chromium ions do not leach out into the medium.
Numerous plastics comply with this constraint but there are further limitations on
material selection that must also be kept in mind. Materials must be usable at 37°C in a
humid atmosphere. They must be sterilisable if they are to be re-used. Bioreactor parts
can be sterilised by autoclaving or disinfected by submersion in alcohol. If they are to
be autoclaved, materials that can withstand numerous cycles of high temperature and
pressure must be used in bioreactor manufacture. Alternatively, some non-sterilisable
disposable bioreactor parts may be used which can be replaced after each use of the
bioreactor. Other material choices are between transparent or opaque and flexible or
inflexible materials. Materials with different properties are needed for various
components in the bioreactor. For example, transparent materials can be of benefit in
allowing the construct to be monitored in the bioreactor during culture while flexible
tubing can help with assembly of the bioreactor.
The design of the bioreactor should be as simple as possible, avoiding the
introduction of, for example, machined recesses which could collect condensed steam
during autoclaving and become breeding grounds for micro-organisms. Simplicity in
design should also mean that the bioreactor is quick to assemble and disassemble.
Apart from being more efficient, this ensures that cell-seeded constructs inserted into
the bioreactor are out of the incubator for the minimum amount of time possible. This
minimises the risk to the cells and the experiment being undertaken.
The specific application of the bioreactor must be kept in mind during the design
process to ensure that all the design constraints are met. If various parameters such as
pH, nutrient concentration or oxygen levels are to be monitored, these sensors should
be incorporated into the design. If a pump or motor is to be used, it must be small
enough to fit into an incubator and also be usable at 37°C and in a humid environment.
The forces needed for cellular stimulation are very small so it is important to ensure
that the pump/motor has the capability to apply small forces accurately. In any design
involving fluids, problems can arise with leaking fluid seals and, if possible, the need
for seals should be removed. However, in most cases, fluid seals are necessary and
good design should decrease the problems with them. If a prototype bioreactor is being
designed, it is worthwhile thinking about scale up opportunities for the bioreactor from
the outset. This may mean designing a device that is relatively easy to enlarge without
changing its characteristics or designing a simple device of which many more can be
made so that numerous scaffolds can be cultured at one time.
N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering 217
Figure 2. A spinner flask bioreactor. Scaffolds are suspended in medium and the medium stirred using a
magnetic stirrer to improve nutrient delivery to the scaffold (Kim et al., 2005)
The rotating wall bioreactor was developed by NASA [14]. It was originally designed
with a view to protecting cell culture experiments from high forces during space shuttle
take off and landing. However, the device has proved useful in tissue engineering here
218 N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering
on earth. In a rotating wall bioreactor (Figure 3), scaffolds are free to move in media in
a vessel. The wall of the vessel rotates, providing an upward hydrodynamic drag force
that balances with the downward gravitational force, resulting in the scaffold remaining
suspended in the media. Fluid transport is enhanced in a similar fashion to the
mechanism in spinner flasks and the devices also provide more homogeneous cell
distribution than static culture [3]. Gas exchange occurs through a gas exchange
membrane and the bioreactor is rotated at speeds of 15-30 rpm. Cartilage tissue of 5
mm thickness has been grown in this type of bioreactor after seven months of culture
[15]. As tissue grows in the bioreactor, the rotational speed must be increased in order
to balance the gravitational force and ensure the scaffold remains in suspension.
Figure 3. A rotating wall vessel bioreactor. Scaffolds are suspended in medium due to opposing gravitational
and drag forces (www.che.msstate.edu/research/MDERL/demos/DEMoBioRctrExperiment)
Another widely used type of bioreactor is the compression bioreactor. This class of
bioreactor is generally used in cartilage engineering and can be designed so that both
static loading and dynamic loading can be applied. This is because static loading has
been found to have a negative effect on cartilage formation while dynamic loading,
which is more representative of physiological loading, has provided better results than
many other stimuli [16].
In general, compression bioreactors consist of a motor, a system providing linear
motion and a controlling mechanism. An example of such a system is shown in Figure
4, where a cam-follower system is used to provide displacements of different
magnitudes and frequencies. A signal generator can be used to control the system and
load cells and linear variable differential transformers can be used to measure the load
response and imposed displacement respectively [8, 17]. The load can be transferred to
the cell-seeded constructs via flat platens which distribute the load evenly, however in a
device for stimulating multiple scaffolds simultaneously, care must be taken that the
constructs are of similar height or the compressive strain applied will vary as the
N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering 219
Figure 4. A compression bioreactor. Constructs are housed in a petri dish and subjected to compressive
forces exerted on them via a cam and follower mechanism [41].
chamber can be pressurised using a piston controlled by an actuator [16]. For sterility,
the piston can apply pressure via an impermeable membrane so that the piston itself
does not come into contact with the culture media. Variations on this design include a
water-filled pressure chamber which pressurises a media-filled chamber via an
impermeable film and is controlled using a variable backpressure valve and an actuator
[20].
Figure 5. A hydrostatic pressure bioreactor. Constructs are placed in a chamber which is subsequently
pressurised to the required level [16].
Flow perfusion bioreactors generally consist of a pump and a scaffold chamber joined
together by tubing. A media reservoir may also be present. The scaffold is kept in
position across the flow path of the device (Figure 6). Media is perfused through the
scaffold, thus enhancing fluid transport.
Culture using flow perfusion bioreactors has been shown to provide more
homogeneous cell distribution throughout scaffolds. Collagen sponges have been
seeded with bone marrow stromal cells and perfused with flow. This has resulted in
greater cellularity throughout the scaffold in comparison to static controls, implying
that better nutrient exchange occurs due to flow [21]. Using a biphasic calcium-
phosphate scaffold, abundant ECM with nodules of CaP was noted after 19 days in
steady flow culture [22].
In comparisons between flow perfusion, spinner flask and rotating wall bioreactors,
flow perfusion bioreactors have proved to be the best for fluid transport. Using the
same flow rate and the same scaffold type, while cell densities remained the same using
all three bioreactors, the distribution of the cells changed dramatically depending on
which bioreactor was used. Histological analysis showed that spinner flask and static
culture resulted in the majority of viable cells being on the periphery of the scaffold. In
N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering 221
contrast, the rotating wall vessel and flow perfusion bioreactor culture resulted in
uniform cell distribution throughout the scaffolds [3, 23]. After 14 days in culture, the
perfusion bioreactor had higher cell density than all other culture methods [3].
In our laboratory, a flow perfusion bioreactor has been developed to examine the
effects of different flow profiles on cell-seeded collagen-glycosaminoglycan scaffolds
[24]. The scaffold chamber was specifically designed to ensure that the compliant
scaffold was under perfusive flow. This involved using a scaffold of larger diameter
than the flow path and using spacers to ensure the scaffold was under 10% compression
during culture. A programmable syringe pump was used in order to stimulate the cell-
seeded constructs using different flow profiles. As discussed below, this device
demonstrated that intermittent flow perfusion is advantageous for mechanically
stimulating osteoblasts while maintaining cell viability.
Figure 6. A flow perfusion bioreactor (A). Media is forced through the scaffold in the scaffold chamber (B)
by the syringe pump [24].
attempt to develop the tissue in the laboratory (Figure 7). Vocal fold tissue experiences
vibrations of 100-1000 Hz at an amplitude of approximately 1 mm, which is a unique
stimulus in the human body [27].
Figure 7. A bioreactor for vocal fold tissue engineering. Constructs can be stimulated using both tensile and
vibrational forces [26].
2.1. Bone
Bone is comprised of a mineral phase, collagen and cells. The mineral phase of bone is
a hard, brittle material (hydroxyapatite) that is strong in compression but cannot
withstand large shear or tensile loads. In contrast, collagen fibres are strong in tension
and flexible in bending but weak in compression. Bone grafts are required to aid bone
defect and non-union healing. The aim in using them is to restore function to the
damaged area as quickly and completely as possible [29]. They are required in a
number of procedures including, for example, replacing diseased bone, filling bone
voids after non-unions or cyst removal, reconstructive surgery and in spinal fusion
operations. The most commonly used graft in bone replacement is the autograft.
Autografts are grafts taken from the patients themselves, usually from the iliac crest,
N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering 223
although the distal femur, the greater trochanter, and the proximal tibia can also be used
[29]. This process is expensive and the size of graft that can be obtained is limited.
Morbidity of the site the graft is acquired from is another problem and complications
can arise due to infection and chronic pain [29]. An alternative to the autograft is the
allograft. The term allograft is used for bone grafts which are obtained from an organ
donor. A drawback to this option is the danger of infection. Xenografts, acquired from
animal bone, are another alternative. However, the morphology of the bone is different
to that of humans and the risk of cross-species infection exists. Coral has also been
used but its structure is very different to that of bone and hence osteointegration of this
type of graft has proved difficult [30].
The use of natural bone grafts has proved problematic and therefore attention has
turned to tissue engineering. Engineered tissue must be strong enough to allow load
bearing after implantation. Resorbability of the scaffold is another important issue.
Ideally, as the scaffold is resorbed, bone should be deposited in its place, thus ensuring
no loss in mechanical strength. A morphology that allows movement of cells and
supports vascularisation is important for a scaffold material. A high porosity and
controllable pore size can provide this. Cells must be able to penetrate into the core of
the scaffold so that they are homogeneously distributed throughout the graft. There
must be space to allow for transport of nutrients to and waste removal from cells, so the
pores must be interconnected [31]. Thus, a balance must be struck between mechanical
strength and porosity to develop an ideal scaffold for bone tissue engineering. The use
of a bioreactor with such a scaffold should provide a homogeneous distribution of
stimulated cells.
It has been shown that fluid flow can stimulate bone cells to increase levels of
bone formation markers [4, 32-37] and its use could improve mineralisation of the
scaffold on which cells are seeded. Flow perfusion bioreactors increase alkaline
phosphatase (ALP) expression after 7 and 14 days more than spinner flasks or rotating
wall vessels [3] and are more commonly used than any other bioreactor for use in 3-D
stimulation studies. In one study, MC3T3-E1 cells were seeded on decalcified human
trabecular bone, the flow rate of perfusion altered and the mRNA expression of Runx2,
OC and ALP measured [4]. It was found that using a steady flow rate of only 1 mL/min
killed nearly all the cells on the scaffold after 7 days in culture. However, a flow rate of
0.01 mL/min led to a high proportion of viable cells both on the surface and inside the
scaffold. This compared favourably to static culture, where cells were predominantly
on the periphery [4].
Using a CaP scaffold and a flow rate of 0.025 mL/min in a flow perfusion
bioreactor, PGE2 levels were found to increase over static controls. When a stimulus of
30 minutes of oscillatory flow at 1 Hz with a 40 mL/min peak was superimposed on the
steady flow, PGE2 levels increased further. The number of cells left residing on the
scaffolds decreased due to this large dynamic stimulus but this decrease was not found
to be statistically significant [38]. When different flow profiles were used intermittently
to stimulate cells on a highly porous collagen-glycosaminoglycan scaffold in our
laboratory, it was found that intermittent flow caused greater stimulation than a
continuous low flow rate without a loss in cell number [39]. Cyclyoxygenase-2 and
osteopontin expression increased due to culture in the bioreactor, as did prostaglandin
E2 production. This lends further backing to the hypothesis that the combination of a
perfusion period (for nutrient delivery and waste removal) and a stimulation period
may deliver enhanced fluid transport with enhanced stimulation of cells and may yet
prove to be the optimum regime for bioreactor culture of bone cells.
224 N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering
2.2. Cartilage
There are three main types of tissue that are encompassed by the heading
cardiovascular tissue, namely: vascular, cardiac muscle and heart valve tissue. These
different tissues experience different mechanical environments in the body and
therefore, in tissue engineering, they are cultured using different bioreactors and under
N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering 225
Figure 8. Structure of an artery. The three layers of the tissue (the tunical intima, the tunica media and the
tunica externa) can be clearly seen [40].
muscle cells. The smooth muscle cells are therefore more likely to be quiescent and this
decreases the risk of thrombosis formation and luminal occlusion [6]. In order to
recreate the structure of blood vessels in vitro, co-cultures of different cell types must
be used. Scaffolds can be cultured with smooth muscle cells initially and then a layer of
endothelial cells can be seeded as the culture period nears its end.
The arteries nearest the heart experience the highest pressures and they have
resilient walls in order to cope with these changing pressures. They are known as
elastic arteries and their tunica media contains a high proportion of elastic fibres. This
enables them to expand as the pressure inside the lumen increases and recoil when the
pressure decreases, thus allowing them to deal with pressure changes but also to damp
the pressure oscillations to make blood flow continuous rather than oscillatory in nature
[40]. It is this expansion and recoil of blood vessels that is the inspiration for a number
of bioreactors for use in vascular tissue engineering.
Some success has been achieved when culturing vascular tissue in both rotating
wall vessels [13] and biomimetic bioreactors [6, 25]. Biomimetic bioreactors use
combinations of strain and perfusion to stimulate vascular tissue development. In one
such bioreactor, a tube of biodegradable polymer was seeded with smooth muscle cells
and silicone tubing placed through the centre of the tube. The construct was maintained
in a bath of stirred media and the highly distensible silicone tubing was pulsed at 165
beats per minute so that the scaffold experienced 5% radial strain over an eight week
culture period [6]. The silicone tubing was then removed, an endothelial cell lining
applied to the inner layer of the construct and culture medium perfused through the
centre of the scaffold. This set up resulted in vessels with a rupture strength similar to
that of native tissue and with a morphology that showed migration of smooth muscle
cells throughout the construct. In contrast, scaffolds kept in static culture over the same
culture period showed no migration of smooth muscle cells and therefore an
inhomogeneous structure. A bioreactor has also been designed to apply both radial and
axial strain to cell-seeded constructs, to mimic the environment in the body more
closely [25]. The bioreactor has two perfusion loops that feed the internal and external
parts of the scaffold. The internal perfusion can be pulsatile in nature and is of variable
frequency.
2.3.2 Cardiac Muscle Tissue
Cardiac muscle cells are called cardiocytes or cardiomyocytes. They form extensive
connections with each other and are almost totally dependent on aerobic metabolism
for energy. When a myocardial infarction (heart attack) occurs, part of the blood supply
to the heart muscle is blocked and cardiac muscle cells die due to lack of oxygen
delivery [40]. Cardiac muscle only has a limited ability to regenerate and scar tissue
formed may create additional problems by constricting other vessels in the damaged
region [40, 43]. Cardiac muscle tissue engineering is therefore of interest in developing
a method for myocardial repair. Oxygen delivery to cardiac cells is vital in order that a
sufficiently thick layer of tissue can be grown in vitro. This means that the use of
bioreactors that increase oxygen delivery to cells is important in cardiac muscle tissue
engineering. The ideal tissue-engineered cardiac muscle tissue graft should have a
dense, uniform distribution of cardiomyocytes, should contract in response to an
electrical stimulus and have the mechanical integrity to allow implantation [43].
Polyglycolic acid (PGA) scaffolds have been seeded with cardiomyocytes and
cultured in spinner flasks and rotating wall bioreactors [44, 45]. In both these culturing
systems, a peripheral region densely populated with cells and a central region with a
N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering 227
The use of bioreactors has brought us a step closer to engineering numerous tissue
types in the laboratory. At present, most bioreactors are specialised devices with low
volume output. Their assembly is often time consuming and labour intensive. Many
also exhibit operator dependent variability. While scaled-up versions of some devices
may be useful for developing larger amounts of tissue, problems with process
consistency and process contamination may persist. A better understanding of the
different effects of mechanical stimulation on cell signalling and mechanotransduction
is also needed. This can be achieved through the use of existing simple bioreactors in
conjunction with numerical simulation of culture conditions to minimise the number of
experiments needed.
228 N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering
In the future, ways of minimising the time and effort needed in order to form tissue
must be found if costs are to be minimised and the use of engineered tissue is to
become routine clinically. One way to do this is to automate the process. The ideal
bioreactor would need autologous material and a scaffold as inputs and, after a defined
culture period, would output the required amount of the required tissue. Automated
systems for culturing cells already exist and work has begun on extending them to
incorporate mechanical stimulation into the culturing process. Aastrom has developed a
system that takes bone marrow as an input and expands the stem and progenitor cell
population. The Automation Partnership has numerous systems for expanding cells. If
systems such as these are used in conjunction with monitoring systems and a feedback
loop, so that factors such as the temperature, oxygen level and pH can be regulated, cell
culture can be optimised. Closed bioreactor systems for seeding and culturing skin
grafts under perfusion were developed by Advanced Tissue Sciences before the
company’s liquidation in 2002. As this system demonstrated, the technology exists to
enable the incorporation of mechanical stimulus into an automated cell culture system
and this may be the future for bioreactors in tissue engineering.
References
[1] I. Martin, D. Wendt and M. Heberer, The role of bioreactors in tissue engineering, Trends Biotechnol
22 (2004), 80-6.
[2] E. M. Darling and K. A. Athanasiou, Biomechanical strategies for articular cartilage regeneration, Ann
Biomed Eng 31 (2003), 1114-24.
[3] A. S. Goldstein, T. M. Juarez, C. D. Helmke, M. C. Gustin and A. G. Mikos, Effect of convection on
osteoblastic cell growth and function in biodegradable polymer foam scaffolds,Biomaterials 22 (2001),
1279-88.
[4] S. H. Cartmell, B. D. Porter, A. J. Garcia and R. E. Guldberg, Effects of medium perfusion rate on cell-
seeded three-dimensional bone constructs in vitro, Tissue Eng 9 (2003), 1197-203.
[5] S. L. Ishaug, G. M. Crane, M. J. Miller, A.W. Yasko, M. J. Yaszemski and A. G. Mikos, Bone
formation by three-dimensional stromal osteoblast culture in biodegradable polymer scaffolds, J.
Biomed. Mater. Res. 36 (1997), 17-28.
[6] L. E. Niklason, J. Gao, W. M. Abbott, K. K. Hirschi, S. Houser, R. Marini and R. Langer, Functional
arteries grown in vitro, Science 284 (1999), 489-93.
[7] S. E. Carver and C. A. Heath, Semi-continuous perfusion system for delivering intermittent
physiological pressure to regenerating cartilage, Tissue Eng 5 (1999), 1-11.
[8] R. L. Mauck, M. A. Soltz, C. C. Wang, D. D. Wong, P. H. Chao, W. B. Valhmu, C. T. Hung and G. A.
Ateshian, Functional tissue engineering of articular cartilage through dynamic loading of chondrocytes
seeded agarose gels, J Biomech Eng 122 (2000), 252-60.
[9] G. H. Altman, R. L. Horan, I. Martin, J. Farhadi, P. R. Stark, V. Volloch, J. C. Richmond, G. Vunjak-
Novakovic and D. L. Kaplan, Cell differentiation by mechanical stress, Faseb J 16 (2002), 270-2.
[10] V. Barron, E. Lyons, C. Stenson-Cox, P. E. McHugh and A. Pandit, Bioreactors for cardiovascular cell
and tissue growth: A review, Ann Biomed Eng 31 (2003), 1017-30.
[11] Y. Martin and P. Vermette, Bioreactors for tissue mass culture: Design, characterization, and recent
advances, Biomaterials 26 (2005), 7481-503.
[12] R. I. Freshney, Culture of animal cells, Wiley-Liss, New York, 2000.
[13] L. E. Freed and G. Vunjak-Novakovic, Tissue engineering bioreactors. in: R. P. Lanza, R. Langer and
J. Vacanti, (Eds.), Principles of tissue engineering, Academic Press, San Diego, CA, 2000, pp. 143-
156.
[14] R. P. Schwarz, T. J. Goodwin and D. A. Wolf, Cell culture for three-dimensional modeling in rotating
wall vessels: An application of simulated microgravity, J Tissue Cult Methods 14 (1992), 51-7.
[15] L. E. Freed, R. Langer, I. Martin, N. R. Pellis and G. Vunjak-Novakovic, Tissue engineering of
cartilage in space, Proc Natl Acad Sci U.S.A 94 (1997), 13885-90.
[16] E. M. Darling and K. A. Athanasiou, Articular cartilage bioreactors and bioprocesses, Tissue Eng 9
(2003), 9-26.
N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering 229
[17] C. Y. Huang, K. L. Hagar, L. E. Frost, Y. Sun and H. S. Cheung, Effects of cyclic compressive loading
on chondrogenesis of rabbit bone-marrow derived mesenchymal stem cells, Stem Cells 22 (2004), 313-
23.
[18] L. A. McMahon, A. J. Reid, V. A. Campbell and P. J. Prendergast, Regulatory effects of mechanical
strain on the chondrogenic differentiation of mscs in a collagen-gag scaffold: Experimental and
computational analysis, Ann Biomed Eng 36 (2008), 185-94.
[19] J. Garvin, J. Qi, M. Maloney and A. J. Banes, Novel system for engineering bioartificial tendons and
application of mechanical load, Tissue Eng 9 (2003), 967-79.
[20] S.Watanabe, S. Inagaki, I. Kinouchi, H. Takai, Y. Masuda and S. Mizuno, Hydrostatic
pressure/perfusion culture system designed and validated for engineering tissue, J Biosci Bioeng 100
(2005), 105-11.
[21] J. Glowacki, S. Mizuno and J. S. Greenberger, Perfusion enhances functions of bone marrow stromal
cells in three-dimensional culture, Cell Transplant 7 (1998), 319-26.
[22] F. W. Janssen, J. Oostra, A. Oorschot and C. A. van Blitterswijk, A perfusion bioreactor system
capable of producing clinically relevant volumes of tissue-engineered bone: in vivo bone formation
showing proof of concept, Biomaterials 27 (2006), 315-23.
[23] X. Yu, E. A. Botchwey, E. M. Levine, S. R. Pollack and C. T. Laurencin, Bioreactor-based bone tissue
engineering: The influence of dynamic flow on osteoblast phenotypic expression and matrix
mineralization, Proc Natl Acad Sci U S A 101 (2004), 11203-8.
[24] M. J. Jaasma, N. A. Plunkett and F. J. O’Brien, Design and validation of a dynamic flow perfusion
bioreactor for use with compliant tissue engineering scaffolds, J Biotechnol. 133 (2008), 490.
[25] K. Bilodeau, F. Couet, F. Boccafoschi and D. Mantovani, Design of a perfusion bioreactor specific to
the regeneration of vascular tissues under mechanical stresses, Artif Organs 29 (2005), 906-12.
[26] I. R. Titze, R. W. Hitchcock, K. Broadhead, K. Webb, W. Li, S. D. Gray and P. A. Tresco, Design and
validation of a bioreactor for engineering vocal fold tissues under combined tensile and vibrational
stresses, J Biomech 37 (2004), 1521-9.
[27] I. R. Titze, On the relation between subglottal pressure and fundamental frequency in phonation, J
Acoust Soc Am 85 (1989), 901-6.
[28] M. C. Wallis, H. Yeger, L. Cartwright, Z. Shou, M. Radisic, J. Haig, M. Suoub, R. Antoon and W. A.
Farhat, Feasibility study of a novel urinary bladder bioreactor, Tissue Eng (2007).
[29] C. R. Perry, Bone repair techniques, bone graft, and bone graft substitutes, Clin Orthop Relat Res 360
(1999), 71-86.
[30] J. Y. de la Caffiniere, E. Viehweger and A. Worcel, [long-term radiologic evolution of coral implanted
in cancellous bone of the lower limb. Madreporic coral versus coral hydroxyapatite], Rev Chir Orthop
Reparatrice Appar Mot 84 (1998), 501-7.
[31] F. J. O’Brien, B. A. Harley, I. V. Yannas and L. Gibson, Influence of freezing rate on pore structure in
freeze-dried collagen-gag scaffolds, Biomaterials 25 (2004), 1077-86.
[32] N. N. Batra, Y. J. Li, C. E. Yellowley, L. You, A. M. Malone, C. H. Kim and C. R. Jacobs, Effects of
short-term recovery periods on fluid-induced signaling in osteoblastic cells, J Biomech 38 (2005),
1909-17.
[33] M. R. Kreke, W. R. Huckle and A. S. Goldstein, Fluid flow stimulates expression of osteopontin and
bone sialoprotein by bone marrow stromal cells in a temporally dependent manner, Bone 36 (2005),
1047-55.
[34] Y. J. Li, N. N. Batra, L. You, S. C. Meier, I. A. Coe, C. E. Yellowley and C. R. Jacobs, Oscillatory
fluid flow affects human marrow stromal cell proliferation and differentiation, J Orthop Res 22 (2004),
1283-9.
[35] K. M. Reich and J. A. Frangos, Effect of flow on prostaglandin E2 and inositol trisphosphate levels in
osteoblasts, Am J Physiol 261 (1991), C428-32.
[36] J. You, G. C. Reilly, X. Zhen, C. E. Yellowley, Q. Chen, H. J. Donahue and C. R. Jacobs, Osteopontin
gene regulation by oscillatory fluid flow via intracellular calcium mobilization and activation of
mitogen activated protein kinase in MC3T3-E1 osteoblasts, J Biol Chem 276 (2001), 13365-71.
[37] J. Klein-Nulend, E. H. Burger, C. M. Semeins, L. G. Raisz and C. C. Pilbeam, Pulsating fluid flow
stimulates prostaglandin release and inducible prostaglandin g/h synthase mrna expression in primary
mouse bone cells, J. Bone Min. Res. 12 (1997), 45-51.
[38] J. Vance, S. Galley, D. F. Liu and S. W. Donahue, Mechanical stimulation of mc3t3 osteoblastic cells
in a bone tissue-engineering bioreactor enhances prostaglandin E2 release, Tissue Eng 11 (2005), 1832-
9.
[39] M. J. Jaasma and F. J. O’Brien, Mechanical stimulation of osteoblasts using steady and dynamic fluid
flow, Tissue Eng Part A 14 (2008), 1213-23.
[40] F. Martini, Fundamentals of anatomy and physiology, Prentice Hall, 2002.
230 N. Plunkett and F.J. O’Brien / Bioreactors in Tissue Engineering
[41] R. L. Mauck, B. A. Byers, X. Yuan and R. S. Tuan, Regulation of cartilaginous ecm gene transcription
by chondrocytes and mscs in 3d culture in response to dynamic loading, Biomech Model Mechanobiol
6 (2007), 113-25.
[42] P. X. Ma and R. Langer, Morphology and mechanical function of long-term in vitro engineered
cartilage, J Biomed Mater Res 44 (1999), 217-21.
[43] M. Radisic, L. Yang, J. Boublik, R. J. Cohen, R. Langer, L. E. Freed and G. Vunjak-Novakovic,
Medium perfusion enables engineering of compact and contractile cardiac tissue, Am J Physiol Heart
Circ Physiol 286 (2004), H507-16.
[44] N. Bursac, M. Papadaki, R. J. Cohen, F. J. Schoen, S. R. Eisenberg, R. Carrier, G. Vunjak-Novakovic
and L. E. Freed, Cardiac muscle tissue engineering: Toward an in vitro model for electrophysiological
studies, Am J Physiol 277 (1999), H433-44.
[45] M. Papadaki, N. Bursac, R. Langer, J. Merok, G. Vunjak-Novakovic and L. E. Freed, Tissue
engineering of functional cardiac muscle: Molecular, structural, and electrophysiological studies, Am J
Physiol Heart Circ Physiol 280 (2001), H168-78.
[46] S. P. Hoerstrup, R. Sodian, S. Daebritz, J. Wang, E. A. Bacha, D. P. Martin, A. M. Moran, K. J.
Guleserian, J. S. Sperling, S. Kaushal, J. P. Vacanti, F. J. Schoen and J. E. Mayer, Jr., Functional living
trileaflet heart valves grown in vitro, Circulation 102 (2000), III44-9.
[47] R. J. Levy, F. J. Schoen,W. B. Flowers and S. T. Staelin, Initiation of mineralization in bioprosthetic
heart valves: Studies of alkaline phosphatase activity and its inhibition by AlCl3 or FeCl3
preincubations, J Biomed Mater Res 25 (1991), 905-35.
Basic Engineering for Medics and Biologists 231
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-231
1. Introduction
The investigation of the structural mechanics of the entire body system is despite it’s
rather long history in science, compared to more recently developed research field like
genetics, still extremely important and by no means outdated. Improved innovative
treatments of the musculoskeletal system will require specifically designed implants
and implant materials to guarantee optimal function of the injured segment and fast
rehabilitation. As the actual loading situation of the implant-biomaterial compound is
highly sensitive to implant-induced changes in the specific in vivo situation, the
mechanical properties of living biological materials still require intense research
activities, especially due to the large scatter in terms of biological variation.
Structural failure of biomaterials by plastic deformation or fracture is observed if
the material dependent threshold strength is exceeded. This load inducing failure can be
the result of a single load case or a superposition of different load combinations (i.e.
monotonic or static loading). Furthermore, material failure can also be induced by
repeated (cyclic) loading at loads which are significantly below the material or
compound stability characterizing static strength. This failure type is generally the most
common (already repeated load application in the order of 20 cycles can induce
dynamic failure). It should be noted that the characteristic material or compound
parameters are strongly affected and directly related to the material (micro-) structure.
For biological materials this is generally described by biological scatter in terms of the
mechanical morphogenesis. Therefore, any analysis concerning mechanical parameters
or properties has to be interpreted in terms of the specific microstructure of the material.
The focus of this chapter is to give an introduction into the basic principles of load
induced material behavior and failure under static and dynamic conditions with respect
to specific material structure. Further information can be found in standard text books
on biomechanics and solid mechanics [4], [14], [12],[13], [19], [20], [10], [24], [15],
[16].
232 S. Dendorfer et al. / Characterisation and Testing of Biomaterials
Under in vivo conditions, materials and structures are exposed to a variety of different
superimposed forces. Unfortunately, these in vivo loading conditions are rather
complex and cannot be applied reproducibly in laboratory experiments. Therefore, it is
common scientific practice to investigate the load bearing capability of materials or
structures under idealized and standardized experimental conditions. These results are
transferred to real components or segments with individual geometries by specific
equations and computational methods.
How do biomaterials respond to specific types of loading due to external forces?
Generally, all materials initially react on external forces (i.e. mechanical loads) by
deformation followed by damage due to increasing deformation. The strength of
materials is thus defined as the resistance of a material against deformation or fracture
due to its atomic composition and its specific microstructure. The ratio of strength and
acting forces has two aspects: The main function of hard materials (e.g. bones) is load
bearing and thus they have to withstand high loads. For this purpose maximum stability
is essential. On the other hand, soft materials (e.g. muscles) may react on external
forces by high forming work and deformation without any risk of fracture. In this
aspect maximum deformation potential is of pronounced interest.
To compare different materials concerning their mechanical properties (e.g.
stability or ductility) and to exclude geometrical size effects, it is essential to normalise
external forces and deformations: A bolt doubled in cross sectional area can carry
doubled loads – therefore, it is appropriate to describe the material property strength by
impingement forces, defined as stresses: Stress, MPa (N/mm2) = Force, N per unit of
area, mm2; σ = F/A. Stresses are directed and therefore, act as vectors affecting the
planes of the loaded structure (Figure 1a). Axial stresses react normal to the surface
while shear stresses are acting as tangential paired forces (Figure 1b).
Equivalent to the forces appropriately normalized by the corresponding area,
deformations are referred to the initial dimensions, e.g. the elongation Δl to the initial
length l0. The obtained relation for the deformation is non-dimensional: Strain =
Elongation, mm per initial length, mm; ε = Δl/l0.
With respect to the direction of the induced forces the following types of
mechanical loading are differentiated (and Table 1).
Figure 1: Systematic of mechanical loading classified according to the acting load vectors: axial stresses, a);
shear stresses, b); bending and torsion loading, c); hydrostatic pressure, d).
Under small loads, each solid body at first deforms in a purely elastic manner. This
implicates that for subsequent unloading the deformation is completely reversible. In
most technical cases a linear relationship between the acting stress and the
corresponding deformation is observed. This constant of proportionality between stress
and strain is defined as Young´s modulus (Equation 1), for planar shear stress and the
corresponding shear strain as shear modulus (Equation 2) and for hydrostatic pressure
and the resulting reduction in volume as compression modulus (Equation 3). As strain,
shear and volume reduction have no dimension, the corresponding moduli have the
dimensions of a stress and, analogous to Hooke´s law can be considered as the spring
constant of the material.
234 S. Dendorfer et al. / Characterisation and Testing of Biomaterials
Figure 2: Inelastic deformation: rectangular stress application (load signal) and delayed deformation
response ε, a) and time compensation of σ (t), ε (t) in terms of a stress-strain diagram, b).
Δσ
= E (1)
Δε e
Δτ
= G (2)
Δγ e
Δp
= K (3)
− ΔV / V 0
2.2. Viscoelasticity
Elastic materials which are deformed to a certain strain ε = σ/E (comp. Equation 1) by
a stress σ immediately recedes to ε = 0 after unloading. In many cases and also for
biomaterials, deformation under small applied stresses can be regarded as elastic in
S. Dendorfer et al. / Characterisation and Testing of Biomaterials 235
terms of reversibility, but is associated with a phase shift (i.e. retarded reaction)
between loading and unloading. These cases where elastic deformation is retarded to a
measurable extent are defined as viscoelastic deformations (Figure 2a). This delay in
material response ε(t) with respect to loading σ(t) is also acting if the loading function
σ(t) is not rectangular but has a sinusoidal shape with the latter being of pronounced
technical importance (as in- or decreases in load never happen infinitely fast). This
phase shift between loading and deformation can be expressed in time-compensated
stress-strain hysteresis (Figure 2b). For strictly elastic behaviour each stress value
would corresponds to exactly one equivalent strain. In terms of viscoelastic material
behaviour, for each stress, two strain values (one for loading and one for unloading) are
equivalent. The inclination of this mechanical hysteresis corresponds to E and the
width describes the phase shift between stress and strain.
Each deformation induced by external forces acting is in a first step basically elastic.
With further increase in loading, changes in the material response occur. In this case,
two basic phenomena are of pronounced importance: plastic flow and fracture.
a) Above a certain threshold stress σy the material deforms by plastic flow (yield
stress, related strain εy = σy/E). This strain contribution is irreversible and
characteristic for ductile materials, e.g. most biomaterials, most metals and
some polymers (Figure 3a).
b) Exceeding a specific threshold stress σr, failure of the material occurs without
any previous plastic deformation (rupture stress, corresponding strain εr =
σr/E). This brittle behaviour is observed especially for glasses, ceramic
materials and hardened metals (Figure 3a).
To characterise plastic behaviour experimentally the monotonic stress-strain test is
suitable (Figure 1). For the analysis of extremely brittle materials (glasses, ceramic
materials) bending experiments are more appropriate.
Plastic deformation can be generally expressed in terms of a complex function
(Equation 4), which describes the material behaviour as a relation of five parameters:
applied stress σ, deformation ε, time t, microstructure S and temperature T, respectively
(please do not worry about the complexity and the mathematical solution of this
“cactus” equation! Things are easier than expected):
Ψ = Ψ (σ ,ε , t , S , T ) (4)
ε versus time t under constant stress or load is defined as creep experiment (Figure 3b),
and finally, constant deformation ε leads to a functional relation between stress σ and
time t which is generally defined as stress relaxation test (Figure 3c).
Figure 3: Basic types of results describing the mechanical behaviour under monotonic loading: Stress-strain
curves for ductile and brittle materials and corresponding fracture morphology (schematic), a); creep
experiment, b); stress relaxation test with time dependent stress decrease until a constant relaxed stress is
reached (internal stress), c).
2.4. Summary
Viscoelastic Behaviour
The material response to loading or unloading is basically reversible, but deformation
is retarded with respect to the actual load value.
• characterizes time dependent elastic deformation
• results in damping of externally induced oscillations
• strength and modulus depend on the strain rate at which the material is
deformed
Plastic Deformation
• occurs after a material dependent threshold stress is exceeded
• is always combined with the irreversibility of deformation
• no linear relation between stress and deformation, characteristic line for
the plasticity of materials is the stress-strain curve
• characteristics of the stress-strain curve are:
o work hardening and thus more stress is required for further
deformation
o maximum defining the tensile strength
o decrease until fracture
Ductile/Brittle Behaviour
• ductile materials exhibit measurable (visible) plastic deformations
• brittle material behaviour is defined for fracture without (practical) plastic
deformation
• fracture occurs directly in the elastic region
Creep
As often (falsely) assumed, creep does not necessarily implicate deformation for very
long times, but only basically describes plastic deformation under constant stresses.
• time dependent plastic deformation, usually under constant load or stress
3. Experimental Methods
3.1. Stress-Strain-Experiments
Figure 4: Schematic test rig for analysing mechanical properties, a) and standardized specimen geometry for
tensile experiments, b).
In most cases tensile experiments are performed. With respect to the use of
different specimen geometries and dimensions, the ratio of cross section to gauge
length is specified to: l0 = 5.65 √A0 (Figure 4b). (This value is used in traditional
material science. Even so, it may not be possible to create arbitrary specimen
geometries with biomaterial it is of high importance to use standardised sizes.) The
specimen is deformed under a constant extension rate dε/dt while the applied load F
and the corresponding deformation Δl are continuously recorded and plotted in force vs.
displacement curves. Normalizing the data by the initial specimen cross section A0 and
the initial length l0, respectively, gives the geometry independent stress-strain curve
σ = σ(ε), e.g. Figure 3. This curve represents the characteristic deformation of the
specific material. It depends not only on the chemical composition of the material, but
also on microstructure (and processing conditions).
How to interpret stress-strain diagrams ? Starting from the origin, the first section
of the curve corresponds to the linear increase in stress (slope E). The transition from
elastic to plastic material behaviour is not clearly obvious in many cases, but is more or
less continuous. Due to this, it is often advantageous to define an additional material
parameter: The flow stress Rp0.2 (Yield Strength YS) is defined as the stress under
which a plastic deformation ε = 0.2 % is obtained after complete unloading. Strain
hardening is characterized by an increase in stress for further deformation (strain). This
can be interpreted that for additional strain, the internal resistance against this
deformation increases (work hardening, characteristic for the deformation of metals).
Consequently, this hardening reaches a peak value Rm (Ultimate Tensile Strength, UTS).
In the following, the force needed for further strain decreases and signalizes “that the
end is near”. The material loses strength until fracture occurs. Strictly spoken, it is
impossible to load a material with loads exceeding Rm. It should also be mentioned that
in the case of work hardening effects, the elastic regime of the material is also
S. Dendorfer et al. / Characterisation and Testing of Biomaterials 239
prolonged. In a second load cycle the maximum stress/strain of the first load cycle may
be reached.
The area below the stress-strain curve provides another important parameter.
Mathematically, the area corresponds to the integral and is equivalent to the work
required for the material deformation and thus, can be regarded as energy density,
usually measured in J/m2:
εf l
1
W = ∫ σ dε
0
=
A0 ⋅ l0 ∫ F dl
l0
(5)
The practical sense of the integral, W, is the work required for a deformation
process, e.g. in terms of the energy absorption, i.e. compensation of kinetic energy due
to impact loading. Further characteristic parameters which can be obtained from the
stress-strain curve are: uniform deformation, describing homogeneous elongation of a
specimen under simultaneous reduction in cross section until Rm. After Rm necking is
initiated, leading to a local reduction in cross section and all subsequent deformation is
confined at this neck. Finally, fracture occurs at the neck.
Testing of certain biomaterials may not result in as nicely shaped curves. E.g. the
tensile testing of cartilage results in a stress-strain curve which exhibits a smooth
transition from initial deformation to the elastic regime with rather large deformations
[11]. This transition zone is often referred to as the toe region and is mostly affected
due to the un-crimping of collagen fibres and elasticity of elastin.
3.2. Creep
Creep is generally defined as time controlled plastic deformation at constant stress. The
initial deformation rate dε/dt continuously decreases, although, the deformation is not
stopped (it does not approach zero) and the material still deforms (creeps) with slow
deformation rates. For constant stress, deformation ε follows a function of time t and
for describing the material behaviour the stress-strain curve is replaced by a strain vs.
time curve ε(t) (creep curve, Figure 5).
Generally, creep can be characterized by three classic regions: Immediately after
mechanical loading and elastic deformation equivalent to the magnitude of the applied
stress εelsat = σ/E, a transition to plastic deformation is observed (primary creep).
During the second stage a constant, stationary creep rate is acting. In the third stage
creep rates increase again and failure is initiated (Figure 5). Secondary creep is
characterized by a constant microscopic structure of the material. This steady state can
be regarded as dynamic equilibrium: While the material reacts on the applied load by
continuous work hardening (i.e. (dσ/dε)dε > 0) competitive softening effects are acting
(dσ/dε)dε < 0).
Steady state creep is therefore observed, if the increase in deformation resistance is
compensated by softening effects of the material (relaxation). After sufficient
deformation (tertiary creep), an acceleration in creep rate is observed until fracture
occurs. This increase in creep is substantially driven by rate controlling damaging
mechanisms which already take place during secondary creep but become more
pronounced in the tertiary creep regime, indicating that the end is near. Creep fracture
occurs at sufficient high strains which only slightly depend on loading and temperature.
240 S. Dendorfer et al. / Characterisation and Testing of Biomaterials
Figure 5: Typical creep curve for loading under constant stress: Strain vs. time, a) and differentiated form
strain rate vs. strain.
Figure 6: Schematic S-N curve and determination of fatigue life and fatigue strength.
4. In-Vitro Characterisation
stress-strain space) at the beginning of the compression. The linear regime is followed
by a decreasing slope, indicating failure and plastic collapse, and an increase
(densification) at higher strain values. While the magnitude of the maximum
compressive stress varies up to tenfold between the groups, the characteristics of the
deformation behaviour are similar. Even if the initial deformation behaviour appears to
be rather linear, non-linearities can be found by analysing the deformations with a
higher local resolution. For this reason optical surface deformation measurements are
applied. The surface deformations of the specimen reveal these inhomogeneous
components. Localised strain concentrations can be found already at a small percentage
of the maximum stress. These strain concentrations do further localise and increase in
value until failure occurs in the same region. Macroscopic damage appears in form of a
slip line across the whole specimen cross section.
The early appearance of these strain localisations indicates that structural damage
is already induced at rather low stress levels, even at the beginning of the elastic regime.
Figure 7 shows also data for other types of cancellous bone, human vertebral bone
cored in the main physiological axis and human vertebral bone cored perpendicular to
the main axis. A remarkably different behaviour is obvious. An explanation of these
differences can be found by looking at the underlying structure. Firstly, there is a large
difference in density between bovine and human specimens. Furthermore, the bovine
specimens contain a combination of plate and rod like trabeculae (Figure 8), whereas
the human bone is strictly composed of rods (Figure 9). A rather large influence of the
orientation of the main material axis with respect to the applied load vector is also
244 S. Dendorfer et al. / Characterisation and Testing of Biomaterials
observed. Stiffness and strength decrease enormously by changing the load axis from 0
degree (perfectly aligned with the main material axis) to 90 degree (perpendicular).
This difference can be explained with the high anisotropy of the trabecular network.
Furthermore, the influence of non-optimal boundary conditions is shown in Figure 7.
The dotted line shows a distinctive toe region, but in contrast to tests on cartilage where
this behaviour is a result of the underlying microstructure, here it is an experimental
artefact due to early trabecula failure on the boundaries and small misalignments. In all
the other shown curves, the specimens were embedded on both sides in order to
achieve a homogeneous load transfer from the test rig to the specimen.
Figure 7: Stress-strain curves for different types of cancellous bone, a). The dotted line shows a specimen
without embedding at the ends and therefore experimental artifacts in the beginning of the test. Von-Mises
surface strains (analyzed with an optical deformation measurement system) are shown at different stages
during the experiment, b)-d). The angle value refers to the deviation between load vector and main
physiological axis. [7]
Only taking into account all the experimental and specific material-structure
properties results in comparable, reproducible and valuable data.
S. Dendorfer et al. / Characterisation and Testing of Biomaterials 245
Figure 8: Bovine vertebral cancellous bone, Figure 9: Human vertebral cancellous bone, main
main physiological axis in vertical direction. [7] physiological axis in vertical direction. [7]
Nowadays, image based methods are increasingly used to characterise the materials
microstructure. For example, micro-CT based methods set a new standard in the
quantitative evaluation of the structure. These techniques provide high resolution
images of the structures and image processing methods can create a 3D computational
model from the data. An active research field is the derivation of quantitative measures
for both, the description of the structure but also for the linkage between structure and
mechanical properties [25], [5]. Furthermore, the damage and deformation behaviour of
biomaterials can be studied using high resolution scans which results in a much deeper
understanding of the ongoing processes. Imaging in combination with numerical
methods like Finite Element Analysis and additional data from mechanical testing has
proven to be a very promising way forward. [21], [17].
With ongoing research on the microstructures of the materials a basis can be set to
improve the relationships, originally derived for engineering materials to the specific
need and challenges of biomaterials [23]. Recently, novel methods in musculoskeletal
simulations allow insight in the in-vivo forces acting on many biomaterials [6]. This
information can be used to adjust experimental conditions to approach the relevant in-
vivo scenario.
Besides the technological advances, a standardisation of load cases and
experimental conditions will be a challenge for the future. The need for an improved
comparability of data from various sources as well as the increasingly complex man-
made biomaterials has to be drawn to standards, comparable to engineering design
codes.
Acknowledgements
The authors would like to thank the staff at the Laboratory for Materials Science at the
University of Applied Sciences Regensburg. Special thanks are due to R. Mai for
technical and experimental support, to N. Fischer for preparing the images and to H.-P.
Bräu for providing the electron microscopic images.
246 S. Dendorfer et al. / Characterisation and Testing of Biomaterials
References
[1] O. Beuf, D. C. Newitt, L. Mosekilde, and S. Majumdar. Trabecular structure assessment in lumbar
vertebrae specimens using quantitative magnetic resonance imaging and relationship with mechanical
competence. J Bone Miner Res, 16(8):1511–9, 2001.
[2] C. Yanping, M. Duancheng, and D. Raabe. The use of flat punch indentation to determine the
viscoelastic properties in the time and frequency domains of a soft layer bonded to a rigid substrate.
Acta Biomater, 5(1):240–248, Jan 2009.
[3] D. Chappard, P. Guggenbuhl, E. Legrand, M. F. Basle, and M. Audran. Texture analysis of X-ray
radiographs is correlated with bone histomorphometry. J Bone Miner Metab, 23(1):24–9, 2005.
[4] S.C. Cowin. Bone mechanics handbook. CRC Press Inc., Boca Raton, 2001.
[5] S.C. Cowin. The relationship between the elasticity tensor and the fabric tensor. Mechanics of Materials,
4(2):137–147, 1985.
[6] M. Damsgaard, J. Rasmussen, S. Tørholm, E. Surma, and M. deZee. Analysis of musculoskeletal
systems in the anybody modeling system. Simulation Modelling Practice and Theory., 14:1100–1111,
2006.
[7] S. Dendorfer. Cyclic deformation and fatigue behaviour in cancellous bone. PhD thesis, University
Paderborn, 2008.
[8] S. Dendorfer, H. J. Maier, D. Taylor, and J. Hammer. Anisotropy of the fatigue behaviour of cancellous
bone. J Biomech, 41(3):636–641, 2008.
[9] S. Dendorfer, H. J Maier, and J. Hammer. How do anisotropy and age affect fatigue and damage in
cancellous bone? Stud Health Technol Inform, 133:68–74, 2008.
[10] [edited by] Mow, Van C. Basic orthopaedic biomechanics. Lippincott Williams and Wilkins,
Philadelphia, 2nd edition, 1997.
[11] J. Fierlbeck, J. Hammer, C. Englert, and R. L. Reuben. Biomechanical properties of articular cartilage
as a standard for biologically integrated interfaces. Technol Health Care, 14(6):541–547, 2006.
[12] Y. C. Fung. Biomechanics: motion, flow, stress, and growth. Springer-Verlag, New-York, 1990.
[13] Y. C. Fung and Pin Tong. Classical and computational solid mechanics. World Scientific, 2001.
[14] L. J. Gibson and M. F. Ashby. Cellular solids: structure & properties. Cambridge University Press,
Cambridge, 2nd, 1997.
[15] R.W. Hertzberg. Deformation and fracture mechanics of engineering materials. John Wiley and Sons,
Inc., 1995.
[16] R. B. Martin, D. B. Burr, and N. A. Sharkey. Skeletal Tissue Mechanics. Springer Verlag, New York,
1998.
[17] A. Nazarian, J. Muller, D. Zurakowski, R. Mueller, and B. D. Snyder. Densitometric, morphometric
and mechanical distributions in the human proximal femur. J Biomech, Jan 2007.
[18] G. L. Niebur, M. J. Feldstein, and T. M. Keaveny. Biaxial failure behavior of bovine tibial trabecular
bone. J Biomech Eng, 124(6):699–705, 2002.
[19] M. Nordin, Oe. Nihat. Fundamentals of Biomechanics. Springer-Verlag New York, Inc., 1999.
[20] S. Suresh. Fatigue of Materials. University Press, Cambridge, 2 edition, 1998.
[21] P. J. Thurner, P. Wyss, R. Voide, M. Stauber, M. Stampanoni, U. Sennhauser, and R. Mueller. Time-
lapsed investigation of three-dimensional failure and damage accumulation in trabecular bone using
synchrotron light. Bone, 39(2):289–299, Aug 2006.
[22] W. J. Whitehouse. The quantitative morphology of anisotropic trabecular bone. Journal of Microscopy,
101(2):153–168, 1974.
[23] G.M Williams, K. R Gratz, and R. L Sah. Asymmetrical strain distributions and neutral axis location of
cartilage in flexure. J Biomech, Dec 2008.
[24] H. Yamada. Strength of biological material. Williams & Wilkins, Baltimore, 1970.
[25] P. K. Zysset. A review of morphology-elasticity relationships in human trabecular bone: theories and
experiments. J Biomech, 36(10):1469–1485, 2003.
Basic Engineering for Medics and Biologists 247
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-247
Introduction to
Chapter V: Medical Imaging
Peter NIEDERER and T. Clive LEE (eds.)
According to a German proverb, a picture says more than a thousand words. This
statement is certainly true in clinical diagnostics. With the advent of the X-ray
projection technique, somewhat more that 100 years ago, it became for the first time
feasible to obtain insight into the human body without opening it. Since then, various
methods of medical imaging have developed into major tools in clinical diagnosis
which cover an enormously wide field of pathologic situations. The various modalities
are still in partly spectacular development.
Ultrasound, X-ray and magnetic spin resonance are the most often applied
modalities in medical imaging performed on the intact human body. The basic physics
underlying the procedures derived thereof are well known and established, and they are
outlined in the next three sections. The aim is such that the reader is in a position to
understand and appreciate present and novel developments and clinical applications for
which there exists a vast literature.
Mathematicl methods, in particular Fourier analysis, along with elementary
physics are essential prerequisites in the understanding of the methods under
consideration in the following. An attempt has nevertheless been made in agreement
with the goals of this book such that the basics can be understood without a thorough
mathematical background. The relevant mathematics are given in appendices.
There are a number of further imaging modalities with an increasing importance in
medical imaging, viz, Positron Emission Tomography (PET), Single Photon Emission
Computed Tomography (SPECT) and methods based on the application of infrared, to
mention the most important. In part, similar (in particular mathematical) methods are
applied in these technologies. Since this book is limited to the treatment of the basic
elements of the most important imaging methods, these procedures are not included.
A further important area in medical imaging is devoted to image analysis with the
aid of the computer. At least some numerical treatment is necessary to present the
results of measurements made with ultrasound, X-ray, or magnetic spin resonance on a
computer-driven screen in a clinically useful form. However, medical image analysis,
regardless whether it is performed automatically or with human intervention, is a wide
field in its own right and is not approached here.
This page intentionally left blank
Basic Engineering for Medics and Biologists 249
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-249
Introduction
results in 1938 (Figure 1). Doppler methods for the measurement of blood flow
velocities were finally developed by the Japanese physicist S. Satomura in 1957.
reflections occurring at layered structures or speckle formation are therefore among the
major difficulties.
Typical pressure amplitudes associated with diagnostic ultrasound applications in
the MHz range are of the order of a few MPa. Such pressures are sufficiently high to
induce nonlinear scattering processes involving, e.g., mode conversion. This
phenomenon can be used for image enhancement which is usually denoted as 2nd
harmonic imaging.
Bony or air-filled structures are, for practical purposes, non-transparent for
diagnostic ultrasound waves because of almost complete reflection (an exception is the
transcranial Doppler method). Accordingly, one cannot “see” behind bones, the lung
or air bubbles in the intestine. Due to the inhomogeneous composition of typical
biological materials, furthermore, the propagation of ultrasound waves is not directed
along exactly straight lines (such as x-rays). For this reason, amongst others, all
attempts to develop ultrasound CT imaging have not been successful to date.
According to a rule of thumb, diagnostic ultrasound in the MHz range is damped
by about 1 db per MHz and cm penetration depth (Figure 2). In applications such as
cardiology where distances up to some 25 cm should be covered to create a useful
image, frequencies higher than about 8 MHz cannot be used because of signal/noise
limitation. The resolution, in turn, should at least reach around 1 mm for practical
applications such that frequencies below 2 MHz are not useful because the resolution is
essentially given by the wavelength which increases with decreasing frequency. In
applications requiring a short penetration distance only (eye, skin), frequencies up to 20
MHz can be used. Even higher frequencies are applied in ultrasound microscopy,
where GHz waves with a wavelength below 1 μm are utilized. Penetration depth is
however limited to a few tens of μm because of damping.
d d : damping (db)
z·f z : distance (cm)
f : frequency (MHz)
2.5
2.0 muscle
1.5
liver
1.0
brain
0.5 fat
From the point of view of signal generation and analysis, medical ultrasound
technology has much in common with RADAR (Radio Detection and Ranging) used in
general aviation; major differences derive from the fact that RADAR waves are
electromagnetic waves and operate in the GHz range.
1. Ultrasound Transducers
Ultrasound waves are mechanical by nature whereas the preparation of suitable pulses,
as well as the signal analysis and representation of measured results, is entirely based
on electronics. Accordingly, electromechanical transducers are necessary for the
generation of ultrasound waves and for the recording of echoes. Piezoelectricity
provides the most useful tool for this purpose: When certain crystals are exposed to an
electric tension, they deform because of rearrangement of electric charges in the
interior; in turn, upon deformation, charges are produced at the surface. This effect is
denoted as piezoelectricity, and it can be applied because of its reciprocity for emitting
as well as for receiving ultrasound waves.
A piezoelectric ultrasound transducer consists of an arrangement where one or
more small bars or plates made of piezoelectric material are placed on a rigid,
electrically inactive backing material mounted in a hand-held application device. Both
surfaces of the piezo-elements are electrically connected, much as an electric condenser,
to an oscillator for transmission and switched to a receiver for recording echoes. While
the oscillator sets the transducer into vibration such that ultrasound waves (mostly in
the form of short pulses) are emitted, the returning echoes deform the piezo-material
passively whereby electric signals are generated that can be recorded. Typical materials
used for medical ultrasound transducers are based on sintered lead zirconate titanate
(PZT) where the piezoelectric effect is particularly strong.
1
The principle is valid for linear problems only. The nonlinearities mentioned in the introduction,
however, are of importance only in the interaction of the waves with scattering centers and not in the
generation and propagation of the wave field itself.
P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement 253
introduce a scalar velocity potential, < ( x , y , z ,t ) , from which the velocity field
&
v ( x , y , z , t ) to be calculated is obtained as the gradient ( t denotes the time)
& &
v ( x , y , z ,t ) < ( x , y , z ,t ) (1)
. z
r
K P x
vibrating surface
Figure 3. Vibrating disk on a rigid backing support. The frontal surface is assumed to vibrate uniformly with
a constant amplitude while the one in the back is at rest.
w<
v0 e i Z t for K d a
wx (2)
0 outside
a
v0 e i Z t ª1 i k r º
<P ³ «r e » 2 S K dK (3)
2S 0 ¬ ¼
dK
a
After integration the relative intensity (energy K
Density I at point P can be calculated from
the real part of <P squared. One arrives at
ªS º
I v sin 2 « §¨ a 2 x 2 x ·¸» (4)
¬ O © ¹ ¼
Figure 4 exhibits this result qualitatively. It is important to note the typical near-field
and far-field characteristics; from a practical point of view, the useful area of such a
transducer is at a distance of a 2 / O . This aspect will further be discussed below.
Intensity
x
a 2/O - O /4
Figure 4. Near field and far field characteristics of a circular ultrasound transducer. The intensity varies
rapidly close to the transducer.
2a
Figure 5. Main and side lobes of a circular ultrasound transducer. Note the influence of the a/O ratio on the
beam characteristics.
Note: Since linearity is assumed, these results hold for ultrasound propagation in
continuous wave or pulsed mode. As we shall see, most medical ultrasound
applications use pulsed mode techniques however; the similarity with RADAR has
been mentioned before. A typical ultrasound pulse as used in medical applications is
shown in Figure 6. (Schlieren imaging method).
Figure 6. Schlieren image of a 5 MHz ultrasound pulse consisting of a package of 8 sinusoidal waves,
wavelength 0.3 mm.
256 P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement
One of the methods adopted from RADAR is the “phased array” technique which
is in widespread usage in pulsed ultrasound devices. Thereby, the Huygens’ principle is
approximated by an array of small transducers which produce each a quasi-spherical
wave package having the envelope of a short pulse. The timing of each pulse along
with the amplitude and phase of the center frequency are controlled such that, upon
superposition, a combined wave forming a main lobe with a defined spatial orientation
results (Figure 7). The problem how the amplitudes and phases of the individual
transducer elements of a phased array have to be controlled in order to obtain a
predefined far field is addressed in Appendix 3.
wave front
transducer elements
Figure 7. Phased array technique. Each individual transducer element is excited independently with a pulse
whose timing is such that by superposition a beam with a defined direction of propagation is formed. By
dynamic variation with time, e.g., an oscillating beam can be produced which scans an area of interest much
like a windshield wiper. This technique is used for producing real-time imaging of selected cross-sections of
the body. Most phased arrays presently in use are linear, one-dimensional arrangements; two-dimensional
arrays which allow scanning of a volume can also be made, however.
The more transducer elements are used, the better a desired wave form can be
obtained. In turn, transducer elements have to be made smaller when more elements are
implemented which may cause mechanical sensitivity and electrical impedance
matching problems. Furthermore, individual transducer elements cannot be positioned
too close to each other because of mechanical crosstalk. If too far away, however, the
Huygens’ principle my not be properly approximated in that the distance of adjacent
elements has to be smaller than O 2 , otherwise, the sampling theorem is violated (not
worked out here).
In the analysis of propagating ultrasound waves considered so far, we have
assumed a vibrating surface consisting theoretically of an infinite number of point
sources according to Huygens’ principle. Yet, the transducer is a mechanical oscillator
in its own right associated with its particular mechanical characteristics, viz., mass,
elasticity and damping. In the simplest approximation, a transducer can be modeled as
a linear oscillator. Its eigenfrequency is given by the mass and deformation parameters,
while the width of the resonance curve is determined by the damping. Transducers are
usually operated at their basic eigenfrequency and the bandwidth of the response
corresponds to the resonance curve. Without going into mathematical detail, one finds
P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement 257
that the resonance curve becomes higher and narrower as the damping is smaller. Such
a transducer exhibits on the one hand a high sensitivity; on the other, undesired
transient vibration behavior (“ringing”) renders a transducer with insufficient damping
unsuitable. The higher the damping, in turn, the larger the bandwidth becomes and the
better a pulse form is reproduced mechanically, this at the cost of sensitivity. In
conclusion, the fabrication of ultrasound transducers requires a high amount of
experience, precision work and skill and is a true industrial specialty.
The acoustic impedance Uc (density of the material times speed of sound, see
Appendix 1) of transducers is typically a few orders of magnitude higher than the one
of biological tissues (the density of water is 1 g/cm3, of PZT around 10 g/cm3, the
speed of sound in water is 1500 m/sec, in PZT ten times higher). At interfaces between
different materials, waves are reflected and refracted much as in optics. The amount of
reflexion and the angle of refraction thereby depends on the ratio of the impedances of
the adjacent materials (corresponding to the indices of refraction in optics). The
solution of the wave equation (Appendix 1) shows that in case of large impedance
jumps, waves are mostly reflected. Accordingly, ultrasound waves cannot be
propagated through the skin from a transducer without appropriate impedance
matching measures; in particular, if there is air (impedance four orders of magnitude
smaller than that of water) in between. It has to be kept in mind, thereby, that in
medical ultrasound echo techniques each impedance jump occurs twice, first during
emission, second during backscatter.
Impedance matching is achieved by positioning suitable interface materials
between transducer and skin in the form of surface layers on the transducer and
coupling gels. The rule according to which the properties of the ideal matching material
are determined can be found on the basis of a simplified model (Figure 8). We consider
three layers, viz., the transducer, the matching layer (thickness D ) and the medium
into which the sound wave is to be propagated:
Z1 , Z 2 , Z 3 acoustic impedances ( Ui ci , see Appendix 1), Z1 !! Z 3
<1 velocity potential in the transducer, primary wave
<1' velocity potential in the transducer, reflected wave
<2 , <2' same for the matching layer
<3 medium, outgoing wave
Z1 Z2 Z3
<1
<2
<3
<2’
<1’
The matching is ideal, when <1' 0 (no reflexion into the transducer, all of the
energy is transmitted into the medium). Upon solving the set of linear equations
associated with these conditions, one finds that the thickness of the interface layer, D ,
has to be equal to
D O / 4 while the impedance Z 2 Z1 Z 3 . (5)
The first result (D = O/4) can be obtained without calculation: Taking into account the
phase jump of S when the wave is reflected at an interface with higher impedance,
considering positive and negative interference of the forward and backward wave will
lead to the desired result.
The refraction characteristics of acoustic waves can be used to design ultrasound
lenses for focusing purposes (Figure 9). If the speed of sound in the lens material is
higher than in biological tissue, a concave lens will produce focusing (in contrast to
optical lenses). With phased arrays, focusing can be achieved by suitable timing of the
individual pulses.
z
Lens Focal region
Transducer
Phased array
Body surface transducer
An echo image exhibits in essence all impedance jumps which a beam encounters
as it propagates through the tissue of interest. If the propagation speed of the sound
package can assumed to be constant (1500 m/sec such as in water) the conversion of
the echo sequence into an image is straightforward. Damping is compensated by
exponential amplification.
As mentioned earlier, backscatter may be noncoherent (advantageous for imaging)
or coherent (image deterioration due to mirroring and speckle formation which is due
to interferences). Nonlinear effects during scattering cause the formation of higher
harmonics of the basic ultrasound frequency. The evaluation of the 2nd harmonic can be
used to reduce speckle noise (“2nd harmonic imaging”).
All body parts which are not covered by bone or air volumes where the large
impedance jump usually prevents sufficient wave transmission can be imaged. Primary
applications are in cardiology (heart) and obstetrics (foetus). Physiological fluids
containing minute air bubbles (μm diameter) are sometimes injected systemically to
serve as contrast media to enhance image contrast.
We consider again a pulsed ultrasound method where pulses are emitted with a pulse
repetition frequency fPR. Pulse characteristics (frequency, length, amplitude) are largely
the same as those implemented in echo (imaging) applications.
If ultrasound waves are scattered by moving targets, Doppler shifts are observed
which depend on the speed of the target and the direction of incidence and observation.
Accordingly, this effect can be used to measure blood flow velocities since clouds of
red blood cells act as scattering objects. (The diameter of red blood cells, 8 μm, is very
small with respect to the ultrasound wavelength of around 0.5 mm. The wave
backscattered from one single cell is therefore much too weak to be determined. Blood
cells are inhomogeneously distributed within a blood vessel such that a net effect
remains from the myriads of backscattered waves which would cancel themselves by
interference if the scattering centers would be arranged homogeneously. Nevertheless,
this net effect that makes up the Doppler signal, is usually small.)
Upon transmission of a pulse, the echoes return continuously to the transducer
which, after completion of the transmission cycle, is immediately switched to receiving.
Each echo is processed sequentially whereby it is subdivided into contiguous constant
time steps of duration G ' t , numbered as m 1, 2 ,..., mmax . Accordingly, each
individual step corresponds to a fixed distance from the transducer (Figure 11).
We consider a pulse, ni (i = 1, …) from the sequence of pulses regularly emitted
by the transducer with the pulse repetition frequency fPR, which is emitted at time
i
ti ti 1 . ti 1 thereby denotes the instant in time when the measurement
f PR
begins. The echo associated with this pulse is processed in consecutive channels m
according to the sampling function (Figure 12)
ª n G n Gº
:m
n t 1 for t « ti 1 m G , ti 1 m G »
¬ f PR 2 f PR 2¼ (6)
0 otherwise
P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement 261
'x v 1
'W 2 cos M 2 cos M (7)
c f PR c
v 1
'D 2 Z0 cos M (8)
f PR c
between consecutive echoes arriving at the transducer (Z0 denotes the frequency of the
ultrasound, c the speed of sound).
262 P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement
Time
n = 1 2 3 4 ...
G
m·G + G/2
m·G - G/2
Time
t1 t2 t3 t4
1/fPR
Figure 12. Echo processing sequence and sampling function. G denotes the duration of each channel, m
A1 cos (Z0 t) echo from stationary objects, e.g., the vessel wall within
the channel volume m,
A2 cos (Z0 t + n·'D) Doppler-shifted signals,
ª § n v ·º
S ' mk ¦ :m
n t « A1 A2 cos ¨¨ 2
k Z0 cos M¸¸» (10)
n «¬ © f PR c ¹»¼
n
The step function can be Time
f PR t
5. Further Developments
It has been mentioned previously that nonlinear effects associated with the propagation
and scattering of ultrasound involve „mode conversion“, i.e. the generation of higher
harmonics. This effect can be used mainly to reduce the influence of speckle noise (2nd
harmonic imaging or tissue harmonic imaging).
Intravasal ultrasound imaging using circular transducers emitting beams in a radial
direction have been developed. This method allows us to document, e.g., pathologic
vessel wall conditions such as intimal thickening, fatty streaks or stenoses.
Catheterization is of course necessary. Full 3D noninvasive flow imaging is presently
only possible with the aid of expensive and partially non-real time MR methods.
Two dimensional phased arrays allow the real-time imaging in volumes as, with
the aid of such transducers, a volume can be scanned without moving the transducer.
Figure 16 shows an application in cardiology.
Heart chamber
(schematically) Single beams
Figure 13. Principle of duplex scanning (schematically). The area of interest for a Doppler flow
measurement (heart chamber) is covered only by three beams during each imaging cycle. Accordingly, the
Doppler signal has to be obtained from two phase measurements only unless a more sophisticated scanning
scheme is applied.
P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement 265
Figure 14. Imaging of the heart with overlaid Doppler information in color
Figure 15. Imaging of a carotic artery bifurcation with stenosis. A recirculation zone behind the stenosis
manifests itself.
266 P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement
Figure 16. Three-dimensional real-time imaging of the heart valves by utilizing a two dimensional phased
array transducer
Further Reading
[1] P.N.T. Wells, Ultrasound Imaging, Phys Med Biol 51 (2006), R83-R98
[2] M.F. Hamilton, D.T. Blackstock, Nonlinear Acoustics, Academic Press, New York, 1998
[3] F.A. Duck, Nonlinear Acoustics in Diagnostic Ultrasound, Ultrasound Med Biol, 28 (2002), 1-18
[4] A. Kurjak, S. Kupesic, Clinical Application of 3D Sonography, Parthenon Publ., New York, 2000
& &
wv
x momentum equation U p (2)
wt
dU
p thereby denotes the (hydrostatic) pressure and U p the compressibility of the
dp
fluid. The second equation represents a linear approximation of the Navier-Stokes
equation for an ideal (frictionless) fluid.
&
Upon introducing a scalar velocity potential < r ,t (irrotational flow field, i.e.,
& & & &
no eddies and no turbulence) such that the velocity field v r ,t \ r ,t one
w<
finds p U , furthermore the wave equation
wt
w2 < 1
'< Up with the wave speed c2 (3)
2 Up
wt
&
In this derivation, a term containing p has been neglected which is in agreement
with the acoustic approximation.
We now introduce a Cartesian coordinate system such that the velocity field
& & & &
v r ,t can be represented as v
v x , v y , v z (the independent variables r and t
are thereby omitted for brevity). Forward and backward running plane waves
propagating in direction z are solutions of equation (3)
vx vy 0
vz v0 e i k z r Z t (4)
v0 i k z r Z t
< e
ik
with the frequency Z and the wave number k = 2S/O (wavelength O).
In analogy with electric systems, the acoustic impedance of a medium is defined as
the ratio between the pressure (analogous to electric tension) and the flow (analogous
to electric current)
p
I Uc (5)
v
At locations where the impedance changes, scattering, reflexion and refraction set
in as can be found from general solutions of the above equations. If the impedance
change occurs over a well-defined surface (e.g., a blood vessel wall) mirror effects and
coherent scattering are also observed. The latter leads to interference phenomena which
are associated with speckle formation. Further effects (not included in the foregoing
equations) are damping and nonlinear contributions.
268 P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement
§ wu ww· &
³ u 'w w 'u dV ³ ¨¨ w w n& u & ¸ dB
w n ¸¹
(1)
V B©
³
1 i k r w \ &
e & dB ³<
> &
@
w 1 r e i k r &
dB 0 (3)
B
r wn B
wn
(The term ei Z t is omitted for brevity and only included at the end again.) In order to
determine the function \ in point P (Figure 1), we proceed as follows: For simplicity
(without loss of generality), we assume that the vibrating surface (transducer) is plane,
bounded, part of the boundary B and located in the y – z plane according to Figure 1.
The virtual space with boundary B is then extended to include the entire right half
space.
P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement 269
z
B P
J
r
y x
transducer
Figure 1. Derivation of Huygens’ principle. The closed surface B includes the vibrating surface (transducer)
and is closed such as to include the entire right half space (the broken line is extended to infinity.
Next, the point P is excluded from the volume by a small sphere (radius J ). The
surface B now includes also the surface of this sphere. Later, we let J o 0. (That this
procedure leads to the desired result, is in essence due to the particular properties of the
Green’s function.) We can assume (since we are seeking a physically realistic solution)
that \ is finite and almost constant (in particular, for J o 0 ) within the small sphere
w< & ªw < º 2
such that the term & d B equals « » J cos M dM d-
wn ¬ w r ¼J
( J 2 cos M dM d- is the surface element on the sphere with polar angles M and - ).
Upon evaluation of equation (3) for the sphere, we obtain, for J o 0,
1 i k J ªw < º 2 ª w § 1 i k r ·º
³ e « » J cos M dM d- < J « ¨ e ¸» 4 S J 0 4 S <P (4)
B
J ¬ w r ¼J ¬w r © r ¹¼ J
Accordingly,
4 S <P ³
1 i k r w\ &
e & dB ³ < &
>
w 1 r e i k r &
dB
@ (5)
B
r w n B
w n
whereby the integral now extends over the outer boundary only.
w< &
& vn& denotes the component of the velocity to be prescribed along the surface
wn
B which is perpendicular to the surface. Since only the surface of the transducer is in
vibratory motion, edge effects and components of the velocity which are directed along
w<
the surface are neglected and < is assumed to go to zero at infinity, & is different
wn
270 P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement
from zero only on the transducer surface. The integral (5) therefore extends over the
transducer surface only,
4 S <P ³
1 i k r w\ &
e & dB ³<
> &
@
w 1 r e i k r &
dB (6)
Tr
r wn Tr
wn
The integral (6) can further be simplified by making the special assumption that
w< &
the gradient & vn& is symmetric with respect to the surface, i.e., both sides of
wn
the transducer (the surface facing the + x direction and the one oriented towards - x ,
Figure 1) vibrate in phase but in opposite directions. The same procedure can now be
applied to the negative half space except for the small sphere which is absent since the
point P of interest is on the right side. We therefore obtain a similar expression (6) for
the left half space, but the left hand side is zero,
0 ³
1 i k r w\ &
e & dB ³<
> &
@
w 1 r e i k r &
dB (7)
Tr
r wn Tr
wn
Again, the integral extends only over the transducer surface. However, comparing
the two sides of the transducer, we observe that
w w & &
& o & and d B o d B
wn wn
w< w<
while we assume that & o & since both sides exhibit the same vibration
wn wn
component. Accordingly, the relative sign in equation (7) is different from the one in
equation (6). Both integrals are therefore equal, and we obtain the usual formulation of
Huygens’ principle,
1 1 i ( k r Zt ) w \ &
<P ³ e & dB (8)
2S B r wn
Transducer
Transducer
3 cm
15 cm 1 cm
Figure 2. Intensity distribution of a circular transducer measured (top) with the aid of a hydrophone and
calculated using Huygens’ principle (bottom).
z
z’
P (x0 , z0 )
r
dK
K
2a x
b
Linear Phased Array (length 2a, width b)
Figure 1. Linear phased array. The elements (width b and height dK) are aligned along the coordinate z.
ªw\ º
« » v0 K e i MK (2)
¬ w x ¼x 0
is to be determined. Thereby, v0(K) denotes the amplitude and exp [iM(K)] the phase
function. Outside z = r a we have \ = 0 , and within z = r a there is only the
derivative with respect to x different from zero. K denotes a coordinate along the array
in direction z.
The field \P in point P (x0 , z0 ) (Figure 1) is therefore given by
a
b ei Z t § i k r
i MK ¨ e
·
< x0 , z0 , t ³ 0
v K e ¸ dK (3)
2S ¨ r ¸
a © ¹
with r
x02 z 0 K 2
Upon expansion of the square root K, z 0 r one arrives at Fresnel’s
approximation,
P.F. Niederer / Ultrasound Imaging and Doppler Flow Velocity Measurement 273
ª § k z2 · º
b exp « i ¨ k x0 0 Z t ¸ »
«¬ ¨© 2 x0 ¸
¹ »¼
< x 0 , z 0 ,t x
2 S x0
§ ª 2º · (4)
¨ i k «z K K » ¸
a ¨ « 0 ¸
2 ¼»
³ v0 K e i MK exp ¨ ¬ ¸ dK
a
¨ x0 ¸
¨ ¸
¨ ¸
© ¹
In addition, in the far-field, we have x >> K , such that a further step can be performed
to arrive finally at the Fraunhofer approximation, viz.,
i k z0K
b e i k x0 Z t
a
< x0 , z0 ,t ³ v0 K e i MK e x0
dK (5)
2 S x0 a
Accordingly, the farfield and the distribution of the amplitude and phase on the
transducer are connected by a Fourier-transform, i.e., the excitation v0 Ke i MK to
be determined can be obtained from a Fourier-transform of the desired field.
In a practical application, the integral in equation (5) is replaced by a sum
according to the size, shape and number of individual transducer elements making up
the phased array. The smaller the elements, the more elements can be accommodated
within a given transducer surface and the better the integral can be approximated, yet,
mechanical and electrical impedance problems along with manufacturing procedures
prevent the implementation of too small elements. In order to avoid crosstalk,
furthermore, the individual elements have to be separated by a certain distance which
has to be determined empirically. It can be shown (not worked out here, but easily
understandable qualitatively) that the distance between adjacent transducer elements
must not be larger than half the wavelength of the desired ultrasound wave (Nyquist
limit). The larger the transducer elements (however observing the Nyquist limit), the
less elements can be accommodated; it turns out that the undesired side lobe activity
increases at the same time. Manufacturing ultrasound transducers is a great specialty,
scientific considerations and practical manufacturing aspects are both important.
274 Basic Engineering for Medics and Biologists
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-274
Abstract. X-ray projection imaging, introduced in the early 20th century, was for a
long time the only and still is a major routine diagnostic procedure that can be
performed on the intact human body. With the advent of Computed Tomography
in 1973, for the first time, a true 3D imaging of anatomical structures became
feasible. Besides, digital recording and processing allowed further developments
such as subtraction angiography in real-time or direct x-ray imaging without wet
photographic methods.
Introduction
In 1896 W.C. Röntgen announced the discovery of the potential of an at that time yet
unknown type of radiation to penetrate materials and to cause opacification of
photographic plates (after development). It turned out that various biological tissues
attenuated this radiation differently such that the generation of projections of the
human body onto a photographic plate was possible (Figure 1). As the physical nature
of the radiation was unknown, it was said to consist of „X-rays“ – a designation which
is still in use today.
A further milestone was the introduction of computed x-ray tomography (CT) by
G.N. Hounsfield in 1973 which for the first time allowed to produce cross sections of
the human body for diagnostic purposes in vivo and noninvasively apart from the
application of a radiation dose (Figure 2).
The x-rays used for diagnostic purposes in medicine are electromagnetic waves
with a photon energy between 30 - 200 KeV typically. “X-ray” was until a few years
ago the most often applied medical imaging technology. Since then, methods on the
basis of ultrasound (in contrast to x-ray non-ionizing) have become more often used. In
addition, MRI (Magnetic Resonance Imaging) is of rapidly increasing significance in
medical imaging, in particular as with the combination MRI/MRS (MR Imaging and
P.F. Niederer / X-Ray-Based Medical Imaging 275
The radiation dose which is delivered to the patient is always of concern; it should
be as small as possible. An enormous progress has been achieved through the years
such that today a “simple” x-ray image can be produced with the aid of a radiation dose
which can be considered harmless (Figure 3).
energy dose
(mGy)
8 intensifying
universal intesifying screen
screen (CaWO4)
storage screen
(luminescence)
6
image intensifyer
specialized intesifying + TV imaging
screen (CaWO4)
4 rare earth intensifying screen
Ti2Y2O2S Gd2O2S
2 image intensifyer
storage screen
1. Generation of X-Rays
X-rays for diagnostic purposes in clinical medicine are usually generated with the aid
of a x-ray tube (in order to produce the much higher-energy therapeutic radiation for
tumor treatment in the MeV-range, linear accelerators are mostly used). In such a tube
whose interior is highly evacuated (the pressure is typically 10-6 Torr ~ 10-4 Pa),
electrons are „evaporated“ from a wolfram cathode by heating (the melting temperature
of wolfram is 3400° C) and exposed to an anode voltage, V, of some 30’000 to 200’000
Volt for acceleration towards the anode. The electrons thereby obtain a kinetic energy
of 30 to 200 keV (electron charge, e, times potential difference, V). The geometric
arrangement of the cathode along with the spatial characteristics of the accelerating
field are designed such that the electron beam hits the target (anode) in a carefully
outlined distribution (focal spot, see later).
When the accelerated electrons are being absorbed by the target material, x-ray is
generated due to two effects, viz.,
• bremsstrahlung
• characteristic radiation
Besides that, most of the kinetic energy of the electrons is converted into heat (see
below).
P.F. Niederer / X-Ray-Based Medical Imaging 277
Bremsstrahlung is emitted, when fast electrons interact with the strong (positive)
electric field of atomic nuclei. The electrons are deflected in this field and radiate
electromagnetic waves in the form of photons. The electrons thereby loose kinetic
energy, i.e., are braked down (therefore the German expression „bremsstrahlung“). The
spectrum of this radiation is continuous because the effect depends essentially on the
distance at which the electron passes the nucleus and each distance has the same
probability. The spectrum therefore extends from (theoretically) zero up to the maximal
possible photon energy, Emax , which occurs when the electron looses its entire kinetic
energy in one step.
The spectrum as function of energy associated with a thin target is constant,
because most electrons, while traversing the thin layer, are involved in only one
scattering process, thereby loosing part of their energy such that each portion of energy
from zero to the maximum has the same probability. In a thick target, in turn, the
spectrum is linearly decreasing with E → Emax , because except for a few electrons
which are backscattered at the surface layer all electrons always loose their entire
kinetic energy in various single steps of different intensity (there are more small steps
possible than large ones). According to
Emax = e V = h νmax (1)
(h: Planck’s constant, ν: photon frequency, e: electron charge), the minimal wavelength,
λmin , which is associated with the maximal photon energy Emax , is (speed of light
c = λν )
λmin = h c / e V (2)
with h = 6.63⋅10-34 W sec2
c = 3⋅108 m/sec
V = 100 kV
e = 1.6⋅10-19 Clb
one obtains λmin = 12.4⋅10-2 Å = 1.24 nm . Typical spectra are shown in Figure 4.
The decrease at low energies is on the one hand due to self-absorption in the target
material, and on the other to prefiltering (low energy photons are removed purposely
because they do not contribute to image formation but just increase the radiation dose
imparted to the patient, see below and 2. Absorption of X-Rays).
The electrons hitting and penetrating the target not only interact with the nuclei of
the target material, but also with the electrons thereof. Electrons may thereby be ejected
from their orbit, and radiation quanta are emitted during subsequent recombination.
Since the orbits have defined energy levels, the quanta have energies corresponding to
the various orbits (K, L, ..., Figure 4) and the spectrum consists accordingly of lines
(characteristic x-ray).
Most of the kinetic energy is however converted into heat (∼99%) in the target
material (electromagnetic waves with long wavelengths, largely due to interactions
with free and weakly bound electrons of the target material). As mentioned above, the
low-energy part of the spectrum is usually removed with the aid of prefilters (Al or Cu
sheet metal plates) in order to reduce the radiation dose.
The radiation beam is strictly speaking not continuous as it is composed of a large
number of incoherent electromagnetic wave packets (photons; the scattering processes
in the target material occur independently from one another). The number of photons is
278 P.F. Niederer / X-Ray-Based Medical Imaging
a K-lines of
wolfram
an
od
e
vo
l ta
ge
b
photon energy
There is a great variety of x-ray tubes available for different medical applications;
yet, the basic physical effects are always the same. Material problems are prominent
(hot cathode, extreme local heating of the anode). At a typical anode current ( =
electron current) of I = 100 mA and a voltage U = 100 kV the power deposited in
the anode is
E = I⋅ U = 10 kW (3)
A projection image with a good resolution is only obtained if the focal spot (Figure
5) is small (another possibility is to locate the x-ray tube as far away from the object as
possible; this is however limited due to practical reasons). X-rays propagate along
straight paths, there exist no x-ray „lenses“ for practical medical imaging purposes 1.
The electrons are focused on the focal spot by the suitably shaped accelerating field.
Since all the heat is produced in the focal spot, a small spot is difficult to achieve.
Therefore, the anode is usually inclined with respect to the electron beam such that the
focal spot can be extended (Figure 6) and the heat production is distributed accordingly.
In the direction of the projection, the focal area still has a localized aspect (for special
1
New developments in x-ray technology include the usage of diffraction (phase contrast imaging,
much as in optical microscopy); likewise, Bragg reflection can be used for special purposes.
P.F. Niederer / X-Ray-Based Medical Imaging 279
purposes, also tubes with a line focus are used). In addition, in high power tubes the
anode rotates (5000 min-1 typ.) such that the focal spot is not static.
focal spot
focussing and
accelerating field
electron emitter
(heated wolfram wire)
Figure 5. Focusing of electrons, originating from the cathode (-) onto the focal spot of the anode (+)
Figure 6. Schematic view of static (top) and rotating (bottom) anode tub
280 P.F. Niederer / X-Ray-Based Medical Imaging
The heat is disposed of by diffusion through the anode as well as by radiation. The
latter obeys the Stefan-Boltzmann law according to which a black body radiates the
energy per unit area, W, as
W = σ T4 (4)
(σ = Stefan-Boltzmann constant, 5.7⋅10-12 W/cm2K4, T = temperature)
At T = 2500° K and an emission efficiency of 0.7 (graphite; the emission
efficiency of a black body is 1) the radiated power per unit area is W ∼ 200 W/cm2.
Even under normal operating conditions, the anode may be extremely loaded thermally
(Fig.ure 7).
The further away the tube is positioned from the object, the sharper the images
become since the focal spot has always a certain size. In contrast, the recording
medium, e.g., a x-ray film cassette should be located as closely as possible behind the
object.
For certain applications, in particular computed tomography (see later), the high
voltage used in the tube for the acceleration of the electrons has to be extremely stable
in order that the x-ray spectrum is constant in time. For the same reason, the stability of
the electron beam hitting the target is an important aspect.
P.F. Niederer / X-Ray-Based Medical Imaging 281
2. Absorption of X-Rays
In the energy range which is used for diagnostic medical imaging (ca. 30 - 200 KeV)
two absorption effects occur when the photons interact with material, viz.,
• photo effect (Figure 8, a)
• Compton effect (Figure 8, b).
a b
Figure 8. Photo effect ( a ), Compton effect ( b )
cm 2 /g r
cm 2 /g r
10
100
w a te r
1
10
0 .1
w ater
0 .1
0 .0 1
lead 0 .0 1
0 .00 1
p h o to n en erg y p h o to n en erg y
Figure 9. a: Mass absorption coefficient (photoeffect alone) for lead and water.
b: Total mass absorption coefficient μ’ = μ /ρ (the definition of μ is given below, Beer-Lambert’s law) as
function of the photon energy, E . It is composed of the contribution of the photoeffect ( τ / ρ ), of the
Comptoneffect ( σ / ρ ) as well as the one of pair production ( κ / ρ ). The latter effect is not considered here
because for the production of an electron-positron pair an energy of at least 2 m0 c2 = 1.022 MeV is
necessary (m0 = electron mass, 9.11⋅10-31 kg). Since this energy is much higher than what is used in
medical x-ray imaging, this effect is not of importance here (the additional curve, η / ρ , in the graph, is
related to the energy conversion).
In total, the relation holds μ’ = μ /ρ = τ /ρ + σ /ρ + κ /ρ
Since the Compton effect requires on the average a higher amount of photon
energy, at lower energies the photoeffect prevails while at higher energies the
Comptoneffect sets in.
The cross section of an atom with respect to a certain interaction process is defined
as the (theoretical) area σ (in cm2) perpendicular to the incoming radiation which –
much like the target in a shooting range – is available for the initiation of the process in
question. If the photon hits this area, the effect is triggered, if not, no interaction takes
place. In order to obtain the specific absorption capacity (per unit of mass) of a material,
the cross section has to be multiplied by the number of atoms per unit of mass.
Accordingly, the mass absorption coefficient is defined as the total cross section per
gram, i.e.,
μ’ = σ NA / A ( NA Avogadro’s number, A atomic weight) (5)
Since NA indicates the number of atoms per A grams of a substance (6.023⋅1023),
the quotient NA / A corresponds to the number of atoms per gram, i.e., μ’ has the
dimension cm2/g.
With regard to medical imaging, the dependence of the mass absorption coefficient
on the atomic number, Z (= number of protons in the nucleus), is decisive. Because of
this dependence, the various biological tissues in the body exhibit different x-ray
absorption characteristics such that their appearance in the projection image (grey
level) varies according to their chemical composition. In case of the photo effect, σ is
P.F. Niederer / X-Ray-Based Medical Imaging 283
bone
muscle
fat
Figure 10. Dependence of the linear attenuation coefficient on the energy for various tissues
284 P.F. Niederer / X-Ray-Based Medical Imaging
From a technical point of view the resolution provided by an entire x-ray imaging chain
(x-ray tube → image presentation) has to be considered. One has thereby to
distinguish between
a) spatial
b) contrast
c) temporal resolution.
Only linear mappings are considered here.
a) Spatial Resolution: The spatial resolution of an imaging chain is often given in terms
of the normalized intensity amplitude as a function of line pairs/mm (lp/mm), denoted
as Modular Transfer Function (MTF). The practical procedure to measure the MTF for
x-ray systems consists of imaging lead grids of variable spacing (Figure 11). The closer
the lines are together, the smaller is the amplitude of the image and the average grey
level increases until the lines cannot be discerned any more. The maximal amplitude
(arbitrary unit 1) is defined as the maximal intensity step that can be represented.
Typical MTF curves are shown in Figure 12.
Instead of the MTF, the line spread or point spread function (LSF, or PSF),
respectively, are sometimes given. These functions indicate onto what image a line or
point is mapped, in particular, how a point is blurred (Figure 13).
Figure 11. Imaging of a lead grid. With increasing spatial frequency (lp/mm)
the amplitude of the image (grey level) decreases.
P.F. Niederer / X-Ray-Based Medical Imaging 285
1: 25 cm image intensifyer,
CsJ screen
2: 25 cm image intensifyer,
15 cm setting
1
2 3: 17 cm image intensifyer,
MTF in
3 CsJ screen
lp/mm 4
5 4: 17 cm image intensifyer,
ZnS or CdS screen
Figure 12. MTF of various x-ray systems (system definition see later)
imaging
object image
With μ = 0.35 cm-1 (average for biological tissue devoid of bone and air) we
obtain a contrast of exp (-7), i.e., about 1/1000. In the presence of bone and air (lung,
intestine) the contrast is even considerably higher. Neither with photographic film nor
with conventional TV such a contrast range can be recorded (film, as shown below, has
a linear gray scale range of about 2.5 - 3, corresponding to a contrast range of about
1/300 to 1/1000, while conventional TV has less). Depending on whether, e.g., the lung,
other soft tissues or bones are to be imaged, the radiologist has to choose the
parameters anode voltage, anode current, exposure time in order to keep the desired
contrast within the available range.
Digital imaging techniques allow to record images with an extended contrast range
(12 – 16 bit resolution). Typical TV monitors, however, visualize only a limited grey
level range of about 6 bit. Since images are stored in digital form, the visible contrast
range can be selected and altered up and down in real time („windowing“).
c) Temporal Resolution: The temporal resolution is of importance whenever dynamic
processes are to be investigated (phlebogram, ventriculogram, stomach-intestine
passage, etc.). Cassette changers (see later) reach a frequency of up to about 6
images/sec which is sufficient, e.g., for phlebograms. In order to resolve a heart cycle
adequately, up to 50 images/sec are necessary.
2
α γ = tan α
0
-3 -2 -1 0 log E
The γ value of a film is a measure of the maximum change of film density for a
certain change of exposure. This corresponds to the part of the characteristic curve with
the steepest slope, i.e., at the point where the maximum derivative is found ( = tangent
of the inclination angle α). Typical photographic film has a linear range of about
Δ S = 2.5. The “steepness” γ is also used for other imaging modalities.
To eliminate scattered radiation, anti-scatter grids (Figure 15) are used. Such grids
are characterized by the distance and thickness of the lamellae (typ. 25 – 50
lamellae/cm) as well as the grid ratio (height of the lamellae/distance, typ. 5 - 15).
X-ray films are mostly used for single-shot static exposures. If slow processes are
to be examined (e.g., in case of a phlebogram), cassette changers are applied which
allow to take up to about 6 images/sec. The advantage of this procedure consists of the
high image quality which corresponds to the one of x-ray film. If higher frame rates are
required (e.g., for imaging of the heart), other procedures have to be applied, in
particular in order to avoid excessive radiation doses (see later, image intensifier).
source
object
scattered
radiation
film
Photographic film is an analog storage medium and preparation for computer analysis
and archiving including networking necessitates off-line digitization. In addition, film
development requires chemistry (cost and waste). A number of methods have therefore
been developed which allow for a direct digital image acquisition and which are mostly
in use today.
Digital Luminiscence Radiography (DLR), Storage Screen Radiography,
Computer Radiography (CR, Figure 16): On a screen made of BaFBr:Eu2+ (Europium-
doted barium halegonide) a latent image is created by irradiation (excited electron
states) which is converted off-line into visible fluorescence by laser scanning. One of
the advantages of this technology consists of its large dynamic range.
Selenium-based detectors: Prior to the exposure to x-ray, the surface of the
detector is charged electrically. The photons are converted into electrical signals in that
they create charges in the interior of the selenium layer which neutralize the surface
charges. Again, a latent image is formed which is scanned off-line.
Flat screen detectors on the basis of amorphous silicium: On a silicium wafer a
pixel matrix containing the necessary electrical components (transistors) is
implemented which allows for a direct digital conversion of the charges created by the
incident x-rays Because the absorption of x-ray by silicium is only weak, additional
layers containing heavy atoms have to be overlaid on the silicium screen.
Scintillator-fiberoptic-CCD chip arrangement (in the future also the much simpler
and less expensive CMOS technology will be used instead of CCD): For particular
applications, (e.g., μCT) a scintillator screen serves for spatially resolved photon
acquisition which is imaged on a CCD chip by fiber optic connection. The possibility
thereby exists for image enlargement by cutting the fiber optic bundle obliquely.
incident x-ray
exposure
laser beam
stimulated
light luminiscence
excited state
reset read-out
ground state
remaining excitation is
removed by exposure
to intensive light
In a x-ray image intensifyer (Figure 17) the radiation image is converted into an
equivalent electron image in the entrance screen. The conversion occurs in two steps:
First, light flashes are created by x-ray photons in thin CsJ needle crystals (scintillator)
which are about 0.5 - 1 cm long (good absorption of x-ray) and densely arranged
perpendicular to the screen in order to suppress scattered radiation as well as lateral
diffusion of light (veiling glare). These light flashes cause the creation of free electrons
in the photocathode which is connected to the scintillating screen. In addition, on the
entrance side of the scintillator a thin aluminum foil is located which reflects light back
onto the photocathode. The electrons whose lateral diffusion in the photocathode adds
to the veiling glare and has therefore to be suppressed, are accelerated by 25 kV
typically and directed towards the exit screen. Thereby, an electrostatic field acts as an
imaging electron-“optics” such that a bright image on the exit screen appears. The
intensifying effect is due to the kinetic energy which the electrons obtain during
acceleration. The exit screen is recorded digitally; in the past, video or cinéfilm was
used (up to 50 images per sec). Thanks to the good absorption properties of the
entrance screen, the efficiency of present intensifiers is about 40 - 50 %, i.e.,
substantially larger than film/intensifying screen combinations. The efficiency is given
by the conversion factor G in (cd/m2) / (mR/sec) whereby cd (candela) denotes the
light intensity dI/dΩ (light intensity or light flux per solid angle), mR (milli-Röntgen),
the radiation dose (old unit).
incident
radiation
observation
screen
Figure 17. Cross section through image intensifyer showing electron paths and field lines
290 P.F. Niederer / X-Ray-Based Medical Imaging
The lumina of blood vessels, stomach, intestine, ventricles of the heart or of the brain
are only faintly or not visible on a routine x-ray projection image. In order to obtain a
diagnostically useful visualization, the lumen in question has to be filled with an
appropriate contrast agent (e.g., bariumsulfate in case of stomach-intestine, or a
compound containing iodine for blood vessels or the heart which is well tolerated, i.e.,
which causes no immune reactions).
To image a section of an arterial vessel (e.g., if an aneurysm, a stenosis or a
dissection is suspected) or of a ventricle of the heart it is necessary to administer the
contrast agent at the very location under consideration (selective catheterization) in
order to obtain sufficient image contrast. This procedure is associated with a
considerable expense (catheter laboratory) since an arterial vessel has to be opened,
furthermore, there is always the danger of an afterbleeding. In contrast, a venous
infusion of a contrast agent is largely without problems. Although part of the contrast
agent arrives at the location to be imaged on the arterial side also if a venous
administration is made (e.g., through the vena cava), the resulting image contrast is
insufficient because of the dilution of the contrast agent after the lung passage
necessary to reach the arterial side and the subsequent distribution in the entire
vasculature of the body. In addition, not only the location of interest contains therefore
contrast agent, but all adjacent, overlying and underlying vessels including capillaries
and veins such that the image of the vessel section in question is embedded and
covered by other vessels.
With the aid of image subtraction techniques, however, an increase of contrast can
be reached such that at least in the case of not too small arterial vessels a sufficiently
good representation for diagnostic purposes can be obtained also with venous
administration of contrast agent. Since for this technology on-line digital image
treatment is applied, the method is called Digital Subtraction Angiography (DSA).
(Note: Image subtraction can also be made easily with photographic film in that a
negative can be obtained from contact exposure; for medical purposes, this procedure
has been in use since about 1935). With DSA, an image called “mask” is first made (a
digital projection image before the contrast agent is administered) and stored in the
computer. Then, the contrast agent is applied transvenously, mostly through the vena
cava. After about 20 sec, the exact time delay thereby depending on the location to be
imaged, the contrast agent appears in diluted form at the desired section on the arterial
side. Image sequences can now be acquired whereby the mask is subtracted in an on-
line fashion (video, Figure 18). Providing that the patient has not moved (among other,
breathhold is necessary) between the acquisition of the mask and the later images, only
the difference between the images, i.e., the shadow of the contrast agent should be
visible.
Overlay is still present, but above all, noise amplification occurs since images
characterized by small differences associated with noise are subtracted and
subsequently amplified. Since minimal exposure to radiation is desired, the discrete
nature of x-ray may become apparent (“quantum noise”). This can partially be
compensated by using more contrast agent and/or a higher radiation dose, both of
course undesired (on the one hand, the incidence of immune reactions increases with an
increasing amount of contrast agent, on the other, a higher radiation dose is undesired
anyway). In spite of these drawbacks, the advantage of avoiding an arterial
catheterization is by far dominating such that DSA is applied whenever possible.
P.F. Niederer / X-Ray-Based Medical Imaging 291
kidney
Iliac artery
Ilio-femoral
bypass
femoral
artery
Figure 18. DSA image (real-time video subtraction) of a highly pathologic abdominal vessel situation
The „classical“ tomographic procedure in radiology was performed in such a way that
the x-ray source (tube) and the photographic film were shifted during the exposure
continuously in opposite directions. Only one plane is thereby imaged sharply,
structures which are located underneath or above this plane are smeared. The images
have therefore a low contrast, however. Computed tomography (derived from greek
τομειν, to cut), in turn, was introduced in 1973 by G. N. Hounsfield in England (Figure
2). A. M. Cormack in South Africa had in fact already earlier investigated methods
which allowed to reconstruct objects from a series of suitably chosen x-ray projections
(Hounsfield made his later work however independently from Cormack’s results).
Since at the time when Cormack made his investigations there were no reasonably
priced computer-assisted imaging methods available, he could not represent the results
of his calculations as true tomograms, i.e., images of cross sections, but just in the form
of numbers and curves; accordingly, his work had no practical consequences
(nevertheless, both researchers received the Nobel prize in 1979). What both
researchers did not know, however, was the fact that the mathematical problem of
292 P.F. Niederer / X-Ray-Based Medical Imaging
calculating a n-dimensional object from n-1 dimensional projections had been solved
much earlier by the Austrian mathematician Johann Radon (1887 – 1956) already in
1917. This was recognized only after CT had been introduced into clinical routine. Yet,
this was not of importance because Radon’s formal mathematical solution was found to
be unsuitable for practical applications for numerical reasons.
The method of CT consists of the acquisition of a sequence of x-ray projections of
a predetermined cross section of the body under different angles. In scanners of the first
generation, a single projection was taken by translating a collimated x-ray beam
(“pencil beam”) over the chosen cross section. This procedure required a synchronous
movement of the source (x-ray tube) and the detector. Later, the “fan” beam technology
was introduced whereby an entire projection is taken at the same time with the aid of a
detector array (consisting, e.g., of 512 detectors) without translation of the x-ray source.
Subsequent rotation of the x-ray source and detector array around 180° in equally
spaced angle intervals (e.g., 180°/512) further projections are taken (Figure 19). A total
set of (one dimensional) projections then allows to reconstruct the interior of the cross
section, i.e., the two-dimensional distribution of the attenuation coefficients μ (x,y) (x,
y are Cartesian coordinates in the cross section) The image is represented in the form
of a grey level distribution D(x,y). In present scanners, the fan beam – detector array is
rotated and advanced in a spiraling motion around the patient such that entire spatial
body sections can be imaged within a few seconds (Figure 20).
Quelle
x-ray source
Objekt
object
Projektion
projection
Detektor
detector
Figure 19. Principle of first generation CT scanner: A narrow x-ray beam (pencil beam) is generated by
collimators at the x-ray source as well as at the detector (the latter primarily suppresses scattered radiation).
One-dimensional projections are obtained by translating the unit (embedded in a gantry) over the object. By
successive rotation of the gantry after each translation by, e.g., π/512, a complete set of projections is
recorded.
P.F. Niederer / X-Ray-Based Medical Imaging 293
where I0 denotes the intensity of the unattenuated beam (before impinging on the
object) and s[r,ϕ] is the line defined by r and ϕ between source and detector in the
x - y plane. The integration is performed along the line s from -s0 to +s0 (distance
between source and detector).
base (a)
atlas (b)
b carotid
arteries (c)
c vertebral
d arteries (d)
c
d
[
Pϕ (r ) = log I ϕ (r ) I 0 ] = −∫
s0
− s0
μ[s(r , ϕ) ds ] (10)
is defined. The task now consists of the determination of the linear attenuation
coefficients, μ(x,y), from a complete set of projections Pϕ (r).
294 P.F. Niederer / X-Ray-Based Medical Imaging
Detector [ Iϕ(r) ]
(+s0)
y
s
r
x
ϕ
pencil beam
object
μ(x,y)
collimator source [ I0 ]
(-s0)
yields the desired result. The function |u| thereby compensates the 1/r dependence as
well as the fact that the Fourier space associated with the scanned area is not uniformly
sampled (Appendix 2, this follows from the Fourier-slice theorem): High spatial
frequencies are sampled less densely than low frequencies.
Particular problems which have to be given careful attention in x-ray CT machine
design are related to:
Beam hardening: X-ray sources as used in clinical CT are not monochromatic (see
paragraph 1, Generation of X-Ray). Since x-ray absorption is energy-dependent, the
spectrum of the radiation changes continuously when traversing an object. This effect
is nonlinear and object dependent and can therefore not be corrected mathematically.
Careful, object-dependent calibration procedures are necessary for highest precision.
Partial volume effect: For practical computational reasons, the area of interest has to be
subdivided into (usually rectangular) volume elements. Oblique rays cover most such
elements only partially, however. For each direction and ray, the partial volumes
covered by the rays have to be determined and taken into account in the reconstruction
process.
object
Objekt
projections
Projektionen
Figure 22. Schematic representation of four projections (top) of a cylindrical, homogeneous object and
corresponding simple backprojection (bottom). The measured projection values are thereby distributed
homogeneously (“backprojected”) over the projection path. The resulting “reconstruction” is characterized by
star-like artifacts, moreover, the image of the cylindrical cross section extends over the entire scanned area. If
the radius of the cylinder is assumed to approach zero and the number of projections is increased to infinite,
the point spread function (PSF) is theoretically obtained.
scanners are furthermore being applied in the analysis of biological samples (Figure
23). It should be noted in this context that there are also significant non-medical
applications of CT, in particular for material testing purposes.
Figure 23. Micro-CT measurement of cortical bone. The long cleft appearing in the middle has a width of 3
μm. The lacunae (openings where the osteocytes are housed) can furthermore be seen.
Further Reading
[1] Elliott K. Fishman and R. Brooke Jeffrey Jr., Spiral CT: Principles, Techniques and Clinical
Applications (2nd edition), Lippincott-Raven Publishers, Philadelphia, PA (1998)
[2] Marc Kachelriess, Clinical X-Ray Computed Tomography, in: Wolfgang Schlegel, Thomas Bortfeld
and Anca-Ligia Grosu, New Technologies in Radiation Oncology, Springer, Berlin, 2006
[3] Herman GT, Correction for Beam Hardening in Computed Tomography, Phys. Med. Biol. 24 (1979),
81-106
[4] Yu Zou, Emil Y Sidky, Xiaochuan Pan, Partial Volume and Aliasing Artefacts in Helical Cone-Beam
CT, Phys. Med. Biol. 49 (2004), 2365-2375
[5] Thorsten M. Buzug, Computed Tomography: From Photon Statistics to Modern Cone-Beam CT,
Springer, Berlin 2008
P.F. Niederer / X-Ray-Based Medical Imaging 297
by making use of the Dirac function δ(x - X). We assume now that the Line Spread
Function (LSF), L(x’ ), which denotes the mapping of a line at location x = 0 in the
object space, i.e., of δ(x) , onto the image space with coordinate x’ , is known. A line
at location X in the object space with intensity O(x) will then yield an image
O(x ) δ(x − X ) → O( X ) L(x' − X ) (2)
In case of an optical imaging system with lenses, e.g., the LSF will be a diffraction
pattern, for x-ray projection images, the line will be blurred2.
A general object, O(x), can be decomposed into Fourier components (providing
that it fulfils the usual mathematical conditions) according to
∞
O(x ) = ∫−∞ [a(k ) sin(kx ) + b(k ) cos(kx)] dk (3)
is imaged by way of the LSF into the image space as Jk’(x’) in the form
∞
J k ' (x' ) = a(k ) ∫ L (x' − X ) sin(kX ) dX (5)
−∞
Upon performing the transformation x’’ = x’ – X and making use of the addition
formulas for circular function, one arrives at
∞ ∞
J k ' (x' ) = a(k ) ⎡ sin(kx' ) ∫ L(x' ' ) cos(kx' ' ) dx' ' − cos(kx' ) ∫ L(x' ' ) sin(kx' ' ) dx' ' ⎤
⎢⎣ −∞ −∞ ⎥⎦
= a(k ) [sin(kx' ) a1 − cos(kx' ) a 2 ] (6)
with
2
For two dimensional images, the Point Spread Function (PSF) is used instead of the line spread
function, in that a two-dimensional image can be decomposed into points, using the Dirac function in both
dimensions.
298 P.F. Niederer / X-Ray-Based Medical Imaging
1/ 2
2 2
⎧⎪
η(k ) = ⎨ ⎡
∞ ⎤ + ⎡
∞ ⎤ ⎫⎪
⎪⎩ ⎣⎢
∫−∞ cos(kx' ' ) L(x' ' ) dx' ' ⎥⎦ ⎢⎣ ∫−∞ sin(kx' ' ) L(x' ' ) dx' ' ⎥⎦ ⎬⎪ (7)
⎭
and using the addition property of circular functions once more, the result can be
written as
J k ' (x' ) = a(k ) η(k ) sin(kx' − ϕ) (8)
The Fourier component Jk(x) = a(k) sin(kx) is therefore imaged onto the image
space as harmonic function of the same frequency, k , but displaced by the (spatial)
phase ϕ and with the amplitude a(k) η(k) . The relation holds for every k .
The function η(k) is denoted as Modular Transfer Function (MTF), and it is seen
that the relation holds 0 ≤ MTF ≤ 1 . This function describes the dependence of the
amplitude of the original intensity function in the image space (which is always
reduced) as function of the spatial frequency k . Since the relation holds for every k ,
one can conclude that
F (image) = MTF ⋅ F (original) (F : Fourier transformation) (9)
respectively, that the image corresponds to the convolution of the original with the
MTF according to the convolution theorem.
The relation with the LSF (for two-dimensional images with the PSF) is seen from
the definition of the MTF.
(As mentioned above, the integration over s is extended to infinite for formal reasons,
as outside the source - detector area we can assume μ = 0.) Upon transformation of the
variables r = x cos ϕ + y sin ϕ , s = - x sin ϕ + y cos ϕ one obtains
P.F. Niederer / X-Ray-Based Medical Imaging 299
∞ ∞
− P * (u ) = ∫ ∫ μ(s[r , ϕ]) e
−iur
dr ds
−∞ −∞
∞ ∞
−iu [x cos ϕ + y sin ϕ]
= ∫ ∫ μ (x , y ) e dx dy
−∞ −∞
(2)
∞ ∞
−ix [u cos ϕ] −iy [u sin ϕ ]
= ∫ ∫ μ(x , y ) e e dx dy
−∞ −∞
∞ ∞
∫ ∫ μ(x , y ) e
−ix p
= e −iy q dx dy p = u cos ϕ , q = u sin ϕ
−∞ −∞
This implies that the Fourier transform of the projection Pϕ(r) corresponds to the
two-dimensional Fourier transform of μ(x,y) on a straight line which runs (in the two-
dimensional Fourier space (p, q) under the angle ϕ through the origin (Fourier - slice -
theorem). The Fourier space is therefore sampled inhomogeneously in a star-like
fashion in case of translation-rotation scanning (Figure 24). In particular, the high
frequencies (image sharpness ! ) are covered less densely than the low frequencies. In
order to recover the Fourier transform P* (p,q) of μ(x,y) with uniform density in the
entire Fourier space of interest from the Fourier-transformed projections P*ϕ(u),
interpolations are necessary.
From the relation (1) can be seen that the reconstruction problem, i.e., the
calculation of μ(x,y) can be solved by a reverse Fourier-transformation (after having
determined the Fourier-transformed projections and interpolated P*(p,q) ). A two
dimensional reverse Fouriertransform has however to be executed numerically for this
purpose which is an unsuitable procedure. This can be prevented in the following
fashion.
From the formula for the reverse Fourier-transformation
∞ ∞
1
− μ(x , y ) = ∫ ∫ P* ( p, q) e
ixp
e iyq (3)
2
2π −∞ −∞
since the same projection is recorded after a rotation of 180°), this expression can be
written as
π ∞
1 iu [ r ]
− μ(x , y ) = ∫ ∫ P * ϕ (u ) e u du dϕ (7)
2π 2 0 −∞
300 P.F. Niederer / X-Ray-Based Medical Imaging
and
π π
− μ(x , y ) = ∫ P' ϕ(r ) dϕ = ∫ P' ϕ( x cos ϕ + y sin ϕ) dϕ (9)
0 0
u
ϕ
p
On the basis of (2) and (3) the question relating to the PSF can finally be answered.
According to chapter 3.1. the convolution theorem reads ( “ * ” denotes a convolution)
Here, “Image” denotes the reconstruction which is provided by the PSF and
“Object” the true distribution of the linear absorption coefficients μ(x,y). The (two-
dimensional) Fourier-transform of μ(x,y) is known, if measurements are made under
all angles and the entire two-dimensional Fourier-space (p,q) is interpolated (it is
however inhomogeneously covered by the one-dimensional Fourier-transforms
P*ϕ(u) ). Eq. (2) indicates that the factor |u| compensates the artifacts. The inverse
transform of |u| would therefore correspond to the corrected PSF. Upon performing
P.F. Niederer / X-Ray-Based Medical Imaging 301
an inverse transform of the function |u| exp (-ε |u| ) for ε → 0 ( |u| itself cannot be
transformed !), a dependence ∼1/r2 is found. This can be interpreted such that the 1/r
– dependence of the PSF (in simple backprojection) is additionally compensated by a
further 1/r factor which is due to the inhomogeneous coverage of the Fourier space
which also is characterized by a 1/r dependence.
If therefore for the backprojection (3) not the original projections Pϕ(r), but the
corrected “projections” P’ϕ(r) are used, the desired results are obtained. In the object
space, the procedure involves a convolution such that the method is denoted as
convolution-backprojection method. It is the method of choice in computed
tomography technology. If further effects which cause image deterioration are taken
into account by adaptation of the kernel function |u| still further improvements can be
achieved.
302 Basic Engineering for Medics and Biologists
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-302
Introduction
is also of interest. In comparison with other clinical imaging methods (CT, ultrasound),
MRI/MRS has a number of distinct advantages and some disadvantages.
Compared to CT, MRI yields a better contrast of soft tissues, allows the imaging of
cross sections which exhibit an arbitrary spatial orientation and there is no ionizing
radiation - the strong constant and time-dependent magnetic fields have not been found
to cause negative effects in humans. Some disadvantages, however, derive from the
high cost of the technology and installation, the access surveillance (no ferromagnetic
objects must come close to the magnet, patients with pacemakers, surgical clips and
some types of implant cannot be treated or, at least, extreme care has to be exercised).
Claustrophobia is sometimes a problem. For the investigation of the skeleton (bone) CT
remains the imaging method of choice, since mineralized, hard tissue cannot be imaged
directly with MRI. Likewise, in emergency units where rapid examinations are required,
CT scanners are advantageous because of speed, accessibility as well as the fact that
hemorrhages show a good CT contrast due to the iron content of hemoglobin.
Figure 1. First in vivo human MR image (cross section, thorax) made by Damadian et al. 1977
Ultrasound is even more rapid and real-time imaging can be performed. In addition,
ultrasound equipment is considerably less expensive and smaller that MRI or CT
installations. Yet, MR image quality is largely superior compared to ultrasound,
moreover, ultrasound can only be applied through acoustic windows and bony
structures and air-filled cavities are almost intransparent to ultrasound.
Of great future interest is real-time MR imaging which is under development and
may become feasible within the next few years. A further interesting development is
associated with “open” systems which allow direct patient access and provide the
possibility for MRI-guided surgery.
304 P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics
MRI and MRS have established themselves as significant research tools also in a
wide range of non-clinical applications relating to human behavior in general and
psychology in that brain activity can be monitored under various real and artificial
exposure situations. Due to its noninvasive character, MR technology is furthermore
widely used in animal research.
At this time, MR technology in the life sciences is by far not exhausted and new
applications emerge constantly. Accordingly, the literature on MRI/MRS is vast; this
introductory text is limited to basic physical aspects which are presented in a partly
simplified form. Quantum mechanics and thermodynamics are only used as far as
necessary.
The possible values of |m| are furthermore limited; the relation holds
m = I , I - 1 , I - 2 , ... , - I (2)
I represents the spin quantum number. It is specific for each nucleus and can be integer
or half-integer. A nucleus can therefore assume 2I + 1 stationary states with respect to
angular momentum. In the following, the nucleus 1H with I = 1/2 will mainly be
considered (this nucleus is abundant in the human body), such that in this case
Jz = ±h 2 (3)
As atomic nuclei are electrically charged, the spin is associated with a magnetic
dipole moment. This amounts to
r r
μ = γJ (4)
P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics 305
γ is denoted as gyromagnetic ratio and is likewise specific for each type of nucleus.
For 1H , γ = 4.26⋅107 (T⋅sec)-1 (the unit T stands for „Tesla“, the unit used for the
r
magnetic B - field).
r
In a magnetic field with flux density B (also denoted as magnetic induction) the
axis of reference is given in a straightforward way by the direction of the vector-field
which is identified with the z – axis. Accordingly,
μ z = γ h m ; or, in case of 1H , we have μz = ± γh 2 . (5)
in the states m = ± 1 2 , if the potential energy of the system in the field-free space is
normalized to zero
E
m = - 1/2
ΔE = γh B
m = + 1/2
In thermal equilibrium, the state associated with lower energy (m = +1/2, spin
parallel and oriented along the same direction as the field) is more populated than the
state with higher energy. The population density thereby obeys the Boltzmann
distribution according to the probability
306 P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics
⎡ −E ⎤ ⎡γ h B m⎤
W = N exp ⎢ ⎥ (general), resp. W m = N exp ⎢ ⎥ (7)
k T
⎣ B ⎦ ⎣ kB T ⎦
with Wm probability for the system being in state m ,
N normalization factor (sum of all probabilities = 1),
kB Boltzmann constant ( 1.38 ⋅ 10-23 W sec / ° ),
T absolute temperature
In the case considered here, the system has only two states (m = ± 1/2). Upon
⎡γ h B m⎤
calculating the Boltzmann factor, exp ⎢ ⎥ , one finds a value around
⎣ kB T ⎦
[ ]
exp 10 −6 ⋅ B m (at T = 300°K) , i.e., the populations of the two states differ little
even in the presence of high fields. The normalization factor, N, can therefore be
approximated by 0.5, the population difference Δn can furthermore be determined by
linearization as
Δn = n +1 2 − n −1 2 ≈ n γ h B [2 k B T ] (8)
n thereby denotes the number of magnetic moments per unit volume. The total nuclear
r
magnetization M results as the sum of all dipole moments per unit volume
r r r r
M = Σ μ i , M = Δn μ z = n γ 2 h 2 B [4 k B T ] = χ 0 (T ) B (9)
d Mx dMy d Mz
= γ B0 M y , = − γ B0 M x , = 0 (12)
dt dt dt
P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics 307
(the values for mi 0 are thereby given by the initial conditions) represents a precession
r
(i.e., a rotation around the z-axis) of M with frequency ω0 = γ B0 . The direction of
r
the vectorial rotation is opposed to B . ω0 is denoted as Larmor frequency. According
to
ΔE = hν = h ω 0 = γ h B0 (14)
dM x
dt
= [
γ M y B0 − M z B y (t ) − ] Mx
T2
dM y My
= γ [M z B x (t ) − M x B0 ] − (15)
dt T2
dM z
dt
= [
γ M x B y (t ) − M y B x (t ) − ] M z − M0
T1
We calculate the solution of these equations for the time after the excitation is
terminated (for simplicity, we assume that at t = 0 both fields Bx(t) and By(t) are
shut down). The equations then are
dM x Mx
= γ M y B0 −
dt T2
dM y My
= − γ M x B0 − (16)
dt T2
dM z M z − M0
= −
dt T1
d 2M x 2 dM x ⎛ 1 ⎞
+ ⋅ + ⎜ ω0 2 + ⎟M = 0 (17)
dt 2 T2 dt ⎜ T2 2 ⎟ x
⎝ ⎠
The solution of this equation is a damped oscillation with frequency ω0 and damping
constant 1 / T2
⎡⎛ 1 ⎞ ⎤
M x (t ) = M x 0 exp ⎢⎜⎜ i ω 0 − ⎟ t⎥ (18)
⎢⎣⎝ T2 ⎟⎠ ⎥⎦
⎛ t ⎞
M z (t ) = (M z 0 − M 0 ) exp ⎜⎜ − ⎟⎟ + M 0 (19)
⎝ T1 ⎠
and shows that Mz goes towards M0 for t → ∞ ( Mz0 is the magnetization in z-
direction at time t = 0 ).
Nuclear spin- “resonance” is reached when the components Bx(t) and By(t) of the
time-dependent excitation field initializing a measurement oscillates harmonically such
P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics 309
r
[ ] r
that the resulting field B1 = B x (t ), B y (t ), 0 rotates with the frequency ω 0 , i.e. with
the Larmor frequency. This can be shown by solving Bloch’s equations as follows.
r
First, we remember that the Larmor frequency ω 0 has a sense of rotation which is
r
opposite to the one of the field direction B0 . In order to avoid unnecessary calculations,
we assume that the exciting field with for the time being arbitrary frequency ωz has a
sense of rotation which likewise is antiparallel to this field (otherwise no resonance is
r r
possible), i.e. ω z = (0, 0, − ω z ) . The associated field B1 reads
r
[ ]
B1 = B x (t ), B y (t ), 0 = B1 (cos ω x t , − sin ω x t , 0) (20)
This corresponds to a field rotating around the z-axis counterclockwise with constant
r
amplitude B1 = B1 which at time t = 0 has x-direction.
r
r r r r
In order to solve Bloch’s equation,
dM
dt
[ ]
= γ M × B0 + B1 (t ) + R (the relaxation
r
terms are included for brevity in the vector R ), a transformation is now made into a
r
system rotating with ω z . This procedure will lead to a quite illustrative solution
r
because the time dependence of the field B1 disappears: In the rotating system, both
axes x’ and y’ (primes denote the new axes in the transformed system) rotate
counterclockwise around the z-axis with frequency - ωz , while the z-axis rotates on
r
itself (z’ = z). Since the amplitude of the field B1 is constant, a time-independent set
of equations results. Upon application of the general formula for a transformation into a
rotating system,
d' d r
= − ω× (21)
dt' dt
Bloch’s equation read
r r
r r r r r ⎡r r ⎤ r
d' M
[ ]r ω
= γ M × B0 + B1 − ω z' × M + R' = γ M × ⎢ B0 + z' + B1 ⎥ + R'
γ
(22)
dt' ⎣ ⎦
r
The transformed relaxation terms, R' , are first not considered in a simplified analysis,
they are therefore not calculated in their transformed form. Note that the term
r r r
B0 + ω z γ has only a component in the z’-direction, while the vector B1 is constant
after the transformation and is perpendicular to z’ and oriented in the x’- direction. The
r r r r
quantity Beff = B0 + ω z' γ + B1 represents a constant, “effective” field in the
rotating system such that the equation can be written as
r
d' M r r r
= γ M × Beff + R' (23)
dt'
310 P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics
z’ Beff
M0 B0 + ω z’ / γ
M α y’
nutation cone
B1
x’
Figure 3. Nutation cone in rotating system
r r
Before the field B1 is activated, the magnetization M has only a z’-component,
r
M0 , which is constant in time. Once B1 is turned on, the equations then valid can
r
easily be solved if the relaxation terms R' are disregarded. The solution describes a
r
nutation or secondary precession (i.e., rotation) of the vector M around the axis
r
defined by Beff with the angular velocity ω1 = γ Beff (Figure 3). The opening angle of
r r
the nutation cone, α, can be changed at fixed B0 and B1 by variation of ωz’. The
surface line M0 remains thereby unchanged, i.e., the cone always includes the z = z’ -
axis. If one now chooses for - ωz’ the Larmor frequency, - ωz’ = γ B0 , resonance is
r r
obtained. The opening angle α becomes 90°, since Beff coincides with B1 (note
r r
ω
that B0 + z' = 0 ). The cone becomes a plane lying in the y’ - z’ – plane; the tip of
γ
r
the magnetization vector M follows a circle. Accordingly, at resonance a virtual
r r
observer in the rotating system “sees” a constant field B1 = Beff in x’- direction
r
because he rotates with the same angular velocity as B1 . He furthermore can observe
the magnetization vector to rotate with constant angular velocity in the y’- z’- plane
(Figure 4).
The angular velocity of nutation (or secondary precession) ω1 at resonance equals
r
γB1, since Beff is equal to B1 . If the field B1 is therefore activated during the time
r
period Δt , the macroscopic magnetization M will nutate or precess around the angle
θ = Δt ω1 = Δt γ B1. Accordingly, by an appropriate choice of Δt and B1 it is
possible to reach any desired flip angle. In practice, often 90°- or 180°- pulses are
r
applied, which cause a re-orientation of M around 90° into the -y’ – axis, respectively
around 180° into the -z’ – axis. The term “pulse” is used in this context because Δt is
P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics 311
typically on the order of a few msec while the Larmor frequency is in the MHz-range.
The measurement fields are therefore high frequency (radiofrequency, RF) pulses.
z’
M0
y’
Beff = B1
M
α x’
α: flip angle
After termination of the excitation pulse the magnetization vector precesses with
constant angular frequency ω0 around the z’-axis. Since this vector is a macroscopic
quantity in non-equilibrium conditions, it emits a radio signal with the same frequency
which is sufficiently strong that it can be measured in a receiving coil. In order to
analyze this signal, the relaxation terms have now to be taken into account.
The spin-spin interaction derives from the fact that each spin represents a magnetic
dipole associated with a minute magnetic field which is superimposed over the external
magnetic field. In typical biological tissues, the individual spins are inhomogeneously
distributed. This causes small local inhomogeneities of the total field implying that the
elementary spins have slightly different Larmor frequencies. The spins which are
contained in a macroscopic volume element execute initially (i.e., immediately after the
excitation) a synchronous precession. Yet, with increasing time, the spins get gradually
out of phase because of the slightly different Larmor frequencies such that the
component of the magnetization vector which is perpendicular to the z’-axis disappears.
This can be described with sufficient accuracy as a linear process and the associated
relaxation decay is of the first order with a characteristic time constant T2. This effect
is in practice significantly amplified because the external ‘homogeneous’ and constant
magnetic field exhibits small local inhomogeneities itself for technical reasons in spite
of correction measures (shim coils), thereby in addition causing locally varying Larmor
frequencies. Accordingly, a relaxation time constant, T2*, is found which is
significantly smaller than T2 and which in particular is machine-dependent. For
biological tissues and field strengths around 1-3 T, T2* is typically on the order of a
few tens of msec (see later).
The spin-lattice relaxation, in turn, is due to interactions (collisions) with
surrounding atoms whereby energy is exchanged and dissipated. These processes cause
the magnetization vector to relax back into its thermodynamic equilibrium orientation
in z’-axis. This can also be described by a first order relaxation process with an
312 P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics
associated time constant T1. The latter is usually much larger than T2* (a few hundred
msec).
Due to the relaxation phenomena and field inhomogeneities, a macroscopic
volume induces a signal in a receiving coil, which oscillates almost harmonically, has a
frequency ω close to the theoretical Larmor frequency ω0 and whose amplitude decays
exponentially with the characteristic time constant T2* . (Since the time constants T2
and T1 are significantly larger than T2* they manifest themselves only marginally.)
This signal is denoted as Free Induction Decay (FID) and has the form, as shown above
(only real part),
T2*
U = U 0 e −t cos ωt (24)
U" F
~ exp (-t/T2*)
t
ω0 - ω p
The real part of the Fourier-transform (independent variable p) of this signal reads
T2*
RT [F (U ' ' )] ∝ (26)
(
1 + T2* 2 ω p − ω − p )2
and approaches under optimal conditions (signal-to-noise ratio) a line (Figure 5, right).
The determination of such lines in the spectrum (NMR - spectroscopy) is of importance
P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics 313
Once the magnetization has been rotated into the y’-axis by a 90° pulse (Figure 6a), the
spins start immediately to de-phase because of the slightly different precession
velocities, such that the measured signal – as previously shown – decays exponentially
with the time constant T2* . In the rotating system this means that the spins slowly fan
out in both directions from the y’-axis (b). If after a time interval TI a 180° pulse is
applied (Figure 7), all spins are rotated in a plane parallel to the y’-z’-plane around
180° (c).
There are two reasons for the de-phasing of the spins, viz., (1) the spin-spin
interaction which is in essence of a stochastic nature, (2) the inhomogeneities of the
static magnetic field which are constant and systematic (but machine-dependent). By
application of a 180° pulse the second effect can be compensated in that spins that
rotate slower or faster due to field inhomogeneities continue to do so after the 180°
pulse (d). Accordingly, the spins will rephrase after the time period 2TI (e). In the
receiving coil a signal will therefore build up and subsequently decay (f, Figure 7). The
stochastic spin-spin interaction is not affected by this maneuver, such that the
characteristic signal times are T2*. T2 can e.g. be determined from repeating the
procedure (repetition time TR).
314 P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics
90° 180°
U" t
~ exp (-t/T2 )
~ exp (-t/T2*)
TI TI
TR
// t
By application of a 180° pulse the magnetization is turned in -z’ – direction (Figure 8).
Due to the spin-lattice interaction it relaxes back to the original orientation. The
momentary size of the z’- component of the magnetization during the relaxation
process can be determined by triggering FID signals that are obtained from 90° pulses.
T1 is determined by fitting the exponential curve.
180° 90°
t
~ [1 – 2 exp ( -t/T1 )]
M
U"
TI
TR
// t
Figure 8. Inversion-recovery method
P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics 315
T1 (msec) T2 (msec)
Brain: Pons 445 75
Brain: Cerebellum 585 90
Bone marrow 320 80
Due to the gradient fields, the Larmor frequency is dependent on the spatial location.
The gradients can furthermore be combined arbitrarily, such that arbitrary cross
sections through the object can be examined
In the following, the most often applied Fourier-imaging method is considered in a
basic, simplified form. For further simplicity, a cross section perpendicular to the z –
axis is chosen. The imaging procedure consists of three steps (Figure 9).
In a first step, the gradient Gz is activated during the duration TP which is
necessary to induce a 90° flip angle (excitation field B1, frequency ωP ). The stationary
field during this period of time is
316 P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics
r
B0 = ( 0, 0, B0 + G z ⋅ z ) (28)
This implies that only a layer z = z0 = const. is excited. The thickness of the layer,
Δz, depends on the gradient strength and the bandwidth of the applied pulse TP. Outside
this layer the resonance condition is not fulfilled such that associated volume elements
will not contribute to the signals to be recorded later.
After the interval TP, the excitation B1 and gradient Gz fields are shut down
concurrently. The magnetization in the chosen layer Δz which has been flipped around
90° starts to precess in the x - y – plane around the z – axis. Now, in a second step, the
gradient Gy is activated. Due to this gradient, the local Larmor frequencies depend on
the coordinate y. For a given yi , however, the Larmor frequency does not depend on
the coordinate x, i.e., in a slab extending in x - direction at location z0, yi , the
precession frequencies are the same. The size of the slab perpendicular to the x – axis,
Δy Δz, depends on the gradient strength and the bandwidth of the read-out system
active in the third step.
The local precession frequency is
ω ( yi ) = γ ( B0 + Gz z + Gy yi ) = ωP + γ Gy yi (30)
After the time interval Ty the orientation of the local magnetisation vector has an
orientation which exhibits a relative phase δ depending linearly on yi :
δ ( yi ) = [ω ( yi ) - ωP ] Ty = γ Gy yi Ty (31)
The y – coordinate is therefore „packed“ into a phase; accordingly, the gradient
Gy is often denoted as phase-encoding gradient.
Figure 9. Principle of image acquisition, in particular, the phase encoding along the y-axis
P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics 317
After the interval Ty the gradient Gy is turned off and the measurement is started
(third step). In order to achieve a resolution also in the x – direction, the gradient Gx
is activated. The signal recorded in the measurement coil represents then a
superposition of FID-pulses with different frequencies since the frequency of the FID
signal ω(xj) depends on x because of the gradient Gx (Figure 10). The recoded
signal S, originating from the planar area at z0 (without taking into account the
relaxation terms), reads
[( ) ]
∑ ρ(x j , yi , z 0 )e
i ω x j t + δ( yi )
S (z 0 , t ) ∝ with (32)
i, j
ω ( xj ) = γ ( B0 + Gx xj ) = ω0 + γ Gx xj (33)
Gz
t
B1
90o
t
Gy
t
Gx
t
measurement
Messung
Figure 10. Sequence of gradients in the Fourier imaging scheme described here
The resolution Δx is again determined by the strength of the gradient and the
bandwidth of the measurement system. After demodulation (multiplication with cos ω0
and low-pass filtering) the signal reads
[ ]
∑ ρ(x j , yi , z 0 )e
i γ Gx x j t + γ G y yi Ty
S' (z 0 , t ) ∝ (34)
i, j
S ijmn (z 0 ) ∝ ∑ ρ(x j , yi , z0 )e
[
i γ Gx x j Txm + γ G yn yi Ty ] (35)
i, j
The summands represent the components of a discrete Fourier-transform such that the
quantity ρ ( xj , yi , z0 ) , i.e., the image, can be obtained by application of Fourier
methods.
Once the entire Fourier space (denoted as k-space in MR-jargon) is sampled, the
image can be calculated. By an appropriate choice of the parameters Gx , Txm , Gyn , Ty
as well as of their temporal sequence the k-space can be sampled in various fashions
which are adapted to the analysis to be performed.
Figure 11. Phosphor spectrum. α, β, γ correspond to the three different positions of the phosphor atom in the
ATP molecule. Due to the different chemical surrounding, Larmor frequencies are different (in ppm, parts
per million deviation from frequency). PCr denotes phosphocreatinine, where the phosphor atom is again in
different surroundings. The frequency of this strong line is chosen as reference (zero deviation).
By the application of pulse sequences which cause local saturation (no signal is
thus obtained), patterns can be overlaid on biological structures (tagging, Figure 12).
Muscular deformations can be visualized in this fashion.
Small-bore, high-field systems (up to 19 T) are used in small animal research
(mice, rats). In combination with MR spectroscopy, fluorescence diffusion optical
tomography (FDOT) or bioluminescence recording (both latter technologies are based
on infrared diffusion) physiological and pathological processes are studies in animal
models.
Figure 12. Tagging: The pattern induced by an appropriate pulse sequence allows the monitoring of tissue
deformations as function of time, here of the heart (left: short axis section; right: long axis section).
320 P.F. Niederer / Basic Elements of Nuclear Magnetic Resonance for Use in Medical Diagnostics
RV LV
Figure 13. Fiber architecture of the human ventricles as determined by MR-DTI (ex vivo). RV: right
ventricle, LV: left ventricle.
Further Reading
[1] C.N. Chen, D.H. Hoult, Biomedical Magnetic Resonance Technology. Medical Sciences, Taylor &
Francis, 1989. ISBN 978-0852741184.
[2] A. Oppelt, Imaging Systems for Medical Diagnostics: Fundamentals, Technical Solutions and
Applications for Systems Applying Ionizing Radiation, Nuclear Magnetic Resonance and Ultrasound,
Wiley-VCH, 2006, ISBN 978-3895782268.
[3] M. Blaimer, F. Breuer, M. Mueller, R.M. Heidemann, M.A. Griswold, P.M. Jakob, SMASH, SENSE,
PILS, GRAPPA: How to Choose the Optimal Method, Topics in Magnetic Resonance Imaging 15
(2004), 223–236. http://cfmriweb.ucsd.edu/ttliu/be280a_05/blaimer05.pdf.
[4] A.G. Filler, The history, development, and impact of computed imaging in neurological diagnosis and
neurosurgery: CT, MRI, DTI, Nature Precedings. doi:10.1038/npre.2009.3267.2.
Basic Engineering for Medics and Biologists 321
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-321
Introduction to
Chapter VI: Rehabilitation
Engineering
Tadej BAJD and T. Clive LEE (eds.)
Introduction
application of general equations of motion in the inverse dynamics model of the human
body (Koopman paper, Chap1). These models constitute a part of the software of
modern movement tracking systems. The ground reaction forces and the net joint
torques are termed kinetics.
The patterns of muscular activity are termed electromyographs and can be
obtained from surface electrodes applied over the skin of the muscle of interest.
Kinematics, kinetics and electromyography represent important variables that help the
clinicians to understand a particular gait pattern and to discriminate primary gait
abnormalities, from the compensatory changes that are the consequence of the former.
Based on an understanding of a pathological gait, a treatment procedure or application
of suitable orthotics and prosthetics can be devised. However, understanding complex
gait patterns is a challenging task for clinicians, as walking is largely defined by
biomechanical laws. While understanding normal gait patterns can be learned through
descriptions of events and observations during walking, the comprehension of
individual pathological cases is challenging because they can be very diverse. In order
to correctly interpret measured gait patterns, a basic knowledge of the biomechanics of
walking is required.
The aim of this paper is to establish the biomechanical principles that underlie
human walking. This is done by using a range of simplified biomechanical models of
bipedal walking to explain the laws of movement and associated energetic
requirements. Based on these simplified models, the measurements of normal walking
are described. Selected pathological cases are used to illustrate the changes that occur
in abnormal walking patterns. Finally, the basic design principles used when applying
orthotics and prosthetics to enhance or restore impaired or missing function in walking
are described for these case studies.
Walking is a cyclical activity that is composed of several succeeding cycles. The gait
cycle is most frequently defined as a period from heel contact of one leg until the next
heel contact of the same leg. Figure 1 illustrates the sequence of silhouettes of a
walking person that constitute a gait cycle. The gait cycle consists of two distinctive
phases: the stance phase in which the foot of the leg is placed on the ground and the
swing phase in which the leg is in the air and advances to the position of the next heel
contact. Each stance phase has a period of time at the beginning and at the end, termed
double stance, in which both legs are on the ground. In normal walking, the stance
phase extends over 60% of the gait cycle, while the swing phase occupies the
remaining 40%. The duration of each double stance is approximately 10%. The stance
phase is further sub-divided into: initial contact, loading response, mid-stance, push-off
and pre-swing. The swing phase is sub-divided into: initial swing, mid-swing and
terminal swing.
Step length is defined as the distance between the same points on each foot during
double stance. Stride length is the distance between two successive heel strikes by the
same foot and therefore represents the distance travelled within the gait cycle. Walking
speed is the average speed over multiple strides, while cadence is the number of steps
per time unit.
Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics 325
Double stance
Figure 1. The gait cycle is divided into a stance phase, when the leg is in contact with the ground, and a
swing phase, when the leg is in the air. Within the stance phase there are periods of double stance when both
legs are on the ground.
system. When the vertical position is reached, only a minimal horizontal velocity
needed for passing the vertical position is present. Afterwards, when the gravitational
force accelerates the mass particle along the arc, the potential energy level is decreasing
while the kinetic energy level is increasing. The inverted pendulum model from Figure
3 can be considered as the simplest model of bipedal walking during the stance phase.
The model is completed by adding another weightless rod articulated with the first one
through a simple hinge joint at the location of the mass particle which constitutes a hip
joint.
r
Wp = m ⋅ g ⋅ h
r2
m⋅ v
Wp WK =
m 2
h r
v
WK W P + W K = const .
Figure 2. Qualitative model of a simple pendulum. Pendulum movement is illustrated with a sequence of
three positions. When the pendulum is displaced from equilibrium it possesses the largest value of potential
energy WK , velocity of movement equals zero. When passing vertical position the pendulum possesses the
r
largest velocity v WK , which equals potential energy at the initial displaced
and thus also kinetic energy
r
position. (m = mass of pendulum bob; g = acceleration due to gravity, v indicates that velocity, v, is a
vector quantity, as it has both magnitude and direction)
Figure 4 shows such a bipedal model, where one leg is in stance, providing a
support to the mass particle on the top, while the other leg is in swing, in preparation
for the next stance phase of the gait cycle. In the preceding treatment of bipedal
walking we have considered only the gravitational force, which is considered an
external force. Another external force that acts upon the inverted pendulum is a ground
reaction force (GRF), that acts between the tip of the leg and the ground. In the simple
inverted pendulum model, the GRF is entirely passive and acts along the leg
throughout the circular motion. While the model from Figure 4 is conceptually clear, it
misses a very important aspect of bipedal walking - the impact of the swinging leg and
the ground at the moment of initial contact.
Figure 5 shows the situation at the moment of impact. In this passive inverted
pendulum model, the impact time is very short and can be regarded as instantaneous.
r
Just before impact, the mass particle possesses certain velocity v− that is perpendicular
with respect to the stance leg. When the leading leg impacts the ground a GRF L ,
which is directed along the leg, acts on the leg contacting the ground. In the short time
of impact the force impulse, which equals the product of the magnitude of GRF L and
the time duration of the impact, acts in such a way that the direction as well as the
amplitude of the velocity of the mass particle is changed. This change is depicted by a
Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics 327
r
vector composition, which results in a new velocity of the mass particle v+ , which is
directed perpendicularly with respect to the leg that just entered stance in order to be
able to perform circular motion.
WP
WK
r
v
Figure 3. Qualitative model of an inverted pendulum. Movement of the inverted pendulum is illustrated with
a sequence of five positions with indicated velocity vectors. Squares denote a share of potential and kinetic
energy in each of five positions. Note that the sum of both energies remains constant in each case.
Figure 4. The simplest model of bipedal walking presented by two inverted pendulums articulated with a
simple rotational joint.
Due to the geometry of both legs in double support phase, the consequence of the
force impulse of the GRF L at impact always decreases the total mechanical energy of
the system. The change in the mechanical energy before and after the impact is
proportional to the squared change in the velocities before and after the impact. The
lost energy is transformed into a sound that we hear at the time of the impact and the
deformation (which is so small that can not be noticed with a naked eye) of the leg that
impacted the ground. The situation displayed on Figure 5 demonstrate one of the most
important features of the bipedal walking; in every step there is a loss of mechanical
energy due to the impact of the swinging leg with the ground. We can easily imagine
that due to this loss of energy the simple mechanism would not be able to make another
328 Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics
step, because the kinetic energy of the system after the impact would be insufficient to
reach the vertical position. It is clear that lost energy needs to be recovered within each
step; this means that another important feature of bipedal walking is that in every step,
work needs to be performed to recover the level of mechanical energy that is necessary
for alternating circular motion of both legs.
r
v+ ΔW−
r
v−
GRF L
Figure 5. Energetics in double stance. A force impulse of GRF L changes the direction of movement of
mass particle m as well as reducing kinetic energy.
the three snapshots are used to show the gradual changes in the direction of the velocity
of the mass particle throughout the double support phase. We can see that, instead of
the rather abrupt change in the trajectory of the mass particle as shown in Figure 4, a
more fluid and continuous trajectory curve is achieved when simultaneous extension of
the trailing leg and retraction of the leading leg are utilized. Also, the changes in the
mechanical energy are not instantaneous but depend on i) the absorption of power by
the leading leg ( PL ), which in every time instant equals the vector dot product of the
r
GRF L and the v and ii) the generation of power by the trailing leg ( PT ), which in
r
every time instant equals to the vector dot product of the GRFT and the v . The time
integral of both powers relate to changes in the mechanical energy of the system.
r r
v+ v+ ΔW−
r
v− r ΔW+
v−
GRFT GRFL
Figure 6. Active push-off is achieved by a force impulse of the trailing leg just before the impact of the
leading leg with the ground. The overall effect of both force impulses is an appropriate change in the
direction of velocity of the mass particle.
To summarize the above derivation of the simple bipedal walking model, the
essential components are the two legs that can change their length, a mass particle
representing the mass of the whole body and the two external forces; the gravitational
force acting directly on the mass particle and the GRF acting on the mass particle
through both legs. Most of the movement in the steady state is passive while energy
absorption is performed by the leading leg following impact and energy generation is
performed by the trailing leg just prior to entering the double support phase.
While this conceptual simple model of bipedal walking captures all the essential
features of bipedal walking, it is only a first step toward understanding human walking,
since our legs are not simple rods that extend and retract; rather, our legs are made of
segments (pelvis, thighs, shanks and feet) that are articulated with joints (hips, knees
and ankles). It is therefore necessary to extend our treatment to human-like
biomechanical model in order to fully comprehend the mechanisms of bipedal walking.
However, a proper and rigorous treatment of the biomechanics of such a model from
the aspects of mathematics and physics is rather challenging and requires detailed
knowledge of several engineering disciplines. Still, we can make certain simplifications
330 Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics
that will allow us a simplified treatment leading to useful results that relate to human
walking.
r
v
GRFT GRFL
= r
PT GRFT ⋅ v
= r
PL GRFL ⋅ v
Figure 7. Realistic presentation of a double stance, where both ground reaction forces act in a way that
gradually changes the direction of velocity of the mass particle. The leading leg is retracting and thus
absorbing power and energy, while the trailing leg is extending and thus generating power and energy.
To summarize the above derivation of the simple bipedal walking model, the
essential components are the two legs that can change their length, a mass particle
representing the mass of the whole body and the two external forces; the gravitational
force acting directly on the mass particle and the GRF acting on the mass particle
through both legs. Most of the movement in the steady state is passive while energy
absorption is performed by the leading leg following impact and energy generation is
performed by the trailing leg just prior to entering the double support phase.
While this conceptual simple model of bipedal walking captures all the essential
features of bipedal walking, it is only a first step toward understanding human walking,
since our legs are not simple rods that extend and retract; rather, our legs are made of
segments (pelvis, thighs, shanks and feet) that are articulated with joints (hips, knees
and ankles). It is therefore necessary to extend our treatment to human-like
biomechanical model in order to fully comprehend the mechanisms of bipedal walking.
However, a proper and rigorous treatment of the biomechanics of such a model from
the aspects of mathematics and physics is rather challenging and requires detailed
knowledge of several engineering disciplines. Still, we can make certain simplifications
that will allow us a simplified treatment leading to useful results that relate to human
walking.
Figure 8 shows a model that resembles human skeletal system in a sequence of
five snapshots that illustrate the conditions throughout the stance phase from initial
contact until push-off. The leg segments are weightless as was the case in our previous
models, while the mass particle is replaced by a large ellipsoid representing the head,
arms and trunk. In the middle of this ellipsoid is a small circle that represents a centre
of mass (COM). Since the legs have no weight, the COM remains in the same relative
position within the ellipsoid. We shall first observe the sequence of humanoids from
Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics 331
Figure 8 in the same way as was presented in previous figures. We need to concentrate
only on the COM, GRF L and a “virtual” leg, which is represented by the dashed line
connecting the point on the ground where GRF L acts on the leg; this point is called the
centre of pressure (COP). We can see that there is considerable resemblance with a
simple inverted pendulum.
r
r r MH r r
MH MH MH MH
r r
MK MK
r r r
MA MA MA
r r
MA MA
GRF L GRF L
GRF L GRF L GRF L
Figure 8. Human-like simple biomechanical model of walking in stance phase. Depending on the external
moment in each joint produced by the GRF, the necessary balancing muscular moments are shown in the
r r r
ankle ( M A ), the knee ( M K ) and the hip ( M H ) in all sub phases of stance: a.) initial contact, b.) loading
response, c.) midstance, d.) push-off, e.) pre-swing.
The GRF always acts in the direction of the COM, while the ‘virtual’ leg rotates
similar to the inverted pendulum in Figure 3. The length of the ’virtual’ leg depends on
the position of the thighs, shanks and feet of both legs and changes throughout the
whole stance phase. In energetically efficient walking, these changes are such that the
oscillations of the COM in the vertical direction are minimal. So far, the forces that we
associated with bipedal walking were the external forces (gravity and GRF). Here, we
need to introduce also the so-called internal forces that are produced by the leg muscles
and result in equivalent net joint torques. The relationship between the external and
internal forces in our model is rather simple. In each of the lower extremity joints, the
net joint torques must equal to the product of the magnitude of GRF L and the
perpendicular distance from the line of action of the GRF L and each individual joint.
For example, at the initial contact, the GRF L acts posterior to the ankle joint, thus
producing the moment around the ankle that tends to plantarflex the foot onto the
ground. Thereby, ankle dorsiflexors must be active to produce a net joint torque that
nearly balances the external ankle moment produced by the GRF L . Similarly, in the
332 Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics
hip the GRF L passes in front of the hip joint thus producing external moment that
tends to flex the hip. Therefore, hip extensors must produce a net joint torque that
nearly balances the external hip moment produced by the GRF L . In the knee joint no
net joint torque is needed from either knee extensors or flexors, as the GRF L passes
through the joint, thereby producing no external moment. In the subsequent phases of
stance, different demands arise for each joint muscles, depending on the requirements
imposed by the GRF L . Here we need to stress that the only external force that is
independent is gravitational force, while the GRF L depends on gravity, mechanism
dynamics and internal forces and moments produced by the muscles and other soft
tissue. Through loading response, the greatest demand is on the knee extensors that act
eccentrically (producing tension while being lengthened), thereby absorbing the energy
due to the impact of the leading leg with the ground. Also, considerable activity is
needed from the gluteal muscles to maintain the trunk erect. In midstance, there are
rather small net joint torques needed, as the ‘inverted’ pendulum passively rotates
around the ankle joint and the GRF L passes in the vicinity of all three joints. During
push-off, a strong concentric activity (producing tension while shortening) of
plantarflexors is needed (to extend the ‘virtual’ leg), thereby generating the power and
energy needed to compensate for the energy loss that will follow during the impact of
the other leg. Here also, activity of hip flexors is needed to counteract the external
GRF L moment. During pre-swing, the activity of hip flexors, knee extensors and
plantarflexors are needed due to the orientation of the GRF L
The double support phase is illustrated in Figure 9. The left side of Figure 9 shows
both ‘virtual’ legs, both GRFs and all the moments in hips, knees and ankles of both
extremities. The right side of Figure 9, similar to Figure 7, shows the evolution of the
COM trajectory that results from shortening of the leading leg (flexion of all three
joints) and the trailing leg extending (extension of all three joints). The relationships
for power absorption ( PL ) in the leading leg and power generation in the trailing leg
( PT ) in the model from Figure 7 (simple rod shortening and lengthening) and in the
model from Figure 9 (‘virtual’ leg shortening and lengthening resulting from changes
in the angular positions of the leg segments) are given in the following equations:
r r r r r r r
PT = GRF T ⋅ v = M HT ⋅ ω HT + M KT ⋅ ω KT + M AT ⋅ ω AT (1)
r r r r r r r
PL = GRF L ⋅ v = M HL ⋅ ω HL + M KL ⋅ ω KL + M AL ⋅ ω AL (2)
The generated power/absorbed power in each joint depends on the net joint torques
produced by the muscles and the angular velocity. If the direction of net joint torque
and angular velocity is the same as is the case in Figure 9 for the trailing leg, the power
and energy is generated in a particular joint, if the two directions are opposite as is the
case in Figure 9 for the leading leg the power and energy is absorbed in a particular
joint. Therefore, while the requirements that the leading leg is absorbing power and
energy during loading while the trailing leg is generating power and energy during
push-off is an inherent, non-negotiable feature of bipedal walking, this can be achieved
in many different ways (with different joints contributing differently depending, for
Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics 333
r
ω HT r r r
M HL ω HL
r M HT
ω KT r r
r ω KL
M KT M KL
r
ω AT r r
r ω AL
M AT M AL
GRF L
GRF T GRF T GRF L
a.) b.)
Figure 9. Human-like simple biomechanical model of walking in double stance phase. a.) Depending on the
external moment at each joint produced by the ground reaction forces ( GRF L , GRF T ), also the
r r
necessary balancing muscular moments and angular velocities are shown in the ankle ( M AT , M AL and
v v r r v v r r v v
ω AT , ω AL ), the knee ( M KT , M KL and ω KT , ω KL ) and the hip ( M HT , M HL and ω HT , ω HL ).
b.) Realistic presentation of double stance where both ground reaction forces act in a way that gradually
changes the direction of velocity of the mass particle.
The swing phase, which completes our theoretical treatment of bipedal walking, is
depicted in Figure 10. For clarity, only the swinging leg is shown. Immediately after
the leg leaves the ground (initial swing) net joint torques are needed in the hip (flexor
moment) to propel the leg forward, while knee extensors need to eccentrically arrest the
movement of the knee in the flexion that resulted from preceding push-off. During
mid-swing, almost no muscular activity is needed as the leg segments are moving
balistically. Terminal swing is characterized with eccentric activity of hip extensors and
knee flexors to decelerate hip flexion and knee extension in order to prepare the leg for
the next stance phase.
r r r r r
M H ωH ωH MH ωH
r r
ωK r ωK
r r MK
MK ωK
r r
MA MA
Figure 11, which is similar to Figure 8, shows five snapshots of stance phase of
walking in a neurologically and orthopaedically intact individual at a velocity of 1 m/s.
Figure 11 shows the displacements of body segments and the magnitude and direction
of GRF along the dashed line pointing into the COM. Figure 12 shows the horizontal
and vertical components of GRF throughout the gait cycle. Consistent with our simple
biomechanical model, we can notice that the first 30% of the stance phase is
characterized by a negative value of horizontal GRF component (braking the
movement of an inverted pendulum) while the remaining 30% is characterized by a
positive value of horizontal GRF (propelling the movement of the inverted pendulum
forward in the direction of progression). The first 20% of stance is characterized by a
first peak value of vertical GRF component that exceeds the gravitational force, which
is associated with the deceleration of the downward movement of COM and its
redirection into an upward movement, as predicted by our simple energetic model of
the double support phase. In the next 20% of stance, the inverted pendulum is almost
passively rotating around the ankle joint, therefore the magnitude of vertical GRF
component is lower than the gravitational force (body weight). In the last 20% of the
stance the vertical GRF component again exceeds the gravitational force which is
associated with strong forward and upward propulsion (leg is extending and generating
positive power).
Figure 11. Representation of pelvis and lower extremities movement together with a GRF in five
consecutive snapshots within the stance phase in normal walking.
Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics 335
Figure 12. Graphs of horizontal component of GRF ( Rx ) and vertical component of GRF ( Rz )
throughout the gait cycle. Stars above the curves denote time instants that correspond to the snapshots from
Fig. 11. Mean values and standard deviations are for normal walking (adapted from Winter, 1991). Both
periods of double stance are also shown.
Figure 13 shows the joint kinematics, kinetics and energetics throughout the whole
gait cycle. Consistent with the simple biomechanical model, we can observe that the
ankle undergoes dorsiflexion (rotation of the inverted pendulum) throughout the
majority of stance, ankle plantarflexion net joint torque slowly increases due to the
requirements imposed by the GRF. During the loading response, the knee performs
controlled movement into flexion and then returns to a neutral position, which is
associated with leg shortening. This shortening is opposed by a strong knee extensor
moment, resulting in considerable power absorption, which agrees with our simple
biomechanical model. The hip undergoes a movement from flexed position at the initial
contact into extension to facilitate the vertical position of the trunk. This is
accompanied by a noticeable hip extensor moment required by the GRF passing the hip
anteriorly. In this period there is a small burst of positive power generated in the hip
that propels the COM forward. At the end of the stance phase, within the second double
support phase the ankle goes rapidly into plantarflexion, which generates considerable
positive power and energy as predicted by our simple biomechanical model. With a
slight delay the power generation in the ankle is accompanied by a considerable power
generation in the hip, which propels the leg into swing phase.
Figure 14 of the recorded EMG activity in the major lower limb muscles, shows
timing and intensity of activity that corresponds well with the magnitude of joint
moments throughout the gait cycle.
The comparison of the simple biomechanical model of bipedal walking and the
measurement of human walking shows close resemblance. The magnitude of joint
moments closely relates to the external moments generated around each joint by the
GRF, while power generation and absorption just before and during the double support
phase corresponds to energetic considerations of the inverted pendulum model.
336 Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics
Figure 13. Graphs of joint goniograms, muscular moments and powers in the ankle, knee and hip joints
throughout the the gait cycle of normal walking. Negative values in goniograms represent extension; negative
values of moments represent flexor muscular moments; negative values of power represent absorption of
energy. Stars above the curves denote time instants that correspond to the snapshots from Fig. 11. Mean
values and standard deviations for normal walking were adapted from Winter (1991). Both periods of double
stance are also shown.
Figure 14. Graphs of electromyograms from the selected muscles of lower extremity throughout the gait
cycle. Mean values and standard deviations are for normal walking were adapted from Winter, 1991. Both
periods of double stance are also shown.
Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics 337
2. Orthotics
Figure 15. Presentation of pelvis and lower extremities movement together with a GRF in five consecutive
snapshots within the stance phase of walking in a child with cerebral palsy.
338 Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics
By providing the above example, we can state the primary function of an orthosis
is to change the biomechanics of the lower extremity. Application of an orthosis can
have the immediate effect of substituting for the lost muscle function (in our case the
weakened knee extensors). Also, an orthosis can bring about improvement indirectly by
providing an opportunity to enhance or acquire motor skills. In our example, this means
that the child whose walking pattern is enhanced could make initial contact on the heel,
followed by the gradual weight transfer over the entire sole of the foot, which is not the
case in his regular walking. Walking on tiptoes is rather unstable, therefore by enabling
a more normal weight transfer the orthosis would also give the opportunity to the child
to practice and improve his balance skills.
When designing an optimal orthosis for a given pathological case, one first needs
to identify the primary cause for the observed changes in the walking patterns in order
to determine what kind of orthosis would be appropriate, i.e. foot orthosis, ankle-foot
orthosis, ankle orthosis etc. When this is determined, the mechanical properties of an
orthosis need to be determined; the form, resting position-angle, the stiffness. Two
major design principles need to be considered when making an orthosis. The first
principle relates to GRF control. Basically, this principle derives directly from our
simple biomechanical model of walking where we have shown that the GRF
determines the required joint moments in the hip, knee and ankle joints. Alternatively,
we can say that the joint moments developed by muscle activity shift the COP and GRF.
Therefore, when designing an orthosis we need to anticipate what kind of changes the
mechanical action of an orthosis will introduce in the kinematics and kinetics of
walking in each particular case.
The second design principle relates to a so called three-point pressure control. The
purpose of this design principle is to influence joint stability and/or mechanical
properties by providing the first point of pressure above the axis of rotation, the second
point of pressure below the axis of rotation and finally to provide the third point of
pressure that acts in the opposite direction at or near the axis of rotation. This is
illustrated in the Figure 16 where a KAFO (knee-ankle-foot orthosis) and AFO (ankle-
foot orthosis) are shown with the indicated three points of pressure in each case.
Special care must be given to the mentioned three points of pressure where the orthotic
forces are transmitted onto the skeletal system. The areas needs to be large enough and
shaped such as to enable even pressure distribution over the skin in contact with an
orthosis. The mechanical action in each of the artificial joints of an orthosis may be
different. For example, a KAFO may be designed in such a way as to lock the
movement in the ankles and the knees thus providing stability during the stance phase
to a person who is paralyzed but with some remaining function in the hips which are
used together with the crutches for propulsion during ambulation. On the other hand,
the action of a knee orthosis in the above example of crouch walking in a CP child
must be such as to provide just enough stiffness to adequately modify the COP and
GRF during walking.
Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics 339
Figure 16. Three-point pressure examples in KAFO (knee-ankle-foot orthosis) and AFO (ankle-foot
orthosis).
3. Prostethics
Unlike orthoses, which are generally used to augment function of the existing
biological structures, prostheses are used to completely replace, not only lost function,
but also a lost part of the biological structure – part of a limb. Although some
similarities exist between the design principles for orthoses and prostheses, there are
also major differences. Let us consider the application of an above-knee prosthesis,
which is also called a trans-femoral prosthesis. Such a prosthesis is fitted after the
amputation of the leg at the level between the hip and the knee joint. The remaining
thigh is called a stump. Figure 17 illustrates the remaining leg and a trans-femoral
prosthesis. The interface between the stump and the prosthesis is the socket, which is a
critical element of prosthesis. It needs to provide a firm contact between the stump and
prosthesis as well as even distribution of the pressure over the entire stump-socket
contact area in order to avoid damage to the skin of the stump. Below the socket there
is an artificial knee joint that can be a simple hinge joint or a more sophisticated
polycentric knee that resembles the movement in a human knee joint. Below the knee
joint there is a pylon that connects the ankle-foot complex with the knee.
Figure 18 shows the swing phase of walking with a trans-femoral prosthesis. Since
the knee and the ankle-foot complex are passive, the user of a trans-femoral prosthesis
needs to learn to control the movement appropriately. First, the hip flexor muscles
accelerate the hip in flexion as in normal walking. However, this also initiates the
acceleration of the artificial knee joint into flexion. The artificial knee joint has certain
damping, which can be mechanical friction or viscous damping of the hydraulic fluid in
the joint, that is set-up such to assure appropriate swing time of the artificial lower leg.
Immediately after the leg is in the air, a brief activity of hip extensors is needed to
reverse the motion in the knee joint in order for the leg to be fully extended when
impacting the ground, which is vital for the stability of the limb in the following stance
phase.
340 Z. Matjačić / Gait Analysis and Synthesis: Biomechanics, Orthotics, Prosthetics
pelvis
stump
hip socket
knee
pylon
foot
Figure 17. Schematic presentation of various components of a trans-femoral prosthesis – sagittal plane view.
r r
MH MH
acc knee
damping damping locked
acc
acc
acc
r r
r MH MH
MH
knee knee
locked locked
impact
References
[1] Gage, J., R., (editor), The Treatment of Gait Problems in Cerebral Palsy, Mac Keith Press, London, UK,
2004
[2] Inman, V. T., Ralson, H. J., Todd F., Human Walking, Williams and Wilkins, Baltimore, USA, 1981
[3] Perry, J., Gait Analysis: Normal and Pathological Function, SLACK Incorporated, Thorofare, USA,
1992
[4] Winter, D. A., The Biomechanics and Motor Control of Human Gait: Normal, Elderly and Pathological,
University of Waterloo Press, Waterloo, Canada, 1991
[5] Zatsiorsky, V., M., Kinematics of Human Motion. Human Kinetics, Champaign, USA, 2002
[6] Zatsiorsky, V., M., Kinetics of Human Motion. Human Kinetics, Champaign, USA, 2002
[7] D. Popović and T. Sinkjaer, Control of Movement for the Physically Disabled, Springer, 2000
[8] Seymour, R., (editor), Prosthetics and Orthotics: Lower Limb and Spinal, Lippincott Williams &
Wilkins, Baltimore, USA, 2002
Basic Engineering for Medics and Biologists 343
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-343
Medical applications of the discharge of electric fish for the treatment of headache and
arthritis were first recorded by Scribonius Largus in the second century AD. However,
the first electric devices making use of static electricity were developed in the
eighteenth century. These were based on the principle of using friction of insulators,
such as glass, to develop high voltages due to charge separation. The ninteenth century
brought a more practical stimulator with an induction coil. The so-called 'Faradic
stimulator' consisted of a mechanical interrupter in series with a battery and the primary
winding of a transformer. The output from the secondary winding of the transformer
was a series of pulses which were similar to the stimuli of the present day electrical
stimulators. The Faradic stimulator was the first device that could produce controlled
and sustained muscle contractions. The output frequency could be adjusted to easily
exceed the stimulation rate of 100 pulses per second and the output level was also
controllable. Finally, the twentieth century brought transistors, integrated circuits, and
microprocessors permitting sophisticated electronic circuits to be incorporated into
very small devices.
344 T. Bajd and M. Munih / Basic Functional Electrical Stimulation (FES) of Extremites
1. FES Parameters
The ionic flow across the nerve influences transmembrane potential and can
generate an action potential. The action potential propagates along the nerve causing
contraction of a paralyzed muscle. In this way, FES provides restoration of movement
or function, such as standing or walking by a person with a spinal cord injury.
FES is performed in a series of rectangular monophasic or biphasic (symmetrical
or asymmetrical) electric pulses described by the following parameters: amplitude or
intensity of pulses, frequency or pulse repetition rate, duration of single pulse, and
duration of a pulse train. In most cases of surface FES applications, periodic
monophasic or unidirectional pulses are used. Biphasic or bidirectional pulses prevent a
slow deterioration of the electrodes, while the chemical conditions on the skin and in
the muscular tissue remain unchanged.
With respect to the stimulator output circuit, electric pulses are either voltage- or
current- controlled. Stimulators providing a constant output voltage can maintain the
voltage desired, irrespective of resistance changes in the stimulated muscle tissue.
Stimulators with current output stages make constant current pulses possible. An
important difference between the two stimulator types becomes evident in the case of
an improper contact between the electrode and the skin. In the case of a constant
current stimulator, a smaller effective electrode surface results in greater current
density, which can cause skin burns. With a constant voltage source, the resistance
T. Bajd and M. Munih / Basic Functional Electrical Stimulation (FES) of Extremites 345
Figure 2. Muscle recruitment curves assessed in 10 paraplegic subjects after a FES restrenghtening
programme.
346 T. Bajd and M. Munih / Basic Functional Electrical Stimulation (FES) of Extremites
diameter, and those supplying fast muscles have axons of a larger diameter. In a
voluntary contraction of a normally innervated muscle, the slow fibers are recruited
first, and as increased muscle force is required, the fast fibers are recruited. Slow fibers
are, therefore, activated frequently, while fast fibers are employed only infrequently,
during a burst of intense activity. When applying electrical stimulation, fibers with a
greater diameter respond earlier. This are motoneurons innervating fast muscles. The
normal order of recruitment is, therefore, reversed resulting in an increased fatiguing of
electrically stimulated muscle. In addition, by electrical stimulation, the same nerve
fibers are stimulated all the time, whereas with a healthy muscle the work is divided
among different motor units of the same muscle. Also, due to relatively high
stimulation frequency, the transmitter in the neuromuscular junctions is being
exhausted, so the muscle stimulated soon shows signs of fatigue.
In a fashion similar to amplitude, pulse duration also exerts a direct effect upon the
intensity of contraction. Here again, this is determined by threshold value and response
saturation. When applying surface stimulation electrodes, the accompanying unpleasant
sensation (when preserved) or even skin damage is due mainly to an excessively long
duration of a stimulus, therefore short durations are used (0.1 to 0.5 ms), while the
force of a paralyzed extremity is controlled by increasing the stimulus amplitude.
Changing the pulse duration has little or no effect on stimulated muscle fatigue.
Functional movement of a paralyzed extremity cannot be obtained by a single
electric stimulus, but requires a series of stimuli of a certain duration, following one
another at an appropriate frequency. Such a series of stimuli is called a stimulation
pulse train. In FES training of atrophied muscles, a stimulation pulse train is followed
by a pause, and then by another stimulation train. The relationship of train duration and
pause is often called the duty cycle and exerts an influence upon the fatigue of a
stimulated muscle.
2. FES Electrodes
Besides distinguishing between positive and negative electrodes, we also speak about
unipolar and bipolar electrical stimulation techniques. With unipolar stimulation, one
electrode is often considerably smaller than the other, whereas the electrodes used in
bipolar stimulation both have the same size. In unipolar stimulation the smaller
electrode is negative and is also called an active electrode due to the fact that in its
vicinity there occurs depolarization of the membrane of nerve fibers. In motor nerve
stimulation, the active electrode is positioned as closely to the motor point of the
muscle as possible. The motor point is a site on the skin, where the amplitude of the
stimulus required to fully activate the muscle is at a minimum, and where all of the
motor nerve fibers are closest to the stimulating electrode. In multichannel electrical
stimulation systems, it is possible to have a single anode and several independent
cathodes or to have anodes and cathodes that are galvanically separated.
Let us examine four properties of surface stimulation electrodes and electrodes
positioning, which influence the effectiveness of electrical stimulation: electrode size,
polarity of electrodes, resistance, and distance between the electrodes.
(i) Electrode size: Electrical stimulation is applied to a nerve fiber, since muscle
fibers have a considerably higher stimulation threshold. Thus we can say that larger
electrodes are used to stimulate the nerve endings spreading all over the underlying
tissue, whereas smaller electrodes are applied to influence the nerve when the latter
come closer to the skin. Using larger electrodes, stronger contraction is obtained along
with a reduced current density and a less pronounced unpleasant sensation on the skin.
However, large electrodes permit no selective choice of a desired movement of the
stimulated paralyzed extremity. The active areas of electrodes range between 2 cm2 and
50 cm2. Electrodes of 2 cm2 to 4 cm2 are used to stimulate the nerves near the surface,
those of about 8 cm2 for the stimulation of smaller muscles, while electrodes of 25 cm2
or more are used in case of larger muscles.
(ii) Polarity:A positive and a negative electrode are placed along the muscle to be
stimulated. Considering their polarity, the electrodes are positioned so as to provoke an
optimal movement from the functional point of view. Stronger movement is usually
obtained by placing the positive electrode distally.
(iii) Resistance:It is desirable that the resistance should be as low as possible in
order to avoid energy losses before the stimulation has reached the neuromuscular
tissue. The impedance between the electrode and the skin is frequency dependent. The
DC (or low frequency) impedance tends to be several orders of magnitude larger than
the impedance at higher frequencies. Nominal values of 2 kȍ are encountered. Contact
conduction is increased by moistening the electrodes with water or special conductive
electrode gels. Adipose tissue offers high resistance to electrical currents and so higher
stimulation amplitudes must be used, causing pain in the skin as a side effect. Bones,
too, are very bad conductors of electric current; electrical stimulation cannot reach
muscles which are behind them.
(iv) Distance between electrodes:The greatest current density appears at the skin-
electrode contact and tends to decrease with distance from the electrodes as the flow
spreads out over a larger area. Closely spaced, small electrodes generally make the
effective area of stimulation rather superficial due to the lower impedance of the
current path through proximal tissue. Deeper tissues will be stimulated by using a
greater distance between the electrodes. Increasing the electrode separation leads in
general to an increase of the maximal achievable force. If the skin between the
T. Bajd and M. Munih / Basic Functional Electrical Stimulation (FES) of Extremites 349
electrodes is moist, this causes the current between the electrodes to flow to the skin
which results in a burning sensation and a slight or no muscle contraction at all.
The design criteria for surface stimulation electrodes are: physical comfort to the skin,
electrical surface area greater than four square centimeters to prevent skin irritation, use
of hypo-allergenic materials, flexibilty to follow body surface, ease of attachement and
ability to remain in position for a duration of at least one active day, reusable, low cost,
reliable means of connection to stimulator, resistant to medical solvents and electrode
gels, low and stable electrical resistance.
The simplest surface electrodes consists of a metal plate or metal wire mesh coated
with fabric or sponge. Common materials used are stainless steel, silver-silver chloride,
platinum, or gold. For safety purposes, the upper part of the electrode is covered with a
non-conductive material. The electrode is applied after having been moistened with
water. Such electrodes are usually fixed on the extremity by means of velcro or elastic
bands. With the development of longer term surface electrodes, these electrodes are
often used for site evaluation either in pain treatment or stimulation provoking muscle
contraction. Small, button-shaped electrodes of similar design are highly suitable for
the stimulation of a single nerve. Here, the metal plate is coated with several layers of
gauze so that the electrode might retain the moisture required for as long as possible.
Surface electrodes made of silicone impregnated with carbon are applied to the
skin surface with conductive gels and held in place with adhesive patches. A too thinly
or unevenly spread gel increases current density at certain points, thereby bringing
about a danger of burns. Electrodes may be left on the skin for several days a time.
Another important property of electrodes made of conductive rubber is their flexibility
making them adaptable to any part of the body. These electrodes can be shaped by
cutting, so as to adapt them as much as possible to a proper stimulation site.
Conductive adhesive gel electrodes provide self-adhesion when applied (Figure 4).
Karaya gum was the first of these adhesive gels. Later a variety of electrically
conductive polymers were developed, enabling good contact with the irregularities of
the skin surface and a more uniform current density at the electrode-skin interface.
These electrodes can be used for extended periods with minimal skin irritation. An
important breakthrough in electrode design was made by Axelgaard. This is a flexible
electrode which adheres well to the patient's skin, is easily removed therefrom, and is
able to move with the patient's skin ensuring proper placement of the electrode. The
electrode is in the form of a knit conductive fabric. Conductive fibers include a blend of
stainless steel and polyester. The fabric can be stretched up to about 20%. A conductive
adhesive fills interstitial areas of the knit fabric and adheres the electrode to the
patient's skin. A non-conductive sheet on the other side of the knit fabric prevents
undesired electrical contacts.
When improperly handled, electrodes can damage the skin in the contact area. Burns
typically occur underneath the anode, but not the cathode, when using identical surface
electrodes. Another problem resides in a precise electrodes positioning along a muscle.
Sometimes a displacement of a few millimeters completely changes the muscle
response. This happens when a selected nerve (e.g. peroneal nerve) ought to be
350 T. Bajd and M. Munih / Basic Functional Electrical Stimulation (FES) of Extremites
stimulated by surface electrodes. Surface electrodes may excite pain receptors in the
skin, although patients sensibility may be reduced to such an extent that the sensation
of pain is not critical. Another problem is undesired motion of the skin with respect to
the neuromuscular tissue. Even though an electrode seemingly occupies the same place
all the time, its distance from the nerve is not constant. This is one of the reasons why
the movements caused by electrical stimulation cannot be easily repeated. Another
limitation is that small muscles usually cannot be selectively activated and deep
muscles cannot be stimulated without first exciting the superficial muscles. Relatively
high voltages, sometimes in excess of 100V, between electrode pairs cause hazards for
the patients and the personnel that treat them. Finally, the applicability of the surface
stimulation electrodes depends on fixation problems. Stretchable garments with
electrodes already mounted in appropriate locations have been developed by several
manufacturers to simplify the application of electrodes to the skin surface. In the case
of lower limb stimulation, fixation problems can be overcome by specially designed
trousers carrying stimulation electrodes and cables. Such stimulation equipment is
comfortable and easy to handle. In the non-invasive, upper limb neuroprosthesis, the
surface stimulation electrodes were built into an elegant, self-aligning, and flexible
splint. The splint provides additional fixation of the wrist joint and allows the entire
electrode array to be positioned within a few seconds.
Figure 4. A conductive adhesive gel electrode with a portion of nonconductive sheet peeled back showing
the knit fabric.
3. Electrical timulators
Electrical stimulators comprise an input circuit, pulse generator, output stage, and
power supply (Figure 5). The input into the electrical stimulator is represented by a
control signal automatically switching electrical stimulation pulses on or off or may be
under the patient's voluntary control. The output electric pulse current is led, via
T. Bajd and M. Munih / Basic Functional Electrical Stimulation (FES) of Extremites 351
electrodes, to the selected stimulation site. The type of stimulation pulse is determined
by a pulse generator. The output pulses provided by the circuit are considerably lower
than the surface stimulation pulses required for a functional movement of an extremity
but are of an appropriate frequency and duration. The output stage ensures energy for
the electrical stimulation of paralyzed muscle in either constant current or the constant
voltage outputs. In case of current stimulation, the internal resistance of the end stage is
considerably higher than the tissue resistance between the electrodes. The current
source of stimulation pulses provides a constant current irrespective of the resistance of
the skin and the tissue between the electrodes. In the case of constant voltage output
stages, skin resistance is lower than that between the electrodes, and the stimulator
provides a constant voltage independent of the skin and tissue resistance. A power
supply provides the energy necessary for the operation of particular electronic circuits
(low voltage) and the electrical stimulation itself (high voltage). Electrical stimulators
are usually battery powered with stimulation pulses with an amplitude of more than
100 V. A high voltage for the output stage is obtained from low battery voltage by
means of a voltage convertor.
The development of electronics made it possible for Wladimir Liberson to develop
a stimulator to preventing foot-drop in hemiplegic patients. It was triggered by a heel
switch in the shoe of the affected leg. Stimulation electrodes were positioned above the
peroneal nerve in the popliteal fossa, behind the knee joint. Each time the patient raised
the heel, the heel switch triggered the stimulator, causing the nerve to cause the
extensor group of muscles to contract and dorsiflex the ankle and so lift the foot.
Advanced versions of these dropped foot stimulators were developed in Ljubljana,
Slovenia and Salisbury, UK. Both stimulators were applied to a large group of stroke
patients. In the WalkAide peroneal nerve stimulator, a tilt sensor triggers ankle
dorsiflexion.
Control Electrodes
Power
Supply
A minimum of four channels of FES were used for synthesis of a simple reciprocal
gait pattern in completely paralyzed paraplegic subjects. During reciprocal walking, the
stimulator must be controlled through three different phases of walking: right swing
352 T. Bajd and M. Munih / Basic Functional Electrical Stimulation (FES) of Extremites
phase, double stance phase, and left swing phase. This is achieved by two hand-push
buttons built into the handle of the walker or crutches. When niether of the push
buttons is pressed, both knee extensors are stimulated providing support to the body.
On pressing the push button in the right hand, the peroneal nerve is stimulated in the
right leg eliciting the flexion reflex presented by simultaneous hip and knee flexion and
ankle dorsiflexion. The same is true for the left leg. The Ljubljana FES walking system
consists of two small two-channel stimulators attached to each leg. Only three
electrodes are applied to a single leg in order to produce knee extension and flexion
responses. As both activities never occur simultaneously, the distal electrode placed
over the knee extensors represents the common electrode for both stimulation channels.
Using the same principles as the Ljubljana FES system, and adding two channels of
stimulation to both hip extensors, the FDA-approved Parastep surface stimulation
system was developed. A new multipurpose programmable transcutaneous electric
stimulator, Compex Motion, was developed to allow users to design various custom-
made neuroprostheses, neurological assessment devices, muscle exercise systems, and
experimental setups for physiological studies. The Compex Motion stimulator can
generate any arbitrary stimulation sequence, which can be controlled in real-time using
any external sensor.
Despite the fact that correct application of a stimulator implies no danger either for
the patient or therapist, we wish to emphasise some important safety points. When
buying a new electrical stimulator, be sure that it carries the CE mark. Every stimulator
has its own characteristics which must be indicated in the instructions of use, an
obligatory accompaniment of any commercially-available stimulator. Special attention
should be paid to stimulation parameters: pulse duration, frequency, and maximal
current or voltage. Also of utmost importance is the information on the proper use of
surface electrodes. The use of electrical stimulators might be dangereous in case of
patients with an implanted pacemaker. Simultaneous use of electrical stimulators and
high frequency surgical devices is prohibited. Electrical stimulators might not work
properly in close proximity to microwave devices. Transthoracic positioning of FES
electrodes may cause fibrillation of the heart. The stimulator should not be switched on
in case of short-circuit of the electrodes. It is true, that electrical stimulators are made
in such a way that a short-circuit of several minutes does not damage them. However,
one should not take unfair advantage of this property.
Further Reading
[1] L.A. Benton, L.L. Baker, B.R. Bowman, R. Waters, Functional electrical stimulation – A practical
guide, Rancho Los Amigos Hospital, 1980
[2] L. Vodovnik, T. Bajd, A. Kralj, F. Graþanin, and P. Strojnik, Functional electrical stimulation for
control of locomotor systems, CRC Critical Rev. Bioeng. 6 (1981), 63-131
[3] A.R. Kralj and T. Bajd, Functional electrical stimulation: Standing and walking after spinal cord
injury, CRC Press Inc., 1989
[4] L.S. Illis (ed.), Neurological Rehabilitation, Blackwell Scientific Publications, 1994
[5] D. Popoviü and T. Sinkjaer, Control of movement for the physically disabled, Springer, 2000
[6] P.J. Rosch and M. Markov (eds.), Bioelectromagnetic Medicine, Marcel Dekker Inc., 2004
[7] K.W. Horch, GS Dhillon (eds.), Neuroprosthetics – Theory and Practice, World Scientific, 2004
[8] J.D. Bronzino (ed.), The Biomedical Engineering Handbook, CRC Press and IEEE Press, 2000
[9] M. Akay (ed.), Wiley Encyclopedia of Biomedical Engineering, John Wiley&Sons, 2006
[10] L. R. Sheffler, J Chae, Neromuscular electrical stimulation in neurorehabilitation, Muscle Nerve 35
(2007), 562-590
Basic Engineering for Medics and Biologists 353
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
doi:10.3233/978-1-60750-527-3-353
Abstract. The paper presents the background, main achievements and components
of rehabilitation robotics in a simple way, using non-technical terms. The
introductory part looks at the development of robotic approaches in the
rehabilitation of neurological patients and outlines the principles of robotic device
interactions with patients. There follows a section on virtual reality in
rehabilitation. Hapticity and interaction between robot and human are presented in
order to understand the added value of robotics that cannot be exploited in other
devices. The importance of passive exercise and active tasks is then discussed
using the results of various clinical trials, followed by the place of upper and lower
extremity robotic devices in rehabilitation practice. The closing section refers to
the general importance of measurements in this area and stresses quantitative
measurements as one of the advantages in using robotic devices.
Introduction
in gradual steps to finally learn complete movement, for example, in spasticity. This
approach highlights and relies on the plasticity of the nervous system. Solving
problems includes activities focused at a specific task, motivation, planning, interaction
with the environment, selective placement of attention by stimuli and biomechanical
aspects for effective and functional direction of movement. Thus, haptic interaction
between a patient and a therapist must not be invasive and unilateral. On the contrary, it
must evoke all the capabilities of a person leading to functional movement by
minimum intervention.
To enable the best possible imitation of natural human interaction, while using a
machine as a rehabilitator assistant, a robotic therapist must be highly consistent with
human interaction. It is no coincidence that the original ideas which brought these
systems to life did not arise in the rehabilitation field, but rather in neurophysiology
and haptic perception, including sensimotor learning [7-9]. Robotic therapist devices
differ in a number of ways from industrial robots. These are traditionally moved
between two known points along a defined trajectory. Maintaining a known and highly
accurate position is essential. In contrast, the robotic therapist must be programmable
and adjustable. This can be achieved by impedance control algorithms or by passive
manual control as in backdrivable robot manipulators including MIT-Manus and
Braccio di Ferro. Other devices designed for robotic therapy might use industrial
manipulator technology and have high manipulator arm mechanical impedance due to
high gearing ratio and admittance control scheme.
Robotic devices may be used in positional trajectory mode for enforcing passive
movements, for simple positional control of extremities, which, does not involve active
participation of a patient on either neuromuscular or sensory levels. Such passive
exercise, quite the opposite of the Bobath concept, using a robot contributes to
rehabilitation, at least in specific clinical conditions [10-12].
Other studies [13-15] indicate better results with the application of techniques that
consider the adaptive nature of the nervous system by solving problems with a suitable
adaptability level. These techniques include active assisted exercises in which the robot
moves the extremity along a predetermined trajectory. In the active constrained
exercises, the robot provides higher opposing forces. The movement may be limited
inside some 3D virtual space, with opposing forces applied when the subject tries to
move outside this region. Similarly, one could imagine a ball object that limits
movement toward the center (inside), with completely empty space all around.
Furthermore, in active resistive exercises, the robot opposes the intended movement.
In adaptive exercises, the robot is providing a previously unknown dynamic
environment to which the subjects have to react and adapt.
Passive exercises need no input from the patient, while active constrained
exercises require at least residual movement cababilities or subconscious sensory-
motor co-operation. Active resistive and adaptive exercises require cooperation with
sufficient volitional motor activity. Active resistive and adaptive exercises are therefore
not a suitable choice for persons with larger movement deficits, since they usually
cannot use them independently. Still, such persons may exploit robot-therapeutic
exercises in which minimum co-operation on their side will help them in using and
strengthening their remaining abilities.
M. Munih and T. Bajd / Rehabilitation Robotics 355
Most of us are familiar with virtual reality (VR) technology from entertainment (e.g.
games) and military simulations. Lately, its use has expanded to other areas, for
instance computer-aided design (CAD), architecture, general virtual presentation of
data and to medical applications. Medical applications of virtual reality include training
in medicine and surgery, modelling of hard and soft tissue, image displays, remote
surgery, ergonomy and rehabilitation. In rehabilitation, virtual reality is used for
training and measuring the motor ability (e.g. in hemiplegia, paraplegia, Parkinson's
disease).
Virtual reality is usually understood as a three-dimensional computer model which
primarily defines the geometrical model (kinematic or dynamic model) of different
virtual objects and their environment. It is possible to define not only the image of
objects, but also their inherent physical characteristics. Such a virtual world can change
in time - not only the position and orientation of particular items, but also dynamic
characteristics of the surroundings, such as friction and gravity. Realistic visualisation
and deep presence of person in VR are mostly dependent on high-quality graphics
which take the user into the virtual world. The impression of virtual reality depends on
the realistic appearance of the scene and activity. Simpler non-immersive
visualisations include a 2D computer screen, a projection screen and 3D projection
techniques in the environment. The immersive method is more realistic, enabling the
user to percieve a full 3D visual field by use of special glasses.
The dynamic characteristics of objects, including mass, moment of inertia, forces,
torques, and torques and forces resulting from the interaction of objects and, for
instance, compliance (stiffness) of objects, roughness and smoothness of objects cannot
be visually detected or presented. These require the sense of touch.
2. Hapticity
The verb ‘ĮʌIJȦ’ - hapto is originating from Greek for reaching, holding and touching.
If a person exerts a force F onto a mass m, which is integrated in the environment via
the damper h and the spring k, the movement x is determined by the differential
equation:
§ d2 d ·
F ¨m h k¸x m x h x k x (1)
¨ dt 2 dt ¸
© ¹
If the coefficients m, h and k are constant, the force depends on the position
(distance to origin), velocity and acceleration. The first force term in the equation
depends on mass and acceleration m x , with m representing the mass of e.g. a cube on
icy surface. Significant force will be needed only during acceleration and deceleration
with high m, while constant velocity movement requires no force. Only a small force is
needed for the acceleration and deceleration of a light styrofoam cube with a small m.
The force can, as a second example, depend on the term h x , where h stands for
damping of e.g. of an oar in water. If the velocity x is low, only a small force is
needed, whereas at higher velocities, a considerably greater force is required. Another
medium, like oil or a spoon in honey or air, represents a different damping coeficient
356 M. Munih and T. Bajd / Rehabilitation Robotics
value h and thus a different force at the same velocity x . The damping force is small
when an oar moves in air, but if the velocity is high, in the case of an airplane propeller,
the pull/push force becomes significant. The impact of stiffness k is presented by
means of a spring. The greater the deviation x, the greater the force. Also, a higher
spring stiffness coeficient k value, with x unchanged, requires greater force.
A general body in a virtual environment, onto which a person exerts a force, is not
a point mass, pure damper or spring but rather a combination of all three terms in the
above equation that becomes more complex. This is typical for haptic touch in a virtual
environment, where none of the three coeficients of the equation is constant. The
values of m, h and k change locally in the virtual environment. Actually, one of the
three parameters, e.g. k, has six values at the same point (or position) in space: kx, ky
and kz along individual axes of space, and kRx, kRy and kRz around individual axes of
rotation – thus it has six degrees of freedom (DOF) (Fig. 1). An empty space has nil or
small values of all six k. A flat wall in a virtual environment is represented as a high k
horizontally. A wall can also be slippery as ice or non-slippery if coated with rubber,
which is set with the other two (transversal) k parameters of the wall. The remaining
three k elements stand for the three rotational stiffness coeficients. Similar six-
dimensional understanding applies to the other two parameters - h and m.
hZ kZ
hY
kRZ
hRZ kY
hRY
zO kRY
yO
O
xO
FO kRX
hRX
y hX
kX
x
Figure 1. Second-order mechanical system with six DOF, three translations and three rotations
and m, h and k
3. Passive Exercise
CPM (Continuous Passive Motion) devices have been used in the postoperative
rehabilitation of joints since approximately 1960. It was later established that joint
immobilisation in a test group of rabbits led to great problems in restored mobility [16].
After five to six weeks of immobilisation, most joints developed moderate to extreme
changes, including those in joint cartilage and bone mass, with changes already
apparent in the second week. In clinical practice, this presented as joint contractures
and reduced range of motion (ROM). To avoid this, they tried to maintain a good ROM
by simple CPM devices which moved the limb over an arc corresponding to joint
movement. Such devices consist of a simple motor with a mechanically or
electronically adjustable range and velocity of movement. Most modern CPM devices
adopt ideal joints with a fixed point of rotation, and can provide movement in one
vertical plane (2D). Usually they cannot be reprogrammed and are controlled using the
open loop principle, which means that their movement angles and forces are not
measured immediately but are corrected and reset a few 100-times in a second.
Such CPM devices are very efficient in preserving range of movement, they reduce
stiffness in joints, decrease the need for drug administration and shorten the length of
hospitalization. Comparison of CPM and physiotherapy reveals no major differences as
regards the above parameters, however in the majority of cases, CPM results in weaker
muscles, delays in activation of extensors and stiffness of flexors [17]. It would be
358 M. Munih and T. Bajd / Rehabilitation Robotics
difficult to expect anything else, since the CPM devices only move a person's
extremities without activating the muscles.
4. Active Exercises
Several research groups developed robotic devices for rehabilitation of the upper and
lower extremities. These devices, designed for exercise, use various methods of
operation. Active constrained movements are made possible by most of the mentioned
devices, while movements with active resistance are provided by MIT-Manus, Bi-
Manu-Track and MIME [18-20]. Adaptive exercises are possible using Bi-Manu-Track
and MIME [19, 20], in which case the healthy arm leads and the injured arm imitates
movement, both moving simultaneously.
Finteraction
Finteraction
Figure 4. Example of active exercise with active resistance (torus), HIFE project
Figure 5. Example of active exercise with active resistance (pipe with a spring), I-Match project
might be more successful than others. Only Fasoli et al. [27] and Stein et al. [28] tried
to specifically evaluate particular exercise versions. The findings of the Fasoli trial
disclose that active resistive exercises are more useful than active constrained exercises
as regards the upper extremities. Still, the repeated trial covering the same group did
not reveal any differences between the exercises involving active constrained
movements and those involving active resistive movements, perhaps because of the
manner of data calculation. In the first trial Fasoli et al. [27] included in the group of an
active constrained movement those who did not have sufficient motor function for the
active resistive group. This was not the case in the trial conducted by Stein et al. [28].
Thus, the results of active resistive exercises were overrated in advance. These findings
point to the fact that some robot-assisted therapies are less suitable for specific groups
of patients. Therefore, the precise contribution of individual exercise methods to the
rehabilitation of upper extremities in a hemiplegic person remains unexplained.
Other methods of exercise are perhaps equally or even more important for robotic
rehabilitation, but they have not been investigated sufficiently, or at all. One such
method is gravity compensation of the upper extremity. Most robotic devices provide
some sort of support for the arm. The evolution of gravity compensation devices has
been going on for decades – for example Sanches [29]. Beer et al. specifically
investigated the implementation and reach of gravity compensation for the upper
extremity [30-32]. Their preliminary research showed that poor coordination of muscle
activation led to unexpected torques in the hemiplegic arm joints (e.g. shoulder
abduction prevented the ability of elbow extension). Further trials revealed immediate
improvement of motor abilities in persons after a stroke, if gravity did not influence the
upper extremity. The active abduction of the shoulder was reduced and consequently
extension in the elbow increased in static conditions. The latest results point to a
similar mechanism functioning in dynamic circumstances. Such findings indicate
improved motor abilities due to the use of gravity compensation. The second trial
involving exercise with a sling revealed no significant difference. Even though the
principle of gravity compensation is at present rather unexploited, faster recovery can
be expected along with new trials. Recently, three arm gravity compensators were put
on the market: Armon1, Dynamics Arm Support2 (DAS) and Armeo3. The first two
devices provide lower arm support, while the third is an exoskeleton type mechanism
allowing free arm movement for people with little muscular arm strength, using passive
principles, thus providing for static and not accounting for dynamic compensation.
More research in this field can clarify the mechanisms of the compensation impact, as
well as indicate where it would be reasonable to apply these mechanisms.
Besides the above approaches, other methods of rehabilitation can additionally
promote fast recovery in hemiplegic patients. Interesting assertions were published by
Kahn et al. [35], claiming that equally positive changes in motor function were
observed in a group undergoing robot-assisted reaching exercise (target is always
reached) and in a group undergoing robot-assisted reaching exercise without obligatory
reaching of the target. Some other possible techniques deserve mention: functional
electrical stimulation, pharmacology, intensive exercise, including many repetitive
exercises, loading therapy - automated use of this method - and sensor-motor exercise.
1
www.mginside.info
2
www.exactdynamics.com
3
www.hocoma.ch
M. Munih and T. Bajd / Rehabilitation Robotics 361
A number of research or commercial platforms exist today designed specifically for the
tasks of rehabilitation robotics, while in some other platforms the primary design goal
has been aimed at something else, but the resulting device also suits the needs of
rehabilitation. Regardless of their origin, the devices can be of either the exoskeleton or
the end-effector type. An exoskeleton is an external framework, close to, or in contact
with, human limbs.
Exoskeletons contain rigid and resistant components for each of their segments,
being able to providing passive or active translation or rotation in a particular joint. The
rotation axes of the robot must correspond to the rotation axes of the human skeleton,
and the limb may be connected to the exoskeleton at several points. Powered
exoskeletons may be designed, for example, to assist and protect soldiers and
construction workers, or to aid the survival of people in other dangerous environments.
Wider medical markets exist for providing mobility assistance and rehabilitation for
aged and infirm people. A weak point in exoskeleton devices is the limited amount of
Z-bandwidth and the transparency of the haptic interface – the parameters that tell how
good is the touch (impedance) compared to the ideal model in the virtual environment.
A strength is the good fit to each human body segment and the ability to move each
limb and interact with supervised known parameters.
Perry et al. [39] is using an exoskeleton robot with seven degrees of freedom in the
most important joints of human arm. L-Exos is a tendon-driven, wearable haptic
interface with 5 DoF, optimized to obtain a solution with reduced mass and high
stiffness, by employing special mechanical components and carbon fiber structural
parts [41]. Neural control of an upper limb powered exoskeleton system has 8 DoF
[44]. ARMin represents an interesting later design that at the moment enables
movements in 6 DoF [26].
In most cases, robotic haptic arms that interact with the environment only via end-
effectors are in contact with the human arm via the wrist or with the human leg via the
foot only. As a consequence, the trajectory of each body segment under motion is not
supervised and determined by a machine, as above, but also depends on the person.
This can be good, but may also lead to a less determined environment. Good points in
362 M. Munih and T. Bajd / Rehabilitation Robotics
the end-effector approach are better values for Z-bandwidth and transparency of the
haptic device.
Examples of end-effector upper extremity devices include MIT Manus [18],
Assisted Rehabilitation and Measurement (ARM) Guide [19], Mirror Image Motion
Enabler (MIME) [20], Bi-Manu-Track [21], GENTLE/S [22], Neurorehabiltation
(NeReBot) [23], REHAROB [24], Arm Coordinating Training 3-D (ACT3D) [25],
Braccio di Ferro, [42], NEDO project device [43] and Baltimore Therapeutic
Equipment Co. (BTE), produce several training devices.
Most devices have been designed for training of the proximal joints (shoulders and
elbows) of the hemiplegic arm. Some allow for two-dimensional plane movement only,
while in most cases, movement is possible in three dimensions within the limited
section of the entire working area of the arm. On the contrary, Bi-Manu-Track includes
the forearm and wrist, which is enabled also by the recent version of the MIT-Manus
device. Over and over again we encounter new robotic devices and evolutional
upgrades of the existing ones (e.g. Furusho et al. [33] and Colombo et al. [34]).
There are also devices available for the lower extremities, Gait Trainer GT I [37]
and HapticWalker [38] most frequently in combination with treadmills, and also
separately for ankles. In the lower extremities, the end-effector devices only guide foot
motion, while the multiple degrees of freedom of the rest of the body (e.g. leg, hip)
remain completely unrestricted. In the case of a patient with an unstable knee joint, the
physiotherapist has to stabilize the knee manually. Studies on the Gait trainer GT I
revealed comparable muscle activation patterns as in treadmill walking.
Research devices are also available for exercising some other joints of the body,
like ankle, wrist, individual fingers [45] or several fingers simultaneously.
Figure 6. ARMin robot is one of the recent exoskeleton devices (ETH Zürich)
M. Munih and T. Bajd / Rehabilitation Robotics 363
6. Measurements in Rehabilitation
In neurology, standard rating scales are used for measurements, mainly for the purpose
of early systemisation. On the other hand, uniform, quantitative, traceable procedures
for measurements in rehabilitation have not yet been introduced. One of the models
supposedly establishing a logical, understandable and thorough system in rehabilitation
was adopted by the World Health Organisation in 1980. Even though the model, known
as ICIDH (International Classification of Impairments, Disabilities and Handicaps) is
very detailed and subjective (rheumatologic diseases), it provides a solid framework for
understanding and treating neurological diseases and injuries. The most important
concept of the model is the assumption that every disease can be evaluated on four
levels, namely:
o pathology level,
o impairment level,
o disability level,
o handicap level.
In practice, the border between individual levels can be quite blurred. In general, a
distinction has to be made between measurements and scoring scales. Measurement
involves the use of a specific, traceable standard and the comparison of an actual value
of a certain quantity against this standard. Scales often rely on subjective observations
or rough measurements leading to a very few quantum descriptors. In rehabilitation,
measurements are much rarer than scoring scales. The simplest scoring of motor
364 M. Munih and T. Bajd / Rehabilitation Robotics
abilities and impairment of e.g. the upper extremity includes the squeezing of the
dynamometer for estimating muscle strength and measuring the range of movement of
individual joints. Self-scoring by patients is also present. The criteria of a good scoring
scale include:
o validity (the result actually refers to the aspects for which the test was carried
out),
o reliability (competent observer and time stability of results of such observer),
o sensitivity (detection, differentiation of sufficiently small changes that are still
relevant).
Standardised rehabilitation tests are for the most part tailored to a specific disease
and consider also the condition of other parts of the body and activity in general. Some
of these tests are Fugl-Meyr, Barthel Index and the widely-used Barthel Index, Katz
index, Nottingham Adl Index, Jebsen, FIM-test, Box and Block Test, Nine-Hole Peg
Test, Action Research Arm Test, Franchay Arm Test and others.
Unlike the scoring approach, haptic, robotic technology provides for the next stage
in rehabilitation, objective measurement during exercise. Quantities that vary with time
(position and orientation, velocities, accelerations, forces, torques) can often be
measured directly or derived. The presentation of these quantities for individual,
relevant points of the body offers an objective basis for evaluation, as in kinesiology,
and derived indexes can answer significant questions. Observation of measured data in
a frequency domain can result in new insights, for instance, in the case of Parkinson's
disease, the amplitude and frequency of shaking are determined by both place and time.
The measured quantities can be used as entries into various physical or
physiological models and through them new important parameters may be acquired, e.g.
active torques of the muscles of individual joints, passive torques in joints or even mass
and evaluation of moment of inertia of some body segments. In doing so, we have to be
aware of the fact that the more complex models usually comprise a higher number of
parameters, some roughly evaluated, carrying high measurement uncertainly, which in
the end may contribute to an even more uncertain final result.
7. Conclusion
Recent technological developments have made many things possible for the first time.
However, actual systems are only now being introduced into our lives in the fields of
computing, virtual reality visualisation, medicine in general and a narrow field of
rehabilitation. Previously most simple devices, accessories and physiotherapists’ hands
will in the future be complemented by computerized devices. These will help in
existing and, hopefully, also some new aspects of rehabilitation. Robotic approaches
undoubtedly allow for equally fast, perhaps even faster, but certainly more interesting
and entertaining rehabilitation methods. It has to be expected that at least some good
laboratory prototypes will develop into widely-available products. These products are
expected to have a suitable modular design, and developments in other areas, e.g.
improvement in speed of computer processing, improvement in video techniques,
software support, etc., will further contribute to the advancement in this area.
Unfortunately, the size of the rehabilitation robotic market will be considerably smaller
than the size of the market for computer games, which is why the volume of investment
will be smaller and the pace of development is expected to be slower.
M. Munih and T. Bajd / Rehabilitation Robotics 365
Acknowledgements
The described research work is financed by the Slovenian Research Agency in the form
of a programme group Analysis and Synthesis of Movement in Man and Machine, and
by providing grants to young researchers.
References
[1] D. Khalili, M. Zomlefer, An intelligent robotic system for rehabilitation of joints and estimation of body
segment parameters, IEEE Trans Biomed Eng 35 (1988), 138-46
[2] B. Gjelsvik, E. Bassoe, Form und Function. Stuttgart, New York: Thieme Verlag, 2002.
[3] J. Howle, Neuro-Developmental Treatment Approach: Theoretical Foundations and Principles of
Practice, Laguna Beach, CA, NDTA, 2002
[4] B. Bobath, Observations on adult hemiplegia and suggestions for treatment, Physiotherapy 45 (1959),
279-289
[5] E. Panturin, The Bobath concept, Clin. Rehabil. 15 (2001), 111-113
[6] S. Raine, Defining the Bobath concept using the Delphi technique, Physiother. Res. Int. 11 (2006), 4-13
[7] F.A. Mussa-Ivaldi, N. Hogan, E. Bizzi, Neural, mechanical, and geometric factors subserving arm
posture in humans, J. Neurosci.. 5 (1985), 2732-2743
[8] R. Shadmehr, F.A. Mussa-Ivaldi, Adaptive representation of dynamics during learning of a motor task,
J. Neurosci. 14 (1994), 3208-3224
[9] T. Tsuji, P.G. Morasso, K. Goto, K. Ito, Human hand impedance characteristics during maintained
posture, Biol. Cybern..72 (1995), 475-485
[10] P. Lum, D. Reinkensmeyer, R. Mahoney, W.Z. Rymer, C. Burgar, Robotic devices for movement
therapy after stroke: current status and challenges to clinical acceptance, Top Stroke Rehabil 8 (2002),
40-53
[11] G. Fazekas, M. Horvath, A. Toth, A novel robot training system designed to supplement upper limb
physiotherapy of patients with spastic hemiparesis, Int. J. Rehabil. Res. 29 (2006), 251-254
[12] S. Hesse, G. Schulte-Tigges, M. Konrad, A. Bardeleben, C. Werner, Robot-assisted arm trainer for the
passive and active practice of bilateral forearm and wrist movements in hemiparetic subjects, Arch.
Phys. Med. Rehabil. 84 (2003) 915-920
[13] C.D. Takahashi, D.J. Reinkensmeyer, Hemiparetic stroke impairs anticipatory control of arm movement,
Exp. Brain Res. 149 (2003), 131-140
[14] J.L. Patton, F.A. Mussa-Ivaldi, Robot-assisted adaptive training: custom force fields for teaching
movement patterns, IEEE Trans. Biomed. Eng. 51 (2004), 636-646
[15] J.L. Patton, M.E. Stoykov, M. Kovic, F.A. Mussa-Ivaldi, Evaluation of robotic training forces that
either enhance or reduce error in chronic hemiparetic stroke survivors, Exp. Brain Res. 168 (2006),
368-383
[16] R.B. Salter et al., The effects of continuous compression on living articular cartilage, J. Bone Joint Surg.
42-A (1960)
[17] M.A. Ritter, V.S. Gandolf, K.S. Holston, Continuous passive motion versus physical therapy in total
knee arthroplasty, Clin. Orthop. Relat. Res. 244 (1989), 239-243
[18] H.I. Krebs, N. Hogan, B.T. Volpe, M.L. Aisen, L. Edelstein, C. Diels, Overview of clinical trials with
MIT-MANUS: A robot-aided neuro-rehabilitation facility, Technol. Health Care 7 (1999), 419–423
[19] D.J. Reinkensmeyer, L.E. Kahn, M. Averbuch, A. McKenna-Cole, B.D. Schmit, W.Z. Rymer,
Understanding and treating arm movement impairment after chronic brain injury: progress with the
ARM guide, J. Rehabil. Res. Dev., 37 (2000), 653–662
[20] C.G. Burgar, P.S. Lum, P.C. Shor, H.F. Machiel Van der Loos, Development of robots for rehabilitation
therapy: the Palo Alto VA/Stanford experience, J. Rehabil. Res. Dev., 37 (2000), 663–673
[21] S. Hesse, G. Schulte-Tigges, M. Konrad, A. Bardeleben, C. Werner, Robot-assisted arm trainer for the
passive and active practice of bilateral forearm and wrist movements in hemiparetic subjects, Arch.
Phys. Med. Rehabil. 84 (2003), 915–920
[22] S. Coote, E. Stokes, B. Murphy, W. Harwin, The effect of GENTLE/s robot-mediated therapy on upper
extremity dysfunction post stroke, Proc. 8th International Conference on Rehabilitation Robotics; 2003,
Daejeon, Korea, Human-Friendly Welfare Robot System Engineering Research Center, 59–61, 2003
[23] C. Fanin, P. Gallina, A. Rossi, U. Zanatta, S. Masiero, NeRe-Bot: A wire-based robot for
neurorehabilitation, Proc. 8th International Conference on Rehabilitation Robotics, Daejeon, Korea,
Human-Friendly Welfare Robot System Engineering Research Center, 23–26, 2003.
366 M. Munih and T. Bajd / Rehabilitation Robotics
[24] A. Toth, G. Fazekas, G. Arz, M. Jurak, M. Horvath, Passive robotic movement therapy of the spastic
hemiparetic arm with REHAROB: Report of the first clinical test and the follow-up system improvement,
Proc. 9th International Conference on Rehabilitation Robotics, Chicago, Illinois. Madison (WI),
Omnipress, 127–130, 2005.
[25] J. Dewald, M.D. Ellis, B.G. Holubar, T. Sukal T, A.M. Acosta, The robot application in the
rehabilitation of stroke patients, Neurol. Rehabil. 4 (2004), 7
[26] T. Nef, R. Riener, ARMin—Design of a novel arm rehabilitation robot, Proc. 9th International
Conference on Rehabilitation Robotics, Chicago, Illinois. Madison (WI), Omnipress, 57–60, 2005
[27] S.E. Fasoli, H.I. Krebs, J. Stein, W.R. Frontera, N. Hogan, Effects of robotic therapy on motor
impairment and recovery in chronic stroke, Arch. Phys. Med. Rehabil. 84 (2003), 477–482
[28] J. Stein, H.I. Krebs, W.R. Frontera, S.E. Fasoli, R. Hughes, N. Hogan, Comparison of two techniques of
robot-aided upper limb exercise training after stroke, Am. J. Phys. Med. Rehabil. 83 (2003), 720–728
[29] R.J. Sanchez, E. Wolbrecht, R. Smith, J. Liu, S. Rao, S. Cramer, T. Rahman, J.E. Bobrow, D.J.
Reinkensmeyer, A pneumatic robot for re-training arm movement after stroke: rationale and
mechanical design, Proc. 9th Int. Conf. Rehabilitation Robotics, Chicago, Illinois, Madison (WI),
Omnipress, 500–504, 2005
[30] R.F. Beer, J.D. Given, J.P. Dewald, Task-dependent weakness at the elbow in patients with hemiparesis,
Arch. Phys. Med. Rehabil. 80 (1999), 766–772
[31] R.F. Beer, J.P. Dewald, W.Z. Ryme, Deficits in the coordination of multijoint arm movements in
patients with hemiparesis: evidence for disturbed control of limb dynamics, Exp. Brain Res. 131 (2000),
305–319
[32] R.F. Beer, J.P. Dewald, M.L. Dawson, W.Z. Rymer, Targetdependent differences between free and
constrained arm movements in chronic hemiparesis, Exp. Brain Res. 156 (2004), 458–470
[33] J. Furusho, K. Koyanagi, Y. Imada, Y. Fujii, K. Nakanishi, K. Domen, K. Miyakoshi, U. Ryu, S.
Takenaka, A. Inoue, A 3-D rehabilitation system for upper limbs developed in a 5-year NEDO project
and its clinical testing, Proc. 9th Int. Conf. Rehabilitation Robotics, Chicago, Illinois. Madison (WI),
Omnipress, 53–56, 2005
[34] R. Colombo, F. Pisano, S. Micera, A. Mazzone, C. Delconte, M.C. Carrozza, P. Dario, G. Minuco,
Upper limb rehabilitation and evaluation of stroke patients using robot-aided techniques. Proc. 9th Int.
Conf. Rehabilitation Robotics, Chicago, Illinois, Madison (WI), Omnipress, 515–518, 2005
[35] L.E. Kahn, M. Averbuch, W.Z. Rymer, D.J. Reinkensmeyer, Comparison of robot-assisted reaching to
free reaching in promoting recovery from chronic stroke, In: M. Mokhtari (ed.), Integration of assistive
technology in the information age, Amsterdam (the Netherlands), IOS Press, 39–44, 2001
[36] G.B. Prange, M.J.A. Jannink, C.G.M. Groothuis-Oudshoorn, H.J. Hermens, M.J. Ijzerman, Systematic
review of the effect of robot-aided therapy on recovery of the hemiplegic arm after stroke, J.
Rehabilitation Research & Development, 43 (2006), 171–184
[37] S. Hesse, D. Uhlenbrock, A mechanized gait trainer for restoration of gait, J. Rehabilitation Research &
Development, 37 (2000), 701–708
[38] H. Schmidt, D. Sorowka, F. Piorko, N. Marhoul, R. Bernhardt, Entwicklung eines robotergestützten
Laufsimulators zur Gangrehabilitation (Development of a Robotic Walking Simulator for gait
rehabilitation), Biomed. Eng, 40 (2003), 281-286
[39] J.C. Perry, J. Rosen, Design of a 7-Degree-of-freedom upper limb powered exoskeleton, BioRob 2006,
The IEEE/RAS-EMBS Int. Conf. Biomedical Robotics and Biomechatronics, Pisa, Toscany, Italy, 805-
810, 2006
[40] H. Schmidt, D. Sorowka, F. Piorko, N. Marhoul, R. Bernhardt, Entwicklung eines robotergestützten
Laufsimulators zur Gangrehabilitation (Development of a robotic walking simulator for gait
rehabilitation), Biomedical Engineering 40 (2003), 281-286
[41] A. Frisoli, F. Rocchi, S. Marcheschi, A. Dettori, F. Salsedo, M. Bergamasco, A new force-feedback arm
exoskeleton for haptic interaction in Virtual Environments, First Joint Eurohaptics Conference and
Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems WHC, 2005.
[42] M. Casadio, V. Sanguineti, P.G. Morasso, V. Arrichiello, Braccio di Ferro: A new haptic workstation
for neuromotor rehabilitation, Technol. Health Care 14 (2006), 123–142
[43] K. Koyanagi, J. Furusho, U. Ryu, A. Inoue, Development of rehabilitation system for the upper limbs in
a NEDO project, Proc. 2003 IEEE Int. Conf. Robotics & Automation, Taipei, Taiwan (2003), 14-19
[44] J.C. Perry, J. Rosen, S. Burns, Upper-limb powered exoskeleton design, IEEE Transactions on
Mechatronics 12 (2007), 408-417
[45] U. Mali, M. Munih, HIFE-haptic interface for finger exercise, IEEE/ASME Transactions on
Mechatronics 11 (2006), 93-102
Basic Engineering for Medics and Biologists 367
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
Subject Index
bending 13 kinetics 323
Bernoulli 45 laminar flow 45
bioimpedance 69 magnetic resonance imaging 302
biopotential 90 magnetic resonance spectroscopy 302
bipedal locomotion 323 measurement 353
brittle 13 mechanics 58
cardiogram 90 medical imaging 249
color Doppler 249 membrane potentials 90
computed tomography 274 moment 3
couple 3 muscle fatigue 343
diagnostic imaging 302 muscle recruitment 343
dielectric permittivity 69 myogram 90
differential equation 58 node 58
digital imaging 274 oculogram 90
Doppler flow measurement 249 orthotics 323
Doppler measurement 45 pressure gradient 45
ductile 13 prevention 81
effective orifice area and regurgitation 45
performance index 45 Reynolds number 45
electric conduction 69 rigid body dynamics 27
electrical injury 81 robot 353
electrical safety 81 shock 81
electrical stimulation parameters 343 static equilibrium 3
electrodes 343 stiffness 13
electrograms 90 stiffness matrix 58
encephalogram 90 stimulators 343
finite element analysis 58 strain 13
first aid 81 stress 13
force 3 tissue characterisation 69
free-body diagram 3 torsion 13
gait analysis 27 turbulent flow 45
haptic interface 353 ultrasound 249
heart valve 45 virtual reality 353
human movement 27 walking 323
internal force 3 X-ray 274
kinematics 323
This page intentionally left blank
Basic Engineering for Medics and Biologists 369
T.C. Lee and P.F. Niederer (Eds.)
IOS Press, 2010
© 2010 The authors and IOS Press. All rights reserved.
Author Index
Bajd, T. 343, 353 Mawson, S. 121
Black, N.D. 121, 140, 172 McCullagh, P.J. 121, 158
Campbell, V.A. 202 McGlade, K. 121
Craig, D. 172 Munih, M. 343, 353
Davies, R. 121, 172 Niederer, P.F. v, 249, 279, 302
Dendorfer, S. 231 Nugent, C.D. 158, 172
Donnelly, M.P. 158, 172 O’Brien, F.J. 187, 214
Dumont, K. 45 O’Connell, B. 202
Finlay, D.D. 158, 172 Partap, S. 187
Hallberg, J. 172 Plunkett, N. 214
Hammer, J. 231 Reilly, R.B. 90, 109
Hazenberg, J.G. 58 Schmid, J. 58
Jossinet, J. 69, 81 Sharma, P.K. 13
Koopman, H.F.J.M. 27 Verdonck, P. 45
Lee, T.C. v, 3, 58, 90, 109 Verkerke, G.J. 3, 58
Lenich, A. 231 Wang, H. 140, 158
Lyons, F. 187 Winder, J. 140
Matjačić, Z. 323 Zheng, H. 121, 140
This page intentionally left blank
This page intentionally left blank
This page intentionally left blank