Robot Sensors and Transducers
Robot Sensors and Transducers
Edited by
Titles in preparation
HALSTED PRESS
John Wiley & Sons
New York - Toronto
and
OPEN UNIVERSITY PRESS
Milton Keynes
Open University Press
Open University Educational Enterprises Limited
12 Cofferidge Close
Stony Stratford
Milton Keynes MKll 1BY, England
First Published 1987
Copyright 1987 S. R. Ruocco
Ruocco, S. R.
Robot sensors and transducers.
Chapter 1 Introduction 1
1.1 Historical notes 1
1.2 Definitions of sensor and transducer 2
1.3 Generalities 2
PART I TRANSDUCERS
PART II SENSORS
controlled robot axes, and for 'external' sensing to relate the robot to its
task. A thorough and systematic presentation of the basic physical principles
used in different single-cell devices leads the student to understand how the
same principles underlie the more complex operation of linear and area
array devices. The final section outlines standard methods of processing
array data to produce reliable control signals for robot and other machines.
The book will have earned its place in the Open University Press
Robotics Series if it achieves one thing: better training of robot engineers
not just in what the sensing systems available to them do, but in how they
work and thus how they may best be matched to future applications.
P. G. Davey
Preface
The requirements for the book were generated, initially, by the Middlesex
Polytechnic decision to run on an Industrial Robotics course for the
Manpower Services Commission based on the B. Tech. Industrial Robotics
syllabus which includes approximately 25% of it on sensors. An added
requirement for the book came from the need to integrate Robotics, rapidly
becoming a subject area in its own right, into the curriculum of both the full
time and part time Electronics degree courses.
The lack, at the time, of suitable books covering the required area of
interest was a prime motive behind the decision to write extensive lecture
notes which, after integration with the experience gained during the years of
industrial employment, were edited into a format suitable for publishing.
The book is aimed at the student of an electronic discipline studying the
traditional subjects of Systems, Control and Instrumentation as well as the
more recently established one of Robotics. The level catered for is equivalent
to the 2nd and 3rd year of a degree course, but since no prior knowledge of the
principles involved is assumed the book is thought to be suitable for Diploma
students as well.
A feature of this book is that it covers only the sensors and transducers
necessary for computer interfacing as related to robot control. It therefore
differs from similar books by omitting the sections on fluid-mechanical
(density, flow, humidity and level), nuclear, temperature and chemical
transducers and concentrating on the recent research and developments in the
sensors and transducers suitable for interfacing to the 2nd and 3rd generation
of robots. It also differs from other books on robot technology by omitting the
sections on robot characteristics in favour of a section on Image Processing,
thus providing a broad picture of modern sensing equipment and its
applications in the modern field of Robotics.
Acknowledgements
The author wishes to thank all who have contributed to the compilation and
the successful completion of this book, in particular my wife Patricia for her
unfailing support and my Middlesex Polytechnic colleagues in the faculty of
Engineering, Science and Mathematics for their constructive contributions.
The author wishes to credit the following sources for some of the
figures, tables and other material used in this book; a reminder of each
contribution is also included in the appropriate parts of the text: Academic
Press Inc. (London) Ltd., 24/26 Oval Road, London NW1, UK; Apple User
and Apple Computer Inc., 10260 Bandley Drive, Cupertino, CA, 95014,
USA; Drews, S. P. et aI., Institut fur Prozesstenerung in der Schweissteck-
nik, RWTH Achen, Rentershagweg 4, D-5100, Aachen, Federal Republic of
Germany; Electronic Automation, Haworth House, 202 High Street, Hull,
HU1 1HA, UK; Elsevier Science Publ. B.V., P.O. Box 1991, 1000 BZ,
Amsterdam, The Netherlands; Fairchild Semiconductors Ltd., 230 High
Street, Potters Bar, Herts, UK; Ferranti Electronics Ltd., Fields New Road,
Chadderton, Oldham, OL9 8NP, UK; General Electric Semiconductors,
Belgrave House, Basing view, Basingstoke, Hants, RG21 2YS, UK;
Hewlett-Packard (UK) Ltd., Nine Mile Ride, Easthampstead, Wokingham,
Berks, RGll 3LL, UK; Honeywell-Visitronics, P.O. Box 5077, Englewood,
CO, USA; Joyce-Loebl, Marquis way, Team valley, Gateshead, Tyne and
Wear, NEll OQW, UK; Kanade, T. and Somner, T., Carnegie-Mellon
University, Pittsburgh, PENN USA; PERA (Production Eng. Res. Assoc.),
Melton Mowbray, Leicestershire, LE13 OPB, UK; Polaroid (UK) Ltd.,
Ashley Road, St. Albans, Herts, ALI 5PR, UK; Ramakant Nevatia,
'Machine perception', 1982, pp. 159-163, Adapted by permission of
Prentice-Hall Inc., Englewood Cliffs, NJ, USA; Robertson, B. and Walk-
den, A., GEC-Hirst Research Labs., East End Lane, Wembley, Middlesex,
UK; Rosenfeld, A. et al., University of Maryland, College Park, Washin-
gton D.C., USA; SPIE, The Int. Soc. of Opt. Engineers, P.O. Box 10,
Bellingham, Washington, USA; Texas Instruments Ltd., Manton Lane,
Bedford, MK41 7PA, UK; West, G. A. W. and Hill, W. J., City University,
Northampton Square, London EC1, UK.
Chapter 1
Introduction
1.3 Generalities
100~ - - - - - - - - -
75 a
Output 50
(% FSO)
Linearity error
+ 8.6%-3.4% FSO
25
25 50 75 100
Measurand (% range)
100 - - - - - - - - - --.;Yr
/ I
Output Term
../
/ / I
I
(%FSO) line /
/ Theoretical I
/ slope I
// I
100
Measurand (% range)
Figure 1.2
100 - - -----,
Output
(%FSO)
Figure 1.3
Introduction 5
Output
(%FSO)
Measurand (% range)
Figure 1.4
100 ----------------,
I
Output
(%FSO)
,... ,...
,... /' Least squares
v"" line
100 ----------------,
o X~
o x
Run l O X
ox
Ox
I 0 Ox X
Output
+
o
0
X
X Run 2
(% FSQ) o X
o
o
o X
X ~16.6% FSO
Precision within
o X
oX
oX
o X
~0 X
100 --------------------
Output
(%FSO)
100 --------------------
Output
(% FSO)
_ _ _L- Resolution
- - - - , 2.5% FSO
a Measurand (% range)
Figure 1.8
100~--------------~~
/
Percentage
amplitude
ratio /
/
~ndPass
~~aracteristic
/
Measurand frequency (Hz)
Figure 1.9
100
90
I
I
I
I
Input I Output
Amplitude
(%FSO)
I
I
I
I
-
10
- -
I
~ - --11- - -
I
-----i - -
0
Time(s)
Figure 1.10
Introduction 9
step change in the input measurand value. The time required for the
output to reach a specified percentage of the final value is termed the
'response time'. These specified percentages are usually 10% and 90% of
the final value.
As shown in Figure 1.10 there are four main types of response time:
Delay time-the time taken for the output to rise to 10% of the final
value once the input step has been applied.
Rise time-the time taken for the output to rise from 10% to 90% of the
final value.
Storage time-the time taken for the output to fall to 90% of the final
value once the input step has been removed.
Fall time-the time taken for the output to fall from 90% to 10% of the
final value.
Other important elements of the output characteristic are noise and noise
margins, which can be defined as:
Noise-the level of any spurious signal(s) appearing at the output of the
device due to any cause other than the driving input physical quantity.
Noise margin-the maximum noise level that can be tolerated by the device
before it has any significant effect on the output signal.
PART I
TRANSDUCERS
Chapter 2
Position transducers
2.1 Overview
The control structure of a robot needs to know the position of each joint in
order to calculate the position of the end effector (e.g. gripper) thus
enabling the successful completion of the programmed task. The movements
of the joints can be angular and/or linear depending on the type of robot;
they are illustrated in Figure 2.1 for each of the four main robot coordinate
systems. A suitable algorithm for calculating the end effector position in any
of the coordinate systems can then easily be deduced, as shown in Figure 2.2
for the polar coordinates type robot, thus permitting the development of the
necessary robot control software.
It should be noted that future developments in the robotics field may
one day allow the use of external sensory feedback, or exteroceptors, as
End effector
position
(X e Y e Ze)
I
I
'P: Y
--1- 7 --"
I /
,/ e ..... I
............... ""-:...Y1/
_______
//
,L.----
,/
,/ Xe = L2COS'P.cose
/x L, Y e = L2cos'P.sine
Ze = L2sintp + L,
Coordinates
reference point
(X o Yo. Zo)
Figure 2.2 Schematic representation of polar coordinates robot and relative end
effector position formulae
USER
INTERACTION
ROBOT
r+ CONTROLLER ....... ROBOT
ACTUATORS r--+ END
ROBOT
~
EFFECTOR
~ lI
PROPRIOCEPTIVE I
FEEDBACK I
DEVICES I
(e.g. position
transducers) I
I
I
I I
EXTEROCEPTIVE I I
f.--j--~ __ J
FEEDBACK
DEVICES
(e.g. vision sensory In formation
tactile sensors)
Figure 2.3 Block diagram of second generation robot system, showing the use of
exteroceptive feedback
14 Robot sensors and transducers
operates, is still largely at the research stages and may not be implemented
generally on commercial machines for some time.
Internal position transducers (proprioceptors) therefore remain, at least
for the present, the most accurate and reliable way of determining the end
effector position within a robot control structure. There are two main types
of position transducers: absolute and incremental (or 'relative'). The
absolute position transducers are themselves further subdivided into resistive
and optical types, as illustrated in Table 2.1. Other types of position
transducer are also employed in industry, such as differential transformers
and synchros, but their larger size and/or higher cost have precluded their
use within robot structures and will not therefore be discussed here.
Device
Type Class description
Absolute Resistive Potentiometer
Optical Coded optical
encoder disks
Incremental Optical Slotted optical
encoder disks
2.2 Potentiometers
2
2
Physical Schematic
construction representation
r::::;::o 12 3
(2.1)
(2.2)
(Given that the supply voltage v.,the total resistance value R tou the total
angular displacement 8 to t. and the total linear displacement X tot . are known
constants. )
The resistive track can be made of a carbon film to reduce cost, or a
cermet (conductive ceramic) film to increase the resolution and reduce
noise, or a wound metal wire to allow higher power dissipation. The main
advantages of potentiometers as position transducers are low cost, small size
and versatility of operation (for example, it can easily provide logarithmic or
quadratic functions), whereas the two main drawbacks are, firstly, that it is
an inherently analogue device and therefore requires additional hardware,
usually an analogue-to-digital converter (ADC), to interface it to a computer
and, secondly, that its principle of operation requires the wiper to actually
touch the main resistive medium, which makes it prone to mechanical wear
and tends to limit its operational life.
Optical transducers are therefore slowly replacing the potentiometer in
those position measuring applications which require very long operational
life. Nevertheless the potentiometer is still widely used as a position
transducer in most engineering applications in view of its ease of operation
and versatility, small size and low cost.
An example of the potentiometer's popularity can be found in its use
within the control structure of some educational and light industrial robots.
The potentiometer's low cost and small size, in fact, allows full exploitation
of the high torque-weight ratio of permanent-magnet d.c. motors, which has
led some manufacturers to produce integral, lightweight and powerful pulse
position (PWP) controlled d.c. servo units. For further details on such a
device see Section 2.4 on Interfacing.
16 Robot sensors and transducers
JUUUl
V OUT
Light
source
opaque line (or clear slot if a metal disk is used) which then provides the
detection of each complete revolution, as required. The output signal from
this third channel is in fact a pulse train whose period is equivalent to one
complete shaft revolution; each pulse can be used to 'zero' the angular
position count as well as providing the signal for a shaft revolution count, as
shown in Figure 2.16 in the section on interfacing.
One drawback of the incremental encoder systems thus far described is
their poor angular resolution, ~a. This depends on the number n, of opaque
lines on the glass disk (or the number of transparent slots on a metal disk),
as shown in eqn (2.3), and the width Wp of the photodiode active area Wp on
the disk plane, which in turn depends on the photodiode mounting distance
from the disk centre, as illustrated in eqn (2.4) and Figure 2.8.
Note that, since most photodiodes have a square active area, the
required dimension Wp is simply the square root of the active area given in
Photodiode
Reference
slot
Photodiodes
10mm
360
~a= (2.3)
n
.
Wp = r SIn (~a) .
2 = r SIn (360)
----;: (2.4)
(HAhNH ,.
~~....
C..AN"-Iu'. " I, VoA
.J!.) QftOVNO
-l!.. 00:0.
I
\ Ii ,-.~.,.
cor. <0'
CH..,N [\.11111 V
I
I
I
CHA,IIIW(ll
I
10
..
~-_~_::'_
----=
_=_=_
==_=_=_==_=_=_-_-_-_..L::
-l ' G 0
{Mil TlA (hO "LAU. COOl "'"HL [NCODE" IIOov AO'ATIOH
100
80
60
.mmm"""'IIIIIIIIIIII"""",,,,,mmHH
MOIRE FRINGES AT 3% SCALE DIFFERENCE
~m"'II'I'm"~II""'IIIII'm~'''ffllll'''ml
" 1.5 " SCALE FACTOR SENSOR CIRCUIT OUTPUT
80
80
FIXED GRATING
::: I:: r: I:: It: ::::::11 .... ,' ...........'...........'.\.\\\\ \ .. ,,, ... \ .. , t\ \, 1t:: t 11
Figure 2.11 Optical position transducer based on Moire fringe pattern (courtesy of
General Electric)
12 :3
11 4
B 7
Figure 2.12 Absolute optical encoder based on a 4-bit Gray code (after McCloy
and Harris, courtesy of Open University Press)
15 0
12 3
11 4
8 7
Figure 2.13 Absolute optical encoder based on a 4-bit binary code
To computer
,-------,
c ADC
Potentiometer
however, there is no real need for hardware interfacing between the optical
absolute transducer and the computer since any code conversion is usually
done in software using a look-up table stored in ROM.
To interface the other absolute position transducer, namely the
potentiometer, is also straightforward. As shown previously in Section 2.2,
the output of this transducer is a voltage which represents the robot arm
absolute position, which means it can be interfaced to the computer with a
simple ADC, as shown in Figure 2.14.
An alternative to this technique can be found in small robots such as
those encountered in the educational market and other distributed control
systems such as remotely controlled model aeroplanes. The potentiometer in
this case is integrated with the actuator, a small permanent-magnet d.c.
motor, and some interface electronics to produce an integral, lightweight
d.c. servo unit. The principle of operation of such a device is illustrated in
Figure 2.15.
In these units the potentiometer provides the resistance value for a
monostable RC charging constant; in other words, the monostable pulse
output depends on the potentiometer shaft position. The monostable output
is compared with the input signal provided by the microcomputer controller
and an error signal is generated. The width of this error signal determines
the length of the motor shaft rotation, that is the size of the adjustment
necessary to maintain the shaft position under computer control.
This interface technique allows the computer to 'delegate' the control of
the robot arm to these servo units therefore freeing it for other tasks, and is
eminently suitable for use in robot systems with limited computer power
such as microcomputer-controlled robots. However, the computer does not
have a direct measurement of the robot arm position so that, in robot
control terms, this is an open-loop system which is undesirable when
Position transducers 23
1---------------1
Input ..r-t..
pulse
--+-i~
(P.W.P. signal
Comparator A JL....n..
I Error
from micro
controller)
I signal
I Feedback
pulse
I .s-L..
I
I
Mono
I stable
I
I Potentiometer
I I
L ________________ J
Figure 2.15 Pulse width position (PWP) controlled D.C. servo unit
>-<.-----1 ClK
Count
channel Counter
(angular f-"""?"'o::--,
position)
.....----'-'---1 Reset
Single rev.
reference
channel
~~-----I ClK
Counter DATA
(No. of 1-----'
Incremental Revs)
encoder disk
Reset Clear
input
from compo
Figure 2.16 Example of computer interface circuit for optical incremental trans-
ducer system
interface circuit requires a form of local memory (so that the transducer
output data can be held until the computer needs it) and a sequence
detection circuit (to determine the direction of rotation by comparison of the
two transducer outputS). An example of such a circuit is shown in Figure
2.17.
This is based on the high-resolution Moire fringe type of optical encoder
disk which provides directional movement information by virtue of the phase
relationship of Yt and V2 as discussed previously in Section 2.3.1; note that,
for clarity's sake, the 'number of revolutions' counter and associated
components have been omitted from Figure 2.17.
The measurement resolution of the position transducer thus described
depends on the line width of the Moire fringe pattern which can be obtained
from the manufacturer data sheet; it should be noted, however, that the
counter hardware must match this high resolution in terms of position count
and may therefore need to be bigger than the 'standard' 8-bit width.
2.5 Conclusions
Sequence Strobe
detection 1 - - - - - ,
circuit
Up/down
Load
To
computer
Clock Counter
(angular latch
position)
Reset
Clear
Reset
Incremental
encoder disk
the most popular posItion transducer in the field of machine control and
robotics, replacing the potentiometer in many applications requiring longer
operational life.
Optical absolute position transducers have been restricted in their use to
high reliability, non-space-critical applications in view of their larger size and
cost.
Light transducers
3.1 Overview
Designation of
Wavelength range radiation
10
~~ -~ ---- ------'
UIoJIU 11C'411C'lh.ICS
L'ln~tl('\.'llh.JI trSl.lllalhlllS
iii
10-- 10 10 ' 0 ' 10 ' 10 fro 10' 10'" 10 I 10 1 10 I I 10 10J 10' 10 10' 10" 10 10" 10'011 10'11 10" 10 1 : 10'4 10 1... t).lllll
l
Conduction band ~Electron
E, -;----:..::-- E, t } B,ndg,p
Photoeffect
enabled
(E).> Egl
Valence ban~ole
Ec----- Photoeffect
not enabled
(E).< <Egl
3.2 Photoresistors
10- 2
1.0 10 10 2 103 104
Cell resistance (ohms)
These are photovoltaic devices based on the junction photoeffect, that is the
creation of optically induced electron-hole pairs within a P-N junction. It is
therefore useful at this point to retrace some of the basic steps of
semiconductor physics.
Electron-hole pairs are generated by thermal agitation throughout a
semiconductor material and are, by their nature, in continuous random
movement. Upon formation of a P-N junction (e.g. by diffusion of an
element of opposite type to that of the bulk dopant) some of the mobile
carriers from one side cross the junction and recombine with some of the
mobile carriers on the other side (i.e. electrons from the N region recombine
with holes from the P region) leaving behind fixed positive and negative
ions. The process continues and this thin region either side of their junction,
now depleted of mobile carriers (and thus termed 'depletion' region), widens
until the electrical potential created by the fixed ions is sufficient to prevent
Light transducers 31
any further diffusion (i.e. junction crossing) by the mobile carriers. The
junction is now in thermal equilibrium and has acquired a so called 'barrier
potential' (this however cannot be measured since it is cancelled out by the
external connections contact potential). If the material is now irradiated, any
electron-hole pairs produced by the internal photoelectric effect within the
depletion region are swept across the junction by the electric field associated
with the barrier potential thereby producing a current flow, as illustrated in
Figure 3.4.
This injection of optically induced majority carriers therefore reduces
the barrier potential and, since the connections contact potential is still the
same, it produces a measurable voltage across the junction. The device thus
created is therefore an electrical generator capable, as shown in Figure 3.5,
of driving a current through a suitably connected external load, with the P
terminal becoming positive with respect to the N terminal.
,-" .. r _
I
I
I I o I~""
I I
I L_~ _ _ __ f '"""
t-----+-~
I
I
I
'orl4~l_r'"
I .~
l - '(tt, ' I
"1"
"'f~I If'\f.
~.,.t 4' ,
___________...,1__
I lh
~
"
"
I '
I
\'\ - Rl
I
Vo -
C Ro
RL
1
Vo
I
+
IL j I
Ie lOlL
Figure 3.S Photocell diagram and equivalent circuit (courtesy of Ferranti)
The photocurrent which has been caused to flow across the junction
(Le. the rate of change of the majority carriers crossing it) is approximately
proportional to the rate at which the light quanta impinge on the
photovoltaic device and therefore it increases with the intensity of the
illumination. This is shown diagrammatically in Figure 3.6 under different
lighting conditions.
When the device has no load connected to it the external photocurrent
must, naturally, be zero and therefore all the current flows through the
diode in the forward direction. This produces an 'open circuit voltage' v"c
which is a logarithmic function of the photocurrent Ip and has an upper limit
of 0.6 volts for silicon devices (a photovoltaic device used with a very large
load resistor will therefore act as a logarithmic voltage source). When the
device is short circuited, on the other hand, the maximum current that can
flow from the P to the N region is the short-circuit current Isc which is equal
to the photocurrent Ip (a photovoltaic device used with a small load resistor
I I ,
I
I
I
I
voe, (sst) I
,1:.:;::::;;;;;;;;Hr-V'
VQC.L(85C) I
Voc2 (25-C) ,I
I
,
I I
/ I
---
--~
I
,
,,,
,- I
--
(I) DARK CHARACTERISTIC. (b) LIGHT INTENSITY I. (e) LIGHT INTENSITY 2.
-,
'+~
-
~ sttJ,
~
3000LUMENS '>I./Ft
t:=Joo , 1/ 1/ ,,'
,
1/ /
, ,
\
/\
~. /
,
/' \ /
\oJ
,/ \ /
'"~I
--'
o
>
,
.1/
, "
10
10 100 1000
CURRENT (fA)
Optical
transducers
External Internal
photoeffect photoeffect
transducers transducers
+
Junction
Photoconductor photo detector
(photovoltaic)
+
Non
+
amplifying
(vac. photocell)
Amplifying
(photomultiplier)
limage conv. I
Doped
(lR detector)
Intrinsic
Iphotoresistor)
Non
amplifying
(photocell
photodiodel
Amplifying
+
Iphototransistor)
lAval. Photocell)
based on the junction photo effect but which provide a higher current output
for the same illuminance input value. These devices, in other words, have an
amplification function associated with the basic junction photoeffect and, as
shown in Figure 3.8, are known as amplifying photodetectors. The most
common examples of such devices are the photo transistor and the
avalanchephotodiode.
3.3.1 Phototransistors
Photons
Wire to
49
C
Wire to N diffusion
base terminal
E Photo
sensitive N-substrate
Schematic junction
Collector terminal
proportional to the algebraic sum of the bias current IB and the photocurrent
Ip.
Phototransistors are therefore more sensitive than photo diodes but do
have worse output current tolerances and a less linear transfer characteristic
due to the hFE variations with temperature, bias voltage and collector
current. The speed of response of a phototransistor also differs from that of
a photodiode due to the dominant effect of the input R-C time constant on
the output rise and fall times (i.e. the bandwidth of a phototransistor is
governed by the effective 'input' R-C value rather than the actual transistor
switching speed) as well as the effect of the current amplification factor hFE'
as shown in eqn (3.4):
Output time constant = T = hFE . R/N . C/. (3.4)
Phototransistors are therefore mostly used in low cost, medium speed
and medium sensitivity light detection but are slowly being replaced by
photodiode-operational amplifier combinations due to the market availabi-
lity of low cost, high gain operational amplifiers whose performance can be
selected to suit the application.
3.4 Photomultipliers
01 02 03 04 05 06 07 08 A
PK I I I I I I I I
I I I I I I I I
I I I I I I I I Vo
I I I I I I I I
I I I I I I I I la
I I I I I I I I
I I I I I I I I
'pK
-Vs
As well as the basic devices thus described, there are others that embody
more than one elementary transducer in their construction and are thus
referred to as 'composite' or, more specifically, 'array' transducers. These
devices are in fact based on arrays of photosensitive elements known as
photosites, with each array element working on the same principles as the
elementary transducers described in the previous paragraphs. The array can
be two dimensional, and is then known as an area array or area imaging
device (AID) as shown in Figure 3.11, or it can be one dimensional in which
case it is called a linear array or linear imaging device (LID).
PHOTOSENSITIVE AREA
/
\ /
\0 }.
\\ /
/
Viewed ~ Column
scene
[Y-------V scan
! \
/ Optics \
/ \ "-
\ \
]1
\
Photosite
,).-
) \ Amplifiers V
Row scan & control ~
~ OUT
Vacuum optical array transducers are based on a 2-D photosite array, that is
there is no vacuum 1-D equivalent of the solid state linescan camera, and
have a typical construction as shown in Figure 3.12. These devices include
Target - e
---..-.-----
Scan coils
~
Figure 3.U Schematic diagram of vacuum optical array transducer (television
camera)
the television cameras and have traditionally been very popular as optical
transducers even though, as shown later in Section 3.5.3, they are
increasingly being challenged by solid state cameras in the robotics field.
Vacuum television cameras include the recently introduced high sensitivity
Newvicon and Chalnicon as well as the longer established Vidicon and
Plumbicon cameras. These latter two types are by far the most popular and
will therefore be described in more detail in this Section.
The main difference between the various vacuum television cameras lies
in the photosensitive area (usually known as the 'target') which is coated
with lead oxide (PbO) for the Plumbicon, selenium or antimony compounds
(e.g. Sb 2 0 3 ) for the Vidicon and zinc and cadmium tellurides for the
Newvicon and Chalnicon.
The scanning of the photosites is achieved via an electron beam in the
same way as in cathode ray tubes, namely by the use of magnetic deflection
40 Robot sensors and transducers
RL
+ Vs Plumbicon
Figure 3.13 Vidicon and Plumbicon principle of operation (after West and Hill)
as the electron beam scans across the target area it 'closes' the switch S on
each photosite (i.e. it connects it to the cathode which is at ground potential)
thereby charging its equivalent capacitor Ceq to a specific voltage "' . While
the beam is scanning another area of the target (i.e. switch S is open), the
capacitor discharges through its equivalent parallel component and attains a
new, lower voltage V2
As the electron beam scans across the target area again, it 'recharges'
the photosites' capacitors back to "', at a rate Ie proportional to the voltage
difference (both the capacitor value Ceq and the time interval l!.t are in fact
constants), as shown in eqn (3.8) :
dV
Ie = Ceq ' dt = Ceq ' (VI - V2 )/ fl.t = K . (VI - V2 ) (3.8)
Solid state optical array transducers have been derived mainly from the
silicon memories technology, indeed their development has been possible
only because of the great strides that microelectronics has made during the
70s and 80s. As a result solid state cameras have a good geometric stability
within a flexible configuration. Solid state cameras, in fact, unlike vacuum
types, can have either a 2-D photosite array (area scan cameras) or a 1-D
photosite array (linescan cameras), the latter geometric configuration having
the advantage of using a much smaller silicon area thus producing a higher
manufacturing yield and a lower unit cost.
The solid state camera can be based on 4 different photosite arrays:
(i) the Photodiode array
42 Robot sensors and transducers
I
I
I
L ________ ~
Figure 3.14 Principle of operation of photodiode array
J
q = cJ . v.s = ie . dt (3.15)
When the switch S is open the charge is reduced at a rate proportional to the
incident radiation (the photocurrent Ipe) and the thermally generated
hole-electron pairs (the dark current Ide). For modem, high quality P-N
junction (where Ipe Ide) the discharge time constant is of the order of
several seconds for zero incident radiation and in the milliseconds region for
typical irradiation levels (Hill, 1982): the charge stored within each photosite
P-N junction is therefore inversely proportional to the incident irradiation
level. Monitoring the rate of change of this charge (i.e. the capacitor current
ie) while scanning the photosites provides a voltage Yout proportional to the
Light transducers 43
grey scale image of the scene focussed on the array, as required. This
technology favours uniformity and stability of response but does allow a
lower array resolution than other solid state cameras in view of the more
complex 'on chip' support electronics required.
The most common geometry for industrial photodiode arrays is in fact
the linear array (or linescan) camera. This is due, as previously mentioned,
to the technological difficulties involved with making large 2-D photodiode
arrays and their relative high cost.
Linescan cameras offer a simple, low cost method of detecting edge
position which, in turn, can be used to measure object size parameters such
as width or length (IPL, 1984; Electronic Automation, 1985; Hollingum,
1984; Drews et aI., 1986), as shown in Figure 3.15 (for further information
on how to separate the object image from the background and obtain its
edges positions, please refer to Chapter 8 on image processing) .
V OO1
V
(OBJECT') I - - --
1=-:--tF~m
VT.. c - - - - - - ----
lens Object V 2 I -_ _~"-"I~
Photosites (BACKGROUND) I
array J i I i I
linescan Camera 10 20 Photosite 30
(a) (b) position
Height
measurement
Linescan
Figure 3.16 (a) Example of vision sensor based on a linear array solid state
transducer (after Drews et al., courtesy of Robotics); (b) Non-contact inspection in
the food processing industry (courtesy of Electronic Automation)
(a)
__ _ J-, __~~~;t~E~(!~I'1
OV +V OV OV +V OV
~-&h
\@J
'- -,
~_ ~ () J 'eee'
\ - - -J - - Silicon ~ubs~rate
'-'- -- -- --- (p-type In this case)
(b)
'"
...
AD , II cs
PIXEL ..
'7 OS
'i .,.
14 0,.
I
- 1lA 13 ...
TP3 ~ TP'
Pin"""": CD 00
0 AD
CD
PG
.... ...
"""'ogo'.
Tranoft< Clock
A4,.. Drain (after We t and Hill 19 2)' (b) CCDlll line can camera (cour-
TP THt PoInt
v.. Substrat. (G'ound) te y of Fairchild, 19 4)
M o +v 0 0 v 0 o 0 0
~ \~~~:Ito<
~~
---, ee8r-- - - --..888,--- --- --
~~
eee
---
\ ()
~ _ ()
__ ()I / '\:: - =.!
I +\ ~ Substrate
Figure 3.18 Charge Injection Devices principle of operation (after West and Hill)
Since the MOS capacitor value C and the time interval !:!.t (necessary to
take the electrode voltage from V to 0) are both constants the height of this
'injected' photocurrent pulse is proportional to the stored charge q and
Light transducers 47
therefore to the incident irradiation E in :
(3.17)
By monitoring the displacement current peak value during the photo sites
scan an analogue signal Vout proportional to the image intensity is obtained,
as required. CID solid state cameras have similar features to CCD ones,
being based on the same MOS technology, namely stable and accurate
geometric configuration with a good input dynamic range and a high unit
cost. CID cameras are, however, more tolerant of excess light input levels
and suffer from less cross-talk than CCD cameras because of the absence of
the charge transfer mechanism.
(a) DRAM fully encapsulated. (b) DRAM chip exposed after lid
removal.
fact, aims to reduce the cell size, the stray capacitances and the signal path
lengths in order to obtain the highest cell density with the lowest access
time.
This may lead to a physical cell layout unsuitable for use as a light
transducer and therefore not all DRAM chips available on the market are
suitable for use as a solid state camera. Even those that have been
commercially used require some image pre-processing in order to produce
an acceptable image, such as the removal of the black line caused by the
memory cells power supply rail. Two other drawbacks of these devices are
the limited input dynamic range (and the relative long exposure times
required) and the inherent binary nature of the image (since each cell output
can only be a logic 1 or 0).
A grey scale image using this device can, however, be produced at the
expense of image acquisition time by overlaying several binary images taken
with different exposure times. DRAM cameras are inexpensive in comparison
with other optical array transducers because the optical array transducers
used (the DRAM memory chip) is manufactured in large numbers for the
computing industry. This feature has produced wide acceptance of the
DRAM camera in low cost vision systems, such as those used on educational
robots and in light industrial automation (Lambert, 1984; Owen, 1982).
" ~ .J..
X\ '-/'~/'.\
100
V IdIcon I " \ .
\1 / A', \., \\
1/ \ / . '.
f)',Ii ~ \
75
>
.....
Newvicon
.
1/ I .,/ \
;; \
.':;
.iii
c 50 I
Ql
Cfl
Ql 1.1 \ , I \
>
\
.':;
ro
Qi Sensitivity \~ I
a:
25
curve of
human eye ~, \
540/Green
~ " "
(a) Vacuum optical array transducers
[Courtesy JOYCE LOEBL, 19851
465/Blue 610/Red "
0 .......
300 400 500 600 700 800
Wavelength (n m)
(b) 3.5
3.0
2.5
E
u
J
:::.
:> 2.0
>
.....
;;
.iii
c
8. 1.5
Cfl
Ql
a:
1.0
the case of imaging objects moving on a conveyer belt where strobed I.R.
illumination may need to be used in order to avoid creating unpleasant
conditions for factory workers (see Section 6.2).
Solid state cameras have a very low image lag and can therefore be used
in higher speed applications than vacuum cameras. They also allow a greater
flexibility in vision system design since solid state cameras come in different
geometries (1-0 and 2-0 array devices are available) and with different
resolutions (64 to 4096 pixels is a typically resolution range for linescan
cameras). These features can be used to help improve specific vision system
parameters, such as system cost or image processing complexity and speed,
as demonstrated in a food processing plant where the use of medium
resolution, linescan cameras enables the high speed inspection of corn cobs
(Hollingum, 1984).
The smaller size and more rugged construction of solid state cameras,
afforded by the absence of a vacuum envelope, also allows a wider range of
applications compared to vacuum cameras, such as in the case of eye-in-
hand vision sensors (Pugh, 1982; Loughlin and Hudson, 1982; Van de Stadt,
1985).
Solid state cameras have a higher sensitivity than the Vidicon but,
typically, a lower one than the Newvicon and Chalnicon. Solid state cameras
do, however, have a less homogeneous sensitivity across the photosite due to
the photosite physical size tolerances. A typical CCO photosite is shown in
Figure 3.21; its dimensional tolerances are generated mainly by inaccuracies
in the electron beam lithographic masks used during the device manufacture
and yield a variable photosite active area which, in turn, produce variations
in the individual photosite sensitivity: in the CCO case, for instance, the
worst case photosite size error is 0.065 J.lm for 1024 pixels at 13 J.lm pitch
(Fairchild, 1984) which, as shown by eqn 3.18, produces a photo site
sensitivity error !l.Sp of 0.5%:
!l.Sp = !l.Lp.
L
100 = 0.065.
13 100 =
05% ( )
3.18
p
Photo Serpentine
element channel
stop
Figure 3.21 Typical photoelement (or photosite) dimensions for the CCDll1
linescan camera (courtesy of Fairchild, 1984)
Light transducers 51
changed by adjusting the target voltage (other devices have a fixed
sensitivity) and a higher dynamic range; this latter, however, is afforded
largely by the gamma factor non-linearity and the lower sensitivity and might
not always be an advantage.
Electrical noise tends to be higher in vacuum devices because of the
thermoionic, and therefore random, nature of the electron emission from
the vacuum cameras cathode. Some noise is, however, also generated within
CCD cameras as a consequence of the cross-talk that takes place between
contigous photosites and during the charge transfers.
Ultimately cost may also playa part in the device selection even though
the camera cost is typically only a small proportion of the overall vision
system cost. The Vidicon is generally the cheapest area scan camera on the
market (with the exception of DRAM cameras which however have a worse
performance and do not find large use in the industrial robotics field) due
mainly to the large number of devices manufactured for the television
market, whereas the Plumbicon tends to be the most expensive device
among those described. It should be noted that solid state linescan cameras
are, in fact, cheaper than Vidicons but that a direct price comparison is
somewhat meaningless in view of their different array geometries and
specifications.
3.6 Conclusions
Optical transducers are based on the photoelectric effect and vary in cost
and size from the subminiature photodiode to the wide angle photomul-
tiplier. The most popular optical transducers in the robotic field are usually
spectrally matched Light Emitting Diode/Photodiode combinations, used
mostly in proximity detectors, and the solid state array cameras, used for
scene imaging and subsequent processing.
(a) The circuit shown in Figure 3.a uses the photoresistor whose transfer
characteristic was shown in Figure 3.3. Calculate a suitable set of
r-----~-----.--_.---.----~~-- + 1 2 v
Figure 3.a
52 Robot sensors and transducers
~
Figure 3.b
(c) Carry out a survey on the use of solid state cameras in robotics
application of your choice. What conclusions can you draw from this
mini-survey?
Force transducers
4.1 Overview
Force
transducer ~F
'%W~SUPPORT STRUCTURE ~
-~
Figure 4.1 Direct measurement, force deforms transducer only
Figure 4.2 Indirect measurement, force deforms both support and transducer
The unit change in the transducer parameter value (L\ ~/~) is the quantity
usually measured. The force applied can be obtained if the supporting
structure cross-sectional area A, its Young's modulus E and the transducer's
sensitivity Gf are known, as shown by eqn (4.1), Young's modulus being the
ratio E = (stress/strain)
F=EA (~~) (4.1)
Gf V.
The unit change in the transducer parameter value (L\ ~/~) can be measured
using a variety of techniques, some of which will be described later in this
chapter.
Transducer
+~L~V/
Figure 4.4 Bending force measurement principle
Mf = (~L)EW = EW(~~)
Gf V. (4.2)
56 Robot sensors and transducers
Fixed point F
(joint brakes
are on)
o
limb subject
to bending
force
Figure 4.5 Example of robot limb subject to bending force caused by weight of
forearm and gripper
This force, as the name suggests and Figure 4.6 illustrates, is present in all
rotating structures, such as the drive shaft of a robot actuator. The
measurement of the twisting force on such a shaft may, for instance, be
required to monitor the mechanical resistance to the shaft rotation. The
twisting moment Mt can be derived as shown in eqn (4.3):
M=
t
(4Srlp)
dGf
(~ ~)
~'
(4.3)
where St is the tangential stiffness of the shaft, d is the shaft diameter and Ip
is the moment of polar inertia of the shaft cross-section which, for a
cylindrical shaft, is given by 0.1 d4; (~~/~) still represents the measurement
of the unit change in the transducer parameter value.
and have therefore been included in this chapter. Resistive and semiconduc-
tor transducers are generally referred to as strain gauges.
Electrical
connections
Etched
Insulating metal
backing pattern
The gauge factor of metal strain gauges is given by eqn (4.5), where fl is
the Poisson ratio, which relates the increase in the length of a solid object
under the effect of a traction force to the reduction of its diameter.
dR/R
Gf(metal) = dL/ L = 1 + 21J (4.5)
Table 4.1 lists typical gauge factors for metal strain gauges.'
58 Robot sensors and transducers
\ ---
Doped Electrical
Si or Ge
,,--
Resistivity
Material (Qcm)
(4.7)
relationship:
Rx Rz
R4 R3
Any vanatlOns in Rx subsequently produce a non-zero output voltage Vo
proportional to the resistance change !:l.Rx; the value of Vo therefore provides
a measure of the force acting on the structure under test.
A practical bridge circuit is composed of four equal resistances, that is
the three fixed resistances are chosen to be of the same value as the strain
gauge unstressed resistance, plus a bridge balancing network. This latter can
be of three basic types, as shown in Figure 4.10. For these types of circuit
where Rx(unstressed) = Rl = Rz = R 3 , the output voltage variation !:l. Va is
(a) Using a
single variable
resistance R4
given by:
~v, = V.~R V. ~R
(4.8)
o (4R + 2~R) 2R (2+ ~RIR)
This shows that the ~ Vol ~R relationship is only linear for small resistance
changes, that is small ~R/R. This is in fact the case in practice and we can
therefore approximate eqn (4.8) to:
~ V, = v. . ~R = V. . Gf ~L (4.9)
o 4 R 4 L
where L is the unstressed length of the strain gauge and ~L its increment
under stress. The ratio ~L/ L is a measurement of the mechanical strain and
can therefore be inserted in Young's modulus formula E = stress/strain:
S . E~L
tress = E . stram = L (4.10)
The level of the required mechanical stress on the structure under test can
thus be measured by the out-of-balance output voltage ~ Vo, given that the
bridge supply voltage v.,
Young's modulus for the structure material, E and
the force transducer gauge factor, Gf are known. To reduce the Wheatstone
bridge temperature dependence a second strain gauge Rx(comp) can be
inserted in the circuit in place of R2 and mounted on the structure
perpendicular to the first gauge so as to minimize any longitudinal
elongation, as shown in Figure 4.11.
~F
Figure 4.12 Example of traction force measured using four strain gauge in a
dual-active temperature-compensated configuration
Y:Gt (t::..L)
t::..Vo=-2- L (1+!l) (4.12)
RIGID BODY
Figure 4.13 Example of bending force measured using four strain gauges in a
dual-acting temperature-compensated configuration
case all the gauges are 'active', that is they all receive dimensional variations
along their length and there the overall bridge sensitivity is four times that
of the single-gauge uncompensated circuit shown in Figure 4.9. Also it
should be pointed out that any bending moments acting on the structure will
introduce an error since the mounting configuration shown does not
compensate for them.
RIGID BODY
Figure 4.14 Example of twisting force measured using four strain gauges in a
temperature-compensated configuration
64 Robot sensors and transducers
The most widely used non-resistive force transducer is the capacitive type
which proeuces a change in capacitance proportional to the pressure exerted
on the plates. The principle of operation is similar to that of a compression
capacitor used to provide small values of variable capacitance in electrical
circuits. This principle is illustrated in Figure 4.15. The variation in
Plate 1 0--------.
~
DIELECTRIC _m~mmm~[1d
Plate 2 0 - - - - - - '
Figure 4.15 Simplified diagram of a compression-type variable capacitor
-
To robot arm
Figure 4.17 Example of a load cell using strain gauges to measure wrist forces and
torques (after McCloy and Harris, courtesy of Open University press, 1986)
Vs + 12V + 5V
lMrl
120rl
lK{l
Robot
controller
- 12V highway
Figure 4.18 Typical low-cost computer interface circuit for strain gauge bridge
Force transducers 67
its connector are machined from a single block for stability and are available
from several sources, such as the Lords Corporation in the USA and Meta
Machines in the UK.
4.7 Conclusions
Strain gauges are the most widely used force transducer, in both the metal
foil (resistive) and semiconductor (piezoresistive) types, the latter having a
higher sensitivity but lower dynamic and temperature ranges. The Wheat-
stone bridge is the most effective technique for measuring the transducer's
resistance change under the effect of the mechanical strain and therefore the
force involved.
(a) Show the physical configuration of a strain gauge used to measure the
twisting force on the structure shown in Figure 4.a.
(b) Figure 4.b shows two strain gauges used to measure twisting force. Show
their electrical connections within a bridge circuit such as to reduce the
effects of bending moments. (You can ignore temperature variations and
axial movement.)
(c) The four gauges in Figure 4.12 are used to measure traction force; the
electrical connections are as shown in Figure 4.13. The bridge operates
68 Robot sensors and transducers
Structure
under test (Plan view)
Figure 4.a Figure 4.b
1. Transducers, Sensors and Detectors, R. G. Seippel, Reston Publishing Co. Inc., 1983.
2. Transducers and Interfacing, B. R. Bannister and D. G. Whitehead, Van Nostrand
Reinhold (UK) Co. Ltd., 1986.
3. Robotics-An Introduction, D. McCloy and M. Harris, Open University Press, 1986.
4. Industrial Robots: Computer interfacing and control, W. E. Snyder, Prentice-Hall Inc.,
1985.
5. Industrial Robots: Design, operation and application, T. W. Polet, Glentop Publishers
Ltd., 1985.
Chapter 5
Velocity transducers
5.1 Overview
da
V=- (5.1)
dt
5.2 Tachogenerator
R=oo
20
R = 10k
18 R = 3, 5k
16
R = 1K
14
12
v tach . 10
8
6
4
2
(rev/min)
0 600 1200 1800 2400 3000
(5.2)
where B is the magnetic flux density, 'I' the length of the active portion of
the conductor linking the flux, w is the relative velocity between the
conductor and the magnetic field and 10+ 8 is the number of lines a single
loop must link per second in order to induce a voltage of 1 V.
In the tachogenerator case the magnetic flux density B and the loop
length I are constants, so the output voltage Vo is proportional to the shaft
angular velocity W, as required and a shown previously in eqn (5.1).
This phenomena is therefore the same which gives rise to the back
electromotive force (b.e.m.f.) in electric motors. The equivalent circuit of a
tachogenerator is also the dual of an electric motor and is shown in Figure
5.2:
F I--...L....-/ M
where T;n is the torque input whereas Fa, M a, La and Ra are all armature
parameters, respectively friction, mass, winding self inductance and winding
resistance. RL represents the electric load connected to the tachometer.
Ignoring any transients, and therefore the effects of all reactive com-
ponents, we can show that the output voltage is given by eqn (5.3):
R L K . W.
Vo = (Kaws - Raid = (R L ; Ra) s (5.3)
which, once again, can be simplified back to eqn (5.1) by lumping Ka and the
resistances in the total constant K t The output voltage Vo can be in direct
current form in the case of the dynamo (a d.c. tachogenerator) or alternating
current form in the case of the alternator (an a.c. tachogenerator).
72 Robot sensors and transducers
w.1 Encoder
5// 000 disk
o 0
000 A COUNTER
o 0 + .JUUlA.
To
OerO IlC
The light source is usually an Infra Red Light Emitting Diode (IR LED)
whereas the light detector is a spectrally matched photodiode (IR PD); for
further information on both these devices please refer to Chapter 3 on
optical transducers.
The modulation of the light from the source to the detector can be
achieved in different ways, as applies also to position transducers; the
perforated disk shown in Figure 5.3 can in fact be substituted by a
transparent (i.e. glass or clear plastic) disk with a pattern of light and dark
stripes printed on it. These light stripes can, in turn, be transparent or
reflective which, in the latter case, permits the use of single packaged
LED-PD pairs. A comparison of these two alternatives is shown in Figure
5.4:
Velocity transducers 73
GEAR
BOX
Load
shaft
Figure 5.5 Example of direct load shaft speed monitoring using reflective optical
transducer
74 Robot sensors and transducers
,---,
I I
( DATA
!.:
I c I I
I ADC ....Qj::J
I L . -_ _
c.
--I E
I I I I .--_ _--1 uo
c:
I ~ I I ~ I "iij
L __ -.-l
L _____ J ~
Low pass Attenuator
filter network
'Clear'
signal
from
Reset computer
COUNTER DATA
ClK To
computer
High
frequency
clock gen
Figure 5.7 Simplified circuit diagram of low speed measurement using single-line
pulse technique
The latter is a useful technique when measuring low speeds with high
resolution and is based on gating a high-frequency clock with the photodiode
output pulse corresponding to a single line transition (Figure 5.7).
The former technique is, however, more useful because in most cases
robot shaft actuators rotate at medium to high speeds (that is in excess of
1 rev Is) and will therefore be described in more detail.
To measure speed by the former technique, one pulse per revolution
(i.e. a single line or slot on the encoder disk) would be sufficient. But digital
counters have an inherent count resolution of one clock pulse so, in order
to allow the accurate measurement of low angular velocities, more than one
pulse per revolution needs to be counted per unit time and therefore several
dark lines (or slots) are required on the encoder disk. The velocity
measurement time can thus be reduced at the expense of having more lines
on the disk and vice versa.
The three main components of the processing hardware are therefore an
amplifier, a digital counter and a clock circuit. A real-time clock is not
strictly necessary since the measurement can be calibrated on any time
interval to give the output in revolutions per minute as required. However,
with the advent of very stable and highly accurate quartz crystal oscillators a
real-time clock is quite easily built. A circuit is illustrated in Figure 5.8
Crystal Divider
oscillator circuit
Figure 5.8 Typical block diagram of real-time clock based on a quartz oscillator
76 Robot sensors and transducers
Mains
S1SL
II
voltage Comparator
(240V)
(50Hz)
I. .1
20ms
ClK
---.lIIlllJ1 JIlIlL IP
TOIL C
8 bit
counter
8
Reset
Comparator \------'
5.5 Conclusions
The d.c. tachogenerator and the optical incremental encoder are the most
popular types of velocity transducer in the field of machine control and
robotics.
The tachogenerator can be interfaced easily to a computer via an A/D
converter, requires no power supply nor, typically, any voltage amplification
and its measurement resolution depends only on the A/D number of bits.
The optical incremental encoder requires output amplification and a
digital counter for interfacing to a computer and its measurement resolution
depends on the number of lines on the encoder disk as well as the
measurement interval. However, it has the inherent advantage of providing
both position and velocity measurement from a single transducer, though
78 Robot sensors and transducers
requmng the use of two separate interface circuits to achieve it, and is
therefore a more versatile transducer.
(a) Draw the interface circuit between a computer and the tachogenerator
whose characteristics are shown in Figure 5.1 using an A/D with an
input range of 0-2 V. State suitable component values for a speed range
of 0-1900 rev/min assuming the input of the A/D to be negligible and
that no low-pass filtering is required.
(b) A photodiode circuit with a risetime of 10 /lS is sued to detect the light
modulation through an optical encoder disk with 1000 lines on it.
Calculate the maximum shaft speed possible to maintain measurement
reliability. Assume that the minimum output swing required for reliable
operation of the circuit is 20% of full output.
(c) A robot control system requires a shaft velocity measurement in the
range 60-3600 rev/min with a resolution of 10%. The handshaking
protocol allows communication to the computer only every 10 ms.
Calculate the minimum number of lines required on the encoder disk if
an optical incremental encoder is to be used, and draw the block
diagram of a suitable system.
1. Transducers, Sensors and Detectors, R. G. Seippel, Reston Publishing Co. Inc., 1983.
2. Transducers and Interfacing, B. R. Bannister and D. G. Whitehead, Van Nostrand
Reinhold (UK) Co. Ltd., 1986.
3. Robotics-An Introduction, D. McCloy and M. Harris, Open University Press, 1986.
4. Industrial Robots: Computer interfacing and control, W. E. Snyder, Prentice-Hall Inc.,
1985.
5. Industrial Robots: Design, operation and application, T. W. Polet, Glentop Publishers
Ltd., 1985.
PART II
SENSORS
Chapter 6
6.1 Overview
The last few years have seen the increasing use of robots in the industrial
world and the emergence of robotics as a subject area with its own identity.
The development of robot technology has been identified as following three
major conceptual stages which, in turn, have helped to identify the robots
belonging to these stages as respectively, robots of the first, second and third
generations (Pera, 1981).
First generation robots: These are robots without any external (i.e. ex-
teroceptive) sensors or transducers. They do not therefore have the means
(i.e. the sensors) nor, usually, the computing power to interact with their
environment. These robots control the end-effector by calculating its
location using the data supplied by the internal (i.e. proprioceptive) position
transducers present within each robot joint. For further information on the
position transducers please refer back to Chapter 2.
At present the majority of the commercially used robots belong to this
category.
power which allows them to interact fully with the environment, that is: to
make decisions, plan and execute the tasks that circumstances require.
These robots are not yet in existence.
Second and third generation robots therefore require sensors that can
provide the necessary environmental feedback and help to increase the
robots' accuracy and/or their flexibility. The development of these sensors
can follow two main paths:
(i) a long term strategy aimed at the development of a general-purpose,
flexible sensor
(ii) a shorter term strategy aimed at providing specific, albeit inflexible,
solutions to current industrial automation problems.
The former option is more suitable for the development of the future
Flexible Manufacturing Systems but does require a larger investment, not
only in terms of capital but also, quite significantly, in terms of manpower
and planning. This type of research therefore tends to be limited to
academic institutions and large corporate companies.
The latter option, by contrast, reflects the needs of western industries to
be both competitive and efficient, achievements which are measured on
relatively short time scales.
It comes as no surprise, therefore, to find that in the field of Robotic
research the majority of USA and, to a lesser extent, European research
institutions, are following a plan of 'task driven' research based on a close
collaboration between the academic institutions and individual 'pools' of
industrial sponsors. This arrangement, first pioneered by Rosen et al. at the
Stanford Research Institute in the early 70s, is believed to be one of the
underlying causes for the present USA success in this field.
The choice of which type of sensor is incorporated in the robot control
structure depends, on course, on the application. It is generally accepted,
however, that vision is the most powerful and yet flexible type of
environmental feedback available, which has led to considerable research
and development in to this field. Indeed Robot Vision and Sensory control is
now an established conference topic in its own right. Table 6.1 shows the
main areas of interest in the field of robotic vision.
n
2-D 3-D
I
BINARY GREY DIRECT RANGEPIC INDIRECT RANGEPICS
LEVELS MEAS. METHODS MEAS. METHODS
OPTICAL ULTRASONIC
RANGEFINDER RANGEFINDER
Reflected Through
illumination illumination
technique technique
Principal Possibility of Easy separation of
advantages surface features objects from background;
for image analysis; minimum amount
grey scale gradations of data to be handled
are possible;
additional information
about object's
shape is available;
use of code markings
possible
Principal Extracting the Only contour and
disadvantages relevant picture 'through' features of
detail may be object are accessible
difficult and slow; to image analysis;
shadows and highlights transparency of
have to be considered; object support,
extraneous light viewing and
sources may illuminating
influence system elements can be
functioning impaired by dirt, etc.
Vision sensors are so called because they possess, in their make-up and
functioning, a certain analogy with the human eye and vision. The analogy is
somewhat easier to see in the case of vacuum and solid-state cameras,
because they also possess the 'equivalent' of the human retina in the form of
a photosensitive array. In the case of some active 3-D vision sensors, such as
scanning laser range finders , the data is acquired by mechanical scanning
using a single optical transducer.
84 Robot sensors and transducers
Through
illumination
Reflective
illumination
Light 'sections'
...
A ... Illumination ... ~ontrolline~ l
system
I
I
, __ ..l __ ,
I-
zw
~ Data Data
z acquisition pre-processing
0 (intelligent
a:: system
5> sensors only)
Z
w
L--T--.J
I
... I
.. ... _____ -.J
.
Illumination
A
system
Main computer
highway
thought that prefers 2-D to 3-D vision systems in view of their simpler and
more trouble-free operation (Keller, 1983; Corby, 1983)-and as the input
stage for 'model-based' vision processing systems that aim to derive a 3-D
image from the 2-D data using a suitable model of the object in question
(Meyer, 1984). The research emphasis, however, is on 3-D robot vision as a
means of achieving faster, more accurate and more flexible robot operation.
2-D vision sensors are based on optical array transducers (both vacuum
and solid-state types, such as camera tubes, CCD, DRAM, and photodiode
arrays), dealt with in Chapter 3, and specially designed artificial lighting.
Some 'intelligent' 2-D vision sensors also possess limited computer power to
enable them to carry out a certain amount of picture pre-processing. This
may be used both to relieve the data-processing burden on the main control
86 Robot sensors and transducers
computer and to interact with the illumination system (e.g. to control the
lighting level) and so provide the best possible picture quality to the main
computer.
Most optical array transducers (e.g. vacuum and solid state cameras)
have the capability, though via different means, to provide a 'grey level'
output image. The only exception to this guideline is the DRAM camera
(for further details, see Section 3.5.2.4) which, being based on a memory
device, produces an inherently binary output image. It should be pointed
out, however, that a grey level image can be produced with a binary device
at the expense of the image acquisition time by overlaying several binary
images obtained with different exposure times.
The lighting conditions and/or the computing power limitations may,
"'" r-- ~
Illumination
system
~
2-D . Thresholding
camera
V
/I
Address "
\
f\ V
Illumination
system
~"<:
,..
Main computer
highway
This class of sensor produces images whose pixel (pixel = picture element)
values are either a black or white luminosity level equivalent to a logic 0 or
1, hence the name 'binary'. Figure 6.3 shows a typical block diagram of a
binary vision sensor:
The complete picture is therefore only a series of logic 1 and O. This
allows easy distinction of dark objects on light background (and vice versa)
and, in view of the low visual data content, fast image data manipulation
such as object perimeter and/or area calculations.
Figure 6.4 shows the typical output of a complete binary vision system:
I(-:=--==~:\---------~;----------------------I
l .....-........--...-.-......- ....-----....
J ....-",-.-.,
If
1,..1_.. _.,1
II
(:~~~~~~:.)
I
_..__.._.._.._...... _. __.._..__ .. _.........._......_............_.._.......... _......._.._.. _. __...._......._.......___...__.__.._._....__._.._..__............._......_..._........ _...... _ ............. __ ....._................_..........____....1
In the example shown the vision system has learnt the characteristics of
three objects, based on the number of holes, object size and shape features.
Thereafter, whenever the vision system is presented with an image contain-
ing one of these objects it will be able to 'recognize' it, that is acknowledge
its presence within the field of view. This is achieved by comparing the
features of the object in the new image with those stored in memory,
looking for a match. In practice a certain amount of tolerance needs to be
built into the matching algorithm because the sensing process is not perfect
and will not always produce the same output for the same object, so that a
'perfect' match would hardly ever be possible. Too much tolerance,
however, may result in the vision system 'recognizing' two similar objects as
being one and the same. Further details on image degradation and noise are
given in Chapter 8 on image processing.
This technique of image processing is sometimes referred to as the SRI
method (or one of its derivatives) because it was first introduced by Rosen et
al. at the Stanford Research Institute (now SRI International) and imposes
severe restrictions on the lighting conditions and on the position of the
object for a successful operation. The object must, in fact, show up as an
isolated silhouette and its stable states must be known in advance (that is the
object can be rotated within the image plane but not tilted with respect to
it). Thus overlapping objects and situations with any 3-D freedom of
movement are difficult to deal with using binary vision sensors.
This class of sensor produces an output image whose pixel values are
quantized into a number of discrete levels, achieved by converting the
optical transducer analogue output into the appropriate number of digital
levels.
An alternative to this method is necessary when using a DRAM camera
as the imaging device because its output is inherently binary; a grey-level
image can be produced with the help of some computing power (either
within, as in the case of intelligent sensors, or without, by using the vision
system main computer). This procedure is based on processing an appropri-
ate number of binary images obtained using different exposure times, whose
difference provides a measure of each pixel intensity in the object image.
Figure 6.5 shows the typical grey-level image of a bicycle chain link
using eight intensity levels as provided by a 2-D grey level vision sensor
based on a 256 x 128 cells DRAM camera. A comparison of this image with
the one obtained using a binary sensor, as seen in Section 6.4.1, shows how
grey-level vision sensors (in conjunction with reflected illumination) provide
more details of the object surface features. This, however, increases the
amount of visual data and makes its processing more difficult, slower and
computationally more expensive.
Robot vision sensors 89
"i'!iii:im:';:i'j::;':'mmmm::::!:!::::::!!:!!!
iillll~!!!;IWi~[~f~/( .
Figure 6.6 Stereo imaging diagram (after Nevatia, courtesy of Prentice-Hall, 1982)
consisting of three dark squares each as marked. Each square in one image
is similar to any of the three in the other. If we now correspond Ll and R 1 ,
Lz and R 2 , L3 and R 3, the three squares will be computed to be at the same
height above the background, as shown by the filled squares. If Ll were to
be matched with R 2 , Lz with R3 and L3 with Rl then the computed heights
would be shown by the empty triangles. Another possible interpretation is
shown by the unfilled circles, thereby giving an indication of how critical the
correspondence problem can become in the absence of any known and
unique object features.
In spite of these difficulties and the relatively high expected price tag,
robot stereo vision is a desirable goal. Stereo vision has the highest inherent
3-D image resolution (limited only by the type of camera and its optics) and
flexibility (for instance it is the only method that can provide colour images
relatively easily) and as such it comes closest to the aforementioned
definition of a general-purpose, flexible vision sensor (see Section 6.1). This
makes it a desirable goal but does require large investments and long project
lead times.
The USA financial investment in stereo vision research, for instance,
has already been considerable (approx. $3,000,000 to date), but the results
so far have been limited mostly to laboratory prototypes. The reasons are
thought to be many and varied, ranging from the aforementioned difficulty
of solving the disparity problem in a sufficiently short time to the sheer
complexity of a system that is essentially trying to emulate a major function
of the human brain. Recently, however, there have been reports of
successful industrial applications, such the positioning of car bodies, using
stereo vision (Rooks, 1986).
There are 3 main methods of using the 2-D vision sensors to obtain
multiple views as required for stereo vision:
(a) Disparity method 1. Use of two stationary imaging devices. This
Illumination
system
~
Figure 6.8 Stereo vision, disparity 1 method
Robot vision sensors 93
could be defined as the 'classical' stereo vision method because of its analogy
to the human vision system. As shown in Figure 6.8 it consists of an
illumination system and two stationary cameras which provide the required
two 2-D images. This method is inherently more expensive than the other
two because it uses 2 cameras but dOt!s not require any mechanical
movement and, therefore, compared to method 'b' is faster and can provide
more accurate measurement of the cameras positions as required for the
disparity calculations.
(b) Disparity method 2. Use of one imaging device moved to different
known positions. This is essentially a cost variation on the method 'a' since,
as shown in Figure 6.9, it only differs by the use of a single camera which, to
provide images from a different angle, is mechanically moved to a different
known position.
Illumination
system
Move
/ '<,
A ,---
<
/
) Camera
position
" / 2
"V /
I Came" I
Iliumina~ion 011umination
system' system 2
A A/
Jt. /
~jJ /
~ Jt.
Structured lighting research has also received wide support both in Europe
and in the USA, particularly in the areas of inspection and quality control.
This method is based on the idea of using geometric information, encoded in
the illumination, to help extract the required geometric information from the
object 2-D image. This is achieved by projecting a suitable light pattern
(using a high-power light projector or a laser) on the target object and
observing the deformations that the object shape produces on the pattern,
using a suitable 2-D optical transducer such as a camera and triangulation
calculations to obtain the depth map.
The light pattern can be a series of geometrically arranged dots, parallel
lines or more simply a sheet of light, depending on the shape of the object
and the application. Figure 6.11 shows how a plane of light is projected on
to the target object (in this case an optically opaque cube) and produces the
2-D image shown in Figure 6.12. This image, often referred to as the 'raw
data', is then processed to extract the 3-D information about the object.
As only that part of the object which is illuminated is sensed by the
camera so, in the instance of a single plane of light (as shown in Figures
6.11, 6.12 and 6. 13(a , the image is restricted to an essentially one-
dimensional entity, thereby simplifying the pixel correspondence problem.
The light plane itself has a known position and every point in the object
Figure 6.11 Structured lighting generalized diagram (light stripe pattern shown)
Robot vision sensors 95
image must also lie on the light plane in the 3-D space. To find out where
exactly this point lies (i.e. how far from the projector and the camera) one
must consider that the light plane and the line of sight of the camera
intersect in just one point (as long as the camera's focal point is not in the
light plane). Thus, by computing the intersection of the line of sight with the
light plane we can calculate the 3-D position of any object point illuminated
by the stripe.
It is important to notice that the only points that can be 'seen' by both
the light source and the camera at once can be computed in 3-D. Since the
triangulation calculations require a non-zero baseline, the camera cannot be
too close to the light source and thus concavities in the scene are potentially
difficult to measure, since both the camera and the light source may not be
able to see into them at the same time. Another potential problem is created
by those object surfaces which are parallel with the light piane since they will
have only a relatively small number of lines projected on to them.
This, and similar problems, can be improved by using different light
patterns as shown in Figure 6.13(a), (b) and (d). However, the image
processing is easier when the light pattern is a single plane of light
(Popplestone et at., 1975; Agin, 1972; Sugihara, 1977) than when it is a
series of geometrically arranged dots or stripes, in view of the higher visual
data content provided by the two latter techniques.
Robot vision based on structured lighting has recently produced some
industrial applications worthy of note both in the USA and the UK (Nelson,
1984; Meta Machines, 1985; Edling, 1986).
Dynamic focusing has not received as much attention as the other two
indirect 3-D vision methods but represents an important alternative particu-
larly in terms of price and speed of operation. This technique relies upon
correlating the signature between two linear photodiode arrays to obtain the
range information (Stauffer and Wilwerding 1984) and is therefore in-
96 Robot sensors and transducers
(a) (b)
-I-- - -
f-- --
f---'-
'-- ./
herently less flexible than structured lighting or stereo vision (it relies on
conveyor belt movement to provide the third dimensional axis) but is
eminently suitable for object dimensioning, edge detection and tracking.
Vision sensors based on dynamic focusing make use of the automatic
focusing technology first developed in the mid-1970s for the 35 mm SLR and
video camera markets. They sense the relative position of the plane of focus
by analysing the image phase shift which occurs when a picture is not in
focus.
The principle of operation is as follows: when an image is in focus, all
the light arriving at a single point on the image plane is produced by a
corresponding single point in the scene. That is to say all the light collected
by the lens from this single point in the scene is focused on the
corresponding single point (and this one only) on the image plane. The
whole lens, viewed from a point in the image, must therefore appear to be
one uniform colour and brightness, like the corresponding point in the
scene. It thus follows that, if every point in the image were divided into two
halves, they would both have the same values of colour and brightness (i.e.
their 'signature' would be the same). When the scene image is not in focus
the same two halves of the image point would not have the same values of
colour and brightness (i.e. a different 'signature'), the degree of difference
providing a measure of how far out of focus the scene image is.
This is the same mechanism upon which dynamic focusing vision
sensors, such as the Honeywell HDS-23, are based. The front-end optical
transducer, as shown in Figure 6.14(a), is a single row of twenty-three light
IR FILTER
CORRECTOR LENS
COVER
, / APERTURE
LENSLET
ARRAY
SENSOR
DIE
I--------l
25mm
TCL SENSOR
Figure 6.14(a) Exploded view of Honeywell-HD23 Figure 6.14(b) Electron microscope photograph of
dynamic-focusing vision sensor (courtesy of two light detector cells (divided clearly into two
Honeywell Inc.) halves) and the adjacent CCD shift register array
(courtesy of Honeywell Inc.)
98 Robot sensors and transducers
cells which are similar to the cells in a solid-state camera, but with two major
differences: each of the cells has its own miniature lens, about the diameter
of human hair, and is made up of two halves, a right half and a left half.
Thus, each of the cells in the HDS-23 sensor can look at the surface of the
lens from a single point on the image plane and determine if the light
brightness from the right half of the lens is the same as that from the left
half. Local intelligence within the sensor then computes a 'signature' of the
scene on the image plane, as seen by the twenty three left and right light
cells, and determines if these two signatures are the same (image in focus) or
are not the same (image not in focus).
5 Detectors 15
(a): In-focus signatures
--, 1 ,- r---
Light
, \
,
1.J
' R i g h t halves -
\.. -
intensity
Left halves
~11:'111--
5 Detectors 15
(b): Out-of-focus signatures
It turns out, as shown in Figure 6.15, that for a 2-D scene, for instance a
black square on a white background, the out-of-focus signatures are actually
the same but shifted in phase. By computer processing the direction and the
magnitude of the signature phase shift, the sensor can determine the
Sensor plane
Image plane
14---3.77 (Displacement)
magnitude and the direction of the image plane displacement from the focus
condition and, as shown in Figure 6.16, determine the object distance
(Iversen, 1983).
Honeywell-Visitronics has recently commercially released a range of
robotic sensors based on the dynamic focusing technique (Honeywell-
Visitronics, 1984).
Amplitude Optical
transducer Data
modulated pre-processing
laser and data
I--------~ acquisition
Referencing phase
'------'
Main computer
highway
Static rangefinders
Static LED arrays offer two main advantages over scanning laser range-
finders : they have no moving parts and employ cheaper light sources. They
do, however, suffer from a lower optical launch power and, potentially, also
a lower x-y resolution which tends to limit their use to low-cost, short-range
applications. Thus they are eminently suitable for eye-in-hand robot vision
applications particularly in the case of multisensor robot systems where they
can fill the natural gap between the visual sensing of an overhead camera
and the pressure sensing of a tactile sensor.
Figure 6.18 Typical output of scanning laser rangefinder (after Lewis and Johnson)
Robot vision sensors 101
The principle of operation of a static LED array rangefinder is
essentially the same as that of the scanning laser device, that is each LED
output is focused on to the target object whose reflected light is processed to
provide the range data, but the scan is achieved electronically by multiplex-
ing the drive to the LED matrix so that only one LED is 'on' at any point in
time, thereby avoiding the need for a potentially troublesome mechanical
scan. Two main techniques have been developed to achieve this goal-an
x-y LED array sensor and a circular LED array sensor (Ruocco, 1986;
Kanade and Somner, 1983).
A 2-D LED array sensor has been developed by Ruocco at Middlesex
Polytechnic based on the phase measurement principle. As shown in Figure
6.19 this sensor has the light-emitting diodes arranged in a 2-D array with
the same x-y resolution as the intended object range image.
The principle of operation, as illustrated in Figures 6.20 and 6.21, is as
follows: a single LED is modulated at an appropriate frequency and is
focused on to the target object. The subsequent reflected/scattered optical
flux CPs is coupled to the secondary detectors whose output signal V. is
compared with the reference signal v,. provided by the primary detectors
which receive light only (and directly) from the LED array. The phase
difference between these two signals provides a measure of the object
distance, as shown by eqns (6.1) and (6.2):
.
SIgna time 0 f fl19h t
I = td = -2 . d (6.1)
c
. 2WM~
SIgnal phase lag = /).cP = WM td = = k .d (6.2)
c
where OJM is the LED modulating frequency and c the speed of light.
The procedure is repeated for all LEDs in the array and a 3-D scan of
the target object is thus achieved.
Local intelligence provides a degree of 3-D image pre-processing such
as 'closest object point' and 'holdsites location' which helps to reduce
communication with and processing times by the main robot computer, a
desirable goal in most applications such as obstacle avoidance and pick-and-
place operations.
The features of such a sensor are low cost, high speed and medium x-y
resolution. This latter furthermore increases as the sensor approaches the
target object thus making it ideal for eye-in-hand operation and integration
within a robot Multisensory Feedback System. Embryonic examples of such
systems (like the one based on an overhead camera, a 3-D eye-in-hand
vision sensor and a tactile sensor) have been shown to provide a consider-
able increase in the robot flexibility of operation and are the subject of wide
ranging research interest (Ruocco and Seals, 1986; Ruocco, 1986, Dillman
1982; Van Gerwen and Vleeskens, 1985; Andre, 1985).
A circular LED array sensor has been developed by Prof. Kanade at
102 Robot sensors and transducers
Figure 6.19 Orthogonal projections of 3-D vision sensor (front end only)
I A
I
--!---
......
......
......
syst.
Light
source
matrix
Beat
freq.
asciI!.
Phase
camp.
Intelligent
interface
......
......
"-
"-
"-
"- Second Low pass
.......
detector Mixer and
system limiter
.....
Signal
rv 'i'Ot./ light
r - - - ..... Optical
interface
--- detector
circuit
Light
source
circuit
Reference
+
Phase
light
/~L ___ ..... detector
detector
'V circuit
circuit
'Local intelligence' highway ~.>
ADDRESS BUS
Figure 6.21 Functional diagram of 3-D vision sensor (front-end data acquisition
section)
X,
LSI11
Figure 6.22 Rangefinder based on circular LED array (after Kanade and Somner,
courtesy of SPIE 1983)
104 Robot sensors and transducers
~
Electrode
Y2 Ro
Electrode
X2
PLayer
N Layer X,
Electrode
Y,
(a) (b)
Figure 6.23 Optical transducer functional diagram (after Kanade and Somner,
courtesy of SPIE 1983)
spot position according to eqns (6.1) and (6.2), where r is the resistance of
the homogeneous resistive sheet, L is the length of the transducer and I is
the electrode current:
x = Ix! - Ix2 = 2rxc (6.1)
c Ix! + IX2 rL2R o
Knowledge of the light spot x-y position, the camera optics and the
trajectory of the light beam allows triangulation calculations to be per-
formed, in accordance with eqn (6.3), to find the object point coordinates in
3-D. From the measurement of multiple close object points, the surface
orientation can be calculated.
d
Z=----- (6.3)
1 ( Xc )
dMtan ()
The features of this sensor are: the simplicity of the operating principle,
fast speed (it can generate up to 1000 range measurements per second) and
precision (0.07 mm for distance and 1.5 for surface orientation). Its
drawbacks, compared with the 2-D linear LED array technique, are: the
larger size (Figure 6.24 shows the typical size for a six LEDs sensor), and the
smaller range of operation (4-5 cm compared with 10-80 cm typically), both
thought to be caused indirectly by the inherent construction requirement
that the LEDs be angled with respect to the transducer optical axis.
Development of this sensor is still proceeding, with Prof. Kanade and
his team currently working on a prototype that enables the use of LED
Robot vision sensors 105
o
Laser Linear
photodiode
array
~
Le~ ~I Lens
1--1.
11
l/j7
1//1 - - -
- - - - - TT/- - - Work sur f ace pOSition
.. 1
These, however, are single point devices, that is they are only capable
of providing the distance measurement to one point on the object surface
and, because of their size, are unsuitable for mounting as an x-y array which
would be necessary in order to provide a 3-D image. This sensor geometry
does, however, allow high precision single point range measurements (for
instance, Tanwar quoted a precision of 0.01 micrometers for polished
surfaces) albeit within a small range of operation (75 micrometers for the
Tanwar sensor).
Ultrasonic rangefinders
Ultrasonic imaging has come to the fore in recent years as an important
alternative to X-rays in the field of medicine (Karrer, 1983; Waaft and
Gramiak, 1976). In spite of this success, however, the ultrasonic technique
has yet to provide any new solutions to the problem of range finding in the
field of robotics. This is thought to be due mainly to the large mismatch
between the impedance of the ceramic transducers and that of the air
medium, as well as the wide emission angle of most ultrasonic transducers.
An interesting research program is, however, being carried out at the
University of Canterbury (NZ) by Prof. L. Kay and his team on the
development of an ultrasonic sensor suitable for robotic applications.
The present prototype is only a single point rangefinder but the small
size of the transducers indicate the possible expansion to a limited 2-D linear
array.
Because of their wide angle of operation, single point ultrasonic
rangefinders have, in fact, useful applications in obstacle detection in such
fields as mobile robotics (Seals, 1984). The best known single point distance
sensor currently on the market is the Polaroid ultrasonic sensor developed
Robot vision sensors 107
primarily for photographic camera applications. Since the Polaroid sensor is,
in principle, very similar to others in this field, it will be used as the mean of
explanation. The basic structure of the front end ultrasonic transducer is
shown in Figure 6.26.
)))
TRANSDUCER
TRANSMITTED)
PULSE
,,-
, REFLECTED
I ECHO
"
causing faulty operation, since this object would not be 'detected' by the
ultrasonic sensor.
Vision sensors are a relatively new technology with few common features.
Interfacing these sensors therefore requires an in depth knowledge of the
individual sensor being used and commands strict adherence to the
manufacturer or, more generally the research establishment, data sheets.
Although internally very complex these sensors, however, usually
require few external control signals and, as a general guideline, need the less
handshaking the more local intelligence they have. An example of basic
interfacing is shown in Figure 6.29 for the Fairchild ISCAN 1-D vision
sensor as extracted from the manufacturer CCD catalog (Fairchild, 1984).
Illumination
system
Supplies
connections
OBJECT
Figure 6.29 Interfacing the I-SCAN vision sensor (based on the CCDlll linescan
camera; courtesy of Fairchild, 1984)
This sensor is based on the 1-D optical array transducer CCDll1 (i.e. a
linescan camera) and generates internally all the necessary waveforms shown
in Figure 6.30, thus requiring only an external clock for synchronization
purposes, an exposure signal (if the image exposure needs to be under
software control) and the supply lines in order to provide the output image
on a serial line (for further information on CCD cameras, see Section
3.5.2.3).
Another example of specific yet simple vision sensor interfacing is
provided by the EV1 sensor manufactured by Micro Robotics. This device is
based on the IS-32 DRAM chip and, once again, generates internally all the
necessary waveforms thus requiring only 3 signals: a clock signal of between
110 Robot sensors and transducers
MC
EXPSYNC ------------------.u'--------- , !
cf>XA = cf>XB
cf>1A = cf>1B
cf>2A = cf>2B
4===~~~~~==~----~
.--------------i F; ~----'
cf>R
Output CS
Output OS Video
255
NOTE: FOR DETAILS OF PULSE SHAPES, 2 254 256
SEE CCDDlll DATA SHEET. (FAIRCHILD, 1984)
FOR DETAILS OF CCD 111 PIN LAYOUT
AND DESCRIPTIONS REFER TO FIGURE 3.17(b)
Figure 6.30 I-SCAN vision sensor timing diagram (courtesy of Fairchild, 1984)
1 and 10 f..l,S, a reset signal and a separate logic signal to determine which
half of the DRAM is being used (this latter is only needed because of the
particular optical arrangement of the IS-32 chip). Figure 6.31 shows the
interfacing of the EV1 to the Applelle microcomputer as reported by Marr
(1986); for further references on the DRAM, see Section 3.5.2.4.
6.7 Conclusions
Vision sensors provide the most powerful yet flexible environmental sensing
available for robot systems development. Tough 2-D vision is still used in
industry, 3-D vision sensors are currently being developed in the laboratories
throughout the world and will soon enable the full exploitation of the second
and eventually third generation of robots.
+ 5v + 5v
2K2
U1
Q 8
C040 Clock
STROBE out
I "'L...J -, r
I IV 600nS 11-'S
I
I (T= O.45RC)
AN 2 $--------------------......j$ Reset
: I
AN 0 $ $ High-half
I I
GND;nrfJ ~GNDout
Apple
games
L--$
~
socket EV 1
lead
Figure 6.31 Interfacing the EV1 vision sensor (based on the IS-32 DRAM camera)
to the Apple lIe microcomputer (after Marr, 1986 and courtesy of Apple User)
112 Robot sensors and transducers
Tactile sensors
7.1 Overview
Tactile sensors, like vision systems, have been the object of intense research
throughout the world. Indeed it is increasingly evident that the limitations
suffered by vision systems could be overcome by the cooperative use of
tactile sensors and that Robot Multisensory Feedback Systems, comprising
at least one 2-D vision system, a range finder and a tactile sensor, are a good
base for the development of future generations of robots (Ruocco and Seals,
1986).
As mentioned earlier in Chapter 4, tactile sensors are based on direct
measurement force transducers, that is devices which measure the mechani-
cal deformation produced by a force acting directly on the transducer itself.
This principle was illustrated in Figure 4.1 and is shown again below in
Figure 7.1. Tactile sensors are based on 2-D arrays of such measuring 'cells'.
Like their human analogue, in fact, robotic tactile sensors work by
measuring the contact pressure between the object surface and the robot
gripper; by dividing the contact area into an array of several measurement
points (typically 16 x 16), a 2-D image of the contact pressure can be
mapped, thereby producing a 3-D view, though partial, of the object being
manipulated by the robot gripper.
It should be noted at this point that, following a generally accepted
convention, 'tactile sensing' is herewith defined as the continuous measure-
ment of contact pressure within an array of so called 'tactels', or tactile
elements, as distinct from 'touch sensing' which is defined as a single contact
pressure measurement, as was the case with the force transducers discribed
in Chapter 4.
114 Robot sensors and transducers
Pressure exerted
by object surface
Tactile
pad
Tactile pads
(al Generalised view of a robot (b) Principle of operation
gripper fitted with tactile of tactile sensor
pads
Figure 7.1
Resistive devices are the most popular tactile sensors at present, because of
their higher sensitivity and x-y resolution. They can be based on a variety of
resistive compounds such as carbon fibres, conductive rubbers and various
specially developed media impregnated with conductive dopants. Although
they all exhibit the required property of a change in resistance in response to
an applied pressure, the last category has achieved the highest popularity
because of its lower hysteresis and larger dynamic range.
There are two main ways to measure the resistance change-in line with
the applied force vector, hence measuring vertical resistance changes, and
perpendicular to the applied force vector, hence measuring horizontal
resistance changes. An example of a tactile sensor which measures resistance
change on a vertical plane is shown in Figure 7.3(a). The resistive mat used
Top electrodes
I V '.' ~:->-----"'..,......,.------"'....-.::r-:;oV"'------':>-._ T'"
r----'------.,
Resistive mat
(a) Physical construction (b) Interfacing principle
Figure 7.3 Piezoresistive tactile sensor (after Robertson and Walkden, courtesy of
SPIE, 1983)
116 Robot sensors and transducers
~ 100
Q)
E
Q)
U5
~
Q;
u
c:
co
ti::J
"0
c:
o
U a 2 3 4 5
Pressure/ bar
Figure 7.4 Piezoresistive tactile sensor characteristic (after Robertson and Walk-
den, courtesy of SPIE, 1983)
Figure 7.5 Typical tactile sensor output image (after Robertson and Walkden ,
courtesy SPIE, 1983)
This is not a popular technology for tactile sensors at present but may be
developed more in future in view of its potential for measuring higher
Tactile sensors 117
Total of 8 cells
,-------,II 11_____
r---,
I I
L _ _ ...J
,-------,I I I1,--------, I
L , ---, ,---,
I
L _ _ ---1
r--,r---'
'/'0" I
L
I I
_ _ ...J L _ _ -I
I
L
I
_ _ -I
II
L _ _ --l
I
Force
Dielectric
Figure 7.7 Side view of tactile sensor based on capacitive tactels array
118 Robot sensors and transducers
Light A
~~______~
flFo'ce __-------- ___~
A Light
source ----l~ detector
Several clear plastics and glass fibres exhibit this property, so one
implementation of an optical tactile sensor is based on an x-y array of light
sources and detectors communicating via glass fibres or arranged on the edge
of a square sheet of the aforementioned clear plastic, as shown respectively
in Figure 7.9(a) and (b). Another optical technique which can be used to
build a tactile array is shown in Figure 7.10 and is based on the slotted
F
F
Figure 7.9 Optical tactile sensors operating on light scattering (note: represents
a light source and 0 represents a detector)
Tactile sensors 119
Plunger
Opto switch
These devices are still mostly at the research stage, their interfacing de'pend
largely on the technology used and therefore needs to be referred to the
manufacturer or research institution data sheets and device manuals.
The few tactile sensors commercially available, however, are usually
fully interfaced to one of the popular bus systems and can therefore be
accessed like any computer peripheral. An example of such a tactile sensor
is the device manufactured by the GEC-Hirst Research Laboratories
(Robertson and Walkden, 1983) which already contains an IEEE Multibus
interface.
120 Robot sensors and transducers
7.5 Conclusions
(a) Discuss the application of a robot to pick eggs from a carton using only
vision or tactile sensors (not both at the same time) and a standard
gripper configuration, that is not a suction gripper.
(b) Describe the principle of operation of an optical tactile sensor. Can you
suggest other suitable designs other than those found in this book?
(c) Analyse the action of picking a cup full of liquid and design, in block
diagram form, a suitable multisensory system that would help a robot to
carry out the operation reliably at high speed.
IMAGE PROCESSING
Chapter 8
Image processing
8.1 Overview
The previous chapters have dealt solely with the data acquisition aspect of
sensory feedback, that is they described how to obtain the sensory data
(which will be referred to as the 'image data') and how to interface the
transducer or sensor to the robot control computer. This image data must,
however, be processed in order to extract the required object parameters
and supply them to the robot controller so that the necessary action can be
planned, thus closing the sensory feedback loop, as shown in Figure 8.l.
This processing can be carried out by the robot controller or, as is
becoming increasingly more common, by some local microprocessor con-
tained within the sensor interface, thereby creating what is known as an
'intelligent' sensor. In general terms this combination of the sensory data
acquisition and the subsequent processing is referred to as machine
perception (Nevatia, 1982) or, more specifically as in the case of vision
sensors, as machine/computer vision (Zeuch and Miller, 1986; Ballard and
Brown, 1982).
Typically a robotics vision system would be used for one or more of the
following purposes:
ROBOT
USER ROBOT
INTERACTION f-+ CONTROLLER r--+
ROBOT
ACTUATORS r-. END
EFFECTOR
action
r- ---,
I
~ : I
I
PROPRIOCEPTIVE
I
I +
FEEDBACK I
DEVICES I
(e.g. position
transducers) I
I
I
I I
EXTEROCEPTIVE I I
FEEDBACK
DEVICES
~_i __ ~ ___ J
(e.g. vision sensoryrn formation
tactile sensors)
Enhancement
Correction
I I
r.~-l
-r-
~nalysisJ
Segmentation
Features Extraction
(statistical analysis)
I
Primitives Extraction
(syntactic analysis)
v ou,
I-
v
(OBJECT) - - - - - -, . - - - ' " "
~-~
11.
-~if---~
V TH t - - - - - - - ----
,---------,
Ideal
I I Real
I image
,.---------, Pre-processed
image
---f-(X-=ly-)-IIH"~ h (x,Y)
~r=\
I ~~t l g (x,Y) ..
Image pre- image
L-~_~O_s~_:_~_Si_ng---, r (x,Y)
: I
I n (x,y) I
L ________ --.l
Sensor system
ideally f' (x, y) = r (x, y)
Figure 8.3 Block diagram of typical vision system showing effect of noise and
sensor degradation
(b) Spatial degradation, which affects the intensity value of several image
points at the same time. Typical causes of spatial degradation are poorly
focused optics, object motion (both of which result in blurred images) and
geometric distortion (which results in focused but distorted images, such as
in the case of pincushion and barrel distortion).
It is worth pointing out that this is not a mere academic distinction:
identifying the cause of the image degradation can sometimes provide a
prompt solution to the problem without the use of extensive (and therefore
expensive) software. For instance if the image degradation is due to spatial
distortion caused by improperly aligned optics (e.g. oblique camera viewing
angles) then it may be possible to re-design the sensor system to reduce the
distortion (i.e. reposition the camera). In general, however, image pre-
processing requires the use of considerable computing power. There are two
Image processing 127
. (t,,)
f (a,b) h (a, b)
h ( - a, - b) h (x - a, V - b)
Integral of the
shaded area is
the convolution
for this x and y
f (a,b) h (x - a, y - b)
proposed by Rosenfeld and Kak and is shown in Figure 8.4. The convolution
of tea, b) and h(a, b), in fact, was shown by eqn (8.1) to be equal to the
product of the two functions tea, b) and h(a, b) with this latter rotated by
180 and shifted by an amount x and y along the two respective orthogonal
0
aXlS.
Moreover if we consider an image as a function varying in two dimensions,
then by analogy with time varying signal analysis, we can perform a
harmonic or Fourier analysis on the image . The resultant Fourier transform
of the image will give information on the spatial frequency components from
which the image is composed.
It is common for some forms of image degradation, for example noise,
to occupy a band of spatial frequencies distinct from that occupied by the
features of interest. In this case the degradation can be simply removed by
transforming the image to the spatial frequency domain, deleting those
regions of the transformed image occupied by the degradation component
frequencies by multiplying with a suitable mask function, and transforming
the result back to the spatial domain.
Image processing 129
where
where hr(x,y) is the restoration filter and the symbol '*' indicates 'the
convolution of' the two functions that it links. This is illustrated by Figure
8.5.
Noise
n (x,v)
In order to design this restoration filter we need to solve eqn (8.4) for
which we can use the F.T. as previously described and as shown in eqs (8.5)
and (8.6)
~[r(x, y)] = {~[f(x, y)]. ~[hd(X, y)] + ~[n(x, y)]} . ~[hr(x, y)] (8.5)
R(u, v) = [F(u, v) . Hd(u, v) + N(u, v)] . Hr(u, v). (8.6)
This leads to the derivation of the wanted restoration filter Hr(u, v) in
the spatial frequency domain. This technique is used, for instance, to obtain
the Inverse Filter, that is a function which aims to remove any image
degradation in order to produce a restored image rex, y) = I(x, y). This
technique is therefore a form of image correction as will be described later in
Section 8.3.1.
,-------1
I CORRECTION I ENHANCEMENT I
+
MODELLING
MODELLING
THE
I HYSTOGRAM MAPPING
THE IMAGE DEGRADATICN I OPTIMISATION OPTIMISATION
PROCESS
IMAGE
POINTS SPATIAL ALTERATION
DEGRADATION DEGRADATION
CORRECTION CORRECTION
I
I
1-----, I
I I
I REMAPPING I BACKGROUND
SUBTRUCTION
FILTERING
(CONVOLUTION)
FILTERING
(CONVOLUTION) II AND
GEOMETRIC
INTERPOLATION
I
I
I
I
LlNEARISATION
II I
I r----.,-----------.:r
I (LOOK UP TABLES) I ,
I I I I LOW HIGH
L--~-_--'--J PASS
FILTERING
PASS
FILTERING
J
SIMPLE
AVERAGE
MASK
(UNITY
MATRIX) 71 78 72 77 7~
1 1 1] CENTRE 72 73 70 7~ '2 0
(.[~ : ~j
CENTRE
79- '22 20 23 24
WITH 4
NEIGHBOURS / GREY LEVEL 5 x 5 I MAGE
MASK BEFORE LOW PASS FILTERS
r
I
/
I
73 70 5}
V37
: Li..~ . ..... ~ <[
68 5~
5)3 "38 27*
mask contains negative coefficients. The intensity of the 'filtered' pixel (i.e.
the one at the centre of the mask) is therefore obtained by subtracting the
contributions of the pixel's neighbours from the original pixel at the centre
of the mask, as shown in Figure 8.9. This technique is used to highlight rapid
changes in image intensities such as edges and contours, it closely resembles
mathematical differentiation and tends to sharpen any variation of contrast.
The convolution mask size most commonly used is the 3 x 3 matrix,
since this produces the largest image improvement with the lowest computa-
tionalload. Larger masks, such as 5 x 5 or 9 x 9, are sometimes used in the
larger systems but do impose a prohibitive load on the image processing
system (typically a 512 x 512 8-bit grey level image filtered with a 9 x 9
136 Robot sensors and transducers
STEP MEAN FILTERED MEDIAN FILTERED
I I I IIII II II I I I I I [ III
SINGLE
PULSE
I I I I I I
TRIPLE
PULSE
[-~-:-~]
70 70 70 72 7~
70 71 72 7~ V60
0-1 0 71 72 1~ vS1 60
TYPICAL
EXAMPLE OF
72 7~ 1.-61 60 61
HIGH-PASS
FILTERING 7~ /62 60 61 60
MASK
GREY LEVEL 5 x 5 1M AGE
OBJECT (WITH POOR CONTRAST)
EDGE BEFORE HIGH PASS FILTERING
ENHANCED
OBJECT EDGE
L EXAM PLE: TO
CALC ULATE PIXEL *
71 73 lOy!;"
(BY A PPL YING CONVOLUTION
72 10J 34 MASK TO AREA OF
IMAGE SHOWN SHADED)
9~ 39* 57 ( - 1)(7 3) + ( -1 )(60) +
+(- 1 )(60)+( - 1)(73)+
+ (5)(6 1) = 39
(the kind that produces 'snowy' images even in stationary scenes) which is
quite common within electronic systems. The technique is based on
averaging the intensity values of corresponding pixels p(x, y) within a
sequence of M images; the averaged image g(x, y) will thus be given by:
1 M
g(x,y)= M~g(x,y), (8.11)
Input image
Grey
level
value
(a) Row x x
Grey Grey
level level
value value
x
(b) Row x after low (e) Row x after high
pass (averaging) pass filter: edges
filter: edges are are highlighted
blurred
technique can help to correct for some sensor non-linearities such as the
spatial degradation found in the output image of some sensor whose
sensitivity in the centre of the image matrix is different from that of the
outer part thus causing what is known as the 'halo' effect, as shown in Figure
8.11:
Image processing 139
,
-.
,
Image corrupted by Background (ie . image
lens dirt with no object)
Filtered image
Region determination
As shown in Table 8.3, there are currently two main techniques for region
determination: point dependent and neighbourhood dependent techniques.
Point dependent techniques aim to find and sort out regions with similar
,
Region determination
~
,
Point dependent Neighbourhood dependent
Global
thresholding
Adaptive
thresholding
Edge
enhancement
Edge detection
~
Boundary
mapping
Image processing 141
grey level properties (e.g. to separate all 'white' regions from the 'black'
ones), this process is commonly known as thresholding.
Neighbourhood dependent techniques, on the other hand, aim to find
and sort out regions with rapidly changing grey level values (e.g. edges and
boundaries), this process is commonly known as edge mapping.
8.4.1.1 Thresholding
Thresholding is a point dependent segmentation technique and produces two
different types of output images. This tends to lead to a further subdivision
in terms of thresholding and semi-thresholding. Given that Ln represents the
light intensity of the nth pixel in the image, these segmentation terms can be
defined as follows:
Thresholding produces a binary image whose white pixels (logic '1' in
optical binary terms) represent inputs above the threshold value Th
Conversely the black pixels represent inputs below Th , that is:
THRESHOLDING
(binary output) {I for Ln(in) 2:T,.
Ln(out)
o for Ln(in) < T,.
.
The selection of the threshold value T" is quite critical and often
requires a human operator to make the choice, through either experience or
trial-and-error techniques. This is particularly true in medical image proces-
sing where the very complex images encountered require a great deal of skill
to reliably extract the wanted image features.
In those fields where a human operator is not part of the sensory
feedback loop, as in most systems based on Artificial Intelligence, the
computer needs to select the threshold value T". This can be done in two
different ways:
(i) Global thresholding, when the value of T" is constant throughout the
processed image. This technique is usually applied when the wanted
object is clearly distinguishable from the background as in the case of
images whose grey level histogram is BIMODAL as shown in Figure
8.12. In this instance the choice of threshold value is straightforward
since T" can be placed in the valley between the two peaks produced
by the background and object pixel counts, as shown by the shaded
area of Figure 8.12.
(ii) Adaptive thresholding, when the value of T" needs to be varied within
142 Robot sensors and transducers
Pixel
count
255
Grey level value (8 bits)
Chain link
Grey
level
Row Th
2
X
Th l
Coin
o
(a) Original grey (b) Corresponding
level image edge map
Row
X
o
Grey
level
value L-+--_-+_ _+--+----.--+-+--~
Second
derivative L....JL--+--+1I-L....I..1r-hr-+----'~
x
Figure 8.15 Edge detection using 1st and 2nd order derivatives
but are finite in computer terms because in the digital world of computer
vision dx is equivalent to the image horizontal resolution which cannot
become zero and therefore always produces a finite dy / dx derivative.
One drawback of differentiation is that it tends to highlight any noise
present within the image because it treats the noise as a signal and produces
an 'edge' pixel wherever a noise spike occurs: this can however be corrected
by a limited form of low pass filtering prior to the differentiation stage. The
amount of low pass filtering required is small because the edge mapping
technique can tolerate a little random noise, since this tends to produce
isolated spots which can be eliminated during the next stage in the edge
mapping process, namely template matching or boundary tracking.
The edge pixels thus obtained are, in fact, further processed using
template matching and boundary tracking masks to check which object they
belong to.
Template matching masks are used in the same way as convolution masks
and aim to detect a particular configuration of edge pixels. Figure 8.16
shows an example of a template matching mask used to identify edges
with a 45 orientation.
Image processing 145
-1 -1 2
-1 2 -1
2 -1 -1
Boundary tracking is applied to the edge data, that is to the image after
the differentiation stage, in order to find all the pixels which belong to the
same object boundary thus producing the necessary segmented image, as
shown in Figure 8.17.
Boundary tracking is therefore based on identifying a pixel on a
boundary and on following the boundary line, using a tracking mask, until
returning to the same pixel, thus having traced the closed boundary of the
object in question. This is achieved in four main steps:
(i) A starting boundary pixel is selected by choosing the highest value in
the edge data, that is the highest grey level gradient in the original
image,
(ii) A tracking mask, usually a 3 x 3 matrix, is centred on this starting
pixel,
(iii) The next boundary point in the chain is found, as illustrated in Figure
8.18, by examining the pixel diametrically opposite to the current
point and its neighbours either side of it,
(iv) The tracking mask is then moved on to this new boundary point and
steps (iii) and (iv) are repeated until returning to the starting
boundary point as chosen in step (i) thus having traced the object
closed boundary.
Figure 8.17 Example of boundary tracking algorithm: 1 denotes the starting pixel
and 2 the next point found on the boundary
146 Robot sensors and transducers
L J l
(a) Image with single (b) Image with multiple
boundaries only boundaries due to the
effect of the shadows
(a)
Figure 8.19 Region labelling: connectivity. (a) with 8-way connectivity the two
sub-regions are joined (b) with 4-way connectivity the two subregions are not joined
Once the image has been segmented and the wanted regions within it have
been selected, a number of features can be measured. These can then be
used for automated object inspection and/or robot closed loop control. A
form of object recognition based on the statistical method of object
classification can, of course, be carried out if required and is described later
in Section 8.4.3.
A wide range of different features may be measured and which are used
depends very heavily on the application. For illustrative purposes this
section outlines some common measures.
where cxp is the angle of the pth vector to x axis. For simply coded
boundary cos p can only adopt values of -1, 0 or 1.
The amount of computation included here is considerably
reduced as only the boundary vectors are used instead of the
complete set of pixels in the image. However note that there is no
method of computing the IOD from a boundary description.
(ii) Perimeter
The perimeter of a region is simple to compute for either form of
region representation. For labeled regions the count of boundary
pixels can be taken. This may be the set of region pixels each of
which has at least one background neighbour or vice-versa. Neither
of these will produce the same measure as that for the boundary
description, which is obtained from the arithmetic sum of boundary
vector sizes.
Perimeter, although simple to compute, is an unsatisfactory
Image processing 149
(iii) Envelopes
The simplest region enclosing envelope is the rectangle. The mini-
mum enclosing rectangle (MER) is the smallest rectangle enclosing
all the region of interest. Typically the best fitting rectangle is
computed for a number of orientations and the smallest area
rectangle taken. From the MER can be taken a measure of size
(MER area) and a shape measure,
Max Side Length
the aspect ratio = - - - - - - = - -
Min Side Length
Higher order envelopes can be used but are less simple to compute.
(iv) Derived shape measures
New measures can obviously be produced by combining those
considered so far. Commonly used here are the dimensionless, hence
size independent, shape measures:
Circularity = Perimeter2 / Area
Rectangularity = Area/MER area.
The circularity measure approaches 4n as the region shape ap-
proaches circular. Likewise the rectangularity approaches 1 as the
shape approaches a rectangle.
i ff
= x . f(x, y) dx . dy / ff f(x, y) dx . dy
etc.
150 Robot sensors and transducers
Mo, 1 = II y . dx . dy = Y . Area
Clearly these properties can be simply evaluated in the case of discrete
images. Thus for a labelled, thresholded image
n
Ml,o = 2: Xi
i=l
n
Mo,l = 2: Yi'
i=l
II
This axis can be computed from the axis aligned second central
moments: M . = (x - i)2 dx . dy
20
Principle
axis
Some care must be taken in cases where there may be two competing
axes between which the result may switch in successive images through
simple noise induced variability (see Figure 8.21).
Area
100
IMAGE A
o ~ rl A
(
/\ /\ //\\ /\
) I \
IMAGE, ~
//\\ / \ /\ / \
C) 1\/1'/'1-1
8.5 Conclusions
6r---~---r---'----~--~--~
5~--~---+--- +----r---;---~
Y 3 1---- +---+---i--+--.+---....,
2 r- -
O ~--~--~--~----L---~--~
2 3 4 5 6
X
(c) Calculate the output from the image shown in Figure 8.R.c using a low
pass convolution filter.
24 25 23 26 85 82
24 21 24 90 88 85
25 25 85 87 86 87
, 26 84 91 85 83 85
82 89 83 84 85 82
I
i
92 87 84 83 81 84
Wirewound are typically cheaper. have the same approximate size and a higher
resolution (compared to an optical position transducer with 120 lines) but do suffer
from a lower operational life.
156 Solutions to Revision Questions
2 (b)
Using a radial encoder as shown in Figure 2.8 and applying eqn (2.3) we have:
No. of lines required = 360/6 = 60.
We can use a standard photo diode with an active area of Immz (e.g. a square diode
geometry) but, in order to maintain the 6 degrees resolution we need to mount it at a
distance r from the centre, that is:
Wp :;:., 1
r~ ~ 19.1mm = 20mm.
sin (a oJL) -- sin 3
2(c)
The answer is as shown in Figure 2.13 with 5 channels since, using eqn (2.5):
angular resolution = 360/2c = 15 degrees,
therefore
2c ~ 24 thus giving c ~ 5 channels.
3(a)
We need to turn the relay ON when the object breaks the light beam (that is when the
cell is in the dark, e.g. for an illumination < 10 lum). Therefore we need to have V z >
VI when the cell has a resistance greater than lOKfl. That is:
R3 RI
----<
Rz + R3 RI + Rp
let RI be 5 Kfl and therefore V z > 4 V
12 R3 > 4
Rz + R3
Suitable values R z = R3 = 1.5Kfl with a 22Kfl preset resistor used for fine
adjustment of the threshold value.
3(b)
Since op. amp. is ideal ignore the input capacitance and resistance. From eqn (3.2) we
can extract R v
RL = 1/(2 . 'IT fbw . Cj ) = 318 fl.
As op. amp. is ideal it has no bias current, therefore:
Vo = ip . RL = ('TJUf' <l>OIODIJ R L ,
which leads to the required optical flux coupled to the photo diode active area:
<l>OIOOE = Vo = 6.29 mW, 'TJEFf- RL
coupled into Immz or, as is more commonly quoted, 629 mW/cmz.
Solutions to Revision Questions 157
4(a)
_._ .~:::::::~45 -
- ~-
FTW1ST
4(b)
4(c)
From a standard material's data book Young's modulus for steel is:
Esteel = Stress/Strain = 2.1 x 105 MN/m,
whereas the nichrome gauge factor can be obtained from Table 4.1:
Gf(NiCr) = 2.0.
Using eqn (4.12):
Vs Gf ilL
IlVo = 2 T(1 + 1-'-),
1lV:() = 10 x 2.0 (
2
0.4
2.1 x 10
s) (1 + 0.3) = 24.7 "V.
,...
158 Solutions to Revision Questions
Sea)
As shown, in Figure 5.1 we have that for RL = 1 Kfl the tachogenerator output goes
from 0 to 10 V, therefore a 1:5 attenuation network is required for interfacing to the
AID converter. A suitable circuit is:
w ( ~ Tacho (-+--:}:,~
~~
800 n
Ji
200 n ~
Cony. To Cornpu,"'
S(b)
A photo diode equivalent circuit is essentially capacitive (see 3.3.1 and 3.3.2) and
therefore its output voltage follows a capacitive charge and discharge behaviour:
Vo = VF (1 - e- tiT ),
since 20% of the full output voltage V F can be assumed to be sufficient for reliable
operation we have:
Vo tI
V p = (1 - e- T) = 02;
This is the minimum transition time from light to dark that the photo diode circuit will
tolerate and translates to a maximum frequency of 448 kHz. A disk with 1000 lines on
it will therefore achieve this frequency at 448 rev/sec, that is 26880 rev/min.
S(c)
The minimum speed is
6(a)
Refer to para. 6.2 and, in particular, Table 6.2.
6(b)
Refer to para. 6.S.4(iii). As the speed of light is much greater than that of sound, the
time of flight of a light pulse is much smaller than that of a sound wave.
Solutions to Revision Questions 159
8(b)
Using the zero moments formulae shown in Section 8.4.2.3 and expanding them into
discrete form
/l
= I I (-2,0,0,4,5,2,2, -1)
= l = 5.
8(c)
11~
Using a simple averaging 3 x 3 convolution matrix [ 1 1 1/ +9
1 1 IJ
we have:
31 45 66 80
45 66 80 86
65 79 85 85
71 85 84 83
Agin, G. J. (1972) Representation and description of curved objects (Ph.D. diss.) AIM-173,
Stanford AI Lab.
Andre, G. (1985) A multiproximity sensor system for the guidance of a robot end-effector,
Proc. 5th Int. Conf. on Robot Vision and Sensory Control.
Ballard, D. H. and Brown, C. M. (1982) Computer Vision, Prentice-Hall, New Jersey.
Batchelor, B. G., Hill, D. A. and Hodgson, D. C. (1984) Automated Visual Inspection, IFS
(Publications) Ltd.
Chappel, A. (1976) Optoelectronics-Theory and Practice, Texas Instruments Ltd.
Corby, N. R. Jr. (1983) Machine vision for robotics, IEEE Transactions on Industrial
Electronics 30, 3.
Dillman, R. (1982) A sensor controlled gripper with tactile and non-tactile sensor environ-
ment, Proc. 2nd Int. Conf. on Robot Vision and Sensory Control.
Drews, P. et al. (1986) Optical sensor system for automated arc welding, Robotics 2, l.
Edling, G. (1986) New generation of seamtacking systems, Proc. 9th Annual Conf. of British
Robot Association.
Electronic Automation Ltd. (1984) The EA laser ranger manual.
Fairchild (1984) Charge Coupled Devices (CCD) catalog.
Ferranti Ltd. (1982) Silicon Photocells Manual.
Hamamatsu (1983) Photomultipliers Booklet.
Hill, J. and West, G. A. W. (1982) Digital image processing-short course notes, City
University.
Hollingum, J. (1984) Machine Vision: the eyes of automation, IFS Publications Ltd.
Honeywell-Visitronics (1984) The HDS-24 manual.
IPL (1984) The 2000 series linescan camera manual.
Ikeuchi, K. (1980) Numerical shape from shading and occluding contours in a single view, AI
Memo 566, AI Lab., MIT, Cambridge, MA.
Iversen, W. R. (1983) Low cost system adds vision, Electronics, (November), p. 54-56.
Jarvis J. F. (1982) Research directions in industrial machine vision, Computer, (December),
p.55-6l.
Jarvis, R. A. (1983) A laser time of flight range scanner for robotic vision, IEEE Transactions
on Pattern Analysis and Machine Intelligence.
References 161
Kanade, T. and Somner, T. (1983) An optical proximity sensor for measuring surface position
and orientation for robot manipulation, SPIE 449.
Kaisto, I. et al. (1983) Optical range finder for l.5-1O m distances, Applied Optics 22, no. 20.
Karrer, H. E. and Dickey, A. M. (1983) Ultrasound imaging: an overview, Hewlett-Packard
10urnal34, no. 10.
Keller, E. L. (1983) Clever robots set to enter industry en masse, Electronics, (November), p
116-18.
Kulyasov, A. G. et al. (1973) Optical phase difference rangefinder with digital readout, Soviet
lournal of Optical Technology 40, no. 6.
Joyce-Loebl (1986) Choosing a video input, Sensor Review 6, no. 2.
Lambert, J. (1984) Snap sight system is a snip, Practical Robotics, Sept/Oct.
Lewis, R. A. and Johnson, A. R. (1977) A scanning laser range finder for a robotic vehicle,
Proc. 5th Int. loint Conf. on Artificial Intelligence.
Loughlin, C. and Hudson, E. (1982) Eye-in-hand robot vision, Proc. 2nd Int. Conf. on Robot
Vision and Sensory Control.
Marr, D. and Poggio, T. (1976) Cooperative computation of stereo disparity, Science 194.
Marr, D. and Poggio, T. (1977) A theory of human stereo vision, Report no. AIM451,
Artificial Intelligence Lab., MIT, Cambridge, MA.
Marr, J. (1986) Put your Apple in the picture, Apple User, February.
Matthys, R. J. et al. (1980) CO 2 TEA laser rangefinder, SPIE 227.
Meta Machines Ltd. (1985) Advancing sensor technology into industry, The Industrial Robot,
March.
Nelson, F. Jr. (1984) 3-D robot vision, Robotics World, (March), p. 28-29.
Nevatia, R. (1981) Machine Perception, Prentice-Hall, New Jersey.
Nimrod, N., Margalith, A. and Mergler, H. (1982) A laser based scanning rangefinder for
robotic applications, Proc. 2nd Int. Conf. on Robot Vision and Sensory Control.
Nitzan, D., Brain, A. and Duda, R. (1977) The measurement and use of registered
reflectance and range data in scene analysis. Proc. of the IEEE 65, no. 2.
Okada, T. (1978) A short range finding sensor for manipulators. Bull. Electrotech. Lab.
42(6), pp. 492-504.
Okada, T. (1982) Development of an optical distance sensor for robots, Int. lournal of
Robotics Research 1, no. 4.
Oldeft. Seampilot: a 3-D laser based seam tacking system for arc welding automation.
Owen, G. and Eckstein, R. (1982) Automating tedious tasks in the laboratory, Analytical
Chemistry
Page, C. J. and Hassan, H. (1981) Non-contact inspection of complex components using a
rangefinder vision system, Proc. 1st Int. Conf. on Robot Vision and Sensory Control.
Page, C. J., Hassan, H. and Payne, D. B. (1983) An integrated rangefinding and intensity
sensing system, Proc. lEE Seminar on UK Robotics Research, Digest no. 1983/104.
PERA (1981) Robots, report no. 361.
PERA (1982) Vision systems, report no. 366.
Polaroid. Ultrasonic rangefinder manual.
Popplestone, R. J. et al. (1975) Forming models of plane-and-cylinder faceted bodies from
light stripes, Proc. 4th Int. IICAI.
Pratt, W. K. (1978) Digital image processing, John Wiley & Sons, New York.
Pugh, A. (1982) Second generation robotics, Proc. 12th Int. Symp. on Industrial Robots.
Pugh, A. (1983) Robot vision, IFS (Publications) Ltd.
Robertson, B. E. and Walkden, A. J. (1983) Tactile sensor system for robotics, Proc. 3rd Int.
Conf. on Robot Vision and Sensory Control.
Rooks, B. (1986) Advanced technology shines on the Rover 800 line, The Industrial Robot
13, no. 3.
Rosenfeld, A. and Kak, A. C. (1976) Digital Picture Processing, Academic Press, New York.
Ruocco, S. R. (1986) The design of a 3-D vision sensor suitable for robot multisensory
feedback, Proc. 6th Int. Conf. on Robot Vision and Sensory Control.
Ruocco, S. R. and Seals, R. C. (1986) A robot multisensory feedback system, Proc. 9th
Annual Conf. of the British Robot Association.
162 References
disparity method 89, 91, frame averaging 137-8 image degradation 88,126-
92-3 Freeman chain coding 153 7
dynamic characteristics 7-9 frequency response 7, 8 correction 129, 130-33
dynamic focusing 95-9 full-scale output (FSO) 4, image pre-processing 48,
dynamic random access 6, 7 88,124-6
memories (DRAM) convolution and filtering
47-8,51,85,86,88 gauge factor 57 127-39
dynodes 36-7 geometric distortion 48, image analysis 139-53
132-3 image processing 43,88,
edge detection 143-4 global thresholding 141 122-7
edge mapping 141, 142-6 Gray coded patterns 19, 21 phases see convolution
electron-hole pairs 28-31, Grey-scale modification and filtering; image
42,45-6 133-4 analysis
encoders grey level image segmentation 124,
absolute 16, 19-21 image 48, 82-3, 86, 87, 125
incremental 16-19, 24-5, 88-9 region description 146-
69, 72-8 vision sensors 88-9 51
end effector position 12-14 grippers 12, 53 region determination
energy transduction 1 140-46
energy transformers 1 Hall effect 69 incremental encoders 16-
EVI vision sensor 109-10, 'halo' effect 137, 139 19,24-5,69,72-8
111 handshaking 77 incremental position
exposure time (DRAM high pass filtering 134-7, transducers 14, 18,
camera) 48 143 23-4
external photoeffect 29 histogram modification 134 indirect measurement force
external position feedback hole-electron pairs 28-31, transducers 54
12-13,80 42,45-6 Infra Red Light Emitting
exteroceptive feedback 12- hysteresis 6-7 Diode 72-3
13 Infra Red photo diode 72-3
exteroceptive sensors 80 I-SCAN vision sensor 109- input characteristics 2
10 Integrated Optical Density
FS04, 6, 7 illumination techniques 81- 147,151
fall time 8, 9 3,84,86 intelligent sensors 2
Faraday's law of image interfacing
electromagnetic classification 124, 125, force transducers 67
induction 71 151-3 position transducers 21-
'feature extraction' 2 correction 129, 130-33 4
filtering data 122,123,124 tactile sensors 119
high pass 134-7, 143 descriptions 146 velocity transducers 74-
low pass 134, 135, 144 enhancement 127, 133-9 7,78
see also convolution and feature extraction 151 vision sensors 109-10
filtering grey-level 48, 82-3, 86, internal photoeffect 28-9
first generation robots 80 87,88-9 internal position
Flexible Manufacturing measurement 125 transducers 14, 80
Systems 81 segmentation 124,140-51 inverse filter 130, 132
force transducers signature 96, 98
interfacing 66, 67 transformation/filtering junction photoeffect 28
load cells 65, 67 134-7
measurement image analysis 124, 139 kinematic forces 53
generalities 53-7 classification 124, 125,
non-resistive 56-7, 64-5 140,151-3 LDR 29-30,41
strain gauges 57-63,65- feature extraction 151 LED see Light Emitting
6 segmentation 124, 140- Diode
Fourier analysis 128, 129 51 LID 38
Index 165