Introduction-To-Robotics - Compress (1) - 98-137
Introduction-To-Robotics - Compress (1) - 98-137
configurations, i.e., position and orientation of the parts, before the robot starts
welding.
There could be other scenarios like identifying the color code of a particular car
model before painting with that color is done by the robot, etc.
wiper and end leads of the device changes in proportion to the displacement, x and q
for linear and angular potentiometers, respectively.
3. LVDT The Linear Variable Differential Transformer (LVDT) is one of the most
used displacement transducers, particularly when high accuracy is needed. It generates
an ac signal whose magnitude is related to the displacement of a moving core, as
indicated in Fig. 4.4. The basic concept is that of a ferrous core moving in a magnetic
field, the field being produced in a manner similar to that of a standard transformer.
There is a central core surrounded by two identical secondary coils and a primary
coil, as shown in Fig. 4.4. As the core changes position with respect to the coils, it
changes the magnetic field, and hence the voltage amplitude in the secondary coil
changes as a linear function of the core displacement over a considerable segment. A
Rotary Variable Differential Transformer (RVDT) operates under the same principle
as the LVDT is also available with a range of approximately ±40°.
4. Synchros and Resolver While encoders give digital output, synchros and
resolvers provide analog signal as their
What are arc-minute
output. They consist of a rotating shaft
and arc-second?
(rotor) and a stationary housing (stator).
Their signals must be converted into the They are the measures of small angles.
one degree = 60 arc-minutes and
digital form through an analog-to-digital
one arc-minute = 60 arc-seconds.
converter before the signal is fed to the
computer.
As illustrated in Fig. 4.5, synchros and resolvers employ single-winding rotors
that revolve inside fixed stators. In a simple synchro, the stator has three windings
oriented 120° apart and electrically connected in a Y-connection. Resolvers differ
from synchros in that their stators have only two windings oriented at 90°. Because
synchros have three stator coils in a 120° orientation, they are more difficult than
resolvers to manufacture and are, therefore, more costly.
Modern resolvers, in contrast, are available in a brushless form that employ a
transformer to couple the rotor signals from the stator to the rotor. The primary
winding of this transformer resides on the stator, and the secondary on the rotor.
Other resolvers use more traditional brushes or slip rings to couple the signal into the
rotor winding. Brushless resolvers are more rugged than synchros because there are
no brushes to break or dislodge, and the life of a brushless resolver is limited only by
its bearings. Most resolvers are specified to work over 2 V to 40 V rms (root mean
square) and at frequencies from 400 Hz to 10 kHz. Angular accuracies range from 5
arc-minutes to 0.5 arc-minutes.
of the angle q between the rotor-coil axis and the stator-coil axis. In the case of a
synchro, the voltage induced across any pair of stator terminals will be the vector
sum of the voltages across the two connected coils. For example, if the rotor of a
synchro is excited with a reference voltage, V sin (w t), across its terminal R1 and R2,
the stator’s terminal will see voltages denoted as V0 in the form:
V0(S1 − S3) = V sin (w t) sin q (4.2a)
V0(S3 − S2) = V sin (w t) sin (q + 120°) (4.2b)
V0(S2 − S1) = V sin (w t) sin (q + 240°) (4.2c)
where S1, S2, etc., denotes the stator terminals. Moreover, V and w are the input
amplitude and frequency, respectively, whereas q is the shaft angle. In the case of a
resolver, with a rotor ac reference voltage of V sin (w t), the stator’s terminal voltages
will be
V0(S1 − S3) = V sin (w t) sin q (4.3a)
V0(S4 − S2) = V sin (w t) sin (q + 90°) = V sin (w t) cos q (4.3b)
As said earlier, the output of these synchros and resolvers must be first digitized.
To do this, analog-to-digital converters are used. These are typically 8-bit or 16-
bit. An 8-bit means that the whole range of analog signals will be converted into a
maximum of 28 = 256 values.
4.2.2 Velocity Sensors
Velocity or speed sensors measure by taking consecutive position measurements at
known time intervals and computing the time rate of change of the position values or
directly finding it based on different principles.
1. All Position Sensors Basically, all position sensors when used with certain
time bounds can give velocity, e.g., the number of pulses given by an incremental
position encoder divided by the time consumed in doing so. But this scheme puts
some computational load on the controller which may be busy in some other
computations.
2. Tachometer Such sensors can directly find the velocity at any instant of time,
and without much of computational load. This measures the speed of rotation of
an element. There are various types of tachometers in use but a simpler design is
based on the Fleming’s rule, which states ‘the voltage produced is proportional to the
rate of flux linkage.’ Here, a conductor (basically a coil) is attached to the rotating
element which rotates in a magnetic field (stator). As the speed of the shaft increases,
the voltage produced at the coil terminals also increases. In other ways, as shown in
Fig. 4.6, one can put a magnet on the rotating shaft and a coil on the stator. The
voltage produced is proportional to the speed of rotation of the shaft. This information
is digitized using an analog-to-digital converter and passed on to the computer.
3. Hall-effect Sensor Another velocity-measuring device is the Hall-effect
sensor, whose principle is described next. If a flat piece of conductor material, called
Hall chip, is attached to a potential difference on its two opposite faces, as indicated
in Fig. 4.7 then the voltage across the perpendicular faces is zero. But if a magnetic
field is imposed at right angles to the conductor, the voltage is generated on the two
84 Introduction to Robotics
other perpendicular faces. Higher the field value, higher the voltage level. If one
provides a ring magnet, the voltage produced is proportional to the speed of rotation
of the magnet.
where F is force, DR is the change in resistance of the strain gauge, A is the cross-
sectional area of the member on which the force being applied, E is the elastic
modulus of the strain-gauge material, R is the original resistance of the gauge, and
G is gauge factor of the strain gauge. Then, the acceleration a is the force divided by
mass of the accelerating object m, i.e.,
F DRAE
a= = (4.5)
m RCm
It is pointed out here that the velocities
What is Gauge Factor?
and accelerations that are measured using
It is a measure of sensitivity for the strain
position sensors require differentiations.
gauges, and defined by
It is generally not desirable, as the 1 DR
noise in the measured data, if any, will G=
e R
be amplified. Alternatively, the use of
where G is the gauge factor, and e is
integrators to obtain the velocity from strain.
the acceleration, and consequently the
position, are recommended. Integrators tend to suppress the noise.
Example 4.2 Change in Resistance
If the gauge factor G = 2, resistance of the unreformed wire R = 100 W, and strain
e = 10–6, then change in resistance is given by
DR = GeR = 2 × 10–6 × 100 = 0.0002 W (4.6)
capacitive switch has the same four elements as the inductive sensor, i.e., sensor (the
dielectric media), oscillator circuit, detector circuit, and solid-state output circuit.
The oscillator circuit in a capacitive switch operates like one in an inductive
switch. The oscillator circuit includes capacitance from the external target plate
and the internal plate. In a capacitive sensor, the oscillator starts oscillating when
sufficient feedback capacitance is detected. Major characteristics of the capacitive
proximity sensors are as follows:
∑ They can detect non-metallic targets.
∑ They can detect lightweight or small objects that cannot be detected by
mechanical limit switches.
∑ They provide a high switching rate for rapid response in object counting
applications.
∑ They can detect limit targets through nonmetallic barriers (glass, plastics, etc.).
∑ They have long operational life with a virtually unlimited number of operating
cycles.
∑ The solid-state output provides a bounce-free contact signal.
Capacitive proximity sensors have two major limitations.
∑ The sensors are affected by moisture and humidity, and
∑ They must have extended range for effective sensing.
Capacitive proximity sensors have a greater sensing range than inductive
proximity sensors. Sensing distance for capacitive switches is a matter of plate
area, as coil size is for inductive proximity sensors. Capacitive sensors basically
measure a dielectric gap. Accordingly, it is desirable to be able to compensate for
the target and application conditions with a sensitivity adjustment for the sensing
range. Most capacitive proximity sensors are equipped with a sensitivity adjustment
potentiometer.
2. Semiconductor Displacement Sensor As shown in Fig. 4.14, a
semiconductor displacement sensor uses a semiconductor Light Emitting Diode
(LED) or laser as a light source, and a Position-Sensitive Detector (PSD). The laser
[Courtesy: http://www.sensorcentral.com/displacement/
laser02.php]
beam is focused on the target by a lens. The target reflects the beam, which is then
focused on to the PSD forming a beam spot. The beam spot moves on the PSD as
the target moves. The displacement of the workpiece can then be determined by
detecting the movement of the beam spot.
4.4 VISION
Vision can be defined as the task of
extracting information about the external Computer Vision vs. Computer Graphics
world from light rays imaged by a camera Computer vision can be thought of as
or an eye. Vision, also referred in the ‘inverse computer graphics.’ Computer
literature as computer vision or machine graphics deals with how to generate images
from a specification of the visual scene (e.g.,
vision or robot vision, is a major subject objects, scene structures, light sources),
of research and many textbooks, e.g., by whereas computer vision inverts this process
Haralick and Shapiro (1992, 1993), and to infer the structure of the world from the
others. A good coverage on the topic has observed image(s).
also appeared in Niku (2001). There are
also dedicated journals, e.g., Computer Vision, Graphics, and Image Processing, and
conferences in the area of robot vision. The area is so vast that it cannot be covered
in one section or chapter of a book. However, an attempt is made here to introduce
the basic concepts and techniques so that one is able to understand the systems and
methodologies used in robot vision. For detailed study and research, other references
in the area should be consulted.
Note in Fig. 4.1 that the vision systems or vision sensors are classified as external
noncontact type. They are used by robots to let them look around and find the parts,
for example, picking and placing them at appropriate locations. Earlier, fixtures
were used with robots for accurate positioning of the parts. Such fixtures are very
expensive. A vision system can provide alternative economic solution. Other tasks
of vision systems used with robots include the following:
1. Inspection Checking for gross surface defects, discovery of flaws in labeling,
verification of the presence of components in assembly, measuring for dimensional
accuracy, checking the presence of holes and other features in a part.
2. Identification Here, the purpose is to recognize and classify an object rather
than to inspect it. Inspection implies that the part must be either accepted or rejected.
3. Visual Servoing and Navigation Control The purpose here is to direct the
actions of the robot based on its visual inputs, for example, to control the trajectory
of the robot’s end-effector toward an object in the workspace. Industrial applications
of visual servoing are part positioning, retrieving parts moving along a conveyor,
seam tracking in continuous arc welding, etc.
All of the above applications someway require the determination of the
configuration of the objects, motion of the objects, reconstruction of the 3D geometry
of the objects from their 2D images for measurements, and building the maps of the
environments for a robot’s navigation. Coverage of vision system is from a few
millimetres to tens of metres with either narrow or wide angles, depending upon the
Sensors, Vision and Signal Conditioning 91
system needs and design. Figure 4.15 shows a typical visual system connected to an
industrial robot.
The task of the camera as a vision sensor is to measure the intensity of the light
reflected by an object, as indicated in Fig. 4.16, using a photosensitive element
termed pixel (or photosite). A pixel is capable of transforming light energy into
electric energy. The sensors of different types like CCD, CMOS, etc., are available
depending on the physical principle exploited to realize the energy transformation.
Depending on the application, the camera could be RS-170/CCIR, NTSC/PAL (These
are American RS-170 monocolor, European/Indian CCIR monocolor, NTSC color,
PAL color television standard signal produced by the video cameras, respectively)
progressive scan, variable scan, or line scan. Five major system parameters which
govern the choice of camera are field of view, resolution, working distance, depth of
field, and image data acquisition rate. As a rule of thumb, for size measurement, the
sensor should have a number of pixels at least twice the ratio of the largest to smallest
object sizes of interest.
92 Introduction to Robotics
Both the CCDs and CID chips use large transfer techniques to capture an image.
In a CCD camera, light impinges on the optical equivalent of a Random Access
Memory (RAM) chip. The light is absorbed in a silicon substrate, with charge build-
up proportional to the amount of light reaching the array. Once sufficient amount of
energy has been received to provide a picture, the charges are read out through built-
in control registers. Some CCD chips use an interline charge-transfer technique.
Others use frame-transfer approach, which is more flexible for varying the integration
period.
The CID camera works on a similar principle. A CID chip is a Metal Oxide
Semiconductor (MOS) based device with multiple gates similar to CCDs. The video
signal is the result of a current pulse from a recombination of carriers. CIDs produce
Sensors, Vision and Signal Conditioning 95
a better image (less distortion) and use a different read-out technique than CCDs
which require a separate scanning address unit. CIDs are, therefore, more expensive
than CCDs. The principle difference between a CCD and a CID camera is the method
of generating the video signal.
2. Lighting Techniques One of the key questions in robot vision is what
determines how bright the image of some surface on the object will be? It involves
radiometry (measurement of the flow and transfer of radiant energy), general
illumination models, and surface having both diffuse and specular reflection
components. Different points on the objects in front of the imaging system will
have different intensity values on the image, depending on the amount of incident
radiance, how they are illuminated, how they reflect light, how the reflected light is
collected by a lens system, and how the sensor camera responds to the incoming light.
Figure 4.20 shows the basic reflection phenomenon. Hence, proper illumination of
the scene is important. It also affects the complexity level of the image-processing
algorithm required. The lighting techniques must avoid reflections and shadow unless
they are designed for the purpose of image processing. The main task of lighting is to
create contrast between the object features to be detected. Typical lighting techniques
are explained below.
Direct Incident Lighting This simple lighting technique can be used for non-
reflective materials which strongly scatter the light due to their matte, porous, fibrous,
non-glossy surface. Ideally, a ring light is chosen for smaller illuminated fields that
can be arranged around the lens. Shadows are avoided to the greatest extent due to the
absolutely vertical illumination. Halogen lamps and large fluorescence illumination
can be used too.
Diffuse Incident Lighting Diffused light is necessary for many applications,
e.g., to test reflective, polished, glossy, or metallic objects. It is particularly difficult
if these surfaces are not glossy, perfectly flat, but individually shaped, wrinkled,
curved, or cylindrical. To create diffused lighting, one may use incident light with
diffusers, coaxial illumination, i.e., light is coupled into the axis of the camera by
means of a beam splitter or half-mirror, or the dome-shaped illumination where light
is diffused by means of a diffused coated dome in which the camera looks through
an opening in the dome onto the workpiece.
96 Introduction to Robotics
Lateral Lighting Light from the side can be radiated at a relatively wide or narrow
angle. The influence on the camera image can be significant. In an extreme case, the
image information can almost be inverted.
Dark Field Lighting At first sight, images captured using dark field illumination
seem unusual to the viewer. The light shines at a shallow angle. According to the
principle of angle of incidence equals the angle of reflection, all the light is directed
away from the camera. The field of view, therefore, remains dark. Inclined edges,
scratches, imprints, slots, and elevations interfere with the beam of light. At these
anomalies, the light is reflected towards the camera. Hence, these defects appear
bright in the camera image.
Backlighting Transmitted light illumination is the first choice of lighting when it
is necessary to measure parts as accurately as possible. The lighting is arranged on
the opposite side of the camera, the component itself is put in the light beam.
Example 4.5 Pixel Processing
Assume a vidicon tube generates a signal for 512 lines of a faceplate. If the sampling
capability of an ADC is 100 nanoseconds, i.e., the time required to process one pixel,
and 1/30 seconds are required to scan all 512 lines then scanning rate for each line is,
1
= 65.1 microsecond per line (4.11)
(30 ¥ 512)
Hence, the number of pixels which can be processed per line is given below:
65.1 ¥ 10 -6
= 651 pixels per line (4.12)
100 ¥ 10 -9
recognize an object within the image field. It is divided into several sub-processes,
which are discussed below:
Image Data Reduction Here, the objective is to reduce the volume of data.
As a preliminary step in the data analysis, the schemes like digital conversion or
windowing can be applied to reduce the data. While the digital conversion reduces
the number of gray levels used by the vision system, windowing involves using
only a portion of the total image stored in the frame buffer for image processing
and analysis. For example, in windowing, to inspect a circuit board, a rectangular
window is selected to surround the component of interest and only pixels within that
window are analyzed.
Histogram Analysis A histogram is a representation of the total number of pixels
of an image at each gray level. Histogram information is used in a number of different
processes, including thresholding. For example, histogram information can help in
determining a cut-off point when an image is to be transformed into binary values.
Thresholding It is the process of dividing an image into different portions or
levels by picking a certain grayness level as a threshold. Comparing each pixel value
with the threshold, and then assigning the pixel to the different portions or level,
depending on whether the pixel’s grayness level is below the threshold (‘off’ or 0, or
not belonging) or above the threshold (‘on’ or 1, or belonging).
Masking A mask may be used for many different purposes, e.g., filtering operations
and noise reduction, and others. It is possible to create masks that behave like a low-
pass filter such that higher frequencies of an image are attenuated while the lower
frequencies are not changed very much. This is illustrated in Example 4.8. Thereby,
the noise is reduced. Masking an image considers a portion of an imaginary image
shown in Fig. 4.23(a), which has all the pixels at a gray value of 20 except the one at
a gray level of 100. The one with 100 may be considered noise. Applying the 3 ¥ 3
mask shown in Fig. 4.23(b) over the corner of the image yields the following value:
(m1 A + m2 B + m3C + m4 E + m5 F + m6 G + m7 I + m8 J + m9 K )
X= = 29 (4.14)
S
where S ∫ m1 + m2 + � + m9 = 9, and values of A, B, …, and m1, m2, … are shown
in Figs. 4.23 (a) and (b), respectively. As a result of application of the mask, the pixel
with the value of 100 changes to 29, as indicated in Fig. 4.23(c). The large difference
between the noisy pixel and the surrounding pixels, i.e., 100 vs. 20, becomes much
smaller, namely, 29 vs. 20, thus reducing the noise. With this characteristic, the mask
acts as a low-pass filter. Note that the above reduction of noise has been achieved
using what is referred as neighborhood averaging, which causes the reduction of the
sharpness of the image as well.
Edge Detection Edge detection is a general name for a class of computer programs
and techniques that operate on an image and result in a line drawing of the image.
The lines represent changes in values such as cross section of planes, intersections
of planes, textures, lines, etc. In many edge-detection techniques, the resulting edges
are not continuous. However, in many applications, continuous edges are preferred,
which can be obtained using the Hough transform. It is a technique used to determine
the geometric relationship between different pixels on a line, including the slope
of the line. Consider a straight line in the xy-plane, as shown in Fig. 4.24, which is
expressed as
y = mx + c (4.15)
where m is the slope and c is the intercept. The line represented by Eq. (4.15) can
be transformed into a Hough plane of m – c with x and y as its slope and intercept,
respectively. Thus, a line in the xy-plane with a particular slope and intercept will
transform into a point in the Hough plane. For the Hough or m-c plane, if a line
passes through (m1, c1), as shown in Fig. 4.24(b), it represents a point (x1, y1) in the
x-y plane of Fig. 4.24(a). In that way, all lines through a point in the Hough plane
will transform into a single line in the x-y plane. Alternatively, if a group of points
is collinear, their Hough transform will all intersect. By examining these properties
in a particular case, it can be determined whether a cluster of pixels is on a straight
line or not, whereupon the orientation of an object in a plane can be determined by
calculating the orientation of a particular line in the object.
above, region growing and splitting, and others. While region growing works based
on the similar attributes, such as gray-level ranges or other similarities, and then try
to relate the regions by their average similarities, region splitting is carried out based
on thresholding in which an image is split into closed areas of neighborhood pixels
by comparing them with thresholding value or range.
Morphology Operations Morphology operations are a family of operations
which are applied on the shape of subjects in an image. They include many different
operations, both for binary and gray images, such as thickening, dilation, erosion,
skeletonization, opening, closing, and filing. These operations are performed on an
image in order to aid in its analysis, as well as to reduce the ‘extra’ information that
may be present in the image. For example, Fig. 4.25(a) shows the object which after
skeletonization is shown in Fig. 4.25(b).
Similarly, one can find n0,1 as the summation of y-coordinates which is same as
the summation of each pixel area multiplied by its distance from the x-axis. Hence, it
is similar to the first moment of the area relative to the x-axis. Then, the location of
the center of the area relative to the x-axis can be calculated immediately as
1 n0,1
y– = Â y= (4.19a)
Area n0,0
Following Eq. (4.19a), one can also find the location of the center of the area
relative to the y-axis as
1 n1,0
x– = Â x= (4.19b)
Area n0,0
In a similar manner, one can also find the second moment of area with respect to
x and y axes by putting p = 2; q = 0, and p = 0; q = 2, respectively. Note here that the
first moments are independent of orientation, whereas the second moments are not.
Object Recognition The next in image analysis is to identify the object that the
image represents based on the extracted features. The recognition algorithm should
be powerful enough to uniquely identify the object. Typical techniques used in the
industries are template matching and structural technique.
In template matching, the features of the object in the image, e.g., its area,
diameter, etc., are compared to the corresponding stored values. These values
constitute the stored template. When a match is found, allowing for certain statistical
variations in the comparison process, the object has been properly classified. For
example, the area of an image given by the moment n0,0 can be used to determine
nature of the object, and to distinguish it from other objects that have different
areas. Similarly, the first and second moments can be used to identify an object’s
position and orientation so that it can be picked up by the robot. Note that, instead of
calculating the moment values which change as the object moves, a look-up table can
be prepared. For example, the moment of inertia of an object varies significantly as
the object rotates about its center. Hence, the values of n2,0 and n0,2 will also change.
Since each orientation creates a unique value, a look-up table that contains these
values can later be used to identify the orientation of the object.
An important aspect here is that similar shapes,
as shown in Fig. 4.27 cannot be distinguished with
their first and second moments, as they are similar
or close to each other. In such a situation, a small
difference between the two objects may be
exaggerated through higher-order moments,
Fig. 4.27 Two similar objects
making identification of the object possible.
Structural techniques of pattern recognition rely on the relationship between
features or edges of an object. For example, if the image of an object can be subdivided
in four straight lines connected at their end points, and the connected lines are at right
angles then the object is a rectangle. This kind of technique is known as syntactic
pattern recognition.
In case depth information of an object is required, it can be obtained either by
using a range finder in conjunction with a vision system, or using a stereo vision. The
details of these systems are beyond the scope of this book. Readers are encouraged to
look at Haralick and Shapiro (1992) and others.
4.4.3 Hierarchy of a Vision System
The collection of processes involved in visual perception are often perceived as a
hierarchy spanning the range from ‘low’ via ‘intermediate’ to ‘high-level’ vision. The
notion of ‘low’ and ‘high’ vision are used routinely, but there is no clear definition of
Sensors, Vision and Signal Conditioning 103
user. Further grouping is essential at this stage since one may still need to be able
to decide which lines group together to form an object. To do this, it is necessary to
further distinguish lines which are part of the object structure, from those which are
part of a surface texture, or caused by shadows. High-level systems are, therefore,
object oriented, and sometimes called ‘top-down’. High-level visual processes are
applied to a selected portion of the image, rather than uniformly across the entire
image, as done in low- and intermediate-level visions. They almost always require
some form of knowledge about the objects of the scene to be included.
4.4.4 Difficulties in Vision and Remedies
A vision system cannot uniquely represent or process all available data because of
computational problem, memory, and processing-time requirements imposed on the
computer. Therefore, the system must compromise. Other problems include variation
of light, part-size, part placement, and limitations in the dynamic ranges available in
typical vision sensors. A vision system requires specialized hardware and software.
It is possible to purchase just the hardware with little or no vision application
programming. In fact, a few third-party programs are available. A hardware-only
approach is less expensive and can be more flexible for handling usual vision
requirements. But, since this approach requires image processing expertise, it is only
of interest to users who wish to retain the responsibility of image interpretation. It
is usual practice to obtain the hardware and application software together from the
supplier. However, the user might still need custom programming for an application.
Sensors, Vision and Signal Conditioning 105
Major vision system suppliers specialise in providing software for only a few
application areas.
Every vision system requires a sensor to cover the visual image into an electronic
signal. Several types of video sensors are used, including vidicon cameras, vacuum
tube devices, and solid-state sensors. Many of these vision systems were originally
designed for other applications, such as television so the signal must be processed to
extract the visual image and remove synchronization information before the signal
is sent to the computer for further processing. The computer then treats this digital
signal as the array pixels, and processes this data to extract the desired information.
Image processing can be very time consuming. For a typical sensor of 200,000
or more pixels, a vision system can take many seconds, even minutes, to analyze
the complete scene and determine the action to be taken. The number of bits to be
processed is quite large, for example, a system with 512 × 512 pixels array and an
8-bit intensity per pixel yields over two million of bits to be processed. If continuous
image at a 30 FPS frame rate were being received, data bytes would be received at
an 8 MHz rate. Few computers can accept inputs at these data rates and, in any case,
there would be no time left to process the data. When higher resolution system, color
system, or multiple camera systems are considered, data-handling requirements
become astronomical. Several methods can be used to reduce the amount of data
handled and, therefore, the processing time. They are explained as follows:
(a) One approach is the binary vision, which is used when only black-and-white
information is processed (intensity variations and shades of gray are ignored).
In binary vision, a picture is converted into a binary image by thresholding, as
illustrated in Fig. 4.29. In thresholding, a brightness level is selected. All data
with intensities equal to or higher than this value are considered white, and all
other levels are considered black.
(b) Another method of shortening process time is to control object placement so
that objects of interest cannot overlap in the image. Complicated algorithms
to separate images are then not necessary, and the image-processing time is
reduced.
(c) A third approach reduces data handling by processing only a small window of
the actual data; that is, the object is located in a predefined field of view. For
example, if the robot is looking for a mark on the printed circuit board, the
system can be held in such a way that the mark is always in the upper right
corner.
(d) A fourth approach takes a statistical sample of data and makes decisions on this
data sample. Unfortunately, all of these approaches ignore some of the available
data and, in effect, produce a less robust system. Processing time is saved, but
some types of complex objects cannot be recognized.
Note that some of the above steps are actually a part of image processing explained
in Section 4.4.2.
the voltages at the two input leads are nearly equal. If a large voltage differential vi
(say, 10 V) at the input then according to Eq. (4.22), the output voltage should be
extremely high. This never happens in practice, because the device saturates quickly
beyond moderate output voltages in the order of 15 V. From Eqs (4.22–4.23), it is
clear that if the –ve input lead is grounded, i.e., vin = 0 then
vo = Kvip (4.24a)
and if the +ve input lead is grounded, i.e., vip = 0 then
vo = –Kvin (4.24b)
This is the reason why vip is termed non-inverting input and vin is termed inverting
input.
4.5.2 Filters
A filter is a device that allows through only
Dynamic System vs. Analog Filter
the desirable part of a signal, rejecting
the unwanted part. Unwanted signals can The way an analog sensor is defined, any
dynamic system can be interpreted as an
seriously degrade the performance of a
analog filter.
robotic system. External disturbances,
error components in excitations, and noise generated internally within system
components and instrumentation are such spurious signals, which may be removed
by a filter. There are four broad categories of filters, namely, low-pass filters, high-
pass filters, band-pass filters, and band-reject (or notch) filters, which are shown in
Fig. 4.31.
1 1 1 1
the carrier signal is varied in proportion to the amplitude of the data signal
(modulating signal), while keeping the amplitude of the carrier signal constant. This
is called Frequency Modulation (FM). An FM is less susceptible to noise than AM.
In Pulse-Width Modulation (PWM), the carrier signal is a pulse sequence, and its
width is changed in proportion to the amplitude of the data signal, while keeping
the pulse spacing constant. This is shown in Fig. 4.32(d). The PWM signals can be
used directly in controlling a process without having to demodulate them. There
also exists Pulse-Frequency Modulation (PFM), where the carrier signal is a pulse
sequence. Here, the frequency of the pulses is changed in proportion to the value
of the data signal, while keeping the pulse width constant. It has the advantages of
ordinary FM.
Demodulation or detection is the process of extracting the original data signal
from a modulated signal. A simple and straightforward method of demodulation is
by detection of the envelope of the modulated signal.
4.5.4 Analog and Digital Conversions
Most sensors have analog output while much data processing is done using digital
computers. Hence, devices for conversion between these two domains have to be
performed. These can be achieved using an analog-to-digital converter (ADC) and a
digital-to-analog converter (DAC), as mentioned in Chapter 2. Explanation on DAC
is given first as some ADC uses DAC as its component.
1. Digital-to-Analog Converter (DAC) The function of a digital-to-analog-
convertor or DAC is to convert a sequence of digital words stored in a data register,
typically in the straight binary form, into an analog signal. A typical DAC unit is
an active circuit in the integrated circuit (IC) form and consists of a data register
(digital circuits), solid-state switching circuits, resistors, and op-amps powered by an
external power supply. The IC chip that represents the DAC is usually one of many
components mounted on a Printed Circuit Board (PCB), which is the I/O board or
card. This board is plugged into a slot of the PC having DAQ. A typical commercial
DAC is shown in Fig. 2.10.
2. Analog-to-Digital Converter (ADC) An analog-to-digital converter or ADC,
on the other hand, converts an analog signal into the digital form, according to an
appropriate code, before the same is used by a digital processor or a computer. The
process of analog-to-digital conversion is more complex and time consuming than
the digital-to-analog conversion. ADCs are usually more costly, and their conversion
rate is slower than DACs. Several types of ADCs are commercially available.
However, their principle of operation may vary depending on the type. A typical
commercial ADC is shown in Fig. 2.8.
Note that the most fundamental property of any DAC or ADC is their number of
bits for which it is designed, since this is a basic limit on resolution. Units of 8 to
12 bits are most common even though higher bits are also available. Both DAC and
ADC are elements in typical input/output board (or I/O board, or data acquisition
and control card, i.e., DAC or DAQ), which are usually situated on the same digital
interface board.
110 Introduction to Robotics
If one adds sine and cosine functions together, the original signal can be
reconstructed. Equation (4.25) is called a Fourier series, and the collection of
different frequencies present in the equation is called the frequency spectrum or
frequency content of the signal. Even though the signal is in amplitude-time domain,
the frequency spectrum is in the amplitude-frequency domain. For example, for the
function, f(t) = sin(t) of Fig. 4.34(a), which consists of only one frequency with
Sensors, Vision and Signal Conditioning 111
constant amplitude, the plotted signal would be represented by a single line at the
given frequency, as shown in Fig. 4.34(b). If the plot with given frequency and
amplitude is represented by the arrow in Fig. 4.34(b), the same sine function can be
reconstructed. The plots in Fig. 4.35 are similar and represent the following:
1
f(t) = Â sin(nt ) (4.26)
n =1,3,...,15 n
where w = 1.
Amplitude Amplitude
1 1
Time 1 Frequency
Amplitude Amplitude
Time 1 3 5 7 15 Frequency
of the mass that must be accelerated and decelerated by the drive motors of the wrist
and arm. So, it directly affects the performance of the robot. It is a challenge to
sensor designers to reduce size and weight. An early wrist force-torque sensor, for
example, was about 125 mm in diameter but was reduced to about 75 mm in diameter
through careful redesign.
12. Environmental Conditions Power requirement and its easy availability
should be considered. Besides, conditions like chemical reactions including
corrosion, extreme temperatures, light, dirt accumulation, electromagnetic field,
radioactive environments, shock and vibrations, etc., should be taken into account
while selecting a sensor or considering how to shield them.
13. Reliability and Maintainability Reliability is of major importance in all
robot applications. It can be measured in terms of Mean Time To Failure (MTTF) as
the average number of hours between failures that cause some part of the sensor to
become inoperative. In industrial use, the total robot system is expected to be available
as much as 98 or 99% of the working days. Since there are hundreds of components
in a robot system, each one must have a very high reliability. Some otherwise good
sensors cannot stand the daily environmental stress and, therefore, cannot be used
with robots. Part of the requirement for reliability is ease of maintenance. A sensor
that can be easily replaced does not have to be as reliable as one that is hidden in the
depths of the robot. Maintainability is a measure in terms of Mean Time To Repair
(MTTR).
14. Interfacing Interfacing of sensors with signal-conditioning devices and the
controller of the robot is often a determining factor in the usefulness of sensors.
Nonstandard plugs or requirements for nonstandard voltages and currents may make
a sensor too complex and expensive to use. Also, the signals from a sensor must be
compatible with other equipment being used if the system is to work properly.
15. Others Other aspects like initial cost, maintenance cost, cost of disposal and
replacement, reputation of manufacturers, operational simplicity, ease of availability
of the sensors and their spares should be taken into account. In many occasions, these
nontechnical considerations become the ultimate deciding factor in the selection of
sensors for an application.
SUMMARY
Necessity of sensors in robotics, different types of sensors used in robotic applications
and the selection criteria are presented in this chapter. Functional aspects, rather than
their mathematical models, are emphasized so that the reader is able to decide which
sensor to use and when. Different sensors are classified from their functional uses, for
example, position, velocity, acceleration, force, sensors, etc. Special attention has been
paid to vision system, its elements, how it can be processed, etc. Signal-conditioning
units, their characteristics, and when to use what are also explained. Finally guidelines
are provided to select appropriate sensors.
Sensors, Vision and Signal Conditioning 115
EXERCISES
4.1 Define sensitivity and linearity.
4.2 Distinguish accuracy from repeatability.
4.3 State the physical characteristics in sensor selection.
4.4 What are the essential components of a sensor?
4.5 Why are the terms internal and external used to classify sensors?
4.6 Classify internal sensors.
4.7 Name some velocity sensors?
4.8 Why are position sensors not preferred to obtain velocity and acceleration?
4.9 Is there any advantage of external sensors over internal types?
4.10 Name some contact and noncontact type sensors.
4.11 What are the advantages of capacitive proximity sensors?
4.12 What is a machine vision?
4.13 What are the major components in a vision system?
4.14 What is the sensor component in a vision system?
4.15 What are different types of cameras? Explain the vidicon camera.
4.16 What are the typical difficulties in a vision system?
4.17 What is thresholding in vision?
4.18 Why is signal processing important?
4.19 Name at least three devices used in signal conditioning.
4.20 Define different types of filters.
4.21 What are the characteristics one should check while selecting a sensor?
4.22 What is Hough transform?
4.23 What are the static characteristics a sensor should have?
WEB-BASED EXERCISES
Based on Web search, find the answers for the following questions:
4.24 Name at least five commercial companies who make different types of sensors. List the
sensors against each company’s name.
4.25 What are the typical prices of encoders, tachogenerators, and accelerometers?
4.26 What is the price of a CCD camera?
4.27 Name two camera companies and mention which camera you will select for industrial
robot system involved in accurate pick-and-place operation?
4.28 Name a robot manufacturing company who implements vision system in their robots.
4.29 Give a typical significance of a force/torque sensor.
4.30 What command MATLAB uses for digital filtering and image processing?