0% found this document useful (0 votes)
668 views204 pages

Robotics - Sensors and Methods For Autonomous Mobile Robot P

This document provides a summary of a technical report from the University of Michigan titled "Where am I? Sensors and Methods for Autonomous Mobile Robot Positioning". The report examines various sensors and techniques used for determining the position and orientation of autonomous mobile robots. It covers sensors and methods for dead reckoning, heading sensing, active beacons, and map-based positioning.

Uploaded by

Reginaldo Alves
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
668 views204 pages

Robotics - Sensors and Methods For Autonomous Mobile Robot P

This document provides a summary of a technical report from the University of Michigan titled "Where am I? Sensors and Methods for Autonomous Mobile Robot Positioning". The report examines various sensors and techniques used for determining the position and orientation of autonomous mobile robots. It covers sensors and methods for dead reckoning, heading sensing, active beacons, and map-based positioning.

Uploaded by

Reginaldo Alves
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 204

The University of Michigan

Volume III:

"Where am I?"
Sensors and Methods for Autonomous Mobile Robot Positioning
by

L. Feng , J. Borenstein 2, and H. R. Everett3


Edited and compiled by J. Borenstein

December 1994
Copies of this report are available from the University of Michigan as: Technical Report UM-MEAM-94-21

Prepared by the University of Michigan For the Oak Ridge National Lab (ORNL) D&D Program and the United States Department of Energy's Robotics Technology Development Program Within the Environmental Restoration, Decontamination and Dismantlement Project

1)

Dr. Liqiang Feng The University of Michigan Department of Mechanical Engineering and Applied Mechanics Mobile Robotics Laboratory 1101 Beal Avenue Ann Arbor, MI 48109 Ph.: (313) 936-9362 Fax: (313) 763-1260 Email: [email protected]

Dr. Johann Borenstein The University of Michigan Department of Mechanical Engineering and Applied Mechanics Mobile Robotics Laboratory 1101 Beal Avenue Ann Arbor, MI 48109 Ph.: (313) 763-1560 Fax: (313) 944-1113 Email:
[email protected]

2)

Commander H. R. Everett Naval Command, Control, and Ocean Surveillance Center RDT&E Division 5303 271 Catalina Boulevard San Diego CA 92152-5001 Ph.: (619) 553-3672 Fax: (619) 553-6188 Email: [email protected]

3)

Please direct all inquiries to Johann Borenstein


D:\WP\DOE_94\ORNL\POSITION.RPT\POSITION.WP6, February 25, 1995

Acknowledgments:
This work was done under the direction and on behalf of the

Department of Energy
Robotics Technology Development Program Within the Environmental Restoration, Decontamination, and Dismantlement Project.

Parts of this report were adapted from: H. R. Everett, "Sensors for Mobile Robots." A. K. Peters, Ltd., Wellesley, expected publication date Spring 1995.

The authors wish to thank Professors David K. Wehe and Yoram Koren for their support in preparing this report. The authors also wish to thank Dr. William R. Hamel, D&D Technical Coordinator and Dr. Linton W. Yarbrough, DOE Program Manager, for their continuing support in funding this report. The authors further wish to thank A. K. Peters, Ltd., for granting permission to publish (for limited distribution within Oak Ridge National Laboratories and the Department of Energy) selected parts of their soon-to-be published book "Sensors for Mobile Robots" by H. R. Everett. Thanks are also due to Todd Ashley Everett for making most of the line-art drawings, and to Photographer David A. Kother who shot most of the artful photographs on the front cover of this report. Last but not least, the authors are grateful to Mr. Brad Holt for proofreading the manuscript and providing many useful suggestions.

Table of Contents
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 1 Part I: Sensors for Mobile Robot Positioning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 5 Chapter 1: Sensors for Dead Reckoning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 7 1.1 Optical Encoders . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 8 1.1.1 Incremental Optical Encoders . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 8 1.1.2 Absolute Optical Encoders . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 10 1.2 Doppler Sensors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 12 1.2.1 Micro-Trak Trak-Star Ultrasonic Speed Sensor . . . . . . . . . . . . . . . . . . . . . . Page 13 1.2.2 Other Doppler Effect Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 13 1.3 Typical Mobility Configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 14 1.3.1 Differential Drive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 14 1.3.2 Tricycle Drive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 15 1.3.3 Ackerman Steering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 16 1.3.4 Synchro-Drive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 17 1.3.5 Omni-Directional Drive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 20 1.3.6 Multi-Degree-of Freedom Vehicles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 21 1.3.7 Tracked Vehicles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 22 Chapter 2: Heading Sensors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1 Gyroscopes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.1 Mechanical Gyroscopes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.1.1 Space-Stable Gyroscopes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.1.2 Gyrocompasses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.2 Optical Gyroscopes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.2.1 Active Ring Laser Gyros . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.2.2 Passive Ring Resonator Gyros . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.2.3 Open-Loop Interferometric Fiber Optic Gyros . . . . . . . . . . . . . . . . . . . . 2.1.2.4 Closed-Loop Interferometric Fiber Optic Gyros . . . . . . . . . . . . . . . . . . . 2.1.2.5 Resonant Fiber Optic Gyros . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 Geomagnetic Sensors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.1 Mechanical Magnetic Compasses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Dinsmore Starguide Magnetic Compass . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.2 Fluxgate Compasses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.2.1 Zemco Fluxgate Compasses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.2.2 Watson Gyro Compass . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.2.3 KVH Fluxgate Compasses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.3 Hall Effect Compasses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.4 Magnetoresistive Compasses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
ii

Page 24 Page 24 Page 24 Page 25 Page 26 Page 27 Page 28 Page 31 Page 32 Page 35 Page 35 Page 36 Page 37 Page 38 Page 39 Page 43 Page 45 Page 46 Page 47 Page 49

2.2.4.1 Philips AMR Compass . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 49 2.2.5 Magnetoelastic Compasses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 50 Chapter 3: Active Beacons . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.1 Navstar Global Positioning System (GPS) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2 Ground-Based RF Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2.1 Loran . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2.2 Kaman Sciences Radio Frequency Navigation Grid . . . . . . . . . . . . . . . . . . . 3.2.3 Precision Location Tracking and Telemetry System . . . . . . . . . . . . . . . . . . . 3.2.4 Motorola Mini-Ranger Falcon . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2.5 Harris Infogeometric System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 4: Sensors for Map-based Positioning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1 Time-of-Flight Range Sensors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.1 Ultrasonic TOF Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.1.1 National Semiconductors LM1812 Ultrasonic Transceiver . . . . . . . . . 4.1.1.2 Massa Products Ultrasonic Ranging Module Subsystems . . . . . . . . . . . 4.1.1.3 Polaroid Ultrasonic Ranging Modules . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.2 Laser-Based TOF Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.2.1 Schwartz Electro-Optics Laser Rangefinders . . . . . . . . . . . . . . . . . . . . . 4.1.2.2 RIEGL Laser Measurement Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2 Phase Shift Measurement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.1 ERIM 3-D Vision Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.2 Odetics Scanning Laser Imaging System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.3 ESP Optical Ranging System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.4 Acuity Research AccuRange 3000 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.5 TRC Light Direction and Ranging System . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3 Frequency Modulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.3.1 VRSS Automotive Collision Avoidance Radar . . . . . . . . . . . . . . . . . . . . . . . . 4.3.2 VORAD Vehicle Detection and Driver Alert System . . . . . . . . . . . . . . . . . . 4.3.3 Safety First Systems Vehicular Obstacle Detection and Warning System . . . 4.3.4 Millitech Millimeter Wave Radar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

Page 53 Page 53 Page 60 Page 60 Page 61 Page 62 Page 62 Page 64

Page 66 Page 66 Page 68 Page 68 Page 69 Page 71 Page 73 Page 73 Page 77 Page 82 Page 86 Page 89 Page 90 Page 91 Page 92 Page 94 Page 95 Page 96 Page 98 Page 98

Part II: Systems and Methods for Mobile Robot Positioning . . . . . . . . . . . . . . . . . . Page 100 Chapter 5: Dead-reckoning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1 Systematic and Non-systematic Dead-reckoning Errors . . . . . . . . . . . . . . . . . . . . 5.2 Reduction of Dead-reckoning Errors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2.1 Auxiliary Wheels and Basic Encoder Trailer . . . . . . . . . . . . . . . . . . . . . . . . 5.2.2 The Basic Encoder Trailer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2.3 Mutual Referencing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2.4 MDOF vehicle with Compliant Linkage . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2.5 Internal Position Error Correction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
iii

Page 102 Page 103 Page 104 Page 105 Page 105 Page 106 Page 106 Page 107

5.3 Automatic Vehicle Calibration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.4 Inertial Navigation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.4.1 Accelerometers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.4.2 Gyros . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 6: Active Beacon Navigation Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.1 Discussion on Triangulation Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2 Ultrasonic Transponder Trilateration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.1 IS Robotics 2-D Location System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.2 Tulane University 3-D Location System . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3 Optical Positioning Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.1 Cybermotion Docking Beacon . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.2 Hilare . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.3 NAMCO LASERNET . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.4 Intelligent Solutions EZNav Position Sensor . . . . . . . . . . . . . . . . . . . . . . . . 6.3.5 TRC Beacon Navigation System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.5 Siman Sensors & Intelligent Machines Ltd., "ROBOSENSE" . . . . . . . . . . . . 6.3.7 Imperial College Beacon Navigation System . . . . . . . . . . . . . . . . . . . . . . . . 6.3.8 MacLeod Technologies CONAC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.9 Lawnmower CALMAN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 7: Landmark Navigation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.1 Natural Landmarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.2 Artificial Landmarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3 Artificial Landmark Navigation Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3.1 MDARS Lateral-Post Sensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3.2 Caterpillar Self Guided Vehicle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.4 Line Navigation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 8: Map-based Positioning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.1 Map-building . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.1.1 Map-building and sensor-fusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.1.2 Phenomenological vs. geometric representation . . . . . . . . . . . . . . . . . . . . . . 8.2 Map matching . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.1 Schiele and Crowley [1994] . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.2 Hinkel and Knieriemen [1988] the Angle Histogram . . . . . . . . . . . . . . . . 8.2.3 Siemens' Roamer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3 Geometric and Topological Maps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.1 Geometric Maps for Navigation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.1.1 Cox [1991] . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
iv

Page 109 Page 110 Page 111 Page 111 Page 112

Page 113 Page 115 Page 116 Page 116 Page 117 Page 119 Page 119 Page 121 Page 122 Page 123 Page 124 Page 125 Page 126 Page 127 Page 128 Page 129

Page 130 Page 131 Page 131 Page 133 Page 134 Page 135 Page 135 Page 136

Page 138 Page 139 Page 140 Page 141 Page 141 Page 142 Page 144 Page 145 Page 147 Page 148 Page 148

8.3.1.2 Crowley [1989] . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.2 Topological Maps for Navigation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.2.1 Taylor [1991] . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.2.2 Courtney and Jain [1994] . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3.2.3 Kortenkamp and Weymouth [1993] . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

Page 150 Page 153 Page 153 Page 154 Page 154 Page 157

Part III: References and "Systems-at-a-Glance" Tables . . . . . . . . . . . . . . . . . . . . . . Page 158 References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 160

Systems-at-a-Glance Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Page 188

Introduction
Leonard and Durrant-Whyte [1991] summarized the problem of navigation by three questions: "where am I?", "where am I going?", and "how should I get there?" This report surveys the stateof-the-art in sensors, systems, methods, and technologies that aim at answering the first question, that is: robot positioning in its environment. Perhaps the most important result from surveying the vast body of literature on mobile robot positioning is that to date there is no truly elegant solution for the problem. The many partial solutions can roughly be categorized into two groups: relative and absolute position measurements. Because of the lack of a single, generally good method, developers of automated guided vehicles (AGVs) and mobile robots usually combine two methods, one from each category. The two categories can be further divided into the following sub-groups. Relative Position Measurements: 1. Dead-reckoning uses encoders to measure wheel rotation and/or steering orientation. Deadreckoning has the advantage that it is totally self-contained and it is always capable of providing the vehicle with an estimate of its position. The disadvantage of dead-reckoning is that the position error grows without bound unless an independent reference is used periodically to reduce the error [Cox, 1991]. 2. Inertial navigation uses gyroscopes and sometimes accelerometers to measure rate of rotation, and acceleration. Measurements are integrated once (or twice) to yield position. Inertial navigation systems also have the advantage that they are self-contained. On the downside, inertial sensor data drifts with time because of the need to integrate rate-data to yield position; any small constant error increases without bound after integration. Inertial sensors are thus unsuitable for accurate positioning over extended period of time. Another problem with inertial navigation is the high equipment cost. For example, highly accurate gyros, used in airplanes are inhibitively expensive. Very recently fiber-optics gyros (also called laser-gyros), which are said to be very accurate, have fallen dramatically in price and have become a very attractive solution for mobile robot navigation. Absolute Position Measurements: 3. Active beacons This methods computes the absolute position of the robot from measuring the direction of incidence of three or more actively transmitted beacons. The transmitters, usually using light or radio frequencies, must be located at known locations in the environment. 4. Artificial Landmark Recognition In this method distinctive artificial landmarks are placed at known locations in the environment. The advantage of artificial landmarks is that they can be designed for optimal detectability even under adverse environmental conditions. As with active beacons, three or more landmarks must be "in view" to allow position estimation. Landmark positioning has the advantage that the position errors are bounded, but detection of external landmarks and real-time position fixing may not always be possible. Unlike the usually pointPage 1

shaped beacons, Artificial Landmarks may be defined as a set of features, e.g., a shape or an area. Additional information, for example distance, can be derived from measuring the geometrical properties of the landmark, but this approach is computationally intensive and not very accurate. 5. Natural Landmark Recognition Here the landmarks are distinctive features in the environment. There is no need for preparations of the environment, but the environment must be known in advance. The reliability of this method is not as high as with artificial landmarks. 6. Model matching In this method information acquired from the robot's on-board sensors is compared to a map or world model of the environment. If features from the sensor-based map and the world model map match, then the vehicle's absolute location can be estimated. Mapbased positioning often includes improving global maps based on the new sensory observations in a dynamic environment and integrating local maps into the global map to cover previously unexplored area. The maps used in navigation include two major types: geometric maps and topological maps. Geometric maps represent the world in a global coordinate system, while topological maps represent the world as a network of nodes and arcs. The nodes of the network are distinctive places in the environment and the arcs represent paths between places [Kortenkamp and Weymouth, 1994]. There are large variations in terms of the information stored for each arc. Brooks [Brooks, 1985] argues persuasively for the use of topological maps as a means of dealing with uncertainty in mobile robot navigation. Indeed, the idea of a map that contains no metric or geometric information, but only the notion of proximity and order, is enticing because such an approach eliminates the inevitable problems of dealing with movement uncertainty in mobile robots. Movement errors do not accumulate globally in topological maps as they do in maps with a global coordinate system since the robot only navigate locally, between places. Topological maps are also much more compact in their representation of space, in that they represent only certain places and not the entire world [Kortenkamp and Weymouth, 1994]. However, this also makes a topological map unsuitable for any spatial reasoning over its entire environment, e.g., optimal global path planning. In the following survey we present and discuss the state-of-the-art in each one of the above categories. We compare and analyze different methods based on technical publications and on commercial product and patent information. Mobile robot navigation is a very diverse area, and a useful comparison of different approaches is difficult because of the lack of a commonly accepted test standards and procedures. The equipment used varies greatly and so do the key assumptions used in different approaches. Further difficulty arises from the fact that different systems are at different stages in their development. For example, one system may be commercially available, while another system, perhaps with better performance, has been tested only under a limited set of laboratory conditions. Our comparison will be centered around the following criteria: accuracy of position and orientation measurements, equipment needed, cost, sampling rate, effective range, computational power required, processing needed, and other special features. We present this survey in three parts. Part I deals with the sensors used in mobile robot positioning, while Part II discusses the methods and techniques that use these sensors. The report is organized in 9 chapters. Part I: Sensors for Mobile Robot Positioning
Page 2

1. Sensors for Dead-reckoning 2. Heading Sensors 3. Active Beacons 4. Sensors for Map-based Positioning Part II: Systems and Methods for Mobile Robot Positioning 5. Reduction of Dead-reckoning Errors 6. Active Beacon Navigation Systems 7. Landmark Navigation 8. Map-based positioning 9. Other Types of Positioning Part III: References and Systems-at-a-Glance Tables

Page 3

Part I: Sensors for Mobile Robot Positioning

CARMEL, the University of Michigan's first mobile robot, has been in service since 1987. Since then, CARMEL has served as a reliable testbed for countless sensor systems. In the extra "shelf" underneath the robot is an 8086 XT compatible singleboard computer that runs U of M's ultrasonic sensor firing algorithm. Since this code was written in 1987, the computer has been booting up and running from floppy disk. The program was written in FORTH and was never altered: Should anything ever go wrong with the floppy, it will take a computer historian to recover the code... Page 5

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

Chapter 1: Sensors for Dead Reckoning

Dead reckoning (derived from deduced reckoning of sailing days) is a simple mathematical procedure for determining the present location of a vessel by advancing some previous position through known course and velocity information over a given length of time [Dunlap & Shufeldt, 1972]. The vast majority of land-based mobile robotic systems in use today rely on dead reckoning to form the very backbone of their navigation strategy, and like their nautical counterparts, periodically null out accumulated errors with recurring fixes from assorted navigation aids. The most simplistic implementation of dead reckoning is sometimes termed odometry; the term implies vehicle displacement along the path of travel is directly derived from some onboard odometer. A common means of odometry instrumentation involves optical encoders directly coupled to the motor armatures or wheel axles. Since most mobile robots rely on some variation of wheeled locomotion, a basic understanding of sensors that accurately quantify angular position and velocity is an important prerequisite to further discussions of odometry. There are a number of different types of rotational displacement and velocity sensors in use today: Brush Encoders Potentiometers Synchros Resolvers Optical Encoders Magnetic Encoders Inductive Encoders Capacitive Encoders. A multitude of issues must be considered in choosing the appropriate device for a particular application. Avolio [1993] points out that over 17 million variations on rotary encoders are offered by one company alone. For mobile robot applications incremental and absolute optical encoders are the most popular type. We will discuss those in the following sections.

Page 7

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

1.1 Optical Encoders


The first optical encoders were developed in the mid-1940s by the Baldwin Piano Company for use as tone wheels that allowed electric organs to mimic other musical instruments [Agent, 1991]. Todays contemporary devices basically embody a miniaturized version of the break-beam proximity sensor. A focused beam of light aimed at a matched photodetector is periodically interrupted by a coded opaque/transparent pattern on a rotating intermediate disk attached to the shaft of interest. The rotating disk may take the form of chrome on glass, etched metal, or photoplast such as Mylar [Henkel, 1987]. Relative to the more complex alternating-current resolvers, the straightforward encoding scheme and inherently digital output of the optical encoder results in a lowcost reliable package with good noise immunity. There are two basic types of optical encoders: incremental and absolute. The incremental version measures rotational velocity and can infer relative position, while absolute models directly measure angular position and infer velocity. If non volatile position information is not a consideration, incremental encoders generally are easier to interface and provide equivalent resolution at a much lower cost than absolute optical encoders. 1.1.1 Incremental Optical Encoders The simplest type of incremental encoder is a single-channel tachometer encoder, basically an instrumented mechanical light chopper that produces a certain number of sine or square wave pulses for each shaft revolution. Adding pulses increases the resolution (and subsequently the cost) of the unit. These relatively inexpensive devices are well suited as velocity feedback sensors in medium-to high-speed control systems, but run into noise and stability problems at extremely slow velocities due to quantization errors [Nickson, 1985]. The tradeoff here is resolution versus update rate: improved transient response requires a faster update rate, which for a given line count reduces the number of possible encoder pulses per sampling interval. A typical limitation for a 5 cm (2-inch) diameter incremental encoder disk is 2540 lines [Henkel, 1987]. In addition to low-speed instabilities, single-channel tachometer encoders are also incapable of detecting the direction of rotation and thus cannot be used as position sensors. Phase-quadrature incremental encoders overcome these problems by adding a second channel, displaced from the first, so the resulting pulse trains are 90o out of phase as shown in Fig. 1.1. This technique allows the decoding electronics to determine which channel is leading the other and hence ascertain the direction of rotation, with the added benefit of increased resolution. Holle [1990] provides an indepth discussion of output options (single-ended TTL or differential drivers) and various design issues (i.e., resolution, bandwidth, phasing, filtering) for consideration when interfacing phasequadrature incremental encoders to digital control systems.

Page 8

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

State I S1 A S2 S3 B S4 1 2 3 4

Ch A High High Low Low

Ch B Low High High Low

Figure 1.1: The observed phase relationship between Channel A and B pulse trains can be used to determine the direction of rotation with a phase-quadrature encoder, while unique output states S1 - S4 allow for up to a four-fold increase in resolution. The single slot in the outer track generates one index pulse per disk rotation [Everett, 1995].

The incremental nature of the phase-quadrature output signals dictates that any resolution of angular position can only be relative to some specific reference, as opposed to absolute. Establishing such a reference can be accomplished in a number of ways. For applications involving continuous 360-degree rotation, most encoders incorporate as a third channel a special index output that goes high once for each complete revolution of the shaft (see Fig. 1.1 above). Intermediate shaft positions are then specified by the number of encoder up counts or down counts from this known index position. One disadvantage of this approach is that all relative position information is lost in the event of a power interruption. In the case of limited rotation, such as the back-and-forth motion of a pan or tilt axis, electrical limit switches and/or mechanical stops can be used to establish a home reference position. To improve repeatability this homing action is sometimes broken into two steps. The axis is rotated at reduced speed in the appropriate direction until the stop mechanism is encountered, whereupon rotation is reversed for a short predefined interval. The shaft is then rotated slowly back into the stop at a specified low velocity from this designated start point, thus eliminating any variations in inertial loading that could influence the final homing position. This two-step approach can usually be observed in the power-on initialization of stepper-motor positioners for dot-matrix printer heads. Alternatively, the absolute indexing function can be based on some external referencing action that is decoupled from the immediate servo-control loop. A good illustration of this situation involves an incremental encoder used to keep track of platform steering angle. For example, when the K2A Navmaster [CYBERMOTION] robot is first powered up, the absolute steering angle is unknown, and must be initialized through a referencing action with the docking beacon, a nearby wall, or some other identifiable set of landmarks of known orientation. The up/down count output from the decoder electronics is then used to modify the vehicle heading register in a relative fashion. A growing number of very inexpensive off-the-shelf components have contributed to making the phase-quadrature incremental encoder the rotational sensor of choice within the robotics research and development community. Several manufacturers now offer small DC gear motors with
Page 9

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

incremental encoders already attached to the armature shafts. Within the U.S. automated guided vehicle (AGV) industry, however, resolvers are still generally preferred over optical encoders for their perceived superiority under harsh operating conditions, but the European AGV community seems to clearly favor the encoder [Manolis, 1993]. Interfacing an incremental encoder to a computer is not a trivial task. A simple state-based interface as implied in Fig. 1.1 is inaccurate if the encoder changes direction at certain positions, and false pulses can result from the interpretation of the sequence of state-changes [Pessen, 1989]. Pessen describes an accurate circuit that correctly interprets directional state-changes. This circuit was originally developed and tested by Borenstein [1987]. A more versatile encoder interface is the HCTL 1100 motion controller chip made by Hewlett Packard [HP]. The HCTL chip performs not only accurate quadrature decoding of the incremental wheel encoder output, but it provides many important additional functions, among others + Closed-loop position control + Closed-loop velocity control in P or PI fashion + 24-bit position monitoring. At the University of Michigan's Mobile Robotics Lab, the HCTL 1100 has been tested and used in many different mobile robot control interfaces. The chip has proven to work reliably and accurately, and it is used on commercially available mobile robots, such as TRC LabMate and HelpMate. The HCTL 1100 costs only $40 and it comes highly recommended.

1.1.2 Absolute Optical Encoders Absolute encoders are typically used for slower rotational applications that require positional information when potential loss of reference from power interruption cannot be tolerated. Discrete detector elements in a photovoltaic array are individually aligned in break-beam fashion with concentric encoder tracks as shown in Fig. 1.2, creating in effect a non-contact implementation of a commutating brush encoder. The assignment of a dedicated track for each bit of resolution results in larger size disks (relative to incremental designs), with a corresponding decrease in shock and vibration tolerance. A general rule of thumb is that each additional encoder track doubles the resolution but quadruples the cost [Agent, 1991]. Instead of the serial bit streams of incremental designs, absolute optical encoders provide a parallel word output with a unique code pattern for each quantized shaft position. The most common coding schemes are Gray code, natural binary, and binary-coded decimal [Avolio, 1993]. The Gray code (for inventor Frank Gray of Bell Labs) is characterized by the fact that only one bit changes at a time, a decided advantage in eliminating asynchronous ambiguities caused by electronic and mechanical component tolerances. Binary code, on the other hand, routinely involves multiple bit changes when incrementing or decrementing the count by one. For example, when going from position 255 to position 0 in Fig. 1.3B, eight bits toggle from 1s to 0s. Since there is no guarantee all threshold detectors monitoring the detector elements tracking each bit will toggle at the same
Page 10

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

precise instant, considerable ambiguity can exist during state transition with a coding scheme of this form. Some type of handshake line signaling valid data available would be required if more than one bit were allowed to change between consecutive encoder positions.
Detector Array

LED Source

Beam Expander

Collimating Lens

Cylindrical Lens

Multi-track Encoder Disk

Figure 1.2: A line-source of light passing through a coded pattern of opaque and transparent segments on the rotating encoder disk results in a parallel output that uniquely specifies the absolute angular position of the shaft (adapted from [Agent, 1991]).

Absolute encoders are best suited for slow and/or infrequent rotations such as steering angle encoding, as opposed to measuring high-speed continuous (i.e., drive wheel) rotations as would be required for calculating displacement along the path of travel. Although not quite as robust as resolvers for high-temperature, high-shock applications, absolute encoders can operate at temperatures over 125o Celsius, and medium-resolution (1000 counts per revolution) metal or Mylar disk designs can compete favorably with resolvers in terms of shock resistance [Manolis, 1993]. A potential disadvantage of absolute encoders is their parallel data output, which requires a more complex interface due to the large number of electrical leads. A 13-bit absolute encoder using complimentary output signals for noise immunity would require a 28-conductor cable (13 signal pairs plus power and ground), versus only 6 for a resolver or incremental encoder [Avolio, 1993].

Figure 1.3: Rotating an 8-bit absolute Gray code disk (A) counterclockwise by one position increment will cause only one bit to change, whereas the same rotation of a binary-coded disk (B) will cause all bits to change in the particular case (255 to 0) illustrated by the reference line at 12 oclock [Everett, 1995]. Page 11

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

1.2 Doppler Sensors


The rotational displacement sensors discussed above derive navigation parameters directly from wheel rotation, and are thus subject to problems arising from slippage, tread wear, and/or improper tire inflation. In certain applications, Doppler and inertial navigation techniques are sometimes employed to reduce the effects of such error sources. Doppler navigation systems are routinely employed in maritime and aeronautical applications to yield velocity measurements with respect to the earth itself, thus eliminating dead-reckoning errors introduced by unknown ocean or air currents. The principle of operation is based on the Doppler shift in frequency observed when radiated energy reflects off a surface that is moving with respect to the emitter. Maritime systems employ acoustical energy reflected from the ocean floor, while airborne systems sense microwave RF energy bounced off the surface of the earth. Both configurations typically involve an array of four transducers spaced 90 o apart in azimuth and inclined downward at a common angle with respect to the horizontal plane [Dunlap & Shufeldt, 1972]. Due to cost constraints and the reduced likelihood of transverse drift, most robotic implementations employ but a single forward-looking transducer to measure ground speed in the direction of travel. Similar configurations are sometimes used in the agricultural industry, where tire slippage in soft freshly plowed dirt can seriously interfere with the need to release seed or fertilizer at a rate commensurate with vehicle advance. The M113-based Ground Surveillance Vehicle [Harmon, 1986] employed an off-the-shelf unit of this type manufactured by John Deere to compensate for track slippage. The microwave radar sensor is aimed downward at a prescribed angle (typically 45o) to sense ground movement as shown in Fig. 1.4. Actual ground speed V A is derived from the measured velocity VD according to the following equation [Schultz, 1993]: VA = VD c FD = cos 2 Fo cos actual ground velocity along path measured Doppler velocity angle of declination speed of light observed Doppler shift frequency transmitted frequency.

where VA = VD = " = c = FD = FO =

V Vt
Figure 1.4: A Doppler ground speed sensor inclined at an angle  as shown measures the velocity component VD of true ground speed VA (adapted from [Schultz, 1993]).

Errors in detecting true ground speed arise due to side-lobe interference, vertical velocity components introduced by vehicle reaction to road surface anomalies, and uncertainties in the actual angle of incidence due to the finite width

Page 12

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

of the beam. Byrne et al. [1992] point out another interesting scenario for potentially erroneous operation, involving a stationary vehicle parked over a stream of water. The Doppler ground-speed sensor in this case would misinterpret the relative motion between the stopped vehicle and the running water as vehicle travel.

1.2.1 Micro-Trak Trak-Star Ultrasonic Speed Sensor One commercially available speed sensor that is based on Doppler speed measurements is the Trak-Star Ultrasonic Speed Sensor [MICRO-TRAK]. This device, originally designed for agricultural applications, costs $420. The manufacturer claims that this is the most accurate Doppler speed sensor available. The technical specifications are listed in Table 1.1. 1.2.2 Other Doppler Effect Systems Harmon [1986] describes a system using a Doppler effect sensor, and a Doppler effect system based on radar is described in [Patent 1].
Fig. 1.5: The Trak-Star Ultrasonic Speed Sensor is based on the Doppler effect. This device is primarily targeted at the agricultural market Reproduced from [Micro-Trak].

Another Doppler Effect device is the Monitor 1000, a distance and speed monitor for runners. This device was temporarily marketed by the sporting goods manufacturer [NIKE]. The Monitor 1000 was worn by the runner like a front-mounted fanny pack. The small and lightweight device used ultrasound as the carrier, and was said to have an accuracy of 2-5%, depending on the ground characteristics. The manufacturer of the Monitor 1000 is Applied Design Laboratories [ADL]. A microwave radar Doppler effect distance sensor has also been developed by ADL. This radar sensor is a prototype and is not commercially available. However, it differs from the Monitor 1000 only in its use of a radar sensor head as opposed to the ultrasonic sensor head used by the Monitor 1000. The prototype radar sensor measures 15105 cm (6"4"2"), weighs 250 gr, and consumes 0.9 W.
Specification Speed range Speed Resolution Accuracy Transmit Frequency Temperature range Weight Power Requirements Value 0-40 MPH (17.7 m/s) 0.04 MPH (1.8 cm/s) 1.5%+0.04 MPH 62.5 KHz -20o F to 120o F 3 lb 12 Volt [email protected] Amp

1.3 Typical Mobility Configurations


The accuracy of odometry measurements for dead-reckoning is to a great extent a direct function of the kinematic design of a vehicle. Because of this close relation between kinematic design and positioning accuracy, one must consider

Table 1.1: Specifications for the Trak-Star Ultrasonic Speed Sensor Page 13

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

the kinematic design closely before attempting to improve dead-reckoning accuracy. For this reason, we will briefly discuss some of the more popular vehicle designs in the following sections. In Part II of this report, we will discuss some recently developed methods for reducing dead-reckoning errors (or the feasibility of doing so) for some of these vehicle designs. 1.3.1 Differential Drive Figure 1.6 shows a typical differential drive mobile robot, the LabMate platform, manufactured by [TRC]. In this design incremental encoders are mounted onto the two drive motors to count the wheel revolutions. The robot can perform dead reckoning by using simple geometric equations to compute the momentary position of the vehicle relative to a known starting position. For completeness, we rewrite the wellknown equations for dead-reckoning below (also, see [Klarer, 1988; Crowley and Reignier, 1992]).

dead re05.ds 4, .wmf, 10/19/94

Figure 1.6: A typical differential-drive mobile robot (bottom view).

Suppose that at sampling interval I the left and right wheel encoders show a pulse increment of NL and NR, respectively. Suppose further that cm = BDn/nCe where cm = Dn = Ce = n = Conversion factor that translates encoder pulses into linear wheel displacement Nominal wheel diameter (in mm) Encoder resolution (in pulses per revolution) Gear ratio of the reduction gear between the motor (where the encoder is attached) and the drive wheel.

We can compute the incremental travel distance for the left and right wheel, )UL,i and )UR,i , according to

)UL/R, I = cm NL/R, I
and the incremental linear displacement of the robot's centerpoint C, denoted )Ui , according to

)Ui = ()UR + )UL)/2


Next, we compute the robot's incremental change of orientation

)2i = ()UR - )UL)/b


Page 14

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

where b is the wheelbase of the vehicle, ideally measured as the distance between the two contact points between the wheels and the floor. The robot's new relative orientation 2i can be computed from

2i = 2i-1 + )2i
and the relative position of the centerpoint is xi = xi-1 + )Ui cos2i yi = yi-1 + )Ui sin2i where xi, yi - relative position of the robot's centerpoint c at instant I. 1.3.2 Tricycle Drive Tricycle drive configurations (see Fig. 1.7) employing a single driven front wheel and two passive rear wheels (or vice versa) are fairly common in AGV applications because of their inherent simplicity. For odometry instrumentation in the form of a steering angle encoder, the dead reckoning solution is equivalent to that of an Ackerman-steered vehicle, where the steerable wheel replaces the imaginary center wheel discussed in Section 1.3.3. Alternatively, if rear-axle differential odometry is used to determine heading, the solution is identical to the differential-drive configuration discussed in Section 1.3.1. One problem associated with the tricycle drive configuration is the vehicles center of gravity tends to move away from the front wheel when traversing up an incline, causing a loss of traction. As in the case of Ackerman-steered designs, some surface damage and induced heading errors are possible when actuating the steering while the platform is not moving.

Page 15

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

Steerable Driven Wheel

l d

Y X

Passive Wheels
Figure 1.7: Tricycle-drive configurations employing a steerable driven wheel and two passive trailing wheels can derive heading information directly from a steering angle encoder or indirectly from differential odometry [Everett, 1995].

1.3.3 Ackerman Steering Used almost exclusively in the automotive industry, Ackerman steering is designed to ensure the inside front wheel is rotated to a slightly sharper angle than the outside wheel when turning, thereby eliminating geometrically induced tire slippage. As seen in Fig. 1.8, the extended axes for the two front wheels intersect in a common point that lies on the extended axis of the rear axle. The locus of points traced along the ground by the center of each tire is thus a set of concentric arcs about this centerpoint of rotation P1, and (ignoring for the moment any centrifugal accelerations) all instantaneous velocity vectors will subsequently be tangential to these arcs. Such a o SA i steering geometry is said to satisfy the Ackerman equation [Byrne et al, 1992]: P 2 d cot i cot o = l where 2i = relative steering angle of inner wheel 2o = relative steering angle of outer wheel l = longitudinal wheel separation d = lateral wheel separation.

l d
Y

P 1
Figure 1.8: In an Ackerman-steered vehicle, the extended axes for all wheels intersect in a common point (adapted from [Byrne et al, 1992]). Page 16

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

For the sake of convenience, the vehicle steering angle 2 SA can be thought of as the angle (relative to vehicle heading) associated with an imaginary center wheel located at a reference point P 2 as shown in the figure above. 2SA can be expressed in terms of either the inside or outside steering angles (2i or 2o) as follows [Byrne et al, 1992]: cot SA = d + cot i 2l

or, alternatively, cot SA = cot o d 2l .

Ackerman steering provides a fairly accurate dead-reckoning solution while supporting the traction and ground clearance needs of all-terrain operation, and is generally the method of choice for outdoor autonomous vehicles. Associated drive implementations typically employ a gasoline or diesel engine coupled to a manual or automatic transmission, with power applied to four wheels through a transfer case, differential, and a series of universal joints. A representative example is seen in the HMMWV-based prototype of the USMC Tell-operated Vehicle (TOV) Program [Aviles et al, 1990]. From a military perspective, the use of existing-inventory equipment of this type simplifies some of the logistics problems associated with vehicle maintenance. In addition, reliability of the drive components is high due to the inherited stability of a proven power train. (Significant interface problems can be encountered, however, in retrofitting off-the-shelf vehicles intended for human drivers to accommodate remote or computer control). 1.3.4 Synchro-Drive An innovative configuration known as synchro-drive features three or more wheels (Fig. 1.9) mechanically coupled in such a way that all rotate in the same direction at the same speed, and similarly pivot in unison about their respective steering axes when executing a turn. This drive and steering synchronization results in improved dead-reckoning accuracy through reduced slippage, since all wheels generate equal and parallel force vectors at all times. The required mechanical synchronization can be accomplished in a number of ways, the most common being chain, belt, or gear drive. Carnegie Mellon University has implemented an electronically synchronized version on one of their Rover series robots, with dedicated drive motors for each of the three wheels. Chain- and belt-drive configurations experience some degradation in steering accuracy and alignment due to uneven distribution of slack, which varies as a function of loading and direction of rotation. In addition, whenever chains (or timing belts) are tightened to reduce such slack, the individual wheels must be realigned. These problems are eliminated with a completely enclosed gear-drive approach. An enclosed gear train also significantly reduces noise as well as particulate generation, the latter being very important in clean-room applications.

Page 17

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning


Upper Torso Rotation Shaft Steering Sprocket Power Sprocket

Steering Chain Wheel (Foot) Drive Chain

Steering Motor Shaft

Drive Motor Shaft

Figure 1.9: Bottom (A) and top (B) views of a four-wheel synchro-drive configuration (adapted from [Holland, 1983]).

An example of a three-wheeled belt-drive implementation is seen in the Denning Sentry formerly manufactured by Denning Mobile Robots, Woburn, MA [Kadonoff, 1985]. Referring to Fig. 1.9, drive torque is transferred down through the three steering columns to polyurethane-filled rubber tires. The drive-motor output shaft is mechanically coupled to each of the steering-column power shafts by a heavy-duty timing belt to ensure synchronous operation. A second timing belt transfers the rotational output of the steering motor to the three steering columns, allowing them to synchronously pivot throughout a full 360o range [Everett, 1985]. The Sentrys upper head assembly is mechanically coupled to the steering mechanism in a manner similar to that illustrated in Fig. 1.9, and thus always points in the direction of forward travel. The three-point configuration ensures good stability and traction, while the actively driven large-diameter wheels provide more than adequate obstacle climbing capability for indoor scenarios. The disadvantage of this particular implementation include dead-reckoning errors introduced by compliance in the drive belts as well as by reactionary frictional forces exerted by the floor surface when turning in place. To overcome these problems, the Cybermotion K2A Navmaster robot employs an enclosed geardrive configuration with the wheels offset from the steering axis as shown in Fig. 1.10 and Fig. 1.11. When a foot pivots during a turn, the attached wheel rotates in the appropriate direction to minimize floor and tire wear, power consumption, and slippage. Note that for correct compensation, the miter gear on the wheel axis must be on the opposite side of the power shaft gear from the wheel as illustrated. The governing equation for minimal slippage is [Holland, 1983] A r' = B r

Page 18

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

Figure 1.10: The Denning Sentry (foreground) incorporates a three-point synchrodrive configuration with each wheel located directly below the pivot axis of the associated steering column. In contrast, the Cybermotion K2A (background) has wheels that swivel around the steering column. Both robots were extensively tested at the University of Michigan's Mobile Robotics Lab.

where A = number of teeth on the power shaft gear B = number of teeth on the wheel axle gear r = wheel offset from steering pivot axis r = wheel radius.

Power Shaft

A 90 Miter Gear

One drawback of this approach is seen in the decreased lateral stability that results when one wheel is turned in under the vehicle. Cybermotions improved K3A design solves this problem (with an even smaller wheelbase) by incorporating a dualwheel arrangement on each foot [Fisher, et al, 1994]. The two wheels turn in opposite directions in differential fashion as the foot pivots during a turn, but good stability is maintained in the forgoing example by the outward swing of the additional wheel.

r'
Figure 1.11: Slip compensation during a turn is accomplished through use of an offset foot assembly on the three-wheeled K2A Navmaster robot (adapted from [Holland, 1983]).

Page 19

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

The dead-reckoning calculations for synchro-drive are almost trivial; vehicle heading is simply derived from the steering angle encoder, while displacement in the direction of travel is given as follows:
D= 2 N Re Ce

where D = vehicle displacement along path N = measured counts of drive motor shaft encoder Ce = encoder counts per complete wheel revolution Re = effective wheel radius. 1.3.5 Omni-Directional Drive The dead-reckoning solution for most multiple-degree-of-freedom (MDOF) configurations is done in similar fashion to that for differential drive, with position and velocity data derived from the motor (or wheel) shaft-encoders. For the three-wheel example illustrated in Fig. 1.12, the equations of motion relating individual motor speeds to velocity components V x and V y in the reference frame of the vehicle are given by [Holland, 1983]
V1 = 1r = Vx + p R V2 = 2 r = 0. 5 Vx + 0. 867 Vy + p R V3 = 3r = 0 . 5 Vx 0 . 867 Vy + p R

where V1 = tangential velocity of wheel number 1 V2 = tangential velocity of wheel number 2 V3 = tangential velocity of wheel number 3 T1 = rotational speed of motor number 1 T2 = rotational speed of motor number 2 T3 = rotational speed of motor number 3 Tp = rate of base rotation about pivot axis Tr = effective wheel radius TR = effective wheel offset from pivot axis.

Page 20

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

Motor 1 Forward Top View Of Ba se

Motor 3

Motor 2

Figure 1.12: (A) Schematic of the wheel assembly used by the Veterans Administration [La et al., 1981] on an omnidirectional wheelchair; (B) Top view of base showing relative orientation of components in the three-wheel configuration (adapted from [Holland, 1983]).

1.3.6 Multi-Degree-of Freedom Vehicles Multi-degree-of-freedom (MDOF) vehicles have multiple drive and steer motors. Different designs are possible. For example, HERMIES III, a sophisticated platform designed and built at the Oak Ridge National Laboratory [Pin et al., 1989; Reister et al., 1991; Reister, 1991] has two powered wheels that are also individually steered (see Fig. 1.13). With four independent motors, HERMIES-III is a 4-degree-of-freedom vehicle. MDOF configurations display exceptional maneuverability in tight quarters in comparison to conventional 2DOF mobility systems, but have been found to be difficult to control due to their overconstrained nature [Reister et al., 1991; Killough & Pin, 1992; Pin and Killough, 1994; Borenstein, 1994c]. Resulting problems include increased wheel slippage and thus reduced dead-reckoning accuracy. Recently, Reister and Unseren [1992; 1993] introduced a new control algorithm based on Force Control. The researchers reported on a substantial reduction in wheel slippage for their two-wheel drive/two-wheel steer platform, resulting in a reported 20-fold improvement of accuracy. However, the experiments on which these results were based avoided simultaneous steering and driving of the two steerable drive wheels. This way, the critical
Page 21

Castors

V1
Drive wheels

V2

T1

T2

Castors
mdof 01. ds4, mdof01.wmf, 5/19/94

Figure 1.13: A 4-degree-of-freedom vehicle platform can travel in all directions, including sideways and diagonally. The difficulty lies in coordinating all four motors such as to avoid slippage.

Part I: Sensors for Mobile Robot Positioning

Chapter 1: Sensors for Dead Reckoning

problem of coordinating the control of all four motors simultaneously and during transients was completely avoided.

Unique Mobility, Inc., built an 8-DOF vehicle, shown in Fig. 1.14. In personal correspondence engineers from that company mentioned to us difficulties in controlling and coordinating all eight motors. 1.3.7 Tracked Vehicles Yet another drive configuration for mobile robots uses tracks instead of wheels. This very special implementation of a differential drive is known as skid-steering, and is routinely implemented in track form on bulldozers and armored vehicles. Such skid-steer configurations intentionally rely on track or wheel slippage for normal operation (Fig. 1.15), and as a consequence provide rather poor dead-reckoning information. For this reason, skid-steering is generally employed only in teleoperated as opposed to autonomous robotic applications, where the ability to surmount significant floor discontinuities is more desirable than accurate dead-reckoning information. An

Figure 1.14: An 8-DOF platform with 4 wheels individually driven and steered. This platform was designed and built by Unique Mobility Inc (Courtesy [UNIQUE]) Page 22

Part I: Sensors for Mobile Robot Positioning



Chapter 1: Sensors for Dead Reckoning

example is seen in the track drives popular with remote-controlled robots intended for explosive ordnance disposal.

dmin
Track Footprint

dmax

Figure 1.15: The effective point of contact for a skidsteer vehicle is roughly constrained on either side by a rectangular zone of ambiguity corresponding to the track footprint. As is implied by the concentric circles, considerable slippage must occur in order for the vehicle to turn [Everett, 1995].

Page 23

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

Chapter 2: Heading Sensors


Heading sensors are of particular importance to mobile robot positioning because they can help compensate for the foremost weakness of odometry: In an odometry-based positioning method, any small momentary orientation error will cause a constantly growing lateral position error. For this reason it would be of great benefit if orientation errors could be detected and corrected immediately.

2.1 Gyroscopes
Gyroscopes can be classified into two broad categories:
+ Mechanical gyroscopes + Optical gyroscopes.

The mechanical gyroscope, a well known and reliable rotation sensor based on the inertial properties of a rapidly spinning rotor, has been around since the early 1800s. The first known gyroscope was built in 1810 by G.C. Bohnenberger of Germany. In 1852, the French physicist Leon Foucault showed that a gyroscope could detect the rotation of the earth [Carter, 1966]. In the following sections we discuss the principle of operation of various gyroscopes. Later, in Section 5.4 (Inertial Navigation) we discuss the performance of commercially available gyros with respect to dead-reckoning in mobile robots. 2.1.1 Mechanical Gyroscopes Anyone who has ever ridden a bicycle has experienced (perhaps unknowingly) an interesting characteristic of the mechanical gyroscope known as gyroscopic precession. If the rider leans the bike over to the left around its own horizontal axis, the front wheel responds by turning left around the vertical axis. The effect is much more noticeable if the wheel is removed from the bike, and held by both ends of its axle while rapidly spinning. If the person holding the wheel attempts to yaw it left or right about the vertical axis, a surprisingly violent reaction will be felt as the axle instead twists about the horizontal roll axis. This is due to the angular momentum associated with a spinning flywheel, which displaces the applied force by 90o in the direction of spin. The rate of precession S is proportional to the applied torque T [Fraden, 1993]:

Page 24

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

T = I TS where T = applied input torque I = rotational inertia of rotor T = rotor spin rate S = rate of precession. Gyroscopic precession is a key factor involved in the concept of operation for the north-seeking gyrocompass, as will be discussed later. Friction in the support bearings, external influences, and small imbalances inherent in the construction of the rotor cause even the best mechanical gyros to drift with time. Typical systems employed in inertial navigation packages by the commercial airline industry may drift about 0.1o during a 6-hour flight [Martin, 1986]. 2.1.1.1 Space-Stable Gyroscopes The earths rotational velocity at any given point on the globe can be broken into two components: one that acts around an imaginary vertical axis normal to the surface, and another that acts around an imaginary horizontal axis tangent to the surface. These two components are known as the vertical earth rate and the horizontal earth rate, respectively. At the North Pole, for example, the component acting around the local vertical axis (vertical earth rate) would be precisely equal to the rotation rate of the earth, or 15 o-per-hour. The horizontal earth rate at the pole would be zero. As the point of interest moves down a meridian toward the equator, the vertical earth rate at that particular location decreases proportionally to a value of zero at the equator. Meanwhile, the horizontal earth rate, (i.e., that component acting around a horizontal axis tangent to the earths surface) increases from zero at the pole to a maximum value of 15o -per-hour at the equator. There are two basic classes of rotational sensing gyros: 1) rate gyros, which provide a voltage or frequency output signal proportional to the turning rate, and, 2) rate integrating gyros, which indicate the actual turn angle [Udd, 1991]. Unlike the magnetic compass, however, rate integrating gyros can only measure relative as opposed to absolute angular position, and must be initially referenced to a known orientation by some external means. A typical gyroscope configuration is shown in Fig. 2.1. The electrically-driven rotor is suspended in a pair of precision low-friction bearings at either end of the rotor axle. The rotor bearings are in turn supported by a circular ring, known as the inner gimbal ring; this inner gimbal ring pivots on a second set of bearings that attach it to the outer gimbal ring. This pivoting action of the inner gimbal defines the horizontal axis of the gyro, which is perpendicular to the spin axis of the rotor as shown in the figure. The outer gimbal ring is attached to the instrument frame by a third set of

Page 25

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

bearings that define the vertical axis of the gyro. The vertical axis is perpendicular to both the horizontal axis and the spin axis.

Outer Pivot

Wheel Outer Gimbal Inner Pivot Wheel Bearing Inner Gimbal

Figure 2.1: Typical two-axis mechanical gyroscope configuration [Everett, 1995].

Notice that if this configuration is oriented such that the spin axis points east-west, the horizontal axis is aligned with the north-south meridian. Since the gyro is space-stable (i.e., fixed in the inertial reference frame), the horizontal axis thus reads the horizontal earth rate component of the planets rotation, while the vertical axis reads the vertical earth rate component. If the spin axis is rotated 90o to a north-south alignment, the Earths rotation does not affect the gyross horizontal axis, since that axis is now orthogonal to the horizontal earth rate component. 2.1.1.2 Gyrocompasses The gyrocompass is a special configuration of the rate integrating gyroscope, employing a gravity reference to implement a north-seeking function that can be used as a true-north navigation reference. This phenomenon, first demonstrated in the early 1800s by Leon Foucault, was patented in Germany by Herman Anschutz-Kaempfe in 1903, and in the U.S. by Elmer Sperry in 1908 [Carter, 1966]. The U.S. and German navies had both introduced gyrocompasses into their fleets by 1911 [Martin, 1986]. The north-seeking capability of the gyrocompass is directly tied to the horizontal earth rate component measured by the horizontal axis. As mentioned earlier, when the gyro spin axis is oriented in a north-south direction, it is insensitive to earth rotation, and no tilting occurs. From this it follows that if tilting is observed, the spin axis is no longer aligned with the meridian. The direction and magnitude of the measured tilt are directly related to the direction and magnitude of the misalignment between the spin axis and true north.
Page 26

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

2.1.2 Optical Gyroscopes Optical rotation sensors have been under development now as replacements for mechanical gyros for over three decades. With little or no moving parts, such devices are virtually maintenance free and display no gravitational sensitivities, eliminating the need for gimbals. Fueled by a large anticipated market in the automotive industry, highly linear fiber-optic versions are now evolving that have wide dynamic range and very low projected costs. The principle of operation of the optical gyroscope, first discussed by Sagnac [1913], is conceptually very simple, although several significant engineering challenges had to be overcome before practical application was possible. In fact, it was not until the demonstration of the heliumneon laser at Bell Labs in 1960 that Sagnacs discovery took on any serious implications; the first operational ring laser gyro was developed by Warren Macek of Sperry Corporation just two years later [Martin, 1986]. Navigation quality ring laser gyroscopes began routine service in inertial navigation systems for the Boeing 757 and 767 in the early 1980s, and over half a million fiber optic navigation systems have been installed in Japanese automobiles since 1987 [Reunert, 1993]. Many technological improvements since Maceks first prototype make the optical rate gyro a potentially significant influence mobile robot navigation in the future. The basic device consists of two laser beams traveling in opposite directions (i.e., counter propagating) around a closed-loop path. The constructive and destructive interference patterns formed by splitting off and mixing parts of the two beams can be used to determine the rate and direction of rotation of the device itself. Schulz-DuBois [1966] idealized the ring laser as a hollow doughnut-shaped mirror in which light follows a closed circular path. Assuming an ideal 100-percent reflective mirror surface, the optical energy inside the cavity is theoretically unaffected by any rotation of the mirror itself. The counter propagating light beams mutually reinforce each other to create a stationary standing wave of intensity peaks and nulls as depicted in Fig. 2.2, regardless of whether the gyro is rotating [Martin, 1986]. A simplistic visualization based on the Schulz-DuBois idealization is perhaps helpful at this point in understanding the fundamental concept of operation before more detailed treatment of the subject is presented. The light and dark fringes of the nodes are analogous to the reflective stripes or slotted holes in the rotating disk of an incremental optical encoder, and can be theoretically counted in similar fashion by a light detector mounted on the cavity wall. (In this analogy, however, the standing-wave disk is fixed in the inertial reference frame, while the normally stationary detector revolves around it.) With each full rotation of the mirrored doughnut, the detector would see a number of node peaks equal to twice the optical path length of the beams divided by the wavelength of the light.

Page 27

Part I: Sensors for Mobile Robot Positioning


Lossless Cylindrical Mirror

Chapter 2: Heading Sensors

Nodes Observer Moves Around Ring With Rotation EM Field Pattern Is Stationary In Inertial Frame

Figure 2.2: Standing wave created by counter-propagating light beams in an idealized ring laser gyro (adapted from [Schulz-DuBois, 1966]).

Obviously, there is no practical way to implement this theoretical arrangement, since a perfect mirror cannot be realized in practice. Furthermore, the introduction of light energy into the cavity (as well as the need to observe and count the nodes on the standing wave) would interfere with the mirror performance, should such an ideal capability even exist. However, many practical embodiments of optical rotation sensors have been developed for use as rate gyros in navigation applications. Five general configurations will be discussed in the following subsections: + active optical resonators (2.1.2.1) + passive optical resonators (2.1.2.2) + open-loop fiber optic interferometers (analog) (2.1.2.3) + closed-loop fiber optic interferometers (digital) (2.1.2.4) + fiber optic resonators (2.1.2.5). Aronowitz [1971], Menegozzi & Lamb [1973], Chow, et al. [1985], Wilkinson [1987], and Udd [1991] provide in-depth discussions of the theory of the laser ring gyro and its fiber optic derivatives. A comprehensive treatment of the technologies and an extensive bibliography of preceding works is presented by Ezekial and Arditty [1982] in the proceedings of the First International Conference on Fiber Optic Rotation Sensors held at MIT in November, 1981. An excellent treatment of the salient features, advantages, and disadvantages of ring laser gyros versus fiber optic gyros is presented by Udd [1985, 1991]. 2.1.2.1 Active Ring Laser Gyros The active optical resonator configuration, more commonly known as the ring laser gyro, solves the problem of introducing light into the doughnut by filling the cavity itself with an active lazing medium, typically helium-neon. There are actually two beams generated by the laser, which travel around the ring in opposite directions. If the gyro cavity is caused to physically rotate in the
Page 28

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

counterclockwise direction, the counterclockwise-propagating beam will be forced to traverse a slightly longer path than under stationary conditions. Similarly, the clockwise-propagating beam will see its closed-loop path shortened by an identical amount. This phenomenon, known as the Sagnac effect, in essence changes the length of the resonant cavity. The magnitude of this change is given by the following equation [Chow et al., 1985]: L = 4 r2 c

where L = change in path length r = radius of the circular beam path S = angular velocity of rotation c = speed of light. Note that the change in path length is directly proportional to the rotation rate S of the cavity. Thus, to measure gyro rotation, some convenient means must be established to measure the induced change in the optical path length. This requirement to measure the difference in path lengths is where the invention of the laser in the early 1960s provided the needed technological breakthrough that allowed Sagnacs observations to be put to practical use. For lazing to occur in the resonant cavity, the round-trip beam path must be precisely equal in length to an integral number of wavelengths at the resonant frequency. This means the wavelengths (and therefore the frequencies) of the two counter propagating beams must change, as only oscillations with wavelengths satisfying the resonance condition can be sustained in the cavity. The frequency difference between the two beams is given by [Chow et al., 1985]

f =

2 f r 2 r = c

where f = frequency difference r = radius of circular beam path S = angular velocity of rotation 8 = wavelength. In practice, a doughnut-shaped ring cavity would be hard to realize. For an arbitrary cavity geometry, the expression becomes [Chow et al., 1985]

f =

4A P

Page 29

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

where f = frequency difference A = area enclosed by the closed-loop beam path S = angular velocity of rotation P = perimeter of the beam path 8 = wavelength. For single-axis gyros, the ring is generally formed by aligning three highly reflective mirrors to create a closed-loop triangular path as shown in Fig. 2.3. (Some systems, such as Maceks early prototype, employ four mirrors to create a square path). The mirrors are usually mounted to a monolithic glass-ceramic block with machined ports for the cavity bores and electrodes. Most modern three-axis units employ a square block cube with a total of six mirrors, each mounted to the center of a block face as shown in Fig. 2.3. The most stable systems employ linearly polarized light and minimize circularly polarized components to avoid magnetic sensitivities [Martin, 1986]. The approximate quantum noise limit for the ring laser gyro is due to spontaneous emission in the gain medium [Ezekiel and Arditty, 1982]. Yet, the ring laser gyro represents the best-case scenario of the five general gyro configurations outlined above. For this reason the active laser ring gyro offers the highest sensitivity and is perhaps the most accurate implementation to date.

The fundamental disadvantage associated with the active ring laser C D is a problem called frequency lockFigure 2.3: Six-mirror configuration of three-axis ring laser gyro in, which occurs at low rotation (adapted from [Koper, 1987]). rates when the counter propagating beams lock together in frequency [Chao et al., 1984]. This lock-in is attributed to the influence of a very small amount of back scatter from the mirror surfaces, and results in a deadband region (below a certain threshold of rotational velocity) for which there is no output signal. Above the lock-in threshold, output approaches the ideal linear response curve in a parabolic fashion. The most obvious approach to solving the lock-in problem is to improve the quality of the mirrors to reduce the resulting back scatter. Again, however, perfect mirrors do not exist, and some finite amount of back scatter will always be present. Martin [1986] reports a representative value as 10 -12

Page 30

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

of the power of the main beam; enough to induce frequency lock-in for rotational rates of several hundred degrees per hour in a typical gyro with a 20-cm (8-inch) perimeter. An additional technique for reducing lock-in is to incorporate some type of biasing scheme to shift the operating point away from the deadband zone. Mechanical dithering is the least elegant but most common biasing means, introducing the obvious disadvantages of increased system complexity and reduced mean-time-between-failures due to the moving parts. The entire gyro assembly is rotated back and forth about the sensing axis in an oscillatory fashion. State-of-the-art dithered active ring laser gyros have a scale factor linearity that far surpasses the best mechanical gyros. Dithered biasing, unfortunately, is too slow for high-performance systems (i.e., flight control), resulting in oscillatory instabilities [Martin, 1986]. Furthermore, mechanical dithering can introduce crosstalk between axes on a multi-axis system, although some unibody three-axis gyros employ a common dither axis to eliminate this possibility [Martin, 1986]. Buholz and Chodorow [1967], Chesnoy [1989], and Christian and Rosker [1991], discuss the use of extremely short duration laser pulses (typically 1/15 of the resonator perimeter in length) to reduce the effects of frequency lock-in at low rotation rates. The basic idea is to reduce the cross coupling between the two counter propagating beams by limiting the regions in the cavity where the two pulses overlap. Wax and Chodorow [1972] report an improvement in performance of two orders of magnitude through the use of intra cavity phase modulation. Other techniques based on nonlinear optics have been proposed, including an approach by Litton that applies an external magnetic field to the cavity to create a directionally dependent phase shift for biasing [Martin, 1986]. Yet another solution to the lock-in problem is to remove the lazing medium from the ring altogether, effectively forming what is known as a passive ring resonator. 2.1.2.2 Passive Ring Resonator Gyros The Passive Ring Resonator Gyro makes use of a laser source external to the ring cavity (Fig. 2.4), and thus avoids the frequency lock-in problem which arises when the gain medium is internal to the cavity itself. The passive configuration also eliminates problems arising from changes in the optical path length within the interferometer due to variations in the index of refraction of the gain medium [Chow, et al., 1985]. The theoretical quantum noise limit is determined by photon shot noise, and is slightly higher (i.e., worse) than the theoretical limit seen for the Active Laser Ring Gyro [Ezekiel & Arditty, 1982].

Page 31

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

The fact that these devices use mirrored resonators patterned after their active ring predecessors means that their packaging is inherently bulky. However, fiber optic technology now offers a low volume alternative. The fiber optic derivatives also allow longer length multiturn resonators, for increased sensitivity in smaller, rugged, and less expensive packages. As a consequence, the Resonant Fiber Optic Gyro (RFOG), to be discussed in Section 2.1.2.5, has emerged as the most popular of the resonator configurations [Sanders, 1992].

Light Source

Highly Reflective Mirror

Partially Transmissive Mirror

Detector

Figure 2.4: Passive ring resonator gyro with laser source external to the ring cavity (adapted from [Udd, 1991]).

2.1.2.3 Open-Loop Interferometric Fiber Optic Gyros The concurrent development of optical fiber technology, spurred mainly by the communications industry, presented a potential lowcost alternative to the high-tolerance machining and clean-room assembly required for ring laser gyros. The glass fiber in essence forms an internally reflective waveguide for optical energy, along the lines of a small-diameter linear implementation of the doughnut-shaped mirror cavity conceptualized by Schulz-DuBois [1966]. Recall the refractive index n relates the speed of light in a particular medium to the speed of light in a vacuum as follows: n= c cm

where n = refractive index of medium c = speed of light in a vacuum cm = speed of light in medium. Step-index multi-mode fiber (Fig. 2.5) is made up of a core region of glass with index of refraction nco, surrounded by a protective cladding with a lower index of refraction cl [Nolan & n Blaszyk, 1991]. The lower refractive index in the cladding is necessary to ensure total internal reflection of the light propagating through the core region. The terminology step-index refers to this stepped discontinuity in the refractive index that occurs at the core-cladding interface. Referring now to Fig. 2.5, as long as the entry angle (with respect to the waveguide axis) of an incoming ray is less than a certain critical angle 2 c, the ray will be guided down the fiber, virtually
Page 32

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

without loss. The numerical aperture of the fiber quantifies this parameter of acceptance (the lightcollecting ability of the fiber), and is defined as follows [Nolan & Blaszyk, 1991]:
n cl

NA = sin c =

nco 2 ncl 2
Waveguide Axis

n co

where NA = 2c = nco = ncl =

numerical aperture of the fiber critical angle of acceptance index of refraction of glass core index of refraction of cladding.

Figure 2.5:

As illustrated in Fig. 2.6, a number of rays following different-length paths can simultaneously propagate down the fiber, as long as their respective entry angles are less than the critical angle of acceptance 2c . Multiple-path propagation of this nature occurs where the core diameter is much larger than the wavelength of the guided energy, giving rise to the term multi-mode fiber. Such multi-mode operation is clearly undesirable in gyro applications, where the objective is to eliminate all non-reciprocal conditions other than that imposed by the Sagnac effect itself. As the diameter of the core is reduced to approach the operating wavelength, a cutoff condition is reached where just a single mode is allowed to propagate, constrained to travel only along the waveguide axis [Nolan & Blaszyk, 1991].
Numerical Aperture

Step-index multi-mode fiber (adapted from [Nolan et al., 1991]).

Light can randomly change polarization states as it propagates through standard single-mode fiber. The use of special polarization-maintaining fiber, such as PRSM Corning, maintains the original polarization state of the light along the path of travel [Reunert, 1993]. This is important, since light of different polarization states travels through an optical fiber at different speeds.

Waveguide Axis

2 1

A typical block diagram of the minimum-reciprocal IFOG configuration is presented in Fig. 2.7. Polarization-maintaining single-mode fiber [Nolan & Blaszyk, 1991] is employed to ensure the two counterpropagating beams in the loop follow identical paths in the absence of rotation.

Figure 2.6: Entry angles of incoming rays 1 and 2 determine propagation paths in fiber core (adapted from [Nolan et al., 1991]).

Page 33

Part I: Sensors for Mobile Robot Positioning


SOURCE SPLITTER COIL SPLITTER

Chapter 2: Heading Sensors

SOURCE POLARIZER FILTER

DETECTOR

FIBER COIL

PHASE MODULATOR

Figure 2.7: Block diagram of "minimum-reciprocal" integrated fiber optic gyro (adapted from [Lefevre, 1992]).

An interesting characteristic of the IFOG is the absence of any laser source [Burns, et al, 1983], the enabling technology allowing the Sagnac effect to reach practical implementation in the first place. A low-coherence source, such as a super-luminescent diode (SLD), is typically employed instead to reduce the effects of noise [Tai et al., 1986], the primary source of which is backscattering within the fiber and at any interfaces. As a result, in addition to the two primary counterpropagating waves in the loop, there are also a number of parasitic waves that yield secondary interferometers [Lefevre, 1992]. The limited temporal coherence of the broadband SLD causes any interference due to backscattering to average to zero, surpressing the contrast of these spurious interferometers. The detection system becomes sensitive only to the interference between waves that followed identical paths [Ezekiel & Arditty, 1982; Lefevre, 1992]. The Sagnac phase shift introduced by rotation is given by [Ezekiel & Arditty, 1982]

2 LD c

where )N = measured phase shift between counterpropagating beams L = length of fiber optic cable in loop D = diameter of loop 8 = wavelength of optical energy c = speed of light in a vacuum. The stability of the scale factor relating )N to the rotational velocity in the equation above is thus limited to the stability of L, D, and 8 [Ezekiel & Arditty, 1982]. Practical implementations usually operate over plus or minus half a fringe (i.e., B radians of phase difference), with a theoretical sensitivity of 10-6 radians or less of phase shift [Lefevre, 1992].

Page 34

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

IFOG sensitivity may be improved by increasing L (i.e., adding turns of fiber in the sensing loop). This effect peaks at an optimal length of several kilometers, after which the fiber attenuation (1 dB per kilometer typical) begins to degrade performance. This large amount of fiber represents a significant percentage of overall system cost. In summary, the open-loop IFOG is attractive from the standpoint of reduced manufacturing costs. Additional advantages include high tolerance to shock and vibration, insensitivity to gravity effects, quick start-up, and good sensitivity in terms of bias drift rate and random walk coefficient. Coil geometry is not critical, and no path length control is needed. Some disadvantages are that a long optical cable is required, dynamic range is limited with respect to active ring laser gyros, and the scale factor is prone to vary [Adrian, 1991]. Open loop configurations are therefore most suited to the needs of lowcost systems in applications that require relatively low accuracy (i.e., automobile navigation). For applications demanding higher accuracy, such as aircraft navigation (0.01 to 0.001 degreesper-hour), the Closed-Loop IFOG to be discussed in the next section offers significant promise. 2.1.2.4 Closed-Loop Interferometric Fiber Optic Gyros This new implementation of a fiber optics gyro provides feedback to a frequency or phase shifting element. The use of feedback results in the cancellation of the rotationally induced Sagnac phase shift. However, closed-loop digital signal processing is considerably more complex than the analog signal processing employed on open-loop IFOG configurations [Adrian, 1991]. Nonetheless, it now seems that the additional complexity is justified by the improved stability of the gyro: closed-loop IFOGs are now under development with drifts in the 0.001-0.01 degree-per-hour range, and scalefactor stabilities greater than 100 ppm (parts per million) [Adrian, 1991]. 2.1.2.5 Resonant Fiber Optic Gyros The Resonant Fiber Optic Gyro (RFOG) evolved as a solid-state derivative of the Passive Ring Resonator Gyro discussed in Section 2.1.2.2. In the solid-state implementation, a passive resonant cavity is formed from a multi-turn closed loop of optical fiber. An input coupler provides a means for injecting frequency-modulated light from a laser source into the resonant loop in both the clockwise and counterclockwise directions. As the frequency of the modulated light passes through a value such that the perimeter of the loop precisely matches an integral number of wavelengths at that frequency, input energy is strongly coupled into the loop [Sanders, 1992]. In the absence of loop rotation, maximum coupling for both beam directions occurs in a sharp peak centered at this resonant frequency. If the loop is caused to rotate in the clockwise direction, of course, the Sagnac effect causes the perceived loop perimeter to lengthen for the clockwise-traveling beam, and to shorten for the counter-clockwise-traveling beam. The resonant frequencies must shift accordingly, and as a result, energy is coupled into the loop at two different frequencies and directions during each cycle of the
Page 35

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

sinusoidal FM sweep. An output coupler samples the intensity of the energy in the loop by passing a percentage of the two counter-rotating beams to their respective detectors. The demodulated output from these detectors will show resonance peaks, separated by a frequency difference f given by the following [Sanders, 1992]: f = D n

where f = frequency difference between counterpropagating beams D = diameter of the resonant loop S = rotational velocity 8 = freespace wavelength of laser n = refractive index of the fiber. Like the IFOG, the all-solid-state RFOG is attractive from the standpoint of high reliability, long life, quick start up, and light weight. The principle advantage of the RFOG, however, is that it requires significantly less fiber (from 10-100 times less) in the sensing coil than the IFOG configuration, while achieving the same shot-noise-limited performance [Sanders, 1992]. Sanders attributes this to the fact that light traverses the sensing loop multiple times, as opposed to once in the IFOG counterpart. On the down side are the requirements for a highly coherent source and extremely low-loss fiber components [Adrian, 1991].

2.2 Geomagnetic Sensors


Vehicle heading is the most significant of the navigation parameters (X, Y, and 2) in terms of its influence on accumulated dead reckoning errors. For this reason, sensors which provide a measure of absolute heading or relative angular velocity are extremely important in solving the real world navigation needs of an autonomous platform. The most commonly known sensor of this type is probably the magnetic compass. The terminology normally used to describe the intensity of a magnetic field is magnetic flux density B, measured in gauss (G). Alternative units are the Tesla (T), and the gamma ((), where 1 Tesla = 104 gauss = 109 gamma The average strength of the earths magnetic field is 0.5 gauss, and can be represented as a dipole that fluctuates both in time and space, situated roughly 440 kilometers off center and inclined 11o to the planets axis of rotation [Fraden, 1993]. This difference in location between true north and magnetic north is known as declination, and varies with both time and geographical location. Corrective values are routinely provided in the form of declination tables printed directly on the maps or charts for any given locale.

Page 36

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

Instruments which measure magnetic fields are known as magnetometers. For application to mobile robot navigation, only those classes of magnetometers which sense the magnetic field of the earth are of interest. Such geomagnetic sensors, for purposes of this discussion, will be broken down into the following general categories: + Mechanical magnetic compasses + Fluxgate compasses + Hall effect compasses + Magnetoresistive compasses + Magnetoelastic compasses Before we introduce different types of compasses, a word of warning: The earth's magnetic field is often distorted near power lines or steel structures [Byrne et al., 1992]. This makes the straightforward use of geomagnetic sensors difficult for indoor applications. However, it may be possible to overcome this problem in the future by fusing data from geomagnetic compasses with data from other sensors. 2.2.1 Mechanical Magnetic Compasses The first recorded use of a magnetic compass was in 2634 B.C., when the Chinese suspended a piece of naturally occurring magnetite from a silk thread, and used it to guide a chariot over land [Carter, 1966]. Much controversy surrounds the debate over whether the Chinese or the Europeans first adapted the compass for marine applications, but by the middle of the 13th century such usage was fairly widespread around the globe. William Gilbert [1600] was the first to propose that the earth itself was the source of the mysterious magnetic field that provided such a stable navigation reference for ships at sea. The early marine compasses were little more that magnetized needles floated in water on small pieces of cork. These primitive devices evolved over the years into the reliable and time proven systems in use today, which consist of a ring magnet or pair of bar magnets attached to a graduated mica readout disk. The magnet and disk assembly floats in a mixture of water and alcohol or glycerine, such that it is free to rotate around a jeweled pivot. The fluid acts to both support the weight of the rotating assembly, and to dampen its motion under rough conditions. The sealed vessel containing the compass disk and damping fluid is typically suspended from a two-degree-of-freedom gimbal to decouple it from the ships motion. This gimbal assembly is mounted in turn atop a floor stand or binnacle. On either side of the binnacle are massive iron spheres that, along with adjustable permanent magnets in the base, are used to compensate the compass for surrounding magnetic abnormalities that alter the geomagnetic lines of flux. The error resulting from such external influences (i.e., the angle between indicated and actual bearing to magnetic north) is known as compass deviation, and along with local declination, must be added or subtracted as appropriate for true heading: Ht = Hi CFdev CFdec
Page 37

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

where Ht = true heading Hi = indicated heading CFdev = correction factor for compass deviation CFdec = correction factor for magnetic declination. Another potential source of error which must be taken into account is magnetic dip, a term arising from the dipping action observed in compass needles attributed to the vertical component of the geomagnetic field. The dip effect varies with latitude, from no impact at the equator where the flux lines are horizontal, to maximum at the poles where the lines of force are entirely vertical. For this reason, many swing-needle instruments have small adjustable weights that can be moved radially to balance the needle for any given local area of operation. Marine compasses ensure alignment in the horizontal plane by floating the magnet assembly in an inert fluid. Dinsmore Starguide Magnetic Compass An extremely lowcost configuration of the mechanical magnetic compass suitable for robotic applications is seen in a product recently announced by the Dinsmore Instrument Company, Flint, MI. The heart of the Starguide compass is the Dinsmore model 1490 digital sensor [Dinsmore Instrument Company, 1991], which consists of a miniaturized permanent-magnet rotor mounted in low friction jewelled bearings. The sensor is internally damped such that if momentarily displaced 90o, it will return to the indicated direction in 2.5 seconds, with no overshoot. Four Hall-effect switches corresponding to the cardinal headings (N, E, W, S) are arranged around the periphery of the rotor, and activated by the south pole of the magnet as the rotor aligns itself with the Earths magnetic field. Intermediate headings (NE, NW, SE, SW) are indicated through simultaneous activation of the adjacent cardinal-heading switches. The Dinsmore Starguide is not a true Hall-effect compass (see Section 2.2.3), in that the Hall-effect devices are not directly sensing the geomagnetic field of the earth, but rather the angular position of a mechanical rotor. The model 1490 digital sensor measures 12.5 mm (0.5 inches) in diameter by 16 mm (0.63 inches) high, and is available separately from Dinsmore for around $12. Current consumption is 30 milliamps, and the open-collector NPN outputs can sink 25 milliamps per channel. Grenoble [1990] presents a simple circuit for interfacing the device to eight indicator LEDs. An alternative analog sensor (model 1525) with a ratiometric sine-cosine output is also available for around $35. Both sensors may be subjected to unlimited magnetic flux without damage. 2.2.2 Fluxgate Compasses There currently is no practical alternative to the popular fluxgate compass for portability and long missions [Fenn, et al., 1992]. The term fluxgate is actually a trade name of Pioneer Bendix for the saturable-core magnetometer, derived from the gating action imposed by an AC-driven excitation coil that induces a time varying permeability in the sensor core. Before discussing the principle of
Page 38

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

operation, it is probably best to review briefly the subject of magnetic conductance, or permeability. The permeability of a given material is a measure of how well it serves as a path for magnetic lines of force, relative to air, which has an assigned permeability of 1. Some examples of highpermeability materials are listed in Table 2.1 below. Permeability is the magnetic circuit analogy to electrical conductivity, and relates magnetic flux density to the magnetizing force as follows: B = H where B = magnetic flux density = permeability H = magnetizing force.
Material Supermalloy Pure iron Mumetal Permalloy Cast iron Permeability 100,000 - 1,000,000 25,000 - 300,000 20,000 - 100,000 2,500 - 25,000 100 - 600

Table 2.1: Permeability ranges for selected materials. Values vary with proportional makeup, heat treatment, and mechanical working of the material [Bolz & Tuve, 1979].

Since the magnetic flux in a magnetic circuit is analogous to current i in an electrical circuit, it follows that magnetic flux density B is the parallel to electrical current density. A graphical plot of the above equation is known as the normal magnetizing curve, or B-H curve, and the permeability is the slope. In actuality, due to hysteresis, depends not only on the current value of H, but also the history of previous values, and the sign of dH/dt, as will be seen later. The important thing to note at this point in the discussion is the B-H curve is not linear, but rather starts off with a fairly steep slope, and then flattens out suddenly as H reaches a certain value. Increasing H beyond this knee of the B-H curve yields little increase in B; the material is effectively saturated. When a highly permeable material is introduced into a uniform magnetic field, the lines of force are drawn into the lower resistance path presented by the material as shown in Fig. 2.8. However, if the material is forced into saturation by some additional magnetizing force H, the lines of flux of the external field will be relatively unaffected by the presence of the saturated material, as indicated in Fig. 2.8B. The fluxgate magnetometer makes use of this saturation phenomenon in order to directly measure the strength of a surrounding static magnetic field. Various core materials have been employed in different fluxgate designs over the past 50 years, with the two most common being permalloy (an alloy of iron and nickel) and mumetal (iron, nickel, copper, and chromium). The permeable core is driven into and out of saturation by a gating signal applied to an excitation coil wound around the core. For purposes of illustration, lets assume for the moment a square-wave drive current is applied. As the core moves in and out of saturation, the flux lines from the external B field to be measured are drawn into and out of the core, alternating
Page 39

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

in turn between the two states depicted in Fig. 2.8. (This is somewhat of an oversimplification, in that the B-H curve does not fully flatten out with zero slope after the knee).

Drive

Sense

Drive

Sense

Figure 2.8: External lines of flux for: A) unsaturated core, B) saturated core (adapted from [Lenz, 1990]).

These expanding and collapsing flux lines will induce positive and negative e.m.f. surges in a sensing coil properly oriented around the core. The magnitude of these surges will vary with the strength of the external magnetic field, and its orientation with respect to the axis of the core and sensing coil of the fluxgate configuration. The fact that the permeability of the sensor core can be altered in a controlled fashion by the excitation coil is the underlying principle which enables the DC field being measured to induce a voltage in the sense coil. The greater the differential between the saturated and unsaturated states (i.e., the steeper the slope), the more sensitive the instrument will be. The effective permeability a of a material is influenced to a significant extent by its geometry. Bozorth and Chapin [1942] showed how a for a cylindrical rod falls off with a decrease in the length-to-diameter ratio. This relationship can be attributed to the so-called demagnetization factor [Hine, 1968]. When a ferrous rod is coaxially aligned with the lines of flux of a magnetic field, a magnetic dipole is developed in the rod itself. The associated field introduced by the north and south poles of this dipole opposes the ambient field, with a corresponding reduction of flux density through the rod. The lowered value of a results in a less sensitive magnetometer, in that the fluxgathering capability of the core is substantially reduced. Consider again the cylindrical rod sensor presented in Fig. 2.8, now in the absence of any external magnetic field Be. When the drive coil is energized, there will be a strong coupling between the drive coil and the sense coil. Obviously, this will be an undesirable situation since the output signal is supposed to be related to the strength of the external field only. One way around this problem is seen in the Vacquier configuration developed in the early 1940s, where two parallel rods collectively form the core, with a common sense coil [Primdahl, 1979]. The two rods are simultaneously forced into and out of saturation, excited in antiphase by identical but oppositely wound solenoidal drive windings. In this fashion, the magnetization fluxes of the two drive windings effectively cancel each other, with no net effect on the sense coil.

Page 40

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

Bridges of magnetic material may be employed to couple the ends of the two coils together in a closed-loop fashion for more complete flux linkage through the core. This configuration is functionally very similar to the ring-core design first employed in 1928 by Aschenbrenner and Goubau [Geyger, 1957]. An alternative technique for decoupling the pickup coil from the drive coil is to arrange the two in an orthogonal fashion. In practice, there are a number of different implementations of various types of sensor cores and coil configurations as described by Stuart [1972] and Primdahl [1979]. These are generally divided into two classes, parallel and orthogonal, depending on whether the excitation H-field is parallel or perpendicular to the external B-field being measured. Alternative excitation strategies (sine wave, square wave, sawtooth ramp) also contribute to the variety of implementations seen in the literature. Hine [1968] outlines four different classifications of saturable inductor magnetometers based on the method of readout (i.e., how the output e.m.f. is isolated for evaluation): + Fundamental frequency + Second harmonic + Peak-output + Pulse-difference. Unambiguous 360 resolution of Toroidal the earths geomagnetic field re- Drive Winding Core quires two sensing coils at right angles to each other. The ring core geometry lends itself to such dualaxis applications in that two orthogonal pickup coils can be configured in a symmetrical fashion Sensing Winding 2 around a common core. A followon version developed by Gordon and Lundsten [1970] employed a toroidal excitation winding as shown in Fig. 2.9. Since there are no distinct poles in a closed-ring Drive Winding design, demagnetization effects, although still present [Stuart, 1972], Figure 2.9: Two channel ring core fluxgate with toroidal excitation are less severe. The use of a ring (adapted from (Acuna & Pellerin, 1969]).. geometry also leads to more complete flux linkage throughout the core, implying less required drive excitation for lower power operation, and the zero offset can be minimized by rotating the circular core. For these reasons, along with ease of manufacture, toroidal ring-core sensors are commonly employed in many of the lowcost fluxgate compasses available today. The integrated DC output voltages Vx and Vy of the orthogonal sensing coils vary as sine and cosine functions of 2, where 2 is the angle of the sensor unit relative to the earths magnetic field.
Page 41
o

Sensing Winding 1

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

The instantaneous value of 2 can be easily derived by performing two successive A/D conversions on these voltages, and taking the arctangent of their quotient:
= arctan Vx Vy

Another popular two-axis core design is seen in the Flux Valve magnetometer developed by Sperry Corp. [SPERRY] and shown in Fig. 2.10. This three-legged spider configuration employs three horizontal sense coils 120o apart, with a common vertical excitation coil in the middle [Hine, 1968]. Referring to Fig. 2.10B, the upper and lower arms of the sense coil S are excited by the driving coil D, so that a magnetizing force Hx develops. In the absence of an external field He, the flux generated in the upper and lower arms by the excitation coil is equal and opposite due to symmetry. When this assembly is placed in an axial magnetic field He, however, the instantaneous excitation field Hx complements the flow in one arm, while opposing the flow in the other. This condition is periodically reversed in the arms, of course, due to the alternating nature of the driving function. A second-harmonic output is induced in the sensing coil S, proportional to the strength and orientation of the ambient field. By observing the relationships between the magnitudes of the output signals from each of the three sense coils, the angular relationship of the Flux Valve with respect to the external field can be unambiguously determined.

He

Figure 2.10: The Flux-Valve magnetometer developed by Sperry Corporation uses a spider core configuration (adapted from [Lenz, 1990]).

When maintained in a level attitude, the fluxgate compass will measure the horizontal component of the earths magnetic field, with the decided advantages of low power consumption, no moving parts, intolerance to shock and vibration, rapid start up, and relatively low cost. If the vehicle is expected to operate over uneven terrain, the sensor coil should be gimbal-mounted and mechanically dampened to prevent serious errors introduced by the vertical component of the geomagnetic field. 2.2.2.1 Zemco Fluxgate Compasses

Page 42

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

The Zemco Fluxgate compass [ZEMCO] was used in earlier work by Everett et al. [1990] on their robot called ROBART II. The sensor was a fluxgate compass manufactured by Zemco Electronics, San Ramon, CA, model number DE-700. This very lowcost (around $40) unit featured a rotating analog dial, and was originally intended for 12-volt DC operation in automobiles. A system block diagram is presented in Fig. 2.11. The sensor consists of two orthogonal pickup coils arranged around a toroidal excitation coil, driven in turn by a local oscillator. The outputs V x and Vy of amplifier channels A and B are applied across an air-core resolver to drive the display indicator. The standard resolver equations [ILC Corporation, 1982] for these two voltages are
V x = K x sin sin(t + a x ) Vy = K y cos sin(t + a y )

where 2 = the resolver shaft angle T = 2Bf, where f is the excitation frequency Kx and Ky are ideally equal transfer-function constants, and ax and ay are ideally zero time-phase shifts. Thus, for any static spatial angle 2, the equations reduce to
V x = K x sin Vy = K y cos

which can be combined to yield


Vx sin = = tan Vy cos

Amplifier

Driver

Fluxgate Sensor Driver

Oscillator

The magnetic heading 2 therefore is simply the arctangent of Vx over Vy. Everett [1995] recounts his experience with two models of the Zemco fluxgate compass on Robart II as follows: Problems associated with the use of this particular fluxgate compass on ROBART, however, included a fairly high current con-

Air Core Resolver

Amplifier

Driver

Figure 2.11: Block diagram of ZEMCO Model DE-700 fluxgate compass (courtesy ZEMCO, Inc.).

Page 43

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

sumption (250 milliamps), and stiction in the resolver reflecting back as load into the drive circuitry, introducing some error for minor changes in vehicle heading. In addition, the sensor itself was affected by surrounding magnetic anomalies, some that existed onboard the robot (i.e., current flow in nearby cable runs, drive and head positioning motors), and some present in the surrounding environment (metal desks, bookcases, large motors, etc.). The most serious interference turned out to be the fluctuating magnetic fields due to power cables in close proximity on the order of 30 cm (12 inches) to the fluxgate sensor. As various auxiliary systems onboard the robot were turned on when needed and later deactivated to save power, the magnetic field surrounding the sensor would change accordingly. Serious errors could be introduced as well by minor changes in the position of cable runs, which occurred as a result of routine maintenance and trouble shooting. These problems were minimized by securing all cable runs with plastic tie-downs, and adopting a somewhat standardized protocol regarding which auxiliary systems would be activated when reading the compass. There was no solution, however, for the interference effects of large metallic objects within the operating environment, and deviations of approximately 4o were observed when passing within 30 cm (12 inches) of a large metal cabinet, for example. A final source of error was introduced by virtue of the fact that the fluxgate compass had been mounted on the robots head, so as to be

Amplifier

Driver

Fluxgate Sensor Driver

Oscillator

Analog To Digital Convertor

Microprocessor

Display

Amplifier

Driver

Figure 2.12: Block diagram of ZEMCO model DE-710 fluxgate compass (courtesy ZEMCO, Inc.).

Page 44

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

as far away as possible from the effects of the drive motors and power distribution lines discussed above. The exact head position could only be read to within 0.82o due to the limited resolution of the 8-bit A/D converter. In any event, an overall system error of 10o was typical, and grossly insufficient for reliable dead reckoning calculations, but that was not the original intent of the compass. This analog compass was later replaced by a newer digital version produced by Zemco, model DE-710, which cost approximately $90. The system block diagram is shown in Fig. 2.12. This unit contained a built-in ADC0834 A/D converter to read the amplified outputs of the two sensor channels, and employed its own COP 421-MLA microprocessor, which drove a liquid crystal display (LCD). All communication between the A/D converter, microprocessor, and display driver was serial in nature, with a resulting slow update rate of 0.25 Hz. The built-in LCD simulated an analog dial with an extremely coarse resolution of 20o between display increments, but provision was made for serial output to an optional external shift register and associated three-digit numerical display. All things considered, it was determined to be more practical to discard the built-in microprocessor, A/D converter, and LCD display, and interface an external A/D converter directly to the amplifier outputs as before with the analog version. This resulted in a decrease in supply current from 168 to 94 milliamps. Power consumption turned out to be less of a factor when it was discovered the circuitry could be powered up for a reading, and then deactivated afterwards with no noticeable effect on accuracy. Overall system accuracy for this configuration was typically 6o, although a valid comparison to the analog version is not possible since the digital model was mounted in a different location to minimize interference from nearby circuitry. The amount of effort put into calibration of the two systems must also be taken into account; the calibration procedure as performed was an iterative process not easily replicated from unit to unit with any quantitative measure. 2.2.2.2 Watson Gyro Compass A combination fluxgate compass and solid-state rate gyro package (part number FGM-G100DHSRS232) is available from Watson Industries, Eau Claire, WI [WATSON]. The system contains its own microprocessor that is intended to integrate the information from both the rate gyro and the compass to provide a more stable output less susceptible to interference, with an update rate of 40 Hz. An overall block diagram is presented in Fig. 2.13. The Watson fluxgate/rate gyro combination balances the shortcomings of each type of device: the gyro serves to filter out the effects of magnetic anomalies in the surrounding environment, while the compass counters the long-term drift of the gyro. Furthermore, The toroidal ring-core fluxgate sensor is gimbal-mounted for improved accuracy.

Page 45

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

The Watson unit measures 6.34.47.6 cm (2.51.753.0 inches), and weighs only 10 ounces. This integrated package is a much more expensive unit ($2,500) than the lowcost Zemco fluxgate compass, but is advertised to have higher accuracy (2 o). Power supply requirements are 12-volts DC at 200 milliamps, and the unit provides an analog voltage output as well as a 12-bit digital output over a 2400-baud RS-232 serial link.
HDG Select HDG Angular Rate Sensor D/A Bias A/D RS-232 Interface Damping Function Error

Fluxgate Sensor

A/D

HDG Hold HDG Trim (+) HDG Trim (-)

3 /Sec

Figure 2.13: Block diagram of Watson fluxgate compass and rate gyro combination (courtesy [WATSON]).

2.2.2.3 KVH Fluxgate Compasses KVH Industries, Inc., Middletown, RI, offers a complete line of fluxgate compasses and related accessories, ranging from inexpensive units targeted for the individual consumer up through sophisticated systems intended for military applications [KVH]. The C100 COMPASS ENGINE (see Fig. 2.14) is a versatile lowcost (less than $700) developer's kit that includes a microprocessor-controlled standalone fluxgate sensor subsystem based on a two-axis toroidal ring core sensor. Two different sensor options are offered with the C-100: 1) the SE-25 sensor, recFigure 2.14: The C-100 fluxgate compass engine was tested at the University of Michigan in a flying robot prototype (Photo: Adapted from [KVH]). Page 46

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

ommended for applications with a tilt range of 16o , and, 2) the SE-10 sensor, for applications anticipating a tilt angle of up to 45o. The SE-25 sensor provides internal gimballing by floating the sensor coil in an inert fluid inside the lexan housing. The SE-10 sensor provides an additional twodegree-of-freedom pendulous gimbal in addition to the internal fluid suspension. The SE-25 sensor mounts on top of the sensor PC board, while the SE-10 is suspended beneath it. The sensor PC board can be separated as much as 122 cm (48 inches) from the detachable electronics PC board with an optional cable if so desired. The resolution of the C100 is 0.1o, with an advertised accuracy of 0.5o (after compensation, with the sensor card level) and a repeatability of 0.2o. Separate 180o adjustments are provided for declination as well as index offset (in the event the sensor unit cannot be mounted in perfect alignment with the vehicles axis of travel). System damping can be user-selected, anywhere in the range of 0.1 to 24 seconds settling time to final value. An innovative automatic compensation algorithm employed in the C100 is largely responsible for the high accuracy obtained by such a relatively low-priced system. This software routine runs on the controlling microprocessor mounted on the electronics board, and corrects for magnetic anomalies associated with the host vehicle. Three alternative user-selectable procedures are offered:
+ Eight-Point Auto-Compensation starting from an arbitrary heading, the platform turns full

circle, pausing momentarily at approximately 45-degree intervals. No known headings are required.
+ Circular Auto-Compensation starting from an arbitrary position, the platform turns slowly

through a continuous 360-degree circle. No known headings are required.


+ Three-Point Auto-Compensation starting from an arbitrary heading, the platform turns and

pauses on two additional known headings approximately 120o apart. Correction values are stored in a look-up table in non-volatile EEPROM memory. The automatic compensation routine also provides a quantitative indicator of the estimated quality of the current compensation, and the magnitude of any magnetic interference present [KVH Industries, 1993]. The C100 configured with an SE-25 coil assembly weighs just 2.25 ounces, and draws 40 milliamps at 8-18 volts DC (or 18-28 volts DC). The combined sensor and electronics boards measure 4.611 cm (1.84.5 inches). RS-232 (300-9600 baud) and NMEA 0183 digital outputs are provided, as well as linear and sine/cosine analog voltage outputs. Display and housing options are also available.

Page 47

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

2.2.3 Hall Effect Compasses Hall-effect sensors are based on E. H. Hall's observation (in 1879) that a DC voltage develops across a conductor or semiconductor when in the presence of an external magnetic field. One advantage of this technology (i.e., relative to the fluxgate) is the inherent ability to directly sense a static flux, resulting in much simpler readout electronics. Early Hall magnetometers could not match the sensitivity and stability of the fluxgate [Primdahl, 1979], but the sensitivity of Hall devices has improved significantly. The more recent indium-antimonide devices have a lower sensitivity limit of 10-3 gauss [Lenz, 1990]. The U.S. Navy in the early 1960s showed considerable interest in a small solid-state Hall-effect compass for low-power extended operations in sonobuoys [Wiley, 1964]. A number of such prototypes were built and delivered by Motorola for evaluation. The Motorola Hall-effect compass employed two orthogonal Hall elements for temperature-nulled non-ambiguous resolution of the geomagnetic field vector. Each sensor element was fabricated from a 220.1-millimeter indiumarsenide-ferrite sandwich, and inserted between two wing-like Mumetal flux concentrators. It is estimated the 5 cm (2-inch) magnetic concentrators increased the flux density through the sensing elements by two orders of magnitude [Wiley, 1964]. The output of the Motorola unit was a variablewidth pulse train, the width of the pulse being proportional to the sensed magnetic heading. Excellent response linearity was reported down to flux densities of 0.001 gauss [Keller, 1962]. Maenaka et al [1990] report on the development of a monolithic silicon magnetic compass at the Toyohashi University of Technology in Japan, based on two orthogonal Hall-effect sensors. Their use of the terminology magnetic compass is perhaps an unfortunate misnomer, in that the prototype device was tested with an external field of 1,000 gauss. Contrast this with the strength of the Earths magnetic field, which varies from only about 0.1 gauss at the equator to about 0.9 gauss at the poles. Silicon-based Hall-effect sensors have a lower sensitivity limit of around 10 gauss [Lenz, 1990]. It is likely the Toyohashi University device was intended for other than geomagnetic applications, such as remote position sensing of rotating mechanical assemblies. This prototype Hall-effect magnetometer is still of interest in that it represents a fully selfcontained implementation of a two-axis magnetometer in integrated circuit form. Two vertical Hall cells [Maenaka et al, 1987] are arranged at right angles on a 4.7-mm square chip, with their respective outputs coupled to a companion signal processing IC of identical size. (Two separate chips were fabricated for the prototype instead of a single integrated unit to enhance production yield.) The sensor and signal processing ICs are interconnected (along with some external variable resisters for calibration purposes) on a glass-epoxy printed circuit board. The dedicated signal-processing circuitry converts the B-field components B x and B y measured by the Hall sensors into an angle 2 by means of the analog operation [Maenaka et al, 1990]:
= arctan Bx By
Page 48

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

where 2 = angle between B-field axis and sensor Bx = x-component of B-field By = y-component of B-field. The analog output of the signal-processing IC is a DC voltage which varies linearly with vector orientation of the ambient magnetic field in a plane parallel to the chip surface. Reported test results show a fairly straight-line response (i.e., 2 percent full scale) for external field strengths ranging from 8,000 gauss down to 500 gauss; below this level performance begins to degrade rapidly [Maenaka et al., 1990]. A second analog output on the IC provides an indication of the absolute value of field intensity. While the Toyohashi magnetic compass prototype based on silicon Hall-effect technology is incapable of detecting the Earths magnetic field, it is noteworthy nonetheless. A two-axis monolithic device of similar nature employing the more sensitive indium-antimonide Hall devices could potentially have broad appeal for lowcost applications on mobile robotic platforms. An alternative possibility would be to use magnetoresistive sensor elements, which will be discussed in the next section. 2.2.4 Magnetoresistive Compasses The general theory of operation for AMR and GMR magnetoresistive sensors for use in shortrange proximity detection is beyond the scope of this text. However, there are three specific properties of the magnetoresistive magnetometer that make it well suited for use as a geomagnetic sensor: 1) high sensitivity, 2) directionality, and in the case of AMR sensors, 3) the characteristic flipping action associated with the direction of internal magnetization. AMR sensors have an open-loop sensitivity range of 10-2 gauss to 50 gauss (which easily covers the 0.1- to 1.0-gauss range of the earths horizontal magnetic field component), and limitedbandwidth closed-loop sensitivities approaching 10-6 gauss [Lenz, 1990]. Excellent sensitivity, low power consumption, small package size, and decreasing cost make both AMR and GMR sensors increasingly popular alternatives to the more conventional fluxgate designs used in robotic vehicle applications. 2.2.4.1 Philips AMR Compass One of the earliest magnetoresistive sensors to be applied to a magnetic compass application is the KMZ10B offered by Philips Semiconductors BV, The Netherlands [Dibburn and Petersen, 1983; Kwiatkowski & Tumanski, 1986; Petersen, 1989]. The limited sensitivity of this device (approximately 0.1 mV/A/m with a supply voltage of 5V DC) in comparison to the earths maximum horizontal magnetic field (15 A/m) means that considerable attention must be given to error-inducing effects of temperature and offset drift [Petersen, 1989].

Page 49

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

One way around these problems is to exploit the flipping phenomenon by driving the device back and forth between its two possible magnetization states with square-wave excitation pulses applied to an external coil. This switching action toggles the sensors axial magnetic field, resulting in alternating response characteristics. Since the sensor offset remains unchanged while the signal output due to the external magnetic field Hy is inverted, the undesirable DC offset voltages can be easily isolated from the weak AC signal. A typical implementation of this strategy is shown in Fig. 2.15. A 100-Hz square wave generator is capacitively coupled to the external excitation coil L which surrounds two orthogonally mounted magneto-resistive sensors. The sensors output signals are amplified and AC-coupled to a synchronous detector driven by the same squarewave source. The rectified DC voltages VH1 and VH2 are thus proportional to the measured magnetic field components H1 and H2 . The applied field direction is dependant on the ratio of V to H, not their absolute values. This means that as long as the two channels are calibrated to the same sensitivity, no temperature correction is required [Fraden, 1993].
H1 Sensors

H2

R C

VB

Square-wave Generator

Coil L

V(H) 1

Sensor

Amplifier

Synchronous Detector V(H) 2

Figure 2.15: Block diagram of a two-axis magnetic compass system based on a commercially available anisotropic magnetoresistive sensor from Philips [Petersen, 1989).

2.2.5 Magnetoelastic Compasses A number of researchers have recently investigated the use of magnetoelastic (also known as magnetostrictive) materials as sensing elements for high-resolution magnetometers. The principle of operation is based on the changes in Youngs modulus experienced by magnetic alloys when exposed to an external magnetic field. The modulus of elasticity E of a given material is basically a measure of its stiffness, and directly relates stress to strain as follows: E= where E = Youngs modulus of elasticity F = applied stress
Page 50

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

, = resulting strain.
Any ferromagnetic material will experience some finite amount of strain (expansion or shrinkage) in the direction of magnetization due to this magnetostriction phenomenon. It stands to reason that if the applied stress F remains the same, strain , will vary inversely with any change in Youngs modulus E. In certain amorphous metallic alloys, this effect is very pronounced. Barrett et al. [1973] proposed a qualitative explanation, wherein individual atoms in the crystal lattice are treated as tiny magnetic dipoles. The forces exerted by these dipoles on one another depend upon their mutual orientation within the lattice; if the dipoles are aligned end-to-end, the opposite poles attract, and the material shrinks ever so slightly. The crystal is said to exhibit a negative magnetostriction constant in this direction. Conversely, if the dipoles are rotated into sideby-side alignment through the influence of some external field, like poles will repel, and the result is a small expansion. It follows the strength of an unknown magnetic field can be accurately measured if a suitable means is employed to quantify the resulting change in length of some appropriate material displaying a high magnetostriction constant. There are currently at least two measurement technologies with the required resolution allowing the magnetoelastic magnetometer to be a realistic contender for high-sensitivity lowcost performance: 1) interferometric displacement sensing, and, 2) tunneling-tip displacement sensing. Lenz [1990] describes a magnetoelastic magnetometer which employs a Mach-Zender fiber-optic interferometer to measure the change in length of a magnetostrictive material when exposed to an external magnetic field. A laser source directs a beam of light along two optical fiber paths by way of a beam splitter. One of the fibers is coated with a material (nickel iron was used) exhibiting a high magnetostrictive constant. The length of this fiber is stretched or compressed in conjunction with any magnetoelastic expansion or contraction of its coating. The output beam from this fiber optic cable is combined in a light coupler with the output beam from the uncoated reference fiber, and fed to a pair of photodetectors. Constructive and destructive interferences caused by differences in path lengths associated with the two fibers will cause the final output intensity as measured by the photodetectors to vary accordingly. This variation is directly related to the change in path length of the coated fiber, which in turn is a function of the magnetic field strength along the fiber axis. The prototype constructed by Lenz [1990] at McDonnell Douglas Corporation measured 102.5 cm (41 inch), and was able to detect fields ranging from 10-7 gauss up to 10 gauss. Researchers at the U. S. Naval Research Laboratory (NRL) have developed a prototype magnetoelastic magnetometer capable of detecting a field as small as 610-5 gauss [Brizzolara et al., 1989] using the tunneling-tip approach. This new displacement sensing technology, invented in 1982 at IBM Zrich, is based on the measurement of current generated by quantum mechanical tunneling of electrons across a narrow gap. An analog feedback circuit compares the measured
Page 51

Part I: Sensors for Mobile Robot Positioning



Chapter 2: Heading Sensors

tunnel current with a desired value, and outputs a drive signal to suitably adjust the distance between the tunneling electrodes with an electromechanical actuator [Kenny et al., 1991]. The instantaneous tunneling current is directly proportional to the exponential of electrode displacement. The most common actuators employed in this role are piezoelectric and electrostatic, the latter lending itself more readily to silicon micro-machining techniques. The active sense element in the NRL magnetometer is a 10-cm metallic glass ribbon made from METGLAS 2605S2, annealed in a transverse magnetic field to yield a high magnetomechanical coupling [Brizzolara et al., 1989]. (METGLAS is an alloy of iron, boron, silicon, and carbon, and is a registered trademark of Allied Chemical.) The magnetoelastic ribbon elongates when exposed to an axial magnetic field, and the magnitude of this displacement is measured by a tunneling tip. An electrochemically etched gold tip is mounted on a tubular piezoelectric actuator and positioned within about 1 nanometer of the free end of the METGLAS ribbon. The ribbon and tip are electrically biased with respect to each other, establishing a tunneling current that is fed back to the piezo actuator to maintain a constant gap separation. The degree of magnetically induced elongation of the ribbon can thus be inferred from the driving voltage applied to the piezoelectric actuator. The solenoidal coil shown in the diagram supplies a bias field of 0.85 oersted to shift the sensor into its region of maximum sensitivity. Fenn et al. [1992] propose an alternative tunneling-tip magnetoelastic configuration with a predicted sensitivity of 210-11 gauss, along the same order of magnitude as the cryogenically cooled SQUID. A small cantilevered beam of METGLAS 2605S2, excited at its resonant frequency by a gold-film electrostatic actuator, is centered between two high-permeability magnetic flux concentrators. Any changes in the modulus of elasticity of the beam will directly affect its natural frequency; these changes in natural frequency can then be measured and directly related to the strength of the ambient magnetic field. The effective shift in natural frequency is rather small, however (Fenn et al. [1992] report only a 6-Hz shift at saturation), again necessitating a very precise method of measurement. A second (nonmagnetic) cantilever element is employed to track the displacement of the METGLAS reed with sub-angstrom resolution using tunneling-tip displacement sensing. A pair of electrostatic actuator plates dynamically positions the reed follower to maintain a constant tunneling current in the probe gap, thus ensuring a constant lateral separation between the probe tip and the vibrating reed. The frequency of the excitation signal applied to the reed-follower actuator is therefore directly influenced by any resonant frequency changes occurring in the METGLAS reed. The magnetometer provides an analog voltage output which is proportional to this excitation frequency, and therefore indicative of the external magnetic field amplitude.

Page 52

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

Chapter 3: Active Beacons


In this chapter we discuss sensors used for active beacon navigation. Active beacons have been used for many centuries as a reliable and accurate means for navigation. Stars can be considered as active beacons with respect to navigation; and light-houses were early man-made beacon systems. Typical non-robotics applications for active beacon navigation include marine navigation, aircraft navigation, race car performance analysis, range instrumentation, unmanned mobile target control, mine localization, hazardous materials mapping, dredge positioning, geodetic surveys, and most recently, position location and range information for golfers [Purkey, 1994] Modern technology has vastly enhanced the capabilities of active beacon systems with the introduction of laser, ultrasonic and radio-frequency (RF) transmitters. Possibly the most promising approach to outdoor navigation is the recently completed Global Positioning System (GPS). This subject will be addressed in Section 3.1, while Section 3.2 introduces ground-based RF systems. It should be noted that according to our conversations with manufacturers, none of the RF systems can be used reliably in indoor environments. Non-RF beacon systems, based on ultrasonic and light sources will be discussed in Sections 3.3 and 3.4.

3.1 Navstar Global Positioning System (GPS)


The recent Navstar Global Positioning System (GPS) developed as a Joint Services Program by the Department of Defense uses a constellation of 24 satellites (including three spares) orbiting the earth every 12 hours at a height of about 10,900 nautical miles. Four satellites are located in each of six planes inclined 55o with respect to the plane of the earths equator [Getting, 1993]. The absolute 3-D location of any GPS receiver is determined through simple trilateration techniques based on time-of-flight for uniquely coded spread-spectrum radio signals transmitted by the satellites. Precisely measured signal propagation times are converted to pseudo ranges representing the line-ofsight distances between the receiver and a number of reference satellites in known orbital positions. The measured distances have to be adjusted for receiver clock offset, as will be discussed later, hence the term pseudo ranges. Knowing the exact distance from the ground receiver to three satellites theoretically allows for calculation of receiver latitude, longitude, and altitude. Although conceptually very simple (see [Hurn, 1993]), this design philosophy introduces at least four obvious technical challenges:
Page 53

Part I: Sensors for Mobile Robot Positioning


+ + + +

Chapter 3: Active Beacons

Time synchronization between individual satellites and GPS receivers Precise real-time location of satellite position Accurate measurement of signal propagation time Sufficient signal-to-noise ratio for reliable operation in the presence of interference and possible jamming.

The first of these problems is addressed through the use of atomic clocks (relying on the vibration period of the cesium atom as a time reference) on each of the satellites to generate time ticks at a frequency of 10.23 MHZ. Each satellite transmits a periodic pseudo-random code on two different frequencies (designated L1 and L2) in the internationally assigned navigational frequency band. The L1 and L2 frequencies of 1575.42 and 1227.6 MHZ are generated by multiplying the cesium-clock time ticks by 154 and 128, respectively. The individual satellite clocks are monitored by dedicated ground tracking stations operated by the Air Force, and continuously advised of their measured offsets from the ground master station clock. High precision in this regard is critical since electromagnetic radiation propagates at the speed of light, roughly 0.3 m (1 foot) per nanosecond. To establish the exact time required for signal propagation, an identical pseudo code sequence is generated in the GPS receiver on the ground and compared to the received code from the satellite. The locally generated code is shifted in time during this comparison process until maximum correlation is observed, at which point the induced delay represents the time of arrival as measured by the receivers clock. The problem then becomes establishing the relationship between the atomic clock on the satellite and the inexpensive quartz-crystal clock employed in the GPS receiver. This )T is found by measuring the range to a fourth satellite, resulting in four independent trilateration equations with four unknowns. Details of the mathematics involved are presented by Langley [1991]. Precise real-time location of satellite position is determined by a number of widely distributed tracking and telemetry stations at surveyed locations around the world. Referring to Fig. 3.1, all measured and received data are forwarded to a master station for analysis, and referenced to universal standard time. Change orders and signal-coding corrections are generated by the master station and then sent to the satellite control facilities for uploading [Getting, 1993]. In this fashion the satellites are continuously advised of their current position as perceived by the earth-based tracking stations, and encode this ephemeris information into their L1 and L2 transmissions to the GPS receivers. (Ephemeris is the space vehicle orbit characteristics, a set of numbers that precisely describe the vehicle's orbit when entered into a specific group of equations.) In addition to its own timing offset and orbital information, each satellite transmits data on all other satellites in the constellation to enable any ground receiver to build up an almanac after a cold start. Diagnostic information with respect to the status of certain on board systems and expected range-measurement accuracy is also included. This collective housekeeping message is superimposed on the pseudo-random code modulation at a very low (50-bits/second) data rate, and requires 12.5 minutes for complete downloading [Ellowitz, 1992]. Timing offset and ephemeris

Page 54

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

information is repeated at 30-second intervals during this procedure to facilitate initial pseudo-range measurements.
SPACE

CONTROL

USER

Monitor Stations

Uploading Station

Master Station

Figure 3.1: The Navstar Global Positioning System consists of three fundamental segments: Space, Control, and User (adapted from [Getting, 1993]).

To further complicate matters, the sheer length of the unique pseudo-code segment assigned to each individual Navstar Satellite (i.e., around 6.2 trillion bits) for repetitive transmission can potentially cause initial synchronization by the ground receiver to take considerable time. For this and other reasons, each satellite broadcasts two different non-interfering pseudo-codes. The first of these is called the coarse acquisition, or C/A code, and transmitted on the L1 frequency to assist in acquisition. There are 1023 different C/A codes, each having 1023 chips (code bits) repeated 1000 times a second [Getting, 1993] for an effective chip rate of 1.023 MHZ (i.e., one-tenth the cesium clock rate). While the C/A code alone can be employed by civilian users to obtain a fix, the resultant positional accuracy is understandably somewhat degraded. The Y code (formerly the precision or P code prior to encryption on January 1st, 1994) is transmitted on both the L1 and L2 frequencies, and scrambled for reception by authorized military users only with appropriate cryptographic keys and equipment. This encryption also ensures bona fide recipients cannot be "spoofed" (i.e., will not inadvertently track false GPS-like signals transmitted by unfriendly forces). Another major difference between the Y and C/A code is the length of the code segment. While the C/A code is 1023 bits long and repeats every millisecond, the Y code is 2.351014 bits long and requires 266 days to complete [Ellowitz, 1992]. Each satellite uses a 1-week segment of this total code sequence; there are thus 37 unique Y codes (for up to 37 satellites) each consisting of
Page 55

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

6.181012 code bits set to repeat at midnight on Saturday of each week. The higher chip rate of 10.23 MHZ (equal to the cesium clock rate) in the precision Y code results in a chip wavelength of 30 meters for the Y code as compared to 300 meters for the C/A code [Ellowitz, 1992], and thus facilitates more precise time-of-arrival measurement for military purposes. Brown and Hwang [1992] discuss a number of potential pseudo-range error sources as summarized below in Table 3.1. Positional uncertainties related to the reference satellites are clearly a factor, introducing as much as 3 meters (9.8 feet) standard deviation in pseudo-range measurement accuracy. As the radiated signal propagates downward toward the earth, atmospheric refraction and multipath reflections (i.e., from clouds, land masses, water surfaces) can increase the perceived timeof-flight beyond that associated with the optimal straight-line path Error Source Standard Deviation (Fig. 3.2). [meters] [feet] Additional errors can be attribIonospheric Refraction 5 16.4 uted to group delay uncertainties Tropospheric Refraction 2 6.6 introduced by the processing and Multipath Reflection 5 16.4 passage of the signal through the Selective Availability 30 98.4 satellite electronics. Receiver noise and resolution must also be taken into account. Motazed [1993] re- Table 3.1: Summary of potential error sources for measured pseudo-ranges [Brown and Hwang, 1992]. ports fairly significant differences of 0.02 to 0.07 arc minutes in calculated latitudes and longitudes for two identical C/A-code receivers placed side by side. And finally, the particular dynamics of the mobile vehicle that hosts the GPS receiver plays a noteworthy role, in that best-case conditions are associated with a static platform, and any substantial velocity and acceleration will adversely affect the solution. For commercial applications using the C/A code, small errors in timing and satellite position have been deliberately introduced by the master station to prevent a hostile nation from using GPS in support of precision weapons delivery. This intentional degradation in positional accuracy to around 100 m (328 feet) best-case and 200 m (656 feet) typical spherical error probable (SEP) is termed selective availability [Gothard, 1993]. Selective availability has been on continuously (with a few exceptions) since the end of Operation Desert Storm. It was turned off during the war from August 1990 until July 1991 to improve the accuracy of commercial hand-held GPS receivers used by coalition ground forces. There are two aspects of selective availability: epsilon and dither. Epsilon is intentional error in the navigation message regarding the location (ephemeris) of the satellite. Dither is error in the timing source (carrier frequency) that creates uncertainty in velocity measurements (Doppler). Some GPS receivers (for example, the Trimble ENSIGN) employ running-average filtering to statistically reduce the epsilon error over time to a reported value of 15 meters SEP [Wormley, 1994].
Satellite Position 3 29

Page 56

Part I: Sensors for Mobile Robot Positioning


A

Chapter 3: Active Beacons

All of the error sources listed in B Table 3.1 are further influenced by the particular geometry of the four reference satellites at time of sighting. Ignoring time synchronization needs for the moment (i.e., so only Figure 3.2: Atmospheric refraction (A) and multipath reflections (B) three satellites are required), the contribute to pseudo-range measurement errors [Everett, 1995]). most accurate 3-D trilateration solutions will result when the bearing or sight lines extending from the receiver to the respective satellites are mutually orthogonal. If the satellites are spaced closely together in a tight cluster or otherwise arranged in a more or less collinear fashion with respect to the receiver as shown in Fig. 3.3, the desired orthogonality is lost and the solution degrades accordingly. This error multiplier, which can range from acceptable values of 2 or 3 all the way to infinity, is termed geometric dilution of precision [Byrne, 1993]. Kihara and Okada [1984] show where the minimum achievable (best-case) value for GDOP is 1.5811, and occurs when the four required GPS satellites are symmetrically located with an angle of 109.47o between adjacent bearing lines as shown in Fig. 3.4. With the exception of multipath effects, all of the error sources depicted in Table 3.1 above can be essentially eliminated through use of a practice known as differential GPS (DGPS). The concept is based on the premise that a second GPS receiver in fairly close proximity (i.e., within 10 kilometers 6.2 miles) to the Figure 3.3: Worst case geometric dilution of first will experience basically the same error precision (GDOP) errors occur when the receiver and effects when viewing the same reference satel- satellites approach a collinear configuration as shown [Everett, 1995]. lites. If this second receiver is fixed at a precisely surveyed location, its calculated solution can be compared to the known position to generate a composite error vector representative of prevailing conditions in that immediate locale. This differential correction can then be passed to the first receiver to null out the unwanted effects, effectively reducing position error for commercial systems to well under 10 meters.

Page 57

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

The fixed DGPS reference station transmits these correction signals every 2 to 4 minutes to any differential-capable receiver within range. Many commercial GPS receivers are available with differential capability, and most now follow the RTCM-104 standard developed by the Radio Technical Commission for Maritime Services to promote interoperability. Prices for DGPS-capable mobile receivers run about $2K, while the reference stations cost somewhere between $10K and $20K. Magnavox is working with CUE Network Corporation to market a nationwide network to pass differential corrections over an FM link to paid subscribers [GPS Report, 1992].

Z
_ e4 _ e3

_ e2 _ e1

Typical DGPS accuracies are around 4 to 6 m (13 to 20 feet) SEP, with better performance seen as the distance between the mobile receivers and the fixed reference station is decreased. For example, the Coast Guard is in the process of implementing differential GPS in all major US harbors, with an expected accuracy of around 1 m (3.3 feet) SEP [Getting, 1993]. A differential GPS system already in operation at OHare International Airport in Chicago has demonstrated that aircraft and service vehicles can be located to 1 m (3.3 feet). Surveyors use differential GPS to achieve centimeter accuracy, but this practice requires significant post-processing of the collected data [Byrne, 1993]. An interesting variant of conventional DGPS is reported by Motazed [1993] in conjunction with the Non-Line-of-Sight Leader/Follower (NLOSLF) program underway at RedZone Robotics, Inc., Pittsburgh, PA. The NLOSLF operational scenario involves a number of vehicles in a convoy configuration that autonomously follow a lead vehicle driven by a human operator, both on-road and off-road at varying speeds and separation distances. A technique Motazed refers to as intermittent stationary base differential GPS is used to provide global referencing for purposes of bounding the errors of a sophisticated Kalman-filter-based GPS/INS position estimation system. Under this innovative concept, the lead and final vehicle in the convoy alternate as fixedreference differential GPS base stations. As the convoy moves out from a known location, the final vehicle remains behind to provide differential corrections to the GPS receivers in the rest of the vehicles. After traversing a predetermined distance in this fashion, the convoy is halted and the lead vehicle assumes the role of differential reference station, providing enhanced accuracy to the trailing vehicle as it catches up to the pack. During this time, the lead vehicle takes advantage of on-site dwell to further improve the accuracy of its own fix. Once the last vehicle joins up with the rest, the base-station roles are reversed again, and the convoy resumes transit in inch-worm fashion along
Page 58

Figure 3.4: GDOP error contribution is minimal for four GPS satellites symmetrically situated with respect to the receiver (at origin) along bearing lines 109.47o apart [Kihara and Okada, 1984].

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

its intended route. Disadvantages to this approach include the need for intermittent stops, and the accumulating ambiguity in actual location of the appointed reference station. Recall the Y-code chip rate is directly equal to the satellite cesium clock rate, or 10.23 MHZ. Since the L1 carrier frequency of 1575.42 MHZ is generated by multiplying the clock output by 154, there are consequently 154 carrier cycles for every Y-code chip. This implies even higher measurement precision is possible if time-of-arrival is somehow referenced to the carrier instead of the pseudo-code itself. Such codeless interferometric differential GPS schemes measure the phase of the L1 and L2 carrier frequencies to achieve centimeter accuracies, but must start at a known geodetic location and typically require long dwell times. The Armys Engineer Topographic Laboratories (ETL) is in the process of developing a carrier-phase-differential system of this type that is expected to provide 1 to 3 cm (0.4 - 1.2 inches) accuracy at a 60-Hz rate when finished sometime in 1996 [McPherson, 1991]. A reasonable extraction from the open literature of achievable position accuracies for the various GPS configurations is presented in Table 3.2. The Y code has dual-frequency estimation for atmospheric refraction, and no S/A error component, so accuracies are better than stand-alone single-frequency C/A systems. Commercial DGPS accuracy, however, exceeds stand-alone military Y-code accuracy, particularly for small-area applications such as airports. Differential Y code is currently under consideration, and may involve the use of a satellite to disseminate the corrections over a wide area. A typical non-differential GPS was tested by Cooper and DurrantWhite [1994] and yielded an accumulated position error of over 40 m after extensive filtering. Systems likely to provide the best accuracy are those that combine GPS with Inertial Navigation Systems (INS), because the INS position drift is bounded by GPS corrections [Motazed, 1993]. Similarly, the combination of GPS with dead-reckoning and a compass has been proposed by Byrne [1993].
GPS Implementation Method C/A-Code Stand Alone Y-Code Stand Alone Differential (C/A-Code) Differential (Y-Code) Phase Differential (Codeless) Position Accuracy 100 meters SEP (328 feet) 16 meters SEP (52 feet) 3 meter SEP (10 feet) unknown (TBD) 1 centimeter SEP (0.4 inches)

Table 3.2: Summary of achievable position accuracies for various implementations of GPS.

In summary, the fundamental problems associated with using GPS for mobile robot navigation are as follows:
+ Periodic signal blockage due to foliage and hilly terrain + Multipath interference
Page 59

Part I: Sensors for Mobile Robot Positioning


+ Insufficient position accuracy for primary (stand-alone) navigation system.

Chapter 3: Active Beacons

Arradondo-Perry [1992] provides a comprehensive listing of GPS receiver equipment, while Byrne [1993] presents a detailed evaluation of performance for five popular models.

3.2 Ground-Based RF Systems


Ground-based RF position location systems are typically of two types: 1. Passive hyperbolic line-of-position phase-measurement systems that compare the time-of-arrival phase differences of incoming signals simultaneously emitted from surveyed transmitter sites. 2. Active radar-like trilateration systems that measure the round trip propagation delays for a number of fixed-reference transponders. Passive systems are generally preferable when a large number of vehicles must operate in the same local area, for obvious reasons. 3.2.1 Loran An early example of the first category is seen in Loran (short for long range navigation). Developed at MIT during World War II, such systems compare the time of arrival of two identical signals broadcast simultaneously from high-power transmitters located at surveyed sites with a known separation baseline. For each finite time difference (as measured by the receiver) there is an associated hyperbolic line-of-position as shown in Fig. 3.5. Two or more pairs of master-slave stations are required to get intersecting hyperbolic lines resulting in a two-dimensional (latitude and longitude) fix. The original implementation (Loran A) was aimed at assisting convoys of liberty ships crossing the North Atlantic in stormy winter weather. Two 100-kilowatt slave transmitters were located about 200 miles either side of the master station. Non-line-of-sight ground-wave propagation at around 2 MHZ was employed, with pulsed as opposed to continuous-wave transmissions to aid in sky-wave discrimination. The time-of-arrival difference was simply measured as the lateral separation of the two pulses on an oscilloscope display, with a typical accuracy of around 1 microsecond. This numerical value was matched to the appropriate line-of-position on a special Loran chart of the region, and the procedure then repeated for

Vehicle C

A Master Transmitter

B Slave Transmitter

Figure 3.5: For each hyperbolic line-of-position, length ABC minus length AC equals some constant K (adapted from [Dodington, 1989]).

Page 60

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

another set of transmitters. For discrimination purposes, four different frequencies were used, 50 KHz apart, with 24 different pulse repetition rates in the neighborhood of 20 to 35 pulses per second [Dodington, 1989]. In situations where the hyperbolic lines intersected more or less at right angles, the resulting (best-case) accuracy was about 1.5 kilometers. Loran A was phased out in the early 80s in favor of Loran C, which achieves much longer overthe-horizon ranges through use of 5-megawatt pulses radiated from 1300-foot towers at a lower carrier frequency of 100-KHz. For improved accuracy, the phase differences of the first three cycles of the master and slave pulses are tracked by phase-lock-loops in the receiver and converted to a digital readout, which is again cross-referenced to a pre-printed chart. Effective operational range is about 1000 miles, with best-case accuracies in the neighborhood of 100 meters. Coverage is provided by about 50 transmitter sites to all US coastal waters and parts of the North Atlantic, North Pacific, and the Mediterranean. 3.2.2 Kaman Sciences Radio Frequency Navigation Grid The Unmanned Vehicle Control Systems Group of Kaman Sciences Corporation, Colorado Springs, CO, has developed a scaled down version of a Loran-type hyperbolic position location system known as the Radio Frequency Navigation Grid (RFNG). The original application in the late 1970s involved autonomous route control of unmanned mobile targets used in live-fire testing of the laser-guided Copperhead artillery round [Stokes, 1989]. The various remote vehicles sense their position by measuring the phase differences in received signals from a master transmitter and two slaves situated at surveyed sites within a 30-square-kilometer (18.75 square-miles) area as shown in Fig. 3.6. System resolution is 3 cm (1.5 inches) at a 20-Hz update rate, resulting in a vehicle positioning repeatability of 1 m (3.3 feet). Path trajectories are initially taught by driving a vehicle over the desired route and recording the actual phase differences observed. This file is then played back at run time and compared to measured phase difference values, with vehicle steering servoed in an appropriate manner to null any observed error signal. Velocity of advance is directly controlled by the speed of file playback. Vehicle speeds in excess of 50 km/h (30 miles per hour) are supported over path lengths of up to 15 km (9.4 miles) [Stokes, 1989]. Multiple canned paths can be stored and changed remotely, but vehicle travel must always begin from a known start point due to an inherent 6.3-meter (20 feet) phase ambiguity interval associated with the grid [Byrne, et al, 1992]. The Threat Array Control and Tracking Information Center (TACTIC) is offered by Kaman Sciences to augment the RFNG by tracking and displaying the location and orientation of up to 24 remote vehicles [Kaman, 1991]. Real-time telemetry and recording of vehicle heading, position, velocity, status, and other designated parameters (i.e., fuel level, oil pressure, battery voltage) are supported at a 1-Hz update rate. The TACTIC operator has direct control over engine start, automatic path playback, vehicle pause/resume, and emergency halt functions. Non-line-of-sight operation is supported through use of a 23.825-MHZ grid frequency in conjunction with a 72-MHZ

Page 61

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

Figure 3.6: Kaman Sciences 1500-watt navigation grid is a scaled-down version of the LORAN concept, covering an area 8-15 kilometers on a side with a position location repeatability of 1 meter (courtesy Kaman Sciences Corporation).

control and communications channel. 3.2.3 Precision Location Tracking and Telemetry System Precision Technology, Inc., of Saline, MI, has recently introduced to the automotive racing world an interesting variation the conventional phase-shift measurement approach (type 1 RF system). The companys Precision Location tracking and telemetry system employs a number of receive-only antennae situated at fixed locations around a race track to monitor a continuous sine wave transmission from a moving vehicle. By comparing the signals received by the various antennae to a common reference signal of identical frequency generated at the base station, relative changes in vehicle position with respect to each antenna can be inferred from resulting shifts in the respective phase relationships. The 40.58-MHZ VHF signal allows for non-line-of-sight operation, with a resulting precision of approximately 1-10 cm (0.4 - 4 inches) [Duchnowski, 1992]. From a robotics perspective, problems with this approach arise when more than one vehicle must be tracked. The system costs $200,000 - $400,000, depending on the number of receivers used. According to Duchnowski, the system is not suitable for indoor operations. 3.2.4 Motorola Mini-Ranger Falcon An example of the active-transponder category of ground-based RF position-location techniques is seen in the Mini-Ranger Falcon series of range positioning systems offered by the Government and Systems Technology Group of Motorola, Inc, Scottsdale, AZ [MOTOROLA]. The Falcon 484 configuration depicted in Fig. 3.7 is capable of measuring line-of-sight distances from 100 m (328 feet) out to 75 km (47 miles). An initial calibration is performed at a known location to determine the turn-around delay (TAD) for each transponder (i.e., the time required to transmit a response back
Page 62

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

to the interrogator after receipt of interrogation). The actual distance between the interrogator and a given transponder is found by [Byrne et al., 1992] D= ( Te Td ) c 2

where D = separation distance Te = total elapsed time Td = transponder turn-around delay c = speed of light.
Site 1

Display Transeiver Unit

Site 4

Range Processor

Optional Computer

Optional Plotter Site 2 Site 3

Figure 3.7: Motorola's Mini-Ranger Falcon 484 R position location system provides 2 m (6.5 feet) accuracy over ranges of 100 m (328 feet) to 75 km (47 miles) (courtesy [MOTOROLA]).

The MC6809-based range processor performs a least-squares position solution at a 1-Hz update rate, using range inputs from 2, 3 or 4 of 16 possible reference transponders. The individual reference stations answer only to uniquely coded interrogations, and operate in C-band (5410-5890 MHZ) to avoid interference from popular X-band marine radars [Motorola, undated]. Up to 20 mobile users can time share the Falcon 484 system (50 milliseconds per user maximum). System resolution is in tenths of units (meters, feet, or yards) with a range accuracy of 2 m (6.5 feet) probable. Power requirements for the fixed-location reference stations are 22-32 volts DC at 13 watts nominal, 8.5 watts standby, while the mobile range processor and its associated transmitter/receiver and display unit draw 150 watts at 22-32 volts DC. The Falcon system comes in different, customized configurations. Complete system cost is $75,000 - $100,000.

Page 63

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

3.2.5 Harris Infogeometric System Harris Technologies, Inc., [HTI], Clifton, VA, is developing a ground-based R position location and communications strategy wherein moderately priced infogeometric (IG) devices cooperatively form self-organizing instrumentation and communication networks [Harris, 1994]. Each IG device in the network has full awareness of the identity, location, and orientation of all other IG devices, and can communicate with other such devices in both party-line and point-to-point communication modes. The IG devices employ digital code-division-multiple-access (CDMA) spread-spectrum R hardware that provides the following functional capabilities: + Network level mutual autocalibration + Associative location and orientation tracking + Party-line and point-to-point data communications (with video and audio options) + Distributed sensor data fusion. Precision position location on the move is based on high-speed range trilateration from fixed reference devices, a method commonly employed in many instrumentation test ranges and other tracking system applications. In this approach, each beacon has an extremely accurate internal clock that is carefully synchronized with all other beacon-clocks. A time-stamped (coded) R signal is periodically transmitted by each transmitter. The receiver is also equipped with a precision clock, so that it can compare the timing information and time of arrival of the incoming signals to its internal clock. This way, the system is able to accurately measure the signals' time of flight and thus the distance between the receiver and the three beacons. This method, known as "differential location regression" [Harris, 1994] is essentially the same as the locating method used in global positioning systems (GPS). To improve accuracy over current range-lateration schemes, the HTI system incorporates mutual data communications, permitting each mobile user access to the time-tagged range measurements made by fixed reference devices and all other mobile users. This additional network-level range and timing information permits more accurate time synchronization among device clocks, and automatic detection and compensation for uncalibrated hardware delays. Each omnidirectional CDMA spread spectrum geometric transmission uniquely identifies the identity, location, and orientation of the transmitting source. Typically the available geometric measurement update rate is in excess of 1000 KHz. Harris quotes a detection radius of 500 m (1640 feet) with 100 mW peak power transmitters. Larger ranges can be achieved with stronger transmitters. Harris also reports on "centimeter-class repeatability accuracy" obtained with a modified transmitter called an "Interactive Beacon." Tracking and communications at operating ranges of up to 20 km (12.5 miles) are also supported by higher transmission power levels of 1 to 3 watts. Typical raw data measurement resolution and accuracies are cited in Table 3.3 below.

Page 64

Part I: Sensors for Mobile Robot Positioning



Chapter 3: Active Beacons

Parameter Range Bearing (Az, El) Orientation (Az)

Resolution 1 3.3 2 2

Biasing 5 16.4 2 2

Dimension meters feet degrees degrees

Table 3.3: Raw data measurement resolution and accuracy [Everett, 1995].

Enhanced tracking accuracies for selected applications can be provided as cited in Table 3.4 below. This significant improvement in performance is provided by sensor data fusion algorithms that exploit the high degree of relational redundancy characteristic of infogeometric network measurements and communications.
Parameter Range Bearing Orientation Resolution 0.1 - 0.3 0.3 - 0.9 0.5 - 1.0 0.5 - 1.0 Biasing 0.1 - 0.3 0.3 - 0.9 0.5 - 1.0 0.5 - 1.0 Dimension meters feet degrees degrees

Table 3.4: Enhanced tracking resolution and accuracies obtained through sensor data fusion [Everett, 1995].

Infogeometric enhancement algorithms also provide the following capabilities: + Enhanced tracking in multipath and clutter permits precision robotics tracking even when operating indoors. + Enhanced near/far interference reduction permits shared-spectrum operations in potentially large user networks (i.e., hundreds to thousands). Operationally, mobile IG networks support precision tracking, communications, and command and control among a wide variety of potential user devices. A complete Infogeometric Positioning System is commercially available from [HTI], at a cost of $30,000 or more (depending on the number of transmitters required). In conversation with HTI we learned that the system requires an almost clear "line-of-sight" between the transmitters and receivers. In indoor applications, the existence of walls or columns obstructing the path will dramatically reduce the detection range and may result in erroneous measurements, due to multi-path reflections.

Page 65

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

Chapter 4: Sensors for Map-based Positioning


Most sensors used for the purpose of map-building involve some kind of distance measurement. There are three basically different approaches to measuring range: 1. Sensors based on measuring the time-of-flight (TOF) of a pulse of emitted energy traveling to a reflecting object, then echoing back to a receiver 2. The phase-shift measurement (or phase-detection) ranging technique involves continuous wave transmission as opposed to the short pulsed outputs used in TOF systems 3. Sensors based on frequency-modulated (FM) radar. This technique is somewhat related to the (amplitude-modulated) phase-shift measurement technique.

4.1 Time-of-Flight Range Sensors


Many of today's range sensors use the time-of-flight (TOF) method. The measured pulses typically come from an ultrasonic, RF, or optical energy source. Therefore, the relevant parameters involved in range calculation are the speed of sound in air (roughly 0.3 m/ms 1 foot per millisecond), and the speed of light (0.3 m/ns 1 foot per nanosecond). Using elementary physics, distance is determined by multiplying the velocity of the energy wave by the time required to travel the round-trip distance: d = vt where d = round trip distance v = speed of propagation t = elapsed time. The measured time is representative of traveling twice the separation distance (i.e., out and back), and must therefore be reduced by half to result in actual range to the target.
Page 66

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

The advantages of TOF systems arise from the direct nature of their straight-line active sensing. The returned signal follows essentially the same path back to a receiver located coaxially with or in close proximity to the transmitter. In fact, it is possible in some cases for the transmitting and receiving transducers to be the same device. The absolute range to an observed point is directly available as output with no complicated analysis required, and the technique is not based on any assumptions concerning the planar properties or orientation of the target surface. The missing parts problem seen in triangulation does not arise because minimal or no offset distance between transducers is needed. Furthermore, TOF sensors maintain range accuracy in a linear fashion as long as reliable echo detection is sustained, while triangulation schemes suffer diminishing accuracy as distance to the target increases. Potential error sources for TOF systems include the following: + Variations in the speed of propagation, particularly in the case of acoustical systems + Uncertainties in determining the exact time of arrival of the reflected pulse + Inaccuracies in the timing circuitry used to measure the round-trip time-of-flight + Interaction of the incident wave with the target surface. Each of these areas will be briefly addressed below, and discussed later in more detail. Propagation Speed - For mobile robotics applications, changes in the propagation speed of electromagnetic energy are for the most part inconsequential and can basically be ignored, with the exception of satellite-based position location systems as presented in Chapter 3. This is not the case, however, for acoustically based systems, where the speed of sound is markedly influenced by temperature changes, and to a lesser extent by humidity. (The speed of sound is actually proportional to the square root of temperature in degrees Rankine). An ambient temperature shift of just 30o F can cause a 0.3 m (1-foot) error at a measured distance of 10 m (35 feet). Detection Uncertainties - So-called time-walk errors are caused by the wide dynamic range in returned signal strength due to varying reflectivities of target surfaces. These differences in returned signal intensity influence the rise time of the detected pulse, and in the case of fixed-threshold detection will cause the more reflective targets to appear closer. For this reason, constant fraction timing discriminators are typically employed to establish the detector threshold at some specified fraction of the peak value of the received pulse [Vuylsteke, et al., 1990]. Timing Considerations - Due to the relatively slow speed of sound in air, compared to light, acoustically-based systems face milder timing demands than their light-based counterparts and as a result are less expensive. Conversely, the propagation speed of electromagnetic energy can place severe requirements on associated control and measurement circuitry in optical or RF implementations. As a result, TOF sensors based on the speed of light require sub-nanosecond timing circuitry to measure distances with a resolution of about a foot [Koenigsburg, 1982]. More specifically, a desired resolution of 1 millimeter requires a timing accuracy of 3 picoseconds [Vuylsteke, et al., 1990]. This capability is somewhat expensive to realize and may not be cost effective for certain applications, particularly at close range where high accuracies are required.
Page 67

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

Surface Interaction - When light, sound, or radio waves strike an object, any detected echo represents only a small portion of the original signal. The remaining energy reflects in scattered directions, and can be absorbed by or pass through the target, depending on surface characteristics and the angle of incidence of the beam. Instances where no return signal is received at all can occur because of specular reflection at the object surface, especially in the ultrasonic region of the energy spectrum. If the transmission source approach angle meets or exceeds a certain critical value, the reflected energy will be deflected outside of the sensing envelope of the receiver. Scattered signals can reflect from secondary objects as well, returning to the detector at various times to generate false signals yielding questionable or otherwise noisy data (Fig. 4.1). To compensate, repeated measurements are often averaged to bring the signal-to-noise ratio within acceptable levels, but at the expense of additional time required to determine a single range value. 4.1.1 Ultrasonic TOF Systems
C Barrel

Ultrasonic TOF ranging is today the most common technique employed on indoor mobile robotics systems, primarily due to the ready availability of A B low cost systems and their ease of interface. Over Robot Box the past decade, much research has been conducted investigating applicability in such areas as world modeling and collision avoidance, position estima- Figure 5-1: The measured range will reflect the tion, and motion detection. Several researchers round-trip distance through points A, B, and C as opposed to the actual distance from A to B have more recently begun to assess the effective- (adapted from [Everett, 1985]). ness of ultrasonic sensors in exterior settings [Pletta, et al., 1992; Langer & Thorpe, 1992; Pin & Watanabe, 1993; Hammond, 1994]. In the automotive industry, BMW now incorporates four piezoceramic transducers (sealed in a membrane for environmental protection) on both front and rear bumpers in its Park Distance Control system [Siuru, 1994]. Four of the most popular commercially available ultrasonic ranging systems will be reviewed in the following sections. 4.1.1.1 National Semiconductors LM1812 Ultrasonic Transceiver The LM1812 is a general purpose ultrasonic transceiver IC originally designed to support fish- and depth-finding products in the recreational electronics industry [Frederiksen & Howard, 1974]. The 18-pin chip contains a pulse-modulated class-C transmitter, a high-gain receiver, a pulse modulation detector, and noise rejection circuitry [National, 1991]. Maximum range is 30 m (100 feet) in water and 6 m (20 feet) in air, at typical operating frequencies of 20-350 KHz. The chips specifications [National, 1989] list the following features: + Monostatic (single transducer) or bistatic (dual transducer) operation + Transducers interchangeable without realignment + No external transistors required
Page 68

Part I: Sensors for Mobile Robot Positioning


+ + + +

Chapter 4: Sensors for Map-based Positioning

Impulse noise rejection No heat sinking required 12 watts peak transmit power Power consumption 50 milliamps at 18 volts DC.

Two different types of ultrasonic transducers, electrostatic and piezo-ceramic (also known as piezo-electric), are commonly used with the LM1812 [Everett, 1982; Pletta et al., 1992]. Electrostatic transducers transmit an outgoing signal and act as an electrostatic microphone in order to receive the reflected waveform [National, 1991]. Piezo-ceramic transducers are electrically similar to quartz crystals and resonate at only two frequencies: the resonant and antiresonant frequencies. Transmission is most efficient at the resonant frequency while optimum receiving sensitivity occurs at the antiresonant frequency [National, 1989]. In bistatic systems, the resonant frequency of the transmitting transducer is matched to the antiresonant frequency of the receiver for optimal performance. The majority of practical applications, however, use a monostatic configuration for which the maximum echo sensitivity occurs at a frequency close to resonance. 4.1.1.2 Massa Products Ultrasonic Ranging Module Subsystems Massa Products Corporation, Hingham, MA, offers a full line of ultrasonic ranging subsystems with maximum detection ranges from 0.6 to 9.1 m (2 to 30 feet) [MASSA]. The E-201B series sonar operates in the bistatic mode with separate transmit and receive transducers, either side-by-side for echo ranging or as an opposed pair for unambiguous distance measurement between two uniquely defined points. This latter configuration is sometimes used in ultrasonic position location systems, and provides twice the effective operating range with respect to that advertised for conventional echo ranging. The E-220B series (see Fig. 4.2) is designed for monostatic (single transducer) operation, but is otherwise functionally identical to the E-201B. Either version can be externally triggered on command, or internally triggered by a free-running oscillator at a repetition rate determined by an external resistor (see Fig. 4.3). Selected specifications for the four operating frequencies available in the E-220B series are listed in Table 4.1 below. A removable focusing horn is provided for the 26- and 40-KHz models that decreases the effective beam width (when installed) from 35o to 15o. The horn must be in place to achieve the maximum listed range.

Page 69

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

Transmit Driver Trig In Trig Out PRR Internal Oscillator

Receiver

AC AMP

Analog

Threshold

Vcc
+V GND Filter Digital Timing G

Latch

Figure 4.2: The single-transducer Massa E-220B-series ultrasonic ranging module can be internally or externally triggered, and offers both analog and digital outputs (courtesy Massa Products Corp.).
Pulse Repitition Rate Period TRIGGER

Ring Down ANALOG DIGITAL

1st Echo

2nd Echo

Figure 4.3: Timing diagram for the E-220B series ranging module showing analog and digital output signals in relationship to the trigger input (courtesy Massa Products Corp.).

Parameter Range Beam width Frequency Max rep rate Resolution Power Weight

E-220B/215 10 - 61 4 - 24 10 215 150 0.076 0.03 8 - 15 4-8

E-220B/150 20 - 152 8 - 60 10 150 100 0.1 0.04 8 - 15 4-8

E-220B/40 61 - 610 24 - 240 35 (15) 40 25 0.76 0.3 8 - 15 4-8

E-220B/26 61 - 914 24 - 360 35 (15) 26 20 1 0.4 8 - 15 4-8

Units cm inches degrees KHz Hz cm inches volts DC ounces

Table 4.1: Specifications for the monostatic E-220B Ultrasonic Ranging Module Subsystems. The E-201 series is a bistatic configuration with very similar specifications (courtesy Massa Products Corp.). Page 70

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

4.1.1.3 Polaroid Ultrasonic Ranging Modules The Polaroid ranging module is an active TOF device developed for automatic camera focusing, which determines the range to target by measuring elapsed time between transmission of an ultrasonic waveform and the detected echo [Biber et al., 1987, POLAROID]. This system is the most widely found in mobile robotics literature [Koenigsburg, 1982; Moravec & Elfes, 1985; Everett, 1985; Kim, 1986; Moravec, 1988; Elfes, 1989; Arkin, 1989; Borenstein & Koren, 1990; 1991; 1994; Borenstein et al., 1995], and is representative of the general characteristics of such ranging devices. Polaroid offers both the transducer and ranging module circuit board for less than $50, which partially accounts for the widespread usage. The most basic configuration consists of two fundamental components: 1) the ultrasonic transducer, and 2) the ranging module electronics. A choice of transducer types is now available. In the original instrument-grade electrostatic version, a very thin metal diaphragm mounted on a machined backplate formed a capacitive transducer as illustrated in Fig. 4.4 [POLAROID, 1991]. The system operates in the monostatic transceiver mode so that only a single transducer is necessary to acquire range data. A smaller diameter electrostatic transducer (7000-series) has also been made available, developed for the Polaroid Spectra camera [POLAROID, 1987]. A more rugged piezo-electric (9000-series) environmental transducer for applications in severe environment conditions including vibration is able to meet or exceed the SAE J1455 January 1988 specification for heavy duty trucks. Table 4.2 lists the technical specifications for the different Polaroid transducers. The original Polaroid ranging module functioned by transmitting a chirp of four discrete frequencies at about of 50 KHz. The SN28827 module was later developed with reduced parts count, lower power consumption, and simplified computer interface requirements. This second-generation board transmits only a single frequency at 49.1 KHz. A third-generation board (6500 series) introduced in 1990 provided yet a further reduction in interface circuitry, with the ability to detect and report multiple echoes [Polaroid, 1990]. An Ultrasonic Ranging Developers Kit based on the Intel 80C196 microprocessor is now available for use with the 6500 series ranging module that allows

Figure 4.4: The Polaroid instrument grade electrostatic transducer consists of a gold-plated plastic foil stretched across a machined backplate (Reproduced from [Polaroid], 1991]).

Page 71

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

software control of transmit frequency, pulse width, blanking time, amplifier gain, and maximum range [Polaroid, 1993].
Parameter Maximum range Minimum range* Number of pulses Blanking time Resolution Gain steps Multiple echo Programmable frequency Power
*) with custom electronics (see [Borenstein et al., 1995])

Original 10.5 35 25 10.5 56 1.6 1 16 no no 4.7 - 6.8 200

SN28827 10.5 35 20 6 16 2.38 2 12 yes no 4.7 - 6.8 100

6500 10.5 35 20 6 16 2.38 1 12 yes yes 4.7 - 6.8 100

Units m feet cm inches milliseconds percent

volts milliamps

Table 4.2: Specifications for the various Polaroid ultrasonic ranging modules (courtesy Polaroid).

The range of the Polaroid system runs from about 41 cm to 10.5 m (1.33 feet to 35 feet). However, using custom circuitry suggested in [POLAROID, 1991] the minimum range can be reduced reliably to about 20 cm (8") [Borenstein et al., 1995]. The beam dispersion angle is approximately 30o. A typical operating cycle is as follows. 1. The control circuitry fires the transducer and waits for indication that transmission has begun. 2. The receiver is blanked for a short period of time to prevent false detection due to ringing from residual transmit signals in the transducer. 3. The received signals are amplified with increased gain over time to compensate for the decrease in sound intensity with distance. 4. Returning echoes that exceed a fixed threshold value are recorded and the associated distances calculated from elapsed time. Figure 4.5 [Polaroid, 1990] illustrates the operation of the sensor in a timing diagram. In the single-echo mode of operation for the 6500-series module, the blank (BLNK) and blank-inhibit (BINH) lines are held low as the initiate (INIT) line goes high to trigger the outgoing pulse train. The internal blanking (BLANKING) signal automatically goes high for 2.38 milliseconds to prevent transducer ringing from being misinterpreted as a returned echo. Once a valid return is received, the echo (ECHO) output will latch high until reset by a high-to-low transition on INIT. For multipleecho processing, the blanking (BLNK) input must be toggled high for at least 0.44 milliseconds after detection of the first return signal to reset the echo output for the next return.

Page 72

Part I: Sensors for Mobile Robot Positioning


INIT 16 Pulses TRANSMIT (INT) BLNK BINH BLANKIN G (INT) ECHO

Chapter 4: Sensors for Map-based Positioning

Parameter Maximum range Minimum range Accuracy Range jitter Wavelength Diameter Length

Value 10 to 300 3.3-98 2-5 6.5-16.4 0.3 1 11.25 902 89 3.5 178 7 8 to 24

Units meters feet meters feet meter foot centimeters nanometers mm inches mm inches volts DC

Figure 4.5: Timing diagrams for the 6500-Series Sonar Ranging Module executing a multiple-echomode cycle with blanking input (courtesy Polaroid Corp.).

4.1.2 Laser-Based TOF Systems


Power

Laser-based TOF ranging systems, also known as laser radar or lidar, first appeared in Table 4.3: Selected specifications for the LRF-Xwork performed at the Jet Propulsion Labora- series laser rangefinder (courtesy Schwartz Electrotory, Pasadena, CA, in the 1970s [Lewis & Optics, Inc.). Johnson, 1977]. Laser energy is emitted in a rapid sequence of short bursts aimed directly at the object being ranged. The time required for a given pulse to reflect off the object and return is measured and used to calculate distance to the target based on the speed of light. Accuracies for early sensors of this type could approach a few centimeters over the range of 1 to 5 m (3.3 to 16.4 feet) [NASA, 1977; Depkovich & Wolfe, 1984]. 4.1.2.1 Schwartz Electro-Optics Laser Rangefinders Schwartz Electro-Optics, Inc. (SEO), Orlando, FL, produces a number of laser TOF rangefinding systems employing an innovative time-to-amplitude-conversion scheme to overcome the subnanosecond timing requirements necessitated by the speed of light. As the laser fires, a precision capacitor begins discharging from a known set point at a constant rate. An analog-to-digital conversion is performed on the sampled capacitor voltage at the precise instant a return signal is detected, whereupon the resulting digital representation is converted to range using a look-up table. SEO LRF-X Series Rangefinders The LRF-X series rangefinder features compact size, high-speed processing, and the ability to acquire range information from most surfaces (i.e., minimum 10-percent Lambertian reflectivity) out to a maximum of 300 m (984 feet). The basic system uses a pulsed InGaAs laser diode in conjunction with an avalanche photodiode detector, and is available with both analog and digital (RS-232) outputs. Table 4.3 lists general specifications for the sensor's performance [SEO, 1991a]. The High Accuracy Altitude Measurement System (HAAMS) is an enhanced variation of the basic LRF-X concept intended as a lidar altimeter for aircraft (see Table 4.4). The HAAMS system
Page 73

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

operates at a 3-KHz update rate with a maximum range of 76 m (250 feet), and is available in the same 178 mm (7-inch) long by 89 mm (3.5-inch) diameter cylindrical package as the LRF-X. An inclinometer was added to automatically compensate for aircraft angle of bank. In addition, peak detection feedback was incorporated to reduce time-walk errors for an increased accuracy of 7.6 to 10.0 cm (3 to 4 inches). Another adaptation of the LRF-X series involved the addition of a mechanical single-DOF beam scanning capability. Originally developed for use in submunition sensor research, the Scanning Laser Rangefinder is currently installed on board a remotely piloted vehicle. For this application, the sensor is positioned so the forward motion of the RPV is perpendicular to the vertical scan plane, since three-dimensional target profiles are required [SEO, 1991b]. In a second application, the Scanning Laser Rangefinder was used by the Field Robotics Center at Carnegie Mellon University as a terrain mapping sensor on their unmanned autonomous vehicles. SEO Helicopter Obstacle Proximity Sensor System The Helicopter Obstacle Proximity Sensor System (HOPSS) was developed for the U.S. Army [SEO, 1991c] as an on board pilot alert to the presence of surrounding obstructions. Located on the main-rotor driveshaft (Fig. 4.9), the HOPSS system provides continuous distance and azimuth measurements in the horizontal plane of the helicopter.
Laser Collision Av oidance System

Parameter Maximum range Minimum range Accuracy Scan angle Scan rate Samples per scan Wavelength Diameter Length Weight Power

Value 256 840 10 33 15 5 30 24.5- 30.3 175 920 127 5 444 17.5 11.8 8-25

Units meters feet meters feet cm inches degrees KHz nanometers mm inches mm inches pounds volts DC

Laser Altimeter

Figure 4.9: Placement of the Helicopter Optical Proximity Sensor System in a U.S. Army helicopter (courtesy Schwartz Electro-Optics, Inc.).

A high-pulse-repetition-rate GaAs laser-diode emitter shares a common aperture with a sensitive avalanche photodiode detector. The transmit and return beams are reflected from a motor-driven prism rotating at 300 rpm. Range measurements are taken at 0.085o (1.5 milliradian) intervals and correlated with the azimuth angle using an optical encoder. Detected obstacles are displayed in a format similar to a radar plan-position indicator. Table 4.5 lists the key specifications of HOPPS. To achieve broader three-dimensional sensor

Table 4.4: Selected specifications for the SEO Scanning Laser Range finder (courtesy Schwartz Electro-Optics, Inc.).

Page 74

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

coverage, a concept employing two counter-rotating wedge-prisms is under investigation [SEO, 1991d].
Parameter Maximum range Value 2.4-122 8 to 400 Units m feet

SEO TreeSense

TreeSense was developed by SEO for autoAccuracy 61 cm mating the selective application of pesticides to 2 feet orange trees, where the goal was to enable Wavelength 905 nanometers individual spray nozzles only when a tree was Scan angle 360 degrees detected within their associated field of coverScan rate 5 (300?) Hz age. The sensing subsystem consists of a horiSamples per scan 4096 zontally oriented HAAMS unit mounted on the Diameter 178 mm 7 inches back of an agricultural vehicle, suitably Length 300 mm equipped with a rotating mirror arrangement 11.75 inches that scans the beam in a vertical plane orthogoWeight 10 pounds nal to the direction of travel. The scan rate is Power 18 to 36 volts DC controllable up to 40 revolutions per second (35 typical). The ranging subsystem is gated on and off twice during each revolution to illuminate Table 4.5: Selected specifications for the Helicopter two 90o fan-shaped sectors to a maximum range Obstacle Avoidance Proximity Sensor System of 7.6 m (25 feet) either side of the vehicle. The (courtesy Schwartz Electro-Optics, Inc.). existing hardware is theoretically capable of ranging to 30 m (100 feet) using a PIN photodiode, and can be extended further through an upgrade option that incorporates an avalanche photodiode detector. The TreeSense system is hard-wired to a valve manifold to enable/disable a vertical array of nozzles for the spraying of insecticides, but analog as well as digital (RS-232) output can easily be made available for other applications. The system is housed in a rugged fiberglass enclosure with a total weight of only 5 pounds. Power requirements are 12 watts at 12 volts DC. Further details on the system are contained in Table 4.6. SEO AutoSense
Weight Parameter Maximum range Accuracy Wavelength Pulse repetition rate Length Width Height Value 30.48 100 7.5-10 3-4 902 18 229 9 229 9 115 4.5 5 12 Units m feet cm inches nanometers KHz mm inches mm inches mm inches pounds volts DC

The AutoSense I system was developed by SEO under a Department of Transportation Small Business Innovative Research (SBIR)

Power

Table 4.6: Selected specifications for the TreeSense system (courtesy Schwartz Electro-Optics, Inc.).

Page 75

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

effort as a replacement for buried inductive loops for traffic signal control. (Inductive loops dont always sense motorcyclists and some of the smaller cars with fiberglass or plastic body panels, and replacement or maintenance can be expensive as well as disruptive to traffic flow). The system is configured to look down at about a 30 o angle on moving vehicles in a traffic lane. AutoSense I uses a PIN photodiode detector and a pulsed (8 nanosecond) InGaAs near-infrared laser-diode source with peak power of 50 watts. The laser output is directed by a beam splitter into a pair of cylindrical lenses to generate two fan-shaped beams 10o apart in elevation for improved target detection. (The original prototype projected only a single spot of light, but ran into problems due to target absorption and specular reflection). As an added benefit, the use of two separate beams makes it possible to calculate the speed of moving vehicles to an accuracy of 1.6 km/h (1 mph). In addition, a 2-D image (i.e., length and width) is formed of each vehicle as it passes through the sensors field of view, opening the door for numerous vehicle classification applications under the Intelligent Vehicle Highway Systems concept. An improved second-generation unit (AutoSense II) uses an avalanche photodiode detector instead of the PIN photodiode for greater sensitivity, and a multi-faceted rotating mirror with alternating pitches on adjacent facets to create the two beams. Each beam is scanned across the traffic lane 720 times per second, with 15 range measurements made per scan. This azimuthal scanning action generates a precise 3-D profile to better facilitate vehicle classification in automated toll booth applications. An abbreviated system block diagram is depicted in Fig. 4.12.
Lens Laser Diode Laser Driver Trigger Circuit

Scanner Interface

Laser Trigger Detector

Threshold Detector

FO Line

Ref

Lens

Optical Filter Detector Start Stop

Peak Detector MicroRange Gate Time to Amplitude Converter Processor

RS 422

RS 232

Figure 4.12: Simplified block diagram of the AutoSense II time-of-flight 3-D ranging system (courtesy Schwartz Electro-Optics, Inc.).

Intensity information from the reflected signal is used to correct the time walk error in threshold detection resulting from varying target reflectivities, for an improved range accuracy of 7.6 cm
Page 76

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

(3 inches) over a 1.5 to 15 m (5 to 50 foot) field of regard. The scan resolution is 1o , and vehicle velocity can be calculated with an accuracy of 3.2 km/h (2 mph) at speeds up to 96 km/h (60 mph). High-speed RS-232 and RS-422 outputs are provided. A third-generation AutoSense III is now under development for an application in Canada that requires 3-D vehicle profile generation at speeds up to 160 km/h (100 mph). Selected specifications for the AutoSense II package are provided in Table 4.7. 4.1.2.2 RIEGL Laser Measurement Systems

Parameter Maximum range Accuracy Wavelength Pulse repetition rate Power

Value 50 3 904 86.4 12/24

Units feet inches nanometers KHz volts DC

Table 4.7: Selected specifications for the AutoSense II ranging system (courtesy Schwartz Electro-Optics, Inc.).

RIEGL Laser Measurement Systems, Horn, Austria, offers a number of commercial products (i.e., laser binoculars, surveying systems, speed guns, level sensors, profile measurement systems, and tracking laser scanners) employing short-pulse TOF laser ranging. Typical applications include lidar altimeters, vehicle speed measurement for law enforcement, collision avoidance for cranes and vehicles, and level sensing in silos. All RIEGL products are distributed in the United States by RIEGEL USA, Orlando, FL. LD90-3 Laser Rangefinder The RIEGL LD90-3 series laser rangefinder employs a near-infrared laser diode source and a photodiode detector to perform TOF ranging out to 500 meters with diffuse surfaces, and to over 1000 meters in the case of cooperative targets. Round-trip propagation time is precisely measured by a quartz-stabilized clock and converted to measured distance by an internal microprocessor using one of two available algorithms. The clutter suppression algorithm incorporates a combination of range measurement averaging and noise rejection techniques to filter out backscatter from airborne particles, and is therefore useful when operating under conditions of poor visibility [Riegl, 1994]. The standard measurement algorithm, on the other hand, provides rapid range measurements without regard for noise suppression, and can subsequently deliver a higher update rate under more favorable environmental conditions. Worst case range measurement accuracy is 5 cm (2 inches), with typical values of around 2 cm (0.8 inches). See Table 4.8 for a complete listing of the LD90-3's features.

Page 77

Part I: Sensors for Mobile Robot Positioning


Parameter Maximum range (diffuse) (cooperative) Minimum range Accuracy LD90-3100 150 492 >1000 >3,280 1 2 0.3 2 (digital) (analog) Power Size Weight RS-232, -422 0-10 11-18 10 22137.6 8.75.13 3.3 RS-232, -422 0-10 11-18 10 22137.6 8.75.13 3.3 volts DC volts DC watts centimeters inches pounds LD90-3300 400 1,312 >1000 >3,280 3-5 5 2 0.5 2.8 Units meters feet meters feet meters centimeters inches meters/sec milliradians

Chapter 4: Sensors for Map-based Positioning

(distance) (velocity)

Beam divergence Output

Table 4.8: Selected specifications for the RIEGL LD90-3 series laser rangefinder (courtesy RIEGL Laser Measurement Systems).

The pulsed near-infrared laser is Class-1 eye safe under all operating conditions. A nominal beam divergence of 0.1 o (2 milliradians) for the LD90-3100 unit (see Table 4.9 below) produces a 20 cm (8 inches) footprint of illumination at 100 meters (328 feet) [Riegl, 1994]. The complete system is housed in a small light-weight metal enclosure weighing only 1.5 kg (3.3 lb), and draws 10 watts at 11-18 volts DC. The standard output format is serial RS-232 at programmable data rates up to 19.2 kilobits per second, but RS-422 as well as analog options (0-10 volts DC and 4-20 milliamps currentloop) are available upon request. Scanning Laser Rangefinders The LRS90-3 Laser Radar Scanner is an adaptation of the basic LD90-3 electronics, fiberoptically coupled to a remote scanner unit. The scanner package contains no internal electronics and is thus very robust under demanding operating conditions typical of industrial or robotics scenarios. The motorized scanning head pans the beam back and forth in the horizontal plane at a o 10-Hz rate, resulting in 20 data-gathering sweeps per second. Beam divergence is 0.3 o (5 milliradians), with the option of expanding in the vertical direction if desired up to 2 . The LSS390 Laser Scanning System is very similar to the LRS90-3, but scans a more narrow field-of-view (10o) with a faster update rate (2000 Hz) and a more tightly focused beam. Range accuracy is 10 cm (4 inches) typically and 20 cm (8 inches) worst case. The LSS390 unit is available

Page 78

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

with an RS-422 digital output (19.2 kilobits standard, 150 kilobits optional) or a 20-bit parallel TTL interface.
Parameter Maximum range Minimum range Accuracy Beam divergence Sample rate Scan range Scan rate Output Power Size (electronics) (scanner) Weight (electronics) (scanner) (digital) LRS90-3 80 262 2 6.5 3 1.2 5 1000 18 10 RS-232, -422 11-15 880 22137.6 8.75.13 181010 744 7.25 3.52 LSS390 60 197 1 3.25 10 4 3.5 2000 10 10 parallel, RS-422 9-16 22137.6 8.75.13 181010 744 2.86 2 volts DC milliamps cm inches cm inches pounds pounds Units meters feet meters feet cm feet milliradians Hz degrees scans/second

Table 4.9: Typical specifications for the LRS90-3 Laser Radar Scanner and the LSS390 Laser Scanner System (courtesy RIEGL USA).

4.1.2.3 RVSI Long Optical Ranging and Detection System Robotic Vision Systems, Inc., Haupaugue, NY, has conceptually designed a laser-based TOF ranging system capable of acquiring three-dimensional image data for an entire scene without scanning. The Long Optical Ranging and Detection System (LORDS) is a patented concept incorporating an optical encoding technique with ordinary vidicon or solid state camera(s), resulting in precise distance measurement to multiple targets in a scene illuminated by a single laser pulse. The design configuration is relatively simple, and comparable in size and weight to traditional TOF and phase-shift measurement laser rangefinders (Fig. 4.13).
Cone Shaped Object Laser Range Gate Timing Generator CCD Array
Figure 4.13: Simplified block diagram of a three-camera configuration of the LORDS 3-D laser TOF rangefinding system (courtesy Robotics Vision Systems, Inc.). Page 79

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

Major components will include a single laser-energy source; one or more imaging cameras, each with an electronically implemented shuttering mechanism; and the associated control and processing electronics. In a typical configuration, the laser will emit a 25-millijoule pulse lasting 1 nanosecond, for an effective transmission of 25 megawatts. The anticipated operational wavelength will lie between 532 and 830 nanometers, due to the ready availability within this range of the required laser source and imaging arrays. The cameras will be two-dimensional CCD arrays spaced closely together with parallel optical axes resulting in nearly identical, multiple views of the illuminated surface. Lenses for these cameras will be of the standard photographic varieties between 12 and 135 millimeters. The shuttering function will be performed by microchannel plate image intensifiers (MCPs) 18 or 25 millimeters in size, which will be gated in a binary encoding sequence, effectively turning the CCDs on and off during the detection phase. Control of the system will be handled by a single-board processor based on the Motorola MC-68040. LORDS obtains three-dimensional image information in real time by employing a novel time-offlight technique requiring only a single laser pulse to collect all the information for an entire scene. The emitted pulse journeys a finite distance over time; hence, light traveling for 2 ms will illuminate a scene a greater distance away than light traveling only 1 ms. The entire sensing range is divided into discrete distance increments, each representing a distinct range plane. This is accomplished by simultaneously gating the MCPs of the observation cameras according to their own unique on-off encoding pattern over the duration of the detection phase. This binary gating alternately blocks and passes any returning reflection of the laser emission off objects within the field-of-view. When the gating cycles of each camera are lined up and compared, there exists a uniquely coded correspondence which can be used to calculate the range to any pixel in the scene. For instance, in a system configured with only one camera, the gating MCP would be cycled on for half the detection duration, then off the remainder of the time. Figure 4.14 shows any object detected by this camera must be positioned within the first half of the sensors overall range (half the distance the laser light could travel in the allotted detection time). However, significant distance ambiguity exists because the exact time of detection of reflected energy could have occurred anywhere within this relatively long interval.

Page 80

Part I: Sensors for Mobile Robot Positioning


Transmitted Pulse Schematic Of Portion Illuminated vs Time Schematic Of Portion Received vs Time (Delayed) Range Gate 1 (A) Range Gate 2 (B) Range Gate 3 (C)
7 65 4 3 2 1

Chapter 4: Sensors for Map-based Positioning

Object to Lens Delay

Figure 4.14: Range ambiguity is reduced by increasing the number of binary range gates (courtesy Robotic Vision Systems, Inc.).

This ambiguity can be reduced by a factor of two through the use of a second camera with its associated gating cycled at twice the rate of the first. This scheme would create two complete on-off sequences, one taking place while the first camera is on and the other while the first camera is off. Simple binary logic can be used to combine the camera outputs and further resolve the range. If the first camera did not detect an object but the second did, then by examining the instance when the first camera is off and the second is on, the range to the object can be associated with a relatively specific time frame. Incorporating a third camera at again twice the gating frequency (i.e., two cycles for every one of camera two, and four cycles for every one of camera one) provides even more resolution. As Figure 4.15 shows, for each additional CCD array incorporated into the system, the number of distance divisions is effectively doubled.
Range Gate 1 Range Gate 2 Range Gate 3 Composite

1 2 3 4 5 6 7

Figure 4.15: Binary coded images from range gates 1-3 are combined to generate the composite range map on the far right (courtesy Robotics Vision Systems, Inc.).

Alternatively, the same encoding effect can be achieved using a single camera when little or no relative motion exists between the sensor and the target area. In this scenario, the laser is pulsed multiple times, and the gating frequency for the single camera is sequentially changed at each new transmission. This creates the same detection intervals as before, but with an increase in the time required for data acquisition.

Page 81

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

LORDS is designed to operate over distances between one meter and several kilometers. An important characteristic is the projected ability to range over selective segments of an observed scene to improve resolution, in that the depth-of-field over which a given number of range increments is spread can be variable. The entire range of interest is initially observed, resulting in the maximum distance between increments (coarse resolution). An object detected at this stage is thus localized to a specific, abbreviated region of the total distance. The sensor is then electronically reconfigured to cycle only over this region, which significantly shortens the distance between increments, thereby increasing resolution. A known delay is introduced between transmission and the time when the detection/gating process is initiated. The laser light thus travels to the region of interest without concern for objects positioned in the foreground.

4.2 Phase Shift Measurement


The phase-shift measurement (or phase-detection) ranging technique involves continuous wave transmission as opposed to the short pulsed outputs used in TOF systems. A beam of amplitudemodulated laser, RF, or acoustical energy is directed towards the target. A small portion of this wave (potentially up to 6 orders of magnitude less in amplitude) is reflected by the object surface back to the detector along a direct path [Chen, et al., 1993]. The returned energy is compared to a simultaneously generated reference that has been split off from the original signal, and the relative phase shift between the two is measured as illustrated in Fig. 4.16 to ascertain the round-trip distance the wave has traveled. For high-frequency RF- or laser-based systems, detection is usually preceded by heterodyning the reference and received signals with an intermediate frequency (while preserving the relative phase shift) to allow the phase detector to operate at a more convenient lower frequency [Vuylsteke, 1990].

Page 82

Part I: Sensors for Mobile Robot Positioning


n=8 Tx n=7 n=6 n=5

Chapter 4: Sensors for Map-based Positioning

x n=1 Rx n=2 n=3 n=4

Liquid Surface

Figure 4.16: Relationship between outgoing and reflected waveforms, where x is the distance corresponding to the differential phase (adapted from [Woodbury, et al., 1993]).

The relative phase shift expressed as a function of distance to the reflecting target surface is [Woodbury, et al., 1993] = 4 d

where N = phase shift d = distance to target 8 = modulation wavelength. The desired distance-to-target d as a function of the measured phase shift N is therefore given by
d= c = 4 4 f

where f = modulation frequency. For square-wave modulation at the relatively low frequencies typical of ultrasonic systems (20200 KHz), the phase difference between incoming and outgoing waveforms can be measured with the simple linear circuit shown in Fig. 4.17 [Figueroa & Barbieri, 1991]. The output of the exclusiveor gate goes high whenever its inputs are at opposite logic levels, generating a voltage across
Page 83

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

capacitor C that is proportional to the phase shift. For example, when the two signals are in phase (i.e., N = 0), the gate output stays low and V is zero; maximum output voltage occurs when N reaches 180o. While easy to implement, this simplistic approach is limited to very low frequencies, and may require frequent calibration to compensate for drifts and offsets due to component aging or changes in ambient conditions [Figueroa & Lamancusa, 1992]. At higher frequencies, the phase shift between outgoing and reflected sine waves can be measured by multiplying the two signals together in an electronic mixer, then averaging the product over many modulation cycles [Woodbury et al., 1993]. This integration process can be relatively time consuming, making it difficult to achieve extremely rapid update rates. The result can be expressed mathematically as follows [Woodbury et al., 1993]:

lim

1 T

sin
0

2c t + 4d sin 2 c dt

Phase V1 Vp

T
which reduces to 4 d A cos

R V C
Figueroa.ds4 ,.wp g

V2

XOR Gate

where t = time T = averaging interval A = amplitude factor from gain of integrating amplifier.

Figure 4.17: At low frequencies typical of ultrasonic systems, a simple phase-detection circuit based on an exclusive-or gate will generate an analog output voltage proportional to the phase difference seen by the inputs (adapted from [Figueroa & Barbieri, 1991]).

From the earlier expression for N, it can be seen that the quantity actually measured is in fact the cosine of the phase shift, and not the phase shift itself [Woodbury et al., 1993]. This situation introduces a so called ambiguity interval for scenarios where the round-trip distance exceeds the modulation wavelength (i.e., the phase measurement becomes ambiguous once N exceeds 360 o). Conrad and Sampson [1990] define this ambiguity interval as the maximum range that allows the phase difference to go through one complete cycle of 360 o: Ra = c 2f

where: Ra = ambiguity range interval f = modulation frequency c = speed of light.

Page 84

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

Referring again to Fig. 4.16, it can be seen that the total round trip distance 2d is equal to some integer number of wavelengths n8 plus the fractional wavelength distance x associated with the phase shift. Since the cosine relationship is not single valued for all of N, there will be more than one distance d corresponding to any given phase shift measurement [Woodbury, et al., 1993]:
2 (x + n ) 4 d cos = cos = cos

where: d = (x + n8 ) / 2 = true distance to target. x = distance corresponding to differential phase N. n = number of complete modulation cycles. The potential for erroneous information as a result of this ambiguity interval reduces the appeal of phase-detection schemes. Some applications simply avoid such problems by arranging the optical path so that the maximum possible range is within the ambiguity interval. Alternatively, successive measurements of the same target using two different modulation frequencies can be performed, resulting in two equations with two unknowns, allowing both x and n to be uniquely determined. Kerr [1988] describes such an implementation using modulation frequencies of 6 and 32 MHZ. Advantages of continuous-wave systems over pulsed time-of-flight methods include the ability to measure the direction and velocity of a moving target in addition to its range. In 1842, an Austrian by the name of Johann Doppler published a paper describing what has since become known as the Doppler effect. This well known mathematical relationship states that the frequency of an energy wave reflected from an object in motion is a function of the relative velocity between the object and the observer. This subject was discussed in detail in Chapter 1. As with TOF rangefinders, the paths of the source and the reflected beam are coaxial for phaseshift-measurement systems. This characteristic ensures objects cannot cast shadows when illuminated by the energy source, preventing the missing parts problem. Even greater measurement accuracy and overall range can be achieved when cooperative targets are attached to the objects of interest to increase the power density of the return signal. Laser-based continuous-wave (CW) ranging originated out of work performed at the Stanford Research Institute in the 1970s [Nitzan et al., 1977]. Range accuracies approach those of pulsed laser TOF methods. Only a slight advantage is gained over pulsed TOF rangefinding, however, since the time-measurement problem is replaced by the need for fairly sophisticated phase-measurement electronics [Depkovich & Wolfe, 1984]. Because of the limited information obtainable from a single range point, laser-based systems are often scanned in one or more directions by either electromechanical or acousto-optical mechanisms.

Page 85

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

4.2.1 ERIM 3-D Vision Systems The Adaptive Suspension Vehicle (ASV) developed at Ohio State University [Patterson et al., 1984] and the Autonomous Land Vehicle (ALV) developed by Martin Marietta Denver Aerospace were the premier mobile robot projects sponsored by the Defense Advanced Research Projects Agency (DARPA) in the 1980s under the Strategic Computing Program. In support of these efforts, the Environmental Research Institute of Michigan (ERIM) was tasked to develop an advanced threedimensional vision system to meet the close-in navigation and collision avoidance needs of a mobile platform. The initial design, known as the Adaptive Suspension Vehicle Sensor operates on the principle of optical radar and determines range to a point through phase-shift measurement using a CW laser source [Beyer et al., 1987]. The ranging sequence begins with the transmission of an amplitude-modulated laser beam that illuminates an object and is partially reflected back to the detector, generating a representative signal that is amplified and filtered to extract the 70-MHZ modulation frequency. The amplitude of the signal is picked off at this point to produce a reflectance video image for viewing or for twodimensional image processing. A reference signal is output by the modulation oscillator; both the detector and reference signals are then sent to the comparator electronics. The resulting phase difference is determined by a time-measurement technique, where the leading edge of the reference signal initiates a counting sequence that is terminated when the leading edge of the returned signal enters the counter. The resulting count value is a function of the phase difference between the two signals and is converted to an 8-bit digital word representing the range to the scene. Three-dimensional images are produced by the ASV sensor through the use of scanning optics. The mechanism consists of a nodding mirror and a rotating polygonal mirror with four reflective surfaces as shown in Fig. 4.18. The polygonal mirror pans the transmitted laser beam in azimuth across the ground, creating a scan line at a set distance in the front of the vehicle. The scan line is deflected by the objects and surfaces in the observed region and forms a contour of the scene across the sensor's horizontal field of view. The third dimension is added by the nodding mirror which tilts the beam in discrete elevation increments. A complete image is created by scanning the laser in a left-to-right and bottom-to-top raster pattern. The returning signals share the same path through the nodding mirror and rotating polygon (actually slightly offset), but are split off through a separate glass optical chain to the detector. The scan rate of 180 lines per second is a function of the field of view and the desired frame rate. The scan rate is determined by the vehicle's maximum forward velocity, 3 m/s (10 feet/sec) in this case. The size, weight, and required velocities of the mirrors preclude the use of galvanometers in the system design; the rotating and nodding mirrors therefore are servo driven.

Page 86

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

Nodding Mirror The output beam is produced by a gallium arsenide (GaAs) laser diode emitting at a wavelength of Output 820 nanometers. The small size, Beam internal modulation, low temperature, and limited eye hazard of this Folding Mirror source resulted in its selection over an alternative carbon dioxide laser used in a trade-off analysis. The only significant drawbacks for the GaAs device were reduced effectiveness in bright sunlight, and only Receiver Transmitter fair penetration capabilities in fog Figure 4.18: Scanning and nodding mirror arrangement in the ERIM and haze. Collimating and expansion optics were required by the laser rangefinder for the Adaptive Suspension Vehicle (courtesy Environmental Research Institute of Michigan). system to produce a 15 cm (6-inch) diameter laser footprint at 9 m (30 feet), which is the effective operating range of the sensor. (The major factor limiting the useful range of the system is the measurement ambiguity that occurs when the phase difference between the reference and returned energy exceeds 360 o.)

The detector is a silicon avalanche photodiode optically matched to the laser wavelength. The laser source, detector, scanning optics, and drive motors are housed in a single enclosure, designed to be situated at a height of 2.44 m (8 feet) to look down upon the field of view. From this vantage point, the laser strikes the ground between 0.6 and 9 m (2 and 30 feet) in front of the vehicle; at 9 m (30 feet), the horizontal scan line is 6.7 m (22 feet) wide. The 2-Hz frame rate for the system creates a new image of the scene for every 1.5 m (5 feet) of forward motion at the vehicle's maximum speed of 3 m/sec (10 feet/sec). Following the design and fabrication of the ASV sensor, ERIM undertook the task of developing a similar device known as the ALV sensor for DARPA's autonomous land vehicle. The two instruments are essentially the same in configuration and function, but with modified performance specifications to meet the needs of the individual mobile platforms (Table 4.10). An advanced ranging device known as the Multispectral ALV Sensor (MS-ALV) was later developed for exterior applications addressing rugged cross-country traversal as opposed to the relatively uniform road surfaces seen in the initial tests of the autonomous land vehicle concept. The variations in terrain, surface cover, and vegetation encountered in off-road scenarios require an effective means to distinguish between earth, rocks, grass, trees, water, and other natural features. The scanner for the MS-ALV is essentially identical to the scanners developed for the earlier ASV and ALV sensors, the only significant difference being the substitution of a hexagonal rotating mirror instead of a square mirror for panning the beam in azimuth. This configuration causes the transmitted and returned signals to impinge on separate mirrored surfaces, resulting in reduced cross
Page 87

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

Parameter Horizontal FOV Vertical FOV Beam width Frame rate Scan lines per frame Pixels per scan line Maximum range Vertical scan Wavelength Power Size Weight

ASV 80 60 1 2 128 128 10 32 10 820 24 450 356656 142622 85

ALV 80 30 0.5 2 64 256 20 64 20 820 24 450 357456 142922 85

Units degrees degrees degrees Hz

m feet degrees nanometers volts watts cm inches pounds

Table 4.10: Selected specifications for the Adaptive Suspension Vehicle and Autonomous Land Vehicle scanning laser rangefinders (courtesy Environmental Research Institute of Michigan).

talk and simplified sensor alignment (Fig. 4.19). The nodding mirror that tilts the beam in elevation remains largely unchanged.
Planar Mirror

Transmitter Beam Expander

Azimuth Scan Mirror

Field Stop

Receiver Objective

Elevation System Mirror Window Figure 4.19: Hexagonal rotating mirror used in the multispectral scanner reduces cross talk and simplifies mirror alignment (courtesy Environmental Research Institute of Michigan).

Planar Mirror

Page 88

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

The mass of the scanning mechanism plus the plurality of lasers, optics, and detectors made the multispectral sensor large 3.60.90.6 m (1232 feet) and heavy (600 pounds), increasing the complexity of the control and analysis required to produce results. The multiple frequency sources, corresponding detectors, detector cooling system, and scanner resulted in significant power consumption: 15 kilowatts! 4.2.2 Odetics Scanning Laser Imaging System Odetics, Inc., Anaheim, CA, developed an adaptive and versatile scanning laser rangefinder in the early 1980s for use on ODEX 1, a six-legged walking robot [Binger & Harris, 1987; Byrd & DeVries, 1990]. The system determines distance by phase-shift measurement, constructing threedimensional range pictures by panning and tilting the sensor across the field of view. The phase-shift measurement technique was selected over acoustic ranging, stereo vision and structured light alternatives because of the inherent accuracy and fast update rate. The imaging system is broken down into the two major sub-elements depicted in Fig. 4.20: the scan unit and the electronics unit. The scan unit houses the laser source, the photodetector, and the scanning mechanism. The laser source is a GaAlAs laser diode emitting at a wavelength of 820 nanometers; the power output is adjustable under software control between 1 to 50 milliwatts. Detection of the returned energy is achieved through use of an avalanche photodiode whose output is routed to the phase-measuring electronics.
Scan Unit 60 FOV Raster Scan Electronics Unit Range Adp Cw Diode Laser Range/ Video Processor Programmable Frame Buffer Interface Video Sync

Scan Mechanism Phaselock Processor

Figure 4.20: Block diagram of the Odetics scanning laser rangefinder (courtesy Odetics, Inc.).

The scanning hardware consists of a rotating polygonal mirror which pans the laser beam across the scene, and a planar mirror whose back-and-forth nodding motion tilts the beam for a realizable field-of-view of 60o in azimuth and 60o in elevation. The scanning sequence follows a raster-scan pattern, and can illuminate and detect an array of 128128 pixels at a frame rate of 1.2 Hz [Boltinghouse et al., 1990].

Page 89

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

The second sub-element, the electronics unit, contains the range calculating and video processor as well as a programmable frame buffer interface. The range and video processor is responsible for controlling the laser transmission, activation of the scanning mechanism, detection of the returning energy, and determination of range values. Distance is calculated through a proprietary phasedetection scheme, reported to be fast, fully digital, and self-calibrating with a high signal-to-noise ratio. The minimum observable range is 0.46 m (1.5 feet), while the maximum range without ambiguity due to phase shifts greater than 360o is 9.3 m (30 feet). For each pixel, the processor outputs a range value and a video reflectance value. The video data are equivalent to that obtained from a standard black and white television camera, except that interference due to ambient light and shadowing effects are eliminated. The reflectance value is compared to a pre-specified threshold to eliminate pixels with insufficient return intensity to be properly processed, thereby eliminating potentially invalid range data; range values are set to maximum for all such pixels [Boltinghouse & Larsen, 1989]. A 33 neighborhood median filter is used to further filter out noise from data qualification, specular reflection, and impulse response [Larson & Boltinghouse, 1988]. The output format is a 16-bit data word consisting of the range value in either 8 or 9 bits, and the video information in either 8 or 7 bits, respectively. The resulting range resolution for the system is 3.66 cm (1.44) inches for the 8-bit format, and 1.83 cm (0.72 inch) with 9 bits. A buffer interface provides interim storage of the data and can execute single-word or whole-block direct-memoryaccess transfers to external host controllers under program control. Information can also be routed directly to a host without being held in the buffer. Currently, the interface is designed to support VAX, Center of Reflected Rotation VME-Bus, Multibus, and IBMLight Back PC/AT equipment. The scan and Light electronics unit together weigh 31 Out Mirror pounds and require 2 amps at 28 volts DC.
Lens

4.2.3 ESP Optical Ranging System A lowcost near-infrared range finder (see Fig. 4.21 and Table 4.11) was developed in 1989 by ESP Technologies, Inc., Lawrenceville, NJ, for use in autonomous robot cart navigation in factories and similar environments. An eye-safe 2-mW, 820-nanometer LED source is 100 percent modulated at 5-MHZ and used to form a
Motor LED Lens

Detector 6.0" max.

Figure 4.21: Schematic drawing of the ORS-1 ranging system (courtesy ESP Technologies, Inc.). Page 90

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

collimated 2.5 cm (1-inch) diameter transmit Parameter Value Units beam that is unconditionally eye-safe. Reflected Accuracy < 6 inches radiation is focused by a 10 cm (4-inch) diameter AGC output 1-5 volts coaxial Fresnel lens onto the photodetector; the Output power 2 milliwatts measured phase shift is proportional to the roundBeam width 2.5 cm trip distance to the illuminated object. The 1 inch Optical Ranging System (ORS-1) provides three Dimensions 151530 cm outputs: range and angle of the target, and an 6612 inches automatic gain control (AGC) signal [Miller & Weight pounds Wagner, 1987]. Range resolution at 6.1 m (20 Power 12 volts DC feet) is approximately 6 cm (2.5 inches), while 2 amps angular resolution is about 2.5 cm (1 inch) at a range of 1.5 m (5 feet). The ORS-1 AGC output Table 4.11: Selected specifications for the LEDsignal is inversely proportional to the received based near-infrared Optical Ranging System signal strength and provides information about a (courtesy ESP Technologies, Inc.). targets near-infrared reflectivity, warning against insufficient or excessive signal return [ESP, 1992]. Usable range results are produced only when the corresponding gain signal is within a predetermined operating range. A rotating mirror mounted at 45o to the optical axis provides 360o polar-coordinate coverage. It is driven at 1 to 2 revolutions per second by a motor fitted with an integral incremental encoder and an optical indexing sensor that signals the completion of each revolution. The system is capable of simultaneous operation as a wideband optical communication receiver [Miller & Wagner, 1987]. The ORS-1 AGC output signal is inversely proportional to the received signal strength and provides information about a targets near-infrared reflectivity, warning against insufficient or excessive signal return [ESP, 1992]. Usable range results are produced only when the corresponding gain signal is within a predetermined operating range. A rotating mirror mounted at 45o to the optical axis provides 360o polar-coordinate coverage. It is driven at 1 to 2 revolutions per second by a motor fitted with an integral incremental encoder and an optical indexing sensor that signals the completion of each revolution. The system is capable of simultaneous operation as a wideband optical communication receiver [Miller & Wagner, 1987].

4.2.4 Acuity Research AccuRange 3000 Acuity Research, Inc., Menlo Park, CA, has recently introduced an interesting product capable of acquiring unambiguous range data from 0 to 20 m (0 to 66 feet) using a proprietary technique similar to conventional phase-shift measurement (see Table 4.12). The AccuRange 3000 [ACUITY] projects a collimated beam of near-infrared or visible laser light, amplitude modulated with a nonsinusoidal waveform at a 50-percent duty cycle. A 2.5-inch collection aperture surrounding the laser diode emitter on the front face of the cylindrical housing gathers any reflected energy returning from the target, and compares it to the outgoing reference signal to produce a square-wave output with

Page 91

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

a period of oscillation proportional to the measured range. The processing electronics reportedly are substantially different, however, from heterodyne phase-detection systems [Clark, 1994]. The frequency of the output signal varies from approximately 50 MHZ at zero range to 4 MHZ at 20 meters. Distance to target can be determined through use of a frequency-to-voltage converter, or by measuring the period with a hardware or software timer [Clark, 1994]. Separate 0- to 10volt analog outputs are provided for returned signal amplitude, ambient light, and temperature to facilitate dynamic calibration for optimal accuracy in demanding applications. The range output changes within 250 nanoseconds to reflect any change in target distance, and all outputs are updated within a worst-case time frame of only 3 microseconds. This rapid response rate (up to 312.5 KHz for all outputs with the optional SCSI interface) allows the beam to be manipulated at a 1000- to 2000-Hz rate with a mechanical-scanner option. A 45o balanced-mirror arrangement is rotated under servo-control to deflect the coaxial outgoing and incoming beams for full 360o planar coverage.

Parameter Laser output Beam divergence Wavelength Maximum range Minimum range Accuracy Sample rate Response time Diameter Length Weight Power

Value 5 0.5 780/670 20 65 0 2 up to 312.5 3 7.6 3 14 5.5 18 5 and 12 250 and 50

Units milliwatts milliradians nanometers meters feet meters millimeters KHz microseconds cm inches cm inches ounces volts DC milliamps

Table 4.12: Selected specifications for the AccuRange 3000 distance measurement sensor (courtesy Acuity Research, Inc.).

4.2.5 TRC Light Direction and Ranging System Transitions Research Corporation (TRC), Danbury, CT, offers a lowcost lidar system (see Fig. 4.22) for detecting obstacles in the vicinity of a robot and/or estimating position from local landmarks, based on the previously discussed Acuity Research AccuRange 3000 unit. TRC adds a 2-DOF scanning mechanism employing a gold front-surfaced mirror specially mounted on a vertical pan axis that rotates between 200 and 900 rpm. The tilt axis of the scanner is mechanically synchronized to nod one complete cycle (down 45 o and back to horizontal) per 10 horizontal scans, effectively creating a protective spiral of detection coverage around the robot [TRC, 1994] (see

Page 92

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

Fig. 4.23). The tilt axis can be mechanically disabled if so desired for 360o azimuthal scanning at a fixed elevation angle. A 68HC11 microprocessor automatically compensates for variations in ambient lighting and sensor temperature, and reports range, bearing, and elevation data via an Ethernet or RS-232 interface. Power requirements are 500 milliamps at 12 volts DC and 100 milliamps at 5 volts DC. Typical operating parameters are listed in Table 4.13.
Parameter Maximum range Minimum range Laser output Wavelength Beam divergence Modulation frequency Accuracy (range) Resolution (range) (azimuth) Sample rate Scan rate Size (scanner) (electronics) Weight Power Value 12 39 0 6 780 0.5 2 25 1 5 0.2 0.18 25 200-900 131335 5513.7 30265 12102 4.4 12 and 5 500 and 100 Units meters feet meters milliwatts nanometers milliradians MHZ mm inch mm inches degrees KHz rpm cm inches cm inches pounds volts DC milliamps

Table 4.13: Selected specifications for the TRC Light Direction and Ranging System (courtesy Transitions Research Corp.).

Figure 4.23: LightRanger data plotted from scans of a room. An open door at the upper left and a wall in the corridor detected through the open doorway are seen in the image to the left. On the right a trail has been left by a person walking through the room (courtesy Transitions Research Corp.). Page 93

Figure 4.22: The TRC Light Direction and Ranging System incorporates a two-axis scanner to provide full-volume coverage sweeping 360 o in azimuth and 45o in elevation (courtesy Transitions Research Corp. TRC).

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

4.3 Frequency Modulation


A closely related alternative to the amplitude-modulated phase-shift-measurement ranging scheme is frequency-modulated (FM) radar. This technique involves transmission of a continuous electromagnetic wave modulated by a periodic triangular signal that adjusts the carrier frequency above and below the mean frequency f 0 as shown in Fig. 4.21. The transmitter emits a signal that varies in frequency as a linear function of time:
f (t ) = f o + at
f 2d/c

where: a = constant t = elapsed time This signal is reflected from a target and arrives at the receiver at time t + T. T= 2d c

fo

Figure 4.21: The received frequency curve is shifted along the time axis relative to the reference frequency.

where: T = round-trip propagation time d = distance to target c = speed of light The received signal is compared with a reference signal taken directly from the transmitter. The received frequency curve will be displaced along the time axis relative to the reference frequency curve by an amount equal to the time required for wave propagation to the target and back. (There might also be a vertical displacement of the received waveform along the frequency axis, due to the Doppler effect.) These two frequencies when combined in the mixer produce a beat frequency F b:
Fb = f (t ) f ( T + t ) = aT

where: a = constant This beat frequency is measured and used to calculate the distance to the object: d= Fb c 4 Fr Fd

Page 94

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

where: d = range to target c = speed of light Fb = beat frequency Fr = repetition (modulation) frequency Fd = total FM frequency deviation Distance measurement is therefore directly proportional to the difference or beat frequency, and as accurate as the linearity of the frequency variation over the counting interval. Advances in wavelength control of laser diodes now permit this radar ranging technique to be used with lasers. The frequency or wavelength of a laser diode can be shifted by varying its temperature. Consider an example where the wavelength of an 850-nanometer laser diode is shifted by 0.05 nanometers in 4 seconds: the corresponding frequency shift is 5.17 MHZ per nanosecond. This laser beam, when reflected from a surface 1 meter away, would produce a beat frequency of 34.5 MHZ. The linearity of the frequency shift controls the accuracy of the system; a frequency linearity of one part in 1000 yards yields an accuracy of 1 millimeter. The frequency-modulation approach has an advantage over the phase-shift-measurement technique in that a single distance measurement is not ambiguous. (Recall phase-shift systems must perform two or more measurements at different modulation frequencies to be unambiguous.) However, frequency modulation has several disadvantages associated with the required linearity and repeatability of the frequency ramp, as well as the coherence of the laser beam in optical systems. As a consequence, most commercially available FMCW ranging systems are radar-based, while laser devices tend to favor TOF and phase-detection methods. 4.3.1 VRSS Automotive Collision Avoidance Radar One of the first practical IVHS collision avoidance radar systems was developed by Vehicle Radar Safety Systems (VRSS) of Mt. Clemens, MI. This modified Doppler radar unit is intended to alert automobile drivers to potentially dangerous situations. A grill-mounted miniaturized microwave horn antenna sends out a narrow-beam signal that detects only those objects directly in the path of the vehicle, ignoring targets (such as road signs and parked cars) on either side. When the radar signal is reflected from a slower moving or stationary target, it is detected by the antenna and passed to an under-the-hood electronic signal processor [VRSS, 1983]. The signal processor continuously computes the host vehicle speed and acceleration, distance to the target, its relative velocity, and its acceleration. If these parameters collectively require the driver to take any corrective or precautionary action, a warning buzzer and signal light are activated on a special dashboard monitor. An alert signal lights up when an object or slower moving vehicle is detected in the path of the host vehicle. If the target range continues to decrease, and the system determines that a collision is possible, a warning light and buzzer signal the driver to respond

Page 95

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

accordingly. If range continues to decrease with no reduction in relative velocity, then a danger light illuminates indicating the need for immediate action. A filter in the signal processor provides an optimum operating range for the system, based on the relative velocity between the vehicle and the perceived object. The response window corresponds to a calculated difference in speed of between 0.16 and 48 km/h (0.1 and 30 mph) [VRSS, 1983]. If the speed differential exceeds 30 miles per hour, the filter circuit delays signals to the dashboard monitor. This helps to eliminate false signals and signals that might otherwise be caused by approaching vehicles when passing another vehicle on a two-lane highway. Specifications for the VRSS system are given in Table 4.14.
Parameter Effective range Accuracy Update rate Operating frequency RF power Beam width Size (horizontal) (vertical) (antenna) (electronics unit) Weight (total) Power Value 0.3-91 1-300 1.5 200 24 10 Units m feet percent Hz GHz milliwatts

6 degrees 6 degrees 7.615.2 34 1012.75 452 4 12 12 cm inches cm inches pounds volts DC watts

Table 4.14: Selected specifications for the VRSS automotive

The VRSS collision warning system radar (courtesy Vehicle Radar Safety Systems). has been tested in over a million miles of driving conditions in fog, rain, snow, and ice with good results. The present model was perfected in 1983 after 36 years of research, and approved by the FCC in 1985. Although aimed initially at the bus and trucking industries, the low-cost unit offers convincing proof that small, low power radar systems offer a practical alternative to ultrasonic rangefinders for the collision avoidance needs of a mobile robot operating in outdoor scenarios. 4.3.2 VORAD Vehicle Detection and Driver Alert System VORAD (Vehicle Onboard Radar) Safety Systems, Inc., San Diego, CA, has also developed a commercial millimeter-wave FMCW Doppler radar system designed for use onboard a motor vehicle [VORAD]. The Vehicle Collision Warning System employs a 12.712.7 cm (55 inch) antenna/transmitter/receiver package mounted on the front grill of a vehicle to monitor speed of and distance to other traffic or obstacles on the road. The flat etched-array antenna radiates approximately 0.5 milliwatts of power at 24.725 GHz directly down the roadway in a narrow directional beam. A GUNN diode is used for the transmitter, while the receiver employs a balancedmixer detector [Woll, 1993]. The Electronics Control Assembly located in the passenger compartment or cab can individually distinguish up to 20 moving or stationary objects [Siuru, 1994] out to a maximum range of 106 m (350 feet); the closest three targets within a prespecified warning distance are tracked at a 30-Hz
Page 96

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

rate. A Motorola DSP 56001 and an Intel 87C196 microprocessor calculate range and range-rate information from the RF data, and analyze the results in conjunction with vehicle velocity, braking, and steering angle information. If necessary, the Control Display Unit alerts the operator if warranted of potentially hazardous driving situations with a series of caution lights and audible beeps. As an optional feature, the Vehicle Collision Warning System offers blind-spot detection along the right hand side of the vehicle out to 4.5 m (15 feet). The Side Sensor transmitter employs a dielectric resonant oscillator operating in pulsed-Doppler mode at 10.525 GHz, using a flat etched-array antenna with a beam width of about 70o [Woll, 1993]. The system microprocessor in the Electronics Control Assembly analyzes the signal strength and frequency components from the Side Sensor subsystem in conjunction with vehicle speed and steering inputs, and activates audible and visual LED alerts if a dangerous condition is thought to exist. (Selected specifications are listed in Table 4.15).

Parameter Effective range Accuracy Update rate Host platform speed Closing rate Operating frequency RF power Beam width (horizontal) (vertical) Size (antenna) (electronics unit) Weight (total) Power MTBF

Value 0.3-107 1-350 3 30 0.5-120 0.25-100 24.725 0.5 4 5 15203.8 681.5 201512.7 865 6.75 12-24 20 17,000

Units m feet percent Hz miles per hour miles per hour GHz milliwatts degrees degrees cm inches cm inches pounds volts DC watts hours

Among other features of interest is a recording feature, which stores 10 minutes of the most recent historical data on a removable EEPROM memory card for postaccident reconstruction. This data includes steering, braking, and idle time. Greyhound Bus Lines recently completed installation of the VORAD radar on all of its 2,400 buses [Bulkeley, 1993], and subsequently reported a 25-year low accident record [Greyhound, 1994]. The entire system weighs just 6.75 pounds, and operates from 12 or 24 volts DC with a nominal power consumption of 20 watts. An RS-232 digital output is available.

Table 4.15: Selected specifications for the VORAD collision warning system (courtesy VORAD Safety Systems, Inc.).

4.3.3 Safety First Systems Vehicular Obstacle Detection and Warning System Safety First Systems, Ltd. has developed a 10.525-GHz FMCW microwave unit as a blind-spot alert for drivers when backing up or changing lanes. The system uses dual bistatic micro-strip antennae mounted on the rear side panels and rear bumper, and can detect stationary and moving
Page 97

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

objects out to 12 m (40 feet) [Siuru, 1994]. Some coarse range information about reflecting targets is obtained over this interval through use of four discrete range bins. Average power consumption is about 50 microwatts with a 3-percent duty cycle.

4.3.4 Millitech Millimeter Wave Radar Millitech Corporation, Deerfield, MA, has designed a millimeter-wave FMCW sensor system aimed at satisfying the short-distance noncontact ranging needs for robot collision avoidance [Millitech, 1989]. An innovative feature of Millitechs design is the use of a computer-controlled oscillator to generate the basic transmitter waveform. The approach calls for the coherent processing of the return signal, averaged over many successive sweeps. The relatively narrow bandwidth of the sweep serves to achieve unambiguous range resolution at a cost less than that of traditional FMCW sensors. Using a modular design concept, Millitech has developed two baseline systems with the operational range and waveform parameters listed in Table 4.16. Sensor A is typical of a robotics application where the range to the nearest obstruction can be monitored within a 30- by 30-degree field-of-view. Using a bistatic antenna configuration, this configuration is designed to function as a proximity sensor that will respond to the single strongest return from any object within some predetermined range. Sensor B employs a single rotating antenna in a monostatic configuration to scan a fan beam 360 o in azimuth, and can determine range and bearing for multiple targets.
Parameter Range: maximum minimum Radar cross-section Resolution: range azimuth elevation Center frequency Sweep width Sweep time Blank time Number of sweeps Sample time Range bins Baseband signal Sensor A 10 33 10 4 -30 1 0.4 ~30 ~30 95 320 100 100 100 10 1000 2 - 200 Sensor B 100 328 10 4 -40 10 4 2 5 35 120 400 1000 25 10 1000 5 - 500 KHz Units meters feet cm inches dBsm cm inches degrees degrees GHz MHZ microseconds nanoseconds milliseconds

Table 4.16: Selected specifications for Millitech millimeter wave radar prototypes (courtesy Millitech Corp.).

Page 98

Part I: Sensors for Mobile Robot Positioning



Chapter 4: Sensors for Map-based Positioning

Page 99

Part II: Systems and Methods for Mobile Robot Positioning

Tech-Team leaders Chuck Cohen, Frank Koss, Mark Huber, and David Kortenkamp (left to right) fine-tune CARMEL in preparation of the 1992 Mobile Robot Competition in San Jose, CA. The efforts paid off: Despite his age, CARMEL proved to be the most agile among the contestants, winning first price honors for the University of Michigan.

Page 100

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

Chapter 5: Dead-reckoning Navigation

Dead-reckoning is the most widely used navigation method for mobile robot positioning. It is well known that dead-reckoning provides good short-term accuracy, is inexpensive, and allows very high sampling rates. However, the fundamental idea of dead-reckoning is the integration of incremental motion information over time, which leads inevitably to the accumulation of errors. Particularly, the accumulation of orientation errors will cause large position errors which increase proportionally with the distance traveled by the robot. Despite these limitations, most researchers agree that deadreckoning is an important part of a robot navigation system, and that navigation tasks will be simplified if dead-reckoning accuracy can be improved. Dead-reckoning is used in almost all mobile robots, for various reasons: 1. Dead-reckoning data can be fused with absolute position measurements to provide better and more reliable position estimation [Cox, 1991; Hollingum, 1991; Byrne et al., 1992; Chenavier and Crowley, 1992; Evans, 1994; Borenstein and Feng, 1995a;b]. 2. Dead-reckoning can be used in-between absolute position updates with landmarks. Given a required positioning accuracy, increased accuracy in dead-reckoning allows for less frequent absolute position updates. As a result, fewer landmarks are needed for a given travel distance. 3. Many mapping and landmark matching algorithms (for example: [Gonzalez et al., 1992; Chenavier and Crowley, 1992]) assume that the robot can maintain its position well enough to allow the robot to look for landmarks in a limited area, and to match features in that limited area to achieve short processing time and to improve matching correctness [Cox, 1991]. 4. In some cases, dead-reckoning is the only navigation information available. Examples are when no external reference is available, when circumstances preclude the placing or selection of landmarks in the environment, or when another sensor subsystem fails to provide usable data.

Page 102

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

5.1 Systematic and Non-systematic Dead-reckoning Errors


Dead-reckoning is based on simple equations (see Chapter 1) that are easily implemented, and that utilize data from inexpensive incremental wheel encoders. However, dead-reckoning is based on the assumption that wheel revolutions can be translated into linear displacement relative to the floor. This assumption is only of limited validity. One extreme example is wheel slippage: If one wheel was to slip on, say, an oil spill, then the associated encoder would register wheel revolutions even though these revolutions would not correspond to a linear displacement of the wheel. Along with the extreme case of total slippage, there are several other more subtle reasons for inaccuracies in the translation of wheel encoder readings into linear motion. All of these error sources fit into one of two categories: (1) systematic errors and (2) non-systematic errors. 1. Systematic errors a. Unequal wheel diameters b. Average of actual wheel diameters differs from nominal wheel diameter c. Actual wheelbase differs from nominal wheelbase d. Misalignment of wheels e. Finite encoder resolution f. Finite encoder sampling rate 2. Non-systematic errors a. Travel over uneven floors b. Travel over unexpected objects on the floor c. Wheel-slippage due to + slippery floors + over-acceleration + fast turning (skidding) + external forces (interaction with external bodies) + internal forces (castor wheels) + non-point wheel contact with the floor The clear distinction between systematic and non-systematic errors is of great importance for the effective reduction of dead-reckoning errors. For example, systematic errors are particularly grave because they accumulate constantly. On most smooth indoor surfaces systematic errors contribute much more to dead-reckoning errors than non-systematic errors. However, on rough surfaces with significant irregularities, non-systematic errors are dominant. One particularly grave problem with non-systematic errors is that they may appear unexpectedly (for example, when the robot traverses an unexpected object on the ground), and they can cause large position errors. Typically, when a mobile robot system is installed with a hybrid dead-reckoning/landmark navigation system, then the frequency of the landmarks is determined empirically, based on the worst case systematic errors. Such systems are likely to fail when one or more large non-systematic errors occur.

Page 103

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

Uncertainty It is noteworthy that many researchers error elipses Estimated trajectory develop algorithms that estimate the position Start position of robot uncertainty of a dead-reckoning robot (e.g., [Tonouchi et al., 1994; Komoriyah and Oyama, 1994]). With this approach each computed robot position is surrounded by a characteristic "error ellipse," which indicates a region of uncertainty for the robot's actual position (see Figure 5.1: Growing "error ellipses" indicate the Fig. 5.1) [Tonouchi et al., 1994; Adams et al., growing position uncertainty with dead-reckoning (adapted from Tonouchi et al., 1994) 1994]. Typically, these ellipses grow with travel distance, until an absolute position measurement reduces the growing uncertainty and thereby "resets" the size of the error ellipse. These error estimation techniques must rely on error estimation parameters derived from observations of the vehicle's dead-reckoning performance. Clearly, these parameters can take into account only systematic errors, because the magnitude of non-systematic errors is unpredictable.

5.2 Reduction of Dead-reckoning Errors


The accuracy of dead-reckoning in commercial mobile platforms depends to some degree on their kinematic design and on certain critical dimensions. Here are some of the design-specific considerations that affect dead-reckoning accuracy:
+ Vehicles with a small wheelbase are more prone to orientation errors than vehicles with a larger

wheelbase. For example, the differential drive LabMate robot from TRC has a relatively small wheelbase of 340 mm (13.4 inches). According to [Gourley and Trivedi, 1994], the manufacturer recommends that dead-reckoning be limited to about 10 m (33 feet), before a new "reset" becomes necessary.
+ Vehicles with castor wheels that bear a significant portion of the overall weight are likely to

induce slippage when reversing direction (the "shopping cart effect"). Conversely, if the castor wheels bear only a small portion of the overall weight, then slippage will not occur when reversing direction [Borenstein and Koren, 1985].
+ It is widely known that ideally, wheels used for dead-reckoning should be "knife-edge" thin, and

not compressible. The ideal wheel would be made of aluminum with a thin layer of rubber for better traction. In practice, this design is not feasible for all but the most lightweight vehicles, because the dead-reckoning wheels are usually also load-bearing drive wheels, which require a somewhat larger ground contact surface.
+ Typically the synchro-drive design (see Section 1.3.4) provides better dead-reckoning accuracy

than differential-drive vehicles. This is especially true when traveling over floor irregularities: Arbitrary irregularities will affect only one wheel at a time. Thus, since the two other drive wheels stay in contact with the ground, they provide more traction and force the affected wheel to slip.
Page 104

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

Therefore, overall distance traveled will be reflected properly by the amount of travel indicated by dead-reckoning. Other attempts at improving dead-reckoning accuracy are based on more detailed modeling. For example, Larsson et al. [1994] used circular segments to replace the linear segments in each sampling period. The benefits of this approach are relatively small. Boyden and Velinsky [1994] compared (in simulations) conventional dead-reckoning techniques, based on kinematics only, to solutions based on the dynamics of the vehicle. They presented simulation results to show that for both differentially and conventionally steered wheeled mobile robots, the kinematic model was accurate only at slower speeds up to 0.3 m/s when performing a relatively tight turn. This result agrees with experimental observations, which suggest that errors due to wheel slippage can be reduced to some degree by limiting the vehicle's speed during turning, and by limiting accelerations. Specific methods for reducing dead-reckoning errors are described in the following sections. 5.2.1 Auxiliary Wheels and Basic Encoder Trailer It is generally possible to improve deadreckoning accuracy by adding a pair of "knifeedge," non-load-bearing encoder wheels, as shown conceptually in Fig. 5.2. Since these wheels are not used for transmitting power, they can be made to be very thin and with only a thin layer of rubber as a tire. Such a design is feasible for differential drive, tricycle drive, and Ackerman vehicles. Hongo et al. [1987] had built such a set of encoder wheels, to improve the accuracy of a Figure 5.2: Conceptual drawing of a set of encoder large differential drive mobile robot weighing wheels for a differential drive vehicle. 350 kg (770 lb). Hongo et al. report that, after careful calibration, their vehicle had a position error of less than 200 mm (8 inches) for a travel distance of 50 m (164 feet). The ground surface on which this experiment was carried out was a "well paved" road.
trc2 nsf .d s4 , trc2 nsf .w mf, 1 1/2 9/ 93

5.2.2 The Basic Encoder Trailer An alternative approach is the use of a trailer with two encoder wheels. Such an encoder trailer was recently built and tested at the University of Michigan (see Fig. 5.3). This encoder trailer was design to be attached to a Remotec Andros V tracked vehicle. As was explained in Section 1.3, it is virtually impossible to use dead-reckoning with tracked vehicles, because of the large amount of slippage between the tracks and the floor during turning. The idea of the encoder trailer is to perform dead-reckoning whenever the ground characteristics allow it. Then, when the Andros has to move
Page 105

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

over small obstacles, stairs, or otherwise uneven ground, the encoder trailer would be "pulled up." 5.2.3 Mutual Referencing Sugiyama [1993] proposed two robots be used to recognize their positions mutually. When one of the robots moves to another place, the other remains still and observes the motion and determines the first robot's new position. In other words, at any time one robot localizes itself with reference to a fixed object: the standing robot. However, this stop and go approach will limit the efficiency of the robots. 5.2.4 MDOF vehicle with Compliant Linkage Researchers at the University of Michigan designed a unique Multi-Degree-of-Freedom (MDOF) vehicle (see [Borenstein, 1992; 1993; 1994c; 1994d] for various details on this design). This vehicle was designed specifically to overcome the problems of excessive wheelslippage reported for alternative MDOF designs as discussed in Section 1.3.6 [Reister, 1991; Kilough & Pin, 1992; Pin and Kilough, 1994]. The University of Michigan's MDOF vehicle, shown in Fig. 5.4, comprises two differentialdrive LabMate robots from [TRC]. The two LabMates, here referred to as "trucks," are connected by a compliant linkage and two rotary joints, for a total of three internal degrees of freedom. The purpose of the compliant linkage is to accommodate momentary controller errors without transferring any mutual force reactions between the trucks, thereby eliminating the excessive wheel slippage reported for other MDOF vehicles. Because it eliminates excessive wheel slippage, the MDOF vehicle with

Figure 5.3: A simple encoder trailer. The trailer here was designed and built at the University of Michigan for use with the Remotec Andros tracked vehicle. (In the photograph here, the trailer is attached to a TRC LabMate for initial testing.)

Figure 5.4: The compliant linkage is instrumented with two absolute rotary encoders and a linear encoder to measure the relative orientations and separation distance between the two trucks.

Page 106

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

compliant linkage is 1 - 2 orders of magnitude more accurate than other MDOF vehicles, and as accurate as conventional, 2-DOF vehicles. The MDOF vehicle with compliant linkage is shown in Fig. 5.5.

Figure 5.5: The University of Michigan's MDOF vehicle is a dual-differential-drive multiple-degree-of-freedom platform comprising two TRC LabMates. These two "trucks" are coupled together with a compliant linkage, designed to accommodate momentary controller errors that would cause excessive wheel slippage in other MDOF vehicles (courtesy University of Michigan).

5.2.5 Internal Position Error Correction A unique way for reducing dead-reckoning errors even further is Internal Position Error Correction (IPEC). With this approach two mobile robots mutually correct their dead-reckoning errors. However, unlike the approach described in Section 5.2.3 above, the IPEC method works while both robots are in continuous, fast motion [Borenstein, 1994a], but the method requires that both trucks can measure their relative distance and bearing continuously and accurately. Coincidentally, the MDOF vehicle with compliant linkage (described in Section 5.2.4 above) offers exactly these features, and the IPEC method was therefore implemented and demonstrated on that MDOF vehicle. This implementation is named Compliant Linkage Autonomous Platform with Position Error Recovery (CLAPPER). The CLAPPER's compliant linkage instrumentation is illustrated in Fig. 5.4. This setup provides real-time feedback on the relative position and orientation of the two trucks. An absolute encoder
Page 107

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

at each end measures the rotation of each truck (with respect to the linkage) with a resolution of 0.3 o, while a linear encoder is used to measure the separation distance to within 5 mm (0.2 inches). Each differentially-steered truck computes its own dead-reckoned position and heading in conventional fashion, based on displacement and velocity information derived from its left and right drive wheel encoders. By examining the perceived dead-reckoning solutions of the two robot platforms in conjunction with their known relative orientations, the CLAPPER system can detect and significantly reduce heading errors for both trucks (see video clip in Borenstein, 1995V). The principle of operation is centered on the concept of error growth rate presented by Borenstein (1994a), who makes a distinction between fast-growing and slow-growing dead-reckoning errors. For example, when a differentially-steered robot traverses a floor discontinuity it will immediately experience an appreciable orientation error (i.e., a fast-growing error). The associated lateral displacement error, however, is initially very small (i.e., a slow-growing error), but grows in an unbounded fashion as a consequence of the orientation error. The internal error correction algorithm performs relative position measurements with a sufficiently fast update rate (20 ms) to allow each LabMate to detect fast-growing errors in orientation, while relying on the fact that the lateral position errors accrued by both platforms during the sampling interval were very small. Figure 5.6 explains how this method works. After traversing a bump Truck A's orientation will change (a fact unknown to Truck A's deadreckoning computation). Truck A is therefore expecting to "see" Truck B along the extension of line Le. However, because of the physically incurred rotation of Truck A, the absolute encoder on truck A will report that truck B is now actually seen along line Lm. The angular difference between Le and Lm is the thus measured dead-reckoning orientation error of Truck A, which can be corrected immediately.

Figure 5.6: After traversing a bump, the resulting change of orientation of Truck A can be measured relative to Truck B.

Page 108

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

One should note that even if Truck B encountered a bump at the same time, the resulting rotation of Truck B would not affect the orientation error measurement. The compliant linkage in essence forms a pseudo-stable heading reference in world coordinates, its own orientation being dictated solely by the relative translations of its end points, which in turn are affected only by the lateral displacements of the two trucks. Since the lateral displacements are slow growing, the linkage rotates only a very small amount between encoder samples. The fastgrowing azimuthal disturbances of the bases, on the other hand, are not coupled up through the rotational joints to the linkage, thus allowing the rotary encoders to detect and quantify the instantaneous orientation errors of the bases, even when both are in motion. Borenstein (1994a) provides a more complete description of this innovative concept, and reports experimental results indicating improved dead-reckoning performance of up to two orders of magnitude over conventional mobile robots.

5.3 Automatic Vehicle Calibration


Another approach to improving dead-reckoning accuracy without any additional devices or sensors is based on the careful calibration of the vehicle. As was explained in Section 5.1 above, systematic errors are inherent properties of each individual robot. They change very slowly as the result of wear or of different load distributions. Thus, these errors remain almost constant over extended periods of time [Tsumura et al., 1981]. One way to reduce such errors is vehicle-specific calibration. However, calibration is difficult because even minute deviations in the geometry of the vehicle or its parts (e.g., a change in wheel diameter due to a load) may cause substantial deadreckoning errors. Such calibrations are often performed in a non-systematic, trial and error fashion. For example, the vehicle is run through a certain path and the resulting errors at the end are recorded. Then, the experimenter will change a parameter, say, the effective wheelbase of the vehicle, and observe how the resulting error changes. This method is time-consuming and yields non-optimal results. Borenstein and Feng [1995] have developed a systematic procedure for the measurement and correction of dead-reckoning errors. This work is based on a simplified error model, in which there are only two dominant systematic errors: The error due two unequal wheel diameters, denoted E d, and the error due to the uncertainty in the effective wheelbase, denoted E b. The error correction procedure requires that the vehicle be run 5 times in a clockwise (cw) square path, and 5 times in a counter-clockwise (ccw) square path. The resulting errors are recorded and inserted into a set of equations detailed in [Borenstein & Feng, 1995]. Based on these equations two correction factors, one for error E d and one for error E b can be computed. These correction factors are then inserted into the vehicle's dead-reckoning algorithm. The result is a 10 to 20-fold reduction in systematic errors. Fig. 5.7 shows the result of a typical calibration session. DR and DL are the

Page 109

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

effective wheel diameters, and b is the effective wheelbase.


100

Y [mm]

This calibration procedure can be performed with nothing more than an ordinary tape measure. It takes about 30 minutes to run the complete calibration procedure and measure the individual return errors with a tape measure. However, one can also use a set of ultrasonic sensors, controlled by the onboard computer. With this setup, it is possible to perform the complete calibration procedure automatically.

50 X [mm] -50 -50 -100 -150 -200


Before correction, cw Before correction, ccw After correction, cw After correction, ccw

50

100

150

200

250

5.4 Inertial Navigation


-250 An alternative method for en- Figure 5.7: Position Errors after completion of the Bi-directional hancing dead-reckoning is inertial Square Path Experiment (4 x 4 m). navigation, initially developed for Before calibration b=340.00 mm, DR/DL = 1.00000 deployment on aircraft. The tech- After calibration: b=336.17, DR/DL = 1.00084 nology was quickly adapted for use on missiles and in outer space, and found its way to maritime usage when the nuclear submarines Nautilus and Skate were suitably equipped in support of their transpolar voyages in 1958 [Dunlap & Shufeldt, 1972]. The principle of operation involves continuous sensing of minute accelerations in each of the three directional axes, and integrating over time to derive velocity and position. A gyroscopically stabilized sensor platform is used to maintain consistent orientation of the three accelerometers throughout this process.
DEADRE81.DS4, DEADRE81.W MF, 10/08/94

Although fairly simple in concept, the specifics of implementation are rather demanding. This is mainly caused by error sources that adversely affect the stability of the gyros used to ensure correct attitude. The resulting high manufacturing and maintenance costs have effectively precluded any practical application of this technology in the automated guided vehicle industry [Turpin, 1986]. For example, a high quality inertial navigation system (INS) such as would be found in a commercial airliner will have a typical drift of about 1 nautical mile per hour of operation, and cost between $50K and $70K [Byrne et al., 1992]. High-end INS packages used in ground applications have shown performance of better than 0.1 percent of distance traveled, but cost in the neighborhood of $100K to $200K, while lower performance versions (i.e., 1 percent of distance traveled) run between $20K to $50K [Dahlin & Krantz, 1988].

Page 110

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

Experimental results by the Universite Montpellier in France [Vaganay et al., 1993a; 1993b], Barshan and Durrant-Whyte [1993], and the University of Michigan [Borenstein, 1994a] indicate the inertial navigation approach is not realistically advantageous for the above reasons. As a consequence, the use of INS hardware in robotics applications has to date been generally limited to scenarios that arent readily addressable by more practical alternatives. An example of such a situation is presented by Sammarco [1990; 1994], who reports preliminary results in the case of an INS used to control an autonomous vehicle in a mining application. Inertial navigation is attractive mainly because it is self-contained and no external motion information is needed for positioning. One important advantage of inertial navigation is its ability to provide fast, low latency dynamic measurements. Furthermore, inertial navigation sensors typically have independent noise and error sources from the external sensors [Parish and Grabbe, 1993]. For example, the noise and error from an inertial navigation system should be quite different from that of, say, a landmark based system. Inertial navigation sensors are self-contained, non-radiating, and non-jammable. Fundamentally, gyros provide angular rate and accelerometers provide velocity rate information. Dynamic information is provided through direct measurements. However, the main disadvantage is that the angular rate data and the linear velocity rate data must be integrated once and twice (respectively), to provide orientation and linear position, respectively. Thus, even very small errors in the rate information can cause an unbounded growth in the error of integrated measurements. In recent years the prices for very accurate laser gyros and optical fiber gyros have come down significantly. With prices of $1,000 to $5,000, these devices are suitable for many mobile robot applications.

5.4.1 Accelerometers
The suitability of accelerometers for mobile robot positioning was evaluated at the University of Michigan. n this unpublished study it was found that there is a very poor signal-to-noise ratio at lower accelerations (i.e., during low-speed turns). Accelerometers also suffer from extensive drift, and they are sensitive to uneven grounds, because any disturbance from a perfectly horizontal position will cause the sensor to detect the gravitational acceleration g. One low-cost inertial navigation system was aimed at overcoming the latter problem included a tilt sensor [Barshan & Durrant-Whyte, 1993, 1994]. The tilt information provided by the tilt sensor was supplied to the accelerometer to cancel the gravity component projecting on each axis of the accelerometer. Nonetheless, the results obtained from the tilt-compensated system indicate a position drift rate of 10-80 mm/sec, depending on the frequency of acceleration changes. This is an unacceptable error rate for most mobile robot applications.

5.4.2 Gyros
Gyros have long been used in robots to augment the sometimes erroneous dead-reckoning information of mobile robots. One such sensor is the Gyrostar, used in [Barshan and Durrant-Whyte, 1993, 1994]. Barshan and Durrant-White show that an Extended Kalman Filter (EKF) can improve

Page 111

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 5: Reduction of Dead-reckoning Errors

the accuracy of the Gyrostar by a factor of 5. However, even then a drift rate of 1-3 o/minute can be expected. As explained in Chapter 2, there is great promise in the performance of Laser or Optical Fiber Gyros. Especially as prices for these devices are now falling into the $1K - $5K price-range. In one study with such relatively inexpensive gyros, Komoriya and Oyama [1994] used an optical fiber gyro called OFG-3. The researchers measured a drift rate of 0.19o/minute (0.003o /sec or 11.4o/hr) a level that is substantially better than that provided by conventional devices. However, in evaluating the results from Komoriya and Oyama [1994] one finds an interesting point: The OFG-3 was quoted to have a Minimum Detectable Rotation Rate of 0.05o/sec. Note that if this gyro was installed on a robot with a systematic error (e.g., due to unequal wheel diameters) of 1 o per 10 m linear travel, then the OFG-3 would not detect this systematic error at speeds lower than 0.5 m/sec. Some commercially available gyros include 1. The "AUTOGYRO"Andrew Corp. [ANDREW] offers this rotation sensor for terrestrial navigation. It is a single-axis interferometric fiber optic gyroscope, based on polarizationo maintaining fiber and precision fiber-optic gyroscope technology. It drifts about 0.005/sec. (18o/hr) and the minimum detectable rotation rate is 0.02o /sec. 2. The GYRATIONThis open-loop directional GyroEngine is offered by Gyration, Inc. It has a typical drift (with software compensation) of about 1 o/min or 60o/hr. Software compensation was shown to reduce the location error by as much as a factor of six (in a 6.6 miles and 12 minutes road test).

5.5 Summary
+ Dead-reckoning is a central part of almost all mobile robot navigation systems. + Improvements in dead-reckoning techniques will not change their incremental nature, i.e., even

for improved dead-reckoning, periodic absolute position updates are necessary.


+ More accurate dead-reckoning will reduce the requirements on absolute position updates and will

facilitate the solution of landmark and map-based positioning.


+ Inertial navigation systems alone are generally inadequate for periods of time that exceed a few

minutes. However, inertial navigation can provide accurate short-term information, for example orientation changes during a robot maneuver. Software compensation, usually by means of a Kalman Filter, can significantly improve heading measurement accuracy.

Page 112

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

Chapter 6: Active Beacon Navigation Systems

Active beacon navigation systems are the most common navigation aids on ships and airplanes. Active beacons can be detected reliably and provide very accurate positioning information with minimal processing. As a result, this approach allows high sampling rates and yields high reliability, but it does also incur high cost in installation and maintenance. Accurate mounting of beacons is required for accurate positioning. For example, land surveyors' instruments are frequently used to install beacons in a high-accuracy application [Maddox, 1994]. Kleeman [1992] notes that: "Although special beacons are at odds with notions of complete robot autonomy in an unstructured environment, they offer advantages of accuracy, simplicity, and speed - factors of interest in industrial and office applications, where the environment can be partially structured." One can distinguish between two different types of active beacon systems: Trilateration and Triangulation. 1. Trilateration Trilateration is the determination of a vehicle's position based on distance measurements to known beacon sources. In trilateration navigation systems there are usually three or more transmitters mounted at known locations in the environment and one receiver onboard the robot. Conversely, there may be one transmitter on board and the receivers are mounted on the walls. Using time-of-flight information, the system computes the distance between the stationary transmitters and the onboard receiver. Global Positioning Systems (GPS), discussed in Section 3.1, are an example for trilateration. Beacon systems based on ultrasonic sensors (see Section 6.2, below) are another example. 2. Triangulation In this configuration there are three or more active transmitters (usually infra-red) mounted at known locations in the environment, as shown in Fig. 6.1. A rotating sensor onboard the robot registers the angles 8 1, 8 2, and 8 3 at which it "sees" the transmitter beacons relative to the vehicle's longitudinal axis. From these three measurements the unknown x and y coordinates and the unknown
Page 113

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

vehicle orientation 2 can be computed. Simple Y navigation systems of this kind can be built 1 very inexpensively [Borenstein and Koren, 2 1986]. One problem with this configuration is that the active beacons need to be extremely 2 1 powerful to insure omnidirectional transmission over large distances. Since such powerful beacons are not very practical it is necessary to o focus the beacon within a cone-shaped propagation pattern. As a result, beacons are not 3 3 visible in many areas, a problem that is particularly grave because at least three beacons must X be visible for triangulation. A commercially 0 available sensor system based on this configura- Figure 6.1: The basic triangulation problem: a rotating tion (manufactured and marketed by Denning) sensor head measures the three angles &1, &2, and &3 the vehicle's was tested at the University of Michigan in betweenS , S , and S longitudinal axes and the three sources 1 2 3 1990. The system provided an accuracy of approximately 5 cm, but the aforementioned limits on the area of application made the system unsuitable for precise navigation in large open areas.

course9 .ds4; .wmf 19/17/94

Rotating transmitter/receiver, stationary reflectors In this implementation of triangulation, there is one rotating laser beam onboard the vehicle and three or more stationary retro-reflectors are mounted at known locations in the environment. Rotating transmitter, stationary receivers Here the transmitter, usually a rotating laser beam, is used onboard the vehicle. Three or more stationary receivers are mounted on the walls. The receivers register the incident beam, which may also carry the encoded azimuth of the transmitter. For either one of the above methods, we will refer to the stationary devices as "beacons," even though they may physically be receivers, retro-reflectors, or transponders.

6.1 Discussion on Triangulation Methods


Most of the active beacon positioning systems discussed in Section 6.3 below include computers capable of computing the vehicle's position. The algorithms used for this computation are usually proprietary because the solutions are non-trivial. In this section we discuss some aspects of triangulation algorithms. In general, it can be shown that triangulation is sensitive to small angular errors when either the observed angles are small, or when the observation point is on or near a circle which contains the
Page 114

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

three beacons. Assuming reasonable angular measurement tolerances, it was found that accurate navigation is possible throughout a large area, although error sensitivity is a function of the point of observation and the beacon arrangements [McGillem and Rappaport, 1988]. Cohen and Koss [1992] performed a detailed analysis on three-point triangulation algorithms and ran computer simulations to verify the performance of different algorithms. The results are summarized as follows:
+ The geometric triangulation method works consistently only when the robot is within the triangle

formed by the three beacons. There are areas outside the beacon triangle where the geometric approach works, but these areas are difficult to determine and are highly dependent on how the angles are defined.
+ The Geometric Circle Intersection method has large errors when the three beacons and the robot

all lie on, or close to, the same circle.


+ The Newton-Raphson method fails when the initial guess of the robot' position and orientation is

beyond a certain bound.


+ The heading of at least two of the beacons was required to be greater than 90o . The angular

separation between any pair of beacons was required to be greater than 45 o. In summary, it appears that none of the above methods alone is always suitable, but an intelligent combination of two or more methods helps overcome the individual weaknesses. Yet another variation of the triangulation method is the so-called running fix, proposed by Case [1986]. The underlying principle of the running fix is that an angle or range obtained from a beacon at time t-1 can be utilized at time t, as long as the cumulative movement vector recorded since the reading was obtained is added to the position vector of the beacon, thus creating a virtual beacon.

6.2 Ultrasonic Transponder Trilateration


Ultrasonic trilateration schemes offer a medium- to high-accuracy, low-cost solution to the position location problem for mobile robots. Because of the relatively short range of ultrasound, these systems are suitable for operation in relatively small work areas, and only if no significant obstructions are present to interfere with wave propagation. The advantages of a system of this type fall off rapidly, however, in large multi-room facilities due to the significant complexity associated with installing multiple networked beacons throughout the operating area. Two general implementations exist: 1) a single transducer transmitting from the robot, with multiple fixed-location receivers, and, 2) a single receiver listening on the robot, with multiple fixed transmitters serving as beacons. The first of these categories is probably better suited to applications involving only one or at most a very small number of robots, whereas the latter case is basically

Page 115

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

unaffected by the number of passive receiver platforms involved (i.e., somewhat analogous to the Navstar GPS concept). 6.2.1 IS Robotics 2-D Location System IS Robotics, Inc. [ISR], Somerville, MA, a spin-off company from MIT's renown Mobile Robotics Lab, has introduced a beacon system based on an inexpensive ultrasonic trilateration system. This system allows their Genghis series robots to localize position to within 12.7 mm (0.5 inches) over a 9.19.1 m (3030-foot) operating area [ISR, 1994]. The ISR system consists of a base station master hard-wired to two slave ultrasonic pingers positioned a known distance apart (typically 90 inches) along the edge of the operating area as shown in Fig. 6.2. Each robot is equipped with a receiving ultrasonic transducer situated beneath a cone-shaped reflector for omni-directional coverage. Communication between the base station and individual robots is accomplished using a Proxim spread-spectrum (902-928 MHZ) RF link. The base station alternately fires Base Station the two 40-KHz ultrasonic pingers every half second, each time transmitting a two-byte radio packet in "A" "B" broadcast mode to advise all robots Pinger Pinger of pulse emission. Elapsed time between radio packet reception and detection of the ultrasonic wave front is used to calculate distance Pinger Side View between the robots current position and the known location of the active beacon. Inter-robot communication is accomplished over the Figure 6.2: The ISR Genghis series of legged robots localize X-Y same spread-spectrum channel us- position with a master-slave trilateration scheme using two 40 KHz ing a time-division-multiple-access ultrasonic pingers (adapted from [ISR, 1994]). scheme controlled by the base station. Principle sources of error include variations in the speed of sound, the finite size of the ultrasonic transducers, non-repetitive propagation delays in the electronics, and ambiguities associated with time-of-arrival detection. The cost for this system is $10,000.

6.2.2 Tulane University 3-D Location System Researchers at Tulane University in New Orleans, LA, have come up with some interesting methods for significantly improving the time-of-arrival measurement accuracy for ultrasonic transmitter-receiver configurations, as well as compensating for the varying effects of temperature and humidity. In the hybrid scheme illustrated in Fig. 6.3, envelope peak detection is employed to establish the approximate time of signal arrival, and consequently eliminate ambiguity interval
Page 116

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

problems for a more precise phase-measurement technique that provides final resolution [Figueroa & Lamancusa, 1992]. The desired 0.04 micrometer (0.001-inch) range accuracy required a time unit discrimination of 75 nanoseconds at the receiver, which can easily be achieved using fairly simplistic phase measurement circuitry, but only within the interval of a single wavelength. The actual distance from transmitter to receiver is the summation of some integer number of wavelengths (determined by the coarse time-of-arrival measurement), plus that fractional portion of a wavelength represented by the phase measurement results.

Digital I/O In PC 40 kHz reference Phase Difference

Phase Detection TTL of Received Waveform


From Receiver

Amplified Waveform

Envelope of Squared Wave After Differentiation

Rough TOF End of RTOF

Figure 6.3: A combination of threshold adjusting and phase detection is employed to provide higher accuracy in time-of-arrival measurements in the Tulane University ultrasonic position location system [Figueroa & Lamancusa, 1992].

Details of this time-of-arrival detection scheme and associated error sources are presented by Figueroa and Lamancusa [1992]. Range measurement accuracy of the prototype system was experimentally determined to be 0.15 mm (0.006 inches) using both threshold adjustments (based on peak detection) and phase correction, as compared to 0.53 mm (0.021 inches) for threshold adjustment alone. The high-accuracy requirements were necessary for an application that involved tracking the end-effector of a 6-DOF industrial robot [Figueroa, et al, 1992]. The system incorporates seven 90-degree Massa piezo-electric transducers operating at 40-KHz, interfaced to a 33-MHZ IBM-compatible PC. The general position-location strategy was based on a trilateration method developed by Figueroa and Mahajan [1994]. The set of equations describing time-of-flight measurements for an ultrasonic pulse propagating from a mobile transmitter located at point (u, v, w) to various receivers fixed in the inertial reference frame can be listed in matrix form as follows [Figueroa & Mahajan, 1994]:

Page 117

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

( t1 td )2 2 ( t2 td ) * * * ( tn td )2

1 r12 2 1 r2 * * * 1 rn2

2 x1 2 x2

2 y1 2 y2

2 xn

2 yn

2 z1 2 z2 2 zn

p2 2 c 1 c2 u 2 c v c 2 w c2

where: ti = measured time of flight for transmitted pulse to reach ith receiver td = system throughput delay constant ri2 = sum of squares of ith receiver coordinates (xi, yi, zi) = location coordinates of ith receiver (u, v, w) = location coordinates of mobile transmitter c = speed of sound p2 = sum of squares of transmitter coordinates The above equation can be solved for the vector on the right to yield an estimated solution for the speed of sound c, transmitter coordinates (u, v, w), and an independent term p 2 that can be compared to the sum of the squares of the transmitter coordinates as a checksum indicator [Figueroa & Mahajan, 1994]. An important feature of this representation is the use of an additional receiver (and associated equation) to enable treatment of the speed of sound itself as an unknown, thus ensuring continuous on-the-fly recalibration to account for temperature and humidity effects. (The system throughput delay constant td can also be automatically determined from a pair of equations for 1/c2 using two known transmitter positions. This procedure yields two equations with td and c as unknowns, assuming c remains constant during the procedure). A minimum of five receivers is required for an unambiguous 3-D position solution, but more can be employed to achieve higher accuracy using a least squares estimation approach. Care must be taken in the placement of receivers to avoid singularities as defined by Mahajan [1992]. Figueroa and Mahajan [1994] report a follow-up version intended for mobile robot positioning that achieves 0.25 mm (0.01-inch) accuracy with an update rate of 100 Hz. The prototype system tracks a TRC LabMate over a 2.743.66 m (9-by 12-foot) operating area with five ceiling-mounted receivers, and can be extended to larger floor plans with the addition of more receiver sets. An RF link will be used to provide timing information to the receivers and to transmit the subsequent X-Y position solution back to the robot. Three problem areas are being further investigated to increase the effective coverage and improve resolution:
Page 118

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

+ Actual transmission range does not match the advertised operating range for the ultrasonic

transducers, probably due to a resonant frequency mismatch between the transducers and electronic circuitry. + The resolution of the clocks (6-MHz) used to measure time-of-flight is insufficient for automatic compensation for variations in the speed of sound. + The phase-detection range-measurement correction sometimes fails when there is more than one wavelength of uncertainty. This problem can likely be solved using the frequency division scheme described by Figueroa and Barbieri [1991].

6.3 Optical Positioning Systems


Optical positioning systems typically involve some type of scanning mechanism operating in conjunction with fixed-location references strategically placed at pre-defined locations within the operating environment. A number of variations on this theme are seen in practice: + Scanning detectors with fixed active-beacon emitters + Scanning emitter/detectors with passive retro-reflective targets + Scanning emitter/detectors with active transponder targets + Rotating emitters with fixed detector targets. One of the principal problems associated with optical beacon systems, aside from the obvious requirement to modify the environment, is the need to preserve a clear line-of-sight between the robot and the beacon. Preserving an unobstructed view is sometimes difficult if not impossible in certain applications such as congested warehouse environments. In the case of passive retroreflective targets, problems can sometimes arise from unwanted returns from other reflective surfaces in the surrounding environment, but a number of techniques exists for minimizing such interference. 6.3.1 Cybermotion Docking Beacon The automated docking system used on the Cybermotion Navmaster robot incorporates the unique combination of a structured-light beacon (to establish bearing) along with a one-way ultrasonic ranging system (to determine standoff distance). The optical portion consists of a pair of near-infrared transceiver units, one mounted on the front of the robot and the other situated in a known position and orientation within the operating environment. These two optical transceivers are capable of full-duplex data transfer between the robot and the dock at a rate of 9600 bits per second. Separate modulation frequencies of 154 and 205 KHz are employed for the uplink and downlink respectively to eliminate cross talk. Under normal circumstances, the dock-mounted transceiver waits passively until interrogated by an active transmission from the robot. If the interrogation is specifically addressed to the assigned ID number for that particular dock, the dock control computer activates the beacon transmitter for 20 seconds. (Dock ID's are jumper selectable at time of installation).

Page 119

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

Figure 6.4 shows the fixed-location beacon illuminating a 90o field-of-regard broken up into two uniquely identified zones, designated for purposes of illustration here as the Left Zone and Right Zone. An array of LED emitters in the beacon head is divided by a double-sided mirror arranged along the optical axis and a pair of lenses. Positive zone identification is initiated upon request from the robot in the form of a NAV Interrogation byte transmitted over the optical data link. LED's on opposite sides of the mirror respond to this NAV Interrogation with slightly different coded responses. The robot can thus determine its relative location with respect to the optical axis of the beacon based on the response bit pattern detected by the onboard receiver circuitry. Once the beacon starts emitting, the robot turns in the appropriate direction and executes the steepest possible (i.e., without losing sight of the beacon) intercept angle with the beacon optical axis. Crossing the optical axis at point B is flagged by a sudden change in the bit pattern of the NAV Response Byte, whereupon the robot turns inward to face the dock. The beacon optical axis establishes the nominal path of approach, and in conjunction with range offset information uniquely defines the robots absolute location. This situation is somewhat analogous to a TACAN station [Dodington, 1989] but with a single defined radial. The offset distance from vehicle to dock is determined in rather elegant fashion by a dedicated non-reflective ultrasonic ranging configuration. This high-frequency (>200 KHZ) narrow-beam (15o ) sonar system consists of a piezo-electric transmitter mounted on the docking beacon head and a complimentary receiving transducer mounted on the front of the vehicle. A ranging operation is initiated upon receipt of the NAV Interrogation Byte from the robot; the answering NAV Response Byte from the docking beacon signals the simultaneous transmission of an ultrasonic pulse. The difference at the robot end between time of arrival for the NAV Response Byte over the optical link and subsequent ultrasonic pulse detection is used to calculate separation distance. This dualtransducer master/slave technique assures an unambiguous range determination between two well defined points and is unaffected by any projections on or around the docking beacon and/or face of the robot.

Page 120

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

Sonar Receiver

Right Zone
Sonar Transmitter Docking Beacon Controller Optical Beacon Head
B

Optical Axis

Beacon Sensor

Left Zone

Figure 6.4: The structured-light near-infrared beacon on the Cybermotion battery recharging station defines an optimal path of approach for the K2A Navmaster robot [Everett, 1995].

During transmission of a NAV Interrogation Byte, the left and right sides of the LED array located on the robot are also driven with uniquely identifiable bit patterns. This feature allows the docking beacon computer to determine the robots actual heading with respect to the nominal path of approach. Recall the docking beacons structured bit pattern establishes (in similar fashion) the side of the vehicle centerline on which the docking beacon is located. This heading information is subsequently encoded into the NAV Response Byte and passed to the robot to facilitate course correction. The robot closes on the beacon, halting at the defined stop range (not to exceed 8 feet) as repeatedly measured by the docking sonar. Special instructions in the path program can then be used to reset vehicle heading and/or position. 6.3.2 Hilare Early work incorporating passive beacon tracking at the Laboratoire dAutomatique et dAnalyse des Systemes, Toulouse, France, involved the development of a navigation subsystem for the mobile robot Hilare [Banzil et al., 1981]. The system consisted of two near-infrared emitter/detectors mounted with a 25 cm (10 inches) vertical separation on a rotating mast, used in conjunction with passive reflective beacon arrays at known locations in three corners of the room.
Page 121

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

Figure 6.5: Retro-reflective beacon array configuration used on the mobile robot Hilare (adapted from [Banzil, et al, 1981]].

Each of these beacon arrays was constructed of retroreflective tape applied to three vertical cylinders, spaced in a recognizable configuration as shown in Fig. 6.5. One of the arrays was inverted so as to be uniquely distinguishable for purposes of establishing an origin. The cylinders were vertically spaced to intersect the two planes of light generated by the rotating optical axes of the two emitters on the Figure 6.6: A confirmed reflection pattern as depicted robots mast. A detected reflection pattern as in above was required to eliminate potential interference Fig. 6.6 confirmed beacon acquisition. Angular from other highly specular surfaces [Banzil et al., 1981]. orientation relative to each of the retro-reflective arrays was inferred from the stepper motor commands that drove the scanning mechanism; lateral position was determined through simple triangulation.

6.3.3 NAMCO LASERNET The NAMCO LASERNET beacon tracking system employs retro-reflective targets distributed throughout the operating area in order to measure range and angular position. A servo-controlled rotating mirror pans a near-infrared laser beam through a horizontal arc of 90 degrees at a 20-Hz update rate. When the beam sweeps across a target of known dimensions, a return signal of finite duration is sensed by the detector. Since the targets are all the same size, the signal generated by a close target will be of longer duration than from a distant one (see Fig. 6.7).
Page 122

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

Angle measurement is initiated when the scanner begins its sweep from right to left; the laser strikes an internal synchronization photodetector that starts a timing sequence. The beam is then panned across the scene until returned by a retro-reflective target in the field-of-view. The reflected signal is detected by the sensor, terminating the timing sequence (Figure 6.7). The elapsed time is used to calculate the angular position of the target in the equation [NAMCO, 1989]
= V Tb 45
+45 0 -45 +45 0 -45

where = target angle V = scan velocity (7200 deg/sec) Tb = interval between scan initiation and target detection.

Figure 6.7: (A) Perceived width of a retro-reflective target of known size is used to calculate range; (B) while the elapsed time between sweep initiation and leading edge detection yields target bearing (courtesy NAMCO Controls).

This angle calculation determines either the leading edge of the target, the trailing edge of the target, or the center of the target, depending upon the option selected within the LASERNET software option list. The angular accuracy is 1 percent, and the angular resolution is 0.1o for the analog output; accuracy is within .05 percent with a resolution of 0.006o when the RS-232 serial port is used. The analog output is a voltage ranging from 0 to 10 volts over the range of from -45 to +45o degrees, whereas the RS-232 serial port reports a proportional count value from 0 to 15360 over this same range. The system costs only $3,400 in its basic configuration, but it has only ,a limited range of 15 m (50 feet).

6.3.4 Intelligent Solutions EZNav Position Sensor Intelligent Solutions, Inc. (ISI1 ), Marblehead, MA, offers a laser-based scanning beacon system that computes vehicle position and heading out to 183 m (600 feet) using cooperative electronic transponders. A range of 30.5 m (100 feet) is achieved with passive reflectors. The EZNav Position Sensor, shown in Fig. 6.8, is a non-ranging triangulation system with an absolute bearing accuracy of 0.03o at a scan rate of 600 rpm [ISI, 1994]. The fan-shaped beam is spread 4o vertically to ensure target detection at long range while traversing irregular floor surfaces, with horizontal divergence limited to 0.017o. Each target is uniquely coded, and up to 32 separate targets can be processed in a single scan, with vehicle X-Y position calculated every 100 milliseconds [Maddox, 1994]. The sensor package weighs 10 pounds, measures 38 cm (15 inches) high and 30 cm (12 inches) in diameter, and has a power consumption of only 300 milliamps at 12 volts. The eye-safe near-infrared

This sensor was originally developed and marketed by Denning Mobile Robotics, Inc. Page 123

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

laser generates a 1 milliwatt output at a wavelength of 810 nanometers. One potential source of problems with this device is the relatively small vertical divergence of the beam: 2o. Another problem mentioned by the developer [Maddox, 1994] is that "the LaserNav sensor ... is subject to rare spikes of wrong data." This undesirable phenomenon is likely due to reflections off shiny surfaces other than the passive reflectors. This problem affects Figure 6.8: The EZNav (formerly LaserNav, made by probably all light-based beacon navigation Denning) is a laser-based scanning beacon system. systems to some degree. Another source of erroneous beacon readings is bright sun-light entering the workspace through wall openings. 6.3.5 TRC Beacon Navigation System Transitions Research Corporation [TRC], Danbury, CT, has incorporated their LEDbased LightRanger, discussed in Section 4.2, into a compact, low-cost navigational referencing system for open-area autonomous platform control. The TRC X Beacon Navigation System calculates vehicle position and heading at ranges up to 24.4 m (80 feet) within a quadrilateral area defined by four passive retro-reflective Y beacons [TRC, 1994] (see Fig. 6.9). A static 15-second unobstructed view of all four beacons is required for initial acquisition Figure 6.9: The TRC Beacon Navigation System and set up, after which only two beacons calculates position and heading based on ranges and must remain in view as the robot moves bearings to two of four passive beacons defining a quadrilateral operating area (courtesy TRC). about. No provision currently exists to periodically acquire new beacons along a continuous route, so operation is currently constrained to a single zone roughly the size of a small building (i.e., 24.424.4 m or 8080 feet). System resolution is 120 mm (4-3/4 inches) in range and 0.125o in bearing for full 360o coverage in a horizontal plane. The scan unit (less processing electronics) is a cube approximately 100 mm (4 inches) on a side, with a maximum 1-Hz update rate dictated by the 60-rpm scan speed. A dedicated 68HC11 microprocessor continuously outputs navigational parameters (X,Y, 2) to the vehicles onboard controller via an RS-232 serial port. Power requirements are 0.5 amps at 12 volts DC and 0.1 amp at 5 volts DC. The system costs $11,000.
Page 124

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

6.3.5 Siman Sensors & Intelligent Machines Ltd., "ROBOSENSE" The ROBOSENSE is an eye-safe, scanning laser rangefinder developed by Siman Sensors & Intelligent Machines Ltd., Misgav, Israel ( see Fig. 6.10). The scanner illuminates retro-reflective targets mounted on walls in the environment. According to the manufacturer [Siman, 1994], the system is designed to operate under severe or adverse conditions, such as the partial occlusion of the reflectors. The measurement range of the ROBOSENSE is 0.3 - 30 m (1-100 feet). The position accuracy is better than 2 cm (3/4 inch) and the accuracy in determining the orientation is better than 0.17 o . The system can communicate with an onboard computer via serial link and it updates the position and heading information at a rate of 10-40 Hz. A pre-production unit costs $12,500. Production models will be available in the first quarter of 1995, at a price of $6,500 ea. for a minimum order of three units.

Figure 6.10: The ROBOSENSE scanning laser rangefinder was developed by Siman Sensors & Intelligent Machines Ltd., Misgav, Israel. The system measures azimuth and range with an accuracy of 0.17o and 20 mm, respectively (courtesy Siman Sensors & Intelligent Machines).

6.3.7 Imperial College Beacon Navigation System Premi and Besant [1983] of the Imperial College of Science and Technology, London, England, describe an AGV guidance system that incorporates a vehicle-mounted laser beam rotating in a horizontal plane that intersects three fixed-location reference sensors as shown in Fig. 6.11. The photoelectric sensors are arranged in collinear fashion with equal separation, and individually wired to a common FM transmitter via appropriate electronics so that the time of arrival of laser energy is relayed to a companion receiver onboard the vehicle. A digitally-coded identifier in the data stream identifies the activated sensor that triggered the transmission, thus allowing the onboard

Page 125

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

computer to measure the separation angles "1 and "2. AGV position P(x,y) is given by the equations [Premi & Besant, 1983] x = x1 + r cos y = y1 + r sin where
r= a sin( 1 + ) sin 1

Y T 3 a T 2 a T1 Low Power Laser Beam 2 1 r P(x,y) AGV

(X1,Y ) 1

Figure 6.11: Three equidistant collinear photosensors are employed in lieu of retro-reflective beacons in the Imperial College laser triangulation system for AGV guidance (adapted from [Premi & Besant, 1983].

2 tan 1 tan 2 = arctan 1 tan 2 tan 1 =

An absolute or indexed incremental position encoder that monitors laser scan azimuth is used to establish platform heading. This technique has some inherent advantages over the use of passive retro-reflective targets, in that false acquisition of reflective surfaces is eliminated, and longer ranges are possible since target reflectivity is no longer a factor. More robust performance is achieved through elimination of target dependencies, allowing a more rapid scan rate to facilitate faster positional updates. The one-way nature of the optical signal significantly reduces the size, weight, and cost of the onboard scanner with respect to that required for retro-reflective beacon acquisition. Tradeoffs, however, include the increased cost associated with installation of power and communications lines and the need for significantly more expensive beacons. This can be a serious drawback in very-large-area installations, or scenarios where multiple beacons must be incorporated to overcome line-of-sight limitations. 6.3.8 MacLeod Technologies CONAC A similar type system using a predefined network of fixed-location detectors is currently being marketed by MacLeod Technologies, Inc., (MTI) Chelmsford, MA [MTI]. MTIs CONAC (Computerized Optoelectronic Navigation and Control) is a relatively low-cost high-performance navigational referencing system employing a vehicle-mounted laser STROAB (STRuctured Optoelectronic Acquisition Beacon). The scanning laser beam is spread vertically to eliminate critical alignment, allowing the reference NOADs (Networked Opto-electronic Acquisition Datums) to be mounted at arbitrary heights (see Fig. 6.12). Detection of incident illumination by a NOAD triggers a response over the network to a host PC, which in turn calculates the implied angles " 1 and " 2. An
Page 126

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

index sensor built into the STROAB generates a special rotation reference pulse to facilitate heading measurement. Indoor accuracy is advertised at 0.002 mm (0.05 inches) for position and 0.05o for heading.
Stationary Noads 3000+ RPM 2 1 Cable Link Radio Link To Host PC Mobile STROAB Laser Line Projection 3

Optional Heading Data Link

Figure 6.12: The Computerized Optoelectronic Navigation and Control (CONAC) system employs a number of networked detectors tracking a rapidly scanning laser emitter mounted on the moving vehicle (courtesy MacLeod Technologies, Inc.)

The reference NOADs are strategically installed at known locations throughout the area of interest, and daisy chained together with ordinary 4-conductor modular telephone cable. Alternatively the NOADS can be radio linked to eliminate cable installation problems, as long as power is independently available to the various NOAD sites. STROAB acquisition range is sufficient to where three NOADS can effectively cover an area of 4,046 m 2 (a square of 6363 m or one acre) assuming no interfering structures block the view. Additional NOADS are typically employed to increase fault tolerance and minimize ambiguities when two or more robots are operating in close proximity. The optimal set of three NOADS is dynamically selected by the host PC, based on the current location of the robot and any predefined visual barriers. The selected NOADS are individually addressed over the network in accordance with assigned codes (set into DIP switches on the back of each device at time of installation). An interesting and rather unconventional aspect of CONAC is that no fall-back dead reckoning capability is incorporated into the system [MacLeod & Chiarella, 1993]. The 3000-rpm angular rotation speed of the laser STROAB facilitates rapid position updates at a 25-Hz rate, which MTI claims is sufficient for safe automated transit at highway speeds, provided line-of-sight contact is preserved with at least three fixed NOADS. To minimize chances of occlusion, the light-weight (less than 9 ounces) STROAB is generally mounted as high as possible on a supporting mast.

Page 127

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 6: Active Beacon Navigation Systems

Long-range exterior position accuracy is specified as 1.3 mm (0.5 inches) and the heading accuracy as (0.05o). The nominal maximum distance is 36.5 m (120 feet), but larger distances can be covered with a more complex system. The system was successfully demonstrated in an outdoor environment when MacLeod engineers outfitted a Dodge Caravan with electric actuators for steering, throttle, and brakes, then drove the unmanned vehicle at speeds up to 80 km/h (50 mph) [Baker, 1993]. Absolute position and heading accuracies were sufficient to allow the Caravan to maneuver among parked vehicles and into a parking place using a simple AutoCad representation of the environment. Position computations are updated at a rate of 20 Hz. This system represents the current state-of-the-art in terms of active beacon positioning [Fox, 1993; Baker, 1993; Gunther, 1994]. A basic system with one STROAB and four NOADS costs on the order of $4,000. 6.3.9 Lawnmower CALMAN Larsson et al. [1994] from the University of Lulea, Sweden, have converted a large riding lawn mover to fully autonomous operation. This system, called CALMAN, uses an onboard rotating laser scanner to illuminate strategically placed vertical retro-reflector stripes. These reflectors are attached to tree stems or vertical poles in the environment. Larsson et al. report experimental results from running the vehicle in a parking lot. According to these results, the vehicle had a positioning error of less than 2 cm (3/4 inch) at speeds of up to 0.3 m/sec (1 foot/sec). The motion of the vehicle was stable at speeds of up to 1 m/sec (3.3 feet/sec) and became unstable at 1.5 m/sec (5 feet/sec).

6.4 Summary
We summarize the general characteristics of active beacon systems as follows:
+ The environment needs to be modified, and some systems require electric outlets or battery

maintenance for stationary beacons.


+ A line-of-sight between transmitter and detector needs to be maintained, i.e., there must be at

least two or three visible landmarks in the environment.


+ Triangulation-based methods are subject to the limitations of triangulation as discussed by Cohen

and Koss [1992].


+ Active beacon systems have been proven in practice, and there are several commercial systems

available using laser, infrared, and ultrasonic transducers.


+ In practice, active beacon systems are the choice when high accuracy and high reliability are

required.

Page 128

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 7: Landmark Navigation

Chapter 7: Landmark Navigation

Landmarks are distinct features that a robot can recognize from its sensory input. Landmarks can be geometric shapes (e.g., rectangles, lines, circles) and they may include additional information (e.g., in the form of bar-codes). In general, landmarks have a fixed and known position, relative to which a robot can localize itself. Landmarks are carefully chosen to be easy to identify; for example, there must be sufficient contrast to the background. Before a robot can use landmarks for navigation, the characteristics of the landmarks must be known and stored in the robot's memory. The main task in localization is then to recognize the landmarks reliably and to calculate the robot's position. In order to simplify the problem of landmark acquisition it is often assumed that the current robot position and orientation are known approximately, so that the robot only needs to look for landmarks in a limited area. For this reason good dead-reckoning accuracy is a prerequisite for successful landmark detection. The general procedure for performing landmark-based positioning is shown in Fig. 7.1.

1. Use special beacons 2. Use distinct landmarks

1. Constrain search by assuming initial estimate is close to the true position and orientation 2. Currently detection and correspondence are bottlenecks in landmark positioning

1. Triangulation: measurement error is a function of the relative position between the robot and landmarks 2. Geometric shape: measurement error is a function of the distance & angle between the robot and the feng1 pos .ds 4, .w m f landmark

Figure 7.1: General procedure for landmark-based positioning

Page 129

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 7: Landmark Navigation

Some approaches fall between landmark and map-based positioning (see Chapter 8). They use sensors to sense the environment and then extract distinct structures that serve as landmarks for navigation in the future. These approaches will be discussed in the chapter on map-based positioning techniques. Our discussion in this chapter addresses two types of landmarks: "artificial" and "natural". It is important to bear in mind that "natural" landmarks work best in highly structured environments such as corridors, manufacturing floors, or hospitals. Indeed, one may argue that "natural" landmarks work best when they are actually man-made (as is the case in highly structured environments). For this reason, we shall define the terms "natural landmarks" and "artificial landmarks" as follows: natural landmarks are those objects or features that are already in the environment and have a function other than robot navigation; artificial landmarks are specially designed objects or markers that need to be placed in the environment with the sole purpose of enabling robot navigation.

7.1 Natural Landmarks


The main problem in natural landmark navigation is to detect and match characteristic features from sensory inputs. The sensor of choice for this task is computer vision. Most computer visionbased natural landmarks are long vertical edges, such as doors and wall junctions, and ceiling lights. As stated in the introduction to this survey, computer vision is an area that is too large and too diverse for the scope of this report. For this reason computer vision-based landmark detection will not be addressed. When range sensors are used for natural landmark navigation, distinct signatures, such as those of a corner or an edge, or of long straight walls, are good feature candidates. The selection of features is important since it will determine the complexity in feature description, detection, and matching. Proper selection of features will also reduce the chances for ambiguity and increase positioning accuracy. A natural landmark positioning system generally has the following basic components: + A sensor (usually computer vision) for detecting landmarks and contrasting them against their background + A method for matching observed features with a map of known landmarks + A method of computing location and localization errors from the matches.

7.2 Artificial Landmarks


Detection is much easier with artificial landmarks [Atiya & Hager, 1993], which are designed for optimal contrast. In addition, the exact size and shape of artificial landmarks is known in advance. Size and shape can yield a wealth of geometric information when transformed under the perspective projection.

Page 130

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 7: Landmark Navigation

Researchers have used different kinds of patterns or marks, and the geometry of the method and the associated techniques for position estimation vary accordingly [Talluri and Aggarwal, 1993]. Many artificial landmark positioning systems are based on computer vision. We will not discuss these systems in detail, but we will mention some of the typical landmarks used with computer vision. Fukui [1981] used a diamond shaped landmark and applied a least-square method to find line segments in the image plane. Borenstein [1987] used a black rectangle with four white dots in the corners. Kabuka and Arenas [1987] used a half white and half black circle with a unique bar code for each landmark. Magee and Aggarwal [1984] used a sphere with horizontal and vertical calibration circles to achieve 3D localization from a single image. Other systems use reflective material patterns and strobe light to ease the segmentation and parameter extraction [Lapin, 1992; Mesaki & Masuda, 1992]. There are also systems that use active (i.e., LED) patterns to achieve the same effect [Fleury & Baron, 1992]. The accuracy achieved by the above methods depends on the accuracy with which the geometric parameters of the landmark images are extracted from the image plane, which in turn, depends on the relative position and angle between the robot and the landmark. In general, the accuracy decreases with the increase in relative distance. Normally there is a range of relative angles in which good accuracy can be achieved, while accuracy drops significantly once the relative angle moves out of the "good" region. There is also a variety of landmarks used in conjunction with non-vision sensors. Most often used are bar-coded reflectors for laser scanners. For example, currently ongoing work by Everett on the Mobile Detection Assessment and Response System (MDARS) [DeCorte, 1994] uses bar-code reflectors, and so does the commercially available system from Caterpillar on their Self Guided Vehicle [Gould, 1990]. The shape of these landmarks is usually unimportant. By contrast, a unique approach taken by Feng et al. [1992] used a circular landmark and applied an optical Hough transform to extract the parameters of the ellipse on the image plane in real time. 7.2.1 Global Vision Yet another approach is the so-called global vision that refers to the use of cameras placed at fixed locations in a work space to extend the local sensing available on board each AGV [Kay & Luo, 1993]. Fig. 7.2 shows a block diagram of the processing functions for vehicle control using global vision. In global vision methods, characteristic points forming a pattern on the mobile robot are identified and localized from a single view. A probabilistic method is used to select the most probable matching according to geometrical characteristics of those percepts. From this reduced search space a prediction-verification loop is applied to identify and to localize the points of the pattern [Fleury & Baron, 1992]. One advantage of this approach is that it allows the operator to monitor robot operation at the same time.

Page 131

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 7: Landmark Navigation

Off-line central processing


Load Database Transport Equip. DB Camera Placement 3-D World Model

Fixed Obj. DB Create 3-D Model

Facility layout

System

On-line central processing status


Mobile Object Tracking 2-D World Model Free path verification Collision avoidance Path planning

Dispatch. request

Controls
Camera Transmit/ Receive AGV controller

FM radio

Vehicle status

Infrared

Vehicle processing

Transmit/ Receive

Ultrasonic sensors

Infrared sensors

Odometers

Bumpers

kay_luo.ds4, .w mf, 11/12/94

Figure 7.2: Block diagram of the processing functions for vehicle control using global vision (adapted from [Kay and Luo, 1993]).

7.3 Artificial Landmark Navigation Systems


It may not be immediately clear why we are distinguishing in this report between retro-reflective bar-codes as artificial landmarks, and retro-reflective poles, as for example those used by TRC as "beacons." The answer is that if retro-reflective markers (with or without bar-code) are attached to the walls of a room and their function is merely to aid in determining the location of the wall, then these markers do not function as beacons. By contrast, if markers are used on arbitrarily placed poles (even if the location of these poles is carefully surveyed), then they act as beacons. A related distinction is the method used for computing the vehicle's position: if triangulation is used, then the reflectors act as beacons.

Page 132

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 7: Landmark Navigation

7.3.1 MDARS Lateral-Post Sensor Currently ongoing work by Everett on the Mobile Detection Assessment and Response System (MDARS) [Everett et al., 1994; DeCorte, 1994] uses bar-code reflectors in conjunction with a rotating laser onboard the robot. This technique, called lateral post detection, was incorporated on MDARS to significantly reduce costs by exploiting the forward motion of the robot for scanning purposes. Short vertical strips of 2.5 cm (1 inch) retro-reflective tape are placed on various immobile objects (usually structural-support posts) on either side of a virtual path segment. The exact X-Y locations of these tape markers are encoded into the virtual path program. Installation takes only seconds, and since the flat tape does not protrude into the aisle way at all there is little chance of damage from a passing fork truck. A pair of Banner Q85VR3LP retro-reflective proximity sensors mounted on the turret of the Navmaster robot face outward to either side as shown in Fig. 7.3 These inexpensive sensors respond to reflections from the tape markers along the edges of the route, triggering a snapshot virtual path instruction that records the current side-sonar range values. The longitudinal position of the platform is updated to the known marker coordinate, while lateral position is inferred from the sonar data, assuming both conditions fall within specified tolerances.

Retro-reflective Markers

The accuracy of the marker correction is much higher (and therefore assigned greater credibility) than that of the lateral sonar readings due to the markedly different uncertainties associated with the respective targets. The circularly-polarized Banner sensor responds only to the presence of a retro reflector while ignoring even highly specular surrounding surfaces, whereas the ultrasonic energy from the sonar will echo back from any reflective surface encountered by its relatively wide beam. Protruding objects in the vicinity of the tape (quite common in a warehouse environment) result in a shorter measured range value than the reference distance for the marker itself. The overall effect on X-Y bias is somewhat averaged out in the long run, as each time the vehicle executes a 90o course change the association of X and Y components with tape versus sonar updates is interchanged. 7.3.2 Caterpillar Self Guided Vehicle Caterpillar Industrial, Inc., Mentor, OH, manufactures a free-ranging AGV for materials handling that relies on a scanning laser triangulation scheme to provide positional updates to the vehicles onboard dead-reckoning system. The Class-I laser rotates at 2 rpm to illuminate passive retroreflective bar-code targets affixed to walls or support columns at known locations up to 15 m (50
Page 133

Figure 7.3: Circularly polarized retro-reflective proximity sensors are used to locate vertical strips of retro-reflective tape attached to shelving support posts in the Camp Elliott warehouse installation of the MDARS security robot [Everett et al, 1994].

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 7: Landmark Navigation

feet) away [Gould, 1990; Byrne, et al., 1992]. The bar codes serve to positively identify the reference target and eliminate ambiguities due to false returns from other specular surfaces within the operating area. An onboard computer calculates X-Y position updates through simple triangulation to null out accumulated dead-reckoning errors (Fig. 7.4). Some target occlusion problems have been 10 to 15 meters apart are used by the Caterpillar SGV experienced in exterior applications where to triangulate position (adapted from [Caterpillar, there is heavy fog, as would be expected, and 1991a]]. minor difficulties have been encountered as well during periods when the sun was low on the horizon [Byrne, 1993]. The Self Guided Vehicle relies on dead reckoning under such conditions to reliably continue its route for distances of up to 10 m (33 feet) before the next valid fix. The robot platform is a hybrid combination of tricycle and differential drive, employing two independent series-wound DC motors powering 45 cm (18-inch) rear wheels through sealed gearboxes [CATERPILLAR, 1991]. High resolution resolvers attached to the single front wheel continuously monitor steering angle and distance traveled. A pair of mechanically scanned nearinfrared proximity sensors sweeps the path in front of the vehicle for potential obstructions. Additional near infrared sensors monitor the area to either side of the vehicle, while ultrasonic sensors cover the back.
Figure 7.4: Retro-reflective bar-code targets spaced

7.4 Line Navigation


Another major type of landmark navigation that has been widely used in the industry is line navigation. Line navigation can be thought of as a continuous landmark, although in most cases the sensor used in this system needs to be very close to the line, so that the range of the vehicle is limited to the immediate vicinity of the line. There are different implementations for line navigation: 1. Electromagnetic Guidance or Electro-magnetic Leader Cable. 2. Reflecting Tape Guidance (also called Optical Tape Guidance). 3. Ferrite Painted Guidance, which uses ferrite magnet powder painted on the floor [Tsumura, 1986]. These techniques have been in use for many years in industrial automation tasks. Vehicles using these techniques are generally called Automatic Guided Vehicles (AGVs).

Page 134

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 7: Landmark Navigation

In this report we don't address these methods in detail, because they do not allow the vehicle to move freely the main feature that sets mobile robots apart from AGVs. However, a recently introduced variation of the above technique is of interest for mobile robots. This variation is based on laying down and detecting trails of volatile chemicals. The implementation of this technique involves an improved odor sensing system which allows odor markings on the floor to be reliably detected and accurately localized. The sensor system uses controlled flows of air to draw odor laden air over a sensor crystal [Russel et al., 1994]. This approach is suitable for temporarily marking a path taken by a robot in a prior run. One typical group of applications can the defined as floor coverage applications. Examples are floor cleaning, lawn moving, or floor surveillance. In such applications it is important for the robot to travel along adjacent paths on the floor, with minimal overlap and without "blank" spots. With the method discussed here, the robot could conceivably mark the outside border of the path, and trace that border line in a subsequent run. Another variation is to use short-lived navigational markers (SLNM) which represent a group of techniques for indicating the path taken by a mobile robot. The short lived nature of the markers has the advantage that it is not necessary to remove them after use. However, the varying intensity of the SLNM's complicates the process of detecting and following the robot path. Kleeman [1992] reports on a pyroelectric sensor that has been developed to detect thermal paths created by heating the floor with a quartz halogen bulb. Kleeman reported that thermal paths can be tracked up to ten minutes after being laid on a vinyl tiled floor.

7.5 Summary
Artificial landmark detection methods are well developed and reliable. By contrast, natural landmark navigation is not sufficiently developed yet for reliable performance under a variety of conditions. A survey of the market of commercially available natural landmark systems, produces only a few. One of them is TRC's vision system that allows the robot to localize itself using rectangular and circular ceiling light [King & Weiman, 1990]. Cyberworks has a similar system [Cyberworks]. It is generally very difficult to develop a feature-based landmark positioning system capable of detecting different natural landmarks in different environments. It is also very difficult to develop a system that is capable of using many different types of landmarks. We summarize the characteristics of landmark-based navigation as follows:
+ Natural landmark offer flexibility and require no modifications to the environment. + Artificial landmarks are inexpensive and can have additional information encoded as patterns or

shapes.
+ The maximal distance between robot and landmark is substantially shorter than in active beacon

systems.

Page 135

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 7: Landmark Navigation

+ The positioning accuracy depends on the distance and angle between the robot and the landmark.

Landmark navigation is rather inaccurate when the robot is further away from the landmark. A higher degree of accuracy is obtained only when the robot is near a landmark.
+ Substantially more processing is necessary than with active beacon systems. + Ambient conditions, such as lighting, can be problematic; in marginal visibility, landmarks may

not be recognized at all or other objects in the environment with similar features can be mistaken for a legitimate landmark.
+ Landmarks must be available in the work environment around the robot; + Landmark-based navigation requires an approximate starting location so that the robot knows

where to look for landmarks. If the starting position is not known, the robot has to conduct a timeconsuming search process.
+ A database of landmarks and their location in the environment must be maintained. + There is only limited commercial support for this type of technique.

Page 136

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

Chapter 8: Map-based Positioning

Map-based positioning, also known as "map-matching," is a technique in which the robot uses its sensors to create a map of its local environment. This local map is then compared to a global map previously stored in memory. If a match is found, then the robot can compute its actual position and orientation in the environment. The pre-stored map can be a CAD model of the environment, or it can be constructed from prior sensor data. The basic procedure for map-based positioning is shown in Fig. 8.1.

Establish correspondence between local map and stored global map


1. Filtering 2. Sensor fusion 3. Sensor data and robot position error modeling 1. Constrained search by assuming the initial position estimate is close to the true robot position 2. Sensory and robot motion error models are also used to constrain search area 3. Matching: least square and correlation 4. Measurement of matching goodness 5. Currently correspondence is the main problem in map-based positioning

fengpos2.ds4, .wm f

Figure 8.1: General procedure for map-based positioning

The main advantages of map-based positioning are


+ This method uses the naturally occurring structure of typical indoor environments to derive

position information without modifying the environment.

Page 137

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

+ Map-based positioning can be used to generate an updated map of the environment. Environment

maps are important for other mobile robot tasks, such as global path planning or the avoidance of "local minima traps" in some local obstacle avoidance methods.
+ Map-based positioning allows a robot to learn a new environment, and to improve positioning

accuracy through exploration. Disadvantages of map-based positioning are the specific requirements for satisfactory navigation. For example, map-based positioning requires that + there be enough stationary, easily distinguishable features that can be used for matching + the sensor map be accurate enough (depending on the tasks) to be useful + a significant amount of sensing and processing power be available. One should note that currently most work in map-based positioning is limited to laboratory settings and to relatively simple environments.

8.1 Map-building
There are two fundamentally different starting points for the map-based positioning process. Either there is a pre-existing map, or the robot has to build its own environment map. Rencken [1993] defined the map building problem as the following: "Given the robot's position and a set of measurements, what are the sensors seeing"? Obviously, the map building ability of a robot is closely related to its sensing capacity. Talluri and Aggarwal [1993] explained: "The position estimation strategies that use map-based positioning rely on the robot's ability to sense the environment and to build a representation of it, and to use this representation effectively and efficiently. The sensing modalities used significantly affect the map making strategy. Error and uncertainty analyses play an important role in accurate position estimation and map building. It is important to take explicit account of the uncertainties; modeling the errors by probability distributions and using Kalman filtering techniques are good ways to deal with these errors explicitly." Talluri and Aggarwal [1993] also summarized the basic requirements for a map: "The type of spatial representation system used by a robot should provide a way to incorporate consistently the newly sensed information into the existing world model. It should also provide the necessary information and procedures for estimating the position and pose of the robot in the environment. Information to do path planning, obstacle avoidance, and other navigation tasks must also be easily extractable from the built world model."

Page 138

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

Hoppen et al. [1990] listed the three main steps of sensor data processing for map building: 1. feature extraction from raw sensor data 2. fusion of data from various sensor types 3. automatic generation of an environment model with different degrees of abstraction. And Crowley [1989] summarized the construction and maintenance of a composite local world model as a three-step process: 1. building an abstract description of the most recent sensor data (a sensor model) 2. matching and determining the correspondence between the most recent sensor models and the current contents of the composite local model 3. modifying the components of the composite local model and reinforcing or decaying the confidences to reflect the results of matching. 8.1.1 Map-building and sensor-fusion Many researchers believe that no single sensor modality alone can adequately capture all relevant features of a real environment. To overcome this problem, it is necessary to combine data from different sensor modalities, a process known as sensor-fusion. Here are a few examples:
+ Buchenberger et al. [1993] developed a mechanism that utilizes heterogeneous information

obtained from a laser-radar and a sonar system in order to construct a reliable and complete world model.
+ Courtney and Jain [1994] integrated three common sensing sources (sonar, vision, and infra-red)

for sensor-based spatial representation. They implemented a feature-level approach to sensor fusion from multisensory grid maps using a mathematical method based on spatial moments and moment invariants, which are defined as follows: The two-dimensional (p+q)th order spacial moments of a grid map G(x,y) are defined as

mpq = x p y qG( x, y )
x y

p, q = 012,... ,,

Using the centroid, translation-invariant central moments (moments don't change with the translation of the grid map in the world coordinate system) are formulated:

pq = ( x x ) p ( y y)q G( x, y )
x y

Page 139

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

From the second- and third-order central moments, a set of seven moment invariants that are independent of translation, rotation, and scale can be derived. A more detailed treatment of spatial moments and moment invariants is given in [Gonzalez & Wintz, 1977].

8.1.2 Phenomenological vs. geometric representation, Engelson and McDermott [1992] Most research in sensor-based map building attempts to minimize mapping errors at the earliest stage when the sensor data is entered into the map. Engelson and McDermott [1992] suggest that this methodology will reach a point of diminishing returns, and hence further research should focus on explicit error detection and correction. The authors observed that the geometric approach attempts to build a more-or-less detailed geometric description of the environment from perceptual data. This has the intuitive advantage of having a reasonably well-defined relation to the real world. However, there is, as yet, no truly satisfactory representation of uncertain geometry, and it is unclear whether the volumes of information that one could potentially gather about the shape of the world are really useful. To overcome this problem Engelson and McDermott suggested the use of a topological approach that constitutes a phenomenological representation of the robot's potential interactions with the world, and so directly supports navigation planning. Positions are represented relative to local reference frames to avoid unnecessary accumulation of relative errors. Geometric relations between frames are also explicitly represented. New reference frames are created whenever the robot's position uncertainty grows too high; frames are merged when the uncertainty between them falls low enough. This policy ensures locally bounded uncertainty. Engelson and McDermott showed that such error correction can be done without keeping track of all mapping decisions ever made. The methodology makes use of the environmental structure to determine the essential information needed to correct mapping errors. The authors also showed that it is not necessary for the decision that caused an error to be specifically identified for the error to be corrected. It is enough that the type of error can be identified. The approach has been implemented only in a simulated environment, where the effectiveness of the phenomenological representation was demonstrated.

8.2 Map matching


One of the most important and challenging aspects of map-based navigation is map matching, i.e., establishing the correspondence between a current local map and the stored global map [Kak et al., 1990]. Work on map-matching in the computer vision community is often focused on the general problem of matching an image of arbitrary position and orientation relative to a model (for example, [Talluri & Aggarwal, 1993]. In general, matching is achieved by first extracting features, followed by determination of the correct correspondence between image and model features, usually by some form of constrained search [Cox, 1991]. Such matching algorithms can be classified as either icon-based or feature-based. Schaffer et al. [1992] summarized these two approaches:
Page 140

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

"Iconic-based pose estimation pairs sensory data points with features from the map, based on minimum distance. The robot pose is solved for that minimizes the distance error between the range points and their corresponding map features. The robot pose is solved [such as to] minimize the distance error between the range points and their corresponding map features. Based on the new pose, the correspondences are recomputed and the process repeats until the change in aggregate distance error between points and line segments fall below a threshold. This algorithm differs from the feature-based method in that it matches every range data point to the map rather than corresponding the range data into a small set of features to be matched to the map. The feature-based estimator, in general, is faster than the iconic estimator and does not require a good initial heading estimate. The iconic estimator can use fewer points than the feature-based estimator, can handle less-than-ideal environments, and is more accurate. Both estimators are robust to some error in the map." Kak et al. [1990] pointed out that one problem in map-matching is that the sensor readings and the world model may be of different formats. One typical solution to this problem is that the approximate position based on dead-reckoning is utilized to generate (from the pre-stored global model), an estimated visual scene that would be "seen" by robot. This estimated scene is then matched against the actual scene viewed by the on-board sensors. Once the matches are established between the features of the two images (expected and actual), the position of the robot can be estimated with reduced uncertainty. This approach is also supported by Rencken [1994], as will be discussed in more detail below. In order to match the current sensory data to the stored environment model reliably, several features must be used simultaneously. This is particularly true for a range image-based system since the types of features are limited to a range image map. Long walls and edges are the most commonly used features in a range image-based system. In general, the more features used in one match, the less likely a mismatch will occur, but the longer it takes to process. A realistic model for the odometry and its associated uncertainty is the basis for the proper functioning of a map-based positioning system. This is so because the feature detection as well as the updated position calculation rely on odometric estimates [Chenavier & Crowley, 1992]. 8.2.1 Schiele and Crowley [1994] Schiele and Crowley [1994] discussed different matching techniques for matching two occupancy grids. The first grid is the local grid that is centered on the robot and models its vicinity using the most recent sensor readings. The second grid is a global model of the environment furnished either by learning or by some form of computer aided design tool. Schiele and Crowley propose that two representations be used in environment modeling with sonars: parametric primitives and an occupancy grid. Parametric primitives describe the limits of free space in terms of segments or surfaces defined by a list of parameters. However, noise in the sensor signals can make the process of grouping sensor readings to form geometric primitives unreliable. In particular, small obstacles such as table legs are practically impossible to distinguish from noise.
Page 141

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

Figure 8.2: Schiele and Crowley's robot models its position in a hallway: a) Raw ultrasonic range data projected onto external coordinates around the robot. b) Local grid and the edge segments extracted from this grid. c) The robot with its uncertainty in estimated position within the global grid. d) The local grid imposed on the global grid at the position and orientation of best correspondence. (Reproduced and adapted from [Schiele and Crowley, 1994])

Schiele and Crowley discuss four different matches: 1. Matching segment to segment as realized by comparing segments in (1) similarity in orientation, (B) colinearity, and (C) overlap 2. Matching segment to grid 3. Matching grid to segment 4. Matching grid to grid as realized by generating a mask of the local grid. This mask is then transformed into the global grid and correlated with the global grid cells lying under this mask. The value of that correlation increases when the cells are of the same state and decreases when

Page 142

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

the two cells have different states. Finally finding the transformation that generates the largest correlation value. Schiele and Crowley pointed out the importance of designing the updating process to take into account the uncertainty of the local grid position. The correction of the estimated position of the robot is very important for the updating process particularly during exploration of unknown environments. Figure 8.2 shows an example of one of the experiments with the robot in a hallway. Experimental results obtained by Schiele and Crowley show that the most stable position estimation results are obtained by matching segments to segments or grids to grids. 8.2.2 Hinkel and Knieriemen [1988] the Angle Histogram Hinkel and Knieriemen [1988] from the University of Kaiserslautern, Germany, developed a powerful tool for locating a robot's position relative to the walls of a room. At first, the robot takes a 360o-scan of the room with a laser-scanner, and the resulting "hits" are recorded in a map. Then the algorithm measures the relative angle * between any two adjacent hits. After compensating for noise in the readings (caused by the inaccuracies in position between adjacent hits), an angle histogram (see Fig. 8.3a) can be built. The uniform direction of the main walls are clearly visible as peaks in the angle histogram. Computing the histogram modulo B results in only two main peaks: one for each pair of parallel walls. This algorithm is very robust with regard to openings in the walls, such as doors and windows, or even cabinets lining the walls. After computing the angle histogram, all angles of the hits can be normalized, resulting in the representation shown in Fig. 8.3b. After this transformation, two additional histograms, one for the X and one for the Y direction can be constructed. This time, peaks show the distance to the walls in X and Y direction. During operation, new orientation and position data is updated at a rate of 4 Hz. (In conversation with Prof. Von Puttkammer, Director of the Mobile Robotics Laboratory at the University of Kaiserslautern, we learned that this algorithm had since been improved to yield a reliable accuracy of 0.5o).

Page 143

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

20o

Figure 8.3:

pos1 0rpt. ds4 , .w m f

Readings from a rotating laser scanner generate the contour of a room. (a) An angle histogram allows the robot to determine its orientation relative to the walls. (b) After normalizing the orientation of the room relative to the robot, an x-y histogram can be built form the same data points. (Adapted from [Hinkel and Knieriemen, 1988])

8.2.3 Siemens' Roamer Rencken [1993; 1994] at the Siemens Corporate Research and Development Center in Munich, Germany, has made substantial contributions toward solving the bootstrap problem resulting from the uncertainty in position and environment. This problem exists when a robot must move around in an unknown environment, with uncertainty in the robot's odometry-derived position. For example, when building a map of the environment, all measurements are necessarily relative to the carrier of the sensors (i.e., the mobile robot). Yet, the position of the robot itself is not known exactly, because of the errors accumulating in dead-reckoning. Rencken addresses the problem as follows: In order to represent features "seen" by its 24 ultrasonic sensors, the robot constructs hypotheses about these features. To account for the typically unreliable information from ultrasonic sensors, features can be classified as hypothetical, tentative, or confirmed. Once a feature is "confirmed," it is used for constructing the map as shown in Fig. 8.4. Before the map can be updated, though, every new data point must be associated with either a plane, a corner, or an edge (and some variations of those). Rencken devices a "hypothesis tree" which is a data structure that allows tracking of different hypotheses until a sufficient amount of data has been accumulated to make a final decision.
Page 144

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

Map Building
Delete Features

Localisation

no

Plausible and Certain?

Map
Update Confirmed Features Confirmed Features Plausible and yes Certain?

yes

no

Implausible?

yes

no

Large enough?

yes

Update Tentative Features

Tentative Features

yes

Large enough?

no

Too old?

yes

Generate Hypothesis

Hypothetical Features

Cluster Hypothesis

Robot Position

Observation

Sensor Measurements
pos20r ep.D S4, . WMF

Figure 8.4: The basic map building algorithm maintains a hypothesis tree for the three sensor reading categories: hypothetical, tentative, and confirmed (adapted from [Rencken, 1994]).

One further important aspect in making this decision is "feature visibility." Based on internal models for different features, the robot's decisions are aided by a routine check on visibility. For example, the visibility of edges is smaller than that of corners. The visibility check further reduces the uncertainty and improves the robustness of the algorithm. Based on the above methods, Rencken [1993] summarizes his method with the following procedure: 1. Predict the robot's position using dead-reckoning. 2. Predict the associated covariance of this position estimate. 3. Among the set of given features, test which feature is visible to which sensor and predict the measurement. 4. Compare the predicted measurements to the actual measurements. 5. Use the error between the validated measured and predicted measurements to estimate the robot's position. 6. The associated covariance of the new position estimate is also determined.

Page 145

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

The algorithm was implemented on Siemens' experimental robot Roamer. In an endurance experiment, Roamer traveled through a highly cluttered office environment for approximately 20 minutes. During this time, the robot updated its internal position only by means of odometry and its map-building capabilities. At a relatively slow travel speed of 12 cm/sec (4 inches/sec) Roamer's position accuracy was periodically recorded, as shown in Table 8.1.
Time [min:sec] 5:28 11:57 14:53 18:06 20:12 Pos Error [cm] (inches) 5.8 (2-1/4) 5.3 (2) 5.8 (2-1/4) 4.0 (1-1/2) 2.5 (1) Orientation error [o] -7.5 -6.2 0.1 -2.7 3.0

Table 8.1: Position and orientation errors of Siemens' Roamer robot in an map-building "endurance test" (Adapted from [Rencken, 1994]).

8.3 Geometric and Topological Maps


In map-based positioning there are two common representations: geometric and topological maps. A geometric map represents objects according to their absolute geometric relationships. It can be a grid map, or a more abstracted map, such as a line map or a polygon map. In map-based positioning, sensor-derived geometric maps must be matched against a global map of a large area. This is often a formidable difficulty because of the robot's position error. By contrast, the topological approach is based on recording the geometric relationships between the observed features rather than their absolute position with respect to an arbitrary coordinate frame of reference. The resulting presentation takes the form of a graph where the nodes represent the observed features and the edges represent the relationships between the features. Unlike geometric maps, topological maps can be built and maintained without any estimates for the position of the robot. This means that the errors in this representation will be independent of any errors in the estimates for the robot position [Taylor, 1991]. This approach allows one to integrate large area maps without suffering from the accumulated dead-reckoning position error since all the connections are relative between nodes instead of absolute. After the map has been established, the positioning process is essentially the process of matching a local map to the appropriate location on the stored map. 8.3.1 Geometric Maps for Navigation There are different ways for representing geometric map data. Perhaps the simplest way is an occupancy grid-based map. The first such map (in conjunction with mobile robots) was the Certainty Grid, developed by Moravec and Elfes, [1985]. In the Certainty Grid approach, sensor
Page 146

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

readings are placed into the grid by using probability profiles that describe the algorithm's certainty about the existence of objects at individual grid cells. Based on the Certainty Grid approach, Borenstein and Koren [1991] refined the method with the Histogram Grid, which derives a pseudoprobability distribution out of the motion of the robot [Borenstein & Koren, 1991]. The Histogram Grid method is now widely used in many mobile robots (see for example [Buchenberger et al., 1993; Congdon et al., 1993; Courtney & Jain, 1994; Stuck et al., 1994; Wienkop et al., 1994]). A measure of the goodness of the match between two maps and a trial displacement and rotation is found by computing the sum of products of corresponding cells in the two maps [Elfes, 1987]. Range measurements from multiple points of view are symmetrically integrated into the map. Overlapping empty volumes reinforce each other and serve to condense the range of the occupied volumes. The map definition improves as more readings are added. The method deals effectively with clutter and can be used for motion planning and extended landmark recognition. The advantages of occupancy grid-based maps are that they + allow higher density than stereo maps, + require less computation and can be built more quickly, + allow for easy integration of data from different sensors, and + can be used to express statistically the confidence in the correctness of the data [Raschke & Borenstein, 1990]. The disadvantages of occupancy grid-based maps are that they + have large uncertainty areas associated with the features detected + have difficulties associated with active sensing [Talluri and Aggarwal, 1993] + have difficulties associated with modeling of dynamic obstacles + require a more complex estimation process for the robot vehicle [Schiele and Crowley, 1994]. In the following sections we discuss some specific examples for occupancy grid-based map matching. 8.3.1.1 Cox [1991] One typical grid-map system was implemented on the mobile robot Blanche [Cox, 1991]. This positioning system is based on matching a local grid map to a global line segment map. Blanche is designed to operate autonomously within a structured office or factory environment without active or passive beacons. Blanche's positioning system consists of 1. an a priori map of its environment, represented as a collection of discrete line segments in the plane 2. a combination of odometry and a rotating optical range sensor to sense the environment 3. an algorithm for matching the sensory data to the map, where matching is constrained by assuming that the robot position is roughly known from dead-reckoning

Page 147

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

4. an algorithm to estimate the precision of the corresponding match/correction that allows the correction to be combined optimally (in a maximum likelihood sense) with the current odometric position to provide an improved estimate of the vehicle's position. The operation of Cox' map matching algorithm (item 2, above) is quite simple. Assuming that the sensor hits are near the actual objects (or rather, the lines that represent the objects), the distance between a hit and the closest line is computed. This is done for each point, according to the procedure in Table 8.2 (from [Cox, 1991]):
1. For each point in the image, find the line segment in the model that is nearest to the point. Call this the target. 2. Find the congruence that minimizes the total squared distance between the image points and their target lines. 3. Move the points by the congruence found in step 2. 4. Repeat steps 1-3 until the procedure converges. Table 8.2: Procedure for implementing Cox' [1991] map-matching algorithm.

Figure 8.5 shows how the algorithm work on a set of real data: Figure 8.5a shows the line model of the contour of the office environment (solid lines). The dots show hits by the range sensor. This scan was taken while the robot's position estimate was offset from its true position by 2.75 m (9 ft) in x and 244 m (8 feet) in y direction. A very small orientation error was also present. After running the map-matching procedure in Table 8.2, the robot corrected its internal position, resulting in the very good match between sensor data and line model, shown in Fig. 8.5b. In a longer run through corridors and junctions Blanche traveled at various slow speeds, on the order of 5 cm/sec (2 inches/sec). The maximal deviation of its computed position from the actual position was said to be 15 cm (6 inches).

Page 148

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

Figure 8.5: Map and range data (a) before registration and (b) after registration. (Reproduced and adapted from [Cox, 1991]).

Discussion With the grid-map system used in Blanche, generality has been sacrificed for robustness and speed. The algorithm is intrinsically robust against incompleteness of the image. Incompleteness of the model is dealt with by deleting any points whose distance to their target segments exceed a certain limit . In Cox's approach, a reasonable heuristic used for determining correspondence is the minimum Euclidean distance between model and sensed data. Gonzalez et al. [1992] comment that this assumption is valid only as long as the displacement between the sensed data and the model is small enough. However, this minimization problem is inherently nonlinear, but is linearized assuming that the rotation angle is small. To compensate for the error introduced due to linearization, the computed position correction is applied to the data points and the process is repeated until no significant improvement can be obtained [Jenkin et al., 1993]. 8.3.1.2 Crowley [1989] Crowley's [1989] system is based on matching a local line segment map to a global line segment map. Crowley develops a model for the uncertainty inherent in ultrasonic range sensors, and he describes a method for the projection of range measurements onto external Cartesian coordinates. Crowley develops a process for extracting line segments from adjacent co-linear range measurements and he presents a fast algorithm for matching these line segments to a model of the geometric limits for the free-space of the robot. A side effect of matching sensor-based observations to the model is a correction to the estimated position of the robot at the time that the observation was made. The projection of a segment into the external coordinate system is based on the estimate of the position of the vehicle. Any uncertainty in the vehicle's estimated position must be included in the uncertainty
Page 149

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

of the segment before matching can proceed. This uncertainty affects both the position and orientation of the line segment. As each segment is obtained from the sonar data, it is matched to the composite model. Matching is a process of comparing each of the segments in the composite local model against the observed segment, to allow detection of similarity in orientation, colinearity, and overlap. Each of these tests is made by comparing one of the parameters in the segment representation: 1. Orientation: the square of the difference in orientation of the two candidates must be smaller than the sum of the variances. 2. Alignment: the square of the difference of the distance from the origin to the two candidates must be smaller than the sum of the corresponding variance. 3. Overlap: the difference of the distance between center-points to the sum of the half lengths must be smaller than a threshold. The longest segment in the composite local model which passes all three tests is selected as the matching segment. The segment is then used to correct the estimated position of the robot and to update the model. An Figure 8.5: Model of the ultrasonic range sensor and its uncertainties explicit model of uncer- (adapted from [Crowley, 1989]). tainty using covariance and Kalman filtering provides a tool for integrating noisy and imprecise sensor observations into a model of the geometric limits for the free-space of a vehicle. Such a model provides a means for a vehicle to maintain an estimate of its position as it travels, even in the case where the environment is unknown. Figure 8.5 shows the model of the ultrasonic range sensor and its uncertainties (shown as the hatched area A). The length of the A is given by the uncertainties in robot orientation, F w and the width is given by the uncertainty in depth, F D. This area is approximated by an ellipse with the major and minor axis given by Fw and FD. Figure 8.6a shows a vehicle with a circular uncertainty in position of 40 cm detecting a line segment. The ultrasonic readings are illustrated as circles with a radius determined by its uncertainty as defined in Fig. 8.5. The detected line segment is illustrated by a pair of parallel lines (the actual line segment can fall anywhere between that two lines, only uncertainties associated with sonar readings are considered here).

Page 150

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

Fig.8.6b shows the segment after the uncertainty in the robot's position has been added to the segment uncertainties. Fig.8.6c shows the uncertainty in position after correction by matching a model segment. The position uncertainty of the vehicle is reduced to an ellipse with a minor axis of approximately 8 cm (3.15 inches). In another experiment, the robot was place inside the simple environment shown in Fig. 8.7. Segment 0 corresponds to a wall covered with a textured wall paper. Segment 1 corresponds to a metal cabinet with a sliding plastic door. Segment 2 corresponds to a set of chairs pushed up against two tables. The robot system has no a-priori knowledge of its environment. The location and orientation at which the system was started were taken as the origin and x-axis of the world coordinate system. After the robot has run three cycles of ultrasonic acquisition, both the estimated position and orientation of the vehicle were set to false values. Instead of the correct position x = 0, y = 0, and 2 = 0, the position was set to x = 0.10 m, y = 0.10 m, and the orientation was set to 5o. The uncertainty was set to a standard deviation of 0.2 m in x and y, with a uncertainty in orientation of 10o. The system was then allowed to detect the "wall"-segments around it. The resulting estimated position and covariance is listed in Table 8.3.

Segment 0
Figure 8.6: a) A vehicle with a position uncertainty of 40 cm (15.7 inches), as shown by the circle around the center point (cross), is detecting a line segment. b) The boundaries for the line segment grow after adding the uncertainty for the robot's position. c) After correction by matching the segment boundaries with a stored map-segment, the uncertainty of the robot's position is reduced to about 8 cm (3.15 inches) as shown by the ellipse squat ellipse around the robot's center (cross). Page 151

Robot

crowle y3.d s4, .wm f

Segment 2

Figure 8.7: Experimental setup for testing Crowley's map-matching method. Initially, the robot is intentionally set-off from the correct starting position.

Part II: Systems and Methods for Mobile Robot Positioning


Initial estimated position (with deliberate initial error) covariance x,y, = (0.100, 0.100, 5.0) 0.040 0.000 0.000 After match with segment 0 estimated position: covariance 0.000 0.040 0.000 0.000 0.000 100.0

Chapter 8: Map-based Positioning

x,y, = (0.102, 0.019, 1.3) 0.039 0.000 0.000 0.000 0.010 0.000 0.000 0.000 26.28

After match with segment 1 estimated position: covariance x,y, = (0.033, 0.017, 0.20) 0.010 0.000 0.000 0.000 0.010 0.000 0.000 0.000 17.10

Table 8.3: Experimental results with Crowley's map-matching method. Although initially placed in an incorrect position, the robot corrects its position error with every additional wall segment scanned.

8.3.2 Topological Maps for Navigation Topological maps are based on recording the geometric relationships between the observed features rather than their absolute position with respect to an arbitrary coordinate frame of reference. Kortenkamp and Weymouth [1994] defined the two basic functions of a topological map: 1. Place recognition With this function, the current location of the robot in the environment is determined. In general, a description of the place, or node in the map, is stored with the place. This description can be abstract or it can be a local sensory map. At each node, matching takes place between the sensed data and the node description. 2. Route selection With this function, a path from the current location to the goal location is found. Following are brief descriptions of specific research efforts related to topological maps. 8.3.2.1 Taylor [1991] Taylor, working with stereo vision, observed that each local stereo map may provide good estimates for the relationships between the observed features. However, because of errors in the estimates for the robot position, local stereo maps don't necessarily provide good estimates for the coordinates of these features with respect to the base frame of reference. The recognition problem in a topological map can be reformulated as a graph matching problem where the objective is to find
Page 152

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

a set of features in the relational map such that the relationships between these features match the relationships between the features on the object being sought. Reconstructing Cartesian maps from relational maps involves minimizing a nonlinear objective function with multiple local minimal. 8.3.2.2 Courtney and Jain [1994] A typical example of a topologi330 cal map-based approach is given by Courtney and Jain [1994]. In this Hallway Hallway work the coarse position of the ro312 323 325 327 335 bot is determined by classifying the 350 map description. Such classification allows to recognize the workspace region that a given map represents. 352 Using data collected from ten differ354 ent rooms and ten different doorways in a building (see Fig. 8.8), Courtney and Jain estimated a 94% 360 recognition rate of the rooms and a 98% recognition rate of the doorcollected from ten ways. Courtney and Jain concluded Figure 8.8: Based on datasets building, Courtney different rooms and ten different doorways in a and Jain estimate that coarse position estimation, or a 94% recognition rate of the rooms and a 98% recognition rate of place recognition, in indoor domains the doorways (adapted from [Courtney and Jain, 1994]). is possible through classification of grid-based maps. They developed a paradigm wherein pattern classification techniques are applied to the task of mobile robot localization. With this paradigm the robot's workspace is represented as a set of grid-based maps interconnected via topological relations. This representation scheme was chosen over a single global map in order to avoid inaccuracies due to cumulative dead reckoning error. Each region is represented by a set of multisensory grid maps, and feature-level sensor fusion is accomplished through extracting spatial descriptions from these maps. In the navigation phase, the robot localizes itself by comparing features extracted from its map of the current locale with representative features of known locales in the environment. The goal is to recognize the current locale and thus determine the workspace region in which the robot is present.
courtney.ds4, .wmf ,11/13/94

8.3.2.3 Kortenkamp and Weymouth [1993] Kortenkamp and Weymouth implemented a cognitive map that is based on a topological map. In their topological map, instead of looking for places that are locally distinguishable from other places and then storing the distinguishing features of the place in the route map, their algorithm looks for places that mark the transition between one space in the environment and another space (gateways). In this algorithm sonar and vision sensing is combined to perform place recognition for better accuracy in recognition, greater resilience to sensor errors, and the ability to resolve ambiguous places. Experimental results show excellent recognition rate in a well structured
Page 153

Part II: Systems and Methods for Mobile Robot Positioning



Chapter 8: Map-based Positioning

environment. In a test of seven gateways, using either sonar or vision only, the system correctly recognized only four out of seven places. However, when sonar and vision were combined, all seven places were correctly recognized. Fig. 8.9 shows the experimental space for place recognition. Key-locations are marked in capital letters. Table 8.4a and Table 8.4b show the probability for each place using only vision and sonar, respectively. Table 8.4c shows the combined probabilities (vision and sonar) for each place.

G F

Figure 8.9: An experiment to determine if the robot can detect the same place upon return at a later time. In this case, multiple paths through the place can be "linked" together to form a network (adapted from [Kortenkamp and Weymouth, 1994]). Page 154

Part II: Systems and Methods for Mobile Robot Positioning


Stored Places A A B C D E F G 0.43 0.05 0.10 0.14 0.14 0.14 0.14 B 0.09 0.52 0.12 0.05 0.14 0.14 0.14 C 0.22 0.21 0.36 0.24 0.14 0.14 0.14 D 0.05 0.06 0.2 0.43 0.14 0.16 0.14 E 0.05 0.05 0.04 0.05 0.14 0.14 0.14 F 0.1 0.05 0.13 0.04 0.14 0.14 0.14

Chapter 8: Map-based Positioning

G 0.06 0.05 0.04 0.05 0.14 0.14 0.14

Table 8.4a: Probabilities for each place using only vision. Stored Places A A B C D E F G 0.82 0.02 0.02 0.02 0.04 0 0 B 0.04 0.31 0.31 0.31 0.12 0 0 C 0.04 0.31 0.31 0.31 0.12 0 0 D 0.04 0.31 0.31 0.31 0.12 0 0 E 0.04 0.06 0.06 0.61 0.61 0 0 F 0 0 0 0 0 0.90 0.10 G 0 0 0 0 0 0.10 0.90

Table 8.4b: Probabilities for each place using only sonar. Stored Places A A B C D E F G 0.95 0 0 0.01 0.04 0 0 B 0.01 0.65 0.17 0.07 0.12 0 0 C 0.02 0.26 0.52 0.33 0.12 0 0 D 0.01 0.07 0.29 0.58 0.12 0 0 E 0.01 0.01 0.01 0.01 0.61 0 0 F 0 0 0 0 0 0.90 0.09 G 0 0 0 0 0 0.1 0.91

Table 8.4c: Combined probabilities (vision and sonar) for each place.

In spite of the good results evident from Table 8.4c, Kortenkamp and Weymouth pointed out several drawbacks of their system:

Page 155

Part II: Systems and Methods for Mobile Robot Positioning


+

Chapter 8: Map-based Positioning

the robot requires several initial, guided traversals of a route in order to acquire a stable set of location cues so that it can navigate autonomously acquiring, storing, and matching visual scenes is very expensive, both in computation time and storage space, the algorithm is restricted to highly structured, orthogonal environments.

8.4 Summary
Map-based positioning is still in the research stage. Currently, this technique is limited to laboratory settings and good results have been obtained only in well structured environments. It is difficult to judge how the performance of a laboratory robot scales up to a real world application. Kortenkamp and Weymouth [1994] noted that very few systems that are tested on real robots are tested under realistic conditions with more than a handful of places. We summarize relevant characteristics of map-based navigation systems as follows: Map-based navigation systems + are still in the research stage and are limited to laboratory settings; + have not been tested extensively in real world environments; + require a significant amount of processing and sensing capability; + need extensive processing, depending on the algorithms and resolution used; + require initial position estimates from dead-reckoning in order to limit the initial search for features to a smaller area. There are several critical issues that need to be developed further:
+ + + +

Sensor selection and sensor fusion for specific applications and environments. Accurate and reliable algorithms for matching local maps to the stored map. Good error models of sensors and robot motion. Good algorithms for integrating local maps into a global map.

Page 156

Part III:

References and "Systems-at-a-Glance" Tables

Grad. student Ulrich Raschke verifies the proper alignment of ultrasonic sensors. All three robots in this picture use 15o-angular spacing between the sensors. Many researchers agree that 15o spacing assures complete coverage of the area around the robot.

Page 157

Part III: References and "Systems at-a-Glance" Tables

References

References
1. Abidi, M. and Chandra, T., 1990, "Pose Estimation for Camera Calibration and Landmark Tracking." Proceedings of IEEE International Conference on Robotics and Automation, Cincinnati, OH, May 13-18, pp. 420-426. Acuna, M.H. and Pellerin, C.J., 1969, "A Miniature Two-Axis Fluxgate Magnetometer." IEEE Transactions on Geoscience Electronics, Vol. GE-7, pp 252-260. Adams, M. et al., 1994, "Control and Localisation of a Post Distributing Mobile Robot." 1994 International Conference on Intelligent Robots and Systems (lROS '94). Mnchen, Germany, September 12-16, pp. 150-156. Adrian, P., 1991, "Technical Advances in Fiber-Optic Sensors: Theory and Applications." Sensors, pp. 23-45, September. Agent, A., 1991, "The Advantages of Absolute Encoders for Motion Control." Sensors, pp. 19-24, April. Arkin, R.C., 1989, "Motor-Schema-Based Mobile Robot Navigation." International Journal of Robotics Research, Vol. 8., No. 4, pp. 92-112, August. Aronowitz, F., "The Ring Laser Gyro," Laser Applications, Vol. 1, M. Ross, ed., Academic Press, 1971. Arradondo-Perry, J., 1992, "GPS World Receiver Survey." GPS World, January, pp. 46-58. Atiya, S. and Hager, G., 1993, "Real-time Vision-based Robot Localization." IEEE Transactions on Robotics and Automation, Vol. 9, No. 6, pp. 785-800.

2.

3.

4.

5.

6.

7.

8. 9.

Page 159

Part III: References and "Systems at-a-Glance" Tables

References

10.

Aviles, W.A. et al., 1991, "Issues in Mobile Robotics: The Unmanned Ground Vehicle Program Teleoperated Vehicle (TOV)" Proceedings of the SPIE - The International Society for Optical Engineering, Vol: 1388 p. 587-97. Avolio, G., 1993, "Principles of Rotary Optical Encoders." Sensors, pp. 10-18, April. Baker, A., 1993, "Navigation System Delivers Precision Robot Control." Design News, Dec. 20, p. 44. Banzil, G., et. al., 1981, "A Navigation Subsystem Using Ultrasonic Sensors for the Mobile Robot Hilare." Proceedings of 1st Conference on Robot Vision and Sensory Control, Stratford/Avon, U.K., 13 April. Barrett, C.R., Nix, W.D., and Tetelman, A.S., 1973, "The Principles of Engineering Materials." Prentice Hall, Englewood Cliffs, NJ. Barshan, B. and Durrant-Whyte, H. F., 1993, "An Inertial Navigation System for a Mobile Robot." Proceedings of the 1993 IEEE/RSJ International Conference on Intelligent Robotics and Systems, Yokohama, Japan, July 26-30, pp. 2243-2248. Barshan, B. and Durrant-Whyte, H. F., 1994, "Orientation Estimate for Mobile Robots Using Gyroscopic Information." 1994 International Conference on Intelligent Robots and Systems (lROS '94). Mnchen, Germany, September 12-16, pp. 1867-1874. Benayad-Cherif, F., Maddox, J., and Muller, L., 1992, "Mobile Robot Navigation Sensors." Proceedings of the 1992 SPIE Conference on Mobile Robots, Boston, MA, Nov. 18-20, pp. 378-387. Beyer, J., Jacobus, C., Pont, F., 1987, "Autonomous Vehicle Guidance Using Laser Range Imagery." SPIE Vol. 852, Mobile Robots II, Cambridge, MA, p. 34-43, November. Biber, C., Ellin, S., Shenk, E., 1987, "The Polaroid Ultrasonic Ranging System." Audio Engineering Society, 67th Convention, New York, NY, October-November. Binger, N, Harris, S.J., 1987, "Applications of Laser Radar Technology." Sensors, April, 1987, pp. 42-44. Boltinghouse, S., Burke, J., Ho, D., 1990, "Implementation of a 3D Laser Imager Based Robot Navigation System with Location Identification." SPIE Vol. 1388, Mobile Robots V, Boston, MA, pp. 14-29, November. Boltinghouse, S., Larsen, T., 1989, "Navigation of Mobile Robotic Systems Employing a 3D Laser Imaging Radar." ANS Third Topical Meeting on Robotics and Remote Systems, Section 2-5, pp. 1-7, Charleston, SC, March.
Page 160

11. 12.

13.

14.

15.

16.

17.

18.

19.

20.

21.

22.

Part III: References and "Systems at-a-Glance" Tables

References

23.

Bolz, R.E., Tuve, G.L., Editors,, 1979, CRC Handbook of Tables for Applied Engineering Science, CRC Press, Boca Raton, FL. Borenstein, J. and Koren, Y., 1985, "A Mobile Platform For Nursing Robots." IEEE Transactions on Industrial Electronics, Vol. 32, No. 2, pp. 158-165. Borenstein, J. and Koren, Y., 1986, "Hierarchical Computer System for Autonomous Vehicle." Proceedings of the 8th Israeli Convention on CAD/CAM and Robotics, Tel-Aviv, Israel, December 2-4. Borenstein, J., 1987, "The Nursing Robot System." Ph. D. Thesis, Technion, Haifa, Israel, June 1987, pp. 146-158. Borenstein, J. and Koren, Y., 1990, "Real-Time Obstacle Avoidance for Fast Mobile Robots in Cluttered Environments." IEEE International Conference on Robotics and Automation, Vol. CH2876-1, Cincinnati, OH, pp. 572-577, May. Borenstein, J. and Koren, Y., 1991, "The Vector Field Histogram  Fast Obstacle-Avoidance for Mobile Robots." IEEE Journal of Robotics and Automation, Vol. 7, No. 3., June 1991, pp. 278-288. Borenstein, J., 1992, "Compliant-linkage Kinematic Design for Multi-degree-of-freedom Mobile Robots ." Proceedings of the SPIE Symposium on Advances in Intelligent Systems, Mobile Robots VII, Boston, MA, Nov. 15-20, 1992, pp. 344-351. Borenstein, J., 1993, "Multi-layered Control of a Four-Degree-of-Freedom Mobile Robot With Compliant Linkage." Proceedings of the 1993 IEEE International Conference on Robotics and Automation, Atlanta, Georgia, May 2-7, pp. 3.7 - 3.12. Borenstein, J., 1994a, "The CLAPPER: A Dual-drive Mobile Robot with Internal Correction of Dead-reckoning Errors." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 3085-3090. Borenstein, J., 1994b, "Internal Correction of Dead-reckoning Errors With the Smart Encoder Trailer." 1994 International Conference on Intelligent Robots and Systems (lROS '94). Mnchen, Germany, September 12-16, pp. 127-134. Borenstein, J., 1994c, "Control and Kinematic Design for Multi-degree-of-freedom Mobile Robots With Compliant Linkage." Accepted for publication in the IEEE Transactions on Robotics and Automation, May.

24.

25.

26.

27.

28.

29.

30.

31.

32.

33.

Page 161

Part III: References and "Systems at-a-Glance" Tables

References

34.

Borenstein, J., 1994d, "Four-Degree-of-Freedom Redundant Drive Vehicle With Compliant Linkage." Video Proceedings of the 1994 IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13. Borenstein, J. and Koren, Y., 1994, "Error Eliminating Rapid Ultrasonic Firing for Mobile Robot Obstacle Avoidance." To appear in the IEEE Transactions on Robotics and Automation, December. Borenstein, J., 1995, "IPEC: A new Dead-reckoning Method for the Operation of AGVs Without Guide Wires." To be presented at the International Robots & Vision Automation Conference, Detroit, MI, May 8-11. Borenstein, J. and Feng, L., 1995a, "Correction of Dead-reckoning Errors in Mobile Robots and AGVs." To be presented at the International Robots & Vision Automation Conference, Detroit, MI, May 8-11. Borenstein, J. and Feng, L., 1995b, "Measurement and Correction of Systematic Deadreckoning Errors in Mobile Robots." Submitted to the IEEE Transactions on Robotics and Automation, November 1994. Borenstein, J., Wehe, D., Feng, C., Koren, Y., 1995, "Mobile Robot Navigation in Narrow Aisles with Ultrasonic Sensors." Accepted for presentation at the ANS 6th Topical Meeting on Robotics and Remote Systems," Monterey, California, February 5-10. Borenstein, J., 1995V, Video, "The CLAPPER: A Dual-drive Mobile Robot With Internal Correction of Dead-reckoning Errors." Submitted for inclusion in the Video Proceedings of the 1995 IEEE International Conference on Robotics and Automation, Nagoya, Japan, May 21-27, 1995. Brooks, R., 1985, "Visual Map Making for a Mobile Robot." Proceedings of IEEE International Conference on Robotics and Automation, St. Louis, MO, March 25-28, pp. 824829. Brown, R.G., Hwang, P.Y.C., 1992, "Introduction to Random Signals and Applied Kalman Filtering." 2nd ed., New York, NY, John Wiley and Sons. Buchenberger, M., Jorg, K., and Puttkamer, E., 1993, "Laserradar and Sonar Based World Modeling and Motion Control for Fast Obstacle Avoidance of the Autonomous Mobile Robot MOBOT-IV." Proceedings of IEEE International Conference on Robotics and Automation, Atlanta, GA, May 10-15, pp. 534-540. Buholz, N., Chodorow, M., 1967, "Acoustic Wave Amplitude Modulation of a Multimode Ring Laser." IEEE Journal of Quantum Electronics, Vol. QE-3, No. 11, pp. 454-459, November.

35.

36.

37.

38.

39.

40.

41.

42.

43.

44.

Page 162

Part III: References and "Systems at-a-Glance" Tables

References

45. 46.

Bulkeley, D., 1993, "The Quest for Collision-Free Travel." Design News, October 4. Burns, W.K., Chen, C.L., Moeller, R.P., 1983, "Fiber-Optic Gyroscopes with Broad-Band Sources." IEEE Journal of Lightwave Technology, Vol. LT-1, p. 98. Byrd, J.S., DeVries, K.R., 1990, "A Six-Legged Telerobot for Nuclear Applications Development, International Journal of Robotics Research, Vol. 9, pp. 43-52, April. Byrne, R.H., 1993, "Global Positioning System Receiver Evaluation Results." Sandia Report SAND93-0827, Sandia National Laboratories, Albuquerque, NM, September. Byrne, R.H., Klarer, P.R., Pletta, J.B., 1992, "Techniques for Autonomous Navigation." Sandia Report SAND92-0457, Sandia National Laboratories, Albuquerque, NM, March. Cao, Z., Roning, J., and Hall, E., 1986, "Omnidirectional Vision Navigation Integrating Beacon Recognition with Positioning." Proceedings of the 1986 SPIE Conference on Mobile Robots, Cambridge, MA, Oct. 30-31, pp. 213-220. Carter, E.F., ed., 1966, Dictionary of Inventions and Discoveries, Crane, Russak, and Co., NY. Case, M., 1986, "Single Landmark Navigation by Mobile Robot." Proceedings of the 1986 SPIE Conference on Mobile Robots, Cambridge, MA, Oct. 30-31, pp. 231-237. Chao, S., Lim, W.L., Hammond, J.A., 1984, "Lock-in Growth in a Ring Laser Gyro." Proceedings, Physics and Optical Ring Gyros Conference, SPIE Vol 487, Snowbird, UT, pp. 50-57, January. Chen, S. and Tsai, W., 1991, "Determination of Robot Locations by Common Object Shapes." IEEE Transactions on Robotics and Automation, Vol. 7, No. 1, pp. 149-156. Chen, Y.D., Ni, J., and Wu, S.M., 1993, "Dynamic Calibration and Compensation of a 3D Lasar Radar Scanning System." IEEE International Conference on Robotics and Automation, Atlanta, GA, Vol. 3, May, pp. 652-664. Chenavier, F. and Crowley, J., 1992, "Position Estimation for a Mobile Robot Using Vision and Odometry." Proceedings of IEEE International Conference on Robotics and Automation, Nice, France, May 12-14, pp. 2588-2593. Chesnoy, J., 1989, "Picosecond Gyrolaser." Optics Letters, Vol 14, No. 18, pp. 990-992, September.

47.

48.

49.

50.

51.

52.

53.

54.

55.

56.

57.

Page 163

Part III: References and "Systems at-a-Glance" Tables

References

58.

Chow, W.W., Gea-Banacloche, J., Pedrotti, L.M., Sanders, V.E., Schleich, W., Scully, M.O., 1985, "The Ring Laser Gyro." Reviews of Modern Physics, Vol. 57, No. 1, January, pp. 61104. Christian, W.R., Rosker, M.J., 1991, "Picosecond Pulsed Diode Ring Laser Gyroscope." Optics Letters, Vol. 16, No. 20, October, pp. 1587-1589. Clark, R.R., 1994, "A Laser Distance Measurement Sensor for Industry and Robotics." Sensors, pp. 43-50, June. Cohen, C. and Koss, F., 1992, "A Comprehensive Study of Three Object Triangulation." Proceedings of the 1993 SPIE Conference on Mobile Robots, Boston, MA, Nov. 18-20, pp. 95-106. Congdon, C., Huber, M., Kortenkamp, D., et al, 1993, CARMEL Versus FLAKEY -- A Comparison of Two Winners. AI Magazine, Winter, pp. 49-56. Conrad, D.J., Sampson, R.E., 1990, "3D Range Imaging Sensors." in Traditional and NonTraditional Robotic Sensors, T.C. Henderson, ed., NATO ASI Series, Vol. F63, SpringerVerlag, pp. 35-47. Cooper, S., and Durrant-Whyte, H., 1994, "A Kalman Filter for GPS Navigation of Land Vehicles." 1994 International Conference on Intelligent Robots and Systems (lROS '94). Mnchen, Germany, September 12-16, 1994, pp. 157-163. Courtney, J. and Jain, A., 1994, "Mobile Robot Localization via Classification of Multisensor Maps." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 1672-1678. Cox, I.J., 1991, "Blanche - An Experiment in Guidance and Navigation of an Autonomous Mobile Robot." IEEE Transactions Robotics and Automation, 7(3), pp. 193-204. Crowley, J.L. and Reignier, P., 1992, "Asynchronous Control of Rotation and Translation for a Robot Vehicle." Robotics and Autonomous Systems, Vol. 10, 1992, pp. 243-251. Crowley, J., 1989, "World Modeling and Position Estimation for a Mobile Robot Using Ultrasonic Ranging." Proceedings of IEEE International Conference on Robotics and Automation, Scottsdale, AZ, May 14-19, pp. 674-680. DeCorte, C., 1994, "Robots Train for Security Surveillance." Access Control, pp. 37-38, June. D'Orazio, T., Capozzo, L., Ianigro, M., and Distante, A., 1993, "Model Based Vision System for Mobile Robot Position Estimation." Proceedings of the 1993 SPIE Conference on Mobile Robots, Boston, MA, Sept. 9-10, pp. 38-49.
Page 164

59.

60.

61.

62.

63.

64.

65.

66.

67.

68.

69. 70.

Part III: References and "Systems at-a-Glance" Tables

References

71.

D'Orazio, T., Distante, A., Attolico, G., Caponetti, L., and Stella, E., 1991, "A Vision Algorithm for Mobile Vehicle Navigation." Proceedings of the 1991 SPIE Conference on Mobile Robots, Boston, MA, Nov. 14-15, pp. 302-309. D'Orazio, T., Ianigro, M., Stella, E., Lovergine, F., and Distante, A., 1993, "Mobile Robot Navigation by Multi-Sensory Integration." Proceedings of IEEE International Conference on Robotics and Automation, Atlanta, GA, May 10-15, pp. 373-379. Dahlin, T., Krantz, D., 1988, "Low-Cost, Medium Accuracy Land Navigation System." Sensors, pp. 26-34, February. Depkovich, T., Wolfe, W., 1984, "Definition of Requirements and Components for a Robotic Locating System." Final Report No. MCR-83-669, Martin Marietta Aerospace, Denver, CO, February. Dibburn, U., Petersen, A., 1983, The Magnetoresistive Sensor - a Sensitive Device for Detecting Magnetic Field Variations. Electronic Components and Applications, Vol. 5, No. 3, June, 1983. Dodington, S.H., 1989, "Electronic Navigation Systems." Electronic Engineers Handbook, D. Christiansen and D. Fink, eds., 3rd edition, New York, McGraw Hill, pp. 76-95. Dunlap, G.D., Shufeldt, H.H., Duttons Navigation and Piloting, Naval Institute Press, pp. 557-579. Durieu, C., Clergeot, H., and Monteil, F., 1989, "Localization of a Mobile Robot with Beacons Taking Erroneous Data Into Account." Proceedings of IEEE International Conference on Robotics and Automation, Scottsdale, AZ, May 14-19, pp. 1062-1068. Duchnowski, L.J., 1992, "Vehicle and Driver Analysis with Real-Time Precision Location Techniques." Sensors, pp. 40-47, May. Elfes, A., 1987, "Sonar-Based Real-World Mapping and Navigation." IEEE Journal of Robotics and Automation, Vol. RA-3, No. 3, pp. 249-265. Elfes, A., 1989, "Using Occupancy Grids for Mobile Robot Perception and Navigation." Computer, June, pp. 46-57. Ellowitz, H.I., 1992, "The Global Positioning System." Microwave Journal, pp. 24-33, April. Engelson, S. and McDermott, D., 1992, "Error Correction in Mobile Robot Map Learning." Proceedings of IEEE International Conference on Robotics and Automation, Nice, France, May 12-14, pp. 2555-2560.
Page 165

72.

73.

74.

75.

76.

77.

78.

79.

80.

81.

82. 83.

Part III: References and "Systems at-a-Glance" Tables

References

84.

Evans, J. M., 1994, "HelpMate: An Autonomous Mobile Robot Courier for Hospitals." 1994 International Conference on Intelligent Robots and Systems (lROS '94). Mnchen, Germany, September 12-16, 1994, pp. 1695-1700. Everett, H.R., 1982, "A Computer Controlled Autonomous Sentry Robot." Masters Thesis, Naval Postgraduate School, Monterey, CA, October. Everett, H.R., 1985, "A Multi-Element Ultrasonic Ranging Array." Robotics Age, pp. 13-20, July. Everett, H.R., Gilbreth, G.A., Tran, T., and Nieusma, J.M., 1990, "Modeling the Environment of a Mobile Security Robot." Technical Document 1835, Naval Command Control and Ocean Surveillance Center, San Diego, CA, June. Everett, H.R., Gage, D.W., Gilbreath, G.A., Laird, R.T., Smurlo, R.P., 1994, "Real-World Issues in Warehouse Navigation." Proceedings SPIE Mobile Robots IX, Volume 2352, Boston, MA, 2-4 November. Everett, H. R., 1995, Sensors for Mobile Robots: Theory and Application, ISBN 1-56881-048-2, A. K. Peters, Ltd., Wellesley, MA. Ezekial, S., Arditty, H.J., Editors, Fiber Optic Rotation Sensors and Related Technologies, Proceedings of the First International Conference, MIT, Springer-Verlag, New York. Feng, L., Koren, Y., and Borenstein, J., 1994, "A Model-Reference Adaptive Motion Controller for a Differential-Drive Mobile Robot." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 3091-3096. Feng, L., Fainman, Y., and Koren, Y., 1992, "Estimate of Absolute Position of Mobile Systems by Opto-electronic Processor,'' IEEE Transactions on Man, Machine and Cybernetics, Vol. 22, No. 5, pp. 954-963. Fenn, R.C., Gerver, M.J., Hockney, R.L., Johnson, B.G., 1992, "Microfabricated Magnetometer Using Youngs Modulous Changes in Magnetoelastic Materials." SPIE Vol. 1694. Figueroa, J.F., Lamancusa, J.S., 1992, "A Method for Accurate Detection of Time of Arrival: Analysis and Design of an Ultrasonic Ranging System." Journal of the Acoustical Society of America, Vol. 91, No. 1, pp. 486-494, January. Figueroa, J.F., Doussis, E., Barbieri, E., 1992, "Ultrasonic Ranging System for 3-D Tracking of a Moving Target." 92-WA/DSC-3, Proceedings, Winter Annual Meeting, American Society of Mechanical Engineers, Anaheim, CA, November.

85.

86.

87.

88.

89.

90.

91.

92.

93.

94.

95.

Page 166

Part III: References and "Systems at-a-Glance" Tables

References

96.

Figueroa, J.F., Barbieri, E., 1991, "Increased Measurement Range Via Frequency Division in Ultrasonic Phase Detection Methods." Acustica, Vol. 73, pp. 47-49. Figueroa, J.F., Mahajan, A., 1994, "A Robust Navigation System for Autonomous Vehicles Using Ultrasonics." Control Engineering Practice, Vol. 2, No. 1, pp. 49-59. Fisher, D., Holland, J.M., Kennedy, K.F., 1994, "K3A Marks Third Generation SynchroDrive." [conf, loc, pages, month]. Fleury, S. and Baron, T., 1992, "Absolute External Mobile Robot Localization Using a Single Image." Proceedings of the 1992 SPIE Conference on Mobile Robots, Boston, MA, Nov. 1820, pp. 131-143.

97.

98.

99.

100. Fox, K., 1993, "Indoor Robots Starts Flying Blind." Science, Vol. 261, Aug. 6, pp. 685. 101. Fraden, J., 1993, AIP Handbook of Modern Sensors, ed., Radebaugh, R., American Institute of Physics, New York. 102. Frederiksen, T.M., Howard, W.M., 1974, "A Single-Chip Monolithic Sonar System." IEEE Journal of Solid State Circuits, Vol. SC-9, No. 6, December. 103. Fukui, I., 1981, "TV Image Processing to Determine the Position of a Robot Vehicle." Pattern Recognition, Vol. 14, pp. 101-109. 104. Getting, I.A., 1993, "The Global Positioning System," IEE Spectrum, pp. 36-47, December. 105. Geyger, W.A., 1957, Magnetic Amplifier Circuits, 2nd ed., McGraw-Hill, New York. 106. Gilbert, W., 1992, "De Magnete." 1600. (Translation: P.F. Mottelay, John Wiley, 1893.) 107. Gonzalez, J., Stentz, A., and Ollero, A., 1992, "An Iconic Position Estimator for a 2D Laser RangeFinder." Proceedings of IEEE International Conference on Robotics and Automation, Nice, France, May 12-14, pp. 2646-2651. 108. Gonzalez, R. and Wintz, P., 1977, "Digital Image Processing." Addison-Welsey: Reading, MA. 109. Gonzalez, J., Ollero, A., and Reina, A., 1994, "Map Building for a Mobile Robot Equipped with a 2D Laser Rangefinder." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 1904-1909. 110. Gothard, B.M., Etersky, R.D., Ewing, R.E., 1993, "Lessons Learned on a Low-Cost Global Navigation System for the Surrogate Semi-Autonomous Vehicle." Proceedings, SPIE Vol. 2058, Mobile Robots VIII, pp. 258-269.
Page 167

Part III: References and "Systems at-a-Glance" Tables

References

111. Gould, L., 1990, "Is Off-Wire Guidance Alive or Dead?" Managing Automation, pp. 38-40, May. 112. Gourley, C. and Trivedi, M., 1994, "Sensor Based Obstacle Avoidance and Mapping for Fast Mobile Robots." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 1306-1311. 113. Grenoble, B., 1990, "Sensor and Logic Form Digital Compass." Electronic Design News, pp. 228-229, 6 December. 114. Gunther, J., 1994, "Robot Asks, Where Am I?" Popular Science, Feb., pp. 32. 115. Hammond, W., 1993, "Smart Collision Avoidance Sonar Surpasses Conventional Systems." Industrial Vehicle Technology 93: Annual Review of Industrial Vehicle Design and Engineering, UK and International Press, pp. 64-66. 116. Harmon, S.Y., 1986, "USMC Ground Surveillance Robot (GSR): Lessons Learned." Mobile Robots, SPIE Vol. 727, Cambridge, MA, pp. 336-343. 117. Harris, J.C., 1994, "An Infogeometric Approach to Telerobotics," Proceedings, IEEE National Telesystems Conference, San Diego, CA, pp. 153- 156. 118. Henkel, S. L., 1987, "Optical Encoders: A Review." Sensors, pp. 9-12, September. 119. Henkel, S. L., 1994, "GMR Materials Advance Magnetic Field Detection." Sensors, p. 8, June. 120. Hine, A., 1968, Magnetic Compasses and Magnetometers, Adam Hilger Ltd., London. 121. Hinkel, R. and Knieriemen, T., 1988, "Environment Perception with a Laser Radar in a Fast Moving Robot." Symposium on Robot Control 1988 (SYROCO '88), October 5-7, Karlsruhe, West Germany, pp. 68.1 - 68.7 122. Holenstein, A., Muller, M., and Badreddin, E., 1992, "Mobile Robot Localization in a Structured Environment Cluttered with Obstacles." Proceedings of IEEE International Conference on Robotics and Automation, Nice, France, May 12-14, pp. 2576-2581. 123. Holland, J.M., 1983, Basic Robotics Concepts, Indianapolis, IN. Howard W. Sams, Macmillan, Inc.,

124. Holle, S., 1990, "Incremental Encoder Basics." Sensors, pp. 22-30, April.

Page 168

Part III: References and "Systems at-a-Glance" Tables

References

125. Hollingum, J., 1991, "Caterpillar make the earth move: automatically." The Industrial Robot, vol. 18, no. 2, pp. 15-18. 126. Hongo, T., Arakawa, H., Sugimoto, G., Tange, K., and Yamamoto, Y., 1987, "An Automated Guidance System of a Self-Controlled Vehicle." IEEE Transactions on Industrial Electronics, Vol. IE-34, No. 1, 1987, pp. 5-10. 127. Hoppen, P., Knieriemen, T., and Puttkamer, E., 1990, "Laser-Radar Based Mapping and Navigation for an Autonomous Mobile Robot." Proceedings of IEEE International Conference on Robotics and Automation, Cincinnati, OH, May 13-18, pp. 948-953. 128. Hurn, J., 1993, GPS, A Guide to the Next Utility, No. 16778, Trimble Navigation, Sunnyvale, CA, November. 129. Janet, J., Luo, R., Aras, C., and Kay, M., 1993, "Sonar Windows and Geometrically Represented Objects for Mobile Robot Self-Referencing." Proceedings of the 1993 IEEE/RSJ International Conference on Intelligent Robotics and Systems, Yokohama, Japan, July 26-30, pp. 1324-1331. 130. Jenkin, M., Milios, E., Jasiobedzki, P., Bains, N., and Tran, K., 1993, "Global Navigation for ARK." Proceedings of the 1993 IEEE/RSJ International Conference on Intelligent Robotics and Systems, Yokohama, Japan, July 26-30, pp. 2165-2171. 131. Kabuka, M and Arenas, A., 1987, "Position Verification of a Mobile Robot Using Standard Pattern." IEEE Journal of Robotics and Automation, Vol. RA-3, No. 6, pp. 505-516. 132. Kadonoff, M.B., 1985, "Navigation Techniques for the Denning Sentry." [where published?], December. 133. Kak, A., Andress, K., Lopez-Abadia, and Carroll, M., 1990, "Hierarchical Evidence Accumulation in the PSEIKI System and Experiments in Model-driven Mobile Robot Navigation." in Uncertainty in Artificial Intelligence, Vol. 5, Elsevier Science Publishers B. V., North-Holland, pp. 353-369. 134. Kaman, 1991, "Threat Array Control and Tracking Information Center, Product Literature, PM1691, Kaman Sciences Corporation, Colorado Springs, CO. 135. Kay, M. and Luo, R., 1993, "Global Vision for the Control of Free-Ranging AGV Systems." Proceedings of IEEE International Conference on Robotics and Automation, Atlanta, GA, May 10-15, pp. 14-19. 136. Keller, E., 1962, "[Title Missing]." Electronics, p. 40, October 5.

Page 169

Part III: References and "Systems at-a-Glance" Tables

References

137. Kenny, T.W., Waltman, S.B., Reynolds, J.K., Kaiser, W.J., 1991, "Micromachined Silicon Tunnel Sensor for Motion Detection." Applied Physics Letters, Vol. 58, No. 1, January. 138. Kerr, J.R., 1988, "Real Time Imaging Rangefinder for Autonomous Land Vehicles." SPIE Vol. 1007, Mobile Robots III, Cambridge, MA, pp. 349-356, November. 139. Kihara, M., Okada, T., 1984, "A Satellite Selection Method and Accuracy for the Global Positioning System." Navigation: Journal of the Institute of Navigation, Vol. 31, No. 1, pp. 8-20, Spring. 140. Kilough, S.M., Pin, F.G., 1992, "Design of an Omnidirectional Holonomic Wheeled Platform Prototype." Proceedings of the IEEE Conference on Robotics and Automation, Nice, France, pp. 84-90, May. 141. Kim, E.J., "Design of a Phased Sonar Array for a Mobile Robot." Bachelor's Thesis, MIT, Cambridge, MA, May, 1986.

142. King, S. and Weiman, C., 1990, "HelpMate Autonomous Mobile Robot Navigation System." Proceedings of the 1990 SPIE Conference on Mobile Robots, Boston, MA, Nov. 8-9, pp. 190198. 143. Klarer, P.R., 1988, "Simple 2-D Navigation for Wheeled Vehicles." Sandia Report SAND880540, Sandia National Laboratories, Albuquerque, NM, April. 144. Kleeman, L., 1992, "Optimal Estimation of Position and Heading for Mobile Robots Using Ultrasonic Beacons and Dead-reckoning." Proceedings of IEEE International Conference on Robotics and Automation, Nice, France, May 12-14, pp. 2582-2587. 145. Kleeman, L. and Russel, R., 1993, "Thermal Path Following Robot Vehicle: Sensor Design and Motion Control." Proceedings of the 1993 IEEE/RSJ International Conference on Intelligent Robotics and Systems, Yokohama, Japan, July 26-30, pp. 1319-1323. 146. Koenigsburg, W.D., 1982, "Noncontact Distance Sensor Technology." GTE Laboratories, Inc., 40 Sylvan Rd., Waltham, MA, 02254, pp. 519-531, March. 147. Komoriya, K. and Oyama, E., 1994, "Position Estimation of a mobile Robot Using Optical Fiber Gyroscope (OFG)." International Conference on Intelligent Robots and Systems (lROS '94). Munich, Germany, September 12-16, pp. 143-149. 148. Koper, J.G., 1987, "A Three-Axis Ring Laser Gyroscope," Sensors, pp. 8-21, March.

Page 170

Part III: References and "Systems at-a-Glance" Tables

References

149. Kortenkamp, D. and Weymouth, T., 1994, "Combining Sonar and Vision Sensing in the Construction and Use of Topological Maps for Mobile Robots." Submitted to IEEE Transactions on Robotics and Automation. 150. Krotkov, E., 1991b, "Mobile Robot Localization Using a Single Image." Proceedings of IEEE International Conference on Robotics and Automation, Sacramento, CA, April 9-11, pp. 978983. 151. Kwiatkowski, W., Tumanski, S., 1986, "The Permalloy Magnetoresistive Sensors - Properties and Applications." J. Phys. E: Sci. Instrum., Vol. 19, pp. 502-515. 152. La, W.H.T., Koogle, T.A., Jaffe, D.L., Leifer, L.J., 1981, "Microcomputer-Controlled Omnidirectional Mechanism for Wheelchairs." Proceedings, IEEE Frontiers of Engineering in Health Care, CH1621-2/81/0000-0326. 153. Langer, D., Thorpe, C., 1992, "Sonar Based Outdoor Vehicle Navigation and Collision Avoidance." IROS 92. 154. Langley, R.B., 1991, "The Mathematics of GPS." GPS World, pp. 45-49, July/August. 155. Lapin, B., 1992, "Adaptive Position Estimation for an Automated Guided Vehicle." Proceedings of the 1992 SPIE Conference on Mobile Robots, Boston, MA, Nov. 18-20, pp.82-94. 156. Larson, T.R., Boltinghouse, S., 1988, "Robotic Navigation Within Complex Structures." SPIE Vol. 1007, Mobile Robots III, Cambridge, MA, pp. 339-348, November. 157. Larsson, U., Zell, C., Hyyppa, K., and Wernersson, A., 1994, "Navigating an Articulated Vehicle and Reversing with a Trailer." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 2398-2404. 158. Lefevre, H.C., 1992, "The Interferometric Fiber-Optic Gyroscope." in Fiber Optic Sensors, Udd, E., Editor, Vol. CR44, SPIE Optical Engineering Press, Bellingham, WA, September. 159. Lenz, J.E., 1990, "A Review of Magnetic Sensors." Proceedings of the IEEE, Vol. 78, No. 6, June. 160. Leonard, J. and Durrant-Whyte, H. F., 1991, "Mobile Robot Localization by Tracking Geometric Beacons." IEEE Transactions on Robotics and Automation, Vol. 7, No. 3, pp. 376382. 161. Lewis, R.A., Johnson, A.R., 1977, "A Scanning Laser Rangefinder for a Robotic Vehicle." 5th International Joint Conference on Artificial Intelligence, pp. 762-768.

Page 171

Part III: References and "Systems at-a-Glance" Tables

References

162. MacLeod, E. and Chiarella, M., 1993, "Navigation and Control Breakthrough for Automated Mobility." Proceedings of the 1993 SPIE Conference on Mobile Robots, Boston, MA, Sept. 9-10, pp. 57-68. 163. Maddox, J., 1994, "Smart Navigation Sensors for Automatic Guided Vehicles." Sensors, pp. 48-50, April. 164. Maenaka, K., Ohgusu, T., Ishida, and M., Nakamura, T., 1987, "Novel Vertical Hall Cells in Standard Bipolar Technology, Electronic Letters, Vol. 23, pp. 1104-1105. 165. Maenaka, K., Tsukahara, M., and Nakamura, T., 1990, "Monolithic Silicon Magnetic Compass." Sensors and Actuators, pp. 747-750. 166. Magee, M. and Aggarwal, J., 1984, "Determining the Position of a Robot Using a Single Calibrated Object." Proceedings of IEEE International Conference on Robotics and Automation, Atlanta, GA, March 13-15, pp. 140-149. 167. Mahajan, A., 1992, "A Navigation System for Guidance and Control of Autonomous Vehicles Based on an Ultrasonic 3-D Location System." Masters Thesis, Mechanical Engineering Department, Tulane University, July. 168. Manolis, S., 1993, "Resolvers vs. Rotary Encoders For Motor Commutation and Position Feedback." Sensors, pp. 29-32, March. 169. Martin, G. J., 1986, "Gyroscopes May Cease Spinning." IEEE Spectrum, pp. 48-53, February. 170. Mataric, M., 1990, "Environment Learning Using a Distributed Representation." Proceedings of IEEE International Conference on Robotics and Automation, Cincinnati, OH, May 13-18, pp. 402-406. 171. McGillem, C. and Rappaport, T., 1988, "Infra-red Location System for Navigation of Autonomous Vehicles." Proceedings of IEEE International Conference on Robotics and Automation, Philadelphia, PA, April 24-29, pp. 1236-1238. 172. McPherson, J.A., 1991, "Engineering and Design Applications of Differential Global Positioning Systems (DGPS) for Hydrographic Survey and Dredge Positioning." Engineering Technical Letter No 1110-1-150, US Army Corps of Engineers, Washington, DC,, July 1. 173. Menegozzi, L.N., Lamb, W.E., 1973, "Theory of a Ring Laser." Physical Review A, Vol. 1, No. 4, pp. 2103-2125, October. 174. Mesaki, Y. and Masuda, I., 1992, "A New Mobile Robot Guidance System Using Optical Reflectors." Proceedings of the 1992 IEEE/RSJ International Conference on Intelligent Robots and Systems, Raleigh, NC, July 7-10, pp. 628-635.
Page 172

Part III: References and "Systems at-a-Glance" Tables

References

175. Miller, G.L. and Wagner, E.R., 1987, "An Optical Rangefinder for Autonomous Robot Cart Navigation." Proceedings of the Advances in Intelligent Robotic Systems: SPIE Mobile Robots II. 176. Moravec, H. P. and Elfes, A., 1985, "High Resolution Maps from Wide Angle Sonar." Proceedings of the IEEE Conference on Robotics and Automation, Washington, D.C., pp. 116-121. 177. Moravec, H. P., 1988, "Sensor Fusion in Certainty Grids for Mobile Robots." AI Magazine, Summer, pp. 61-74. 178. Motazed, B., 1993, "Measure of the Accuracy of Navigational Sensors for Autonomous Path Tracking." Proceedings, SPIE Vol. 2058, Mobile Robots VIII, pp. 240-249. 179. Murray, C., 1991, "AGVs Go Wireless." Design News, 6-3-91, pp. 27-28. 180. Nickson, P., 1985, "Solid-State Tachometry." Sensors, pp. 23-26, April. 181. Nishide, K., Hanawa, M., and Kondo, T., 1986, "Automatic Position Findings of Vehicle by Means of Laser." Proceedings of IEEE International Conference on Robotics and Automation, San Francisco, CA, Apr. 7-10, pp. 1343-1348. 182. Nitzan, D. et al. 1977, "The Measurement and Use of Registered Reflectance and Range Data in Scene Analysis." Proceedings of IEEE, vol. 65, no. 2, pp. 206-220, February. 183. Nolan, D.A., Blaszyk, P.E., Udd, E., 1991, "Optical Fibers", in Fiber Optic Sensors: An Introduction for Engineers and Scientists, E. Udd, Editor, John Wiley, New York, pp. 9-26. 184. Parish, D. and Grabbe, R., 1993, "Robust Exterior Autonomous Navigation." Proceedings of the 1993 SPIE Conference on Mobile Robots, Boston, MA, Sept. 9-10, pp. 280-291. 185. Patterson, M.R., Reidy, J.J., Rudolph, R.C., 1984, "Guidance and Actuation Systems for an Adaptive-Suspension Vehicle." Final Technical Report, Battelle Columbus Division, OH, AD#A139111, 20 March. 186. Pessen, D. W., 1989, "Industrial Automation." John Wiley & Sons, Inc., ISBN 0-471-60071-7. 187. Petersen, A., 1989, "Magnetoresistive Sensors for Navigation." Proceedings, 7th International Conference on Automotive Electronics, London, England, pp. 87-92, October. 188. Pin, F. G. and Killough, M., 1994, "A New Family of Omnidirectional and Holonomic Wheeled Platforms for Mobile Robots." IEEE Transactions on Robotics and Automation, Vol. 10, No. 4, August 1994, pp. 480-489.
Page 173

Part III: References and "Systems at-a-Glance" Tables

References

189. Pin, F. G. et al., 1989, "Autonomous Mobile Robot Research Using the Hermies-III Robot." IROS International Conference on Intelligent Robot and Systems, Tsukuba, Japan, Sept. 190. Pin, F. G., Watanabe, Y., 1993, "Using Fuzzy Behaviors for the Outdoor Navigation of a Car with Low-Resolution Sensors." IEEE International Conference on Robotics and Automation, Atlanta, Georgia, May 2-7, pp. 548-553. 191. Pletta, J.B., Amai, W.A., Klarer, P., Frank, D., Carlson, J., Byrne, R., 1992, "The Remote Security Station (RSS) Final Report." Sandia Report SAND92-1947 for DOE under Contract DE-AC04-76DP00789, Sandia National Laboratories, Albuquerque, NM, October. 192. Premi, K. S. and Besant, C. B.: "A Review of Various Vehicle Guidance Techiques That Can be Used by Mobile Robots or AGVS." A.G.V.S. and 16th IPA Conference, June. 193. Primdahl, F., 1970, "The Fluxgate Mechanism, Part I: The Gating Curves of Parallel and Orthogonal Fluxgates." IEEE Transactions on Magnetics, Vol. MAG-6, No. 2, June. 194. Primdahl, F., 1979, "The Fluxgate Magnetometer." J. Phys. E: Sci. Instrum., Vol. 12, pp. 241253. 195. Purkey, M., 1994, "On Target." Golf Magazine, pp. 120-121, May, 1994. 196. Raschke, U. and Borenstein, J., 1990, "A Comparison of Grid-type Map-building Techniques by Index of Performance." Proceedings of IEEE International Conference on Robotics and Automation, Cincinnati, CA, May 13-18, pp. 1828-1832. 197. Reister, D. B., 1991, "A New Wheel Control System for the Omnidirectional HERMIES-III Robot." Proceedings of the IEEE Conference on Robotics and Automation, Sacramento, California, April 7-12, pp. 2322-2327. 198. Reister, D. B. et al., 1991, "DEMO 89 The Initial Experiment With the HERMIES-III Robot." Proceedings of the 1991 IEEE Conference on Robotics and Automation Sacramento, California, April 1991, pp. 2562-2567. 199. Reister, D.B. and Unseren, M.A., 1992, "Position and Force Control of a Vehicle with Two or More Steerable Drive Wheels." Internal Report ORNL/TM-12193, Oak Ridge National Laboratories. 200. Reister, D.B. and Unseren, M.A., 1993, "Position and Constraint Force Control of a Vehicle with Two or More Steerable Drive Wheels." IEEE Transactions on Robotics and Automation.Vol. 9, No. 6, December, pp. 723-731.

Page 174

Part III: References and "Systems at-a-Glance" Tables

References

201. Rencken, W.D., 1993, "Concurrent Localization and Map Building for Mobile Robots Using Ultrasonic Sensors." Proceedings of the 1993 IEEE/RSJ International Conference on Intelligent Robotics and Systems, Yokohama, Japan, July 26-30, pp. 2192-2197. 202. Rencken, W.D., 1994, "Autonomous Sonar Navigation in Indoor, Unknown, and Undstructured Environments."1994 International Conference on Intelligent Robots and Systems (lROS '94). Mnchen, Germany, September 12-16, 1994, pp. 127-134. 203. Reunert, M.K., 1993, "Fiber Optic Gyroscopes: Principles and Applications." Sensors, pp. 3738, August, 1993. 204. Russel, A., Thiel, D., and Mackay-Sim, A., 1994, "Sensing Odour Trails for Mobile Robot Navigation." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 2672-2677. 205. Sabatini, A. and Benedetto, O., 1994, "Towards a Robust Methodology for Mobile Robot Localization Using Sonar." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 3142-3147. 206. Sagnac, G.M., 1913, "Lether lumineux demontre par leffet du vent relatif dether dans un interferometre en rotation uniforme." C.R. Academy of Science, 95, pp. 708-710, 1913. 207. Sammarco, J.J., 1994, "A Navigational System for Continuous Mining Machines." Sensors, pp. 11-17, January, 1994. 208. Sammarco, J.J., 1990, "Mining Machine Orientation Control Based on Inertial, Gravitational, and Magnetic Sensors." Report of Investigations 9326, US Bureau of Mines, Pittsburgh, PA, 1990. 209. Sanders, G.A., 1992, "Critical Review of Resonator Fiber Optic Gyroscope Technology." in Fiber Optic Sensors, Udd, E., Editor, Vol. CR44, SPIE Optical Engineering Press, Bellingham, WA, September, 1992. 210. Schaffer, G., Gonzalez, J., and Stentz, A., 1992, "Comparison of Two Range-based Pose Estimators for a Mobile Robot." Proceedings of the 1992 SPIE Conference on Mobile Robots, Boston, MA, Nov. 18-20, pp. 661-667. 211. Schiele, B. and Crowley, J., 1994, "A Comparison of Position Estimation Techniques Using Occupancy Grids." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 1628-1634. 212. Schultz, W., 1993, "Traffic and Vehicle Control Using Microwave Sensors." Sensors, pp. 3442, October, 1993.

Page 175

Part III: References and "Systems at-a-Glance" Tables

References

213. Schulz-DuBois, E.O., 1966, "Alternative Interpretation of Rotation Rate Sensing by Ring Laser." IEEE Journal of Quantum Electronics, Vol. QE-2, No. 8, pp. 299-305, August, 1966. 214. Siuru, B., 1994, "The Smart Vehicles Are Here." Popular Electronics, Vol. 11, No. 1, pp. 4145, January, 1994. 215. Stokes, K.W., 1989, "Remote Control Target Vehicles for Operational Testing." Association for Unmanned Vehicles Symposium, Washington, DC, July, 1989. 216. Stuart, W.F., 1972, "Earths Field Magnetometry, Reports on Progress in Physics, J.M. Zinman, Editor, Vol. 35, Part 2, pp. 803-881. 217. Stuck, E. R., Manz, A., Green, D. A., and Elgazzar, S., 1994, "Map Updating and Path Planning for Real-Time Mobile Robot Navigation."1994 International Conference on Intelligent Robots and Systems (lROS '94). Muenchen, Germany, September 12-16, pp. 753760. 218. Sugiyama, H., 1993, "A Method for an Autonomous Mobile Robot to Recognize its Position in the Global Coordinate System when Building a Map." Proceedings of the 1993 IEEE/RSJ International Conference on Intelligent Robotics and Systems, Yokohama, Japan, July 26-30, pp. 2186-2191. 219. Tai, S., Kojima, K., Noda, S., Kyuma, K., Hamanaka, K., Nakayama, T., 1986, "All-Fibre Gyroscope Using Depolarized Superluminescent Diode." Electronic Letters, Vol. 22, p. 546, 1986. 220. Takeda, T., Kato, A., Suzuki, T., and Hosoi, M., 1986, "Automated Vehicle Guidance Using Spotmark." Proceedings of IEEE International Conference on Robotics and Automation, San Francisco, CA, Apr. 7-10, pp. 1349-1353. 221. Talluri, R., and Aggarwal, J., 1993, "Position Estimation Techniques for an Autonomous Mobile Robot - a Review". In Handbook of Pattern Recognition and Computer Vision, chapter 4.4, pp. 769-801, World Scientific: Singapore. 222. Taylor, C., 1991, "Building Representations for the Environment of a Mobile Robot from Image Data." Proceedings of the 1991 SPIE Conference on Mobile Robots, Boston, MA, Nov. 14-15, pp. 331-339. 223. Tonouchi, Y., Tsubouchi, T., and Arimoto, S., 1994, "Fusion of Dead-reckoning Positions With a Workspace Model for a Mobile Robot by Bayesian Inference." International Conference on Intelligent Robots and Systems (lROS '94). Munich, Germany, September 12-16, 1994, pp. 1347-1354.

Page 176

Part III: References and "Systems at-a-Glance" Tables

References

224. Tsumura, T. and Hashimoto, M., 1986, "Positioning and Guidance of Ground Vehicle by Use of Laser and Corner Cube." Proceedings of IEEE International Conference on Robotics and Automation, San Francisco, CA, Apr. 7-10, pp. 1335-1342. 225. Tsumura, T., 1986, "Survey of Automated Guided Vehicle in Japanese Factory." Proceedings of IEEE International Conference on Robotics and Automation, San Francisco, CA, April 710, pp. 1329-1334. 226. Tsumura T., Fujiwara, N., Shirakawa, T. and Hashimoto, M.: "An Experimental System for Automatic Guidance of Roboted Vehicle Following the Route Stored in Memory." Proc. of the 11th Int. Symp. on Industrial Robots, Tokyo 1981, pp. 18-193. 227. Tsumura, T., Hashimoto, M., and Fujiwara, N., 1988, "A Vehicle Position and Heading Measurement System Using Corner Cube and Laser Beam." Proceedings of IEEE International Conference on Robotics and Automation, Philadelphia, PA, April 24-29, pp. 47-53. 228. Turpin, D.R., 1986, "Inertial Guidance: Is It a Viable Guidance System for AGVs?" 4th International Conference on AGVs (AGVS4), pp. 301-320, June, 1986. 229. Udd, E., 1985, "Fiberoptic vs. Ring Laser Gyros: An Assessment of the Technology." in Laser Focus/Electro Optics, December, 1985. 230. Udd, E., 1991, "Fiberoptic Sensors Based on the Sagnac Interferometer and Passive Ring Resonator." in Fiber Optic Sensors: An Introduction for Engineers and Scientists, E. Udd, Editor, John Wiley, New York, pp. 233-269, 1991. 231. Vaganay, J., Aldon, M.J., and Fournier, A., 1993a, "Mobile Robot Attitude Estimation by Fusion of Inertial Data." Proceedings of IEEE International Conference on Robotics and Automation, 1993, Atlanta, GA, May 10-15, pp. 277-282. 232. Vaganay, J., Aldon, M.J., and Fournier, A., 1993b, "Mobile Robot Localization by Fusing Odometric and Inertial Measurements." Fifth Topical Meeting on Robotics and Remote Systems, Knoxville, TN, Vol. 1, pp. 503-510, April, 1993. 233. Vuylsteke, P., Price, C.B., Oosterlinck, A., 1990, "Image Sensors for Real-Time 3D Acquisition, Part 1." ." in Traditional and Non-Traditional Robotic Sensors, T.C. Henderson, ed., NATO ASI Series, Vol. F63, Springer-Verlag, pp. 187-210, 1990. 234. Wax, S.I. and Chodorow, M., 1972, "Phase Modulation of a Ring-Laser Gyro - Part II: Experimental Results," IEEE Journal of Quantum Electronics, pp. 352-361, March, 1972. 235. Wienkop, U., Lawitzky, G., and Feiten, W., 1994, "Intelligent Low-cost Mobility." 1994 International Conference on Intelligent Robots and Systems (lROS '94). Muenchen, Germany, September 12-16, 1994, pp. 1708-1715.
Page 177

Part III: References and "Systems at-a-Glance" Tables

References

236. Wiley, C.M., 1964, "Navy Tries Solid-State Compass." Electronics, pp. 57-58, February 14, 1964. 237. Wilkinson, J.R., 1987, "Ring Lasers." Progress in Quantum Electronics, edited by Moss, T.S., Stenholm, S., Firth, W.J., Phillips, W.D., and Kaiser, W., Vol. 11, No. 1, Pergamon Press, Oxford, 1987. 238. Woll, J.D., 1993, "A Review of the Eaton VORAD Vehicle Collision Warning System." Reprinted from International Truck and Bus Meeting and Exposition, Detroit, MI, SAE Technical Paper Series 933063, ISSN 0148-7191, pp. 1-4, November, 1993. 239. Wong, A. and Gan, S., "Vision Directed Path Planning, Navigation, and Control for An Autonomous Mobile Robot." Proceedings of the 1992 SPIE Conference on Mobile Robots, Boston, MA, Nov. 18-20, pp. 352-360. 240. Woodbury, N., Brubacher, M., Woodbury, J.R., 1993, "Noninvasive Tank Gauging with Frequency-Modulated Laser Ranging." Sensors, pp. 27-31, September, 1993. 241. Wormley, S., 1994, "A Little GPS Background." Internet message, swormley@ thrl.cnde.iastate.edu to Newsgroup sci.geo.satellite-nav, 15 March, 1994. 242. Wun-Fogle, M., Savage, H.T., 1989, "A Tunneling-tip Magnetometer." Sensors and Actuators, Vol. 20, pp. 199-205.

Cited Product Information from Commercial Companies


243. ACUITY - Acuity Research, Stevens Creek Blvd, Cupertino, CA 95014, (408) 252-9639 244. ADL - Applied Design Laboratories, P. O. Box 2405, Grass Valley, CA 95945, Ph.: (916) 2728206 245. ANDREW Andrew Corporation, 10500 W. 153rd Street, Orland Park, IL 60462. 246. BENTHOS, Inc., Edgerton Drive, North Falmouth, MA 02556 247. CATERPILLAR, Product Literature, SGV-1106/91, Caterpillar Self Guided Vehicle Systems, Mentor, OH. 248. CONTROL - Control Engineering Company, 8212 Harbor Spring Road, Harbor Spring, MI 49740. 249. CYBERMOTION, 115 Sheraton Drive, Salem, VA 24153, 703-562-7626.
Page 178

Part III: References and "Systems at-a-Glance" Tables

References

250. CYBERWORKS, Camera Vision Robot Position and Slippage Control System, Cyberworks, Inc., 31 Ontario Street, Orillia, Ontario, L3V 6H1 Canada, 705-325-6110. 251. DINSMORE Instrument Company, Product Literature, 1490 and 1525 Magnetic Sensors, Flint, MI, 313-744-1330. 252. EATON Eaton-Kenway, Inc., 515 East 100 South, Salt Lake City, UT 84102 253. ESP, "ORS-1 Optical Ranging System." Product Literature, ESP Technologies, Inc., Lawrenceville, NJ 08648, 609-275-0356. 254. GPS Report, 1992, Phillips Business Information, Potomac, MD, November. 255. GREYHOUND, 1994, "Accident Rate Keeps Falling." Greyhound Connections, Vol. 4, No. 2, March/April. 256. GYRATION Gyration, Inc., 12930 Saratoga Ave., Building C, Saratoga, CA 95070. 257. HP - Hewlett Packard, "Optoelectronics Designer's Catalog, 1991-1992." 258. HTI - Harris Technologies, Inc., Clifton, VA, 703-266-0904. 259. ILC Data Device Corporation, 1982, "Synchro Conversion Handbook", Bohemia, NY. 260. ISI - Intelligent Solutions, Inc., EZNav Position Sensor, One Endicott Avenue, Marblehead, MA 01945, 508-664-5655. 261. ISR - IS Robotics, Inc., RR-1/BS-1 System for Communications and Positioning. Preliminary Data Sheet." IS Robotics, Twin City Office Center, Suite 6, 22 McGrath Highway, Somerville, MA 02143, Ph.: 617-629-0055. 262. KVH Industries, C100 Compass Engine Product Literature, Middletown, RI, 401-847-3327. 263. MASSA, "E-201B & E-220B Ultrasonic Ranging Module Subsystems Product Selection Guide." Product Literature 891201-10M, Massa Products Corporation, Hingham, MA 02043, 617-749-4800. 264. MICRO-TRAK - Micro-Trak Systems, Inc., "Trak-Star Ultrasonic Speed Sensor." Product Information. P.O. Box 3699, Mankato, MN 56002, 507-257-3600. 265. MILLITECH, 1989, "Modular Millimeter Wave FMCW Sensor." Final Report for NSSC Contract No. N00024-88-C-5114, Millitech Corporation, South Deerfield, MA 01373, 413665-8551.
Page 179

Part III: References and "Systems at-a-Glance" Tables

References

266. MTI - MacLeod Technologies, Inc., Computerized Opto-electronic Navigation and Control (CONAC), 313 Littleton Road, Chelmsford, MA 01824., 508-250-4949. 267. MOTOROLA, Mini-Ranger Falcon, Product Literature, Motoroloa Government and Systems Technology Group, 8220 E. Roosevelt Road, PO Box 9040, Scottsdale, AZ 85252, 602-4417685. 268. NAMCO Controls, 7567 Tyler Blvd. Mentor, OH 44060, 1-800-NAMTECH (626-8324) 269. NASA, 1977, "Fast, Accurate Rangefinder, NASA Tech Brief, NPO-13460. 270. NATIONAL, 1991, "Electrostatic Transducers Provide Wide Range Ultrasonic Measurement." Linear Applications Handbook, pp. 1172-1173. National Semiconductor Corp., 2900 Semiconductor Drive, P.O. Box 58090, Santa Clara, CA 95052,408-721-5000. 271. NATIONAL, 1989, "LM1812 Ultrasonic Transceiver." Special Purpose Linear Devices Databook, Section 5, pp. 103-110. National Semiconductor Corp., 2900 Semiconductor Drive, P.O. Box 58090, Santa Clara, CA 95052, 408-721-5000. 272. NIKE, Dept. of Fitness Electronics, 9000 South-West Nimbus, Beaverton, Oregon 97005, 503644-9000. 273. PATENT1 Doppler-effect Vehicle Ground Speed Detecting Apparatus with Means for Adjusting Frequency of Generated Wave, US Patent #5287329, Issue year 1994. 274. POLAROID, 1981, "Polaroid Ultrasonic Ranging System Users Manual." Publication No. P1834B, Polaroid Corporation, 784 Memorial Drive, Cambridge, MA 02139, 617-386-3964. 275. POLAROID, 1991, "Ultrasonic Ranging System." Product Literature, Polaroid Corporation, 784 Memorial Drive, Cambridge, MA 02139, 617-386-3964. 276. POLAROID, 1990, "6500-Series Sonar Ranging Module." Product Specifications PID 615077,Polaroid Corporation, 784 Memorial Drive, Cambridge, MA 02139, 617-386-3964. 277. POLAROID, 1993, "Polaroid Ultrasonic Ranging Developers Kit." Publication No. PXW6431 6/93, Polaroid Corporation, 784 Memorial Drive, Cambridge, MA 02139, 617-3863964. 278. POLAROID, 1987, "Technical Specifications for Polaroid Electrostatic Transducer." 7000Series Product Specification ITP-64, Polaroid Corporation, 784 Memorial Drive, Cambridge, MA 02139, 617-386-3964.

Page 180

Part III: References and "Systems at-a-Glance" Tables

References

279. RIEGL 1994, "Laser Distance, Level, and Speed Sensor LD90-3." Product Data Sheet 3/94, RIEGL Laser Measurement Systems, RIEGL USA, 4419 Parkbreeze Court, Orlando, FL 32808, 407-294-2799. 280. SEO, 1991b, Schwartz Electro-Optics, Inc, "Scanning Laser Rangefinder." Product Literature, 3404 N. Orange Blossom Trail, Orlando, FL 32804, 407-298-1802. 281. SEO, 1991a, Schwartz Electro-Optics, Inc, "LRF-X Laser Rangefinder Series." Product Literature, Schwartz Electro-Optics, Inc., 3404 N. Orange Blossom Trail, Orlando, FL 32804, 407-298-1802. 282. SEO, 1991c, , Schwartz Electro-Optics, Inc., "Helicopter Optical Proximity Sensor System." Product Literature, 3404 N. Orange Blossom Trail, Orlando, FL 32804, 407-298-1802. 283. SEO, 1991d, Schwartz Electro-Optics, Inc, Process Report for US Army Contract DAAJ02-91C-0026, 3404 N. Orange Blossom Trail, Orlando, FL 32804, 407-298-1802. 284. SIMAN - Siman Sensors & Intelligent Machines Ltd., MTI-Misgav, D.N. Misgav 20179, Israel.

285. SPERRY - Sperry Marine Inc., Seminole Trail, Charlottesville, VA, 22901, Ph.: (804) 974-2000, Contact: Peter Arnold, Head of Engineering, ext. 2213 286. TRC - Transitions Research Corp., "TRC Light Ranger," Product Literature, Danbury, CT. 287. TRC - Transitions Research Corp., "Beacon Navigation System," Product Literature, Shelter Rock LaneDanbury, CT 06810, 203-798-8988. 288. UNIQUE -Unique Mobility, Inc., Product Literature, 425 Corporate Circle, Golden, CO 80401 303-278-2002. 289. VORAD Safety Systems, Inc., "The VORAD Vehicle Detection and Driver Alert System." Product Literature, 10802 Willow Ct, San Diego, CA 92127, 619-674-1450. 290. VRSS - Vehicle Radar Safety Systems, Inc., "Rashid Radar Safety Brake." Product Literature, Town Square Building, 10 South Gratiot, Mount Clemens, MI, 48043, 810-463-7883 291. WATSON - Watson Industries, Inc., Melby Rd., Eau Claire, WI 54703, 715-839-0628. 292. ZEMCO Group, Inc., 3401 Crow Canyon Road, Suite 201, San Ramon, CA 94583, 415-8667266.

Page 181

Part III: References and "Systems at-a-Glance" Tables

References

Relevant Research Articles, to be included in the 1995 Extended Positioning Report:

293. Adams, M., 1993, "Amplitude Modulated Optical Range Data Analysis in Mobile Robots." Proceedings of IEEE International Conference on Robotics and Automation, Atlanta, GA, May 10-15, pp. 8-13. 294. Chen, Q., Asada, M., and Tsuji, S., 1988, "A New 2-D World Representation System for Mobile Robots." Proceedings of IEEE International Conference on Robotics and Automation, Philadelphia, April 24-29, pp. 604-606. 295. Crowley, J., 1986, "Navigation for an Intelligent Mobile Robot." IEEE Journal of Robotics and Automation, Vol. RA-1, No. 1, pp. 31-41. 296. Curran, A. and Kyriakopoulos, K., 1993, "Sensor-based Self-localization for Wheeled Mobile Robots." Proceedings of IEEE International Conference on Robotics and Automation, Atlanta, GA, May 10-15, pp. 8-13. 297. Drake, K., McVey, E., and Inigo, R., 1985, "Sensing Error for a Mobile Robot Using Line Navigation." IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. PAMI-7, No. 4, pp. 485-490. 298. Drake, K., McVey, E., and Inigo, R., 1987, "Experimental Position and Ranging Results for a Mobile Robot." IEEE Journal of Robotics and Automation, Vol. RA-3, No. 1, pp. 31-42. 299. Fennema, C., Hanson, A., Riseman, E., Beveridge, J., and Kumar, R., 1990, "Model-Directed Mobile Robot Navigation." IEEE Transactions on Systems, Man, and Cybernetics, Vol. 20, No. 6, pp. 1352-1369. 300. Harmon, S.Y., 1987, "The Ground Surveillance Robot (GSR): An Autonomous Vehicle Designed to Transit Unknown Terrain." IEEE Journal of Robotics and Automation, Vol. RA-3, No. 3, pp. 266-279. 301. Holcombe, W., Dickerson, S., Larsen, J., and Bohlander, R., 1988, "Advances in Guidance Systems for Industrial Automated Guided Vehicle." Proceedings of the 1988 SPIE Conference on Mobile Robots, Cambridge, MA, Nov. 10-11, pp. 288-297. 302. Huang, Y., Cao, Z., Oh, S., Katten, E., and Hall, E., 1986, "Automatic Operation for a Robot Lawn Mower." Proceedings of the 1986 SPIE Conference on Mobile Robots, Cambridge, MA, Oct. 30-31, pp. 344-354. 303. Kanbara, T., Miura, J., and Shirai, Y., 1993, "Selection of Efficient Landmarks for an Autonomous Vehicle." Proceedings of the 1993 IEEE/RSJ International Conference on Intelligent Robotics and Systems, Yokohama, Japan, July 26-30, pp. 1332-1338.
Page 182

Part III: References and "Systems at-a-Glance" Tables

References

304. Bhanu, B., Roberts, B., and Ming, J., 1990, "Inertial Navigation Sensor Integrated Motion Analysis for Obstacle Detection." Proceedings of IEEE International Conference on Robotics and Automation, Cincinnati, OH, May 13-18, pp. 954-959. 305. Kortenkamp, D., 1993, "Cognitive Maps for Mobile Robots: A Representation for Mapping and Navigation." Ph.D. Thesis, The University of Michigan. 306. Krotkov, E., 1991a, "Laser Rangefinder Calibration for a Walking Robot." Proceedings of IEEE International Conference on Robotics and Automation, Sacramento, CA, April 9-11, pp. 2568-2573. 307. Kuipers, B. and Byun, Y., 1988, "A Robust Qualitative Method for Robot Spatial Learning." The Seventh National Conference on Artificial Intelligence, pp. 774-779. 308. Kurazume, R. and Nagata, S., 1994, "Cooperative Positioning with Multiple Robots." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 1250-1257. 309. Lebegue, X. and Aggarwal, J., 1994, "Generation of Architectural CAD Models Using a Mobile Robot." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 711-717. 310. Levitt, T., Lawton, D., Chelberg, D., and Nelson, P., 1987, "Qualitative Navigation." Proc. DARPA Image Understanding Workshop, pp. 447-465. 311. Blais, F., Rioux, M., and Domey, J., 1991, "Optical Range Image Acquisition for the Navigation of a Mobile Robot." Proceedings of IEEE International Conference on Robotics and Automation, Sacramento, CA, April 9-11, pp. 2574-2580. 312. Lu, Z., Tu, D., Li, P., Hong, Z., and Wu, B., 1992, "Range Imaging Sensor for Auto - Vehicle Guidance Applying an Optical Radar." Proceedings of the 1992 SPIE Conference on Mobile Robots, Boston, MA, Nov. 18-20, pp. 456-465. 313. MacKenzie, P. and Dudek, G., 1994, "Precise Positioning Using Model-Based Maps." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 1615-1621. 314. Malik, R. and Polkowski, E., 1990, "Robot Self-Location Based on Corner Detection." Proceedings of the 1990 SPIE Conference on Mobile Robots, Boston, MA, Nov. 8-9, pp. 306-316. 315. Malik, R. and Prasad, S., 1992, "Robot Mapping with Proximity Detectors." Proceedings of the 1992 SPIE Conference on Mobile Robots, Boston, MA, Nov. 18-20, pp. 614-618.
Page 183

Part III: References and "Systems at-a-Glance" Tables

References

316. McGillem, C. and Rappaport, T., 1989, "A Beacon Navigation Method for Autonomous Vehicles." IEEE Transactions on Vehicular Technology, Vol. 38, No. 3, pp. 132-139. 317. McKendall, R., 1988, "Robust Fusion of Location Information." Proceedings of IEEE International Conference on Robotics and Automation, Philadelphia, PA, April 24-29, pp. 1239-1243. 318. McVey, E., Drake, K., and Inigo, R., 1986, "Range Measurements by a Mobile Robot Using a Navigation Line." IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. PAMI-8, No. 1, pp. 105-109. 319. Moravec, H. P., 1983, "The Stanford Cart and the CMU Rover." Proceeding of IEEE, Vol. 71, No. 7, pp. 872-884. 320. Ohya, A., Nagashima, Y., and Yuta, S., 1994, "Exploring Unknown Environment and Map Construction Using Ultrasonic Sensing of Normal Direction of Walls." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 485492. 321. Parker, K., 1993, "'Bots Struggle to Learn Basics." Manufacturing Systems, Oct. 12, pp. 13-14. 322. Partaatmadja, O., Benhabib, A., Sun, A., and Goldenberg, A., 1992, "An Electrooptical Orientation Sensor for Robotics." IEEE Transactions on Robotics and Automation, Vol. 8, No. 1, pp. 111-119. 323. Pears, N. and Probert, P., 1993, "An Optical Range Sensor for Mobile Robot Guidance." Proceedings of IEEE International Conference on Robotics and Automation, Altanta, GA, May 10-15, pp. 659-664. 324. Roth-Tabak, Y. and Jain, R., 1989, "Building an Environment Model Using Depth Information." Computer, June, pp. 85-90. 325. Roth-Tabak, Y. and Weymouth, T., 1990, "Environment Model for Mobile Robot Indoor Navigation." Proceedings of the 1990 SPIE Conference on Mobile Robots, Boston, MA, Nov. 8-9, pp. 453-463. 326. Safaee-Rad, R., Tchoukanov, I., Smith, K., and Benhabib, B., 1992, "Three-dimensional Location Estimation of Circular Features for Machine Vision." IEEE Transactions on Robotics and Automation, Vol. 8, No. 5, pp. 624-640. 327. Santos, V., Goncalves, J., and Vaz, F., 1994, "Perception Maps for the Local Navigation of a Mobile Robot: a Neural Network Approach." Proceedings of IEEE International Conference on Robotics and Automation, San Diego, CA, May 8-13, pp. 2193-2198.
Page 184

Part III: References and "Systems at-a-Glance" Tables

References

328. Schwind, G., 1994, "Controls Offer Non-Wire Guidance Without a Costly Premium." Material Handling Engineering, March, pp. 31. 329. Shertukde, H. and Bar-Shalom, Y., 1988, "Target Parameter Estimation in the Near Field with Two Sensors." IEEE Transactions on Acoustics, Speech, and Signal Processing, Vol. 36, No. 8, pp. 1357-1360. 330. Singh, K. and Fujimura, K., 1993, "Map Making by Cooperating Mobile Robots." Proceedings of IEEE International Conference on Robotics and Automation, Atlanta, GA, May 10-15, pp. 254-259. 331. Xu, H. and Chi, X., 1993, "Calibration and Parameter Identification for Laser Scanning Sensor." Proceedings of IEEE International Conference on Robotics and Automation, Atlanta, GA, May 10-15, pp. 665-670. 332. Yoo, J. and Sethi, 1991, "Mobile Robot Localization with Multiple Stationary Cameras." Proceedings of the 1991 SPIE Conference on Mobile Robots, Boston, MA, Nov. 14-15, pp. 155-170. 333. Zheng, J., Barth, M., and Tsuji, S., 1991, "Autonomous Landmark Selection for Route Recognition by a Mobile Robot." Proceedings of IEEE International Conference on Robotics and Automation, Sacramento, CA, April 9-11, pp. 2004-2009. 334. Bourbakis, N., 1988, "Design of an Autonomous Navigation System." IEEE Control Systems Magazine, Oct., pp. 25-28.

Page 185

Part III: References and "Systems at-a-Glance" Tables

References

Systems-at-a-Glance

Page 186

Systems-at-a-Glance Tables and "Systems at-a-Glance" Tables Part III: References


Name Computer On-board Equipment Accuracy -position [mm] Accuracy orientation [o ] S a mp l i n g Rate [Hz] Features

De

General

0.01%-5% of traveled distance

100-10,000 or analog

Error accumulation

TRC Labmate

486-33MHz

Each quad-encoder pulse corresponds to 0.012 mm wheel displacement Drive and steer encoders

44 m bi-directional square path2: 310 mm

on smooth concrete1 : 6o With 10 bumps1 : 8o on smooth concrete1 :1-3.8o With 10 bumps1: 4o

Very high ~ 1 KHz

Short wheelbase

Cybermotion

Onboard proprietory

44 m bi-directional square path1 : 63 mm

Synchro-drive design

Blanche

MC68020

Uses a pair of knifeedge nonload-bearing wheels for odometry Wheel encoders and sonars for orientation measurements Two cooperative robots, one moves and one stays still and measures the motion of the moving one Average after a 22m square path: 20 mm Simulation: 8 mm after 100 m movement at 2m step Average after 22m square path: 0.5 20 Hz Can only compensate for systematic error Capable of maintaining good position estimate over long distance on smooth concrete1: 22o With 10 bumps1: 0.4o 25 Hz Capable of compensating for random disturbance

Model-reference adaptive motion control Multiple robots

386-20 MHz TRC Labmate

CLAPPER: Dual-drive robot with Internal Correction of Deadreckoning

486-33MHz

2 TRC Labmates, connected by a compliant linkage. Two absolute rotary encoders, one linear encoder

44 m square path: no bumps: 22 mm With 10 bumps1: 44 mm

This result is based on running the University of Michigan Benchmark (UMBmark) test for dead-reckoning accuracy. This test is des of this report [Borenstein and Feng, 1994]. Page 187

Systems-at-a-Glance Tables and "Systems at-a-Glance" Tables Part III: References


Name Computer On-board Equipment Accuracy -position [mm] Accuracy orientation [o ] S a mp l i n g Rate [Hz] Features

De

Fluxgate Magnetometer

1 - 4

10-1000 or analog

External, global, $1002000 Prone to magnetic disturbance Internal, local, $1K20K.

Angular Rate Gyro (Laser or Optical Fiber) Radar Velocimeter (Doppler) Filtered/Inertial Sensor Suite (direction gyros and accelerometer based) MiniRover MKI Underwater vehicle RS232 interface

Very accurate models available at $1K-5K. Problems are time dependent drift, and minimum detectable rate of rotation. Gyro will not "catch" slow rotation errors. 0.01%-5% of full scale rate

0.01%-5% of full scale rate.

10-1000 or analog

100-1000 or analog

Internal, local, $1K10K

0.01%-5% of distance traveled, also time dependent drift

10-1000 or analog

Internal, local, $3K$150K+

Flux gate magnetic sensor

Accuracy: 2% max. Resultion: 2 Drift: 0.005/sec minimum detectable rotation rate: 0.02/sec Drift: 1/min (actual road test, with software compensation) 10Hz $1000

Autogyro

Gyration Gyro

RS232 interface VCC-2 Vehicle control computer Solid state gyroscope, position code reader Position codes (landmarks)

$1000

Non-Wire Guidance System for AGV's

Page 188

Systems-at-a-Glance Tables and "Systems at-a-Glance" Tables Part III: References


Name Computer On-board Equipment Accuracy -position [mm] Accuracy orientation [o ] S a mp l i n g Rate [Hz] Features

De

Inertial navigation system

3 ENV-O5S Gyrostar rate gyros by Murata 1 triaxial linear accelerometer 2 inclinometers

Position drift rate 1-8 cm/s depending on the freq. of acceleration change

Gyro drift 11/min. After compensation: drift 0.75/min

100-1000 or analog

Internal, global

Page 189

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer On-board Components Stationary Components Accuracy - position [mm] Accuracy orientation [o ] Sampling rate [Hz]

Beacon Position
Features

CONAC (computerized Opto-Electronic Navigation and Control) ROBOSENSE

486-33MHz

Structured Opto-electronic Acquisition Beacon (STROAB) Scanning laser rangefinder

Networked Opto-electronic Acquisition Datum (NOAD) Retro-reflective targets

Indoor 1.3 mm outdoor 5 mm

Indoor and outdoor 0.05

25 Hz

3D - At least 3 NOADS for one acre. Networkable for unlim. area 2D - Measure both angle and distance to target Derives distance from computing time of sweep over target of known width Currently limited to single work area of 8080' 2D

System measures direction and distance to beacons with accuracy <0.17 and <20 mm, respectively. Accuracy for robot location and orientation not specified. Angular accuracy is within 0.05 percent with a resolution of 0.006o . Accuracy for robot location and orientation not specified.

10-40 Hz

NAMCO LASERNET beacon tracking system

RS-232 serial interface provided

Rotating mirror pans a near-infrared laser beam through a horizontal arc of 90o Rotating mirror for scanning laser beam Laser scanner

Retro-reflective targets of known dimensions

20 Hz

TRC Beacon Navigation System LASERNAV

6808 integrated computer, RS232 interface 64180 microcomputer

Retro-reflective targets, usually mounted on standalone poles

Resolution is 120 mm (4-3/4 inches) in range and 0.125 o in bearing for full 360 o coverage in a horizontal plane 1" moving at 2'/sec; 0.5" stationary 0.03.

1 Hz

Reflective targets

90 Hz

EZNAV (Probably identical to LASERNAV, above)

Laser-based scanning beacon.

Retro-reflective bar codes. Up to 32 can be distinguished..

Angular beacon measuring accuracy = 0.03 o. Accuracy for robot location and orientation not specified.

10 Hz

Measures only angles to reflectors. Can distinguish between multiple barcoded reflectors. 2D

LASERNAV

64180 microcomputer

Laser scanner

Reflective target

1" moving at 2'/sec; 0.5" stationary

0.03.

90 Hz

Page 190

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer On-board Components Stationary Components Accuracy - position [mm] Accuracy orientation [o ] Sampling rate [Hz]

Beacon Position
Features

BNS (Beacon Navigation System); 30.5 m

Optical IR detector (10 field of view in horizontal and vertical axes). 8086 Laser scanner

Infrared beacon transmitter (uniquely identifiable, 128 possible codes) Three corner cubes LN-10: 500 LN-20: 20 LN-30: 500 LN-40: 20

0.3 in the 5 central area and 1 out to the periphery of the sensitive area LN-10: 1 LN-20: 0.1 LN-30: 1 LN-40: 0.1

10 Hz

Laser Scanner + corner cubes

0.5 Hz

Laser scanner + Bar code Magnetic markers

Laser scanner

Bar coded target Magnetic markers buried under path (50' apart)

0.033 Hz

Page 191

Systems-at-a III: References and "Systems at-a-Glance" Tables Part Glance Tables
Name Computer On-board Components Stationary Components Accuracy position [mm] Accuracy orientation [o ] Sampling rate [Hz]

Beacon Navigati
Features

Three object triangulation

486-33MHz

Vision

Mean Error (I) x=234 y=225 (G) x=304 y=301 (N) x=17 y=17 (C) x=35 y=35 Two corner cube reflectors on both sides of the path 6 sonar beacons in a 12 m2 space Infrared beacons x=30 y= 2

Mean Error (I) 4.75 (G) 141.48 (N) 2.41 (C) 5.18

Mean time (I) 3048.7 (G) 3.8 (N) 33.5 (C) 4.8

Computer simulation for comparative study of four triangulation algorithms Accuracies are sensitive to landmark location

Laser Beam + Corner Cube Ultrasonic Beacons Infrared Beacons

8086

Four laser transceiver (transmitter and receiver) Eight sonar receiver array (45 deg apart) One optical infrared scanner

10 Hz

Measured standard dev. of path error of 40 mm 25 m2 test area, beacons (0,0), (5,0) and (5,4) worst error = 70 Inside DABC: Mean=57,..=25 Outside DABC: mean=140, .=156 On line AB or AC mean=74, .=57 Path error within 10mm, at 1m/s 100 with very noisy measurement 0.2

150 ms

Laser Scanner + corner cube

Z80

Laser scanner

Retro-reflector 4545m space, 3 reflectors at A(0,0),B(45,0), C(0,45) Retro-reflectors on the path Active beacon

Inside DABC: mean=0.07 .=0.06 Outside DABC: mean=0.13,.=0.16 On line AB or AC mean=0.12,.=0.05 10 Hz

Vision camera + retro-reflectors Three target triangulation

Vision camera + light source Detector

Optimize using all beacon data, reweighted least square criterion

Page 192

Systems-at-a III: References and "Systems at-a-Glance" Tables Part Glance Tables
Name Computer On-board Components Stationary Components Accuracy position [mm] Accuracy orientation [o ] Sampling rate [Hz]

Beacon Navigati
Features

Direction measure of several identical beacons

Laser scanner

strips of reflective tapes

At 0.3 m/s, error <2 cm At 1 m/s, is stable At 1.5 m/s, instable

Can navigate on wet rainy field, even when the drive wheels were spinning

Page 193

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer On-board Components Features used Accuracy position [mm] Accuracy orientation [o ] Sampling Rate [Hz] Features

Camera Vision Robot Position and Slippage Control System Absolute positioning using a single image

PC

Vision Camera

Rectangular ceiling lights, concentric circle Known pattern composed of coplanar points (IR diodes) Test pattern: 1.02.8 m. 84 uniformly distributed points Vertical edges matching using stored map

<100 mm

> 1 Hz

68030, 25MHz

Fixed Vision Camera (6m high) discretization 9.56.0mm for one pixel

Accuracy: mean=2,max:10 repeatability X: mean=0.7,max: 2 .= 0.8 Y: mean: 2 max: 5, std. 2 15 mm

Repeatability mean: 0.3 max: 0.7 std. 0.4

4 Hz

Can monitor robot o same time. 3D oper

Real-time Visionbased Robot Localization

Sun 4/280 computer Karlsruhe mobile robot (KAMRO) Sun workstation

780580 CCDcamera, f=8mm VISTA real-time image processing system 6404004b CCD camera, PC-EYE imaging interface Vision camera with fish-eye lens

0.1

2 Hz

Correspondence be landmarks and a sto bond on the localiza 2D operation. Sensitive at certain orientations

Robot localization using common object shapes Omnidirectional vision navigation with beacon recognition Vision algorithm for mobile vehicle navigation

Objects with a polygon-shaped top and a lateral surface perpendicular to the top A light array (3x3)

<5%

40 mm

0.3

TRC Labmate

Vision camera

Two sets of four coplanar points are necessary

7m distance 10%

Page 194

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer On-board Components Features used Accuracy position [mm] Accuracy orientation [o ] Sampling Rate [Hz] Features

Adaptive position estimation

Litton S-800 486 control MC68000 positioning Sun

Camera, strobe, landmark

Two circles of different radii

5 mm

Convergence 120 measurements

Guidance system using optical reflectors

Camera, strobe light (only on 0.3 sec)

Reflector pattern mounted on the ceiling 2m high

Positioning using a single calibrated object

Camera

A sphere with horizontal and vertical calibration great circles Corners of the room

5%

3D. Angle error inc circles approach the sphere. Distanc with the distance be and landmark

Model based vision system

TRC LabMate 68040

512x512 graylevel CCD camera, f=6mm

100 mm middle error 2%

3D. Orientation erro corner is in the cen Large error when co center and angle co R are too small

Pose estimation

9200 image processor

Fairchild 3000 CCD camera (256256), f=13mm Perceptics

Quadrangular target
s12=77.5,s13=177.5 s14=162,s23=160 s24=191,s34=104

at 1500mm 11 mm

at 1500 mm 1.5.

3D. Volume measu dra composed of fe extracted from an a gular target and the Errors increase with increasing distance, angle between landmark and camera too small or too large

Positioning using standard pattern

Relative displacement pattern: circle, half white & half black Identification pattern: bar code

at 5000 mm 2.2%

tLarges error:2

Page 195

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer On-board Components Features used Accuracy position [mm] Accuracy orientation [o ] Sampling Rate [Hz] Features

TV image processing for robot positioning

Diamond shape, 90 angle and 23 cm each side

at 4000 mm 70mm

at 4000 mm 2

90 sec. processing time

2D

Single landmark navigation

ARCTEC Gemini robot

Infrared detector (angular resolution 4 deg)

Infrared beacons

at 4000 mm 400 mm at 2400 mm 200 mm

2D, Error increases with the increase of distance between the vehicle and beacon 30 Hz 2D, The result is the fusion of dead-reckoning and observed Large range over which obstacles can be detected, allows global path planning

Robot positioning using optoelectronic processor Global vision

386 PC Image-100 image processing board

256256 Camera, f=16mm Hough Transform filter (128128) Camera mounted at fixed points in the environment Sony CCD camera, f=8.5mm resolution = 0.12/pixel at image center

Circle (R=107mm)

at 2000 mm 35 mm

Robot localization using a single image

Vertically oriented parts of fixed objects, e.g., doors, desks and wall junctions. Stored map.

min distance to landmark 1000 mm 0.2

2D

Page 196

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer & Robot On-board Components Maps and Features Accuracy position [mm] Accuracy orientation [o] Sampling Rate [Hz] Features

Map-based P

E R

Scanning Laser Rangefinder

0.5%-5%

1KHz10KHz or analog 100-1000 or analog

External, local, $10K-100K

Scanning IR Rangefinder

1%-10%

External, local, $5K20K

Scanning (or arrayed) Ultrasonic Rangefinder Visual

1%-10%

1-100

External, local, $100-5K

1%-20%

0.1-100

External, local, $500-50K

Navigation by multi-sensory integration

TRC Labmate

Cohu CCD Camera, f=16mm dead-reckoning

Integrates position estimates from vision system with odometry using Kalman filter framework Utilize heterogeneous info from laserradar and sonars

Laserradar and sonar based world modeling

Tricycle robot

24 sonars. 4 laser rangefinders, rotate at 360o/s, each scan 720 range points Vision Camera Doors, columns 5.0 cm 2.0 2 Hz Convex and concave polygons

Vision Directed Navigation

Sun Sparc for vision, MicroVAX as host, ROBMAC100 tricycle type vehicle

3D

Page 197

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer & Robot On-board Components Maps and Features Accuracy position [mm] Accuracy orientation [o] Sampling Rate [Hz] Features

Map-based P

E R

Robot localization by tracking geometric beacons Position estimation using vision and odometry

Sun-3 for localization Sun-4 vehicle control Differentialdrive vehicle 386 PC

One rotating sonar or Six fixed Sonars

Geometric beacon - naturally occurring environment feature Vertical edges and stored map 40 mm 0.5

1 Hz

EKF utilizes matches between observed geometric beacons and a priori map of beacon locations 2D - Realistic odometry model and its uncertainty is used to detect and calculate position update fused with observation

756x581 CCD camera f=12.5 mm

E K c c t tw s m e

Recognize world location with stereo vision Environment learning using a distributed repre-sentation Omnidirectional 3wheeled base

Stereo cameras

Long, near vertical stereo features

1000 real world data recognition test, under 10% false negative, zero false positive dynamic landmark detection utilizing robot's motion

L f o a f

a ring of 12 sonars and a compass

Left wall, right wall, corridors

l s o b p t

Localization in structured environment

Motorola M68020

A ring of 24 sonars

Classify objects into edges, corners, walls, and unknown objects

0.1 Hz

Positions resulting from all possible mappings are calculated and then analyzed for clusters. The biggest cluster is assumed to be at the true robot pos.

E o o t o s t

Page 198

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer & Robot On-board Components Maps and Features Accuracy position [mm] Accuracy orientation [o] Sampling Rate [Hz] Features

Map-based P

E R

Localization using sonar

SUN 4

Linear array of three sonars: A. reduce the angular uncertainty, B. help identify the target's class Sonars

Local map: feature map (extended reflectors, e.g., wall, and point reflectors) Probability based occupancy grid map

<10 mm

<1

Local map: feature extrac Matching: least square. EKF for estimating the g rameters of different targ uncertainty

Sonar-based real-world mapping

Neptune mobile robot

Map with 3000 6" cells made from 200 well spaced readings of a cluttered 20'20' room can be matched with 6" displacement and 3 o rotation in 1 sec. of VAX time HIMM results in a sensor grid in which entries in close proximity to actual object locations have a a favorable (low) Index of Performance value Best result obtained by matching segment to segment 6" in path following
Position update every 8 sec. for a 180 points image and a map of 24 lines. 2D map.

map matching by convol gives the displacement an best brings one map into with the other, with a me goodness of match

Comparison of Grid-type map building by index of performance (IOP)

Cybermotion K2A synchrodrive robot 386 20MHz PC

A ring of 24 sonars

Histogramic In Motion Mapping (HIMM) & Heuristic Probability Function (HPF)

Index of performance (IO the correlation between t sition of objects, as comp map-building algorithm, object position, as measu manuallyThe IOP gives q measure of the difference grid maps produced by ea type Grid to seg. matching: generating a mask for the segment and correlating it with the grid map
(1) Least-square for data & model matching (2) Combine odometry and matching to better pos. estimate using maximum likelihood

Comparison of position estimation using occupancy grid Blanch MC68020 Tricycle type mobile robot Optical rangefinder, res.=1" at 5', 1000 samples/rev. in one sec. Odometry

Local map: grid map Global map: grid map 24 line segment environments map for a 300"x200" room

S m o B C

S A d b & s

Page 199

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer & Robot On-board Components Maps and Features Accuracy position [mm] Accuracy orientation [o] Sampling Rate [Hz] Features

Map-based P

E R

Range map pose estimation

SPARC1+

1D Laser range finder 1000 oints/rev

Line segment, corner

Mean error Feature-based: 60 Iconic estimator: 40 In a 1010m space

Max under 1.2 for both

Featurebased: 0.32 sec. Iconic: 2 sec.

1000 points/rev. Iconic approach matches every range data point to the map rather than condensing data into a small set of features to be matched to the map Classification of data points Weighted voting of correction vectors Kalman filter position estimation Line fitting Matching, only good matches are accepted

Positioning using modelbased maps

A rotatable ring of 12 polaroid sonars

Line segments

3-5 cm Coverge if initial estimate is within 1 m of the true pos. The variance never exceeds 6 cm

C s L

Positioning using optical range data

INMOS-T805 transputer

Infrared scanner

Line segment

W w e m te

World modeling and localization using sonar ranging

A ring of 24 sonars

Line segments

x=33 mm covariance: 1 y=17 mm covariance: 1

0.20 co-variance: 17.106

A model for the uncertainty in sonars, and the projection of range measurement into external Cartesian coordinate On SUN Sparc, 80 ms for local map building and 135 ms for global map update

Extracting line segments from adjacent co-linear range measurements and matching these line segments to a stored model Matching: remove segment already in the global map from local map, add new segment

M c t a c o t r

2D Laser rangefinder map building

Sun Sparc

Cyclone 2D laser rangefinder accuracy 20 cm, range 50m

Local map: line segment map Global map: line segments

Max. 5 cm average 3.8 cm

L C C m L

Page 200

Systems-at-a-Glance Part III: References and "Systems at-a-Glance" Tables


Name Computer & Robot On-board Components Maps and Features Accuracy position [mm] Accuracy orientation [o] Sampling Rate [Hz] Features

Map-based P

E R

Iconic position estimator

Locomotion Emulator, allwheel drive & all-wheel steer, Sun Sparc 1

Cyclone laser range scanner, res=10 cm range=50m 1000 readings per rev.

In general, has a large number of short line segments

Max. 36 mm mean 19.9 mm

Max. 1.8 mean 0.73

Iconic method works directly on the raw sensed data, minimizing the discrepancy between it and the model A graph where the nodes represent the observed features and edges represent the relationships between features

A p s m T m s m

Environment representation from image data

geometrical relationships between observed features rather than their absolute position Sonars Lateral motion vision Infrared proximity sensor Local map: multi-sensor 100x100 grid maps, cell 20x20 cm Using datasets from 10 rooms and hallways, estimate a 94% recognition rate for rooms, and 98% for hallways
Local grid maps Feature-level sensor fusion by extracting spatial descriptions from these maps

T p f g p

Localization via classification of multisensor maps

Positioning by classfying the map descriptions to recognize the workspace region that a given map represents

M n a M d

Page 201

Systems-at-a-Glance Tables and "Systems at-a-Glance" Tables Part III: References


Name Computer On-board Components Maps and Features Accuracy position [mm] Accuracy orientation [o] Sampling Rate [Hz] Features

Guide path Sensor (magnetic, optical, inductive, etc.) Odour Trails for Navigation Applicator for laying volatile chemicals on the floor Alfactory sensor Quartz halogen bulb and pyroelectric sensor

0.01-0.1 m

100-1000 or analog

External, local, or waypoint indication, $100-$5K

Thermal path following

0.833

No need to remove markers after use

Page 202

You might also like