0% found this document useful (0 votes)
15 views26 pages

BBGN Minor

Uploaded by

darlami939
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
15 views26 pages

BBGN Minor

Uploaded by

darlami939
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 26

NATIONAL COLLEGE OF ENGINEERING (NCE)

(Affiliated to Tribhuwan University)


Talchikhel, Lalitpur

Minor Project Proposal

On

“AUTONOMOUS ROBOT USING VSLAM FOR RESCUE AND

SERVEILLANCE”

SUBMITTED BY:

Bibek Shrestha (NCE-076-BEI-002)


Bishal Singh (NCE-076-BEI-004)
Gaurav Darlami (NCE-076-BEI-005)
Nabin Upadhayaya (NCE-076-BEI-006)

SUBMITTED TO:

“Department of Computer & Electronics Engineering”

AUGUST, 2022

LALITPUR, NEPAL

1
NATIONAL COLLEGE OF ENGINEERING (NCE)
(Affiliated to Tribhuwan University)
Talchikhel, Lalitpur

Minor Project Proposal

On

“AUTONOMOUS ROBOT USING VSLAM FOR RESCUE AND

SERVEILLANCE”

SUBMITTED BY:

Bibek Shrestha (NCE-076-BEI-002)

Bishal Singh (NCE-076-BEI-004)

Gaurav Darlami (NCE-076-BEI-005)

Nabin Upadhayaya (NCE-076-BEI-006)

THE MINOR PROJECT PROPOSAL SUBMITTED TO DEPARTMENT

SUBMITTED TO ELECTRONICS AND COMPUTER ENGINEERING TO THE

PARTIAL FULFILLMENT OF BACHELOR DEGREE IN ELECTRONICS

ENGINEERING.

AUGUST, 2022

LALITPUR, NEPAL

1
ii
ABSTRACT

Simultaneous Localization and Mapping (SLAM) technology is a well-study settlement.


Light Detection and Ranging (LIDAR) and camera sensors are commonly used for
localization and perception. However, through ten or twenty years of evolution, the
LIDAR-SLAM method does not seem to have changed much. Compared with the LIDAR
based schemes, the visual SLAM has strong scene recognition ability with the advantages
of low cost and easy installation. Indeed, people are trying to replace LIDAR sensors with
camera only, or integrating other sensors on the basis of camera in the field of autonomous
driving. Based on the current research situation of visual SLAM, this review covers the
visual SLAM technologies. In particular, we firstly illustrated the typical structure of visual
SLAM. Secondly, the state-of-the-art studies of visual and visual-based (i.e. visual-inertial,
visual-LIDAR, visual-LIDAR-IMU) SLAM is completely reviewed, as well the
positioning accuracy of our previous work is compared with the well-known frameworks
on the public datasets.

Keywords: self-driving cars; lane detection; scene detection; obstacle detection;


Visualization simultaneous localization and mapping (VSLAM) technology.

iii
2
TABLE OF CONTENTS

ABSTRACT....................................................................................................................... iii
TABLE OF CONTENTS ................................................................................................... iv
LIST OF FIGURES ............................................................................................................ v
LIST OF ABBREVIATIONS ............................................................................................ vi
1. INTRODUCTION ....................................................................................................... 1
1.1 Background .......................................................................................................... 1
1.2 Problem statement ................................................................................................ 2
1.3 Aims and Objectives ............................................................................................ 3
1.3.1 Aims .............................................................................................................. 3
1.3.2 Objectives ............................................................................................................... 3
1.4 Scope .................................................................................................................... 4
2. LITERATURE REVIEW ............................................................................................ 5
3. METHODOLOGY ...................................................................................................... 7
3.1 Diagram of Proposed System ............................................................................... 7
3.2 Tool used ............................................................................................................ 10
3.2.1 Hardware requirements..................................................................................... 10
3.2.2 Software requirements ...................................................................................... 10
3.3 SYSTEM REQUIREMENTS ............................................................................ 11
3.4 FEASIBILITY STUDY ..................................................................................... 14
4. EPILOGUE ................................................................................................................ 15
4.1 EXPECTED OUTPUT....................................................................................... 15
4.2 GANTT CHART ................................................................................................ 16
REFERENCE AND BIBLIOGRAPHY ........................................................................... 17
5.REFERENCES………………………………………………………………………17

3
iv i
LIST OF FIGURES

Figure 1: Proposed Diagram of Robot 10

Figure 2: Block Diagram of the Proposed system 11

Figure 3: Level of Autonomy 12

4v
LIST OF ABBREVIATIONS

ROS (Robot Operating System)


VSLAM (Visual Simulation Localization and Mapping)
SLAM (Simulation Localization and Mapping)
GPS (Global Positioning System)
DARPA (Defense Advanced Research Project Agency)
SAE (Society of Automotive Engineers)
LIDAR (Light Detection and Ranging)

5vi
1. INTRODUCTION

1.1 Background

The second edition of the autonomous vehicle “Grand Challenge” of the U.S. Defense
Advanced Research Projects Agency (DARPA), in 2005, and the Urban Challenge two
years later, revived interest by making the technology for self-driving cars seem within
reach. That stimulated technology companies to jump in notably Google, which launched
its program in 2009. Major programs have followed at giant, traditional carmakers
including General Motors, Ford, and Toyota. But what does “self-driving” really mean?
The Society of Automotive Engineers (SAE) defines five levels of automation, beyond
fully manual control (Level 0). These range from Level 1 (“feet off”) automation, typified
by cruise control; to Level 2 (“hands off”),systems such as Tesla’s “Autopilot,” which
assume the driver is poised to take immediate control; all the way to Level 4 (“mind off”),
which can stop and turn over control when needed, and completely automated, Level5
vehicles requiring no driver at all. Level 4 and Level 5 constitute the Holy Grail for
autonomous-vehicle development—and also the most difficult targets to reach.

Recently, the rapid development of the artificial intelligence has greatly promoted the
progress of unmanned driving ,such as self-driving cars, unmanned aerial vehicles and so
on.[1,2] Among these unmanned driving technologies, self-driving cars have attracted
more and more attention for their important economic effect. However, there are lots of
challenges in. More and more solutions based on VSLAM for self-driving cars have been
presented, including obstacle-detection, scene recognition, lane detection, and so on.

1
1.2 Problem statement

We focus on self-driving cars which are categorized as level 3 or above. The environment
perception system utilizes the prior knowledge of the environment to establish an
environmental model including obstacles, road structures, and traffic signs through
obtaining surrounding environmental information.

The main function of the environment perception system is to realize functions like lane
detection, traffic signal detection, and obstacle detection, by using some hardware devices
such as cameras and laser radars. The main function of the autonomous decision system is
to make some decisions for the self-driving car, including obstacle avoidance, path
planning, navigation, and so on.

For example, in the path planning, the autonomous decision system plans a global path
according to the current location and the target location firstly, then reasonably plans a
local path for the self-driving car by combining the global path and the local environment
information provided by the environment perception system.

2
1.3 Aims and Objectives

1.3.1 Aims

▪ To design and run an autonomous robot for rescue and surveillance using
VSLAM.

1.3.2 Objectives

▪ To visualize sensor data in Robot operating system visualization (Rviz.) and


interface real sensors with Robot Operating System (ROS).
▪ To use collected data for localization and mapping using visual simultaneous
localization and mapping (VSLAM).
▪ To test the robot for real-time purposes in rescue and surveillance.

3
1.4 Scope

Visual SLAM is still in infancy, commercially speaking .while it has enormous


potential in a wide range of setting, and it’s still an emerging technology. Accurately
projecting virtual image onto the physical world require a precise mapping of the
physical environment, only visual SLAM technology is capable of providing this level
of accuracy.

Visual SLAM systems are also used in a wide variety of field robots. For Example,
rover and Landers for exploring mars use visual SLAM systems to navigate
autonomously. Field robots in agriculture, as well as drones, can use the same
technology to independent travel around crop fields. Autonomous Vehicles could
potentially use visual SLAM system for mapping and understand world around them
for rescue and surveillance.

One major potential opportunity for Visual SLAM systems is to replace GPS tracking
and navigation in certain applications.GPS systems aren’t useful indoors or in big cities
where the view of the sky is obstructed, and they’re only accurate within a few meters.
Visual SLAM systems solve these problems as they aren’t dependent on satellite
information and they’re taking accurate measurement of the physical world around
them.

4
2. LITERATURE REVIEW

Kim et al. used SSD algorithm using LIDAR 3D for autonomous driving application. They
used RGB dataset. The method used is GFU SSD method which showed better result than
baseline SSD.

In [110], Kim et al. used the SSD algorithm for general object detection in the autonomous
driving applications. LIDAR 3D point clouds were converted into 2D images, and then
these images were used along with RGB images as inputs for two separate SSD networks.
Finally, gated fusion units (GFU) were used to assign selective weights to fuse both feature
maps produced by the two SSD networks through a feature fusion level. The experimental
results showed that the proposed GFU SSD method outperformed the baseline SSD [1]

Sighandhupe et. al proposed and reviewed SLAM algorithm for localization and mapping
in his paper presented in IEE conference. SLAM relies on vision-based sensors which
separate it from use of GPS, LIDAR system. However, the aforementioned system can also
be included in SLAM.

SLAM is an algorithm that combines a set of sensors to build a map of the AV and its
surroundings, while simultaneously keeping track of the vehicles current position in
reference to the built map. Although SLAM algorithms were initially applied in the field
of mobile robots, researchers have put a noticeable effort into adjusting the algorithms to
suit autonomous vehicle applications. This was done by taking into consideration different
key challenges, such as the need for faster processing, the outdoor lighting conditions, and
the dynamic road obstacles. It is important to point out that while SLAM mainly relies on
vision-based sensors, other sensors such as GPS, LIDAR, and sonar have also been used
to implement SLAM algorithms. Surveys on recent SLAM methods have been done. [2]

Hrisham Ali et. al in his book has proposed using DOT to improve accuracy in dynamic
environments. The book also proposes edge detection in SLAM for localization and
mapping.

5
Localization and navigation play a key role in many location-based services and have
attracted numerous research efforts from both academic and industrial community.
However, the ever-growing computation resource demanded by SLAM impedes its
application to resource-constrained mobile devices. The design, implementation, and
evaluation of edge SLAM, an Ed assisted real-time semantic visual SLAM service running
on mobile devices is provided. It presents DOT (Dynamic Object Tracking); a front-end
that added to existing SLAM systems can significantly improve their robustness and
accuracy in highly dynamic environments. [3]

Irene Ballaster et. al has developed a project using Dot-technology in VSLAM. Using
ORB-SLAM 2 dataset, the result was significantly shown to be improved.

Segmentation and multi-view geometry to generate masks for dynamic objects in order to
allow SLAM systems based on rigid scene models to avoid such image areas in their
optimizations. This short-term tracking improves the accuracy of the segmentation with
respect to other approaches. In the end, only actually dynamic masks are generated.
Evaluation of DOT with ORB-SLAM 2 in three public datasets is done. Results show that
the approach improves significantly the accuracy and robustness of ORB-SLAM2,
especially in highly dynamic scenes [4].

Lu. W et. al in his research has shown the disadvantages the VSLAM offers where it is
shown to be prone to ambient lighting.

The Visual SLAM is sensitive to ambient lighting and optical texture and not stable in the
outdoor environment, so it cannot be used for all-weather unmanned vehicle within a short
time. [5]

6
3. METHODOLOGY

3.1 Diagram of Proposed System

Field of
Vision

U PI U
CAMERA
W W
TCR
T
Legend
Raspberry PI 4
W-Wheels
W TCR W TCRT-Line following sensor
T
PI U-Ultra sonic Sensor
CAMERA
U U

Field of
Vision

Figure 1: Proposed Diagram of Robot

7
ROBOT CONTROLS
WHEELS,
ARDUINO, MOTOR
Raspberry PI DRIVER
4

SENSOR Ubu Raspberry


Ubuntu VSLAM
DATA PI 4 18.04/20.04

Level 2ROS

Some automation

PI
CAMER
A DATA
Rviz
(Simulation)

Figure 2: Block Diagram of the proposed


system

Level 5
Level 4
Level 3 Fully
Level 1 Level 2
High Autonomous
Nominal
Level 0 Some Automation
Intelligent Autonomy
automation
Manual Features
Driver

Figure 3: level of automation

8
LEVEL Automation Features

0 Zero automation Manual Driver

1 Intelligent Features Car can alert driver to


issues, conditions.

2 Some automation Car can perform multiple


assisted functions but driver
still in control.

3 Nominal Autonomy Car can perform most safety


critical functions in known
conditions, but driver must
be ready to take control.

4 High Automation Car can perform all safety-


critical driving functions
within limited areas and can
stop to transfer control when
needed.

5 Fully Autonomous Car can proceed without a


driver needed in any
circumstances, steering
wheel and brake pedal no
longer necessary.

Table 1: Level of Automation

9
3.2 Tool used

3.2.1 Hardware requirements

S.N Particulars Quantity Price (Rs.)

1. Raspberry Pi 4 1 9000

2. Wheel (10 cm diameter) 4 1350

3. Ply (5 mm) 5 400

4. Monster Shield 1 900

5. Motor (300-500 rpm) 4 2000

6. Pi Camera 2 1500

7. Ultrasonic Sensor 4 1500

8. TCRT 5000 2 700

9. ATMEGA328 1 700

TOTAL 18050

3.2.2 Software requirements


For development:

▪ Ubuntu LTS 18.04/20.04


▪ ROS Melodic/ROS Noel
▪ VSLAM
▪ Robot Operating System Visualization (Rviz).
▪ Arduino IDE

10
3.3 SYSTEM REQUIREMENTS

✓ Open VSLAM

Open VSLAM is a well-engineered, module-structured implementation of an ORB features


based Visual graph VSLAM. It contains optimized implementations of features extractors
and stereo matcher. Open VSLAM is a monocular, stereo, and RGBD visual SLAM
system. The notable features are: It is compatible with various types of camera models and
can be easily customized for other camera models. Created maps can be stored and loaded,
and then Open VSLAM can localize new images based on the prebuilt maps.

Figure 1 Source: https://dl.acm.org

✓ Raspberry PI 4

Raspberry pi 4 Model B is the latest product in popular Raspberry pi range of computers.


It offers ground-breaking increase in processor speed, multimedia performance, memory,
and connectivity compared to the prior-generation Raspberry pi 3 model B+ while retaining
backwards capability and similar power consumption.

Figure 2 source: https://www.conrad.com

11
✓ Ubuntu

Ubuntu is Linux distribution based on Debian and composed mostly of free and open-
source software. Due to compatibility issues, we use LTS Version 18.04 or 20.04. It is
officially released in three editions: Desktop, Server, and Core for Internet of things devices
and robots. All the editions can run on the computer alone, or in a virtual machine.

Figure 3 Source:https://logo-worlds.net

✓ RViz

Rviz is a 3D visualization software tool for robots, sensors, and algorithms. It enables you
to see the robot's perception of its world (real or simulated). The purpose of it is to enable
you to visualize the state of a robot. If an actual robot is communicating with a workstation
that is running rviz, it will display the robot's current configuration on the virtual robot
model. ROS topics will be displayed as live representations based on the sensor data
published by any cameras, infrared sensors, and laser scanners that are part of the robot's
system. This can be useful to develop and debug.

Figure 4 Source :https//ros//roswiki.com

12
✓ ATMEGA328

ATmega328 is a simple,low-cost and low-powered microcontroller used in


autonomous system.It is a single chip microcontroller created by Atmel in
megaAVRfamily.It has a modified Harvard architecture 8-bit RISC processor core.It uses
flash as program type memory.It has a program memory of 32kb.It has 23 I/O pins and 6
PMW pin.

Figure 5 www.microchip.com

13
3.4 FEASIBILITY STUDY

➢ Open VSLAM is in infancy, economically speaking. However, it can replace GPS


system and LIDAR system. Indoor mapping can be used for surveillance and
mapping of indoor.

➢ Autonomous system is a booming multi-billion dollar industry that will surely


revolutionize mobility and transportation. The use of Autonomous System will
reduce deadly toll of manual driver prone to human error.

➢ Visual simultaneous localization and mapping (SLAM) is quickly becoming an


important advancement in embedded vision and refers to the process of
determining the position and orientation of a sensor with respect to its surroundings
while simultaneously mapping the environment around that sensor.

➢ One major potential opportunity for visual SLAM systems is to replace GPS
tracking and navigation in certain applications. GPS systems aren’t useful indoors,
or in big cities where the view of the sky is obstructed, and they’re only accurate
within a few meters. Visual SLAM systems solve each of these problems as they’re
not dependent on satellite information and they’re taking accurate measurements of
the physical world around them.

14
4. EPILOGUE

4.1 EXPECTED OUTPUT

• Rich and mature visual SLAM approaches have been provided across May
operating domains of service robots. The adoption of VSLAM approaches on
service robots is a critical evolution to grow the industry away from prohibitively
expensive Lidars service robots are deployed in many environments, with any
different sensors, on the ground and in the air.

• The goal of this analysis is to identify general purpose techniques which may be
used to support innumerable services robot applications. Through this
experimentation, it was concluded the open SLAM was the overall best general
purpose technique for the broadest range of service robot types, environments and
sensors. It performed well in all three studies, showcasing superior re-localization,
variable lighting performance with high reality.

15
4.2 GANTT CHART

Days
1 21 41 61 81 101

Requirement gathering

Analysis

Design

Coding

Simulation and Testing

Prototyping

Implementation

Documentation

Start day Duration in days

16
REFERENCE AND BIBLIOGRAPHY

[1]110.Kim, J.; Choi, J.; Kim, Y.; Koh, J.; Chung, C.C.; Choi, J.W. Robust Camera Lidar
Sensor Fusion Via Deep Gated Information Fusion Network. In Proceedings of the 2018
IEEE Intelligent Vehicles Symposium (IV), Changshu, China, 26 30 June 2018; IEEE:
Changshu, China, 2018; pp. 1620 1625

[2]127. Singandhupe, A.; La, H.M. A Review of SLAM Techniques and Security in
Autonomous Driving. In Proceedings of the 2019 Third IEEE International Conference
on Robotic Computing (IRC), Naples, Italy, 25 27 February 2019; pp. 602 607. [Google
Scholar]
[3] Htisham Ali Ahmed Durmush Olli SuominenJariYli-Hietanen Sari PeltonenJussi Collin
Atanas Gotchev, Robotics and Autonomous system 132(2020) 103610

[4] Edge Assisted Mobile Semantic Visual SLAM ,Jingao Xu, Hao Cao, Danyang Li,
Kehong Huang, Chen Qian, LongfeiShangguan, Zheng Yang 2020

[4][DOT: Dynamic Object Tracking for Visual SLAM, Irene Ballester, Alejandro Fontan,
Javier Civera, Klaus H. Strobl, Rudolph Triebel 2021]

[5]166.Lu, W.; Zhou, Y.; Wan, G.; Hou, S.; Song, S. L3-Net: Towards Learning Based
LiDAR Localization for Autonomous Driving. In Proceedings of the 2019 IEEE/CVF
Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA,
16 21 June 2019; pp. 6382 6391. [Google Scholar]

17
18
19
20

You might also like