Adaptive Real-Time Road Detection Using Neural Networks
Adaptive Real-Time Road Detection Using Neural Networks
The Intelligent Systems Division of the National Institute 2.1 Training of the Neural Network
of Standards and Technology has been engaged for
several years in developing real-time systems for Figure 1 gives an overview of the Neural Network
autonomous driving. Road detection systems have been training step. From a number of road images, which are
developed using a number of different algorithms, but acquired as RGB images, features are extracted. As
real-time processing is always the key. Algorithms for introduced in [6], an “independent” color histogram
paved road detection [2, 4] have been extensively consisting of 8 bins per channel is used [9]. Additionally,
developed. Most are based on detecting road markings we put the normalized x and y position values of the
and lanes and cannot be applied to roads without any current point of consideration into the set of features. As
markings or lanes. shown in [3], the integration of x and y has improved
classification results. Each feature vector is then manually
We present a real-time road detection application using annotated as either road or non-road area in the image.
Neural Networks [7], which is independent of road
markings and lanes. During a short initialization step,
feature data is automatically collected based on the
conforming road structure in images captured from the
driver’s point of view (road images). Then, the new
Neural Network is trained and applied to the new
environment. This procedure allows the system to detect
road adaptively.
Figure 1: Neural Network Training Overview
In this paper, we will discuss the Neural Network-based
road detection in Section 2. Section 3 explains the In the next step, a Neural Network is trained based on the
adaptability and real-time processing of the system for extracted set of feature vectors and the annotated
detecting various road and terrain environments. road/non-road labels as targets.
the images captured from this perspective typically forms
the shape as illustrated in Figure 4. Each road image
consists of the road area and non-road areas as well as
named borders between these areas, e.g. road boundary
and horizon. For the approach described here, we take
Figure 2: Neural Network based advantage of the fact that the road usually covers a
Road Detection Overview trapezoidal area, which is centered in the lower part of the
image.
2.2 Road Detection
1
Certain commercial equipment, instruments, or materials are
3.4 Post Processing identified in this paper in order to adequately specify the
experimental procedure. Such identification does not imply
Road segmentation with Neural Network provides recommendation or endorsement by NIST, nor does it imply that
relatively reliable results in a real-time environment. the materials or equipment identified are necessarily best for the
purpose.
5 Results
Overall, the two peaks are mainly due to false negatives.
We compared the system’s performance with manually This is caused by shadow on the road, which was not
annotated frames of video files in order to measure the trained by the Neural Network and is therefore classified
accuracy. This allowed us to compute the false positive as non-road.
and false negative ratios. False positives refer to actual
non-road areas in the image, which were classified by the
system as road, while false negatives refer to actual road 5.2 Example “Road with Shadow”
areas classified as non-road.
Followings are the results of four videos with different The Neural Network applied for this scene was trained on
road scenes processed by the system. Each section frames where some road areas are covered with shadow
consists of four images and a graph. We trained the (see Figure 10). This condition consistently appears
Neural Network on selected frames of the first four throughout the video.
seconds of the video represented by the first image. The
other three images are captured during the course of
process.
The results are depicted by a graph in which we show the
classification errors – false negatives, false positives and
the sum of both – for every 25th frame of the video. Figure 10: Sample Frames for Example
"Road with Shadow"
5.1 Example “Straight Road” A rise of the overall classification error as seen in Figure
11 is mainly due to a higher false positive rate.
The road scene of this example (see Figure 8) consists 30
mainly of a straight road showing the same appearance
Classification Errors (%)
25
regarding pavement and lane markings. Some parts of the
video show shadows on the road. 20
15
10
0
1
11
13
15
17
19
21
23
25
27
29
31
Figure 8: Sample Frames for Example
Frames
"Straight Road"
False positives False negatives Sum
The Neural Network training frames did not contain any
shadows on the road. The classification results Figure 11: Sample Frames and Classification Results
summarized in Figure 9 show a very low misclassification for Shadow Road Scene
rate for most parts of the video. Since the dark areas of shadow on the road and in non-
30 road areas appear similar, the system tends to misclassify
dark non-road samples as road.
Classification Errors (%)
25 Shadow
20
15 5.3 Example “Slight Changes in Environment”
10
5
The road in this video follows a curvy course but is still
simple in structure (see Figure 12).
0
1
5
9
13
17
21
25
29
33
37
41
45
49
53
Frames
False positives False negatives Sum
25
the course, where it was exposed to limited types of road
20
and scenery, variations of the road (e.g. concrete, regular
pavement or gravel) cause higher false negative rates.
15
Higher false positives rates are due to changes in the
10
environment (buildings, trees etc.).
5
0
1
3
5
9
11
13
15
17
19
21
23
25
27
29
31
Frames
5.5 Summery
False positives False negatives Sum
Figure 16 gives an overview of the minimum, average
Figure 13: Sample Frames and Classification Results and maximum classification errors for the processed road
for a Scene with slight Changes scene samples. The left-most column represents the result
from the straight road scene with no shadows only. The
misclassification rate of this part is lower than the column
Smooth changes occur in the environment (non-road area) second from left, which depicts the result for the entire
while the road itself appears similar. These changes and straight road scene.
the changing direction of the sun cause an increase in
30
false positives over the time.
20
5.4 Example “Road and Environment with 15
drastic Variations” 10
5
In this video, the environment and the road itself change
0
drastically (see Figure 14). Straight
Straight Slight Massive
Road (No Shadow
Road Changes Changes
Shadow)
Min 0.93506494 0.93506494 2.23376623 2.28571429 1.4025974
Average 1.9554731 3.27272727 4.51948052 4.62987013 8.21872864
Max 3.11688312 17.3506494 11.9480519 14.8571429 27.8961039
25 example).
20 In general, the classification error increases over time as
15
new environments are entered. Variations in the non-road
areas lead to more false positives and changes in the road
10
area lead to more false negatives.
5
0
1
10
13
16
19
22
25
28
31
34
37
Acknowledgments
Thanks to Christopher Rasmussen, who originally
developed the feature extraction system and determined
the methods for SVM and neural Network generation, and
Patrick Conrad, who has enhanced the system by adding
more features. We are also grateful for the support of the
Army Research Laboratory and the Defense Advanced
Research Projects Agency MARS program.