Lane Detection and Tracking Using A New Lane Model
Lane Detection and Tracking Using A New Lane Model
net/publication/37988009
Lane Detection and Tracking Using a New Lane Model and a Distance Transform
CITATIONS READS
75 849
4 authors, including:
Tobi Vaudrey
Auckland University of Technology
51 PUBLICATIONS 963 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Tobi Vaudrey on 30 May 2014.
1 Introduction
Lane detection plays a significant role in driver assistance systems (DAS), as it can help
estimate the geometry of the road ahead, as well as the lateral position of the ego-vehicle
on the road. Lane detection and tracking have been widely studied for driving on a free-
way or an urban road, for single or multiple lanes, with or without marks, based on
region (texture or color) or edge features. Various models have been applied to describe
the borders of a lane, such as piecewise linear segments, clothoids, parabola, hyperbola,
splines, or snakes. For a complete review of lane detection algorithms, please refer to
[4]. There are even some commercial lane detection systems available, working mainly
on highways. Loose et al (Daimler AG) state in [3]: “Despite the availability of lane de-
parture and lane keeping systems for highway assistance, unmarked and winding rural
roads still pose challenges to lane recognition systems.” Generally, it is a challenging
task to robustly detect lanes in varying situations, especially in complex urban environ-
ment.
This paper introduces a new weak (i.e., with no assumption about the global shape
of a lane) road model for lane detection and tracking. Instead of modeling global road
geometry, this new model only constrains relations between points on the left and right
lane boundaries. Tracking based on these points in the birds-eye image (using a particle
filter) provides lane detection results. Furthermore, a modified version of a standard
Euclidean Distance Transform (EDT) is applied on the edge map of the birds-eye image.
Utilizing some beneficial properties of this distance transform for lane detection and
tracking, this paper also specifies an innovative initialization method for the particle
filter.
This paper is organized as follows: Section 2 describes a new lane model. Section 3
explains a modified version of a standard Euclidean distance transform, and its useful-
ness for lane detection. Lane detection and tracking methods are introduced in Section 4
2 R. Jiang, R. Klette, T. Vaudrey, and S. Wang
PZ
aa a
H
b1 b2
y b1 b2
PL PR
Pc
L2 PL Pc(xc,yc) PR
L1
Ground manifold PL Pc(xc,yc) PR
x
(a) (b) (c)
Fig. 1. Lane model as used in this paper. (a) 3D lane view; boundaries are drawn in bold. (b)
Perspective 2D lane view in the input image. (c) Birds-eye image of the lane. Slope angles β1
and β2 are shown in the 3D view and the birds-eye images; the zenith angle α in the 3D and the
projective view. See text for further explanations.
and Section 5. Experimental results are given in Section 6. Finally, conclusions are pro-
vided in Section 7.
3 Distance transform
The distance transform is applied to the binary edge map, which labels each pixel with
the distance to the nearest edge pixel. The Euclidean distance transform (EDT) is, in
general, the preferred option; using the Euclidean metric for measuring the distance be-
tween pixels. A modified EDT was proposed in [7], called orientation distance trans-
form (ODT). This divides the Euclidean distance into a contributing component in the
row and column direction. A complex number is assigned to each pixel by the ODT,
with the distance component in the row direction as the real part, and the distance com-
ponent in the column direction as the imaginary part. Note that distance component in
row direction is signed, with a positive value indicating that the nearest edge point lies
New Lane Model and Distance Transform 3
E1
d
i
r row direction
P2 P1
E2
Fig. 2. Euclidean distance and Orientation distance. P1 and P2 are two neighboring non-edge
pixels. E1 is the nearest edge point to P1 , and E2 is the nearest edge point to P2 .
(d) (e)
Fig. 3. EDT and ODT on a birds-eye road image. (a) Birds-eye road image. (b) Binary edge map.
(c) Imaginary part of ODT. (d) Real part of ODT (absolute value). (e) EDT. (c)(d)(e) have been
contrast adjusted for better visibility. The square root of distance value in (c) and (d) will be the
Euclidean distance value as (e). The rectangle in (b) indicates a small window for comparisons
of dashed lane mark with EDT and RODT(see below).
to the right, and a negative value if it is to the left. See Figure 2 for the relationship
between ODT and EDT.
This paper uses only the Euclidean distance in the row direction, and we call this
the real orientation distance transform (RODT). An example of EDT and ODT in our
application is shown in Figure 3. The RODT of our edge map offers various benefits.
First, the initialization of lane detection becomes much easier (to be discussed in Sec-
tion 4.1). Second, dashed lane marks will make no difference with continuous ones in
the RODT, as illustrated in the comparison of the rectangle area in Figure 3. Third, more
information about the centerline is provided by the distance transform compared to the
edge map. Generally, a (non-edge) pixel on the centerline of a lane will have a local
maximum in distance to the lane boundaries. Thus, combined with the lane model in-
troduced in Section 2, a point with a high distance value is likely to be a centerline point
PC . The usefulness of these properties of the RODT will be discussed in the following
sections.
4 R. Jiang, R. Klette, T. Vaudrey, and S. Wang
4.1 Initialization
The aim of the initialization step is to find an initial value (e.g., the x-coordinate of a
point PL and point PR in a selected image row) for the specified model. In [5], a clus-
tered particle filter is used in order to find a start point on a lane boundary. In distinction
to this, we fully utilize the distance map to find the first left and right boundary points.
In a pre-defined start row (near to the bottom) of the birds-eye image, a search is con-
ducted, starting at the middle of the row, for a pixel which has a positive distance value
but a negative distance value at its left neighbor (see Figure 5). When such a pixel is
found, the left and right boundary points in the start row are instantly known using the
distance value of the found pixel and of its left neighbor.
For the initial state X0 (xc0 , α0 , β10 , β20 ) of the particle filter, xc0 and α0 are ini-
tialized by using the detected left and right start points, while β10 and β20 are simply
set to be zero.
row yc0 in the birds-eye image. For the application of a particle filter, two models are
discussed in the following.
The dynamic model. The dynamic model A is used to define the motion of particles in
the image. The prediction value X̂n is generated from Xn−1 by using X̂n = A · Xn−1 .
We simply take A as being the identity matrix, because of the assumed smoothness of
the lane boundary.
The observation model. The observation model determines each particle’s importance
factor for re-sampling. Based on the RODT information combining new model, it is
reasonable to assume points (xcn , ycn ) will have large distance values, and L1 and L2
coincide with short lines of pixels which all only have small distance values.
Tracking step n is identified by ycn = (yc0 + n · ∆). We calculate the lateral
position of the left boundary point of the lane from the predicted state vectors, with
X̂ni (x̂icn , α̂ni , β̂1i n , β̂2i n ) for the ith particle.
From now on, PL and PR only represent the lateral position of boundary points, for
simplicity. The left position is calculated as follows:
Here, d(·, ·) is the distance value of the RODT. Calculating SLi 2 in the analogous way,
we obtain the ith importance factor
!
i 1 (SLi 1 − µ1 )2 (SLi 2 − µ2 )2
ωdist = exp − −
2πσ1 σ2 2σ1 2σ2
For the centerline point (xicn , ycn ), the importance factor is equal to
2
1
− µ
1 i
d(xcn ,ycn ) 3
i
ωcenter = √ exp −
2σ3
σ3 2π
where µk and σk are constants, for k = 1, 2, 3 . The final observation model is given by
the factors
i i
ωi = ωdist · ωcenter
0 -1 -d d 1 0
Fig. 5. Illustration of the search procedure in the start row of the distance map. Note that the
distance values are signed, as described in Section 3.
6 R. Jiang, R. Klette, T. Vaudrey, and S. Wang
t=t+1 Adjustment of
Lane detection Detection of
in frame t {PLi, PRi, i=0,...,N-k} {PLi, PRi, i=N-k+1,...,N}
Here, k is determined by the driven distance between time t and t + 1, and is usually a
small number. Furthermore, points
(t+1) (t+1)
{PLn : n = 0, 1, . . . , N − k} and {PRn : n = 0, 1, . . . , N − k}
are obtained by adding some translation (according to n) caused by the variation in
driving direction between t and t + 1.
(t+1) (t+1)
For the detection of {PLn : n = N − k + 1, . . . , N } and {PRn : n = N −
k + 1, . . . , N }, note that k is small and we also assume smoothness of lane boundaries.
Thus, we simply start as follows:
(t+1) (t+1) (t+1) (t+1)
PL n = PLn−1 , PRn = PRn−1 , n = N − k + 1, . . . , N
(t+1) (t+1)
For further refinement, those predictions {PLn } and {PRn } from the previous
result at frame t are likely to be already located near the true points on the bound-
aries, as the variation of a lane is usually minor between two subsequent frames. – The
New Lane Model and Distance Transform 7
Fig. 7. Experimental results for lane detection. (a) Input images. (b) Lanes detected in the birds-
eye image. Note that the centerline of a lane is also marked. (c) Lanes detected in input images.
Fig. 8. Lane detection on roads with unmarked or blocked lane marks. (a) Input images. (b) Edge
map. (c) RODT. (d) Lane detected. Note that RODT contains more information than edge map
for lane detection as discussed in Sec. 3.
adjustment
(t+1) (t+1) (t+1)
PLn = PLn + d(PLn , ycn ), n = 0, 1, . . . , N
(t+1) (t+1) (t+1)
PRn = PRn + d(PRn , ycn ), n = 0, 1, . . . , N
of all N +1 points is finally achieved by information available from values of the RODT
of the current birds-eye edge map.
6 Experiments
Experiments were conducted on images and sequences recorded with the test vehicle
“HAKA1” of the .enpeda.. project [8].
Experimental results for lane detection are shown in Figure 7 and Figure 8. Differ-
ent scenarios are considered. Note that detected lane boundaries are sometimes locally
slightly curved. This is due to the fact that the distance transform of dashed lane marks
is slightly unaligned in column direction in the birds-eye image (see Figure 3).
Experimental results while using the efficient lane tracking method are illustrated in
Figure 9. The results are really acceptable except for some outliers for lane boundaries.
With an off-the-shelf computer, the computation time for the adjustment steps of effi-
cient lane tracking is negligible. The only measurable time needed is that for birds-eye
8 R. Jiang, R. Klette, T. Vaudrey, and S. Wang
SEQ1
SEQ2
view mapping, edge detection and distance transform, and all three sub-processes can
be computed highly efficiently.
7 Conclusions
This paper introduced a new weak model of a lane, and a possible lane detection scheme
using a particle filter based on a monocular camera. Furthermore, an efficient lane track-
ing method was proposed and discussed.
A (simple and easy to calculate) distance transform was used in this paper for lane
detection and tracking. It shows that the distance transform is a powerful method to
exploit information in lane detection situations. The distance transform can deal with
dashed lane marks, provides information for the detection of the centerline of a lane,
finds initial values for the particle filter, and adjusts the tracking results conveniently.
References
1. M. Bertozzi, A. Broggi: GOLD - A parallel real-time stereo vision system for generic obsta-
cle and lane detection. IEEE Trans. Image Processing, 7:62–81, 1998.
2. Z. Kim: Robust lane detection and tracking in challenging scenarios. IEEE Trans. Intelligent
Transportation System, 9:16-26, 2008
3. H. Loose, U. Franke, and C. Stiller: Kalman particle filter for lane recognition on rural roads.
In Proc. IEEE Intelligent Vehicle Symp., to appear, 2009.
4. J. C. McCall and M. M. Trivedi: Video-based lane estimation and tracking for driver assis-
tance: survey, system, and evaluation. IEEE Trans. Intelligent Transportation System, 7:20–
37, 2006.
5. S. Sehestedt, S. Kodagoda, A. Alempijevic, and G. Dissanayake: Efficient lane detection and
tracking in urban environments. In Proc. European Conf. Mobile Robots, pages 126–131,
2007.
6. Y. Wang, L. Bai, M. Fairhurst: Robust road modeling and tracking using condensation. IEEE
Trans. Intelligent Transportation Systems, 9:570–579, 2008.
7. T. Wu, X. Q. Ding, S. J. Wang, and K. Q. Wang: Video object tracking using improved
chamfer matching and condensation particle filter. In Proc. SPIE-IS & T Electronic Imaging.
volume 6813, pages 04.1-04.10, 2008
8. http://www.mi.auckland.ac.nz