tmpDD23 TMP
tmpDD23 TMP
1 King Abdullah II School for Information Technology, University of Jordan, Amman 11942, Jordan
College of Engineering and Computer Science, Australian National University, Canberra, ACT 0200, Australia
Introduction
620
that was developed by Oh et al. [7] which looks for three categories in video images acquired from a single camera, which
are trac speed, trajectory, and conflict. In another work [8],
Cheng and Hsu applied a time-varying adaptive system state
transition matrix in Kalman filter for vehicle tracking and
also utilised a regression analysis model for estimating traffic flow. Houben et al. extracted maximum phase congruency
and edges from stereo images and matched together with local matching algorithm, then processed by maximum spanning tree clustering algorithm to group the points into vehicle objects [9]. Lane trac characteristics, number of vehicles in each lane, and mandatory lane-changing fractions
in lanes with trac congestion were employed in this work.
Bishop and Casabona used GPS-enabled networked mobile
phones for collecting and analysing location based data set,
and followed by sample filtering and route machining for
trac congestion determination [10]. Others like Ozkurt and
Camci used a machine learning approach based on neural
networks for trac density estimation [11]. Based on maximum phase congruency and edges features matching and a
3D tracking, the system detects vehicles and determines their
dimensions. Jain et al. applied a simulation-based analysis on
a simple network topology where a local congestion protocol
that controls the flow of trac into near-congested regions for
preventing congestion collapse sustaining time variant trac
bursts [12]. Chen et al. developed a night-time vehicle detection and trac surveillance system, where vehicles head
and tail-lights are located and processed by a spatial clustering for analysing the spatial and temporal features of vehicle
light patterns [13]. While Yu and Chen used a simple consecutive temporal dierence approach for trac monitoring after pre-processing with a square neighbourhood algorithm for
compensation of camera disturbances [14]. Also, Marfia and
Roccetti proposed a short-term trac congestion forecasting
approach that requires no prior. knowledge on road condition,
by estimating the time interval of how long would a congestion last [15]. Another work implements a vehicle counting
method through blob analysis and tracking, and eventually
determining the speed of vehicles [16]. A review on computer vision techniques utilised in trac scene analysis can
be found in [17].
Since image texture has shown in previous work its usefulness in dierentiating between dierent image patterns
[18,19], it could as well assist in representing a better understanding of road trac scenes, if we take into account that the
vehicles distribution are the patterns of interest. Therefore,
and to the best of our knowledge, road trac density from
a texture perspective has not been classified before into ad-
Fig. 1
Materials
621
Monitored roadway
Capture
Road
Capture
Road
Capture
location
time
location
time
location
time
A1-R1a
16:00
A2-R2a
13:00
A4-R3a
14:00
A1-R2a
16:00
A2-R2b
13:00
A4-R3b
14:00
A1-R2b
16:30
A3-R1a
15:15
A5-R1a
08:00
A1-R1b
16:30
A3-R1b
15:15
A5-R1b
08:00
A1-R3a
15:45
A3-R2a
15:00
A6-R1
12:30
A1-R3b
15:00
A3-R2b
15:00
A7-R1a
08:30
A1-R4a
15:00
A3-R3
14:00
A7-R1b
08:30
A1-R4b
15:45
A4-R1a
13:30
A7-R2a
08:30
A1-R5
15:45
A4-R1b
13:30
A7-R2b
08:30
A2-R1a
11:45
A4-R2a
14:00
A8-R1a
15:30
A2-R1b
11:45
A4-R2b
14:00
A8-R1b
15:30
In order to preserve quality, the acquired video files were initially processed at full size with a resolution of RGB 720480
pixels and 29 frames per second (fps), and the recording process was for duration of 3 minutes. This will result in a very
large file size that would be deemed computationally intensive. Thus a preprocessing stage would be required to reduce
the video size and frame rate while maintaining acceptable
quality, a process that would have minimal side eect on the
vehicle segmentation.
The captured videos from the fixed camcorder were in the
MOD file extension, a JVCs implementation of MPEG-2
transport stream, and for the successive video frames to be
appropriately read and processed, they were converted to AVI
format, a common multimedia container used by dierent
codecs and known for its good video quality and wide applicability. Xilisoft video converter http://www.xilisoft.com
was used to convert the recorded MOD video files to its corresponding AVI format after setting appropriate resolution,
frame rate and aspect ratio.
All acquired videos were converted to grey scale in order
to reduce processing time since we are interested in detecting
any car irrespective of its colour. Then each video frame rate
and size was adjusted from 29 fps, 720480 pixels having
622
VA
,
Vmax
(3)
Methodology
D
,
FA
(1)
FA =
(2)
where fr is the frame rate and C corresponds to the total number n of m detected vehicles c per video frame.
For example in case A1-R1b in Table 1, suppose the monitored part of the lane D was 25 meters long and the video
displayed the frames at a rate of 15 per second for a duration
of 1 minute and the number of detected objects (i.e., vehicles)
was 740. The number of detected vehicles (F A ) in the video
can be calculated by substituting in Eq. (2), so the average
velocity (VA ) can be easily estimated by substituting in Eq.
(1) which yields 74 km/hr.
(4)
0, otherwise,
where k (k > 1) is the time of video frame capture, and T is
an appreciably dierent threshold set to the mid grey-level of
the each video frame according to the following function:
1
(max( f (x, y, tk )) + min( f (x, y, tk ))) .
(6)
2
Regional maxima
Regional maxima can be defined as a certain connected components of pixels with a certain height attribute, such that all
surrounding grey level pixels in the external boundary have
strictly lower values. This could be utilised in removal of
isolated low-valued points in the absolute dierence image
Di j (x, y). Isolated small structures are likely to be considered
as noise, thus regarded as an essential procedure for reduction of false positives in the segmentation process. The elimination of such case can be done by arranging a certain set
of n-connected regions (8 pixels was applied in this work)
in Di j (x, y) and ignoring detected structures that are smaller
in size. Nevertheless, small size and/or slow moving vehicles
might be erroneously eliminated, yet the appropriate selection of the threshold value T could contribute in mitigating
this drawback, taking into consideration the camera field of
view, depth of field, spatial resolution, focused distance, position and orientation.
Connected component extraction
Connectivity is an important concept that defines regions and
boundaries in images. To establish connectivity between two
pixels in an image, they have to be neighbours and their corresponding grey levels satisfy a criterion of similarity.
Let Y represent a connect component contained in a set A
and assume that a point of Y is known. Then,
T=
Xk = (Xk1 B) A,
k = 1, 2, . . . , n
(7)
623
(8)
(10)
624
Fig. 4 Example of vehicle detection and tracking in a video frame of road A1R5. (a) Acquired frame, (b) intensity adjustment, (c) background
removal, (d) morphological processing and object detection, and (e) labeling detected vehicles on original frame for tracking subsequently speed
measurement
1, x 0,
s(x) =
(12)
0, otherwise.
If the coordinates of gc where (xc , yc ) the coordinates
of
2p
, yc
the neighbours g p are given by xc + R cos
P
2p
.
R sin
P
The spatial structure of the local image texture can be characterised by
LBPP,R =
P1
s(I x p y p I xc yc ) 2P .
(13)
p=0
eavg =
Fig. 5
image
(14)
i=0
scale, normalised from 0 to 1, by setting four dierent thresholds T s , (s = 0, 1, 2, 3) relevant to the road average velocity. The threshold describes how much the road capacity is
utilised, where the congestion rate is represented as 1 minus
the detected vehicles divided by the value of 800, which is
the number of detected vehicles equivalent to the maximum
road speed limit of 80 km/hr for a monitored distance of 25
meters for duration of 1 minute, which is indicated as:
Let T s = [0.1, 0.3, 0.6, 0.8];
if CR < T 3 then RF = 5;
else if CR T 3 AND CR < T 2 then RF = 4;
else if CR T 2 AND CR < T 1 then RF = 3;
else if CR T 1 AND CR < T 0 then RF = 2;
else if CR T 0 then RF = 1;
end if.
(15)
The feature vectors with the morphological operations are
used to determine the road congestion rate can be summarised
in Algorithm 1.
Algorithm 1
for i = 1, 2, . . ., m do
C = 0;
for j = 1, 2, . . ., n do
D xy = |Hq Ir |;
if D xy > S then D xy = 1;
else D xy = 0;
end if
M = Dos
for k = 1, 2, . . ., l do
arg max(M(k))
end
C = C + M;
FA = f /C;
Table 2
625
speed
Road
rate /%
condition
Yellow
/kmhr1
A1-R1a
423
0.50
3.70
42.3
47
A1-R2a
626
0.32
3.98
62.6
22
Green
A1-R2b
710
0.42
3.49
71.0
11
Green
A1-R1b
740
1.10
3.93
74.0
07
Blue
A1-R3a
81
5.19
4.79
08.1
90
Red
A1-R3b
449
0.22
4.12
44.9
44
Yellow
A1-R4a
738
0.00
4.21
73.8
08
Blue
A1-R4b
687
2.23
4.18
68.7
14
Green
A1-R5
774
0.77
3.70
77.4
03
Blue
A2-R1a
214
3.38
4.48
21.4
73
Orange
A2-R1b
392
1.26
3.22
39.2
51
Yellow
A2-R2a
357
0.83
2.88
35.7
55
Yellow
A2-R2b
702
1.81
4.10
70.2
12
Green
A3-R1a
553
2.47
3.68
55.3
31
Yellow
Yellow
Va = D/FA ;
A3-R1b
540
1.10
3.73
54.0
32
A3-R2a
53
1.42
4.30
05.3
93
Red
A3-R2b
118
1.72
4.07
11.8
85
Red
A3-R3
227
0.88
3.76
22.7
72
Orange
% Set threshold
A4-R1a
433
0.92
3.00
43.3
46
Yellow
if CR < T 3 then RF = 5;
% Free flow
A4-R1b
541
1.99
3.17
54.1
32
Yellow
else if CR T 3 AND CR
< T 2 then RF = 4;
% Moderate flow
A4-R2a
279
0.36
3.74
27.9
65
Orange
else if CR T 2 AND CR
< T 1 then RF = 3;
% Restricted flow
A4-R2b
306
2.00
3.82
30.6
62
Orange
else if CR T 1 AND CR
< T 0 then RF = 2;
% Slow flow
else if CR T 0 then RF = 1;
% Stopped flow
end if
end for
end for
Results
A4-R3a
192
2.67
3.80
19.2
76
Orange
A4-R3b
236
2.48
3.62
23.6
71
Orange
A5-R1a
690
0.00
3.40
69.0
14
Green
A5-R1b
357
3.48
4.03
35.7
55
Yellow
A6-R1
210
0.00
4.31
21.0
74
Orange
A7-R1a
186
1.59
3.29
18.6
77
Orange
A7-R1b
663
2.21
3.76
66.3
17
Green
A7-R2a
597
1.97
3.27
59.7
25
Green
A7-R2b
538
0.00
3.27
53.8
33
Yellow
A8-R1a
735
1.94
4.14
73.5
08
Blue
A8-R1b
360
4.00
3.33
36.0
55
Yellow
Table 3
Colour code
Description
Blue
Free flow
Less than 10
Green
Moderate flow
Between 10 and 29
Yellow
Restricted flow
Between 30 and 59
Orange
Slow flow
Between 60 and 79
Red
Stopped flow
Black
No data
626
Fig. 8 Time series LBP entropy for (from top to bottom) A1-R3a (stopped),
A1-R1a (restricted), A1-R2a (moderate), A1-R2b (moderate), A1-R3b (restricted), and A1-R5 (free), respectively
Fig. 6 Road flow conditions for 33 roads in eight dierent busy areas in
Amman city at dierent times recorded at daytime
Table 4
Colour code
Peak cases /%
Free
50.00
Moderate
42.86
Restricted
54.54
Slow
85.71
Stopped
66.67
As a second measure that can assist in improving road condition estimation, the LBP entropy can measure the degree
of randomness in the monitored roads without counting the
number of vehicles in the assigned two minutes period. For
instance, Figure 7 shows how the LBP average entropy can
be used to dierentiate between restricted and moderate flow
road conditions for road A1-R1. Also when applied to six different roads in area A1 as illustrated in Fig. 8, it could assist
in distinguishing between stopped, restricted, moderate and
free flow road conditions.
two other statistical-based texture methods, namely greylevel co-occurrence matrix (GLCM) and run-length matrix
(RLM), and with non-textural processed video frames for investigating robustness. The errors for all cases are shown in
the boxplot of Fig. 9.
Fig. 9 Boxplot of vehicle error detection for the proposed LBP method
as compared to two other statistical-based texture method and non-textural
processed video frames
Discussion
Road trac usually exhibit stochastic behaviour due to conditions beyond our control, such as accidents, bad weather,
and urgent road works that could aect the general trac
flow. Thus this work tends to provide a system that can automatically indicate the severity of road congestion, and hence
improving the road flow by giving drivers an option to use
an alternative way in case if the trac situation is congested.
As humans tend to be more attentive to colours, and are also
mainly used in trac lights to indicate road status and control
trac flow, this work utilises colour codes as well to further
discriminate between dierent levels of congestion and to fa-
627
edges of the vehicle structures would increase the randomness of the road background and hence give an indication of
road occupancy. In other words, a higher entropy value would
be noticed as compared when the road is empty. This is apparent from the comparison results in Fig. 9 where the LBP average entropy had an improved performance as compared to
other statistical textural methods and to the non-textural processed video frames. This implies that the decrease in the vehicles detection error in respect to the other compared methods had a positive eect on the estimated average road speed,
and hence the accuracy of the estimated congestion rate.
As there is no system free from disadvantages, the system currently does not indicate explicitly the reason for the
increase in congestion (i.e., is it due to an accident, road
work, double parking, etc.), and other trac conditions and
their possible eect on the approach proposed, such as bad
weather, would be investigated in future work. Knowing the
reasons behind congestion would give road trac control authorities more information on the problem in order to take
fast and suitable measures to deal with this situation. Also the
level of road noise detected by the surveillance cameras microphones could be employed for improving congestion detection.
Conclusions
References
1.
Villagra J, Milanes V, Perez J, Godoy J. Smooth path and speed planning for an automated public transport vehicle. Robotics and Autonomous Systems, 2012, 60: 25226
2.
3.
628
4. Schrank D, Lomax T, Eisele B. 2011 Urban Mobility Report and Appendices. Texas Transportation Institute, 2011
14.
Yu Z, Chen Y P. A real-time motion detection algorithm for trac monitoring systems based on consecutive temporal dierence. In: Proceedings of the 7th Asian Control Conference. 2009, 15941599
15.
16.
Brahme Y B, Kulkarni P S. An implementation of moving object detection, tracking and counting objects for trac surveillance system. In:
Proceedings of the 2011 International Conference on Computational
Intelligence and Communication Networks. 2011, 143148
17.
Buch N, Velastin S A, Orwell J. A review of computer vision techniques for the analysis of urban trac. IEEE Transactions on Intelligent Transportation Systems, 2011, 12: 920939
18.
19.
20.
21.
Ojala T, Pietikainen M, Maenpaa T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 2002, 24:
971987
22.
Guo Y, Zhao G, Pietik M. Discriminative features for texture description. Pattern Recognition, 2012, 45: 38343843
7. Oh J, Min J, Kim M, Cho H. Development of an automatic trac conflict detection system based an image tracking technology. Transportation Research Record, 2009, 4554
8. Cheng H Y, Hsu S H. Intelligent highway trac surveillance with
self-diagnosis abilities. IEEE Transactions on Intelligent Transportation Systems, 2011, 12: 14621472
9. Houben Q, Diaz J C T, Warzee N, Debeir O, Czyz J, Insticc. Multifeature stereo vision system for road trac analysis. In: Proceedings
of the 4th International Conference on Computer Vision Theory and
Applications. 2009, 554559
10.
Bishop B, Casabona J. Automatic congestion detection and visualization using networked GPS unit data. In: Proceedings of the 47th Annual Southeast Regional Conference. 2009, 79
11.
12.
Jain V, Sharma A, Subramanian L. Road trac congestion in the developing world. In: Proceedings of the 2nd ACM Symposium on Computing for Development. 2012, 11
13.