Fast Anomaly Detection in Traffic Surveillance Video Based On Robust Sparse Optical Flow
Fast Anomaly Detection in Traffic Surveillance Video Based On Robust Sparse Optical Flow
College of Information System and Management, National University of Defense Technology, China
hanlin [email protected], [email protected], [email protected], [email protected]
Fig. 1. Optical flow computation and feature extraction. (a) illustrates the Extracted sparse optical flow. The head and length
of red arrows denotes the direction and speed of extracted optical flow. (b) demonstrates four major steps to compute robust
optical flow with high speed. (c) illustrates feature extraction and aggregation. Pixel feature consists of HOF with additional
channel of foreground [4]. To further improve robustness, a spacial Gaussian blur is performed on the aggregated feature (on
each channel separately). (d) demonstrates a scene image and the red block circled marks out one of the feature blocks. (e)
illustrates its feature variation along time with a plot of six channels of the feature. The horizontal axis is the frame number; the
vertical axis is the feature value.
The most time-consuming parts of our algorithms lie in the The proposed method is tested in numerous real-life situa-
computation of optical flow and aggregation of features. For tions to validate its effectiveness. Figure 2 illustrates the re-
optical flow, there are two methods for acceleration: sults. Given a few minutes of normal video for training, our
algorithm accurately detect various anomalies such as pedes-
a) Foreground mask. Computation of optical flow in the trian across the road at wrong location and reversely running
background area is neither nonsense nor wrong. There- motorcycles and trunks at different scenes, day and night.
fore, foreground is used as a mask to increase both effi-
ciency and robustness.
b) Spacial sampling. Sparse optical flow extracted from fixed 3.2. UCSD Ped1 Dataset
grid pixels are sufficient for feature aggregation.
The algorithm is tested on one of the most evaluated datasets:
For aggregation, integral images are used to compute the UCSD Ped1 Dataset [15]. This dataset provide training se-
sum of a rectangle area with a constant time cost. quences with only pedestrians and marks non-pedestrians as
anomalies. Note the ground truth is not fully consistent with
our definition of anomaly since this paper only takes low-
3. EXPERIMENTS possibility motion patterns as anomalies.
The results are illustrated in Figure 3. (a), (b) is the
The method is evaluated on both real-life situations and frame-level and pixel-level ROC curve, respectively. It can
a benchmark dataset. Parameter setting are as follows: be seen that our algorithm performs comparable with state-
the block size is 16 · 16 with a time window of 5 frames. of-art methods at frame-level and outperforms state-of-art
The threshold vector is different according to scenes. How- methods at pixel-level when the false positive rate is less than
ever, 0.1·num of block pixels/spacial sample distance· around 24%. This improvement is important because it is not
time window typically produces an acceptable result. practical to tolerate high false alarm rate. And (c) is the run-
Fig. 2. Detection results of real-life surveillance video. Group (a) shows two detected anomalies of one scene: the pedestrian
crossing the road at wrong location and a car entering the side-road at wrong location. Group (b) shows two detected reversely
running motorcycles of another scene. Group (c) shows two anomalies of different scenes: detected trunks running on the
wrong side of road at night and the pedestrian across the road at wrong location.
Fig. 3. Comparison of ROC curves on UCSD ped1 dataset. Here Adam refers to [1]; MDT refers to [3]; social force refers to
[14]; sparse and sparse combination refer to [1] and [6], respectively. (a) is the frame-level ROC curve; Note that frame-level
AUC is not the bigger the better if the pixel-level AUC is not in accord with it. (b) is the pixel-level comparison. (c) is a table
of running time comparison.
ning time comparison. Our algorithm runs real-time on this The sparse optical flow computation is improved in both ro-
benchmark dataset, which is much faster than existing algo- bustness and efficiency. A foreground channel is added to
rithms [3, 1]. Algorithm [6] reaches an impressing speed by HOF feature to detect long-term static objects. The algo-
resizing frames to 30 · 30 and other small resolutions. How- rithm is validated on real-life traffic surveillance and a bench-
ever, this will not work if the abnormal objects in original mark dataset to prove its effectiveness. The algorithm runs
frame are small. real-time and is hundreds of times faster than a number of
comparative algorithms. The speed gain is achieved by a
4. CONCLUSION fast feature extraction design and a simple detection model.
This research was partially supported by the National Natu-
We propose an efficient algorithm for detecting anomaly in ral Science Foundation of China under Grant 61403403 and
traffic surveillance video based on robust sparse optical flow. 61405252.
5. REFERENCES [12] C. Tomasi and J. Shi, “Good features to track,” Com-
puter Vision and Pattern Recognition (CVPR), IEEE
[1] Y. Cong, J. Yuan, and J. Liu, “Sparse reconstruction Conference on, pp. 593–593, 1994.
cost for abnormal event detection,” Computer Vision
and Pattern Recognition (CVPR), IEEE Conference on, [13] Z. Kalal, “Tracking learning detection,” PhD The-
pp. 3449–3456, 2011. sis, University of Surrey, vol. 30, pp. 555–560, January
2011.
[2] A. Adam, E. Rivlin, I. Shimshoni, and D. Reinitz, “Ro-
bust real-time unusual event detection using multiple [14] R. Mehran, A. Oyama, and M. Shah, “Abnormal crowd
fixed-location monitors,” Pattern Analysis and Machine behavior detection using social force model,” Computer
Intelligence, IEEE Transactions on, vol. 30, pp. 555– Vision and Pattern Recognition (CVPR), IEEE Confer-
560, 2008. ence on, pp. 935–942, 2009.