Visual Object Tracking Based On Dynamic Weights and Gaze Density
Visual Object Tracking Based On Dynamic Weights and Gaze Density
ABSTRACT
This study proposes an adaptive algorithm for
tracking objects based on the dynamic weights
scheme and gaze density. The proposed adaptive
algorithm is to amend the deficiency on tracking
moving objects that are not driven by inertia. Using
the dynamic weights scheme and the information of
the gaze point density, possible areas where the
objects may appear are constantly researched to
target and trace the aimed objects. It is proven to be
effective on tracking non-inertia driven moving
objects.
KEYWORDS
gaze tracking, object tracking, adaptive, gaze
density, dynamic weights
1 INTRODUCTION
Object tracking is a popular research
subject and applied widely in the domain of
computer vision for applications, such as
automatic video surveillance, e-healthcare,
vehicle detection, human-computer interaction,
and railway cross obstructions. There are
numerous algorithms proposed for object
tracking. Whereas, there are some general
obstacles when the tracked objects are partially
obscured, the observable perspectives are
converted, the observable appearance is altered,
or there are multiple objects to be tracked.
Mostly as seen in prior research studies, the
ascertainment of the moving model of the
objects is highlighted in terms of machine
learning, fixed-rate moving model, appearance-
152
Proceedings of The Fourth International Conference on Informatics & Applications, Takamatsu, Japan, 2015
Frame 314
Frame 315
Frame 316
Frame 317
Frame 318
Frame 319
Frame 320
Frame 321
(a) The Inconsistent Objects
Frame 322
Frame 323
Frame 251
Frame 252
Frame 253
Frame 254
Frame 255
Frame 256
Frame 257
Frame 258
Frame 259
(b) Non-Inertia Driven Moving Objects
Frame 260
sourcehttps://www.youtube.com/watch?v=ikpdmeQ1Dq8
153
Proceedings of The Fourth International Conference on Informatics & Applications, Takamatsu, Japan, 2015
154
Proceedings of The Fourth International Conference on Informatics & Applications, Takamatsu, Japan, 2015
2 PROPOSED METHOD
The proposed adaptive object tracking
algorithm is derived from Camshift in increasing
the tracking accuracy rate in terms of utilizing
the comparisons between the characteristics of
the video images as well as incorporating the
gaze density to adjust the prediction on the
central positions of the objects. The designated
flow of the adaptive object tracking algorithm is
shown as in Figure 2 and the corresponding
parameter specifications are listed in Table 1.
Parameters
n
WC
WS
WT
PC
PS
PT
FV
TF
Th
WA
N
(XAE,YAE)
dij
MGD
TGD
R
155
Proceedings of The Fourth International Conference on Informatics & Applications, Takamatsu, Japan, 2015
During
the
tracking
process,
the
coordination of users fixations are estimated and
the distances between gaze points (dij, distance
between gaze point i and j) are calculated, as
shown in Equation (1). When FV is smaller than
TF (TF, similarity threshold of tracked object) a
loss problem will be identified. Calculating the
average coordination (XAE,YAE) of the visual
fixations in the close N frames, as shown in
Equation (2) and Equation (3), as well as
calculating the density of gaze points as in
Equation (4), when the MGD (MGD, density of
gaze points) is smaller than TGD (TGD, threshold
of MGD ), the average coordination (XAE,YAE) of
the visual fixations will serve as the central
position to predict the location of the object. The
edges (a and b) of the search box, as in Equation
(5), will be calculated to fulfill the propose on
adjusting and predicting the central position of
the tracked object.
d ij
X AE
X1 X 2 X 3 X N
N
(2)
Y1 Y2 Y3 YN
N
(3)
YAE
M GD
dij
N
X j Yi Y j
2
X j Yi Y j
2
(1)
(4)
a b R M GD
(5)
3 EXPERIMENTAL RESULTS
The accuracy rate of this proposed system is
shown as in Figure 5. The terms, true positive
and true negative, refer to the correct notation of
the detected object (the green rectangle in slash
pattern) and the ground truth object (the blue
rectangle). The terms, false positive and false
negative, refer to the incorrect indication of the
detected object and the ground truth object. The
Figure 4. Adjust Prediction Center
156
Proceedings of The Fourth International Conference on Informatics & Applications, Takamatsu, Japan, 2015
(6)
Pre TP TP FN 100%
(7)
(8)
157
Proceedings of The Fourth International Conference on Informatics & Applications, Takamatsu, Japan, 2015
Frame 21
Frame 28
Frame 31
Figure 7. Object Tracking with Camshift
Frame 112
Frame 21
Frame 28
Frame 31
Figure 8. Object Tracking Dynamic Weights Scheme
Frame 112
158
Proceedings of The Fourth International Conference on Informatics & Applications, Takamatsu, Japan, 2015
Frame 91
Frame 92
Frame 93
Figure 9. Object Tracking with Camshift
Frame 103
Frame 91
Frame 92
Frame 93
Figure 10. Object Tracking with Adaptive Algorithm
Frame 103
Frame 93
Frame 94
Frame 95
Frame 96
Frame 97
Frame 98
Frame 99
Frame 100
Frame 101
Frame 102
4 CONCLUSIONS
From the above experimental results, it is
learned that using the color, shape, and texture
features of the video images to enhance the
comparisons between the characteristic features
and to adjust the weights dynamically can return
better performance on tracking objects. In
addition, adding in the coordinate of users visual
fixations, the adaptive tracking algorithm can
solve the tracking problem on the loss. The
proposed adaptive tracking algorithm can
complement the insufficiency of Camshift for
Camshift using of color feature to make the
judgment on tracking the object. Especially when
159
Proceedings of The Fourth International Conference on Informatics & Applications, Takamatsu, Japan, 2015
ACKNOWLEDGEMENTS
This paper is supported by Ministry of
Science and Technology (MOST), Taiwan, under
Contract of MOST 103-2221-E-130 -011 -.
REFERENCES
[1] H.T. Chen, T.L. Liu, and C.S. Fuh. Learning
Effective Image Metrics from Few Pairwise
Examples, 10th IEEE International Conference on
Computer Vision (ICCV), vol. 2, pp. 1371-1378, 2005.
[2] E. Emami, and M. Fathy. Object tracking using
improved CamShift algorithm combined with motion
segmentation, Machine Vision and Image Processing
(MVIP), 7th Iranian. IEEE, pp. 1-4, 2011..
[3] P. Hidayatullah and H. Konik, CAMSHIFT
Improvement on Multi-Hue and Multi-Object
Tracking, IEEE International Conference on
Electrical Engineering and Informatics (ICEEI), pp.16, 2011.
[4] Z. Zhou, X. Peng, D. Wu, Z. Zhu, C. Wu and J. Wu,
Object Tracking Based on Particle Filter with
Improved Camshift, Journal of Applied Sciences,
Asian Network for Scientific Information, vol. 14, no.
2, pp. 121-128, Oct. 21-23, 2014
[5] J. M. Henderson, and A. Hollingworth, High-Level
Scene Perception, Annual Review of Psychology, vol.
50, pp. 243-271, 1999.
[6] W. Osberger, and A. J. Maeder, Automatic
Identification of Perceptually Important Regions in an
Image, IEEE Proceedings, Fourteenth International
Conference on Pattern Recognition, vol. 1, pp. 701704, 1998.
[7] S.W. Chou, C.H. Hsieh, B.J. Hwang and H.W. Chen,
A Modified Mean Shift Algorithm for Visual Object
Tracking, Signal and Information Processing
Association Annual Summit and Conference
(APSIPA), pp. 1-5, 2013.
[8] S.M. Singh and K. Hemachandran, Content-Based
Image Retrieval using Color Moment and Gabor
Texture Feature, International Journal of Computer
Science Issues (IJCSI), vol. 9, no. 1, pp. 299-309,
2012.
[9] C.W. Kao, Y.W. Chen, C.W. Yang, K.C. Fan, B.J.
Hwang and C.P. Huang, Eye Gaze Tracking based on
Pattern Voting Scheme for Mobile Device, First
International
Conference
on
Instrumentation,
Measurement, Computer, Communication and Control
(IMCCC), pp. 337-340, 2011.
[10] C.W. Kao, Y.J. Huang, K.C. Fan, H.H. Chen, W.C.
Chung, B.J. Hwan and C.H. Hsieh, The integrated
gaze, web and object tracking techniques for the webbased e-learning platform, IEEE International
Conference on Teaching, Assessment and Learning for
Engineering (TALE), pp. 720-724, Bali Dynasty
Resort, Kuta, Indonesia, 2013.
[11] C.W. Kao, B.J. Hwang, C.W. Yang, K.C. Fan and C.P.
Huang, A Novel with Low Complexity Gaze Point
Estimation Algorithm, First International Conference
on
Instrumentation,
Measurement,
Computer,
Communication and Control (IMECS 2012), pp. 204208, March 14-16, Hong Kong, 2012.
[12] Y. Sugano, Y. Matsushita and Y. Sato, AppearanceBased Gaze Estimation Using Visual Saliency, IEEE
Transactions on Pattern Analysis and Machine
Intelligence, vol. 35, no. 2, pp. 329-341, 2013.
[13] K.H. Tan, D. Kriegman and N. Ahuja, Appearancebased eye gaze estimation, In Applications of
Computer Vision (WACV), Proceedings. Sixth IEEE
Workshop, pp. 191-195, 2002.
[14] D.W. Hansen and Q. Ji, In the eye of the beholder: A
survey of models for eyes and gaze, Pattern Analysis
and Machine Intelligence, IEEE Transactions on, vol.
32, no. 3, pp. 478-500, 2010.
[15] H.C. Lu, G.L. Fang, C. Wang and Y.W. Chen, A
novel method for gaze tracking by local pattern model
and support vector regressor, Signal Processing, pp.
1290-1299, 2010.
[16] C.W. Kao, B.J. Hwang, C.W. Yang, K.C. Fan and C.P.
Huang, A Novel with Low Complexity Gaze Point
Estimation Algorithm, Proceedings of the
International MultiConference of Engineers and
Computer Scientists, vol. 1, 2012.
[17] W.T. Li, H.S. Chang, K.C. Lien, H.T. Chang and Y.F.
Wang, Exploring Visual and Motion Saliency for
Automatic Video Object Extraction, IEEE
Transactions on Image Processing, vol. 22, no. 7, pp.
2600-2610, 2013.
160