Smart Cards
Smart Cards
Object tracking is an important task within the field of computer vision. The proliferation
of high-powered computers, the availability of high quality and inexpensive video
cameras, and the interesting need for automated video analysis has generated a great deal
of interest in object tracking.
Numerous approaches for object tracking have been proposed. These primarily differ
from each other based on the way they approach the following question: which object
representation is suitable for tracking? Which image features should be used? How
should the appearance and shape of object be modelled? The answers to these questions
depend on the context/environment in which the tracking is performed. A large number of
tracking methods have been proposed which attempt to answer these questions for variety
of scenarios.
Figure 1 shows the schematic of a generic object tracking system. As can be seen, visual
input is usually achieved through digitized images obtained from a camera connected to a
digital computer. This camera can be either stationary or moving depending on the
application. Beyond image acquisition, the computer performs the necessary tracking and
any higher-level tasks using the tracking result.
Fig 1: schematic of a generic object tracking system. The camera obtains visual images,
and the computer tracks the seen objects.
OBJECT DETECTION
Detecting regions that corresponds to moving objects in video sequence plays a very
important role in many computer vision applications.
In simplest form, Object detection from video sequence is the process of detecting the
moving objects in frame sequence using digital image processing techniques. Moving
object detection is the basis of moving object identification and tracking.
The survey present here covers object detection algorithms appeared in recent literature.
We present the taxonomy of object detection algorithms in which the algorithms are
classified into five major categories. The advantages and disadvantage of the algorithms
considered are tabulated.
CHAPATER 2
LITERATURE SURVEY
2.1. Literature Review
The techniques stated in [3] ranges from very basic algorithm to state of the art published
techniques categorized based on speed, memory requirements and accuracy. They used
methods such as frame difference technique, Real time background subtraction and
shadow detection technique, adaptive background mixture model for real time tracking
technique. They used algorithms ranges from varying levels of accuracy and
computational complexity. Some of them can also deal with real time challenges like
snow, rain, moving branches, objects overlapping, light intensity or slow moving objects.
The problems of achieving high detection rate with low false alarm rate for human
detection and tracking in video sequence is to maximize performance and improve
response time. The stated causes are the effect of scene complexity, scale changes and
scene background-human interactions. A two-step processing solution which is, human
detection, and human tracking with two novel pattern classifiers presented in [4]. There
are three basic phases in video examination: detection of interesting objects in video
scene, tracking of such objects from frame to frame, and analysis of object tracks to
recognize their activities. Detecting humans from video is a challenging problem owing
to the motion of the subjects. In [6] they developed a detector for moving people in
videos with possibly moving cameras and backgrounds, testing several different coding
schemes of moving object and showing that orientated histograms of differential optical
flow give the maximum performance. Motion-based descriptors are combined with
Histogram of Oriented Gradient appearance descriptors. Achieved detector is tested on
several databases includes a challenging test set taken from video and containing wide
ranges of position, motion and background imbalance, including rotating cameras and
backgrounds. [14] In [7], they have analyzed moving object detection techniques, frame
difference and the approximate median method. The frame differentiating has been
adopted for the reference frame and the step length. They have suggested the moving
object detection and object tracking by using the modified frame difference method. In
the surveillance system for video captured by single camera is considered for the space
under the observation. This method is experiment on almost ten videos and the results are
quite satisfactory.
CHAPATER
TECHNICAL ACTIVITIES
Working
The background subtraction method which is used for framing the moving object from its
background which requires a following steps:
a) Reference frame selection (RFS): In it the initial frame is selected as the reference
frame.
b) Step Length: Appropriate step length has been selected on the basis of experimental
results.
c) Removing Noise: Noise is affecting the accuracy and performance of system so it has
to be removed.
d) Moving object detection (MOD): to detect the moving object from the frame
difference with the help of background subtraction methods like Frame difference,
approximate median and Modified frame difference methods.
e) Suspicious Activity: The bounding box is constructed in the isolated area of interest
from video sequence and the object is tracked according to its movement.
f) Rise alert: After tracking object the recorded sound will be generate for alert. Below
image is the example of these above steps.
In [5] cascade-of-rejectors approach with the Histograms of Oriented Gradients features
to achieve a fast and accurate human detection system. The features used are Histograms
of Oriented Gradients of variable-size blocks that capture salient features of humans
automatically. Using algorithm for feature selection, it identifies the appropriate set of
blocks, from a large set of possible blocks. It uses the integral image representation and a
rejection cascade which significantly speed up the computation. For an image, the system
can process 5 to 30 frames per second depending on the density in which it scans the
image, while maintaining an accuracy level similar to existing methods. In [1], author
specified new algorithm for detecting moving objects from a static background scene to
detect moving object based on background subtraction.
OBJECT REPRESENTATION
In a tracking scenario, an object can be defined as anything that is of interest for further
analysis. For instance, boats on the sea, fish inside an aquarium, vehicles on a road,
planes in the air etc are a set of objects that may be important to track in a specific
domain. Objects can be represented by their shapes.
1. Points:
The object is represented by a point, that is, centroid (fig 2(a)) or by a set of points
(fig 2(b)). The point representation is suitable for tracking objects that occupy small
regions in an image.
Fig 2, object representations (a) centroid, (b) multiple points, (c) rectangular patch, (d)
elliptical patch, (e) part-based multiple patches, (f) object skeleton, (g) complete object
contour, (h) control points on object contour, (i) object silhouette.
CHAPATER 4
FEATURE SELECTION FOR TRACKING
Selecting the right features plays a critical role in tracking. The most desirable property of
visual feature is its uniqueness so that the objects can be easily distinguished in the
feature space. In general many tracking algorithms use these features. The details of
visual features are:
1. Color: The apparent color of an object is influenced primarily by two physical factors,
1) the spectral power distribution of the illuminant and 2) the surface reflectance
properties of the objects. In image processing, the RGB (red, green, blue) color space is
usually used to represent color.
2. Edges: Object boundaries usually generate strong changes in image intensities. Edge
detection is used to identify these changes. Algorithms that track the boundary of the
objects usually use edge as the representative feature.
3. Optical Flow: Is a dense field of displacement vectors which defines the translation of
each pixel in a region. It is computed using the brightness constraints, which assumes
brightness constancy of corresponding pixels in the consecutive frames.
4. Texture: Texture is the measure of the intensity variation of the surface whitch
quantifies properties such as smoothness and regularity.
ALGORITHM FOR OBJECT TRACKING
An accurate and fast background subtraction technique for object tracking in still
camera videos. Regions of motion in a frame are first estimated by comparing the current
frame to a previous one. A sampling re-sampling based Bayesian learning technique is
then used on the estimated regions to perform background subtraction and accurately
determine the exact pixels which correspond to moving objects. An obvious advantage in
terms of processing time is gained as the Bayesian learning steps are performed only on
the estimated motion regions, which typically constitute only a small fraction of the
frame. The technique has been used on a variety of indoor and outdoor sequences, to
track both slow and fast moving objects, under different lighting conditions and varying
object-background contrast.
This algorithm presents robust system that achieves both (1) high speed and (2) high
degrees of sensitivity compared to existing techniques. To achieve these objectives a 2
step tracking system has been used.
1) Motion Region Estimation
2) Bayesian Sampling Resampling
Motion Region Estimation: The Block Matching Algorithm (BMA) is a standard way of
encoding video frames. A simplified variation of the BMA algorithm is used for
determining regions of each frame which have had motion relative to a reference frame.
Such regions have been called regions of motion. Each incoming frame is divided into
non-overlapping blocks of equal size. Each block is compared to the corresponding block
in the reference frame and the Sum of Absolute Difference (SAD) is determined for the
block. The reference frame may be chosen to be a few frames before the current frame, to
account for slow moving objects.
Fig 1. Result of motion region estimation
Figure 2: The first row shows original frames from a video sequence. The second row
shows the results of motion region estimation. The third row shows the final Bayesian
Sampling-Resampling results.
1. FRAME DIFFERENCE:
In this method a background image without any moving objects of interest is taken as
reference image. Pixel value for each co-ordinate (x, y) for each color channel of the
background image is subtracted from the corresponding pixel value of the input image. If
the resulting value is greater than a particular threshold value, then that is foreground
pixel otherwise background. This method is simple and easy to implement, but the results
are not accurate enough, because the changes taking place in the background brightness
cause misjudgment.
Edge map (EDGE k-1) of frame (k-1) edge map (EDGE k) of frame (k)
Edge difference image D(x, y) highlighting the difference of Edge maps of frame (k-1)
and frame(k).
Obtained motion areas are then mapped to the original image and appropriate edge pixels
(white pixels) are highlighted.
1.2 A Moving object Detection Algorithm for Smart Cameras
Yongseok Yoo suggested a new frame differencing method for moving object detection
using signed difference and Earth Mover’s Distance (EMD). First, a signed difference
image is acquired by subtracting two consecutive frames. For each fixed block in the
signed difference image, a motion pattern is calculated by EMD. The EMD is defined as
the minimum total amount of cost to move piles of earth to holes until all the earth is
moved or all the holes are filled. The neighboring blocks are then linked to detect moving
object regions.
The main idea behind this algorithm is to calculate matching costs for given
directions separately rather then to calculate exact EMD by linear programming. Here
block-based motion is used to locate moving object regions. An input image is divided
into blocks of fixed size and pairing vectors are calculated for each block. Blocks with
large pairing vectors indicate that there are motions in them. By combining these blocks,
moving objects can be detected.
Flow diagram:
Original image Signed difference
Result
1.3 An Automatic Moving Object Detection Algorithm for Video
Surveillance Applications
Xiaoshi Zheng proposed an automatic moving object detection algorithm based on frame
difference and region combination for video surveillance applications. Initially an
automatic threshold calculation method is used to 0obtain moving pixels of video frames.
Frame difference is obtained by absolute difference value of two frames. Moving pixels
and static background pixels can be distinguished by a threshold value. In order to make
all moving pixels continuous and filter isolated pixels, moving regions are obtained by
morphological CLOSE operation.
In this algorithm we have three phases i.e. 1) moving object detection phase 2)
moving object extraction phase 3) moving object recognition phase
The second phase is the extraction of a detected object. In this phase the output
image which was obtained at the end of the first phase is used. A simple but efficient way
of extracting the object is to use horizontal and vertical scanning of the output image. In a
two-dimensional matrix an object could be addressed by finding its vertical and
horizontal coordinates. The width and height of the extracted object image can be found
by using its starting coordinates and ending coordinates.
The implementation of this extraction method has two steps. Firstly, the output
image matrix is scanned horizontally, starting at the first x-direction coordinate value, the
pixel values in the corresponding column are summed up. The x direction coordinate
value is incremented by one and total pixel value in the next column is calculated. This
process is repeated until the last value in the x-direction. As result the total pixel values of
each column is calculated. Each total value is compared to a certain threshold value in
order to determine the x-coordinate where an object starts or ends within the image.
Secondly, the horizontal scanning method is repeated vertically, thus calculating the total
pixel value in each row, and then apply thresholding to determine the y coordinate where
an object starts or ends within the image. The background of the image that contains the
object is uniform as it has already been set to white or black at the end of the first phase.
2. BACKGROUND SUBTRACTION:
In this method, the moving regions are detected by subtracting the current image
pixel-by-pixel from a reference background image. The pixels where the difference is
above a threshold are classified as foreground otherwise background. Some
morphological post processing operations are performed to reduce noise and enhance the
detected region.
FLOW DIAGRAM:
Fig 1 Fig 2
EXPERIMENT:
Lianqiang Niu describes an algorithm for moving object dtection based on combining of
improved background subtraction. This method can improve the detection speed and
overcome the lack of sensitivity to light changes. Considering the pixels relativity,
Gaussian Mixture Model in background subtraction is used. To extract a motion region,
the differences between the current frame and its previous frame is calculated.
After getting the motion scene background by improved Gaussian Mixture Model,
the foreground image is extracted. Foreground image is obtained by subtracting the
current image frame from background image. Symmetrical differencing is used to detect
the undetected regions. At each position of the pixel, the foreground images which are
achieved by using background subtraction and symmetrical differencing are processed by
a logical OR operation to obtain an accurate foreground image.
In background updating, the background of the selected pixels are replaced by the
average of the current and background pixels.
Shows the result of origin MOG detection and legs which are overlap with floor can’t be
detected completely.
Shows the detected result of improved MOG, and legs which were overlap with floor can
be detected complete
Shows the color image segmentation result with the edged image
Background model
Show the final detected result of joint color image segmentation and background
5. CROSS CORRELATION:
Manoj S Nagmode described a method to detect and track the moving objects to detect
and track the moving objects by using Normalized Cross Correlation algorithm (NCC). In
this an approach is proposed for the detection and tracking of moving object in an image
sequence. Two consecutive frames from image sequence are partitioned into four
quadrants and then the Normalized Cross Correlation (NCC) is applied to each sub frame.
The sub frame which has minimum value of NCC indicates the presence of moving
object. Next step is to identify the location of the moving object. Location of the moving
object is obtained by performing component connected analysis and morphological
processing. After that the centroid calculation is used to track the moving object. Number
of experiments performed using indoor and outdoor image sequences. The results are
compared with Simple Difference (SD) method. The proposed algorithm gives better
performance in terms of Detection Rate (DR) and processing time per frame.
The tracking sequence of a walking person. This walking person is pointed by a red star.
Algorithm based on FD & Higher recognition rate and False detection under
edge detection higher detection speed complicated background
Algorithm for smart cameras Reject false motions due to Falsely detect specular
illumination changes reflections from moving
objects
Real time moving object Extract moving object regions Processing time is strictly
detection for video accurately and completely depends on the quality of
monitoring systems moving points and on the
image dimension
Partitioning & normalized Poor lighting conditions Average processing time per
cross correlation algorithm giving better results frame is high
CHAPATER
APPLICATIONS
Traffic information
SURVEILLANCE
CHAPATER
CONCLUSION
Object tracking means tracing the progress of objects as they move about in visual scene.
Object tracking, thus, involves processing spatial as well as temporal changes. Certain
features of those objects have to be selected for tracking. These features need to be
matched over different frames. Significant progress has been made in object tracking.
Taxonomy of moving object detection is been proposed. Performance of various object
detection is also compared. It is not possible to consider a single method for all type of
images, nor can all methods perform well for particular types of image. The background
subtraction method detects object with noise and output is not accurate. Object behind
object is not detected. Problem occurs during identification of object when any obstacles
come before the object. If the position of camera is not proper and object in image is not
captured properly then it cannot be identified. To solve above problems and bring some
accuracy and richness by combining multiple methods and make use of it together
according to the application.
REFERENCES
[1] Rupali S.Rakibe, Bharati D.Patil,”Background Subtraction Algorithm Based Human
Motion Detection - published at: "International Journal of Scientific and Research
Publications (IJSRP), Vol. 3, Issue 5, May 2013 Edition".
[3] Mr. Deepjoy Das and Dr. Sarat Saharia,” Implementation and Performance
Evaluation of Background Subtraction Algorithms”, International Journal on
Computational Sciences & Applications (IJCSA) Vol.4, No.2, April 2014.
[4] Thomas Andzi-Quainoo Tawiah,” Video Content Analysis For Automated Detection
And Tracking Of Humans In cctv Surveillance Applications”, School of Engineering and
Design Brunel University, August 2010.
[5] Qiang Zhu, Shai Avidan, Mei-Chen Yeh, Kwang-Ting Cheng,” Fast Human Detection
Using a Cascade of Histograms of Oriented Gradients”, June 2006.
[6] Navneet Dalal, Bill Triggs, and Cordelia Schmid,” Human Detection Using Oriented
Histograms of Flow and Appearance”, April 2006.
[7] Seema Kumari, Manpreet Kaur, Birmohan Singh,” Detection And Tracking Of
Moving Object In Visual Surveillance System”, International Journal of Advanced
Research in Electrical Electronics and Instrumentation Engineering (IJAREEIE) Vol. 2,
Issue 8, August 2013.
[8] Khushboo Khurana, Reetu Awasthi, “Techniques for Object Recognition in Images
and Multi-Object Detection”, International Journal of Advanced Research in Computer
Engineering & Technology (IJARCET) Volume 2, Issue 4, April 2013.