0% found this document useful (0 votes)
39 views5 pages

Hand Gestures

Uploaded by

shaileshpawark11
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
39 views5 pages

Hand Gestures

Uploaded by

shaileshpawark11
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 5

International Journal of Science and Research (IJSR)

ISSN (Online): 2319-7064


Index Copernicus Value (2013): 6.14 | Impact Factor (2013): 4.438

Navigation of PowerPoint Using Hand Gestures


Dnyanada R Jadhav1, L. M. R. J Lobo2
1
M.E Department of Computer Science & Engineering, Walchand Institute of technology, Solapur, India
2
Associate Professor in Department of Computer Science & Engineering, Walchand Institute of technology, Solapur, India

Abstract: Computer applications require interaction between human and computer. This interaction needs to be unrestricted and it
made challengeable accounts as compared to traditional input devices such as keyboard, mouse, pen etc. Hand gestures are used in
daily life while communication. They are more natural in interaction, compared with those devices mentioned above. Human computer
interaction becomes easy, with use of hand as a device. Use of hand gestures to operate machine would make interaction interesting.
Gesture recognition has gained a lot of importance. Hand gestures are used to control various applications like VLC media player,
robot control, gaming, virtual mouse etc. Use of gesture makes interaction easy, convenient and does not require any extra device.
Vision and audio recognition can be used together. But audio commands may not work in noisy environment. The presented developed
system uses hand gestures to control power point navigation. Here both static and dynamic gestures are used together for navigation.
Use of gestures would make the presenter to handle the presentation more easily.

Keywords: Hand gestures, skin segmentation, active fingers, finger count.

1. Introduction system uses both static and dynamic gestures for slide show
control. In this system the images are captured through the
Gesture recognition has gained a lot of importance since few webcam. It is segmented to recognize the hand region. A skin
years. Various applications can be controlled using gestures. color detection algorithm is used for hand region detection.
Face gestures like lip movements are used to recognize the The binary image that is generated is given to the distance
language spoken, hand gestures are used in various transform method that calculates the centroid of the hand and
applications like gaming, mouse control etc. In an application using this it recognizes the gesture. Accordingly the slide
like robot control using hand gestures, the robot responds to show is controlled.
hand gestures given by the human. This hand sign of humans
is visually observed by a robot through a camera. The 2. Related Work
algorithm that enables the robot to identify the hand gesture
from the image is of interest. Each gesture corresponds to a Many applications have been developed that are controlled
particular command. The command that is identified will be through gestures that include gaming, sign language
used to control the robot to perform certain action or to recognition, control through facial gestures, controlling
execute a certain task. Different gestures will have different mouse, VLC media player etc. In 2012 [1] Ruize Xu, Shengli
meaning associated with them. For example, count one could Zhou and Wen J. Li developed a system that recognizes
mean stop, two for move forward, three, four and five for seven hand gestures like up, down, right, left, cross and
turn right, turn left and reverse respectively. The hand circle. Three different modules were built that recognized
gesture recognition system makes use of gloves, markers etc. hand gesture. Signals from MEMS 3-axes accelerometers
Though the use of gestures increases the interactivity were given as input. The motion of the hand in three
between man and machine, the use of such gloves or markers perpendicular direction is detected by 3 accelerometers and
increases the cost of the system. Some applications require transmitted to the system through Bluetooth. Segmentation
the use of two cameras to obtain a 3D view of hand and from algorithm was applied and finally the gestures are recognized
which a hand gesture is recognized. by comparing gestures that are already stored in the system.
People always use the internet to get daily information about
Two types of hand gestures are used, static and dynamic. weather, news etc. For this they have to repeat same
Static gestures make gestures by keeping the hand stable. For keyboard and mouse actions. In 2011 [2] Kuan-Ching Li,
example, by using the finger without moving the hand, the Hwei-Jen Lin, Sheng-Yu Peng, Kanoksak Wattanachote used
system would perform the specified function. Dynamic hand moments to retrieve information from the internet
gestures are those that involve the movement of the hand. which reduces time and also convenient to use. Once the user
Like in VLC media player that is controlled by hand gestures, provides the gesture, appropriate function is selected, then
moving the hand to the right side may indicate increasing the the system will report the action to the user in form of
volume. For some application hand gesture recognition speech. This system also uses face recognition to identify and
system may require to store images in the database. personalize each user as requirements are different for
Execution of these applications may require the use of different users. They used the PCA method to recognize hand
complex algorithm to compare images already stored in the and face. Hand gestures were acquired and stored in the
database with that of images taken from the camera and then system and the later compared this with input gestures and
perform necessary tasks. For such applications gestures perform tasks. This system results better with the small scale
should be known prior the use as they are already stored in of face recognition and hand gesture recognition. In 2011 [3]
the database. In a Hand Gesture Recognition System to Ginu Thomas presented an article on A Review of Various
control Slide Navigation there is no use of database. The Hand Gesture Recognition Techniques where he compared

Volume 4 Issue 1, January 2015


www.ijsr.net
Paper ID: SUB15298 833
Licensed Under Creative Commons Attribution CC BY
International Journal of Science and Research (IJSR)
ISSN (Online): 2319-7064
Index Copernicus Value (2013): 6.14 | Impact Factor (2013): 4.438
the results obtained by different techniques. The different control systems, test and measurement, and computational
methods used are edges method, pixel by pixel comparison biology. Array is basic data element of MATLAB that does
and orientation histogram. Image database was used that not require dimensioning thus allows solving many technical
contained various static hand gestures images. These images computing problems. The reason for choosing MATLAB for
are subset of American sign languages. Filtering of the image development is its toolboxes. Toolboxes allow to learn and
was done to reduce noise present in it and then segmentation to apply specialized technology. Toolboxes are collections
to analyze it. It was then transformed into feature vector and of MATLAB functions i.e M-files that are used to solve
then compared with a trained set of gestures. In 2010 [4] particular classes of problems.
Anupam Agrawal and Siddharth Swarup Rautaray used hand
gestures to control the VLC media player. The K nearest 4. Methodology
neighbor algorithm was used to recognize the gesture. A
VLC media player function that were controlled by hand The System architecture is as shown in Fig (1); it consists of
gesture includes play, pause, Fullscreen, stop, increase image acquisition, segmentation of hand region, finger count
volume, decrease volume. Lucas Kanade Pyramidical Optical recognition and motion recognition and finally slide show
Flow algorithm was used to detect hand from video. This control.
algorithm detects moving points in the image. After this K
_MEAN was used to find a center of the hand. Using this
centre motion of the hand is recognized. This system used
database that contains different gestures and then input was
compared with this stored image and accordingly VLC media
player was controlled. In 2007 [5] Yikai Fang, Jian Cheng
and Hanqing Lu, Kongqiao Wang recognized hand through
trigger followed by tracking and segmentation and used this
gesture for image browsing. Segmentation is done during
detection and tracking using motion and color cues. Palm and
finger like structures are then determined using Scale Scape
features. Using this palm finger configuration, hand gesture is Figure 1: System architecture
determined. In 2007 [6] Yikai Fang, Jian Cheng, Hanqing
Lu, Kongqiao Wang speed up hand gesture recognition. They Image Acquisition
used integral image approximation for Gaussian derivation in The user makes gestures by positioning hand parallel to
image convolution. Fast multi-scale feature detectors were webcam. Images are continuously captured and then given as
constructed to speed up computation. In 2006 [7] Erol Ozgur, Input for segmentation.
santerabi Malima controlled robot using hand gestures but
considered limited gestures. Firstly segmentation of hand Hand Detection
region was carried followed by locating the fingers and then The primary step in gesture recognition systems is the
finally classifying the gestures. detection of hands and the segmentation of the corresponding
image regions. This segmentation is crucial because it
In 2002 [9] Lars Bretzner, Ivan Laptev, Tony Lindeberg isolates the task-relevant data from the image background,
presented algorithms for hand tracking and hand posture before passing them to the subsequent tracking and
recognition. In this on each image multi-scale color feature recognition stages. A large number of methods have been
detection is performed. Using particle filtering, with an proposed in the literature that utilize a several types of visual
extension of layeredsampling referred to as hierarchical features and, in many cases, their combination. Such features
layered sampling Hands are detected and tracked. In 1994 are skin color, shape, motion and anatomical models of
[10] William T. Freeman, Craig D. Weissman used hand hands.
gestures and controlled the television. Here a user makes a
gesture facing the camera. On the display the hand icon Segmentation of hand region
appears that follows the user. User can move his hands on The images captured are given for analysis which is done
various graphical controls using this hand icon and control using segmentation. Here skin detection algorithm is used
the television. Gestures for a television control like television that detects the skin region from the input image as the
off, television on, channel control, channel change were used. background may consist of many other things along with the
hand region. The video obtained through a webcam is in the
3. Matlab 2012 RGB color model. This video is converted to HSV color
MATLAB means matrix laboratory. It is a high-level model because the regions which belong to the skin can be
language and it is interactive environment for visualization, easily identified in HSV model. Following this, the rules for
numerical computation and programming. MATLAB is used skin segmentation are applied. After recognizing hand it is
to develop algorithms, analyze data, create models and converted into a binary image. The skin regions are
applications. Availability of tools and built in math functions represented using white color and all other non-skin regions
helps to achieve solution faster than with traditional are black. The largest connected region, which is detected as
programming languages such as Java or C/C++. MATLAB skin is taken as the hand region. This gives the segmented
can be used for image and video processing, signal hand region and this is the region of interest. The recognition
processing and communications, computational finance, of the gestures depends on this region.

Volume 4 Issue 1, January 2015


www.ijsr.net
Paper ID: SUB15298 834
Licensed Under Creative Commons Attribution CC BY
International Journal of Science and Research (IJSR)
ISSN (Online): 2319-7064
Index Copernicus Value (2013): 6.14 | Impact Factor (2013): 4.438
RGB to HSV conversion • Check for all the pixels that are on boundary and assign it
some value
ALGORITHM • For each pixel in image an image other than boundary
• Load image. image
• Read each pixel from image. a. Calculate distance of that pixel from every pixel on
• Separate RGB color for each pixel. boundary
R = col & 0xff; • Assign some value for the pixel whose distance from every
G = (col >> 8) & 0xff; boundary pixel is maximum
B= (col >> 16) & 0xff; • This pixel is the centroid of the image. Using this centroid
• Find minimum value and maximum value from R,G,B. slide show is controlled.
• Assign max to value.
• If value equal to zero then 5. Results
assign hue=saturation =0.
set pixel in image again Following snapshots illustrates the results achieved.
if valueof (Lum) = 0 then output color is black i.e there is Initially we get the screen shown in snapshot 1.
no brightness in the color hence Hue and Saturation will It contains various buttons, like Master initialization, start,
be zero. stop, previous slide, previous two slides, centre slide, next
slide, and next two slides. In the screenshot can also see the
Else button as Set PPT, it is used to select the ppt that we want to
//Formula for finding Saturation. navigate.
Find saturation= 255 * (Max-Min)/value.
if saturation = 0 then
assign hue=0.
set pixel
end if
Else
if max equal to R then
//Formula for finding Hue.
hue = 0 + 43*(G-B)/(max-min).
End if.
If max is equals G then
hue = 85 + 43 * (B-R) / (max-min).
End if
If max is equal to B then
hue = 171 + 43*(B-R)/(max-min).
End if.
If hue<0 then
hue=hue+255. Snapshot 1: Initial Screen
End if.
End if. We can see three segments showing three different screens.
• Set each pixel again on image. First screen is used for master initialization, next is actual
end. image captured through webcam and the last shows the
After this we get binary image binary image that is obtained using skin color segmentation
algorithm. When the hand is detected, immediately is marked
Distance transform method for gesture recognition with circle as shown in snapshot 2.

After segmentation the binary image is given to the distance


transform method that recognizes the gesture. In this method
firstly the centroid of palm is calculated by considering each
pixel and calculating distance from nearest boundary.
Therefore the pixel that is far from every boundary is chosen
as centroid. Using this centroid active fingers are counted
and if there is motion of hand, this is detected by motion of
centroid from original position from a set of continuously
captured images and the slide show is controlled that is
PowerPoint presentation either goes to the next slide,
previous slide or start after recognizing static gesture.

ALGORITHM
• Consider binary image.
Snapshot 2: Detection of hand

Volume 4 Issue 1, January 2015


www.ijsr.net
Paper ID: SUB15298 835
Licensed Under Creative Commons Attribution CC BY
International Journal of Science and Research (IJSR)
ISSN (Online): 2319-7064
Index Copernicus Value (2013): 6.14 | Impact Factor (2013): 4.438
Initially we do master initialization as shown in below
snapshot 3.

For this we drag to the point that is concentrated on the hand


region.

This master initialization is done so that it can recognize


hand of different skin tone.

Snapshot 5: Displays that the next slide button is triggered


after gesture recognition

Snapshot 5, Displays that the next slide button is triggered


after gesture recognition

Comparison between distance transform method and circular


profiling method for centroid detection is shown in below
Snapshot 3: Initialization of hand region graph that results distance transform method is better in
performance than circular profiling method.
We can see in snapshot 3 that we have segmented the region,
as shown in green color.

6. Conclusion
The presented gesture recognition system recognizes both
static and dynamic gestures. Gestures are recognized using
distance transform method along with skin color
segmentation algorithm. The presented method does not
require any training phase to identify the hand gestures.
Hence does not require storage of images in database to
recognize the hand gestures. A recommended future work
Snapshot 4: Displays the binary image of the hand would include, increasing the number of gestures along with
speech recognition that make it interesting and easy to
As shown in snapshot 4, whenever we drag our hand in navigate power point which will include creating slides,
segment as shown we can see that Next slide button is adding contents etc. Furthermore gestures can be used to
triggered. This means we navigate to next slide. control various applications like adobe reader, paint and add
further by controlling computer using gestures.

Volume 4 Issue 1, January 2015


www.ijsr.net
Paper ID: SUB15298 836
Licensed Under Creative Commons Attribution CC BY
International Journal of Science and Research (IJSR)
ISSN (Online): 2319-7064
Index Copernicus Value (2013): 6.14 | Impact Factor (2013): 4.438
References
[1] Dnyanada Jadhav, L.M.R.J. Lobo, “Hand Gesture
Recognition System To Control Slide Show
Navigation”, IJAIEM, 2014
[2] Ruize Xu, Shengli Zhou, Wen J. Li, “MEMS
Accelerometer Based Nonspecific-UserHand Gesture
Recognition”, IEEE, 2011.
[3] Sheng-Yu Peng, Kanoksak Wattanachote, Hwei-Jen Lin,
Kuan-Ching Li, “A Real time hand gesture recognition
system for daily information retrieval from Internet”,
IEEE Fourth International Conference on Ubi-Media
Computing, 2011
[4] Ginu Thomas, “A Review of Various Hand Gesture
Recognition Techniques”, 2011.
[5] Siddharth Swarup Rautaray and Anupam Agrawal, “A
Vision based Hand Gesture Interface for Controlling
VLC Media Player”, International Journal of Computer
Applications, 2010.
[6] Yikai Fang, Kongqiao Wang, Jian Cheng, Hanqing Lu,
“A Real-Time Hand Gesture Recognition Method”,
IEEE, 2007.
[7] Yikai Fang, Jian Cheng, Kongqiao Wang, Hanqing Lu,
“Hand Gesture Recognition Using Fast Multi-scale
Analysis”, IEEE, 2007.
[8] Asanterabi Malima, Erol Ozgur, “A Fast Algorithm For
Vision-Based Hand Gesture Recognition For Robot
Control”, IEEE, 2006.
[9] Ahmed Elgammal, Vinay Shet, Yaser Yacoob, Larry S.
Davis, “Learning Dynamics for Exemplar-based Gesture
Recognition”, IEEE, 2003.
[10] Lars Bretzner, Ivan Laptev, Tony Lindeberg, “Hand
Gesture Recognition using Multi-Scale Color Features,
Hierarchical Models and Particle Filtering”, IEEE, 2002.
[11] William T. Freeman, Craig D. Weissman, “Television
Control by Hand Gestures”, 1994.

Author Profile
Miss Dnyanada Jadhav, received B.E degree in Computer Science
and Engineering in 2012 from Walchand Institute Of Technology
Solapur. Solapur University, Maharashtra, India and pursuing the
M. E. degree in Computer Science and Engineering in Walchand
Institute of Technology, Solapur, India. She is doing her
dissertation work under the guidance of Mr. Lobo L.M.R.J,
Associate Professor& Head, Department of IT, Walchand Institute
of Technology, Solapur Maharashtra, India. Her research area is
Image Processing.

Mr. Lobo L. M. R. J received the B.E degree in Computer


Engineering in 1989 from Shivaji University, Kolhapur, India and
M. Tech degree in Computer and Information Technology in 1997
from IIT, Kharagpur, India. He is registered for Ph.D in Computer
Science and Engineering at SGGS, Nanded of Sant Ramanand
Teerth Marathawada University, Nanded, India. Under the guidance
of Dr. R.S. Bichkar. He is presently working as an Associate
Professor & Head, Department of IT Walchand Institute of
Technology, Solapur, Maharashtra, India. His research interests
include Evolutionary Computation, Genetic Algorithms and Data
Mining.

Volume 4 Issue 1, January 2015


www.ijsr.net
Paper ID: SUB15298 837
Licensed Under Creative Commons Attribution CC BY

You might also like