Gesture Recognition for Interactive Presentation Control a Deep Learning and Edge Computing Approach on Raspberry Pi
Gesture Recognition for Interactive Presentation Control a Deep Learning and Edge Computing Approach on Raspberry Pi
Prashanth N
Assistant Professor
Dept of Artificial Intelligence and Machine Learning Vidyavardhaka College of Engineering, Gokulum, Mysore
BE Students
Dept of Artificial Intelligence and Machine Learning Vidyavardhaka College of Engineering, Gokulum, Mysore
Abstract Keywords
3.2 A Deep Learning Framework for The system encounters challenges in maintain- ing
Recognizing Both Static and Dynamic accuracy under non-ideal lighting and is con-
Gestures[2] strained by a reliance on predefined gestures.
Mediapipe’s performance varies under certain
3.2.1 Methodologies conditions, which can impact gesture recognition
accuracy and usability in diverse environments.
taDNet uses pose-driven spatial attention with
the use of OpenPose for extracting pose and
3.3.2 Learnings
depth estimation. The former is captured by a
CNN where LSTMs are utilized for modeling the This paper provides useful methods for gesture
temporal aspects of dynamic gestures. Fur- detection, particularly through Mediapipe and
thermore, cropping RGB frames around hand OpenCV, which you can use for real-time hand
regions fosters gesture detection in a sensor-less tracking in your presentation system. Imple-
scenario while normalizing joint velocities and menting similar gesture-mapping techniques al- lows
accelerations improve accuracy. you to customize commands for slides and
multimedia control. Leveraging TensorFlow Lite
models on Raspberry Pi can enhance gesture
3.2.2 Limitations
recognition speed, while data augmentation can
Since the system uses RGB images, it is light address lighting variability for more robust per-
sensitive and best suited for indoor use. It does formance
not support continuous gesture recognition but
instead works only on single/isolated gestures. 3.4 Human-machine interactions based on
hand gesture recog- nition using deep learning
methods[4]
3.2.3 Learnings
3.4.1 Methodologies
The approach of StaDNet can thus adapt for
the presentation control system with pose-driven The paper uses CNNs for extracting key ges- ture
spatial attention and architecture CNN-LSTM features and LSTMs to capture temporal patterns,
with TensorFlow Lite running on the Raspberry improving gesture accuracy. Medi- aPipe detects
Pi. In this way, gesture recognition may be en- hand landmarks (e.g., fingertips) in real-time
abled in real time, with a reliability improve- video, enabling gesture-based con- trol. Data on
ment of variable lighting conditions using intu- hand movements is mapped to specific actions,
itive gestures to support the control of slides in a and these mappings train CNN and LSTM models
presentation. for tasks like cursor control.
3.4.2 Limitations
3.3 Exploring The Techniques and
Challenges exist in achieving big labeled datasets
Applications of Hand Ges- ture Recognition for
for training while, at the same time, there are real-
Human- Computer Interaction [3]
time processing challenges on de- vices like the
3.3.1 Methodologies Raspberry Pi. Gesture recognition performance
may also be sensitive to environ- mental factors,
The paper utilizes OpenCV and Mediapipe for such as lighting and background, which can impact
real-time video capture and hand tracking. Me- the accuracy.
5 Implications and Consider- ations with a cost-effective tool. The work has very
for Future Work important implications that can be exploited to
transform technology which is more widely ac-
• Challenging Theoretical Frameworks and cessible to all sorts of users, therefore improv- ing
Assumptions: The findings may cast doubt on the systems in human-computer interaction sys- tems.
viability of utilising edge com- puting and deep
learning for real-time ges- ture detection on
devices with limited re- sources, such as
Raspberry Pi.
References
• Suggestions for Subsequent Research:
Future research could concentrate on inves- [1] Ali A, Abed SA, Rahman O. Python-based
tigating hybrid models (such as CNNs and RNNs) Raspberry Pi for Hand Gesture Recognition.
for improved dynamic gesture recog- nition and International Journal of Computer Applica- tions.
optimising models to lower la- tency and 2017;173(4):18–24.
computational load.
[2] Osama M, Sofiane R, Andrea C. A Deep
• Implications for Practitioners: Practi- Learning Framework for Recognizing Both Static
tioners could use these findings to create more and Dynamic Gestures. Sensors. 2021;21(6):2227.
efficient and accessible gesture-based control [3] Anonymous. Exploring The Techniques and
systems, particularly for hands-free Applications of Hand Gesture Recognition for
environments. Human-Computer Interaction. Indian Scientific
Journal Of Research In Engineer- ing And
• Influence on Programs, Methods, and
Management. 2023;07(05).
Interventions: This study could lead to more
inclusive and user-friendly technolo- gies for [4] Zholshiyeva L, Manbetova Z, Kaibassova D,
individuals with disabilities or for applications in Kassymova A, Tashenova Z, Baizhumanov S, et al.
interactive settings. Human-machine interactions based on hand gesture
recognition using deep learning methods.
• Contribution to Solving Problems: It International Journal of Electrical and Computer
may help solve accessibility issues by devel- oping Engineering. 2024.
technology that doesn’t rely on tradi- tional input
devices, improving inclusivity. [5] Magar P, Singh S, Mishra S, Gaikwad P.
Carry Out Computer Tasks with Gesture us- ing
• Improvements or Changes Needed: Image Pre-Processing and TensorFlow
Improvements should focus on enhancing gesture Framework. International Journal of Engi- neering
recognition robustness, addressing environmental Research and. 2020;9(09).
challenges, and optimizing real-time processing on
devices like Rasp- berry Pi. [6] Bansal JL. Gesture-based Human-Computer
Interaction using Wearable Devices. Interna- tional
Journal for Research Publication and Seminar.
6 Conclusions
2023.
This work applies deep learning and edge com-
puting on a Raspberry Pi to exhibit the signif-
icance of gesture detection in interactive pre-
sentation control. Currently, the study focuses on
the requirement of hands-free control sys- tems
applied in many areas, including assistive
technology and public speaking. This disser-
tation aims to look into the implementation of
deep learning models, namely CNNs and RNNs, on
edge devices so as to provide real-time ges- ture
recognition facilities for interactive presen- tations.
The chosen approach employs deep learning
frameworks and picture pre-processing utilizing
Raspberry Pi to achieve real-time performance