Vision Based Gesture in Embedded Devices
Vision Based Gesture in Embedded Devices
Vision Based Gesture in Embedded Devices
GESTURE RECOGNITION
-- Mukesh Kumar
OUTLINE
Introduction
Input devices
How System Works
Computer Vision Frameworks
OpenCV
Challenges
Few Apps Idead
Demo
Conclusions
1. INTRODUCTION
Computer vision is the science where machine is able to
extract information from an image that is necessary to
solve some task. It enables computer to understand
visual input
The classical problem in computer vision is that of determining whether or not the image data
contains some specific object, feature, activity, gesture, optical character, Motion e.t.c.
This can be achieved using following steps :
Image Acquisition :
Image Acquisition is done to generate a 2D image, 3D depth data, Image sequence from
video feeds at real time or from any other sources using specific hardware.
Pre processing of Images
Noise reduction, Contrast Enhancements, Gray scale conversion, Matrix of Image
creation, Histogram comparison, Image color Inversion.
Few of these steps can also be done during Image Acquisition phase inside the cameras.
Feature Extraction
Line, Edge, Corner Detection
Blobs and shape detection and extraction
Color recognition, like skin color
High Level Processing
High Level Processing is done on the extracted features from the image for Gesture, pose, Motion detection.
Pixel Classification
Image Correlation
Facial Recognition
Hand detection
Feature Tracking
In this phase Neural Network and Artificial intelligence is used to teach the system about the
gesture, kind of motions, faces. Adaptive algorithms are used to adapt the system for working under
different kind of scenarios and environments which learn itself and become more intelligent system after
learning from the various kind of inputs.
Training samples
Training samples
Negative samples: images that must not
contain object representations. We collected 500
random images as negative samples.
Positive samples: hand posture images that are
collected from humans hand, or generated with a
3D hand model. For each posture, we collected
around 450 positive samples. As the initial test, we
use the white wall as the background.
AVAILABLE FRAMEWORKS
Few Computer Vision Frameworks available:
OpenNI – Works with Microsoft Kinect and other 3D sensors generating 3D depth
Images
OpenCV – (started by Intel) Have a C/C++ interface for complex Image Processing
and Applying Neural network Algorithm. Works with 2D or 3D image samples or live
video feeds.
Aforge.Net – Same as OpneCV, with a .NET interface. Have complex Image
Processing and Neural network Algorithm, genetic algorithms and machine learning
libraries for Applications. Subset of this project has a Glyph recognition system.
NokiaCV Library
EmbedCV - An Embeddable Computer Vision Library
The OpenSURF Computer Vision Library
OPENCV
INTEL® OPEN SOURCE COMPUTER VISION LIBRARY
Few of the libraries of Image processing, data structures and XML Support are already ported.
CHALLENGES
Accuracy of gesture recognition software.
Image noise - not necessary be under consistent lighting,
or in the same location. Items in the background or
distinct features of the users may make recognition more
difficult.
Hardware Requirement like 3D depth Mapper.