I know this has been all over the net. I browsed a lot and found a lot of information about Gesture recognition .However i got more confused after reading most of them seem to work on Kinect Official SDK , Microsoft's SDK and i don't want to use them.
I need to develop it for Linux and was wondering if OpenNI, NITE gives me the flexibility to recognize a gesture.
I am trying to develop a sign language program, which recognizes the gestures and draws stuff(spheres,cubes etc) on screen,
Could any one give me a clear guidelines to begin this project.I have no clue where to begin it from.
Any help is welcomed.
Thanks
For getting starting with understanding gestures I suggest checking out a blog post I made a while back: http://www.exceptontuesdays.com/gestures-with-microsoft-kinect-for-windows-sdk-v1-5/
This, plus the post it links to, go into what makes up a "gesture" and how to implement them using Kinect for Windows SDK.
You can also check out thine Kinect Toolbox, which also does gestures for the official Kinect SDK: http://kinecttoolbox.codeplex.com
These will give you a good understanding of how to deal with gestures. You can then move those concepts and ideas into an OpenNI environment.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With