Aviator Hand Tracking Based on Depth Images

Author(s):  
Xiaolong Wang ◽  
Shan Fu
Keyword(s):  
2013 ◽  
Vol 765-767 ◽  
pp. 2822-2825 ◽  
Author(s):  
Lin Song ◽  
Rui Min Hu ◽  
Yu Lian Xiao ◽  
Li Yu Gong

In this paper, we propose a depth image based real-time 3D hand tracking method. Our method is based on the fact that human hand is an end point of human body. Therefore, we locate human hand by finding the end point from a predicted position of hand based on the hand position of the previous frame. We iteratively grow a region around the predicted position. The end point on the major axis of the region which stops moving with region growing is selected as the final position of human hand. Experiments on Microsoft Kinect for Xbox captured sequences show the effectiveness and efficiency of our proposed method.


Author(s):  
Chia-Ping Chen ◽  
Yu-Ting Chen ◽  
Ping-Han Lee ◽  
Yu-Pao Tsai ◽  
Shawmin Lei
Keyword(s):  

Author(s):  
Sukhendra Singh ◽  
G. N. Rathna ◽  
Vivek Singhal

Introduction: Sign language is the only way to communicate for speech-impaired people. But this sign language is not known to normal people so this is the cause of barrier in communicating. This is the problem faced by speech impaired people. In this paper, we have presented our solution which captured hand gestures with Kinect camera and classified the hand gesture into its correct symbol. Method: We used Kinect camera not the ordinary web camera because the ordinary camera does not capture its 3d orientation or depth of an image from camera however Kinect camera can capture 3d image and this will make classification more accurate. Result: Kinect camera will produce a different image for hand gestures for ‘2’ and ‘V’ and similarly for ‘1’ and ‘I’ however, normal web camera will not be able to distinguish between these two. We used hand gesture for Indian sign language and our dataset had 46339, RGB images and 46339 depth images. 80% of the total images were used for training and the remaining 20% for testing. In total 36 hand gestures were considered to capture alphabets and alphabets from A-Z and 10 for numeric, 26 for digits from 0-9 were considered to capture alphabets and Keywords. Conclusion: Along with real-time implementation, we have also shown the comparison of the performance of the various machine learning models in which we have found out the accuracy of CNN on depth- images has given the most accurate performance than other models. All these resulted were obtained on PYNQ Z2 board.


Sign in / Sign up

Export Citation Format

Share Document