Design and Implementation of User Interface through Hand Movement Tracking and Gesture Recognition

Author(s):  
Jeong Hoon Seong ◽  
Younggeun Choi
2021 ◽  
Vol 297 ◽  
pp. 01030
Author(s):  
Issam Elmagrouni ◽  
Abdelaziz Ettaoufik ◽  
Siham Aouad ◽  
Abderrahim Maizate

Gesture recognition technology based on visual detection to acquire gestures information is obtained in a non-contact manner. There are two types of gesture recognition: independent and continuous gesture recognition. The former aims to classify videos or other types of gesture sequences that only contain one isolated gesture instance in each sequence (e.g., RGB-D or skeleton data). In this study, we review existing research methods of visual gesture recognition and will be grouped according to the following family: static, dynamic, based on the supports (Kinect, Leap…etc), works that focus on the application of gesture recognition on robots and works on dealing with gesture recognition at the browser level. Following that, we take a look at the most common JavaScript-based deep learning frameworks. Then we present the idea of defining a process for improving user interface control based on gesture recognition to streamline the implementation of this mechanism.


Author(s):  
Chris Scogings ◽  
Chris Phillips

The primary focus in UML has been on support for the design and implementation of the software comprising the underlying system. Very little support is provided for the design or evolution of the user interface. This chapter commences with a brief review of UML and its support for user interface modeling. Lean Cuisine+, a notation capable of modeling both dialogue structure and high-level user tasks, is described. It is shown through a case study that Lean Cuisine+ can be used to augment UML and provide the user interface support that is currently lacking.


Sign in / Sign up

Export Citation Format

Share Document