The Infona portal uses cookies, i.e. strings of text saved by a browser on the user's device. The portal can access those files and use them to remember the user's data, such as their chosen settings (screen view, interface language, etc.), or their login data. By using the Infona portal the user accepts automatic saving and using this information for portal operation purposes. More information on the subject can be found in the Privacy Policy and Terms of Service. By closing this window the user confirms that they have read the information on cookie usage, and they accept the privacy policy and the way cookies are used by the portal. You can change the cookie settings in your browser.
Quality of food and agricultural products is vital for farmers and consumers. Quality based classification of these products is being carried out manually in the industry which is tedious and expensive. Computer Vision systems can be used to automate the classification process. Automation can reduce the production cost and improve the overall quality. A computer vision system captures the image of...
In this study we have developed a real time gestural interface based on 3D dynamic hand gesture recognition using Hidden Markov Models (HMM). We developed a system, which captures and recognizes hand gestures of the user wearing colored gloves, where the hand coordinates are obtained via 3D reconstruction from stereo. The gestural interface provides supplementary features such as an interactive training...
Using sign language, speech, and haptics as communication modalities, a virtual treasure-hunting game serves as an entertainment and educational tool for visually and hearing impaired users. In recent years, there has been an increasing interest in human-computer interaction (HCI) for multimodal interfaces. Since Sutherland's SketchPad in 1961 and Xerox's Alto in 1973, computer users have long been...
In this paper we present a review of human social interaction analysis based on audio and visual nonverbal cues. Furthermore, as an example study, we present our study on automatic dominance estimation in small group conversations. We extracted low level audio and visual features, defined in parallel to the nonverbal cues displayed by dominant people, as stated in social psychology literature. We...
This paper presents a method to extract isolated signs from continuous sign language videos. We use sequences that approximately contain the sign that we are interested in and align the sequences to find the exact start and end frames. We compare different feature extraction methods, different alignment methods, and assess the performance of our system on a database from Turkish sign language.
Language learning can only advance with practice and corrective feedback. The interactive system, SignTutor, evaluates users' signing and gives multimodal feedback to help improve signing.
In this work, we propose a multi-class classification strategy based on Fisher kernels. Fisher kernels combine the powers of discriminative and generative classifiers by mapping variable-length sequences to a new fixed length feature space. The mapping is based on a single generative model and the classifier is intrinsically binary. We apply a multi-class classification, instead of a binary classification,...
This paper presents a particle filter based algorithm for tracking face and hands of a signer. During signing, the hands and the face occlude each other frequently and a proper multiple object tracking algorithm is needed for accurate results. We use separate particle filters for the two hands and the face, where each filter effects the particle weights of others. This methodology ensures that the...
We present a system that integrates gesture recognition and 3D talking head technologies for a patient communication application at a hospital or healthcare setting for supporting patients treated in bed. As a multimodal user interface, we get the input from patients using hand gestures and provide feedback by using a 3D talking avatar.
Multimodal speech and speaker modeling and recognition are widely accepted as vital aspects of state of the art human-machine interaction systems. While correlations between speech and lip motion as well as speech and facial expressions are widely studied, relatively little work has been done to investigate the correlations between speech and gesture. Detection and modeling of head, hand and arm gestures...
In this study, using a two camera setup, we designed a system that recognizes 3D gestures. When 3D reconstruction is not possible or infeasible, combining 2D hand trajectories at feature or decision level increases the system performance drastically. The trajectories are extracted by tracking the center-of-mass of the hand and the width, height and orientation of the enclosing ellipse. Trajectories...
Set the date range to filter the displayed results. You can set a starting date, ending date or both. You can enter the dates manually or choose them from the calendar.