An Implementation of Multimodal User Interface using Speech, Image and EOG
نویسندگان
چکیده
There have been many recent studies on gaze recognition system in the field of HCI (Human Computer Interaction). This system will be the most natural and intuitive HCI system due to the application of gaze direction or biomedical Signals. We propose a multimodal user interface system using the nine directional gaze recognition based on image, EOG (Electrooculography) signal and speech recognition. In this paper, we use DFA (Deterministic Finite Accepter), Haar-like feature and Adaboost algorithm, SVM (Support Vector Machine) for gaze recognition based on the EOG signal and image. Furthermore, the CHMM (Continuous Hidden Markov Model) based speech recognition system has been linked for inputting speech commands. The proposed multimodal user interface system solves the problem of constraint of single modal recognition systems. As a result, the proposed system achieves higher recognition performance and more natural interface using speech commands.
منابع مشابه
Design and Implementation of a Semantic Dialogue System for Radiologists
This chapter describes a semantic dialogue system for radiologists in a comprehensive case study within the large-scale MEDICO project. MEDICO addresses the need for advanced semantic technologies in the search for medical image and patient data. The objectives are, first, to enable a seamless integration of medical images and different user applications by providing direct access to image sema...
متن کاملHMM Based Continuous EOG Recognition for Eye-input Speech Interface
To provide an efficient means of communication for those who cannot move muscles of the whole body except eyes due to amyotrophic lateral sclerosis (ALS), we are developing a speech synthesis interface that is based on electrooculogram (EOG) input. EOG is an electrical signal that is observed through electrodes attached on the skin around eyes and reflects eye position. A key component of the s...
متن کاملA Multimodal Lego Robot
The goal of the project was to develop a robot and a multimodal user interface. The robot, designed as a digital cat, can show complex behaviours such as move, speak, touch, listen, and read. The input command interface is based on text, icons, and speech. A prototype of the robot is implemented using the Lego MindstormsTM System. The design and implementation of the robot are presented in this...
متن کاملSpeech and Gesture Multimodal Control of a Whole Earth 3D Visualization Environment
A growing body of research shows several advantages to multimodal interfaces including increased expressiveness, flexibility, and user freedom. This paper investigates the design of such an interface that integrates speech and hand gestures. The interface has the additional property of operating relative to the user and can be used while the user is in motion or standing at a distance from the ...
متن کاملiVIEW: An Intelligent Video over InternEt and Wireless Access System
We describe the design and implementation of a digital video content management system, iVIEW, for intelligent searching and access of video contents over Internet and wireless devices. The iVIEW system allows full content indexing, searching and retrieval of multilingual text, audio and video material. iVIEW integrates image processing techniques for scenes and scene changes analyses, speech p...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2011