Multimodal Interaction Control

نویسندگان

  • Jonas Beskow
  • Rolf Carlson
  • Jens Edlund
  • Björn Granström
  • Mattias Heldner
  • Anna Hjalmarsson
  • Gabriel Skantze
چکیده

No matter how well hidden our systems are and how well they do their magic un-noticed in the background, there are times when direct interaction between system and human is a necessity. As long as the interaction can take place unobtrusively and without techno-clutter, this is desirable. It is hard to picture a means of interaction less obtrusive and techno-cluttered than spoken communication on human terms. Spoken face-to-face communication is the most intuitive and robust form of communication between humans imaginable. In order to exploit such human spoken communication to its full potential as an interface between human and machine, we need a much better understanding of how the more human-like aspects of spoken communication work. A crucial aspect of face-to-face conversation is what people do and what they take into consideration in order to manage the flow of the interaction. For example, participants in a conversation have to be able to identify places where it is legitimate to begin to talk, as well as to avoid interrupting their interlocutors. The ability to indicate that you want to say something, that somebody else may start talking, or that a dialog partner should refrain from doing so is of equal importance. We call this interaction control. Examples of the features that play a part in interaction control include the production and perception of auditory cues such as intonation patterns, pauses, voice quality, and various disfluencies; visual cues such as gaze, nods, facial expressions, gestures, and visible articulatory movements; and content cues like pragmatic and semantic (in)completeness. People generally seem to use these cues in combination, and to mix them or shift between them seamlessly. By equipping spoken dialog systems with more human-like interaction control abilities, we aim to move interaction between system and human toward the intuitive and robust communication among humans. The bulk of work on interaction control in CHIL has been focused on auditory prosodic cues, but visual cues have also been explored, and especially through the use of embodied conversational agents (ECAs) – human-like representations of a system, for example, animated talking heads that are able to interact with a user in a natural way using speech, gesture, and facial expression. ECAs are one way of leveraging

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Achieving Multimodal Cohesion during Intercultural Conversations

How do English as a lingua franca (ELF) speakers achieve multimodal cohesion on the basis of their specific interests and cultural backgrounds? From a dialogic and collaborative view of communication, this study focuses on how verbal and nonverbal modes cohere together during intercultural conversations. The data include approximately 160-minute transcribed video recordings of ELF interactions ...

متن کامل

Multimodal Interaction: Intuitive, Robust, and Preferred?

We investigated if and under which conditions multimodal interfaces (touch, speech, motion control) fulfil the expectation of being superior to unimodal interfaces. The results show that the possibility of multimodal interaction with a handheld mobile device turned out to be more intuitive, more robust, and more preferred than the interaction with the individual modalities speech and motion con...

متن کامل

Multimodal Pain Management Protocol Versus Patient Controlled Narcotic Analgesia for Postoperative Pain Control after Shoulder Arthroplasty

  Background: Our institution’s traditional pain management strategy after shoulder arthroplasty has involved the utilization of postoperative patient-controlled narcotic analgesia. More recently, we have implemented a protocol (TLC) that utilizes a multimodal approach. The purpose of this study was to determine whether this change has improved pain control and decreased narcotic utilization. ...

متن کامل

Speech and Gesture Interaction in an Ambient Assisted Living Lab

In this paper we describe our recent and future research on multimodal interaction in an Ambient Assisted Living Lab. Our work combines two interaction modes, speech and gesture, for multiple device control in Ambient Assisted Living environments. We conducted a user study concerning multimodal interaction between participants and an intelligent wheelchair in a smart home environment. Important...

متن کامل

Bringing User Experience empirical data to gesture-control and somatic interaction in virtual reality videogames: an Ex- ploratory Study with a multimodal interaction prototype

With the emergence of new low-cost gestural interaction devices various studies have been developed on multimodal human-computer interaction to improve user experience. We present an exploratory study which analysed the user experience with a multimodal interaction game prototype. As a result, we propose a set of preliminary recommendations for combined use of such devices and present implicati...

متن کامل

Designing robust multimodal systems for diverse users and environments

Multimodal interfaces are being developed that permit our highly skilled and coordinated communicative behavior to control system interactions in a more transparent and flexible interface experience than ever before. The presence of modality choice per se is an important feature and design issue for multimodal interfaces. As applications become more complex, a single modality does not permit va...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009