The Efficiency of Multimodal Interaction for a Map-based Task

作者: Philip Cohen , David McGee , Josh Clow

DOI: 10.3115/974147.974193

关键词: Control (management)Human–computer interactionComputer scienceInterface (computing)Graphical user interfaceTask (project management)Mean time to repairMultimodal interaction

摘要: This paper compares the efficiency of using a standard direct-manipulation graphical user interface (GUI) with that QuickSet pen/voice multimodal for supporting military task. In this task, places units and control measures (e.g., various types lines, obstacles, objectives) on map. Four personnel designed entered their own simulation scenarios via both interfaces. Analyses revealed led to an average 3.5-fold speed improvement in entity creation time, including all error handling. The mean time repair errors also was 4.3 times faster when interacting multimodally. Finally, subjects reported strong preference interaction. These results indicate substantial advantage over GUI-based interaction during map-based tasks.

参考文章(19)
Josh Clow, Ira A. Smith, Philip R. Cohen, Michael Johnston, Sharon L. Oviatt, David McGee, The efficiency of multimodal interaction: a case study. conference of the international speech communication association. ,(1998)
Christopher D. Wickens, Diane L. Sandry, Michael Vidulich, Compatibility and Resource Competition between Modalities of Input, Central Processing, and Output: Human Factors. ,vol. 25, pp. 227- 248 ,(1983) , 10.1177/001872088302500209
Alphonse Chapanis, Robert B. Ochsman, Robert N. Parrish, Gerald D. Weeks, Studies in interactive communication. I - The effects of four communication modes on the behavior of teams during cooperative problem-solving. Human Factors. ,vol. 14, pp. 487- 509 ,(1972) , 10.1177/001872087201400601
R. C. Moore, J. Dowding, H. Bratt, J. M. Gawron, Y. Gorfu, A. Cheyer, CommandTalk: A Spoken-Language Interface for Battlefield Simulations conference on applied natural language processing. pp. 1- 7 ,(1997) , 10.3115/974557.974558
Sharon Oviatt, Mutual disambiguation of recognition errors in a multimodel architecture human factors in computing systems. pp. 576- 583 ,(1999) , 10.1145/302979.303163
P. R. Cohen, S. L. Oviatt, The role of voice input for human-machine communication Proceedings of the National Academy of Sciences of the United States of America. ,vol. 92, pp. 9921- 9927 ,(1995) , 10.1073/PNAS.92.22.9921
Bernhard Suhm, Brad Myers, Alex Waibel, Model-based and empirical evaluation of multimodal interactive error correction human factors in computing systems. pp. 584- 591 ,(1999) , 10.1145/302979.303165
Sharon Oviatt, Phil Cohen, Lizhong Wu, Lisbeth Duncan, Bernhard Suhm, Josh Bers, Thomas Holzman, Terry Winograd, James Landay, Jim Larson, David Ferro, Designing the user interface for multimodal speech and pen-based gesture applications: state-of-the-art systems and future research directions Human-Computer Interaction. ,vol. 15, pp. 263- 322 ,(2000) , 10.1207/S15327051HCI1504_1
Clare-Marie Karat, Christine Halverson, Daniel Horn, John Karat, Patterns of entry and correction in large vocabulary continuous speech recognition systems human factors in computing systems. pp. 568- 575 ,(1999) , 10.1145/302979.303160
P. R. Cohen, M. Dalrymple, D. B. Moran, F. C. Pereira, J. W. Sullivan, Synergistic use of direct manipulation and natural language human factors in computing systems. ,vol. 20, pp. 227- 233 ,(1989) , 10.1145/67449.67494