Speech perception in virtual environments

作者: Johan Verwey

DOI:

关键词: Surround soundSpeech perceptionAnimationVisual perceptionSound cardPerceptionSpeech recognitionComputer scienceVirtual machineAmbient noise level

摘要: Many virtual environments like interactive computer games, educational software or training simulations make use of speech to convey important information the user. These applications typically present a combination background music, sound effects, ambient sounds and dialog simultaneously create rich auditory environment. Since allow users roam freely among different producing objects, designers do not always have exact control over what user will perceive at any given time. This dissertation investigates factors that influence perception in under adverse listening conditions. A environment was created study hearing performance audio-visual conditions. The two main areas investigation were contribution "spatial unmasking" lip animation perception. Spatial unmasking refers benefit achieved when target masking are presented from locations. Both visual influencing considered. The capability modern hardware produce spatial release using real-time 3D spatialization compared with pre-computed method creating spatialized sound. It found could be consumer card either headphone surround speaker display. Surprisingly, less effective subjects better than used. Most research on has been conducted pure environments. an additional cue first investigated determine whether this provided benefit. No difference observed visible objects same location as stimuli. Because inherent limitations display devices, often perfectly aligned, causing sound-producing object seen where it is heard. integration between conflicting see had noise. significant regardless stimuli correct matching position, spatially disparate source. Lastly rudimentary investigated. results showed animations significantly contribute also incorrect result worse no used all. The conclusions are: That capabilities can should speech; Perfectly align very for perception; Even enhance

参考文章(75)
Barbara G. Shinn-Cunningham, Antje Ihlefeld, Selective and Divided Attention: Extracting Information from Simultaneous Sound Sources. international conference on auditory display. ,(2004)
Virginia Best, Andre van Schaik, Simon Carlile, 7TWO-POINT DISCRIMINATION IN AUDITORY DISPLAYS Georgia Institute of Technology. ,(2003)
Claudia Hendrix, Woodrow Barfield, The sense of presence within auditory virtual environments Presence: Teleoperators & Virtual Environments. ,vol. 5, pp. 290- 301 ,(1996) , 10.1162/PRES.1996.5.3.290
Durand R. Begault, Elizabeth M. Wenzel, Mark R. Anderson, Direct comparison of the impact of head tracking, reverberation, and individualized head-related transfer functions on the spatial perception of a virtual speech source. Journal of The Audio Engineering Society. ,vol. 49, pp. 904- 916 ,(2001)
Vijay Madisetti, The Digital Signal Processing Handbook CRC Press , IEEE Press. ,(2009) , 10.1201/9781482275193
Ville Pulkki, Virtual Sound Source Positioning Using Vector Base Amplitude Panning Journal of The Audio Engineering Society. ,vol. 45, pp. 456- 466 ,(1997)
A.S. Bregman, Auditory Scene Analysis International Encyclopedia of the Social & Behavioral Sciences. ,vol. 3, pp. 861- 870 ,(2008) , 10.1016/B978-012370880-9.00009-8
Perceptual consequences of interpolating head-related transfer functions during spatial synthesis Proceedings of IEEE Workshop on Applications of Signal Processing to Audio and Acoustics. pp. 102- 105 ,(1993) , 10.1109/ASPAA.1993.379986