de.mpg.escidoc.pubman.appbase.FacesBean
Deutsch
 
Hilfe Wegweiser Impressum Kontakt Einloggen
  DetailsucheBrowse

Datensatz

DATENSATZ AKTIONENEXPORT

Freigegeben

Poster

Can auditory cues influence the visually induced self-motion illusion?

MPG-Autoren
http://pubman.mpdl.mpg.de/cone/persons/resource/persons84199

Schulte-Pelkum,  J
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

http://pubman.mpdl.mpg.de/cone/persons/resource/persons84170

Riecke,  BE
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

http://pubman.mpdl.mpg.de/cone/persons/resource/persons83846

Caniard,  F
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

http://pubman.mpdl.mpg.de/cone/persons/resource/persons83839

Bülthoff,  HH
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

Externe Ressourcen
Es sind keine Externen Ressourcen verfügbar
Volltexte (frei zugänglich)
Es sind keine frei zugänglichen Volltexte verfügbar
Ergänzendes Material (frei zugänglich)
Es sind keine frei zugänglichen Ergänzenden Materialien verfügbar
Zitation

Schulte-Pelkum, J., Riecke, B., Caniard, F., & Bülthoff, H. (2005). Can auditory cues influence the visually induced self-motion illusion?. Poster presented at 28th European Conference on Visual Perception, A Coruña, Spain.


Zitierlink: http://hdl.handle.net/11858/00-001M-0000-0013-D4D1-1
Zusammenfassung
It is well known that a moving visual stimulus covering a large part of the visual field can induce compelling illusions of self-motion ('vection'). Lackner (1977 Aviation Space and Environmental Medicine 48 129 - 131) showed that sound sources rotating around a blindfolded person can also induce vection. In the current study, we investigated visuo-auditory interactions for circular vection by testing whether adding an acoustic landmark that moves together with the visual stimulus enhances vection. Twenty observers viewed a photorealistic scene of a market place that was projected onto a curved projection screen (FOV 54 deg × 40 deg). In each trial, the visual scene rotated at 30° s-1 around the Earth's vertical axis. Three conditions were randomised in a within-subjects design: no-sound, mono-sound, and spatialised-sound (moving together with the visual scene) played through headphones using a generic head-related transfer function (HRTF). We used sounds of flowing water, which matched the visual depiction of a fountain that was visible in the market scene. Participants indicated vection onset by deflecting the joystick in the direction of perceived self-motion. The convincingness of the illusion was rated on an 11-point scale (0 - 100). Only the spatialised-sound that moved according to the visual stimulus increased vection significantly: convincingness ratings increased from 60.2 for mono-sound to 69.6 for spatialised-sound (t19 = -2.84, p = 0.01), and the latency from vection onset until saturated vection decreased from 12.5 s for mono-sound to 11.1 s for spatialised-sound (t19 = 2.69, p = 0.015). In addition, presence ratings assessed by the IPQ presence questionnaire were slightly but significantly increased. Average vection onset times, however, were not affected by the auditory stimuli. We conclude that spatialised-sound that moves concordantly with a matching visual stimulus can enhance vection. The effect size was, however, rather small (15). In a control experiment, we will investigate whether this might be explained by a ceiling effect, since visually induced vection was already quite strong. These results have important implications for our understanding of multi-modal cue integration during self-motion.