de.mpg.escidoc.pubman.appbase.FacesBean
Deutsch
 
Hilfe Wegweiser Impressum Kontakt Einloggen
  DetailsucheBrowse

Datensatz

DATENSATZ AKTIONENEXPORT

Freigegeben

Poster

Cross-modal integration of visual and haptic information for object recognition: Effects of view changes and shape similarity

MPG-Autoren
http://pubman.mpdl.mpg.de/cone/persons/resource/persons84940

Lawson,  R
Max Planck Institute for Biological Cybernetics, Max Planck Society;

http://pubman.mpdl.mpg.de/cone/persons/resource/persons83839

Bülthoff,  HH
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

Externe Ressourcen
Es sind keine Externen Ressourcen verfügbar
Volltexte (frei zugänglich)
Es sind keine frei zugänglichen Volltexte verfügbar
Ergänzendes Material (frei zugänglich)
Es sind keine frei zugänglichen Ergänzenden Materialien verfügbar
Zitation

Lawson, R., & Bülthoff, H. (2008). Cross-modal integration of visual and haptic information for object recognition: Effects of view changes and shape similarity. Poster presented at 9th International Multisensory Research Forum (IMRF 2008), Hamburg, Germany.


Zitierlink: http://hdl.handle.net/11858/00-001M-0000-0013-C857-1
Zusammenfassung
Four studies contrasted cross-modal object matching (visual to haptic and haptic to visual) with uni-modal matching (visual-visual and haptic-haptic). The stimuli were hand-sized, plastic models of familiar objects. There were twenty pairs of similarly-shaped objects (cup/jug; frog/lizard, spoon/knife, etc.) and a morph midway in shape between each pair. Objects at fixed orientations were presented sequentially behind an LCD screen. The screen was opaque for haptic inputs and clear for visual presentations. We tested whether a 90º depth rotation from the first to the second object impaired people’s ability to detect shape changes. This achievement of object constancy over view changes was examined across different levels of task difficulty. Difficulty was varied between groups by manipulating shape similarity on mismatch trials. First, view changes from the first to the second object impaired performance in all conditions except haptic to visual matching. Second, for visual-visual matches only, these disruptive effects of task-irrelevant rotations were greater when the task was harder due to increased shape similarity on mismatches. Viewpoint thus influenced both visual and haptic object identification but its effects differed across modalities and for unimodal versus crossmodal matching. These results suggest that the effects of view changes are caused by modality-specific processes.