English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Poster

Analyzing haptic and visual object categorization of parametrically-defined shapes

MPS-Authors
/persons/resource/persons83925

Gaissert,  N
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

/persons/resource/persons84298

Wallraven,  C
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

/persons/resource/persons83839

Bülthoff,  HH
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

External Resource
No external resources are shared
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)
There are no public fulltexts stored in PuRe
Supplementary Material (public)
There is no public supplementary material available
Citation

Gaissert, N., Wallraven, C., & Bülthoff, H. (2008). Analyzing haptic and visual object categorization of parametrically-defined shapes. Poster presented at 9th Conference of the Junior Neuroscientists of Tübingen (NeNa 2008), Ellwangen, Germany.


Cite as: https://hdl.handle.net/11858/00-001M-0000-0013-C6D3-E
Abstract
To investigate multi-sensory, perceptual representations of three-dimensional object spaces, we generated complex, shell-shaped objects by altering three parameters defining shell shape. For haptic experiments, 3D-printed plastic models were freely explored by blindfolded participants with both hands. For visual experiments, we used 2D images of these objects. Previously, we reported results of a similarity rating task in which we split the three-dimensional object space into three orthogonal planes. Multidimensional scaling (MDS) of the pair-wise similarity ratings showed that participants reproduced the three planes almost exactly both visually and haptically. Here, we report results of a categorization task in which all objects were presented simultaneously either visually or haptically to ten participants who then categorized the objects in as many groups as they liked to. MDS analyses revealed a three-dimensional perceptual space underlying both visual and haptic data. Interestingly, the three dimensions corresponded to the parameters of shell shape with a different weighting of the dimensions in the visual and the haptic condition. Our results show that humans are able to reproduce the underlying parameters of a complex, three-dimensional object space in a similarity and categorization task using either visual or haptic modalities surprisingly well.