de.mpg.escidoc.pubman.appbase.FacesBean
English
 
Help Guide Disclaimer Contact us Login
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Poster

A Computational Mid-Level Vision Approach For Shape-Specific Saliency Detection

MPS-Authors
http://pubman.mpdl.mpg.de/cone/persons/resource/persons83871

Curio,  C
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

http://pubman.mpdl.mpg.de/cone/persons/resource/persons83902

Engel,  D
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;

Locator
There are no locators available
Fulltext (public)
There are no public fulltexts available
Supplementary Material (public)
There is no public supplementary material available
Citation

Curio, C., & Engel, D. (2010). A Computational Mid-Level Vision Approach For Shape-Specific Saliency Detection. Poster presented at 10th Annual Meeting of the Vision Sciences Society (VSS 2010), Naples, FL, USA.


Cite as: http://hdl.handle.net/11858/00-001M-0000-0013-C05E-6
Abstract
We present a novel computational approach to visual saliency detection in dynamic natural scenes based on shape centered image features. Mid-level features, such as medial features, have been recognized as important entities in both human object recognition and in computational vision systems [Tarr Buelthoff 1998, Kimia 2003]. [Kienzle et al 2009] have shown how image driven gaze predictors can be learned from fixations during free viewing of static natural images and result in center-surround receptive fields. Method: Our novel shape-centered vision framework provides a measure for visual saliency, and is learning free. It is based on the estimation of singularities of long ranging gradient vector flow (GVF) fields that have originally been developed for the alignment of image contours [Xu Prince 1998]. The GVF uses an optimization scheme to guarantee preservation of gradients at contours and, simultaneously, smoothness of the flow field. The specific properties are similar to filling-in processes in the human brain. Our method reveals the properties of medial-feature shape transforms and provides a mechanism to detect shape specific information, local scale, and temporal change of scale, in clutter. The approach generates a graph which encodes the shape across a scale-space for each image. Results: We have made medial-feature transforms amenable to work in cluttered environments and have demonstrated temporal stability thus providing a mechanism to track shape over time. The approach can be used to model eye tracking data in dynamic scenes. A fast implementation will provide a useful tool for predicting shape-specific saliency at interactive framerates.