de.mpg.escidoc.pubman.appbase.FacesBean
English
 
Help Guide Disclaimer Contact us Login
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Journal Article

A saliency-based bottom-up visual attention model for dynamic scenes analysis

MPS-Authors
http://pubman.mpdl.mpg.de/cone/persons/resource/persons84847

Ramirez-Villegas,  JF
Max Planck Institute for Biological Cybernetics, Max Planck Society;

Locator
There are no locators available
Fulltext (public)
There are no public fulltexts available
Supplementary Material (public)
There is no public supplementary material available
Citation

Ramirez-Moreno, D., Schwartz, O., & Ramirez-Villegas, J. (2013). A saliency-based bottom-up visual attention model for dynamic scenes analysis. Biological Cybernetics, 107(2), 141-160. doi:10.1007/s00422-012-0542-2.


Cite as: http://hdl.handle.net/11858/00-001M-0000-0013-B484-F
Abstract
This work proposes a model of visual bottom-up attention for dynamic scene analysis. Our work adds motion saliency calculations to a neural network model with realistic temporal dynamics [(e.g., building motion salience on top of De Brecht and Saiki Neural Networks 19:1467–1474, (2006)]. The resulting network elicits strong transient responses to moving objects and reaches stability within a biologically plausible time interval. The responses are statistically different comparing between earlier and later motion neural activity; and between moving and non-moving objects. We demonstrate the network on a number of synthetic and real dynamical movie examples. We show that the model captures the motion saliency asymmetry phenomenon. In addition, the motion salience computation enables sudden-onset moving objects that are less salient in the static scene to rise above others. Finally, we include strong consideration for the neural latencies, the Lyapunov stability, and the neural properties being reproduced by the model.