CaltechAUTHORS
  A Caltech Library Service

Modeling feature sharing between object detection and top-down attention

Walther, Dirk B. and Serre, Thomas and Poggio, Tomaso and Koch, Christof (2005) Modeling feature sharing between object detection and top-down attention. Journal of Vision, 5 (8). Art. No.1041. ISSN 1534-7362. http://resolver.caltech.edu/CaltechAUTHORS:20130816-103301458

[img]
Preview
PDF - Published Version
See Usage Policy.

1807Kb

Use this Persistent URL to link to this item: http://resolver.caltech.edu/CaltechAUTHORS:20130816-103301458

Abstract

Visual search and other attentionally demanding processes are often guided from the top down when a specific task is given (e.g. Wolfe et al. Vision Research 44, 2004). In the simplified stimuli commonly used in visual search experiments, e.g. red and horizontal bars, the selection of potential features that might be biased for is obvious (by design). In a natural setting with real-world objects, the selection of these features is not obvious, and there is some debate which features can be used for top-down guidance, and how a specific task maps to them (Wolfe and Horowitz, Nat. Rev. Neurosci. 2004). Learning to detect objects provides the visual system with an effective set of features suitable for the detection task, and with a mapping from these features to an abstract representation of the object. We suggest a model, in which V4-type features are shared between object detection and top-down attention. As the model familiarizes itself with objects, i.e. it learns to detect them, it acquires a representation for features to solve the detection task. We propose that by cortical feedback connections, top-down processes can re-use these same features to bias attention to locations with higher probability of containing the target object. We propose a model architecture that allows for such processing, and we present a computational implementation of the model that performs visual search in natural scenes for a given object category, e.g. for faces. We compare the performance of our model to pure bottom-up selection as well as to top-down attention using simple features such as hue.


Item Type:Article
Related URLs:
URLURL TypeDescription
http://dx.dpoi.org/10.1167/5.8.1041 DOIConference proceedings abstract
Additional Information:Received September 15, 2005. Thanks to Xinpeng Huang for labeling the training and test images. The model figures are modified from Serre and Poggio, Cosyne 2005. This research is funded by grants from NSF, NIH, and NIMH.
Group:Koch Laboratory, KLAB
Funders:
Funding AgencyGrant Number
NSFUNSPECIFIED
NIHUNSPECIFIED
NIMHUNSPECIFIED
Record Number:CaltechAUTHORS:20130816-103301458
Persistent URL:http://resolver.caltech.edu/CaltechAUTHORS:20130816-103301458
Official Citation:Dirk Walther, Thomas Serre, Tomaso Poggio, and Christof Koch. J Vis September 23, 2005 5(8): 1041; doi:10.1167/5.8.1041
Usage Policy:No commercial reproduction, distribution, display or performance rights in this work are provided.
ID Code:40608
Collection:CaltechAUTHORS
Deposited By: KLAB Import
Deposited On:16 Jan 2008 02:54
Last Modified:09 Sep 2013 21:47

Repository Staff Only: item control page