Computational Models of Visual Attention: Bottom-Up and Top-Down. By: Soheil Borhani

Similar documents
Computational Cognitive Science

Computational Cognitive Science. The Visual Processing Pipeline. The Visual Processing Pipeline. Lecture 15: Visual Attention.

Top-down Attention Signals in Saliency 邓凝旖

Overview of the visual cortex. Ventral pathway. Overview of the visual cortex

Control of Selective Visual Attention: Modeling the "Where" Pathway

Models of Attention. Models of Attention

Lateral Geniculate Nucleus (LGN)

On the implementation of Visual Attention Architectures

Advertisement Evaluation Based On Visual Attention Mechanism

Pre-Attentive Visual Selection

In: Connectionist Models in Cognitive Neuroscience, Proc. of the 5th Neural Computation and Psychology Workshop (NCPW'98). Hrsg. von D. Heinke, G. W.

A Model of Saliency-Based Visual Attention for Rapid Scene Analysis

The Integration of Features in Visual Awareness : The Binding Problem. By Andrew Laguna, S.J.

Modeling the Deployment of Spatial Attention

Top-Down Control of Visual Attention: A Rational Account

Measuring the attentional effect of the bottom-up saliency map of natural images

An Information Theoretic Model of Saliency and Visual Search

Selective Attention. Inattentional blindness [demo] Cocktail party phenomenon William James definition

Adding Shape to Saliency: A Computational Model of Shape Contrast

Compound Effects of Top-down and Bottom-up Influences on Visual Attention During Action Recognition

Neurally Inspired Mechanisms for the Dynamic Visual Attention Map Generation Task

Single cell tuning curves vs population response. Encoding: Summary. Overview of the visual cortex. Overview of the visual cortex

intensities saliency map

A context-dependent attention system for a social robot

EDGE DETECTION. Edge Detectors. ICS 280: Visual Perception

Real-time computational attention model for dynamic scenes analysis

Relevance of Computational model for Detection of Optic Disc in Retinal images

A Neural Network Architecture for.

The Role of Top-down and Bottom-up Processes in Guiding Eye Movements during Visual Search

Computational Cognitive Science

M Cells. Why parallel pathways? P Cells. Where from the retina? Cortical visual processing. Announcements. Main visual pathway from retina to V1

Biologically Motivated Local Contextual Modulation Improves Low-Level Visual Feature Representations

Introduction to Computational Neuroscience

Validating the Visual Saliency Model

Theories of Visual Search and Their Applicability to Haptic Search

Efficient Visual Search without Top-down or Bottom-up Guidance: A Putative Role for Perceptual Organization

On the role of context in probabilistic models of visual saliency

Can Saliency Map Models Predict Human Egocentric Visual Attention?

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian

An Evaluation of Motion in Artificial Selective Attention

Neural codes PSY 310 Greg Francis. Lecture 12. COC illusion

Visual Attention. International Lecture Serie. Nicolas P. Rougier. INRIA National Institute for Research in Computer Science and Control

Neuroscience Tutorial

Attention and Scene Understanding

Recurrent Refinement for Visual Saliency Estimation in Surveillance Scenarios

VIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING

A Neural Theory of Visual Search: Recursive Attention to Segmentations and Surfaces

A Computational Model of Saliency Depletion/Recovery Phenomena for the Salient Region Extraction of Videos

Development of goal-directed gaze shift based on predictive learning

An Attentional Framework for 3D Object Discovery

Neurobiological Models of Visual Attention

Goal-directed search with a top-down modulated computational attention system

(Visual) Attention. October 3, PSY Visual Attention 1

The synergy of top-down and bottom-up attention in complex task: going beyond saliency models.

PSYC20007 READINGS AND NOTES

Video Saliency Detection via Dynamic Consistent Spatio- Temporal Attention Modelling

Contribution of Color Information in Visual Saliency Model for Videos

Object detection in natural scenes by feedback

Prof. Greg Francis 7/31/15

How does the ventral pathway contribute to spatial attention and the planning of eye movements?

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

SUN: A Model of Visual Salience Using Natural Statistics. Gary Cottrell Lingyun Zhang Matthew Tong Tim Marks Honghao Shan Nick Butko Javier Movellan

Theoretical Neuroscience: The Binding Problem Jan Scholz, , University of Osnabrück

The discriminant center-surround hypothesis for bottom-up saliency

Spontaneous Cortical Activity Reveals Hallmarks of an Optimal Internal Model of the Environment. Berkes, Orban, Lengyel, Fiser.

Bottom-up and top-down processing in visual perception

Visual Attention Framework: Application to Event Analysis

LISC-322 Neuroscience Cortical Organization

Attention and Scene Perception

Experiences on Attention Direction through Manipulation of Salient Features

A new path to understanding vision

Cognitive Modelling Themes in Neural Computation. Tom Hartley

Parallel streams of visual processing

Biologically Inspired Autonomous Mental Development Model Based on Visual Selective Attention Mechanism

using deep learning models to understand visual cortex

Psychophysical tests of the hypothesis of a bottom-up saliency map in primary visual cortex

Object recognition and hierarchical computation

Keywords- Saliency Visual Computational Model, Saliency Detection, Computer Vision, Saliency Map. Attention Bottle Neck

PSY 310: Sensory and Perceptual Processes 1

Visual Search: A Novel Psychophysics for Preattentive Vision

PHY3111 Mid-Semester Test Study. Lecture 2: The hierarchical organisation of vision

A coincidence detector neural network model of selective attention

SUN: A Bayesian Framework for Saliency Using Natural Statistics

Reading Assignments: Lecture 5: Introduction to Vision. None. Brain Theory and Artificial Intelligence

Object vision (Chapter 4)

Spatial Attention: Unilateral Neglect

Contextual Influences in Visual Processing

CS/NEUR125 Brains, Minds, and Machines. Due: Friday, April 14

REVIEWS COMPUTATIONAL MODELLING OF VISUAL ATTENTION. Laurent Itti* and Christof Koch

Basics of Computational Neuroscience

Recognizing Scenes by Simulating Implied Social Interaction Networks

Attention, short-term memory, and action selection: A unifying theory

Control of visuo-spatial attention. Emiliano Macaluso

Properties of V1 neurons tuned to conjunctions of visual features: application of the V1 saliency hypothesis to visual search behavior

2/3/17. Visual System I. I. Eye, color space, adaptation II. Receptive fields and lateral inhibition III. Thalamus and primary visual cortex

Perception & Attention

Fundamentals of Cognitive Psychology, 3e by Ronald T. Kellogg Chapter 2. Multiple Choice

A saliency map in primary visual cortex

The Eye. Cognitive Neuroscience of Language. Today s goals. 5 From eye to brain. Today s reading

A Dynamical Systems Approach to Visual Attention Based on Saliency Maps

Transcription:

Computational Models of Visual Attention: Bottom-Up and Top-Down By: Soheil Borhani

Neural Mechanisms for Visual Attention 1. Visual information enter the primary visual cortex via lateral geniculate nucleus (LGN). 2. Then, visual information progresses along two parallel streams: Dorsal stream (PPC): Mainly involved in spatial localization and directing attention. Control of attentional deployment takes place in this area. Ventral stream (IT): Mainly concerned with the recognition and identification of visual stimuli. Although not directly involved in the control of attention, ventral stream receives attentional feedback modulation and involved in the representation of attended locations and objects. 2

Classifying computational models of visual attention Saliency map Visual attention models Existence of specific task, targets or intentions Computational models of human visual attention A given image is the only resource, independent of targets and task. Bottom-up models Signal detection theory Knowledge-based model Salient region extraction Top-down models Using prior knowledge Feature gain modulation Weighting BU and TD saliency 3

Visual Search Feature search requires identification of a pop-out target, defined by a single feature such as intensity, color, edge orientation. (e.g., search for the only triangle among other stimuli) Conjunction search requires identification of a target defined by a combination of two or more features (e.g., search for a green triangle among green squares and red triangles & red squares). Feature search Conjunction search We try to examine visual perception via visual search with various relationship of target and distractors 4

Feature Integration Theory (FIT)-(Bottom-up) üproposed by Treisman [1980], the methodology is to measure reaction time of subjects who search for a target among several distractors under the condition of the feature and conjunction searches. üshe argued that the feature search is conducted in parallel and the conjunction search is conducted sequentially. Assumptions: üa constant reaction time independent of the number of distractors is a sign of parallel search. üan increased reaction time with increasing the number of distractors is a sign of sequential search. 5

Critiques on FIT Reaction times can differ greatly according to the given tasks, even with the same sets of visual stimuli. This phenomenon is interpreted in terms of trade-off between the spatial coverage and resolution of attention. Several studies have reported the results indicating that both feature and conjunction search have a difficulty depending on the similarity between targets and distractors as well as the similarity between distractors. Asymmetry between searching result of A among B stimulus and B among A stimulus. 6

Saliency Map (Bottom-Up) Saliency map model proposed by Itti, Koch and Niebur [1998] The proposed architecture introduces a multiresolution structure The structure computes multi-scale spatial contrasts for fine and coarse scales. The saliency map model introduces a winner takesall (WTA) mechanism, which selects the location where the pixel value of the saliency map is greater than at any other locations. 7

Critique on the Saliency map model A deterministic extraction of the Saliency map which implies that all humans would focus on the same location for the same input image. However, according to the findings, humans may focus on different locations in the same input image (Caused by top-down intention, knowledge and preferences) Original scene Itti saliency map Extracted saliency map from eye tracker on one subject 8

Signal Detection Theory (SDT) The theory has been employed in communication theory. Signal detection theory described a mechanism that causes attention ambiguity using only bottom-up processing. Eckstein [2001] applied SDT to model visual attention. FIT à Human eyes cannot make a mistake! SDT à Assumes distractors might be also recognized as a target 9

Signal Detection Theory Takahiko Koike and Jun Saiki [2006] first introduced a stochastic mechanism of human visual attention into a computational model and verified that with psychophysical settings. Pang et al. [2008, 2010] extended this model to video inputs, and constructed a dynamic Bayesian network that considered the stochastic ambiguity and temporal smoothness of visual saliency simultaneously Miyazato et al. [2009] achieved real-time computing of this model with the full use of parallel processors in GPUs. 10

Saliency in temporal changes Itti and Baldi [2009 & 2010] incorporated the temporal dynamics of image features into computational models of human visual attention, and proposed the Bayesian Surprise model that regards the difference between the visual features that are expected to be obtained and those that are actually obtained as indicating saliency. Sophie Marat [2009] utilized the motion features in a computational model of human visual attention. 11

Guided Search Model (Bottom-Up &Top-Down) Proposed by Jeremy Wolfe [1989], it has introduced top-down knowledge on characteristics of target stimuli in visual search. The model employs both bottom-up and top-down activation. The bottom-up features come from saliency in each feature (e.g., color, orientation). The feature maps obtained from feature search. The top-down activation is obtained based on the correlation between input and target stimuli with regard to each of the features. The activation maps is achieved by summing up the top-down and bottom-up activation maps. 12

Top-Down Models Many researchers proposed top-down models including Wolfe s guided search model The top-down models fall into two categories: 1. Computational models based on prior knowledge of the search target in a visual search task. 2. Computational models based on contextual knowledge in a specific task situation. 13

Top-Down Models Top-Down Models Prior knowledge of the search target Contextual knowledge in a specific task Weight modulation of bottom-up features Weighted combination of bottom-up and top-down saliency maps 14

Weight Modulation of Bottom-Up features using Top-down features The weighted response of each channel to the target is compared with its average response to the distractors. The channel with the greatest positive difference is selected to compute the top-down saliency map. The signal-to-noise ratio (SNR), i.e., the ratio between target salience and distractor salience, is effective information for controlling the weights of feature channels. (Frintrop et al. [2006] ) 15

Weighted combination of bottom-up and topdown saliency maps In the original guided search model, Bottom-up saliency is combined using equal weight with top-down saliency calculated using the selected feature channels. The model has evolved through several versions which give rise to Weighted combination of bottom-up and top-down. It has an additional top-down mechanism to handle a scene context and the inhibition tagging mechanism to simulate visual attention more accurately. 16

Thank you! 17