The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian

Similar documents
Computational Cognitive Science

On the implementation of Visual Attention Architectures

Validating the Visual Saliency Model

Computational Cognitive Science

A Model of Saliency-Based Visual Attention for Rapid Scene Analysis

Computational Cognitive Science. The Visual Processing Pipeline. The Visual Processing Pipeline. Lecture 15: Visual Attention.

Can Saliency Map Models Predict Human Egocentric Visual Attention?

Advertisement Evaluation Based On Visual Attention Mechanism

Biologically Inspired Autonomous Mental Development Model Based on Visual Selective Attention Mechanism

Real-time computational attention model for dynamic scenes analysis

An Evaluation of Motion in Artificial Selective Attention

A Model for Automatic Diagnostic of Road Signs Saliency

Reading Assignments: Lecture 18: Visual Pre-Processing. Chapters TMB Brain Theory and Artificial Intelligence

Development of goal-directed gaze shift based on predictive learning

Video Saliency Detection via Dynamic Consistent Spatio- Temporal Attention Modelling

VENUS: A System for Novelty Detection in Video Streams with Learning

Computational Models of Visual Attention: Bottom-Up and Top-Down. By: Soheil Borhani

Keywords- Saliency Visual Computational Model, Saliency Detection, Computer Vision, Saliency Map. Attention Bottle Neck

Application of Tree Structures of Fuzzy Classifier to Diabetes Disease Diagnosis

Finding Saliency in Noisy Images

Relevance of Computational model for Detection of Optic Disc in Retinal images

A Computational Model of Saliency Depletion/Recovery Phenomena for the Salient Region Extraction of Videos

Learning Utility for Behavior Acquisition and Intention Inference of Other Agent

Compound Effects of Top-down and Bottom-up Influences on Visual Attention During Action Recognition

Experiences on Attention Direction through Manipulation of Salient Features

Goal-directed search with a top-down modulated computational attention system

Neurally Inspired Mechanisms for the Dynamic Visual Attention Map Generation Task

Comparison of Mamdani and Sugeno Fuzzy Interference Systems for the Breast Cancer Risk

An Attentional Framework for 3D Object Discovery

Control of Selective Visual Attention: Modeling the "Where" Pathway

Oscillatory Neural Network for Image Segmentation with Biased Competition for Attention

HUMAN VISUAL PERCEPTION CONCEPTS AS MECHANISMS FOR SALIENCY DETECTION

Visual Task Inference Using Hidden Markov Models

M.Sc. in Cognitive Systems. Model Curriculum

Lung Cancer Diagnosis from CT Images Using Fuzzy Inference System

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China

Fusing Generic Objectness and Visual Saliency for Salient Object Detection

Detection of terrorist threats in air passenger luggage: expertise development

MEMORABILITY OF NATURAL SCENES: THE ROLE OF ATTENTION

Dynamic Visual Attention: Searching for coding length increments

Sensory Cue Integration

Object detection in natural scenes by feedback

A HMM-based Pre-training Approach for Sequential Data

Computational Cognitive Neuroscience Approach for Saliency Map detection using Graph-Based Visual Saliency (GBVS) tool in Machine Vision

USING AUDITORY SALIENCY TO UNDERSTAND COMPLEX AUDITORY SCENES

The discriminant center-surround hypothesis for bottom-up saliency

Skin color detection for face localization in humanmachine

Realization of Visual Representation Task on a Humanoid Robot

Minimum Feature Selection for Epileptic Seizure Classification using Wavelet-based Feature Extraction and a Fuzzy Neural Network

Object-based Saliency as a Predictor of Attention in Visual Tasks

Introduction to Computational Neuroscience

Empirical Validation of the Saliency-based Model of Visual Attention

Increasing Spatial Competition Enhances Visual Prediction Learning

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation

LUNG NODULE SEGMENTATION IN COMPUTED TOMOGRAPHY IMAGE. Hemahashiny, Ketheesan Department of Physical Science, Vavuniya Campus

Available online at ScienceDirect. Procedia Computer Science 54 (2015 )

Incorporating Audio Signals into Constructing a Visual Saliency Map

A context-dependent attention system for a social robot

A Bayesian Hierarchical Framework for Multimodal Active Perception

Knowledge-driven Gaze Control in the NIM Model

A Scoring Policy for Simulated Soccer Agents Using Reinforcement Learning

[EN-A-022] Analysis of Positive and Negative Effects of Salience on the ATC Task Performance

Attention Estimation by Simultaneous Observation of Viewer and View

Computational Saliency Models Cheston Tan, Sharat Chikkerur

Gender Based Emotion Recognition using Speech Signals: A Review

Neuromorphic convolutional recurrent neural network for road safety or safety near the road

Introduction to Machine Learning. Katherine Heller Deep Learning Summer School 2018

A Visual Saliency Map Based on Random Sub-Window Means

Models of Attention. Models of Attention

Spatial Cognition for Mobile Robots: A Hierarchical Probabilistic Concept- Oriented Representation of Space

International Journal for Science and Emerging

Available online at ScienceDirect. Procedia Computer Science 93 (2016 )

Recurrent Refinement for Visual Saliency Estimation in Surveillance Scenarios

Top-down Attention Signals in Saliency 邓凝旖

A Dynamical Systems Approach to Visual Attention Based on Saliency Maps

Segmentation of Cell Membrane and Nucleus by Improving Pix2pix

Eye movements, recognition, and memory

Edge Detection Techniques Using Fuzzy Logic

Subjective randomness and natural scene statistics

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING

Modeling the Deployment of Spatial Attention

A Decision-Theoretic Approach to Evaluating Posterior Probabilities of Mental Models

Detection of Inconsistent Regions in Video Streams

Cognitive Maps-Based Student Model

A Hierarchical Visual Saliency Model for Character Detection in Natural Scenes

Practical Bayesian Optimization of Machine Learning Algorithms. Jasper Snoek, Ryan Adams, Hugo LaRochelle NIPS 2012

PathGAN: Visual Scanpath Prediction with Generative Adversarial Networks

VIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING

A Knowledge Driven Computational Visual Attention Model

Relative Influence of Bottom-up & Top-down Attention

A Fuzzy Expert System for Heart Disease Diagnosis

Local Image Structures and Optic Flow Estimation

Online Speaker Adaptation of an Acoustic Model using Face Recognition

Type-2 fuzzy control of a fed-batch fermentation reactor

Neuroinformatics. Ilmari Kurki, Urs Köster, Jukka Perkiö, (Shohei Shimizu) Interdisciplinary and interdepartmental

Cancer Cells Detection using OTSU Threshold Algorithm

Adaptation of Classification Model for Improving Speech Intelligibility in Noise

EEL 6586, Project - Hearing Aids algorithms

A prediction model for type 2 diabetes using adaptive neuro-fuzzy interface system.

Errol Davis Director of Research and Development Sound Linked Data Inc. Erik Arisholm Lead Engineer Sound Linked Data Inc.

Transcription:

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) A Fuzzy Inference Method Based on Saliency Map for Prediction Mao Wang, Yoichiro Maeda 2, Yasutake Takahashi Graduate School of Engineering, University of Fukui 2 Faculty of Engineering, Osaka Institute of Technology Abstract: Visual attention region prediction has been paid much attention by researchers in intelligent systems recent years because it can make the interaction between human and intelligent agents to be more convenient. In this paper, the prediction method of the visual attention region inferred by using fuzzy inference after extracting and computing of graphics s feature maps and saliency map is proposed. A user experiment was also conducted to evaluate the prediction effect of proposed method by making surveys for the prediction results. The results indicated that prediction method proposed by us has a better performance in the level of attention regions position prediction according to different graphics. Introduction Recently, the intention recognition, recognizing the intention of a user or an agent by analyzing their actions or changes of state, is becoming an important issue in various research fields of intelligent systems. Especially, the intention recognition can make the Human-Computer Interaction (HCI) more convenient. So far, many intention recognition approaches have been proposed. Much of early work is in the context of speech understanding and response automatically[]. For example, Pynadath et al. achieved plan recognition on a problem in traffic monitoring through exploited the context by using a general Bayesian framework [2]. More recently, Pereira et al.[3] described an approach to tackle intention recognition by combining dynamically configurable and situation-sensitive Causal Bayes Networks plus plan generation techniques[4][]. Mao et al. have presented a utilitybased approach to solve recognition of intention, which is realized by incrementally using plan knowledge and observations to change state probabilities [6]. In their researches, probability is the main factor which used to infer the human intention. Another method is use an automatic capture of bottom-up salient stimuli and volitional shifts guided by and topdown context factors[7][8], where bottom-up salient stimuli are the external factors to the user and topdown contexts are the internal factors. The characteristic of all the methods mentioned above is that one or several characteristic values of the graphics showing before user have been extracted, calculated and combined as their basis for inferring user s intention. In this paper, a visual attention region prediction system inspired on saliency map is described. The aim of this work is to present a new approach that improves the performance of attention prediction based on saliency map by using fuzzy inference. The fuzzy inference employing features of graphics as input allows us to combine features and infer with great flexibility some intuitive decision rules based on the visual perception principles. 2 Overview of Proposed Approach We propose a new approach to predict visual attention region based on graphics s saliency map which got by using fuzzy inference. The overall procedural flow of proposed approach is summarized in Fig.. Firstly, intensity, color (red, green, blue and yellow) and orientation (degrees of 0, π/4, π/2, 3π/4) are extracted in multi-resolution from the Gaussian pyramid by linear filtering. Then a saliency map is generated for each of them by computing the difference between the layers of the pyramids, which imitates 49

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian Pyramid (9 levels) C I O Intensity Feature Maps Center-Surround Difference Normalization Intensity Saliency Map Fuzzy Inference Saliency Map Orientation Feature Maps Orientation Saliency Map Fig. : Overall Procedure for Proposed System the center-surround type receptive field. Finally, the saliency map is generated by fuzzy inference using the three saliency maps as inputs. 3 Saliency Map by Fuzzy Inference Most attention models are based on a saliency map and a dynamical process for visiting saliency maxima. Itti et al.[7] introduced a model for bottomup selective attention based on serially scanning a saliency map, which is computed from local feature contrasts, for salient locations in the order of decreasing saliency. The saliency map is entirely based on features of a graphics and was originally designed to explain covert attention on simple stimuli. A lot of researches on saliency map are getting some features of graphics and combining them by simply sum in mathematics[8]. But this method also has its weakness. For example, if the saliency map of a graphics is based on the three features which are color, intensity and orientation. The method of simply sum of them gives them the same importance at the same time. But based on experiments[9], most people do not pay equal attention to all of them. Actually, for example, the color feature in a graphics takes more attention from observers than other two features, which shows that the method may not very reasonable in some situation. In order to solve this problem, we proposed a method to compute saliency map by fuzzy inference based on the features of graphics. In this way, the importance of all features can be reflected in fuzzy rule respectively. 3. Feature Maps In Fig., the linear filter is used in order to compute center-surround different of various features at 9 scales. In this paper, the input image is sub-sampled into a dyadic Gaussian pyramid by convolution with a linearly separable Gaussian filter and decimation by a factor of two[]. After filtering, the three features of graphics have their values at each position according to the input image, which are divided into 9 levels of pyramid ready to be calculated. Then, in this paper, the color feature is reflected by two values defined by us, which are red-green and blue-yellow opponencies. If r, g, and b are the red, green, and blue values of the input color image respectively. Then the color map of one level can be calculated according to the following equation. M r g = r g max(r, g, b) () b min(r, g) M b y = (2) max(r, g, b) Where M r g, M b y stand for red-green and blueyellow opponencies. And note that the definitions deviate from the original model by[]. The intensity map of one level is calculated as: M i = r + g + b (3) 3 These operations are repeated for each level of the input pyramid to obtain an intensity pyramid with also 9 levels. Local orientation maps M o is obtained by applying steerable filters to the intensity pyramid levels M i [2]. After getting M r g, M b y, M i and M o, in order to yield the feature maps, we simulate the centersurround receptive fields by subtraction between two maps at the center (c) and the surround (s) levels in these pyramids. They can be calculated as F l,c,s = N( M l (c) M l (s) ) l L = L C L I L O (4) 496

0 0 0 0 0 0 The 29th Fuzzy System Symposium (Osaka, September 9-, 3) 0 0 0 0 0 2 4 6 8 2 4 6 (a)original Graphics IntensitiesCM (c)intensity Feature Map ColorCM 2 4 6 8 2 4 6 (b)color Feature Map OrientationsCM 2 4 6 8 2 4 6 (d)orientation Feature Map Table : Fuzzy Inference Rules C O I OL OM OH IL SVL SL SLL CL IM SL SL SLL IH SLL SLL SM IL SL SLL SM CM IM SLL SLL SM IH SM SM SLH IL SM SLH SH CH IM SLH SLH SH IH SH SH SVH Fig. 2: Example of Feature Maps Got from Image where L C = {I} L I = {r g, b y} () L O = {0, π 4, π 2, 3π 4 } Note that N is an iterative operator for local nonlinear iterative competition between salient locations within each feature map. Finally, by summing over the center-surround combinations and normalizing again according the results obtained in Eq.(4), the feature maps of color, intensity and orientation can be obtained according to Eq.(6) as C c, C i, C o, respectively. Here, the feather maps all we need for building region saliency map are obtained already. Fig.2 shows an example of three feature maps mentioned above. C c = F l C i = N( l L c F c ) C o = N( l L o F o ) (6) 3.2 Fuzzy Inference Fuzzy inference is based on fuzzy logic and resembles human reasoning in its use of approximate information and uncertainty to generate decisions. It consists of one or several rules (implications), a set of facts, and a conclusion. In section, we have pointed out the defect of ordinary combination method of feature maps. It has not an importance distinction be- tween various features, especially when a feature is more important comparing with others. In the feature maps building stage, it has little sense to use fuzzy theory as a classifier, but in the combination stage, using fuzzy theory to infer visual attention region can lead to the results more approach the saliency feature that graphics have. The greatest difference between mathematical sum method and fuzzy inference method is that the importance reflected in every feature map is different. These are tuned by fuzzy rule according to feature values. In fuzzy inference method, the importance of color feature map is higher than intensity and orientation ones, to avoid getting bad result when met the situation that color feature is very low while other two are high. In this study, We use the feature variables from color feature map(c c ), intensity feature map(c i ) and orientation feature map (C o ) in the IF part while the output value in THEN part is value of region saliency map(s m ). Every value of region saliency map is decided by fuzzy inference rules as shown in Table., where C, I, O stand for color, intensity, and orientation respectively. Fig.3 shows the membership functions and singletons. In Table., CL, CM and CH are fuzzy labels that represent each state of value from color feature map is low, medium and high, respectively. And so on to IL, IM, IH, OL, OM and OH. The following shows 497

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) CL CM CH 0 0 0 0 0 0 0 0 0 0 l l 2 l 3 Cc 40 00 0 0 0 00 600 700 800 00 600 0 0 0 00 600 700 800 IL IM IH (a)original Graphics ColorCM ColorCM 0 m m 2 m 3 Ci OL OM OH 3 40 4 0 3 3 40 4 0 (b) Color Feature Maps 0 n n 2 n 3 Co (a) Memebership Functions in IF Part IntensitiesCM IntensitiesCM SVL SL SLL SM SLH SH SVH 3 40 4 0 3 3 40 4 0 (c) Intensity Feature Maps OrientationsCM OrientationsCM 0 S S 2 S 3 S 4 S S 6 S 7 S m (b) Singletons in THEN Part Fig. 3: Memebership Functions and Singletons 3 40 4 0 3 3 40 4 0 the fuzzy labels that express the state in each level of region saliency map. Output Fuzzy Labels of Region Saliency Map(S m) SVL : Value of Region Saliency Map Very Small SL : Value of Region Saliency Map Small SLL : Value of Region Saliency Map Little Small SM : Value of Region Saliency Map Medium SLH: Value of Region Saliency Map Little Large SH : Value of Region Saliency Map Large SVH : Value of Region Saliency Map Very Large 4 Experimental Results We conduct several experiments to demonstrate the inference result of the proposed method and also compare with the performance of the proposed method with Itti s model[8]. As mentioned in last section, we get the various feature saliency maps at first. Here, two different graphics are used as the input images. (d) Orientation Feature Maps Fig. 4: Two Examples of Feature Saliency Maps The input graphics is processed for low-level features at multiple scales, and center-surround differences are computed according to Eq.(4). Then, the resulting feature maps are combined into feature saliency maps according to Eq.(6), which is shown in Fig.4. After getting the feature saliency maps, the region locations in the saliency map compete for the highest saliency value by fuzzy inference proposed by us. After segmentation around the most salient region location, this saliency map is used for obtaining a smooth object mask at image resolution and for object-based inhibition of return. The parameters of fuzzy rules (l i, m i, n i ) shown in (a) of Fig.3 are chosen based on the average value of each feature map. The results 498

SaliencyMap The 29th Fuzzy System Symposium (Osaka, September 9-, 3) 3 40 4 0 SaliencyMap 3 3 40 4 0 (a) Saliency Maps by Sum Feature Maps Method SaliencyMap SaliencyMap 3 3 40 4 0 3 40 4 0 (b) Saliency Maps by Fuzzy Inference 0 0 0 0 0 0 0 0 0 0 40 0 00 0 0 0 00 600 700 800 0 0 0 0 0 0 (c) Attention Region by Sum Feature Maps Method 0 0 0 0 0 0 0 0 0 0 40 0 00 0 0 0 00 600 700 800 0 0 0 0 0 0 (d) Attention Region by Fuzzy Inference Fig. : Two Examples of Feature Saliency Maps of saliency map and attention region by bother sum feature maps method and fuzzy inference method are shown in Fig.. The attention regions are marked by yellow lines while red lines express the order easy to be paid attention of them. As we can see in the figure, there are only little differences between the saliency maps of two methods. And for the first example, the approximate locations of attention regions and the orders are basically the same between two methods while the little difference is the shape and size of region. This is because the color, intensity and orientation feature are all reflect obviously in regions marked of it compared with the rest regions, which also means that the differences of importance for the three features are small. So the method we proposed has not functioned very efficiently. But from the result of the second example we can see that the attention regions of both two methods are basically the same while the order is different. But only from these results we cannot yet say whether our proposed method is better than Itti s[8] or not. So we conduct another experiment to verify it. The following experiment is ask males who are because to years old to look over graphics. After all the experiments they will be showed the results of attention region got by the two methods and asked to compare with the ones they actually attending and looking at in the experiment process. Finally, an evaluation of user s attitude to the results is carried on and shown in Fig.6. Every factor of them has five ranks and represented by from worst to best in these figures. We can see from the evaluation results that the performance of our proposed method is higher at the order of visual attention region while a little lower at the accuracy of region detection. This also illustrated that proposed fuzzy inference method can improve the performance of attention region prediction at some aspect. Conclusions In this paper, we proposed a fuzzy inference method based on color, intensity and orientation feature maps of graphics to predict the visual attention regions. We also conducted a series of attention region predict experiments. The prediction accuracy of our proposed approach was evaluated in experiments, and the results confirmed the effectiveness of our method in visual attention region prediction. The problem still existing is the method in this paper will be effective when the input graphics performance significantly in color feature. In the future, we will work on proposing a method to get the conspicuity feature of graphics and apply to fuzzy rules. Reference [] F. Sadri, Logic-Based Approaches to Intention Recognition, Handbook of Research on Ambient Intelligence: Trends and Perspectives,. 499

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) [7] L. Itti, C. Koch and E. Niebur, A Model of Saliency-Based Visual Attention for Rapid Scene Analysis, IEEE Trans. Pattern Analysis and Machine Intelligence, Vol., No., pp.4-9, 998. (a) Evaluation for Sum Feature Maps Method [8] L. Itti and C. Koch, A saliency-based search mechanism for overt and covert shifts of visual attention, Vision Research, Vol.40, pp.489-06, 00. [9] D. Walther, U. Rutishauser, C. Koch and P. Perona, On the usefulness of attention for object recognition, Workshop on Attention and Perfromance in Computational Vision, pp.96-3, 04. (b) Evaluation for Fuzzy Inference Method Fig. 6: Standard Deviation for Evaluation Results of Attention Regions Obtained by Two Methods [2] D. V. Pynadath and M. P. Wellman. Accounting for Contextin Plan Recognition, with Application to Traffic Monitoring, Proceedings of the Eleventh International Conference on Uncertainty in Artificial Intelligence, pp.472-48, 99. [3] L. M. Pereira and H. T. Anh, Intention Recognition via Causal Bayes Networks Plus Plan Generation, Progress in Artificial Intelligence, pp. 38-49, 09. [4] K. A. Tahboub, Intelligent Human-Machine Interaction Based on Dynamic Bayesian Networks Probabilistic Intention Recognition, Journal of Intelligent and Robotic Systems, Vol.4, pp.3-2, 06. [] L. Itti, Models of bottom-up and top-down visual attention, PhD thesis, California Institute of Technology, 00. [] R. Manduchi, P. Perona and D. Shy, Efficient deformable filter banks, IEEE Transactions on Signal Processing, Vol.46, No.4, pp.68-73, 998. [2] W. O. Lee, J. W. Lee, K. R. Park, E. C. Lee and M. Whang, Object recognition and selection method by gaze tracking and SURF algorithm, International Conference on Multimedia and Signal Processing, pp.26-26,. Contact Address Mao Wang Graduate School of Engineering, University of Fukui 3-9-, Bunkyo, Fukui 9-807 Japan E-mail: mwang@ir.his.u-fukui.ac.jp [] J. W. Harris and H. Stocker, Handbook of Mathematics and Computational Science, Springer- Verlag New York, 998. [6] W. Mao and J. Gratch, A Utility-Based Approach to Intention Recognition, Proceedings of the AAMAS 04 Workshop on Agent Tracking: Modeling Other Agents from Observations, 04. 00