Saliency Inspired Modeling of Packet-loss Visibility in Decoded Videos

Size: px
Start display at page:

Download "Saliency Inspired Modeling of Packet-loss Visibility in Decoded Videos"

Transcription

1 1 Saliency Inspired Modeling of Packet-loss Visibility in Decoded Videos Tao Liu*, Xin Feng**, Amy Reibman***, and Yao Wang* *Polytechnic Institute of New York University, Brooklyn, NY, U.S. **Chongqing University, Chongqing, P.R. China ***AT&T Labs-Research, Florham Park, NJ, U.S. Abstract The visibility of packet loss in decoded video depends on various factors and their complicated interactions, such as loss severity and duration, and characteristics of background signal. However, visual attention or saliency may play an important role as well. In this work, we investigate how to improve visibility prediction by incorporating the saliency information. Based on earlier findings about how saliency affects the perceptual quality of video with packet losses, we propose several saliency-based factors and incorporate them into a Generalized Linear Model (GLM) to predict loss visibility. Test results with 1080 MPEG-2 packet losses indicate that saliency information can help improve the prediction accuracy about 12% over nonsaliency-based model, and that saliency-weighted mean-squareerror and variation of saliency information are promising metrics. Index Terms Packet loss visibility, perceptual video quality, saliency information, GLM important component of the Human Visual System (HVS), describes features of an image (for example, color, intensity and orientation) which cause one region to stand out relative to other regions [6]. Because of both physiological and psychological evidence that humans have high selectivity on exposed visual information [6], there is recent interest in designing image and video quality metrics that weight signal errors based on their visual sensitivities [4][5][10]. Some evidence suggests incorporating saliency during spatial pooling may not always improve quality metrics [5]. However, in this paper, we demonstrate that adding saliency information to the problem of predicting the visibility of impairments due to packet losses can dramatically improve prediction performance. We focus here on individual packet losses in high-quality video that has few other visual artifacts. I. INTRODUCTION In video transmission, the decoder may not receive all the encoded video data because of the losses occurred in packet networks. Generally, different packet losses may not cause equal video quality degradation, and in fact some may not be visible to human viewers. On the other hand, assessing the impact of packet loss is crucial for network providers to monitor and control their networks to guarantee satisfactory video quality for end users. However, because of significant difficulty in assessing the perceptual quality of video affected by packet loss directly, investigating the visibility of packet losses is a natural and appropriate starting point. In the literature, some researchers reported their findings about the visibility of packet loss in video transmission. The work in [1][7] accurately modeled the visibility of single packet loss by a Generalized Linear Model (GLM) with various factors, such as signal error, video motion, loss position and etc. The authors of [8] extended this to predict visibility of multiple losses. The work in [3] introduced an intuitive concept Mean Time Between Failures (MTBF) that video quality can be interpreted as the probability of visible artifacts, and estimate it with several existing quality metrics. In [9], the visibility of single packet loss in compressed video is evaluated by loss severity (in PSNR drop) caused by lost frame, and the duration of its propagated error. In this work, we explore whether saliency information can improve the prediction of packet-loss visibility. Saliency, an Specifically, we still use the GLM method to predict loss visibility, but introduce new factors that incorporate saliency information. Motivated by an earlier finding [10] that weighting the pixel-wise errors by the visual saliency map can predict the perceptual quality better than using non-weighted errors, we propose to use saliency weighted error as a factor. Furthermore, we also investigate factors summarizing the saliency map itself, including how the saliency map changes when there is a packet loss, and the temporal variation of the saliency map. In this work, we explore packet loss visibility using the subjective data in [1]. We first design a GLM with previously identified non-saliency factors as a benchmark, and then we incorporate our aforementioned saliency-based factors into the GLM. We compute the saliency map using the computational visual attention model developed by Itti et al. [6]. Our tests show that both saliency-weighted error and changes in saliency map are good metrics to predict loss visibility, and they can significantly improve the prediction accuracy of GLM over the one using non-saliency factors. The remainder of this paper is organized as follows. All factors we consider, both with and without saliency, are discussed in Section II. The subjective data is described briefly in Section III. In Section IV, we design two GLMs, with and without saliency factors, and we analyze and compare their performances. We conclude this paper with Section V.

2 2 A. Non-saliency factors II. FACTORS AFFECTING VISIBILITY In [1][7], there were totally 20 non-saliency factors (and their variations) proposed, which covers both characteristics of videos and packet loss impairments. Furthermore, the interaction between them on a scene-level was considered as well. Here we only briefly discuss them in each category; refer to [1][7] for more detailed descriptions. (a) Error characteristics: Mean Squared Error (MSE) and Structural Similarity Index Metric (SSIM) are two widely accepted quality metrics. In the loss-visibility scenario, for the sake of easy calculation, two simplified variations of each metric are used to predict the visibility of packet loss: those measurements of initial frame in loss-affected segment, IMSE and ISSIM; and the extreme values of IMSE and ISSIM in macroblock level, MaxIMSEmb and MinISSIMmb. When there is a single slice loss (as opposed to the loss of an entire frame), the impact of discontinuities caused by lost slices on video quality can be measured by the Slice-Boundary Mismatch (SBM), first proposed in [13] and modified in minor details in [7]. Only SBM on the initial frame of loss-affected segment, ISBM, is considered. Additionally, some important content-independent measures, such as spatial extent, or SXTNT (the number of slices lost in one frame), HGT (the average height of the lost slices), Duration (duration of the loss-affected segment), are also considered. (b) Video characteristics: Motion is one of the most important characteristics of videos. Therefore,the mean and variance of the magnitudes of the motion vectors across all macroblocks initially affected by a loss, MotMean and MotVar can also be used to predict the visibility of a loss. SigMean and SigVar, the mean and variance of intensity values of the initial frame of loss-affected segment, and ResidEng (residual energy after motion compensation) of that frame are also effective. (c) Scene-level characteristics: In addition to the factors in (a) and (b), some high-level characteristics are also considered. The authors of [7] show that the relative position between scene change and packet loss impairment has great influence on its visibility. Therefore, D2R (the distance between the current frame (with packet loss) and the reference frame used for concealment)(farconceal is set when values of D2R are equal to or greater than 3), DistFromCut (the distance in time between the first frame affected by the packet loss and the nearest scene cut, either before or after) and its threshold versions, AtScene, BeforeScene, and AfterScene, are considered in this work. Since these non-saliency factors were proved capable to predict the loss visibility, we use them as candidate factors to design a GLM. B. Saliency-based factors In the previous subsection, we discuss a score of factors covering many attributes of packet loss impairment. However, these are only heuristically linked to properties of the HVS. We hereby investigate factors associated with saliency in this subsection. We adopt the widely accepted saliency-based visual attention model (SVAM) [6] proposed by Itti. et al. to calculate the saliency information of our test videos. Itti s SVAM decomposes an input image into a set of multiscale feature maps of color, intensity and orientation. By using a center-surround mechanism, SVAM accurately simulates which location in the image will automatically and unconsciously attract visual attention. Fig.1 shows the saliency maps of original and loss-imaired video frames (for demonstration purpose only, not the actual test video). Figure 1. (a) (c) Saliency maps of original and loss-affected video frames. We consider two basic attributes of the computed saliency information for their ability to predict loss visibility. First, inspired by the results in our previous work [9] that weighting the pixel-wise errors by the saliency values correlates well with the perceptual quality of loss-impaired videos, we propose to supplement the IMSE factor by saliency-weighted IMSE, denoted by IMSE_Sal. We also consider the saliencyweighted MSE computed over all loss-affected frames, yielding MSE_Sal. Second, in addition to the impact of saliency on pixel-wise error, we investigate properties of the saliency information itself. If we examine the saliency maps of both the original and the loss-impaired video frames from Fig. 1, we discover that packet losses not only distort the video frames but also alter the distribution of salient regions across the affected frames spatially and temporally. We also observ that packet losses are more visible in videos where the saliency map changes rapidly in time. These two observations lead us to propose two additional factors: SMSE, which measures the changes (in terms of MSE) between saliency maps of original and loss-impaired frames (only in the position where loss happens); STV, which measures temporal variation of the saliency map of loss-impaired frames, respectively. They are defined mathematically as follows with SMSE = E t [ E(x,y) ( S 1 (x, y, t) S 2 (x, y, t) 2 ) ] (1) (b) (d) ST V = STD t (SM 2 (t)) (2) SM 2 (t) = E (x,y) (S 2 (x, y, t)) (3) where E (x,y) () is the 2-D mean operator averaging over all pixels in frame t, E t () is the mean operator averaging over

3 3 time in the segment of loss-impaired sequence, and STD t () is the standard deviation operator over time for that segment. S i (x, y, t) denotes the saliency value at position (x, y) in frame t, and i = 1or2 refers to the original or distorted video sequence. Note that, for saliency computation, we tested two methods, one using color, orientation, and intensity information only, as in the original Itti s model [6]; another one further using motion information with the motion features computed following [11]. From our previous study, the second method produces saliency maps that are more consistent with our visual inspections, although it requires extra compuation over the former one in saliency detection. Therfore, we focus on the latter one in this work. Table I summarizes all the aforementioned factors. Table I LIST OF ALL NON-SALIENCY AND SALIENCY FACTORS 1 IMSE 14 D2R 2 MaxIMSEmb 15 SigVar 3 ISSIM 16 DistFromCut 4 MinSSIMmb 17 AtScene 5 ISBM 18 BeforeScene 6 SXTNT 19 AfterScene 7 HGT 20 FarConceal 8 Duration 21 IMSE_Sal (no motion) 9 ResidEng 22 IMSE_Sal 10 CameraMotion 23 MSE_Sal 11 SigMean 24 S_MSE 12 MotMean 25 STV 13 MotVar III. SUBJECTIVE TEST The subjective data used in this work was first presented in [1]; to be self-contained, we describe it briefly here. This subjective test was designed not to assess the quality of video at a given packet loss rate, but instead to learn about what affects the visibility of impairments caused by individual packet losses. The test videos shown were compressed with MPEG-2 at 720x480 resolution and 30 fps frame rate, with various scene contents and different camera motions, using 13-frame GOPs with 2 B-frames before every P-frame at a bitrate of around 4Mbps. One isolated packet loss was randomly inserted into the video in every 4-second window. The 1080 packet losses affected either one slice, two slices, or an entire frame. The decoder applied Zero-motion error concealment (copying macroblocks from the closest reference frame) when losses occurred. Each packet loss was viewed by 12 viewers, whose task was to indicate each time they saw an artifact while were watching a 6-minute continuous video clip. The ground truth of visibility of each packet loss was defined as the percentage of viewers who indicated they saw the loss. For the detailed information about this subjective test, please refer to [1]. A. GLM fitting method IV. GENERALIZED LINEAR MODELS As [1][7], we model the probability of visibility using a GLM [2], which is a development of linear models to accommodate both non-normal response distributions and transformations to linearity in a straightforward way. We use logistic regression to fit our model, with logit() as the link function. With the help of statistical software R [12], the model is fit with an iteratively re-weighted least-square method to generate a maximum-likelihood estimate. The GLM fittings suggested in [2] can be summarized to the following 3 steps: Step 1: Before designing the GLM with multiple variables, each individual factor is analyzed first. By examining the Lowess smoothed scatter plot between each factor and logit function of subjective visibility, the best (in prediction error) form of each factor, e.g. log(), can be determined. Step 2: Since some of the factors are correlated, it is possible to overfit the model by using all of them together. To select factors into the model, we took a step-wise approach by adding one factor at a time. We begin with Null model, and in each step, the one factor that generates maximum reduction of prediction error in the cross-validation process is added to the model, so that the efficiency of each inclusion is maximized. When we can no longer reduce the prediction error by adding a single factor, the model is preliminarily established. Step 3: We check if there are any interactions (products) between any two selected factors that can further improve the prediction of the model. If there are multiple such pairs, the decisions of their inclusion are made in the similar fashion as Step 2. Note that we perform 10-fold cross-validation in the process of building up the model. Specifically, we randomly divide the entire data set of 1080 losses into 10 groups of equal size and choose the data from 9 out of the 10 sets as a training set. The remaining data set is used for testing. We repeat this process 10 times, each time choosing a different set for training. The average prediction error is used as the performance measure. B. GLMs of two models In order to test the impact of saliency information on modeling the loss visibility, we fit the same subjective data with two different set of factors, one containing only nonsaliency-based factors (Model 1); the other one containing all the aforementioned factors (Model 2). We note that our model differs from that in [7] because it uses just one subjective dataset. The factors and their coefficients of both models are summarized in Table II. To test the significance of each factor in the model, including the interaction terms added in the third step, we re-perform Step 2 with all the selected factors and interaction terms to update their inclusion order. The new order provides a ranking of the significance of each factor and interaction term. This allows the two models to be compared if we limit each to having the same number of predictive factors. We draw the bar plots of factors and model prediction errors of this stepwise procedure, in the order of their inclusions, in Fig. 2 (a) and 2 (b). C. Comparison Model 1 v.s. Model 2 To compare the prediction performances of Model 1 and Model 2, we show the relationship between the number of

4 4 (a) (b) Figure 2. Factor inclusions of (a) Model 1; and (b) Model 2.

5 5 Table II COEFFICIENTS OF MODEL 1 AND MODEL 2 Model 1 Model 2 factor coef. factor coef. MinSSIMmb e+00 MinSSIMmb e-01 FarConceal* FarConceal* 1.929e-01 log(imse) (IMSE_Sal) e-01 StillCamera e-01 StillCamera e+00 log(reseng) e-01 log(reseng) e+00 log(imse) 7.209e-01 log(imse) 7.090e-01 log(isbm)* (SXTNT!=30) 6.590e-02 SXTNT e-01 HGT e-02 log(reseng)* 3.027e+00 SigMean e-02 StillCamera* e+00 AtScene 1.404e+00 HGT e-02 Duration< e-01 SigMean e-03 BeforeScene e e+01 MotMean> e-01 BeforeScene e-01 SigVar e-04 log(1-issim) 2.915e-01 DistFromCut e-02 Duration< e-01 Duration 1.578e+01 SigVar* FarConceal*(IMSE_Sal) e-04 SigVar 8.905e-04 log(stv) 1.715e-01 DistFromCut e-03 log(isbm)* SXTNT!= e-02 MotMean> e-01 (MotMean>0.707)* FarConceal*(IMSE_Sal) e-01 AtScene * 1.472e+01 AtScene e+00 factors used in the model and prediction error reduction ratio (Model 2 to Model 1) in Fig.3. We can see that the overall prediction error of Model 2 ( ) is about 12% less than that of Model 1( ). When both models are limited to use 15 factors, Model 2 still outperforms Model 1 by about 9%. In addition, we can see that, no matter how many factors are used to fit the models, our model with saliency-based factors always outperform those without, except for the case of only one factor, since that factor (MinSSIMmb) is the same for both models. Therefore, we conclude that saliency information significantly boosts the visibility prediction performance. To gain a clearer picture of the contribution of each individual saliency-based factor, we examine the inclusion orders (or significance ranks) of saliency-based factors in Model 2 in Fig. 2 (b). There are 4 saliency factors in the first half of the 23 factors in the model: FarConceal*(IMSE_Sal) 1 4, log(resideng)*, StillCamera*, and. If we expand our focus to the first 15 (the number of factors used in Model 1) factors, SigVar*FarConceal*(IMSE_Sal) 1 4 is also present. Note that the saliency-weighted pixel-wise error (IMSE_Sal) and the difference of saliency maps of original and distorted video caused by packet loss (S_MSE) are two very helpful factors in modeling packet loss visibility. Admittedly, this improvement is at the expense of complex computation of the saliency detection system, unless there is any salinecy or foveal information avaiable, such as in some Figure 3. Performance comparison between Model 1 and Model 2. ( The number of factors used in Model 1 is limited to 15.) particular types of video contents, e.g. news broadcasting or soccer games, where salient regions are obvious. V. CONCLUSIONS In this paper, we proposed several saliency-based factors, which can be used to improve the prediction of the visibility of packet losses. Together with a variety of non-saliencybased factors, we fit the GLMs with and without our proposed factors using existing subjective data, and the results show that saliency-based factors significantly improve the performance in loss-visibility modeling. Our prior work in perceptual quality prediction and the current work in prediciting packet loss have shown that saliency information is helpful for both. One interesting question is whether we can predict the perceptual quality from the packet loss visibility. This is one of our current research directions. REFERENCES [1] S. Kanumuri, et al, Modeling Packet-loss Visibility in MPEG-2 Video IEEE Trans. Multimedia, vol.8, pp , Apr [2] P. McCullagh, et al., Generalized Linear Models, 2nd Edition, London, U.K.: Chapman & Hall. [3] N. Suresh, et al., Mean Time Between Failures: A Subjectively Meaningful Video Quality Metric, ICASSP,2006 [4] Z. Wang, et al., Image Quality Assessment:From Error Visibility to Structure Similarity IEEE Trans. Im. Proc. Vol.13, Apr [5] A. Ninassi, et al., Does Where You Gaze on An Image Affect Your Perception of Quality? Applying Visual Attention To Image Quality Metric, Intl. Cof. Im. Proc., 2007 [6] L. Itti, et al., A Model of Saliency-Based Visual Attention for Rapid Scene Analysis, IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 20, no.11, pp , Nov [7] A. Reibman, and D. Poole, Predicting packet-loss visibility using scene characteristics, Packet Video, 2007 [8] S. Kanumuri, et al., Predicting H.264 Packet Loss Visibility using a Generalized Linear Model, ICIP, 2006 [9] T. Liu, et al.. Subjective Quality Evaluation of Decoded Video in The Presence of Packet Losses ICASSP 2007 [10] X. Feng, et al., Saliency Based Objective Quality Assessment of Decoded Video Affected by Packet Losses, ICIP, [11] D. Walther, Interactions of Visual Attention and Object Recognition: Computational Modeling, Algorithms, and Psychophysics, PhD thesis, California Institute of Technology, Pasadena, CA, Feb [12] R software is available at [13] H. Rui et al., Evaluation of packet loss impairment on streaming video, J. of Zhejiang University SCIENCE, vol. 7, April 2006.

Saliency Inspired Full-Reference Quality Metrics for Packet-Loss-Impaired Video

Saliency Inspired Full-Reference Quality Metrics for Packet-Loss-Impaired Video IEEE TRANSACTIONS ON BROADCASTING, VOL. 57, NO. 1, MARCH 2011 81 Saliency Inspired Full-Reference Quality Metrics for Packet-Loss-Impaired Video Xin Feng, Tao Liu, Member, IEEE, Dan Yang, and Yao Wang,

More information

INCORPORATING VISUAL ATTENTION MODELS INTO IMAGE QUALITY METRICS

INCORPORATING VISUAL ATTENTION MODELS INTO IMAGE QUALITY METRICS INCORPORATING VISUAL ATTENTION MODELS INTO IMAGE QUALITY METRICS Welington Y.L. Akamine and Mylène C.Q. Farias, Member, IEEE Department of Computer Science University of Brasília (UnB), Brasília, DF, 70910-900,

More information

Validating the Visual Saliency Model

Validating the Visual Saliency Model Validating the Visual Saliency Model Ali Alsam and Puneet Sharma Department of Informatics & e-learning (AITeL), Sør-Trøndelag University College (HiST), Trondheim, Norway er.puneetsharma@gmail.com Abstract.

More information

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian The 29th Fuzzy System Symposium (Osaka, September 9-, 3) A Fuzzy Inference Method Based on Saliency Map for Prediction Mao Wang, Yoichiro Maeda 2, Yasutake Takahashi Graduate School of Engineering, University

More information

Video Saliency Detection via Dynamic Consistent Spatio- Temporal Attention Modelling

Video Saliency Detection via Dynamic Consistent Spatio- Temporal Attention Modelling AAAI -13 July 16, 2013 Video Saliency Detection via Dynamic Consistent Spatio- Temporal Attention Modelling Sheng-hua ZHONG 1, Yan LIU 1, Feifei REN 1,2, Jinghuan ZHANG 2, Tongwei REN 3 1 Department of

More information

VIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING

VIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING VIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING Yuming Fang, Zhou Wang 2, Weisi Lin School of Computer Engineering, Nanyang Technological University, Singapore 2 Department of

More information

FROM H.264 TO HEVC: CODING GAIN PREDICTED BY OBJECTIVE VIDEO QUALITY ASSESSMENT MODELS. Kai Zeng, Abdul Rehman, Jiheng Wang and Zhou Wang

FROM H.264 TO HEVC: CODING GAIN PREDICTED BY OBJECTIVE VIDEO QUALITY ASSESSMENT MODELS. Kai Zeng, Abdul Rehman, Jiheng Wang and Zhou Wang FROM H. TO : CODING GAIN PREDICTED BY OBJECTIVE VIDEO QUALITY ASSESSMENT MODELS Kai Zeng, Abdul Rehman, Jiheng Wang and Zhou Wang Dept. of Electrical and Computer Engineering, University of Waterloo, Waterloo,

More information

MobileASL: Making Cell Phones Accessible to the Deaf Community. Richard Ladner University of Washington

MobileASL: Making Cell Phones Accessible to the Deaf Community. Richard Ladner University of Washington MobileASL: Making Cell Phones Accessible to the Deaf Community Richard Ladner University of Washington American Sign Language (ASL) ASL is the preferred language for about 500,000-1,000,000 Deaf people

More information

Can You See Me Now? Communication Over Cell Phones for Deaf People. Eve Riskin Richard Ladner Computer Engineering University of Washington

Can You See Me Now? Communication Over Cell Phones for Deaf People. Eve Riskin Richard Ladner Computer Engineering University of Washington Can You See Me Now? Communication Over Cell Phones for Deaf People Eve Riskin Richard Ladner Computer Engineering University of Washington Co-PI Thanks Sheila Hemami (Cornell) Graduate Students Anna Cavender,

More information

ANALYSIS OF MOTION-BASED CODING FOR SIGN LANGUAGE VIDEO COMMUNICATION APPLICATIONS

ANALYSIS OF MOTION-BASED CODING FOR SIGN LANGUAGE VIDEO COMMUNICATION APPLICATIONS ANALYSIS OF MOTION-BASED CODING FOR SIGN LANGUAGE VIDEO COMMUNICATION APPLICATIONS Benjaporn Saksiri*, Supavadee Aramvith**, Chai Phongphanphanee* *Information System Engineering Laboratory, Department

More information

MobileASL: Making Cell Phones Accessible to the Deaf Community

MobileASL: Making Cell Phones Accessible to the Deaf Community American Sign Language (ASL) MobileASL: Making Cell Phones Accessible to the Deaf Community Richard Ladner University of Washington ASL is the preferred language for about 500,000-1,000,000 Deaf people

More information

Computational Cognitive Science

Computational Cognitive Science Computational Cognitive Science Lecture 19: Contextual Guidance of Attention Chris Lucas (Slides adapted from Frank Keller s) School of Informatics University of Edinburgh clucas2@inf.ed.ac.uk 20 November

More information

Computational Cognitive Science

Computational Cognitive Science Computational Cognitive Science Lecture 15: Visual Attention Chris Lucas (Slides adapted from Frank Keller s) School of Informatics University of Edinburgh clucas2@inf.ed.ac.uk 14 November 2017 1 / 28

More information

Computational Cognitive Science. The Visual Processing Pipeline. The Visual Processing Pipeline. Lecture 15: Visual Attention.

Computational Cognitive Science. The Visual Processing Pipeline. The Visual Processing Pipeline. Lecture 15: Visual Attention. Lecture 15: Visual Attention School of Informatics University of Edinburgh keller@inf.ed.ac.uk November 11, 2016 1 2 3 Reading: Itti et al. (1998). 1 2 When we view an image, we actually see this: The

More information

Seamless Audio Splicing for ISO/IEC Transport Streams

Seamless Audio Splicing for ISO/IEC Transport Streams Seamless Audio Splicing for ISO/IEC 13818 Transport Streams A New Framework for Audio Elementary Stream Tailoring and Modeling Seyfullah Halit Oguz, Ph.D. and Sorin Faibish EMC Corporation Media Solutions

More information

Framework for Comparative Research on Relational Information Displays

Framework for Comparative Research on Relational Information Displays Framework for Comparative Research on Relational Information Displays Sung Park and Richard Catrambone 2 School of Psychology & Graphics, Visualization, and Usability Center (GVU) Georgia Institute of

More information

An Evaluation of Motion in Artificial Selective Attention

An Evaluation of Motion in Artificial Selective Attention An Evaluation of Motion in Artificial Selective Attention Trent J. Williams Bruce A. Draper Colorado State University Computer Science Department Fort Collins, CO, U.S.A, 80523 E-mail: {trent, draper}@cs.colostate.edu

More information

An Attentional Framework for 3D Object Discovery

An Attentional Framework for 3D Object Discovery An Attentional Framework for 3D Object Discovery Germán Martín García and Simone Frintrop Cognitive Vision Group Institute of Computer Science III University of Bonn, Germany Saliency Computation Saliency

More information

Emotion Recognition using a Cauchy Naive Bayes Classifier

Emotion Recognition using a Cauchy Naive Bayes Classifier Emotion Recognition using a Cauchy Naive Bayes Classifier Abstract Recognizing human facial expression and emotion by computer is an interesting and challenging problem. In this paper we propose a method

More information

Facial expression recognition with spatiotemporal local descriptors

Facial expression recognition with spatiotemporal local descriptors Facial expression recognition with spatiotemporal local descriptors Guoying Zhao, Matti Pietikäinen Machine Vision Group, Infotech Oulu and Department of Electrical and Information Engineering, P. O. Box

More information

A NOVEL VARIABLE SELECTION METHOD BASED ON FREQUENT PATTERN TREE FOR REAL-TIME TRAFFIC ACCIDENT RISK PREDICTION

A NOVEL VARIABLE SELECTION METHOD BASED ON FREQUENT PATTERN TREE FOR REAL-TIME TRAFFIC ACCIDENT RISK PREDICTION OPT-i An International Conference on Engineering and Applied Sciences Optimization M. Papadrakakis, M.G. Karlaftis, N.D. Lagaros (eds.) Kos Island, Greece, 4-6 June 2014 A NOVEL VARIABLE SELECTION METHOD

More information

Natural Scene Statistics and Perception. W.S. Geisler

Natural Scene Statistics and Perception. W.S. Geisler Natural Scene Statistics and Perception W.S. Geisler Some Important Visual Tasks Identification of objects and materials Navigation through the environment Estimation of motion trajectories and speeds

More information

SUPPLEMENTARY INFORMATION. Table 1 Patient characteristics Preoperative. language testing

SUPPLEMENTARY INFORMATION. Table 1 Patient characteristics Preoperative. language testing Categorical Speech Representation in the Human Superior Temporal Gyrus Edward F. Chang, Jochem W. Rieger, Keith D. Johnson, Mitchel S. Berger, Nicholas M. Barbaro, Robert T. Knight SUPPLEMENTARY INFORMATION

More information

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation

Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation Aldebaro Klautau - http://speech.ucsd.edu/aldebaro - 2/3/. Page. Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation ) Introduction Several speech processing algorithms assume the signal

More information

Research Methods in Forest Sciences: Learning Diary. Yoko Lu December Research process

Research Methods in Forest Sciences: Learning Diary. Yoko Lu December Research process Research Methods in Forest Sciences: Learning Diary Yoko Lu 285122 9 December 2016 1. Research process It is important to pursue and apply knowledge and understand the world under both natural and social

More information

No-Reference Video quality assessment of H.264 video streams based on semantic saliency maps

No-Reference Video quality assessment of H.264 video streams based on semantic saliency maps No-Reference Video quality assessment of H.264 video streams based on semantic saliency maps Hugo Boujut, Jenny Benois-Pineau, Toufik Ahmed, Ofer Hadar, Patrick Bonnet To cite this version: Hugo Boujut,

More information

A Visual Saliency Map Based on Random Sub-Window Means

A Visual Saliency Map Based on Random Sub-Window Means A Visual Saliency Map Based on Random Sub-Window Means Tadmeri Narayan Vikram 1,2, Marko Tscherepanow 1 and Britta Wrede 1,2 1 Applied Informatics Group 2 Research Institute for Cognition and Robotics

More information

SUPPLEMENTAL MATERIAL

SUPPLEMENTAL MATERIAL 1 SUPPLEMENTAL MATERIAL Response time and signal detection time distributions SM Fig. 1. Correct response time (thick solid green curve) and error response time densities (dashed red curve), averaged across

More information

Learning to Generate Long-term Future via Hierarchical Prediction. A. Motion-Based Pixel-Level Evaluation, Analysis, and Control Experiments

Learning to Generate Long-term Future via Hierarchical Prediction. A. Motion-Based Pixel-Level Evaluation, Analysis, and Control Experiments Appendix A. Motion-Based Pixel-Level Evaluation, Analysis, and Control Experiments In this section, we evaluate the predictions by deciles of motion similar to Villegas et al. (2017) using Peak Signal-to-Noise

More information

Outline. Teager Energy and Modulation Features for Speech Applications. Dept. of ECE Technical Univ. of Crete

Outline. Teager Energy and Modulation Features for Speech Applications. Dept. of ECE Technical Univ. of Crete Teager Energy and Modulation Features for Speech Applications Alexandros Summariza(on Potamianos and Emo(on Tracking in Movies Dept. of ECE Technical Univ. of Crete Alexandros Potamianos, NatIONAL Tech.

More information

Reading Assignments: Lecture 18: Visual Pre-Processing. Chapters TMB Brain Theory and Artificial Intelligence

Reading Assignments: Lecture 18: Visual Pre-Processing. Chapters TMB Brain Theory and Artificial Intelligence Brain Theory and Artificial Intelligence Lecture 18: Visual Pre-Processing. Reading Assignments: Chapters TMB2 3.3. 1 Low-Level Processing Remember: Vision as a change in representation. At the low-level,

More information

A Model for Automatic Diagnostic of Road Signs Saliency

A Model for Automatic Diagnostic of Road Signs Saliency A Model for Automatic Diagnostic of Road Signs Saliency Ludovic Simon (1), Jean-Philippe Tarel (2), Roland Brémond (2) (1) Researcher-Engineer DREIF-CETE Ile-de-France, Dept. Mobility 12 rue Teisserenc

More information

(Visual) Attention. October 3, PSY Visual Attention 1

(Visual) Attention. October 3, PSY Visual Attention 1 (Visual) Attention Perception and awareness of a visual object seems to involve attending to the object. Do we have to attend to an object to perceive it? Some tasks seem to proceed with little or no attention

More information

MEMORABILITY OF NATURAL SCENES: THE ROLE OF ATTENTION

MEMORABILITY OF NATURAL SCENES: THE ROLE OF ATTENTION MEMORABILITY OF NATURAL SCENES: THE ROLE OF ATTENTION Matei Mancas University of Mons - UMONS, Belgium NumediArt Institute, 31, Bd. Dolez, Mons matei.mancas@umons.ac.be Olivier Le Meur University of Rennes

More information

Actions in the Eye: Dynamic Gaze Datasets and Learnt Saliency Models for Visual Recognition

Actions in the Eye: Dynamic Gaze Datasets and Learnt Saliency Models for Visual Recognition Actions in the Eye: Dynamic Gaze Datasets and Learnt Saliency Models for Visual Recognition Stefan Mathe, Cristian Sminchisescu Presented by Mit Shah Motivation Current Computer Vision Annotations subjectively

More information

arxiv: v1 [cs.lg] 4 Feb 2019

arxiv: v1 [cs.lg] 4 Feb 2019 Machine Learning for Seizure Type Classification: Setting the benchmark Subhrajit Roy [000 0002 6072 5500], Umar Asif [0000 0001 5209 7084], Jianbin Tang [0000 0001 5440 0796], and Stefan Harrer [0000

More information

Mammogram Analysis: Tumor Classification

Mammogram Analysis: Tumor Classification Mammogram Analysis: Tumor Classification Term Project Report Geethapriya Raghavan geeragh@mail.utexas.edu EE 381K - Multidimensional Digital Signal Processing Spring 2005 Abstract Breast cancer is the

More information

Can Saliency Map Models Predict Human Egocentric Visual Attention?

Can Saliency Map Models Predict Human Egocentric Visual Attention? Can Saliency Map Models Predict Human Egocentric Visual Attention? Kentaro Yamada 1, Yusuke Sugano 1, Takahiro Okabe 1 Yoichi Sato 1, Akihiro Sugimoto 2, and Kazuo Hiraki 3 1 The University of Tokyo, Tokyo,

More information

NMF-Density: NMF-Based Breast Density Classifier

NMF-Density: NMF-Based Breast Density Classifier NMF-Density: NMF-Based Breast Density Classifier Lahouari Ghouti and Abdullah H. Owaidh King Fahd University of Petroleum and Minerals - Department of Information and Computer Science. KFUPM Box 1128.

More information

Comparative Evaluation of Color Differences between Color Palettes

Comparative Evaluation of Color Differences between Color Palettes 2018, Society for Imaging Science and Technology Comparative Evaluation of Color Differences between Color Palettes Qianqian Pan a, Stephen Westland a a School of Design, University of Leeds, Leeds, West

More information

IMAGE QUALITY ASSESSMENT BY USING AR PREDICTION ALGORITHM WITH INTERNAL GENERATIVE MECHANISM

IMAGE QUALITY ASSESSMENT BY USING AR PREDICTION ALGORITHM WITH INTERNAL GENERATIVE MECHANISM Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 8, August 2014,

More information

Computational modeling of visual attention and saliency in the Smart Playroom

Computational modeling of visual attention and saliency in the Smart Playroom Computational modeling of visual attention and saliency in the Smart Playroom Andrew Jones Department of Computer Science, Brown University Abstract The two canonical modes of human visual attention bottomup

More information

Comparative Study of K-means, Gaussian Mixture Model, Fuzzy C-means algorithms for Brain Tumor Segmentation

Comparative Study of K-means, Gaussian Mixture Model, Fuzzy C-means algorithms for Brain Tumor Segmentation Comparative Study of K-means, Gaussian Mixture Model, Fuzzy C-means algorithms for Brain Tumor Segmentation U. Baid 1, S. Talbar 2 and S. Talbar 1 1 Department of E&TC Engineering, Shri Guru Gobind Singhji

More information

Identification of Tissue Independent Cancer Driver Genes

Identification of Tissue Independent Cancer Driver Genes Identification of Tissue Independent Cancer Driver Genes Alexandros Manolakos, Idoia Ochoa, Kartik Venkat Supervisor: Olivier Gevaert Abstract Identification of genomic patterns in tumors is an important

More information

Viewpoint Dependence in Human Spatial Memory

Viewpoint Dependence in Human Spatial Memory From: AAAI Technical Report SS-96-03. Compilation copyright 1996, AAAI (www.aaai.org). All rights reserved. Viewpoint Dependence in Human Spatial Memory Timothy P. McNamara Vaibhav A. Diwadkar Department

More information

Recurrent Refinement for Visual Saliency Estimation in Surveillance Scenarios

Recurrent Refinement for Visual Saliency Estimation in Surveillance Scenarios 2012 Ninth Conference on Computer and Robot Vision Recurrent Refinement for Visual Saliency Estimation in Surveillance Scenarios Neil D. B. Bruce*, Xun Shi*, and John K. Tsotsos Department of Computer

More information

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES

MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES 24 MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES In the previous chapter, simple linear regression was used when you have one independent variable and one dependent variable. This chapter

More information

Reveal Relationships in Categorical Data

Reveal Relationships in Categorical Data SPSS Categories 15.0 Specifications Reveal Relationships in Categorical Data Unleash the full potential of your data through perceptual mapping, optimal scaling, preference scaling, and dimension reduction

More information

Overview of the visual cortex. Ventral pathway. Overview of the visual cortex

Overview of the visual cortex. Ventral pathway. Overview of the visual cortex Overview of the visual cortex Two streams: Ventral What : V1,V2, V4, IT, form recognition and object representation Dorsal Where : V1,V2, MT, MST, LIP, VIP, 7a: motion, location, control of eyes and arms

More information

Bootstrapped Integrative Hypothesis Test, COPD-Lung Cancer Differentiation, and Joint mirnas Biomarkers

Bootstrapped Integrative Hypothesis Test, COPD-Lung Cancer Differentiation, and Joint mirnas Biomarkers Bootstrapped Integrative Hypothesis Test, COPD-Lung Cancer Differentiation, and Joint mirnas Biomarkers Kai-Ming Jiang 1,2, Bao-Liang Lu 1,2, and Lei Xu 1,2,3(&) 1 Department of Computer Science and Engineering,

More information

CHAPTER 6 HUMAN BEHAVIOR UNDERSTANDING MODEL

CHAPTER 6 HUMAN BEHAVIOR UNDERSTANDING MODEL 127 CHAPTER 6 HUMAN BEHAVIOR UNDERSTANDING MODEL 6.1 INTRODUCTION Analyzing the human behavior in video sequences is an active field of research for the past few years. The vital applications of this field

More information

G5)H/C8-)72)78)2I-,8/52& ()*+,-./,-0))12-345)6/3/782 9:-8;<;4.= J-3/ J-3/ "#&' "#% "#"% "#%$

G5)H/C8-)72)78)2I-,8/52& ()*+,-./,-0))12-345)6/3/782 9:-8;<;4.= J-3/ J-3/ #&' #% #% #%$ # G5)H/C8-)72)78)2I-,8/52& #% #$ # # &# G5)H/C8-)72)78)2I-,8/52' @5/AB/7CD J-3/ /,?8-6/2@5/AB/7CD #&' #% #$ # # '#E ()*+,-./,-0))12-345)6/3/782 9:-8;;4. @5/AB/7CD J-3/ #' /,?8-6/2@5/AB/7CD #&F #&' #% #$

More information

Measuring Focused Attention Using Fixation Inner-Density

Measuring Focused Attention Using Fixation Inner-Density Measuring Focused Attention Using Fixation Inner-Density Wen Liu, Mina Shojaeizadeh, Soussan Djamasbi, Andrew C. Trapp User Experience & Decision Making Research Laboratory, Worcester Polytechnic Institute

More information

Performance and Saliency Analysis of Data from the Anomaly Detection Task Study

Performance and Saliency Analysis of Data from the Anomaly Detection Task Study Performance and Saliency Analysis of Data from the Anomaly Detection Task Study Adrienne Raglin 1 and Andre Harrison 2 1 U.S. Army Research Laboratory, Adelphi, MD. 20783, USA {adrienne.j.raglin.civ, andre.v.harrison2.civ}@mail.mil

More information

Two Themes. MobileASL: Making Cell Phones Accessible to the Deaf Community. Our goal: Challenges: Current Technology for Deaf People (text) ASL

Two Themes. MobileASL: Making Cell Phones Accessible to the Deaf Community. Our goal: Challenges: Current Technology for Deaf People (text) ASL Two Themes MobileASL: Making Cell Phones Accessible to the Deaf Community MobileASL AccessComputing Alliance Advancing Deaf and Hard of Hearing in Computing Richard Ladner University of Washington ASL

More information

Comparison of Two Approaches for Direct Food Calorie Estimation

Comparison of Two Approaches for Direct Food Calorie Estimation Comparison of Two Approaches for Direct Food Calorie Estimation Takumi Ege and Keiji Yanai Department of Informatics, The University of Electro-Communications, Tokyo 1-5-1 Chofugaoka, Chofu-shi, Tokyo

More information

Gray Scale Image Edge Detection and Reconstruction Using Stationary Wavelet Transform In High Density Noise Values

Gray Scale Image Edge Detection and Reconstruction Using Stationary Wavelet Transform In High Density Noise Values Gray Scale Image Edge Detection and Reconstruction Using Stationary Wavelet Transform In High Density Noise Values N.Naveen Kumar 1, J.Kishore Kumar 2, A.Mallikarjuna 3, S.Ramakrishna 4 123 Research Scholar,

More information

Salient Object Detection in Videos Based on SPATIO-Temporal Saliency Maps and Colour Features

Salient Object Detection in Videos Based on SPATIO-Temporal Saliency Maps and Colour Features Salient Object Detection in Videos Based on SPATIO-Temporal Saliency Maps and Colour Features U.Swamy Kumar PG Scholar Department of ECE, K.S.R.M College of Engineering (Autonomous), Kadapa. ABSTRACT Salient

More information

FEATURE EXTRACTION USING GAZE OF PARTICIPANTS FOR CLASSIFYING GENDER OF PEDESTRIANS IN IMAGES

FEATURE EXTRACTION USING GAZE OF PARTICIPANTS FOR CLASSIFYING GENDER OF PEDESTRIANS IN IMAGES FEATURE EXTRACTION USING GAZE OF PARTICIPANTS FOR CLASSIFYING GENDER OF PEDESTRIANS IN IMAGES Riku Matsumoto, Hiroki Yoshimura, Masashi Nishiyama, and Yoshio Iwai Department of Information and Electronics,

More information

Chapter 1. Introduction

Chapter 1. Introduction Chapter 1 Introduction 1.1 Motivation and Goals The increasing availability and decreasing cost of high-throughput (HT) technologies coupled with the availability of computational tools and data form a

More information

The Attraction of Visual Attention to Texts in Real-World Scenes

The Attraction of Visual Attention to Texts in Real-World Scenes The Attraction of Visual Attention to Texts in Real-World Scenes Hsueh-Cheng Wang (hchengwang@gmail.com) Marc Pomplun (marc@cs.umb.edu) Department of Computer Science, University of Massachusetts at Boston,

More information

Selection and Combination of Markers for Prediction

Selection and Combination of Markers for Prediction Selection and Combination of Markers for Prediction NACC Data and Methods Meeting September, 2010 Baojiang Chen, PhD Sarah Monsell, MS Xiao-Hua Andrew Zhou, PhD Overview 1. Research motivation 2. Describe

More information

Saliency aggregation: Does unity make strength?

Saliency aggregation: Does unity make strength? Saliency aggregation: Does unity make strength? Olivier Le Meur a and Zhi Liu a,b a IRISA, University of Rennes 1, FRANCE b School of Communication and Information Engineering, Shanghai University, CHINA

More information

Group-Wise FMRI Activation Detection on Corresponding Cortical Landmarks

Group-Wise FMRI Activation Detection on Corresponding Cortical Landmarks Group-Wise FMRI Activation Detection on Corresponding Cortical Landmarks Jinglei Lv 1,2, Dajiang Zhu 2, Xintao Hu 1, Xin Zhang 1,2, Tuo Zhang 1,2, Junwei Han 1, Lei Guo 1,2, and Tianming Liu 2 1 School

More information

BIOSTATISTICAL METHODS AND RESEARCH DESIGNS. Xihong Lin Department of Biostatistics, University of Michigan, Ann Arbor, MI, USA

BIOSTATISTICAL METHODS AND RESEARCH DESIGNS. Xihong Lin Department of Biostatistics, University of Michigan, Ann Arbor, MI, USA BIOSTATISTICAL METHODS AND RESEARCH DESIGNS Xihong Lin Department of Biostatistics, University of Michigan, Ann Arbor, MI, USA Keywords: Case-control study, Cohort study, Cross-Sectional Study, Generalized

More information

Gene expression analysis. Roadmap. Microarray technology: how it work Applications: what can we do with it Preprocessing: Classification Clustering

Gene expression analysis. Roadmap. Microarray technology: how it work Applications: what can we do with it Preprocessing: Classification Clustering Gene expression analysis Roadmap Microarray technology: how it work Applications: what can we do with it Preprocessing: Image processing Data normalization Classification Clustering Biclustering 1 Gene

More information

Valence-arousal evaluation using physiological signals in an emotion recall paradigm. CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry.

Valence-arousal evaluation using physiological signals in an emotion recall paradigm. CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry. Proceedings Chapter Valence-arousal evaluation using physiological signals in an emotion recall paradigm CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry Abstract The work presented in this paper aims

More information

Cancer Cells Detection using OTSU Threshold Algorithm

Cancer Cells Detection using OTSU Threshold Algorithm Cancer Cells Detection using OTSU Threshold Algorithm Nalluri Sunny 1 Velagapudi Ramakrishna Siddhartha Engineering College Mithinti Srikanth 2 Velagapudi Ramakrishna Siddhartha Engineering College Kodali

More information

Neuromorphic convolutional recurrent neural network for road safety or safety near the road

Neuromorphic convolutional recurrent neural network for road safety or safety near the road Neuromorphic convolutional recurrent neural network for road safety or safety near the road WOO-SUP HAN 1, IL SONG HAN 2 1 ODIGA, London, U.K. 2 Korea Advanced Institute of Science and Technology, Daejeon,

More information

Applied Machine Learning in Biomedicine. Enrico Grisan

Applied Machine Learning in Biomedicine. Enrico Grisan Applied Machine Learning in Biomedicine Enrico Grisan enrico.grisan@dei.unipd.it Algorithm s objective cost Formal objective for algorithms: - minimize a cost function - maximize an objective function

More information

Intelligent Edge Detector Based on Multiple Edge Maps. M. Qasim, W.L. Woon, Z. Aung. Technical Report DNA # May 2012

Intelligent Edge Detector Based on Multiple Edge Maps. M. Qasim, W.L. Woon, Z. Aung. Technical Report DNA # May 2012 Intelligent Edge Detector Based on Multiple Edge Maps M. Qasim, W.L. Woon, Z. Aung Technical Report DNA #2012-10 May 2012 Data & Network Analytics Research Group (DNA) Computing and Information Science

More information

Assigning B cell Maturity in Pediatric Leukemia Gabi Fragiadakis 1, Jamie Irvine 2 1 Microbiology and Immunology, 2 Computer Science

Assigning B cell Maturity in Pediatric Leukemia Gabi Fragiadakis 1, Jamie Irvine 2 1 Microbiology and Immunology, 2 Computer Science Assigning B cell Maturity in Pediatric Leukemia Gabi Fragiadakis 1, Jamie Irvine 2 1 Microbiology and Immunology, 2 Computer Science Abstract One method for analyzing pediatric B cell leukemia is to categorize

More information

Local Image Structures and Optic Flow Estimation

Local Image Structures and Optic Flow Estimation Local Image Structures and Optic Flow Estimation Sinan KALKAN 1, Dirk Calow 2, Florentin Wörgötter 1, Markus Lappe 2 and Norbert Krüger 3 1 Computational Neuroscience, Uni. of Stirling, Scotland; {sinan,worgott}@cn.stir.ac.uk

More information

Bangor University Laboratory Exercise 1, June 2008

Bangor University Laboratory Exercise 1, June 2008 Laboratory Exercise, June 2008 Classroom Exercise A forest land owner measures the outside bark diameters at.30 m above ground (called diameter at breast height or dbh) and total tree height from ground

More information

Quantifying the Effect of Disruptions to Temporal Coherence on the Intelligibility of Compressed American Sign Language Video

Quantifying the Effect of Disruptions to Temporal Coherence on the Intelligibility of Compressed American Sign Language Video Quantifying the Effect of Disruptions to Temporal Coherence on the ntelligibility of Compressed American Sign Language Video Frank M. Ciaramello and Sheila S. Hemami Visual Communication Laboratory School

More information

Automated Assessment of Diabetic Retinal Image Quality Based on Blood Vessel Detection

Automated Assessment of Diabetic Retinal Image Quality Based on Blood Vessel Detection Y.-H. Wen, A. Bainbridge-Smith, A. B. Morris, Automated Assessment of Diabetic Retinal Image Quality Based on Blood Vessel Detection, Proceedings of Image and Vision Computing New Zealand 2007, pp. 132

More information

Jitter-aware time-frequency resource allocation and packing algorithm

Jitter-aware time-frequency resource allocation and packing algorithm Jitter-aware time-frequency resource allocation and packing algorithm The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As

More information

Introduction to Machine Learning. Katherine Heller Deep Learning Summer School 2018

Introduction to Machine Learning. Katherine Heller Deep Learning Summer School 2018 Introduction to Machine Learning Katherine Heller Deep Learning Summer School 2018 Outline Kinds of machine learning Linear regression Regularization Bayesian methods Logistic Regression Why we do this

More information

Computational Models of Visual Attention: Bottom-Up and Top-Down. By: Soheil Borhani

Computational Models of Visual Attention: Bottom-Up and Top-Down. By: Soheil Borhani Computational Models of Visual Attention: Bottom-Up and Top-Down By: Soheil Borhani Neural Mechanisms for Visual Attention 1. Visual information enter the primary visual cortex via lateral geniculate nucleus

More information

Experiences on Attention Direction through Manipulation of Salient Features

Experiences on Attention Direction through Manipulation of Salient Features Experiences on Attention Direction through Manipulation of Salient Features Erick Mendez Graz University of Technology Dieter Schmalstieg Graz University of Technology Steven Feiner Columbia University

More information

Learning Spatiotemporal Gaps between Where We Look and What We Focus on

Learning Spatiotemporal Gaps between Where We Look and What We Focus on Express Paper Learning Spatiotemporal Gaps between Where We Look and What We Focus on Ryo Yonetani 1,a) Hiroaki Kawashima 1,b) Takashi Matsuyama 1,c) Received: March 11, 2013, Accepted: April 24, 2013,

More information

International Journal of Computational Science, Mathematics and Engineering Volume2, Issue6, June 2015 ISSN(online): Copyright-IJCSME

International Journal of Computational Science, Mathematics and Engineering Volume2, Issue6, June 2015 ISSN(online): Copyright-IJCSME Various Edge Detection Methods In Image Processing Using Matlab K. Narayana Reddy 1, G. Nagalakshmi 2 12 Department of Computer Science and Engineering 1 M.Tech Student, SISTK, Puttur 2 HOD of CSE Department,

More information

Discovering Meaningful Cut-points to Predict High HbA1c Variation

Discovering Meaningful Cut-points to Predict High HbA1c Variation Proceedings of the 7th INFORMS Workshop on Data Mining and Health Informatics (DM-HI 202) H. Yang, D. Zeng, O. E. Kundakcioglu, eds. Discovering Meaningful Cut-points to Predict High HbAc Variation Si-Chi

More information

Estimating Multiple Evoked Emotions from Videos

Estimating Multiple Evoked Emotions from Videos Estimating Multiple Evoked Emotions from Videos Wonhee Choe (wonheechoe@gmail.com) Cognitive Science Program, Seoul National University, Seoul 151-744, Republic of Korea Digital Media & Communication (DMC)

More information

Perceptual-Based Objective Picture Quality Measurements

Perceptual-Based Objective Picture Quality Measurements Perceptual-Based Objective Picture Quality Measurements Introduction In video systems, a wide range of video processing devices can affect overall picture quality. Encoders and decoders compress and decompress

More information

Notes for laboratory session 2

Notes for laboratory session 2 Notes for laboratory session 2 Preliminaries Consider the ordinary least-squares (OLS) regression of alcohol (alcohol) and plasma retinol (retplasm). We do this with STATA as follows:. reg retplasm alcohol

More information

IAT 355 Visual Analytics. Encoding Information: Design. Lyn Bartram

IAT 355 Visual Analytics. Encoding Information: Design. Lyn Bartram IAT 355 Visual Analytics Encoding Information: Design Lyn Bartram 4 stages of visualization design 2 Recall: Data Abstraction Tables Data item (row) with attributes (columns) : row=key, cells = values

More information

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Midterm, 2016

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Midterm, 2016 UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Midterm, 2016 Exam policy: This exam allows one one-page, two-sided cheat sheet; No other materials. Time: 80 minutes. Be sure to write your name and

More information

Fusing Generic Objectness and Visual Saliency for Salient Object Detection

Fusing Generic Objectness and Visual Saliency for Salient Object Detection Fusing Generic Objectness and Visual Saliency for Salient Object Detection Yasin KAVAK 06/12/2012 Citation 1: Salient Object Detection: A Benchmark Fusing for Salient Object Detection INDEX (Related Work)

More information

Sparse Coding in Sparse Winner Networks

Sparse Coding in Sparse Winner Networks Sparse Coding in Sparse Winner Networks Janusz A. Starzyk 1, Yinyin Liu 1, David Vogel 2 1 School of Electrical Engineering & Computer Science Ohio University, Athens, OH 45701 {starzyk, yliu}@bobcat.ent.ohiou.edu

More information

ANALYSIS OF FACIAL FEATURES OF DRIVERS UNDER COGNITIVE AND VISUAL DISTRACTIONS

ANALYSIS OF FACIAL FEATURES OF DRIVERS UNDER COGNITIVE AND VISUAL DISTRACTIONS ANALYSIS OF FACIAL FEATURES OF DRIVERS UNDER COGNITIVE AND VISUAL DISTRACTIONS Nanxiang Li and Carlos Busso Multimodal Signal Processing (MSP) Laboratory Department of Electrical Engineering, The University

More information

Empirical Analysis of Object-Oriented Design Metrics for Predicting High and Low Severity Faults

Empirical Analysis of Object-Oriented Design Metrics for Predicting High and Low Severity Faults IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, VOL. 32, NO. 10, OCTOBER 2006 771 Empirical Analysis of Object-Oriented Design Metrics for Predicting High and Low Severity Faults Yuming Zhou and Hareton Leung,

More information

Quantitative Evaluation of Edge Detectors Using the Minimum Kernel Variance Criterion

Quantitative Evaluation of Edge Detectors Using the Minimum Kernel Variance Criterion Quantitative Evaluation of Edge Detectors Using the Minimum Kernel Variance Criterion Qiang Ji Department of Computer Science University of Nevada Robert M. Haralick Department of Electrical Engineering

More information

On Shape And the Computability of Emotions X. Lu, et al.

On Shape And the Computability of Emotions X. Lu, et al. On Shape And the Computability of Emotions X. Lu, et al. MICC Reading group 10.07.2013 1 On Shape and the Computability of Emotion X. Lu, P. Suryanarayan, R. B. Adams Jr., J. Li, M. G. Newman, J. Z. Wang

More information

Object-Level Saliency Detection Combining the Contrast and Spatial Compactness Hypothesis

Object-Level Saliency Detection Combining the Contrast and Spatial Compactness Hypothesis Object-Level Saliency Detection Combining the Contrast and Spatial Compactness Hypothesis Chi Zhang 1, Weiqiang Wang 1, 2, and Xiaoqian Liu 1 1 School of Computer and Control Engineering, University of

More information

Development of goal-directed gaze shift based on predictive learning

Development of goal-directed gaze shift based on predictive learning 4th International Conference on Development and Learning and on Epigenetic Robotics October 13-16, 2014. Palazzo Ducale, Genoa, Italy WePP.1 Development of goal-directed gaze shift based on predictive

More information

Object-based Saliency as a Predictor of Attention in Visual Tasks

Object-based Saliency as a Predictor of Attention in Visual Tasks Object-based Saliency as a Predictor of Attention in Visual Tasks Michal Dziemianko (m.dziemianko@sms.ed.ac.uk) Alasdair Clarke (a.clarke@ed.ac.uk) Frank Keller (keller@inf.ed.ac.uk) Institute for Language,

More information

Using Dynamic Time Warping for Intuitive Handwriting Recognition

Using Dynamic Time Warping for Intuitive Handwriting Recognition Using Dynamic Time Warping for Intuitive Handwriting Recognition Ralph Niels and Louis Vuurpijl Nijmegen Institute for Cognition and Information (Radboud University Nijmegen) P.O. Box 9104, 6500 HE Nijmegen,

More information

Firefighter safety: How visible is your protective clothing?

Firefighter safety: How visible is your protective clothing? News Article ID: 4028 Fire international 23 April 2003 Firefighter safety: How visible is your protective clothing? TNO Human Factors Research Institute has developed a method to determine visual conspicuity

More information

Age (continuous) Gender (0=Male, 1=Female) SES (1=Low, 2=Medium, 3=High) Prior Victimization (0= Not Victimized, 1=Victimized)

Age (continuous) Gender (0=Male, 1=Female) SES (1=Low, 2=Medium, 3=High) Prior Victimization (0= Not Victimized, 1=Victimized) Criminal Justice Doctoral Comprehensive Exam Statistics August 2016 There are two questions on this exam. Be sure to answer both questions in the 3 and half hours to complete this exam. Read the instructions

More information