Saliency Inspired Modeling of Packet-loss Visibility in Decoded Videos
|
|
- Paul Casey
- 6 years ago
- Views:
Transcription
1 1 Saliency Inspired Modeling of Packet-loss Visibility in Decoded Videos Tao Liu*, Xin Feng**, Amy Reibman***, and Yao Wang* *Polytechnic Institute of New York University, Brooklyn, NY, U.S. **Chongqing University, Chongqing, P.R. China ***AT&T Labs-Research, Florham Park, NJ, U.S. Abstract The visibility of packet loss in decoded video depends on various factors and their complicated interactions, such as loss severity and duration, and characteristics of background signal. However, visual attention or saliency may play an important role as well. In this work, we investigate how to improve visibility prediction by incorporating the saliency information. Based on earlier findings about how saliency affects the perceptual quality of video with packet losses, we propose several saliency-based factors and incorporate them into a Generalized Linear Model (GLM) to predict loss visibility. Test results with 1080 MPEG-2 packet losses indicate that saliency information can help improve the prediction accuracy about 12% over nonsaliency-based model, and that saliency-weighted mean-squareerror and variation of saliency information are promising metrics. Index Terms Packet loss visibility, perceptual video quality, saliency information, GLM important component of the Human Visual System (HVS), describes features of an image (for example, color, intensity and orientation) which cause one region to stand out relative to other regions [6]. Because of both physiological and psychological evidence that humans have high selectivity on exposed visual information [6], there is recent interest in designing image and video quality metrics that weight signal errors based on their visual sensitivities [4][5][10]. Some evidence suggests incorporating saliency during spatial pooling may not always improve quality metrics [5]. However, in this paper, we demonstrate that adding saliency information to the problem of predicting the visibility of impairments due to packet losses can dramatically improve prediction performance. We focus here on individual packet losses in high-quality video that has few other visual artifacts. I. INTRODUCTION In video transmission, the decoder may not receive all the encoded video data because of the losses occurred in packet networks. Generally, different packet losses may not cause equal video quality degradation, and in fact some may not be visible to human viewers. On the other hand, assessing the impact of packet loss is crucial for network providers to monitor and control their networks to guarantee satisfactory video quality for end users. However, because of significant difficulty in assessing the perceptual quality of video affected by packet loss directly, investigating the visibility of packet losses is a natural and appropriate starting point. In the literature, some researchers reported their findings about the visibility of packet loss in video transmission. The work in [1][7] accurately modeled the visibility of single packet loss by a Generalized Linear Model (GLM) with various factors, such as signal error, video motion, loss position and etc. The authors of [8] extended this to predict visibility of multiple losses. The work in [3] introduced an intuitive concept Mean Time Between Failures (MTBF) that video quality can be interpreted as the probability of visible artifacts, and estimate it with several existing quality metrics. In [9], the visibility of single packet loss in compressed video is evaluated by loss severity (in PSNR drop) caused by lost frame, and the duration of its propagated error. In this work, we explore whether saliency information can improve the prediction of packet-loss visibility. Saliency, an Specifically, we still use the GLM method to predict loss visibility, but introduce new factors that incorporate saliency information. Motivated by an earlier finding [10] that weighting the pixel-wise errors by the visual saliency map can predict the perceptual quality better than using non-weighted errors, we propose to use saliency weighted error as a factor. Furthermore, we also investigate factors summarizing the saliency map itself, including how the saliency map changes when there is a packet loss, and the temporal variation of the saliency map. In this work, we explore packet loss visibility using the subjective data in [1]. We first design a GLM with previously identified non-saliency factors as a benchmark, and then we incorporate our aforementioned saliency-based factors into the GLM. We compute the saliency map using the computational visual attention model developed by Itti et al. [6]. Our tests show that both saliency-weighted error and changes in saliency map are good metrics to predict loss visibility, and they can significantly improve the prediction accuracy of GLM over the one using non-saliency factors. The remainder of this paper is organized as follows. All factors we consider, both with and without saliency, are discussed in Section II. The subjective data is described briefly in Section III. In Section IV, we design two GLMs, with and without saliency factors, and we analyze and compare their performances. We conclude this paper with Section V.
2 2 A. Non-saliency factors II. FACTORS AFFECTING VISIBILITY In [1][7], there were totally 20 non-saliency factors (and their variations) proposed, which covers both characteristics of videos and packet loss impairments. Furthermore, the interaction between them on a scene-level was considered as well. Here we only briefly discuss them in each category; refer to [1][7] for more detailed descriptions. (a) Error characteristics: Mean Squared Error (MSE) and Structural Similarity Index Metric (SSIM) are two widely accepted quality metrics. In the loss-visibility scenario, for the sake of easy calculation, two simplified variations of each metric are used to predict the visibility of packet loss: those measurements of initial frame in loss-affected segment, IMSE and ISSIM; and the extreme values of IMSE and ISSIM in macroblock level, MaxIMSEmb and MinISSIMmb. When there is a single slice loss (as opposed to the loss of an entire frame), the impact of discontinuities caused by lost slices on video quality can be measured by the Slice-Boundary Mismatch (SBM), first proposed in [13] and modified in minor details in [7]. Only SBM on the initial frame of loss-affected segment, ISBM, is considered. Additionally, some important content-independent measures, such as spatial extent, or SXTNT (the number of slices lost in one frame), HGT (the average height of the lost slices), Duration (duration of the loss-affected segment), are also considered. (b) Video characteristics: Motion is one of the most important characteristics of videos. Therefore,the mean and variance of the magnitudes of the motion vectors across all macroblocks initially affected by a loss, MotMean and MotVar can also be used to predict the visibility of a loss. SigMean and SigVar, the mean and variance of intensity values of the initial frame of loss-affected segment, and ResidEng (residual energy after motion compensation) of that frame are also effective. (c) Scene-level characteristics: In addition to the factors in (a) and (b), some high-level characteristics are also considered. The authors of [7] show that the relative position between scene change and packet loss impairment has great influence on its visibility. Therefore, D2R (the distance between the current frame (with packet loss) and the reference frame used for concealment)(farconceal is set when values of D2R are equal to or greater than 3), DistFromCut (the distance in time between the first frame affected by the packet loss and the nearest scene cut, either before or after) and its threshold versions, AtScene, BeforeScene, and AfterScene, are considered in this work. Since these non-saliency factors were proved capable to predict the loss visibility, we use them as candidate factors to design a GLM. B. Saliency-based factors In the previous subsection, we discuss a score of factors covering many attributes of packet loss impairment. However, these are only heuristically linked to properties of the HVS. We hereby investigate factors associated with saliency in this subsection. We adopt the widely accepted saliency-based visual attention model (SVAM) [6] proposed by Itti. et al. to calculate the saliency information of our test videos. Itti s SVAM decomposes an input image into a set of multiscale feature maps of color, intensity and orientation. By using a center-surround mechanism, SVAM accurately simulates which location in the image will automatically and unconsciously attract visual attention. Fig.1 shows the saliency maps of original and loss-imaired video frames (for demonstration purpose only, not the actual test video). Figure 1. (a) (c) Saliency maps of original and loss-affected video frames. We consider two basic attributes of the computed saliency information for their ability to predict loss visibility. First, inspired by the results in our previous work [9] that weighting the pixel-wise errors by the saliency values correlates well with the perceptual quality of loss-impaired videos, we propose to supplement the IMSE factor by saliency-weighted IMSE, denoted by IMSE_Sal. We also consider the saliencyweighted MSE computed over all loss-affected frames, yielding MSE_Sal. Second, in addition to the impact of saliency on pixel-wise error, we investigate properties of the saliency information itself. If we examine the saliency maps of both the original and the loss-impaired video frames from Fig. 1, we discover that packet losses not only distort the video frames but also alter the distribution of salient regions across the affected frames spatially and temporally. We also observ that packet losses are more visible in videos where the saliency map changes rapidly in time. These two observations lead us to propose two additional factors: SMSE, which measures the changes (in terms of MSE) between saliency maps of original and loss-impaired frames (only in the position where loss happens); STV, which measures temporal variation of the saliency map of loss-impaired frames, respectively. They are defined mathematically as follows with SMSE = E t [ E(x,y) ( S 1 (x, y, t) S 2 (x, y, t) 2 ) ] (1) (b) (d) ST V = STD t (SM 2 (t)) (2) SM 2 (t) = E (x,y) (S 2 (x, y, t)) (3) where E (x,y) () is the 2-D mean operator averaging over all pixels in frame t, E t () is the mean operator averaging over
3 3 time in the segment of loss-impaired sequence, and STD t () is the standard deviation operator over time for that segment. S i (x, y, t) denotes the saliency value at position (x, y) in frame t, and i = 1or2 refers to the original or distorted video sequence. Note that, for saliency computation, we tested two methods, one using color, orientation, and intensity information only, as in the original Itti s model [6]; another one further using motion information with the motion features computed following [11]. From our previous study, the second method produces saliency maps that are more consistent with our visual inspections, although it requires extra compuation over the former one in saliency detection. Therfore, we focus on the latter one in this work. Table I summarizes all the aforementioned factors. Table I LIST OF ALL NON-SALIENCY AND SALIENCY FACTORS 1 IMSE 14 D2R 2 MaxIMSEmb 15 SigVar 3 ISSIM 16 DistFromCut 4 MinSSIMmb 17 AtScene 5 ISBM 18 BeforeScene 6 SXTNT 19 AfterScene 7 HGT 20 FarConceal 8 Duration 21 IMSE_Sal (no motion) 9 ResidEng 22 IMSE_Sal 10 CameraMotion 23 MSE_Sal 11 SigMean 24 S_MSE 12 MotMean 25 STV 13 MotVar III. SUBJECTIVE TEST The subjective data used in this work was first presented in [1]; to be self-contained, we describe it briefly here. This subjective test was designed not to assess the quality of video at a given packet loss rate, but instead to learn about what affects the visibility of impairments caused by individual packet losses. The test videos shown were compressed with MPEG-2 at 720x480 resolution and 30 fps frame rate, with various scene contents and different camera motions, using 13-frame GOPs with 2 B-frames before every P-frame at a bitrate of around 4Mbps. One isolated packet loss was randomly inserted into the video in every 4-second window. The 1080 packet losses affected either one slice, two slices, or an entire frame. The decoder applied Zero-motion error concealment (copying macroblocks from the closest reference frame) when losses occurred. Each packet loss was viewed by 12 viewers, whose task was to indicate each time they saw an artifact while were watching a 6-minute continuous video clip. The ground truth of visibility of each packet loss was defined as the percentage of viewers who indicated they saw the loss. For the detailed information about this subjective test, please refer to [1]. A. GLM fitting method IV. GENERALIZED LINEAR MODELS As [1][7], we model the probability of visibility using a GLM [2], which is a development of linear models to accommodate both non-normal response distributions and transformations to linearity in a straightforward way. We use logistic regression to fit our model, with logit() as the link function. With the help of statistical software R [12], the model is fit with an iteratively re-weighted least-square method to generate a maximum-likelihood estimate. The GLM fittings suggested in [2] can be summarized to the following 3 steps: Step 1: Before designing the GLM with multiple variables, each individual factor is analyzed first. By examining the Lowess smoothed scatter plot between each factor and logit function of subjective visibility, the best (in prediction error) form of each factor, e.g. log(), can be determined. Step 2: Since some of the factors are correlated, it is possible to overfit the model by using all of them together. To select factors into the model, we took a step-wise approach by adding one factor at a time. We begin with Null model, and in each step, the one factor that generates maximum reduction of prediction error in the cross-validation process is added to the model, so that the efficiency of each inclusion is maximized. When we can no longer reduce the prediction error by adding a single factor, the model is preliminarily established. Step 3: We check if there are any interactions (products) between any two selected factors that can further improve the prediction of the model. If there are multiple such pairs, the decisions of their inclusion are made in the similar fashion as Step 2. Note that we perform 10-fold cross-validation in the process of building up the model. Specifically, we randomly divide the entire data set of 1080 losses into 10 groups of equal size and choose the data from 9 out of the 10 sets as a training set. The remaining data set is used for testing. We repeat this process 10 times, each time choosing a different set for training. The average prediction error is used as the performance measure. B. GLMs of two models In order to test the impact of saliency information on modeling the loss visibility, we fit the same subjective data with two different set of factors, one containing only nonsaliency-based factors (Model 1); the other one containing all the aforementioned factors (Model 2). We note that our model differs from that in [7] because it uses just one subjective dataset. The factors and their coefficients of both models are summarized in Table II. To test the significance of each factor in the model, including the interaction terms added in the third step, we re-perform Step 2 with all the selected factors and interaction terms to update their inclusion order. The new order provides a ranking of the significance of each factor and interaction term. This allows the two models to be compared if we limit each to having the same number of predictive factors. We draw the bar plots of factors and model prediction errors of this stepwise procedure, in the order of their inclusions, in Fig. 2 (a) and 2 (b). C. Comparison Model 1 v.s. Model 2 To compare the prediction performances of Model 1 and Model 2, we show the relationship between the number of
4 4 (a) (b) Figure 2. Factor inclusions of (a) Model 1; and (b) Model 2.
5 5 Table II COEFFICIENTS OF MODEL 1 AND MODEL 2 Model 1 Model 2 factor coef. factor coef. MinSSIMmb e+00 MinSSIMmb e-01 FarConceal* FarConceal* 1.929e-01 log(imse) (IMSE_Sal) e-01 StillCamera e-01 StillCamera e+00 log(reseng) e-01 log(reseng) e+00 log(imse) 7.209e-01 log(imse) 7.090e-01 log(isbm)* (SXTNT!=30) 6.590e-02 SXTNT e-01 HGT e-02 log(reseng)* 3.027e+00 SigMean e-02 StillCamera* e+00 AtScene 1.404e+00 HGT e-02 Duration< e-01 SigMean e-03 BeforeScene e e+01 MotMean> e-01 BeforeScene e-01 SigVar e-04 log(1-issim) 2.915e-01 DistFromCut e-02 Duration< e-01 Duration 1.578e+01 SigVar* FarConceal*(IMSE_Sal) e-04 SigVar 8.905e-04 log(stv) 1.715e-01 DistFromCut e-03 log(isbm)* SXTNT!= e-02 MotMean> e-01 (MotMean>0.707)* FarConceal*(IMSE_Sal) e-01 AtScene * 1.472e+01 AtScene e+00 factors used in the model and prediction error reduction ratio (Model 2 to Model 1) in Fig.3. We can see that the overall prediction error of Model 2 ( ) is about 12% less than that of Model 1( ). When both models are limited to use 15 factors, Model 2 still outperforms Model 1 by about 9%. In addition, we can see that, no matter how many factors are used to fit the models, our model with saliency-based factors always outperform those without, except for the case of only one factor, since that factor (MinSSIMmb) is the same for both models. Therefore, we conclude that saliency information significantly boosts the visibility prediction performance. To gain a clearer picture of the contribution of each individual saliency-based factor, we examine the inclusion orders (or significance ranks) of saliency-based factors in Model 2 in Fig. 2 (b). There are 4 saliency factors in the first half of the 23 factors in the model: FarConceal*(IMSE_Sal) 1 4, log(resideng)*, StillCamera*, and. If we expand our focus to the first 15 (the number of factors used in Model 1) factors, SigVar*FarConceal*(IMSE_Sal) 1 4 is also present. Note that the saliency-weighted pixel-wise error (IMSE_Sal) and the difference of saliency maps of original and distorted video caused by packet loss (S_MSE) are two very helpful factors in modeling packet loss visibility. Admittedly, this improvement is at the expense of complex computation of the saliency detection system, unless there is any salinecy or foveal information avaiable, such as in some Figure 3. Performance comparison between Model 1 and Model 2. ( The number of factors used in Model 1 is limited to 15.) particular types of video contents, e.g. news broadcasting or soccer games, where salient regions are obvious. V. CONCLUSIONS In this paper, we proposed several saliency-based factors, which can be used to improve the prediction of the visibility of packet losses. Together with a variety of non-saliencybased factors, we fit the GLMs with and without our proposed factors using existing subjective data, and the results show that saliency-based factors significantly improve the performance in loss-visibility modeling. Our prior work in perceptual quality prediction and the current work in prediciting packet loss have shown that saliency information is helpful for both. One interesting question is whether we can predict the perceptual quality from the packet loss visibility. This is one of our current research directions. REFERENCES [1] S. Kanumuri, et al, Modeling Packet-loss Visibility in MPEG-2 Video IEEE Trans. Multimedia, vol.8, pp , Apr [2] P. McCullagh, et al., Generalized Linear Models, 2nd Edition, London, U.K.: Chapman & Hall. [3] N. Suresh, et al., Mean Time Between Failures: A Subjectively Meaningful Video Quality Metric, ICASSP,2006 [4] Z. Wang, et al., Image Quality Assessment:From Error Visibility to Structure Similarity IEEE Trans. Im. Proc. Vol.13, Apr [5] A. Ninassi, et al., Does Where You Gaze on An Image Affect Your Perception of Quality? Applying Visual Attention To Image Quality Metric, Intl. Cof. Im. Proc., 2007 [6] L. Itti, et al., A Model of Saliency-Based Visual Attention for Rapid Scene Analysis, IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 20, no.11, pp , Nov [7] A. Reibman, and D. Poole, Predicting packet-loss visibility using scene characteristics, Packet Video, 2007 [8] S. Kanumuri, et al., Predicting H.264 Packet Loss Visibility using a Generalized Linear Model, ICIP, 2006 [9] T. Liu, et al.. Subjective Quality Evaluation of Decoded Video in The Presence of Packet Losses ICASSP 2007 [10] X. Feng, et al., Saliency Based Objective Quality Assessment of Decoded Video Affected by Packet Losses, ICIP, [11] D. Walther, Interactions of Visual Attention and Object Recognition: Computational Modeling, Algorithms, and Psychophysics, PhD thesis, California Institute of Technology, Pasadena, CA, Feb [12] R software is available at [13] H. Rui et al., Evaluation of packet loss impairment on streaming video, J. of Zhejiang University SCIENCE, vol. 7, April 2006.
Saliency Inspired Full-Reference Quality Metrics for Packet-Loss-Impaired Video
IEEE TRANSACTIONS ON BROADCASTING, VOL. 57, NO. 1, MARCH 2011 81 Saliency Inspired Full-Reference Quality Metrics for Packet-Loss-Impaired Video Xin Feng, Tao Liu, Member, IEEE, Dan Yang, and Yao Wang,
More informationINCORPORATING VISUAL ATTENTION MODELS INTO IMAGE QUALITY METRICS
INCORPORATING VISUAL ATTENTION MODELS INTO IMAGE QUALITY METRICS Welington Y.L. Akamine and Mylène C.Q. Farias, Member, IEEE Department of Computer Science University of Brasília (UnB), Brasília, DF, 70910-900,
More informationValidating the Visual Saliency Model
Validating the Visual Saliency Model Ali Alsam and Puneet Sharma Department of Informatics & e-learning (AITeL), Sør-Trøndelag University College (HiST), Trondheim, Norway er.puneetsharma@gmail.com Abstract.
More informationThe 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian
The 29th Fuzzy System Symposium (Osaka, September 9-, 3) A Fuzzy Inference Method Based on Saliency Map for Prediction Mao Wang, Yoichiro Maeda 2, Yasutake Takahashi Graduate School of Engineering, University
More informationVideo Saliency Detection via Dynamic Consistent Spatio- Temporal Attention Modelling
AAAI -13 July 16, 2013 Video Saliency Detection via Dynamic Consistent Spatio- Temporal Attention Modelling Sheng-hua ZHONG 1, Yan LIU 1, Feifei REN 1,2, Jinghuan ZHANG 2, Tongwei REN 3 1 Department of
More informationVIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING
VIDEO SALIENCY INCORPORATING SPATIOTEMPORAL CUES AND UNCERTAINTY WEIGHTING Yuming Fang, Zhou Wang 2, Weisi Lin School of Computer Engineering, Nanyang Technological University, Singapore 2 Department of
More informationFROM H.264 TO HEVC: CODING GAIN PREDICTED BY OBJECTIVE VIDEO QUALITY ASSESSMENT MODELS. Kai Zeng, Abdul Rehman, Jiheng Wang and Zhou Wang
FROM H. TO : CODING GAIN PREDICTED BY OBJECTIVE VIDEO QUALITY ASSESSMENT MODELS Kai Zeng, Abdul Rehman, Jiheng Wang and Zhou Wang Dept. of Electrical and Computer Engineering, University of Waterloo, Waterloo,
More informationMobileASL: Making Cell Phones Accessible to the Deaf Community. Richard Ladner University of Washington
MobileASL: Making Cell Phones Accessible to the Deaf Community Richard Ladner University of Washington American Sign Language (ASL) ASL is the preferred language for about 500,000-1,000,000 Deaf people
More informationCan You See Me Now? Communication Over Cell Phones for Deaf People. Eve Riskin Richard Ladner Computer Engineering University of Washington
Can You See Me Now? Communication Over Cell Phones for Deaf People Eve Riskin Richard Ladner Computer Engineering University of Washington Co-PI Thanks Sheila Hemami (Cornell) Graduate Students Anna Cavender,
More informationANALYSIS OF MOTION-BASED CODING FOR SIGN LANGUAGE VIDEO COMMUNICATION APPLICATIONS
ANALYSIS OF MOTION-BASED CODING FOR SIGN LANGUAGE VIDEO COMMUNICATION APPLICATIONS Benjaporn Saksiri*, Supavadee Aramvith**, Chai Phongphanphanee* *Information System Engineering Laboratory, Department
More informationMobileASL: Making Cell Phones Accessible to the Deaf Community
American Sign Language (ASL) MobileASL: Making Cell Phones Accessible to the Deaf Community Richard Ladner University of Washington ASL is the preferred language for about 500,000-1,000,000 Deaf people
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 19: Contextual Guidance of Attention Chris Lucas (Slides adapted from Frank Keller s) School of Informatics University of Edinburgh clucas2@inf.ed.ac.uk 20 November
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 15: Visual Attention Chris Lucas (Slides adapted from Frank Keller s) School of Informatics University of Edinburgh clucas2@inf.ed.ac.uk 14 November 2017 1 / 28
More informationComputational Cognitive Science. The Visual Processing Pipeline. The Visual Processing Pipeline. Lecture 15: Visual Attention.
Lecture 15: Visual Attention School of Informatics University of Edinburgh keller@inf.ed.ac.uk November 11, 2016 1 2 3 Reading: Itti et al. (1998). 1 2 When we view an image, we actually see this: The
More informationSeamless Audio Splicing for ISO/IEC Transport Streams
Seamless Audio Splicing for ISO/IEC 13818 Transport Streams A New Framework for Audio Elementary Stream Tailoring and Modeling Seyfullah Halit Oguz, Ph.D. and Sorin Faibish EMC Corporation Media Solutions
More informationFramework for Comparative Research on Relational Information Displays
Framework for Comparative Research on Relational Information Displays Sung Park and Richard Catrambone 2 School of Psychology & Graphics, Visualization, and Usability Center (GVU) Georgia Institute of
More informationAn Evaluation of Motion in Artificial Selective Attention
An Evaluation of Motion in Artificial Selective Attention Trent J. Williams Bruce A. Draper Colorado State University Computer Science Department Fort Collins, CO, U.S.A, 80523 E-mail: {trent, draper}@cs.colostate.edu
More informationAn Attentional Framework for 3D Object Discovery
An Attentional Framework for 3D Object Discovery Germán Martín García and Simone Frintrop Cognitive Vision Group Institute of Computer Science III University of Bonn, Germany Saliency Computation Saliency
More informationEmotion Recognition using a Cauchy Naive Bayes Classifier
Emotion Recognition using a Cauchy Naive Bayes Classifier Abstract Recognizing human facial expression and emotion by computer is an interesting and challenging problem. In this paper we propose a method
More informationFacial expression recognition with spatiotemporal local descriptors
Facial expression recognition with spatiotemporal local descriptors Guoying Zhao, Matti Pietikäinen Machine Vision Group, Infotech Oulu and Department of Electrical and Information Engineering, P. O. Box
More informationA NOVEL VARIABLE SELECTION METHOD BASED ON FREQUENT PATTERN TREE FOR REAL-TIME TRAFFIC ACCIDENT RISK PREDICTION
OPT-i An International Conference on Engineering and Applied Sciences Optimization M. Papadrakakis, M.G. Karlaftis, N.D. Lagaros (eds.) Kos Island, Greece, 4-6 June 2014 A NOVEL VARIABLE SELECTION METHOD
More informationNatural Scene Statistics and Perception. W.S. Geisler
Natural Scene Statistics and Perception W.S. Geisler Some Important Visual Tasks Identification of objects and materials Navigation through the environment Estimation of motion trajectories and speeds
More informationSUPPLEMENTARY INFORMATION. Table 1 Patient characteristics Preoperative. language testing
Categorical Speech Representation in the Human Superior Temporal Gyrus Edward F. Chang, Jochem W. Rieger, Keith D. Johnson, Mitchel S. Berger, Nicholas M. Barbaro, Robert T. Knight SUPPLEMENTARY INFORMATION
More informationFrequency Tracking: LMS and RLS Applied to Speech Formant Estimation
Aldebaro Klautau - http://speech.ucsd.edu/aldebaro - 2/3/. Page. Frequency Tracking: LMS and RLS Applied to Speech Formant Estimation ) Introduction Several speech processing algorithms assume the signal
More informationResearch Methods in Forest Sciences: Learning Diary. Yoko Lu December Research process
Research Methods in Forest Sciences: Learning Diary Yoko Lu 285122 9 December 2016 1. Research process It is important to pursue and apply knowledge and understand the world under both natural and social
More informationNo-Reference Video quality assessment of H.264 video streams based on semantic saliency maps
No-Reference Video quality assessment of H.264 video streams based on semantic saliency maps Hugo Boujut, Jenny Benois-Pineau, Toufik Ahmed, Ofer Hadar, Patrick Bonnet To cite this version: Hugo Boujut,
More informationA Visual Saliency Map Based on Random Sub-Window Means
A Visual Saliency Map Based on Random Sub-Window Means Tadmeri Narayan Vikram 1,2, Marko Tscherepanow 1 and Britta Wrede 1,2 1 Applied Informatics Group 2 Research Institute for Cognition and Robotics
More informationSUPPLEMENTAL MATERIAL
1 SUPPLEMENTAL MATERIAL Response time and signal detection time distributions SM Fig. 1. Correct response time (thick solid green curve) and error response time densities (dashed red curve), averaged across
More informationLearning to Generate Long-term Future via Hierarchical Prediction. A. Motion-Based Pixel-Level Evaluation, Analysis, and Control Experiments
Appendix A. Motion-Based Pixel-Level Evaluation, Analysis, and Control Experiments In this section, we evaluate the predictions by deciles of motion similar to Villegas et al. (2017) using Peak Signal-to-Noise
More informationOutline. Teager Energy and Modulation Features for Speech Applications. Dept. of ECE Technical Univ. of Crete
Teager Energy and Modulation Features for Speech Applications Alexandros Summariza(on Potamianos and Emo(on Tracking in Movies Dept. of ECE Technical Univ. of Crete Alexandros Potamianos, NatIONAL Tech.
More informationReading Assignments: Lecture 18: Visual Pre-Processing. Chapters TMB Brain Theory and Artificial Intelligence
Brain Theory and Artificial Intelligence Lecture 18: Visual Pre-Processing. Reading Assignments: Chapters TMB2 3.3. 1 Low-Level Processing Remember: Vision as a change in representation. At the low-level,
More informationA Model for Automatic Diagnostic of Road Signs Saliency
A Model for Automatic Diagnostic of Road Signs Saliency Ludovic Simon (1), Jean-Philippe Tarel (2), Roland Brémond (2) (1) Researcher-Engineer DREIF-CETE Ile-de-France, Dept. Mobility 12 rue Teisserenc
More information(Visual) Attention. October 3, PSY Visual Attention 1
(Visual) Attention Perception and awareness of a visual object seems to involve attending to the object. Do we have to attend to an object to perceive it? Some tasks seem to proceed with little or no attention
More informationMEMORABILITY OF NATURAL SCENES: THE ROLE OF ATTENTION
MEMORABILITY OF NATURAL SCENES: THE ROLE OF ATTENTION Matei Mancas University of Mons - UMONS, Belgium NumediArt Institute, 31, Bd. Dolez, Mons matei.mancas@umons.ac.be Olivier Le Meur University of Rennes
More informationActions in the Eye: Dynamic Gaze Datasets and Learnt Saliency Models for Visual Recognition
Actions in the Eye: Dynamic Gaze Datasets and Learnt Saliency Models for Visual Recognition Stefan Mathe, Cristian Sminchisescu Presented by Mit Shah Motivation Current Computer Vision Annotations subjectively
More informationarxiv: v1 [cs.lg] 4 Feb 2019
Machine Learning for Seizure Type Classification: Setting the benchmark Subhrajit Roy [000 0002 6072 5500], Umar Asif [0000 0001 5209 7084], Jianbin Tang [0000 0001 5440 0796], and Stefan Harrer [0000
More informationMammogram Analysis: Tumor Classification
Mammogram Analysis: Tumor Classification Term Project Report Geethapriya Raghavan geeragh@mail.utexas.edu EE 381K - Multidimensional Digital Signal Processing Spring 2005 Abstract Breast cancer is the
More informationCan Saliency Map Models Predict Human Egocentric Visual Attention?
Can Saliency Map Models Predict Human Egocentric Visual Attention? Kentaro Yamada 1, Yusuke Sugano 1, Takahiro Okabe 1 Yoichi Sato 1, Akihiro Sugimoto 2, and Kazuo Hiraki 3 1 The University of Tokyo, Tokyo,
More informationNMF-Density: NMF-Based Breast Density Classifier
NMF-Density: NMF-Based Breast Density Classifier Lahouari Ghouti and Abdullah H. Owaidh King Fahd University of Petroleum and Minerals - Department of Information and Computer Science. KFUPM Box 1128.
More informationComparative Evaluation of Color Differences between Color Palettes
2018, Society for Imaging Science and Technology Comparative Evaluation of Color Differences between Color Palettes Qianqian Pan a, Stephen Westland a a School of Design, University of Leeds, Leeds, West
More informationIMAGE QUALITY ASSESSMENT BY USING AR PREDICTION ALGORITHM WITH INTERNAL GENERATIVE MECHANISM
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 8, August 2014,
More informationComputational modeling of visual attention and saliency in the Smart Playroom
Computational modeling of visual attention and saliency in the Smart Playroom Andrew Jones Department of Computer Science, Brown University Abstract The two canonical modes of human visual attention bottomup
More informationComparative Study of K-means, Gaussian Mixture Model, Fuzzy C-means algorithms for Brain Tumor Segmentation
Comparative Study of K-means, Gaussian Mixture Model, Fuzzy C-means algorithms for Brain Tumor Segmentation U. Baid 1, S. Talbar 2 and S. Talbar 1 1 Department of E&TC Engineering, Shri Guru Gobind Singhji
More informationIdentification of Tissue Independent Cancer Driver Genes
Identification of Tissue Independent Cancer Driver Genes Alexandros Manolakos, Idoia Ochoa, Kartik Venkat Supervisor: Olivier Gevaert Abstract Identification of genomic patterns in tumors is an important
More informationViewpoint Dependence in Human Spatial Memory
From: AAAI Technical Report SS-96-03. Compilation copyright 1996, AAAI (www.aaai.org). All rights reserved. Viewpoint Dependence in Human Spatial Memory Timothy P. McNamara Vaibhav A. Diwadkar Department
More informationRecurrent Refinement for Visual Saliency Estimation in Surveillance Scenarios
2012 Ninth Conference on Computer and Robot Vision Recurrent Refinement for Visual Saliency Estimation in Surveillance Scenarios Neil D. B. Bruce*, Xun Shi*, and John K. Tsotsos Department of Computer
More informationMULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES OBJECTIVES
24 MULTIPLE LINEAR REGRESSION 24.1 INTRODUCTION AND OBJECTIVES In the previous chapter, simple linear regression was used when you have one independent variable and one dependent variable. This chapter
More informationReveal Relationships in Categorical Data
SPSS Categories 15.0 Specifications Reveal Relationships in Categorical Data Unleash the full potential of your data through perceptual mapping, optimal scaling, preference scaling, and dimension reduction
More informationOverview of the visual cortex. Ventral pathway. Overview of the visual cortex
Overview of the visual cortex Two streams: Ventral What : V1,V2, V4, IT, form recognition and object representation Dorsal Where : V1,V2, MT, MST, LIP, VIP, 7a: motion, location, control of eyes and arms
More informationBootstrapped Integrative Hypothesis Test, COPD-Lung Cancer Differentiation, and Joint mirnas Biomarkers
Bootstrapped Integrative Hypothesis Test, COPD-Lung Cancer Differentiation, and Joint mirnas Biomarkers Kai-Ming Jiang 1,2, Bao-Liang Lu 1,2, and Lei Xu 1,2,3(&) 1 Department of Computer Science and Engineering,
More informationCHAPTER 6 HUMAN BEHAVIOR UNDERSTANDING MODEL
127 CHAPTER 6 HUMAN BEHAVIOR UNDERSTANDING MODEL 6.1 INTRODUCTION Analyzing the human behavior in video sequences is an active field of research for the past few years. The vital applications of this field
More informationG5)H/C8-)72)78)2I-,8/52& ()*+,-./,-0))12-345)6/3/782 9:-8;<;4.= J-3/ J-3/ "#&' "#% "#"% "#%$
# G5)H/C8-)72)78)2I-,8/52& #% #$ # # &# G5)H/C8-)72)78)2I-,8/52' @5/AB/7CD J-3/ /,?8-6/2@5/AB/7CD #&' #% #$ # # '#E ()*+,-./,-0))12-345)6/3/782 9:-8;;4. @5/AB/7CD J-3/ #' /,?8-6/2@5/AB/7CD #&F #&' #% #$
More informationMeasuring Focused Attention Using Fixation Inner-Density
Measuring Focused Attention Using Fixation Inner-Density Wen Liu, Mina Shojaeizadeh, Soussan Djamasbi, Andrew C. Trapp User Experience & Decision Making Research Laboratory, Worcester Polytechnic Institute
More informationPerformance and Saliency Analysis of Data from the Anomaly Detection Task Study
Performance and Saliency Analysis of Data from the Anomaly Detection Task Study Adrienne Raglin 1 and Andre Harrison 2 1 U.S. Army Research Laboratory, Adelphi, MD. 20783, USA {adrienne.j.raglin.civ, andre.v.harrison2.civ}@mail.mil
More informationTwo Themes. MobileASL: Making Cell Phones Accessible to the Deaf Community. Our goal: Challenges: Current Technology for Deaf People (text) ASL
Two Themes MobileASL: Making Cell Phones Accessible to the Deaf Community MobileASL AccessComputing Alliance Advancing Deaf and Hard of Hearing in Computing Richard Ladner University of Washington ASL
More informationComparison of Two Approaches for Direct Food Calorie Estimation
Comparison of Two Approaches for Direct Food Calorie Estimation Takumi Ege and Keiji Yanai Department of Informatics, The University of Electro-Communications, Tokyo 1-5-1 Chofugaoka, Chofu-shi, Tokyo
More informationGray Scale Image Edge Detection and Reconstruction Using Stationary Wavelet Transform In High Density Noise Values
Gray Scale Image Edge Detection and Reconstruction Using Stationary Wavelet Transform In High Density Noise Values N.Naveen Kumar 1, J.Kishore Kumar 2, A.Mallikarjuna 3, S.Ramakrishna 4 123 Research Scholar,
More informationSalient Object Detection in Videos Based on SPATIO-Temporal Saliency Maps and Colour Features
Salient Object Detection in Videos Based on SPATIO-Temporal Saliency Maps and Colour Features U.Swamy Kumar PG Scholar Department of ECE, K.S.R.M College of Engineering (Autonomous), Kadapa. ABSTRACT Salient
More informationFEATURE EXTRACTION USING GAZE OF PARTICIPANTS FOR CLASSIFYING GENDER OF PEDESTRIANS IN IMAGES
FEATURE EXTRACTION USING GAZE OF PARTICIPANTS FOR CLASSIFYING GENDER OF PEDESTRIANS IN IMAGES Riku Matsumoto, Hiroki Yoshimura, Masashi Nishiyama, and Yoshio Iwai Department of Information and Electronics,
More informationChapter 1. Introduction
Chapter 1 Introduction 1.1 Motivation and Goals The increasing availability and decreasing cost of high-throughput (HT) technologies coupled with the availability of computational tools and data form a
More informationThe Attraction of Visual Attention to Texts in Real-World Scenes
The Attraction of Visual Attention to Texts in Real-World Scenes Hsueh-Cheng Wang (hchengwang@gmail.com) Marc Pomplun (marc@cs.umb.edu) Department of Computer Science, University of Massachusetts at Boston,
More informationSelection and Combination of Markers for Prediction
Selection and Combination of Markers for Prediction NACC Data and Methods Meeting September, 2010 Baojiang Chen, PhD Sarah Monsell, MS Xiao-Hua Andrew Zhou, PhD Overview 1. Research motivation 2. Describe
More informationSaliency aggregation: Does unity make strength?
Saliency aggregation: Does unity make strength? Olivier Le Meur a and Zhi Liu a,b a IRISA, University of Rennes 1, FRANCE b School of Communication and Information Engineering, Shanghai University, CHINA
More informationGroup-Wise FMRI Activation Detection on Corresponding Cortical Landmarks
Group-Wise FMRI Activation Detection on Corresponding Cortical Landmarks Jinglei Lv 1,2, Dajiang Zhu 2, Xintao Hu 1, Xin Zhang 1,2, Tuo Zhang 1,2, Junwei Han 1, Lei Guo 1,2, and Tianming Liu 2 1 School
More informationBIOSTATISTICAL METHODS AND RESEARCH DESIGNS. Xihong Lin Department of Biostatistics, University of Michigan, Ann Arbor, MI, USA
BIOSTATISTICAL METHODS AND RESEARCH DESIGNS Xihong Lin Department of Biostatistics, University of Michigan, Ann Arbor, MI, USA Keywords: Case-control study, Cohort study, Cross-Sectional Study, Generalized
More informationGene expression analysis. Roadmap. Microarray technology: how it work Applications: what can we do with it Preprocessing: Classification Clustering
Gene expression analysis Roadmap Microarray technology: how it work Applications: what can we do with it Preprocessing: Image processing Data normalization Classification Clustering Biclustering 1 Gene
More informationValence-arousal evaluation using physiological signals in an emotion recall paradigm. CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry.
Proceedings Chapter Valence-arousal evaluation using physiological signals in an emotion recall paradigm CHANEL, Guillaume, ANSARI ASL, Karim, PUN, Thierry Abstract The work presented in this paper aims
More informationCancer Cells Detection using OTSU Threshold Algorithm
Cancer Cells Detection using OTSU Threshold Algorithm Nalluri Sunny 1 Velagapudi Ramakrishna Siddhartha Engineering College Mithinti Srikanth 2 Velagapudi Ramakrishna Siddhartha Engineering College Kodali
More informationNeuromorphic convolutional recurrent neural network for road safety or safety near the road
Neuromorphic convolutional recurrent neural network for road safety or safety near the road WOO-SUP HAN 1, IL SONG HAN 2 1 ODIGA, London, U.K. 2 Korea Advanced Institute of Science and Technology, Daejeon,
More informationApplied Machine Learning in Biomedicine. Enrico Grisan
Applied Machine Learning in Biomedicine Enrico Grisan enrico.grisan@dei.unipd.it Algorithm s objective cost Formal objective for algorithms: - minimize a cost function - maximize an objective function
More informationIntelligent Edge Detector Based on Multiple Edge Maps. M. Qasim, W.L. Woon, Z. Aung. Technical Report DNA # May 2012
Intelligent Edge Detector Based on Multiple Edge Maps M. Qasim, W.L. Woon, Z. Aung Technical Report DNA #2012-10 May 2012 Data & Network Analytics Research Group (DNA) Computing and Information Science
More informationAssigning B cell Maturity in Pediatric Leukemia Gabi Fragiadakis 1, Jamie Irvine 2 1 Microbiology and Immunology, 2 Computer Science
Assigning B cell Maturity in Pediatric Leukemia Gabi Fragiadakis 1, Jamie Irvine 2 1 Microbiology and Immunology, 2 Computer Science Abstract One method for analyzing pediatric B cell leukemia is to categorize
More informationLocal Image Structures and Optic Flow Estimation
Local Image Structures and Optic Flow Estimation Sinan KALKAN 1, Dirk Calow 2, Florentin Wörgötter 1, Markus Lappe 2 and Norbert Krüger 3 1 Computational Neuroscience, Uni. of Stirling, Scotland; {sinan,worgott}@cn.stir.ac.uk
More informationBangor University Laboratory Exercise 1, June 2008
Laboratory Exercise, June 2008 Classroom Exercise A forest land owner measures the outside bark diameters at.30 m above ground (called diameter at breast height or dbh) and total tree height from ground
More informationQuantifying the Effect of Disruptions to Temporal Coherence on the Intelligibility of Compressed American Sign Language Video
Quantifying the Effect of Disruptions to Temporal Coherence on the ntelligibility of Compressed American Sign Language Video Frank M. Ciaramello and Sheila S. Hemami Visual Communication Laboratory School
More informationAutomated Assessment of Diabetic Retinal Image Quality Based on Blood Vessel Detection
Y.-H. Wen, A. Bainbridge-Smith, A. B. Morris, Automated Assessment of Diabetic Retinal Image Quality Based on Blood Vessel Detection, Proceedings of Image and Vision Computing New Zealand 2007, pp. 132
More informationJitter-aware time-frequency resource allocation and packing algorithm
Jitter-aware time-frequency resource allocation and packing algorithm The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As
More informationIntroduction to Machine Learning. Katherine Heller Deep Learning Summer School 2018
Introduction to Machine Learning Katherine Heller Deep Learning Summer School 2018 Outline Kinds of machine learning Linear regression Regularization Bayesian methods Logistic Regression Why we do this
More informationComputational Models of Visual Attention: Bottom-Up and Top-Down. By: Soheil Borhani
Computational Models of Visual Attention: Bottom-Up and Top-Down By: Soheil Borhani Neural Mechanisms for Visual Attention 1. Visual information enter the primary visual cortex via lateral geniculate nucleus
More informationExperiences on Attention Direction through Manipulation of Salient Features
Experiences on Attention Direction through Manipulation of Salient Features Erick Mendez Graz University of Technology Dieter Schmalstieg Graz University of Technology Steven Feiner Columbia University
More informationLearning Spatiotemporal Gaps between Where We Look and What We Focus on
Express Paper Learning Spatiotemporal Gaps between Where We Look and What We Focus on Ryo Yonetani 1,a) Hiroaki Kawashima 1,b) Takashi Matsuyama 1,c) Received: March 11, 2013, Accepted: April 24, 2013,
More informationInternational Journal of Computational Science, Mathematics and Engineering Volume2, Issue6, June 2015 ISSN(online): Copyright-IJCSME
Various Edge Detection Methods In Image Processing Using Matlab K. Narayana Reddy 1, G. Nagalakshmi 2 12 Department of Computer Science and Engineering 1 M.Tech Student, SISTK, Puttur 2 HOD of CSE Department,
More informationDiscovering Meaningful Cut-points to Predict High HbA1c Variation
Proceedings of the 7th INFORMS Workshop on Data Mining and Health Informatics (DM-HI 202) H. Yang, D. Zeng, O. E. Kundakcioglu, eds. Discovering Meaningful Cut-points to Predict High HbAc Variation Si-Chi
More informationEstimating Multiple Evoked Emotions from Videos
Estimating Multiple Evoked Emotions from Videos Wonhee Choe (wonheechoe@gmail.com) Cognitive Science Program, Seoul National University, Seoul 151-744, Republic of Korea Digital Media & Communication (DMC)
More informationPerceptual-Based Objective Picture Quality Measurements
Perceptual-Based Objective Picture Quality Measurements Introduction In video systems, a wide range of video processing devices can affect overall picture quality. Encoders and decoders compress and decompress
More informationNotes for laboratory session 2
Notes for laboratory session 2 Preliminaries Consider the ordinary least-squares (OLS) regression of alcohol (alcohol) and plasma retinol (retplasm). We do this with STATA as follows:. reg retplasm alcohol
More informationIAT 355 Visual Analytics. Encoding Information: Design. Lyn Bartram
IAT 355 Visual Analytics Encoding Information: Design Lyn Bartram 4 stages of visualization design 2 Recall: Data Abstraction Tables Data item (row) with attributes (columns) : row=key, cells = values
More informationUNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Midterm, 2016
UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Midterm, 2016 Exam policy: This exam allows one one-page, two-sided cheat sheet; No other materials. Time: 80 minutes. Be sure to write your name and
More informationFusing Generic Objectness and Visual Saliency for Salient Object Detection
Fusing Generic Objectness and Visual Saliency for Salient Object Detection Yasin KAVAK 06/12/2012 Citation 1: Salient Object Detection: A Benchmark Fusing for Salient Object Detection INDEX (Related Work)
More informationSparse Coding in Sparse Winner Networks
Sparse Coding in Sparse Winner Networks Janusz A. Starzyk 1, Yinyin Liu 1, David Vogel 2 1 School of Electrical Engineering & Computer Science Ohio University, Athens, OH 45701 {starzyk, yliu}@bobcat.ent.ohiou.edu
More informationANALYSIS OF FACIAL FEATURES OF DRIVERS UNDER COGNITIVE AND VISUAL DISTRACTIONS
ANALYSIS OF FACIAL FEATURES OF DRIVERS UNDER COGNITIVE AND VISUAL DISTRACTIONS Nanxiang Li and Carlos Busso Multimodal Signal Processing (MSP) Laboratory Department of Electrical Engineering, The University
More informationEmpirical Analysis of Object-Oriented Design Metrics for Predicting High and Low Severity Faults
IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, VOL. 32, NO. 10, OCTOBER 2006 771 Empirical Analysis of Object-Oriented Design Metrics for Predicting High and Low Severity Faults Yuming Zhou and Hareton Leung,
More informationQuantitative Evaluation of Edge Detectors Using the Minimum Kernel Variance Criterion
Quantitative Evaluation of Edge Detectors Using the Minimum Kernel Variance Criterion Qiang Ji Department of Computer Science University of Nevada Robert M. Haralick Department of Electrical Engineering
More informationOn Shape And the Computability of Emotions X. Lu, et al.
On Shape And the Computability of Emotions X. Lu, et al. MICC Reading group 10.07.2013 1 On Shape and the Computability of Emotion X. Lu, P. Suryanarayan, R. B. Adams Jr., J. Li, M. G. Newman, J. Z. Wang
More informationObject-Level Saliency Detection Combining the Contrast and Spatial Compactness Hypothesis
Object-Level Saliency Detection Combining the Contrast and Spatial Compactness Hypothesis Chi Zhang 1, Weiqiang Wang 1, 2, and Xiaoqian Liu 1 1 School of Computer and Control Engineering, University of
More informationDevelopment of goal-directed gaze shift based on predictive learning
4th International Conference on Development and Learning and on Epigenetic Robotics October 13-16, 2014. Palazzo Ducale, Genoa, Italy WePP.1 Development of goal-directed gaze shift based on predictive
More informationObject-based Saliency as a Predictor of Attention in Visual Tasks
Object-based Saliency as a Predictor of Attention in Visual Tasks Michal Dziemianko (m.dziemianko@sms.ed.ac.uk) Alasdair Clarke (a.clarke@ed.ac.uk) Frank Keller (keller@inf.ed.ac.uk) Institute for Language,
More informationUsing Dynamic Time Warping for Intuitive Handwriting Recognition
Using Dynamic Time Warping for Intuitive Handwriting Recognition Ralph Niels and Louis Vuurpijl Nijmegen Institute for Cognition and Information (Radboud University Nijmegen) P.O. Box 9104, 6500 HE Nijmegen,
More informationFirefighter safety: How visible is your protective clothing?
News Article ID: 4028 Fire international 23 April 2003 Firefighter safety: How visible is your protective clothing? TNO Human Factors Research Institute has developed a method to determine visual conspicuity
More informationAge (continuous) Gender (0=Male, 1=Female) SES (1=Low, 2=Medium, 3=High) Prior Victimization (0= Not Victimized, 1=Victimized)
Criminal Justice Doctoral Comprehensive Exam Statistics August 2016 There are two questions on this exam. Be sure to answer both questions in the 3 and half hours to complete this exam. Read the instructions
More information