Impute vs. Ignore: Missing Values for Prediction

Similar documents
Statistical Analysis Using Machine Learning Approach for Multiple Imputation of Missing Data

Credal decision trees in noisy domains

An Improved Algorithm To Predict Recurrence Of Breast Cancer

Classification of breast cancer using Wrapper and Naïve Bayes algorithms

Predicting Breast Cancer Survivability Rates

A scored AUC Metric for Classifier Evaluation and Selection

Performance Analysis of Different Classification Methods in Data Mining for Diabetes Dataset Using WEKA Tool

Hybridized KNN and SVM for gene expression data classification

International Journal of Pharma and Bio Sciences A NOVEL SUBSET SELECTION FOR CLASSIFICATION OF DIABETES DATASET BY ITERATIVE METHODS ABSTRACT

Comparison of discrimination methods for the classification of tumors using gene expression data

Evaluating Classifiers for Disease Gene Discovery

Response to Mease and Wyner, Evidence Contrary to the Statistical View of Boosting, JMLR 9:1 26, 2008

Colon cancer survival prediction using ensemble data mining on SEER data

A Comparison of Collaborative Filtering Methods for Medication Reconciliation

Roadmap for Developing and Validating Therapeutically Relevant Genomic Classifiers. Richard Simon, J Clin Oncol 23:

Data complexity measures for analyzing the effect of SMOTE over microarrays

Predicting Breast Cancer Survival Using Treatment and Patient Factors

International Journal of Computer Science Trends and Technology (IJCST) Volume 5 Issue 1, Jan Feb 2017

A NOVEL VARIABLE SELECTION METHOD BASED ON FREQUENT PATTERN TREE FOR REAL-TIME TRAFFIC ACCIDENT RISK PREDICTION

Statistical data preparation: management of missing values and outliers

A Deep Learning Approach to Identify Diabetes

How to Create Better Performing Bayesian Networks: A Heuristic Approach for Variable Selection

Keywords Artificial Neural Networks (ANN), Echocardiogram, BPNN, RBFNN, Classification, survival Analysis.

Learning with Rare Cases and Small Disjuncts

Model reconnaissance: discretization, naive Bayes and maximum-entropy. Sanne de Roever/ spdrnl

Data Mining and Knowledge Discovery: Practice Notes

A STUDY OF AdaBoost WITH NAIVE BAYESIAN CLASSIFIERS: WEAKNESS AND IMPROVEMENT

Missing Data and Imputation

COMPARISON OF DECISION TREE METHODS FOR BREAST CANCER DIAGNOSIS

Using AUC and Accuracy in Evaluating Learning Algorithms

Module 14: Missing Data Concepts

Selected Topics in Biostatistics Seminar Series. Missing Data. Sponsored by: Center For Clinical Investigation and Cleveland CTSC

Variable Features Selection for Classification of Medical Data using SVM

Analysis of Classification Algorithms towards Breast Tissue Data Set

The Long Tail of Recommender Systems and How to Leverage It

A DATA MINING APPROACH FOR PRECISE DIAGNOSIS OF DENGUE FEVER

Nearest Shrunken Centroid as Feature Selection of Microarray Data

UvA-DARE (Digital Academic Repository)

S Imputation of Categorical Missing Data: A comparison of Multivariate Normal and. Multinomial Methods. Holmes Finch.

The Nottingham eprints service makes this work by researchers of the University of Nottingham available open access under the following conditions.

PREDICTION OF BREAST CANCER USING STACKING ENSEMBLE APPROACH

Scientific Journal of Informatics Vol. 3, No. 2, November p-issn e-issn

Downloaded from ijbd.ir at 19: on Friday March 22nd (Naive Bayes) (Logistic Regression) (Bayes Nets)

J2.6 Imputation of missing data with nonlinear relationships

Performance Based Evaluation of Various Machine Learning Classification Techniques for Chronic Kidney Disease Diagnosis

Keywords Missing values, Medoids, Partitioning Around Medoids, Auto Associative Neural Network classifier, Pima Indian Diabetes dataset.

Remarks on Bayesian Control Charts

Positive and Unlabeled Relational Classification through Label Frequency Estimation

INDUCTIVE LEARNING OF TREE-BASED REGRESSION MODELS. Luís Fernando Raínho Alves Torgo

Predicting Breast Cancer Recurrence Using Machine Learning Techniques

DEVELOPMENT OF AN EXPERT SYSTEM ALGORITHM FOR DIAGNOSING CARDIOVASCULAR DISEASE USING ROUGH SET THEORY IMPLEMENTED IN MATLAB

Wrapper subset evaluation facilitates the automated detection of diabetes from heart rate variability measures

Logistic Regression with Missing Data: A Comparison of Handling Methods, and Effects of Percent Missing Values

Catherine A. Welch 1*, Séverine Sabia 1,2, Eric Brunner 1, Mika Kivimäki 1 and Martin J. Shipley 1

Statistics 202: Data Mining. c Jonathan Taylor. Final review Based in part on slides from textbook, slides of Susan Holmes.

Propensity scores and causal inference using machine learning methods

CS229 Final Project Report. Predicting Epitopes for MHC Molecules

Positive and Unlabeled Relational Classification through Label Frequency Estimation

A COMPARISON OF IMPUTATION METHODS FOR MISSING DATA IN A MULTI-CENTER RANDOMIZED CLINICAL TRIAL: THE IMPACT STUDY

Selection and Combination of Markers for Prediction

3. Model evaluation & selection

Type II Fuzzy Possibilistic C-Mean Clustering

Recognition of sign language gestures using neural networks

Ineffectiveness of Use of Software Science Metrics as Predictors of Defects in Object Oriented Software

Generalized additive model for disease risk prediction

Lazy Learning of Bayesian Rules

COMMITTEE FOR PROPRIETARY MEDICINAL PRODUCTS (CPMP) POINTS TO CONSIDER ON MISSING DATA

Identifying Parkinson s Patients: A Functional Gradient Boosting Approach

Empirical function attribute construction in classification learning

Using Ensemble-Based Methods for Directly Estimating Causal Effects: An Investigation of Tree-Based G-Computation

arxiv: v1 [cs.lg] 4 Feb 2019

BayesRandomForest: An R

BLOOD GLUCOSE PREDICTION MODELS FOR PERSONALIZED DIABETES MANAGEMENT

EFFICIENT MULTIPLE HEART DISEASE DETECTION SYSTEM USING SELECTION AND COMBINATION TECHNIQUE IN CLASSIFIERS

Challenges of Automated Machine Learning on Causal Impact Analytics for Policy Evaluation

Wrapper subset evaluation facilitates the automated detection of diabetes from heart rate variability measures

Modeling Sentiment with Ridge Regression

Towards More Confident Recommendations: Improving Recommender Systems Using Filtering Approach Based on Rating Variance

Comparative Analysis of Machine Learning Algorithms for Chronic Kidney Disease Detection using Weka

Gene Selection for Tumor Classification Using Microarray Gene Expression Data

Diagnosis of Breast Cancer Using Ensemble of Data Mining Classification Methods

Combining Classifiers in Multimodal Affect Detection

MISSING DATA AND PARAMETERS ESTIMATES IN MULTIDIMENSIONAL ITEM RESPONSE MODELS. Federico Andreis, Pier Alda Ferrari *

Multi Parametric Approach Using Fuzzification On Heart Disease Analysis Upasana Juneja #1, Deepti #2 *

Mammogram Analysis: Tumor Classification

Fuzzy Decision Tree FID

AN EXPERIMENTAL STUDY ON HYPOTHYROID USING ROTATION FOREST

Data mining for Obstructive Sleep Apnea Detection. 18 October 2017 Konstantinos Nikolaidis

Machine Learning to Inform Breast Cancer Post-Recovery Surveillance

White Paper Estimating Complex Phenotype Prevalence Using Predictive Models

Predicting Juvenile Diabetes from Clinical Test Results

Automated Medical Diagnosis using K-Nearest Neighbor Classification

An Empirical and Formal Analysis of Decision Trees for Ranking

Validity and reliability of measurements

Article from. Forecasting and Futurism. Month Year July 2015 Issue Number 11

Data Mining Techniques to Predict Survival of Metastatic Breast Cancer Patients

Mining Big Data: Breast Cancer Prediction using DT - SVM Hybrid Model

4. Model evaluation & selection

What to do with missing data in clinical registry analysis?

Prediction of heart disease using k-nearest neighbor and particle swarm optimization.

Transcription:

Proceedings of International Joint Conference on Neural Networks, Dallas, Texas, USA, August 4-9, 2013 Impute vs. Ignore: Missing Values for Prediction Qianyu Zhang, Ashfaqur Rahman, and Claire D Este Abstract Sensor faults or communication errors can cause certain sensor readings to become unavailable for prediction purposes. In this paper we evaluate the performance of imputation techniques and techniques that ignore the missing values, in scenarios: (i) when values are missing only during prediction phase, and (ii) when values are missing during both the induction and prediction phase. We also investigated the influence of different scales of missingness on the performance of these treatments. The results can be used as a guideline to facilitate the choice of different missing value treatments under different circumstances. I. INTRODUCTION Event detection from multiple environmental sensor data streams can add value to real-time decision support systems for industries such as agriculture and aquaculture [1]. However, sensors placed in the field commonly experience faults and communication errors. This can cause missing values in the data that we require both for training our classifiers and for gaining a classification from them. Frequently failing to provide a classification will result in a reduction of trust in the overall decision support system. This trust will also be based on the accuracy of the classifications, which means we do not want to throw away training examples if we can avoid it. Therefore we require a means for training and providing a classification even when some of the features are not available. Missing value treatments in machine learning are a solution to this problem. There are two alternative approaches to deal with missing values: Impute and Ignore. Imputation treatments [2][3] fill in attributes in the instance vector using statistical techniques and the complete vector is fed to the predictor. Ignore treatments (also called reduced model or ensemble classifier treatments) [4][5] overlook the missing attributes, produce a vector based on the available attributes, and feed that vector to a predictor trained on those particular attributes. In current literature, imputation treatments have been widely investigated [6][7], but ignore treatments draws less attention. Ignore and impute treatments have been previously compared on data sets that are naturally occurring with missing values [5][8]. There are also studies that compare the performance of various missing value treatments under various scales of missingness [8][9][10][11]. Farhangfar [9] compared imputation methods under different scales of missingness. Pelckmans [10] studied supported vector machine Qianyu Zhang, Ashfaqur Rahman, and Claire D Este are with the Intelligent Sensing and Systems Laboratory, CSIRO, Castray Esplanade, Hobart, Australia (email: {Qianyu.Zhang, Ashfaqur.Rahman, Claire.D Este}@csiro.au). The research presented in this paper was supported by ICT division and Food Future Flagship in CSIRO. classifiers tested on a modified version of the Ripley dataset. Melville [11] provides a comparison between the bagging [12], boosting [13] and DECORATE [14] methods. Concluding that DECORATE outperforms the others when features go missing, where DECORATE is an ensemble method that contains classifiers trained on both the original data and some artificial data. Garcia [8] compared most popular missing value treatments, using natural dataset with missing values and using dataset with various scales of missingness introduced to its most relevant feature. We observe that the above studies are all restricted to certain circumstances and more or less focuses on methods involving artificial data. All such studies uses data with missing values into both the training and testing set, some only introduces missing values to a single feature in the data. Motivated by the above fact the research presented in this paper aims to compare the performance of the impute and ignore treatments under a broader and more practical set of situations, in order to provide true comparisons for potential applications. In detail, this paper investigates the performance of the impute and ignore treatments under different scales of missingness, (i) when values are missing only during prediction phase, and (ii) when values are missing during both the induction and prediction phase. For each instance, all features have an equivalent chance of being set missing. Moreover, the largest missing percentage considered in previous studies was 40%, we extended this to 50% in order to ensure the overall trend and to further support our results. We have artificially introduced missingness at different scales on a number of benchmark data sets from the UCI machine learning repository [15] and analysed the performance of these treatments under different scenarios using the decision tree [16], nearest neighbour [17], and Bayesian network [18] classifiers. With different treatments and controlled introduction of missingness, the contribution in this paper can be considered as providing guidelines to facilitate the choice of missing value treatments under different circumstances. The paper is organized as follows: some commonly used missing value treatments are presented in Section II. The experimental set-up and methodology is presented in Section III. Results are analysed in Section IV. Section V concludes the paper. II. MISSING VALUE TREATMENTS The most commonly used imputation treatments are mean, nearest neighbour, and regression [5]. The reduced model/ensemble classifier is the potential ignore treatment. The statistical approaches relies on the assumption that values are missing completely at random (MCAR) [19]. MCAR refers to the case where the probability of the value 978-1-4673-6129-3/13/$31.00 2013 IEEE 2193

TABLE I EXAMPLE USED TO DEMONSTRATE DIFFERENT MISSING VALUE TREATMENTS A B Class 1 2 0? 6 1 4 8 1 of an attribute going missing is independent of both observed and unobserved attribute values. Each of these treatments are detailed with the example presented in Table I: 1) Impute: Value imputation [2][20] replaces missing values with values estimated using existing data, this is most commonly used in the statistics community. Mean imputation treatment replaces a missing entry with the statistical mean of the available values of corresponding attribute. In the example presented in Table I, we can replace all unknown values in attribute A by the mean of the known values in A, which is (1+4)/2=2.5. Nearest Neighbour imputation treatment replaces missing values of an instance using the values of its nearest neighbour. The similarity between instances is determined using Euclidean distance, which is calculated based on the values of the available attributes. In the example presented in Table I, attribute A is missing in the middle instance. In terms of the available attribute B, the difference between the first and the middle instance is 4, whereas the difference is only 2 between the middle and the third instance. So the nearest neighbour is the third instance and the nearest neighbour imputation treatment will replace the missing value with 4. Regression imputation treatment derives a relationship between the missing attribute and other available attributes based on the instances in the training data with the same missing pattern. More precisely, the attribute under consideration is expressed as a linear combination a 1 x 1 + a 2 x 2 +... + a n x n where x 1, x 2,..., x n are the available attributes and the parameters a 1, a 2,..., a n are derived from the available instances. In the example presented in Table I, it is clear from instance one and three that A is equal to half of B. Applying this function, regression imputation estimates A to be 0.5 6 = 3. 2) Ignore: These treatments ignores the missing values instead of replacing them with artificial values. Case deletion treatment simply removes any training example for which we do not have a full set of attributes. This treatment risks losing important information from the remainder of the example at the training stage. It is also not possible to use this technique at the prediction stage. Therefore we do not present results in the paper for this treatment. Ensemble classifier treatment alternatively trains specific classifiers for each possible missing pattern that might be present. During the training/induction phase, we first partition the data into different subsets according to their missing patterns. Each subset of data is then used to train a classifier, together the group of classifiers form the ensemble classifier. Each test instance is classified using the classifier trained on the group of training data with the same missing pattern. The procedure is illustrated in Fig. 1. Fig. 1. Example of ensemble classifier treatment The imputation treatments introduce artificial values leading to estimation error. This will be propagated or even exaggerated during classification. Whereas ignore treatments using the ensemble classifier eliminates this estimation error. However, one limitation is that this treatment requires larger memory for the storage of the multiple classifiers. An alternative could be to train the classifier for a particular missing pattern only when a test instance with such missing pattern is encountered. III. EXPERIMENTAL SETUP In this section we present the data sets used in the experiments, the process of introducing missing values, and two different experiment models: (i) when values are missing during prediction only and (ii) when values are missing during both induction and prediction. A. Data Sets We have used nine benchmark datasets from UCI machine learning repository [15], Table II provides a brief summary to the data sets. All the attributes in these data sets are numeric. 2194

TABLE II SUMMARY OF DATA SETS Data Set Instances Attributes Nominal Attributes Breast Cancer 683 9 2 Diabetes 768 8 2 Ecoli 336 7 8 Glass 218 10 6 Haberman 306 3 2 Iris 150 4 3 Liver 345 6 2 Thyroid 220 5 3 Transfusion 748 4 2 TABLE III SUMMARY OF CLASSIFIER CONFIGURATIONS Classifier Parameter setting Decision Tree Confidence factor: 0.25 Minimum # of instances per leaf: 2 Number of folds used for pruning: 3 Bayesian Network Estimator algorithm: simple estimator alpha(initial count): 0.5 B. Introduction of Missing Values We artificially introduced missing values into the data sets for the purpose of these experiments. We assume that all attributes are important for class prediction and hence do not perform a attribute selection stage. A spectrum of 0 to 50% missingness is considered. Zero missingness is when the full data set is being used. The performance of the classifiers on the full data set can be used as a benchmark for the use of various treatments and classifiers as missingness is introduced. When x% missingness is introduced, each attribute value has x% chance of being set missing, regardless of its value or position. More precisely x% attribute values are removed randomly. Note this way of introducing missingness results in a data set with no missing values in the class attribute.the resulting data set is also in accordance with the MCAR principle i.e. the probability of a value goes missing is independent of both observed and unobserved attribute values. C. Classifiers We investigated the various treatments with a range of classifiers; namely the decision tree classifier, the nearest neighbour classifier, and the Bayesian network classifier. We employed the classifier implementation in WEKA [21], with parameters shown in Table III. D. Experiment Procedure: Values Missing in Test Set Only 1) Randomize and stratify the data. 2) Split the data set into a training set containing the first 90% of instances, and the rest 10% forms the test set. 3) Introduce missing values into the test set 4) Train and test the classifiers a) For the basic classifiers: i) Train the classifiers using the training data. ii) Impute the missing values in the test data. Fig. 2. Experimental model with missing values in test set only iii) Test the classifiers using the imputed test data. iv) Record the resulting classification accuracy. b) For the ensemble classifier treatment: i) Enumerate all possible missing patterns. ii) Sort the training data into groups according to the missing patterns. Note that since there are no missing values in this situation, all instances should be included in all the groups, with the corresponding attributes deleted. iii) Train a sub-classifier for each group of training data. iv) Sort the test data into groups according to the missing patterns. v) Test each group of test data using the subclassifier trained with the training data having the same missing pattern. vi) Record the resulting classification accuracy, averaged across all sub-classifiers used. 5) (10-fold cross validation) Repeat the above procedure for 10 times, each time choose a different subset of test data. 6) Repeat the whole procedure for 10 times, record the average accuracy and standard deviation. 2195

Fig. 3. set Experimental model with missing values in both training and test iii) Train a sub-classifier for each group of training data. iv) Sort the test data into groups according to the missing patterns. v) Test each group of test data using the subclassifier trained with the training data having the same missing pattern. vi) Record the resulting classification accuracy, averaged across all sub-classifiers used. 5) (10-fold cross validation) Repeat the above procedure for 10 times, each time choose a different subset of test data. 6) Repeat the whole procedure for 10 times, record the average accuracy and standard deviation. The above procedure is illustrated in Fig. 3. Those two scenarios are experimented with to simulate practical situations i) when the recorded historical data are full, and the test data is collected through sensors and contain missing values, and ii) when the training data and test data are collected using the same type of sensors with the same failing rate. IV. RESULTS AND DISCUSSIONS In this section we discuss the findings on the performance of different missing value treatments using the above experimental platform. Our measure of performance is the average classification accuracy of 10 rounds of random introduction of missingness. A. Values Missing at Prediction Time Only The above procedure is illustrated in Fig. 2. Note that the imputation step does not apply to the ensemble classifier treatment. E. Experiment Procedure: Values Missing in both Training and Test Set 1) Randomize and stratify the data. 2) Split the data set into a training set containing the first 90% of instances, and the rest 10% forms the test set. 3) Introduce missing values into both the training set and the test set 4) Train and test the classifiers a) For the basic classifiers: i) Impute the missing values in the training data. ii) Train the classifiers using imputed training data. iii) Impute the missing values in the test data. iv) Test the classifiers using the imputed test data. v) Record the resulting classification accuracy. b) For the ensemble classifier treatment: i) Enumerate all possible missing patterns. ii) Sort the training data into groups according to the missing patterns. Fig. 4. Average accuracy of treatments as the percentage of missing values varies in the test set In this experiment values are missing during prediction only. Fig. 4 presents the accuracy of each treatment as the percentage of missing values changes. The accuracy reported at each missing percentage is averaged over all datasets and all classifiers. It is evident from the plot that no matter which treatment is used, the existence of missing values leads 2196

to reduced accuracy. As the percentage of missing values increases, the performance reduction is higher. It can also be observed from the plot that the ensemble classifier treatment remains on top as the best performer. To support the fact that the ensemble classifier treatment is the best performer in this scenario, we have obtained the number of times each treatment is the best performer in terms of absolute accuracy across different missing percentages using different classifiers and data sets. The dominance results are presented in Table IV. Each row in Table IV presents the number of times each missing value treatments are the best performers for a particular scale of missingness. There are three classifiers and nine datasets and thus the dominance values will sum to 3 9 = 27 for each row. Note that ensemble classifier treatment remains the most reliable at each scale of missingness. We hypothesise that the ensemble classifier treatment is the best performer in this scenario, because each sub-classifier is trained with a training subset, containing all training instances with the missing attributes removed. When a test instance with missing values is presented. Imputation treatments will introduce artificial values into the test instance, this approximation error will then propagate to the prediction phase. Whereas the ensemble classifier treatment deals directly with the true observations and does not introduce this approximation error. Fig. 5 presents the plots comparing the performance of the treatments when used with a particular classifier. These plots reveal that the accuracy trend line of the ensemble classifier treatment is always convex. This is an interesting phenomenon comparing to the decrease in accuracy of the imputation treatments, which are more or less straight lines. This phenomenon suggests that the decrease of the performance of the ensemble classifier treatment is becoming faster as the missing percentage increases, and that it suffers more severely than the imputation treatments at very large percentage of missing values. Therefore ensemble classifier treatment works best when no values are missing in training data, and when the missingness is not too large. But which classifier yields the best performance when used as a base classifier for the ensemble classifier treatment? Fig. 6 reveals the fact that under all scales of missingness, the decision tree is the best choice. It is also clear that the performance of each classifier has an explicit ordering, which doesn t hold for the imputation treatments, as shown in Fig. 7. The reason is that the instances passed to the base-classifiers in the ensemble classifier treatment has not been modified and all unknown values are ignored. Hence if the performance of a classifier is better than another under some missing percentage, it will always be the case. On the other hand, the imputation treatments introduce artificial values that brings along errors. Depending on the missing percentage and the imputation treatment used, the bias of the artificial data leads to varying relative performance of the classifiers. Fig. 7 also suggests that the decision tree classifier per- (a) Decision tree classifier (b) Nearest neighbour classifier (c) Bayesian network classifier Fig. 5. The performance of each treatment using various classifiers. Values are missing in test set only. forms best when the missingness is very small (less than 10%), but its performance decreases sharply as the missing percentage increases. It is clear that the Bayesian network classifier works the best with a large missing percentage. As the missing percentage gets large, the value of top level attributes used for branching decisions are more likely to be estimated. This might lead to completely different branches, yielding the wrong prediction. However the Bayesian network does not suffer as much because it calculates the probability of having a particular instance taking into account all attribute values, the estimation error is less likely to affect 2197

TABLE IV SUPERIORITY SUMMARY OF THE MISSING VALUE TREATMENTS WHEN VALUES ARE MISSING IN THE TEST SET ONLY Treatment Missing percentage (%) Mean Nearest Neighbour Regression Ensemble 10 4 0 8 15 20 2 0 3 22 30 2 0 5 20 40 2 1 5 19 50 2 1 9 15 (a) Mean treatment Fig. 6. Performance of various classifiers when used as base classifiers for the ensemble classifier treatment. Values are missing in test set only. the final result as much. B. Values Missing at Both Induction and Prediction Time In this section we present the performance of different missing value treatments when values are missing during both the training and test phase. Fig. 8 shows the accuracy of each treatment across different missing percentage, averaged over all datasets and all classifiers. From the plot it is again evident that no matter which treatment is used, the increase in the percentage of missing values leads to greater performance reductions. It is interesting to observe that the ensemble classifier treatment is performing the worst in this scenario. This is due to the fact that when there are missing values in the training data, the size of many training subsets will reduce. This leads to insufficient training samples for many subclassifiers to generalise. Thus the presence of missing values in the training data reduces the overall accuracy significantly. Fig. 9 presents the performance of the missing value treatments using different classifiers. When the decision tree classifier is used, the ensemble classifier treatment performs the worst, in accordance with our previous findings. What is more clear is that regression tends to perform the best when the percentage of missingness is small (less than (b) Nearest neighbour treatment (c) Regression treatment Fig. 7. Performance of various classifiers with different missing value treatments. Values are missing in test set only 2198

(a) Decision tree classifier Fig. 8. Average accuracy of treatments as the percentage of missing values changes in both training and test set 20 30%). As the missingness increases, the mean imputation performs the best. This is also confirmed by Table V, which presents the dominance summary. This can be explained by considering how the imputation treatments work. Both the nearest neighbour imputation and regression imputation rely on other available attributes in estimating the missing values. As the missing percentage increases, there are less available attributes for each instance. Thus the probability of the most relevant attributes also goes missing is higher. So the estimated value is likely to be more biased than the value estimated using mean imputation, which considers only values of the same attribute across all instances. Looking at Fig. 9(b), we found that the nearest neighbour imputation performs more poorly than the ensemble classifier treatment. The reason is that the nearest neighbour imputation assigns the missing entries the values of its nearest neighbour. Then the nearest neighbour classifier will assign it to the same class as its nearest neighbour, which is the same instance used for imputing the missing values. Therefore this combination is just assigning the test instance the class of the instance having the most similar values for its available attributes. This is much more naive and hence has less predictive power comparing to any other treatments. The plot in Fig. 9(c) shows the performance comparison when using the Bayesian network classifier. The plot shows that even for small missing percentages, the mean imputation performs the best. The reason being that the regression imputation estimates the missing value using the relationship between the missing attribute and other available attributes. Similarly the Bayesian network classifier considers all the attribute values of an instance and the joint probability of them occurring together. Since this classifier already takes into account the relationships between the attributes, the value estimated using regression imputation will not help much in terms of classification accuracy. (b) Nearest neighbour classifier (c) Bayesian network classifier Fig. 9. The performance of each treatment in conjunction with various classifiers. Values are missing in both training and test set. V. CONCLUSIONS In this paper we have done an empirical study that compares the performance of various missing value treatments when (i) values are missing only during prediction phase, and (ii) values are missing during both induction and prediction phase, under different scales of missingness. Based on the findings we can reach the following conclusions: (i) The presence of missing values always leads to performance reductions of classifier, no matter which treatment we use to deal with the missing values. (ii) When there are no values missing in the training 2199

TABLE V DOMINANCE SUMMARY WHEN VALUES ARE MISSING BOTH IN THE TRAINING AND TEST SET Treatment Missing percentage (%) Mean Nearest Neighbour Regression Ensemble 10 6 3 11 7 20 9 3 11 4 30 13 2 7 5 40 13 5 6 3 50 17 3 5 2 data, ensemble classifier treatment is the best choice. And when using the ensemble classifier treatment in such circumstances, using decision tree as the base classifier yields the best performance. Average accuracy as high as 76% can be achieved with missingness scales as large as 50%. (iii) The introduction of missing values into the training data leads to great reduction in the performance of the ensemble classifier treatment. In such cases, the regression treatment works the best when values are missing at low percentage and the mean treatment works the best at when values are missing at high percentage. In this particular study we assume that all the features are important for prediction. We also assume the missing values are MCAR, and the distribution of missing values is uniform across all attributes. In the future we aim to examine the performance of the treatments under the influence of other distributions of missingness over the attributes, which might be the case for many practical situations. Combination of the imputation and ignorance treatments might result in a hybrid treatment that yields consistent good performance. Further more, this study focuses on numerical attributes only. Further experiments can be conducted to see that how does the results differ if categorical attributes come into the picture. Overall, we have demonstrated that a significant improvement in reliability can be achieved by choosing the correct missing value techniques. We also provide a guideline to facilitate the choice of missing value treatments under different scenarios. REFERENCES [1] R. Wang, D.Q. Chen, and Z. Fu, AWQEE-DSS: A Decision Support System for Aquaculture Water Quality Evaluation and Early-warning, International Conference on Computational Intelligence and Security, vol. 2, pp. 959-962, 2006. [2] T. Hastie, R. Tibshirani, and J. Friedman, The Elements of Statistical Learning, Springer Verlag, New York, 2001. [3] M. Sariyar, A. Borg and K. Pommerening, Missing Values in Deduplication of Electronic Patient Data, J Am Med Inform Assoc, vol 19, pp. e76-e82, 2012. [4] M. Saar-Tsechansky and F. Provost, Handling missing values when applying classification models Journal of machine learning research, vol. 8, pp. 1625-1657, 2007. [5] A. Rahman, Claire D Este, and G. Timms, Dealing with Missing Sensor Values in Predicting Shellfish Farm Closure in IEEE Eighth International Conference on Intelligent Sensors, Sensor Networks and Information Processing, 2013. [6] J. Grzymala-Busse and M. Hu, A Comparison of Several Approaches to Missing Attribute Values in Data Mining, Rough Sets and Current Trends in Computing - RSCTC, pp. 378-385, 2000. [7] J. A. Saunders, N. Morrow-Howell, E. Spitznagel, P. Dori, E. K. Proctor, and R. Pescarino, Imputing Missing Data: A Comparison of Methods for Social Work Researchers, Social Work Research, vol 30, pp. 19-31, 2006. [8] Garcia-Laencina, J.Pedro ; Sancho-Gomez, Jose-Luis; Figueiras-Vidal, and R. Anibal, Pattern classification with missing data: a review, Neural Computing and Applications, vol 19, pp. 263-282, 2010. [9] A. Farhangfar, L. Kurgan and J. Dy, Impact of imputation of missing values on classification error for discrete data, Pattern Recognition, vol 41, pp. 3692-3705, 2008. [10] K. Pelckmans, J. De Brabanter, J. Suykens, et al., Handling missing values in support vector machine classifiers, Neural Networks, vol 18, pp. 684-692, 2005. [11] P. Melville, N. Shah; L. Mihalkova, et al., Experiments on ensembles with missing and noisy data, in Proceeding of Multiple Classifier Systems, vol 3077, pp. 293-302, 2004. [12] L. Breiman, Bagging predictors, Machine Learning, vol. 24, no. 2, pp. 123-140, 1996. [13] R. E. Schapire, The strength of weak learnability, Machine Learning, vol. 5, no. 2, pp. 197-227, 1990. [14] P. Melville and R. Mooney, Constructing diverse classier ensembles using artificial training examples, in Proceeding of 18th International Joint Conference. on Artificial Intelligence, pp. 505-510, 2003. [15] A. Frank and A. Asuncion, UCI Machine Learning Repository [Online]. Available: http://www. ics. uci. edu/ml. Irvine, CA: University of California, School of Information and Computer Science, 2010. [16] J.R. Quinlan, Induction of Decision Trees Machine Learning, vol. 1, pp. 81-106, 1986. [17] T. Cover and P. Hart, Nearest neighbor pattern classification IEEE Transactions on Information Theory vol. 13, pp. 21-27, 1967. [18] N. Friedman, D. Greiger, and M. Goldszmidt Bayesian Network Classifiers Machine Learning, vol. 29, pp. 131-163, 1997. [19] R. Little and D. Rubin. Statistical Analysis with Missing Data John Wiley and Sons, 1987. [20] C.K. Enders, Applied Missing Data Analysis, New York: Guilford Press, 2010. [21] M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, and I.H. Witten, The WEKA Data Mining Software: An Update, SIGKDD Explorations, vol 11, 2009. 2200