A Learning Method of Directly Optimizing Classifier Performance at Local Operating Range

Similar documents
METHODS FOR DETECTING CERVICAL CANCER

Modifying ROC Curves to Incorporate Predicted Probabilities

A scored AUC Metric for Classifier Evaluation and Selection

AN EFFICIENT AUTOMATIC MASS CLASSIFICATION METHOD IN DIGITIZED MAMMOGRAMS USING ARTIFICIAL NEURAL NETWORK

When Overlapping Unexpectedly Alters the Class Imbalance Effects

Classification of benign and malignant masses in breast mammograms

LUNG CANCER continues to rank as the leading cause

Gene Selection for Tumor Classification Using Microarray Gene Expression Data

Mammogram Analysis: Tumor Classification

4. Model evaluation & selection

Predicting Breast Cancer Survivability Rates

Outlier Analysis. Lijun Zhang

Machine learning II. Juhan Ernits ITI8600

Classıfıcatıon of Dıabetes Dısease Usıng Backpropagatıon and Radıal Basıs Functıon Network

3. Model evaluation & selection

Lung Cancer Diagnosis from CT Images Using Fuzzy Inference System

QUANTIFICATION OF PROGRESSION OF RETINAL NERVE FIBER LAYER ATROPHY IN FUNDUS PHOTOGRAPH

1 Introduction. st0020. The Stata Journal (2002) 2, Number 3, pp

Chapter 1. Introduction

Multilayer Perceptron Neural Network Classification of Malignant Breast. Mass

Auto-Encoder Pre-Training of Segmented-Memory Recurrent Neural Networks

Intelligent Edge Detector Based on Multiple Edge Maps. M. Qasim, W.L. Woon, Z. Aung. Technical Report DNA # May 2012

Quantitative Evaluation of Edge Detectors Using the Minimum Kernel Variance Criterion

Mammogram Analysis: Tumor Classification

Selection and Combination of Markers for Prediction

A micropower support vector machine based seizure detection architecture for embedded medical devices

EARLY STAGE DIAGNOSIS OF LUNG CANCER USING CT-SCAN IMAGES BASED ON CELLULAR LEARNING AUTOMATE

COMPARATIVE STUDY ON FEATURE EXTRACTION METHOD FOR BREAST CANCER CLASSIFICATION

Question 1 Multiple Choice (8 marks)

Effect of Feedforward Back Propagation Neural Network for Breast Tumor Classification

CHAPTER 5 WAVELET BASED DETECTION OF VENTRICULAR ARRHYTHMIAS WITH NEURAL NETWORK CLASSIFIER

Yeast Cells Classification Machine Learning Approach to Discriminate Saccharomyces cerevisiae Yeast Cells Using Sophisticated Image Features.

Edge Detection Techniques Based On Soft Computing

Automated Detection of Tumors in Mammograms Using Two Segments for Classification

Copyright 2007 IEEE. Reprinted from 4th IEEE International Symposium on Biomedical Imaging: From Nano to Macro, April 2007.

NMF-Density: NMF-Based Breast Density Classifier

Assigning B cell Maturity in Pediatric Leukemia Gabi Fragiadakis 1, Jamie Irvine 2 1 Microbiology and Immunology, 2 Computer Science

arxiv: v2 [cs.cv] 8 Mar 2018

Review. Imagine the following table being obtained as a random. Decision Test Diseased Not Diseased Positive TP FP Negative FN TN

Using Genetic Algorithms to Optimise Rough Set Partition Sizes for HIV Data Analysis

Leukemia Blood Cell Image Classification Using Convolutional Neural Network

Classification of ECG Data for Predictive Analysis to Assist in Medical Decisions.

Classification of mammogram masses using selected texture, shape and margin features with multilayer perceptron classifier.

Information Processing During Transient Responses in the Crayfish Visual System

Artificial Neural Networks in Cardiology - ECG Wave Analysis and Diagnosis Using Backpropagation Neural Networks

Large-Scale Statistical Modelling via Machine Learning Classifiers

Investigating the performance of a CAD x scheme for mammography in specific BIRADS categories

Using AUC and Accuracy in Evaluating Learning Algorithms

Competition Between Objective and Novelty Search on a Deceptive Task

PII: S (96) THE USE OF THE AREA UNDER THE ROC CURVE IN THE EVALUATION OF MACHINE LEARNING ALGORITHMS

IEEE SIGNAL PROCESSING LETTERS, VOL. 13, NO. 3, MARCH A Self-Structured Adaptive Decision Feedback Equalizer

The Optimization Variables of Input Data of Artificial Neural Networks for Diagnosing Acute Appendicitis

Breast Cancer Diagnosis using a Hybrid Genetic Algorithm for Feature Selection based on Mutual Information

Computational Intelligence Lecture 21: Integrating Fuzzy Systems and Neural Networks

IN THE iterated prisoner s dilemma (IPD), two players

MICROWAVE IMAGING FOR THE DETECTION AND LOCALIZATION OF BREAST CANCER USING ARTIFICIAL NEURAL NETWORK

MRI Image Processing Operations for Brain Tumor Detection

An Efficient Attribute Ordering Optimization in Bayesian Networks for Prognostic Modeling of the Metabolic Syndrome

Using stigmergy to incorporate the time into artificial neural networks

Probability-Based Protein Identification for Post-Translational Modifications and Amino Acid Variants Using Peptide Mass Fingerprint Data

Improved Intelligent Classification Technique Based On Support Vector Machines

Comparing Two ROC Curves Independent Groups Design

SimTest. Niklaus E. Zimmermann Swiss Federal Research Institute WSL CH-8903 Birmensdorf

JSM Survey Research Methods Section

CSE Introduction to High-Perfomance Deep Learning ImageNet & VGG. Jihyung Kil

Mining of an electrocardiogram

Mining the Risk Types of Human Papillomavirus (HPV) by AdaCost

Motivation: Fraud Detection

Efficacy of the Extended Principal Orthogonal Decomposition Method on DNA Microarray Data in Cancer Detection

COMPUTERIZED SYSTEM DESIGN FOR THE DETECTION AND DIAGNOSIS OF LUNG NODULES IN CT IMAGES 1

Introduction to diagnostic accuracy meta-analysis. Yemisi Takwoingi October 2015

10CS664: PATTERN RECOGNITION QUESTION BANK

Learning with Rare Cases and Small Disjuncts

RetCAD 1.3.0: White paper

Prediction of Breast Cancer Biopsy Outcomes Using a Distributed Genetic Programming Approach

Technical Specifications

Introduction. We can make a prediction about Y i based on X i by setting a threshold value T, and predicting Y i = 1 when X i > T.

Statistics 202: Data Mining. c Jonathan Taylor. Final review Based in part on slides from textbook, slides of Susan Holmes.

Classification of Cervical Cancer Cells Using HMLP Network with Confidence Percentage and Confidence Level Analysis.

Remarks on Bayesian Control Charts

Prediction of Diabetes by using Artificial Neural Network

An Introduction to ROC curves. Mark Whitehorn. Mark Whitehorn

University of Cambridge Engineering Part IB Information Engineering Elective

Predictive performance and discrimination in unbalanced classification

Deep Learning-based Detection of Periodic Abnormal Waves in ECG Data

Brain Tumor Segmentation Based On a Various Classification Algorithm

Epileptic seizure detection using EEG signals by means of stationary wavelet transforms

WHILE behavior has been intensively studied in social

Early Detection of Lung Cancer

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING

ROC (Receiver Operating Characteristic) Curve Analysis

A Hierarchical Artificial Neural Network Model for Giemsa-Stained Human Chromosome Classification

Supplementary Materials

Machine Learning! Robert Stengel! Robotics and Intelligent Systems MAE 345,! Princeton University, 2017

J2.6 Imputation of missing data with nonlinear relationships

Keywords Artificial Neural Networks (ANN), Echocardiogram, BPNN, RBFNN, Classification, survival Analysis.

CHAPTER 4 CLASSIFICATION OF HEART MURMURS USING WAVELETS AND NEURAL NETWORKS

Analysis of Speech Recognition Techniques for use in a Non-Speech Sound Recognition System

1 Diagnostic Test Evaluation

Transcription:

A Learning Method of Directly Optimizing Classifier Performance at Local Operating Range Lae-Jeong Park and Jung-Ho Moon Department of Electrical Engineering, Kangnung National University Kangnung, Gangwon-Do, 210-702, South Korea {ljpark, itsmoon}@kangnung.ac.kr Abstract This paper addresses an effective learning method that enables us to directly optimize neural network classifier's discrimination performance at a desired local operating range by maximizing a partial area under a receiver operating characteristic (ROC) or domain-specific curve, which is difficult to achieve with classification accuracy or mean squared error (MSE)-based learning methods. The effectiveness of the proposed approach is demonstrated in terms of fraud detection capability in the credit card fraud detection, compared with the MSE-based approach. Keyword: Classification, Receiver Operating Characteristic, And Area Under Curve. I. Introduction In general, design and learning of a classifier for financial real-world two-class classification problems are plagued by severely overlapping class distribution because samples in one class are often similar or even identical to those in the other class due to the nature of the problems. Examples are database marketing, churn prediction, and fraud detection. Classification accuracy or minimization of misclassifications has been conventionally used to evaluate classifier's discrimination ability. For neural network (NN) classifiers, the mean squared error (MSE) between the actual output and the desired target is defined and minimized to reduce the number of misclassifications. In the severely overlapping class distribution problems, a trade-off between true positive rate (TPR) and false positive rate (FPR) is unavoidable and hence should be determined carefully. The MSE minimization is, however, unsuitable to evaluation of a NN classifier for the severely overlapping class problem because the MSE is not one-to-one correspondent to the performance in terms of TPR and FPR of the NN classifier. Receiver operating characteristic (ROC) curve 1 has been recently used to evaluate classifier's discrimination performance in the skewed and overlapping data sets since its introduction by [1] in the data mining communities. The ROC curve makes it possible to visualize a trade-off of classifier's discrimination capability that is indistinguishable in the MSE measure. Moreover, so-called AUC 1 The TPR is plotted on the Y axis and the FPR is plotted on the X axis. The pairs of TPR and FPR are obtained by varying a decision threshold on the single continuous output of a classifier. 74

International Journal of Information Technology Vol. 11 No. 10 2005 (the area under ROC curve) 2 has been proposed as a single performance measure to deal with a problem with specifying classifier's performance in terms of a single operating point on its ROC curve [2]. The AUC reflects an average classifier's performance on the entire operating points. The larger AUC is, the better classifier's average discrimination performance is. The AUC may be an appropriate performance measure if the class ratio and misclassification costs are unknown and/or a single classifier with a fixed decision threshold must be chosen to handle every possible operating points. Recently, much attention have been paid to design and train classifiers by maximizing the AUC in financial and medical applications [3,4,5,6,7]. On the other hand, in some classification applications, it is often desirable or important to evaluate and optimize classifier's discrimination performance at a certain operating range, not in the entire operating range as in the AUC. For example, in medical diagnosis, TPRs of less than, say, 0.7-0.8 would be probably unacceptable, because patients with a disease should be detected even if it turns out that it is a false detection. In credit card fraud detection, a fraud detection system should not operate in a range of high FPRs because it cannot handle the overwhelming number of suspicious transactions. In order to produce a high-quality classifier in those application domains, a method for optimizing classifier's discrimination performance at a desired local operating range, for example, the TPR at a certain range of FPRs is required, but to my knowledge, how to optimize a partial area under a ROC curve or the like has been rarely addressed. In this paper, an effective learning approach is proposed that makes it possible to optimize the discrimination performance of a NN classifier at a desired specific operating range by utilizing a partial area under a ROC or domain-specific curve, which is difficult to achieve with common MSEbased learning methods. The performance of the proposed approach is examined and compared with the MSE-based approach in credit card fraud detection, which is a representative one of the severely overlapping real-world classification problems. II. Optimization of Classifier Performance at Local Operating Range A. ROC, AUC, and Partial AUC The ROC curve of a NN classifier is obtained by varying a threshold θ on the continuous output of the classifier, ranging from [0,1]. An example of histograms of classifier outputs of positive and negative samples is illustrated in Fig. 1. Samples whose their outputs are greater than θ are classified as positives, otherwise they are classified as negatives. The more overlapping the two classes are, the more likely it is that the two histograms overlap each other. Given a value of θ, a point of (FPR,TPR) on the ROC space is determined by negatives incorrectly classified FPR =, total negatives positives correctly classified TPR =. (1) total positives 2 AUC stands for the Area Under the Curve. For a ROC curve rx, ( ) the AUC is represented by 1 0 rx ( ) dx, where x is FPR. 75

Figure 1 Two histograms of classifier s outputs of samples of the two classes Fig. 2 shows ROC curves of three distinct NN classifiers. Classifier B is completely outperformed by classifier A. The less likely it is that the two histograms of classifier outputs overlap each other, the more bowed the ROC curve is toward the left corner of (0,1), which represents the perfect discrimination. The AUC is a single measure to quantify the difference of classifier performances across all decision thresholds. The AUC of classifier A is definitely larger than that of classifier B. The AUC has been recently adopted as an informative classifier performance measure compared with the error rate and therefore several methods to optimize the AUC directly have been proposed. Verrelst et. al. adopted simulated annealing to maximize the AUC to produce a MLP classifier in ovarian tumor malignancy prediction problem [3]. Yan et. al. proposed a gradient-based training algorithm for directly maximizing the AUC by using a differentiable objective function that is approximation to the Wilcoxon-Mann-Whitney statistic, which is equivalent to the AUC [4]. In some applications, it is meaningful to focus on an area under only a portion of the ROC curve, for example, within a specific range of FPRs or TPRs because no interest lies in the entire range of FPRs or TPRs. For example, in a diagnostic test, much attention is generally paid to the portion of the ROC curve where TPR is greater than a predetermined threshold. For those applications, since a classifier with a larger AUC value may have lower discrimination performance than classifiers with smaller AUC values in a specific range of FPRs or TPRs (as shown in Fig. 2, classifier C is slightly better than classifier A at a low FPR range, but poor in terms of AUC values), it is required to train a classifier by optimizing discrimination performance at a desired local operating range that may be indistinguishable by the AUC performance measure. To my knowledge, there have been few attentions on importance of classifier's discrimination performance at a local operating range in machine learning and data mining communities, and therefore, there have been few researches on methods designed to train a NN classifier by directly optimizing a partial area under the ROC curve or an application-specific curve, so-called PAUC (Partial AUC). 76

International Journal of Information Technology Vol. 11 No. 10 2005 Figure 2 ROC curves of three distinct classifiers B. PAUC Maximization Unfortunately, the backpropagation (BP) algorithm that minimizes the MSE of a NN classifier is not suited for optimizing classifier's performance in terms of the PAUC. It is because the MSE minimization at the outputs does not correspond directly to the PAUC-based optimization. A nonstandard training method is required because it is impossible to compute the partial derivatives of the PAUC with respect to the weights. A NN classifier is, therefore, trained by an effective search algorithm or Evolutionary Programming (EP) rather than gradient-descent algorithms so that the weights are chosen in terms of the PAUC optimization. An objective function for the PAUC optimization is given in the form of β α 1 0 Cx ( ) dx, (2) Cx ( ) dx where [α,β] specifies a desired operating range, Cx ( ) is a ROC curve or a domain-specific curve, and x is a FPR or problem-dependent variable. Table 1 shows a procedure for numerically calculating the PAUC value of a NN classifier. In the STEP 3, NN classifiers that the number of (TPR, FPR) points in [α,β] is smaller than κ, is excluded or penalized during the evolutionary search in order not to produce NN classifiers that have a risk of generating an abrupt transition of classifier output histograms, which may be associated with poor discrimination performance. Table 1. A procedure for numerical PAUC calculation of a NN classifier Input: [α,β] : a range of FPRs N θ : The number of thresholds between [0,1] 77

STEP 1: STEP 2: STEP 3: N θ pairs of (TPR, FPR) are obtained by varying the threshold on the NN classifier s output A ROC curve is constructed by using N θ pairs of (TPR, FPR) A partial area under the curve in the range of [α,β] is calculated with the trapezoidal integration rule III. Experiments on Fraud Detection A. Credit Card Fraud Detection Credit card fraud detection [8,9] is one of real-world two-class classification problems where the class distribution is severely overlapping and the class ratio is highly skewed. Fraudulent transactions resemble legitimate ones so that it is impossible to detect fraudulent transactions without a lot of false detection. Hence, a classifier for credit card fraud detection should be operated on a very low range of FPRs in order to reduce the number of legitimate transactions incorrectly detected. Therefore, it is crucial to optimize classifier's discrimination performance at a specific local range of FPRs, which is determined based on the several operational constraints. It should be noted that learning of classifiers is performed not at a single operating point but in a local operating range, because an operating point changes with time, from a month to another, depending on the changing class distribution. Two data sets of credit card transactions labeled as legitimate or fraudulent were provided by a credit card company in Korea. A set consisting of about 51,260 transactions collected selectively during one year is used as a training data set. Another data set of about 7 millions transactions during three months is used to evaluate the NN classifier. A multi-layered perceptron (MLP) with one hidden layer and sigmoidal function is used as the NN classifier. The six features are extracted from each transaction by using machine learning techniques and are then used for the inputs of the NN classifier. B. Application-specific PAUC Maximization For credit card fraud detection, it is practically desirable that an operating range of the NN classifier is represented by a rejection rate (true positives plus false positives) since the rate is a major monitoring variable that is carefully controlled due to the constraint imposed on the capacity of suspicious transaction investigation. Instead of a ROC curve, a domain-specific curve Cx ( ) is introduced that represents the number of correctly detected fraudulent transactions with respect to a rejection rate, x. The EP with Gaussian mutation is used to train a NN classifier in order to maximize Eq. (2). With the domain-specific curve, the objective is to maximize an average of the correctly detected frauds in the chosen range of the rejection rates by maximizing the partial area under the curve. A penalty term is added to the cost function of EP so as to penalize classifiers that have a small number of (TPR, FPR) points falling on [α,β] since their decision boundaries locate the regions on the feature space where samples are distributed densely. The value of κ is set to 20 by trial-and-error. The EP parameter values for all experiments are as follows. The 78

International Journal of Information Technology Vol. 11 No. 10 2005 number of generations and the population size are chosen as 1000 and 20, respectively, for a reasonable convergence speed. The tournament size is 10 and σ in the Gaussian mutation is 1. C. Experimental Results and Discussions Two sets of experiments were performed to evaluate and compare the PAUC-based NN classifier with the MSE-based one. For fair comparison, the MSE minimization was performed by the EP, too. For each criterion, ten NNs were trained with 10, 15, 20 hidden nodes. The desired operating range was chosen as [0.001,0.004] to reflect reality. Table 2 shows the averaged MSE and PAUC values of ten classifiers trained with the PAUC criterion and ten ones with the MSE criterion. As expected, the averaged MSE value of classifiers trained with the MSE criterion is significantly smaller than that of classifiers trained with the PAUC criterion, and averaged PAUC value of classifiers trained with the PAUC criterion is significantly larger than that of classifiers trained with the MSE criterion. An explanation could be that, by maximization of Eq. (2) the decision boundary of a NN classifier is formed in such a way that Cx ( ) within the operating range of [α,β] increases regardless of large MSE contribution of the samples associated with the rejection rates of [ β,1]. Table 2. Averages and standard deviations of MSE and PAUC values of twenty NN classifiers, a half trained with the PAUC criterion, and the others with the MSE criterion Learning criterion (MSE, PAUC) MSE minimization (3.07±0.03, 2.91±0.10) PAUC maximization (4.05±0.85, 3.05±0.02) Figure 3 X-Y plot of twenty NN classifiers performance on the (MSE, PAUC) space, a half trained with the PAUC, the others with the MSE 79

Fig. 3 shows a X-Y plot of the twenty classifiers' performance on the (MSE, PAUC) space. As seen in Fig. 3, classifiers trained with the PAUC criterion (or MSE criterion) have a large σ in the MSE values (or PAUC values) compared with σ in the PAUC values (or MSE values), which implies that the MSE minimization does not correspond to the PAUC maximization one-to-one. The similar results between AUC and MSE performance indices have been observed in [6,7]. It has been revealed analytically in [6] that algorithms designed to minimize the error rate may not lead to the best AUC possible values. In [7], by empirical comparison of error surfaces in the weight space of MSE, AUC, and partial AUC performance measures, they showed that MSE minimization tends to maximize AUC, but not partial AUC defined at a range of high true positive rates. Fig. 4 shows a difference between the two averaged curves of the top five classifiers trained with the PAUC maximization and the top five ones with the MSE minimization on the test set. Note that a coordinate value of 0 in the Y axis represents that the PAUC maximization gives no performance improvement over the MSE minimization. Even though training of a NN classifier with the PAUC criterion generates classifiers with large MSE values, compared with the MSE-based classifiers, it enables a NN classifier to detect more fraudulent transactions at the specified rejection rate of [0.001,0.004] by increasing the partial area under the curve within the operating range. The performance improvement at the local operating range results from the difference of shape of the two histograms of classifier's outputs of legitimate and fraudulent transaction samples. Fig. 5 shows the histograms of classifier's outputs of legitimate (solid line) and fraudulent (dash-dotted line) transaction samples for MSE-based and PAUC-based classifiers with 10 hidden nodes. The fraud score is calculated by 500(1+o), where o is the classifier output ranging between -1 and 1. For the MSE-based classifier, the histogram of fraudulent samples has the spike-like shape near at the fraud score of 800, which is caused by the force to minimize the MSE during training (The target outputs of the fraudulent samples are set to +1). On the other hand, for the PAUC maximization, instead of adjusting the classifier outputs to the associated target values, a classifier is trained in such a way that the output histograms of the two classes are overlapped as less as possible at a high range of fraud scores that corresponds to the specified rejection rate [0.001,0.004]. Hence, it can be seen in Fig. 5(b) that the spike-like shape of the histograms of Fig. 5(a) disappears and that the two histograms flatten and get less overlapped (more separable) around at the fraud score of 800. On the other hand, it is observed that the curve ( ) Cx of the PAUC-based classifier increases more slowly at a medium range of the rejection rates than that of MSE-based classifiers to compensate the increase in the range of [0.001,0.004]. The improvement is consistent with the number of hidden nodes between 10 and 20. 80

International Journal of Information Technology Vol. 11 No. 10 2005 (a) (b) Figure 4 Difference between the two averaged curves of the top five NN classifiers in terms of the PAUC and MSE. (a) 10 hidden nodes. (b) 20 hidden nodes (a) (b) Figure 5 Histograms of classifier outputs of legitimate and fraudulent transaction samples. (a) histograms of MSE-based classifier. (b) histograms of PAUC-based classifier IV. Conclusion A learning method has been proposed that enables us to directly optimize neural network classifier's discrimination performance at a desired specific operating range by maximizing a partial area under a ROC or domain-specific curve, which is difficult to achieve with MSE-based learning methods. The effectiveness of the proposed approach has been demonstrated and compared with the MSEbased approach in terms of discrimination performance in credit card fraud detection in which interests lie in only a range of very low false positive rates. The experimental results with real credit card transactions data have demonstrated that the proposed approach makes it possible to detect more fraudulent transactions in a desired operating range. 81

References [1] F. Provost and T. Fawcett, Analysis and visualization of classifier performance comparison under imprecise class and cost distributions, in Proc. of Conf. Knowledge Discovery and Data Mining, 1997, pp. 43 48. [2] A. P. Bradley, The use of the area under the ROC curve in the evaluation of machine learning algorithms, in Pattern Recognition, vol. 30, 1997, pp. 1145 1159. [3] H. Verrelst, Y. Moreau, and D. Timmerman, Use of a multi-layer perceptron to predict malignancy in ovarian tumors, in Proc. of Conf. on Advances in Neural Information Processing Systems, 1998, pp. 978 984. [4] L. Yan, R. Dodier, M. C. Mozer, and R. Wolniewicz, Optimizing classifier performance via the Wilcoxon-Mann-Whitney statistics, in Proc. of Int. Conf. on Machine Learning, 2003, pp. 848 855. [5] B. Sahnier, H.-P. Chan, N. Petrick, S. S. Gopal, and M. M. Goodsitt, Neural network design for optimization of the partial area under the receiver operating characteristic curve, in Proc. IEEE Conf. on Neural Networks, 1997, pp. 2468 2471 [6] C. Cortes and M. Mohri, AUC optimization vs. error rate minimization, in Proc. of Conf. on Advances in Neural Information Processing Systems, 2003. [7] M. K. Markey, J. Y. Lo, R. Vargas-woracek, G. D. Tourassi, and Jr. C. E. Floyd, Perceptron error surface analysis: A case study breast cancer diagnosis, in Computers in Biology and Medicine, vol. 32, 2002, pp. 99 109. [8] R. J. Bolton and D. J. Hand, Statistical fraud detection: A review, in Statistical Science, vol. 17, 2002, pp. 235 255. [9] P. K. Chan, W. Fan, A. L. Prodromidis, and S. J. Stolfo, Distributed data mining in credit card fraud detection, in IEEE Intelligent Systems, vol. 14, 1999, pp. 67 74. Lae-Jeong Park received the B.S. degree in Electrical Engineering from Seoul National University in 1991, and the M.S. and Ph.D. degrees in Electrical Engineering from the Korea Advanced Institute of Science and Technology in 1993 and 1997, respectively. He was with the Information Technology Lab. at LG Corporate Institute of Technology, Seoul, Korea. He is currently an assistant professor in the Department of Electrical Engineering at Kangnung National University, Korea. His current research interests are machine learning, pattern recognition, and evolutionary and neural computation. Jung-Ho Moon received the B.S. degree in control and instrumentation engineering from Seoul National University in 1991, the M.S. and Ph.D. degrees in electrical engineering from Korea Advanced Institute of Science and Technology in 1993 and 1998, respectively. In 2002, he joined the faculty of the Department of Electrical Engineering, Kangnung National University, Korea, where he is an Assistant Professor. From 1998 to 2000, he worked for Samsung Electronics Co., Ltd as a Senior Research Engineer, and from 2001 to 2002, he worked for Humax Co., Ltd as a Senior H/W Engineer. His research interests include digital control, disk drive servo systems, intelligent control, and embedded systems. 82