Applying One-vs-One and One-vs-All Classifiers in k-nearest Neighbour Method and Support Vector Machines to an Otoneurological Multi-Class Problem

Similar documents
Class discovery in Gene Expression Data: Characterizing Splits by Support Vector Machines

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014

Colon cancer subtypes from gene expression data

Variable Features Selection for Classification of Medical Data using SVM

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Midterm, 2016

Breast Cancer Diagnosis using a Hybrid Genetic Algorithm for Feature Selection based on Mutual Information

Comparison of discrimination methods for the classification of tumors using gene expression data

Predicting Sleep Using Consumer Wearable Sensing Devices

7.1 Grading Diabetic Retinopathy

Data mining for Obstructive Sleep Apnea Detection. 18 October 2017 Konstantinos Nikolaidis

On Signal Analysis of Three-Dimensional Nystagmus

INTRODUCTION TO MACHINE LEARNING. Decision tree learning

Research Article The Classification of Valid and Invalid Beats of Three-Dimensional Nystagmus Eye Movement Signals Using Machine Learning Methods

Assigning B cell Maturity in Pediatric Leukemia Gabi Fragiadakis 1, Jamie Irvine 2 1 Microbiology and Immunology, 2 Computer Science

Applied Machine Learning in Biomedicine. Enrico Grisan

BREAST CANCER EPIDEMIOLOGY MODEL:

An automatic mammogram system: from screening to diagnosis. Inês Domingues

EMOTION CLASSIFICATION: HOW DOES AN AUTOMATED SYSTEM COMPARE TO NAÏVE HUMAN CODERS?

Identification of Neuroimaging Biomarkers

GDP Growth vs. Criminal Phenomena: Data Mining of Japan

Predicting Breast Cancer Survival Using Treatment and Patient Factors

Design of Multi-Class Classifier for Prediction of Diabetes using Linear Support Vector Machine

Brain Tumour Detection of MR Image Using Naïve Beyer classifier and Support Vector Machine

INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY

Data complexity measures for analyzing the effect of SMOTE over microarrays

Comparison Classifier: Support Vector Machine (SVM) and K-Nearest Neighbor (K-NN) In Digital Mammogram Images

Gender Based Emotion Recognition using Speech Signals: A Review

Prediction Models of Diabetes Diseases Based on Heterogeneous Multiple Classifiers

Improved Hepatic Fibrosis Grading Using Point Shear Wave Elastography and Machine Learning

Efficacy of the Extended Principal Orthogonal Decomposition Method on DNA Microarray Data in Cancer Detection

Efficient Classification of Cancer using Support Vector Machines and Modified Extreme Learning Machine based on Analysis of Variance Features

Classification. Methods Course: Gene Expression Data Analysis -Day Five. Rainer Spang

Classification of Mammograms using Gray-level Co-occurrence Matrix and Support Vector Machine Classifier

Predicting Breast Cancer Recurrence Using Machine Learning Techniques

Nature Neuroscience: doi: /nn Supplementary Figure 1. Behavioral training.

Hybridized KNN and SVM for gene expression data classification

Multiclass microarray data classification based on confidence evaluation

Introduction to Discrimination in Microarray Data Analysis

Sound Texture Classification Using Statistics from an Auditory Model

Predicting Diabetes and Heart Disease Using Features Resulting from KMeans and GMM Clustering

Handwriting - marker for Parkinson s Disease

ABSTRACT I. INTRODUCTION. Mohd Thousif Ahemad TSKC Faculty Nagarjuna Govt. College(A) Nalgonda, Telangana, India

Mammogram Analysis: Tumor Classification

Available online at ScienceDirect. Procedia Computer Science 96 (2016 )

Classification of EEG signals in an Object Recognition task

An Empirical and Formal Analysis of Decision Trees for Ranking

Learning Convolutional Neural Networks for Graphs

Kernel Methods and String Kernels for Authorship Analysis

TITLE: A Data-Driven Approach to Patient Risk Stratification for Acute Respiratory Distress Syndrome (ARDS)

MACHINE LEARNING BASED APPROACHES FOR PREDICTION OF PARKINSON S DISEASE

BRAIN TUMOR SEGMENTATION USING DEEP NEURAL NETWORK

A Vision-based Affective Computing System. Jieyu Zhao Ningbo University, China

Data Mining in Bioinformatics Day 4: Text Mining

THE data used in this project is provided. SEIZURE forecasting systems hold promise. Seizure Prediction from Intracranial EEG Recordings

Case Studies of Signed Networks

Self-Advising SVM for Sleep Apnea Classification

Supersparse Linear Integer Models for Interpretable Prediction. Berk Ustun Stefano Tracà Cynthia Rudin INFORMS 2013

Breast Cancer Prevention and Early Detection using Different Processing Techniques

EFFICIENT MULTIPLE HEART DISEASE DETECTION SYSTEM USING SELECTION AND COMBINATION TECHNIQUE IN CLASSIFIERS

Performance Based Evaluation of Various Machine Learning Classification Techniques for Chronic Kidney Disease Diagnosis

HYBRID SUPPORT VECTOR MACHINE BASED MARKOV CLUSTERING FOR TUMOR DETECTION FROM BIO-MOLECULAR DATA

Jacobi Moments of Breast Cancer Diagnosis in Mammogram Images Using SVM Classifier

Winner s Report: KDD CUP Breast Cancer Identification

Learning in neural networks

COMPARISON BETWEEN GMM-SVM SEQUENCE KERNEL AND GMM: APPLICATION TO SPEECH EMOTION RECOGNITION

Adapting Breast Density Classification from Digitized to Full-Field Digital Mammograms

Identification of Thyroid Cancerous Nodule using Local Binary Pattern Variants in Ultrasound Images

A New Approach for Detection and Classification of Diabetic Retinopathy Using PNN and SVM Classifiers

IJESRT. Scientific Journal Impact Factor: (ISRA), Impact Factor: 1.852

Classification of Insulin Dependent Diabetes Mellitus Blood Glucose Level Using Support Vector Machine

Machine Learning for Predicting Delayed Onset Trauma Following Ischemic Stroke

Supporting Information Identification of Amino Acids with Sensitive Nanoporous MoS 2 : Towards Machine Learning-Based Prediction

Confluence: Conformity Influence in Large Social Networks

Prediction of Alternative Splice Sites in Human Genes

Classıfıcatıon of Dıabetes Dısease Usıng Backpropagatıon and Radıal Basıs Functıon Network

Identifying Thyroid Carcinoma Subtypes and Outcomes through Gene Expression Data Kun-Hsing Yu, Wei Wang, Chung-Yu Wang

Pre-mRNA Secondary Structure Prediction Aids Splice Site Recognition

Contents. Just Classifier? Rules. Rules: example. Classification Rule Generation for Bioinformatics. Rule Extraction from a trained network

Multi-atlas-based segmentation of the parotid glands of MR images in patients following head-and-neck cancer radiotherapy

Machine Learning Classifications of Coronary Artery Disease

Supervised Learning Approach for Predicting the Presence of Seizure in Human Brain

EECS 433 Statistical Pattern Recognition

Evaluating Classifiers for Disease Gene Discovery

BraTS : Brain Tumor Segmentation Some Contemporary Approaches

Utilizing Posterior Probability for Race-composite Age Estimation

The Long Tail of Recommender Systems and How to Leverage It

Comparative Study of Classification System using K-NN, SVM and Adaboost for Multiple Sclerosis and Tumor Lesions using Brain MRI

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

MRI Image Processing Operations for Brain Tumor Detection

Gene Selection for Tumor Classification Using Microarray Gene Expression Data

Predicting clinical outcomes in neuroblastoma with genomic data integration

Comparing Decision Support Methodologies for Identifying Asthma Exacerbations

Classification of benign and malignant masses in breast mammograms

Classifying Substance Abuse among Young Teens

Applying Data Mining for Epileptic Seizure Detection

Mammogram Analysis: Tumor Classification

IMPROVED SELF-ORGANIZING MAPS BASED ON DISTANCE TRAVELLED BY NEURONS

Detection of Heart Disease using Binary Particle Swarm Optimization

Transcription:

Oral Presentation at MIE 2011 30th August 2011 Oslo Applying One-vs-One and One-vs-All Classifiers in k-nearest Neighbour Method and Support Vector Machines to an Otoneurological Multi-Class Problem Kirsi Varpa, Henry Joutsijoki, Kati Iltanen, Martti Juhola School of Information Sciences - Computer Science University of Tampere, Finland

MIE 2011 Oslo -- Kirsi Varpa -- 1 Introduction: From a Multi-Class Classifier to Several Two-Class Classifiers We studied how splitting of a multi-class classification task into several binary classification tasks affected predictive accuracy of machine learning methods. One classifier holding nine disease class patterns was separated into multiple two-class classifiers. Multi-class classifier can be converted into One-vs-One (OVO, 1-vs-1) or One-vs-All the rest (OVA,1-vs-All) classifiers.

MIE 2011 Oslo -- Kirsi Varpa -- 2 From a Multi-Class Classifier to Several Two- Class Classifiers OVO 1-2-3-4-5-6-7-8-9 OVA nr of classifiers = 9 = nr of classes nr of classifiers = 36 = nr of classes (nr of classes 1) 2

MIE 2011 Oslo -- Kirsi Varpa -- 3 One-vs-One (OVO) Classifier The results of each classifier are put together, thus having 36 class proposals (votes) for the class of the test sample. The final class for the test sample is chosen by the majority voting method, the max-wins rule: A class, which gains the most votes, is chosen as the final class. [2 3 3 4 5 6 7 8 1 2 5 6 7 8 9 1 5 3 7 5 6 1 2 4 8 5 1 7 3 4 1 8 9 1 2 1] à max votes to class 1 (max-wins) [2 3 3 4 5 6 7 8 1 2 5 6 7 8 9 1 5 3 7 5 6 1 2 4 8 5 6 7 3 4 1 8 9 1 2 9] à max votes to classes 1 and 5 à tie: SVM: 1-NN between tied classes 1 and 5, k-nn: nearest class (1 or 5) from classifiers 5-6, 1-3, 3-5, 1-4, 2-5, 5-8, 1-5, 5-9, 1-7 and 1-8.

MIE 2011 Oslo -- Kirsi Varpa -- 4 One-vs-All (OVA) Classifier Each classifier is trained to separate one class from all the rest of the classes. The class of the rest of the cases is marked to 0. The test sample is input to each classifier and the final class for the test sample is assigned according to the winner-takes-all rule from a classifier voting a class. [ 0 0 0 0 5 0 0 0 0] à vote to a class 5 (winner-takes-all) [ 0 0 0 0 0 0 0 0 0] à tie: find 1-NN from all the classes [ 0 2 0 0 0 6 0 0 0] à votes to classes 2 and 6 à tie: SVM: 1-NN between tied classes 2 and 6, k-nn: nearest class (2 or 6) from classifiers 2-vs-All and 6-vs-All.

MIE 2011 Oslo -- Kirsi Varpa -- 5 Data Classifiers were tested with an otoneurological data containing 1,030 vertigo cases from nine disease classes. The dataset consists of 94 attributes concerning a patient s health status: occurring symptoms, medical history and clinical findings. Disease name N Acoustic Neurinoma 131 12.7 Benign Positional Vertigo 173 16.8 Meniere's Disease 350 34.0 Sudden Deafness 47 4.6 Traumatic Vertigo 73 7.1 Vestibular Neuritis 157 15.2 Benign Recurrent Vertigo 20 1.9 Vestibulopatia 55 5.3 Central Lesion 24 2.3 The data had about 11 missing values, which were imputed.

MIE 2011 Oslo -- Kirsi Varpa -- 6 Methods OVO and OVA classifiers were tested using 10-fold cross-validation 10 times with k-nearest Neighbour (k-nn) method and Support Vector Machines (SVM). Basic 5-NN method (using a classifier with all disease classes) was also run in order to have the baseline where to compare the effects of using multiple classifiers.

MIE 2011 Oslo -- Kirsi Varpa -- 7 k-nearest Neighbour Method (k-nn) k-nn method is a widely used, basic instance-based learning method that searches for the k most similar cases of a test case from the training data. In similarity calculation were used Heterogeneous Value Difference Metric (HVDM).

MIE 2011 Oslo -- Kirsi Varpa -- 8 Support Vector Machine (SVM) The aim of SVM is to find a hyperplane that separates classes C 1 and C 2 and maximizes the margin, the distance between the hyperplane and the closest members of both classes. The points, which are the closest to the separating hyperplane, are called Support Vectors. Kernel functions were used with SVM because the data was linearly non-separable in the input space.

Results Disease Cases 1,030 5-NN 5-NN OVO Classifiers SVM linear SVM RBF MIE 2011 Oslo -- Kirsi Varpa -- 9 5-NN OVA Classifiers SVM linear Acoustic Neurinoma 131 89.5 95.0 91.6 87.2 90.2 90.6 90.7 Benign Positional Vertigo 173 77.9 79.0 70.0 67.0 77.6 73.5 78.6 Meniere s disease 350 92.4 93.1 83.8 90.1 89.8 87.8 91.5 Sudden Deafness 47 77.4 94.3 88.3 79.4 87.4 61.3 58.1 Traumatic vertigo 73 89.6 96.2 99.9 99.3 77.7 79.9 96.7 Vestibular Neuritis 157 87.7 88.2 82.4 81.4 85.0 85.4 84.3 Benign Recurrent Vertigo 20 3.0 4.0 20.0 16.5 8.0 21.0 8.0 Vestibulopatia 55 9.6 14.0 16.5 22.8 15.8 15.3 13.5 Central Lesion 24 5.0 2.1 26.0 28.5 15.0 19.0 15.8 Median of True Positive Rate () 77.9 88.2 82.4 79.4 77.7 73.5 78.6 Total Classification accuracy () 79.8 82.4 77.4 78.2 78.8 76.8 79.4 Linear kernel with box constraint bc = 0.20 (OVO and OVA) Radial Basis Function (RBF) kernel with bc = 0.4 and scaling factor σ = 8.20 (OVO), bc = 1.4 and σ =10.0 (OVA) SVM RBF

MIE 2011 Oslo -- Kirsi Varpa -- 10 Conclusions The results show that in most of the disease classes the use of multiple binary classifiers improves the true positive rates of disease classes. The results show that in most of the disease classes use positive of multiple rates of binary disease classifiers classesimproves the true. with Especially, this data 5-NN than 5-NN with OVO with OVA classifiers classifiers. worked out better the

MIE 2011 Oslo -- Kirsi Varpa -- 11 Thank you for your attention! Questions? Kirsi.Varpa@cs.uta.fi More information about the subject: Questions? Kirsi.Varpa@cs.uta.fi More information about the subject: Allwein EL, Schapire RE, Singer Y. Reducing multiclass to binary: a unifying approach for margin