AIME03, Oct 21, 2003 Classification of Ovarian Tumors Using Bayesian Least Squares Support Vector Machines C. Lu 1, T. Van Gestel 1, J. A. K. Suykens.

Slides:



Advertisements
Similar presentations
Neural Networks and Kernel Methods
Advertisements

(SubLoc) Support vector machine approach for protein subcelluar localization prediction (SubLoc) Kim Hye Jin Intelligent Multimedia Lab
Generative Models Thus far we have essentially considered techniques that perform classification indirectly by modeling the training data, optimizing.
SISTA seminar Feb 28, 2002 Preoperative Prediction of Malignancy of Ovarian Tumors Using Least Squares Support Vector Machines C. Lu 1, T. Van Gestel 1,
ECG Signal processing (2)
Brief introduction on Logistic Regression
SVM - Support Vector Machines A new classification method for both linear and nonlinear data It uses a nonlinear mapping to transform the original training.
An Introduction of Support Vector Machine
Classification / Regression Support Vector Machines
Pattern Recognition and Machine Learning
Pattern Recognition and Machine Learning
An Introduction of Support Vector Machine
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Supervised Learning Recap
Transductive Reliability Estimation for Kernel Based Classifiers 1 Department of Computer Science, University of Ioannina, Greece 2 Faculty of Computer.
Chapter 4: Linear Models for Classification
Computer vision: models, learning and inference
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
PhD Hearing (Oct 15, 2003) Predictive Computer Models for Medical Classification Problems Predictive Computer Models for Medical Classification Problems.
Industrial Engineering College of Engineering Bayesian Kernel Methods for Binary Classification and Online Learning Problems Theodore Trafalis Workshop.
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
Pattern Recognition and Machine Learning
Classification and risk prediction
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Modeling Gene Interactions in Disease CS 686 Bioinformatics.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
PhD defense C. LU 25/01/ Probabilistic Machine Learning Approaches to Medical Classification Problems Probabilistic Machine Learning Approaches to.
UOG Journal Club: January 2013
Introduction to variable selection I Qi Yu. 2 Problems due to poor variable selection: Input dimension is too large; the curse of dimensionality problem.
Prediction model building and feature selection with SVM in breast cancer diagnosis Cheng-Lung Huang, Hung-Chang Liao, Mu- Chen Chen Expert Systems with.
A Comparative Study on Variable Selection for Nonlinear Classifiers C. Lu 1, T. Van Gestel 1, J. A. K. Suykens 1, S. Van Huffel 1, I. Vergote 2, D. Timmerman.
Reduced the 4-class classification problem into 6 pairwise binary classification problems, which yielded the conditional pairwise probability estimates.
Prediction of Malignancy of Ovarian Tumors Using Least Squares Support Vector Machines C. Lu 1, T. Van Gestel 1, J. A. K. Suykens 1, S. Van Huffel 1, I.
ICML2004, Banff, Alberta, Canada Learning Larger Margin Machine Locally and Globally Kaizhu Huang Haiqin Yang, Irwin King, Michael.
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
EMBC2001 Using Artificial Neural Networks to Predict Malignancy of Ovarian Tumors C. Lu 1, J. De Brabanter 1, S. Van Huffel 1, I. Vergote 2, D. Timmerman.
Learning Theory Reza Shadmehr Linear and quadratic decision boundaries Kernel estimates of density Missing data.
Jun-Won Suh Intelligent Electronic Systems Human and Systems Engineering Department of Electrical and Computer Engineering Speaker Verification System.
Machine Learning CUNY Graduate Center Lecture 4: Logistic Regression.
Christopher M. Bishop, Pattern Recognition and Machine Learning.
Multi-Speaker Modeling with Shared Prior Distributions and Model Structures for Bayesian Speech Synthesis Kei Hashimoto, Yoshihiko Nankaku, and Keiichi.
Support Vector Machines in Marketing Georgi Nalbantov MICC, Maastricht University.
University of Texas at Austin Machine Learning Group Department of Computer Sciences University of Texas at Austin Support Vector Machines.
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Additive Models , Trees , and Related Models Prof. Liqing Zhang Dept. Computer Science & Engineering, Shanghai Jiaotong University.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
Blackbox classifiers for preoperative discrimination between malignant and benign ovarian tumors C. Lu 1, T. Van Gestel 1, J. A. K. Suykens 1, S. Van Huffel.
Introduction Background Medical decision support systems based on patient data and expert knowledge A need to analyze the collected data in order to draw.
1 Kernel Machines A relatively new learning methodology (1992) derived from statistical learning theory. Became famous when it gave accuracy comparable.
Institute of Statistics and Decision Sciences In Defense of a Dissertation Submitted for the Degree of Doctor of Philosophy 26 July 2005 Regression Model.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
Canadian Bioinformatics Workshops
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Lecture 1.31 Criteria for optimal reception of radio signals.
Machine Learning – Classification David Fenyő
Sparse Kernel Machines
Glenn Fung, Murat Dundar, Bharat Rao and Jinbo Bi
Ch3: Model Building through Regression
Statistical Learning Dong Liu Dept. EEIS, USTC.
Pattern Recognition and Machine Learning
Robust Full Bayesian Learning for Neural Networks
Parametric Methods Berlin Chen, 2005 References:
Machine Learning with Clinical Data
Machine Learning – a Probabilistic Perspective
MAS 622J Course Project Classification of Affective States - GP Semi-Supervised Learning, SVM and kNN Hyungil Ahn
Presentation transcript:

AIME03, Oct 21, 2003 Classification of Ovarian Tumors Using Bayesian Least Squares Support Vector Machines C. Lu 1, T. Van Gestel 1, J. A. K. Suykens 1, S. Van Huffel 1, D. Timmerman 2, I. Vergote 2 1 Department of Electrical Engineering, Katholieke Universiteit Leuven, Leuven, Belgium, 2 Department of Obstetrics and Gynecology, University Hospitals Leuven, Leuven, Belgium

AIME03, Oct 21, 2003 Overview Introduction Data Bayesian least squares support vector machines (LS-SVMs) for classification LS-SVM classifier Bayesian evidence framework Input variable Selection Experiments Conclusions

AIME03, Oct 21, 2003 Introduction Problem ovarian masses: a common problem in gynecology. ovarian cancer : high mortality rate early detection of ovarian cancer is difficult treatment and management of different types of ovarian tumors differs greatly. develop a reliable diagnostic tool to preoperatively discriminate between benign and malignant tumors. assist clinicians in choosing the appropriate treatment. Preoperative medical diagnostic methods serum tumor maker: CA125 blood test transvaginal ultrasonography color Doppler imaging and blood flow indexing

AIME03, Oct 21, 2003 Logistic Regression Artificial neural networks Support Vector Machines Introduction Attempts to automate the diagnosis Risk of malignancy Index (RMI) (Jacobs et al) RMI= score morph × score meno × CA125 Methematical models Bayesian blief network Hybrid Methods Least Squares SVM Bayesian Framework

AIME03, Oct 21, 2003 Data Patient data collected at Univ. Hospitals Leuven, Belgium, 1994~ records (data with missing values were excluded), 25 features. 291 benign tumors, 134 (32%) malignant tumors Preprocessing: e.g. CA_125->log, Color_score {1,2,3,4} -> 3 design variables {0,1}.. Descriptive statistics

AIME03, Oct 21, 2003 Data Demographic, serum marker, color Doppler imaging and morphologic variables

AIME03, Oct 21, 2003 Data Patient data collected at Univ. Hospitals Leuven, Belgium, 1994~ records (data with missing values were excluded), 25 features. 291 benign tumors, 134 (32%) malignant tumors Preprocessing: e.g. CA_125->log, Color_score {1,2,3,4} -> 3 design variables {0,1}.. Descriptive statistics Visualization: Biplot

AIME03, Oct 21, 2003Data Fig. Biplot of Ovarian Tumor data. The observations are plotted as points (o - benign, x - malignant), the variables are plotted as vectors from the origin. - visualization of the correlation between the variables - visualization of the relations between the variables and clusters.

AIME03, Oct 21, 2003 Bayesian LS-SVM Classifiers Least square support vector machines (LS-SVM) for classification Kernel based method: Map the input data into a higher dimensional feature space x   (x) good generalization performance, unique solution, statistical learning theory

AIME03, Oct 21, 2003 Bayesian LS-SVM Classifiers LS-SVM classifier Given data D = {(x i, y i )} i=1,..,N, with binary targets y i = ±1(+1: malignant, -1: benign } solved in dual space

AIME03, Oct 21, 2003 Bayesian LS-SVM classifiers Integrate Bayesian evidence framework with LS-SVM Need of probabilistic framework Tune the regularization and kernel parameters To judge the uncertainty in predictions, which is critical in medical environment Maximizing the posterior probabilities of the models  marginalizing over the model parameters.

AIME03, Oct 21, 2003 Bayesian Inference Find the maximum a posteriori estimates of model parameters w MP and b MP, using conventional LS-SVM training The posterior probability of the parameters can be estimated via marginalization using Gaussian probability at w MP, b MP Assuming a uniform prior p(H j ) over all model, rank the model by the evidence p(D|H j ) evaluated using Gaussian approximation. Bayesian LS-SVM classifiers

AIME03, Oct 21, 2003 Bayesian LS-SVM classifiers Class probability for LS-SVM classifiers Conditional class probabilities computed using Gaussian distributions. Posterior class probability The probability of tumor being malignant p(y=+1|x,D,H) will be used for final classification (by thresholding). Cases with higher uncertainty can be rejected.

AIME03, Oct 21, 2003 Bayesian LS-SVM Classifiers Input variable selection Select the input variable according to model evidence p(D|H j ) Performs a forward selection (greedy search). Starting from zero variables, Iteratively select the variable which gives the greatest increase in the current model evidence. Stop the selection when addition of any remaining variables can no longer increase the model evidence.

AIME03, Oct 21, 2003 Experiments Performance evaluation Receiver operating characteristic (ROC) analysis Goal: high sensitivity for malignancy  low false positive rate. Providing probability of malignancy for individual ‘Temporal’ cross-validation Training set : 265 data (1994~1997). Test set: 160 data (1997~1999). Compared models Bayesian LS-SVM classifiers Bayesian MLPs : Linear discriminant analysis (LDA)

AIME03, Oct 21, 2003 Experiments – input variable selection Evolution of the model evidence 10 variables were selected based on the training set (first treated 265 patient data), using an RBF kernel.

AIME03, Oct 21, 2003 Model Evaluation Performance on Test Set: ROC curves

AIME03, Oct 21, 2003 Model Evaluation Performance on Test set * Probability cutoff value: 0.5 and 0.3

AIME03, Oct 21, 2003 Model Evaluation Performance (LS-SVM_RBF) on Test set with rejection based on  The rejected patients need further examination by human experts  Posterior probability essential for medical decision making

AIME03, Oct 21, 2003Conclusions Summary Within the Bayesian evidence framework, the hyperparameter tuning, input variable selection and computation of posterior class probability can be done in a unified way, without the need of selecting additional validation set. The proposed forward variable selection procedure which tries to maximize the model evidence can be used to identify the subset of important variables for model building. Posterior class probability enables us to assess the uncertainty in classification, important for medical decision making. Bayesian LS-SVMs have the potential to give reliable preoperative prediction of malignancy of ovarian tumors. Future work Application of the model to the multi-center data in a larger scale. Possibly further subclassify the tumors.