COMPUTER AIDED DIAGNOSIS: CLASSIFICATION Prof. Yasser Mostafa Kadah – www.k-space.org.

Slides:



Advertisements
Similar presentations
Computer Aided Diagnosis: Feature Extraction
Advertisements

Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Notes Sample vs distribution “m” vs “µ” and “s” vs “σ” Bias/Variance Bias: Measures how much the learnt model is wrong disregarding noise Variance: Measures.
CSCE555 Bioinformatics Lecture 15 classification for microarray data Meeting: MW 4:00PM-5:15PM SWGN2A21 Instructor: Dr. Jianjun Hu Course page:
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Machine learning continued Image source:
LECTURE 11: BAYESIAN PARAMETER ESTIMATION
COMPUTER AIDED DIAGNOSIS: FEATURE SELECTION Prof. Yasser Mostafa Kadah –
Chapter 4: Linear Models for Classification
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Active Learning with Support Vector Machines
Anomaly Detection. Anomaly/Outlier Detection  What are anomalies/outliers? The set of data points that are considerably different than the remainder.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation X = {
MACHINE LEARNING 6. Multivariate Methods 1. Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Motivating Example  Loan.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Thanks to Nir Friedman, HU
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Jeff Howbert Introduction to Machine Learning Winter Classification Bayesian Classifiers.
METU Informatics Institute Min 720 Pattern Classification with Bio-Medical Applications PART 2: Statistical Pattern Classification: Optimal Classification.
Exercise Session 10 – Image Categorization
Methods in Medical Image Analysis Statistics of Pattern Recognition: Classification and Clustering Some content provided by Milos Hauskrecht, University.
This week: overview on pattern recognition (related to machine learning)
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
An Introduction to Support Vector Machine (SVM) Presenter : Ahey Date : 2007/07/20 The slides are based on lecture notes of Prof. 林智仁 and Daniel Yeung.
Image Classification 영상분류
CHAPTER 7: Clustering Eick: K-Means and EM (modified Alpaydin transparencies and new transparencies added) Last updated: February 25, 2014.
Learning Theory Reza Shadmehr Linear and quadratic decision boundaries Kernel estimates of density Missing data.
1 E. Fatemizadeh Statistical Pattern Recognition.
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN Modified by Prof. Carolina Ruiz © The MIT Press, 2014 for CS539 Machine Learning at WPI
1Ellen L. Walker Category Recognition Associating information extracted from images with categories (classes) of objects Requires prior knowledge about.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Introduction to Machine Learning Multivariate Methods 姓名 : 李政軒.
CSSE463: Image Recognition Day 11 Due: Due: Written assignment 1 tomorrow, 4:00 pm Written assignment 1 tomorrow, 4:00 pm Start thinking about term project.
Bayesian decision theory: A framework for making decisions when uncertainty exit 1 Lecture Notes for E Alpaydın 2010 Introduction to Machine Learning 2e.
Machine Learning ICS 178 Instructor: Max Welling Supervised Learning.
METU Informatics Institute Min720 Pattern Classification with Bio-Medical Applications Part 9: Review.
The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Classification COMP Seminar BCB 713 Module Spring 2011.
Computer Vision Lecture 7 Classifiers. Computer Vision, Lecture 6 Oleh Tretiak © 2005Slide 1 This Lecture Bayesian decision theory (22.1, 22.2) –General.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Classification of tissues and samples 指導老師:藍清隆 演講者:張許恩、王人禾.
High resolution product by SVM. L’Aquila experience and prospects for the validation site R. Anniballe DIET- Sapienza University of Rome.
CS479/679 Pattern Recognition Dr. George Bebis
Semi-Supervised Clustering
Machine Learning Clustering: K-means Supervised Learning
A Personal Tour of Machine Learning and Its Applications
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
School of Computer Science & Engineering
Basic machine learning background with Python scikit-learn
Machine Learning Basics
Outline Parameter estimation – continued Non-parametric methods.
Prepared by: Mahmoud Rafeek Al-Farra
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Pattern Recognition and Image Analysis
Generally Discriminant Analysis
Support Vector Machines
LECTURE 07: BAYESIAN ESTIMATION
Other Classification Models: Support Vector Machine (SVM)
Parametric Methods Berlin Chen, 2005 References:
Multivariate Methods Berlin Chen
Mathematical Foundations of BME
Multivariate Methods Berlin Chen, 2005 References:
Machine Learning – a Probabilistic Perspective
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
ECE – Pattern Recognition Lecture 10 – Nonparametric Density Estimation – k-nearest-neighbor (kNN) Hairong Qi, Gonzalez Family Professor Electrical.
Presentation transcript:

COMPUTER AIDED DIAGNOSIS: CLASSIFICATION Prof. Yasser Mostafa Kadah –

Recommended Reference  Recent Advances in Breast Imaging, Mammography, and Computer-Aided Diagnosis of Breast Cancer, Editors: Jasjit S. Suri and Rangaraj M. Rangayyan, SPIE Press, 2006.

CAD

Classification Assumptions  There is a visible difference between normal and abnormal images  If you cannot see consistent differences, you cannot program a CAD system to see them  Training set with normal and abnormal cases of interest  Features selected describe such difference effectively  Irrelevant features will only confuse and misguide the CAD system  Normal and abnormal cases form distinct clusters that are somewhat apart according to a distance measure  Cases from the same pathology are represented by points in feature space that with smaller distance separation than with cases from other pathologies  “Intra-cluster” distance is significantly smaller than “Inter-cluster” distance

Classifiers: Model  Parametric classification  Assumed a certain distribution for data clusters (e.g., Gaussian)  Estimates model parameters from the data  Uses this a priori information to design the classification method and estimate its parameters  Good if model is correct but bad if not (difficult to know ahead)  Non-Parametric classification  Does not assume or impose any model on the data  “Model-Free” or “Data-Dependent”

Classifiers: Learning  Supervised classifiers  Classifiers are trained with data samples of known labels  Number of clusters known a priori and cannot be changed  Relies on “trainer” to provide the correct information  Unsupervised classifiers  Classifiers are trained with data samples with unknown labels  Discover underlying clusters according to particular criteria  Interesting to identify new clusters representing sub-classes within large normal or pathological classes

Minimum Distance Classifier  Idea: Compute the distance between unknown case and the center of known clusters and assign the case to the closest cluster  How to compute cluster centers  Distance computation  Simple but not good in many cases

K-Voting Nearest Neighbor  Assumption: an unknown case is likely to belong to the class of its neighbors in space  Idea: Find the K neighbors of the unknown case and make a majority vote among them and assign the case to the pathology of the majority  Select K in such a way to avoid “ties”: e.g., odd number for binary classification  Special case: K=1 – Nearest Neighbor Classifier  Matlab function “knnclassify”

Naïve Bayesian Classifier  Idea: Assign the unknown case to the class of largest posterior probability  P(y) is the a priori probability of y  P(x | y) is the likelihood function;  P(x) is the marginal probability  P(y | x) is the a posteriori probability Classified as Normal Classified as Abnormal Normal Group Abnormal Group Poster Probability

Discriminant Analysis  Idea: Find the best separation surface between clusters  Generalization of Naïve Bayesian classifier  Linear: LDA  Quadratic: QDA  Matlab function “classify”

Other Important Classifiers  Support vector machine (SVM) classifier  Neural Network classifier  Fuzzy logic classifier  K-means clustering  Biclustering Subjects Features

Things to Watch for  Feature normalization is important and can improve training and classification results  Independent training and testing data is a must  Half and half for example  Leave-one-out cross validation can be used for small data sets

Assignments  Apply 2 classification methods to the selected features obtained from previous tasks and compare their results