Results Comparison with existing approaches on benchmark datasets Evaluation on a uveal melanoma datasetEvaluation on the two-spiral dataset Evaluation.

Slides:



Advertisements
Similar presentations
The Robert Gordon University School of Engineering Dr. Mohamed Amish
Advertisements

Algorithms and applications
Original Figures for "Molecular Classification of Cancer: Class Discovery and Class Prediction by Gene Expression Monitoring"
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
Chapter Thirteen Conclusion: Where We Go From Here.
TECHNOLOGY GUIDE 4: Intelligent Systems
Outline Data with gaps clustering on the basis of neuro-fuzzy Kohonen network Adaptive algorithm for probabilistic fuzzy clustering Adaptive probabilistic.
Becerra-Fernandez, et al. -- Knowledge Management 1/e -- © 2004 Prentice Hall Chapter 7 Technologies to Manage Knowledge: Artificial Intelligence.
4 Intelligent Systems.
U NIVERSITY OF D ELAWARE C OMPUTER & I NFORMATION S CIENCES D EPARTMENT Mitigating the Compiler Optimization Phase- Ordering Problem using Machine Learning.
Active subgroup mining for descriptive induction tasks Dragan Gamberger Rudjer Bošković Instute, Zagreb Zdenko Sonicki University of Zagreb.
WRSTA, 13 August, 2006 Rough Sets in Hybrid Intelligent Systems For Breast Cancer Detection By Aboul Ella Hassanien Cairo University, Faculty of Computer.
Overfitting and Its Avoidance
Classification Dr Eamonn Keogh Computer Science & Engineering Department University of California - Riverside Riverside,CA Who.
Analysis of Classification-based Error Functions Mike Rimer Dr. Tony Martinez BYU Computer Science Dept. 18 March 2006.
Machine Learning Reading: Chapter Machine Learning and AI  Improve task performance through observation, teaching  Acquire knowledge automatically.
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
Introduction to Boosting Aristotelis Tsirigos SCLT seminar - NYU Computer Science.
The Pareto fitness genetic algorithm: Test function study Wei-Ming Chen
Boğaziçi University Artificial Intelligence Lab. Artificial Intelligence Laboratory Department of Computer Engineering Boğaziçi University Techniques for.
Kun Yi Li, Young Scholar Student, Quincy High School
Results Classification - Comparison with existing approaches on benchmark datasets Survival Analysis – International multi-centre double-blind study Survival.
Introduction to Data Mining Engineering Group in ACL.
Foundation of High-Dimensional Data Visualization
Data Mining By Andrie Suherman. Agenda Introduction Major Elements Steps/ Processes Tools used for data mining Advantages and Disadvantages.
Evaluating Classifiers
Attention Deficit Hyperactivity Disorder (ADHD) Student Classification Using Genetic Algorithm and Artificial Neural Network S. Yenaeng 1, S. Saelee 2.
Tree-Based Methods (V&R 9.1) Demeke Kasaw, Andreas Nguyen, Mariana Alvaro STAT 6601 Project.
A Simple Method to Extract Fuzzy Rules by Measure of Fuzziness Jieh-Ren Chang Nai-Jian Wang.
Department of Information Technology Indian Institute of Information Technology and Management Gwalior AASF hIQ 1 st Nov ‘09 Department of Information.
COMP3503 Intro to Inductive Modeling
An Introduction to Artificial Intelligence and Knowledge Engineering N. Kasabov, Foundations of Neural Networks, Fuzzy Systems, and Knowledge Engineering,
Project MLExAI Machine Learning Experiences in AI Ingrid Russell, University.
© Negnevitsky, Pearson Education, Will neural network work for my problem? Will neural network work for my problem? Character recognition neural.
Xuhua Xia Slide 1 MANOVA All statistical methods we have learned so far have only one continuous DV and one or more IVs which may be continuous or categorical.
GATree: Genetically Evolved Decision Trees 전자전기컴퓨터공학과 데이터베이스 연구실 G 김태종.
1 Machine Learning 1.Where does machine learning fit in computer science? 2.What is machine learning? 3.Where can machine learning be applied? 4.Should.
IE 585 Competitive Network – Learning Vector Quantization & Counterpropagation.
1 Universidad de Buenos Aires Maestría en Data Mining y Knowledge Discovery Aprendizaje Automático 5-Inducción de árboles de decisión (2/2) Eduardo Poggi.
Patch Based Prediction Techniques University of Houston By: Paul AMALAMAN From: UH-DMML Lab Director: Dr. Eick.
Artificial Intelligence, Expert Systems, and Neural Networks Group 10 Cameron Kinard Leaundre Zeno Heath Carley Megan Wiedmaier.
Applications of Supervised Learning in Bioinformatics Yen-Jen Oyang Dept. of Computer Science and Information Engineering.
Machine Learning Tutorial-2. Recall, Precision, F-measure, Accuracy Ch. 5.
Iowa State University Department of Computer Science Center for Computational Intelligence, Learning, and Discovery Harris T. Lin, Sanghack Lee, Ngot Bui.
Feature Extraction Artificial Intelligence Research Laboratory Bioinformatics and Computational Biology Program Computational Intelligence, Learning, and.
Data Mining and Decision Support
D Nagesh Kumar, IIScOptimization Methods: M8L5 1 Advanced Topics in Optimization Evolutionary Algorithms for Optimization and Search.
Introduction Background Medical decision support systems based on patient data and expert knowledge A need to analyze the collected data in order to draw.
Introduction to Classifiers Fujinaga. Bayes (optimal) Classifier (1) A priori probabilities: and Decision rule: given and decide if and probability of.
An Evolutionary Algorithm for Neural Network Learning using Direct Encoding Paul Batchis Department of Computer Science Rutgers University.
VISUALIZATION TECHNIQUES UTILIZING THE SENSITIVITY ANALYSIS OF MODELS Ivo Kondapaneni, Pavel Kordík, Pavel Slavík Department of Computer Science and Engineering,
Machine Learning Reading: Chapter Classification Learning Input: a set of attributes and values Output: discrete valued function Learning a continuous.
Big data classification using neural network
Machine Learning for Computer Security
Koichi Odajima & Yoichi Hayashi
Computational Intelligence: Methods and Applications
Rule Induction for Classification Using
CS 9633 Machine Learning Inductive-Analytical Methods
Erich Smith Coleman Platt
TECHNOLOGY GUIDE FOUR Intelligent Systems.
Machine Learning Dr. Mohamed Farouk.
What is Pattern Recognition?
Classifiers Fujinaga.
A Fast and Scalable Nearest Neighbor Based Classification
CS621: Artificial Intelligence Lecture 17: Feedforward network (lecture 16 was on Adaptive Hypermedia: Debraj, Kekin and Raunak) Pushpak Bhattacharyya.
Principal Component Analysis (PCA)
CSE 573 Introduction to Artificial Intelligence Decision Trees
Classifiers Fujinaga.
Using Bayesian Network in the Construction of a Bi-level Multi-classifier. A Case Study Using Intensive Care Unit Patients Data B. Sierra, N. Serrano,
Presentation transcript:

Results Comparison with existing approaches on benchmark datasets Evaluation on a uveal melanoma datasetEvaluation on the two-spiral dataset Evaluation on strabismus (squint) dataset Conclusions The evaluation of MERBIS on artificial datasets, real clinical datasets, and benchmark datasets has shown that the system produces comprehensible classifiers of low complexity, which perform well on unseen data. Acknowledgments This research has been supported by the Computer Science Department of the University of Liverpool. MERBIS - A Multi-Objective Evolutionary Rule Base Induction System C Setzkorn*, AFG Taktak, AC Fisher, BE Damato # and A Chandna Dept. of Clinical Engineering, St Paul’s Eye Unit #, Royal Liverpool Children’s NHS Trust ‡, RLBUHT Summary MERBIS is an artificial intelligence approach for classification which simulates the Darwinian principle of the survival of the fittest within a computer. MERBIS allows the extraction of simple, precise, and general classification models from existing data. The ‘fitness’ of a model is determined by its simplicity and precision. Introduction This research focuses upon the induction of classification models from data. Existing methods from disciplines such as statistics, machine learning and artificial neural networks require a great deal of expertise to generate models and to understand them. This can be disadvantageous, especially when the constructed models are used to formulate general hypotheses about the observed domain. To alleviate these problems, we induce fuzzy classification rule systems from datasets using a multi- objective evolutionary algorithm. Fuzzy classification rule systems correspond to a linguistic knowledge representation form that can easily be understood. Multi- objective evolutionary algorithms allow the optimisation of several objectives, such as the fit of the models to the dataset and their complexity. Methods Why Multi-Objective Evolutionary Algorithms? Advantages: Generate a set of trade-off solutions in a single run. Applicable to large and complex search spaces. Deals with incommensurable objectives. Unsusceptible to the shape of trade-off surface Alleviates over-fitting and reduces model complexity Disadvantages: Computational expensive. No guaranteed convergence. Classification using fuzzy rule systems – example iris dataset Validation The approach has been evaluated on artificial datasets and on real clinical data. In addition, it has been compared with existing approaches using benchmark datasets and 10-fold stratified cross-validation. Figure 2: Scatter diagrams of the iris dataset. Figure 3: A rule system generated for the iris data. Figure 4: Decision surfaces and contour plots of the rule system depicted in Figure 3. Decrease Complexity Increase area under receiver-operating curve Figure 1: Set of trade-off solutions. Red points are so- called non-dominated solutions (models) Table 1: Comparison of MERBIS with three existing approaches on several benchmark datasets. Figure 5: The generated model aims to predict the loss of an eye due to uveal melanoma. Decision surface (left) and contour plot (right) for a dataset related to uveal melanoma. Only two rules were used. The final model only deployed two out of the original twelve predictors. Figure 8: Scatter diagrams of the strabismus (squint) data. Figure 9: A rule system generated for the strabismus (squint) data. Figure 10: Decision surfaces and contour plots of the rule system depicted in Figure 9. Figure 6: Decision surface for the spiral dataset. The spiral dataset is a complicated artificial problem C linical E ngineering Royal Liverpool University Hospital The ‘iris dataset’ consists of hundred and fifty samples that represent flowers from the iris species setosa, versicolor, and virginica and is a classic benchmark problem. The dataset contains fifty samples from each flower species and each sample has four feature values: sepal length, sepal width, petal length and petal width. Strabismus problem: Differential diagnosis of squint into eight types (classes: 4 hypertropic and 4 hypotropic) using ten orthoptic measures (features) per patient.