Mining Several Databases with an Ensemble of Classifiers Seppo Puuronen Vagan Terziyan Alexander Logvinovsky 10th International Conference and Workshop.

Slides:



Advertisements
Similar presentations
Similarity Evaluation Techniques for Filtering Problems ? Vagan Terziyan University of Jyvaskyla
Advertisements

Mustafa Cayci INFS 795 An Evaluation on Feature Selection for Text Clustering.
Ensemble Learning Reading: R. Schapire, A brief introduction to boosting.
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Other Classification Techniques 1.Nearest Neighbor Classifiers 2.Support Vector Machines.
Data Mining Classification: Alternative Techniques
Data Mining Classification: Alternative Techniques
Application of Stacked Generalization to a Protein Localization Prediction Task Melissa K. Carroll, M.S. and Sung-Hyuk Cha, Ph.D. Pace University, School.
Navneet Goyal. Instance Based Learning  Rote Classifier  K- nearest neighbors (K-NN)  Case Based Resoning (CBR)
TLI371 – Distributed Artificial Intelligence in Mobile Environment Course Introduction Vagan Terziyan Department of Mathematical Information Technology.
1 MetaPetrinets brief introduction to the concept Vagan Terziyan University of Jyvaskyla, Finland
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Knowledge Management Challenges in Knowledge Discovery Systems Mykola Pechenizkiy, Seppo Puuronen Department of Computer Science University of Jyväskylä.
Jierui Xie, Boleslaw Szymanski, Mohammed J. Zaki Department of Computer Science Rensselaer Polytechnic Institute Troy, NY 12180, USA {xiej2, szymansk,
A Similarity Evaluation Technique for Data Mining with Ensemble of Classifiers Seppo Puuronen, Vagan Terziyan International Workshop on Similarity Search.
An Interval Approach to Discover Knowledge from Multiple Fuzzy Estimations Vagan Terziyan * & **, Seppo Puuronen **, Helen Kaikova * *Department of Artificial.
AAECC’99 Honolulu November 1999 Flexible Arithmetic for Huge Numbers with Recursive Series of Operations Vagan Terziyan*, Alexey Tsymbal**, Seppo Puuronen**
A Technique for Advanced Dynamic Integration of Multiple Classifiers Alexey Tsymbal*, Seppo Puuronen**, Vagan Terziyan* *Department of Artificial Intelligence.
University of Athens, Greece Pervasive Computing Research Group Predicting the Location of Mobile Users: A Machine Learning Approach 1 University of Athens,
Dynamic Integration of Virtual Predictors Vagan Terziyan Information Technology Research Institute, University of Jyvaskyla, FINLAND
CES 514 – Data Mining Lec 9 April 14 Mid-term k nearest neighbor.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
1 Ensembles of Nearest Neighbor Forecasts Dragomir Yankov, Eamonn Keogh Dept. of Computer Science & Eng. University of California Riverside Dennis DeCoste.
A Similarity Evaluation Technique for Cooperative Problem Solving with a Group of Agents Seppo Puuronen, Vagan Terziyan Third International Workshop CIA-99.
Intelligent Web Applications (Part 1) Course Introduction Vagan Terziyan AI Department, Kharkov National University of Radioelectronics / MIT Department,
1 The Law of Semantic Balance brief introduction of the concept Vagan Terziyan University of Jyvaskyla, Finland
CS Instance Based Learning1 Instance Based Learning.
1 A Semantic Metanetwork Vagan Terziyan University of Jyvaskyla, Finland
On Comparing Classifiers: Pitfalls to Avoid and Recommended Approach Published by Steven L. Salzberg Presented by Prakash Tilwani MACS 598 April 25 th.
3 ème Journée Doctorale G&E, Bordeaux, Mars 2015 Wei FENG Geo-Resources and Environment Lab, Bordeaux INP (Bordeaux Institute of Technology), France Supervisor:
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 01: Training, Testing, and Tuning Datasets.
Spatial Statistics and Spatial Knowledge Discovery First law of geography [Tobler]: Everything is related to everything, but nearby things are more related.
Cost-Sensitive Bayesian Network algorithm Introduction: Machine learning algorithms are becoming an increasingly important area for research and application.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
We introduce the use of Confidence c as a weighted vote for the voting machine to avoid low confidence Result r of individual expert from affecting the.
Evaluating Hypotheses Reading: Coursepack: Learning From Examples, Section 4 (pp )
Page 1 Ming Ji Department of Computer Science University of Illinois at Urbana-Champaign.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology A data mining approach to the prediction of corporate failure.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Ensemble Methods: Bagging and Boosting
Achieving High Software Reliability Using a Faster, Easier and Cheaper Method NASA OSMA SAS '01 September 5-7, 2001 Taghi M. Khoshgoftaar The Software.
Institut für Softwarewissenschaft - Universität WienP.Brezany 1 Meta-Learning in Distributed Datamining Systems Peter Brezany Institut für Softwarewissenschaft.
CLASSIFICATION: Ensemble Methods
Stefan Mutter, Mark Hall, Eibe Frank University of Freiburg, Germany University of Waikato, New Zealand The 17th Australian Joint Conference on Artificial.
Ensemble with Neighbor Rules Voting Itt Romneeyangkurn, Sukree Sinthupinyo Faculty of Computer Science Thammasat University.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Fuzzy integration of structure adaptive SOMs for web content.
Feature Extraction Artificial Intelligence Research Laboratory Bioinformatics and Computational Biology Program Computational Intelligence, Learning, and.
Classification Ensemble Methods 1
1 January 24, 2016Data Mining: Concepts and Techniques 1 Data Mining: Concepts and Techniques — Chapter 7 — Classification Ensemble Learning.
… Algo 1 Algo 2 Algo 3 Algo N Meta-Learning Algo.

CS Machine Learning Instance Based Learning (Adapted from various sources)
Classification Cheng Lei Department of Electrical and Computer Engineering University of Victoria April 24, 2015.
Max-Confidence Boosting With Uncertainty for Visual tracking WEN GUO, LIANGLIANG CAO, TONY X. HAN, SHUICHENG YAN AND CHANGSHENG XU IEEE TRANSACTIONS ON.
Reading: R. Schapire, A brief introduction to boosting
Presented by Khawar Shakeel
Can-CSC-GBE: Developing Cost-sensitive Classifier with Gentleboost Ensemble for breast cancer classification using protein amino acids and imbalanced data.
Classification Nearest Neighbor
DEFECT PREDICTION : USING MACHINE LEARNING
Instance Based Learning (Adapted from various sources)
Combining Base Learners
Data Mining Practical Machine Learning Tools and Techniques
Classification Nearest Neighbor
Prepared by: Mahmoud Rafeek Al-Farra
COSC 4335: Other Classification Techniques
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
CS539 Project Report -- Evaluating hypothesis
CSE4334/5334 Data Mining Lecture 7: Classification (4)
Presentation transcript:

Mining Several Databases with an Ensemble of Classifiers Seppo Puuronen Vagan Terziyan Alexander Logvinovsky 10th International Conference and Workshop on Database and Expert Systems Applications August 30 - September 3, 1999 Florence, Italy DEXA-99

Authors Department of Computer Science and Information Systems University of Jyvaskyla FINLAND Seppo PuuronenVagan Terziyan Department of Artificial Intelligence Kharkov State Technical University of Radioelectronics, UKRAINE Alexander Logvinovsky Department of Artificial Intelligence Kharkov State Technical University of Radioelectronics, UKRAINE

Contents The problem of “multiclassifiers” - “multidatabase” mining; Case “One Database - Many Classifiers”; Dynamic integration of classifiers; Case “One Classifier - Many Databases”; Weighting databases; Case “Many Databases - Many Classifiers”; Context-based trend within the classifiers predictions and decontextualization; Conclusion

Introduction

Introduction

Problem

Problem

Case ONE:ONE

Case ONE:MANY

Dynamic Integration of Classifiers Final classification is made by weighted voting of classifiers from the ensemble; Weights of classifiers are recalculated for every new instance; Weighting is based on predicted errors of the classifiers in the neighborhood area of the instance

Sliding Exam of a Classifier (Predictor, Interpolator) Remove an instance y(x i ) from training set; Use a classifier to derive prediction result y’(x i ); Evaluate difference  as distance between real and predicted values Continue for every instance

Brief Review of Distance Functions According to D. Wilson and T. Martinez (1997)

PEBLS Distance Evaluation for Nominal Values (According to ) PEBLS Distance Evaluation for Nominal Values (According to Cost S. and Salzberg S., 1993 ) The distance d i between two values v 1 and v 2 for certain instance is: where C 1 and C 2 are the numbers of instances in the training set with selected values v 1 and v 2, C 1i and C 2i are the numbers of instances from the i-th class, where the values v 1 and v 2 were selected, and k is the number of classes of instances

Interpolation of Error Function Based on Hypothesis of Compactness | x - x i | <  (   0)  |  (x) -  (x i ) |  0

Competence map absolute difference  weight function 

Solution for ONE:MANY

Case MANY:ONE

Integration of Databases Final classification of an instance is obtained by weighted voting of predictions made by the classifier for every database separately; Weighting is based on taking the integral of the error function of the classifier across every database

Integral Weight of Classifier Integral Weight of Classifier

Solution for MANY:ONE

Case MANY:MANY

Weighting Classifiers and Databases Prediction and weight of a databasePrediction and weight of a classifier

Solutions for MANY:MANY

Decontextualization of Predictions Sometimes actual value cannot be predicted as weighted mean of individual predictions of classifiers from the ensemble; It means that the actual value is outside the area of predictions; It happens if classifiers are effected by the same type of a context with different power; It results to a trend among predictions from the less powerful context to the most powerful one; In this case actual value can be obtained as the result of “decontextualization” of the individual predictions

Neighbor Context Trend x prediction in (1,2) neighbor context: “worse context” prediction in (1,2,3) neighbor context: “better context” actual value: “ideal context” y xixi y(x i ) y + (x i ) y - (x i )

Main Decontextalization Formula y Y y - - prediction in worse context y + - prediction in better context y ’ - decontextualized prediction y - actual value y’ y+y+ y-y- ’’ ++ --  ’ = - ·+- ·+- ·+- ·+  - +  +  ’ <  - ;  ’ <  +  + <  -

Decontextualization One level decontextualization All subcontexts decontextualization Decontextualized difference  New sample classification

Physical Interpretation of Decontextualization R1R1 R2R2 R res actual value decontextualized value predicted values Uncertainty is like a “resistance” for precise prediction actual value y+y+ y-y- y’y’ y y y i- - prediction in worse context y + - prediction in better context y ’ - decontextualized prediction y - actual value

Conclusion Dynamic integration of classifiers based on locally adaptive weights of classifiers allows to handle the case «One Dataset - Many Classifiers»; Integration of databases based on their integral weights relatively to the classification accuracy allows to handle the case «One Classifier - Many Datasets»; Successive or parallel application of the two abowe algorithms allows a variety of solutions for the case «Many Classifiers - Many Datasets»; Decontextualization as the opposite to weighted voting way of integration of classifiers allows to handle context of classification in the case of a trend