Lower-Bound Estimate for Cost-sensitive Decision Trees Mikhail Goubko, senior researcher Trapeznikov Institute of Control Sciences of the Russian Academy.

Slides:



Advertisements
Similar presentations
Lindsey Bleimes Charlie Garrod Adam Meyerson
Advertisements

An Introduction to Artificial Intelligence
Chapter 11 Limitations of Algorithm Power Copyright © 2007 Pearson Addison-Wesley. All rights reserved.
Fast Algorithms For Hierarchical Range Histogram Constructions
Decision Tree Approach in Data Mining
Data Mining Classification: Alternative Techniques
CS171 Introduction to Computer Science II Graphs Strike Back.
Network Correlated Data Gathering With Explicit Communication: NP- Completeness and Algorithms R˘azvan Cristescu, Member, IEEE, Baltasar Beferull-Lozano,
Computational problems, algorithms, runtime, hardness
Author: David He, Astghik Babayan, Andrew Kusiak By: Carl Haehl Date: 11/18/09.
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Fall, 2001 Lecture 1 (Part 3) Tuesday, 9/4/01 Greedy Algorithms.
The Theory of NP-Completeness
Ensemble Learning: An Introduction
Visual Querying By Color Perceptive Regions Alberto del Bimbo, M. Mugnaini, P. Pala, and F. Turco University of Florence, Italy Pattern Recognition, 1998.
SubSea: An Efficient Heuristic Algorithm for Subgraph Isomorphism Vladimir Lipets Ben-Gurion University of the Negev Joint work with Prof. Ehud Gudes.
1 MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING By Kaan Tariman M.S. in Computer Science CSCI 8810 Course Project.
Chapter 11: Limitations of Algorithmic Power
UMass Lowell Computer Science Analysis of Algorithms Prof. Karen Daniels Spring, 2002 Lecture 1 (Part 3) Tuesday, 1/29/02 Design Patterns for Optimization.
Chapter 11 Limitations of Algorithm Power Copyright © 2007 Pearson Addison-Wesley. All rights reserved.
Heterogeneous and Grid Computing2 Communication models u Modeling the performance of communications –Huge area –Two main communities »Network designers.
Delay Efficient Sleep Scheduling in Wireless Sensor Networks Gang Lu, Narayanan Sadagopan, Bhaskar Krishnamachari, Anish Goel Presented by Boangoat(Bea)
Learning Chapter 18 and Parts of Chapter 20
1.1 Chapter 1: Introduction What is the course all about? Problems, instances and algorithms Running time v.s. computational complexity General description.
Operations Research Models
Chapter 11 Limitations of Algorithm Power. Lower Bounds Lower bound: an estimate on a minimum amount of work needed to solve a given problem Examples:
A. Levitin “Introduction to the Design & Analysis of Algorithms,” 3rd ed., Ch. 11 ©2012 Pearson Education, Inc. Upper Saddle River, NJ. All Rights Reserved.
Bold Stroke January 13, 2003 Advanced Algorithms CS 539/441 OR In Search Of Efficient General Solutions Joe Hoffert
Computational Complexity Polynomial time O(n k ) input size n, k constant Tractable problems solvable in polynomial time(Opposite Intractable) Ex: sorting,
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
07/21/2005 Senmetrics1 Xin Liu Computer Science Department University of California, Davis Joint work with P. Mohapatra On the Deployment of Wireless Sensor.
Complexity Classes (Ch. 34) The class P: class of problems that can be solved in time that is polynomial in the size of the input, n. if input size is.
Stochastic Algorithms Some of the fastest known algorithms for certain tasks rely on chance Stochastic/Randomized Algorithms Two common variations – Monte.
Introduction to Data Mining Group Members: Karim C. El-Khazen Pascal Suria Lin Gui Philsou Lee Xiaoting Niu.
Time Series Data Analysis - I Yaji Sripada. Dept. of Computing Science, University of Aberdeen2 In this lecture you learn What are Time Series? How to.
Inferring Decision Trees Using the Minimum Description Length Principle J. R. Quinlan and R. L. Rivest Information and Computation 80, , 1989.
Major objective of this course is: Design and analysis of modern algorithms Different variants Accuracy Efficiency Comparing efficiencies Motivation thinking.
1 Learning Chapter 18 and Parts of Chapter 20 AI systems are complex and may have many parameters. It is impractical and often impossible to encode all.
1 Lower Bounds Lower bound: an estimate on a minimum amount of work needed to solve a given problem Examples: b number of comparisons needed to find the.
1Computer Sciences Department. Book: Introduction to Algorithms, by: Thomas H. Cormen Charles E. Leiserson Ronald L. Rivest Clifford Stein Electronic:
1 Short Term Scheduling. 2  Planning horizon is short  Multiple unique jobs (tasks) with varying processing times and due dates  Multiple unique jobs.
The Traveling Salesman Problem Over Seventy Years of Research, and a Million in Cash Presented by Vladimir Coxall.
MACHINE LEARNING 10 Decision Trees. Motivation  Parametric Estimation  Assume model for class probability or regression  Estimate parameters from all.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Data Mining Practical Machine Learning Tools and Techniques By I. H. Witten, E. Frank and M. A. Hall Chapter 6.2: Classification Rules Rodney Nielsen Many.
CSE 589 Part V One of the symptoms of an approaching nervous breakdown is the belief that one’s work is terribly important. Bertrand Russell.
CPS Computational problems, algorithms, runtime, hardness (a ridiculously brief introduction to theoretical computer science) Vincent Conitzer.
A* optimality proof, cycle checking CPSC 322 – Search 5 Textbook § 3.6 and January 21, 2011 Taught by Mike Chiang.
CES 592 Theory of Software Systems B. Ravikumar (Ravi) Office: 124 Darwin Hall.
Data Mining By Farzana Forhad CS 157B. Agenda Decision Tree and ID3 Rough Set Theory Clustering.
Onlinedeeneislam.blogspot.com1 Design and Analysis of Algorithms Slide # 1 Download From
Fast and Efficient Static Compaction of Test Sequences Based on Greedy Algorithms Jaan Raik, Artur Jutman, Raimund Ubar Tallinn Technical University, Estonia.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Data Mining CH6 Implementation: Real machine learning schemes(2) Reporter: H.C. Tsai.
Mustafa Gokce Baydogan, George Runger and Eugene Tuv INFORMS Annual Meeting 2011, Charlotte A Bag-of-Features Framework for Time Series Classification.
Behavior Recognition Based on Machine Learning Algorithms for a Wireless Canine Machine Interface Students: Avichay Ben Naim Lucie Levy 14 May, 2014 Ort.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Advanced Algorithms Analysis and Design
CS 9633 Machine Learning Support Vector Machines
Summary of lectures Introduction to Algorithm Analysis and Design (Chapter 1-3). Lecture Slides Recurrence and Master Theorem (Chapter 4). Lecture Slides.
RE-Tree: An Efficient Index Structure for Regular Expressions
Design and Analysis of Computer Algorithm (CS575-01)
Analysis and design of algorithm
Objective of This Course
Chapter 3: The Efficiency of Algorithms
Approximation Algorithms
Chapter 11 Limitations of Algorithm Power
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
CS639: Data Management for Data Science
INTRODUCTION TO ALOGORITHM DESIGN STRATEGIES
Presentation transcript:

Lower-Bound Estimate for Cost-sensitive Decision Trees Mikhail Goubko, senior researcher Trapeznikov Institute of Control Sciences of the Russian Academy of Sciences, Moscow, Russia 18 th IFAC World Congress, Milan, August 31, 2011

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 from Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Decision tree – a popular classification tool for: machine learning pattern recognition fault detection medical diagnostics situational control Decision is made from a series of tests of attributes The next attribute tested depends on the results of the previous tests Decision trees are learned from data Compact trees are good Expected length of the path in a tree is the most popular measure of tree size

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Set D of decisions chest cold influenza healthy Set of cases N Set of attributes M No High No High No High No High No High Yes High No Low Yes Low No Low Cough t0t0 t0t0 Chronic illness No Yes No Yes No Yes No Wheezing 11 22 33 44 55 66 77  nn

S U M M A R Y Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Set D of decisions chest cold influenza healthy Set of cases N Set of attributes M No Yes No Yes No High Low Cough t0t0 t0t0 Chronic illness No Yes No Yes No Yes No Wheezing 11 22 33 44 55 66 77  nn

S U M M A R Y Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Set D of decisions chest cold influenza healthy Set of cases N Set of attributes M No Yes No Yes No Yes No Yes No 11 22 33 44 55 66 77  nn No Yes No High Low

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Set D of decisions chest cold influenza healthy Set of cases N Set of attributes M No Yes Cough t0t0 t0t0 Chronic illness Yes No Wheezing 11 22 33 44 55 66 77  nn No Yes No High Low

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Tests differ in measurement costs (e.g., a general blood analysis is much cheaper than a computer tomography procedure) The decision tree is grown to minimize the expected cost of classification given zero misclassification rate on a learning set Other types of costs relevant to decision tree growing (Turney, 2000), i.e. misclassification, teaching, intervention costs, are not considered Measurement costs (see Turney, 2000) may depend on: individual case true class of the case side-effects (value of hidden attributes) prior tests prior results (other attributes’ values) correct answer of the current question Studied! Covered! Can be accounted!

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Hyafil and Rivest (1976), and also Zantema and Bodlaender (2000) have shown the decision tree growing problem to be NP-hard Sieling (2008) has shown that the size of an optimal tree is hard to approximate up to any constant factor Quinlan (1979) developed an information-gain-based heuristics ID3. Now it is commercial and has version 5 Heuristic algorithms for cost-sensitive tree construction: CS-ID3 - Tan (1993) IDX - Norton (1989) EG2 - Núñez (1991) Lower-bound estimates for cost-sensitive decision trees: Entropy-based estimate - Ohta and Kanaya (1991) Huffman code length - Biasizzo, Žužek, and Novak (1998) Combinatory estimate - Bessiere, Hebrard, and O’Sullivan (2009) Hyafil and Rivest (1976), and also Zantema and Bodlaender (2000) have shown the decision tree growing problem to be NP-hard Sieling (2008) has shown that the size of an optimal tree is hard to approximate up to any constant factor Quinlan (1979) developed an information-gain-based heuristics ID3. Now it is commercial and has version 5 Heuristic algorithms for cost-sensitive tree construction: CS-ID3 - Tan (1993) IDX - Norton (1989) EG2 - Núñez (1991) Lower-bound estimates for cost-sensitive decision trees: Entropy-based estimate - Ohta and Kanaya (1991) Huffman code length - Biasizzo, Žužek, and Novak (1998) Combinatory estimate - Bessiere, Hebrard, and O’Sullivan (2009)

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Definition 1. A subset of tests Q  M isolates case w in subset of cases S  N ( w  S ) iff sequence of tests Q assures proper decision f(w) given initial uncertainty S and w is the real state of the world. Definition 2. Optimal set of questions Q(w, S)  M is the cheapest of the sets of questions that isolate case w in set S ; Define also minimum cost The lower-bound estimate Unlike known estimates it performs well when: 1)the number of classes is small compared to that of examples 2)there is a small number of examples

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps The problem of unknown case classification is replaced by the problem of proving the true case to a third party. What is the true case? Test this I know the true case! Try testing this Initial problem Simplified problem

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Average time  n 2 m, ( n - the number of cases, m - the number of tests) Calculation of the estimate is reduced to a number of set-covering problems and is NP-hard in the worst case. But experiments show nice performance for the estimate and its linear programming relaxation.

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps 1. Use the lower-bound estimate to evaluate extra costs due to imperfectness of heuristic tree growing algorithms The quality of the lower-bound estimate on real data sets The quality of the estimate varies from 60 to 95% depending on the data set

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Comparing new algorithms with known heuristics (IDX, CS-ID3, EG2) on different data sets (tree costs are shown) 2. Use the lower-bound estimate to build new tree growing algorithms new algorithms perform better on small data sets new algorithms work worse in the presence of “dummy” tests adjacency of results shows that heuristic trees are nearly optimal ones

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps General methods of hierarchy optimization General methods of hierarchy optimization

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Cost function The problem is to find Allowed arbitrary number of layers asymmetric hierarchies multiple subordination several top nodes Not allowed cycles unconnected parts subordinating to the “worker” nodes (black)

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Sectional cost function 1. Covers most of applied problems of hierarchy optimization 2. Analytical methods when the tallest or the flattest hierarchy is optimal when an optimal tree exists when an optimal hierarchy has the shape of a conveyor 3. Algorithms optimal hierarchy search optimal tree search Build an optimal conveyor-belt hierarchy 4. The general problem of hierarchy optimization is complex Homogenous cost function 1. Also has numerous applications 2. Closed-form solution for the optimal hierarchy problem 3. Efficient algorithms of nearly-optimal tree construction

S U M M A R Y 1.New lower-bound estimate is suggested for the decision tree with case-dependent test costs 2.Unlike known estimates it performs well when the number of classes is small 3.Estimate calculation average performance is n 2  m operations for n examples and m tests 4.Use the estimate to evaluate absolute losses of heuristic decision tree algorithms 5.Use the estimate in split criteria of greedy top-down algorithms of decision tree construction 6.Experiments on real data sets show these algorithms to give comparable results with popular cost-sensitive heuristics – IDX, CS-ID3, EG2 7.Algorithms suggested perform better on small data sets with lack of tests Mikhail Goubko Lower-Bound Estimate for Cost-sensitive Decision Trees IFAC’2011, Milan, August 31 Introduction Decision trees Costs Results Motivation Literature Definitions Calculation Applications Bonus: Hierarchy Optimization Bonus: Hierarchy Optimization Model Methods Apps Applications of hierarchy optimization methods Manufacturing planning (assembly line balancing) Networks design communication and computing networks data collection networks structure of hierarchical cellular networks Computational mathematics optimal coding structure of algorithms real-time computation and aggregation hierarchical parallel computing User interfaces design optimizing hierarchical menus building compact and informative taxonomies Data mining decision trees growing structuring database indexes Organization design org. chart re-engineering theoretical models of a hierarchical firm