LEARNING MODEL FOR CONSTRUCTION OF THE BEST DECISION SEQUENCE USING PRIOR KNOWLEDGE Thesis Presentation by Lilit Yenokyan.

Slides:



Advertisements
Similar presentations
FUNCTION FITTING Student’s name: Ruba Eyal Salman Supervisor:
Advertisements

Algorithm Design Techniques
I-DMSS for Bus Rental in Seoul, Korea Katta G. Murty Dept. IOE, U. of Michigan Ann Arbor, MI , USA and Woo-Je Kim Dept.
January 23 rd, Document classification task We are interested to solve a task of Text Classification, i.e. to automatically assign a given document.
Traffic assignment.
Steps of a Design Brief V Design Brief  Problem, identification, and definition Establish a clear idea of what is to be accomplished. Identify.
Problems and Their Classes
Shortest Paths (1/11)  In this section, we shall study the path problems such like  Is there a path from city A to city B?  If there is more than one.
13-Optimization Assoc.Prof.Dr. Ahmet Zafer Şenalp Mechanical Engineering Department Gebze Technical.
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Decision Tree Approach in Data Mining
Types of Algorithms.
Greed is good. (Some of the time)
Algorithms + L. Grewe.
Chapter 3 The Greedy Method 3.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Neural Networks My name is Burleson. Neural Networks vs Conventional Computing Programming is broken into small, unambiguous steps Algorithms must be.
Neural Networks Basic concepts ArchitectureOperation.
3 -1 Chapter 3 The Greedy Method 3 -2 The greedy method Suppose that a problem can be solved by a sequence of decisions. The greedy method has that each.
Brent Dingle Marco A. Morales Texas A&M University, Spring 2002
Basic Data Mining Techniques
Data Mining with Decision Trees Lutz Hamel Dept. of Computer Science and Statistics University of Rhode Island.
(Page 554 – 564) Ping Perez CS 147 Summer 2001 Alternative Parallel Architectures  Dataflow  Systolic arrays  Neural networks.
Backtracking.
The Shortest Path Problem
NetworkModel-1 Network Optimization Models. NetworkModel-2 Network Terminology A network consists of a set of nodes and arcs. The arcs may have some flow.
Rohit Ray ESE 251. What are Artificial Neural Networks? ANN are inspired by models of the biological nervous systems such as the brain Novel structure.
6 Resource Utilization Teaching Strategies
1 Shortest Path Calculations in Graphs Prof. S. M. Lee Department of Computer Science.
Basic Data Mining Techniques
West Virginia University
CHAPTER 12 ADVANCED INTELLIGENT SYSTEMS © 2005 Prentice Hall, Decision Support Systems and Intelligent Systems, 7th Edition, Turban, Aronson, and Liang.
Chapter 14: Artificial Intelligence Invitation to Computer Science, C++ Version, Third Edition.
© The McGraw-Hill Companies, Inc., Chapter 3 The Greedy Method.
Operations Research Assistant Professor Dr. Sana’a Wafa Al-Sayegh 2 nd Semester ITGD4207 University of Palestine.
Inductive learning Simplest form: learn a function from examples
Learning CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
Using Neural Networks in Database Mining Tino Jimenez CS157B MW 9-10:15 February 19, 2009.
SOFTWARE DESIGN AND ARCHITECTURE LECTURE 21. Review ANALYSIS PHASE (OBJECT ORIENTED DESIGN) Functional Modeling – Use case Diagram Description.
 The most intelligent device - “Human Brain”.  The machine that revolutionized the whole world – “computer”.  Inefficiencies of the computer has lead.
An Introduction to Programming and Algorithms. Course Objectives A basic understanding of engineering problem solving process. A basic understanding of.
Fundamentals of Algorithms MCS - 2 Lecture # 7
On Graph Query Optimization in Large Networks Alice Leung ICS 624 4/14/2011.
Testing Methods Carl Smith National Certificate Year 2 – Unit 4.
Algorithm Paradigms High Level Approach To solving a Class of Problems.
George F Luger ARTIFICIAL INTELLIGENCE 6th edition Structures and Strategies for Complex Problem Solving Machine Learning: Symbol-Based Luger: Artificial.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Distributed Models for Decision Support Jose Cuena & Sascha Ossowski Pesented by: Gal Moshitch & Rica Gonen.
Types of Algorithms. 2 Algorithm classification Algorithms that use a similar problem-solving approach can be grouped together We’ll talk about a classification.
An Introduction Student Name: Riaz Ahmad Program: MSIT( ) Subject: Data warehouse & Data Mining.
Data Mining and Decision Support
Decision Maths 1 Shortest path algorithm Dijkstra’s Algorithm A V Ali :
Knowledge Discovery and Data Mining 19 th Meeting Course Name: Business Intelligence Year: 2009.
Artificial Intelligence, simulation and modelling.
Dr. Chen, Data Mining  A/W & Dr. Chen, Data Mining Chapter 3 Basic Data Mining Techniques Jason C. H. Chen, Ph.D. Professor of MIS School of Business.
6 Resource Utilization 4/28/2017 Teaching Strategies
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
Introduction to Machine Learning, its potential usage in network area,
Introduction to Algorithms
Chapter 5. Optimal Matchings
Types of Algorithms.
1.3 Modeling with exponentially many constr.
Chapter 3 Dynamic Programming.
Unit 4: Dynamic Programming
Applied Combinatorics, 4th Ed. Alan Tucker
1.3 Modeling with exponentially many constr.
Haskell Tips You can turn any function that takes two inputs into an infix operator: mod 7 3 is the same as 7 `mod` 3 takeWhile returns all initial.
Introduction to Algorithms
The Greedy Approach Young CS 530 Adv. Algo. Greedy.
Presentation transcript:

LEARNING MODEL FOR CONSTRUCTION OF THE BEST DECISION SEQUENCE USING PRIOR KNOWLEDGE Thesis Presentation by Lilit Yenokyan

PROBLEM STATEMENT The task of this thesis is development of a learning system capable of learning from prior solution instances of a problem and able of developing new solution procedures when similar problem is introduced to the system. System involves multiple decision makers and multiple decision sequences.

PROBLEM STATEMENT (cont.) Applications Medical Field Product Assembly Assisting Drivers in Navigation Other tasks presented in the terms of system (Decision Makers, Decision sequences)

Stages of the Approach Constructing the Knowledge Base 1. Initial approach builds a learning system to document the steps of decision sequences in problem solving and also information/data used for decision making. 2. Directed Acyclic Graph (DAG) is used to represent these decision sequences, where the nodes in the graph represent the actions taken by the problem solver and directed edges represent the order of the decisions (steps). 3. In mathematics DAG, is a directed graph with no directed cycles; such that, for any vertex v, there is no nonempty directed path that starts and ends on v.

Stages of the Approach Finding the Best Decision Sequence 1. Dynamic programming approach is proposed to quickly examine various solution alternatives documented in DAG and develop as new improved solution for the problem. 2. The notion of best solution is based on the problem and the application, e.g. shortest completion time, shortest distance traveled, least load or cost required and so on.

Available Learning Models Artificial Intelligence Agent – the actor in the system. Multi-agent Systems Cooperative Concurrent Supervised Correct output is provided to system. Unsupervised Ability of recognize patterns in streams of input Reinforcement Reward for correct decisions, punishment for incorrect decisions Semi-Supervised Learning Combination of Supervised and Unsupervised Learning

Available Learning Models (cont.) Neural Networks Inspired and modeled based on brain neurons. Feed-Forward Recurrent Neural Networks Decision Trees Classification Learning. Every branch is a chose between a number of alternatives, and each leaf node represents a classification or decision. Explanation Based Learning Uses learning model of an example is sufficient for learning. Relevance Based Learning Generalizes the information from prior knowledge and uses hypothesis to create new learning examples

Differences and similarities Our approach includes elements of Reinforcement Learning Semi-Supervised Learning Decision Trees Using Methods: Graph Theory Dynamic programming

Notations

Types of nodes Knowledge base presented as Directed Acyclic Graph with distinct types of nodes NodeSets and Regular Nodes is called Hypertree.

Introducing concept of NodeSet Decision that results from merging nodes i1 and i2 with attribute x, is the outcome A

Decision that results from merging nodes i3 and i4 with attribute y, is the outcome A Introducing concept of NodeSet (cont.)

NodeSet notation used to illustrate the case: and. Introducing concept of NodeSet (cont.)

Example – Two Decision sequences First Decision Sequence

Second Decision Sequence Example – Two Decision sequences

Theory and Algorithms

Knowledge Base Construction Algorithm Decision made by DM N, where A is resulting decision {e 1, …, e n } are the decision at the input that produce A, {a} is the attribute value of the decision. Decision sequence of decision maker N. NodeSet in the Hypertree, where A is resulting intermediate or final entity, {e 1, …, e n } are the basic or intermediate entities at the input that produce A, {a} is the attribute value of the nodeSet.

Input: Set of decision sequences Step 3: For each entity e i in {e 1, …, e n } do: If e i is not in H_Tree then add new node e i to H_tree Endfor If A is not in H_Tree then add new node A to H_tree Add new nodeSet to H_tree, with references to resulting node {A} and entity nodes {e 1, …, e n } and attribute value {a}. Step 2: For each decision sequence DS m where m = 1, 2,...,S For each decision in decision sequence DS m do: Goto Step 3 Endfor Return H_Tree Knowledge Base Construction Algorithm Output: Hypertree of knowledge base H_tree Step 1: Initialize H_tree=Ø

Example – Constructing Hypertree Hypertree Build based on two decision sequences

Example – KBC (Step 1) First decision of the blue decision sequence

Example – KBC (Step 2) Second decision of blue decision sequence

Example – KBC (Step 3) Final decision of blue decision sequence

Example – KBC (Step 4) First decision of the green decision sequence

Example – KBC (Step 5) Second decision of green decision sequence

Example – KBC (Step 6) Final decision of the green decision sequence

Example – Constructing Hypertree Constructing Knowledge Base – Complete

Best Decision Sequence where: BDS is Best Decision Sequence, N is the number of nodeSets for the entity node A, NodeSet ns i is the i-th nodeSet of node A, a i is the attribute value of nodeSet ns i, {e j } is the j-th entity node in nodeSet ns i resulting to A, and m is the number of decisions sequences that result to A by nodeSet ns i.

Step 2: Calculate BDS for all the solution sequences For each outcome sequence where n = 1, 2,...,m Calculate and store End For Algorithm 2- Construction of Best Decision Sequences for List of Outcomes Input: Set of decisions to examine: Output: Best solution sequence for each final outcome in Out_set Step 1: Initialize array of solution paths Des_array[1,..,m] = Ø Step 3: Return Des_array[1,..,m]

Algorithm 3- Best Decision Sequence Input: Node A Output: Best Decision Sequence Step 1: If BDS(A) has already been calculated then return BDS(A) Step 2: Initialize: current_value = +branch_value = + Step 3:for each NodeSet of node A do Set branch_value = {a} + branch_value = min(current_value, branch_value) Record index k, from {e k } which minimized current_value Endfor Step 4:Return BDS(A)

Example – Illustration of Algorithm 2 & 3 Given the Hypertree from the previous example extended with new Decision sequence: i 6, i 7 Solution 3 (14) Construction of Best Decision Sequences (Algorithm 2) is called with {Solution 1, Solution 2, Solution 3} on its input

Example – BDS Solution 3

Example – BDS Solution 1

Example – BDS Solution 1 (Step 1)

Example – BDS Solution 1 (Step 2)

Example – BDS Solution 1 (Step 3)

Example – BDS Solution 1 (Final)

Example – BDS Solution 2

Example – BDS Solution 2 (Step 1)

Example – BDS Solution 2 (Step 2)

Example – BDS Solution 2 (Step 3)

Example – BDS Solution 2 (Step 4)

Example – BDS Solution 2 (Final)

Applications and Examples

Medical Application Input - Various sequences of tests prescribed by doctors for patients with similar symptoms and their final diagnoses Goal - Identify the best sequence of tests for developing accurate diagnosis of patients with certain symptoms. Sequence is based on specified criteria. Criteria - Cost, accuracy, lead time to diagnose

Product Assembly Application Input - Various feasible assembly sequences are captured in a knowledge base. Goal – Construct assembly sequence based on sequences given at the input optimal with respect to certain criteria. Criteria – Assembly time, cost, number of setups

Assisting Drivers Finding Best Driving Directions Application Input - various driving routes between the origin and destination. Goal - examine those routes and produce a new path that is optimal with respect to specific criteria Criteria - time, distance traveled, cost of gas

Medical Application Study involves medical history of group of patients with the similar initial symptoms and complains. It is possible that during the examination of the patient several doctors prescribe different tests but arrive to same conclusion.

Medical Application Consider the diagnostic history of patients with similar symptoms. The history of symptoms and diagnoses are documented for the patients and later the patients are divided by a medical professional to groups of people who had similar symptoms. During the examination of a patient from a group, every test prescribed by the doctor is recorded, doctor also supplies the system all the possible conclusions or intermediate decisions he/she makes based on results of each test.

Medical Application Operation is performing the test Initial entities in the knowledge base model are medical tests Intermediate entities are the decisions of doctors based on initial tests Final entities are final diagnoses.

Medical Application Construction of knowledge base for tests prescribed for patients with similar symptoms

Medical Application Construction of Best Decision Sequence for Diagnosis 1 and Diagnosis 2

Medical Application Diagnosis 1 - Comparing values (costs) of checking a patient against Diagnosis 1 is 11 versus initial costs 15 and 13. This is an improvement of 15% percent of the initial cost. Diagnosis 2 - The value for checking a patient against Diagnosis 2 is the same as it was initially 15.

Medical Application System is intended to assist the medical professional not replace. In the case if during patient inspection doctor decides to examine a diagnosis that is not in the knowledge base, the learning system records new diagnosis and the tests prescribed by the doctor for checking it. At any step if a test results do not satisfy the condition and diagnose k should not be considered any longer, sequence for performing the tests stops, as negative decision about diagnosis k has been reached.

Assembly Application Given all the necessary mechanical parts of the desk, a number of professionals are assigned to independently assemble it. The professionals perform the assembly in various orders requiring various amounts of time. The assumption is that each professional has same level of competency and that the time required performing a certain task is the same for all of them. The time of each action performed is available, i.e. system is tracking the time it takes the professional to perform each step of assembly.

Assembly Application Operation in the application is joining several parts of product together. Basic entities in the knowledge base are the parts of the final product. Intermediate entities are subassemblies obtained by sequence of operations. Final entity is the assembled product. In this assembly example desk is the final entity.

Assembly Application

RED LINE 1) A, B AB (2) 2) C, AB ABC (2) 3) D, ABC ABCD (4) 4) K, F KF (3) 5) G, KF KFG (6) 6) H, GKF KFGH (6) 7) L, E EL (5) 8) EL, KFGH Drawer (5) 9) M, ABCD ABCDM (7) 10) N, ABCDM Table (7) 11) Table, Drawer Desk (6) Total: 53 minutes

Assembly Application GREEN LINE 1) A, M AM (2) 2) C, N CN (2) 3) AM, B ABM (2) 4) CN, ABM ABCNM (2) 5) D, ABCNM Table (5) 6) F, K KF (3) 7) H, FK KFH (3) 8) G, KFH KFGH (5) 9) E, KFGH Drawer_no_handle (2) 10) Drawer_no_handle, Table Desk_no_handle (6) 11) L, Desk_no_handle Desk (14 ) Total: 47 minutes

Assembly Application BLUE LINE A, M AM (2) 2) C, N CN (2) 3) D, AM ADM (2) 4) ADM, CN ACDNM (3) 5) B, ACDNM Table (1) 6) F, K KF (3) 7) G, KF KFG (6) 8) H, KFG KFGH (6) 9) L, E EL (5) 10) Table, KFGH Desk_no_front (4) 11) Desk_no_front, EL Desk (11) Total: 45 minutes

Assembly Application

blue green red 1) A, M AM (2) 2) C, N CN (2) 3) D, AM ADM (2) 4) ADM, CN ACDNM (3) 5) B, ACDNM Table (1) 6) F, K KF (3) 7) H, FK KFH (3) 8) G, KFH KFGH (5) 9) L, E EL (5) 10) EL, KFGH Drawer (5) 11) Table, Drawer desk (6) Total: 37 minutes

Deadlock Elimination

Software Screenshot (Desk Assembly)

Research at the UofM Dearborn

Mirror symmetric reconstruction and Matching 1. Right leg 2. Left leg missing patella 3. Mirror reconstructed left leg 4. Generated mesh for missing sections 5. Zoomed view

1. Right ribs2. Left ribs 5 th rib broken 3. Mirror reconstructed left ribs 4. Generated mesh for missing rib 5. Generated solid Mirror symmetric reconstruction and Matching

Partial Symmetry Reconstruction 1. Right hand damaged middle finger2. Damaged area intersection with box3.Middle finger separate view in box 4. Candidate reconstruction area5. Damaged bone view6. Reconstructed finger bone

Density Analysis – Damage Area Recognition 1. Ulna bone 2.Ulna vertex mapping 3. Ulna edges mapping 4. Left ribs 5th rib broken 5. Left ribs vertex mapping 6. Left ribs edges mapping

Thank you! Questions and Answers