Situation decomposition method extracts partial data which contains some rules. Hiroshi Yamakawa (FUJITSU LABORATORIES LTD.)

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Experiential Learning Cycle
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
Computer science is a field of study that deals with solving a variety of problems by using computers. To solve a given problem by using computers, you.
Vision Based Control Motion Matt Baker Kevin VanDyke.
COMPUTER AIDED DIAGNOSIS: FEATURE SELECTION Prof. Yasser Mostafa Kadah –
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
Yiannis Demiris and Anthony Dearden By James Gilbert.
SSP Re-hosting System Development: CLBM Overview and Module Recognition SSP Team Department of ECE Stevens Institute of Technology Presented by Hongbing.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Week 9 Data Mining System (Knowledge Data Discovery)
Using Analytic QP and Sparseness to Speed Training of Support Vector Machines John C. Platt Presented by: Travis Desell.
Parameterizing Random Test Data According to Equivalence Classes Chris Murphy, Gail Kaiser, Marta Arias Columbia University.
Intelligent Database Systems Lab Advisor : Dr.Hsu Graduate : Keng-Wei Chang Author : Andrew K. C. Wong Yang Wang 國立雲林科技大學 National Yunlin University of.
Multiple Data Structuring I. N. Skopin Possibilities of working with data possessing several structures are discussed. Such work is shown to require development.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Course Instructor: Aisha Azeem
Introduction to machine learning
Learning Table Extraction from Examples Ashwin Tengli, Yiming Yang and Nian Li Ma School of Computer Science Carnegie Mellon University Coling 04.
System Analysis Overview Document functional requirements by creating models Two concepts help identify functional requirements in the traditional approach.
Kansas State University Department of Computing and Information Sciences CIS 830: Advanced Topics in Artificial Intelligence From Data Mining To Knowledge.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Example Clustered Transformations MAP Adaptation Resources: ECE 7000:
UNIVIRTUAL FOR INSTRUCTIONAL DESIGN Versione 00 del 29/07/2009.
Complex Cognitive Processes Woolfolk, Cluster 8
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Introduction SNR Gain Patterns Beam Steering Shading Resources: Wiki:
Presented by Tienwei Tsai July, 2005
© Yilmaz “Agent-Directed Simulation – Course Outline” 1 Course Outline Dr. Levent Yilmaz M&SNet: Auburn M&S Laboratory Computer Science &
Project MLExAI Machine Learning Experiences in AI Ingrid Russell, University.
CS654: Digital Image Analysis Lecture 3: Data Structure for Image Analysis.
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
Study on Genetic Network Programming (GNP) with Learning and Evolution Hirasawa laboratory, Artificial Intelligence section Information architecture field.
SOFTWARE DESIGN.
My talk describes how the detailed error diagnosis and the automatic solution procedure of problem solving environment T-algebra can be used for automatic.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
GATree: Genetically Evolved Decision Trees 전자전기컴퓨터공학과 데이터베이스 연구실 G 김태종.
Analysis of algorithms Analysis of algorithms is the branch of computer science that studies the performance of algorithms, especially their run time.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Finnish-Russian Doctoral Seminar on Multicriteria Decision Aid and Optimization Iryna Yevseyeva Niilo Mäki Institute University of Jyväskylä, Finland
1 CHAPTER 2 Decision Making, Systems, Modeling, and Support.
University of Windsor School of Computer Science Topics in Artificial Intelligence Fall 2008 Sept 11, 2008.
OOAD Unit – I OBJECT-ORIENTED ANALYSIS AND DESIGN With applications
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
(1) Abstract and Contents New model selection criteria called Matchability which is based on maximizing matching opportunity is proposed. Given data set.
Software Architecture Evaluation Methodologies Presented By: Anthony Register.
Chapter 7. Learning through Imitation and Exploration: Towards Humanoid Robots that Learn from Humans in Creating Brain-like Intelligence. Course: Robots.
Distributed Models for Decision Support Jose Cuena & Sascha Ossowski Pesented by: Gal Moshitch & Rica Gonen.
CpSc 881: Machine Learning
Chapter 8. Learning of Gestures by Imitation in a Humanoid Robot in Imitation and Social Learning in Robots, Calinon and Billard. Course: Robots Learning.
Finite State Machines (FSM) OR Finite State Automation (FSA) - are models of the behaviors of a system or a complex object, with a limited number of defined.
Data Mining and Decision Support
RULES Patty Nordstrom Hien Nguyen. "Cognitive Skills are Realized by Production Rules"
Motor Behavior Chapter 5. Motor Behavior Define motor behavior, motor development, motor control, and motor learning. What is the influence of readiness,
KNOWLEDGE MANAGEMENT UNIT II KNOWLEDGE MANAGEMENT AND TECHNOLOGY 1.
Some Thoughts to Consider 5 Take a look at some of the sophisticated toys being offered in stores, in catalogs, or in Sunday newspaper ads. Which ones.
COURSE AND SYLLABUS DESIGN
Public Policy Process An Introduction.
1 A latent information function to extend domain attributes to improve the accuracy of small-data-set forecasting Reporter : Zhao-Wei Luo Che-Jung Chang,Der-Chiang.
DATA MINING TECHNIQUES (DECISION TREES ) Presented by: Shweta Ghate MIT College OF Engineering.
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
1 Double-Patterning Aware DSA Template Guided Cut Redistribution for Advanced 1-D Gridded Designs Zhi-Wen Lin and Yao-Wen Chang National Taiwan University.
1 Chapter 1 Introduction to Accounting Information Systems Chapter 2 Intelligent Systems and Knowledge Management.
BRAIN SCAN  Brain scan is an interactive quiz for use as a revision/ learning reinforcement tool that accompanies the theory package.  To answer a question.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
1 Tempo Induction and Beat Tracking for Audio Signals MUMT 611, February 2005 Assignment 3 Paul Kolesnik.
Process 4 Hours.
Rule Induction for Classification Using
A Unifying View on Instance Selection
Market-based Dynamic Task Allocation in Mobile Surveillance Systems
Generally Discriminant Analysis
Presentation transcript:

Situation decomposition method extracts partial data which contains some rules. Hiroshi Yamakawa (FUJITSU LABORATORIES LTD.)

Abstract In an infantile development process, the fundamental knowledge about the external world is acquired through learning without clear purposes. An adult is considered to use that fundamental knowledge for various works. The acquisition of the internal model in these early stages may exist as a background of the flexible high order function of human brain. However, research of such learning technology is not progressing to nowadays. The system can improves prediction ability and reusability in the lasting work by using the result of learning without clear purposes. Then, we proposed the situation decomposition technology which chooses the partial information which emphasizes the relation "another attribute value will also change if one attribute value changes." Situation decomposition technology is the technology of performing attribute selection and case selection simultaneously from the data structure from which each example constitutes an attribute vector. The newly introduced Matchability criteria are the amount of evaluations which becomes large, when the explanation range of the selected partial information becomes large and a strong relation exists in the inside. Processing of situation decomposition extracts plural partial situations (result of attribute selection and case selection) of corresponding to the local maximum points over this evaluation. Furthermore, extraction of partial problem space (based on the Markov decision process) is possible using the technology which extended situation decomposition in the direction of time. In action decision task, such as robot control, partial problem space can be assigned as each module of multi-module architecture. Then, it can be efficiently adapted to unknown problem space by combining extracted plural partial problem space.

My strategy for Brain-like processing Brain has very flexible learning ability. The intelligent processes which has more flexible learning abilities are more close to real brain processes. I want introduce learning ability to my system as possible.

Contents 1. Development and Autonomous Learning 2. SOIS (Self-organizing Information Selection) as Pre-task Learning 3. Delivering Matchable Principle 4. Situation Decomposition using Matchability Criterion 5. Application of Situation Decomposition 6. Conclusions & Future works

Autonomous Learning (Framework) Outline of this talk Pre-task learning Self-organizing Information Selection Situation decomposition Task learning Cognitive Development Situation Decomposition using Matchability Criterion Matchable Principle Matchability Criterion

Development and Autonomous Learning

Two aspects of Development “Acquired environmental knowledge without particular goals which helps for problem solving for particular goals”  → “Pre-task Learning” in Autonomous Learning “Calculation process which increases the predictable and/or operable object in the world”  → Enhancing prediction ability

Autonomous Learning: AL Two phases learning (Research in RWC) Task learningExisting Knowledge Acquiring environmental knowledge General fact For design Acquiring solution for goal goal No reaching over the wall Acquiring movable paths Generating path to the goal Environment is given Goal is given Pre-task learning Development Today’s Topic

Pre-task Learning helps Task Learning Autonomous Learning (AL)  Pre-task Learning Acquiring environmental knowledge without particular goal.  Task Learning Environmental knowledge speed up aacquiring solution for goal. In human:  Adult people can solve given task quickly using environmental knowledge acquired for other goal or without particular goal. Development ~ Pre-task Learning Development Today’s Topic

Research topics for AL Pre-task Learning (How to acquire environmental knowledge)  Situation Decomposition using Matchability criterion Situation Decomposition is kind of a Self-organizing Information Selection technology. Task learning (How to use environmental knowledge)  CITTA (Cognition based Intelligent Transaction Architecture) Multi-module architecture which can combining environmental knowledge acquired during Pre-task learning  Cognitive Distance Learning Goal driven problem solver for each environmental knowledge. Development Today’s Topic

Overview of Approaching for AL CITTA Combining environmental knowledge Situation Decomposition Acquiring environmental knowledge Cognitive Distance Learning Problem solver for each environmental knowledge Architecture Learning algorithm Pre-task Learning Task Learning

SOIS (Self-organizing Information Selection) as Pre-task Learning

SOIS: Self-organizing Information Selection Process: Selecting plural partial information from data.  → “Situation Decomposition” Criterion: Evaluation for each partial information.  → Matchability Criterion Knowledge = Set of structure. Partial Information = One kind of structure ※ SOIS could be a kind of knowledge acquiring process in development

Situation Decomposition is kind of SOIS For situation decomposition Partial Information = Situation Extracting plural situations which are combination of selected attributes and cases from spread sheet. MS 4 attributes Cases MS 1 MS 2 MS 3

Delivering Matchable Principle

Two aspects of Development “Acquired environmental knowledge without particular goals which helps solving problem for particular goals”  → “Pre-task Learning” in Autonomous Learning “Calculation process which increases the predictable and/or operable object in the world”  → Enhancing prediction ability

How to enhance prediction ability We needs Criterion for selecting situation.  We wants to extract local structures. Multiplex local structure is mixed in real world data MS 4 MS 1 MS 2 MS 3 Situation Decomposition

Deriving Matchable Principal What is Criterion for each selecting situation. Matchable principle  “S tructures where a matching opportunity is large are selected.” Extracting structure (knowledge) without particular goals. Prediction is based on matching a case with experiences.

Factors in Matchable Principle To increase matching opportunity Simplicity of Structure Ockham’s razor MDL 、 AIC Consistency for Data Coverage for Data Our proposed Matchability criterion Relation in Structure Accuracy Minimize error Case-increasing Attribute-increasing Association rule

SD (Situation Decomposition ) and Implementation

Situation Decomposition Extracting plural situations which are combination of selected attributes and cases from spread sheet. Matchability=This criteria evaluates matching opportunity Matchable Situation = Local maximums of Matchability MS 4 attributes Cases MS 1 MS 2 MS 3

Formalization: Whole situation and Partial situations Whole situation J=(D, N) : Contains N attributes and D cases. Attribute selection vector:  d = (d 1, d 2, …,d D ) Case selection vector :  n = (n 1, n 2, …,n N ) Vector element d i,n i are binary indicator of selection/unselection. Number of selected attributes: d Number of selected cases : n Situation decomposition extracts some matchable situations from whole situation J=(D, N) which potentially contains 2 D+N partial situation.

Case selection using Segment space Segment space is multiplication of separation of each selected attributes. (example: two dimension) n : Number of selected cases S d : Number of total segments r d : Number of selected segments ※ Cases inside the chosen segments are surely chosen. Sd =s1 s2Sd =s1 s2 attribute1 attribute2

[Number of selected cases] n →Make Larger [Number of total segments] Sd →Make Larger Matchability criterion from Matchable Principle nn SdSd rdrd rdrd N: Total number of cases, C 1, C 2, C 3 : Positive constant [Number of selected segments] rd →Make Smaller Simplicity of Structure Coverage for Data

Matchability Focuses in covariance Types of Relations  Coincidence The relation to which two events happen simultaneously  Covariance The relation that another attribute value will also change if one attribute value changes Matchability:  Estimates covariance in selected data for categorical attributes. ABC ⅰ8010 ⅱ 8010 ⅲ 80

How to find situations Algorithms searches local maximums of Matchability Criterion. Algorithm Overview  for each subset of d of D  Search Local maximums  Reject saddle point  end Time complexity ∝ 2 D

Simple example Input situation  Mixture of cases on two plains. Situation A: x + z = 1 Situation B: y +z = 1 Extracted situation  Input Situations MS 1= Input Situation A MS 2= Input Situation B  A New Situation MS 3 :  line x = y, x + z = 1

Generalization ability Multi-valued function φ:(x,y)→z Even if the input situation A (x+z=1) lacks half of its parts, such that no data exists in the range y>0.5, our method outputs φ MS1 (0,1)=1.0.

Applications of Situation Decomposition (SD)

Multi-module Prediction System InputOutput

● Training cases 500 cases are sprayed on each plain in uniform distribution in the range x=[0.0, 1.0] and y=[0.0, 1.0]. ● Test cases 11×11 cases are arranged to notches at a regular interval of 0.1 on each plane Training cases and Test cases q: sampling rate

Prediction Result without Matchable Situation with Matchable Situation

Autonomous Learning: AL Two step learning (Research in RWC) Task learningExisting Knowledge Acquiring environmental knowledge General fact For design Acquiring solution for goal goal No reaching over the wall Acquiring movable paths Generating path to the goal Environment is given Goal is given Pre-task learning Development Today’s Topic

Demonstration of Autonomous Learning Door & Key task with CITTA Start Mobile Agent Door Telephone Key Goal Agent acquire knowledge as situation Door can open by the key.

Input/Output Each Situation is used as Module PositionActionObjectBelongings Matchable Situation i Matchable Situation 1 Go by wall Go straight Matchable Situation 2 Open door by telephone Open door by Key Extracting Matchable Situation Pre-task Learning Combining Matchable Situation Task Learning … Environment Mobile Agent

Situation Decomposition in AL SD in Pre-task learning:  Situation decomposition handles input /output vector of two time step for extracts Markov process. Advantages by SD in Task learning:  Adaptation by combining situations are possible.  Learning data can be reduced, because learning space for each module is reduced.

Conclusions and Future works

Autonomous Learning Conclusions Pre-task learning Self-organizing Information Selection Situation decomposition Task learning Cognitive Development Situation Decomposition using Matchability Criterion Matchable Principle Matchability Criterion

Conclusions & Future work Situation decomposition Matchability is new model selection criterion maximizing matching opportunity, which emphasize Coverage for data. In opposition ockham’s razor emphasize the Consistency for data. Decomposed situations by matchability criterion has powerful prediction ability. Situation decomposition method can be applied to pre-processing of data analysis, self-organization, pattern recognition and so on.

Future work Situation decomposition:  Needs theoretical research on Matchabilty criterion. This intuitively delivered criterion affected unbalanced data.  Needs speed up for large-scale problem. Exponential time complexity for number of attribute is awful.  Advanced Self-organized Information Selection Situation decomposition method only selects set of attributes and cases Autonomous Learning:  Relates with the knowledge of cognitive science.