Bayesian Metanetworks for Context-Sensitive Feature Relevance Vagan Terziyan Industrial Ontologies Group, University of Jyväskylä, Finland.

Slides:



Advertisements
Similar presentations
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
Advertisements

PROBABILITY. Uncertainty  Let action A t = leave for airport t minutes before flight from Logan Airport  Will A t get me there on time ? Problems :
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
Simulation Metamodeling using Dynamic Bayesian Networks in Continuous Time Jirka Poropudas (M.Sc.) Aalto University School of Science and Technology Systems.
Introduction of Probabilistic Reasoning and Bayesian Networks
OMEN: A Probabilistic Ontology Mapping Tool Mitra et al.
Naïve Bayes Classifier
Industrial Diagnostics Using Algebra of Uncertain Temporal Relations Vladimir Ryabov, Vagan Terziyan* IASTED-2003 Innsbruck, Austria.
From: Probabilistic Methods for Bioinformatics - With an Introduction to Bayesian Networks By: Rich Neapolitan.
1 Public Commerce brief introduction of the concept Vagan Terziyan University of Jyvaskyla, Finland
PR-OWL: A Framework for Probabilistic Ontologies by Paulo C. G. COSTA, Kathryn B. LASKEY George Mason University presented by Thomas Packer 1PR-OWL.
Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –One exception: games with multiple moves In particular, the Bayesian.
… Hidden Markov Models Markov assumption: Transition model:
Review: Bayesian learning and inference
TLI371 – Distributed Artificial Intelligence in Mobile Environment Course Introduction Vagan Terziyan Department of Mathematical Information Technology.
1 Augmented Bayesian Classifier Beijie Xu. 2 problem to be solved NAÏVE Bayesian Classifier –Why naïve? - conditional independence among attributes –Is.
Predictive and Contextual Feature Separation for Bayesian Metanetworks Vagan Terziyan Industrial Ontologies Group, University of Jyväskylä,
Part 3 of 3: Beliefs in Probabilistic Robotics. References and Sources of Figures Part 1: Stuart Russell and Peter Norvig, Artificial Intelligence, 2.
Bayes Nets Rong Jin. Hidden Markov Model  Inferring from observations (o i ) to hidden variables (q i )  This is a general framework for representing.
Pattern Classification, Chapter 3 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P.
A Review of Research Topics of the AI Department in Kharkov (MetaIntelligence Laboratory) Vagan Terziyan, Helen Kaikova November 25 - December 5, 1999.
KI2 - 2 Kunstmatige Intelligentie / RuG Probabilities Revisited AIMA, Chapter 13.
Part 2 of 3: Bayesian Network and Dynamic Bayesian Network.
21 21 Web Content Management Architectures Vagan Terziyan MIT Department, University of Jyvaskyla, AI Department, Kharkov National University of Radioelectronics.
Personalisation in Mobile Environment Based on papers and presentations of Catholijn Jonker, Vagan Terziyan, Jan Treur, Oleksandra Vitko and others MIT.
Ai in game programming it university of copenhagen Welcome to... the Crash Course Probability Theory Marco Loog.
Bayesian networks practice. Semantics e.g., P(j  m  a   b   e) = P(j | a) P(m | a) P(a |  b,  e) P(  b) P(  e) = … Suppose we have the variables.
Uncertainty Logical approach problem: we do not always know complete truth about the environment Example: Leave(t) = leave for airport t minutes before.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Intelligent Web Applications (Part 1) Course Introduction Vagan Terziyan AI Department, Kharkov National University of Radioelectronics / MIT Department,
CPSC 422, Lecture 18Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Feb, 25, 2015 Slide Sources Raymond J. Mooney University of.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
1 A Semantic Metanetwork Vagan Terziyan University of Jyvaskyla, Finland
ONTOLOGY-BASED INTERNATIONAL DEGREE RECOGNITION Vagan Terziyan, Olena Kaykova University of Jyväskylä, Finland Oleksandra Vitko, Lyudmila Titova (speaker)
Learning In Bayesian Networks. Learning Problem Set of random variables X = {W, X, Y, Z, …} Training set D = { x 1, x 2, …, x N }  Each observation specifies.
Determining the Significance of Item Order In Randomized Problem Sets Zachary A. Pardos, Neil T. Heffernan Worcester Polytechnic Institute Department of.
Artificial Intelligence CS 165A Tuesday, November 27, 2007  Probabilistic Reasoning (Ch 14)
Made by: Maor Levy, Temple University  Probability expresses uncertainty.  Pervasive in all of Artificial Intelligence  Machine learning 
Cognitive Computer Vision Kingsley Sage and Hilary Buxton Prepared under ECVision Specific Action 8-3
Development of An ERROR ESTIMATE P M V Subbarao Professor Mechanical Engineering Department A Tolerance to Error Generates New Information….
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 26 of 41 Friday, 22 October.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes March 13, 2012.
Visibility Graph. Voronoi Diagram Control is easy: stay equidistant away from closest obstacles.
Artificial Intelligence CS 165A Thursday, November 29, 2007  Probabilistic Reasoning / Bayesian networks (Ch 14)
Empirical Research Methods in Computer Science Lecture 7 November 30, 2005 Noah Smith.
V13: Causality Aims: (1) understand the causal relationships between the variables of a network (2) interpret a Bayesian network as a causal model whose.
Bayesian Classifier. 2 Review: Decision Tree Age? Student? Credit? fair excellent >40 31…40
Probability Course web page: vision.cis.udel.edu/cv March 19, 2003  Lecture 15.
Marginalization & Conditioning Marginalization (summing out): for any sets of variables Y and Z: Conditioning(variant of marginalization):
Uncertainty Chapter 13. Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes' Rule.
Review: Bayesian inference  A general scenario:  Query variables: X  Evidence (observed) variables and their values: E = e  Unobserved variables: Y.
CPSC 422, Lecture 19Slide 1 Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of.
Probabilistic reasoning over time Ch. 15, 17. Probabilistic reasoning over time So far, we’ve mostly dealt with episodic environments –Exceptions: games.
Uncertainty Let action A t = leave for airport t minutes before flight Will A t get me there on time? Problems: 1.partial observability (road state, other.
1 An infrastructure for context-awareness based on first order logic 송지수 ISI LAB.
Kansas State University Department of Computing and Information Sciences CIS 730: Introduction to Artificial Intelligence Lecture 24 of 41 Monday, 18 October.
PROBABILISTIC REASONING Heng Ji 04/05, 04/08, 2016.
Matching ® ® ® Global Map Local Map … … … obstacle Where am I on the global map?                                   
Mohsen Riahi Manesh and Dr. Naima Kaabouch
Pekka Laitila, Kai Virtanen
Availability Availability - A(t)
Pekka Laitila, Kai Virtanen
Bayes Net Learning: Bayesian Approaches
Uncertainty Chapter 13.
A Bayesian Approach to Learning Causal networks
CS621: Artificial Intelligence
CSCI 5822 Probabilistic Models of Human and Machine Learning
Probabilistic Influence & d-separation
CPS 570: Artificial Intelligence Bayesian networks
Chapter 8 Estimation.
Presentation transcript:

Bayesian Metanetworks for Context-Sensitive Feature Relevance Vagan Terziyan Industrial Ontologies Group, University of Jyväskylä, Finland SETN-2006, Heraclion, Crete, Greece 24 May 2006

2 Contents Bayesian Metanetworks Metanetworks for managing conditional dependencies Metanetworks for managing feature relevance Example Conclusions Vagan Terziyan Industrial Ontologies Group Department of Mathematical Information Technologies University of Jyvaskyla (Finland) This presentation:

3 Bayesian Metanetworks

4 Conditional dependence between variables X and Y P(Y) =  X (P(X) · P(Y|X))

5 Bayesian Metanetwork Definition. The Bayesian Metanetwork is a set of Bayesian networks, which are put on each other in such a way that the elements (nodes or conditional dependencies) of every previous probabilistic network depend on the local probability distributions associated with the nodes of the next level network.

6 Two-level Bayesian C-Metanetwork for Managing Conditional Dependencies

7 Contextual and Predictive Attributes Machine Environment Sensors X x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 x7x7 predictive attributes contextual attributes air pressure dust humidity temperature emission

8 Contextual Effect on Conditional Probability (1) X x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 x7x7 predictive attributes contextual attributes xkxk xrxr Assume conditional dependence between predictive attributes (causal relation between physical quantities)… xtxt … some contextual attribute may effect directly the conditional dependence between predictive attributes but not the attributes itself

9 Contextual Effect on Conditional Probability (2) X ={x 1, x 2, …, x n } – predictive attribute with n values; Z ={z 1, z 2, …, z q } – contextual attribute with q values; P(Y|X) = {p 1 (Y|X), p 2 (Y|X), …, p r (Y|X)} – conditional dependence attribute (random variable) between X and Y with r possible values; P(P(Y|X)|Z) – conditional dependence between attribute Z and attribute P(Y|X);

10 Contextual Effect on Conditional Probability (3) xkxk xrxr xtxt X k 1 : order flowers X k 2 : order wine X r 1 : visit football match X r 2 : visit girlfriend P 1 (X r |X k ) Xk1Xk1 Xk2Xk2 Xr1Xr Xr2Xr P 2 (X r |X k ) Xk1Xk1 Xk2Xk2 Xr1Xr Xr2Xr X t 1 : I am in Paris X t 2 : I am in Moscow X k : Order present X r : Make a visit

11 Contextual Effect on Conditional Probability (4) xtxt P 1 (X r |X k ) Xk1Xk1 Xk2Xk2 Xr1Xr Xr2Xr P 2 (X r |X k ) Xk1Xk1 Xk2Xk2 Xr1Xr Xr2Xr X t 1 : I am in Paris X t 2 : I am in Moscow xrxr xkxk P( P (X r |X k ) | X t ) Xt1Xt1 Xt2Xt2 P 1 (X r |X k ) P 2 (X r |X k )

12 Contextual Effect on Unconditional Probability (1) X x1x1 x2x2 x3x3 x4x4 x5x5 x6x6 x7x7 predictive attributes contextual attributes xkxk Assume some predictive attribute is a random variable with appropriate probability distribution for its values… xtxt … some contextual attribute may effect directly the probability distribution of the predictive attribute x1x1 x2x2 x3x3 x4x4 X P(X)

13 Contextual Effect on Unconditional Probability (2)  X ={x 1, x 2, …, x n } – predictive attribute with n values; · Z ={z 1, z 2, …, z q } – contextual attribute with q values and P(Z) – probability distribution for values of Z; P(X) = {p 1 (X), p 2 (X), …, p r (X)} – probability distribution attribute for X (random variable) with r possible values (different possible probability distributions for X) and P(P(X)) is probability distribution for values of attribute P(X); · P(Y|X) is a conditional probability distribution of Y given X; · P(P(X)|Z) is a conditional probability distribution for attribute P(X) given Z

14 Contextual Effect on Unconditional Probability (3) xkxk xtxt XkXkXkXk P 1 (X k ) Xk1Xk1 Xk2Xk XkXkXkXk P 2 (X k ) Xk1Xk1 Xk2Xk X t 1 : I am in Paris X t 2 : I am in Moscow X k 1 : order flowers X k 2 : order wine X k : Order present P( P (X k ) | X t ) Xt1Xt1 Xt2Xt2 P 1 (X k ) P 2 (X k )

15 Causal Relation between Conditional Probabilities xkxk xrxr xmxm xnxn P 1 (X n |X m ) P(X n | X m ) P(P(X n | X m )) P 2 (X n |X m )P 3 (X n |X m ) P 1 (X r |X k ) P(X r | X k ) P(P(X r | X k )) P 2 (X r |X k ) P(P(X r | X k )|P(X n | X m )) There might be causal relationship between two pairs of conditional probabilities

16 Two-level Bayesian C-Metanetwork for managing conditional dependencies

17 Example of Bayesian C-Metanetwork The nodes of the 2 nd -level network correspond to the conditional probabilities of the 1 st -level network P(B|A) and P(Y|X). The arc in the 2 nd - level network corresponds to the conditional probability P(P(Y|X)|P(B|A))

18 Two-level Bayesian R-Metanetwork for Modelling Relevant Features’ Selection

19 Feature relevance modelling (1) We consider relevance as a probability of importance of the variable to the inference of target attribute in the given context. In such definition relevance inherits all properties of a probability.

20 Feature relevance modelling (2) X: {x 1, x 2, …, x nx }

21 Example (1) Let attribute X will be “state of weather” and attribute Y, which is influenced by X, will be “state of mood”. X (“state of weather”) ={“sunny”, “overcast”, “rain”}; P(X=”sunny”) = 0.4; P(X=”overcast”) = 0.5; P(X=”rain”) = 0.1; Y (“state of mood”) ={“good”, “bad”}; P(Y=”good”|X=”sunny”)=0.7; P(Y=”good”|X=”overcast”)=0.5; P(Y=”good”|X=”rain”)=0.2; P(Y=”bad”|X=”sunny”)=0.3; P(Y=”bad”|X=”overcast”)=0.5; P(Y=”bad”|X=”rain”)=0.8; P(X) P(Y|X) Let:  X =0.6

22 Example (2) Now we have: One can also notice that these values belong to the intervals created by the two extreme cases, when parameter X is not relevant at all or it is fully relevant: !

23 General Case of Managing Relevance (1) Predictive attributes: X1 with values {x1 1,x1 2,…,x1 nx1 }; X2 with values {x2 1,x2 2,…,x2 nx2 }; … XN with values {xn 1,xn 2,…,xn nxn }; Target attribute: Y with values {y 1,y 2,…,y ny }. Probabilities: P(X1), P(X2),…, P(XN); P(Y|X1,X2,…,XN). Relevancies:  X1 = P(  (X1) = “yes”);  X2 = P(  (X2) = “yes”); …  XN = P(  (XN) = “yes”); Goal: to estimate P(Y).

24 General Case of Managing Relevance (2) Probability P(XN)

25 Example of Relevance Bayesian Metanetwork (1) Conditional relevance !!!

26 Example of Relevance Bayesian Metanetwork (2)

27 Example of Relevance Bayesian Metanetwork (3)

28 When Bayesian Metanetworks ? 1. Bayesian Metanetwork can be considered as very powerful tool in cases where structure (or strengths) of causal relationships between observed parameters of an object essentially depends on context (e.g. external environment parameters); 2. Also it can be considered as a useful model for such an object, which diagnosis depends on different set of observed parameters depending on the context.

29 Conclusion We are considering a context as a set of contextual attributes, which are not directly effect probability distribution of the target attributes, but they effect on a “relevance” of the predictive attributes towards target attributes. In this paper we use the Bayesian Metanetwork vision to model such context-sensitive feature relevance. Such model assumes that the relevance of predictive attributes in a Bayesian network might be a random attribute itself and it provides a tool to reason based not only on probabilities of predictive attributes but also on their relevancies.

30 Read more about Bayesian Metanetworks in: Terziyan V., A Bayesian Metanetwork, In: International Journal on Artificial Intelligence Tools, Vol. 14, No. 3, 2005, World Scientific, pp Terziyan V., Vitko O., Bayesian Metanetwork for Modelling User Preferences in Mobile Environment, In: German Conference on Artificial Intelligence (KI-2003), LNAI, Vol. 2821, 2003, pp Terziyan V., Vitko O., Learning Bayesian Metanetworks from Data with Multilevel Uncertainty, In: M. Bramer and V. Devedzic (eds.), Proceedings of the First International Conference on Artificial Intelligence and Innovations, Toulouse, France, August , 2004, Kluwer Academic Publishers, pp