Presentation is loading. Please wait.

Presentation is loading. Please wait.

Graphical Models - Learning -

Similar presentations


Presentation on theme: "Graphical Models - Learning -"— Presentation transcript:

1 Graphical Models - Learning -
Based on J. A. Bilmes,“A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models“, TR , U.C. Berkeley, April 1998; G. J. McLachlan, T. Krishnan, „The EM Algorithm and Extensions“, John Wiley & Sons, Inc., 1997; D. Koller, course CS-228 handouts, Stanford University, 2001., N. Friedman & D. Koller‘s NIPS‘99. Advanced I WS 06/07 PCWP CO HRBP HREKG HRSAT ERRCAUTER HR HISTORY CATECHOL SAO2 EXPCO2 ARTCO2 VENTALV VENTLUNG VENITUBE DISCONNECT MINVOLSET VENTMACH KINKEDTUBE INTUBATION PULMEMBOLUS PAP SHUNT ANAPHYLAXIS MINOVL PVSAT FIO2 PRESS INSUFFANESTH TPR LVFAILURE ERRBLOWOUTPUT STROEVOLUME LVEDVOLUME HYPOVOLEMIA CVP BP Graphical Models - Learning - Structure Learning Wolfram Burgard, Luc De Raedt, Kristian Kersting, Bernhard Nebel Albert-Ludwigs University Freiburg, Germany

2 Learning With Bayesian Networks
Fixed structure Fixed variables Hidden variables observed fully Partially Easiest problem counting Selection of arcs New domain with no domain expert Data mining Numerical, nonlinear optimization, Multiple calls to BNs, Difficult for large networks Encompasses to difficult subproblem, „Only“ Structural EM is known Scientific discouvery ? ? ? A B A B A B H - Learning Stucture learing? Parameter Estimation

3 Why Struggle for Accurate Structure?
Earthquake Alarm Set Sound Burglary Truth Missing an arc Adding an arc Earthquake Alarm Set Sound Burglary Earthquake Alarm Set Sound Burglary - Learning Cannot be compensated for by fitting parameters Wrong assumptions about domain structure Increases the number of parameters to be estimated Wrong assumptions about domain structure

4 Unknown Structure, (In)complete Data
E, B, A <Y,N,N> <Y,N,Y> <N,N,Y> <N,Y,Y> . Network structure is not specified Learnerr needs to select arcs & estimate parameters Data does not contain missing values E B A ? e b B E P(A | E,B) .9 .1 e b .7 .3 .99 .01 .8 .2 B E P(A | E,B) Learning algorithm E B A - Learning E, B, A <Y,?,N> <Y,N,?> <N,N,Y> <N,Y,Y> . <?,Y,Y> Network structure is not specified Data contains missing values Need to consider assignments to missing values

5 Score-based Learning Define scoring function that evaluates how well a structure matches the data E, B, A <Y,N,N> <Y,Y,Y> <N,N,Y> <N,Y,Y> . score E - Learning E B E A A B A B Search for a structure that maximizes the score

6 Structure Search as Optimization
Input: Training data Scoring function Set of possible structures Output: A network that maximizes the score - Learning

7 Traverse space looking for high-scoring structures Search techniques:
Heuristic Search Define a search space: search states are possible structures operators make small changes to structure Traverse space looking for high-scoring structures Search techniques: Greedy hill-climbing Best first search Simulated Annealing ... Theorem: Finding maximal scoring structure with at most k parents per node is NP-hard for k > 1 - Learning

8 Typically: Local Search
Start with a given network empty network, best tree , a random network At each iteration Evaluate all possible changes Apply change based on score Stop when no modification improves score S C E D - Learning

9 Typically: Local Search
Start with a given network empty network, best tree , a random network At each iteration Evaluate all possible changes Apply change based on score Stop when no modification improves score S C E D Add C D S C E D - Learning

10 Typically: Local Search
Start with a given network empty network, best tree , a random network At each iteration Evaluate all possible changes Apply change based on score Stop when no modification improves score S C E D Add C D Reverse C E S C E D - Learning S C E D

11 Typically: Local Search
Start with a given network empty network, best tree , a random network At each iteration Evaluate all possible changes Apply change based on score Stop when no modification improves score S C E D Add C D Reverse C E S C E D Delete C E - Learning S C E D S C E D

12 Typically: Local Search
If data is complete: To update score after local change, only re-score (counting) families that changed S C E D Add C D Reverse C E S C E D Delete C E - Learning S C E D S C E D If data is incomplete: To update score after local change, reran parameter estimation algorithm

13 Local Search in Practice
Local search can get stuck in: Local Maxima: All one-edge changes reduce the score Plateaux: Some one-edge changes leave the score unchanged Standard heuristics can escape both Random restarts TABU search Simulated annealing - Learning

14 Local Search in Practice
Using LL as score, adding arcs always helps Max score attained by fully connected network Overfitting: A bad idea… Minimum Description Length: Learning  data compression Other: BIC (Bayesian Information Criterion), Bayesian score (BDe) < > < ?? ??> < ?? > <?? ??> ………………. - Learning DL(Data|model) DL(Model)

15 Local Search in Practice
Perform EM for each candidate graph Parameter space G3 G2 G1 Parametric optimization (EM) Local Maximum G4 Gn - Learning

16 Local Search in Practice
Perform EM for each candidate graph Parameter space G3 G2 G1 Parametric optimization (EM) Local Maximum Computationally expensive: Parameter optimization via EM — non-trivial Need to perform EM for all candidate structures Spend time even on poor candidates  In practice, considers only a few candidates G4 Gn - Learning

17 Structural EM [Friedman et al. 98]
Recall, in complete data we had Decomposition  efficient search Idea: Instead of optimizing the real score… Find decomposable alternative score Such that maximizing new score  improvement in real score - Learning

18 Structural EM [Friedman et al. 98]
Idea: Use current model to help evaluate new structures Outline: Perform search in (Structure, Parameters) space At each iteration, use current model for finding either: Better scoring parameters: “parametric” EM step or Better scoring structure: “structural” EM step - Learning

19 Structural EM [Friedman et al. 98]
Reiterate Score & Parameterize Computation Expected Counts N(X1) N(X2) N(X3) N(H, X1, X1, X3) N(Y1, H) N(Y2, H) N(Y3, H) X1 X2 X3 H Y1 Y2 Y3 X1 X2 X3 H Y1 Y2 Y3 - Learning N(X2,X1) N(H, X1, X3) N(Y1, X2) N(Y2, Y1, H) Training Data X1 X2 X3 H Y1 Y2 Y3

20 Structure Learning: incomplete data
S X D C B <? > <1 1 ? 0 1> < ? ?> <? ? 0 ? 1> ……… Data E A Expectation B Inference: P(S|X=0,D=1,C=0,B=1) Current model S X D C B ……….. Expected counts Maximization Parameters - Learning EM-algorithm: iterate until convergence

21 Structure Learning: incomplete data
B S X D C B <? > <1 1 ? 0 1> < ? ?> <? ? 0 ? 1> ……… Data E A A Expectation B Inference: P(S|X=0,D=1,C=0,B=1) Current model S X D C B ……….. Expected counts Maximization Parameters Maximization Structure - Learning E E B E SEM-algorithm: iterate until convergence A A A B B

22 Structure Learning: Summary
Expert knowledge + learning from data Structure learning involves parameter estimation (e.g. EM) Optimization w/ score functions likelihood + complexity penality = MDL Local traversing of space of possible structures: add, reverse, delete (single) arcs Speed-up: Structural EM Score candidates w.r.t. current best model - Learning


Download ppt "Graphical Models - Learning -"

Similar presentations


Ads by Google