Stochastic Markov Processes and Bayesian Networks

Slides:



Advertisements
Similar presentations
Probability.
Advertisements

Introduction to Probability
MARKOV ANALYSIS Andrei Markov, a Russian Mathematician developed the technique to describe the movement of gas in a closed container in 1940 In 1950s,
Part 3 Probabilistic Decision Models
Copyright © 2007 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Slide
Markov models and HMMs. Probabilistic Inference P(X,H,E) P(X|E=e) = P(X,E=e) / P(E=e) with P(X,E=e) = sum h P(X, H=h, E=e) P(X,E=e) = sum h,x P(X=x, H=h,
Probability How likely is an event to occur?
CS188: Computational Models of Human Behavior
Ideal Parent Structure Learning School of Engineering & Computer Science The Hebrew University, Jerusalem, Israel Gal Elidan with Iftach Nachman and Nir.
40S Applied Math Mr. Knight – Killarney School Slide 1 Unit: Probability Lesson: PR-L1 Intro To Probability Intro to Probability Learning Outcome B-4 PR-L1.
Markov Processes Aim Higher. What Are They Used For? Markov Processes are used to make predictions and decisions where results are partly random but may.
Probability: The Study of Randomness Chapter Randomness Think about flipping a coin n times If n = 2, can have 2 heads (100% heads), 1 heads and.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 14 From Randomness to Probability.
Markov chains Assume a gene that has three alleles A, B, and C. These can mutate into each other. Transition probabilities Transition matrix Probability.
Review and Preview and Random Variables
A Tutorial on Learning with Bayesian Networks
Probabilistic Reasoning over Time
Probabilistic sequence modeling II: Markov chains Haixu Tang School of Informatics.
Many useful applications, especially in queueing systems, inventory management, and reliability analysis. A connection between discrete time Markov chains.
Section 5.1 and 5.2 Probability
Chapter 4 Probability and Probability Distributions
Solve for x. 28 = 4(2x + 1) = 8x = 8x + 8 – 8 – 8 20 = 8x = x Distribute Combine Subtract Divide.
Hidden Markov Models Reading: Russell and Norvig, Chapter 15, Sections
Lesson  In this investigation you will explore the predictability of random outcomes. You will use a familiar random process, the flip of a coin.
1 Chapter 6: Probability— The Study of Randomness 6.1The Idea of Probability 6.2Probability Models 6.3General Probability Rules.
Biostatistics Unit 4 Probability.
1 Markov Chains (covered in Sections 1.1, 1.6, 6.3, and 9.4)
Hidden Markov Models Tunghai University Fall 2005.
. Computational Genomics Lecture 10 Hidden Markov Models (HMMs) © Ydo Wexler & Dan Geiger (Technion) and by Nir Friedman (HU) Modified by Benny Chor (TAU)
Biostatistics Unit 4 - Probability.
Андрей Андреевич Марков. Markov Chains Graduate Seminar in Applied Statistics Presented by Matthias Theubert Never look behind you…
Hidden Markov Models in Bioinformatics Applications
Probability And Expected Value ————————————
Math 310 Section 7.2 Probability. Succession of Events So far, our discussion of events have been in terms of a single stage scenario. We might be looking.
Bayesian Networks Alan Ritter.
CS 188: Artificial Intelligence Fall 2009 Lecture 19: Hidden Markov Models 11/3/2009 Dan Klein – UC Berkeley.
. Markov Chains Tutorial #5 © Ilan Gronau. Based on original slides of Ydo Wexler & Dan Geiger.
Chapter 6 Probability.
stochastic processes(2)
Second Midterm Exam Friday, April minutes You may bring a cheat sheet with you. Charts Regression Solutions to homework 4 are available online.
Let’s flip a coin. Making Data-Based Decisions We’re going to flip a coin 10 times. What results do you think we will get?
BINF6201/8201 Hidden Markov Models for Sequence Analysis
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 6. Probability Theory and the Normal Probability Distribution.
Monte Carlo Methods Versatile methods for analyzing the behavior of some activity, plan or process that involves uncertainty.
Chapter 10: Introducing Probability STAT Connecting Chapter 10 to our Current Knowledge of Statistics Probability theory leads us from data collection.
Markov Models and Simulations Yu Meng Department of Computer Science and Engineering Southern Methodist University.
Courtesy of J. Akinpelu, Anis Koubâa, Y. Wexler, & D. Geiger
Probability Review CSE430 – Operating Systems. Overview of Lecture Basic probability review Important distributions Poison Process Markov Chains Queuing.
1 CS 552/652 Speech Recognition with Hidden Markov Models Winter 2011 Oregon Health & Science University Center for Spoken Language Understanding John-Paul.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
5.1 Probability in our Daily Lives.  Which of these list is a “random” list of results when flipping a fair coin 10 times?  A) T H T H T H T H T H 
CHAPTER 6 Naive Bayes Models for Classification. QUESTION????
POSC 202A: Lecture 4 Probability. We begin with the basics of probability and then move on to expected value. Understanding probability is important because.
Inference: Probabilities and Distributions Feb , 2012.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Bayesian networks and their application in circuit reliability estimation Erin Taylor.
Lecture 7 Dustin Lueker. 2STA 291 Fall 2009 Lecture 7.
1 CMSC 671 Fall 2001 Class #20 – Thursday, November 8.
10.1 Properties of Markov Chains In this section, we will study a concept that utilizes a mathematical model that combines probability and matrices to.
From DeGroot & Schervish. Example Occupied Telephone Lines Suppose that a certain business office has five telephone lines and that any number of these.
Hidden Markov Models BMI/CS 576
Section 5.1 and 5.2 Probability
Markov Chains Tutorial #5
CS 188: Artificial Intelligence Spring 2007
Probability Topics Random Variables Joint and Marginal Distributions
Courtesy of J. Akinpelu, Anis Koubâa, Y. Wexler, & D. Geiger
CS 188: Artificial Intelligence Fall 2008
CS 188: Artificial Intelligence Fall 2007
Markov Chains Tutorial #5
Homework Due Tomorrow mrsfhill.weebly.com.
Presentation transcript:

Stochastic Markov Processes and Bayesian Networks Aron Wolinetz

Bayesian or Belief Network A probabilistic graphical model that represents a set of random variables and their conditional dependencies via a directed acyclic graph (DAG). Will my student loan get funded?

Bayesian Networks All we need to add is some “chance” to the graph Chance of rain given clouds? P(R|C) Chance of wet grass given rain? P(W|R)

Network Example • An example (from California): I’m at work, neighbor John calls to say my alarm is ringing, but neighbor Mary doesn’t call. Sometimes it’s set off by minor earthquakes. Is there a burglar? • Variables: Burglar, Earthquake, Alarm, JohnCalls, MaryCalls • Network topology reflects “causal” knowledge: – A burglar can set the alarm off – An earthquake can set the alarm off – The alarm can cause Mary to call – The alarm can cause John to call

Inference Tasks Simple queries Conjunctive queries Optimality – P(outcome | action, evidence) Value of information – what to do next Sensitivity – which values are most important Explanation – why do I need to do something

Inference by Enumeration Go through every case using every variable with every probability Grows exponentially with size of the network Can become intractable 250 = 1X1015 > 1 day (at 1 billion calculations per second)

Inference by Stochastic Simulation There must be a better way then enumeration. Lets roll the dice and see what happens I can calculate the odds of heads or tails, or I can flip a coin over and over and see what odds turn up.

Stochastic? A system whose behavior is intrinsically non-deterministic. A system’s subsequent state is determined by predictable actions and a random element. Drunken sailor.

Stochastic Matrix also called a probability matrix, transition matrix, substitution matrix, or Markov matrix Matrix of non-negative, real numbers between 0 and 1 ( 0 ≤ X ≤ 1 ) Every row must total to 1

Stochastic Matrix i, j represent: rows, column The probability of going from state i to state j is equal to Xi,j Or we can write P(j|i)= Xi,j Five box cat and mouse game.

Cat and Mouse Five Boxes [1,2,3,4,5] Cat starts in box 1, mouse starts in box 5 Each turn each animal can move left or right, (randomly) If they occupy the same box, game over (for the mouse anyway)

5 box cat and mouse game States: Stochastic Matrix: State 1: cat in the first box, mouse in the third box: (1, 3) State 2: cat in the first box, mouse in the fifth box: (1, 5) State 3: cat in the second box, mouse in the fourth box: (2, 4) State 4: cat in the third box, mouse in the fifth box: (3, 5) State 5: the cat ate the mouse and the game ended: F.

State Diagram of cat and mouse State two in green is the initial state State five in blood red, is an accepting state We can bounce around for a while, but eventually the mouse will die

Instance of a game This game consists of a chain of events State 2 C=1 M=5 State 3 C=2 M=4 State 4 C=3 State 5 C=4 This game consists of a chain of events Lets call it a Markov Chain!

Stochastic System

Properties How many numbers do we need to specify all the necessary probability values for the network? How many would we need if there were no conditional independencies? (without the network) Does the network cut down on our work? 1+2+2+4+2 = 11 25 =32

Frog Cell Cycle Sible and Tyson figure 1 Methods 41 2007

Frog Cell Cycle Concentration or number of each of the molecule is a state. Each reaction serves as a transition from state to state. Whether or not a reaction will occur is Stochastic. Each state contains a complete picture of every species' quantity

Markov? Andrey (Andrei) Andreyevich Markov Russian Mathematician June 14, 1856 – July 20, 1922

Markov Chain Future is independent of the past given the present. Want to know tomorrow’s weather? Don’t look at yesterday, look out the window. Requires perfect knowledge of current state. Very Simple, Very Powerful. P(Future | Present)

Markov Chain Make predictions about future events given probabilities based on the current state. Probability of the future, given the present. Transition from state to state

First Order Markov Chain Make a Markov assumption that the value of the current state depends only on a fixed number of previous states In our case we are only looking back to one previous state Xt only depends on Xt-1

Second Order Markov Chain Value of the current state depends on the two previous states P(Xt|Xt-1,Xt-2) The math starts getting very complicated Can expand to third fourth… Markov chains

Stochastic Markov Chain Drunken sailor. Walk through the number line Flip a coin, Heads +1, Tails -1 (50/50) Stochastic Matrix From any position there are two possible transitions, to the next or previous integer. The transition probabilities depend only on the current position, not on the manner in which the position was reached. For example, the transition probabilities from 5 to 4 and 5 to 6 are both 0.5, and all other transition probabilities from 5 are 0. These probabilities are independent of whether the system was previously in 4 or 6.

Back to the Frog Cell Cycle State 1 Time = 0 #Cyclin = .1 #MPF = .05 #PMPF = .05 State 2 Time = 20 #Cyclin = .25 #MPF = .15 #PMPF = .025 State 3 Time = 40 #Cyclin = .35 #MPF = .25 State 4 Time = 60 #Cyclin = .4 #MPF = .1 #PMPF = .3 State 5 Time = 80 #Cyclin = .15 #MPF = .5 #PMPF = .5

Hidden Markov Models Sometimes we have an incomplete view of the world. However, where there is smoke, there is usually fire. We can use what we observe to make inferences about the present, or the future.

Hidden Markov Models Let (Z1, Z2 … Zn) be our “hidden” variables. Let (X1, X2 … Xn) be what we observe. This is what an HMM looks like Z1 X1 Z2 X2 Z3 X3 Z4 X4 Z5 X5

Components of an HMM States (Hidden) Observations Starting Probability - Where might we begin Transition Probability - From state to state Emission Probability- Given a state, probability of observable actions occurring.

Handwriting analysis using HMM Hidden states – What letter does the writer intend. Observation – What chicken scratch did the person scribble. Starting probability – There are 26 letters, some are more likely to start a word. Emission probability – What letters are likely to follow other letters (stochastic matrix)

Lets predict the weather On day 0 it is sunny and beautiful [1 0] Day 0 times Transition Matrix = Day 1 [1 0] * = [.9 .1] (90% chance of sun) Day 1 times Transition Matrix = Day 2 [.86 .14] What if we took this to ∞? = [.833 .167]

Steady State Requires a regular transition matrix (at least one row of all non zero entries) Is independent of the starting state Represents the probability of all days 83% of days are sunny

Poisson Process A stochastic process which counts the number of events and the time that these events occurred. Independent increments – the numbers of occurrences counted in disjoint intervals are independent from each other. Stationary increments - the probability distribution of the number of occurrences counted in any time interval only depends on the length of the interval. No counted occurrences are simultaneous.

Poisson Process The number of raindrops falling within a specified area The number of particles emitted via radioactive decay by an unstable substance The number of requests for individual documents on a web server Number of goals scored in a hockey Game

Poisson Process

WAKE UP!!! I’M DONE