Stochasticity and Probability. A new approach to insight Pose question and think of the answer needed to answer it. Ask: How do the data arise? What is.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

A Tutorial on Learning with Bayesian Networks
Bayesian Network and Influence Diagram A Guide to Construction And Analysis.
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
CHAPTER 21 Inferential Statistical Analysis. Understanding probability The idea of probability is central to inferential statistics. It means the chance.
Chapter 4 Probability and Probability Distributions
Introduction of Probabilistic Reasoning and Bayesian Networks
Introduction to probability theory and graphical models Translational Neuroimaging Seminar on Bayesian Inference Spring 2013 Jakob Heinzle Translational.
DEPARTMENT OF HEALTH SCIENCE AND TECHNOLOGY STOCHASTIC SIGNALS AND PROCESSES Lecture 1 WELCOME.
. Approximate Inference Slides by Nir Friedman. When can we hope to approximate? Two situations: u Highly stochastic distributions “Far” evidence is discarded.
Review of Probability and Random Processes
Computer vision: models, learning and inference Chapter 10 Graphical Models.
1. An Overview of the Data Analysis and Probability Standard for School Mathematics? 2.
POSC 202A: Lecture 1 Introductions Syllabus R Homework #1: Get R installed on your laptop; read chapters 1-2 in Daalgard, 1 in Zuur, See syllabus for Moore.
Statistical Decision Theory
Statistical Modeling with SAS/STAT Cheng Lei Department of Electrical and Computer Engineering University of Victoria April 9, 2015.
EE325 Introductory Econometrics1 Welcome to EE325 Introductory Econometrics Introduction Why study Econometrics? What is Econometrics? Methodology of Econometrics.
PHYSICS Introduction. What is Science - systematic knowledge of the physical or material world gained through observation and experimentation.
EXPERIMENTAL DESIGN Science answers questions with experiments.
6/3/2016 SCIENTIFIC METHOD PROCESSES OF SCIENTIFIC INQUIRY.
Welcome to Biology Mrs. Webster Room 243. List the steps of the scientific method. List characteristics of life. What is the difference between growth.
Lecture 7: What is Regression Analysis? BUEC 333 Summer 2009 Simon Woodcock.
Ch 8. Graphical Models Pattern Recognition and Machine Learning, C. M. Bishop, Revised by M.-O. Heo Summarized by J.W. Nam Biointelligence Laboratory,
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
Learning In Bayesian Networks. General Learning Problem Set of random variables X = {X 1, X 2, X 3, X 4, …} Training set D = { X (1), X (2), …, X (N)
POSC 202A: Lecture 4 Probability. We begin with the basics of probability and then move on to expected value. Understanding probability is important because.
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
Warsaw Summer School 2015, OSU Study Abroad Program Normal Distribution.
1 Review of Probability and Random Processes. 2 Importance of Random Processes Random variables and processes talk about quantities and signals which.
Statistical NLP: Lecture 4 Mathematical Foundations I: Probability Theory (Ch2)
Chapter 1 Section 1. What is Science? Science is a method for studying the natural world. Science comes from the Latin word “scientia” which means knowledge.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Bayes made simple. Significance is… P (obtaining a test statistic more extreme than the one we observed|H 0 is true) This is not a test of the strength.
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
Bayesian analysis of a conceptual transpiration model with a comparison of canopy conductance sub-models Sudeep Samanta Department of Forest Ecology and.
Prediction and Missing Data. Summarising Distributions ● Models are often large and complex ● Often only interested in some parameters – e.g. not so interested.
CS498-EA Reasoning in AI Lecture #19 Professor: Eyal Amir Fall Semester 2011.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Understanding Populations & Samples
Intro to Research Methods
QMT 3033 ECONOMETRICS QMT 3033 ECONOMETRIC.
Hierarchical Models.
Prepared by Lloyd R. Jaisingh
Point and interval estimations of parameters of the normally up-diffused sign. Concept of statistical evaluation.
PROCESSES OF SCIENTIFIC INQUIRY
Modeling and Simulation (An Introduction)
Appendix A: Probability Theory
Statistical Data Analysis
Introductory Econometrics
Maximum Likelihood Find the parameters of a model that best fit the data… Forms the foundation of Bayesian inference Slide 1.
Chapter 4 – Part 3.
Scientific Methods Science in Practice.
Chapter 1 Section 2: Scientific Inquiry
How will discussion days/note taking work in science class this year?
Steps of answering a scientific question
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Propagation Algorithm in Bayesian Networks
Statistical NLP: Lecture 4
CSCI 5822 Probabilistic Models of Human and Machine Learning
Like all science, biology is a process of inquiry.
Chapter 2: Steps of Econometric Analysis
POSC 202A: Lecture 1 Introductions Syllabus R
Statistical Data Analysis
CS 594: Empirical Methods in HCC Introduction to Bayesian Analysis
Tools of Environmental Science
CS639: Data Management for Data Science
MECH 3550 : Simulation & Visualization
Statistics Review (It’s not so scary).
Lesson Overview 1.1 What Is Science?.
Introduction to Decision Sciences
Presentation transcript:

Stochasticity and Probability

A new approach to insight Pose question and think of the answer needed to answer it. Ask: How do the data arise? What is the hypothesized process that produces them? What are the sources of randomness/uncertainty in the process and the way we observe it? How can we model the process and its associated uncertainty in a way that allows the data to speak informatively? This approach is based on a firm intuitive understanding of the relationship between process models and probability models

Modeling process Identify scientific objectives Collect & understand data Draw upon existing theory/ knowledge Visualize model (DAG) Write model using probability notation Pick appropriate distributions Write down (unnormalized) posterior Derive full-conditional distributions Construct MCMC algorithm Fit model (using MCMC & data) Evaluate models (posterior predictive checks) Use output to make inferences Model selection Program model components in software Problem identification / data collection Model specification Model implementation Model evaluation & inference or

What are deterministic ecological models? Scientific hypothesis = verbal statement of the way nature works. Model = mathematical statement of scientific hypothesis All models reduce detail of natural world to focus on particular quantities and processes of interest

Some lingo and notation Prediction or dependent variable Parameter vector Independent variables

Deterministic Process Models We have a model g(θ) that describes the ecological process of interest: θ is a vector of model parameters and initial conditions. The model may (or may not) depend on x, a set covariates. The model takes values for θ and x and produces values (μ). For the same set of θ and x, the model returns the same values for μ.

What is this course about? A mathematical model of a process A statistical model of the data that arise from the process

Stochasticity Stochasticity refers to things that are unknown. Models are stochastic because we cannot specify everything (e.g., process error, observation error, individual variation). The elements that we leave out of models are treated stochastically. This creates uncertainties that must be quantified.

How do we represent unknowns? Process model Probability model represents stochasticity Data Parameters in the ecological process (deterministic) model Predictions of the ecological process model Parameters in probability model that describe the unknown

Stochastic Models In contrast to deterministic models that predict a scalar for any input, stochastic models predict probability distributions of values. μiμi P(μ i ) g (θ)

Random variables A random or stochastic variable is a quantity that can take on values due to chance. It does not have a single value but instead can take on a range of values, with its chance governed by a probability distribution. A random variable's possible values might represent the possible outcomes of a yet-to-be-performed experiment, or the possible outcomes of a past experiment whose already-existing value is uncertain (for example, as a result of incomplete information or imprecise measurements). Random variables have probability distributions which allow us to gain insight into unobserved quantities.

A general framework for stochasticity which is the same as:

Terminology This notation is equivalent. You will see it all in this course. These are general stand-ins for distributions.

Terminology in H&H This notation is equivalent. You will see it all in this course. These are general stand-ins for distributions.

Probability Concepts An experiment is an operation with uncertain outcome. A sample space is a set of all possible outcomes of an experiment. An event is the random variable, a particular outcome of an experiment, a subset of the sample space.

Concept of Probability Event A S= sample space Examples?

Conditional Probabilities Event A S= sample space Event B Probability of B given that we know A occurred:

Event A Event B Joint Probabilities S= sample space Probability of A and B : Is there another way to write P(A,B)?

Conditional Probabilities Event A S= sample space Event B Probability of A given that we know B occurred:

Conditional Probabilities Event A S= sample space Event B Probability of A given that we know B occurred:

A key concept The probability of the data point y i given the parameters, given that the values of θ are known or fixed. The probability of the data given the parameters.

Intuition for Likelihood Observed data S= sample space Predictions of model with parameters θ

Intuition for Likelihood Observed data S= sample space Predictions of model with new values for parameters θ Better? Worse?

is called a likelihood function (tomorrow).

Law of total probability

Factoring joint probabilities Event A S= sample space Event B What is the probability of A and B?

Factoring joint probabilities: Directed Acyclic Graphs DAG A B Parent node Child node

Specify how joint distributions are factored into conditional distributions. Nodes at head of arrows must be in the LHS of conditioning symbol (|); nodes at tails of the RHS. Example: P(A|,B.C) Nodes without incoming arrows must be expressed unconditionally (e.g., P(C)) Bayesian networks

Exercise What are the joint probabilities of A, B, C, D [and E]?

Exercise

What are the joint probabilities of A, B, C, D?

Exercise

These rules allow us to take complicated joint distributions of random variables and break them down into manageable chunks that can be analyzed one at a time as if all of the other random variables were known and constant. We will take this approach throughout the course. Factoring joint probabilities: Why do we care?

DAG: Hierarchical tree growth models  y x Hierarchical parameter model (species-specific  ’s) α Original model (common  ) Data (stochastic) Parameters Data (fixed)=diameter β  y x Observation level Species level Community level Can you write out the hierarchical model?

Event A Event B Marginal Probabilities S= sample space The marginal distribution of A is the probability of A averaged over probability of B

Why do we care about marginal distributions? Consider 2 jointly distributed random variables: number of offspring by age. Marginal distribution of y

Joint Marginal

Why do we care about marginal distributions? They allow us to represent the univariate distribution of unknown quantities that are parts of joint distributions that might contain many parameters and latent quantities. They are a vital tool for simplification Diamond’s pigeons…

Diamond’s pigeons (1975)

If probabilities of S and R are independent: Pr(R, S) = Pr(R) Pr(S) =11/32*20/32= Diamond interpreted this difference as evidence of niche separation resulting from interspecific competition. The conditional probabilities are: