Zooming in on Probability and Causality (aim & preliminary results)

Slides:



Advertisements
Similar presentations
Review of Probability.
Advertisements

Random Variables ECE460 Spring, 2012.
Week 11 Review: Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution.
9.2a Tests about a Population Proportion Target Goal: I can check the conditions for carrying out a test about a population proportion. I can perform a.
Normal Distribution; Sampling Distribution; Inference Using the Normal Distribution ● Continuous and discrete distributions; Density curves ● The important.
Bayesian inference Gil McVean, Department of Statistics Monday 17 th November 2008.
Probability theory Much inspired by the presentation of Kren and Samuelsson.
A/Prof Geraint Lewis A/Prof Peter Tuthill
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Confidence intervals.
Evaluating Hypotheses
The Analysis of Variance
8-1 Introduction In the previous chapter we illustrated how a parameter can be estimated from sample data. However, it is important to understand how.
Probability and the Sampling Distribution Quantitative Methods in HPELS 440:210.
Chapter 1: Introduction to Statistics
Chapter 1: Introduction to Statistics
Cohort Study.
L7.1b Continuous Random Variables CONTINUOUS RANDOM VARIABLES NORMAL DISTRIBUTIONS AD PROBABILITY DISTRIBUTIONS.
20-Jan-2010electrical, computer and energy engineering Prof. Subramaniam (“Subby”) D. Rajan, Prof. Narayanan Neithalath and Amie Baisley Graduate Students:
Chapter 1: Introduction to Statistics
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Probability Distributions - Discrete Random Variables Outcomes and Events.
Education Research 250:205 Writing Chapter 3. Objectives Subjects Instrumentation Procedures Experimental Design Statistical Analysis  Displaying data.
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
Ch 8 Estimating with Confidence. Today’s Objectives ✓ I can interpret a confidence level. ✓ I can interpret a confidence interval in context. ✓ I can.
1 Chapters 6-8. UNIT 2 VOCABULARY – Chap 6 2 ( 2) THE NOTATION “P” REPRESENTS THE TRUE PROBABILITY OF AN EVENT HAPPENING, ACCORDING TO AN IDEAL DISTRIBUTION.
1 Lecture 4. 2 Random Variables (Discrete) Real-valued functions defined on a sample space are random vars. determined by outcome of experiment, we can.
Geo597 Geostatistics Ch9 Random Function Models.
Introduction to Behavioral Statistics Probability, The Binomial Distribution and the Normal Curve.
Conceptual Modelling and Hypothesis Formation Research Methods CPE 401 / 6002 / 6003 Professor Will Zimmerman.
Maximum Likelihood - "Frequentist" inference x 1,x 2,....,x n ~ iid N( ,  2 ) Joint pdf for the whole random sample Maximum likelihood estimates.
Confidence Intervals: The Basics BPS chapter 14 © 2006 W.H. Freeman and Company.
Sampling Distribution Models Chapter 18. Toss a penny 20 times and record the number of heads. Calculate the proportion of heads & mark it on the dot.
Confidence intervals: The basics BPS chapter 14 © 2006 W.H. Freeman and Company.
Hypothesis testing and effect sizes Sylvain Chartier Laboratory for Computational Neurodynamics and Cognition Centre for Neural Dynamics.
Confidence Interval Estimation For statistical inference in decision making:
Lecture 2 Molecular dynamics simulates a system by numerically following the path of all particles in phase space as a function of time the time T must.
Chapter 3 Discrete Random Variables and Probability Distributions  Random Variables.2 - Probability Distributions for Discrete Random Variables.3.
POSC 202A: Lecture 4 Probability. We begin with the basics of probability and then move on to expected value. Understanding probability is important because.
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Summarizing Risk Analysis Results To quantify the risk of an output variable, 3 properties must be estimated: A measure of central tendency (e.g. µ ) A.
Statistical Inference Drawing conclusions (“to infer”) about a population based upon data from a sample. Drawing conclusions (“to infer”) about a population.
Discrete Random Variables. Introduction In previous lectures we established a foundation of the probability theory; we applied the probability theory.
Confidence Interval Estimation For statistical inference in decision making: Chapter 9.
ECE-7000: Nonlinear Dynamical Systems 2. Linear tools and general considerations 2.1 Stationarity and sampling - In principle, the more a scientific measurement.
Ch 8 Estimating with Confidence 8.1: Confidence Intervals.
From the population to the sample The sampling distribution FETP India.
CHAPTER 2.3 PROBABILITY DISTRIBUTIONS. 2.3 GAUSSIAN OR NORMAL ERROR DISTRIBUTION  The Gaussian distribution is an approximation to the binomial distribution.
In Bayesian theory, a test statistics can be defined by taking the ratio of the Bayes factors for the two hypotheses: The ratio measures the probability.
President UniversityErwin SitompulPBST 10/1 Lecture 10 Probability and Statistics Dr.-Ing. Erwin Sitompul President University
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
Sampling Distributions Chapter 18. Sampling Distributions A parameter is a number that describes the population. In statistical practice, the value of.
Random Variables and Probability Distributions. Definition A random variable is a real-valued function whose domain is the sample space for some experiment.
Theoretical distributions: the Normal distribution.
Aim – How do we analyze a Discrete Random Variable?
1 Chapter 1: Introduction to Statistics. 2 Variables A variable is a characteristic or condition that can change or take on different values. Most research.
AP Statistics: Chapter 7
Confidence Intervals Chapter 10 Section 1.
Sets and Probabilistic Models
Berlin Chen Department of Computer Science & Information Engineering
Discrete Random Variables: Basics
Berlin Chen Department of Computer Science & Information Engineering
Sets and Probabilistic Models
Discrete Random Variables: Joint PMFs, Conditioning and Independence
Discrete Random Variables: Basics
Experiments, Outcomes, Events and Random Variables: A Revisit
Sets and Probabilistic Models
Berlin Chen Department of Computer Science & Information Engineering
Discrete Random Variables: Basics
Sets and Probabilistic Models
Presentation transcript:

Zooming in on Probability and Causality (aim & preliminary results)

A B D C E  ACB A C D + + CB ED Motivation - ‘Building Causal Models from Multiple Databases’ Results from three experiments … One model. (under suitable assumptions)  But how can we verify the validity of causal claims we make?

Causal structures Problem: Verification implies somehow comparing with the ‘true’ causal structure of a system. For a system that is described in terms of (causal) relations, e.g. SEMs and CBNs, this is fairly straightforward: the system is the model. But most real world systems are ultimately dynamical systems, governed by differential equations that describe the detailed evolution through phase space, and at that ground level it is no longer clear where the probabilities and causal relations actually reside. Main Question: What, if anything, actually is ‘the causal structure’ of a system? Approach: Look in detail at a (simple) dynamical system for which you have complete information and try to see if you can understand probability and causality at this level.

Representing a system Three crucial ingredients to capture: A simple dynamical system (T,S,f) T - a time interval S - the state space of the entire system f - a set of coupled differential equations with initial values A set R of (random) variables V = {v 1, …, v n } v i - set of possible outcomes of V A context ( ,D) for the system  - set of possible/allowed initial conditions D - relative weight distribution over  Then we have complete information about the (deterministic) system. Note: (fixed) system parameters can also be included in S and .

Sample system - Throwing a drawing pin Two (stable) outcomes Questions: What is the probability of ‘pin up’? What causes (i.e. effectively influences) this probability? What is the causal structure between various quantities (i.e. parameters, intermediate states and/or other random variables)? Point is: can we calculate the answer these questions from the description alone? Experimental setup

Preliminary results - Probability Defining a quantity P representing the relative proportion of initial states in a given context ( ,D) that lead to the outcome V = v i, as Then this proportion turns out to correspond very closely to our everyday notion of probability. It behaves according to the Kolmogorov axioms In the Bayesian interpretation the context corresponds to the prior (though not as ‘belief’) In a frequentist setting the context represents the average, infinite limit properties of the normal, ‘default’ throw (only now made explicit). But, It does make probability a fundamentally relative notion … except for degenerate cases where there is an inherent symmetry in the system and context, e.g. fairly tossing a fair coin; in that case the proportion is not dependent on the context.

Preliminary results - Causal relations Now what about causality? Take ‘effective manipulability’ as a prime characteristic of a causal relation. If probability relates to proportions of trajectories through phase space, then causality relates to changes in these proportions with respect to changes in system parameters, initial and/or intermediate states of the system. However: Causal nodes have no clearly defined boundaries Causal nodes change with every random variable Causal nodes depend on the level of detail in the manipulation Making causal relations also a fundamentally relative notion, without a clear prospect of obtaining ‘the’ causal structure of a system … because in reality it is much more complicated, … or is it? Hopeful for obtaining a causal structure: Causal nodes represent areas in phase space where these changes are effectively induced (e.g. parameter pin length on ‘pin up’) These areas in phase space are connected via the system trajectories Random variables can represent intermediate stages of trajectories