Introduction We consider the data of ~1800 phenotype measurements Each mouse has a given probability distribution of descending from one of 8 possible.

Slides:



Advertisements
Similar presentations
Copula Representation of Joint Risk Driver Distribution
Advertisements

Pattern Recognition and Machine Learning
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Image Modeling & Segmentation
Copula Regression By Rahul A. Parsa Drake University &
Uncertainty in fall time surrogate Prediction variance vs. data sensitivity – Non-uniform noise – Example Uncertainty in fall time data Bootstrapping.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
The General Linear Model. The Simple Linear Model Linear Regression.
DATA ANALYSIS Module Code: CA660 Lecture Block 6: Alternative estimation methods and their implementation.
The loss function, the normal equation,
Multivariate linear models for regression and classification Outline: 1) multivariate linear regression 2) linear classification (perceptron) 3) logistic.
Visual Recognition Tutorial
Classification and risk prediction
Maximum likelihood (ML) and likelihood ratio (LR) test
Descriptive statistics Experiment  Data  Sample Statistics Sample mean Sample variance Normalize sample variance by N-1 Standard deviation goes as square-root.
An Optimal Learning Approach to Finding an Outbreak of a Disease Warren Scott Warren Powell
Simulation Modeling and Analysis Session 12 Comparing Alternative System Designs.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Gaussian Mixture Example: Start After First Iteration.
Part 4 c Baum-Welch Algorithm CSE717, SPRING 2008 CUBS, Univ at Buffalo.
1 QTL mapping in mice, cont. Lecture 11, Statistics 246 February 26, 2004.
Maximum Likelihood (ML), Expectation Maximization (EM)
Linear and generalised linear models
Linear and generalised linear models
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Maximum likelihood (ML)
Model Inference and Averaging
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Bayesian inference review Objective –estimate unknown parameter  based on observations y. Result is given by probability distribution. Bayesian inference.
MULTIPLE TRIANGLE MODELLING ( or MPTF ) APPLICATIONS MULTIPLE LINES OF BUSINESS- DIVERSIFICATION? MULTIPLE SEGMENTS –MEDICAL VERSUS INDEMNITY –SAME LINE,
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Generalized Linear Models All the regression models treated so far have common structure. This structure can be split up into two parts: The random part:
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Conjugate Priors Multinomial Gaussian MAP Variance Estimation Example.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Forecasting Choices. Types of Variable Variable Quantitative Qualitative Continuous Discrete (counting) Ordinal Nominal.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
: Chapter 3: Maximum-Likelihood and Baysian Parameter Estimation 1 Montri Karnjanadecha ac.th/~montri.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Generalized Linear Models (GLMs) and Their Applications.
Lecture 2: Statistical learning primer for biologists
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Simulation Study for Longitudinal Data with Nonignorable Missing Data Rong Liu, PhD Candidate Dr. Ramakrishnan, Advisor Department of Biostatistics Virginia.
5. Maximum Likelihood –II Prof. Yuille. Stat 231. Fall 2004.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
Over-fitting and Regularization Chapter 4 textbook Lectures 11 and 12 on amlbook.com.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
Tutorial I: Missing Value Analysis
Lecture 23: Quantitative Traits III Date: 11/12/02  Single locus backcross regression  Single locus backcross likelihood  F2 – regression, likelihood,
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Statistical Significance Hypothesis Testing.
Hidden Markov Models. A Hidden Markov Model consists of 1.A sequence of states {X t |t  T } = {X 1, X 2,..., X T }, and 2.A sequence of observations.
Learning Theory Reza Shadmehr Distribution of the ML estimates of model parameters Signal dependent noise models.
Two Types of Empirical Likelihood Zheng, Yan Department of Biostatistics University of California, Los Angeles.
Hierarchical Mixture of Experts Presented by Qi An Machine learning reading group Duke University 07/15/2005.
Probability and Likelihood. Likelihood needed for many of ADMB’s features Standard deviation Variance-covariance matrix Profile likelihood Bayesian MCMC.
Applied statistics Usman Roshan.
BINARY LOGISTIC REGRESSION
Probability Theory and Parameter Estimation I
دانشگاه صنعتی امیرکبیر Instructor : Saeed Shiry
6-1 Introduction To Empirical Models
The regression model in matrix form
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
10701 / Machine Learning Today: - Cross validation,
EM for Inference in MV Data
Parametric Methods Berlin Chen, 2005 References:
EM for Inference in MV Data
Presentation transcript:

Introduction We consider the data of ~1800 phenotype measurements Each mouse has a given probability distribution of descending from one of 8 possible strains We model where Y is phenotype, X – “genotype” matrix, π – probability of descending from strains, z – flanking markers The objective is to estimate β’s and to test for

Linear regression replace the true design matrix with E(X) and the estimator is given by - estimator is unbiased - normally distributed - linear in Y - has large variance due to collinearity Approaches

Approaches (2) Maximum likelihood estimator Maximize with respect to β: - expression simplifies - easy to evaluate point-wise - functional form not known, hence difficult to optimize - properties of the MLE are unknown

The two steps are: E-step, calculate for i-th mouse (only for categorical covariates) M-step, maximize Q w.r.t. β Advantages: Automatic Fast Approximate distribution of estimates allows to perform testing Easily generalised to GLM Approaches (3) Use a stochastic optimiser for finding MLE: the EM

M-step becomes equivalent to a Weighted Least Squares or a weighted GLM model (fitting routines available in R and Matlab): Where Y and X are augmented matrices, the weights matrix constructed using HMM output. Below there are only results for normal distribution of Y but the EM was applied to the binomial and exponential cases as well. Implementation of the EM

Given the phenotypes Y and the weights W we create the model: Augmenting the model with corresponding weights

Simulated example: generated phenotypes Response generated for set variance 0.3 and β = (1,0,0,0,0,0,0,0)

Values of β parameters at the EM iterations. The real values are (1,0,0,0,0,0,0,0). Running the EM

10 seconds - approximate running time for the WLS case - on 1,649 mice - implemented in Matlab - with convergence achieved at 15 iterations for some starting points 60 seconds - For 3,298 mice Running time

Likelihood ratio test performed for - the EM - linear regression with known design matrix - linear regression with the expectation of design matrix. Testing under collinearity

E(X) case null distribution Empirical null distributions EM algorithm null distribution

Description of the power of the LR test All β’s set to 0 except first one Simulate data sets and plot number of rejections For each value of β 500 simulations were performed Power curves

Simulated power curves Most likely combination of progenitor strains Randomly drawn combination of progenitor strains Least likely combination of progenitor strains

Considered OpenArmTime phenotype ~200 mice have zero records and were removed Is it a mixture of normal distributions? Data

Time to event models - Censored data - Cox proportional hazards model Bayesian models Implementation in R Models for multivariate phenotypes Multiple hypothesis testing HMM improvement Future development