Generalized Additive Model

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Linear Regression.
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Ch11 Curve Fitting Dr. Deshi Ye
Simple Linear Regression

Missing at Random (MAR)  is unknown parameter of the distribution for the missing- data mechanism The probability some data are missing does not depend.
Generalised linear models
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Log-linear and logistic models Generalised linear model ANOVA revisited Log-linear model: Poisson distribution logistic model: Binomial distribution Deviances.
Log-linear and logistic models
Generalised linear models Generalised linear model Exponential family Example: Log-linear model - Poisson distribution Example: logistic model- Binomial.
Probability & Statistics for Engineers & Scientists, by Walpole, Myers, Myers & Ye ~ Chapter 11 Notes Class notes for ISE 201 San Jose State University.
Linear and generalised linear models
Chi Square Distribution (c2) and Least Squares Fitting
Linear and generalised linear models
Correlation and Regression Analysis
Review of Lecture Two Linear Regression Normal Equation
Objectives of Multiple Regression
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Chapter 3: Generalized Linear Models 3.1 The Generalization 3.2 Logistic Regression Revisited 3.3 Poisson Regression 1.
Inferences in Regression and Correlation Analysis Ayona Chatterjee Spring 2008 Math 4803/5803.
Introduction to Generalized Linear Models Prepared by Louise Francis Francis Analytics and Actuarial Data Mining, Inc. October 3, 2004.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
Generalized Linear Models All the regression models treated so far have common structure. This structure can be split up into two parts: The random part:
General Linear Models; Generalized Linear Models Hal Whitehead BIOL4062/5062.
Linear Model. Formal Definition General Linear Model.
Logistic Regression Database Marketing Instructor: N. Kumar.
Forecasting Choices. Types of Variable Variable Quantitative Qualitative Continuous Discrete (counting) Ordinal Nominal.
1 GLM I: Introduction to Generalized Linear Models By Curtis Gary Dean Distinguished Professor of Actuarial Science Ball State University By Curtis Gary.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Model Building and Model Diagnostics Chapter 15.
Generalized Linear Models (GLMs) and Their Applications.
Machine Learning 5. Parametric Methods.
Curve Fitting Introduction Least-Squares Regression Linear Regression Polynomial Regression Multiple Linear Regression Today’s class Numerical Methods.
Variance Stabilizing Transformations. Variance is Related to Mean Usual Assumption in ANOVA and Regression is that the variance of each observation is.
Richard Kass/F02P416 Lecture 6 1 Lecture 6 Chi Square Distribution (  2 ) and Least Squares Fitting Chi Square Distribution (  2 ) (See Taylor Ch 8,
Topics, Summer 2008 Day 1. Introduction Day 2. Samples and populations Day 3. Evaluating relationships Scatterplots and correlation Day 4. Regression and.
Remember the equation of a line: Basic Linear Regression As scientists, we find it an irresistible temptation to put a straight line though something that.
Logistic Regression: Regression with a Binary Dependent Variable.
Estimating standard error using bootstrap
Transforming the data Modified from:
Physics 114: Lecture 13 Probability Tests & Linear Fitting
Chapter 7. Classification and Prediction
Regression Analysis AGEC 784.
Probability Theory and Parameter Estimation I
Statistics for the Social Sciences
Non-linear relationships
Generalized Linear Models
Ch3: Model Building through Regression
Statistics in MSmcDESPOT
Drop-in Sessions! When: Hillary Term - Week 1 Where: Q-Step Lab (TBC) Sign up with Alice Evans.
Correct statistics in ecological research
Generalized Linear Models (GLM) in R
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Chapter 12 Curve Fitting : Fitting a Straight Line Gab-Byung Chae
Chi Square Distribution (c2) and Least Squares Fitting
Geology Geomath Chapter 7 - Statistics tom.h.wilson
Model Comparison: some basic concepts
Linear regression Fitting a straight line to observations.
10701 / Machine Learning Today: - Cross validation,
What is Regression Analysis?
Simple Linear Regression
Fixed, Random and Mixed effects
Model generalization Brief summary of methods
Parametric Methods Berlin Chen, 2005 References:
Chapter 6 Logistic Regression: Regression with a Binary Dependent Variable Copyright © 2010 Pearson Education, Inc., publishing as Prentice-Hall.
Multiple Regression Berlin Chen
Presentation transcript:

Generalized Additive Model Shadi Ghasemi 7/27/2019

What is GAM? 7/27/2019

General Linear Model We have some response/variable we wish to study, and believe it to be some function of other variables. y is the variable of interest, assumed to y be normally distributed with mean  and variance σ 2 , and the Xs are the predictor variables/covariates in this scenario. The predictors are multiplied by the coefficients (beta) and summed, giving us the linear predictor, which in this case also directly provides us the estimated fitted values. 7/27/2019

in its basic form it can be very limiting in its assumptions about the data generating process for the variable we wish to study. 7/27/2019

Generalized Linear Model Generalized linear models incorporate other types of distributions (Of the exponential family), and include a link function g(.) relating the mean µ, or stated differently, the estimated fitted values E(y), to the linear predictor Xb, The general form is thus 7/27/2019

In Linear Model we assume a Gaussian (i. e In Linear Model we assume a Gaussian (i.e. Normal) distribution for the response, we assume equal variance for all observations, and that there is a direct link of the linear predictor and the expected value , i.e. Xb = . In fact the typical linear regression model is a generalized linear model with a Gaussian distribution and identity link function. 7/27/2019

Unlike classical linear models, which presuppose a Gaussian (i. e Unlike classical linear models, which presuppose a Gaussian (i.e., normal) distribution and an identity link, the distribution of Y in a GLMs may be any of the exponential family distributions (e.g., Gaussian, Poisson or binomial) and the link function may be any monotonic differentiable function (like logarithm or logit). 7/27/2019

The main improvements of GLMs over LS regression are: (1) the ability to handle a larger class of distributions for the response variable Y. (2) the relationship of the response variable Y to the linear predictor (LP) through the link function g(E(Y)). (3) it incorporates potential solutions (like quasi-likelihood) to deal with overdispersion 7/27/2019

Generalized Additive Model another generalization to incorporate nonlinear forms of the predictors. gives the new setup relating our new, now nonlinear predictor to the expected value, with whatever link function may be appropriate. 7/27/2019

GAMs are parameterized just like GLMs, except that some predictors can be modeled nonparametrically in addition to linear and polynomial terms for other predictors. The probability distribution of the response variable must still be specified, and in this respect, a GAM is parametric. In this sense they are more aptly named semi-parametric models. 7/27/2019

7/27/2019

One can envision the different regression models as being nested within each other, with simple and multiple LS linear regression (SLR and MLR) being the two most limiting cases, and GAMs the most general: 7/27/2019

A crucial step in applying GAMs is to select the appropriate level of the ‘‘smoother’’ for a predictor. This is best achieved by specifying the level of smoothing using the concept of effective degrees of freedom. A reasonable balance must be maintained between the total number of observations and the total number of degrees of freedom used when fitting the model (sum of levels of smoothness used for each predictor). 7/27/2019

What is a Smoother? 7/27/2019

7/27/2019

7/27/2019

For comparison the regular regression fit is also provided. the result from such a fitting process, specifically lowess, or locally weighted scatterplot smoothing. For comparison the regular regression fit is also provided. 7/27/2019

This is the biggest difference from Generalized Linear Model (GLM) This is the biggest difference from Generalized Linear Model (GLM). It allowes an ‘approximation’ with sum of functions, (these functions have separated input variables), not just with one unknown function only. 7/27/2019

Linear functions have , while nonlinear functions produce non-zero values. 7/27/2019

Fitting the Generalized Additive Model There are a number of approaches for the formulation and estimation of additive models. The back-fitting algorithm is a general algorithm that can fit an additive model using any regression-type fitting mechanism. 7/27/2019

Variable selection methods and diagnostics Variable selection is basically the same for all the described regression models, although evaluation criteria like the Akaike Information Criterion can be used with GLMs and GAMs. In all models, one can use predefined rules such as deviance reduction as measured with the chi-square statistic, or approaches that minimize AIC. 7/27/2019

Comparison between GLM and GAM 7/27/2019

7/27/2019

7/27/2019

7/27/2019

7/27/2019

7/27/2019

7/27/2019

7/27/2019