Vector Generalized Additive Models and applications to extreme value analysis Olivier Mestre (1,2) (1) Météo-France, Ecole Nationale de la Météorologie,

Slides:



Advertisements
Similar presentations
A Tale of Two GAMs Generalized additive models as a tool for data exploration Mariah Silkey, Actelion Pharmacueticals Ltd. 1.
Advertisements

Additive Models, Trees, etc. Based in part on Chapter 9 of Hastie, Tibshirani, and Friedman David Madigan.
A. The Basic Principle We consider the multivariate extension of multiple linear regression – modeling the relationship between m responses Y 1,…,Y m and.
Copula Regression By Rahul A. Parsa Drake University &
Regression “A new perspective on freedom” TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAA A A A A AAA A A.
Generalized Additive Models Keith D. Holler September 19, 2005 Keith D. Holler September 19, 2005.
Copyright © 2006 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 ~ Curve Fitting ~ Least Squares Regression Chapter.
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
Ch11 Curve Fitting Dr. Deshi Ye
Model assessment and cross-validation - overview
Nonlinear Regression Ecole Nationale Vétérinaire de Toulouse Didier Concordet ECVPT Workshop April 2011 Can be downloaded at
Data mining and statistical learning - lecture 6
Basis Expansion and Regularization Presenter: Hongliang Fei Brian Quanz Brian Quanz Date: July 03, 2008.
Basis Expansion and Regularization
Kernel methods - overview
Linear Methods for Regression Dept. Computer Science & Engineering, Shanghai Jiao Tong University.
統計計算與模擬 政治大學統計系余清祥 2003 年 6 月 9 日 ~ 6 月 10 日 第十六週:估計密度函數
Lecture Notes for CMPUT 466/551 Nilanjan Ray
Curve-Fitting Regression
Missing at Random (MAR)  is unknown parameter of the distribution for the missing- data mechanism The probability some data are missing does not depend.
Linear Regression Models Based on Chapter 3 of Hastie, Tibshirani and Friedman Slides by David Madigan.
Basis Expansions and Regularization Based on Chapter 5 of Hastie, Tibshirani and Friedman.
Nonparametric Smoothing Methods and Model Selections T.C. Lin Dept. of Statistics National Taipei University 5/4/2005.
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
Statistical analysis and modeling of neural data Lecture 4 Bijan Pesaran 17 Sept, 2007.
Linear and generalised linear models
1 An Introduction to Nonparametric Regression Ning Li March 15 th, 2004 Biostatistics 277.
Linear and generalised linear models
Biostatistics-Lecture 14 Generalized Additive Models Ruibin Xi Peking University School of Mathematical Sciences.
Classification and Prediction: Regression Analysis
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Physics 114: Lecture 15 Probability Tests & Linear Fitting Dale E. Gary NJIT Physics Department.
Nonparametric Regression
Model Inference and Averaging
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Chapter 11 – Neural Networks COMP 540 4/17/2007 Derek Singer.
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
Data Mining Volinsky - Columbia University 1 Chapter 4.2 Regression Topics Credits Hastie, Tibshirani, Friedman Chapter 3 Padhraic Smyth Lecture.
Model Comparison for Tree Resin Dose Effect On Termites Lianfen Qian Florida Atlantic University Co-author: Soyoung Ryu, University of Washington.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Université d’Ottawa - Bio Biostatistiques appliquées © Antoine Morin et Scott Findlay :32 1 Logistic regression.
Additive Models , Trees , and Related Models Prof. Liqing Zhang Dept. Computer Science & Engineering, Shanghai Jiaotong University.
Developing long-term homogenized climate Data sets Olivier Mestre Météo-France Ecole Nationale de la Météorologie Université Paul Sabatier, Toulouse.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Gaussian Process and Prediction. (C) 2001 SNU CSE Artificial Intelligence Lab (SCAI)2 Outline Gaussian Process and Bayesian Regression  Bayesian regression.
Kernel Methods Arie Nakhmani. Outline Kernel Smoothers Kernel Density Estimators Kernel Density Classifiers.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Basis Expansions and Generalized Additive Models Basis expansion Piecewise polynomials Splines Generalized Additive Model MARS.
R. Kass/Sp07P416/Lecture 71 More on Least Squares Fit (LSQF) In Lec 5, we discussed how we can fit our data points to a linear function (straight line)
Chapter 11 Linear Regression and Correlation. Explanatory and Response Variables are Numeric Relationship between the mean of the response variable and.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Physics 114: Lecture 13 Probability Tests & Linear Fitting
Deep Feedforward Networks
Probability Theory and Parameter Estimation I
More General Need different response curves for each predictor
CSE 4705 Artificial Intelligence
The Maximum Likelihood Method
Machine learning, pattern recognition and statistical data modelling
The Maximum Likelihood Method
Linear regression Fitting a straight line to observations.
Biointelligence Laboratory, Seoul National University
Nonlinear Fitting.
More General Need different response curves for each predictor
Basis Expansions and Generalized Additive Models (2)
政治大學統計系余清祥 2004年5月26日~ 6月7日 第十六、十七週:估計密度函數
Generalized Additive Model
Presentation transcript:

Vector Generalized Additive Models and applications to extreme value analysis Olivier Mestre (1,2) (1) Météo-France, Ecole Nationale de la Météorologie, Toulouse, France (2) Université Paul Sabatier, LSP, Toulouse, France Based on previous studies realized in collaboration with : Stéphane Hallegatte (CIRED, Météo-France) Sébastien Denvil (LMD)

SMOOTHER « Smoother=tool for summarizing the trend of a response measurement Y as a function of predictors » (Hastie & Tibshirani) estimate of the trend that is less variable than Y itself  Smoothing matrix S Y*=S  Y The equivalent degrees of freedom (df) of the smoother S is the trace of S. Allows compare with parametric models.  Pointwise standard error bands COV(Y*)=V=S t S  ²given an estimation of  ², this allows approximate confidence intervals (values : ±2  square root of the diagonal of V)

SCATTERPLOT SMOOTHING EXAMPLE  Data: wind farm production vs numerical windspeed forecasts

SMOOTHING  Problems raised by smoothers How to average the response values in each neighborhood? How large to take the neighborhoods?  Tradeoff between bias and variance of Y*

SMOOTHING: POLYNOMIAL (parametric)  Linear and cubic parametric least squares fits: MODEL DRIVEN APPROACHES

SMOOTHING: BIN SMOOTHER  In this example, optimum intervals are determined by means of a regression tree

SMOOTHING: RUNNING LINE  Running line

KERNEL SMOOTHER  Watson-Nadaraya

SMOOTHING: LOESS  The smooth at the target point is the fit of a locally-weighted linear fit (tricube weight)

CUBIC SMOOTHING SPLINES  This smoother is the solution of the following optimization problem: among all functions f(x) with two continuous derivatives, choose the one that minimizes the penalized sum of squares Closeness to the datapenalization of the curvature of f It can be shown that the unique solution to this problem is a natural cubic spline with knots at the unique values x i Parameter can be set by means of cross-validation

CUBIC SMOOTHING SPLINES  Cubic smoothing splines with equivalent df=5 and 10

Additive models  Gaussian Linear Model:IE[Y]=  o +  1 X 1 +  2 X 2  Gaussian Additive model:IE[Y]=S 1 (X 1 )+S 2 (X 2 ) S 1, S 2 smooth functions of predictors X 1, X 2, usually LOESS, SPLINE Estimation of S 1, S 2 : « Backfitting Algorithm »  PRINCIPLE OF THE BACKFITTING ALGORITHM Y=S 1 (X 1 )+e  estimation S 1 * Y-S 1 *(X 1 )=S 2 (X 2 )+e  estimation S 2 * Y-S 2 *(X 2 )=S 1 (X 1 )+e  estimation S 1 ** Y-S 1 **(X 1 )=S 2 (X 2 )+e  estimation S 2 ** Y-S 2 **(X 2 )=S 1 (X 1 )+e  estimation S 1 *** Etc… until convergence

Additive models  Additive models One efficient way to perform non-linear regression, but…  Crucial point ADAPTED WHEN ONLY FEW PREDICTORS 2, 3 predictors at most

Additive models  Philosophy DATA DRIVEN APPROACHES RATHER THAN MODEL DRIVEN APPROACH USEFUL AS EXPLORATORY TOOLS  Approximate inference tests are possible, but full inferences are better assessed by means of parametric models

Generalized Additive models (GAM)  Extension to non-normal dependant variables  Generalized additive models: additive modelling of the natural parameter of exponential family laws (Poisson, Binomial, Gamma, Gauss…). g[µ]=  =S 1 (X 1 )+S 2 (X 2 )  Vector Generalized Additive Models (VGAM): one step beyond…

Example 1 Annual umber and maximum integrated intensity (PDI) of hurricane tracks over the North Atlantic

Number of Hurricanes  Number of Hurricanes in North Atlantic ~ Poisson distribution

Factors influencing the number of hurricanes  GAM applied to number of hurricanes (YEAR,SST,SOI,NAO)

GAM model  Log( )=  o +S 1 (SST)+S 2 (SOI)

PARAMETRIC model  “broken stick model” (with continuity constraint) in SOI, revealed by GAM analysis  log( )=  o+  SOI (1) SOI+  SST SST SOI<K =  o+  SOI (1) SOI+  SOI (2) (SOI-K)+  SST SST SOI  K  The best fit obtained for SOI value K=1 log-likelihood= , to be compared with (linearity) standard deviance test allows reject linearity (p value=0.02)  Expectation of the hurricane number is then straightforwardly computed as a function of SOI and SST

EXPECTATION OF HURRICANE NUMBERS

OBSERVED vs EXPECTED: r=0.6