1 General Iteration Algorithms by Luyang Fu, Ph. D., State Auto Insurance Company Cheng-sheng Peter Wu, FCAS, ASA, MAAA, Deloitte Consulting LLP 2007 CAS.

Slides:



Advertisements
Similar presentations
Value-at-Risk: A Risk Estimating Tool for Management
Advertisements

Advanced Piloting Cruise Plot.
1 Copyright © 2010, Elsevier Inc. All rights Reserved Fig 2.1 Chapter 2.
STATISTICS Sampling and Sampling Distributions
STATISTICS HYPOTHESES TEST (III) Nonparametric Goodness-of-fit (GOF) tests Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering.
Linearization Variance Estimators for Survey Data: Some Recent Work
FDA/Industry Workshop September, 19, 2003 Johnson & Johnson Pharmaceutical Research and Development L.L.C. 1 Uses and Abuses of (Adaptive) Randomization:
1 ESTIMATION IN THE PRESENCE OF TAX DATA IN BUSINESS SURVEYS David Haziza, Gordon Kuromi and Joana Bérubé Université de Montréal & Statistics Canada ICESIII.
Analysis of Algorithms
Business Transaction Management Software for Application Coordination 1 Business Processes and Coordination.
SMA 6304 / MIT / MIT Manufacturing Systems Lecture 11: Forecasting Lecturer: Prof. Duane S. Boning Copyright 2003 © Duane S. Boning. 1.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Title Subtitle.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Multiplying binomials You will have 20 seconds to answer each of the following multiplication problems. If you get hung up, go to the next problem when.
0 - 0.
DIVIDING INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
SUBTRACTING INTEGERS 1. CHANGE THE SUBTRACTION SIGN TO ADDITION
MULT. INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
Addition Facts
Year 6 mental test 5 second questions
MCMC estimation in MlwiN
ZMQS ZMQS
STATISTICAL INFERENCE ABOUT MEANS AND PROPORTIONS WITH TWO POPULATIONS
Chapter 7 Sampling and Sampling Distributions
The Maximum Likelihood Method
ANALYZING AND ADJUSTING COMPARABLE SALES Chapter 9.
On Comparing Classifiers : Pitfalls to Avoid and Recommended Approach
ABC Technology Project
Factors, Prime Numbers & Composite Numbers
Environmental Data Analysis with MatLab
Chapter 4 Inference About Process Quality
Scale Free Networks.
Squares and Square Root WALK. Solve each problem REVIEW:
Do you have the Maths Factor?. Maths Can you beat this term’s Maths Challenge?
Lets play bingo!!. Calculate: MEAN Calculate: MEDIAN
Lecture Unit Multiple Regression.
Chapter 5 Test Review Sections 5-1 through 5-4.
GG Consulting, LLC I-SUITE. Source: TEA SHARS Frequently asked questions 2.
Chapter 5 Discrete Probability Distributions 1 Copyright © 2012 The McGraw-Hill Companies, Inc. Permission required for reproduction or display.
Addition 1’s to 20.
25 seconds left…...
6.4 Logarithmic Functions
Determining How Costs Behave
Week 1.
We will resume in: 25 Minutes.
Chapter 15: Quantitatve Methods in Health Care Management Yasar A. Ozcan 1 Chapter 15. Simulation.
CPSC 322, Lecture 14Slide 1 Local Search Computer Science cpsc322, Lecture 14 (Textbook Chpt 4.8) Oct, 5, 2012.
Chapter 8 Estimation Understandable Statistics Ninth Edition
Copyright © Cengage Learning. All rights reserved.
PSSA Preparation.
IP, IST, José Bioucas, Probability The mathematical language to quantify uncertainty  Observation mechanism:  Priors:  Parameters Role in inverse.
1 PART 1 ILLUSTRATION OF DOCUMENTS  Brief introduction to the documents contained in the envelope  Detailed clarification of the documents content.
Simple Linear Regression Analysis
How Cells Obtain Energy from Food
Multiple Regression and Model Building
What is the experimental unit in premix bioequivalence ? June 2010 Didier Concordet
The Student Handbook to T HE A PPRAISAL OF R EAL E STATE 1 Chapter 23 Yield Capitalization — Theory and Basic Applications.
Chapter 5 The Mathematics of Diversification
1 ECE 776 Project Information-theoretic Approaches for Sensor Selection and Placement in Sensor Networks for Target Localization and Tracking Renita Machado.
Lecture 14 Nonlinear Problems Grid Search and Monte Carlo Methods.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Severity Distributions for GLMs: Gamma or Lognormal? Presented by Luyang Fu, Grange Mutual Richard Moncher, Bristol West 2004 CAS Spring Meeting Colorado.
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Generalized Minimum Bias Models
5. Maximum Likelihood –II Prof. Yuille. Stat 231. Fall 2004.
Practical GLM Analysis of Homeowners David Cummings State Farm Insurance Companies.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Presentation transcript:

1 General Iteration Algorithms by Luyang Fu, Ph. D., State Auto Insurance Company Cheng-sheng Peter Wu, FCAS, ASA, MAAA, Deloitte Consulting LLP 2007 CAS Predictive Modeling Seminar Las Vegas, Oct 11-12, 2007

2 Agenda  History and Overview of Minimum Bias Method  General Iteration Algorithms (GIA)  Conclusions  Demonstration of a GIA Tool  Q&A

3 History on Minimum Bias  A technique with long history for actuaries: Bailey and Simon (1960) Bailey (1963) Brown (1988) Feldblum and Brosius (2002) A topic in CAS Exam 9  Concepts: Derive multivariate class plan parameters by minimizing a specified “bias” function Use an “iterative” method in finding the parameters

4 History on Minimum Bias  Various bias functions proposed in the past for minimization  Examples of multiplicative bias functions proposed in the past:

5 History on Minimum Bias  Then, how to determine the class plan parameters by minimizing the bias function?  One simple way is the commonly used an “iterative” methodology for root finding: Start with a random guess for the values of x i and y j Calculate the next set of values for x i and y j using the root finding formula for the bias function Repeat the steps until the values converge  Easy to understand and can be programmed in almost any tool

6 History on Minimum Bias  For example, using the balanced bias functions for the multiplicative model:

7 History on Minimum Bias  Past minimum bias models with the iterative method:

8 Iteration Algorithm for Minimum Bias  Theoretically, is not “bias”.  Bias is defined as the difference between an estimator and the true value. For example, is bias. If, then xhat is an unbiased estimator of x.  To be consistent with statistical terminology, we name our approach as General Iteration Algorithm.

9 Issues with the Iterative Method  Two questions regarding the “iterative” method: How do we know that it will converge? How fast/efficient that it will converge?  Answers: Numerical Analysis or Optimization textbooks Mildenhall (1999)  Efficiency is a less important issue due to the modern computation power

10 Other Issues with Minimum Bias  What is the statistical meaning behind these models?  More models to try?  Which models to choose?

11 Summary on Historical Minimum Bias  A numerical method, not a statistical approach  Best answers when bias functions are minimized  Use of an “iterative” methodology for root finding in determining parameters  Easy to understand and can be programmed in many tools

12 Connection Between Minimum Bias and Statistical Models  Brown (1988) Show that some minimum bias functions can be derived by maximizing the likelihood functions of corresponding distributions Propose several more minimum bias models  Mildenhall (1999) Prove that minimum bias models with linear bias functions are essentially the same as those from Generalized Linear Models (GLM) Propose two more minimum bias models

13 Connection Between Minimum Bias and Statistical Models  Past minimum bias models and their corresponding statistical models

14 Statistical Models - GLM  Advantages include : Commercial software and built-in procedures available Characteristics well determined, such as confidence level Computation efficiency compared to the iterative procedure

15 Statistical Models - GLM  Issues include: Requires more advanced knowledge of statistics for GLM models Lack of flexibility:  Reliance on commercial software / built-in procedures.  Cannot do the mixed model.  Assumes a pre-determined distribution of exponential families.  Limited distribution selections in popular statistical software.  Difficult to program from scratch.

16 Motivations for GIA  Can we unify all the past minimum bias models?  Can we completely represent the wide range of GLM and statistical models using Minimum Bias Models?  Can we expand the model selection options that go beyond all the currently used GLM and minimum bias models?  Can we fit mixed models or constraint models?

17 General Iteration Algorithm  Starting with the basic multiplicative formula  The alternative estimates of x and y:  The next question is – how to roll up x i,j to x i, and y j,i to y j ?

18 Possible Weighting Functions  First and the obvious option - straight average to roll up  Using the straight average results in the Exponential model by Brown (1988)

19 Possible Weighting Functions  Another option is to use the relativity-adjusted exposure as weight function  This is Bailey (1963) model, or Poisson model by Brown (1988).

20 Possible Weighting Functions  Another option: using the square of relativity-adjusted exposure  This is the normal model by Brown (1988).

21 Possible Weighting Functions  Another option: using relativity-square-adjusted exposure  This is the least-square model by Brown (1988).

22 General Iteration Algorithms  So, the key for generalization is to apply different “weighting functions” to roll up x i,j to x i and y j,i to y j  Propose a general weighting function of two factors, exposure and relativity: W p X q and W p Y q  Almost all published to date minimum bias models are special cases of GMBM(p,q)  Also, there are more modeling options to choose since there is no limitation, in theory, on (p,q) values to try in fitting data – comprehensive and flexible

23 2-parameter GIA  2-parameter GIA with exposure and relativity adjusted weighting function are:

24 2-parameter GIA vs. GLM pqGLM 1Inverse Gaussian 10Gamma 11Poisson 12Normal

25 2-parameter GIA and GLM  GMBM with p=1 is the same as GLM model with the variance function of  Additional special models: 0<q<1, the distribution is Tweedie, for pure premium models 1<q<2, not exponential family -1<q<0, the distribution is between gamma and inverse Gaussian  After years of technical development in GLM and minimum bias, at the end of day, all of these models are connected through the game of “weighted average”.

26 3-parameter GIA  One model published to date not covered by the 2- parameter GMBM: Chi-squared model by Bailey and Simon (1960)  Further generalization using a similar concept of link function in GLM, f(x) and f(y)  Estimate f(x) and f(y) through the iterative method  Calculate x and y by inverting f(x) and f(y)

27 3-parameter GIA

28 3-parameter GIA  Propose 3-parameter GMBM by using the power link function f(x)=x k

29 3-parameter GIA  When k=2, p=1 and q=1  This is the Chi-Square model by Bailey and Simon (1960)  The underlying assumption of Chi-Square model is that r 2 follows a Tweedie distribution with a variance function

30 Additive GIA

31 Mixed GIA  For commonly used personal line rating structures, the formula is typically a mixed multiplicative and additive model: Price = Base*(X + Y) * Z

32 Constraint GIA  In real world, for most of the pricing factors, the range of their values are capped due to market and regulatory constraints

33 Numerical Methodology for GIA  For all algorithms: Use the mean of the response variable as the base Starting points:1 for multiplicative factors; 0 for additive factors Use the latest relativities in the iterations All the reported GIAs converge within 8 steps for our test examples  For mixed models: In each step, adjust multiplicative factors from one rating variable proportionally so that its weighted average is one. For the last multiplicative variable, adjust its factors so that the weighted average of the product of all multiplicative variables is one.

34 Conclusions  2 and 3 Parameter GIA can completely represent GLM and minimum bias models  Can fit mixed models and models with constraints  Provide additional model options for data fitting  Easy to understand and does not require advanced statistical knowledge  Can program in many different tools  Calculation efficiency is not an issue because of modern computer power.

35 Demonstration of a GIA Tool  Written in VB.NET and runs on Windows PCs  Approximately 200 hours for tool development  Efficiency statistics: Efficiency for different test cases Excel DataCSV Data # of Records# of VariablesLoading TimeModel TimeLoading TimeModel Time sec0.5 sec0.1 sec0.5 sec 25, sec6 sec1 sec5 sec 48, sec50 sec1.5 sec50 sec 642,12849N/A 40 sec

36 Q & A