Using Neural Networks to Predict Claim Duration in the Presence of Right Censoring and Covariates David Speights Senior Research Statistician HNC Insurance.

Slides:



Advertisements
Similar presentations
Artificial Neural Networks
Advertisements

Slides from: Doug Gray, David Poole
Neural networks Introduction Fitting neural networks
Linear Regression.
Intelligent Environments1 Computer Science and Engineering University of Texas at Arlington.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Supervised Learning Recap
Chapter 4: Linear Models for Classification
Financial Informatics –XVI: Supervised Backpropagation Learning
Visual Recognition Tutorial
Pattern Recognition and Machine Learning
I welcome you all to this presentation On: Neural Network Applications Systems Engineering Dept. KFUPM Imran Nadeem & Naveed R. Butt &
Data Mining with Neural Networks (HK: Chapter 7.5)
October 28, 2010Neural Networks Lecture 13: Adaptive Networks 1 Adaptive Networks As you know, there is no equation that would tell you the ideal number.
Data mining and statistical learning - lecture 11 Neural networks - a model class providing a joint framework for prediction and classification  Relationship.
Dan Simon Cleveland State University
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Introduction to Directed Data Mining: Neural Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Microsoft Enterprise Consortium Data Mining Concepts Introduction to Directed Data Mining: Neural Networks Prepared by David Douglas, University of ArkansasHosted.
Radial Basis Function Networks
Classification Part 3: Artificial Neural Networks
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
Chapter 11 – Neural Networks COMP 540 4/17/2007 Derek Singer.
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
Introduction to Artificial Neural Network Models Angshuman Saha Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg.
NEURAL NETWORKS FOR DATA MINING
Chapter 7 Neural Networks in Data Mining Automatic Model Building (Machine Learning) Artificial Intelligence.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Using Resampling Techniques to Measure the Effectiveness of Providers in Workers’ Compensation Insurance David Speights Senior Research Statistician HNC.
Artificial Intelligence Techniques Multilayer Perceptrons.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Neural Networks Automatic Model Building (Machine Learning) Artificial Intelligence.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Neural and Evolutionary Computing - Lecture 9 1 Evolutionary Neural Networks Design  Motivation  Evolutionary training  Evolutionary design of the architecture.
Neural Nets: Something you can use and something to think about Cris Koutsougeras What are Neural Nets What are they good for Pointers to some models and.
Predictive Modeling Spring 2005 CAMAR meeting Louise Francis, FCAS, MAAA Francis Analytics and Actuarial Data Mining, Inc
Neural Networks Demystified by Louise Francis Francis Analytics and Actuarial Data Mining, Inc.
Neural Networks Presented by M. Abbasi Course lecturer: Dr.Tohidkhah.
Neural Networks Teacher: Elena Marchiori R4.47 Assistant: Kees Jong S2.22
Image Source: ww.physiol.ucl.ac.uk/fedwards/ ca1%20neuron.jpg
1 Chapter 4: Introduction to Predictive Modeling: Regressions 4.1 Introduction 4.2 Selecting Regression Inputs 4.3 Optimizing Regression Complexity 4.4.
Chapter 8: Adaptive Networks
Chapter 2-OPTIMIZATION G.Anuradha. Contents Derivative-based Optimization –Descent Methods –The Method of Steepest Descent –Classical Newton’s Method.
Lecture 8, CS5671 Neural Network Concepts Weight Matrix vs. NN MLP Network Architectures Overfitting Parameter Reduction Measures of Performance Sequence.
NEURAL NETWORKS LECTURE 1 dr Zoran Ševarac FON, 2015.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Neural Networks The Elements of Statistical Learning, Chapter 12 Presented by Nick Rizzolo.
Computacion Inteligente Least-Square Methods for System Identification.
Chapter 11 – Neural Nets © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
A Presentation on Adaptive Neuro-Fuzzy Inference System using Particle Swarm Optimization and it’s Application By Sumanta Kundu (En.R.No.
Chapter 7. Classification and Prediction
Deep Feedforward Networks
Advanced information retreival
One-layer neural networks Approximation problems
第 3 章 神经网络.
Multimodal Learning with Deep Boltzmann Machines
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
Overview of Supervised Learning
Machine Learning Today: Reading: Maria Florina Balcan
Regression Models - Introduction
Artificial Intelligence Chapter 3 Neural Networks
Neural Networks Geoff Hulten.
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Neural networks (1) Traditional multi-layer perceptrons
Artificial Intelligence Chapter 3 Neural Networks
Artificial Intelligence Chapter 3 Neural Networks
Presentation transcript:

Using Neural Networks to Predict Claim Duration in the Presence of Right Censoring and Covariates David Speights Senior Research Statistician HNC Insurance Solutions Irvine, California Session CPP-53

Presentation Outline Introduction to Neural Networks Introduction to Survival Analysis Neural Networks with Right Censoring Simulated Example Predicting Claim Duration

Introduction to Neural Networks Motivation Complex Classification –Character Recognition –Voice Recognition Humans have no trouble with these concepts –We can read even distorted documents –We can recognize voices over poor telephone lines. Attempt to model human brain

Introduction to Neural Networks Connection to Brain Functionality Brain –made up of millions of neurons sending signals to the body and each other Neural Networks –collection of “neurons” which send “signals” to produce an output

Introduction to Neural Networks Common Representation... X1X1 X2X2 XPXP Y P predictors (inputs) 1 Hidden Layer with M Neurons 1 output 12 M

Introduction to Neural Networks Architecture of the i th Neuron Represents a neuron in the brain X1X1 X2X2 XPXP... O=b i0 + b i1 X 1 + … + b ip X p s(O) S is a function on the interval (0,1) representing the strength of the output 0 1 s O Activation Function

Introduction to Neural Networks Connection to Multiple Regressions Similarities –Both describe relationships between variables –Both can create predictions Differences –Function describing the relationships is more complex –Response variables are typically called outputs –Predictor variables are typically called inputs –Estimating the parameters is usually called training

Introduction to Neural Networks Functional Representation Y = f(X 1, …, X p ) + error Multiple Linear Regression –f() = linear combination of regressors –Forced to model only specified relationships Neural Network –f() = nonlinear combination of regressors –Can deal with nonlinearities and interactions without special designation

Introduction to Neural Networks Functional Specification For a neural network f() is written Here g and s are transformation functions specified in advance

Introduction to Survival Analysis What is Survival Analysis Used to model time to event data (example: time until a claim ends) Usually represented by (1) right skewed data (2) multiplicative error structure (3) right censoring Common in cancer clinical trials, component failure analysis, and AIDS data analysis among other examples

Introduction to Survival Analysis Notation T 1,..., T n - independent failure times with distribution F and density function f C 1,..., C n - independent censoring times with distribution G and density function g Y i = min(T i,C i ) - observed time  i = I(Y i = T i ) - Censoring indicator X i = (X i1,..., X ip ) - vector of known covariates associated with the i th individual

Introduction to Survival Analysis Likelihood Analysis (Parametric Models) (Y i,  i, X i ) i=1, …, n, independent observations Likelihood written f  (Y,  |X)=[f  (Y|X)(1-G(Y|X))]  [g(Y|X)(1-F  (Y|X))]  Here L 2 does not depend on 

Neural Networks with Right Censoring Model Specification Neural Network Model Here  has distribution function F  and density f   = {  0, …,  p,  1, …,  p } The likelihood is

Neural Networks with Right Censoring Fitting Neural Networks without Censoring  estimated by minimizing squared error If  is normal minimizing squared error same as maximizing the likelihood.

Neural Networks with Right Censoring Fitting Neural Networks without Censoring Gradient decent algorithm for estimating  Algorithm updated at each observation is known as the learning rate  j:0 =  j-1:n Known as back-propagation algorithm To generalize to right censored data, replace C(  ) with the likelihood for censored neural networks.

Neural Networks with Right Censoring Fitting Neural Networks with Censoring Step 1 - Estimating  –Fix  and pass through data once using Step 2 - Estimating  –fix  at end of pass through data –iterate until |  j -  j-1 |<  using Newton-Raphson algorithm

Neural Networks with Right Censoring Fitting Neural Networks with Censoring With highly parameterized neural networks we risk over fitting

We need to design the fitting procedure to find a good fit to the data Neural Networks with Right Censoring Fitting Neural Networks with Censoring

The negative of the likelihood is calculated on both sets of data at the same time. Negative Likelihood 75% Training Data25% Testing Data Parameter Estimates Training Cycles Neural Networks with Right Censoring Fitting Neural Networks with Censoring

Potential drawbacks to neural networks –Hard to tell the individual effects of each predictor variable on the response. –Can have poor extrapolation properties Potential Gains from neural networks –Can reduce preliminary analysis in modeling discovery of interactions and nonlinear relationships becomes automatic –Increases predictive power of models Neural Networks with Right Censoring Fitting Neural Networks with Censoring

True Time Model : log(t) = x   Censoring Model: log(c) = x   x ~ U(-3,3)      ~ N(0,1) Censored if c < t ~ 35% censoring 3 node neural network fit Simulated Example

Scatter are true times versus x Solid line represents NN fit to data Simulated Example

Predicting Claim Duration Predictor Variables –NCCI Codes Body Part Code Injury Type Nature of Injury Industry Class Code –Demographic Information Age Gender Weekly Wage Zip Code Response Variable –Time from report until the claim is closed

Predicting Claim Duration Ratio of prediction to actual duration on log 10 scale Difficult to represent open claim results

Conclusions Provides an intuitive method to address right censored data with a neural network Allows for more flexible mean function Can be used with many time to event data situations