Daphne’s Approximate Group of Students. Outline Linear Regression Unregularized L2 Regularized What is a GP? Prediction with a GP Relationship to SVM.

Slides:



Advertisements
Similar presentations
Ordinary Least-Squares
Advertisements

Copula Regression By Rahul A. Parsa Drake University &
4.1: Linearizing Data.
1 Non-Linear and Smooth Regression Non-linear parametric models: There is a known functional form y=  x,  derived from known theory or from previous.
Computer vision: models, learning and inference Chapter 8 Regression.
CS Statistical Machine learning Lecture 13 Yuan (Alan) Qi Purdue CS Oct
1.4 Equations of Lines and Linear Models
Second order cone programming approaches for handing missing and uncertain data P. K. Shivaswamy, C. Bhattacharyya and A. J. Smola Discussion led by Qi.
Pattern Recognition and Machine Learning
Rutgers CS440, Fall 2003 Support vector machines Reading: Ch. 20, Sec. 6, AIMA 2 nd Ed.
Lesson #32 Simple Linear Regression. Regression is used to model and/or predict a variable; called the dependent variable, Y; based on one or more independent.
Prepared By: Kevin Meier Alok Desai
SVM for Regression DMML Lab 04/20/07. SVM Recall Two-class classification problem using linear model:
Support Vector Regression (Linear Case:)  Given the training set:  Find a linear function, where is determined by solving a minimization problem that.
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
CHAPTER 4: Parametric Methods. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Parametric Estimation X = {
Analysis of Covariance Goals: 1)Reduce error variance. 2)Remove sources of bias from experiment. 3)Obtain adjusted estimates of population means.
Study of Sparse Online Gaussian Process for Regression EE645 Final Project May 2005 Eric Saint Georges.
1 Linear Classification Problem Two approaches: -Fisher’s Linear Discriminant Analysis -Logistic regression model.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
Chapter 2 – Simple Linear Regression - How. Here is a perfect scenario of what we want reality to look like for simple linear regression. Our two variables.
1 1 Slide Simple Linear Regression Chapter 14 BA 303 – Spring 2011.
Today Wrap up of probability Vectors, Matrices. Calculus
Section 6.1: Euler’s Method. Local Linearity and Differential Equations Slope at (2,0): Tangent line at (2,0): Not a good approximation. Consider smaller.
Review of Lecture Two Linear Regression Normal Equation
Overview of Kernel Methods Prof. Bennett Math Model of Learning and Discovery 2/27/05 Based on Chapter 2 of Shawe-Taylor and Cristianini.
Gaussian process regression Bernád Emőke Gaussian processes Definition A Gaussian Process is a collection of random variables, any finite number.
Psyc 235: Introduction to Statistics DON’T FORGET TO SIGN IN FOR CREDIT!
University of Southern California Department Computer Science Bayesian Logistic Regression Model (Final Report) Graduate Student Teawon Han Professor Schweighofer,
Statistical Methods Statistical Methods Descriptive Inferential
Physics 270.  y ( x )Data Set #1 t (s) y (cm)
Simple Linear Regression. The term linear regression implies that  Y|x is linearly related to x by the population regression equation  Y|x =  +  x.
Chapter 6 Simple Regression Introduction Fundamental questions – Is there a relationship between two random variables and how strong is it? – Can.
Bayesian Generalized Kernel Mixed Models Zhihua Zhang, Guang Dai and Michael I. Jordan JMLR 2011.
Gaussian Processes Li An Li An
Sparse Kernel Methods 1 Sparse Kernel Methods for Classification and Regression October 17, 2007 Kyungchul Park SKKU.
Review #2.
1 Beginning & Intermediate Algebra – Math 103 Math, Statistics & Physics.
1 Honors Physics 1 Class 11 Fall 2013 Propagation of Uncertainty Rotational quantities as vectors.
Linear Prediction Correlation can be used to make predictions – Values on X can be used to predict values on Y – Stronger relationships between X and Y.
Gaussian Processes For Regression, Classification, and Prediction.
Short Introduction to Particle Filtering by Arthur Pece [ follows my Introduction to Kalman filtering ]
Logistic Regression William Cohen.
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Gaussian Process and Prediction. (C) 2001 SNU CSE Artificial Intelligence Lab (SCAI)2 Outline Gaussian Process and Bayesian Regression  Bayesian regression.
Sparse Approximate Gaussian Processes. Outline Introduction to GPs Subset of Data Bayesian Committee Machine Subset of Regressors Sparse Pseudo GPs /
Gaussian Process Networks Nir Friedman and Iftach Nachman UAI-2K.
Week of March 23 Partial correlations Semipartial correlations
Kernelized Value Function Approximation for Reinforcement Learning Gavin Taylor and Ronald Parr Duke University.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Efficient Logistic Regression with Stochastic Gradient Descent William Cohen 1.
Hypothesis Testing Example 3: Test the hypothesis that the average content of containers of a particular lubricant is 10 litters if the contents of random.
Probability Theory and Parameter Estimation I
Ch3: Model Building through Regression
CSE 4705 Artificial Intelligence
Non-Parametric Models
Lecture 09: Gaussian Processes
Propagating Uncertainty In POMDP Value Iteration with Gaussian Process
Lesson 5.7 Predict with Linear Models The Zeros of a Function
The Multiple Regression Model
FORECASTING 16-Jan-19 Dr.B.Sasidhar.
Ch11 Curve Fitting II.
More Parameter Learning, Multinomial and Continuous Variables
Lecture 10: Gaussian Processes
Test123 blah. hjghgghhg.
Introduction to Radial Basis Function Networks
Linear regression.
I can determine the different sampling techniques used in real life.
X ⦁ X = 64 ±8 ±14 X ⦁ X ⦁ X =
FORECASTING 11-Dec-19 Dr.B.Sasidhar.
Presentation transcript:

Daphne’s Approximate Group of Students

Outline Linear Regression Unregularized L2 Regularized What is a GP? Prediction with a GP Relationship to SVM Implications What does this mean?

Linear Regression Predicting Y given X Y = wtx + n w_ml = argmax y[m+1] = w_mltx[m+1]

L2 Regularized Lin Reg L2 Regularized (Gaussian Prior on w) Y = wtx + n w ~ N(0,S) w_map = argmax blah + ||w||^2

What is a random process? It’s a prior over functions

What is a Gaussian Process? It’s a prior over functions that generalized a Gaussian Random Vector Prior over Y(x) ~ N(0,I)

Alternate Definition The thing with Euler’s equation

This is weird Not used to thinking of prior over Ys Or are we? We ARE used to thining about prior over w What prior over y does this induce

Math P(w) -> P(Y) Wow! This became a Gaussian Process!

Prediction with a GP Predict y*[m+1] given y[1]…y[m] We get a covariance = error bars Wow! This prediction is the same as w_map but we get error bars!

Generalize that shit - Covariance Functions Note that we have a thing here that is defined by C(x1,x2) which can be kernelized Has to be pos semidefinite Is a kernel function

Relationship to SVM

Example

How do we reconcile these views? Does this change anything?