lectures 15,16 – additive models, trees, and related methods rice ece697 farinaz koushanfar fall...

32
Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Upload: stewart-mcdaniel

Post on 17-Jan-2016

216 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Lectures 15,16 – Additive Models, Trees, and Related Methods

Rice ECE697

Farinaz Koushanfar

Fall 2006

Page 2: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Summary

• Generalized Additive Models

• Tree-Based Methods

• PRIM – Bump Hunting

• Mutlivariate Adaptive Regression Splines (MARS)

• Missing Data

Page 3: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Additive Models

• In real life, effects are nonlinear

Note: Some slides are borrowed from Tibshirani

Page 4: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Examples

Page 5: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

The Price for Additivity

Data from a study of Diabetic children, Predicting log C-peptide(a blood measurement)

Page 6: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Generalized Additive Models (GAM)Two-class Logistic Regression

Page 7: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Other Examples

Page 8: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Fitting Additive Models

• Given observations xi,yi, a criterion like the penalized sum of squares can be specified for this problem, where ’s are tuning parameters

p

1jjj)X(fY The mean of error term is zero!

N

1i

p

1jj

2

jjj

2p

1jijji

p1

dt)t("f})x(fy{

)f,...,f,(PRSS

Page 9: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Fitting Additive Models

Page 10: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

The Backfitting Algorithm for Additive Models

• Initialize:

• Cycle: j=1,2,…,p,1,2,…,p,1,…

• Until the functions fj change less than a prespecified threshold

j,i,0f̂;yN

1j

N

1ii

]})x(f̂y[{Sf̂ N

1jk

ikkijj

N

1iijjjj)x(f̂

N

1f̂f̂

Page 11: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Fitting Additive Models (Cont’d)

Page 12: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Example: Penalized Least square

Page 13: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Example: Fitting GAM for Logistic Regression (Newton-Raphson Algorithm)

Page 14: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Example: Predicting Email Spam

• Data from 4601 mail messages, spam=1, email=0, filter trained for each user separately

• Goal: predict whether an email is spam (junk mail) or good

• Input features: relative frequencies in a message of 57 of the commonly occurring words and punctuation marks in all training set

• Not all errors are equal; we want to avoid filtering out good email, while letting spam get through is not desirable but less serious in its consequences

Page 15: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Predictors

Page 16: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Details

Page 17: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Some Important Features

Page 18: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Results

• Test data confusion matrix for the additive logistic regression model fit to the spam training data

• The overall test error rate is 5.3%

Page 19: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Summary of Additive Logistic Fit• Significant predictors from the additive model fit to the spam

training data. The coefficients represent the linear part of f^j,

along with their standard errors and Z-score. • The nonlinear p-value represents a test of nonlinearity of f^

j

Page 20: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Example: Plots for Spam Analysis

Figure 9.1. Spam analysis: estimated functions for significant predictors. The rug plot along the bottom of each frame indicates the observed values of the corresponding predictor. For many predictors, the nonlinearity picks up the discontinuity at zero.

Page 21: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

In Summary

• Additive models are a useful extension to linear models, making them more flexible

• The backfitting procedure is simple and modular

• Limitations for large data mining applications

• Backfitting fits all predictors, which is not desirable when a large number are available

Page 22: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Tree-Based Methods

Page 23: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Node Impurity Measures

Page 24: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Results for Spam Example

Page 25: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Pruned tree for the Spam Example

Page 26: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Classification Rules Fit to the Spam Data

Page 27: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

PRIM-Bump Hunting

Page 28: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Number of Observations in a Box

Page 29: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Basis Functions

Page 30: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

MARS Forward Modeling Procedure

Page 31: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

Multiplication of Basis Functions

Page 32: Lectures 15,16 – Additive Models, Trees, and Related Methods Rice ECE697 Farinaz Koushanfar Fall 2006

MARS on Spam Example