playing with features for learning and prediction jongmin kim seoul national university

Post on 26-Dec-2015

215 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Playing with features forlearning and prediction

Jongmin KimSeoul National University

Problem statement

• Predicting outcome of surgery

Predicting outcome of surgery

• Ideal approach

. . . .

?

Training Data

Predicting out-come

surgery

Predicting outcome of surgery

• Initial approach– Predicting partial features

• Predict witch features?

Predicting outcome of surgery

• 4 Surgery– DHL+RFT+TAL+FDO

flexion of the knee( min / max )

dorsiflexion of the ankle( min )

rotation of the foot( min / max )

Predicting outcome of surgery

• Is it good features?

• Number of Training data– DHL+RFT+TAL : 35 data– FDO+DHL+TAL+RFT : 33 data

Machine learning and feature

DataFeature

representationLearningalgorithm

Featurerepresentation

Learningalgorithm

• Joint position / angle• Velocity / acceleration• Distance between body parts• Contact status• …

Features in motion

Features in computer vision

SIFT Spin image

HoG RIFT

Textons GLOH

Machine learning and feature

Outline

• Feature selection• - Feature ranking• - Subset selection: wrapper, filter, embedded• - Recursive Feature Elimination• - Combination of weak prior (Boosting)• - ADAboosting(clsf) / joint boosting (clsf)/ Gradi-

entboost (regression)

• Prediction result with feature selection

• Feature learning?

Feature selection

• Alleviating the effect of the curse of dimensionality

• Improve the prediction performance• Faster and more cost-effective• Providing a better understanding of

the data

Subset selection

• Wrapper

• Filter

• Embedded

Feature learning?

• Can we automatically learn a good feature represen-tation?

• Known as: unsupervised feature learning, feature learning, deep learning, representation learning, etc.

• Hand-designed features (by human):• 1. need expert knowledge• 2. requires time-consuming hand-tuning.

• When it’s unclear how to hand design features: au-tomatically learned features (by machine)

Learning Feature Representations

• Key idea: • –Learn statistical structure or correlation of the

data from unlabeled data • –The learned representations can be used as fea-

tures in supervised and semi-supervised settings

Learning Feature Representations

EncoderDecoder

Input (Image/ Features)

Output Features

e.g.Feed-back /generative /top-downpath

Feed-forward /bottom-up path

Learning Feature Representations

σ(Wx)Dz

Input Patch x

Sparse Features z

e.g.

• Predictive Sparse Decomposition [Kavukcuoglu et al., ‘09]

Encoder filters W

Sigmoid function σ(.)

Decoder filters D

L1 Spar-sity

Stacked Auto-Encoders

En-coder

De-coder

Input Image

Class label

Features

En-coder

De-coder

Features

En-coder

De-coder

[Hinton & Salakhutdinov Science ‘06]

At Test Time

En-coder

Input Image

Class label

Features

En-coder

Features

En-coder

[Hinton & Salakhutdinov Science ‘06]

• Remove decoders• Use feed-forward

path

• Gives standard(Convolutional)Neural Network

• Can fine-tune with backprop

Status & plan

• Data 파악 / learning technique survey…

• Plan : 11 월 실험 끝• 12 월 논문 writing• 1 월 시그랩 submit• 8 월에 미국에서 발표

• But before all of that….

Deep neural net vs. boost-ing

• Deep Nets:• - single highly non-linear system• - “deep” stack of simpler modules• - all parameters are subject to learning

• Boosting & Forests:• - sequence of “weak” (simple) classifiers that are lin-

early combined to produce a powerful classifier• - subsequent classifiers do not exploit representa-

tions of earlier classifiers, it's a “shallow” linear mix-ture

• - typically features are not learned

Deep neural net vs. boost-ing

top related