foundations of machine learning - nyu computer sciencemohri/mls/ml_introduction.pdf · foundations...

47
Foundations of Machine Learning page Foundations of Machine Learning Introduction to ML 1 Mehryar Mohri Courant Institute and Google Research [email protected]

Upload: ngonhan

Post on 17-Feb-2019

221 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

Foundations of Machine Learning page

Foundations of Machine Learning Introduction to ML

1

Mehryar MohriCourant Institute and Google Research

[email protected]

Page 2: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

LogisticsPrerequisites: basics in linear algebra, probability, and analysis of algorithms.

Workload: about 3-4 homework assignments + project (topic of your choice).

Mailing list: join as soon as possible.

2

Page 3: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Course MaterialTextbook

3

Slides: course web page.

http://www.cs.nyu.edu/~mohri/ml17

Page 4: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

This LectureBasic definitions and concepts.

Introduction to the problem of learning.

Probability tools.

4

Page 5: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Machine LearningDefinition: computational methods using experience to improve performance.

Experience: data-driven task, thus statistics, probability, and optimization.

Computer science: learning algorithms, analysis of complexity, theoretical guarantees.

Example: use document word counts to predict its topic.

5

Page 6: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Examples of Learning TasksText: document classification, spam detection.

Language: NLP tasks (e.g., morphological analysis, POS tagging, context-free parsing, dependency parsing).

Speech: recognition, synthesis, verification.

Image: annotation, face recognition, OCR, handwriting recognition.

Games (e.g., chess, backgammon).

Unassisted control of vehicles (robots, car).

Medical diagnosis, fraud detection, network intrusion.

6

Page 7: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Some Broad ML TasksClassification: assign a category to each item (e.g., document classification).

Regression: predict a real value for each item (prediction of stock values, economic variables).

Ranking: order items according to some criterion (relevant web pages returned by a search engine).

Clustering: partition data into ‘homogenous’ regions (analysis of very large data sets).

Dimensionality reduction: find lower-dimensional manifold preserving some properties of the data.

7

Page 8: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

General Objectives of MLTheoretical questions:

• what can be learned, under what conditions?

• are there learning guarantees?

• analysis of learning algorithms.

Algorithms:

• more efficient and more accurate algorithms.

• deal with large-scale problems.

• handle a variety of different learning problems.

8

Page 9: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

This CourseTheoretical foundations:

• learning guarantees.

• analysis of algorithms.

Algorithms:

• main mathematically well-studied algorithms.

• discussion of their extensions.

Applications:

• illustration of their use.

9

Page 10: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

TopicsProbability tools, concentration inequalities.

PAC learning model, Rademacher complexity, VC-dimension, generalization bounds.

Support vector machines (SVMs), margin bounds, kernel methods.

Ensemble methods, boosting.

Logistic regression and conditional maximum entropy models.

On-line learning, weighted majority algorithm, Perceptron algorithm, mistake bounds.

Regression, generalization, algorithms.

Ranking, generalization, algorithms.

Reinforcement learning, MDPs, bandit problems and algorithm.10

Page 11: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Definitions and TerminologyExample: item, instance of the data used.

Features: attributes associated to an item, often represented as a vector (e.g., word counts).

Labels: category (classification) or real value (regression) associated to an item.

Data:

• training data (typically labeled).

• test data (labeled but labels not seen).

• validation data (labeled, for tuning parameters).

11

Page 12: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

General Learning ScenariosSettings:

• batch: learner receives full (training) sample, which he uses to make predictions for unseen points.

• on-line: learner receives one sample at a time and makes a prediction for that sample.

Queries:

• active: the learner can request the label of a point.

• passive: the learner receives labeled points.

12

Page 13: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Standard Batch ScenariosUnsupervised learning: no labeled data.

Supervised learning: uses labeled data for prediction on unseen points.

Semi-supervised learning: uses labeled and unlabeled data for prediction on unseen points.

Transduction: uses labeled and unlabeled data for prediction on seen points.

13

Page 14: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Example - SPAM DetectionProblem: classify each e-mail message as SPAM or non-SPAM (binary classification problem).

Potential data: large collection of SPAM and non-SPAM messages (labeled examples).

14

Page 15: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Learning Stages

15

labeled data

training sample

validation data

test sample

featuresA(⇥)

algorithm

A(⇥0)parameterselection

evaluation

prior knowledge

Page 16: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

This LectureBasic definitions and concepts.

Introduction to the problem of learning.

Probability tools.

16

Page 17: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

DefinitionsSpaces: input space , output space .

Loss function: .

• : cost of predicting instead of .

• binary classification: 0-1 loss, .

• regression: , .

Hypothesis set: , subset of functions out of which the learner selects his hypothesis.

• depends on features.

• represents prior knowledge about task.

17

X Y

L : Y ⇥Y !R

L(by, y) by y

L(y, y0)=1y 6=y0

l(y, y0)=(y0 � y)2Y ✓R

H✓Y X

Page 18: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Supervised Learning Set-UpTraining data: sample of size drawn i.i.d. from according to distribution :

Problem: find hypothesis with small generalization error.

• deterministic case: output label deterministic function of input, .

• stochastic case: output probabilistic function of input.

18

S mD

X⇥Y

S = ((x1, y1), . . . , (xm, ym)).

h2H

y=f(x)

Page 19: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

ErrorsGeneralization error: for , it is defined by

Empirical error: for and sample , it is

Bayes error:

• in deterministic case,

19

h2H

R(h) = E(x,y)⇠D

[L(h(x), y)].

bR(h) =

1

m

mX

i=1

L(h(xi), yi).

R? = infh

h measurable

R(h).

h2H S

R?=0.

Page 20: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

NoiseNoise:

• in binary classification, for any ,

• observe that

20

x 2 X

noise(x) = min{Pr[1|x],Pr[0|x]}.

E[noise(x)] = R

⇤.

Page 21: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Learning ≠ Fitting

21

Notion of simplicity/complexity. How do we define complexity?

Page 22: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

GeneralizationObservations:

• the best hypothesis on the sample may not be the best overall.

• generalization is not memorization.

• complex rules (very complex separation surfaces) can be poor predictors.

• trade-off: complexity of hypothesis set vs sample size (underfitting/overfitting).

22

Page 23: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Model SelectionGeneral equality: for any ,

Approximation: not a random variable, only depends on .

Estimation: only term we can hope to bound.

23

best in classh2H

H

R(h)�R⇤ = [R(h)�R(h⇤)]| {z }estimation

+ [R(h⇤)�R⇤]| {z }approximation

.

Page 24: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Empirical Risk MinimizationSelect hypothesis set .

Find hypothesis minimizing empirical error:

• but may be too complex.

• the sample size may not be large enough.

24

h = argminh2H

bR(h).

h2H

H

H

Page 25: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Generalization BoundsDefinition: upper bound on

Bound on estimation error for hypothesis given by ERM:

25

h0

Pr

suph2H

|R(h)� bR(h)| > ✏

�.

R(h0)�R(h⇤) = R(h0)� bR(h0) + bR(h0)�R(h⇤)

R(h0)� bR(h0) + bR(h⇤)�R(h⇤)

2 suph2H

|R(h)� bR(h)|.

How should we choose ? (model selection problem)H

Page 26: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Model Selection

26

erro

r

�⇤

estimationapproximationupper bound

H =[

�2�

H� .

Page 27: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Structural Risk MinimizationPrinciple: consider an infinite sequence of hypothesis sets ordered for inclusion,

• strong theoretical guarantees.

• typically computationally hard.

27

(Vapnik, 1995)

H1 ⇢ H2 ⇢ · · · ⇢ Hn ⇢ · · ·

h = argminh2Hn,n2N

bR(h) + penalty(Hn,m).

Page 28: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

General Algorithm FamiliesEmpirical risk minimization (ERM):

Structural risk minimization (SRM): ,

Regularization-based algorithms: ,

28

h = argminh2H

bR(h).

Hn✓Hn+1

��0

h = argminh2H

bR(h) + �khk2.

h = argminh2Hn,n2N

bR(h) + penalty(Hn,m).

Page 29: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

This LectureBasic definitions and concepts.

Introduction to the problem of learning.

Probability tools.

29

Page 30: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Basic PropertiesUnion bound:

Inversion: if , then, for any , with probability at least , .

Jensen’s inequality: if is convex, .

Expectation: if , .

30

f

Pr[A _B] Pr[A] + Pr[B].

Pr[X � ✏]f(✏) �>0

1��

f(E[X])E[f(X)]

E[X]=

Z +1

0Pr[X > t] dtX�0

Xf�1(�)

Page 31: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Basic InequalitiesMarkov’s inequality: if and , then

Chebyshev’s inequality: for any ,

31

X�0 ✏>0

Pr[X�✏] E[X]✏ .

Pr[|X � E[X]| � ✏] �2X✏2 .

✏>0

Page 32: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Hoeffding’s InequalityTheorem: Let be indep. rand. variables with the same expectation and , ( ). Then, for any , the following inequalities hold:

32

X1, . . . , Xm

µ Xi2 [a, b] a<b ✏>0

Pr

µ� 1

m

mX

i=1

Xi > ✏

� exp

✓� 2m✏2

(b� a)2

Pr

1

m

mX

i=1

Xi � µ > ✏

� exp

✓� 2m✏2

(b� a)2

◆.

Page 33: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

McDiarmid’s InequalityTheorem: let be independent random variables taking values in and a function verifying for all ,

33

(McDiarmid, 1989)

Then, for all ,

X1, . . . , Xm

✏>0

U f :Um!Ri2 [1,m]

supx1,...,xm,x

0i

|f(x1, . . . , xi

, . . . , x

m

)�f(x1, . . . , x0i

, . . . , x

m

)|c

i

.

Pr

h��f(X1, . . . , Xm)�E[f(X1, . . . , Xm)]

��>✏i2 exp

✓� 2✏2Pm

i=1 c2i

◆.

Page 34: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

Foundations of Machine Learning page

Appendix

34

Page 35: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Markov’s InequalityTheorem: let be a non-negative random variable with , then, for all ,

Proof:

35

X

E[X]<1 t>0

Pr[X � tE[X]] 1

t.

Pr[X � tE[X]] =X

x�tE[X]

Pr[X = x]

X

x�tE[X]

Pr[X = x]x

tE[X]

X

x

Pr[X = x]x

tE[X]

= E

X

tE[X]

�=

1

t

.

Page 36: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Chebyshev’s InequalityTheorem: let be a random variable with , then, for all ,

Proof: Observe that

The result follows Markov’s inequality.

36

X Var[X]<1t>0

Pr[|X � E[X]| � t�X ] 1

t2.

Pr[|X � E[X]| � t�X ] = Pr[(X � E[X])2 � t2�2X ].

Page 37: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Weak Law of Large NumbersTheorem: let be a sequence of independent random variables with the same mean and variance and let , then, for any ,

Proof: Since the variables are independent,

Thus, by Chebyshev’s inequality,

37

(Xn)n2Nµ �2<1

✏>0Xn=1n

Pni=1 Xi

limn!1

Pr[|Xn � µ| � ✏] = 0.

Var[Xn] =nX

i=1

Var

Xi

n

�=

n�2

n2=

�2

n.

Pr[|Xn � µ| � ✏] �2

n✏2.

Page 38: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Concentration InequalitiesSome general tools for error analysis and bounds:

• Hoeffding’s inequality (additive).

• Chernoff bounds (multiplicative).

• McDiarmid’s inequality (more general).

38

Page 39: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Hoeffding’s LemmaLemma: Let be a random variable with and . Then for any ,

Proof: by convexity of , for all ,

39

Thus,

with,

E[etX ] � E[ b�Xb�a eta + X�a

b�a etb] = bb�aeta + �a

b�aetb = e�(t),

E[X]=0X 2 [a, b]

b 6=a t>0

E[etX ] et2(b�a)2

8 .

x 7!e

tx

axb

e

tx b� x

b� a

e

ta +x� a

b� a

e

tb

.

�(t) = log(

bb�ae

ta+

�ab�ae

tb) = ta+ log(

bb�a +

�ab�ae

t(b�a)).

Page 40: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Taking the derivative gives:

Note that: Furthermore,

with There exists such that:

40

�0(t) = a� aet(b�a)

bb�a� a

b�a et(b�a) = a� ab

b�a e�t(b�a)� ab�a

.

�(0) = 0 and �0(0) = 0.

�00(t) =�abe�t(b�a)

[ bb�ae

�t(b�a) � ab�a ]

2

=↵(1� ↵)e�t(b�a)(b� a)2

[(1� ↵)e�t(b�a) + ↵]2

=↵

[(1� ↵)e�t(b�a) + ↵]

(1� ↵)e�t(b�a)

[(1� ↵)e�t(b�a) + ↵](b� a)2

= u(1� u)(b� a)2 (b� a)2

4,

↵ =�a

b� a. 0✓ t

�(t) = �(0) + t�0(0) +t2

2�00(✓) t2

(b� a)2

8.

Page 41: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Hoeffding’s TheoremTheorem: Let be independent random variables. Then for , the following inequalities hold for , for any ,

Proof: The proof is based on Chernoff ’s bounding technique: for any random variable and , apply Markov’s inequality and select to minimize

41

X1, . . . , Xm

Xi2 [ai, bi]

✏>0Sm =Pm

i=1 Xi

Pr[Sm � E[Sm] � ✏] e�2✏2/Pm

i=1(bi�ai)2

Pr[Sm � E[Sm] �✏] e�2✏2/Pm

i=1(bi�ai)2

.

X t>0

t

Pr[X � ✏] = Pr[etX � et✏] E[etX ]

et✏.

Page 42: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Using this scheme and the independence of the random variables gives

The second inequality is proved in a similar way.

42

choosing

Pr[Sm � E[Sm] � ✏]

e�t✏ E[et(Sm�E[Sm])]

= e�t✏⇧mi=1 E[e

t(Xi�E[Xi])]

(lemma applied to Xi�E[Xi]) e�t✏⇧mi=1e

t2(bi�ai)2/8

= e�t✏et2 Pm

i=1(bi�ai)2/8

e�2✏2/Pm

i=1(bi�ai)2

,

t = 4✏/Pm

i=1(bi � ai)2.

Page 43: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Hoeffding’s InequalityCorollary: for any , any distribution and any hypothesis , the following inequalities hold:

Proof: follows directly Hoeffding’s theorem.

Combining these one-sided inequalities yields

43

✏>0 D

h :X!{0, 1}

Pr[ bR(h)�R(h) � ✏] e�2m✏2

Pr[ bR(h)�R(h) �✏] e�2m✏2 .

Prh�� bR(h)�R(h)

�� � ✏i 2e�2m✏2 .

Page 44: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Chernoff ’s InequalityTheorem: for any , any distribution and any hypothesis , the following inequalities hold:

Proof: proof based on Chernoff ’s bounding technique.

44

✏>0 D

h :X!{0, 1}

Pr[ bR(h) � (1 + ✏)R(h)] e�mR(h) ✏2/3

Pr[ bR(h) (1� ✏)R(h)] e�mR(h) ✏2/2.

Page 45: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

McDiarmid’s InequalityTheorem: let be independent random variables taking values in and a function verifying for all ,

45

(McDiarmid, 1989)

Then, for all ,

X1, . . . , Xm

✏>0

U f :Um!Ri2 [1,m]

supx1,...,xm,x

0i

|f(x1, . . . , xi

, . . . , x

m

)�f(x1, . . . , x0i

, . . . , x

m

)|c

i

.

Pr

h��f(X1, . . . , Xm)�E[f(X1, . . . , Xm)]

��>✏i2 exp

✓� 2✏2Pm

i=1 c2i

◆.

Page 46: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Comments:

• Proof: uses Hoeffding’s lemma.

• Hoeffding’s inequality is a special case of McDiarmid’s with

46

f(x1, . . . , xm) =1

m

mX

i=1

xi and ci =|bi � ai|

m

.

Page 47: Foundations of Machine Learning - NYU Computer Sciencemohri/mls/ml_introduction.pdf · Foundations of Machine Learning page Logistics Prerequisites: basics in linear algebra, probability,

pageFoundations of Machine Learning

Jensen’s InequalityTheorem: let be a random variable and a measurable convex function. Then,

Proof: definition of convexity, continuity of convex functions, and density of finite distributions.

47

X

f(E[X]) E[f(X)].

f