centralesup elec (paris, france) › docs › projects › rmt4graph_midproject... · centralesup...

62
Projet RMT4GRAPH Evaluation ` a mi-parcours Romain COUILLET (en collaboration avec Hafiz Tiomoko Ali) CentraleSup´ elec (Paris, France) 9 septembre 2016 1 / 27

Upload: others

Post on 28-Jun-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Projet RMT4GRAPHEvaluation a mi-parcours

Romain COUILLET(en collaboration avec Hafiz Tiomoko Ali)

CentraleSupelec (Paris, France)

9 septembre 2016

1 / 27

Page 2: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Outline

Project Status

Machine Learning: Community Detection on Graphs

Machine Learning: Kernel Spectral Clustering

Future Investigations

2 / 27

Page 3: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Outline

Project Status

Machine Learning: Community Detection on Graphs

Machine Learning: Kernel Spectral Clustering

Future Investigations

3 / 27

Page 4: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Project definition and objectives

“develop a framework of big data processing analysis (notably graph-basedmethods) relying on random matrix tools”

Applic

ati

ons

Theory

RMT4GRAPH

Applications toBig Data Processing

Random Matrix Modelsfor Random Graphs

Objects of interest Matrix models

Studied quantities SP&ML applications

Deterministic equivalentsRandom equivalentsLimiting spectrumIsolated eigenvaluesExtreme eigenvalues

Kernel random matricesSpiked random matricesRandom Laplacian matricesOther Hermitian modelsNon-Hermitian models

Limiting graph behaviorAsymptotic consistencePerformance analysisAlgorithm stabilityImprovement parameters

Random graph systemsClassification and clusteringSignal Processing on graphsScan statistics on graphsEcho-state neural networks

4 / 27

Page 5: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Work Packages and Timeline

I WP1. Random Matrix Models for Random Graphs.I Task 1.1. Kernel random matrix models.I Task 1.2. Hermitian models and spikes.I Task 1.3. Random matrices with non-linear or recursive entries

(formerly “Task 1.3. Non-hermitian random matrix models)

I WP2. Applications to Big Data Processing.I Task 2.1. Applications to machine learning.I Task 2.2. Signal processing on graphsI Task 2.3. Neural networks

(formerly restricted to ”Task 2.3. Echo-state neural networks)

WP1

WP2

Task 1.1

Task 1.2

Task 1.3

Task 2.1

Task 2.2

Task 2.3

Deliverables D1 D2 D3

5 / 27

Page 6: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Work Packages and Timeline

I WP1. Random Matrix Models for Random Graphs.I Task 1.1. Kernel random matrix models.I Task 1.2. Hermitian models and spikes.I Task 1.3. Random matrices with non-linear or recursive entries

(formerly “Task 1.3. Non-hermitian random matrix models)

I WP2. Applications to Big Data Processing.I Task 2.1. Applications to machine learning.I Task 2.2. Signal processing on graphsI Task 2.3. Neural networks

(formerly restricted to ”Task 2.3. Echo-state neural networks)

WP1

WP2

Task 1.1

Task 1.2

Task 1.3

Task 2.1

Task 2.2

Task 2.3

Deliverables D1 D2 D3

5 / 27

Page 7: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Work Packages and Timeline

I WP1. Random Matrix Models for Random Graphs.I Task 1.1. Kernel random matrix models.I Task 1.2. Hermitian models and spikes.I Task 1.3. Random matrices with non-linear or recursive entries

(formerly “Task 1.3. Non-hermitian random matrix models)

I WP2. Applications to Big Data Processing.I Task 2.1. Applications to machine learning.I Task 2.2. Signal processing on graphsI Task 2.3. Neural networks

(formerly restricted to ”Task 2.3. Echo-state neural networks)

WP1

WP2

Task 1.1

Task 1.2

Task 1.3

Task 2.1

Task 2.2

Task 2.3

Deliverables D1 D2 D3

5 / 27

Page 8: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Taskforce and Actions

TaskforceI Principal Investigator. Romain Couillet.

I Students.I Hafiz Tiomoko Ali (PhD student, RMT4GRAPH grant, 2015-2018): community

detection, neural networks.I Xiaoyi Mai (PhD student, DIGICOSME grant, 2016-2019): semi-supervised learning.I Zhenyu Liao (intern, ERC-MORE grant, 2016): support vector machines.I Cosme Louart (intern, ERC-MORE grant, 2016): neural networks (extreme learning

machines).I Evgeny Kusmenko (PhD student, ERC-MORE grant, jan. 2015-dec. 2015): spectral

clustering.I Harry Sevi (intern, ERC-MORE grant, 2015): echo-state neural networks.

I Collaborators.I Florent Benaych-Georges (professor at Universit Paris Descartes): kernel random

matrices.I Gilles Wainrib (assistant professor at ENS Paris): neural networks.I Abla Kammoun (research scientist at KAUST): subspace clustering.

Actions and Publications.I Publications: 3 journal articles, 7 conference articles

I Dissemination:I Organization of the summer school “Large Random Matrices and High Dimensional

Statistical Signal Processing”, Telecom ParisTech, June 7-8, 2016.I SSP’16 Special Session “Random matrices in signal processing and machine learning”I Distinguished keynote speaker at EUSIPCO 2016I Special Issue on Random Matrices in “Revue du Traitement du Signal”I Several invited talks and contributions to local events

6 / 27

Page 9: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Taskforce and Actions

TaskforceI Principal Investigator. Romain Couillet.

I Students.I Hafiz Tiomoko Ali (PhD student, RMT4GRAPH grant, 2015-2018): community

detection, neural networks.I Xiaoyi Mai (PhD student, DIGICOSME grant, 2016-2019): semi-supervised learning.I Zhenyu Liao (intern, ERC-MORE grant, 2016): support vector machines.I Cosme Louart (intern, ERC-MORE grant, 2016): neural networks (extreme learning

machines).I Evgeny Kusmenko (PhD student, ERC-MORE grant, jan. 2015-dec. 2015): spectral

clustering.I Harry Sevi (intern, ERC-MORE grant, 2015): echo-state neural networks.

I Collaborators.I Florent Benaych-Georges (professor at Universit Paris Descartes): kernel random

matrices.I Gilles Wainrib (assistant professor at ENS Paris): neural networks.I Abla Kammoun (research scientist at KAUST): subspace clustering.

Actions and Publications.I Publications: 3 journal articles, 7 conference articles

I Dissemination:I Organization of the summer school “Large Random Matrices and High Dimensional

Statistical Signal Processing”, Telecom ParisTech, June 7-8, 2016.I SSP’16 Special Session “Random matrices in signal processing and machine learning”I Distinguished keynote speaker at EUSIPCO 2016I Special Issue on Random Matrices in “Revue du Traitement du Signal”I Several invited talks and contributions to local events

6 / 27

Page 10: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Taskforce and Actions

TaskforceI Principal Investigator. Romain Couillet.

I Students.I Hafiz Tiomoko Ali (PhD student, RMT4GRAPH grant, 2015-2018): community

detection, neural networks.I Xiaoyi Mai (PhD student, DIGICOSME grant, 2016-2019): semi-supervised learning.I Zhenyu Liao (intern, ERC-MORE grant, 2016): support vector machines.I Cosme Louart (intern, ERC-MORE grant, 2016): neural networks (extreme learning

machines).I Evgeny Kusmenko (PhD student, ERC-MORE grant, jan. 2015-dec. 2015): spectral

clustering.I Harry Sevi (intern, ERC-MORE grant, 2015): echo-state neural networks.

I Collaborators.I Florent Benaych-Georges (professor at Universit Paris Descartes): kernel random

matrices.I Gilles Wainrib (assistant professor at ENS Paris): neural networks.I Abla Kammoun (research scientist at KAUST): subspace clustering.

Actions and Publications.I Publications: 3 journal articles, 7 conference articles

I Dissemination:I Organization of the summer school “Large Random Matrices and High Dimensional

Statistical Signal Processing”, Telecom ParisTech, June 7-8, 2016.I SSP’16 Special Session “Random matrices in signal processing and machine learning”I Distinguished keynote speaker at EUSIPCO 2016I Special Issue on Random Matrices in “Revue du Traitement du Signal”I Several invited talks and contributions to local events

6 / 27

Page 11: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Taskforce and Actions

TaskforceI Principal Investigator. Romain Couillet.

I Students.I Hafiz Tiomoko Ali (PhD student, RMT4GRAPH grant, 2015-2018): community

detection, neural networks.I Xiaoyi Mai (PhD student, DIGICOSME grant, 2016-2019): semi-supervised learning.I Zhenyu Liao (intern, ERC-MORE grant, 2016): support vector machines.I Cosme Louart (intern, ERC-MORE grant, 2016): neural networks (extreme learning

machines).I Evgeny Kusmenko (PhD student, ERC-MORE grant, jan. 2015-dec. 2015): spectral

clustering.I Harry Sevi (intern, ERC-MORE grant, 2015): echo-state neural networks.

I Collaborators.I Florent Benaych-Georges (professor at Universit Paris Descartes): kernel random

matrices.I Gilles Wainrib (assistant professor at ENS Paris): neural networks.I Abla Kammoun (research scientist at KAUST): subspace clustering.

Actions and Publications.I Publications: 3 journal articles, 7 conference articles

I Dissemination:I Organization of the summer school “Large Random Matrices and High Dimensional

Statistical Signal Processing”, Telecom ParisTech, June 7-8, 2016.I SSP’16 Special Session “Random matrices in signal processing and machine learning”I Distinguished keynote speaker at EUSIPCO 2016I Special Issue on Random Matrices in “Revue du Traitement du Signal”I Several invited talks and contributions to local events

6 / 27

Page 12: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Outline

Project Status

Machine Learning: Community Detection on Graphs

Machine Learning: Kernel Spectral Clustering

Future Investigations

7 / 27

Page 13: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Assume n-node, m-edge graph G, with

I “intrinsic” average connectivity q1, . . . , qn ∼ µ i.i.d.

I k classes C1, . . . , Ck independent of {qi} of (large) sizes n1, . . . , nk, withpreferential attachment Cab between Ca and Cb

I induces edge probability for node i ∈ Ca, j ∈ Cb,

P (i ∼ j) = qiqjCab.

8 / 27

Page 14: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Assume n-node, m-edge graph G, with

I “intrinsic” average connectivity q1, . . . , qn ∼ µ i.i.d.

I k classes C1, . . . , Ck independent of {qi} of (large) sizes n1, . . . , nk, withpreferential attachment Cab between Ca and Cb

I induces edge probability for node i ∈ Ca, j ∈ Cb,

P (i ∼ j) = qiqjCab.

8 / 27

Page 15: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Assume n-node, m-edge graph G, with

I “intrinsic” average connectivity q1, . . . , qn ∼ µ i.i.d.

I k classes C1, . . . , Ck independent of {qi} of (large) sizes n1, . . . , nk, withpreferential attachment Cab between Ca and Cb

I induces edge probability for node i ∈ Ca, j ∈ Cb,

P (i ∼ j) = qiqjCab.

8 / 27

Page 16: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Objective:Understand and improve performance of spectral community detection methods:

I based on adjacency A or modularity A− ddT

2mmatrices (adapted to dense nets)

I based on Bethe Hessian (r2 − 1)In − rA+D (adapted to sparse nets!).

0 spikes

⇓ Eigenvectors ⇓

9 / 27

Page 17: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Objective:Understand and improve performance of spectral community detection methods:

I based on adjacency A or modularity A− ddT

2mmatrices (adapted to dense nets)

I based on Bethe Hessian (r2 − 1)In − rA+D (adapted to sparse nets!).

0 spikes

⇓ Eigenvectors ⇓

9 / 27

Page 18: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Objective:Understand and improve performance of spectral community detection methods:

I based on adjacency A or modularity A− ddT

2mmatrices (adapted to dense nets)

I based on Bethe Hessian (r2 − 1)In − rA+D (adapted to sparse nets!).

0 spikes

⇓ Eigenvectors ⇓

9 / 27

Page 19: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Objective:Understand and improve performance of spectral community detection methods:

I based on adjacency A or modularity A− ddT

2mmatrices (adapted to dense nets)

I based on Bethe Hessian (r2 − 1)In − rA+D (adapted to sparse nets!).

0 spikes

⇓ Eigenvectors ⇓

9 / 27

Page 20: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Eig

env.

1E

igen

v.2

⇓ p-dimensional representation ⇓

Eigenvector 1

Eig

enve

ctor

2

⇓EM or k-means clustering.

10 / 27

Page 21: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Eig

env.

1E

igen

v.2

⇓ p-dimensional representation ⇓

Eigenvector 1

Eig

enve

ctor

2

⇓EM or k-means clustering.

10 / 27

Page 22: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

System Setting

Eig

env.

1E

igen

v.2

⇓ p-dimensional representation ⇓

Eigenvector 1

Eig

enve

ctor

2

⇓EM or k-means clustering.

10 / 27

Page 23: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Limitations of Adjacency/Modularity Approach

Scenario: 3 classes with µ bi-modal (e.g., µ = 34δ0.1 + 1

4δ0.5)

→ Leading eigenvectors of A (or modularity A− ddT

2m) biased by qi distribution.

→ Similar behavior for Bethe Hessian.

(Modularity) (Bethe Hessian)

11 / 27

Page 24: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Limitations of Adjacency/Modularity Approach

Scenario: 3 classes with µ bi-modal (e.g., µ = 34δ0.1 + 1

4δ0.5)

→ Leading eigenvectors of A (or modularity A− ddT

2m) biased by qi distribution.

→ Similar behavior for Bethe Hessian.

(Modularity) (Bethe Hessian)

11 / 27

Page 25: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Regularized Modularity Approach

Connectivity Model: P (i ∼ j) = qiqjCab for i ∈ Ca, j ∈ Cb.

Dense Regime Assumptions: Non trivial regime when, as n→∞,

Cab = 1 +Mab√n

with Mab = O(1) (fixed).

Considered Matrix:For α ∈ [0, 1], (and with D = diag(A1n) = diag(d) the degree matrix)

Lα = (2m)α1√nD−α

[A−

ddT

2m

]D−α.

12 / 27

Page 26: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Regularized Modularity Approach

Connectivity Model: P (i ∼ j) = qiqjCab for i ∈ Ca, j ∈ Cb.

Dense Regime Assumptions: Non trivial regime when, as n→∞,

Cab = 1 +Mab√n

with Mab = O(1) (fixed).

Considered Matrix:For α ∈ [0, 1], (and with D = diag(A1n) = diag(d) the degree matrix)

Lα = (2m)α1√nD−α

[A−

ddT

2m

]D−α.

12 / 27

Page 27: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Asymptotic Equivalence

Theorem (Limiting Random Matrix Equivalent)For each α ∈ [0, 1], as n→∞, ‖Lα − Lα‖ → 0 almost surely, where

Lα = (2m)α1√nD−α

[A−

ddT

2m

]D−α

Lα =1√nD−αq XD−αq + UΛUT

with Dq = diag({qi}), X zero-mean random matrix,

U =[D1−αq

J√n

11TnDq1n

D−αq X1n], rank k + 1

Λ =

[(Ik − 1kc

T)M(Ik − c1Tk) −1k

1Tk 0

]and J = [j1, . . . , jk], ja = [0, . . . , 0, 1T

na, 0, . . . , 0]T ∈ Rn canonical vector of class Ca.

Consequences:I Lα is a well-known spiked random matrixI it is “easy” to study and leads to a full analysis of the spectral clustering

performance!I it helps us correct and optimize classical spectral clustering into a powerful new

algorithm.

13 / 27

Page 28: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Asymptotic Equivalence

Theorem (Limiting Random Matrix Equivalent)For each α ∈ [0, 1], as n→∞, ‖Lα − Lα‖ → 0 almost surely, where

Lα = (2m)α1√nD−α

[A−

ddT

2m

]D−α

Lα =1√nD−αq XD−αq + UΛUT

with Dq = diag({qi}), X zero-mean random matrix,

U =[D1−αq

J√n

11TnDq1n

D−αq X1n], rank k + 1

Λ =

[(Ik − 1kc

T)M(Ik − c1Tk) −1k

1Tk 0

]and J = [j1, . . . , jk], ja = [0, . . . , 0, 1T

na, 0, . . . , 0]T ∈ Rn canonical vector of class Ca.

Consequences:I Lα is a well-known spiked random matrix

I it is “easy” to study and leads to a full analysis of the spectral clusteringperformance!

I it helps us correct and optimize classical spectral clustering into a powerful newalgorithm.

13 / 27

Page 29: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Asymptotic Equivalence

Theorem (Limiting Random Matrix Equivalent)For each α ∈ [0, 1], as n→∞, ‖Lα − Lα‖ → 0 almost surely, where

Lα = (2m)α1√nD−α

[A−

ddT

2m

]D−α

Lα =1√nD−αq XD−αq + UΛUT

with Dq = diag({qi}), X zero-mean random matrix,

U =[D1−αq

J√n

11TnDq1n

D−αq X1n], rank k + 1

Λ =

[(Ik − 1kc

T)M(Ik − c1Tk) −1k

1Tk 0

]and J = [j1, . . . , jk], ja = [0, . . . , 0, 1T

na, 0, . . . , 0]T ∈ Rn canonical vector of class Ca.

Consequences:I Lα is a well-known spiked random matrixI it is “easy” to study and leads to a full analysis of the spectral clustering

performance!

I it helps us correct and optimize classical spectral clustering into a powerful newalgorithm.

13 / 27

Page 30: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Asymptotic Equivalence

Theorem (Limiting Random Matrix Equivalent)For each α ∈ [0, 1], as n→∞, ‖Lα − Lα‖ → 0 almost surely, where

Lα = (2m)α1√nD−α

[A−

ddT

2m

]D−α

Lα =1√nD−αq XD−αq + UΛUT

with Dq = diag({qi}), X zero-mean random matrix,

U =[D1−αq

J√n

11TnDq1n

D−αq X1n], rank k + 1

Λ =

[(Ik − 1kc

T)M(Ik − c1Tk) −1k

1Tk 0

]and J = [j1, . . . , jk], ja = [0, . . . , 0, 1T

na, 0, . . . , 0]T ∈ Rn canonical vector of class Ca.

Consequences:I Lα is a well-known spiked random matrixI it is “easy” to study and leads to a full analysis of the spectral clustering

performance!I it helps us correct and optimize classical spectral clustering into a powerful new

algorithm.

13 / 27

Page 31: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Performance Results (2 masses of qi)

(Modularity) (Bethe Hessian)

(Algo with α = 1) (Algo with αopt)

Figure: Two dominant eigenvectors (x-y axes) for n = 2000, K = 3, µ = 34 δq1 + 1

4 δq2 ,

q1 = 0.1, q2 = 0.5, c1 = c2 = 14 , c3 = 1

2 , M = 100I3.

14 / 27

Page 32: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Performance Results (2 masses of qi)

(Modularity) (Bethe Hessian)

(Algo with α = 1)

(Algo with αopt)

Figure: Two dominant eigenvectors (x-y axes) for n = 2000, K = 3, µ = 34 δq1 + 1

4 δq2 ,

q1 = 0.1, q2 = 0.5, c1 = c2 = 14 , c3 = 1

2 , M = 100I3.14 / 27

Page 33: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Performance Results (2 masses of qi)

(Modularity) (Bethe Hessian)

(Algo with α = 1) (Algo with αopt)

Figure: Two dominant eigenvectors (x-y axes) for n = 2000, K = 3, µ = 34 δq1 + 1

4 δq2 ,

q1 = 0.1, q2 = 0.5, c1 = c2 = 14 , c3 = 1

2 , M = 100I3.14 / 27

Page 34: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Performance Results (2 masses for qi)

0.2 0.4 0.6 0.80

0.2

0.4

0.6

0.8

1

q2 (q1 = 0.1)

Ove

rla

p

α = 0

α = 0.5

α = 1

α = αopt

Bethe Hessian

Figure: Overlap performance for n = 3000, K = 3, µ = 34 δq1 + 1

4 δq2 with q1 = 0.1 and

q2 ∈ [0.1, 0.9], M = 10(2I3 − 131T3), ci =

13 .

15 / 27

Page 35: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Outline

Project Status

Machine Learning: Community Detection on Graphs

Machine Learning: Kernel Spectral Clustering

Future Investigations

16 / 27

Page 36: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Kernel Spectral Clustering

Problem Statement

I Dataset x1, . . . , xn ∈ RpI Objective: “cluster” data in k similarity classes S1, . . . ,Sk.

I Typical metric to optimize:

(RatioCut) argminS1∪...∪Sk={1,...,n}

k∑i=1

∑j∈Sij /∈Si

κ(xj , xj)

|Si|

for some similarity kernel κ(x, y) ≥ 0 (large if x similar to y).

I Can be shown equivalent to

(RatioCut) argminM∈M trMT(D −K)M

where M⊂ Rn×k ∩{M ; Mij ∈ {0, |Sj |−

12 }}

(in particular, MTM = Ik) and

K = {κ(xi, xj)}ni,j=1, Dii =n∑j=1

Kij .

I But integer problem! Usually NP-complete.

17 / 27

Page 37: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Kernel Spectral Clustering

Problem Statement

I Dataset x1, . . . , xn ∈ RpI Objective: “cluster” data in k similarity classes S1, . . . ,Sk.

I Typical metric to optimize:

(RatioCut) argminS1∪...∪Sk={1,...,n}

k∑i=1

∑j∈Sij /∈Si

κ(xj , xj)

|Si|

for some similarity kernel κ(x, y) ≥ 0 (large if x similar to y).

I Can be shown equivalent to

(RatioCut) argminM∈M trMT(D −K)M

where M⊂ Rn×k ∩{M ; Mij ∈ {0, |Sj |−

12 }}

(in particular, MTM = Ik) and

K = {κ(xi, xj)}ni,j=1, Dii =n∑j=1

Kij .

I But integer problem! Usually NP-complete.

17 / 27

Page 38: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Kernel Spectral Clustering

Problem Statement

I Dataset x1, . . . , xn ∈ RpI Objective: “cluster” data in k similarity classes S1, . . . ,Sk.

I Typical metric to optimize:

(RatioCut) argminS1∪...∪Sk={1,...,n}

k∑i=1

∑j∈Sij /∈Si

κ(xj , xj)

|Si|

for some similarity kernel κ(x, y) ≥ 0 (large if x similar to y).

I Can be shown equivalent to

(RatioCut) argminM∈M trMT(D −K)M

where M⊂ Rn×k ∩{M ; Mij ∈ {0, |Sj |−

12 }}

(in particular, MTM = Ik) and

K = {κ(xi, xj)}ni,j=1, Dii =n∑j=1

Kij .

I But integer problem! Usually NP-complete.

17 / 27

Page 39: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Kernel Spectral Clustering

Problem Statement

I Dataset x1, . . . , xn ∈ RpI Objective: “cluster” data in k similarity classes S1, . . . ,Sk.

I Typical metric to optimize:

(RatioCut) argminS1∪...∪Sk={1,...,n}

k∑i=1

∑j∈Sij /∈Si

κ(xj , xj)

|Si|

for some similarity kernel κ(x, y) ≥ 0 (large if x similar to y).

I Can be shown equivalent to

(RatioCut) argminM∈M trMT(D −K)M

where M⊂ Rn×k ∩{M ; Mij ∈ {0, |Sj |−

12 }}

(in particular, MTM = Ik) and

K = {κ(xi, xj)}ni,j=1, Dii =n∑j=1

Kij .

I But integer problem! Usually NP-complete.

17 / 27

Page 40: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Kernel Spectral Clustering

Towards kernel spectral clustering

I Kernel spectral clustering: discrete-to-continuous relaxations of such metrics

(RatioCut) argminM, MTM=IKtrMT(D −K)M

i.e., eigenvector problem:1. find eigenvectors of smallest eigenvalues2. retrieve classes from eigenvector components

I Refinements:I working on K, D −K, In −D−1K, In −D−

12KD−

12 , etc.

I several steps algorithms: Ng–Jordan–Weiss, Shi–Malik, etc.

18 / 27

Page 41: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Kernel Spectral Clustering

Towards kernel spectral clustering

I Kernel spectral clustering: discrete-to-continuous relaxations of such metrics

(RatioCut) argminM, MTM=IKtrMT(D −K)M

i.e., eigenvector problem:1. find eigenvectors of smallest eigenvalues2. retrieve classes from eigenvector components

I Refinements:I working on K, D −K, In −D−1K, In −D−

12KD−

12 , etc.

I several steps algorithms: Ng–Jordan–Weiss, Shi–Malik, etc.

18 / 27

Page 42: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Kernel Spectral Clustering

Figure: Leading four eigenvectors of D−12KD−

12 for MNIST data.

19 / 27

Page 43: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Methodology and objectives

Objectives and Roadmap:

I Develop mathematical analysis framework for BigData kernel spectral clustering(p, n→∞)

I Understand:1. Phase transition effects (i.e., when is clustering possible?)2. Content of each eigenvector3. Influence of kernel function4. Performance comparison of clustering algorithms

Methodology:

I Use statistical assumptions (Gaussian mixture)

I Benefit from doubly-infinite independence and random matrix tools

20 / 27

Page 44: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Methodology and objectives

Objectives and Roadmap:

I Develop mathematical analysis framework for BigData kernel spectral clustering(p, n→∞)

I Understand:1. Phase transition effects (i.e., when is clustering possible?)2. Content of each eigenvector3. Influence of kernel function4. Performance comparison of clustering algorithms

Methodology:

I Use statistical assumptions (Gaussian mixture)

I Benefit from doubly-infinite independence and random matrix tools

20 / 27

Page 45: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Methodology and objectives

Objectives and Roadmap:

I Develop mathematical analysis framework for BigData kernel spectral clustering(p, n→∞)

I Understand:1. Phase transition effects (i.e., when is clustering possible?)2. Content of each eigenvector3. Influence of kernel function4. Performance comparison of clustering algorithms

Methodology:

I Use statistical assumptions (Gaussian mixture)

I Benefit from doubly-infinite independence and random matrix tools

20 / 27

Page 46: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Model and Assumptions

Gaussian mixture model:

I x1, . . . , xn ∈ Rp,

I k classes C1, . . . , Ck,

I x ∈ Ca ⇔ x ∼ N (µa, Ca).

Kernel Matrix:

I Kernel matrix of interest:

K =

{f

(1

p‖xi − xj‖2

)}ni,j=1

for some sufficiently smooth nonnegative f .

I We study the normalized Laplacian:

L = nD−12KD−

12

with D = diag(K1n).

21 / 27

Page 47: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Model and Assumptions

Gaussian mixture model:

I x1, . . . , xn ∈ Rp,

I k classes C1, . . . , Ck,

I x ∈ Ca ⇔ x ∼ N (µa, Ca).

Kernel Matrix:

I Kernel matrix of interest:

K =

{f

(1

p‖xi − xj‖2

)}ni,j=1

for some sufficiently smooth nonnegative f .

I We study the normalized Laplacian:

L = nD−12KD−

12

with D = diag(K1n).

21 / 27

Page 48: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Model and Assumptions

Gaussian mixture model:

I x1, . . . , xn ∈ Rp,

I k classes C1, . . . , Ck,

I x ∈ Ca ⇔ x ∼ N (µa, Ca).

Kernel Matrix:

I Kernel matrix of interest:

K =

{f

(1

p‖xi − xj‖2

)}ni,j=1

for some sufficiently smooth nonnegative f .

I We study the normalized Laplacian:

L = nD−12KD−

12

with D = diag(K1n).

21 / 27

Page 49: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Model and Assumptions

Difficulty: L is a very intractable random matrix

I non-linear f

I non-trivial dependence between entries of L

Strategy:

1. Find random equivalent L (i.e., ‖L− L‖ a.s.−→ 0 as n, p→∞) based on:I concentration: Kij → τ , constant, as n, p→∞ (for all i 6= j)I Taylor expansion around limit point

2. Apply spiked random matrix approach to study:I existence of isolated eigenvalues in L: phase transitionI eigenvector projections on canonical class-basis

22 / 27

Page 50: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Model and Assumptions

Difficulty: L is a very intractable random matrix

I non-linear f

I non-trivial dependence between entries of L

Strategy:

1. Find random equivalent L (i.e., ‖L− L‖ a.s.−→ 0 as n, p→∞) based on:I concentration: Kij → τ , constant, as n, p→∞ (for all i 6= j)I Taylor expansion around limit point

2. Apply spiked random matrix approach to study:I existence of isolated eigenvalues in L: phase transitionI eigenvector projections on canonical class-basis

22 / 27

Page 51: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Model and Assumptions

Difficulty: L is a very intractable random matrix

I non-linear f

I non-trivial dependence between entries of L

Strategy:

1. Find random equivalent L (i.e., ‖L− L‖ a.s.−→ 0 as n, p→∞) based on:I concentration: Kij → τ , constant, as n, p→∞ (for all i 6= j)I Taylor expansion around limit point

2. Apply spiked random matrix approach to study:I existence of isolated eigenvalues in L: phase transition

I eigenvector projections on canonical class-basis

22 / 27

Page 52: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Model and Assumptions

Difficulty: L is a very intractable random matrix

I non-linear f

I non-trivial dependence between entries of L

Strategy:

1. Find random equivalent L (i.e., ‖L− L‖ a.s.−→ 0 as n, p→∞) based on:I concentration: Kij → τ , constant, as n, p→∞ (for all i 6= j)I Taylor expansion around limit point

2. Apply spiked random matrix approach to study:I existence of isolated eigenvalues in L: phase transitionI eigenvector projections on canonical class-basis

22 / 27

Page 53: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Theoretical Findings versus MNIST

Figure: Leading four eigenvectors of D−12KD−

12 for MNIST data (red), versus Gaussian

equivalent model (black), and theoretical findings (blue).

23 / 27

Page 54: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Theoretical Findings versus MNIST

Figure: Leading four eigenvectors of D−12KD−

12 for MNIST data (red), versus Gaussian

equivalent model (black), and theoretical findings (blue).

23 / 27

Page 55: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Theoretical Findings versus MNIST

Figure: Leading four eigenvectors of D−12KD−

12 for MNIST data (red), versus Gaussian

equivalent model (black), and theoretical findings (blue).

23 / 27

Page 56: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Theoretical Findings versus MNIST

−.09 −.08 −.07 −.06

−0.1

0

0.1

Eigenvector 2/Eigenvector 1

−0.1 0 0.1

−0.1

0

0.1

0.2

Eigenvector 3/Eigenvector 2

Figure: 2D representation of eigenvectors of L, for the MNIST dataset. Theoretical means and 1-and 2-standard deviations in blue. Class 1 in red, Class 2 in black, Class 3 in green.

24 / 27

Page 57: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Outline

Project Status

Machine Learning: Community Detection on Graphs

Machine Learning: Kernel Spectral Clustering

Future Investigations

25 / 27

Page 58: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Other Results and Perspectives

Objectives:

I Kernel methods.4 Subspace spectral clustering (dramatically different case of f ′(τ) = 0)

. Spectral clustering with outer product kernel f(xTy)

. Semi-supervised learning, kernel approaches.

. Support vector machines (SVM).

I Community detection.4 Complete study of eigenvector contents in adjacency/modularity methods.� Study of Bethe Hessian approach.� Analysis of non-necessarily spectral approaches (wavelet approaches).

I Neural Networks.. Analysis of non-linear extreme learning machines� non-linear echo-state

I Signal processing on graphs, further graph inference, etc.� Making graph methods random.

26 / 27

Page 59: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Other Results and Perspectives

Objectives:

I Kernel methods.4 Subspace spectral clustering (dramatically different case of f ′(τ) = 0)

. Spectral clustering with outer product kernel f(xTy)

. Semi-supervised learning, kernel approaches.

. Support vector machines (SVM).

I Community detection.4 Complete study of eigenvector contents in adjacency/modularity methods.� Study of Bethe Hessian approach.� Analysis of non-necessarily spectral approaches (wavelet approaches).

I Neural Networks.. Analysis of non-linear extreme learning machines� non-linear echo-state

I Signal processing on graphs, further graph inference, etc.� Making graph methods random.

26 / 27

Page 60: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Other Results and Perspectives

Objectives:

I Kernel methods.4 Subspace spectral clustering (dramatically different case of f ′(τ) = 0)

. Spectral clustering with outer product kernel f(xTy)

. Semi-supervised learning, kernel approaches.

. Support vector machines (SVM).

I Community detection.4 Complete study of eigenvector contents in adjacency/modularity methods.� Study of Bethe Hessian approach.� Analysis of non-necessarily spectral approaches (wavelet approaches).

I Neural Networks.. Analysis of non-linear extreme learning machines� non-linear echo-state

I Signal processing on graphs, further graph inference, etc.� Making graph methods random.

26 / 27

Page 61: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

Other Results and Perspectives

Objectives:

I Kernel methods.4 Subspace spectral clustering (dramatically different case of f ′(τ) = 0)

. Spectral clustering with outer product kernel f(xTy)

. Semi-supervised learning, kernel approaches.

. Support vector machines (SVM).

I Community detection.4 Complete study of eigenvector contents in adjacency/modularity methods.� Study of Bethe Hessian approach.� Analysis of non-necessarily spectral approaches (wavelet approaches).

I Neural Networks.. Analysis of non-linear extreme learning machines� non-linear echo-state

I Signal processing on graphs, further graph inference, etc.� Making graph methods random.

26 / 27

Page 62: CentraleSup elec (Paris, France) › docs › projects › RMT4GRAPH_midproject... · CentraleSup elec (Paris, France) 9 septembre 2016 1/27 Outline Project Status Machine Learning:

The End

Thank you.

27 / 27