most tensor problems are np-hard

64
Most Tensor Problems are NP-Hard C. J. Hillar and L.-H. Lim Betty Shea UBC MLRG 2020 Winter Term 1 14-Oct-2020 UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 1 / 28

Upload: others

Post on 09-Apr-2022

6 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Most Tensor Problems are NP-Hard

Most Tensor Problems are NP-HardC. J. Hillar and L.-H. Lim

Betty Shea

UBC MLRG 2020 Winter Term 1

14-Oct-2020

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 1 / 28

Page 2: Most Tensor Problems are NP-Hard

Today

1 Introduction

2 Tensor Approximation

3 Tensor Eigenvalues

4 Conclusion

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 2 / 28

Page 3: Most Tensor Problems are NP-Hard

Overview and Definitions

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 3 / 28

Page 4: Most Tensor Problems are NP-Hard

Where we are

Second of two papers on the ‘Tensor Basics’ subsection

Mark’s talk: tensor definitions, operations, concepts (e.g. rank, decomposition, etc.)

Today: multilinear algebra complexity

Bahare’s talk: Tensor factorization in graph representational learning

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 4 / 28

Page 5: Most Tensor Problems are NP-Hard

Where we are

Second of two papers on the ‘Tensor Basics’ subsection

Mark’s talk: tensor definitions, operations, concepts (e.g. rank, decomposition, etc.)

Today: multilinear algebra complexity

Bahare’s talk: Tensor factorization in graph representational learning

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 4 / 28

Page 6: Most Tensor Problems are NP-Hard

Where we are

Second of two papers on the ‘Tensor Basics’ subsection

Mark’s talk: tensor definitions, operations, concepts (e.g. rank, decomposition, etc.)

Today: multilinear algebra complexity

Bahare’s talk: Tensor factorization in graph representational learning

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 4 / 28

Page 7: Most Tensor Problems are NP-Hard

Where we are

Second of two papers on the ‘Tensor Basics’ subsection

Mark’s talk: tensor definitions, operations, concepts (e.g. rank, decomposition, etc.)

Today: multilinear algebra complexity

Bahare’s talk: Tensor factorization in graph representational learning

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 4 / 28

Page 8: Most Tensor Problems are NP-Hard

Theme of paper

1 ’The central message of our paper is that many problems in linear algebra that areefficiently solvable on a Turing machine become NP-hard in multilinear algebra.’

2 As much about the ‘tractability of a numerical computing problem using the richcollection of NP-complete combinatorial problems....’

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 5 / 28

Page 9: Most Tensor Problems are NP-Hard

Theme of paper

1 ’The central message of our paper is that many problems in linear algebra that areefficiently solvable on a Turing machine become NP-hard in multilinear algebra.’

2 As much about the ‘tractability of a numerical computing problem using the richcollection of NP-complete combinatorial problems....’

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 5 / 28

Page 10: Most Tensor Problems are NP-Hard

Everything is hard

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 6 / 28

Page 11: Most Tensor Problems are NP-Hard

Matrices

Matrix A over field F is a m × n array of elements of F

A = [aij ]m,ni,j=1 ∈ Fm×n

Given standard basis e1, . . . , ed in Fd , matrices are also bilinear maps.

f : Fm × Fn → F where aij = f (ei , ej ) ∈ F

By linearity, f (u, v) = uᵀAv .

If m = n, A is symmetric means that f is invariant under coordinate exchange:

f (u, v) = uᵀAv = (uᵀAv)ᵀ = vᵀAᵀu = vᵀAu = f (v , u)

where second to last equality made use of A = Aᵀ.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 7 / 28

Page 12: Most Tensor Problems are NP-Hard

Matrices

Matrix A over field F is a m × n array of elements of F

A = [aij ]m,ni,j=1 ∈ Fm×n

Given standard basis e1, . . . , ed in Fd , matrices are also bilinear maps.

f : Fm × Fn → F where aij = f (ei , ej ) ∈ F

By linearity, f (u, v) = uᵀAv .

If m = n, A is symmetric means that f is invariant under coordinate exchange:

f (u, v) = uᵀAv = (uᵀAv)ᵀ = vᵀAᵀu = vᵀAu = f (v , u)

where second to last equality made use of A = Aᵀ.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 7 / 28

Page 13: Most Tensor Problems are NP-Hard

Matrices

Matrix A over field F is a m × n array of elements of F

A = [aij ]m,ni,j=1 ∈ Fm×n

Given standard basis e1, . . . , ed in Fd , matrices are also bilinear maps.

f : Fm × Fn → F where aij = f (ei , ej ) ∈ F

By linearity, f (u, v) = uᵀAv .

If m = n, A is symmetric means that f is invariant under coordinate exchange:

f (u, v) = uᵀAv = (uᵀAv)ᵀ = vᵀAᵀu = vᵀAu = f (v , u)

where second to last equality made use of A = Aᵀ.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 7 / 28

Page 14: Most Tensor Problems are NP-Hard

3-Tensor

3-tensor A over field F is an l ×m × n array of elements of F

A = [aijk ]l,m,ni,j,k=1 ∈ Fl×m×n (1)

Also a trilinear map

f : Fl × Fm × Fn → F where aijk = f (ei , ej , ek ) ∈ F

If l = m = n, A is (super-)symmetric means

aijk = ajik = · · · = akji

ORf (u, v ,w) = f (u,w , v) = · · · = f (w , v , u)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 8 / 28

Page 15: Most Tensor Problems are NP-Hard

3-Tensor

3-tensor A over field F is an l ×m × n array of elements of F

A = [aijk ]l,m,ni,j,k=1 ∈ Fl×m×n (1)

Also a trilinear map

f : Fl × Fm × Fn → F where aijk = f (ei , ej , ek ) ∈ F

If l = m = n, A is (super-)symmetric means

aijk = ajik = · · · = akji

ORf (u, v ,w) = f (u,w , v) = · · · = f (w , v , u)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 8 / 28

Page 16: Most Tensor Problems are NP-Hard

3-Tensor

3-tensor A over field F is an l ×m × n array of elements of F

A = [aijk ]l,m,ni,j,k=1 ∈ Fl×m×n (1)

Also a trilinear map

f : Fl × Fm × Fn → F where aijk = f (ei , ej , ek ) ∈ F

If l = m = n, A is (super-)symmetric means

aijk = ajik = · · · = akji

ORf (u, v ,w) = f (u,w , v) = · · · = f (w , v , u)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 8 / 28

Page 17: Most Tensor Problems are NP-Hard

Cubic form

Given x ∈ Rn, A ∈ Rn×n, define

A(x , x) = xᵀAx =n∑

i,j=1

aijxixj

Extended to 3-tensor A ∈ Rl×m×n when l = m = n, we get cubic form

A(x , x , x) :=n∑

i,j,k=1

aijkxixjxk (2)

In general, the trilinear form is

A(x , y , z) :=

l,m,n∑i,j,k=1

aijkxiyjzk (3)

where x ∈ Rl ,y ∈ Rm,z ∈ Rk

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 9 / 28

Page 18: Most Tensor Problems are NP-Hard

Cubic form

Given x ∈ Rn, A ∈ Rn×n, define

A(x , x) = xᵀAx =n∑

i,j=1

aijxixj

Extended to 3-tensor A ∈ Rl×m×n when l = m = n, we get cubic form

A(x , x , x) :=n∑

i,j,k=1

aijkxixjxk (2)

In general, the trilinear form is

A(x , y , z) :=

l,m,n∑i,j,k=1

aijkxiyjzk (3)

where x ∈ Rl ,y ∈ Rm,z ∈ Rk

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 9 / 28

Page 19: Most Tensor Problems are NP-Hard

Cubic form

Given x ∈ Rn, A ∈ Rn×n, define

A(x , x) = xᵀAx =n∑

i,j=1

aijxixj

Extended to 3-tensor A ∈ Rl×m×n when l = m = n, we get cubic form

A(x , x , x) :=n∑

i,j,k=1

aijkxixjxk (2)

In general, the trilinear form is

A(x , y , z) :=

l,m,n∑i,j,k=1

aijkxiyjzk (3)

where x ∈ Rl ,y ∈ Rm,z ∈ Rk

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 9 / 28

Page 20: Most Tensor Problems are NP-Hard

Inner product, outer product

Let A = [aijk ]l,m,ni,j,k=1, B = [bijk ]l,m,n

i,j,k=1 ∈ Rl×m×n

Inner product is defined as

〈A,B〉 :=

l,m,n∑i,j,k=1

aijkbijk

Outer product of vectors x ∈ Fl , y ∈ Fm, z ∈ Fn, denoted x ⊗ y ⊗ z, gives 3-tensor A where

A = [aijk ]l,m,ni,j,k=1 where aijk = xiyjzk

Note: 〈A, x ⊗ y ⊗ z〉 = A(x , y , z) which is in the trilinear form (3)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 10 / 28

Page 21: Most Tensor Problems are NP-Hard

Inner product, outer product

Let A = [aijk ]l,m,ni,j,k=1, B = [bijk ]l,m,n

i,j,k=1 ∈ Rl×m×n

Inner product is defined as

〈A,B〉 :=

l,m,n∑i,j,k=1

aijkbijk

Outer product of vectors x ∈ Fl , y ∈ Fm, z ∈ Fn, denoted x ⊗ y ⊗ z, gives 3-tensor A where

A = [aijk ]l,m,ni,j,k=1 where aijk = xiyjzk

Note: 〈A, x ⊗ y ⊗ z〉 = A(x , y , z) which is in the trilinear form (3)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 10 / 28

Page 22: Most Tensor Problems are NP-Hard

Norms

Frobenius norm squared of 3-tensor A is defined as

‖A‖2F :=

l,m,n∑i,j,k=1

|aijk |2

Note: ‖A‖2F = 〈A,A〉

Spectral norm of 3-tensor A is defined as

‖A‖2,2,2 := supx,y,z 6=0

A(x , y , z)

‖x‖2‖y‖2‖z‖2(4)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 11 / 28

Page 23: Most Tensor Problems are NP-Hard

Norms

Frobenius norm squared of 3-tensor A is defined as

‖A‖2F :=

l,m,n∑i,j,k=1

|aijk |2

Note: ‖A‖2F = 〈A,A〉

Spectral norm of 3-tensor A is defined as

‖A‖2,2,2 := supx,y,z 6=0

A(x , y , z)

‖x‖2‖y‖2‖z‖2(4)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 11 / 28

Page 24: Most Tensor Problems are NP-Hard

Computation model

Mostly follows Sipser 2012.

Computations on a Turing Machine. Inputs are rational numbers. Outputs are rationalvectors or Yes/No.

Decision problem: the solution is in the form of Yes or No.

A decision problem is decidable if there is a Turing machine that will output a Yes/No forall allowable inputs in finitely many steps. Undecidable otherwise.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 12 / 28

Page 25: Most Tensor Problems are NP-Hard

Computation model

Mostly follows Sipser 2012.

Computations on a Turing Machine. Inputs are rational numbers. Outputs are rationalvectors or Yes/No.

Decision problem: the solution is in the form of Yes or No.

A decision problem is decidable if there is a Turing machine that will output a Yes/No forall allowable inputs in finitely many steps. Undecidable otherwise.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 12 / 28

Page 26: Most Tensor Problems are NP-Hard

Computation model

Mostly follows Sipser 2012.

Computations on a Turing Machine. Inputs are rational numbers. Outputs are rationalvectors or Yes/No.

Decision problem: the solution is in the form of Yes or No.

A decision problem is decidable if there is a Turing machine that will output a Yes/No forall allowable inputs in finitely many steps. Undecidable otherwise.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 12 / 28

Page 27: Most Tensor Problems are NP-Hard

Computation model

Mostly follows Sipser 2012.

Computations on a Turing Machine. Inputs are rational numbers. Outputs are rationalvectors or Yes/No.

Decision problem: the solution is in the form of Yes or No.

A decision problem is decidable if there is a Turing machine that will output a Yes/No forall allowable inputs in finitely many steps. Undecidable otherwise.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 12 / 28

Page 28: Most Tensor Problems are NP-Hard

Complexity

Time complexity measured in units of bit operations, i.e. the number of tape-level instructionson bits. (Input size is also specified in terms of number of bits.)

Measuring whether problems are equivalently difficult.

Reducibility in the Cook-Karp-Levin sense.

Very informally, problem P1 polynomially reduces to P2 if there is a way to solve P1 byfirst solving P2 and then translating the P2-solution into a P1-solution deterministicallyand in polynomial-time. P2 is at least as hard as P1.

NP: Problems where solutions could be certified in polynomial time.

NP-complete: If one can polynomially reduce any particular NP-complete problem P1 to aproblem P2, then all NP-complete problems are so reducible to P2. (Cook–Levin Theorem)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 13 / 28

Page 29: Most Tensor Problems are NP-Hard

Complexity

Time complexity measured in units of bit operations, i.e. the number of tape-level instructionson bits. (Input size is also specified in terms of number of bits.)

Measuring whether problems are equivalently difficult.

Reducibility in the Cook-Karp-Levin sense.

Very informally, problem P1 polynomially reduces to P2 if there is a way to solve P1 byfirst solving P2 and then translating the P2-solution into a P1-solution deterministicallyand in polynomial-time. P2 is at least as hard as P1.

NP: Problems where solutions could be certified in polynomial time.

NP-complete: If one can polynomially reduce any particular NP-complete problem P1 to aproblem P2, then all NP-complete problems are so reducible to P2. (Cook–Levin Theorem)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 13 / 28

Page 30: Most Tensor Problems are NP-Hard

Complexity

Time complexity measured in units of bit operations, i.e. the number of tape-level instructionson bits. (Input size is also specified in terms of number of bits.)

Measuring whether problems are equivalently difficult.

Reducibility in the Cook-Karp-Levin sense.

Very informally, problem P1 polynomially reduces to P2 if there is a way to solve P1 byfirst solving P2 and then translating the P2-solution into a P1-solution deterministicallyand in polynomial-time. P2 is at least as hard as P1.

NP: Problems where solutions could be certified in polynomial time.

NP-complete: If one can polynomially reduce any particular NP-complete problem P1 to aproblem P2, then all NP-complete problems are so reducible to P2. (Cook–Levin Theorem)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 13 / 28

Page 31: Most Tensor Problems are NP-Hard

Complexity

Time complexity measured in units of bit operations, i.e. the number of tape-level instructionson bits. (Input size is also specified in terms of number of bits.)

Measuring whether problems are equivalently difficult.

Reducibility in the Cook-Karp-Levin sense.

Very informally, problem P1 polynomially reduces to P2 if there is a way to solve P1 byfirst solving P2 and then translating the P2-solution into a P1-solution deterministicallyand in polynomial-time. P2 is at least as hard as P1.

NP: Problems where solutions could be certified in polynomial time.

NP-complete: If one can polynomially reduce any particular NP-complete problem P1 to aproblem P2, then all NP-complete problems are so reducible to P2. (Cook–Levin Theorem)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 13 / 28

Page 32: Most Tensor Problems are NP-Hard

Tensor Approximation

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 14 / 28

Page 33: Most Tensor Problems are NP-Hard

Tensor rank

Recall, the rank of a tensor A = [aijk ]l,m,ni,j,k=1 ∈ Fl×m×n is the minimum r for which A is a sum of

r rank-1 tensors.

rank(A) := min

{r : A =

r∑i=1

λixi ⊗ yi ⊗ zi

}

where λi ∈ F, xi ∈ Fl , yi ∈ Fm, and zi ∈ Fn.

1 Rank-1 tensors are tensors that could be expressed as an outer product of vectors.

2 More than one definition of tensor ‘rank’: e.g. symmetric rank, border rank.

3 Unlike matrices, rank of a tensor changes over changing fields.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 15 / 28

Page 34: Most Tensor Problems are NP-Hard

Tensor rank

Recall, the rank of a tensor A = [aijk ]l,m,ni,j,k=1 ∈ Fl×m×n is the minimum r for which A is a sum of

r rank-1 tensors.

rank(A) := min

{r : A =

r∑i=1

λixi ⊗ yi ⊗ zi

}

where λi ∈ F, xi ∈ Fl , yi ∈ Fm, and zi ∈ Fn.

1 Rank-1 tensors are tensors that could be expressed as an outer product of vectors.

2 More than one definition of tensor ‘rank’: e.g. symmetric rank, border rank.

3 Unlike matrices, rank of a tensor changes over changing fields.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 15 / 28

Page 35: Most Tensor Problems are NP-Hard

Tensor rank

Recall, the rank of a tensor A = [aijk ]l,m,ni,j,k=1 ∈ Fl×m×n is the minimum r for which A is a sum of

r rank-1 tensors.

rank(A) := min

{r : A =

r∑i=1

λixi ⊗ yi ⊗ zi

}

where λi ∈ F, xi ∈ Fl , yi ∈ Fm, and zi ∈ Fn.

1 Rank-1 tensors are tensors that could be expressed as an outer product of vectors.

2 More than one definition of tensor ‘rank’: e.g. symmetric rank, border rank.

3 Unlike matrices, rank of a tensor changes over changing fields.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 15 / 28

Page 36: Most Tensor Problems are NP-Hard

Tensor rank

Recall, the rank of a tensor A = [aijk ]l,m,ni,j,k=1 ∈ Fl×m×n is the minimum r for which A is a sum of

r rank-1 tensors.

rank(A) := min

{r : A =

r∑i=1

λixi ⊗ yi ⊗ zi

}

where λi ∈ F, xi ∈ Fl , yi ∈ Fm, and zi ∈ Fn.

1 Rank-1 tensors are tensors that could be expressed as an outer product of vectors.

2 More than one definition of tensor ‘rank’: e.g. symmetric rank, border rank.

3 Unlike matrices, rank of a tensor changes over changing fields.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 15 / 28

Page 37: Most Tensor Problems are NP-Hard

Matrix approximation

Last week: Recommender systems with large matrix A of rank r . Netflix competition:best rank k < r matrix that approximates A.

Singular value decomposition gives a sum of rank-1 matrices

A =r∑

i=1

σiui ⊗ vi where singular values σ1 ≥ · · · ≥ σr

Define Ak =∑k

i=1 σiui ⊗ vi , k < r .

Eckart-Young: If matrix B has rank k, then ‖A− B‖F ≥ ‖A− Ak‖F .(Works also with ‖ · ‖2 = σ1 and ‖ · ‖∗ = σ1 + · · ·+ σr )

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 16 / 28

Page 38: Most Tensor Problems are NP-Hard

Matrix approximation

Last week: Recommender systems with large matrix A of rank r . Netflix competition:best rank k < r matrix that approximates A.

Singular value decomposition gives a sum of rank-1 matrices

A =r∑

i=1

σiui ⊗ vi where singular values σ1 ≥ · · · ≥ σr

Define Ak =∑k

i=1 σiui ⊗ vi , k < r .

Eckart-Young: If matrix B has rank k, then ‖A− B‖F ≥ ‖A− Ak‖F .(Works also with ‖ · ‖2 = σ1 and ‖ · ‖∗ = σ1 + · · ·+ σr )

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 16 / 28

Page 39: Most Tensor Problems are NP-Hard

Matrix approximation

Last week: Recommender systems with large matrix A of rank r . Netflix competition:best rank k < r matrix that approximates A.

Singular value decomposition gives a sum of rank-1 matrices

A =r∑

i=1

σiui ⊗ vi where singular values σ1 ≥ · · · ≥ σr

Define Ak =∑k

i=1 σiui ⊗ vi , k < r .

Eckart-Young: If matrix B has rank k, then ‖A− B‖F ≥ ‖A− Ak‖F .(Works also with ‖ · ‖2 = σ1 and ‖ · ‖∗ = σ1 + · · ·+ σr )

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 16 / 28

Page 40: Most Tensor Problems are NP-Hard

Matrix approximation

Last week: Recommender systems with large matrix A of rank r . Netflix competition:best rank k < r matrix that approximates A.

Singular value decomposition gives a sum of rank-1 matrices

A =r∑

i=1

σiui ⊗ vi where singular values σ1 ≥ · · · ≥ σr

Define Ak =∑k

i=1 σiui ⊗ vi , k < r .

Eckart-Young: If matrix B has rank k, then ‖A− B‖F ≥ ‖A− Ak‖F .(Works also with ‖ · ‖2 = σ1 and ‖ · ‖∗ = σ1 + · · ·+ σr )

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 16 / 28

Page 41: Most Tensor Problems are NP-Hard

Rank-r tensor approximation

Fix F = RRank-r tensor approximation of tensor A = [aijk ]l,m,n

i,j,k=1 solves the problem

minxi ,yi ,zi

‖A− λ1x1 ⊗ y1 ⊗ z1 − · · · − λrxr ⊗ yr ⊗ zr‖F

Generally unsolvable when r > 1 because the set {A : rankR(A) ≤ r} is not closed whenr > 1. (Details for this in Silva and Lim 2008.)

Where r = 1, however, this set is closed. So consider the smaller problem of rank-1 tensorapproximation.

Problem simplifies tominx,y,z‖A− x ⊗ y ⊗ z‖F (5)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 17 / 28

Page 42: Most Tensor Problems are NP-Hard

Rank-r tensor approximation

Fix F = RRank-r tensor approximation of tensor A = [aijk ]l,m,n

i,j,k=1 solves the problem

minxi ,yi ,zi

‖A− λ1x1 ⊗ y1 ⊗ z1 − · · · − λrxr ⊗ yr ⊗ zr‖F

Generally unsolvable when r > 1 because the set {A : rankR(A) ≤ r} is not closed whenr > 1. (Details for this in Silva and Lim 2008.)

Where r = 1, however, this set is closed. So consider the smaller problem of rank-1 tensorapproximation.

Problem simplifies tominx,y,z‖A− x ⊗ y ⊗ z‖F (5)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 17 / 28

Page 43: Most Tensor Problems are NP-Hard

Rank-r tensor approximation

Fix F = RRank-r tensor approximation of tensor A = [aijk ]l,m,n

i,j,k=1 solves the problem

minxi ,yi ,zi

‖A− λ1x1 ⊗ y1 ⊗ z1 − · · · − λrxr ⊗ yr ⊗ zr‖F

Generally unsolvable when r > 1 because the set {A : rankR(A) ≤ r} is not closed whenr > 1. (Details for this in Silva and Lim 2008.)

Where r = 1, however, this set is closed. So consider the smaller problem of rank-1 tensorapproximation.

Problem simplifies tominx,y,z‖A− x ⊗ y ⊗ z‖F (5)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 17 / 28

Page 44: Most Tensor Problems are NP-Hard

Rank-r tensor approximation

Fix F = RRank-r tensor approximation of tensor A = [aijk ]l,m,n

i,j,k=1 solves the problem

minxi ,yi ,zi

‖A− λ1x1 ⊗ y1 ⊗ z1 − · · · − λrxr ⊗ yr ⊗ zr‖F

Generally unsolvable when r > 1 because the set {A : rankR(A) ≤ r} is not closed whenr > 1. (Details for this in Silva and Lim 2008.)

Where r = 1, however, this set is closed. So consider the smaller problem of rank-1 tensorapproximation.

Problem simplifies tominx,y,z‖A− x ⊗ y ⊗ z‖F (5)

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 17 / 28

Page 45: Most Tensor Problems are NP-Hard

Rank-1 tensor approximation

Introduce σ where x ⊗ y ⊗ z = σu ⊗ v ⊗ w and ‖u‖2 = ‖v‖2 = ‖w‖2 = 1.

Problem becomesminu,v,w

‖A− σu ⊗ v ⊗ w‖F

= ‖A‖2F − 2σ〈A, u ⊗ v ⊗ w〉+ σ2‖u ⊗ v ⊗ w‖2

F = ‖A‖2F − 2σ〈A, u ⊗ v ⊗ w〉

Above minimized when σ

σ = max‖u‖2=‖v‖2=‖w‖2=1

〈A, u ⊗ v ⊗ w〉

Because 〈A, u ⊗ v ⊗ w〉 = A(u, v ,w), can rewrite this as finding the spectral norm (4)

σ = ‖A‖2,2,2

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 18 / 28

Page 46: Most Tensor Problems are NP-Hard

Rank-1 tensor approximation

Introduce σ where x ⊗ y ⊗ z = σu ⊗ v ⊗ w and ‖u‖2 = ‖v‖2 = ‖w‖2 = 1.

Problem becomesminu,v,w

‖A− σu ⊗ v ⊗ w‖F

= ‖A‖2F − 2σ〈A, u ⊗ v ⊗ w〉+ σ2‖u ⊗ v ⊗ w‖2

F = ‖A‖2F − 2σ〈A, u ⊗ v ⊗ w〉

Above minimized when σ

σ = max‖u‖2=‖v‖2=‖w‖2=1

〈A, u ⊗ v ⊗ w〉

Because 〈A, u ⊗ v ⊗ w〉 = A(u, v ,w), can rewrite this as finding the spectral norm (4)

σ = ‖A‖2,2,2

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 18 / 28

Page 47: Most Tensor Problems are NP-Hard

Rank-1 tensor approximation

Introduce σ where x ⊗ y ⊗ z = σu ⊗ v ⊗ w and ‖u‖2 = ‖v‖2 = ‖w‖2 = 1.

Problem becomesminu,v,w

‖A− σu ⊗ v ⊗ w‖F

= ‖A‖2F − 2σ〈A, u ⊗ v ⊗ w〉+ σ2‖u ⊗ v ⊗ w‖2

F = ‖A‖2F − 2σ〈A, u ⊗ v ⊗ w〉

Above minimized when σ

σ = max‖u‖2=‖v‖2=‖w‖2=1

〈A, u ⊗ v ⊗ w〉

Because 〈A, u ⊗ v ⊗ w〉 = A(u, v ,w), can rewrite this as finding the spectral norm (4)

σ = ‖A‖2,2,2

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 18 / 28

Page 48: Most Tensor Problems are NP-Hard

Rank-1 tensor approximation

Introduce σ where x ⊗ y ⊗ z = σu ⊗ v ⊗ w and ‖u‖2 = ‖v‖2 = ‖w‖2 = 1.

Problem becomesminu,v,w

‖A− σu ⊗ v ⊗ w‖F

= ‖A‖2F − 2σ〈A, u ⊗ v ⊗ w〉+ σ2‖u ⊗ v ⊗ w‖2

F = ‖A‖2F − 2σ〈A, u ⊗ v ⊗ w〉

Above minimized when σ

σ = max‖u‖2=‖v‖2=‖w‖2=1

〈A, u ⊗ v ⊗ w〉

Because 〈A, u ⊗ v ⊗ w〉 = A(u, v ,w), can rewrite this as finding the spectral norm (4)

σ = ‖A‖2,2,2

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 18 / 28

Page 49: Most Tensor Problems are NP-Hard

Rank-1 tensor approximation

Deciding whether σ is the spectral norm of a tensor is shown in another section of thepaper to be NP-hard.

Say we can solve the rank-1 approximation problem (5) efficiently and we get solution(x , y , z). Then we can also find the spectral norm σ by setting

σ = ‖σu ⊗ v ⊗ w‖F = ‖x‖2‖y‖2‖z‖2

Finding the spectral norm is reducible to best rank-1 approximation.

So, best rank-1 approximation is also NP-hard.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 19 / 28

Page 50: Most Tensor Problems are NP-Hard

Rank-1 tensor approximation

Deciding whether σ is the spectral norm of a tensor is shown in another section of thepaper to be NP-hard.

Say we can solve the rank-1 approximation problem (5) efficiently and we get solution(x , y , z). Then we can also find the spectral norm σ by setting

σ = ‖σu ⊗ v ⊗ w‖F = ‖x‖2‖y‖2‖z‖2

Finding the spectral norm is reducible to best rank-1 approximation.

So, best rank-1 approximation is also NP-hard.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 19 / 28

Page 51: Most Tensor Problems are NP-Hard

Rank-1 tensor approximation

Deciding whether σ is the spectral norm of a tensor is shown in another section of thepaper to be NP-hard.

Say we can solve the rank-1 approximation problem (5) efficiently and we get solution(x , y , z). Then we can also find the spectral norm σ by setting

σ = ‖σu ⊗ v ⊗ w‖F = ‖x‖2‖y‖2‖z‖2

Finding the spectral norm is reducible to best rank-1 approximation.

So, best rank-1 approximation is also NP-hard.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 19 / 28

Page 52: Most Tensor Problems are NP-Hard

Rank-1 tensor approximation

Deciding whether σ is the spectral norm of a tensor is shown in another section of thepaper to be NP-hard.

Say we can solve the rank-1 approximation problem (5) efficiently and we get solution(x , y , z). Then we can also find the spectral norm σ by setting

σ = ‖σu ⊗ v ⊗ w‖F = ‖x‖2‖y‖2‖z‖2

Finding the spectral norm is reducible to best rank-1 approximation.

So, best rank-1 approximation is also NP-hard.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 19 / 28

Page 53: Most Tensor Problems are NP-Hard

Tensor Eigenvalues

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 20 / 28

Page 54: Most Tensor Problems are NP-Hard

Eigenpairs for matrices

Also fix F = R.

Given symmetric A ∈ Rn×n, eigenvalues and eigenvectors are stationary values and pointsof

R(x) =xᵀAx

xᵀx

Equivalently, constrained maximization max‖x‖22=1 x

ᵀAx with Lagrangian

L(x , λ) = xᵀAx − λ(‖x‖2

2 − 1)

The solution give the eigenvalue equation

Ax = λx

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 21 / 28

Page 55: Most Tensor Problems are NP-Hard

Eigenpairs for matrices

Also fix F = R.

Given symmetric A ∈ Rn×n, eigenvalues and eigenvectors are stationary values and pointsof

R(x) =xᵀAx

xᵀx

Equivalently, constrained maximization max‖x‖22=1 x

ᵀAx with Lagrangian

L(x , λ) = xᵀAx − λ(‖x‖2

2 − 1)

The solution give the eigenvalue equation

Ax = λx

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 21 / 28

Page 56: Most Tensor Problems are NP-Hard

Eigenpairs for matrices

Also fix F = R.

Given symmetric A ∈ Rn×n, eigenvalues and eigenvectors are stationary values and pointsof

R(x) =xᵀAx

xᵀx

Equivalently, constrained maximization max‖x‖22=1 x

ᵀAx with Lagrangian

L(x , λ) = xᵀAx − λ(‖x‖2

2 − 1)

The solution give the eigenvalue equation

Ax = λx

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 21 / 28

Page 57: Most Tensor Problems are NP-Hard

Eigenpairs for 3-tensor

Conceptually, for tensor A ∈ Rn×n×n, find the stationary values and points of cubic form (2)

A(x , x , x) :=n∑

i,j,k=1

aijkxixjxk

with some generalization of the unit constraint.

Which generalization?

‖x‖33 = 1?

‖x‖22 = 1?

x31 + · · ·+ x3

n = 1?

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 22 / 28

Page 58: Most Tensor Problems are NP-Hard

Eigenpairs for 3-tensor

Conceptually, for tensor A ∈ Rn×n×n, find the stationary values and points of cubic form (2)

A(x , x , x) :=n∑

i,j,k=1

aijkxixjxk

with some generalization of the unit constraint.

Which generalization?

‖x‖33 = 1?

‖x‖22 = 1?

x31 + · · ·+ x3

n = 1?

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 22 / 28

Page 59: Most Tensor Problems are NP-Hard

Eigenpairs for 3-tensor

Formally, fix F = R or C.

‖x‖22 = 1: The number λ ∈ F is called an l2-eigenvalue of the tensor A ∈ Fn×n×n and

0 6= x ∈ Fn its corresponding l2-eigenvector if

n∑i,j=1

aijkxixj = λxk k = 1, . . . , n holds

‖x‖33 = 1: The number λ ∈ F is called an l3-eigenvalue of the tensor A ∈ Fn×n×n and

0 6= x ∈ Fn its corresponding l3-eigenvector if

n∑i,j=1

aijkxixj = λx2k k = 1, . . . , n holds

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 23 / 28

Page 60: Most Tensor Problems are NP-Hard

Tensor eigenvalue over R is NP-hard

Theorem 1.3 Graph 3-colorability is polynomially reducible to tensor 0-eigenvalue over R. Thus,deciding tensor eigenvalue over R is NP-hard.

Proof outline

Restrict ourselves to the λ = 0 case. Both l2- and l3-eigenpair equations reduce to

n∑i,j=1

aijkxixj = 0 k = 1, . . . , n holds

The above becomes the square quadratic feasibility problem, which is deciding whether there isa 0 6= x ∈ Rn solution to a system of equations {xᵀAix = 0}mi=1.

By a previous result, graph 3-colorability is polynomially reducible to quadratic feasibility.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 24 / 28

Page 61: Most Tensor Problems are NP-Hard

Conclusion

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 25 / 28

Page 62: Most Tensor Problems are NP-Hard

Hard equals interesting

‘Bernd Sturmfels once made the remark to us that “All interesting problems are NP-hard.” Inlight of this, we would like to view our article as evidence that most tensor problems areinteresting.’

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 26 / 28

Page 63: Most Tensor Problems are NP-Hard

References

De Silva, V. & Lim, L.-H. Tensor rank and the ill-posedness of the best low-rank approximationproblem. SIAM J. Matrix Anal. Appl., 30(3): 1084-1127, 2008

Hillar, C.J. & Lim, L.-H.. Most tensor problems are NP-hard. J. ACM, 60(6):45:1-45:39, 2013.

Robeva, Elina. MATH605D Fall 2020 Tensor decomposition and their applications

https://sites.google.com/view/ubc-math-605d/class-overview

Sipser, M. Introduction to the Theory of Computation, 3rd Ed., 2012.

Strang, G. Linear Algebra and Learning From Data, 2019.

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 27 / 28

Page 64: Most Tensor Problems are NP-Hard

Thank you

UBC MLRG 2020 Winter Term 1 Most Tensor Problems are NP-Hard 14-Oct-2020 28 / 28