robust matrix completion via alternating projection · robust matrix completion . via alternating...

Post on 15-Oct-2020

11 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

H. C. So Page 1

Robust Matrix Completion

via Alternating Projection

Hing Cheung So

http://www.ee.cityu.edu.hk/~hcso Department of Electronic Engineering, City University of Hong Kong

X. Jiang, Z. Zhong, X. Liu and H. C. So, "Robust matrix completion via alternating projection," IEEE Signal Processing Letters, vol. 24, no. 5, pp. 579-583, May 2017

H. C. So Page 2

Outline Introduction

Matrix Completion as a Feasibility Problem

Alternating Projection Algorithm

Numerical Examples

Concluding Remarks

List of References

H. C. So Page 3

Introduction What is Matrix Completion? The aim is to recover a low‐rank matrix given only a subset of its possibly noisy entries, e.g.,

1 42 5

5 44 5

?? ?

? ? ?

????? ?

H. C. So Page 4

Denote the known entries of an incomplete matrix as :

where is a subset of the complete set of entries , with being the list while the unknown entries are assumed zero. Basically, matrix completion is to find a matrix , which is an estimate of , given with the use of low‐rank information of , which can be mathematically formulated as:

H. C. So Page 5

Why Matrix Completion is Important? It is a core problem in many applications including: Collaborative Filtering Image Inpainting and Restoration System Identification Node Localization Genotype Imputation It is because many real-world signals can be approximated by a matrix whose rank is . Netflix Prize, whose goal was to accurately predict user preferences with the use of a database of over 100 million movie ratings made by 480,189 users in 17,770 films,

H. C. So Page 6

which corresponds to the task of completing a matrix with around 99% missing entries.

AliceBob

CarolDave

1 42 5

5 44 5

H. C. So Page 7

How to Recover an Incomplete Matrix? Directly solving the noise-free version:

or noisy version:

is difficult because the rank minimization problem is NP-hard. A popular and practical solution is to replace the nonconvex rank by convex nuclear norm:

H. C. So Page 8

or

where equals the sum of singular values of . However, complexity of nuclear norm minimization is still high and this approach is not robust when contains outliers. Another popular direction which is computationally simple is to apply low-rank matrix factorization:

where and . However, generalization of the Frobenius norm to ‐norm for handling impulsive measurements is difficult.

H. C. So Page 9

Matrix Completion as a Feasibility Problem We formulate matrix completion with noise-free entries as: where an estimate or true value of is needed. It is called a feasibility problem because this optimization formulation has no objective function, but two constraints: Low-rank constraint: Fidelity constraint: With Gaussian noise, the fidelity constraint is modified as:

H. C. So Page 10

To achieve robustness, the feasibility problem is:

The rank constraint set is:

and the fidelity constraint set is:

where

is element‐wise ‐norm which is robust to outliers if .

H. C. So Page 11

We may rewrite the robust feasibility problem as:

Remarks: is a generalization as

reduces to noise-free version while reduces to conventional scenario of handling Gaussian noise.

We restrict our study for and since their projections onto have closed‐form expressions and are not difficult to compute.

requires computing projection onto a nonconvex

and nonsmooth -ball, which is difficult to compute.

H. C. So Page 12

Alternating Projection Algorithm

Define the projection of a point onto any constraint set , denoted as :

That is, projection onto rank constraint set is:

and projection onto fidelity constraint set is:

H. C. So Page 13

High-Level Algorithm The proposed alternating projection algorithm (APA) is outlined in Algorithm 1:

H. C. So Page 14

According to Eckart‐Young theorem, the projection of onto can be computed via truncated singular value decomposition (SVD) of :

where , , and are the largest singular values and the corresponding left and right singular vectors of , respectively. Assuming , the complexity is which is much smaller than that of full SVD of required in the nuclear norm minimization based methods, particularly when .

H. C. So Page 15

Noting that projection onto only affects the entries indexed by , we first define , which is a vector that contains the observed entries of , e.g., if

then

Hence has the equivalent vector form:

which is an ‐ball with the observed vector being ball center.

H. C. So Page 16

We consider the following three cases with different values of and :

For , reduces to equality constraint of .

For any vector , the projection is simply calculated as .

For and , is the conventional ‐ball in the

Euclidean space. For any vector , it is not difficult to derive the closed‐form expression of the projection onto

as

With a proper value of , the robustness to Gaussian noise is enhanced.

H. C. So Page 17

For and , is an ‐ball. For any vector , the projection onto is the solution of:

Using the Lagrange multiplier method, we obtain:

where , and is the unique root of the nonlinear equation:

H. C. So Page 18

The computational complexity of projection onto ‐ball is , which is much lower than that of projection onto

. Note that also involves the projection onto a convex ‐ball, which is not difficult to solve but requires an iterative procedure. As is more robust than in the presence of outliers, the latter case will not be considered.

Remarks: For the noise-free case, it is clear that is the

optimal value.

H. C. So Page 19

Roughly speaking, larger noise requires a larger for . If we know the probability density function (PDF)

of the noise, proper value of can be calculated.

Note that the nuclear norm regularized problem:

also faces the issue of selecting the user-defined .

Note also that our APA is different from the iterative hard thresholding (IHT) and its variants although they all use a rank‐ projection.

H. C. So Page 20

More precisely, IHT solves the rank constrained Frobenius norm minimization:

using gradient projection with update step being

where determining with a line search scheme requires computing projection for several times. Hence its computational cost is higher than APA per iteration.

We prove that if initial point is close enough to , then APA locally converges to at a linear rate.

H. C. So Page 21

Numerical Examples Noise‐free of rank is generated by the product of and whose entries satisfy standard Gaussian distribution, where , , and . 45% of the entries of are randomly selected as the known observations. Impulsive noise is modelled as two-term Gaussian mixture model (GMM) whose PDF is

H. C. So Page 22

Signal‐to‐noise ratio (SNR) is defined as:

Normalized root mean square error (RMSE) is defined as:

which is calculated based on 200 independent runs. Comparison with singular value thresholding (SVT) and IHT with , and , are included.

H. C. So Page 23

Figure 1: RMSE versus iteration number in noise-free case

0 20 40 60 80 100

Iteration number

10 -8

10 -6

10 -4

10 -2

10 0

Nor

mal

ized

RM

SE

H. C. So Page 24

Figure 2: RMSE versus iteration number at SNR=6dB

0 20 40 60 80 100

Iteration number

0

0.2

0.4

0.6

0.8

1

Nor

mal

ized

RM

SE

H. C. So Page 25

Figure 3: RMSE versus SNR in GMM noise

0 3 6 9 12 15 18 21

SNR (dB)

10 -1

10 0

Nor

mal

ized

RM

SE

H. C. So Page 26

Figure 4: RMSE versus estimated rank at SNR=12dB

4 5 6 7 8 9 10 11 12 13 14 15 16

Estimated rank

10 -1

10 0

Nor

mal

ized

RM

SE

H. C. So Page 27

Figure 5: RMSE versus with at SNR=12dB

0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.80

0.1

0.2

0.3

0.4

0.5

Nor

mal

ized

RM

SE

H. C. So Page 28

Concluding Remarks The key idea is to formulate matrix completion as a

feasibility problem, where a common point of the low-rank constraint set and fidelity constraint set is found by alternating projection.

The fidelity constraint set is modelled as an -ball, where or , which results in closed-form projection.

The APA achieves robustness against Gaussian noise and

outliers, with and , respectively. The APA is conceptually simpler and computationally

more efficient than the popular methods including the SVT and IHT.

H. C. So Page 29

List of References [1] http://www.netflixprize.com/ [2] E. J. Candès and Y. Plan, “Matrix completion with noise,”

Proc. IEEE, vol. 98, no. 6, pp. 925-936, Jun. 2010. [3] M. A. Davenport and J. Romberg, “An overview of low-

rank matrix recovery from incomplete observations,” IEEE J. Sel. Top. Signal Process., vol. 10, no. 4, pp. 608-622, Jun. 2016.

[4] E. J. Candès and T. Tao, “The power of convex relaxation: Near-optimal matrix completion,” IEEE Trans. Inf. Theory, vol. 56, no. 5, pp. 2053-2080, May 2010.

[5] B. Recht, M. Fazel and P. A. Parrilo, “Guaranteed minimum-rank solutions of linear matrix equations via

H. C. So Page 30

nuclear norm minimization,” SIAM Rev., vol. 52, no. 3, pp. 471-501, 2010.

[6] J.-F. Cai, E. J. Candès and Z. Shen, “A singular value thresholding algorithm for matrix completion,” SIAM J. Opt., vol. 20, no. 4, pp. 1956-1982, 2010.

[7] P. Jain, R. Meka and I. S. Dhillon, “Guaranteed rank minimization via singular value projection,” in Adv. Neural Inf. Process. Syst. (NIPS), pp. 937-945, 2010.

[8] J. Duchi, S. Shalev-Shwartz, Y. Singer and T. Chandra, “Efficient projections onto the L1-ball for learning in high dimensions,” in Proc. 25th Int. Conf. Machine Learning (ICML), pp.272-279, 2008.

[9] L. Condat, “Fast projection onto the simplex and the L1-ball,” Math. Program. Ser. A, vol. 158, no. 1, pp. 575-585, Jul. 2016.

H. C. So Page 31

[10] A. S. Lewis, D. R. Luke and J. Malick, “Local linear convergence for alternating and averaged nonconvex projections,” Found. Comp. Math., vol. 9, no. 4, pp 485-513, Aug. 2009.

[11] D. R. Luke “Prox-regularity of rank constraint sets and implications for algorithms,” J. Math. Imaging and Vision, vol. 47, no. 3, pp. 231-328, 2013.

[12] R. Hesse and D. R. Luke, “Nonconvex notions of regularity and convergence of fundamental algorithms for feasibility problems,” SIAM J. Optim., vol. 23, no. 4, pp. 2397-2419, 2013.

top related