Minimax risk of matrix denoising by singular value thresholding

D. Donoho,M. Gavish

Published 2013 in Annals of Statistics

ABSTRACT

An unknown m by n matrix X0 is to be estimated from noisy measurements Y = X0 + Z, where the noise matrix Z has i.i.d Gaussian entries. A popular matrix denoising scheme solves the nuclear norm penalization problem minXkY − Xk 2 /2 + λkXk∗, where kXk∗ denotes the nuclear norm (sum of singular values). This is the analog, for matrices, of l1 penalization in the vector case. It has been empirically observed that, if X0 has low rank, it may be recovered quite accurately from the noisy measurement Y . In a proportional growth framework where the rank rn, number of rows mn and number of columns n all tend to ∞ proportionally to each other (rn/mn → ρ, mn/n → β), we evaluate the asymptotic minimax MSE M(ρ,β) = lim mn,n→∞ inf λ sup rank(X)≤rn MSE(X, ˆ Xλ). Our formulas involve incomplete moments of the quarter- and semi-circle laws (β = 1, square case) and the Marycenko-Pastur law (β < 1, non square case). We also show that any least-favorable matrix X0 has norm “at infinity”. The nuclear norm penalization problem is solved by applying soft thresholding to the singular values of Y . We also derive the minimax threshold, namely the value λ ∗ (ρ) which is the optimal place to threshold the singular values. All these results are obtained for general (non square, non symmetric) real matrices. Comparable results are obtained for square symmetricnonnegative- definite matrices.

PUBLICATION RECORD

CITATION MAP

EXTRACTION MAP

CLAIMS

  • No claims are published for this paper.

CONCEPTS

  • No concepts are published for this paper.

REFERENCES

Showing 1-32 of 32 references · Page 1 of 1

CITED BY

Showing 1-100 of 129 citing papers · Page 1 of 2