Dataset Viewer
Auto-converted to Parquet Duplicate
text
stringlengths
2
512
embedding
sequencelengths
768
768
Matrix Denoising with Doubly Heteroscedastic Noise: Fundamental Limits and Optimal Spectral Methods Yihan Zhang Institute of Science and Technology Austria zephyr.z798@gmail.com Marco Mondelli Institute of Science and Technology Austria marco.mondelli@ist.ac.at Abstract We study the matrix denoising problem of estimati...
[ -0.011110450141131878, 0.23494981229305267, -0.016464726999402046, 0.19671818614006042, -0.17327246069908142, 0.22296208143234253, -0.12856364250183105, -0.4026138186454773, 0.025135979056358337, -0.3536531925201416, -0.20055367052555084, 0.3331570327281952, -0.4114185869693756, 0.21925400...
works are either unable to pinpoint the exact asymptotic estimation error or, when they do so, the resulting approaches (e.g., based on whitening or singular value shrinkage) remain vastly suboptimal. On top of this, most of the literature has focused on the special case of estimating the left singular vector of the si...
[ 0.10198456794023514, -0.07098908722400665, -0.10941815376281738, 0.17469044029712677, -0.20277489721775055, 0.18335001170635223, -0.08232486248016357, -0.40242359042167664, -0.027722947299480438, -0.3925609588623047, -0.1938684731721878, 0.42303913831710815, 0.030774664133787155, 0.1940275...
contrast, our work establishes the information-theoretic and algorithmic limits of matrix denoising with doubly heteroscedastic noise. We characterize the exact asymptotic minimum mean square error, and design a novel spectral estimator with rigorous optimality guarantees: under a technical condition, it attains positi...
[ 0.002928248606622219, -0.04070032015442848, 0.0004898083861917257, 0.11757022142410278, -0.20508430898189545, 0.4052019417285919, -0.07350008934736252, -0.6545189619064331, -0.08046606928110123, -0.6584310531616211, -0.10916861146688461, 0.6453147530555725, -0.22182048857212067, 0.10613565...
one-sided heteroscedasticity, it also achieves the Bayes-optimal error. Numerical experiments demonstrate the significant advantage of our theoretically principled method with the state of the art. The proofs draw connections with statistical physics and approximate message passing, departing drastically from standard ...
[ -0.17276827991008759, 0.13172709941864014, -0.156040757894516, 0.23118956387043, -0.1899864375591278, -0.14100177586078644, -0.09934858977794647, -0.4196460545063019, 0.052123118191957474, -0.6265851855278015, 0.012861710973083973, 0.3972685933113098, -0.40711909532546997, 0.17071297764778...
Matrix denoising is a central primitive in statistics and machine learning, and the problem is to recover a signal X ∈ Rn×d from an observation A = X + W corrupted by additive noise W. This finds applications across multiple domains of sciences, e.g., imaging [21, 60], biology [13, 42] and astronomy [67, 5]. When X has...
[ 0.054392002522945404, -0.051499638706445694, 0.0685233399271965, 0.07521393895149231, -0.5585178136825562, 0.010524860583245754, -0.2783120274543762, -0.6179121136665344, 0.04216598719358444, 0.06930272281169891, -0.10102586448192596, 0.41388097405433655, -0.713725745677948, 0.302488952875...
component analysis, typically referred to as the Johnstone spiked covariance model [38]. When n, d are both large and proportional, which corresponds to the most sample-efficient regime, its Bayes-optimal limits are well understood [48], and it has been established how to achieve them efficiently [53]. Minimax/non-asym...
[ -0.03025408834218979, -0.4393158257007599, 0.19519872963428497, 0.34050193428993225, -0.18859313428401947, 0.2064838409423828, 0.1624588966369629, -0.5335655212402344, -0.04395497962832451, -0.01554141379892826, -0.2917514145374298, 0.8236665725708008, -0.34752562642097473, -0.090402968227...
sparse PCA [17], Gaussian mixtures [69] and certain joint scalings of (n, d) [54]. However, in most applications, noise is highly structured and correlated, thereby calling for more realistic assumptions on W than having i.i.d. entries. A recent line of work addresses this concern by studying matrix denoising with hete...
[ 0.13035106658935547, -0.1265375167131424, -0.032605089247226715, -0.2511976659297943, -0.21057918667793274, -0.017114590853452682, -0.24153079092502594, -0.4219585359096527, -0.28302666544914246, -0.3220566511154175, -0.16410361230373383, 0.05847584828734398, -0.6699129343032837, 0.4198119...
ideas: whitening and singular value shrinkage. Whitening refers to multiplying the data matrix by the square root of the inverse covariance, in order to reduce the model to one with i.i.d. noise; and singular value shrinkage retains the singular vectors of the data while deflating the singular values to correct for the...
[ 0.006064620800316334, -0.004590221215039492, -0.07696246355772018, -0.009579046629369259, -0.1730508804321289, 0.4136766791343689, -0.12243042141199112, -0.3679742217063904, -0.21569442749023438, -0.23310315608978271, 0.0391596257686615, 0.40722817182540894, -0.5499229431152344, 0.26704210...
prove that whitening and shrinkage are not the correct way to approach Bayes optimality. Preprint. Under review. arXiv:2405.13912v1 [math.ST] 22 May 2024 Main contributions. We focus on the prototypical model A = X + W, where X = λ nu∗v∗⊤ is a rank-1 signal, λ is the signal-to-noise ratio (SNR), and W = Ξ1/2� WΣ1/2 i...
[ -0.1329822987318039, -0.2078491598367691, 0.29808828234672546, -0.08694356679916382, -0.1687515377998352, 0.2347637414932251, 0.20666925609111786, -0.735339343547821, -0.22419233620166779, -0.5295178890228271, -0.1676347851753235, 0.3231945037841797, -0.18935628235340118, 0.382404088973999...
W contains i.i.d. Gaussian entries; the covariance matrices Ξ, Σ capture column and row correlations; and we consider the typical high-dimensional regime in which n, d are both large and proportional. Our main results are summarized below. 1. We design an efficient spectral estimator to recover u∗, v∗, and we provide a...
[ 0.050569143146276474, -0.47831231355667114, 0.25020089745521545, -0.07435595989227295, -0.12667694687843323, 0.18910524249076843, 0.3694501519203186, -0.8951903581619263, 0.2192826122045517, -0.6821157932281494, -0.20034907758235931, 0.45331189036369324, -0.10733465850353241, 0.17449599504...
vectors of a matrix obtained by carefully pre-processing A, see (5.3). 2. When the priors of u∗, v∗ are standard Gaussian, we show in Corollary 5.2 that the spectral estimator above is optimal in the following sense: (i) under a technical condition, it achieves the optimal weak recovery threshold, namely its mean squar...
[ -0.17633366584777832, -0.45110151171684265, 0.4138011038303375, -0.09668873250484467, -0.49817484617233276, 0.45358535647392273, 0.26703715324401855, -0.4758678674697876, -0.1296229362487793, -0.7852433919906616, -0.28897789120674133, 0.5262942314147949, -0.5437855124473572, -0.06056007370...
as this is information-theoretically possible; (ii) it achieves the Bayes-optimal error for u∗ (resp. v∗) when Ξ (resp. Σ) is the identity. These optimality guarantees follow from rigorously obtaining the asymptotic minimum mean square error (MMSE) for the estimation of the whitened signals Ξ−1/2u∗ and Σ−1/2v∗, see The...
[ -0.10381801426410675, -0.360792875289917, 0.31785115599632263, -0.30547964572906494, -0.31002718210220337, 0.24832552671432495, 0.2759808301925659, -0.7398487329483032, 0.045286279171705246, -0.714417576789856, -0.2925715148448944, 0.6374086141586304, -0.2459782212972641, 0.001790857873857...
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
134