*(English)*Zbl 1098.94009

Summary: Suppose we wish to recover a vector ${x}_{0}\in {\mathbb{R}}^{m}$ (e.g., a digital signal or image) from incomplete and contaminated observations $y=A{x}_{0}+e$; $A$ is an $n\times m$ matrix with far fewer rows than columns $(n\ll m)$ and $e$ is an error term. Is it possible to recover ${x}_{0}$ accurately based on the data y?

To recover ${x}_{0}$, we consider the solution ${x}^{\#}$ to the 1-regularization problem

where $\epsilon $ is the size of the error term $e$. We show that if $A$ obeys a uniform uncertainty principle (with unit-normed columns) and if the vector ${x}_{0}$ is sufficiently sparse, then the solution is within the noise level

As a first example, suppose that $A$ is a Gaussian random matrix; then stable recovery occurs for almost all such $A$’s provided that the number of nonzeros of ${x}_{0}$ is of about the same order as the number of observations. As a second instance, suppose one observes few Fourier samples of ${x}_{0}$; then stable recovery occurs for almost any set of coefficients provided that the number of nonzeros is of the order of $n/{(logm)}^{6}$. In the case where the error term vanishes, the recovery is of course exact, and this work actually provides novel insights into the exact recovery phenomenon discussed in earlier papers. The methodology also explains why one can also very nearly recover approximately sparse signals.

##### MSC:

94A12 | Signal theory (characterization, reconstruction, filtering, etc.) |

94A34 | Rate-distortion theory |

62B10 | Statistical information theory |

65K10 | Optimization techniques (numerical methods) |

94A08 | Image processing (compression, reconstruction, etc.) |