Significance
The ability to detect sparse signals from noisy, high-dimensional data is a top priority in modern science and engineering. For optimal results, current approaches need to tune parameters that depend on the level of noise, which is often difficult to estimate. We develop a parameter-free, computationally efficient, -norm minimization approach that has a zero false discovery rate (no false positives) with high probability for any level of noise while it detects the exact location of sparse signals when the noise is not too large.
Keywords: high-dimensional probability, convex geometry, sparsity-promoting algorithms, noisy data
Abstract
The ability to detect sparse signals from noisy, high-dimensional data is a top priority in modern science and engineering. It is well known that a sparse solution of the linear system can be found efficiently with an -norm minimization approach if the data are noiseless. However, detection of the signal from data corrupted by noise is still a challenging problem as the solution depends, in general, on a regularization parameter with optimal value that is not easy to choose. We propose an efficient approach that does not require any parameter estimation. We introduce a no-phantom weight and the Noise Collector matrix and solve an augmented system , where is the noise. We show that the -norm minimal solution of this system has zero false discovery rate for any level of noise, with probability that tends to one as the dimension of increases to infinity. We obtain exact support recovery if the noise is not too large and develop a fast Noise Collector algorithm, which makes the computational cost of solving the augmented system comparable with that of the original one. We demonstrate the effectiveness of the method in applications to passive array imaging.
We want to find sparse solutions for
| [1] |
from highly incomplete measurement data corrupted by noise , where . In the noiseless case, can be found exactly by solving the optimization problem (1)
| [2] |
provided that the measurement matrix satisfies additional conditions (e.g., decoherence or restricted isometry properties) (2, 3) and that the solution vector has a small number of nonzero components or degrees of freedom. When measurements are noisy, exact recovery is no longer possible. However, the exact support of can still be determined if the noise is not too strong. The most commonly used approach is to solve the -relaxed form of Eq. 2:
| [3] |
which is known as Lasso in the statistics literature (4). There are sufficient conditions for the support of to be contained within the true support [e.g., the works of Fuchs (5), Tropp (6), Wainwright (7), and Maleki et al. (8)]. These conditions depend on the signal-to-noise ratio (SNR), which is not known and must be estimated, and on the regularization parameter , which must be carefully chosen and/or adaptively changed (9). Although such an adaptive procedure improves the outcome, the resulting solutions tend to include a large number of “false positives” in practice (10). Belloni et al. (11) proposed to solve the square root Lasso minimization problem instead of Eq. 3, which makes the regularization parameter independent of the SNR. Our contribution is a computationally efficient method for exact support recovery with no false positives in noisy settings. It also does not require an estimate on SNR.
Main Results
Suppose that is an -sparse solution of system [1] with no noise, where the columns of have unit length. Our main result ensures that we can still recover the support of when the data are noisy by looking at the support of found as
| [4] |
with an weight and an appropriately chosen Noise Collector matrix , . The minimization problem [4] can be understood as a relaxation of [2] as it works by absorbing all of the noise and possibly, some signal in .
The following theorem shows that, if the signal is pure noise and the columns of are chosen independently and at random on the unit sphere , then for any level of noise, with large probability.
Theorem 1 (No-Phantom Signal). Suppose that and that is uniformly distributed on . Fix , and draw columns for independently from the uniform distribution on . For any , there are constants and such that, for all , , the solution of Eq. 4, is 0 with probability .
This theorem guarantees with large probability a zero false discovery rate in the absence of signals with meaningful information. The key to a zero false discovery rate is the choice of a no-phantom weight . Next, we generalize this result for the case in which the recorded signals carry useful information.
Theorem 2 (Zero False Discoveries). Let be an -sparse solution of the noiseless system . Assume that , , the Noise Collector, and the noise are the same as in Theorem 1. In addition, assume that the columns of are incoherent in the sense that . Then, there are constants and such that for all with probability .
This theorem holds for any level of noise and the same value of as in Theorem 1. The incoherence conditions in Theorem 2 are needed to guarantee that the true signal does not create false positives elsewhere. Theorem 2 guarantees that the support of is inside the support of . The next theorem shows that, if the noise is not too large, then and have exactly the same support.
Theorem 3 (Exact Support Recovery). Keep the same assumptions as in Theorem 2. Let . There are constants , , and such that, if the noise level satisfies , then for all , with probability .
To elucidate an interpretation of the last theorem, consider a model case where is the identity matrix and all coefficients of are either one or zero. Then, . In this case, an acceptable relative level of noise is
| [5] |
This means that , and it implies that each coefficient of may be corrupted by on average and that some coefficients of may be corrupted by .
Motivation
We are interested in imaging sparse scenes accurately using limited and noisy data. Such imaging problems arise in many areas, such as medical imaging (12), structural biology (13), radar (14), and geophysics (15). In imaging, the -norm minimization method in Eq. 2 is often used (16–21) as it has the desirable property of superresolution: that is, the enhancement of the fine-scale details of the images. This has been analyzed in different settings by Donoho (22), Candès and Fernandez-Granda (23), Fannjiang and Liao (24), and Borcea and Kocyigit (25) among others. We want to retain this property in our method when the data are corrupted by additive noise.
However, noise fundamentally limits the quality of the images formed with almost all computational imaging techniques. Specifically, -norm minimization produces images that are unstable for low SNR due to the ill conditioning of superresolution reconstruction schemes. The instability emerges as clutter noise in the image, or grass, that degrades the resolution. Our initial motivation to introduce the Noise Collector matrix was to regularize the matrix and thus, to suppress the clutter in the images. We proposed in ref. 26 to seek the minimal -norm solution of the augmented linear system . The idea was to choose the columns of almost orthogonal to those of . Indeed, the condition number of becomes when columns of are taken at random. This essentially follows from the bounds on the largest and the smallest nonzero singular values of random matrices (theorem 4.6.1 in ref. 27).
The idea to create a dictionary for noise is not new. For example, the work by Laska et al. (28) considers a specific version of the measurement noise model so that , where is a matrix with fewer (orthonormal) columns than rows and the noise vector is sparse. represents the basis in which the noise is sparse and it is assumed to be known. Then, they show that it is possible to recover sparse signals and sparse noise exactly. We stress that we do not assume here that the noise is sparse. In our work, the noise is large (SNR can be small) and is evenly distributed across the data, and therefore, it cannot be sparsely accommodated.
To suppress the clutter, our theory in ref. 26 required exponentially many columns, and therefore, . This seemed to make the Noise Collector impractical, but the numerical experiments suggested that columns were enough to obtain excellent results. We address this issue here and explain why the Noise Collector matrix only needs algebraically many columns. Moreover, to absorb the noise completely and thus, improve the algorithm in ref. 26, we introduce now the no-phantom weight in Eq. 4. Indeed, by weighting the columns of the Noise Collector matrix with respect to those in the model matrix , the algorithm now produces images with no clutter at all regardless of how much noise is added to the data.
Finally, we want the Noise Collector to be efficient, with almost no extra computational cost with respect to the Lasso problem in Eq. 3. To this end, the Noise Collector is constructed using circulant matrices that allow for efficient matrix vector multiplications using fast Fourier transforms (FFTs).
We now explain how the Noise Collector works and reduce our theorems to basic estimates in high-dimensional probability.
The Noise Collector
The method has two main ingredients: the Noise Collector matrix and the no-phantom weight . The construction of the Noise Collector matrix starts with the following three key properties. First, its columns should be sufficiently orthogonal to the columns of , and therefore, it does not absorb signals with “meaningful” information. Second, the columns of should be uniformly distributed on the unit sphere so that we could approximate well a typical noise vector. Third, the number of columns in should grow slower than exponential with ; otherwise, the method is impractical.
One way to guarantee all three properties is to impose
| [6] |
with and fill out drawing at random with rejections until the rejection rate becomes too high. Then, by construction, the columns of are almost orthogonal to the columns of , and when the rejection rate becomes too high, this implies that we cannot pack more N-dimensional unit vectors into ; thus, we can approximate well a typical noise vector. Finally, the Kabatjanskii–Levenstein inequality (discussed in ref. 29) implies that the number of columns in grows at most polynomially: . The first estimate in Eq. 6 implies that any solution satisfies, for any , . This estimate measures how expensive it is to approximate columns of (i.e., the meaningful signal) with the Noise Collector. In turn, the no-phantom weight should be chosen so that it is expensive to approximate noise using columns of . It cannot be taken too large, however, because we may lose the signal. In fact, one can prove that, if , then for any and any level of noise. Intuitively, characterizes the rate at which the signal is lost as the noise increases. The most important property of the no-phantom weight is that it does not depend on the level of noise, and therefore, it is chosen before we start using the Noise Collector.
It is, however, more convenient for the proofs to use a probabilistic version of Eq. 6. Suppose that the columns of are drawn independently at random. Then, the dot product of any two random unit vectors is still typically of order (27). If the number of columns grows polynomially, we only have to sacrifice an asymptotically negligible event where our Noise Collector does not satisfy the three key properties, and the decoherence constraints in Eq. 6 are weakened by a logarithmic factor only. This follows from basic estimates in high-dimensional probability. We will state them in the next lemma after we interpret problem [4] geometrically.
Consider the convex hulls
| [7] |
| [8] |
and . Theorem 1 states that, for a typical noise vector , we can find such that and for any .
Lemma 1 (Typical Width of Convex Hulls ). Suppose that , ; vectors , , are drawn at random and independently; and . Then, for any , there are constants , , and such that, for all ,
| [9] |
and
| [10] |
with probability .
We sketch the proof of estimates [9] and [10] in Proofs. Estimate [9] can also be derived from Milman’s version (30) of Dvoretzky’s theorem. Informally, inequality [9] states that and are contained in the ball of radius except for a few spikes in statistically insignificant directions (Fig. 1, Left). Inequality [10] states that contains an ball of radius except for a few statistically insignificant directions.
Fig. 1.
(Left) A convex hull is an ball of radius with few spikes. (Right) An intersection of with the span is a rounded rhombus.
These inequalities immediately imply Theorem 1. We just need to explain how to choose the no-phantom weight . There will be no phantoms if is strictly inside the ball of radius . This could be done if .
If columns of are orthogonal to each other, then Theorem 2 follows from Theorem 1. We just need to project the linear system in Eq. 4 on the span of , , and apply Theorem 1 to the projections. If, in addition, we assume that , then Proof of Theorem 3 is illustrated in Fig. 1, Right. In detail, a typical intersection of , and is a rounded rhombus because it is the convex hull of and the ball of radius . If , then there are two options: 1) lies on the curved boundary of the rounded rhombus, and then, ; or 2) lies on the flat boundary of the rounded rhombus, and then, . The second option happens if the vector intersects the flat boundary of . This gives the support recovery estimate in Theorem 3.
In the general case, the columns of the combined matrix are incoherent. This property allows us to prove Theorems 2 and 3 in Proofs using known techniques (26). In particular, we automatically have exact recovery using ref. 2 applied to if the data are noiseless.
Lemma 2 (Exact Recovery). Suppose that is an -sparse solution of and that there is no noise so that . In addition, assume that the columns of are incoherent: . Then, the solution to Eq. 4 satisfies for all
| [11] |
Fast Noise Collector Algorithm
To find the minimizer in Eq. 4, we consider a variational approach. We define the function
| [12] |
for a no-phantom weight and determine the solution as
| [13] |
The key observation is that this variational principle finds the minimum in Eq. 4 exactly for all values of the regularization parameter . Hence, the method has no tuning parameters. To determine the exact extremum in Eq. 13, we use the iterative soft thresholding algorithm GeLMA (generalized Lagrangian multiplier algorithm) (31) that works as follows.
First, pick a value for and . For optimal results, one can calibrate to be the smallest constant such that Theorem 1 holds: that is, we see no-phantom signals when the algorithm is fed with pure noise. In our numerical experiments, we use and .
Second, pick a value for the regularization parameter (e.g., ). Choose step sizes and .* Set , , and , and iterate for :
| [14] |
where .
The Noise Collector matrix is computed by drawing normally distributed -dimensional vectors normalized to unit length. These are the generating vectors of the Noise Collector. From each of them, a circulant matrix , , is constructed. The Noise Collector matrix is obtained by concatenation, and therefore, . Exploiting the circulant structure of the matrices , we perform the matrix vector multiplications and in Eq. 14 using the FFT (32). This makes the complexity associated with the Noise Collector . Note that only the generating vectors are stored and not the entire Noise Collector matrix. In practice, we use , which makes the cost of using the Noise Collector negligible as typically, . The columns of the Noise Collector matrix with this circulant structure are uniformly distributed on , and they satisfy Lemma 1. This implies that the theorems of this paper are still valid for such .
Application to Imaging
We consider passive array imaging of point sources. The problem consists of determining the positions and the complex† amplitudes , , of a few point sources from measurements of polychromatic signals on an array of receivers (Fig. 2). The imaging system is characterized by the array aperture , the distance to the sources, the bandwidth , and the central wavelength .
Fig. 2.
General setup for passive array imaging. The source at emits a signal that is recorded at all array elements , .
The sources are located inside an image window (IW), which is discretized with a uniform grid of points , . The unknown is the source vector , with components that correspond to the complex amplitudes of the sources at the grid points , , with . For the true source vector, we have if for some , while otherwise.
Denoting by Green’s function for the propagation of a signal of angular frequency from point to point , we define the single-frequency Green’s function vector that connects a point in the IW with all points , , on the array as
In three dimensions, if the medium is homogeneous. The data for the imaging problem are the signals recorded at receiver locations , , at frequencies , . These data are stacked in a column vector
| [15] |
with . Then, , with being the measurement matrix with columns that are the multiple-frequency Green’s function vectors
| [16] |
normalized to have length 1. The system relates the unknown vector to the data vector .
Next, we illustrate the performance of the Noise Collector in this imaging setup. The most important features are that 1) no calibration is necessary with respect to the level of noise, that 2) exact support recovery is obtained for relatively large levels of noise [i.e., ], and that 3) we have zero false discovery rates for all levels of noise with high probability.
We consider a high-frequency microwave imaging regime with central frequency GHz corresponding to mm. We make measurements for equally spaced frequencies spanning a bandwidth GHz. The array has receivers and an aperture cm. The distance from the array to the center of the imaging window is cm. Then, the resolution is mm in the cross-range (direction parallel to the array) and mm in range (direction of propagation). These parameters are typical in microwave scanning technology (33).
We seek to image a source vector with sparsity (Fig. 3, Left). The size of the imaging window is 20 60 cm, and the pixel spacing is 5 15 mm. The number of unknowns is, therefore, , and the number of data is . The size of the Noise Collector is taken to be , and therefore, . When the data are noiseless, we obtain exact recovery as expected (Fig. 3, Right).
Fig. 3.
Noiseless data. The exact solution is recovered for any value of in algorithm [14]. (Left) The true image. (Right) The recovered solution vector, , is plotted with red stars, and the true solution vector, , is plotted with green circles.
In Fig. 4, we display the imaging results with and without the Noise Collector when the data are corrupted by additive noise. The SNR, and therefore, the norms of the signals and the noise are equal. In column 1 of Fig. 4, we show the recovered image using -norm minimization without the Noise Collector. There is a lot of grass in this image, with many nonzero values outside the true support. When the Noise Collector is used, the level of the grass is reduced, and the image improves (column 2 of Fig. 4). Still, there are several false discoveries because we use in algorithm [14].
Fig. 4.
High level of noise; SNR. (Column 1) -norm minimization without the noise collector. (Column 2) -norm minimization with a noise collector with columns and in algorithm [14]. (Column 3) -norm minimization with a noise collector and the correct in algorithm [14]. (Column 4) -norm solution restricted to the support. In Upper, we show the images. In Lower, we show the solution vector with red stars and the true solution vector with green circles. NC = noise collector.
In column 3 of Fig. 4, we show the image obtained with a weight in algorithm [14]. With this weight, there are no false discoveries, and the recovered support is exact. This simplifies the imaging problem dramatically as we can now restrict the inverse problem to the true support just obtained and then, solve an overdetermined linear system using a classical approach. The results are shown in column 4 of Fig. 4. Note that this second step largely compensates for the signal that was lost in the first step due to the high level of noise.
In Fig. 5, we illustrate the performance of the Noise Collector for different sparsity levels and values. Success in recovering the true support of the unknown corresponds to a value of one (yellow in Fig. 5), and failure corresponds to a value of zero (blue in Fig. 5). The small phase transition zone (green in Fig. 5) contains intermediate values. The black lines in Fig. 5 are the theoretical prediction Eq. 5. These results are obtained by averaging over 10 realizations of noise. We show results for three values of data sizes , , and . In our experiments, the nonzero components of the unknown take values in , and therefore, .
Fig. 5.
(Left) Algorithm performance for exact support recovery. Success corresponds to a value of one (yellow), and failure corresponds to a value of zero (blue). The small phase transition zone (green) contains intermediate values. The black lines are the theoretical estimate . Ordinate and abscissa are the sparsity and . The data sizes are (Left) , (Center) , and (Right) .
Remark 1: We considered passive array imaging for ease of presentation. The same results hold for active array imaging with or without multiple scattering; ref. 34 discusses the detailed analytical setup.
Remark 2: We have considered a microwave imaging regime. Similar results can be obtained in other regimes.
Proofs
Proof of Lemma 1: Using the rotational invariance of all of our probability distributions, inequality [9] is true if
where , are (possibly dependent) uniformly distributed on , and we can assume that . Denote the event
for each . We obtain using the union bound. Choosing for sufficiently large , we get , where and . Hence, Eq. 9 holds with probability .
If columns , of satisfy
| [17] |
then their convex hull will contain with probability . Therefore, inequality [10] follows if [17] holds with probability . Using the rotational invariance of all of our probability distributions, we can assume that . For each ,
Split the index set into nonoverlapping subsets , of size . For each ,
for . By independence,
Then, . Choosing sufficiently large, we obtain [10].
Proof of Theorem 2: When columns of are not orthogonal, we will choose a smaller than that in Theorem 1 by a factor of two. Suppose that the -dimensional space is the span of the column vectors , with in the support of . Say that is spanned by . Let be the orthogonal complement to . Consider the orthogonal decomposition for all . Incoherence of implies that for all . Indeed, fix any . Suppose that and that . Thus, Then, . Therefore, , and .
Project system [4] on . Then, we obtain a new system [4]. The norms of the columns of new are at least . Otherwise, the new system satisfies all conditions of Theorem 1. Indeed, is projected to zero. All and are projected to vectors uniformly distributed on by the concentration of measure (27). If any , , was used in an optimal approximation of , then its projection is used in an optimal approximation of the projection of on . This is a contradiction to Lemma 1 if we choose and recall that .
Proof of Theorem 3: Choose as in Theorem 2. Incoherence of implies that we can argue as in Proof of Theorem 2 and assume that for , . Suppose that are the two-dimensional (2D) spaces spanned by and for . By Lemma 1, all look like the rounded rhombi depicted in Fig. 1, Right, and with probability , where is a 2D ball of radius . Thus, with probability , where is the convex hull of and a vector , . Then, if there exists so that lies on the flat boundary of for all .
If , then there is a constant such that, if lies on the flat boundary of for some and some , then there exists so that lies on the flat boundary of for all . Suppose that is spanned by and , is the convex hull of and , and where is an ball of radius . If lies on the flat boundary of , then there must be an such that lies on the flat boundary of . If
| [18] |
then lies on the flat boundary of . Since with probability , Eq. 18 holds if .
Data Availability Statement.
There are no data in this paper; we present an algorithm, its theoretical analysis, and some numerical simulations.
Acknowledgments
The work of M.M. was partially supported by Spanish Ministerio de Ciencia e Innovación Grant FIS2016-77892-R. The work of A.N. was partially supported by NSF Grants DMS-1515187 and DMS-1813943. The work of G.P. was partially supported by Air Force Office of Scientific Research (AFOSR) Grant FA9550-18-1-0519. The work of C.T. was partially supported by AFOSR Grants FA9550-17-1-0238 and FA9550-18-1-0519. We thank Marguerite Novikov for drawing Fig. 1, Left.
Footnotes
The authors declare no competing interest.
*Choosing two step sizes instead of the smaller one improves the convergence speed.
†We chose to work with real numbers in the previous sections for ease of presentation, but the results also hold with complex numbers.
References
- 1.Chen S. S., Donoho D. L., Saunders M. A., Atomic decomposition by basis pursuit. SIAM Rev. 43, 12–159 (2001). [Google Scholar]
- 2.Donoho D. L., Elad M., Optimally sparse representation in general (nonorthogonal) dictionaries via l1 minimization. Proc. Natl. Acad. Sci. U.S.A. 100, 2197–2202 (2003). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Candès E. J., Tao T., Decoding by linear programming. IEEE Trans. Inf. Theory 51, 4203–4215 (2005). [Google Scholar]
- 4.Tibshirani R., Regression shrinkage and selection via the lasso. J. R. Stat. Soc. B 58, 267–288 (1996). [Google Scholar]
- 5.Fuchs J. J., Recovery of exact sparse representations in the presence of bounded noise. IEEE Trans. Inf. Theory 51, 3601–3608 (2005). [Google Scholar]
- 6.Tropp J. A., Just relax: Convex programming methods for identifying sparse signals in noise. IEEE Trans. Inf. Theory 52, 1030–1051 (2006). [Google Scholar]
- 7.Wainwright M. J., Sharp thresholds for high-dimensional and noisy sparsity recovery using -constrained quadratic programming (Lasso). IEEE Trans. Inf. Theory 55, 2183–2202 (2009). [Google Scholar]
- 8.Maleki A., Anitori L., Yang Z., Baraniuk R., Asymptotic analysis of complex Lasso via complex approximate message passing (CAMP). IEEE Trans. Inf. Theory 59, 4290–4308 (2013). [Google Scholar]
- 9.Zou H., The adaptive Lasso and its oracle properties. J. Am. Stat. Assoc. 101, 1418–1429 (2006). [Google Scholar]
- 10.Sampson J. N., Chatterjee N., Carroll R. J., Müller S., Controlling the local false discovery rate in the adaptive Lasso. Biostatistics 14, 653–666 (2013). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Belloni A., Chernozhukov V., Wang L., Square-root Lasso: Pivotal recovery of sparse signals via conic programming. Biometrika 98, 791–806 (2011). [Google Scholar]
- 12.Trzasko J., Manduca A., Highly undersampled magnetic resonance image reconstruction via homotopic -minimization. IEEE Trans. Med. Imag. 28, 106–121 (2009). [DOI] [PubMed] [Google Scholar]
- 13.AlQuraishi M., McAdams H. H., Direct inference of protein dna interactions using compressed sensing methods. Proc. Natl. Acad. Sci. U.S.A. 108, 14819–14824 (2011). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Baraniuk R., Steeghs P., “Compressive radar imaging” in 2007 IEEE Radar Conference (IEEE, 2007), pp. 128–133. [Google Scholar]
- 15.Taylor H. L., Banks SC., McCoy J. F., Deconvolution with the l1 norm. Geophysics 44, 39–52 (1979). [Google Scholar]
- 16.Malioutov D., Cetin M., Willsky A. S., A sparse signal reconstruction perspective for source localization with sensor arrays. IEEE Trans. Signal Process. 53, 3010–3022 (2005). [Google Scholar]
- 17.Romberg J., Imaging via compressive sampling. IEEE Signal Process. Mag. 25, 14–20 (2008). [Google Scholar]
- 18.Herman M. A., Strohmer T., High-resolution radar via compressed sensing. IEEE Trans. Signal Process. 57, 2275–2284 (2009). [Google Scholar]
- 19.Tropp J. A., Laska J. N., Duarte M. F., Romberg J. K., Baraniuk R. G., Beyond Nyquist: Efficient sampling of sparse bandlimited signals. IEEE Trans. Inf. Theory 56, 520–544 (2010). [Google Scholar]
- 20.Fannjiang A. C., Strohmer T., Yan P., Compressed remote sensing of sparse objects. SIAM J. Imag. Sci. 3, 595–618 (2010). [Google Scholar]
- 21.Chai A., Moscoso M., Papanicolaou G., Robust imaging of localized scatterers using the singular value decomposition and optimization. Inverse Probl. 29, 025016 (2013). [Google Scholar]
- 22.Donoho D. L., Superresolution via sparsity constraints. SIAM J. Math. Anal. 23, 1303–1331 (1992). [Google Scholar]
- 23.Candès E. J., Fernandez-Granda C., Towards a mathematical theory of super-resolution. Comm. Pure Appl. Math 67, 906–956 (2014). [Google Scholar]
- 24.Fannjiang A. C., Liao W., Coherence pattern-guided compressive sensing with unresolved grids. SIAM J. Imag. Sci. 5, 179–202 (2012). [Google Scholar]
- 25.Borcea L., Kocyigit I., Resolution analysis of imaging with optimization. SIAM J. Imag. Sci. 8, 3015–3050 (2015). [Google Scholar]
- 26.Moscoso M., Novikov A., Papanicolaou G., Tsogka C., Imaging with highly incomplete and corrupted data. Inverse Probl. 36, 035010 (2020). [Google Scholar]
- 27.Vershynin R., High-Dimensional Probability: An Introduction with Applications in Data Science (Cambridge Series in Statistical and Probabilistic Mathematics, Cambridge University Press, 2018). [Google Scholar]
- 28.Laska J. N., Davenport M. A., Baraniuk R. G., “Exact signal recovery from sparsely corrupted measurements through the pursuit of justice” in 2009 Conference Record of the Forty-Third Asilomar Conference on Signals, Systems and Computers (IEEE, Pacific Grove, CA, 2009), pp. 1556–1560. [Google Scholar]
- 29.Tao T., “A cheap version of the Kabatjanskii-Levenstein bound for almost orthogonal vectors.” https://terrytao.wordpress.com/2013/07/18/a-cheap-version-of-the-kabatjanskii-levenstein-bound-for-almost-orthogonal-vectors. Accessed 23 January 2019.
- 30.Milman V. D., A new proof of Dvoretzky’s theorem on cross-sections of convex bodies. Funkcional. Anal. I Priložen. 5, 28–37 (1971). [Google Scholar]
- 31.Moscoso M., Novikov A., Papanicolaou G., Ryzhik L., A differential equations approach to l1-minimization with applications to array imaging. Inverse Probl. 28, 10 (2012). [Google Scholar]
- 32.Gray R. M., Toeplitz and circulant matrices: A review. Commun. Inf. Theory 2, 155–239 (2006). [Google Scholar]
- 33.Laviada J., Arboleya-Arboleya A., Alvarez-Lopez Y., Garcia-Gonzalez C., Las-Heras F., Phaseless synthetic aperture radar with efficient sampling for broadband near-field imaging: Theory and validation. IEEE Trans. Antennas Propag. 63, 573–584 (2015). [Google Scholar]
- 34.Chai A., Moscoso M., Papanicolaou G., Imaging strong localized scatterers with sparsity promoting optimization. SIAM J. Imaging Sci 7, 1358–1387 (2014). [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
There are no data in this paper; we present an algorithm, its theoretical analysis, and some numerical simulations.





