Abstract
We study quantum imaging by applying the resolvable expressive capacity (REC) formalism developed for physical neural networks (PNNs). In this paradigm of quantum learning, the imaging system functions as a physical learning device that maps input parameters to measurable features, while complex practical tasks are handled by training only the output weights, enabled by the systematic identification of well-estimated features (eigentasks) and their corresponding sample thresholds. Using this framework, we analyze both direct imaging and superresolution strategies for compact sources, defined as sources with sizes bounded below the Rayleigh limit. In particular, we introduce the orthogonalized SPADE method—a nontrivial generalization of existing superresolution techniques—that achieves superior performance when multiple compact sources are closely spaced. This method relaxes the earlier superresolution studies’ strong assumption that the entire source must lie within the Rayleigh limit, marking an important step toward developing more general and practically applicable approaches. Using the example of face recognition, which involve complex structured sources, we demonstrate the superior performance of our orthogonalized SPADE method and highlight key advantages of the quantum learning approach—its ability to tackle complex imaging tasks and enhance performance by selectively extracting well-estimated features.
Subject terms: Quantum information, Quantum metrology, Imaging and sensing, Information theory and computation
Current superresolution imaging can beat the Rayleigh limit but struggles with complex imaging tasks. Here, the authors use a quantum-learning framework to address these challenges and introduce an improved method that enhances superresolution of nearby compact sources.
Introduction
The quality of an image formed by a single-lens system depends on several factors, including the lens’s resolution, the measurement strategy employed in the image plane, and the number of collected samples. Notably, it has been shown that by optimizing the measurement design, one can resolve two-point sources within the Rayleigh limit, surpassing the Rayleigh’s criterion1. The concept of superresolution has been extended in various directions2, focusing on imaging a single compact source—an object much smaller than the Rayleigh limit. These extensions include more careful treatment of the measurement and data analysis for two point sources3,4, general sources within the Rayleigh limit5–7, sources beyond the weak-source limit8–10, and point sources in higher dimensions11–13. These theories have been experimentally demonstrated for estimating point sources under various scenarios14–21 and for estimating source moments20.
Superresolution often relies on conventional statistical tools, such as the Fisher-information-matrix (FIM) approach for quantifying parameter-estimation precision2–21, and on the Chernoff bound or likelihood-ratio method for discrimination tasks22–26. However, these conventional statistical tools face at least two fundamental challenges when applied to complex imaging tasks. (i) Modeling complexity and ambiguity. It is often unclear which parameterization is most appropriate for a practical imaging task. Image moments, Fourier coefficients, and pixel intensities all provide complete representations but lead to different interpretations and performance. In addition, the likelihood-ratio method requires a full statistical model of the object, which is extremely difficult to construct for complex tasks—for example, in face recognition. (ii) Finite-sample reliability. With limited data, only a subset of features can be estimated accurately; many others are dominated by noise. Effective use of the data therefore requires identifying and retaining only well-estimated features for downstream analysis—an especially difficult step in imaging, where the number of parameters is, in principle, infinite.
The machine learning approach has achieved tremendous success in imaging applications, enabling the handling of complex practical tasks. In this work, we implement imaging tasks with a paradigm of quantum learning—physical neural networks (PNNs)27–40—following in particular the resolvable expressive capacity (REC) formalism in Ref. 40 to address the limitations of the FIM approach. The quantum learning approach adopted here encompasses both model training and inference using the trained model. PNNs encode inputs—such as the positions of incoherent point sources—into an analog physical system whose evolution is fixed and governed by its underlying physics, mapping the inputs to the measured high-dimensional features. A key advantage is that practical tasks can be performed by training only the output weight—equivalent to applying linear or logistic regression to the measured features—while leaving the internal structure unchanged. The relationship between the specific system dynamics and the class of tasks it can realize is a fundamental question in PNNs. The REC formalism addresses this question by identifying the resolvable features and quantifying their achievable precision in the presence of finite-sample noise40. The imaging system is an analog physical system that naturally functions as a learning device in PNNs, with output weights trained to perform imaging tasks. And applying the REC formalism to imaging identifies well-estimated features through eigentasks that are invariant under parameterization, thereby guiding the formulation of complex practical tasks for a given imaging system and prior information, directly addressing the first challenge faced by conventional statistical tools. Moreover, it provides a way to estimate the sample threshold required to detect each eigentask. This makes it particularly useful for determining which measured features can be reliably used in downstream analysis by selecting low-noise eigentasks based on a threshold. This key advantage, highlighted in the original paper40, can enhance the performance of diverse machine learning tasks, such as classification, regression, and clustering, and addresses the second challenge faced by conventional statistical tools. A more detailed introduction to the PNNs is provided in Supplementary Note 1 B.
Besides adopting a quantum learning perspective to study the imaging problem, we extend the existing discussions on superresolution1–22 to the broader challenge of imaging multiple compact sources, each individually constrained within the Rayleigh limit, or equivalently general sources that exceed the Rayleigh limit while exhibiting clustered substructures. This motivation arises as follows: prior studies typically assume that the whole source lies within the Rayleigh limit, i.e., it is a single compact source. However, in practical imaging, sources are often larger than the Rayleigh limit, while we aim to resolve fine features within the source that are below the Rayleigh limit. To address this, a natural idea is to partition the source into compact regions and apply superresolution techniques locally, effectively treating the problem as imaging multiple compact sources. But a straightforward application of superresolution to individual compact sources, referred to as the separate SPADE method, fails to offer an advantage over direct imaging. This is because nearby sources, when separated by distances not much greater than the width of the point spread function (PSF), introduce additional noise into the measurement. However, our generalized approach to superresolution, called the orthogonalized SPADE method, can achieve superresolution even when the separation between compact sources is as small as the PSF width. This discussion advances superresolution to imaging multiple reasonably nearby compact sources, constituting a nontrivial generalization of the earlier SPADE method and a step toward making the approach more practical.
Results
Preliminary
We now provide a more detailed explanation of how imaging tasks are realized using PNNs. The imaging system can be regarded as an input-output map, where the input is a set of system parameters θ (e.g., the locations of point sources) and the output is a set of measured features, which are the probabilities , where Mj denotes the jth element of the positive operator-valued measure (POVM). A general learning task—such as classification or regression—can be formulated in terms of a target function f(θ) of the input parameters. Realizing an imaging task is then equivalent to approximating this target function using the measured features, which corresponds to training the output weights of the PNNs27–39.
For a given physical system, it is important to determine the class of functions that can be approximated using the measured features. Moreover, due to sampling noise, the quantities can only be estimated approximately. It is also necessary to quantify the precision with which these functions can be approximated. These questions, concerning the capability of a physical system when regarded as a PNN, are addressed by the recently developed REC formalism36,40. REC formalism analyzes the resolvable function space of a physical system under sample noise via the concept of REC,
| 1 |
where we take the expectation value for the output samples and the prior distribution p(θ), f(θ) is approximated by a linear combination of measured functions , Wj represents the weight coefficient in front of to be optimized to achieve the optimal linear approximation of f(θ), where the index j corresponds to different measurement outcomes. REC C[f], which takes values between 0 and 1, can be understood as the normalized mean-squared accuracy of approximating f(θ), where C[f] = 1 represents a perfect approximation using the measured features, and deviation from 1 indicates that the target function cannot be well approximated.
To quantify the overall performance of an imaging system as PNNs, we are interested in identifying the set of functions f(θ) that can be approximated in this way and in quantifying the effective size of this set, given the physical imaging system, the finite number of samples S, and prior knowledge about the input θ. The total REC CT ≔ ∑kC[gk]40 fulfills this purpose, where can be any complete orthonormal basis of functions in the Hilbert space equipped with inner product . The value of total REC can be obtained from the following eigenvalue problem,
| 2 |
| 3 |
| 4 |
where and rk are the kth eigenvalue and eigenvector (in increasing order). The eigenbasis rk correspond to a minimal set of eigentasks fk(θ) ≔ ∑mrkmPm(θ) that saturate the available REC of the system in the space of all functions of input parameters θ. Then
| 5 |
where S is the number of samples. Intuitively, CT quantifies how many independent features of the underlying signal can be captured by the measurement process, and capturing more features improves performance on complex learning tasks by providing greater freedom to approximate the target function f(θ). Treating the imaging system as a PNN involves first identifying the eigentasks in the REC formalism and then training the output weights—essentially performing logistic or linear regression using the values of eigentasks fk(θ)—while including only the well-estimated eigentasks under a finite sample size and discarding noisy ones to ensure optimal performance. Another important property of this formalism is that reparameterization leaves both the total REC and the eigentasks unchanged (see Methods section), making the strategy independent of any artificial parameterization and determined solely by the physical system and the structure of the problem. A more detailed introduction to this formalism, proposed in ref. 40, is provided in Supplementary Note 1 C.
When we apply this quantum learning approach to the superresolution setting of imaging multiple compact sources, as shown in Fig. 1, we observe that the images exhibit two types of features: Rayleigh resolvable features and sub-Rayleigh features. The Rayleigh resolvable features are determined by the intensity of each compact source, which can be measured with a constant number of samples independent of the source size for both direct imaging and superresolution methods. In contrast, the sub-Rayleigh features involve details below the Rayleigh limit, requiring the number of samples to scale inversely with the source size, where carefully designed superresolution methods demonstrate clear advantages. Intuitively, the total REC quantifies the number of reliably estimated features—both large and small—and increases as smaller features become resolvable. We first examine the distinctive behavior of the eigenvalues and the total REC CT in each case of superresolution, and introduce our orthogonalized SPADE method. We then demonstrate the advantages of both this quantum learning approach and our new orthogonalized SPADE method through a concrete learning task.
Fig. 1. Imaging of both Rayleigh resolvable features and sub-Rayleigh features.
a Multiple compact sources of size Li are imaged by a lens with a PSF width of σ. Besides Rayleigh-resolvable features, we would also like to extract information from sub-Rayleigh features, which are associated with the small parameter α = L/σ. b The total REC, CT, which shows a stepwise increase, is plotted as a function of the number of samples, S. The threshold of S for each stepwise increase of CT in the shaded region is determined by the eigenvalue associated with the corresponding eigentask in learning. Each time CT increases by 1, there is a corresponding eigenvalue , with the sample number threshold following . The intensity of each compact source corresponds to resolvable features, which can be imaged with a constant number of samples, scaling as , independent of the source size. In contrast, sub-Rayleigh features that reveal detailed information about each compact source require a number of samples scaling inversely with the source size, following , where α is determined by the compact sources, and m depends on the order of moments.
Resolving two-point sources
As a simplest example, we begin with the imaging of two incoherent point sources in one dimension. A single photon received on the image plane can be described as , where , is the single photon state at position u, and we choose the PSF . Define the separation L, which is the input of learning task θ = L and assume u2 = L/2, u1 = − L/2. To enable analytical analysis, we focus on the binary SPADE measurement, which is capable of achieving superresolution in resolving two-point sources as introduced in ref. 1, where POVM , M1 = I − M0, , . Assuming the prior knowledge about the separation is described as , we can calculate the total REC, CT, which here represents the total number of linearly independent functions f(L) that can be expressed as a linear combination of the measured probabilities . Assuming γ ≪ σ to exhibit the advantage of superresolution within the Rayleigh limit, we find that
| 6 |
where α is roughly the ratio between the separation and width of the PSF and α ≪ 1 when the two-point sources are very close to each others.
We can compare this with the direct imaging case, where we directly project onto each spatial mode . In this case, we find that , , and after discretizing the spatial coordinate. More details of the calculations for both the direct imaging and SPADE methods are provided in Supplementary Note 2. The much larger eigenvalue in direct imaging indicates poorer performance compared to binary SPADE, as it requires a larger number of samples S to achieve the same CT.
Resolving a single compact source
We now consider the problem of imaging a single compact source, defined as a generally distributed source whose spatial extent is bounded well below the Rayleigh limit. This represents the most general setting for applying superresolution in imaging that has been considered in previous works2–22. Assume the normalized source intensity I(u) is confined within the interval [ − L/2, L/2]. We can define the moments as , which completely describe the source and are the input for the learning task . Within the Rayleigh limit, α = L/σ ≪ 1, where σ represents the width of the PSF, the size of the compact source is significantly smaller than the resolution limit. For any prior and PSF, we find that for the direct imaging
| 7 |
where is a trivial eigenvalue which corresponds to the fact that ∑mPm = 1.
For superresolution, we adopt the measurement construction from ref. 7, as reviewed in Supplementary Note 1 A. For simplicity, we still refer to this method as the SPADE method throughout the discussion. Intuitively, the SPADE method isolates higher-order moments by constructing probability distributions that start at higher-order terms of α, which serves as the signal strength, thereby suppressing lower-order terms of α that act as noise. This yields a much better signal-to-noise ratio for estimating those moments, especially in the weak-signal regime. For any prior and PSF, we find that for the SPADE method
| 8 |
Compared to direct imaging, the SPADE method achieves smaller , which significantly reduces the required S to achieve the same CT.
We demonstrate the significance of as the threshold for the stepwise increase in the total REC CT, as shown in Fig. 2. The total REC shows that each sets the sample size at which its eigentask contributes significantly, with contributions near 1 when and negligible when . For direct imaging, CT increases by 1 at each step, while for the SPADE method, CT increases by 2 per step, as expected. As α decreases, the plateau regions expand. All numerical calculations in this work assume a Gaussian PSF ; however, our method is applicable to any PSF. We choose the prior distribution for the moment vectors by randomly generating a set of images and assuming they occur with equal probability, thereby establishing as the empirical distribution of the resulting moment vectors, as detailed in Supplementary Note 6. For illustration purpose, we plot the data for only one instance of the randomly generated prior distributions (and in all figures below where the prior is picked randomly).
Fig. 2. Behavior of total REC for one compact source.

Total REC CT for direct imaging and the SPADE method as a function of S, when imaging one generally distributed compact source with different α.
Note that the threshold of S is not precisely located at α−2n in each case. This deviation arises from a constant prefactor in . This constant prefactor is independent of α and is ~102 in Fig. 2. This is reasonable because, even in the simpler imaging task where sources are extended outside the Rayleigh limit (i.e., α ≫ 1), hundreds or more samples are still required to effectively image a source. The prefactors depend on the imaging strategy and the prior information. We provide a more detailed discussion of these prefactors in Supplementary Note 7.
The nth eigentask corresponds to ∑mrnmPm as a function of , where rnm are the components of the eigenvectors obtained by solving Eq. (4), and the corresponding REC is given by . For direct imaging of a single compact source, we observe that the eigentasks converge to xn in the limit of small α to the leading order. For the SPADE method, we show that rnm becomes an triangular matrix as α → 0. The first two leading-order terms of both the 2k-th and (2k + 1)-th eigentasks have coefficients x2k and x2k+1.
Further details on the derivation of the scaling of and the eigenvectors rn, based on perturbation theory and confirmed by numerical calculations, are provided in Supplementary Note 3. So far, our discussion has focused on sources where the entire source lies within the Rayleigh limit. In this case, the total intensity, trivially equal to 1, is the Rayleigh resolvable feature that contributes to the total REC when only a constant number of samples (smaller than 1/α2) is available. The sub-Rayleigh features contribute to the total REC when Ω(α−2) samples are available. As we will see later, the Rayleigh resolvable features can become nontrivial when dealing with multiple compact sources.
New superresolution methods on multiple compact sources
We now want to consider the scenario where we have multiple compact sources, each with a size within the Rayleigh limit, but collectively distributed over a region larger than the Rayleigh limit. The quantum state from these multiple compact sources is given by
| 9 |
where Q is the number of compact sources, Iq(u) is the intensity distribution for qth compact source. We can expand near the centroid uq of qth source and reorganize the state as
| 10 |
where Lq is the size (diameter) of qth source, is the nth moment for the qth source and are the input for the learning task θ = [x0,1, x1,1, x2,1, ⋯ , x0,Q, x1,Q, x2,Q, ⋯ ]. We find that for the direct imaging
| 11 |
where . Here, we assume that Lq does not differ significantly, allowing the different Lq values to be incorporated into the constant coefficients. It is then clear that there are Q Rayleigh resolvable features corresponding to the intensity of each compact sources, and the number of sub-Rayleigh features also increases by a factor of Q. The scaling is numerically confirmed in Fig. 3(a) for the first six eigenvalues, and we expect to hold for eigenvalues with higher indices with any prior . In the numerical calculation, we assume two compact sources with centroids at − L/4 and L/4, with a random prior distribution for the moments (by randomly generating a set of images). For both L = 2 and L = 20, we observe the same scaling behavior in the direct imaging method.
Fig. 3. Scaling behavior of .

Scaling of the as a function of α for imaging two compact sources with distance L/2. We consider three different cases: (a) direct imaging (b) separate SPADE method (c) orthogonalized SPADE method. Width of PSF σ = 1.
To improve imaging performance, one could apply the SPADE method to each compact source individually—a technique referred to here as the separate SPADE method. Unfortunately, it only achieves the same scaling as direct imaging when the sources are not sufficiently spaced apart. This is because the proximity of other compact sources introduces significant noise when estimating higher-order moments. Alternatively, we can construct the orthonormal basis using the Gram-Schmidt procedure, such that
| 12 |
Choose POVM as the projection onto
| 13 |
where j = 1, 2, 3, ⋯ , Q, l = 0, 1, 2, ⋯ , ∞. The key intuition behind this construction is to ensure that when estimating the xn,q term in the Θ(αn) order, lower-order terms must vanish in the probability distribution, particularly those contributions from nearby compact sources. We refer to this new approach as the orthogonalized SPADE method, as it projects onto a basis that is an orthogonalization of the separate SPADE method. This construction applies analogously to any PSF beyond Gaussian PSF. Note that for a single compact source, the separate SPADE and orthogonalized SPADE methods are identical, both referred to as the SPADE method. We find that for the orthogonalized SPADE method
| 14 |
The scaling is numerically confirmed in Fig. 3 for the first six eigenvalues, and we expect to hold for eigenvalues with higher indices and any prior . In the numerical calculation, we again consider two compact sources with centroids at − L/4 and L/4 and a random prior obtained by randomly generating a set of images. We examine the separate SPADE method in Fig. 3(b) and the orthogonalized SPADE method in Fig. 3(c). When the sources are well separated (L ≫ σ), both methods achieve the expected scaling, with four terms scaling as Θ(α−2), compared to two for direct imaging. The doubling of the number of eigenvalues scaling as Θ(α−2n) for each n aligns with expectations for two compact sources. However, when the sources are closer (L = 2, σ = 1), the performance of the separate SPADE method is strongly degraded, reducing the scaling to that of direct imaging, with only two scaling as Θ(α−2). In contrast, our orthogonalized SPADE method retains four eigenvalues with scaling Θ(α−2).
In Fig. 4, we demonstrate the role of as the thresholds for stepwise increases in total REC CT for two compact sources. When L = 2 (sources close together), direct imaging and separate SPADE show that CT increases by 2 at each step after the initial two Θ(1) eigenvalues. In contrast, for orthogonalized SPADE, CT is increased by 4 after the initial two Θ(1) eigenvalues, highlighting the advantage of our method for two close compact sources. When L = 20 (with sources well separated), both separate SPADE and orthogonalized SPADE yield a CT increase of 4 at each step. Note that in certain regions of the sample number when α = 10−2, the orthogonalized SPADE method may perform slightly worse than the separate SPADE method. This difference arises from the different constant prefactors in . To ensure optimal performance, we can adopt an adaptive approach: for a given sample size, we select either the orthogonalized or separate SPADE method based on the total REC of each method, choosing the one that offers superior performance.
Fig. 4. Behavior of total REC for multiple compact sources.

Total REC CT for direct imaging, the separate SPADE method and the orthogonalized SPADE method as a function of S for imaging of two compact sources with α = 10−1, 10−2, 10−3, number of compact source Q = 2, σ = 1. The distance between the centroid of the two compact sources is L/2. When L = 2, the orthogonalized SPADE method demonstrates a clear advantage over both the separate SPADE method and direct imaging in the shaded region. When L = 20, the performance of SPADE and orthogonalized SPADE is comparable (essentially because they become equivalent when the two compact sources are sufficiently far apart) and both outperform direct imaging. Overall, the orthogonalized SPADE protocol demonstrates excellent performance, achieving a high CT (compared to the best from direct and SPADE protocols) for various choices of L = 2 and L = 20, as well as across a wide range of sample sizes S.
Note that in the imaging of multiple compact sources, the eigentasks identified from the eigenvectors rk generally do not have the simple form seen in the single compact source case. The structure of the eigentasks is strongly influenced by the practical imaging model and the prior knowledge, such as the positions of the individual sources. We explicitly demonstrate this in Supplementary Note 5 E. This observation suggests that, in practical applications, the quantum learning approach offers nontrivial guidance on which features should be incorporated into the downstream analysis.
In Fig. 5, we numerically illustrate the shape of the constructed basis for the separate SPADE method and the orthogonalized SPADE method, considering different distances between the centroids of the two compact sources. Note that the construction of the basis states defined in Eq. (12) does not depend on the size of each source Lq. It is evident that when the two compact sources are close to each other, the basis for the separate and orthogonalized SPADE methods differ significantly. However, when the two compact sources are sufficiently far from each other, the basis for the separate and orthogonalized SPADE methods become nearly identical. Given the complicated form of the basis, the spatial light modulator could serve as a practical tool for its implementation, as previously discussed in the context of superresolution41. In Supplementary Note 5, we demonstrate that the Hermite-Gaussian mode sorter42–46 can be used to implement the orthogonalized SPADE method with some additional steps for a Gaussian PSF, and we also provide more details on the derivation of the scaling of and the eigenvectors rk based on numerical calculations.
Fig. 5. Basis states for the orthogonalized SPADE measurement.
We compare the basis states constructed in Eq. (12) for the orthogonalized SPADE measurement, , with the corresponding basis states for the separate SPADE measurement, . Here, q = 1, 2 corresponds to the case of two compact sources (Q = 2) with centroids located at ± L/4, so that the separation between the two sources is L/2. We examine cases where L = 2, 6, 20. The shaded regions show two Gaussian PSFs of width σ = 1 located at the centers of two compact sources, that are used to construct the basis states and .
Demonstrative example
We now present a demonstrative example to illustrate the advantage of the quantum learning approach and our orthogonalized SPADE method in a face-recognition imaging task. The face images used in our simulation are taken from the Olivetti Faces dataset provided by AT&T Laboratories Cambridge and distributed through scikit-learn, with examples shown in Fig. 6b. Each individual has 64 × 64 grayscale images with different facial expressions and small variations in pose. These images are converted into one-dimensional images by rasterization, a standard preprocessing technique in machine learning, and the resulting images are partitioned into M = 3 segments that serve as compact sources placed over the interval according to the position function ζ(u) in Fig. 6a. Our goal is to determine the identity of each given face image. In the following, we treat the imaging system itself as PNNs—the quantum learning devices—and use it to perform this face-recognition task. We emphasize that this approach can address a wide range of imaging tasks beyond this one, such as regression and clustering.
Fig. 6. Simulation for face recognition.
a The position function ζ(u) specifies the positions of three compact sources, labeled S1, S2, and S3. All three compact sources are confined within the interval with , and each source has a size of 0.1, which is smaller than the PSF width σ = 1. b Example images of two individuals from the Olivetti Faces dataset. c The total REC CT as a function of the sample number S for the three approaches. (d1)-(d3) The success probability Psucc as a function of the highest order of the eigentasks included in the training and testing procedures for face recognition, evaluated at different sample sizes S. The lines represent the mean success probability, and the shaded region shows the maximum and minimum values across the 20 repetitions. We set α = 0.1 in this figure.
We use face images from Nperson = 20 individuals, which form the training set whose statistics are used to compute the prior information. Using the REC formalism in Eq. (4), we then calculate the eigenvectors rk. For the mth image, the measurement yields a probability distribution Pm(j), where j labels the measurement outcome. The kth eigentask is obtained as a linear combination of these distributions with coefficients given by the kth eigenvector rk, yielding ξkm = ∑jrkjPm(j). This defines the eigentask vector for the mth image , where the eigentasks are truncated at order . The truncation is introduced because higher-order eigentasks are noisy and poorly estimated, and including them in downstream analysis, namely training and inference, can degrade the performance. The shapes of the eigenvectors are generally complex, reflecting the complexity of the imaging task, which is common in practical applications. PNNs perform classification by training only the output weights using logistic regression. For each individual, we obtain eigentask vectors , forming datasets that serve as the training inputs for the logistic regression classifier. Before training, each component of is normalized by dividing by its mean absolute value across the training set to ensure balanced feature scaling. During training, we use logistic regression from the standard Python package scikit-learn to fit a multi-class classifier to the labeled eigentask vectors, modeling the probabilities with a multinomial logistic (softmax) model. During testing, we use the new face images for each individual. For each image, we fix the total number of detected photons (sample number) to be S. By counting the number of getting each outcome, we obtain an empirical eigentask vector for the mth face image. These empirical eigentask vectors are then used for inference with the trained model.
In the context of imaging multiple compact sources, we show the performance of face recognition using the three approaches in Fig. 6(d1)-(d3). We observe that the success probability Psucc first increases with and then decreases. This behavior is intuitive: increasing captures more information and improves classification, but beyond a point, higher-order tasks become noisy due to limited sample size S, degrading performance and reducing success probability. Comparing the success probability Psucc of the three approaches, since the source distance is comparable to the PSF width σ, separate SPADE and direct imaging perform similarly, whereas our orthogonalized SPADE achieves higher performance, with its peak Psucc exceeding the others at S = 1010.
This simulation highlights the operational meaning of the total REC CT. First, it estimates the number of eigentasks that can be reliably included in the downstream logistic regression. Since sets the sample threshold for estimating the kth eigentask and , with each term approaching 1 when , CT roughly counts the reliably estimated eigentasks. At S = 106 and S = 108, CT is similar across all approaches, matching the Psucc behavior. At S = 1010, CT for orthogonalized SPADE rises to about 8, aligning with its peak Psucc, while direct imaging and separate SPADE reach about 6, consistent with their peaks. Second, a larger CT allows more well-estimated eigentasks to be included, capturing more information and thus improving the success probability in this example; therefore, a larger CT indicates better imaging performance. Further details of this simulation, as well as additional examples beyond face recognition, are provided in Supplementary Note 8.
In the broader context, we here present an example that directly addresses the imaging task using PNNs. PNNs employ analog systems with a fixed internal structure—here, the imaging system—and performs practical tasks by training only the output weights, which in this example corresponds to logistic regression, though alternatives like linear regression can be used depending on the task. In this sense, the PNNs framework provides a systematic approach to modeling and solving complex imaging problems, which is especially valuable given the infinite degrees of freedom inherent to imaging. The REC formalism, developed for the PNNs paradigm, further guides the training step by identifying eigentasks and selecting those that are well-estimated and low-noise. This formalism has also proven highly effective in superresolution problems involving complex source structures, as demonstrated in our simulation.
Many machine-learning tasks, including face recognition, can be viewed as discrimination problems that are in principle solvable by the likelihood-ratio method and whose performance is bounded by the Chernoff bound. While simple discrimination settings have been analyzed using the Chernoff bound in superresolution22–26, the likelihood-ratio method requires an accurate statistical model of the objects being imaged, whereas learning methods can handle far more complex structures. In our face-recognition example involving tens of individuals, it is infeasible to write down a closed-form likelihood function for an exact likelihood-ratio calculation. Moreover, the Chernoff bound is asymptotic and may appear to be reached even when the success probability is already above 99% in some cases, offering little guidance on the practically relevant regime where we care about reaching moderate performance levels such as 70%. It also does not suggest which measurement strategy should be used when the likelihood-ratio method is inapplicable. By contrast, our learning-based approach yields the total REC as a meaningful figure of merit and, crucially, provides a principled strategy for tackling discrimination tasks in the finite-sample regime. Simulated examples illustrating the above discussion are provided in Supplementary Note 9.
Imaging general sources beyond the Rayleigh limit
An intriguing question is whether, when a general source cannot be split into multiple compact sources, we can still split the source into Q small pieces and apply our orthogonalized SPADE method, improving the imaging performance. Unfortunately, there is a constraint. Each source generates a set of states , which are used to construct via the Gram-Schmidt procedure. When Q ≫ L/σ, the differences between for different q can be vanishingly small, leading to the potentially suboptimal performance of orthogonalized SPADE method due to large prefactors of the eigenvalues . In Supplementary Note 5C, we demonstrate that when Q exceeds roughly L/σ, the stepwise increase in CT is smoothed out, and numerically, we find that both the separate SPADE and the orthogonalized SPADE no longer offer advantages over direct imaging. In conclusion, the superresolution methods from refs. 1,7 can be properly generalized to resolve multiple compact sources but may fail for generally distributed sources. We also present a discussion of imaging such a general source using direct imaging, showing that CT is approximately related to the ratio between the source size and the PSF width, as detailed in Supplementary Note 4.
Discussion
In this work, we treat the imaging systems in superresolution as PNNs, i.e., quantum learning devices, thereby providing a systematic framework for addressing practical imaging tasks with complex structures and operating in the finite-sample regime. Based on the REC formalism, the measurable features of the source and their corresponding sample thresholds can be identified with eigentasks, while the total REC serves as a principled metric for selecting relevant features for downstream analysis and quantifying the performance of an imaging method. We further extend the superresolution framework to handle multiple compact sources and propose the orthogonalized SPADE method—a nontrivial generalization that relaxes the strong assumptions of earlier superresolution studies, thereby improving practical applicability. We show that superresolution exhibits a stepwise increase in total REC, with thresholds determined by the ratio of source size to PSF width. The advantages of this quantum learning approach and our orthogonalized SPADE method are demonstrated through total REC calculations and concrete examples, including face recognition.
It would also be worthwhile to investigate other potentially advantageous imaging protocols, e.g., entangled measurements on multiple copies of photon states, which exhibit advantages over separable measurements in tomography47,48. We may also explore the potential applications of quantum computing schemes49–51, where the quantum advantage inherent in these schemes could benefit specific imaging tasks. Note that the total REC depends on the POVM used in the detection. It may be interesting to explore whether a closed-form expression of the total REC optimized over all possible POVMs can be found, at least in some special cases—for example, when the measurable features require compatible measurements.
Methods
Reparameterization invariance of total REC and eigentasks
We emphasize that a key advantage of the REC formalism is that the identified eigentasks and the total REC are invariant under reparameterization. These eigentasks are determined by the prior information, the physical model of the learning device, and the structure of the learning tasks. They are then used as the actual feature vectors for downstream analysis, including model training and inference with the trained model. This ensures that the results are not artificially influenced by the choice of parameterization. The invariance under reparameterization is formally established in the following proposition.
Proposition 1
Let ρ(θ) be a family of states parameterized by θ with prior p(θ), a fixed POVM , sample size S, and measured features . In the parameterization of θ, the total REC at sample size S is CT(S), with eigenvalues and eigentasks defined by solving Eq. (4), yielding fk(θ) = ∑jrkj ηj(θ). Let ϕ = h(θ) be a bijective differentiable reparameterization, with pushforward prior so that pΦ(ϕ) dϕ = p(θ) dθ, and reparameterized family . As in Eq. (4), form , , , , the generalized spectrum , eigentasks with , and total REC . Then:
- (Total REC invariance)
- (Spectral invariance)
- (Eigentask invariance)
Proof
By definition and the pushforward prior,
| 15 |
Set θ = h−1(ϕ). Then and , so the Jacobians cancel:
This relation implies
| 16 |
Hence . With and , we obviously have
| 17 |
The generalized eigenproblems
are also identical, so and . Using ,
| 18 |
Supplementary information
Acknowledgements
We would like to thank Hakan E. Tureci for helpful discussion. Y.W. and S.Z. acknowledge funding provided by Perimeter Institute for Theoretical Physics, a research institute supported in part by the Government of Canada through the Department of Innovation, Science and Economic Development Canada and by the Province of Ontario through the Ministry of Colleges and Universities. Y.W. also acknowledges funding from the Canada First Research Excellence Fund. L.J. acknowledges support from the ARO(W911NF-23-1-0077), ARO MURI (W911NF-21-1-0325), AFOSR MURI (FA9550-19-1-0399, FA9550-21-1-0209, FA9550-23-1-0338), DARPA (HR0011-24-9-0359, HR0011-24-9-0361), NSF (OMA-1936118, ERC-1941583, OMA-2137642, OSI-2326767, CCF-2312755), NTT Research, Packard Foundation (2020-71479), and the Marshall and Arlene Bennett Family Research Program. J.L. is supported in part by the University of Pittsburgh, School of Computing and Information, Department of Computer Science, Pitt Cyber, Pitt Momentum fund, PQI Community Collaboration Awards, John C. Mascaro Faculty Scholar in Sustainability, Thinking Machines Lab, Cisco Research, funding from IBM Quantum through the Chicago Quantum Exchange, and AFOSR MURI (FA9550-21-1-0209). C.O. was supported by the NRF Grants (No. RS-2024-00431768 and No. RS-2025-00515456) funded by the Korean government (MSIT) and IITP grants funded by the Korean government (MSIT) (No. IITP-2025-RS-2025-02283189 and IITP-2025-RS-2025-02263264).
Author contributions
Y.W. carried out the analytical calculation and the numerical simulation. L.J. conceived the project. S.Z., J.L., and L.J. supervised the project. Y.W., C.O., J.L., L.J., and S.Z. contributed to the development of ideas and the writing of the manuscript.
Peer review
Peer review information
Nature Communications thanks Giacomo Sorelli and the other, anonymous, reviewers for their contribution to the peer review of this work. A peer review file is available.
Data availability
No data sets were generated or analyzed during the current study.
Code availability
All codes used in this paper have been deposited in GitHub at https://github.com/ykwang-phys/quantum-learning-imaging.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Yunkai Wang, Email: ywang10@perimeterinstitute.ca.
Liang Jiang, Email: liangjiang@uchicago.edu.
Sisi Zhou, Email: sisi.zhou26@gmail.com.
Supplementary information
The online version contains supplementary material available at 10.1038/s41467-025-67884-1.
References
- 1.Tsang, M., Nair, R. & Lu, X.-M. Quantum theory of superresolution for two incoherent optical point sources. Phys. Rev. X6, 031033 (2016). [Google Scholar]
- 2.Pirandola, S., Bardhan, B. R., Gehring, T., Weedbrook, C. & Lloyd, S. Advances in photonic quantum sensing. Nat. Photonics12, 724–733 (2018). [Google Scholar]
- 3.Sorelli, G., Gessner, M., Walschaers, M. & Treps, N. Optimal observables and estimators for practical superresolution imaging. Phys. Rev. Lett.127, 123604 (2021). [DOI] [PubMed] [Google Scholar]
- 4.Grace, M. R., Dutton, Z., Ashok, A. & Guha, S. Approaching quantum-limited imaging resolution without prior knowledge of the object location. J. Optical Soc. Am. A37, 1288–1299 (2020). [DOI] [PubMed] [Google Scholar]
- 5.Tsang, M. Quantum limit to subdiffraction incoherent optical imaging. Phys. Rev. A99, 012305 (2019). [Google Scholar]
- 6.Tsang, M. Subdiffraction incoherent optical imaging via spatial-mode demultiplexing. N. J. Phys.19, 023054 (2017). [Google Scholar]
- 7.Zhou, S. & Jiang, L. Modern description of Rayleigh’s criterion. Phys. Rev. A99, 013808 (2019). [Google Scholar]
- 8.Wang, Y., Zhang, Y. & Lorenz, V. O. Superresolution in interferometric imaging of strong thermal sources. Phys. Rev. A104, 022613 (2021). [Google Scholar]
- 9.Nair, R. & Tsang, M. Far-field superresolution of thermal electromagnetic sources at the quantum limit. Phys. Rev. Lett.117, 190801 (2016). [DOI] [PubMed] [Google Scholar]
- 10.Lupo, C. & Pirandola, S. Ultimate precision bound of quantum and subwavelength imaging. Phys. Rev. Lett.117, 190802 (2016). [DOI] [PubMed] [Google Scholar]
- 11.Napoli, C., Piano, S., Leach, R., Adesso, G. & Tufarelli, T. Towards superresolution surface metrology: quantum estimation of angular and axial separations. Phys. Rev. Lett.122, 140505 (2019). [DOI] [PubMed] [Google Scholar]
- 12.Yu, Z. & Prasad, S. Quantum limited superresolution of an incoherent source pair in three dimensions. Phys. Rev. Lett.121, 180504 (2018). [DOI] [PubMed] [Google Scholar]
- 13.Ang, S. Z., Nair, R. & Tsang, M. Quantum limit for two-dimensional resolution of two incoherent optical point sources. Phys. Rev. A95, 063847 (2017). [Google Scholar]
- 14.Yang, F., Tashchilina, A., Moiseev, E. S., Simon, C. & Lvovsky, A. I. Farfield linear optical superresolution via heterodyne detection in a higher-order local oscillator mode. Optica3, 1148–1152 (2016). [Google Scholar]
- 15.Tang, Z. S., Durak, K. & Ling, A. Fault-tolerant and finite-error localization for point emitters within the diffraction limit. Opt. Express24, 22004–22012 (2016). [DOI] [PubMed] [Google Scholar]
- 16.Paúr, M., Stoklasa, B., Hradil, Z., Sánchez-Soto, L. L. & Rehacek, J. Achieving the ultimate optical resolution. Optica3, 1144–1147 (2016). [Google Scholar]
- 17.Tham, W.-K., Ferretti, H. & Steinberg, A. M. Beating Rayleigh’s curse by imaging using phase information. Phys. Rev. Lett.118, 070801 (2017). [DOI] [PubMed] [Google Scholar]
- 18.Parniak, M. et al. Beating the Rayleigh limit using two-photon interference. Phys. Rev. Lett.121, 250503 (2018). [DOI] [PubMed] [Google Scholar]
- 19.Santamaria, L., Sgobba, F. & Lupo, C. Single-photon sub-Rayleigh precision measurements of a pair of incoherent sources of unequal intensity. Opt. Quantum2, 46–56 (2024). [Google Scholar]
- 20.Tan, X.-J. et al. Quantum-inspired superresolution for incoherent imaging. Optica10, 1189–1194 (2023). [Google Scholar]
- 21.Rouvière, C. et al. Ultra-sensitive separation estimation of optical sources. Optica11, 166–170 (2024). [Google Scholar]
- 22.Zanforlin, U. et al. Optical quantum super-resolution imaging and hypothesis testing. Nat. Commun.13, 5373 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Huang, Z. & Lupo, C. Quantum hypothesis testing for exoplanet detection. Phys. Rev. Lett.127, 130502 (2021). [DOI] [PubMed] [Google Scholar]
- 24.Zhang, H., Kumar, S. & Huang, Y.-P. Super-resolution optical classifier with high photon efficiency. Opt. Lett.45, 4968–4971 (2020). [DOI] [PubMed] [Google Scholar]
- 25.Lu, X.-M., Krovi, H., Nair, R., Guha, S. & Shapiro, J. H. Quantum-optimal detection of one-versus-two incoherent optical sources with arbitrary separation. npj Quantum Inf.4, 64 (2018). [Google Scholar]
- 26.Grace, M. R. & Guha, S. Identifying objects at the quantum limit for superresolution imaging. Phys. Rev. Lett.129, 180502 (2022). [DOI] [PubMed] [Google Scholar]
- 27.Boyd, S. & Chua, L. Fading memory and the problem of approximating nonlinear operators with Volterra series. IEEE Trans. Comput. -Aided Des. Integr. Circuits Syst.32, 1150 (1985). [Google Scholar]
- 28.Tanaka, G. et al. Recent advances in physical reservoir computing: a review. Neural Netw.115, 100 (2019). [DOI] [PubMed] [Google Scholar]
- 29.Mujal, P. et al. Opportunities in Quantum Reservoir Computing and extreme learning machines. Adv. Quantum Technol.4, 2100027 (2021). [Google Scholar]
- 30.Wilson, C. M. et al. Quantum kitchen sinks: an algorithm for machine learning on near-term quantum computers. Preprint at arXiv10.48550/arXiv.1806.08321 (2018).
- 31.García-Beni, J., Giorgi, G. L., Soriano, M. C. & Zambrini, R. Scalable photonic platform for real-time quantum reservoir computing. Phys. Rev. Appl.20, 014051 (2023). [Google Scholar]
- 32.Havlíček, V. et al. Supervised learning with quantum-enhanced feature spaces. Nat. (Lond.)567, 209 (2019). [DOI] [PubMed] [Google Scholar]
- 33.Rowlands, G. E. et al. Reservoir computing with superconducting electronics. Preprint at arXiv:2103.02522 (2021).
- 34.Lin, X. et al. All-optical machine learning using diffractive deep neural networks. Science361, 1004 (2018). [DOI] [PubMed] [Google Scholar]
- 35.Pai, S. et al. Experimentally realized in situ backpropagation for deep learning in photonic neural networks. Science380, 398 (2023). [DOI] [PubMed] [Google Scholar]
- 36.Dambre, J., Verstraeten, D., Schrauwen, B. & Massar, S. Information processing capacity of dynamical systems. Sci. Rep.2, 514 (2012). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Sheldon, F. C., Kolchinsky, A. & Caravelli, F. Computational capacity of LRC, memristive and hybrid reservoirs. Phys. Rev. E106, 045310 (2022). [DOI] [PubMed] [Google Scholar]
- 38.Schuld, M., Sweke, R. & Meyer, J. J. Effect of data encoding on the expressive power of variational quantum-machine-learning models. Phys. Rev. A103, 032430 (2021). [Google Scholar]
- 39.Wu, Y., Yao, J., Zhang, P. & Zhai, H. Expressivity of quantum neural networks. Phys. Rev. Res.3, L032049 (2021). [Google Scholar]
- 40.Hu, F. et al. Tackling sampling noise in physical systems for machine learning applications: fundamental limits and eigentasks. Phys. Rev. X13, 041020 (2023). [Google Scholar]
- 41.Ozer, I., Grace, M. R. & Guha, S. Reconfigurable spatial-mode sorter for super-resolution imaging. In Conference on Lasers and Electro-Optics (CLEO), 1–2 (IEEE, 2022).
- 42.Lavery, M. P. et al. Refractive elements for the measurement of the orbital angular momentum of a single photon. Opt. express20, 2110–2115 (2012). [DOI] [PubMed] [Google Scholar]
- 43.Beijersbergen, M. W., Allen, L., Van der Veen, H. & Woerdman, J. Astigmatic laser mode converters and transfer of orbital angular momentum. Opt. Commun.96, 123–132 (1993). [Google Scholar]
- 44.Ionicioiu, R. Sorting quantum systems efficiently. Sci. Rep.6, 25356 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Zhou, Y. et al. Sorting photons by radial quantum number. Phys. Rev. Lett.119, 263602 (2017). [DOI] [PubMed] [Google Scholar]
- 46.Zhou, Y. et al. Hermite–Gaussian mode sorter. Opt. Lett.43, 5263–5266 (2018). [DOI] [PubMed] [Google Scholar]
- 47.O’Donnell, R. & Wright, J. Efficient Quantum Tomography. In Proceedings of the forty-eighth annual ACM symposium on Theory of Computing, 899–912 (2016).
- 48.Haah, J., Harrow, A. W., Ji, Z., Wu, X. & Yu, N. Sample-optimal tomography of quantum states. In Proceedings of the forty-eighth annual ACM symposium on Theory of Computing, 913–925 (2016).
- 49.Metger, T., Poremba, A., Sinha, M. & Yuen, H. Pseudorandom unitaries with non-adaptive security. arXiv preprint arXiv:2402.14803 (2024).
- 50.Ma, F. & Huang, H.-Y. How to construct random unitaries. arXiv preprint arXiv:2410.10116 (2024).
- 51.Schuster, T., Haferkamp, J. & Huang, H.-Y. Random unitaries in extremely low depth. Science389, 92 (2025). [DOI] [PubMed]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
No data sets were generated or analyzed during the current study.
All codes used in this paper have been deposited in GitHub at https://github.com/ykwang-phys/quantum-learning-imaging.



