Skip to main content
Science Advances logoLink to Science Advances
. 2026 Feb 11;12(7):eaea4144. doi: 10.1126/sciadv.aea4144

Experimental quantum state certification by actively sampling photonic entangled states

Michael Antesberger 1,, Mariana M E Schmid 1,, Huan Cao 1,2,*, Borivoje Dakić 1, Lee A Rozema 1, Philip Walther 1,2,3,*
PMCID: PMC12893295  PMID: 41671370

Abstract

Entangled quantum states are essential ingredients for many quantum technologies, but they must be validated before they are used. As a full characterization is prohibitively resource intensive, recent work has focused on developing methods to efficiently extract a few parameters of interest, in a so-called verification framework. Most existing approaches are based on preparing an ensemble of nominally identical and independently distributed (IID) quantum states and then measuring each copy of the ensemble. However, this leaves no states left for the intended quantum tasks and the IID assumptions do not always hold experimentally. To overcome these challenges, we experimentally implement quantum state certification (QSC), which measures only a subset of the ensemble, certifying the fidelity of multiple copies of the remaining states. We use active optical switches to randomly sample from sources of two-photon Bell states and three-photon GHZ (Greenberger-Horn-Zeilinger) states, reporting statistically sound fidelities in real time without destroying the entire ensemble. In addition, our QSC protocol removes the assumption that the states are identically distributed (but still assumes independent copies); can achieve close N−1 scaling, in the number of states measured N; and can be implemented in a device-independent manner. Together, these benefits make our QSC protocol suitable for benchmarking large-scale quantum computing devices and deployed quantum communication setups relying on entanglement in both standard and adversarial situations.


Quantum state certification is implemented with optimal scaling without consuming all the copies.

INTRODUCTION

Quantum technology uses entangled states as resources to implement tasks with an efficiency or security that cannot be accomplished with only classical resources (1, 2). However, before using an entangled state for a given task, the experimentally produced state must be verified. Traditionally, this is done by preparing an ensemble of N identical and independently distributed (IID) states and measuring each state. It is widely appreciated that learning the complete quantum state—i.e., performing quantum state tomography (3)—is a challenging task, requiring N to increase exponentially with the system size. This has led to a variety of resource efficient characterization methods, including adaptive state tomography (46), compressed sensing (7), direct fidelity estimation (8), cross-verification (9, 10), and quantum state verification (QSV) (1120). In all of these approaches, one measures every copy of the initially prepared ensemble, leaving no states for use in a further experiment. One must therefore assume that the source operates identically during the characterization and operation phases. Here, we circumvent this problem by experimentally implementing a quantum state certification (QSC) protocol proposed by Gočanin et al. (21).

In QSC, only a subset of the ensemble is measured, allowing one to certify some property of the remaining states. In more detail, we imagine a “verifier” who randomly extracts a subset of the initial ensemble of physical states and a “user” who receives the remaining states. The verifier performs QSV on their subensemble, allowing her to issue a certificate to the user that his remaining states are close to a promised target state. QSC thus relies on QSV. More precisely, QSC considers a set of N independent (but not necessarily identically distributed) physical states. The verifier then extracts a random subset of μN states, measuring each one and subsequently issues a certificate regarding the unmeasured states of the user. QSC then answers the question, “With what confidence can we conclude that the remaining subset of (1μ)N physical states have an average fidelity of at least 95% with the target state?”

Both QSV and QSC are useful but have different potential applications. For example, QSV could be used in initial source characterizations or where the goal is simply to determine whether a given state has been successfully distributed through a quantum network. However, QSV consumes all of the states. Therefore, QSC would have applications when it is important to ensure that the state has been distributed but there is another goal or protocol to be implemented. In particular, the remaining states can be used for these other protocols.

QSC is experimentally accessible, meaning that all measurements can be made locally and the postprocessing complexity is low. This is because it is based on QSV, which can estimate the fidelity with the optimal N−1 scaling (22, 23). Although much work on state characterization operates in a trusted, device-dependent scenario, there is a growing need to perform device-independent (DI) state characterization with untrusted measurement devices (2426), for applications such as blind quantum computing (16, 17, 27) and quantum cryptography (28, 29). Our work is based on the proposal by Gočanin et al. (21), which introduces DI-QSV and DI-QSC, allowing us to experimentally realize efficient QSC. Although, in our experiment, we do not close the loopholes necessary to claim device independence, we provide a detailed discussion of how to extend our experiment to a DI setting.

Photonic Bell states and Greenberger-Horn-Zeilinger (GHZ) states are of extreme importance for a variety of applications. In particular, two-photon Bell states are an essential primitive for many quantum communication protocols (2931), whereas three-photon GHZ states and probabilistic fusion gates can enable the efficient generation of large-scale cluster states (32) for measurement-based quantum computing (3335). Certifying such resources is essential for these and other applications. To show the applicability of our QSC protocol, we therefore experimentally implement it for both two-photon Bell states and three-photon GHZ states (36, 37).

In our implementation of QSC, some of the produced quantum states are randomly routed to the verifier who performs QSV while the user simultaneously runs his experiment. The user can either adapt to the current confidence level broadcasted by the verifier or wait until sufficient confidence is achieved. If the verifier’s measurements are successful, the states arriving to user are certified. Note that our implementation is slightly different from the proposal in (21), wherein the user does not measure their copies until the verifier issues a report. However, because we do not have a quantum memory, our user and verifier operate simultaneously, making our implementation already applicable with today’s technology. To realize this deterministically, we use active optical switches (OSs) (38), as outlined in Fig. 1, because we use active switches; each individual state is deterministically routed to the user or verifier. This provides a realistic and practical implementation of QSC because, from the user’s point of view, the only effect is a constant reduction in the counting rate, which does not scale with the system size.

Fig. 1. Experimental QSC.

Fig. 1.

(A) Sequences of M-photon states are spontaneously produced and sent to M trusted synchronized OSs. The switches periodically alternate between two settings, directing the inputs to the user or verifier. We drive the switches faster than the photons are produced. Because the emission time is random, this serves to randomly send each state to the user or verifier. Both the user and verifier implement standard polarization measurements on their photons. (B) Two-photon QSV measurement. We randomly select the measurement basis for our two-photon experiment using 50:50 beam splitters to send each of the M photons to one of two polarization measurements. (C) Three-photon QSV measurement. For our three-photon experiment, because the three-photon rate is sufficiently low, we randomly select the measurement basis by rotating the waveplates between two settings. Each subsequent setting is chosen on the basis of the outcome of a commercial QRNG. Figure 1 was created using 3ds Max by H.C.

RESULTS

Device-independent quantum state certification

Our implementation of QSC is based on self-testing (39). Although the specifics depend on the target state, the idea of self-testing is that certain quantum correlations are almost unique to some target states. For example, self-testing for a GHZ state works as follows. One attempts to violate a Mermin inequality (40) with the physical states. This can only be achieved if the physical states are close to the target state. Then, if the violation is sufficiently strong, one can place bounds on the average fidelity F¯ of the physical states. We mention that, in a DI framework, it is not possible to directly assess the fidelity. Instead, one uses the extractability Ξ, which is equivalent to the fidelity up to local isometries (39, 41, 42). However, for simplicity, in the following, we use fidelity F (or infidelity η=1F) in place of the extractability if not otherwise specified. Because self-testing is based on the mean value of observables, it usually does not discuss the behavior for finite N (39, 43). To work in this regime, QSV converts self-testing to a nonlocal game that can only be won by the target state (21).

To explain QSV, consider a source producing an independent sequence of N states S={σ1,σ2,,σN} and a target state ψ that, for now, we take to be a three-photon GHZ state. The Mermin inequality for this target state is (40)

B=o1,o2,o3(1)o1+o2+o3p(o1,o2,o30,0,1)+p(o1,o2,o30,1,0)+p(o1,o2,o31,0,0)p(o1,o2,o31,1,1)2 (1)

Here, the individual terms are probabilities of outcomes to be estimated experimentally. In more detail, p(o1,o2,o3i1,i2,i3) are the probabilities to obtain outcomes o1, o2, and o3 when qubits one, two, and three are measured with a setting defined by i1, i2, and i3. The outputs oi are assumed to be binary, taking values zero or one, which leads to the separable state bounds B2. For an ideal GHZ state, however, the quantum bound is Bq = 4. For the Mermin inequality (Eq. 1), there are four measurement settings: {i1,i2,i3}={0,0,1;0,1,0;1,0,0;1,1,1} (defined in Eq. 3 in Materials and Methods) each with eight possible outcomes. For a perfect GHZ state, only a subset of outcomes will occur with certainty: We group these together, forming the winning outcomes. To make the inequality into a nonlocal game, one then randomly measures each state of S in one of the four settings, records the number of winning events nwins, and computes the experimental winning probability as Pexp=nwins/N.

The intuition is then that, because only states equivalent to the target state can violate the inequality, only they can obtain a Pexp close to 1.

In QSC, the verifier measures Nver=μN states of the initial N state ensemble and uses her measurement results to estimate Pexp. The remaining (1μ)N states are then certified. Previous related works were formulated such that only a single copy was certified, i.e., N−1 states are measured and only one state is certified (25, 26). By doing so, one can completely remove all IID assumptions. Reference (21), instead, formulates QSC for any value of μ, allowing multiple states to be certified. This comes at the cost of still requiring the assumption that the subsequently emitted states are independent but, nevertheless, still removes the identical distribution assumption [see (21) for more details]. Given that state independence is a very reasonable assumption for parametric photon-pair sources, here we implement multicopy certification, which may be applicable for a wider class of applications.

To formalize the relation between Pexp and Ξ¯ the average extractability, consider a sequence of μN physical states with a reduced average extractability Ξ¯1η. In this case, the maximum winning probability is Pη=PQM, where c is a constant depending on the specific Bell inequality (21), and PQM is the quantum mechanical probability to win the game (which is 1 for our present example). If we then observe Pexp>Pη, the physical states likely have an extractability Ξ¯>1η. However, if Pexp is close to Pη, it is possible that finite statistical fluctuations led to an “accidental” win. Thus, we need to know how likely it is that our observation, Pexp>Pη, could be reproduced by a series of states with a fidelity less than F=1η. This is a classical statistical problem (44, 45). It can be shown that the probability for a false win after μN states are measured is bounded by δeD(PexpPη)μN, where D(xy)=xlog(x/y)+(1x)log(1x)/(1y) is the Kullback-Leibler divergence (44). On the basis of this, (21) proves that our confidence C that remaining (1μ)N states have an average fidelity greater than 1η is

C1δ=1eD(PexpPη)μN (2)

which grows to 1 exponentially fast with N. See tSupplementary Materials section G for more details.

The scaling of the confidence that our physical states have an extractability larger than 1η depends on the statistical distance between Pexp and Pη (Eq. 2). The estimated number of measured states needed to verify this lower bound is Nln δln[1μ+μeD(PexpPη)] (21). Experimentally, instead, we often wish to compute the fidelity from the results of a set of measurement results. We can also use QSC for this. It results in two different scaling behaviors related to the specifics of the Bell inequality used to make the nonlocal game. In particular, not all states have Bell inequalities that achieve a perfect success probability when translated to a nonlocal game. This happens when quantum mechanics cannot achieve the maximum algebraic bound of the Bell inequality. For example, for the CHSH inequality the quantum Tsirelson bound Bq=22 is less than the algebraic bound of 4. When this occurs, we also need to consider the ideal winning probability PQM. These scenarios arise when some winning outcomes do not occur 100% of time. We can still define the set of winning outcomes, but now they are given by the most likely outcomes. When PQM<1, the verifiable infidelity scales as ηO[ln δ1(1μ)N], i.e., with a suboptimal N1/2 scaling. However, when PQM=1, we can obtain the optimal N1 scaling ηO[ln δ1(1μ)N]. Note that these scaling behaviors are derived in the limit of large N, and, as we will see later, in our intermediate regime, they are only approximate. Nevertheless, as we will show experimentally, this means that for the two-photon Bell state, using the CHSH inequality, we expect to achieve approximate N1/2 scaling, but the three-photon GHZ state with the Mermin inequality can exceed this scaling.

Experimental apparatus for QSC

We will now present our experimental implementation of QSC [as proposed by Gočanin et al. in (21)] for bipartite and tripartite states, allowing us to demonstrate both scaling behaviors discussed above. For the two-photon case, we use a type 0 spontaneous parametric downconversion (SPDC) source in Sagnac configuration (46) to generate the Bell state Φ+=12(HH1,2+VV1,2), whereas for the tripartite case, we use a three-photon GHZ state ΨGHZ=12(HHH1,2,3+VVV1,2,3) produced by two sandwich-like SPDC sources (47). Our three-photon source is a postselected source, meaning that, to ensure that the source has generated the correct entangled state, we must postselect on three photons arriving at each measurement station (as well as detecting an additional trigger photon in the source). With respect to a DI implementation of QSC, this opens up our source to the postselection loophole (4850), which cannot be closed with our current methods. Doing so would require a source that directly generates three-photon entanglement (51) with high enough coupling efficiency or a heralded three-photon source (52, 53) with sufficiently high heralding efficiency. We stress that this postselection loophole is not related to our use of probabilistic sources, which have been used for other DI protocols (54).

For a full description of the photon sources, see Supplementary Materials sections A and B. In both scenarios, we assume that the subsequent states emitted by the source are independent, but we do not need to assume that they are identically distributed.

To randomly route the multiphoton states between the user and verifier for QSC, one could use passive beam splitters. Although this would work perfectly for a single photon state, it introduces significant loss for M-photon states. For QSC, we need all of the photons to either arrive at the user or verifier; situations where, for example, one photon arrives at the user and another at the verifier count as loss events. Because, with 50:50 beam splitters, each photon is independently routed, all M photons will arrive at either the user or verifier with probability ϵ=(12)M. In other words, the photons are not synchronously routed between the user and verifier. This is the passive synchronization efficiency plotted as the blue line in Fig. 2A. In a fully DI implementation of QSC, the detection efficiency must remain sufficiently high. This synchronization loss serves to decrease the detection efficiency, making DI-QSC impossible with passive beam splitters. Even if a DI implementation is not desired, the loss introduced by sampling with passive elements tends to 0 exponentially fast, making active elements essential for larger entangled states.

Fig. 2. Random sampling characterization.

Fig. 2.

(A) Synchronization efficiency. The blue curve plots the maximum synchronization efficiency using probabilistic routing with 50:50 beam splitters versus the number of photons in the state to be certified. The dashed line at 1 indicates the maximum efficiency attainable with the deterministic switches that we use. Our observed two-photon and three-photon efficiencies of ϵ2=0.9439±0.0041 and ϵ3=0.9997±0.0066, respectively, are plotted as stars. (B) Results of NIST’s statistical test suite for random number generators applied to our sampling of the produced state for two-photon Bell states in dark blue and three-photon GHZ states in light blue. In both cases, the results exceed the significance threshold of 0.01, indicating that our implemented sampling is consistent with truly random sampling.

To circumvent this loss, we use M synchronized OSs (see Fig. 1). Because of the different photon production rates, we use different technologies for our two-photon and three-photon experiments, explained in Materials and Methods (“Deterministic random sampling with OSs”). In any case, in both experiments, all M OSs switch between the user and verifier. Both switches are driven with a fixed 50% duty cycle to achieve μ0.5. Experimentally, the duty cycle is set electronically via the periodic function generator that drives the OSs: The duty cycle of the function generator is nominally the duty cycle of the OSs and thus directly yields the sampling rate. Thus, although we only study μ0.5, it can be easily tuned to implement other sampling rates. Note that changing the duty cycle does not significantly change the synchronization of the OSs. We show below how this can still be used to randomly sample from the states. To characterize the QSC efficiency for both experiments and show that we exceed the passive synchronization efficiency, we measure the two and three rates at the user’s (verifier’s) measurement station Ruse (Rver), as well as all two-photon and three-photon events between the two measurement stations Rcross (nominally, Rcross=0). Then, the experimental synchronization efficiency is ϵ=(Ruse+Rver)/(Ruse+Rver+Rcross). These are the data points shown in Fig. 2A, confirming that we have routed the photons with high efficiency. See Supplementary Materials section C for more details. Although the total loss in our experiment is too high to close the detection loophole, our synchronization efficiency is high enough (0.9439 ± 0.0041 and 0.9997 ± 0.0066 for our two-photon and three-photon states, respectively) that it would not significantly affect the detection loophole in an otherwise loophole-free experiment. In Materials and Methods (“Loophole analysis”), we show that also including the insertion loss of the switches used in our two-photon experiments would decrease the overall efficiency too much to remain loophole-free. Although the switches used in our three-photon experiment are already sufficient in efficiency, the relatively slow speed would make the freedom of choice loophole hard to close.

Our OSs sequentially alternate between the user and verifier, which, on its own, does not implement random sampling. To that end, we make use of SPDC’s inherent randomness (making a device-dependent assumption). In particular, we drive our OSs much faster than our sources produce photons the rate of producing photons. Thus, when a given multiphoton state is randomly emitted, the OSs will randomly be in one configuration or the other. To confirm this, we call an M-photon detection at the user (verifier) a “0” (“1”) and generate a bit string. We test these bit strings with NIST’s statistical test suite for random number generators (55). As elaborated on in Supplementary Materials section D, we obtain an overall confidence of 0.99 that our produced bit strings are random. Figure 2B shows the results of the individual tests. Although this analysis indicates that we have achieved random sampling, doing so requires us to make a device-dependent assumption that the randomness originates from the source. For a fully DI implementation, the switches should be driven randomly, as we discuss in Materials and Methods (“Loophole analysis”).

Once the photons arrive at the verifier’s measurement station, she must randomly choose a measurement setting. For our two-photon and three-photon states, we use the measurement sets MCHSH and MMermin, respectively, which are defined in Eq. 3 in the Materials and Methods section. In both cases, the measurements are implemented with standard methods (waveplates and polarizing beam splitters), but because of the different photon production rates, we randomly sample from the measurements differently (see Fig. 1, B and C).

Experimental results

With our QSC apparatus in place, we drive our OSs with 50% duty cycle pulses so that the verifier takes approximately half of the photons for her measurements. In our two-photon measurements, this results in counting rates of Ruser33 kHz (Rver25 kHz) at the user (verifier), yielding a sampling probability of μBell0.43 for our two-photon experiment. For the three-photon experiment, we have Ruser0.3 Hz (Rver0.25 Hz) at the user (verifier), yielding μGHZ0.45. Note that we have defined μ=Rver/(Ruser+Rver). As mentioned above, QSC can be performed for any value of μ. Here, we set it close to 0.5 so that the user and verifier measure approximately the same number of measurement results to ease their comparison.

Although the verifier implements QSV, the user operates his measurements independently. We implement two different characterizations for the user and check their consistency with the verifier. First, the user implements standard device-dependent characterizations. For our two-photon source, the user performs full quantum state tomography, finding a fidelity of 0.9947 ± 0.0002 with the target Bell state (see Fig. 3D). Given the lower three-photon count rate and the larger number of required measurements, for three photons, the user instead uses a GHZ witness (56, 57) to estimate a fidelity of 0.9678 ± 0.0055 (see Fig. 3E). We also perform standard DI self-testing with the user’s measurement device, finding a fidelity of 0.971 ± 0.005 and 0.9032 ± 0.0066, for our two-photon and three-photon states, respectively. See Supplementary Materials section F for more details. The discrepancy between the device-dependent and DI techniques is well known, arising because it is, in general, easier to place tighter bounds in a device-dependent scenario (58).

Fig. 3. Confidence in two- and three-photon QSC.

Fig. 3.

(A) Confidence versus the number of verifier measurements (bottom axis) and measurement time (top axis). The same dataset is analyzed for different fidelity F=1η lower bounds. Less stringent bounds (higher values of η) converge to a confidence of 1 more quickly. The inset shows the high-confidence region, with the shaded area representing a confidence above 0.99. (B) The two-photon measurement presented in (A) is repeated 12 times and averaged. The data are plotted as in (a). (C) Verifier’s confidence growth for our three-photon GHZ states, plotted as in (a). (D) Real part of the density matrix of our two-photon states measured with standard quantum state tomography, taken by the user concurrently with the verifier’s measurements. The plotted density matrix has a fidelity of 0.9947 ± 0.0002 with the target ϕ+ Bell state. (E) Results of the user’s GHZ witness measurement on our three-photon states, yielding a fidelity of 0.9678 ± 0.0055 with the target GHZ state. The labels on the x axis refer to specific measurement settings, defined in the Supplementary Materials.

Although the user performs his measurements, the verifier implements QSV on her subset. To do so, she simply records the number of winning and losing events and computes the winning probability Pexp as a function of the number of measurements. Pexp for both bipartite and tripartite scenarios is provided in Supplementary Materials section E. To use Pexp to certify the physical states, the verifier sets a minimum fidelity 1 − η and uses Eq. 2 to compute her confidence in that lower bound. In Fig. 3A, we plot this confidence versus the number of measurements made by the verifier μN for a range of infidelities from η = 0.08 to η = 0.2 for the two-photon states. Figure 3C shows the same data for our three-photon states for a range of η = 0.12 to η = 0.2. For each dataset, we observe a series of rapid increases and sharp drops. As μN increases, the drops become less pronounced and the confidence tends to 1. The periods of increasing confidence are caused by successive winning events, whereas the drops are caused by losses.

We also stress the measurement times for these datasets. In the two-photon data, in less than ≈1 s, the verifier reports a fidelity above 0.9 with confidence ≫ 0.99, demonstrating the applicability of our methods. Given our lower three-photon rates, we achieve similar values in ≈30 min for the GHZ states. Nevertheless, in both scenarios, the user and verifier can operate concurrently, with only a constant decrease in the user’s count rate.

The effect of changing μ is apparent from the above discussion. Because the user and verifier act completely independently, changing μ effectively only changes the total elapsed time that it takes for the verifier’s measurements to converge, i.e., it will only change the upper time axis in Fig. 3 (A to C). This is because the confidence scales with μN. Thus, μ may be set differently for different scenarios. For example, if it is important that the user maintains the highest possible count rate, μ can be made very small. In this case, it will take the verifier longer to converge, but if the total runtime of the experiment is sufficiently large, this may not be an issue. On the other hand, if one wishes to issue a certificate quickly, then it may be advantageous to increase μ, at a cost that the user will receive a lower rate. Because, in the present work, we are primarily interested in showing that the user and verifier’s predictions are consistent, we chose a balanced value of μ ≈ 0.5.

To smooth out the statistical fluctuations in the two-photon data, we perform 12 rounds of QSV with 35,000 samples and average the results. This results in the curves in Fig. 3B. We identify the minimal sample number required to reach 99% confidence level. The insets in Fig. 3 (A and B) present zoomed-in plots of the high-confidence region, and the blue-shaded areas denote the area above 99%. From the inset in Fig. 3B, we can estimate the number of samples need to certify infidelities below η = {0.2,0.15,0.1,0.08}, finding that {1420,3106,10019,20982} samples are needed, respectively, for our two-photon Bell states.

The main difference between the two-photon and three-photon datasets is related to the different values of PQM. For our two-photon CHSH results, PQM=2+240.85. For the Mermin inequality, however, PQM=1, leading to fewer failed events and thus smoother data in a single experimental run, obviating the need to average to estimate the required number of measurements. Overall, for the three-photon experiment, we have a success probability Pexp0.97, which leads to {433, 919, 1562, 2111} samples to certify infidelities η = {0.2, 0.15, 0.13, 0.12} with 99% confidence, respectively.

As discussed above, different values of PQM lead to different scaling behaviors for our two-photon and three-photon measurements. To observe this, we again perform QSC, now directly comparing the verifier’s claims on the infidelity to that measured by the user. For the verifier, we then instead fix the confidence level 1 − δ to 0.99 and numerically solve Eq. 2 for the infidelity η. This then represents the highest DI infidelity that our sequence of measurements is consistent with at a 99% confidence level. In other words, it establishes a 99% confidence lower bound on the fidelity of the experimentally generated state. On the basis of this, we plot η versus the number of verifier measurements μN in Fig. 4. Therein, Fig. 4A shows the two-photon data, and Fig. 4B shows the three-photon data. The bounds in both plots represent two alternative methods to estimate the fidelity discussed above. The bounds indicated by the blue lines come from device-dependent measurements [quantum state tomography for the two-photon case and GHZ witness (56, 57) for our three-photon states]. On the other hand, the bounds in gray come from DI self-testing (Supplementary Materials section F). All of these bounds are measured by the user. In both cases, when sufficient data are acquired, QSC converges to the DI self-testing bound. These data show that our QSC protocol is consistent: The infidelities the user measures (with DD or DI methods) are all lower than the infidelities reported by the verifier.

Fig. 4. QSC fidelity scaling at 99% confidence.

Fig. 4.

The estimated infidelities η when setting the confidence C = 1 − δ = 0.99 for (A) two-photon Bell states and (B) three-photon GHZ states plotted versus the verifier’s measurement number. In both cases, the infidelity η certified by the verifier (green points) asymptotically approaches infidelity estimated by the user with self-testing (gray lines). The user estimates lower infidelities when using device-dependent methods (blue lines). Insets: The insets plot averaged results over smaller sample ranges to estimate the scaling s. The dark points are the averaged data, and the green dashed lines are linear fits with slopes of sBell,Avg.=0.549 and sGHZ,Avg.=0.801, respectively. The blue solid line is theory, demonstrating scalings of sBell,Theory=0.566 and sGHZandTheory=0.935. In the inset of (B), we additionally plot the best-case scaling from one data run that achieves the scaling of sGHZ,Best=0.907.

Last, we analyze the scaling of the infidelity as a function μN. To this end, we average several repetitions of the experiments over a fixed sample range. The resulting averaged estimated infidelities are shown in the insets of Fig. 4. For the bipartite case, we take 1398 repetitions for N from 1 to 300 (points in Fig. 4A, inset), performing a linear fit in the log-log scale and obtaining a slope of −0.549 (green dashed line in the same inset). This means that the scaling is ηN0.549. This already exceeds the predicted asymptotic scaling of N0.5. Although this may seem unexpected, both the Heisenberg limit optimal N1 scaling and the suboptimal N1/2 scaling are derived asymptotically and cannot be expected to be literally the same for finite statistics. We numerically predict the expected scaling for our parameter regime by substituting Pexp=PQM0.85 into Eq. 2 and solving for η as a function of N. This is plotted as the blue solid line, which has a slope of −0.566. This agrees quite well with our experimentally measured scaling. For the tripartite case, we use 20 repetitions for N from 1 to 75 (dark points in Fig. 4B, inset), and fitting to these data gives a scaling of ηN0.801 (green dashed line Fig. 4A, inset), which significantly exceeds the suboptimal scaling of N1/2.

Notice that, in both cases (main plots in Fig. 4, A and B), the scaling behavior of the infidelity at larger μN deviates from the ideal case, i.e., it is not linear. This is the correct behavior because our experimental states are not perfect. Therefore, the infidelity should not decrease indefinitely, but it should begin to level off as sufficient data are acquired. To assess the scaling in spite of this, we select a data range from Fig. 4B for small values of μN where all the samples pass the nonlocal game {from N[10,100]} (gray points in the same inset) and fit the result (blue dashed line). Doing so yields a scaling of ηN0.907. We compare this to the numerically predicted scaling behavior for ideal GHZ states as before: We use Eq. 2 with Pexp=1 and obtain ηN0.935 (blue line). This again agrees well with our experimentally measured scaling, with a small discrepancy coming from failed events before the displayed range. Both our two-photon and three-photon data qualitatively show the expected trends.

DISCUSSION

We have presented an experimentally feasible method for QSC. To do so, we used active switches to deterministically implement a random sampling of multiphoton states from two-photon and three-photon sources. A verifier measures the sampled states, issuing a statistically rigorous certificate to the user that the remaining states have average fidelity to a target state above some threshold. From the user’s point of view, the only noticeable effect of the verifier is a slight reduction in the counting rate that does not scale with the number of photons in the entangled state. Moreover, we have demonstrated two distinct scaling behaviors related to different features of nonlocal games, i.e., the winning probability. Although we have shown this technique for Bell states and GHZ states, it can be generalized to any states for which a robust self-testing protocol exists. We also mention that, although the theory underlying our work is DI, our implementation contains loopholes and is thus not DI. However, to guide future DI implementations, Fig. 5 presents a potential extension of our experiment, which could close the loopholes needed to claim device independence. See Materials and Methods (“Loophole analysis”) for a full description of this proposal.

Fig. 5. Proposal for DI-QSC.

Fig. 5.

(A) A heralded N-photon source sends photons to N space-like separated nodes. An RNG also distributes signals to the N nodes, which are used to synchronize trusted switches in each node. In addition to these switches, each node hosts a user and verifier measurement station. The verifier measurement stations operate device independently. (B and C) Spacetime diagrams along the x and y directions, respectively. In Materials and Methods (“Loophole analysis”), a loophole analysis of this setup is presented.

To summarize, we have performed QSC (as opposed to verification), showing how to use active switches randomly sample a subset of the available states to certify the remaining states. We have demonstrated the efficiency of QSC, experimentally beating the standard limit of N0.5 with N0.801. Moreover, QSC removes the assumption of identically distributed states and can be implemented in a DI manner. Although our experiment is not DI, we have proposed how this could be achieved in future work. Last, given the relative ease of our realization, requiring only local measurements and low postprocessing complexity, it would be interesting to tailor QSC for other quantum systems, such as trapped ions (59, 60), cold atoms (61, 62), and superconducting circuits (63). We thus anticipate that our work could be used to benchmark future large-scale quantum devices.

After the completion of our experiment, we became aware of a closely related work (64).

MATERIALS AND METHODS

Deterministic random sampling with OSs

As discussed in the main text, we use OSs to redirect photons to the verifier station. For both our two-photon and three-photon experiments, we use fiber-coupled OSs. For our two-photon experiment, the switches are based on integrated electro-optic modulators (EOMs), whereas our three-photon experiments use microelectromechanical system (MEMS)–based switches. For our two-photon work, we use one Agiltron 2x2 Nanospeed switch and one BATi Inc. 2x2 Nanona fiber switch. These switches feature rise and fall times of ~100 and 50 ns, respectively. Both devices have a maximum repetition rate of 1 MHz and exhibit a cross-talk between output ports of around 20 dB. The transmission of each switch is ≈79%, resulting in a two-photon transmission of 0.792 ≈ 0.62. In the three-photon experiment, we use three identical Agiltron 1x2 fiber optical MEMS switches, each with a rise and fall time of ~2 ms, a maximum repetition rate of 5 Hz, and a cross-talk between output ports ranging from 50 to 75 dB. The transmission of each switch is ≈95%, resulting in a three-photon transmission of 0.953 ≈ 0.85. In both cases, we synchronize the OSs and set them to alternate between the user and verifier stations with a nominal 50% duty cycle.

To ensure that states randomly directed to the verifier, there are two essential conditions need to be satisfied: (i) The verifier (or user) should receive the full copy. For example, for the three-photon experiment, this means that all three photons should be sent to the same measurement station. To this end, we synchronize the OSs with an external trigger signal. This is verified in Fig. 2A and in Supplementary Materials section C. The discrepancy in Fig. 2A between the ideal switching efficiency (ϵ = 1) is due to the “synchronization efficiency,” as discussed in the main text, and this can be reduced by nonzero cross-talk between switch output ports. The EOM-based switches used in the two-photon experiment exhibit higher cross-talk than the MEMS-based switches, leading to a nonzero probability of single photons ending up in the “wrong” output port, thereby reducing ϵ. (ii) Distribution of each copy is random. Although we drive the switches with a deterministic signal, we use the intrinsic randomness from the SPDC process, wherein each photon pair is generated at a random time. Therefore, when each state is generated, it can randomly lie in a time slot that will distribute it to the verifier or user. However, this requires the M-photon generation rate to be substantially lower than the switching rate to avoid having one of more copies of the state in each time slot. Consequently, we set our switches speeds to 800-kHz and 5-Hz repetition rates for the two-photon Bell states and three-photon GHZ states, respectively. In both cases, this is significantly higher than the respective state generation rates of ≈62 and ≈0.5 Hz. Figure 2 verifies the validity of this approach.

Measurement strategy

Our DI-QSC protocol is based on converting the CHSH and Mermin inequalities into two-photon and three-photon nonlocal games, respectively. This works by considering the different observables used to violate the respective inequalities and associating each observable with a binary classical input ascribed to each local party. In particular, for the bipartite case, we have the possible inputs (i1,i2)={(0,0),(0,1),(1,0),(1,1)}, and for the tripartite case, we have (i1,i2,i3)={(0, 0, 1),(0, 1, 0),(1, 0, 0),(1, 1, 1)}. These inputs then refer to the following observables

MCHSH=X1X2+Z22,X1X2Z22,Z1X2+Z22,Z1X2Z22MMermin=Y1Y2X3,Y1X2Y3,X1Y2Y3,X1X2X3 (3)

In both cases, the result is four different multipartite measurements that must be made randomly.

To implement the random selection of measurements for our two-photon experiment, we use a 50:50 beam splitter in the path of each photon to direct that photon to physically different measurement apparatus corresponding to the binary inputs (Fig. 1B). This enables us to randomly select our measurements at very high counting rates, but it comes at the cost of opening the freedom-of-choice loophole. Nevertheless, because our measurement stations are not space-like separated, our experiment is open to the locality loophole. Thus, because our experiment is, in any case, open to loopholes, we chose the measurement bases passively for simplicity. Because our three-photon rate is significantly lower, we instead only have a measurement device per photon. The random classical inputs are then generated by a commercial quantum random number generator (QRNG), the Quantis QRNG USB from ID Quantique. These outputs of this QRNG are set to trigger motorized waveplates that change the measurement settings. To further ensure that our sampling is independent, we switch the settings every second (which is faster than the generation rate) and only analyze measurements time slots that contain no more than one three-photon event.

Loophole analysis

Standard loopholes

DI protocol implementations ensure robust protection against any external manipulation or influence on the protocol, maintaining reliable and trustworthy outcomes. In principle, even if the entire apparatus were controlled by an untrusted party, the protocol would still deliver correct results. Implementing DI-QSC requires a loophole-free violation of a Bell inequality (the CHSH or Mermin inequality in our case), and thus the common loopholes—i.e., the (1) locality, (2) freedom of choice, and (3) detection—must be closed (49). We do not achieve this in our work. Although loophole-free Bell tests have recently been conducted (6567), such demonstrations are rare due to the stringent infrastructure requirements involved. Nevertheless, in Fig. 5A, we sketch an N-partite extension of our experiment that could address these three loopholes, as well additional ones related to multipartite QSC. To close the locality loophole (1), a source at a central location generates the N entangled photons, which are distributed to N space-like separated nodes. The measurement basis must then be set after the state are generated, which can, in principle, be easily enforced (Fig. 5, B and C). To address the freedom-of-choice loophole (2), the measurement basis must be randomly set. In our three-photon experiment, we satisfy this by implementing our basis choice using a QRNG. In our two-photon experiment, the basis is set randomly using 50:50 beam splitters. Although this is a random process, it opens the loophole because the photons could carry hidden variables determining whether they are transmitted or reflected to one measurement setting or the other. Last, to ensure that the detection loophole (3) is closed, the overall transmission between the source and the detectors (including the detection efficiency) must remain over a certain threshold. This is 66.7 and 75% for the two-photon and three-photon experiments, respectively (68).

Preparation postselection loophole

Although the above loopholes apply to all Bell tests, current methods of creating multipartite entangled states open up an additional loophole, which we refer to as (4) the preparation postselection loophole. This loophole is one of the primary challenges for a loophole-free multipartite demonstration of quantum nonlocality, which has not yet been realized (50). Closing this loophole is required for DI-QSC of multiphoton states. The preparation postselection loophole comes from the fact that we use nonheralded gates to build three-photon entanglement from two-photon entangled states. We can only be sure that we have successfully generated the entangled state when one photon is detected in each output port [see (50) for a discussion of the implications]. To close the postselection loophole, one could use heralded sources of multiphoton entanglement, as in (53), which can be achieved with either deterministic sources or nondeterministic SPDC as we use here. Unfortunately, the heralding efficiency in such experiments is still relatively low and thus opens the detection loophole.

Loopholes specific to DI-QSC

As above, we have discussed loopholes only related to loophole-free violations of Bell inequalities, which must also be closed for DI-QSC. However, DI-QSC brings additional loopholes. Here, we identify and discuss two additional loopholes: (5) the switching-loss loophole and (6) the random sampling loophole. Note that we also assume that the switches belong to the measurement nodes and are thus trusted devices. We will now discuss how our proposal in Fig. 5 addresses these loopholes.

We already alluded to the switching-loss loophole in the main text. The intuition is relatively straightforward: Upon insertion of the switches used to implement the random sampling, the transmission of the N-photon state from the source to the verifiers should remain above the threshold needed to close the detection efficiency loophole. There are two contributions that reduce this transmission. The first is simply the transmission loss of the switches. In our case, for the two-photon experiment the transmission is ≈0.79 per switch, whereas for the three-photon experiment, it is ≈0.95 per switch. This results in an overall two-photon transmission of 0.792 = 0.62 and a three-photon transmission of 0.953 = 0.85. Thus, although our two-photon switches are too lossy, our three-photon switches could, from a loss perspective alone, meet the requirements to close the loophole.

The other source of loss is the synchronization efficiency, already discussed in the main text and in Materials and Methods (“Deterministic random sampling with optical switches”). For completeness, we discuss it here in the context of loopholes. When an N-photon state is selected to be sent to the verifier, all N photons must simultaneous be routed. If some photons are sent to the user and others in the verifier, these events introduce loss. If this loss reduces the source-to-verifier transmission too much, the detection loophole is opened. This is the reason that we use synchronized active switches. If one uses “passive switches” (i.e., beam splitters), this loss scales to zero exponentially fast with the number of photons in the entangled state. We call this the passive synchronization efficiency and plot it for the case of 50:50 beam splitters in Fig. 2A. Using synchronized active switches, on the other hand, can, in principle, route the photons without introducing any additional loss. Experimentally, we achieve synchronization efficiencies of 0.9439 ± 0.0041% and 0.9997 ± 0.0066% for our two-photon and three-photon experiments, respectively, which could be sufficiently high to remain above the detection efficiency threshold if other experimental losses were low enough.

The final loophole we discuss is the random sampling loophole (6). If the sampling process is not genuinely random, the validity of the protocol is compromised. The challenge in enforcing this experimentally is that the switches, which are located in space-like separated nodes, must be synchronized and the decision of whether to route the N-photon state to the verifier or to the user must be made after the source produces a state, just as is done in choosing the basis to close the standard freedom-of-choice loophole. Nevertheless, this can be enforced with space-like separation. The switch rise time must also be considered. In our experiment, the three-photon switches exhibit a rise time of 2 ms, corresponding to a fiber delay of several hundred kilometers, which introduces substantial loss. Therefore, OSs suitable for a fully DI implementation must combine low loss with sufficiently high switching speed (69, 70).

A loophole-free proposal

To show how our proposed setup can achieve close these six loopholes, we will now analyze each component shown in Fig. 5A, moving from left to right. Corresponding spacetime diagrams for the x and y directions are presented in Fig. 5 (B and C), respectively. We first assume an ideal photon source capable of generating a heralded multiphoton state, i.e., one that closes loophole 4.

The next element in Fig. 5 is a random number generator, which produces a random binary output (“0” or “1”) to determine whether the OSs will direct the state to the verifier or the user. It is essential that the random number generator operates in a manner such that it selects the user or verifier before the photon source can signal to it (closing the loophole 6), and it can signal to the switches before each of the N photons can arrive at their respective node (so the switches can be synchronized and close loophole 5). Note that, in our experiment, we use randomness in the photon generation time to implement the random sampling, which, although random, does not close this loophole. Achieving this requires aligning the space-like separation with the speed and timing of the electronics involved. It may also be possible to close this loophole by making the appropriate shielding assumptions as done in recent DI experiments (71, 72).

Once the switches are configured, two scenarios arise. If the state is routed to the user, no loopholes need to be addressed (unless the user also wishes to operate device independently, but we will not focus on that here). Conversely, if the states are directed to the verifier, device independence is necessary, i.e., loopholes 1 to 3 must be closed. Doing so requires the N verifiers to select their measurements randomly independently and without correlation to the photon source. To ensure freedom of choice, the measurement settings must be determined before any signal from the source can influence a verifier. In photonics, this is typically achieved using an RNG in combination with a fast Pockels cell.

In addition, all N verifiers must be space-like separated from each other to prevent any causal influence in their measurements. This requirement is represented in the spacetime diagram for the y direction in Fig. 5C. Last, addressing the detection loophole necessitates achieving a system detection efficiency above a specific threshold. To close this loophole, the setup must use low-loss optical components, high-efficiency detectors, and sufficiently high coupling efficiencies between free space and fiber. Overall, implementing DI-QSC is a challenging goal, but no more challenging than implementing most other multipartite DI protocols.

Acknowledgments

We are grateful to W. Zhang for insightful discussions and comments on the manuscript and A. Trenti and M. Achleitner for randomness discussions.

Funding:

This research was funded in whole, or in part, by the Austrian Science Fund (FWF) through 10.55776/COE1 (Quantum Science Austria), through 10.55776/F71 (BeyondC) and 10.55776/FG5 (Research Group 5) and by the European Union (ERC, GRAVITES, no. 101071779) and its Horizon 2020 and Horizon Europe Research and Innovation Programme under grant agreement no. 899368 (EPIQUS) and no. 101135288 (EPIQUE) and the Marie Skłodowska-Curie grant agreement no. 956071 (AppQInfo). Views and opinions expressed are, however, those of the author(s) only and do not necessarily reflect those of the European Union or the European Research Council Executive Agency. Neither the European Union nor the granting authority can be held responsible for them. Further funding was received from the Air Force Office of Scientific Research under award numbers FA9550-21-1-0355 (QTRUST) and FA8655-23-1-7063 (TIQI); the financial support by the Austrian Federal Ministry of Labour and Economy, the National Foundation for Research, Technology and Development, and the Christian Doppler Research Association is gratefully acknowledged. L.A.R. acknowledges support from the Erwin Schrödinger Center for Quantum Science and Technology (ESQ Discovery). For the purpose of open access, the authors have applied a CC BY public copyright license to any Author Accepted Manuscript version arising from this submission.

Author contributions:

L.A.R., B.D., and P.W. initiated the project. H.C., L.A.R., and M.A. conceived the experimental design and measurement strategy with theoretical support from B.D. and discussions with P.W.. M.A., M.M.E.S., and H.C. carried out the experiment. M.A., M.M.E.S., and H.C. analyzed the data and prepared the figures. M.A., M.M.E.S., H.C., and L.A.R. wrote the manuscript with input from all authors. H.C., L.A.R., and P.W. supervised the project.

Competing interests:

The authors declare no competing interests.

Data and materials availability:

All data and code needed to evaluate and reproduce the results in the paper are present in the paper and/or the Supplementary Materials and are available online at https://doi.org/10.5281/zenodo.12759267. This study did not generate new materials.

Supplementary Materials

This PDF file includes:

Supplementary Text

Figs. S1 to S5

References

sciadv.aea4144_sm.pdf (1.6MB, pdf)

REFERENCES

  • 1.Gisin N., Ribordy G., Tittel W., Zbinden H., Quantum cryptography. Rev. Mod. Phys. 74, 145–195 (2002). [Google Scholar]
  • 2.Deutsch D., Ekert A., Quantum computation. Phys. World 11, 47–52 (1998). [Google Scholar]
  • 3.James D. F. V., Kwiat P. G., Munro W. J., White A. G., Measurement of qubits. Phys. Rev. A 64, 052312 (2001). [Google Scholar]
  • 4.Mahler D. H., Rozema L. A., Darabi A., Ferrie C., Blume-Kohout R., Steinberg A. M., Adaptive quantum state tomography improves accuracy quadratically. Phys. Rev. Lett. 111, 183601 (2013). [DOI] [PubMed] [Google Scholar]
  • 5.Chapman R. J., Ferrie C., Peruzzo A., Experimental demonstration of self-guided quantum tomography. Phys. Rev. Lett. 117, 040402 (2016). [DOI] [PubMed] [Google Scholar]
  • 6.Qi B., Hou Z., Wang Y., Dong D., Zhong H.-S., Li L., Xiang G.-Y., Wiseman H. M., Li C.-F., Guo G.-C., Adaptive quantum state tomography via linear regression estimation: Theory and two-qubit experiment. NPJ Quant. Inf. 3, 19 (2017). [Google Scholar]
  • 7.Gross D., Liu Y.-K., Flammia S. T., Becker S., Eisert J., Quantum state tomography via compressed sensing. Phys. Rev. Lett. 105, 150401 (2010). [DOI] [PubMed] [Google Scholar]
  • 8.Flammia S. T., Liu Y.-K., Direct fidelity estimation from few Pauli measurements. Phys. Rev. Lett. 106, 230501 (2011). [DOI] [PubMed] [Google Scholar]
  • 9.Greganti C., Demarie T. F., Ringbauer M., Jones J. A., Saggio V., Calafell I. A., Rozema L. A., Erhard A., Meth M., Postler L., Stricker R., Schindler P., Blatt R., Monz T., Walther P., Fitzsimons J. F., Cross-verification of independent quantum devices. Phys. Rev. X 11, 031049 (2021). [Google Scholar]
  • 10.Zhu D., Cian Z. P., Noel C., Risinger A., Biswas D., Egan L., Zhu Y., Green A. M., Alderete C. H., Nguyen N. H., Wang Q., Maksymov A., Nam Y., Cetina M., Linke N. M., Hafezi M., Monroe C., Cross-platform comparison of arbitrary quantum states. Nat. Commun. 13, 6620 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Hayashi M., Group theoretical study of LOCC-detection of maximally entangled states using hypothesis testing. N. J. Phys. 11, 043028 (2009). [Google Scholar]
  • 12.Takeuchi Y., Morimae T., Verification of many-qubit states. Phys. Rev. X 8, 021060 (2018). [Google Scholar]
  • 13.Pallister S., Linden N., Montanaro A., Optimal verification of entangled states with local measurements. Phys. Rev. Lett. 120, 170502 (2018). [DOI] [PubMed] [Google Scholar]
  • 14.Zhu H., Hayashi M., Optimal verification and fidelity estimation of maximally entangled states. Phys. Rev. A 99, 052346 (2019). [Google Scholar]
  • 15.Li Z., Han Y.-G., Zhu H., Optimal verification of Greenberger-Horne-Zeilinger states. Phys. Rev. Appl. 13, 054002 (2020). [Google Scholar]
  • 16.Hayashi M., Morimae T., Verifiable measurement-only blind quantum computing with stabilizer testing. Phys. Rev. Lett. 115, 220502 (2015). [DOI] [PubMed] [Google Scholar]
  • 17.Fujii K., Hayashi M., Verifiable fault tolerance in measurement-based quantum computation. Phys. Rev. A 96, 030301 (2017). [Google Scholar]
  • 18.Hayashi M., Hajdušek M., Self-guaranteed measurement-based quantum computation. Phys. Rev. A 97, 052308 (2018). [Google Scholar]
  • 19.Dimić A., Dakić B., Single-copy entanglement detection. NPJ Quant. Inf. 4, 11 (2018). [Google Scholar]
  • 20.Saggio V., Dimić A., Greganti C., Rozema L. A., Walther P., Dakić B., Experimental few-copy multipartite entanglement detection. Nat. Phys. 15, 935–940 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Gočanin A., Šupić I., Dakić B., Sample-efficient device-independent quantum state verification and certification. PRX Quant. 3, 010317 (2022). [Google Scholar]
  • 22.Zhang W.-H., Zhang C., Chen Z., Peng X.-X., Xu X.-Y., Yin P., Yu S., Ye X.-J., Han Y.-J., Xu J.-S., Chen G., Li C. F., Guo G. C., Experimental optimal verification of entangled states using local measurements. Phys. Rev. Lett. 125, 030506 (2020). [DOI] [PubMed] [Google Scholar]
  • 23.Eisert J., Hangleiter D., Walk N., Roth I., Markham D., Parekh R., Chabaud U., Kashefi E., Quantum certification and benchmarking. Nat. Rev. Phys. 2, 382–390 (2020). [Google Scholar]
  • 24.Han Y.-G., Li Z., Wang Y., Zhu H., Optimal verification of the Bell state and Greenberger–Horne–Zeilinger states in untrusted quantum networks. NPJ Quant. Inf. 7, 164 (2021). [Google Scholar]
  • 25.Zhu H., Hayashi M., Efficient verification of pure quantum states in the adversarial scenario. Phys. Rev. Lett. 123, 260504 (2019). [DOI] [PubMed] [Google Scholar]
  • 26.Zhu H., Hayashi M., General framework for verifying pure quantum states in the adversarial scenario. Phys. Rev. A 100, 062335 (2019). [DOI] [PubMed] [Google Scholar]
  • 27.Barz S., Kashefi E., Broadbent A., Fitzsimons J. F., Zeilinger A., Walther P., Demonstration of blind quantum computing. Science 335, 303–308 (2012). [DOI] [PubMed] [Google Scholar]
  • 28.Zhang W., van Leent T., Redeker K., Garthoff R., Schwonnek R., Fertig F., Eppelt S., Rosenfeld W., Scarani V., Lim C. C.-W., Weinfurter H., A device-independent quantum key distribution system for distant users. Nature 607, 687–691 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Yin J., Li Y.-H., Liao S.-K., Yang M., Cao Y., Zhang L., Ren J.-G., Cai W.-Q., Liu W.-Y., Li S.-L., Shu R., Huang Y. M., Deng L., Li L., Zhang Q., Liu N. L., Chen Y. A., Lu C. Y., Wang X. B., Xu F., Wang J. Y., Peng C. Z., Ekert A. K., Pan J. W., Entanglement-based secure quantum cryptography over 1,120 kilometres. Nature 582, 501–505 (2020). [DOI] [PubMed] [Google Scholar]
  • 30.Duan L.-M., Lukin M. D., Cirac J. I., Zoller P., Long-distance quantum communication with atomic ensembles and linear optics. Nature 414, 413–418 (2001). [DOI] [PubMed] [Google Scholar]
  • 31.Nadlinger D. P., Drmota P., Nichol B. C., Araneda G., Main D., Srinivas R., Lucas D. M., Ballance C. J., Ivanov K., Tan E. Y.-Z., Sekatski P., Urbanke R. L., Renner R., Sangouard N., Bancal J. D., Experimental quantum key distribution certified by Bell’s theorem. Nature 607, 682–686 (2022). [DOI] [PubMed] [Google Scholar]
  • 32.Gimeno-Segovia M., Shadbolt P., Browne D. E., Rudolph T., From three-photon Greenberger-Horne-Zeilinger states to ballistic universal quantum computation. Phys. Rev. Lett. 115, 020502 (2015). [DOI] [PubMed] [Google Scholar]
  • 33.Walther P., Resch K. J., Rudolph T., Schenck E., Weinfurter H., Vedral V., Aspelmeyer M., Zeilinger A., Experimental one-way quantum computing. Nature 434, 169–176 (2005). [DOI] [PubMed] [Google Scholar]
  • 34.Briegel H. J., Browne D. E., Dür W., Raussendorf R., Van den Nest M., Measurement-based quantum computation. Nat. Phys. 5, 19–26 (2009). [Google Scholar]
  • 35.Raussendorf R., Briegel H. J., A one-way quantum computer. Phys. Rev. Lett. 86, 5188–5191 (2001). [DOI] [PubMed] [Google Scholar]
  • 36.Pan J.-W., Chen Z.-B., Lu C.-Y., Weinfurter H., Zeilinger A., Żukowski M. Z., Multiphoton entanglement and interferometry. Rev. Mod. Phys. 84, 777–838 (2012). [Google Scholar]
  • 37.Pan J.-W., Bouwmeester D., Daniell M., Weinfurter H., Zeilinger A., Experimental test of quantum nonlocality in three-photon Greenberger-Horne-Zeilinger entanglement. Nature 403, 515–519 (2000). [DOI] [PubMed] [Google Scholar]
  • 38.Zanin G. L., Jacquet M. J., Spagnolo M., Schiansky P., Calafell I. A., Rozema L. A., Walther P., Fiber-compatible photonic feed-forward with 99% fidelity. Opt. Express 29, 3425–3437 (2021). [DOI] [PubMed] [Google Scholar]
  • 39.D. Mayers, A. Yao, Self testing quantum apparatus. arXiv:0307205 [quant-ph] (2004).
  • 40.Mermin N. D., Extreme quantum entanglement in a superposition of macroscopically distinct states. Phys. Rev. Lett. 65, 1838–1840 (1990). [DOI] [PubMed] [Google Scholar]
  • 41.Kaniewski J., Analytic and nearly optimal self-testing bounds for the Clauser-Horne-Shimony-Holt and Mermin inequalities. Phys. Rev. Lett. 117, 070402 (2016). [DOI] [PubMed] [Google Scholar]
  • 42.Zhang W.-H., Chen G., Peng X.-X., Ye X.-J., Yin P., Xiao Y., Hou Z.-B., Cheng Z.-D., Wu Y.-C., Xu J.-S., Li C. F., Guo G. C., Experimentally robust self-testing for bipartite and tripartite entangled states. Phys. Rev. Lett. 121, 240402 (2018). [DOI] [PubMed] [Google Scholar]
  • 43.Šupić I., Bowles J., Self-testing of quantum systems: A review. Quantum 4, 337 (2020). [Google Scholar]
  • 44.Chernoff H., A measure of asymptotic efficiency for tests of a hypothesis based on the sum of observations. Ann. Math. Statist. 23, 493–507 (1952). [Google Scholar]
  • 45.W. Hoeffding, The Collected Works of Wassily Hoeffding (Springer Science & Business Media, 2012). [Google Scholar]
  • 46.Neumann S. P., Selimovic M., Bohmann M., Ursin R., Experimental entanglement generation for quantum key distribution beyond 1 Gbit/s. Quantum 6, 822 (2022). [Google Scholar]
  • 47.Zhang C., Huang Y.-F., Wang Z., Liu B.-H., Li C.-F., Guo G.-C., Experimental Greenberger-Horne-Zeilinger-type six-photon quantum nonlocality. Phys. Rev. Lett. 115, 260402 (2015). [DOI] [PubMed] [Google Scholar]
  • 48.Aerts S., Kwiat P., Larsson J.-A., Zukowski M., Two-photon franson-type experiments and local realism. Phys. Rev. Lett. 83, 2872–2875 (1999). [Google Scholar]
  • 49.Larsson J.-Å., Loopholes in Bell inequality tests of local realism. J. Phys. A. Math. Theor. 47, 424003 (2014). [Google Scholar]
  • 50.Huang L., Gu X.-M., Jiang Y.-F., Wu D., Bai B., Chen M.-C., Sun Q.-C., Zhang J., Yu S., Zhang Q., Lu C. Y., Pan J. W., Experimental demonstration of genuine tripartite nonlocality under strict locality conditions. Phys. Rev. Lett. 129, 060401 (2022). [DOI] [PubMed] [Google Scholar]
  • 51.Hamel D. R., Shalm L. K., Hübel H., Miller A. J., Marsili F., Verma V. B., Mirin R. P., Nam S. W., Resch K. J., Jennewein T., Direct generation of three-photon polarization entanglement. Nat. Photonics 8, 801–807 (2014). [Google Scholar]
  • 52.Walther P., Aspelmeyer M., Zeilinger A., Heralded generation of multiphoton entanglement. Phys. Rev. A 75, 012313 (2007). [Google Scholar]
  • 53.Cao H., Hansen L. M., Giorgino F., Carosini L., Zahálka P., Zilk F., Loredo J. C., Walther P., Photonic source of heralded Greenberger-Horne-Zeilinger states. Phys. Rev. Lett. 132, 130604 (2024). [DOI] [PubMed] [Google Scholar]
  • 54.Shalm L. K., Zhang Y., Bienfang J. C., Schlager C., Stevens M. J., Mazurek M. D., Abellán C., Amaya W., Mitchell M. W., Alhejji M. A., Fu H., Ornstein J., Mirin R. P., Nam S. W., Knill E., Device-independent randomness expansion with entangled photons. Nat. Phys. 17, 452–456 (2021). [Google Scholar]
  • 55.L. E. Bassham, A. L. Rukhin, J. Soto, J. R. Nechvatal, M. E. Smid, S. D. Leigh, M. Levenson, M. Vangel, N. A. Heckert, D. L. Banks, A statistical test suite for random and pseudorandom number generators for cryptographic applications (National Institute of Standards and Technology, 2010). [Google Scholar]
  • 56.Huang Y.-F., Liu B.-H., Peng L., Li Y.-H., Li L., Li C.-F., Guo G.-C., Experimental generation of an eight-photon Greenberger–Horne–Zeilinger state. Nat. Commun. 2, 546 (2011). [DOI] [PubMed] [Google Scholar]
  • 57.Yao X.-C., Wang T.-X., Xu P., Lu H., Pan G.-S., Bao X.-H., Peng C.-Z., Lu C.-Y., Chen Y.-A., Pan J.-W., Observation of eight-photon entanglement. Nat. Photonics 6, 225–228 (2012). [Google Scholar]
  • 58.J.-D. Bancal, On the Device-Independent Approach to Quantum Physics: Advances in Quantum Nonlocality and Multipartite Entanglement Detection, Springer Theses (Springer International Publishing, 2014), pp. 73–80. [Google Scholar]
  • 59.Joshi M. K., Kokail C., van Bijnen R., Kranzl F., Zache T. V., Blatt R., Roos C. F., Zoller P., Exploring large-scale entanglement in quantum simulation. Nature 624, 539–544 (2023). [DOI] [PubMed] [Google Scholar]
  • 60.Zhang J., Pagano G., Hess P. W., Kyprianidis A., Becker P., Kaplan H., Gorshkov A. V., Gong Z.-X., Monroe C., Observation of a many-body dynamical phase transition with a 53-qubit quantum simulator. Nature 551, 601–604 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Shaw A. L., Chen Z., Choi J., Mark D. K., Scholl P., Finkelstein R., Elben A., Choi S., Endres M., Benchmarking highly entangled states on a 60-atom analogue quantum simulator. Nature 628, 71–77 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Graham T. M., Song Y., Scott J., Poole C., Phuttitarn L., Jooya K., Eichler P., Jiang X., Marra A., Grinkemeyer B., Kwon M., Ebert M., Cherek J., Lichtman M. T., Gillette M., Gilbert J., Bowman D., Ballance T., Campbell C., Dahl E. D., Crawford O., Blunt N. S., Rogers B., Noel T., Saffman M., Multi-qubit entanglement and algorithms on a neutral-atom quantum computer. Nature 604, 457–462 (2022). [DOI] [PubMed] [Google Scholar]
  • 63.Cao S., Wu B., Chen F., Gong M., Wu Y., Ye Y., Zha C., Qian H., Ying C., Guo S., Zhu Q., Huang H. L., Zhao Y., Li S., Wang S., Yu J., Fan D., Wu D., Su H., Deng H., Rong H., Li Y., Zhang K., Chung T. H., Liang F., Lin J., Xu Y., Sun L., Guo C., Li N., Huo Y. H., Peng C. Z., Lu C. Y., Yuan X., Zhu X., Pan J. W., Generation of genuine entanglement up to 51 superconducting qubits. Nature 619, 738–742 (2023). [DOI] [PubMed] [Google Scholar]
  • 64.L. D. S. Martins, N. Laurent-Puig, I. Šupić, D. Markham, E. Diamanti, Experimental sample-efficient and device-independent GHZ state certification. arXiv:2407.13529 [quant-ph] (2024).
  • 65.Hensen B., Bernien H., Dréau A. E., Reiserer A., Kalb N., Blok M. S., Ruitenberg J., Vermeulen R. F. L., Schouten R. N., Abellán C., Amaya W., Pruneri V., Mitchell M. W., Markham M., Twitchen D. J., Elkouss D., Wehner S., Taminiau T. H., Hanson R., Loophole-free Bell inequality violation using electron spins separated by 1.3 kilometres. Nature 526, 682–686 (2015). [DOI] [PubMed] [Google Scholar]
  • 66.Shalm L. K., Meyer-Scott E., Christensen B. G., Bierhorst P., Wayne M. A., Stevens M. J., Gerrits T., Glancy S., Hamel D. R., Allman M. S., Coakley K. J., Dyer S. D., Hodge C., Lita A. E., Verma V. B., Lambrocco C., Tortorici E., Migdall A. L., Zhang Y., Kumor D. R., Farr W. H., Marsili F., Shaw M. D., Stern J. A., Abellán C., Amaya W., Pruneri V., Jennewein T., Mitchell M. W., Kwiat P. G., Bienfang J. C., Mirin R. P., Knill E., Nam S. W., Strong loophole-free test of local realism. Phys. Rev. Lett. 115, 250402 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Giustina M., Versteegh M. A. M., Wengerowsky S., Handsteiner J., Hochrainer A., Phelan K., Steinlechner F., Kofler J., Larsson J.-Å., Abellán C., Amaya W., Pruneri V., Mitchell M. W., Beyer J., Gerrits T., Lita A. E., Shalm L. K., Nam S. W., Scheidl T., Ursin R., Wittmann B., Zeilinger A., Significant-loophole-free test of Bell’s theorem with entangled photons. Phys. Rev. Lett. 115, 250401 (2015). [DOI] [PubMed] [Google Scholar]
  • 68.Cabello A., Rodríguez D., Villanueva I., Necessary and sufficient detection efficiency for the mermin inequalities. Phys. Rev. Lett. 101, 120402 (2008). [DOI] [PubMed] [Google Scholar]
  • 69.Ma X.-s., Zotter S., Tetik N., Qarry A., Jennewein T., Zeilinger A., A high-speed tunable beam splitter for feed-forward photonic quantum information processing. Opt. Express 19, 22723 (2011). [DOI] [PubMed] [Google Scholar]
  • 70.P. Wang, S. Baek, M. Yabuno, S. Miki, H. Terai, F. Kaneda, Low-loss polarization-maintaining router for single and entangled photons at a telecom wavelength. arXiv:2502.12554 [quant-ph] (2025).
  • 71.Liu W.-Z., Li M.-H., Ragy S., Zhao S.-R., Bai B., Liu Y., Brown P. J., Zhang J., Colbeck R., Fan J., Zhang Q., Pan J. W., Device-independent randomness expansion against quantum side information. Nat. Phys. 17, 448–451 (2021). [Google Scholar]
  • 72.Liu W.-Z., Zhang Y.-Z., Zhen Y.-Z., Li M.-H., Liu Y., Fan J., Xu F., Zhang Q., Pan J.-W., Toward a photonic demonstration of device-independent quantum key distribution. Phys. Rev. Lett. 129, 050502 (2022). [DOI] [PubMed] [Google Scholar]
  • 73.Peres Y., Iterating von Neumann’s procedure for extracting random bits. Ann. Stat. 20, 590 (1992). [Google Scholar]
  • 74.Gühne O., Lu C.-Y., Gao W.-B., Pan J.-W., Toolbox for entanglement detection and fidelity estimation. Phys. Rev. A 76, 030305 (2007). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Text

Figs. S1 to S5

References

sciadv.aea4144_sm.pdf (1.6MB, pdf)

Data Availability Statement

All data and code needed to evaluate and reproduce the results in the paper are present in the paper and/or the Supplementary Materials and are available online at https://doi.org/10.5281/zenodo.12759267. This study did not generate new materials.


Articles from Science Advances are provided here courtesy of American Association for the Advancement of Science

RESOURCES