Skip to main content
Nature Portfolio logoLink to Nature Portfolio
. 2024 Nov 26;10(1):124. doi: 10.1038/s41534-024-00918-6

Can quantum computers do nothing?

Alexander Nico-Katz 1,2,, Nathan Keenan 1,2,3,, John Goold 1,2,4,
PMCID: PMC11599043  PMID: 39611079

Abstract

Quantum computing platforms are subject to contradictory engineering requirements: qubits must be protected from mutual interactions when idling (‘doing nothing’), and strongly interacting when in operation. If idling qubits are not sufficiently protected, information ‘leaks’ into neighbouring qubits, becoming ultimately inaccessible. Candidate solutions to this dilemma include many-body localization, dynamical decoupling, and active error correction. However, no protocol exists to quantify this effect in a similar way to e.g. SPAM errors. We develop a scalable, device non-specific, protocol for quantifying idle information loss by exploiting tools from quantum information theory. We implement this protocol in over 3500 experiments carried out across 4 months (Dec 2023–Mar 2024) on IBM’s entire Falcon 5.11 processor series. After accounting for other error sources, we detect information loss to high degrees of statistical significance. This work thus provides a firm quantitative foundation from which the protection-operation dilemma can be investigated and ultimately resolved.

Subject terms: Quantum information, Superconducting devices, Information theory and computation, Qubits, Quantum simulation

Introduction

Quantum computing is a new paradigm of computation based on the exploitation of quantum phenomena which promises frontier impacts on global energy, health, materials science, and technological innovation1,2. With the firm advent of the noisy intermediate-scale quantum (NISQ) era in the form of accessible quantum processors containing some tens to hundreds of qubits36, out-of-the-box error mitigation7, and the nascent implementation of simple error correction810, this new paradigm is now being mapped out comprehensively. However, far from their conceptualization as ideal systems, real quantum computers are intrinsically programmable many-body quantum systems with complicated internal dynamics exposed to stray interactions and thermal fluctuations (Fig. 1a, b schematically show idealized and realistic ‘empty’ circuits respectively)11. The precise control of these complex many-body physical systems is ultimately the fundamental problem of quantum computing.

Fig. 1. Schematic quantum circuits and devices.

Fig. 1

a Schematic of an ideal quantum circuit that realizes perfect idling. b Schematic of a physical quantum circuit with imperfect idling due to native transport (black arrows), interactions (dotted lines), and precession (oscillating lines). c Schematic showing the qubit geometry of the Falcon 5.11 series of IBM's quantum devices. Example target qubits are shown by yellow diamonds. Shaded regions respectively indicate examples of a contiguous set of nearest-neighbour qubits (blue) and a random set of qubits excluding nearest-neighbours (orange). Measuring the additional information gained about the target qubit when addressing the complementary qubits allows us to quantify idle information loss due to the internal dynamics of the hardware.

This problem is encapsulated by the ‘protection-operation dilemma’ wherein qubits need to be coupled and decoupled from each other at different times during the same computation12. When not in active operation (‘doing nothing’), idling qubits should be well protected from mutual interactions with other qubits in the device. However, multi-qubit operations require strong mutual interactions such that entanglement and correlations can be generated between the active qubits quickly. Both of these processes should also be resistant to environmental decoherence, which typically reduces the requirement that computations should be completed as quickly as possible. This generates contradicting engineering requirements: qubits should be decoupled when idling, but strongly coupled when necessary for computation. The two main ways this contradiction is resolved are (i) many-body localizing the qubits through spatial disorder and only bringing neighbouring qubits into resonance during gate operations (ii) rapidly tuning the mutual interactions in-situ to actively couple and decouple qubits during runtime1214. The potential breakdown of the former resolution has been addressed in detail by recent works which find that current-generation superconducting quantum computers may enter chaotic regimes during operation; calling into question how well-protected qubits in these devices actually are1517. The latter poses a complicated engineering problem that may induce higher-order or time-dependent effects that are not yet fully understood. Idle information leakage can also be reduced by active error correction or dynamical decoupling, but these introduce formidable engineering problems and additional gate-based errors and complexity respectively18,19.

Inspired by the burgeoning many-body perspective on quantum computation, we exploit tools from many-body physics and quantum information theory to address a critical question underlying the entire protection-operation dilemma: can the information loss induced by multi-qubit dynamics during idling be quantified in a similar way to e.g. readout errors or dephasing times? In this work, we address this question directly by presenting and experimentally implementing a protocol for quantifying idle information loss. This protocol, which is developed and discussed in Results, leverages the Holevo quantity to accurately quantify information loss. It is a device non-specific, scalable, and can be easily run during computational downtime. We experimentally implement the protocol in Experimental implementation across the entire range of Falcon 5.11 series of IBM’s quantum devices (IBM has recently discontinued this particular series, but the protocol is not device-specific.). After accounting for other sources of error, we identify a small but measurable amount of information leakage to a high degree of statistical significance. Our accurate quantification of the actual informational impact of many-body effects in real devices represents a decisive step toward identifying and measuring idle information loss in future quantum devices. This result also provides a firm quantitative foundation from which the protection-operation dilemma can be further interrogated and resolved.

Results

Information lost whilst idling due to the native dynamics of a device manifests as the spreading of information that is initially localized. Information initially localized to a single qubit moves coherently into the rest of the system and is distributed non-locally. Thus the basic premise of our protocol is to monitor both a target qubit from which information may leak out, and a complementary set of qubits into which information may have flowed. This is shown schematically in Fig. 1c for two types of complementary qubit set: nearest-neigbour (blue) and random (orange). The basic mathematical object we use to characterize the amount of information lost is the Holevo quantity20: an import from quantum information theory that quantifies the amount of classical information that a channel and an ensemble of messages (an alphabet) can carry21. We discuss the Holevo quantity in detail in The Holevo quantity for the interested reader.

The protocol in an ideal system, with no other sources of error, is discussed in detail in Protocol for quantifying idle information loss. The protocol involves initializing a target qubit in one of two message states 0 and 1 which encode a classical bit. By calculating the Holevo quantity both over the target qubit alone χT, and over a complementary set Q of qubits which includes the target qubit χTQ, we can quantify how much extra information we get about the initial classical binary message by looking at non-local degrees of freedom. If the difference between these two Holevo quantities is finite, then information has leaked out of the target qubit. In Unravelling information leakage from other sources of error we address other sources of error and discuss how to unravel them from true information leakage due to internal dynamics. The major issue we identify is shot noise which can artificially induce a significant signature of information loss. We address this issue in Experimental implementation by exploiting central limiting behaviour, and present an ansatz for extrapolating information loss at zero shot noise during our experimental implementation of the protocol.

We remark on two major elements of the protocol that impose constraints on the types of devices to which it is applicable. The platform must be capable of (i) full-state tomography on at least two qubits (a target qubit, and at least one other qubit in the device) (ii) initializing the target qubit in two definite states. We assume nothing about the microscopic features of the native physics of a device except that (iii) the dynamics are in some way local: qubits are assumed to couple more strongly to their neighbours than qubits further away in an array. This means we can look at nearest-neighbour qubits and random qubits, both of which are equally susceptible to single-qubit errors and shot noise but are not equally sensitive to information leakage affecting a fixed target qubit.

The Holevo quantity

The Holevo quantity χ over an alphabet A={pk,ρ^k} and channel Λ quantifies - in bits - the accessible information that is carried by the protocol wherein states ρ^kHi are selected with probability pk and transmitted through the channel Λ[ρ^k]Hf20,21. The Holevo quantity χ is given by

χ=SkpkΛ[ρ^k]kpkSΛ[ρ^k], 1

where S(ρ) denotes the von Neumann entropy of the state ρ. The Holevo quantity χ has been used to characterize information in many-body systems before and is a natural quantity with which to investigate information loss2225. We critically note that the initial space Hi and final space Hf are not generally identical. Consider e.g. a protocol that prepares a target qubit in a specific state, but performs measurements on the combined final state of the target qubit and its nearest-neighbours.

When maximized over all possible input alphabets A, χ gives the maximum classical capacity of the channel Λ. However, this maximization is often not possible in practice: the precise nature of a device’s underlying Hamiltonian, and thus of the channel Λ, is subject to debate - and the Hamiltonian parameters are subject to random fluctuations and drift. However, quantum computing platforms should approximately realize identities on idling qubits. Thus it is reliable to define an alphabet of equi-probable pk=1/dim(Hi), pure, and orthogonal messages that cover the state space Hi (this choice saturates χ when Λ is the identity channel).

We now discuss another important property of the Holevo quantity which becomes critically relevant when we try to account for shot noise in Experimental implementation. It is intuitively appropriate to consider χ as encapsulating the notion of distinguishability. Direct inspection of the form of χ reveals that an alphabet of pure states that are not all mutually orthogonal will cause the first term of Eq. (1) to fail to saturate, resulting in a low value of χ. Conversely, an alphabet of states which are close to maximally mixed states will saturate the first term, but also saturate the second - resulting in a low value for χ overall. Thus the Holevo quantity can be understood as a careful balancing act between the mixedness of the source (first term) and purity of the individual messages (second term) which together capture distinguishability. This may seem like a pedantic point, but it becomes critical when accounting for the effect of shot noise. Whilst the Holevo quantity is monotonically decreasing under the implementation of a CPTP channel, shot noise can not be represented as such, and can thus artificially increase the Holevo quantity. More precisely, shot noise is realised as random contributions to elements of the final density matrices Λ[ρ^k]. These contributions can increase distinguishability but does not actually increase the amount of information that the channel can bear. This effect becomes exacerbated in the case that the dimension of the initial space is smaller than the dimension of the final space. If dim(Hi)<dim(Hf), there are more elements in density matrices drawn from the final space, and thus shot noise introduces more differences between output density matrices.

In the context of this article (and in the absence of shot noise, which we address separately), the Holevo quantity quantifies how much information we can access about the initial state of a qubit given access to it and access to some other subset of qubits on the device. If no information has coherently leaked out of the target qubit, then access to some other region should not give us additional information about its initial state.

Protocol for quantifying idle information loss

We now leverage Holevo quantities computed over two subsets of qubits on a device into a protocol that quantifies the information lost to other qubits during idling. We first introduce the Holevo quantities χT and χTQ, respectively computed on the final reduced density matrices (i) of the target qubit alone ρ^kT, and (ii) of the target qubit and a complementary set of M − 1 qubits in the array ρ^kTQ. Note here that ρ^kT=TrE[ρ^kTQ], and thus by the monotonicity of the Holevo quantity under the partial trace:

ΔχTQ=χTQχT0 2

with equality only when information is fully localized to the target qubit, or when χTQ = χT = 0 and information has completely left the final space. Since the Holevo quantity yields the average number of bits of information that can be transmitted by messages passed through the channel, ΔχTQ can also be interpreted as the extra bits of information we can access about the initial state of the target qubit T given access to the complementary qubits Q. In ideal systems without shot noise or other sources of error, ΔχTQ suffices to identify and quantify information loss due to information leakage. Finite ΔχTQ indicates that some information has left the target qubit in a coherent fashion, and is distributed non-locally: i.e. that the device can’t ‘do nothing’ perfectly, even in ideal conditions. The protocol in full is as follows:

  • Initialize every qubit except the target qubit T in an arbitrary state ρ^E (we take pure separable product states of the logical single-qubit states throughout).

  • Initialize the target qubit T in the state 0.

  • Wait for a fixed period of time t (we measure immediately to address the ‘best case’ scenario with minimal information leakage such that t is the readout time of the device)26.

  • Perform full state tomography on the combined state ρ^0TQ of the target and complementary qubits (the subscript denotes that this state is conditioned upon the initial state, prepared in step 2).

  • Repeat from step 1, with the target qubit T initialized in the state 1, generating the state ρ^1TQ.

  • Process the states into a sample value for ΔχTQ.

  • Check the condition ΔχTQ > 0.

  • (Optional, to address noise) Repeat from step 1 for a large number of samples to build up statistics for ΔχTQ.

A significant strength of this protocol is that it is of constant complexity in total system size. The state tomography scales exponentially with the number of neighbouring qubits nc. For all physical quantum devices (which have constituent qubits of finite physical extent) the relevant coordination number nc is a constant, simply due to the fact that interactions fall off at long-range. This results in a fixed complexity for fixed architectures. For example, in IBM’s devices with a maximum value of nc = 3, one needs only perform state tomography on a nc + 1 = 4 qubit state, regardless of the size of the device. The value of t is arbitrary in principle, and can be tuned to address different regimes or properties of a device. Taking t to be e.g. fifty two-qubit gate times could give a good indicator of the total information leakage during a complete computation.

Unravelling information leakage from other sources of error

Unravelling information leakage from other sources is, in principle, a serious operational issue. As we discussed in The Holevo quantity, shot noise may artificially increase distinguishability. Moreover, other sources of error may artificially increase the difference ΔχTQ, which is not subject to the same monotonicity conditions as the Holevo quantity alone. In essence, ΔχTQ can be non-zero in realistic systems even if the information is fully localized to the target qubit. To deal with this we benchmark our results on nearest-neighbour qubits Q = P against random (non nearest-neighbour) qubits in the array Q = R, whilst keeping the target qubit T fixed. Both of these sets of complementary qubits should, when averaged over a large number of samples - be equally affected by e.g. shot noise, state-preparation-and-measurement (SPAM) errors due to e.g. inaccuracies in clock times, parameter drift, and other fine-tuning issues, and environmentally-induced decoherence. However, the randomly selected qubits should bear much less information about the initial state of T; and thus they can serve as a benchmark for all other sources of error in the device. Nevertheless, we choose to carry out our experiments at very short timescales, two orders of magnitude less than the dephasing times of the devices we use, to mitigate the effect of the environment as much as possible. Our protocol does not make use of any gates in its implementation, and thus gate errors are vacuously irrelevant. SPAM errors are relevant in the context of our protocol due to the considerable amount of single-qubit measurements required to perform (reduced) quantum state tomography (see Details of state tomography). The crucial point here is that all of the sources of error and noise listed above should effect (on average) all qubits in the array equally; whilst idle information leakage will more strongly affect qubits neigbouring the target qubit. Thus, by taking sufficiently many samples at sufficiently high shot counts on a random set of complementary qubits, background statistics can be determined which accounts for all the above sources of noise and error.

We emphasize here that our specific choice of time t as a single readout time (i.e. immediate measurement), our decision to insist that all qubits in the device are idling (as opposed to a realistic algorithm in which only some qubits are idling at any given time), and our eventual decision to exclude ‘bad qubits’ in Data analysis (see Data analysis and supplementary material for more information), all serve to minimize the impact of everything that isn’t idle information loss. As this work seeks to address fundamental, unavoidable, idle information loss that can’t be avoided by e.g. decreasing gate times and error rates or stronger shielding from environmental effects, we want to limit our experiments to settings in which these effects are minimized.

We implement these benchmarking procedures in Experimental implementation and ultimately find that shot noise is the dominant contribution to Δχ in the system of interest. We suggest, but can not prove, that channels comprised of a linear combination of single-qubit processes can not systematically increase ΔχTQ; as such single-qubit processes can not systematically transfer information from T to Q.

We first invoke the assumption that the underlying dynamics are, in a sense, local and that the idea of ‘nearest-neighbours’ is thus well-defined. In such systems, information should flow from the target qubit to its nearest-neighbours, and then through them to more distant qubits in the array. For a fixed target qubit T, we can then define two types of complementary sets of qubits: P corresponding to the plaquette of the M − 1 nearest-neighbours of T, and R corresponding to the target qubit and M − 1 randomly selected (excluding nearest-neighbour) qubits in the array. Since P and R both contain the same number of qubits they should, after averaging, be equally susceptible to single-qubit errors and shot noise (see Details of state tomography for a detailed discussion on the origin of this shot noise in IBM devices). In an ideal system without other errors or shot noise ΔχTPΔχTR > 0 should be a strict inequality (except at very late times where the information becomes fully delocalized). Individual samples may violate this inequality, but since both P and R are equally susceptible to all other sources of noise, these effects should cancel out after averaging over a large number of samples. We thus relax this condition to a statistical inequality

ΔχTP¯ΔχTR¯>0 3

that should hold in the presence of non-trivial information leakage. Moreover, Eq. (3) can be easily realized as the alternative hypothesis H1 for significance testing; which we do in Experimental implementation.

This inequality can be extended more formally by suggesting an ansatz form for Δχ as follows:

ΔχTQ=fηTQ,ηSHOTSTQ,NSHOTS. 4

Which is a function of the excess accessible information ηTQ, artificial information due to shot noise ηSHOTSTQ, and the shot count NSHOTS. We can aggregate statistics at different shot counts, perform a scaling analysis in NSHOTS, and extrapolate ηTQ as NSHOTS. We develop just such an ansatz based on our experimental data in Experimental implementation. This also formalizes the intuitive notion of the states of both the sets P and Q of complementary qubits being ‘equally susceptible’ to shot noise. We say that both are ‘equally susceptible’ to shot noise if ηSHOTSTPηSHOTSTR after fitting. With shot noise mitigated, the difference ηTPηTR then gives us the amount of information that has leaked from the target qubit into its neighbours.

In actual experiment (see Experimental implementation), we find ηTR → 0 exactly, supporting our previous assertion that other single-qubit sources of error can not systematically transfer information to distant regions of the device. This indicates that ηTP alone suffices to determine the amount of idle information loss in future experiments.

Experimental implementation

The experimental implementation of our protocol was carried out on the 27-qubit Falcon 5.11 series of IBM’s quantum computing devices; the architecture of such devices is shown in Fig. 1c. We incorporate two additional steps into the general protocol given in Protocol for quantifying idle information loss: (i) we only select target qubits with the highest possible coordination number nc = 3, which should yield the strongest signatures of idle information loss. And (ii) in step 1 of the protocol, we also simultaneously randomize over target qubits subject to the coordination constraint, and the Falcon 5.11 devices ibm_algiers, ibm_cairo, ibm_hanoi, and ibmq_kolkata themselves. In short, a single sample realization is formed from simultaneous random realizations of (a) a Falcon 5.11 device (b) a single qubit with coordination number nc = 3 on that device (c) an initial product state (d) nc = 3 non-nearest-neighbour qubits on that device on which the background ΔχTR is computed27. This corresponds to the generation of statistics for Δχ which are device agnostic - i.e. the user is interested only in running their computation job on a 27-qubit Falcon 5.11 device, and doesn’t care about the specific device to which their jobs are assigned. Realizations addressing the same processor and target qubit were made multiple times within the interval Dec 2024–Mar 2024; but never on the same initial product state or complementary set R of random qubits.

We performed readout immediately after state preparation, such that the wait time in step 3 of the protocol was just the readout time of the given device (between 700 ns and 900 ns). This minimizes the effect of the environment, as these readout times are several orders of magnitude lower than typical T1 and T2 times. This also means that our results represent the best-case scenario, and can not be improved by e.g. dynamical decoupling, as such processes are not possible during readout. Due to the existence of shot noise, the resulting density matrices after state tomography can have negative eigenvalues. We use a maximum-likelihood reconstruction method to rephysicalize the density matrices before post-processing28. A detailed discussion of the tomographic process, and of the maximum-likelihood reconstruction of aphysical density matrices, are both given in Details of state tomography and Aphysicality and maximum-likelihood reconstruction, respectively.

As discussed in Protocol for quantifying idle information loss, we consider both nearest-neigbour P and random R sets of complementary qubits for each target qubit, and carry out separate collections of experiments for each at a range of different shot counts NSHOTS. The number of samples for all shot counts NSHOTS on both nearest-neighbour plaquettes and random qubits is summarized in Table 2. All in all, our analysis of the Falcon 5.11 series involves the results of over 3500 experiments taken across four months (Dec 2023–Mar 2024), and represents a broad-spectrum comprehensive investigation of idle information loss on these devices. As the runtime increases proportionally with the shot count, most of this time was spent collating results for higher shot count samples; thus a direct comparison where statistics are compared for lots of low shot count samples, is likely the best approach for characterizing idle information loss on IBM devices in the future.

Table 2.

Total sample counts for all experimental implementations

Shot Number NSHOTS Complementary Set Samples
4000 P 609
R 600
8000 P 507
R 480
16,000 P 324
R 288
32,000 P 252
R 204
64,000 P 157
R 157

Sample counts, before filtering, for each shot count and type of tomographic complementary set: nearest-neigbour P and random (excluding nearest-neighbour) R. Each sample corresponds to a single realization of steps 1–7 of the protocol discussed in Protocol for Quantifying Idle Information Loss.

Data analysis

The experimentally determined statistics of ΔχTP and ΔχTR for 1209 total samples across all devices are shown in Fig. 2. The shot count for both was fixed at the standard value of NSHOTS = 4000. Both take a normally distributed form within certain limits, with ΔχTP exhibiting an additional small bump in the range 0.12−0.14. Were this bump due to any of the other processes discussed: decoherence, SPAM errors, shot noise, we would expect bumps to also appear in the statistics of ΔχTR. As this bump is not present in the statistics of ΔχTR, it stands to reason that this is strictly due to some target qubits coupling strongly to their nearest-neighbours. Moreover, we identify a bump in the statistics of ΔχTP at larger shot counts at the exact same position (see supplementary material), evidencing a true signature of idle information loss. The samples that form this bump can be justifiably interpreted as ‘bad qubits’ which have hybridized with their nearest-neighbours to such an extent that information leakage is dominating other sources of error. Over 10% of the information about the initial state of such a qubit is stored non-locally. We will exclude these bad qubits from our numerical analysis for the rest of the main text for two reasons. Firstly, bad qubits are typically artificially excluded (‘programmed around’) anyway when deciding on the geometry on which a practical algorithm should be run. Secondly, because bad qubits are extreme but rare events that skew the mean and variance of our data in a way that is uninformative in the context of identifying unavoidable idle information loss. We instead defer a parallel analysis of our data with bad qubits included to the supplementary material; where we find significantly increased error rates.

Fig. 2. Experimental sample statistics for Δχ Statistics of ΔχTP and ΔχTR for nearest-neighbour P (blue) and random R (orange) complementary sets of qubits respectively.

Fig. 2

The shot count is NS = 4000. Δχ quantifies how much extra information about the target qubit can be accessed by measuring non-target qubits in the complementary sets P or R, and can be artificially increased by shot noise, which should have an identical effect on both distributions. The ranges determined by Eq. (5) at a tolerance of K = 4 are shown as blue dashed lines for ΔχTP, and orange dotted lines for ΔχTR. Samples that fall outside of this range are classified as outliers, and excluded from the numerical analysis. Inset shows means and standard errors for the non-outlier region of each distribution; evidencing that Eq. (3) is satisfied and idle information loss has been directly detected after accounting for shot noise. A noticeable bump of outliers at larger values of only ΔχTP is evidence of `bad qubits' that have strongly coupled to their neighbours and are leaking lots of information during idling. Statistics are comprised of 1209 total samples, taken across four months (Dec 2023–Mar 2024) from all the Falcon 5.11 series of devices.

A more subtle feature of Fig. 2 is that the statistics of ΔχTP show a thicker tail at larger values than ΔχTR, suggesting that the inequality Eq. (3) is satisfied even when these bad qubits are excluded. To quantify this difference, we first filter the statistics for outliers by simple box-filtering according to the condition29,30,

Q1K×IQRΔχsTQQ3+K×IQR 5

where we have introduced the subscript s to denote individual samples of ΔχTQ. Q1 and Q3 are the first and third quartiles of the full sample statistics ΔχTQ respectively, and IQR is their inter-quartile range. Conservative filtration typically takes K = 1.5, whilst R corresponds to no filtering at all. We take K = 4 throughout as this both reliably contains the large Gaussian part of the distributions, and also excludes the bad qubit bump in the statistics of ΔχTP. We superimpose the boundaries of the box defined by Eq. (5) in Fig. 2 as dashed blue lines for ΔχTP and a dotted orange line for ΔχTR31. It is interesting to note that K can be invoked as a tuning parameter which defines precisely what is meant by a ‘bad qubit’. The part of the distribution to the right of the box defined by K can inform us of the probability that any given qubit will be a ‘bad’ one up to a certain informational tolerance. In cases where the entire device is required to complete a computation, or bad qubits are not identified and excluded before runtime, this probability could be used to place bounds on ultimate ruin: wherein said computation fails completely; a topic we recommend to future study. After filtering, we can compute the means ΔχTQ¯ and standard errors in the means σTQ of the resulting datasets, which are displayed in the inset of Fig. 2. The results satisfy Eq. (3), and provide a clear ‘smoking gun’ for idle information loss in the investigated devices. More formally, we carried out a one-tailed Welch’s t-test with null hypothesis H0:ΔχTP¯ΔχTR¯ and alternative hypothesis H1: Eq. (3). This yields a z-value of z = 3.30 (corresponding to a p-value of p = 0.00049) which implies that the inequality holds to a high degree of statistical significance. We perform similar tests at a range of shot counts NSHOTS, the resulting z-values and corresponding p-values of which are summarized in Table 1. These results indicate that H0 can be rejected to a very high degree of statistical significance, and thus that true signatures of idle information loss have been detected, across all investigated shot counts.

Table 1.

Hypothesis testing for signatures of idle information loss

Shot Count NSHOTS z-value p-value
4000 z = 3.3024 p = 4.9448 × 10−4
8000 z = 6.0996 p = 7.8796 × 10−10
16,000 z = 6.7304 p = 2.3839 × 10−11
32,000 z = 5.4390 p = 4.4893 × 10−8
64,000 z = 5.0447 p = 4.8972 × 10−7

Resulting z-values and corresponding p-values of single-tailed Welch’s t-tests given the null hypothesis H0:ΔχTP¯ΔχTR¯ after K = 4 filtering, for all shot counts NSHOTS.

Scaling analysis

We now interrogate the effect of shot noise more rigorously by investigating how the statistics and mean values change as we vary the number of shots NSHOTS. This will form the basis of developing an ansatz of the form of Eq. (4) with which a scaling analysis can be carried out. The results of this investigation are shown in Fig. 3 which shows the statistics of ΔχTP after K = 4 filtering as a function of shot count NS. The resulting distributions still appear normally distributed but drift to lower mean values with smaller variances as NSHOTS is increased. Interestingly, the mean value of ΔχTP¯ falls by a factor of 1/2 every time NSHOTS is doubled. This is further evidence that shot noise, which should follow central limiting behaviour and scale as 1/NSHOTS, is completely dominating the effects of idle information loss. We formalize this intuition by suggesting the simple ansatz

ΔχTQ¯=ηTQ+ηSHOTSTQNSHOTS 6

which incorporates both a flat information leakage ηTQ and a term that captures the distinguishability introduced by shot noise ηSHOTS. Essentially, ηTQ can be interpreted as the number of additional bits of information we can retrieve about the initial state of the target qubit given access to the other qubits in the complementary set Q in the zero-shot noise limit NSHOTS.

Fig. 3. Statistics of ΔχTP for a range of shot counts.

Fig. 3

Outlier-excluded K = 4 statistics of ΔχTP for a range of shot counts NS. As the shot count is doubled, the mean value of each set of statistics appears to scale by a factor of 1/2, indicating central limiting behaviour in NSHOTS which dominates signatures of idle information loss. A total of 1849 samples are distributed across the five different values of NSHOTS, a breakdown of which is provided in Table 2.

We present a fit of the mean values of the ΔχTP distributions (after K = 4 filtering) of Fig. 3 to the ansatz of Eq. (6) in Fig. 4, where we find very good agreement between the data and our ansatz. We also present a fitting for K = 4 filtered ΔχTR statistics as a benchmark. The fitting procedure allows us to extrapolate values for ηTQ and ηSHOTSTQ, which are shown for both TP and TQ in the inset text of Fig. 4. Standard errors on these extrapolated values are determined by a simple bootstrapping scheme in which each of the data points is randomly drawn from normal distributions determined by their respective means and standard deviations. These extrapolated values reveal that, as expected, both nearest-neighbour plaquettes and randomly selected qubits are equally affected by shot noise ηSHOTSTPηSHOTSTR. However, the underlying idle information loss saturates to a low but finite value ηTP = 0.00168 ± 0.00038 for nearest-neighbour plaquettes, and exactly to zero (to five decimal places) ηTR = 0 for randomly selected qubits. This evidences the conjecture made in Unravelling information leakage from other sources of error that the single-qubit processes that lead to e.g. SPAM errors and decoherence have no systematic effect on ΔχTQ, and that we can treat ηTQ as a true zero shot noise quantification of idle information leakage. Ultimately, the excess information gained by having access to the joint state TP when compared to TR yields, on average, an additional 0.00168 bits of information about the target qubit T. This is a low value, but it represents a direct quantification of the impact of idle information loss, and a fundamental limit on how well qubits in the Falcon 5.11 series of devices can perform.

Fig. 4. Scaling analysis and zero-noise extrapolation.

Fig. 4

Outlier-filtered K = 4 mean values ΔχTP¯ and ΔχTR¯ at different shot counts NSHOTS; error bars are shown where visible. Dashed lines show a fit to these values according to Eq. (6), with inset text showing the fitted values. The extrapolated NSHOTS zero shot noise values of Δχ(TQ)ηTQ quantify non-trivial leakage of information into nearest neighbours ηTP = 0.00168 ± 0.00038, and none into random distant qubits in the array ηTR = 0. Additionally, both complementary sets of qubits P and R seem equally affected by shot noise ηSHOTSTPηSHOTSTR. A total of 3578 samples are distributed across the five different values of NSHOTS and two types of complementary qubit sets, a breakdown of which is provided in Table 2.

Discussion

The foremost finding of this article is a protocol exploiting the Holevo quantity from quantum information theory. This protocol provides a flexible, scalable, device-non-specific solution to the burgeoning problem of quantifying idle information leakage in quantum computing platforms. A major feature of our protocol is that, for a fixed architecture, it has constant scaling in system size; on the IBM devices we investigate the state tomography is carried out on at most nc + 1 = 4 qubits, regardless of system size. Sufficient degradation in single-qubit protection could destroy the ability of a device to actually carry out quantum computations as information propagates, and our protocol serves as a direct quantification of the information lost (in bits) to this effect. A central component of our protocol is simply waiting, i.e. implementing the empty circuit; and it can thus be easily run during downtime with minimal oversight, replacing otherwise wasted time with a valuable characterization of errors in the computing platform.

The secondary finding of this article is the experimental implementation of our protocol on 3500 samples carried out across four months on all four of IBM’s Falcon 5.11 series of devices. The results of this analysis reveal (i) that a measurable amount of information about the state of any given qubit is leaking out during idling and (ii) the existence of ‘bad qubits’ which leak over 10% of a classical bit of information into their immediate surroundings. The ‘smoking gun’ of idle information loss takes the form of a statistical inequality that accounts for the effects of e.g. SPAM errors, decoherence, and shot noise. We find this inequality to be satisfied to remarkably high degrees of statistical significance at all shot counts. We also determine an ansatz from which exact idle information loss at zero shot noise can be extrapolated. The results of this extrapolation indicate that, after filtering for bad qubits, approximately 0.2% of the information stored locally is lost to idle information loss during a single readout time. This is a low, and but crucially non-zero, value which represents a fundamental limit on how well the Falcon 5.11 series of devices can perform.

Overall, our results indicate that in contrast to what has been suggested in the literature15, unwanted many-body effects do not constitute a significant issue on current-generation IBM quantum devices. In the infinite shot extrapolation, ignoring rare outlier events, the corresponding information loss falls to around 0.2%. However, the main finding of our work is that a measurable albeit small amount of information is already being lost in these systems and we can study this accurately with our method. This should allow us to systematically understand the impact of innovating technologies on the protection-operation dilemma. The near future of quantum computing promises dramatic scale-ups of system sizes, nascent error-correcting hardware, novel approaches to localizing information, and fledgling fault-tolerance. Our work provides a flexible, powerful, scalable protocol to quantify idle information loss in all these settings. This represents a decisive step towards addressing the threat many-body effects pose to high-fidelity idling, and thus to the long-term large-scale stability of generic quantum computing platforms.

Methods

Details of state tomography

State tomography in IBM devices is carried out by measuring all M qubits in the output register j1, j2,..., jM, where jr indexes the physical location of the r-th output qubit, simultaneously in the three Pauli bases. Taking the standard Pauli matrices, σ0=I, σ1 = X, σ2 = Y, σ3 = Z, the statistics of each Pauli string Pb=σj1b1σj2b2σjMbM with M non-identity elements (br ≠ 0 such that there are 3M of this family of string in total) is determined by these measurements. The result of each measurement is a bitstring, and the result of a large number NSHOTS of measurements - called ‘shots’ - is a dictionary Db of bitstrings. Taking NSHOTS sufficiently large ensures that the sample statistics of measuring a specific bitstring given the Pauli string Pb are close to the population statistics. These statistics for the reduced space of Pauli strings can then be aggregated into marginal values which yield statistics for the full space of 4M Pauli strings. This is done by simply aggregating shots from different Pauli strings together wherever they coincide everywhere except where identities occur in the desired marginal.

As a concrete example, consider an output register of three qubits, and lets say we are interested in the value of 〈XIY〉. This is given by the statistics of XXY = P1, XYY = P2, and XZY = P3. For each of these Pb statistics, there is a corresponding dictionary of results Db={Bbs} where the Bbs are bitstrings x which are simply the results of any single shot (measurement) given b. The statistics of P1 are given by calculation of the probability distribution Pb(Bb = x) = P(B = xb). An example dictionary for P1 for NSHOTS = 10 shots might be

D1={101,101,101,111,001,101,101,001,101,100} 7

with an associated probability distribution

P1(B1=x)=0.6ifx=1010.1ifx=1110.2ifx=0010.1ifx=1000.0otherwise. 8

Now consider the following example dictionaries for P2 and P3:

D2={111,101,111,111,011,111,101,011,101,101} 9
D3={101,111,011,111,011,110,111,001,101,110}. 10

We then aggregate the dictionaries Db into a single new dictionary D~ which describes the statistics of XIY by simply excluding the central bit of each bitstring and aggregating the dictionaries:

D1{11,11,11,11,01,11,11,01,11,10} 11
D2{11,11,11,11,01,11,11,01,11,11} 12
D3{11,11,01,11,01,10,11,01,11,10} 13
D~=D1+D2+D3 14

where D~ contains 30 elements. The statistics P~(B~=x~) of D~ are then calculated as

P~(B~=x~)=0.66°ifx=110.23°ifx=010.10ifx=100.00otherwise. 15

We can now evaluate 〈XIY〉 explicitly by summing up contributions to the expectation value 1 → 1, 0 → − 1 as follows

XIY=0.66°(1×1)+0.23°(1×1)+0.1(1×0)=0.3° 16

which completes our example.

This may seem a laborious process, but it allows us to determine O(4M) elements of a given state’s density matrix using only O(3MNSHOTS) measurements. By decomposing the state’s density matrix into a sum of 4M Pauli strings (including identity elements),

ρ^=1ZbPbPb 17

where Z is an appropriate normalization factor, we can readily reconstruct the quantum state of the output register using the dictionaries Dk. These dictionaries Dk are ultimately what IBM’s quantum computers return to their users. Where we address shot count NSHOTS, it simply refers the size of these returned dictionaries; where larger dictionaries more accurately yield the statistics of the actual population. The total number of samples of Δχ for both complementary qubit sets is shown in Table 2 for all shot counts NSHOTS we consider.

A single realization of the experiment required 162 circuits (two sets of 81 circuits for full tomography, regarding both the 0 and 1 initially loaded messages). The total job time for such a realization at the maximum shot count of 64,000 amounted to approximately 47 minutes, or under 30 seconds per circuit. Considering this is a very small amount of time which is constant in system size, and that this can only improve with new technology: our setup is not only currently practical, but will only become more practical as devices progress.

Aphysicality and maximum-likelihood reconstruction

Shot noise due to finite NSHOTS can result in aphysical density matrices by introducing negative eigenvalues into their spectra. We correct for this using the maximum-likelihood reconstruction of the density matrix.

The tomographic process discussed in Details of state tomography yields a density matrix μ^ with matrix elements μij which is definitionally of trace unity and hermitian by inspection of Eq. (17). The eigenvalues μj of μ^ can, however, be negative; and thus μ^ is generally unphysical.

We follow and briefly summarize here the maximum-likelihood mixed state reconstruction algorithm given in the work of Smolin et al.28. First we invoke the existence of some density matrix ρ^P with matrix elements ρij and eigenvalues ρj which minimizes the 2-norm,

minρ^Pρ^μ^22=minρ^Pijρijμij2 18

where P is the space of physical density matrices (unit trace, positive semi-definite, hermitian matrices). We remark here that Eq. (18) is invariant under change of basis, and hence we choose to work in the eigenbasis of μ^ with eigenvectors μj such that,

ρ^μ^22=ijρijμjδij2 19

where δij is the Kronecker delta. Clearly, Eq. (19) is minimized when ρ^ is also diagonal in this basis, i.e. the eigenvectors of ρ^ are also μj, as any non-zero off-diagonal terms ρij for ij strictly increases the value of Eq. (19). This reduces the minimization procedure down from an O(Ω2) problem (where Ω is the total dimension of the system) to a minimization problem in the Ω − 1 eigenvalues of ρ^,

minρ^Pρ^μ^22=min{ρj}jρjμj2 20

subject to only two constraints: that ρj ≥ 0, and that ∑jρj = 1. The reconstructed density matrix is finally given by

ρ^=jρjμjμj. 21

For the system sizes that we consider, the O(Ω) minimization problem can be solved quickly by standard numerical minimization packages, which is the procedure we use in this article. For larger problems, Smolin et al. provide a simple algorithm after reducing the complexity of the problem further by noting that the solution to Eq. (20) essentially involves finding a ‘pivot’ j in the (ordered) μj wherein ρj = 0 for j<j and ρj = μj + c where c is a constant for jj. The use of this algorithm is unnecessary for the situations we consider in this article, and we refer the interested reader to28 for more details.

Supplementary information

Acknowledgements

A.N.-K. would like to thank S. Bose, C. Turner, C. Berke, and O. Dial for fruitful discussions at the preliminary research stages. A.N.-K. sincerely thanks S. Shaikh and C. Bleger for their support. N.K. would like to give thanks to the QuSys group for discussions throughout the project, and also to G. García-Pérez, S. Filippov, and E. Borrelli for giving useful feedback and references. J.G. is supported by an SFI-Royal Society University Research Fellowship and is grateful to IBM Ireland and Microsoft Ireland for generous financial support.

Author contributions

All authors conceived of, developed, and co-ordinated the project. N.K. implemented the experimental aspects of the project. All authors discussed and analysed the results, and contributed to writing the manuscript.

Data availability

Data used in this project are available on request.

Competing interests

The authors declare no competing interests.

Footnotes

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Contributor Information

Alexander Nico-Katz, Email: nicokata@tcd.ie.

Nathan Keenan, Email: nakeenan@tcd.ie.

John Goold, Email: gooldj@tcd.ie.

Supplementary information

The online version contains supplementary material available at 10.1038/s41534-024-00918-6.

References

  • 1.Altman, E. et al. Quantum simulators: Architectures and opportunities. PRX Quantum2, 017003 (2021). [Google Scholar]
  • 2.Bravyi, S., Dial, O., Gambetta, J. M., Gil, D. & Nazario, Z. The future of quantum computing with superconducting qubits. J. Appl. Phys.132, 160902 (2022). [Google Scholar]
  • 3.Preskill, J. Quantum computing in the NISQ era and beyond. Quantum2, 79 (2018). [Google Scholar]
  • 4.Arute, F. et al. Quantum supremacy using a programmable superconducting processor. Nature574, 505 (2019). [DOI] [PubMed] [Google Scholar]
  • 5.Chow, J., Dial, O. and Gambetta, J. Ibm quantum breaks the 100-qubit processor barrier, IBM Res. Blog2 (2021).
  • 6.Kim, Y. et al. Evidence for the utility of quantum computing before fault tolerance. Nature618, 500 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Van Den Berg, E., Minev, Z. K., Kandala, A. & Temme, K. Probabilistic error cancellation with sparse pauli–lindblad models on noisy quantum processors. Nat. Phys.19, 1116 (2023). [Google Scholar]
  • 8.Takeda, K., Noiri, A., Nakajima, T., Kobayashi, T. & Tarucha, S. Quantum error correction with silicon spin qubits. Nature608, 682 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.AI, G. Q. Suppressing quantum errors by scaling a surface code logical qubit. Nature614, 676 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Xu, Q. et al. Constant-overhead fault-tolerant quantum computation with reconfigurable atom arrays, 10.1038/s41567-024-02479-zNat. Phys., 1 (2024).
  • 11.Xu, X. et al. Lattice hamiltonians and stray interactions within quantum processors, https://arxiv.org/abs/2402.09145 (2024).
  • 12.Silveri, M. & Orell, T. Many-qubit protection-operation dilemma from the perspective of many-body localization. Nat. Commun.13, 5825 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Qian, P., Xu, H.-Z., Zhao, P., Li, X. and Liu, D. E. Mitigating crosstalk and residual coupling errors in superconducting quantum processors using many-body localization, https://arxiv.org/abs/2310.06618 (2023).
  • 14.Varvelis, E. & DiVincenzo, D. P. Perturbative analysis of quasiperiodic patterning of transmon quantum computers: Enhancement of many-body localization. Phys. Rev. B109, 144201 (2024). [Google Scholar]
  • 15.Berke, C., Varvelis, E., Trebst, S., Altland, A. & DiVincenzo, D. P. Transmon platform for quantum computing challenged by chaotic fluctuations. Nat. Commun.13, 2495 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Börner, S. D., Berke, C., DiVincenzo, D. P., Trebst, S. & and Altland, A. Classical chaos in quantum computers. Phys. Rev. Research6, 033128 (2024).
  • 17.Basilewitsch, D. et al. Chaotic fluctuations in a universal set of transmon qubit gates, 10.48550/arXiv.2311.14592 (2024).
  • 18.Devitt, S. J., Munro, W. J. & Nemoto, K. Quantum error correction for beginners. Rep. Prog. Phys.76, 076001 (2013). [DOI] [PubMed] [Google Scholar]
  • 19.Ezzell, N., Pokharel, B., Tewala, L., Quiroz, G. & Lidar, D. A. Dynamical decoupling for superconducting qubits: A performance survey. Phys. Rev. Appl.20, 064027 (2023). [Google Scholar]
  • 20.Holevo, A. S. Bounds for the quantity of information transmitted by a quantum communication channel. Probl. Inform. Transm.9, 177 (1973). [Google Scholar]
  • 21.Nielsen, M. A. and Chuang, I. L. Quantum Computation and Quantum Information: 10th Anniversary Edition, 10th ed. (Cambridge University Press, USA, 2011).
  • 22.Nico-Katz, A., Bayat, A. & Bose, S. Information-theoretic memory scaling in the many-body localization transition. Phys. Rev. B105, 205133 (2022). [Google Scholar]
  • 23.Nico-Katz, A., Bayat, A. & Bose, S. Memory hierarchy for many-body localization: Emulating the thermodynamic limit. Phys. Rev. Res.4, 033070 (2022). [Google Scholar]
  • 24.Yuan, D., Zhang, S.-Y., Wang, Y., Duan, L.-M. & Deng, D.-L. Quantum information scrambling in quantum many-body scarred systems. Phys. Rev. Res.4, 023095 (2022). [Google Scholar]
  • 25.Zhuang, J.-Z., Wu, Y.-K. & Duan, L.-M. Dynamical phase transitions of information flow in random quantum circuits. Phys. Rev. Res.5, L042043 (2023). [Google Scholar]
  • 26.t is arbitrary in principle, and can be tuned to address different regimes or properties of a device. Taking t to be e.g. fifty two-qubit gate times could give us a good indicator of the total information leakage during a complete computation.
  • 27.Realizations addressing the same processor and target qubit were made multiple times within the interval Dec 2024–Mar 2024; but never on the same initial product state or complementary set R of random qubits.
  • 28.Smolin, J. A., Gambetta, J. M. & Smith, G. Efficient method for computing the maximum-likelihood quantum state from measurements with additive gaussian noise. Phys. Rev. Lett.108, 070502 (2012). [DOI] [PubMed] [Google Scholar]
  • 29.Schwertman, N. C., Owens, M. A. & Adnan, R. A simple more general boxplot method for identifying outliers. Comput. Stat. Data Anal.47, 165 (2004). [Google Scholar]
  • 30.Yang, J., Rahardja, S. and Fränti, P. Outlier detection: how to threshold outlier scores?, in 10.1145/3371425.3371427Proceedings of the International Conference on Artificial Intelligence, Information Processing and Cloud Computing, AIIPCC ’19 (Association for Computing Machinery, New York, NY, USA, 2019).
  • 31.It is interesting to point out that we can thus invoke K as a tuning parameter which defines precisely what is meant by a ‘bad qubit’. The part of the distribution to the right of the box defined by K can inform us of the probability that any given qubit will be a ‘bad’ one up to a certain informational tolerance. In cases where the entire device is required to complete a computation, or bad qubits are not identified and excluded before runtime, this probability could be used to place bounds on ultimate ruin: wherein said computation fails completely. We defer this topic to future study.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Data Availability Statement

Data used in this project are available on request.


Articles from Npj Quantum Information are provided here courtesy of Nature Publishing Group

RESOURCES