Abstract
We introduce a new approach to analyze single-molecule Förster resonance energy transfer (FRET) data. The method recognizes that FRET efficiencies assumed by traditional ensemble methods are unobservable for single molecules. We propose instead a method to predict distributions of FRET parameters obtained directly from the data. Distributions of FRET rates, given the data, are precisely defined using Bayesian methods and increase the information derived from the data. Benchmark comparisons find that the response time of the new method outperforms traditional methods of averaging. Our approach makes no assumption about the number or distribution of underlying FRET states. The new method also yields information about joint parameter distributions going beyond the standard framework of FRET analysis. For example, the running distribution of FRET means contains more information than any conceivable single measure of FRET efficiency. The method is tested against simulated data and then applied to a pilot-study sample of calmodulin molecules immobilized in lipid vesicles, revealing evidence for multiple dynamical states.
INTRODUCTION
Experimental studies of protein dynamics moved from ensemble-based averages to single-molecule measurements for good reasons.1 Single molecule observations probe dynamical features that are washed out in ensembles. Technical challenges of getting single-molecule data are significant, but they are not always the limiting factor. The challenges of making the most effective use of single-molecule data are a new frontier where much remains to be explored.2
FRET, standing for Förster resonance energy transfer,3 is perhaps the most important tool for exploring single-molecule dynamics.4, 5, 6, 7, 8 FRET probes protein configurations via time variations of fluorescent response. However, shot-noise limitations can degrade useful timing resolution by orders of magnitude compared to the intrinsic timing resolution of instruments. Analysis of single-molecule FRET has therefore used increasingly sophisticated statistical approaches to predict the FRET states underlying single-molecule measurements.9, 10, 11, 12, 13, 14, 15, 16 Analysis of dynamics is even harder. In this paper we explore a new method designed to make the most of the information available in FRET data while optimizing the possible time resolution.
The FRET efficiency
The main tool of traditional FRET analysis is the efficiency En, defined by
| (1) |
where na and nb are the number of detected photons in the acceptor or donor channels. The instantaneous estimate of En(t) at time t is quite meaningless, given that na(t) and nb(t) are either 1 or 0 instant by instant. For this reason the values of na and nb are commonly “binned” to make more stable estimators. For justification, it is common to appeal to maximum likelihood estimators.17 Although data binning has a certain appeal, we will show that it can degrade performance without adding reliability.
In general, the concept of “intrinsic values” of na(t), nb(t), and En(t) is flawed, because these quantities are not physically observable for single molecules. To be specific, let μ be the mean for a sample with (for example) Poisson-distributed photon counts. We cannot possibly measure n photons in a time t and know μ. At most we can learn something about the probability of μ via its probability distribution P(μ; t). Rather than the estimation of a single parameter at each time step, our goal is therefore to calculate a probability distribution P(μ; t) evolving in time.
Applying this notion to single-molecule FRET experiments, and considering what is observable, at best one could observe a distribution of FRET efficiencies, and this distribution may depend on time. At best, a single-molecule FRET experiment can only determine a probability distribution for the FRET efficiency of the single molecule system. Our goal is to implement a method that develops P(E, t), given the data. This approach does not merely estimate a parameter value (E), but a time-dependent function, the probability distribution of E. This distribution provides a new way to explore what can be known from the data.
Bayesian updating
The distribution-based method we describe uses the data in real time to update the distribution being estimated. This concept is Bayesian and implies active, real-time adaptation of the updating rules. A conceptual review of the Bayesian updating method is given in Ref. 18. In this paper we extend the method and test it against simulated data, comparing its performance to a simple maximum likelihood method, the running average. We also apply the method for the first time to the experimental FRET data on calmodulin, an important calcium-signaling protein.19
Our method converts running data of photon numbers na(t), nb(t) into a time-dependent joint distribution P(μa, μb; t). Starting from an assumed “prior” distribution (comprising what we already know about the distribution before data collection), the distribution is updated at each point based on the data. It is important that the output is a probability distribution rather than merely a parameter. Integration over the joint means distribution produces a running efficiency distribution P(E, t). As previously emphasized, no particular FRET efficiency can possibly be observable, whereas the efficiency distribution is well defined from the data. Just as a wave function in quantum mechanics contains much more information than an expected value, the distribution P(E, t) contains more information than any single efficiency parameter. The distribution can respond quickly to short term changes. It also provides useful global information about dynamical states of the system.
In this paper we show that the method enhances the timing resolution of the data, because much more information from the data is used from every run. For example, a 1000 point run treated by the time binning method might be reduced to 10 numbers, each representing an independent average over 100 photons. In our method a distribution having 20 μ-bins (say) is produced at each time step. The 1000 points of the data are transformed into 20 000 pieces of information. There is no paradox in generating this much information. Suppose each time step is represented as (red, green) in the set (0, 0), (1, 0), or (0, 1). There are three possible values and 31000 = 1.43 × 10477 different patterns in 1000 points. Using all possible information about the actual state of the molecule, photon by photon can produce better short-term predictions. The task of converting the information received into a running distribution turns out to be straightforward and statistically robust and is described in Sec. 2.
It is interesting that the time evolution predicted by Bayesian updating is not a strict “Markov” process. An important potential feature of single-molecule data is timing patterns or correlations in the data that come from fluctuations in single-molecule conformations.10, 13, 16, 20 This is because hidden degrees of freedom coupled to fluorescence observables may affect the time evolution that is observed, leading to time correlations.10, 20 Such correlations and fluctuations are exactly the topic single-molecule experiments wish to explore. With Markov assumptions the statistical rules of evolution are independent of the history of the state. While attractive, the Markov assumption may be inappropriate to single molecule studies because it ignores information in the history of fluctuations. Allowing a role for fluctuations opens the door to improve single-molecule timing observations traditionally washed out by ensemble methods.
Several recent reports also highlight the utility of Bayesian analysis for single-molecule data. Cao and Witkoskie introduced a Bayesian Markov chain indicator to compare the probability of different kinetic models in accounting for single-molecule trajectories.16 Landes et al. used Bayesian inference to identify periods of photoblinking by comparing the posterior probability of photoblinking and not-photoblinking, given single-molecule data, permitting photoblinking periods to be excluded from further analysis.21 Ensign and Pande introduce a Bayes factor to evaluate the evidence for the existence of change points in single-molecule trajectories.22 While the above methods were applied to single-channel data, a recent paper by Wiggins et al. described the application of a Bayesian approach to single-molecule FRET data.23 Given a general model (e.g., hidden Markov), the method they describe predicts the number of underlying states as well as model parameter values. This approach may be very useful if the general nature of the underlying model is known. Cumulatively, the treatments described above, as well as the approach we report in this paper, illustrate the capability of Bayesian methods to add information beyond that available from maximum likelihood estimations. The approach we present here uses Bayesian methods in an iterative fashion to predict the evolving FRET probability distribution. Our method does not assume a number of underlying FRET states as a part of the model. Instead, using our method, the number of states or the distribution of FRET efficiencies appear as a result of data analysis. As desired, parameters such as the FRET efficiency transition times can be estimated from the probability distribution.
Section 2 describes the distribution-based method and Sec. 3 compares with a benchmark standard. We show that the procedure is not inherently sensitive to fine procedural details. Section 4 describes the experimental method generating the data on calmodulin. The new method is applied to the calmodulin data in Sec. 5. Despite using sample sizes that are rather small, the procedure develops distribution-based conclusions that are quite robust.
THE EFFICIENCY DISTRIBUTION
The fundamental quantity of interest is the efficiency distribution developed from the distribution of rate parameters μa, μb. The efficiency itself is defined by
| (2) |
For comparison, the integer-based efficiency distribution is defined by En in Eq. 1. One cannot use En and Eμ interchangeably. There is little doubt that En creates a bias from integer ratios, and Eμ is physically correct. Indeed the original Förster analysis was based on the underlying rate parameters, which corresponds to using Eμ. In Subsection 2A we show how our method produces the distribution P(μ), and henceforth we use symbol E for “efficiency” consistently meaning Eμ.
The distribution of the mean
We use symbol P(A | B) to be the normalized conditional probability (a distribution) of A, given B. We let P(A) be the corresponding marginal distribution, implying integration over the variables not written. By definition the joint probability P(A, B) is given by
| (3) |
Models commonly specify P(n | μ), namely, the distribution of n, given some mean parameter μ. However, experiments seldom seek the probability of n photons, given a mean. Most experiments seek P(μ | n), which is the probability that the mean is μ given that n photons have been seen, and subject to prior conditions. The concepts of “probability of the mean” and “probability of the rate” are made precise using Eq. 3 and Bayes’ theorem:
| (4) |
The denominator in Eq. 4 amounts to an overall normalization factor. The definition of P(μ) must be self-consistent:
As shown in the Appendix, an appropriate model is a local-plus-broad distribution:
| (5) |
The local distribution Plocal(n|μ) can be chosen by the user, while we use the Poisson distribution Plocal(n|μ)→μne−μ∕n! for much of our data analysis. Symbol α is a parameter that is small compared to 1 and which is derived from the probability the system makes by a significant transition involving broad distributions Pbroad(μ), Pbroad(n). For all practical purposes, Pbroad(μ), Pbroad(n) can be approximated with a flat distribution over the entire range of n, μ. The parameter α sets the sensitivity to the new information in the data.
We will show that updating is rather insensitive to the details of Pbroad and the constant α, which we call the “annealing parameter” by a thermal physics analogy. For one thing, rather small parameters α ∼ 1% suffice to get relatively sensitive response to transitions. We report on calculations using a random positive distribution Pbroad∼Pran(μ) and a flat distribution.
Equation 5 gives us a prescription for an iterative algorithm capable of “learning” from data. With each incoming data point t, the distribution will follow the iteration
| (6) |
We turn to obtaining P(μ, t) from the data using Bayesian updating.
Single-channel updating
The self-consistent determination of the distribution P(μ, t) is done by using the data itself as it arrives at each moment t:
Begin at time step t = 0 with a “prior” or “seed” distribution P(μ) = P0(μ, 0). P(μ) is numerically defined on discrete steps μ1μp, and normalized by . Some reasonable tuning of the μ range and discretization can be expected: indeed, it is part of improving upon a maximum likelihood method, as described below. Over-populating the number of bins in μ has no effect other than increasing the computational load. Unless data sets are very small (a few points), the details of the seed distribution matter very little, and will be overcome by the data after a few steps. We will document this.
At time step 1, get the data n(1) and calculate P1(μ | n(1)) using Eq. 6. This produces the probability of μ, given that n(1) was observed and given the previous history.
At time step 2, get the data n(2) and use P1(μ | n(1)) as the updated prior for step 2.
Iterate, producing P(μ, t) = Pt(μ | n(t)). After a few steps the procedure “learns” the data distribution, and the seed distribution becomes irrelevant.
A detailed example calculation, showing how to implement a numerical version of the updating algorithm can be found in Appendix B.
General scope of the description
Assuming a Poisson process, the time-dependent number distribution P(n, t) is given by integrating over the distribution of μ:
| (7) |
It is interesting that the formula suggests P(n, t) should not generally be distributed by the classic Poisson rule, save the exceptional circumstance . This condition is unlikely to apply except under very narrow (and probably uninteresting) conditions. In general, we expect a distribution of rates μ resulting from heterogeneous and fluctuating interactions of the fluorophore with its surroundings.
One might ask what general class of functions has the representation of Eq. 7, and, in particular, if some data set might be incompatible. A little algebra resolves this. Let P(μ)=e−μP(μ) and P(n)=n!P(n), suppressing label t. Equation 7 becomes
| (8) |
which happens to be a Mellin transform. It is invertible with
| (9) |
The integration of Eq. 9 in the complex n-plane runs over a strip where Eq. 8 converges. When μ and n are reduced to discrete variables the corresponding problem in linear algebra can always be solved. This shows that Eq. 7 is perfectly general, and always “exact,” given the information about the distributions at hand. This fact is reassuring but only a side issue compared to the task of estimating the distribution of μ from data, which is our main topic.
Cumulative distributions
As in previous approaches, one of the main goals of our approach is the reduction of a data set to summaries capable of quantifying the type and number of states observed in the data set. For this we use the time-cumulative distribution of μ, defined by
| (10) |
The time-cumulative distribution is very robust when evaluated with many time steps τ ≫ 1. Each bin in μ tends to be stabilized by the central limit theorem.
Two channel distribution
The two-channel process is a direct generalization of the single-channel case. Let na, nb be photon counts in channels a, b, with (idealized) Poisson means μa, μb. Define the joint probability of two rates:
The subscript on Pt is redundant but emphasizes that the shape of the function has been updated and evaluated at each time slice. At each instant t, the updating rule is
| (11) |
As in the single-channel case, the value of P(μa, μb) at each step comes from the previous step. The distribution of FRET efficiencies is then
| (12) |
We turn to models developing the updating rule from the data.
The updating model
FRET data taken at ordinary data rates and with good timing resolution has donor and acceptor channels that are mutually exclusive. A molecule under FRET excitation either emits a “red” or a “green” photon, but not both. However, typical data, when binned over typical time intervals, does not show perfectly exclusive correlations. Simulations also need to take into account a finite probability that the emitted photon in either channel might not be observed.
We consult the time-cumulative (na, nb) distribution of a typical data set to proceed. Figure 1 shows the distribution for a subset selected as clean, as judged by strong signals and clear demarkation of the point of bleaching. Observe that the distribution of this data set tends to lie in an L-shaped zone. This does not imply that all data have the same distribution because probability may have moved or evolved within the L-shaped zone in any number of ways to give the same time-cumulative result. There is no reason to expect that the L shape is general. Other shapes may characterize other data, and the experimenter is free to adapt an updating model appropriate for the data.
Figure 1.
Motivation for the L-zone distribution. Right panel: contour plot of a typical FRET data set. Left panel: contour plot of an equal number of points from the L-zone distribution described in the text. Parameters β = 0.1 and γ = 0.2 were used for the figure.
Continuing, any ansatz capable of describing the probability inside the time-cumulative zone is suitable for updating purposes. Requirements are not particularly demanding. In general, an updating model should be able to accommodate the following:
The ability to represent anticorrelation between acceptor and donor channels without imposing it automatically.
A description of exclusive conditions where na, nb are as sharp as shot-noise permits.
Capability of representing low numbers of photons in either channel, or both.
Sensible upper limits on the number of photons in each channel; infinite numbers should be suppressed.
“L-Zone” distributions
A simple model satisfying these requirements is
| (13) |
The parameters γ, β are estimated from the data. Parameter γ cuts off the regions of n → ∞. The L-zone distribution meets each of the requirements enumerated above. The model can represent anticorrelation between acceptor and donor channels through the two probability zones represented by the Gaussians in na and nb. The “sharpness” parameter β is chosen based on the data (second requirement), including low numbers in one or both channels (third requirement). We note that the L-zone distribution does not assume either high or low-FRET states because μa, and μb are repeatedly updated and assume values driven by the data. Finally, the decay parameters restrict the upper limits for photon numbers in each channel (fourth requirement). So long as γ is well-chosen it has no effects other than transferring wasted probability at infinity back into the experimental region. The width parameter β is best obtained by fits to marginal distributions. One might introduce more than one β, γ parameters at the cost of complication. The gist of Eq. 13 allows a doubly infinite number of locations of Gaussian bumps.
The Gaussian L-zone model was validated with Monte Carlo simulations. A summary of the simulations can be seen in Fig. 1. For the simulation, 3000 random pairs of numbers (na, nb) were drawn from the L-zone distribution with means μ1 = 15, μ2 = 1. The L-zone model is not unique. Any other localized functions can replace the Gaussians, and the model serves as a convenient starting point to illustrate a large and versatile class. The L-zone distribution also does not a priori assume a particular number of underlying FRET states. As a result, the distribution of FRET states appears as a product of the data-driven analysis, not as an input requirement. The need for assumptions about the number of FRET states are a known handicap of many Markov chain models. In this respect, our method using the L-zone distribution, or any other suitable distribution, has an advantage.
BENCHMARK COMPARISONS
In this section we compare our procedure with the model “data” generated by simulations. Simulated data is used to characterize nominal states, establish the fluctuation of the detection statistic, and quantify response times. We find that the center of the data-derived efficiency distribution faithfully tracks the value of the efficiency put into the simulation. Under wide conditions we can deduce transitions in the data's efficiency values with an enhanced timing resolution compared to a simple running average.
Before reporting on those results we briefly compare standard benchmarks.
Maximum likelihood
The appeal of maximum likelihood lies in its apparent objectivity. We review the simple case of estimating a value of a Poisson mean μ from a data set n1, n2, …, nN.
The likelihood Ln of the data is the probability that uncorrelated data was seen given μ:
| (14) |
The log-likelihood is the sum of the logs:
The maximum likelihood estimate for μ comes from taking the derivative and setting it to zero:
| (15) |
Not surprisingly, the estimator found for the mean parameter is the mean of the data.
It is surprisingly common for the running average to be poorly suited to the experimental data analysis. Maximum likelihood has optimized the probability Ln of the data, given μ. But the point of the experiment is to optimize the probability of μ, given the data. The calculations are not the same. To calculate the likelihood of μ, denoted Lμ (note the subscript) after one event n1, use the identities of Eq. 4 (“Bayes Theorem”), from which
The right hand side contains P(μ), the prior information about the parameter μ. With a certain distribution P(μ), the likelihood of μ, given uncorrelated data n1, n2… is
| (16) |
Compare Eq. 16 to the usual likelihood function [Eq. 14] which lacks a factor of P(μ). In its typical implementation, omitting P(μ) from the calculation is equivalent to asserting P(μ) ∼ constant over the interval 0 < μ < ∞. It would be highly unusual for a flat distribution of mean parameters in the range 0 < μ < ∞ to be a realistic description of any data set. Indeed, the total probability spread over any finite region accessible to an experiment is zero.
Compare a calculation where μ < μ* is known to occur. Optimizing μ subject to μ < μ* will always produce a more accurate estimate than including μ → ∞. This is the weakness of many maximum likelihood estimators cited in the Introduction: No prepackaged statistic, uninformed about the actual data, could be expected to serve all purposes. By exploiting features learned from the data it is always possible to improve on standard maximum likelihood estimates. More sophisticated applications of maximum likelihood are possible.13 In this paper, however, we choose to explore the performance of the Bayesian updating method, and we compare our Bayesian updating method to the maximum likelihood in its naive running average form.
Define a running average of a quantity ξ(t) over smoothing time ΔT by
For discrete data we use the corresponding sum. The formula defines a backward average; other time offsets can also be used. The running average smoothes the data at a price of reducing response to changes over time.
As shown above, the running mean number is a maximum likelihood estimate for μ(t) that assumes a flat prior distribution over 0 < μ < ∞. The central limit theorem implies that under broad conditions the distribution of will be a Gaussian centered at the true mean. The distribution width from averaging N points is predicted to scale like . Apply this to estimate a mean using the running average:
| (17) |
With idealized conditions of a steady system the experimenter can run for an unlimited time and reduce the error Δμrunav to zero.
Yet we are interested in extracting dynamics from the fluctuations, not integrating them away. The uncertainty from making running averages tends to increase the inherent uncertainty that the underlying distribution σ represents. We express this with an uncertainty relation
For any given value of σμ the condition for the smoothing time ΔTnot to seriously degrade resolution is
| (18) |
This is a very modest requirement. It implies the possibility of averaging on the time scale of the single photon waiting time.
Under the impression it is statistically inevitable, experienced observers will impose μΔT ≳ 100 − 1000 to suppress fluctuations. There are consequences from the “ordinary” uncertainty principle. Consider an instantaneous transition between two channels at time t = 0. Model this with short-term exponentials, starting in state 1 at t = 0:
The Fourier resolution is
The characteristic width in frequency is Δω ∼ γ. The frequency–time uncertainty principle says ΔωΔT ≳ 1. It is important that this operates independent of sampling limitations. Smearing data over an averaging time ΔT automatically eliminates the ability to resolve rates γ ≳ 1∕ΔT.
Analysis of a single state
Dependence of the Bayesian updater on the priors
In this section, we study first the dependence of Bayesian updating on the initial priors. We find that Bayesian updating rapidly becomes insensitive to the choice of the prior. To illustrate the point, we first turn to analytic work. Let a data set consist of a set of p numbers (n1, n2…np). Let P(μ|n0) be the initial prior and Pbroad(μ) be a distribution obtained from random numbers, normalized to one, which appears in the term αPbroad(μ) of the Bayesian updating algorithm. In general, after m iterations the sequence looks like
| (19) |
Notice that the recursion relation for the Bayesian updater with annealing is highly nonlinear and does not have a simple closed form solution. The reason is that the number of terms increases with the number of iterations, so that after m iterations the expression contains m terms.
Only the first term in Eq. 19 contains the information about the initial prior, and it is suppressed by a product of m probabilities and m powers of 1 − α. It is also the only term in the updater containing the entire data history. Other terms involving more recent data are typically the dominant ones. As a result, even a delta function prior will eventually be “forgotten.” To illustrate, we perform the following simulation. We draw 40 random numbers from a Poisson distribution with a mean μ* = 3 and analyze the data with the Bayesian updater. We repeat the procedure with different initial priors presented in Fig. 2. Figure 3 shows the result for α = 0.01.
Figure 2.
Different priors used to test the Bayesian updater. The thick, yellow vertical bar represents the mean of the true underlying distribution P(n|μ).
Figure 3.
Evolution of the probability distribution P(μ|x(t)) for different priors given in Fig. 2. α = 0.01 for the purpose of the graphic.
In contrast, the likelihood function [Eq. 14] always contains the entire data history, rendering the method insensitive to abrupt changes in the underlying distribution P(nk|μ), as documented below for the running average. Also, the extreme case of a delta function prior will, in principle, not allow the likelihood function to evolve past it.
The ability of the Bayesian updater to “forget” the prior is weakly dependent on the choice of the annealing parameter α. Figure 4 shows slight differences between three different values of α = 0.01, 0.05, 0.1. Different curves in the figure correspond to different priors. Even the most unfavorable prior becomes irrelevant in less than 10 to 20 time steps, depending on the choice of α.
Figure 4.
Dependence of the Bayesian updater on the initial priors. Curves show the means from the Bayesian updater after each data point is analyzed. Different curves correspond to different initial priors P(μ|x0) shown in Fig. 2. In all cases, the updater shows weak dependence on the initial prior. Annealing parameter (left to right) α = 0.01, 0.05, 0.1.
Fluctuations of a single state
For our distribution-based method, we compute σμ(t) and 〈μ〉(t) defined by
We compare the performance of the running average , for which is the standard deviation of evaluated on the same smoothing time.
Each approach has a free parameter that balances stability against responsiveness. The running average depends on the smoothing time ΔT (the number of points averaged), developing smaller fluctuations but slower response for longer ΔT. Figure 5 shows the running average fluctuation 〈σ〉 as a function of ΔT. This study used a data set of 50 numbers drawn from a Poisson distribution with constant means of μ*1 = 1, 5, 10. Curves are the analytic prediction which fits very well.
Figure 5.
Simulation results: Standard deviation of the running average mean as a function of smoothing time Δt. Simulation parameter μ1* = 10, 5, 1 (top to bottom). 〈σ〉 is the standard deviation of averaged over 50-point simulated trajectories. The vertical spread shows results for 50 runs. Curves are . This study is done using the first half of simulation runs similar to one in Fig. 7, containing no transitions. Data points are slightly shifted horizontally for visual clarity.
In the distribution-based method, P(μ, t) and its responsiveness depend on the annealing parameter α. The distribution of σμ(t) over 50 data points is shown as a function of α in Fig. 6. These studies also use μ*1 = 1, 5, 10. For small μ* ≲ 1 the typical fluctuations tend to be comparable to those of the running average method (Fig. 5). However the fluctuations of the distribution-based method remain small and fixed even when the underlying distribution has μ* ≫ 1. This appears to come from a “memory effect” retaining information about the history, and the larger number of degrees of freedom used in constructing P(μ).
Figure 6.
Simulation results for the distribution-based method. The distribution of 〈σ(t)〉 is shown as a function of the annealing parameter α. Simulation parameters μ1* = 1, 5, 10. The vertical spread represents 50 runs. 〈σ〉 was averaged over the 50 data points.μ = 10 (magenta online), 5 (green online), 1 (blue online). Curves (red online) show the average over the 50 runs.
Analysis of multiple states
Response time for a single transition
We simulated processes with transitions between two states as follows. Fifty random numbers were drawn from a Poisson distribution with mean μ*1. This sequence was joined to 100 or more numbers from a Poisson distribution with mean μ*2. The joint set made 100-point runs (allowing extra points for smoothing delays) that have an instantaneous transition at t = 50. We made 50 runs of every particular study. A few simulation runs suffice to draw conclusions, because every study is repeated many times with nearby parameter choices that probe the statistical fluctuations.
Figure 7 shows the response of both the Bayesian updater and the 10-point running average to a transition. A state is considered to be “detected” at time t if 〈μ(t)〉 lies within 1 σμ(t) of the mean of the distribution generating it. A “transition” is signaled when the mean shifts to detect a different state. The “response time” ttrans is the number of time steps between the actual transition and its detection. The fluctuation has a trivial tendency to jump in the region of a significant transition. The figure illustrates the faster response of the Bayesian updater. Since a transition is signaled by lying within 1 unit of , the procedure exploiting the jump-related produces the most rapid detection signal we were able to construct. The statistical significance of any signal under large fluctuation criteria is naturally reduced, so that the procedure cannot be called “conservative.” Instead the study establishes minimum timing resolutions possible with the running average.
Figure 7.
Left Panel: Raw data for a simulated two state transition. The green horizontal bars represent the states with μ1, 2 = 1, 5. Right panel: Typical response to a transition of the Bayesian distribution P(μ, t) (shown in color scale from blue to white) compared to the 10-point running average (thick curve, red online) plus and minus the standard deviation (yellow online) represented by the dashed line. The running average is “backward-looking” and therefore increases at time points >50. The fluctuation is practically guaranteed to increase in the transition region, which makes for a particularly rapid detection scheme. “Hits” where are shown as stepped curve (thin black line) of 1 (detection) or 0 (no detection). The thick horizontal lines (green online) are the means of the Poisson distribution, μ*1 = 1 before and μ*2 = 5 after the transition. Shaded regions represent the widths of the Poisson distributions, i.e., . The vertical line (white online) shows the time where a transition is detected in the running average, i.e., where the running average is 1 σ from the mean.
The distribution over the first 50 time points serves as a negative control to show that the Bayesian updater does not show jumps or multiple states when none are present in the underlying simulated mean. The point is also made by the portion of the trajectory after the transition. The distribution predicted by the Bayesian updater, in fact, has a smaller standard deviation over these regions than the running average, as discussed above.
Running average response times are shown in Fig. 8. The simulation used full runs with μ*1 = 1 and μ*2 = 5. The average ttrans=ΔT scales accurately, with significant fluctuations about the mean. These simulations conform well to the analysis of the “uncertainty principle” controlling the running average method. In approaching those limits the running average performance is quite satisfactory for data generated by fixed and unique values of μ*. That is, the running average can be tuned into an excellent estimator provided that knowing the mean is entirely equivalent to knowing the underlying P(μ, t) distribution.
Figure 8.
Simulation results: Response times ttrans using the running average method as a function of the smoothing parameter ΔT. Points are the integer-valued response times; red (online) curve is the average of 50 runs.
Our distribution-based method automatically retains information about the shape of P(μ, t). Response times are shown in Fig. 9. Response times are remarkably fast—a few time steps—and also quite insensitive to the value of the annealing parameter α over the entire range shown. Here is an example of superiority of the distribution-based method. To achieve the same average timing response by a running average as ttrans∼2 shown for α ∼ 0.05 requires a smoothing time ΔT ∼ 1 (Fig. 8). This is the most rapid and earliest response possible. Yet it is certain to come with a fluctuation (Fig. 5). For μ ≳ 5 the growing fluctuations of the running average method cannot compete with the higher stability of the distribution-based method.
Figure 9.
Simulation results: Response times ttrans from our distribution-based method as a function of annealing parameter α. Dots are the integer-valued response times; red (online) curve is the average of 50 runs.
Efficiency statistics of multiple state transitions
In this section we use simulations to compare the time response of distributions from the Bayesian updating method to the response of the running average. Figures 1011 show typical time histories of multiple state transitions analyzed with the Bayesian updater and running average methods. The raw simulated data used in both figures is shown in Fig. 12. The simulations were done by switching between μ1 = 0.5 and μ2 = 5 periodically on a time scale of 5 units. Response of the running average on smoothing times ΔT = 10 and ΔT = 3 are shown. The top panels of each figure show the smoothed , oscillating in each channel. Gray bars indicate the parameters . The bottom panels show the distribution predicted by the Bayesian updater as contour plots. The distribution can be compared to the integer-based efficiency En(t) computed from the running average, along with the actual efficiencies of the underlying distribution shown as horizontal dashed bars. The bottom panels show the distribution predicted by the Bayesian updater compared to the integer-based efficiency En(t) running average, along with the actual efficiencies of the underlying distribution.
Figure 10.
Typical time histories of multiple state transitions analyzed with the Bayesian updater and the running average. Top panel shows the running average for , smoothed using ΔT = 10. Gray bars indicate the distribution parameters oscillating on t = 5 units. Bottom panel shows the predicted Bayesian updater probability distributions in contour plot. Dots (red online) show the integer-based efficiency En(t) obtained from the running average. Actual efficiencies are dashed horizontal bars (green online). α = 0.01, β = 2, γ = 0.2 were used for the purpose of the graphic.
Figure 11.
Time history of multiple state transitions, as in Fig. 10, smoothed using ΔT = 3.
Figure 12.
Several features are clear: (a) Resolving the short time variations of with small ΔT produces wild fluctuations in the efficiency En(t) reported by the running average. (b) Smoothing out the short time variations of tames the efficiency fluctuations of the running average while producing a false compromise efficiency that does not represent the underlying distribution. Indeed, it is impossible for the running average method to produce anything other than a compromise when the data has any sort of nontrivial distribution. (c) It is difficult to extract much information from the running average for highly fluctuating En(t). (d) The Bayesian updater is much more successful in tracking the underlying FRET efficiencies, showing distribution “peaks” at or near the correct underlying FRET efficiencies of 0.09 and 0.9. Although it also predicts some probability at intermediate FRET values due to the high interchange frequency, the peaks in the predicted distribution are near the correct values. The response time appears even faster than predicted in Fig. 9. The response time of the Bayes method depends on the data history and is probably shorter here because the transition occurs after a shorter time period at the previous efficiency value. Thus, under interchange rates high enough that the running average cannot resolve any population in the underlying high or low-FRET states, the distribution does detect the presence of these populations.
EXPERIMENTAL METHOD
Here we describe the experimental procedures. CaM T34C∕T110C was expressed, purified, labeled, and separated following methods described previously.24, 25 Purified CaM T34C∕T110C was labeled simultaneously with Alexa Fluor 488 maleimide (AF488) as donor (D) and Texas Red maleimide (TR) as acceptor (A). The double-labeled construct (CaM-DA) was separated by reverse phase high-performance liquid chromatography. Mass spectrometry verified the expected dye labeling.
Calmodulin was encapsulated inside lipid vesicles by techniques described in Refs. 26 and 27. Solutions of phosphatidylcholine (PC) and 16:0 biotinyl cap phosphatidylethanolamine (biotin-PE) (Avanti Polar Lipids, Alabaster, AL) were prepared (10 mg∕ml PC in chloroform and 1 mg∕ml biotin-PE in chloroform). A 100:1 mixture of PC to biotin-PE was prepared and the chloroform was evaporated under nitrogen. CaM-DA was dialyzed into the buffer of choice at a final concentration of ∼25 nM. The desiccated lipid was then hydrated with the CaM solution and subjected to ten freeze∕thaw cycles between a liquid nitrogen bath and a 30 °C water bath. Large unilamellar vesicles were formed by extrusion through a 100 nm pore membrane. The unencapsulated protein was separated from the vesicle entrapped protein with a Sepharose 4B column (GE Healthcare, Piscataway, NJ) equilibrated with the same buffer used to prepare the vesicles. The vesicles were used within 24 h of formation.
The high Ca2 + buffer consisted of 10 mM HEPES, 100 mM KCl, 1 mM MgCl2, and 0.1 mM CaCl2, adjusted to pH 7.4 and filtered through a 0.2 μm syringe filter. The low Ca2 + buffer was made up of 10 mM HEPES, 100 mM KCl, 1 mM MgCl2, and 3 mM EGTA, adjusted to pH 7.4 and filtered.
Vesicles were immobilized in flow cells constructed from microscope cover slips. Two cover slips used as a spacer were sandwiched between cleaned top and bottom cover slips with a gap to form a channel with approximate dimensions 0.3 mm × 4 mm × 22 mm. A lipid solution (20 mg∕ml) was incubated in the flow cell for 30 min to form a lipid bilayer. The excess lipid was washed away with the buffer solution, and the flow cell was further incubated with a solution of 0.2 mg∕ml streptavidin for 10 min, followed by another washing with the buffer. The final step was incubation with the extruded vesicles at an appropriate dilution for 5 min. Nonspecifically bound vesicles were thoroughly washed away with buffer and the finished flow cell was then placed on the microscope scanning stage. A well-made flow cell contained a ∼20 μl volume for at least 30 min. The buffer was replenished in the flow cell regularly to keep the sample hydrated during scanning.
Signals were collected by an inverted fluorescence microscope (Nikon TE2000) with a 60× objective lens (UPLSAPO 60XW Olympus, Center Valley, PA). The 488-nm line of an Ar ion laser (JDS Uniphase, Milpitas, CA) was directed through a Z488∕10× excitation filter to the microscope dichroic filter (500DCXR). Red and green signals were separated with a 565DCLP FRET dichroic. Green emission was detected through an HQ535∕50M green emission filter, and red emission through an HQ620∕75M red emission filter. All filters were obtained from Chroma (Rockingham, VT). Signals in red and green channels were detected by avalanche photodiodes (SPCM AQR-14, Perkin Elmer, Vaudreuil, Quebec). Immobilized molecules were located by raster-scanning a 15 × 15 μm2 region with a piezo-electric scanning stage (Mad City Labs, Madison, WI) with a laser power of 1 μW at 488 nm while detecting the signal from the red channel. Once an image was generated, the laser was shuttered and molecules were identified based on the intensity of the spots over the background. The scanning stage was then moved one molecule at a time to position each in the observation volume. The laser power was adjusted to 5 μW, the data collection software started, and the shutter opened until after the molecule under observation had undergone photobleaching. Counts from each channel were collected in the time-tagged mode and stored in computer memory. Trajectories were viewed in a LABVIEW program. For each single-molecule trajectory the start time (laser on) and the stop time (photobleach event) were ascertained by eye and recorded for further analysis. Multiple tracks were analyzed independently.
RESULTS—IMMOBILIZED CALMODULIN EFFICIENCY DISTRIBUTIONS
Time-dependent distributions of calmodulin
We analyzed 26 trajectories of immobilized calmodulin FRET data. Each trajectory comprises na, nb values sampled at 300 time steps. Data analyzed was raw: no background has been subtracted. Time step “0” was chosen by eye on the basis of the earliest significant rise in rates after applying full laser power. For the purpose of this analysis, we chose β = 0.01 and γ = 0.01, which was approximately fit from the marginal distributions of all samples. An annealing parameter α = 0.05 was used in updating. Consistency checks showed little sensitivity to small variations of α just as suggested by the simulations. Each trajectory was treated independently. Raw data can be seen in Fig. 13.
Figure 13.
Raw data used for analysis. Red lines are photon counts in the donor channel, green in the acceptor. Data is binned in 50μs bins.
Running probabilities P(μa, μb | na, nb, t) were developed over the range 0 ⩽ μ1, μ2 ⩽ 50 in steps of 0.5. Thus, each time step is represented by a 100 × 100 distribution of 104 values. The running efficiency distribution P(E, t) was collected by binning the E-distribution on intervals of ΔE = 0.01. We checked that the small effects of discretization of μ produced negligible bias with this binning.
Time histories of P(E, t) for typical data samples are shown in Fig. 14. Each time history graphically displays considerably more information than it is possible to get from the running average. Meanwhile, the running averages (red curve online, using ΔT = 10 units) also show a high degree of coincidence with the centers of the running distribution. Visual inspection of the plots suggests conformational transitions among multiple states.
Figure 14.
Samples of running efficiency distribution P(E, t) as a function of time for α = 0.05. Red curve shows the running average efficiency averaged over a bin of size 10 time steps. The running average here is forward looking, which shifts the averaged trajectory to the left. The vertical green line shows the approximate number of time steps that were used to calculate the cumulative efficiency distributions of Fig. 15. Notice the wide band at the end of several trajectories characteristic of donor photobleaching. Each time step is 500μs.
Time-cumulative distributions of calmodulin
The time-cumulative efficiency distributions developed by the analysis above were calculated with Eq. 10. Results corresponding to the six samples in Fig. 13 are shown in Fig. 15. Set by set there is a statistically significant signal that CaM FRET data consists of multiple states.
Figure 15.
Non-normalized time-cumulative efficiency distributions derived from the running distributions of Fig. 14. The data have not been corrected for background or cross-talk of donor emission into the red channel.
We collected the time-cumulative distributions of all 26 data sets for 0<t<300 into one grand-cumulative distribution (Fig. 16). The grand cumulative distribution has a higher statistical significance than each particular cumulative distribution. It also tends to wash out features. We believe it would be possible to resolve features even better by more sophisticated processing, either by the Bayesian distribution approach or by other methods described in the literature.8, 9, 10, 11, 12, 13, 14, 15, 16
Figure 16.
Non-normalized sum of cumulative distributions of the CaM data over the range 0<t<300 time steps and all the data files. Each time step is 500μs. Parts of the data which were photobleached are taken out.
The grand cumulative distribution also supports the observation of multiple states in CaM, which has previously been observed in samples approaching 105 molecules.28, 29, 30 The significance of multiple states of CaM has been discussed elsewhere.25, 26, 27
The joint μa, μb distribution
Bleaching of dyes is a common feature of FRET data and can be treated by many methods. Visual inspection of the time histories, together with the overall rates, indicated that the late-time data t ≳ 150 units is significantly contaminated by bleaching.
By correlating this observation with time-histories we were able to make very precise criteria to remove the data from bleaching. Figure 17 summarizes the process using the time-integrated P(μa, μb). The left panel shows the distribution including a significant peak at μa → 0, μb → 0 as well as a few other peaks at high μa and low μb present due to photobleaching. The right panel is the distribution after throwing out the data recorded after the point where (μa + μb) → 0. Careful visual inspection of the raw data coincided with the analytic signature of bleaching to within a few time steps. The numerous peaks seen in P(μa, μb) after bleached areas are removed appear to probe fluorescent states and the inherent dynamics of the molecule. A similar analysis may be useful for identifying different states or environments of the fluorophores. Note that Fig. 17 represents the dynamical history of a single molecule. The studies we conducted found equally interesting effects in almost every molecule, and with great variety. However, further analysis is beyond the scope of the current paper. Dynamic interchange among multiple conformational states may be important for the biological function of the protein.
Figure 17.
Left: The time-integrated distribution P(μa, μb) of a typical molecule that includes a significant period of bleaching. Right: The same study cut off at the time bleaching set in. The numerous peaks seen in P(μa, μb) appear to probe inherent dynamics of the molecule. Overall normalization in both plots is arbitrary.
CONCLUDING REMARKS
Single molecule data demand methods of analysis going beyond the tradition based on ensemble averages. We have shown that the concept of the distribution of underlying FRET rates is powerful and practical. Benchmark comparisons have shown that data-driven distributions of FRET rates are more efficient than some traditional Poisson estimators (i.e., the running average). The time-dependent, running distributions P(μ1, μ2; t) also capture information about the history of correlations that cannot be expressed with an ensemble language. Time-cumulative distributions P(μ1, μ2) and P(E) can always be constructed as convenient summaries of single molecule states. Previous single molecule FRET studies of calmodulin have shown multiple states using more traditional analytical methods with large numbers of molecules. The studies here confirm multiple states by using vastly more information from each data run. It appears that the vast information about configurational dynamics contained in single molecule measurements is becoming possible to obtain in an entirely new and unexplored manner.
ACKNOWLEDGMENTS
We thank a reviewer for helpful suggestions regarding the rationale for the annealing parameter. We thank Matt DeVore for helpful discussions. This work was supported under NSF Grant No. CHE-0710515. E.S.P. acknowledges support from the Pharmaceutical Aspects of Biotechnology Training Grant No. NIGMS 08359.
APPENDIX A: ANNEALING
Bayesian methods are capable of incorporating and representing information about systems that is more general than textbook rules of “independent” probability. Here we give the details underlying our approach. Let us first show that naive updating with a strict Poisson model does not represent an accurate model of a time-dependent system, unless the system is so slowly varying that the approximation of a delta-function distribution would be good.
Given a data sample (n1, n2, … nT), and a prior distribution P0(μ), naive updating with the Poisson distribution produces
| (A1) |
Let denote the cumulative average up to point T. Then for sufficiently large T, the distribution becomes arbitrarily sharp:
The second term is a saddle-point approximation near the distribution's peak at , and the overall normalization was dropped. The width of the Gaussian is , and the Gaussian approaches a delta function for . This recapitulates the central limit theorem. With at good sample resolution, the strict Poisson updater freezes after T ≫ 1 steps:
Upon reaching the stage where the probability P(μ) is completely localized, the weight given to the new data not agreeing with the localized model is driven to zero. Freezing of the distribution is quite difficult for the data to reverse, because it is driven by the cumulative history. Given a new data point, which is the transformation (n1, n2, … .nT) → (n1, n2, ...nT, nT + 1), the effect on the cumulative average is
Note the effect of order 1/T.
The result is an effective bias against new information simply because it is new. Accumulating data for longer and longer periods tends to guarantee no response before an equally long period. The mathematical mishap comes with the first step, Eq. A1. It enforces a hidden assumption of independence onto the data set (n1, n2, … nT) with a model
Inverting that model literally and including a prior gives Eq. A1. Independence is a common assumption of ensemble statistics not seeking time-correlated information. Excess reliance on independence is a mistake in Bayesian terms, because while we may believe P(n1|μ) is nearly Poisson, we do not accept repeating the error of independence literally over a long string of data. Thus naive updating contradicts our information that the distribution of μ may change, which happens to be the focus of the time dependence being studied.
In order to make a fair updating procedure, our prior beliefs and interest in time dependence must be given an opportunity to enter. Many approaches are possible. One approach might make models of correlation in order to use the information possible in P(n1, n2, … nT…). Another approach might introduce a sharp distribution of a time-dependent parameter μ(t). While tempting, the notion of μ(t) is equivalent to a continuously infinite set of parameters (μ(t) = μ0, μ1, μ2, …μt), which leads to a distribution of an infinite set, namely, functional integrals.
We choose the simplest path. Consider a distribution of n depending on two parameters, μ and x. Parameter x, restricted to 0<x<1, describes “transition regions,” where the uncertainties of P(μ) increase. There are many ways to do this. For simplicity let the condition x ≳ x0 represent a transition, and x ≲ x0 be the converse, or steady-state conditions. We will show in a moment that the exact value of x0 and the details defining it are irrelevant. The conditional distribution P(n | x, μ) then leads by Bayes to
| (A2) |
The overall normalization is not written. The “priors” in P(x, μ) = P(μ | x)P(x) describe how μ and x are correlated. To represent our transition features, we have
| (A3) |
Here P(μ) is the marginal μ distribution, updated step-by step, and Pbroad(μ) is a broad distribution in μ appropriate for transition regions. It will become clear that any reasonably disjoint functions can substitute for θ(x ⩽ x0) and θ(x>x0).
The n distribution is also subdivided. During a transition some broad distribution Pbroad(n|μ) is appropriate. Otherwise, the n-dependence will be represented by a local model Plocal(n|μ) for which the Poisson distribution Plocal(n|μ)→μne−μ∕n! is a good example. Thus,
| (A4) |
Combining gives
| (A5) |
Experiments don't observe x, so we integrate over it:
| (A6) |
where
| (A7) |
| (A8) |
So long as α ≪ 1, both the derivation and the performance of the updater are not particularly sensitive to the details of Pbroad(μ)Pbroad(n), so that numerous models collapse into one parameter. Inserting the Poisson model for Plocal and a flat distribution 1∕Nμ−bins for Pbroad gives the formula used in our analysis,
| (A9) |
APPENDIX B: EXAMPLE: SINGLE-CHANNEL UPDATING
We present an example calculation of single-channel updating. The example shows how to calculate P(μ| n(t)), given n(t) distributed by P(n|μ) = μne−μ/n!, called “the model.”
Consider data n(t) given by
| (B1) |
The data were generated by drawing two (2) random numbers from the model with μ1 = 3 followed by two (2) random numbers from the model with μ2 = 7. For the example, let the discrete μ values of the distribution range over 0 ⩽ μ ⩽ 10 in 5 steps μ = 1, 3, 5, 7, 9, representing rather large steps Δμ = 2. In practice, 20 or more steps will be preferred. The example uses α = 0.01 for definiteness.
To calculate the distribution P(μ | n(t)) using the four data points n(t) and conditions given, perform the following steps:
Step 0: Create a blank 5 × 4 array Pμ t = P(μ | n(t)).
Step 1: Initialize Pμ, 0 = 1/5 for all μ.
-
Step 2: Follow the method of the text to calculate
With α = 0.01 and n(1) = 8 we find the row(B2) Here 8! = 4480, 38/8! = 729/4480, etc.
- Step 3: Normalize Pμ 1 by dividing the entire row Pμ 1 by ∑μPμ 1:
(B3) - Step 5: Repeat steps 2, 3 for each point n(t) in the data set. At each subsequent step t, the distribution P( μ | n(t)) will be determined by
(B4)
For purposes of the debugging code, we include a list of 20 random numbers from the model with μ1 = 3 followed by 20 random numbers from the model with μ2 = 7.
| (B5) |
Let μ values run from 0.5 to 10 in 20 steps of Δμ = 0.5. Then at t = 10, 20, the values rounded to two decimal place are:
| (B6) |
Note that t is counted from t = 0. The resulting contour plot of the distribution Pμ t together with the generating parameters μ1, μ2 is shown in Fig. 18.
Figure 18.
Contour plot of the distribution calculated in the example. Horizontal dashed lines represent the mean parameter values.
Extension to two channel updating is straightforward. At each point in time t, one calculates not an array Pμ but a matrix Pμ1,μ2.
References
- Moerner W. E. and Orrit M., Science 283, 1670 (1999). 10.1126/science.283.5408.1670 [DOI] [PubMed] [Google Scholar]
- Yang H. and Xie X. S., J. Chem. Phys. 117, 10965 (2002). 10.1063/1.1521154 [DOI] [Google Scholar]
- Förster T., Ann. Phys. 437, 55 (1948). 10.1002/andp.19484370105 [DOI] [Google Scholar]
- Lamb D. C., in Single Particle Tracking and Single Molecule Energy Transfer, edited by Bräuchle C., Lamb D. C., and Michaelis J. (Wiley-VCH, Weinheim, 2010), pp. 99–129. [Google Scholar]
- Ha T., Enderle T., Ogletree D. F., Chemla D. S., Selvin P. R., and Weiss S., Proc. Natl. Acad. Sci. U.S.A. 93, 6264 (1996); 10.1073/pnas.93.13.6264 [DOI] [PMC free article] [PubMed] [Google Scholar]; Deniz A. A., Dahan M., Grunwell J. R., Ha T., Faulhaber A. E., Chemla D. S., Weiss S., and Schultz P. G., Proc. Natl. Acad. Sci. U.S.A. 96, 3670 (1999). 10.1073/pnas.96.7.3670 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Talaga D. S., Lau W. L., Roder H., Tang J., Jia Y., DeGrado W. F., and Hochstrasser R. M., Proc. Natl. Acad. Sci. U.S.A. 97, 13021 (2000). 10.1073/pnas.97.24.13021 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Schuler B., Lipman E. A., and Eaton W. A., Nature (London) 419, 743 (2002); 10.1038/nature01060 [DOI] [PubMed] [Google Scholar]; Lipman E. A., Schuler B., Bakajin O., and Eaton W. A., Science 301, 1233 (2002). 10.1126/science.1085399 [DOI] [PubMed] [Google Scholar]
- Rothwell P. J., Berger S., Kensch O., Felekyan S., Antonik M., Wörhl B. M., Restle T., Goody R. S., and Seidel C. A. M., Proc. Natl. Acad. Sci. U.S.A. 100, 1655 (2003); 10.1073/pnas.0434003100 [DOI] [PMC free article] [PubMed] [Google Scholar]; Margittai M., Widengren J., Schweinberger E., Schröder G. F., Felekyan S., Haustein E., König M., Fasshauer D., Grubmüller H., Jahn R., and Seidel C. A. M., Proc. Natl. Acad. Sci. U.S.A. 100, 15516 (2003) 10.1073/pnas.2331232100 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Watkins L. P., Chang H., and Yang H., J. Phys. Chem. A 110, 5191 (2006). 10.1021/jp055886d [DOI] [PubMed] [Google Scholar]
- Talaga D. S., J. Phys. Chem. A 110, 9743 (2006). 10.1021/jp062192b [DOI] [PMC free article] [PubMed] [Google Scholar]
- Antonik M., Felekyan S., Gaiduk A., and Seidel C. A. M., J. Phys. Chem. B 110, 6970 (2006); 10.1021/jp057257+ [DOI] [PubMed] [Google Scholar]; Kalinin S., Felekyan S., Antonik M., and Seidel C. A. M., J. Phys. Chem. B 111, 10253 (2007). 10.1021/jp072293p [DOI] [PubMed] [Google Scholar]
- Nir E., Michalet X., Hamadani K. M., Laurence T. A., Neuhauser D., Kovchegov Y., and Weiss S., J. Phys. Chem. B 110, 22103 (2006). 10.1021/jp063483n [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gopich I. and Szabo A., J. Chem. Phys. 122, 014707 (2005); 10.1063/1.1812746 [DOI] [PubMed] [Google Scholar]; J. Phys. Chem. B 113, 10965 (2009). 10.1021/jp903671p [DOI] [PMC free article] [PubMed] [Google Scholar]
- McKinney S. A., Joo C., and Ha T., Biophys. J. 91, 1941 (2006). 10.1529/biophysj.106.082487 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kou S. C., Xie X. S., and Liu J. S., Appl. Stat. 54, 469 (2005). [Google Scholar]
- Witkoskie J. B. and Cao J., J. Chem. Phys. 121, 6361 (2004); 10.1063/1.1785783 [DOI] [PubMed] [Google Scholar]; Witkoskie J. B. and Cao J., J. Chem. Phys. 121, 6373 (2004); 10.1063/1.1785784 [DOI] [PubMed] [Google Scholar]; Witkoskie J. B. and Cao J., J. Phys. Chem. B 112, 5988 (2008). 10.1021/jp075980p [DOI] [PubMed] [Google Scholar]
- See, e.g., “Kendall's Advanced Theory of Statistics” in Classical Inference and the Linear Model, edited by Stuart A., Ord K., and Arnold S. (Wiley, New York, 2009). [Google Scholar]
- Ralston J. P.2010. (unpublished).
- Chin D. and Means A. R., Trends Cell Biol. 10, 322 (2000). 10.1016/S0962-8924(00)01800-6 [DOI] [PubMed] [Google Scholar]
- Xie X. S., J. Chem. Phys. 117, 11024 (2002). 10.1063/1.1521159 [DOI] [Google Scholar]
- Taylor J. N., Makarov D. E., and Landes C. F., Biophys. J. 98(1) 164 (2010). 10.1016/j.bpj.2009.09.047 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ensign D. L. and Pande V. S., J. Phys. Chem. B 114(1) 280 (2010). 10.1021/jp906786b [DOI] [PubMed] [Google Scholar]
- Bronson J. E. et al. , Biophys. J. 97(12) 3196 (2009). 10.1016/j.bpj.2009.09.031 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Allen M. W., Bieber-Urbauer R. J., and Johnson C. K., Anal. Chem. 76, 3630 (2004). 10.1021/ac0497656 [DOI] [PubMed] [Google Scholar]
- Allen M. W., Bieber-Urbauer R. J., Zaidi A., Williams T. D., Urbauer J. L., and Johnson C. K., Anal. Biochem. 325, 273 (2004). 10.1016/j.ab.2003.10.045 [DOI] [PubMed] [Google Scholar]
- Rhoades E., Gussakovsky E., and Haran G., Proc. Natl. Acad. Sci. U.S.A. 100, 3197 (2003). 10.1073/pnas.2628068100 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Okumus B., Wilson T. J., Lilley D. M., and Ha T., Biophys. J. 87, 2798 (2004). 10.1529/biophysj.104.045971 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Slaughter B. D., Allen M. W., Unruh J. R., Bieber-Urbauer R. J., and Johnson C. K., J. Phys. Chem. B 108, 10388 (2004). 10.1021/jp040098u [DOI] [Google Scholar]
- Slaughter B. D., Bieber-Urbauer R. J., and Johnson C. K., J. Phys. Chem. B 109, 12658 (2005). 10.1021/jp051666o [DOI] [PubMed] [Google Scholar]
- Johnson C. K., Biochemistry 45, 14233 (2006). 10.1021/bi061058e [DOI] [PMC free article] [PubMed] [Google Scholar]


















