Skip to main content
Biophysical Journal logoLink to Biophysical Journal
. 2021 Jan 7;120(3):409–423. doi: 10.1016/j.bpj.2020.12.022

Generalizing HMMs to Continuous Time for Fast Kinetics: Hidden Markov Jump Processes

Zeliha Kilic 1, Ioannis Sgouralis 2, Steve Pressé 1,3,
PMCID: PMC7896036  PMID: 33421415

Abstract

The hidden Markov model (HMM) is a framework for time series analysis widely applied to single-molecule experiments. Although initially developed for applications outside the natural sciences, the HMM has traditionally been used to interpret signals generated by physical systems, such as single molecules, evolving in a discrete state space observed at discrete time levels dictated by the data acquisition rate. Within the HMM framework, transitions between states are modeled as occurring at the end of each data acquisition period and are described using transition probabilities. Yet, whereas measurements are often performed at discrete time levels in the natural sciences, physical systems evolve in continuous time according to transition rates. It then follows that the modeling assumptions underlying the HMM are justified if the transition rates of a physical process from state to state are small as compared to the data acquisition rate. In other words, HMMs apply to slow kinetics. The problem is, because the transition rates are unknown in principle, it is unclear, a priori, whether the HMM applies to a particular system. For this reason, we must generalize HMMs for physical systems, such as single molecules, because these switch between discrete states in “continuous time”. We do so by exploiting recent mathematical tools developed in the context of inferring Markov jump processes and propose the hidden Markov jump process. We explicitly show in what limit the hidden Markov jump process reduces to the HMM. Resolving the discrete time discrepancy of the HMM has clear implications: we no longer need to assume that processes, such as molecular events, must occur on timescales slower than data acquisition and can learn transition rates even if these are on the same timescale or otherwise exceed data acquisition rates.

Significance

Hidden Markov models (HMMs) have been a workhorse of single-molecule data analysis for the past 50 years. Yet, HMMs are inappropriate for molecular systems as they must assume, by construction, that single-molecule events occur much more slowly than the timescale of data acquisition. To move beyond fundamental HMM limitations, we must treat single-molecule events in continuous time as they occur in nature. Here, we exploit and generalize inverse methods for Markov jump processes to treat single-molecule events in continuous time. The implications of our work are profound and we can learn 1) the kinetics of single-molecule events without assuming these to be slower than the measurement timescale and 2) the rates on timescales faster than data acquisition.

Introduction

Hidden Markov models (HMMs) have been important tools of time series analysis for over 50 years (1,2). Under some modeling assumptions, detailed shortly, HMMs have been used to self-consistently determine dynamics of physical systems under noise and the properties of the noise obscuring the system’s dynamics itself.

Originally developed for applications in speech recognition (3,4), the relevance of HMMs to single-molecule time series analysis was quickly recognized (5, 6, 7, 8, 9, 10, 11, 12, 13, 14). Since then, HMMs and variants have successfully been used in the interpretation of ion channel patch-clamp data (15, 16, 17, 18, 19), fluorescence resonant energy transfer (FRET) (10,20, 21, 22, 23, 24, 25, 26, 27, 28, 29), force spectroscopy (30, 31, 32), among many other physical applications (3,33,34).

For HMMs to apply to single molecules and other physical systems, the assumptions underlying the HMMs must hold for such systems. There are several such assumptions worthy of consideration.

  • 1.

    HMMs assume that the system under study evolves in a discrete state space (whether physical of conformational). This is a reasonable approximation for biomolecules visiting different conformations (34, 35, 36) or fluorophores visiting different photostates (34,35,37). Of parallel interest to this point is the notion that the number of discrete states is known (though the transition probabilities between states is unknown). The assumption of a known number of states has been lifted thanks to extensions of the HMM (35,38, 39, 40, 41, 42, 43, 44) afforded by nonparametrics that we discuss elsewhere (34,35,39, 40, 41,44).

  • 2. HMMs assume that measurements are obtained at discrete time levels. That is, successive measurements are reported at regular time levels separated by some fixed period Δt. For clarity, we call Δt the “data acquisition period”. This assumption is consistent with a number of experimental biophysical settings (10,11,39, 40, 41,45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57).

  • 3. HMMs assume that physical systems transition between states in discrete time steps. Put differently, HMMs apply under the implicit assumption that the underlying system switches between states “rarely” as compared to the data acquisition period, Δt. This can only really be assured if the transition rates (as required in a continuous-time description) are slow. This assumption is implicit to the very definition of an HMM, which requires that the system’s switching occurs “precisely” at the end of the data acquisition periods (3,39,40,58, 59, 60).

This last assumption is problematic and presents the following conundrum: on the one hand, the transition rates are unknown and their analog for discrete time processes (transition probabilities) are typically to be determined using HMMs. On the other hand, we must assume that these unknowns are slow as compared to the data acquisition rate. Even if, optimistically, transition rates are slow, molecular events themselves are stochastic and, as with all physical processes, occur in continuous time. As such, any one event has a probability of occurring on timescales faster than the data acquisition period.

As an example, Fig. 1 illustrates the types of dynamical measurements that we can and cannot analyze within the HMM paradigm. The top panel shows an example of single-molecule measurements characterized by slow kinetics that can be analyzed within an HMM paradigm. In contradistinction to the above is the bottom panel illustrating an example of fast kinetics, as compared to the data acquisition rate, that cannot currently be analyzed within the HMM paradigm. The reason for this is simple: the fast kinetics give rise to a large number of apparent states that go beyond the two real states. This is because the measurements reported at each time point averages the molecular signal from all states visited in each acquisition period. Yet it is clear that the information on the transition rates between the rapidly switching molecular states is encoded in the time trace however uninterpretable.

Figure 1.

Figure 1

A conceptual illustration of single-molecule continuous-time-switching kinetics between discrete states probed in discrete time. For illustrative purposes only, the trajectory of a single molecule between two states (σ1,σ2) is shown in cyan in (a1) and (b1). For concreteness, we can think of these molecular states as conformational states. The state levels, i.e., signal level in the absence of noise, for these states is μσ1 and μσ2, respectively, and shown by the horizontal gray dotted line in (a1) and (b1). This synthetic experiment starts at time t0=0.05 s, and ends at time tN=20 s, and the data acquisition period, Δt, is 0.1 s. Next, again only for illustration, we assume that the measurements are acquired by a detector that has a fixed integration period τ=90 ms (light gray) for each fixed data acquisition period shown in (a1) and (a2). As a molecule may switch between states during an integration period, the measurements represent the signal levels capturing the average of the amount of time spent at each state levels (μσ1,μσ2) visited (in addition to added noise). (a1) and (a2) are associated with single-molecule kinetics that are slower than the data acquisition rate. Instead, in (b1) and (b2), we show a single-molecule trajectory when a molecule’s kinetics are faster than the data acquisition rate. In (b1), slow kinetics result in well separated state occupancy histograms around the average state levels. In (b2), we do not have well separated histograms centered around the average state levels because of the fast kinetics of the molecule. To see this figure in color, go online.

Indeed, to address the last assumption, a recent method (46), termed H2MM, was proposed. H2MM has been applied to single-molecule FRET photon arrival time series analyses (47,48). This method handles fast switching kinetics within an HMM framework by embedding a finer discrete timescale into the HMM: in this case, one fine enough to avoid the arrival of two photons within the same time bin. The H2MM applies to a scenario different from that provided in Fig. 1 for which the detector model produces measurement that coincide with noise on top of average molecular signal obtained throughout the detector’s exposure time.

Statistical analysis methods exploiting finer time grids, to approximate faster “continuous” time processes, had previously been considered albeit, for applications outside the natural sciences (61,62). Such statistical methods (61,62) have been criticized (63) for two main reasons: 1) they sometimes, though not always, introduce additional computational load because of the finer time grid and, almost certainly, 2) introduce bias by discretely approximating a continuous-time process. In the mathematical literature, these two challenges are what motivated the development of strategies to infer kinetic rates for genuinely continuous-time processes albeit measured in discrete time (63).

It is, therefore, natural to propose an analysis method that treats physical processes as they occur in “continuous” time to extract rates directly from traces with fast kinetics without relying on the artificial assumption that the physical processes involved occur on timescales much slower than the data acquisition period. To do so, we must fundamentally upgrade both key ingredients of the HMM model: 1) the system dynamics must be in continuous time; and 2) the measurement output must realistically reflect an average over the dynamics of the system over the data acquisition period. The output then encodes fast dynamics that can be retrieved (21,57,64, 65, 66, 67).

It is indeed to address processes evolving in continuous time that continuous-time Markov models, so-called Markov jump processes (MJPs), were developed (45,68, 69, 70). MJPs describe continuous-time events using rates (rather than transition probabilities) and recent advances in computational statistics (63,71, 72, 73, 74, 75) have made it possible to learn these rates given data. However, an important challenge remains: how to infer MJP rates under the assumption that the measurement process averages the probed signal over each measurement period? The nature of the measurement process and the inherent continuous dynamics, therefore, suggest a hidden MJPs (HMJPs) framework that we put forward herewith.

In the Methods section below, we start with the formulation of our HMJP model and also, briefly, summarize the HMM. Next, in the Results section, we move on to the head-to-head comparison of HMMs and HMJPs (showing in what limit the HMM exactly reduces to the HMJP). We focus on their respective performance in learning molecular trajectories and transition probabilities. We show how HMJPs successfully outperform HMMs, especially for kinetics occurring on timescales on the order of or exceeding the data acquisition period. Finally, in the Discussion section, we discuss the broader potential of HMJPs to biophysics. Fine details on the implementation of these two methods can be found in the Supporting Materials and Methods Section A.

Methods

In this section, we describe a physical system that evolves in continuous time alongside a measurement model. We also discuss how to generate realistic synthetic data from such a model and subsequently analyze time traces reflecting both fast and slow dynamics. We analyze the traces using two different methods: HMMs, as they are broadly used across the literature, and our proposed HMJPs, which we describe in detail. We compare the analyses in the Results section.

Model description

Using the experimental data, and the model of the experiment that we will describe, our goal is to learn 1) the switching rates between the states of the system (i.e., transition rates), 2) the state of the system at any given time (which we call the trajectory of the system), 3) initial conditions of the system, and 4) parameters describing the measurement process (i.e., parameters of the emission distribution).

Dynamics

We start by defining the trajectory T() that tracks the state of the system over time. Here, T(t) is the state of the system at time t and, as such, it is a “function” over the time interval [t0,tN]. We adopt functional notation and distinguish between T() and T(t) to avoid confusion with the entire trajectory and the value attained at particular time levels, critical to the ensuing presentation.

We label the states to which the system has access with σk and use the subscript k=1,,K to distinguish them. For example, σ1/σ2 may represent a protein in folded/unfolded conformation or an ion channel in an on/off state. With this convention (borrowed from (63)), if the system is at σk at time t, then we write T(t)=σk.

As with most molecular systems (36,39,76,77), the switching dynamics are faithfully modeled as memoryless. That is, the waiting time of the system in a state is exponentially distributed. Such memoryless processes are termed “MJPs” and below, we present their mathematical formulation. Memoryless dynamics often result from kinetic schemes relying on master equations, and in subsequent section, we explore the connections between our model and kinetic schemes in detail.

At the experiment’s onset, we assume the state of the system T(t0) is chosen stochastically among σk. We use ρσk to denote the probability of the system starting at σk and collect all initial probabilities in ρ¯=(ρσ1,ρσ2,,ρσK), which is a probability vector (78, 79, 80, 81).

Memoryless switching kinetics are described by “switching rates” between all possible state pairs. These switching rates are labeled with λσkσk and, in biophysics, are most commonly termed transition rate coefficients. By convention, all self-switching rates are zero λσkσk=0, which, in general, allows for at most K(K1) nonzero rates (36). Although, the switching rates λσkσk fully describe the system’s kinetics, as we will see shortly, it is mathematically more convenient to work with an alternative parametrization. In this alternative parametrization, we keep track of the “escape rates”

λσk=k=1Kλσkσk, (1)

which, for simplicity, we gather in λ¯=(λσ1,λσ2,,λσK). In biophysics, each λσk is understood to correspond to the reciprocal of a mean dwell time. Furthermore, instead of keeping track of each rate λσkσk, we keep track of the rates normalized by the escape rates, namely

πσkσk=λσkσkλσk (2)

Gathering all normalized rates out of the same state in π¯σk=(πσkσ1,πσkσ2,,πσkσM), we see that each π¯σk forms a probability vector (80). The entries of these probability vectors can also be termed “splitting probabilities.”

In summary, instead of K(K1) switching rates λσkσk, we describe the system’s kinetics with K escape rates λσk and K switching probability vectors π¯σk. The latter have, by convention, πσkσk=0, and so, the total number of scalar parameters is the same in both parametrizations. Below, for simplicity, we gather all transition probability vectors into a matrix

π¯¯=(π¯¯σ1π¯¯σ2π¯¯σK) (3)

Measurements

The overall input to our method consists of the measurements x=(x1,x2,,xN) acquired in an experiment. Here, xn indicates the nth measurement and, for clarity, we assume measurements are time ordered, so n=1 labels the earliest acquired measurement and n=N the latest. These measurements may be image values, photon counts, FRET efficiencies, derived intermolecular extensions, or any other quantity determinable in an experiment.

Each xn is reported at a time tn=tn1+Δt, which is Δt later than the time tn1 at which the previous measurement xn1 was reported. For completeness, together with the time levels t1,t2,,tN at which a measurement is reported, we also consider an additional time level t0, that marks the onset of the experiment, which is not associated with any measurement, Fig. 1.

The most common assumption made almost universally by HMMs is that the instantaneous state of the system at tn determines xn. Yet, for realistic detectors, the reported value xn is influenced by the entire trajectory of our system during the nth integration period, which we represent by the time window [tnτ,tn]. Here, τ is the duration of each integration time (such as an exposure period for optical experiments).

We account for detector features in the generation of the measurements via characteristic state levels that we label with μσk and, for simplicity, gather these in μ¯=(μσ1,μσ2,,μσK). Informally, we think of each μσk as corresponding to a distinct signal level. In this formulation, each σk is associated with its own characteristic level μσk. If the system remains at a single state σk throughout an entire integration period [tnτ,tn], then the detector is triggered by μσk and so, provided that the measurement noise is negligible, the reported measurement xn is similar to μσkτ. However, if the system switches multiple states “during” an integration period, the detector is influenced by the levels of every state attained and the time spent in each state.

More specifically, the nth signal level triggering the detector during the nth integration period, [tnτ,tn], is obtained from the time average of μT() over this integration period. Mathematically, this time average equals 1τtnτtndtμT(t) and, provided measurement noise is negligible, the reported measurement xn is similar to the value of this average.

In the presence of measurement noise, such as shot-noise (82, 83, 84, 85, 86), quantification noise (87, 88, 89), or other degrading effects common to detectors currently available, each measurement xn depends “stochastically” upon the signal that triggers the detector (34,90,91). Of course, the precise relationship depends on the detector employed in the experiment and differs between the various types of cameras, single-photon detectors, or other devices used. To continue with our formulation, we assume that measurement noise is additive, which results in

xn|TNormal1τtnτtndtμTt,v. (4)

The latter expression is a statistical shorthand for the following items: the measurement xn is a random variable that is sampled from a normal distribution whose mean is 1τtnτtndtμT(t) and whose variance is v. For the normal distribution, the variance is related to the detector’s full-width-at-half-maximum (FWHM) by v=(FWHM2/8log2).

Our Eq. 4 is general enough to capture the effect of the history of the system during the detector’s integration time or, put differently, to capture the effect of a low pass filter. Of course, our choice of normal distribution itself is incidental and can be modified depending on the type of detector used to obtain the measurements xn. For example, in an accompanying article (92), we adapt Eq. 4 to FRET measurements in separate donor and acceptor channels with shot-noise and background as follows

xnD|TPoissontnτtndtμTtD,
xnA|TPoissontnτtndtμTtA,

where xn=(xnD,xnA) denotes the measurements acquired in the donor’s and the acceptor’s channels, respectively.

Simulation

Given ρ¯ and λ¯,π¯¯, a trajectory T() that mimics real systems may be simulated using the Gillespie algorithm (76), which we describe briefly here only in an effort to introduce necessary notation. Simulations by means of the Gillespie algorithm are also known as “kinetic Monte Carlo simulations.” Often, the simulated dynamics, which are characteristic of MJPs, are obtained by formulations involving “master equations.”

To begin, an initial state s0 is chosen among σ1,σ2,,σK with probability ρσk. Then, the period d1 that the system spends in s0 is chosen from the exponential distribution with mean 1/λs0. Subsequently, the next state s1 is chosen among σ1,σ2,,σK with probability πs0σk. Because πs0s0=0, any chosen s1 is different from s0; therefore, the transition s0s1 is a jump in the system’s time course that occurs at time t0+d1. Next, a new period d2 is sampled from an exponential distribution with mean 1/λs1 and a new state s2 is chosen among σk with probability πs1σk, and so on. These steps are repeated until the end of the experiment, which, in our setup, is the same as the time tN of the last measurement.

More formally, we summarize the sampling of a Gillespie trajectory as follows

s0Categoricalρ, (5)
dm|smExponentialλsm, (6)
sm+1|smCategoricalπsm, (7)

for m=0,1,2,,M1, where M1 is the lowest value such that

t0+m=0M1dmtN (8)

The categorical distribution we use here is the generalization of the Bernoulli distribution for which more than two outcomes are possible (58).

The successive states of the system s0,s1,,sM1 and the associated durations d0,d1,d2,,dM1, which we term “holding states” and “holding times,” respectively, encode T() throughout the experiment’s time course [t0,tN]. Namely,

T(t)={s0ift0t<t0+d0s1ift0+d1t<t0+d0+d1sM1ift0+d0++dM2t<t0+d0+d1++dM1 (9)

For convenience, we summarize the representation of T() in a triplet (S,D,M), where S = {s0, s1, …, sM – 1}, D = {d0, d1, …, dM – 1}, and M is the size of S and D.

Once a trajectory is obtained through the Gillespie algorithm just described, the signal levels tnτtndtμT(t) for each integration period can be computed. For instance, as the trajectory is piecewise constant, the integrals reduce to sums that can be easily calculated. Therefore, given an appropriate detector model, such as Eq. 4, and a trajectory’s triplet (S,D,M), we can obtain simulated measurements by adding noise according to the detector’s distribution. The graphical model, shown in Fig. 2, illustrates all of the dependencies of the parameters discussed above.

Figure 2.

Figure 2

Graphical representation for HMJP framework. Here, we provide a graphical representation for our HMJP framework. The notation followed in this figure is consistent with that presented in the main text. To see this figure in color, go online.

Model inference

Using experimentally obtained measurements x and the model of the experiment that we have just described, our goal is now to learn initial probabilities ρσk, switching rates λσkσk, and state levels μσk for all states as well as the trajectory of the system T() throughout the experiment’s time course [t0,tN]. Below, we attempt to learn these model parameters by using time series analysis with an HMM and then introduce a novel, to our knowledge, time series analysis relying on HMJPs.

Model inference via HMMs

An HMM requires that each measurement xn depends exclusively on the “instantaneous” state of the system, namely T(tn). In view of Eq. 4, this is achieved by the trajectory of the system T() remaining constant during the integration period [tnτ,tn]. To a sufficiently good approximation, this is satisfied provided

τλσk1, (10)

for all σk. Further details for the bound in Eq. 10 are provided in Appendix A.1.9 in the Supporting Materials and Methods. Thereby, the system rarely exhibits switching during periods that last shorter than τ. This approximation allows for 1τtnτtndtμT(t)μT(tn) to be used. Accordingly, in an HMM, Eq. 4 is replaced with

xn|TNormalμTtn,v. (11)

Again, as with Eq. 4, the exact choice of probability distribution (whether normal or otherwise) is incidental. HMMs can treat any emission distribution “provided” xn only depends on T(tn) as opposed to the full history of the trajectory over the integration time.

With the measurements described by Eq. 11, we can use an HMM to learn the probabilities of the transitions T(tn1)T(tn)T(tn+1). For clarity, from now on, we will use cn=T(tn) and denote these transitions with cn1cncn+1. That is, cn is the state of the system “precisely” at the time tn.

For an HMM, transition probabilities are denoted with Pcn1cn. Because the system can attain K different states σk, in general, an HMM possesses K×K transition probabilities Pσkσk. Now, we gather the transition probabilities out of the same σk in a vector P¯k=(Pσkσ1,Pσkσ2,,PσkσK) and, for clarity, gather all of the vectors in a matrix

P¯¯=(P¯¯σ1P¯¯σ2P¯¯σK) (12)

The matrix P¯¯ is related to the system’s switching rates λσkσk and escape rates λσk. Specifically, if we gather them in

G¯¯=[λσ1λσ1σ2λσ1σKλσ2σ1λσ2λσ2σKλσKσ1λσKσ2λσK], (13)

termed the “generator or rate matrix” (80,93), then the time evolution of the probability of being in conformational state σk at time t, pσk(t), is governed by the “master equation,”

dpσk(t)dt=kkpσk(t)λσkσkpσk(t)kkλσkσk, (14)

for all k=1,2,,K. For finite conformational states, the analytical solution to the master equation” is

p¯(t)=p¯(t0)exp(G¯¯(tt0)), (15)

where

p¯(t)=(pσ1(t),pσ2(t),,pσK(t)) (16)

Now, following Eq. 15, we have set p¯(t0)=ρ¯, where

ρ¯=(ρσ1,ρσ2,,ρσK). (17)

For the evolution over Δt, we have p¯(tn+1)=p¯(tn)P¯¯, where we define the transition probability matrix, P¯¯, as

P¯¯=exp(G¯¯Δt), (18)

where exp() denotes the matrix exponential. We point out that π¯¯ and P¯¯ are both probability matrices; however, they assume quite “different” properties. For instance, πσkσk=0, whereas Pσkσk>0.

Although knowing G¯¯ is sufficient to specify P¯¯, the inverse is “not” true: knowing P¯¯ does not necessarily lead us to a unique G¯¯ and so the switching rates “cannot” simply be inferred from P¯¯. This is a consequence of the multivalued nature of the logarithm. As such, one transition probability matrix may corresponds to multiple rate matrices (93). Instead, provided λσkΔt1 for all σk, we may approximate Eq. 18 by

P¯¯I¯¯+G¯¯Δt, (19)

where I¯¯ is the identity matrix of size K×K. Under this approximation, we can estimate transition rates by G¯¯(P¯¯I¯¯)/Δt. Otherwise, when λσkΔt1 for some σk does not hold, then the transition rates estimated with Eq. 19 are inaccurate.

Below, we highlight the steps necessary to estimate the quantities of interest in an HMM. Specifically, an HMM relies on the statistical model

c0Categoricalρ, (20)
cn+1|cnCategoricalPcn, (21)
xn|cnNormalμcn,v. (22)

To model the full distribution over the quantities of interest (e.g., initial probabilities ρ¯, transition probabilities P¯σk, state levels μ¯, and the trajectory of the system T(), which is encoded by c=(c0,c1,,cN)), we follow the “Bayesian paradigm” (78,94). Within this paradigm, we place prior distributions over the parameters, and we discuss the appropriate choices next.

On the transition probabilities P¯σk, we place a Dirichlet prior with concentration parameter A

Pσ1DirichletAK,AK,,AK, (23)
Pσ2DirichletAK,AK,,AK, (24)
PσKDirichletAK,AK,,AK, (25)

which is conjugate to the categorical distribution (39,40,42,81). We consider a similar prior distribution, with concentration parameter α, also for the initial transition probability ρ¯, namely

ρDirichletαK,αK,,αK. (26)

Subsequently, we place priors on the state levels μ¯=(μσ1,μσ2,,μσK). The prior that we choose is the conjugate normal prior

μσkNormalH,V, (27)

with hyperparameters H, V.

Once the choices for the priors are made, we then form the posterior distribution (35,39, 40, 41, 42, 43, 44)

P(ρ¯,P¯¯,μ¯,T()|x)=P(ρ¯,P¯¯,μ¯,c|x), (28)

containing all unknown variables that we wish to learn. Given that the posterior above can be constructed from the likelihood that we have defined in Eq. 11 and the priors that we have defined in Eqs. 20, 21, 22, 23, and 24, Eqs. 26 and 27 can be more explicitly written as follows:

P(ρ¯,P¯¯,μ¯,c|x)P(x|ρ¯,P¯¯,μ¯,c)P(ρ¯,P¯¯,μ¯,c), (29)
P(x|c0:N,μ¯)n=0N1P(cn+1|cn,P¯¯)P(c0|ρ¯)P(P¯¯)P(ρ¯)P(μ¯), (30)
n=1NNormal(xn;μcn,v)P(cn+1|cn,P¯¯)P(c0|ρ¯)P(P¯¯)P(ρ¯)P(μ¯), (31)

However, the posterior distribution in Eq. 29 does not attain an analytical form. Therefore, we develop a specialized computational scheme exploiting Markov Chain Monte Carlo (MCMC) to generate pseudorandom samples from this posterior. We explain the details of this scheme in Computational Inference.

Model inference via HMJPs

HMJP apply directly on the formulation of eq:measure and, unlike with HMM (see Eq. 10), no approximations are required on the system-switching kinetics. Therefore, to proceed with inference, we need only provide appropriate prior distributions on the parameters, namely ρ¯,π¯¯,λ¯,μ¯.

We start with the prior distribution for the escape rates λ¯=(λσ1,λσ2,,λσK). We put priors on each of the λσk for all k=1,2,,K. The prior we select is

λσkGammaη,bη, (32)

for all k=1,2,,K with hyperparameters η,b. We note that this prior is conjugate to the exponential distribution given in Eq. 6. Next, we place a prior on π¯σk for all k=1,2,,K. For this, we place independent conjugate Dirichlet priors with concentration parameter A such that πσkσk=0 holds for all k=1,2,,K

πσ1Dirichlet0,AK1,,AK1, (33)
πσ2DirichletAK1,0,,AK1, (34)
πσKDirichletAK1,AK1,,0. (35)

Finally, on ρ¯ and μ¯, we place the same prior distributions as in Eqs. 27 and 28, respectively.

Once the choices for the priors are made, we then form the posterior distribution

P(ρ¯,π¯¯,λ¯,μ¯,T()|x)=P(ρ¯,π¯¯,λ¯,μ¯,(S,D,M)|x), (36)

containing all unknown variables that we wish to learn. This posterior also can be expanded in a way that is proportional to the product of the likelihood introduced in Eq. 4 and the priors introduced in Eqs. 5, 6, 7, 28, 27, 34, and 36. More explicitly, the form for this posterior distribution is as follows:

P(ρ¯,π¯¯,λ¯,μ¯,(S,D,M)|x)P(x|ρ¯,π¯¯,λ¯,μ¯,(S,D,M))P(ρ¯,π¯¯,λ¯,μ¯,(S,D,M)), (37)
P(x|s0:M1,d0:M1,μ¯)m=0M1P(dm|sm,λ¯)m=0M1P(sm+1|sm,π¯¯)P(s0|ρ¯)P(π¯¯)P(ρ¯)P(λ¯)P(μ¯), (38)
n=1NNormal(xn;1τm=0M1dmμsm,v)m=0M1P(dm|sm,λ¯)m=0M1P(sm+1|sm,π¯¯)P(s0|ρ¯)P(π¯¯)P(ρ¯)P(λ¯)P(μ¯).

However, once more, the posterior distribution does not attain an analytical form. Therefore, we develop a specialized computational scheme exploiting MCMC.

Computational inference

We carry out the analyses, shown in the Results section, evaluating the associated posteriors with an MCMC scheme (95) relying on Gibbs sampling (35,39, 40, 41, 42, 43, 44,63). The overall sampling strategy, for either the HMM or the HMJP, is as follows:

  • 1)

    update the trajectory T(), that is, c for HMM or (S,D,M) for HMJP;

  • 2)

    update the kinetics, that is,P¯σk for HMM or π¯σk and λσk for HMJP;

  • 3)

    update the initial probabilities ρ¯; and

  • 4)

    update state levels μ¯.

We repeat these updates to obtain a large number of samples. The end result is a sampling of the posterior P(ρ¯,P¯¯,μ¯,c|x) for the HMM and P(ρ¯,π¯¯,λ¯,μ¯,(S,D,M)|x) for the HMJP. Each of the conditional probabilities used in the Gibbs sampling scheme that we have defined above is as follows: for step one, we use P(c|x,ρ¯,P¯¯,μ¯) and P((S,D,M)|x,ρ¯,π¯¯,λ¯,μ¯) for the trajectory; for step two, we use P(P¯k|x,c,μ¯) and P(π¯k|x,(S,D,M),λ¯,μ¯) for the transition probabilities; for step three, we use P(λ¯σk|π¯¯,x,(S,D,M),μ¯) for the switching rates for all k=1,2,,K; for step four, we use P(μσk|μσk,c,x,π¯¯,ρ¯) and P(μσk|μσk,(S,D,M),x,π¯¯,λ¯,ρ¯) for the state levels for all k,k=1,2,,K and kk in HMM and HMJP frameworks. The formulae for each conditional probability distribution are expanded in Appendix A.2.3 in the Supporting Materials and Methods. Both samplings can be used to estimate switching rates λσkσk (for example, HMJP by Eq. 2 and HMM by Eq. 19).

Finally, as can be seen, Gibbs sampling for both HMM and HMJPs requires sampling of the corresponding trajectories. This is achieved by means of a forward-filtering backward-sampling algorithm in the HMM and by means of uniformization in HMJPs. The former is well known (3,58, 59, 60,96) and widely applied in biophysical applications (10,11,39, 40, 41,45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57); however, the latter has been only recently achieved (63) and, to this day, it has never been used to solve in biophysics. Because of their importance to this study, we provide a detailed description of both procedures in Supporting Materials and Methods Section A. Here, we emphasize that uniformization is only a computational tool used in Gibbs sampling scheme developed for HMJPs that has nothing to do with the HMM. In the Supporting Materials and Methods Section A.2, we also provide a thorough description of all steps and a working code through the authors’ website.

Results

To demonstrate how HMJPs work and highlight their advantages over HMMs, in this section, we use synthetic data that mimic a single-molecule experiment. Synthetic data are ideal for this purpose because they allow us to benchmark the results against the exact, readily available, “ground truth.” We obtain such data from the Gillespie algorithm described in Simulation and we explain our simulation choices below.

We focus on two data sets: one where the system exhibits slow kinetics and another where the system exhibits fast kinetics as compared with data acquisition, see Fig. 1. We provide the values for the hyperparameters in all analyses, as well as any other choices made, in Supporting Materials and Methods Section A.3. To be clear, we only assume to have access to the data, i.e., the gray dashes of Fig. 1, a1 and b1. The cyan (ground truth) trajectories are assumed unknown and to be determined.

In our results, we first benchmark the HMJP on the easy (i.e., slow kinetics) case shown in Fig. 1 a1; see Fig. 3 and 4. This is the regime where the HMM also works well and the expected (good) results for the HMM are relegated to the appendix; see Supporting Materials and Methods Section A.1. Next, we turn to the more complex case of fast kinetics. A sample time trace is shown in Fig. 1 b1. The results for both the HMJP and HMM are shown in Figs. 5 and 6. Afterwards, in Figs. 7 and 8, we demonstrate the effect of proportion of integration period to the total data acquisition period, called the “duty cycle,” on the HMJP’s performance for estimating state levels (see Fig. 7) and switching kinetics (see Fig. 8) in the case of fast kinetics provided in Fig. 1 b1. Further results demonstrating the effect of the duty cycle size on the HMJP for faster kinetics are provided in Appendix A.1.1 in the Supporting Materials and Methods. in addition, we present the performance of the HMJP for various detector FWHMs in estimating state levels and switching rates for fast kinetics in Appendix A.1.2 in the Supporting Materials and Methods.

Figure 3.

Figure 3

HMJP trajectory estimates for slow state switching. Here, we provide trajectory estimates obtained with the HMJP when the switching rate is slower than the data acquisition rate, 1/Δt=10 (1/s). In this figure’s (a1), the measurements are shown as gray rectangles (the width of the rectangle coincides with the integration period as shown in Fig. 1) generated based on the description provided in Model Description. We superposed the true trajectory (cyan) with the measurements in (a). Next, in (a2), we provide the histogram of all measurements to visualize the system kinetics. For illustrative purposes, we only show the MAP estimates of the HMJP on a zoomed-in region of (a1). Next, we provide that region of the (a1) in (b). In (b), we show the the MAP trajectory estimates of HMJP (magenta) that are superposed with the measurements and the true trajectory (cyan). For visual purposes only, we offset the HMJP MAP trajectory estimate by slightly shifting it downward. We observe that the HMJP MAP trajectory is able to capture switching occurring roughly in the middle of the integration time. This is not something that the HMM can capture. Here, simulated measurements are generated with λσ1σ2,λσ2σ1Eq. 40 where the data acquisition happens at every Δt=0.1 s with τf=0.8 s and τ=0.09 s starting at t0=0.05 s until tN=20 s. To see this figure in color, go online.

Figure 4.

Figure 4

HMJP state level and rate estimates for slow state switching. Here, we provide posterior state level and rate estimates obtained with HMJP whose time trace we discussed in Fig. 3. We expect HMJPs to perform well in estimating the true state levels and rates when these rates are slower than the data acquisition rate. In all figure panels, we superposed the posterior distributions over state levels and rates for HMJP (blue) along with their 95% confidence intervals, the true state levels (dashed green lines) and the corresponding prior distributions (magenta lines). Here, we emphasize that the posterior distributions over each parameter are obtained based on the Gibbs sampling scheme by drawing samples from the full posterior distribution. In all panels of this figure, we histogram the sampled values for each parameter irrespective of all other parameters. As such, we can call these posterior distributions, over each parameters, marginals. We start with the information in (a1) and (a2). We observe in these panels that the HMJP posterior distributions over state levels contain the true state levels within their 95% confidence intervals. Next, we move to the (b1) and (b2), which show the posterior distributions over the rates labeled λσkσk for all k,k=1,2. Again, the HMJP does quite well in estimating these rates as measured by the fact that the ground truth lies within the 95% confidence intervals of the posteriors. In this figure, the analyzed simulated measurements are generated with the same parameters as those provided in Fig. 3. To see this figure in color, go online.

Figure 5.

Figure 5

HMJP with HMM trajectory estimates for fast state switching. Here, we provide trajectory estimates obtained with HMJP and HMM when the switching rate is faster than the data acquisition rate, 1/Δt=10 (1/s). We expect HMMs to perform poorly in estimating the true trajectory when switching is fast. In this figure’s (a1), the measurements are shown with gray rectangles (the width of the rectangle coincides with the integration period as shown in Fig. 1) that are generated based on the description provided in Model Description. We follow the same color scheme and layout as in Fig. 3 except for (a4), where we provide the MAP trajectory estimate provided by the HMJP as well as the HMM. In (a4), the magenta dashed line shows the HMJP MAP trajectory estimate and the blue line shows the HMM MAP trajectory estimate. For visual purposes, we offset the HMJP MAP trajectory estimate and HMM MAP trajectory estimate by shifting these downward. Here, simulated measurements are generated with λσ1σ2,λσ2σ1 (see Eq. 40), where the data acquisition happens at every Δt=0.1 s with τf=1/15 s and τ=0.09 s starting at t0=0.05 s until tN=20 s. To see this figure in color, go online.

Figure 6.

Figure 6

HMJP with HMM state level and transition probability estimates for fast state switching. Here, we provide posterior state level and transition probability estimates obtained with HMJP and HMM when the switching rate is faster than the data acquisition rate, 1/Δt=10 (1/s). We expect HMMs to perform poorly in estimating the true state levels and transition probabilities when the system switching is fast. In all of this figure’s panels, we superposed the posterior distributions over state levels for both HMJP (blue) and HMM (orange) along with their 95% confidence intervals, the true state levels and true transition probabilities (dashed green lines). Next, we move to transition probability estimates provided in (b1)–(b4). In these panels, we wish to test the performance of HMJPs and HMMs in estimating the transition probabilities. In this figure, each panel is corresponding to the posterior distribution of a transition probability labeled as Pσkσk for all k,k=1,2. Here, simulated measurements are generated with the same parameters as those provided in Fig. 5. To see this figure in color, go online.

Figure 7.

Figure 7

HMJP state level estimates for fast state switching with different duty cycles. Here, we present how the duty cycle affects the HMJP state level estimates for fast kinetics. The parameters for fast dynamics are the same as in Figs. 5 and 6. Here, we simulated four data sets with kinetics presented in Figs. 5 and 6 for four different specified duty cycles. These are 90, 50, 5, and 1%. For clarity, a 90% duty cycle represents 90% integration period of the total data acquisition period. In the data set with 90% duty cycle, what we mean is having 0.09-s-long integration period with 0.01-s dead time period of the detector where the total data acquisition period is Δt=0.1 s. In each of this figure’s panels, we have state level histograms (blue) estimated from HMJPs superposed with their 95% confidence intervals (light blue), true state levels (dashed green lines) and prior distributions (magenta line). We emphasize that we use normal prior distributions and the mean of the distribution is set by the data. The FWHM is 0.75 au. In the top panels (a1)–(d1), we have state level estimates for the first physical state with the HMJP. The bottom panels, (a2)–(d2), illustrate the state level estimates of the HMJP for the second physical state. Here, we observe that shorter integration periods lead to sharper state level estimates. To see this figure in color, go online.

Figure 8.

Figure 8

HMJP switching rate estimates for fast state switching with different duty cycles. Here, we show how the duty cycle affects the HMJP switching rate estimates for fast kinetics. The parameters for fast dynamics are the same as in Figs. 5 and 6. In this figure, we have the same color pattern in each panel for the estimated quantities as presented in Fig. 7. Also, we consider the same duty cycles as in Fig. 7. In the top panels, (a1)–(d1), we have HMJP switching rate estimates for the first physical state that is λσ1. In the bottom panels, (a2)–(d2), we show the HMJP λσ2 estimates. In each panel, we have HMJP switching rate estimates superposed with their 95% confidence intervals and their true values. Here, we demonstrate that longer integration periods give rise to sharper switching rate estimates, namely the uncertainty decreases as the integration period approaches the total data acquisition period. To see this figure in color, go online.

Data simulation

To simulate the synthetic data, we assumed K=2 distinct states, such as on/off or folded/unfolded states for illustrative purposes only. We assumed well-separated state levels, which we set at μσ1=1 au and μσ2=7 au where “au” denote arbitrary units. The prescribed detector FWHM was set at 0.75 au.

Additionally, for sake of concreteness only, we assumed an acquisition period of Δt=0.1 s and consider long integration periods by setting τ equal to 90% of Δt. In terms familiar to microscopists, our setting corresponds to a frame rate of 10 Hz with exposure time of 90 ms and a dead time of 10 ms (41). The onset and concluding time of the experiment are the same for all simulated measurements and set at t0=0.05 s and tN=20 s, respectively.

To specify kinetics, we use the following structure for the switching rates λσ1σ2,λσ2σ1, with a parameter τf which sets the timescale of the system kinetics,

λσ1σ2=1.1τf,λσ2σ1=1.6τf. (40)

We simulate a case with τf=0.8 s, which involves system kinetics that are slower than the data acquisition rate and a case with τf=0.067 s, which involves system kinetics that are faster than the data acquisition rate.

Analysis with HMJPs

As a benchmark, we provide the results for the HMJP for those measurements shown on Fig. 1 a1 associated with slow switching rates. These results include estimates of the trajectory (S,D,M) (see Fig. 3), state levels μ¯ (see Fig. 4), and the switching rates λ¯ (see Fig. 4). To obtain these estimates, we generate samples from the posterior distribution P(ρ¯,π¯¯,λ¯,μ¯,(S,D,M)|x) with the HMJP sampler of Computational Inference.

In Fig. 3 a1, the ground truth trajectory is shown in cyan, whereas the measurements are shown in gray. We showed the zoomed trajectory and observations in Fig. 3 b. We also provide the empirical histogram of the observation in Fig. 3 a2, highlighting the slow switching rates of the system. After determining the posteriors over the trajectories with HMJPs, for illustrative purposes, we only show the maximal a posteriori (MAP) trajectory in Fig. 3 b. We observe that the HMJP MAP trajectory (magenta) captures most of the fast switches, shown in Fig. 3 b, in the system trajectory. In Fig. 4, there are four panels. In these four panels, we provide the superposed posterior distributions over the two state levels and two rates estimated by the HMJP along with its associated 95% confidence interval (sometimes called a credible interval in Bayesian analysis) and ground truths (dashed green lines).

In summary, HMJP performs well on this benchmark data. The same is true of the simpler HMM (as would be expected) whose results are shown in Figs. S7 and S8. An important bring-home message for the HMJP, however, is the fact that even if state transitions occur midway through the integration time, the HMJP can discern when these occurred. The same is not true of the HMM that, as mentioned earlier, assumes by construction that state transitions must occur at the end of the data acquisition period.

Comparison of HMJPs with HMMs

We now present a comparison of HMJPs and HMMs on the analysis of the simulated measurements shown in Fig. 1 b1. We expect the HMJP to outperform the HMM as we are now operating in a regime, with switching rates 2.5 times faster than earlier, where the HMM requirement spelled out in Eq. 10 breaks down.

We used these measurements to estimate the posterior distribution over the trajectory T(), initial and switching probabilities ρ¯ and π¯¯ or P¯¯, state levels μ¯, and escape rates, λ¯. To accomplish this, we generate samples from the posterior distributions P(ρ¯,π¯¯,λ¯,μ¯,(S,D,M)|x) and P(ρ¯,P¯¯,μ¯,c|x) using the HMJP and HMM samplers of Computational Inference, respectively.

Following the pattern from the previous section on slow kinetics, we first show the trajectories inferred by HMJPs and HMMs in Fig. 5, then we show estimates of the state levels and transition probabilities in Fig. 6. For clarity, to compare apples to apples, as HMMs infer transition probabilities but not rates, we compared the transition probabilities computed from the rates obtained by HMJPs to the transition probabilities inferred by HMMs. Here, we emphasize that we obtain a unique transition probability matrix from the rate matrix; however, because of the multivalued nature of the logarithm function, the rate matrix cannot be “uniquely” inferred from the transition probability matrix.

In particular, the escape rates estimated from HMJPs are used in Eq. 19 to yield transition probabilities that we subsequently compared with the transition probabilities inferred by HMMs.

Predictably, the HMM performs poorly. For example, we see in Fig. 5 a4 that the HMJP MAP trajectory captures many of the fast switches occurring during integration times. The HMM MAP trajectory is severely constrained to allowing switches at the end of the time period and, as such, cannot accommodate fast kinetics. Although the trajectory inferred by HMJPs is not perfect, this ability to tease out many correct state switches in its MAP trajectory is sufficient for HMJPs to obtain estimates of the transition probabilities and state levels that lie within the 95% confidence interval; see panels Fig. 6, a1, a2, and b1b4. The same does not hold for HMMs where their inability to detect state switches now percolates down to the quality of their estimates for the state levels and transition probabilities. To wit, from panels Fig. 6, a1 and a2 we see that the HMM grossly overestimates (by about 90%) μσ1 and underestimates (about 30%) μσ2. What is more, as can be seen in Fig. 6, b1b4, the HMM provides very wide posterior distributions over transition probabilities. This is in contrast to the much sharper posterior of the HMJP whose mode closely coincides with the ground truth; see Fig. 6, b1b4.

An observation is warranted here. Because the HMM cannot accommodate fast kinetics, it must ascribe the apparent spread around the Pσkσk histogram (see Fig. 6, b1b4) to an increased variance in the posterior distribution of transition probabilities. So, although the breadth of the posteriors of the HMJP are primarily ascribed to the fact that finite data inform the posterior, the origin of the breadth of the histogram of the HMM is an artifact of its inability to accommodate fast kinetics.

Later, in Figs. 7 and 8, we simulated three more data sets with fast dynamics as in Eq. 40 with τf=0.067 s with duty cycles that are 50, 5, and 1% with integration period Δt=0.1 s. We emphasize that in Fig. 4, we presented the 90% duty cycle case. In Figs. 7 and 8, we show the posterior estimates for the HMJP state levels (see Fig. 7) and switching rates (see Fig. 8). In these figures, we see that the HMJP posterior estimates are centered around the ground truth values for both state levels and switching rates. However, posterior distributions for state level estimates of HMJP get narrower as in the case of HMM for smaller duty cycles (see Fig. 7, a1d1 and a2d2). Namely, because of the measurement model, as the duty cycle gets shorter and the region over which kinetics can be learned (the integration time) of the HMJP concomitantly shrinks. This smaller integration time results in narrower state level posterior estimates. In addition, the posterior distributions for switching rate estimates of HMJP get wider for shorter duty cycles (see Fig. 8, a1d1 and a2d2). This result can be attributed to the fact that as the duty cycle gets shorter then the integration period does not provide extra information about the kinetics of the physical system. This leads a highly varying time grid for the dynamics of the physical system. Therefore, the HMJP ends up providing wider posterior-switching rate estimates. In summary, longer duty cycle gives rise to wider posterior state level estimates and narrower posterior-switching rate estimates of HMJP.

This agreement is not surprising. For instance, as τ0, we find that 1τtnτtndtμT(t)μT(tn)=μcn and so Eq. 4, used in the HMJP, reduces to Eq. 11, used in the HMM. This provides the analytical proof that the HMJP “measurement model” simplifies to the measurement model in HMM framework for very small integration time τ.

In the appendix, we first probe the effect of duty cycles (90, 50, 5, and 1% where the integration period is Δt=0.1 s) on the posterior state level and switching rate estimates of the HMJP for fast kinetics set by Eq. 40 with τf=0.04 s (see Figs. S1 and S2). We see that the HMJP provides poor posterior-switching rate estimates as the duty cycle decreases in terms of the increased width of the switching rate posterior distributions. However, HMJP’s posterior-switching rate and state level estimates are still centered around the ground truths unlike what HMM can do, as presented in Fig. 6. That is attributed to the random time grid used in HMJP framework unlike what is inherited from the HMM.

We subsequently investigated the model selection problem in Figs. S5 and S6 by analyzing two state fast dynamics with HMJP as though we had a three-state system. We find that our framework can identify the redundant state. This is observed based on the middle histogram of the diagonal panels of Fig. S6 in which we see that the final posterior distribution revert to the prior distribution. This tells us that the data do not warrant a third state. Next, we looked into the effect of measurement noise on the HMJP posterior state level and -switching rate estimates by analyzing three more data sets with different detector FWHM values (0.75, 1.1, and 1.3 au). We found that the posterior distribution widths for state level and switching rates increased though the HMJP estimates remain robust with respect to various FWHM values.

Afterwards, we revisited finer details of the effect of finiteness of data on the HMM posterior distributions over transition probabilities for fast switching rates in Supporting Materials and Methods Section A; see Fig. S16. In particular, we analyzed a sequence of three data sets using the HMM framework with the same fast switching rates as in Fig. 5 a1 but with differing data set lengths. The amount of data in Fig. S17 beyond∼400 data points (where the data acquisition period is Δt=0.1 s) seems to have a limited effect on the posterior distribution over state levels. Further analysis on the performance of both the HMJP and the HMM is relegated to the Supporting Materials and Methods Section A, in particular, cases in which the rate from one state to another is fast and the other is slow. We also provide HMJP rate estimates in Fig. S15, a1 and a2 for the data set given in Fig. 1 b1 as well as a comparison of the HMJP posterior transition probability estimates associated with the data provided in Fig. 1 a1 with and without learning the trajectory T() simultaneously in Fig. S18. Finally, we compare the posterior trajectory estimates of the HMJP and the HMM based on a metric that is the enclosed area under the learned trajectories in Fig. S19.

Discussion

HMMs have been a hallmark of time series analysis in single-molecule biophysics (10,11,39, 40, 41,45,46,49, 50, 51, 52, 53, 54, 55, 56, 57), but they have a critical limitation: HMMs apply only provided the temporal resolution of the experimental apparatus is faster than the system kinetics under study (32,97, 98, 99). Otherwise, HMMs mistakenly ascribe the signal generated by fast dynamics to misassignments of signal levels. Fundamentally, this limitation arises because HMM detection models link the measurements with the “instantaneous” state of the system that dynamically evolves. Of course, the HMM framework holds provided that measurements are not obtained with an integrative detector as in the case of stroboscopic fluorescence microcopy experiments (100,101) and fast shuttering systems (102).

By contrast, the HMJP we describe here can deal with rapid dynamics and integrative detectors. This is because the HMJP is the continuous-time analog of the HMM. As compared to the HMM, the main novelty underlying the HMJP framework is primarily in the emission model, which accounts for realistic detectors operating in integrative rather than counting mode. Such detectors are common to modern biophysical experiments (35,103, 104, 105).

Other methods have also attempted to tackle the challenge presented by fast dynamics. One such example is the H2MM, although it tackles data derived from a different type of experiment as the HMJP. In particular, the H2MM assumes the data are available as single-photon trajectories while we focus on the fundamental challenge of unraveling processes on timescales faster than those of detectors with finite exposure time.

We briefly highlight two more examples, although these differ from ours in that they hold for very specific cases. For example, (20) provides a method to extract fast kinetics obscured by the detector integration time. However, the analysis differs from what we introduce in our method in two ways: 1) (20) introduces a method to deal with physical systems that only have two conformational states; therefore, the physical system either populates one or the other conformational state, and 2) the method in (20) operates within the Bayesian framework; however, it requires the marginalization of fractional occupancies of the single molecule. Thereby, the framework presented in (20) does not provide a posterior distribution over the entire system trajectory that can be simultaneously estimated alongside the kinetic rates and model parameters. On the other hand, with HMJPs, we can provide a recipe to sample from a full joint posterior over all unknowns simultaneously and self-consistently following Bayesian paradigm. These unknowns include kinetic parameters, model parameters as well as the single-molecule trajectories modeled as MJPs.

As a second and final example, Lee (10) demonstrates a way to address the problem of extracting kinetic information when there is asynchronous switching (namely the conformational state of the physical state does not change at the same time as the time of data acquisition, instead it happens during the integration period) from one conformational state to the other or alternatively when one conformational state is short lived. Here, the main objective was not to extract kinetics faster than the data acquisition rate (and for this reason HMMs were employed therein). Rather, the main goal was to improve the estimation of switching rates based on transition probabilities within HMM framework. As such, their methodology relied on fitting mixtures of Gaussians until fits to FRET efficiency histograms satisfied a predetermined optimality criterion.

The HMJP does have limitations. In the limit that state switching rate grows, the amount of data needed to ascertain a meaningful posterior over the transition kinetics also grows. In the trivial limit that the state switching is extremely fast, no method, whether HMJP or otherwise, would be able to tease out information on the transition kinetics from what appears as a noisy but otherwise horizontal time trace with no discernible transitions. Although, the quality of the data is not a fundamental limitation for the HMJP, it is clear that the duration of the detector dead time affects the performance of all methods of inference. Specifically, the longer the dead time, the worse the HMJP will perform. In particular, when the dead time is as long as the exposure itself, the HMJP reduces to the HMM. A deeper question relates to whether, at fast enough timescales, it even makes sense to speak of discrete states and whether we should be speaking of continuous space and time. At the moment, these questions lie beyond the scope of this study.

Of equal interest, within the discrete state space paradigm, is the possibility to learn the number of states within an HMJP framework. That is, to repitch the HMJP within a Bayesian nonparametric paradigm following the footsteps of the HMM and its nonparametric realization, the infinite HMM (38, 39, 40, 41,106, 107, 108). Methods have been developed to report on point statistics as they pertain to infinite MJPs (70,109). A natural extension for us would be to propose a way to construct and sample from a joint posterior over all unknowns already discussed in this study as well as the state number.

Author Contributions

Z.K. analyzed data and developed analysis tools. Z.K. and I.S. developed computational tools. Z.K., I.S., and S.P. conceived research. S.P. oversaw all aspects of the projects.

Acknowledgments

Z.K. thanks Sina Jazani for his helpful suggestions on the manuscript.

S.P. acknowledges support from NSF CAREER grant MCB-1719537 and NIH NIGMS (R01GM134426). ASU cluster AGAVE and Saguaro are the main computational resources utilized in this study.

Editor: Anatoly Kolomeisky.

Footnotes

Supporting Material can be found online at https://doi.org/10.1016/j.bpj.2020.12.022.

Supporting Material

Document S1. Supporting Materials and Methods, Figs S1–S21, and Tables S1–S2
mmc1.pdf (5MB, pdf)
Document S2. Article plus Supporting Material
mmc2.pdf (6.1MB, pdf)

References

  • 1.Baum L.E., Petrie T. Statistical inference for probabilistic functions of finite state markov chains. Ann. Math Stat. 1966;37:1554–1563. [Google Scholar]
  • 2.Petrie T. Probabilistic functions of finite state Markov chains. Ann. Math. Stat. 1969;40:97–115. doi: 10.1073/pnas.57.3.580. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Rabiner L.R. A tutorial on hidden markov models and selected applications in speech recognition. Proc. IEEE. 1989;77:257–286. [Google Scholar]
  • 4.Levinson S.E., Rabiner L.R., Sondhi M.M. An introduction to the application of the theory of probabilistic functions of a Markov process to automatic speech recognition. Bell Syst. Tech. J. 1983;62:1035–1074. [Google Scholar]
  • 5.Kelly D., Dillingham M., Wiesner K. A new method for inferring hidden markov models from noisy time sequences. PLoS one. 2012;7:e29703. doi: 10.1371/journal.pone.0029703. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Liu Y., Park J., Ha T. A comparative study of multivariate and univariate hidden Markov modelings in time-binned single-molecule FRET data analysis. J. Phys. Chem. B. 2010;114:5386–5403. doi: 10.1021/jp9057669. [DOI] [PubMed] [Google Scholar]
  • 7.Zarrabi N., Ernst S., Börsch M. Analyzing conformational dynamics of single P-glycoprotein transporters by Förster resonance energy transfer using hidden Markov models. Methods. 2014;66:168–179. doi: 10.1016/j.ymeth.2013.07.026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Smith D.A., Steffen W., Sleep J. Hidden-Markov methods for the analysis of single-molecule actomyosin displacement data: the variance-Hidden-Markov method. Biophys. J. 2001;81:2795–2816. doi: 10.1016/S0006-3495(01)75922-X. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Zarrabi N., Düser M., Börsch M. Detecting substeps in the rotary motors of fof1-atp synthase by hidden markov models. In: Enderlein J., Gryczynski Z., editors. Ultrasensitive and Single-Molecule Detection Technologies II. SPIE Proceedings; 2007. p. 64440E. [Google Scholar]
  • 10.Lee T.-H. Extracting kinetics information from single-molecule fluorescence resonance energy transfer data using hidden markov models. J. Phys. Chem. B. 2009;113:11535–11542. doi: 10.1021/jp903831z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.McKinney S.A., Joo C., Ha T. Analysis of single-molecule FRET trajectories using hidden Markov modeling. Biophys. J. 2006;91:1941–1951. doi: 10.1529/biophysj.106.082487. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Keller B.G., Kobitski A., Noé F. Complex RNA folding kinetics revealed by single-molecule FRET and hidden Markov models. J. Am. Chem. Soc. 2014;136:4534–4543. doi: 10.1021/ja4098719. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Andrec M., Levy R.M., Talaga D.S. Direct determination of kinetic rates from single-molecule photon arrival trajectories using hidden markov models. J. Phys. Chem. A. 2003;107:7454–7464. doi: 10.1021/jp035514+. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Tavakoli M., Jazani S., Pressé S. Pitching single-focus confocal data analysis one photon at a time with bayesian nonparametrics. Phys. Rev. X. 2020;10:011021. doi: 10.1103/physrevx.10.011021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Baumann G., Easton G.S. Modeling state-dependent sodium conductance data by a memoryless random process. Math. Biosci. 1982;60:265–276. [Google Scholar]
  • 16.Edeson R.O., Yeo G.F., Madsen B.W. Graphs, random sums, and sojourn time distributions, with application to ion-channel modeling. Math. Biosci. 1990;102:75–104. doi: 10.1016/0025-5564(90)90056-5. [DOI] [PubMed] [Google Scholar]
  • 17.Yeo G.F., Milne R.K., Madsen B.W. Statistical inference from single channel records: two-state Markov model with limited time resolution. Proc. R. Soc. Lond. B Biol. Sci. 1988;235:63–94. doi: 10.1098/rspb.1988.0063. [DOI] [PubMed] [Google Scholar]
  • 18.Milne R.K., Yeo G.F., Edeson R.O. Estimation of single channel kinetic parameters from data subject to limited time resolution. Biophys. J. 1989;55:673–676. doi: 10.1016/S0006-3495(89)82865-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Milne R.K., Yeo G.F., Madsen B.W. Stochastic modelling of a single ion channel: an alternating renewal approach with application to limited time resolution. Proc. R. Soc. Lond. B Biol. Sci. 1988;233:247–292. doi: 10.1098/rspb.1988.0022. [DOI] [PubMed] [Google Scholar]
  • 20.Kinz-Thompson C.D., Gonzalez R.L., Jr. Increasing the time resolution of single-molecule experiments with bayesian inference. Biophys. J. 2018;114:289–300. doi: 10.1016/j.bpj.2017.11.3741. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Bronson J.E., Fei J., Wiggins C.H. Learning rates and states from biophysical time series: a Bayesian approach to model selection and single-molecule FRET data. Biophys. J. 2009;97:3196–3205. doi: 10.1016/j.bpj.2009.09.031. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Lu M., Ma X., Mothes W. Associating HIV-1 envelope glycoprotein structures with states on the virus observed by smFRET. Nature. 2019;568:415–419. doi: 10.1038/s41586-019-1101-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Mazouchi A., Zhang Z., Gradinaru C.C. Conformations of a metastable sh3 domain characterized by smfret and an excluded-volume polymer model. Biophys. J. 2016;110:1510–1522. doi: 10.1016/j.bpj.2016.02.033. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Roy R., Hohng S., Ha T. A practical guide to single-molecule FRET. Nat. Methods. 2008;5:507–516. doi: 10.1038/nmeth.1208. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Zhu Y., He L., Zhang X.C. smfret probing reveals substrate-dependent conformational dynamics of e. coli multidrug mdfa. Biophys. J. 2019;116:2296–2303. doi: 10.1016/j.bpj.2019.04.034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Lerner E., Cordes T., Weiss S. Toward dynamic structural biology: two decades of single-molecule Förster resonance energy transfer. Science. 2018;359:eaan1133. doi: 10.1126/science.aan1133. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Krishnamoorti A., Cheng R.C., Maduke M. Clc conformational landscape as studied by smfret. Biophys. J. 2019;116:555a. [Google Scholar]
  • 28.Hohng S., Zhou R., Ha T. Fluorescence-force spectroscopy maps two-dimensional reaction landscape of the holliday junction. Science. 2007;318:279–283. doi: 10.1126/science.1146113. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Schuler B. Single-molecule FRET of protein structure and dynamics - a primer. J. Nanobiotechnology. 2013;11(Suppl 1):S2. doi: 10.1186/1477-3155-11-S1-S2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Kruithof M., van Noort J. Hidden Markov analysis of nucleosome unwrapping under force. Biophys. J. 2009;96:3708–3715. doi: 10.1016/j.bpj.2009.01.048. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Elms P.J., Chodera J.D., Marqusee S. Limitations of constant-force-feedback experiments. Biophys. J. 2012;103:1490–1499. doi: 10.1016/j.bpj.2012.06.051. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Zhang Y., Jiao J., Rebane A.A. Hidden markov modeling with detailed balance and its application to single protein folding. Biophys. J. 2016;111:2110–2124. doi: 10.1016/j.bpj.2016.09.045. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Ball F.G., Rice J.A. Stochastic models for ion channels: introduction and bibliography. Math. Biosci. 1992;112:189–206. doi: 10.1016/0025-5564(92)90023-p. [DOI] [PubMed] [Google Scholar]
  • 34.Lee A., Tsekouras K., Pressé S. Unraveling the thousand word picture: an introduction to super-resolution data analysis. Chem. Rev. 2017;117:7276–7330. doi: 10.1021/acs.chemrev.6b00729. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Tavakoli M., Taylor J.N., Pressé S. 2016. Single molecule data analysis: an introduction. arXiv, arXiv:1606.00403.https://arxiv.org/abs/1606.00403 [Google Scholar]
  • 36.Van Kampen N.G. vol. 1. Elsevier; Amsterdam, the Netherlands: 1992. Stochastic Processes in Physics and Chemistry. [Google Scholar]
  • 37.Levitus M., Ranjit S. Cyanine dyes in biophysical research: the photophysics of polymethine fluorescent dyes in biomolecular environments. Q. Rev. Biophys. 2011;44:123–151. doi: 10.1017/S0033583510000247. [DOI] [PubMed] [Google Scholar]
  • 38.Van Gael J., Saatci Y., Ghahramani Z. Proceedings of the 25th International Conference on Machine Learning. (Association for Computing Machinery) 2008. Beam sampling for the infinite hidden Markov model; pp. 1088–1095. [Google Scholar]
  • 39.Sgouralis I., Pressé S. An introduction to infinite hmms for single-molecule data analysis. Biophys. J. 2017;112:2021–2029. doi: 10.1016/j.bpj.2017.04.027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Sgouralis I., Pressé S. Icon: an adaptation of infinite hmms for time traces with drift. Biophys. J. 2017;112:2117–2126. doi: 10.1016/j.bpj.2017.04.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Sgouralis I., Madaan S., Pressé S. A bayesian nonparametric approach to single molecule förster resonance energy transfer. J. Phys. Chem. B. 2019;123:675–688. doi: 10.1021/acs.jpcb.8b09752. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Jazani S., Sgouralis I., Pressé S. An alternative framework for fluorescence correlation spectroscopy. Nat. Commun. 2019;10:3662. doi: 10.1038/s41467-019-11574-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Jazani S., Sgouralis I., Pressé S. A method for single molecule tracking using a conventional single-focus confocal setup. J. Chem. Phys. 2019;150:114108. doi: 10.1063/1.5083869. [DOI] [PubMed] [Google Scholar]
  • 44.Sgouralis I., Whitmore M., Pressé S. Single molecule force spectroscopy at high data acquisition: a Bayesian nonparametric analysis. J. Chem. Phys. 2018;148:123320. doi: 10.1063/1.5008842. [DOI] [PubMed] [Google Scholar]
  • 45.Burzykowski T., Szubiakowski J.P., Ryden T. Vol. 5258. SPIE Proceedings; 2003. Statistical analysis of data from single molecule experiment. IV Workshop on Atomic and Molecular Physics; pp. 171–177. [Google Scholar]
  • 46.Pirchi M., Tsukanov R., Nir E. Photon-by-photon hidden markov model analysis for microsecond single-molecule fret kinetics. J. Phys. Chem. B. 2016;120:13065–13075. doi: 10.1021/acs.jpcb.6b10726. [DOI] [PubMed] [Google Scholar]
  • 47.Aviram H.Y., Pirchi M., Haran G. Direct observation of ultrafast large-scale dynamics of an enzyme under turnover conditions. Proc. Natl. Acad. Sci. USA. 2018;115:3243–3248. doi: 10.1073/pnas.1720448115. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Mazal H., Iljina M., Haran G. Tunable microsecond dynamics of an allosteric switch regulate the activity of a AAA+ disaggregation machine. Nat. Commun. 2019;10:1438. doi: 10.1038/s41467-019-09474-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Lee J., Lee S., Hohng S. Single-molecule four-color FRET. Angew. Chem. Int. Ed. Engl. 2010;49:9922–9925. doi: 10.1002/anie.201005402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Mazal H., Haran G. Single-molecule FRET methods to study the dynamics of proteins at work. Curr. Opin. Biomed. Eng. 2019;12:8–17. doi: 10.1016/j.cobme.2019.08.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Shen K., Arslan S., Shan S.O. Activated GTPase movement on an RNA scaffold drives co-translational protein targeting. Nature. 2012;492:271–275. doi: 10.1038/nature11726. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Joo C., McKinney S.A., Ha T. Real-time observation of RecA filament dynamics with single monomer resolution. Cell. 2006;126:515–527. doi: 10.1016/j.cell.2006.06.042. [DOI] [PubMed] [Google Scholar]
  • 53.Joo C., Balci H., Ha T. Advances in single-molecule fluorescence methods for molecular biology. Annu. Rev. Biochem. 2008;77:51–76. doi: 10.1146/annurev.biochem.77.070606.101543. [DOI] [PubMed] [Google Scholar]
  • 54.Cornish P.V., Ermolenko D.N., Ha T. Following movement of the L1 stalk between three functional states in single ribosomes. Proc. Natl. Acad. Sci. USA. 2009;106:2571–2576. doi: 10.1073/pnas.0813180106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Li H., Yang H. Statistical learning of discrete states in time series. J. Phys. Chem. B. 2019;123:689–701. doi: 10.1021/acs.jpcb.8b10561. [DOI] [PubMed] [Google Scholar]
  • 56.Gopich I.V. Likelihood functions for the analysis of single-molecule binned photon sequences. Chem. Phys. 2012;396:53–60. doi: 10.1016/j.chemphys.2011.06.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Chung H.S., Gopich I.V. Fast single-molecule FRET spectroscopy: theory and experiment. Phys. Chem. Chem. Phys. 2014;16:18644–18657. doi: 10.1039/c4cp02489c. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Bishop C.M. Springer; New York: 2006. Pattern Recognition and Machine Learning. [Google Scholar]
  • 59.Rabiner L., Juang B. Vol. 3. 1986. An introduction to hidden markov models. IEEE ASSP Magazine; pp. 4–16. [Google Scholar]
  • 60.Rabiner L.R. 1996. Multirate Digital Signal Processing. [Google Scholar]
  • 61.Frühwirth-Schnatter S. Data augmentation and dynamic linear models. J. Time Ser. Anal. 1994;15:183–202. [Google Scholar]
  • 62.Carter C.K., Kohn R. Markov chain Monte Carlo in conditionally Gaussian state space models. Biometrika. 1996;83:589–601. [Google Scholar]
  • 63.Rao V., Teh Y.W. Fast mcmc sampling for Markov jump processes and extensions. J. Mach. Learn. Res. 2013;14:3295–3320. [Google Scholar]
  • 64.Shuang B., Cooper D., Landes C.F. Fast step transition and state identification (stasi) for discrete single-molecule data analysis. J. Phys. Chem. Lett. 2014;5:3157–3161. doi: 10.1021/jz501435p. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Chen J., Poddar N.K., Landes C.F. Single-molecule FRET studies of HIV TAR-DNA hairpin unfolding dynamics. J. Phys. Chem. B. 2014;118:12130–12139. doi: 10.1021/jp507067p. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Koopmans W.J., Brehm A., van Noort J. Single-pair FRET microscopy reveals mononucleosome dynamics. J. Fluoresc. 2007;17:785–795. doi: 10.1007/s10895-007-0218-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Preus S., Noer S.L., Birkedal V. iSMS: single-molecule fret microscopy software. Nat. Methods. 2015;12:593–594. doi: 10.1038/nmeth.3435. [DOI] [PubMed] [Google Scholar]
  • 68.Metzner P., Dittmer E., Schütte C. Generator estimation of markov jump processes. J. Comput. Phys. 2007;227:353–375. doi: 10.1103/PhysRevE.76.066702. [DOI] [PubMed] [Google Scholar]
  • 69.Hobolth A., Stone E.A. Simulation from endpoint-conditioned, continuous-time Markov chains on a finite state space, with applications to molecular evolution. Ann. Appl. Stat. 2009;3:1204. doi: 10.1214/09-AOAS247. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Huggins J.H., Narasimhan K., Mansinghka V.K. Jump-means: small-variance asymptotics for markov jump processes. arXiv. 2015 https://arxiv.org/abs/1503.00332 arXiv:1503.00332. [Google Scholar]
  • 71.Zhang B., Rao V. Efficient parameter sampling for markov jump processes. J. Comput. Graph. Stat. 2020 [Google Scholar]
  • 72.Beentjes C.H.L., Baker R.E. Uniformization techniques for stochastic simulation of chemical reaction networks. J. Chem. Phys. 2019;150:154107. doi: 10.1063/1.5081043. [DOI] [PubMed] [Google Scholar]
  • 73.Van Dijk N.M. Uniformization for nonhomogeneous Markov chains. Oper. Res. Lett. 1992;12:283–291. [Google Scholar]
  • 74.Diener J.D., Sanders W.H., Ers Z. Computations with Markov Chains. Springer; 1994. Empirical comparison of uniformization methods for continuous-time markov chains; pp. 547–570. [Google Scholar]
  • 75.Van Moorsel A.P., Wolter K. Proceedings of the 12th European Simulation Multiconference on Simulation - Past, Present and Future. SCS Europe; 1998. Numerical solution of non-homogeneous markov processes through uniformization; pp. 710–717. [Google Scholar]
  • 76.Gillespie D.T. A general method for numerically simulating the stochastic time evolution of coupled chemical reactions. J. Comput. Phys. 1976;22:403–434. [Google Scholar]
  • 77.Gillespie D.T. Exact stochastic simulation of coupled chemical reactions. J. Phys. Chem. 1977;81:2340–2361. [Google Scholar]
  • 78.Sivia D., Skilling J. Oxford University Press; Oxford, UK: 2006. Data Analysis: A Bayesian Tutorial. [Google Scholar]
  • 79.Buelens B., Daas P., van den Brakel J. Statistics Netherlands; The Hague, the Netherlands: 2014. Selectivity of Big Data. [Google Scholar]
  • 80.Papoulis A., Pillai S.U. McGraw-Hill; New York: 2002. Probability, Random Variables, and Stochastic Processes. [Google Scholar]
  • 81.Robert C., Casella G. Springer Science & Business Media; Berlin, Germany: 2013. Monte Carlo Statistical Methods. [Google Scholar]
  • 82.Nir E., Michalet X., Weiss S. Shot-noise limited single-molecule FRET histograms: comparison between theory and experiments. J. Phys. Chem. B. 2006;110:22103–22124. doi: 10.1021/jp063483n. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Kalinin S., Sisamakis E., Seidel C.A. On the origin of broadening of single-molecule FRET efficiency distributions beyond shot noise limits. J. Phys. Chem. B. 2010;114:6197–6206. doi: 10.1021/jp100025v. [DOI] [PubMed] [Google Scholar]
  • 84.Mortensen K.I., Churchman L.S., Flyvbjerg H. Optimized localization analysis for single-molecule tracking and super-resolution microscopy. Nat. Methods. 2010;7:377–381. doi: 10.1038/nmeth.1447. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Charrière F., Colomb T., Depeursinge C. Shot-noise influence on the reconstructed phase image signal-to-noise ratio in digital holographic microscopy. Appl. Opt. 2006;45:7667–7673. doi: 10.1364/ao.45.007667. [DOI] [PubMed] [Google Scholar]
  • 86.Gross M., Goy P., Al-Koussa M. Shot-noise detection of ultrasound-tagged photons in ultrasound-modulated optical imaging. Opt. Lett. 2003;28:2482–2484. doi: 10.1364/ol.28.002482. [DOI] [PubMed] [Google Scholar]
  • 87.Huang F., Hartwich T.M., Bewersdorf J. Video-rate nanoscopy using sCMOS camera-specific single-molecule localization algorithms. Nat. Methods. 2013;10:653–658. doi: 10.1038/nmeth.2488. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Krishnaswami V., Van Noorden C.J., Hoebe R.A. Towards digital photon counting cameras for single-molecule optical nanoscopy. Opt. Nanoscopy. 2014;3:1. [Google Scholar]
  • 89.Lin Y., Long J.J., Moonan D.W. Quantifying and optimizing single-molecule switching nanoscopy at high speeds. PLoS one. 2015;10:e0128135. doi: 10.1371/journal.pone.0128135. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90.Little M.A. Oxford University Press; Oxford, UK: 2019. Machine Learning for Signal Processing: Data Science, Algorithms, and Computational Statistics. [Google Scholar]
  • 91.Hamilton B.A. John Wiley & Sons; Hoboken, NJ: 2015. The Field Guide to Data Science. [Google Scholar]
  • 92.Kilic Z., Sgouralis I., Pressé S. Rapid kinetics for smfret: a continuous time treatment. bioRxiv. 2020 doi: 10.1101/2020.08.28.267468. [DOI] [Google Scholar]
  • 93.Israel R.B., Rosenthal J.S., Wei J.Z. Finding generators for Markov chains via empirical transition matrices, with applications to credit ratings. Math. Finance. 2001;11:245–265. [Google Scholar]
  • 94.Lee E.T., Wang J. vol. 476. John Wiley & Sons; Hoboken, NJ: 2003. Statistical Methods for Survival Data Analysis. [Google Scholar]
  • 95.Camassa R., Kilic Z., McLaughlin R.M. On the symmetry properties of a random passive scalar with and without boundaries, and their connection between hot and cold states. Physica D. 2019;400:132124. [Google Scholar]
  • 96.Cappé O., Moulines E., Rydén T. Springer Science & Business Media; Berlin, Germany: 2006. Inference in Hidden Markov Models. [Google Scholar]
  • 97.van de Meent J.-W., Bronson J.E., Gonzalez R.L., Jr. Empirical Bayes methods enable advanced population-level analyses of single-molecule FRET experiments. Biophys. J. 2014;106:1327–1337. doi: 10.1016/j.bpj.2013.12.055. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98.Whitt W. Columbia University; New York: 2006. Continuous-Time Markov Chains. [Google Scholar]
  • 99.Kinz-Thompson C.D., Bailey N.A., Gonzalez R.L., Jr. Methods in Enzymology. vol. 581. Academic Press; 2016. Precisely and accurately inferring single-molecule rate constants; pp. 187–225. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 100.Flors C., Hotta J., Hofkens J. A stroboscopic approach for fast photoactivation-localization microscopy with Dronpa mutants. J. Am. Chem. Soc. 2007;129:13970–13977. doi: 10.1021/ja074704l. [DOI] [PubMed] [Google Scholar]
  • 101.Holton M.D., Silvestre O.R., Summers H.D. Stroboscopic fluorescence lifetime imaging. Opt. Express. 2009;17:5205–5216. doi: 10.1364/oe.17.005205. [DOI] [PubMed] [Google Scholar]
  • 102.Giannini J.P., York A.G., Shroff H. Anticipating, measuring, and minimizing MEMS mirror scan error to improve laser scanning microscopy’s speed and accuracy. PLoS One. 2017;12:e0185849. doi: 10.1371/journal.pone.0185849. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 103.Vartsky D., Mor I., Breskin A. Novel detectors for fast-neutron resonance radiography. Nucl. Instrum. Methods Phys. Res. A. 2010;623:603–605. [Google Scholar]
  • 104.Mikhaylov A., Pimashkin A., Spagnolo B. Neurohybrid memristive cmos-integrated systems for biosensors and neuroprosthetics. Front. Neurosci. 2020;14:358. doi: 10.3389/fnins.2020.00358. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 105.Chua L. Memristor-the missing circuit element. IEEE Trans. Circuit Theory. 1971;18:507–519. [Google Scholar]
  • 106.Beal M.J., Ghahramani Z., Rasmussen C.E. The infinite hidden Markov model. Adv. Neural Inf. Process. Syst. 2002:577–584. [Google Scholar]
  • 107.Nakano M., Le Roux J., Sagayama S. 2011 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA) IEEE; 2011. Bayesian nonparametric spectrogram modeling based on infinite factorial infinite hidden markov model; pp. 325–328. [Google Scholar]
  • 108.Maheu J.M., Yang Q. An infinite hidden markov model for short-term interest rates. J. Empir. Finance. 2016;38:202–220. [Google Scholar]
  • 109.Georgoulas A., Hillston J., Sanguinetti G. Unbiased Bayesian inference for population Markov jump processes via random truncations. Stat. Comput. 2017;27:991–1002. doi: 10.1007/s11222-016-9667-9. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Document S1. Supporting Materials and Methods, Figs S1–S21, and Tables S1–S2
mmc1.pdf (5MB, pdf)
Document S2. Article plus Supporting Material
mmc2.pdf (6.1MB, pdf)

Articles from Biophysical Journal are provided here courtesy of The Biophysical Society

RESOURCES