Skip to main content
BMC Bioinformatics logoLink to BMC Bioinformatics
. 2012 Jul 9;13:160. doi: 10.1186/1471-2105-13-160

ParticleCall: A particle filter for base calling in next-generation sequencing systems

Xiaohu Shen 1, Haris Vikalo 1,
PMCID: PMC3464607  PMID: 22776067

Abstract

Background

Next-generation sequencing systems are capable of rapid and cost-effective DNA sequencing, thus enabling routine sequencing tasks and taking us one step closer to personalized medicine. Accuracy and lengths of their reads, however, are yet to surpass those provided by the conventional Sanger sequencing method. This motivates the search for computationally efficient algorithms capable of reliable and accurate detection of the order of nucleotides in short DNA fragments from the acquired data.

Results

In this paper, we consider Illumina’s sequencing-by-synthesis platform which relies on reversible terminator chemistry and describe the acquired signal by reformulating its mathematical model as a Hidden Markov Model. Relying on this model and sequential Monte Carlo methods, we develop a parameter estimation and base calling scheme called ParticleCall. ParticleCall is tested on a data set obtained by sequencing phiX174 bacteriophage using Illumina’s Genome Analyzer II. The results show that the developed base calling scheme is significantly more computationally efficient than the best performing unsupervised method currently available, while achieving the same accuracy.

Conclusions

The proposed ParticleCall provides more accurate calls than the Illumina’s base calling algorithm, Bustard. At the same time, ParticleCall is significantly more computationally efficient than other recent schemes with similar performance, rendering it more feasible for high-throughput sequencing data analysis. Improvement of base calling accuracy will have immediate beneficial effects on the performance of downstream applications such as SNP and genotype calling.

ParticleCall is freely available at https://sourceforge.net/projects/particlecall.

Background

The advancements of next-generation sequencing technologies have enabled inexpensive and rapid generation of vast amounts of sequencing data [1-3]. At the same time, high-throughput sequencing technologies present us with the challenge of processing and analyzing large data sets that they provide. A fundamental computational challenge encountered in next-generation sequencing systems is the one of determining the order of nucleotides from the acquired measurements, the task typically referred to as base calling. The accuracy of base calling is of essential importance for various downstream applications including sequence assembly, SNP calling, and genotype calling [4]. Moreover, improving base calling accuracy may enable achieving desired performance of downstream applications with smaller sequencing coverage, which translates to a reduction in the sequencing cost.

A widely used sequencing-by-synthesis platform, commercialized by Illumina, relies on reversible terminator chemistry. Illumina’s sequencing platforms are supported by a commercial base-calling algorithm called Bustard. While Bustard is computationally very efficient, its base-calling error rates can be significantly improved by various computationally more demanding schemes [5]. Such schemes include work presented in [6-9]. Among the proposed methods, the BayesCall algorithm [8] has been shown to significantly outperform Bustard in terms of the achievable base calling error rates. By relying on a full parametric model of the acquired signal, BayesCall builds a Bayesian inference framework capable of providing valuable probabilistic information that can be used in downstream applications. However, its performance gains come at high computational costs. A modified version of the BaseCall algorithm named naiveBayesCall [9] performs base calling in a much more efficient way, but its accuracy deteriorates (albeit remains better than Bustard’s). Both BayesCall and naiveBayesCall rely on expectation-maximization (EM) framework that employs a Markov chain Monte Carlo (MCMC) sampling strategy to estimate the parameters of the statistical model describing the signal acquisition process. This parameter estimation step turns out to be very time-consuming, limiting practical feasibility of the proposed schemes. Highly accurate and practically feasible parameter estimation and base-calling remain a challenge that needs to be addressed.

In this paper, we propose a Hidden Markov Model (HMM) representation of the signal acquired by Illumina’s sequencing-by-synthesis platforms and develop a particle filtering (i.e., sequential Monte Carlo) base-calling scheme that we refer to as ParticleCall. When relying on the BayesCall’s Markov Chain Monte Carlo implementation of the EM algorithm (MCEM) to estimate system parameters, ParticleCall achieves the same error rate performance as BayesCall while reducing the time needed for base calling by a factor of 3. To improve the speed of parameter estimation, we develop a particle filter implementation of the EM algorithm (PFEM). PFEM significantly reduces parameter estimation time while leading to a very minor deterioration of the accuracy of base calling. Finally, we demonstrate that ParticleCall has the best discrimination ability among all of the considered base calling schemes.

Methods

In this section, we first review the data acquisition process and the basic mathematical model of the Illumina’s sequencing-by-synthesis platform. Then we introduce a Hidden Markov Model (HMM) representation of the acquired signals. Relying on the HMM model and particle filtering (i.e., sequential Monte Carlo) techniques, we develop a novel base calling and parameter estimation scheme and discuss some important practical aspects of the proposed method.

Illumina sequencing platform

A sequencing task on the Illumina’s platform is preceded by the preparation of a library of single-stranded short templates created by performing random fragmentation of the target DNA sample. Each single-stranded fragment in the library is placed on a glass surface (i.e., the flow cell [10]) and subjected to bridge amplification in order to create a cluster of identical copies of DNA templates [11]. The flow cell contains eight lanes where each lane is divided into a hundred of nonoverlapping tiles. The order of nucleotides in a DNA template is identified by synthesizing its complementary strand while relying on reversible terminator chemistry [3]. Ideally, in every sequencing cycle, a single fluorescently labeled nucleotide is incorporated into the complementary strand on each copy of the template in a cluster. The incorporated nucleotide is a Watson-Crick complement of the first unpaired base of the template. In reversible terminator chemistry, four distinct fluorescent tags are used to label the four bases, and are detected by CCD imaging technology. The acquired images are processed in order to obtain intensity signals indicating the type of nucleotide incorporated in each cycle. These raw signal intensities are then analyzed by a base-calling algorithm to infer the order of nucleotides in each of the templates.

Quality of the acquired raw signals is adversely affected by the imperfections in the underlying sequencing-by-synthesis and signal acquisition processes. The imperfections are manifested as various sources of uncertainties. For instance, a small fraction of the strands being synthesized may fail to incorporate a base, or they may incorporate multiple bases in a single test cycle. These effects are referred to as phasing and pre-phasing, respectively, and they result in an incoherent addition of the signals generated by the synthesis of the complementary strands on the copies of the template. Other sources of uncertainty are due to cross-talk and delay effects in the optical detection process, the residual effects that are readily observed between subsequent test cycles, signal decay, and measurement noise.

Overview of the mathematical model

To describe the signal acquired by the Illumina’s sequencing-by-synthesis platform, a parametric model was proposed in [8]. Basic components of the model are overviewed below.

A length-L DNA template sequence is represented by a 4×L matrix S, where the ithcolumn of S, si, is considered to be a randomly generated unit vector with a single non-zero entry indicating the type of the ithbase in the sequence. We follow the convention where the first component of the vector sicorresponds to the base A, the second to C, the third to G, and the fourth to T and denote them as eA,eC,eG,eT. The goal of base-calling is to infer unknown S from the signals obtained by optically detecting nucleotides incorporated during the sequencing-by-synthesis process.

Let p denote the average fraction of strands that fail to extend in a test cycle. Phasing is modeled as a Bernoulli random variable with probability p. Let q denote the average fraction of strands which extend by more than one base in a single test cycle. Pre-phasing is modeled as a Bernoulli random variable with probability q. Length of the synthesized strand changes from i to j with probability

Pij=p,ifj=i,1pq,ifj=i+1,q,ifj=i+2,0,otherwise. (1)

Let P denote an (L + 1)×(L + 1) transition matrix with entries Pijdefined above, 1≤i,jL + 1. The signal generated over L cycles of the synthesis process is affected by phasing and pre-phasing and can be expressed as X=SH, where H=(Hi,j) is an L×L matrix with entries Hi,j=[Pj]1(i+1), the probability that a synthesized strand is of length i after j cycles. Here Pj denotes the jth power of matrix P. The decay in signal intensities over cycles (caused by DNA loss due to primer-template melting, digestion by enzymatic impurities, DNA dissociation, misincorporation, etc.) is modeled by the per-cluster density random parameter λt,

λt=(1dt)λt1+(1dt)λt1𝜖t, (2)

where 𝜖t𝒩(0,σt2) is a one-dimensional Gaussian random variable and dt is the per-cluster density decay parameter within [0,1]. We represent the tthcolumn of H as htand the tth column of X as xt. Incorporating the decay into the model, the signal generated in cycle t is expressed as

xt=λtSht, (3)

where xt=xtAxtCxtGxtT is the vector of signals generated in each of the optical channels. Assuming Gaussian observation noise, the measured intensities at cycle t are given by

yt=Ktxt+b{A,C,G,T}xtbηtb, (4)

where Ktdenotes the 4×4 crosstalk matrix describing overlap of the emission spectra of the four fluorescent tags, and ηtA,ηtC,ηtG,ηtT are independent, identically distributed (i.i.d.) 4×1 Gaussian random vectors with zero mean and a common 4×4 covariance matrix Σt.

Note that, due to typically small values of p and q, the components of the vector ht around its tth entry are significantly greater than the remaining ones. This observation can be used to simplify the expressions (2) and (3). In particular, let htw denote the vector obtained by windowing ht around its tth entry, i.e., by setting small components of ht to 0. In general, we consider l + r + 1 dominant components of htcentered at position t, Htl,t,Htl + 1,t,…,Ht,t,…Ht + r−1,t,Ht + r,t, and then expression (2) becomes

xtxtw=λtShtw=λti=lrst+iHt+i,t. (5)

Finally, note that the signal measured in cycle t is empirically observed to contain residual effect from the previous cycle. The residual effect is modeled by adding αt(1−dt)yt−1to yt, where the unknown parameter αt∈(0,1). Therefore, the model can be summarized as

λt|λt1𝒩(1dt)λt1,(1dt)2λt12σt2,yt|yt1,S,λt𝒩Ktxtw+αt(1dt)yt1,xtw22Σt,stUnifeA,eC,eG,eT,xtw=λtShtwt=1,2,,L (6)

where ∥·∥2 denotes the l2-norm of its argument, and where y0=0, λ0=1.

Hidden Markov Model of DNA base-calling

In this section, we reformulate the statistical description of the signal acquired by the Illumina’s sequencing-by-synthesis platform as a Hidden Markov Model (HMM) [12]. HMMs comprise a family of probabilistic graphical models which describe a series of observations by a “hidden” stochastic process and are generally suitable for representing time series data. Sequencing data obtained from the Illumina’s platform is a set of time-series intensities y1:L, motivating the HMM representation. HMMs provide a convenient framework for state and parameter estimation, which we exploit to develop a particle filter base-calling scheme in the next section.

For the sake of convenience, we remove the dependency between subsequent observations yt−1 and yt by defining yt=ytαt(1dt)yt1,t=1,2,,L. Therefore, we can write

yt|S,λt𝒩Ktxtw,xtw22Σt. (7)

Components of y1:L are the observations of our HMM, and depend on the underlying signals x1:L. Moreover, let Stw denote the 4×(l + r + 1) windowed submatrix of S, i.e.,

Stw=[stlstl+1stst+r]. (8)

Since xtw=λtShtw=λti=tlt+rsiHi,t, it is clear that yt depends on λtand Stw. Therefore, we define the state of the HMM to be the combination of λt and Stw – the per-cluster density at cycle t and the collection of (l + r + 1) bases around (and including) the base in position t, respectively.

The proposed HMM representation is illustrated in Figure 1. The observation dynamics that characterize the relationship between yt and the hidden states (Stw,λt) are given by the distribution g(yt|Stw,λt). It is straightforward to show from (5) that

Figure 1.

Figure 1

A hidden Markov model of the generated signal in Illumina sequencing-by-synthesis platforms. An illustration of the graphical HMM of the Illumina’s sequencing platform. The observations yrepresent signal intensities after the removal of residual effects. The states are the combinations of Stw and λt, which represent a subsequence of the template centered at position t and per-cluster density, respectively.

gyt|Stw,λt𝒩Ktxtw,xtw22Σt. (9)

On the other hand, the state transition dynamics is described by the transition probability between subsequent states, (St1w,λt1) and (Stw,λt). Since Stw and λt are independent, the transition probability is

f(Stw,λt|St1w,λt1)=f1(Stw|St1w)f2(λt|λt1). (10)

The second term on the right-hand side of (8), f2(λt|λt−1), is known from the density decay model (1),

f2(λt|λt1)𝒩((1dt)λt1,(1dt)2λt12σt2). (11)

For notational convenience, we use st,1w,,st,l+r+1w to denote the set of l + r + 1 column vectors of Stw. Note that for k=2,3,…,l + r + 1, the column vectors st1,kw in St1w and the column vectors st,k1w in Stw actually represent the same base. Therefore, the transition model between them can be represent by a δ function as

p(st,k1w|st1,kw)=δ{st,k1w=st1,kw}=1,ifst,k1w=st1,kw,0,ifst,k1wst1,kw. (12)

Let U({eA,eC,eG,eT}) denote a uniform distribution on the support set of unit vectors ({eA,eC,eG,eT}). We assume no correlation between consecutive bases of the template sequence, i.e., st,l+r+1w is generated from U({eA,eC,eG,eT}). Therefore, f1(Stw|St1w) can be written as

f1(Stw|St1w)=k=2l+r+1δ{st1,kw=st,k1w}ust,l+r+1w, (13)

where u(·)∼U({eA,eC,eG,eT}). Hereby, all the components of the HMM are specified.

ParticleCall base-calling algorithm

The goal of base calling is to determine the order of nucleotides in a template from the acquired signal y1:t. This can be rephrased as the problem of inferring the most likely sequence of states (Stw,λt) of the HMM in (7)-( 8) from the observed sequence y1:t(clearly, s1:L follows directly from Stw). We assume that the parameters Λ={p,q,d1:L,α1:L,σ1:L,K1:L,Σ1:L} are common for all clusters within a tile, and that they are provided by a parameter estimation step discussed in the following section. In this section, we introduce a novel base calling algorithm ParticleCall which relies on particle filtering techniques to sequentially infer (Stw,λt) and, therefore, recover the matrix S.

In general, particle filtering (i.e., sequential Monte Carlo) methods generate a set of particles with associated weights to estimate the posteriori distribution of unknown variables given the acquired measurements [13]. In the proposed HMM framework, we sequentially calculate the posteriori distribution of the columns of Sp(st|y1:t),t=1,2,…,L, and find the maximum a posteriori (MAP) estimates of st by solving

s^t=argmaxst{eA,eC,eG,eT}{p(st|y1:t)}. (14)

Our algorithm relies on a sequential importance sampling/resampling (SISR) particle filter scheme [14] to calculate p(Stw,λt|y1:t). Different choices and approximation methods of proposal densities are considered in [15-17]. We directly use the transition (8) as the proposal density. This sequential importance sampling suffers from degeneracy and the variance of the importance weights will increase over time. To address the degeneracy problem, a resampling step is introduced in order to eliminate samples which have small normalized importance weights. Common resampling methods include multinomial resampling [14], residual resampling [18] and systematic resampling [19,20]. We measure degeneracy of the algorithm using the effective sample size Keff and, for the sake of simplicity, employ multinomial resampling strategy. If we denote the number of particles by Np and associated weights by w, then Keff=(k=1Np(wt(i))2)1 and resampling step is used when Keff is below a fixed threshold Nthreshold. Nthreshold of size O(Np) is typically sufficient [14]. In our implementation, we set Nthreshold=Np/2.

We omit further details for brevity and formalize the ParticleCall algorithm below.

Algorithm 1

ParticleCall base-calling algorithm

1. Initialization:

1.1 Initialize particles:fori=1→Npdo

Sample each column of the submatrix S1w,(i) from U({eA,eC,eG,eT}); Sample λ1(i) from a Gaussian distribution with mean 1, and the variance calculated using Bustard’s estimates of λin the first 10 test cycles.end for

1.2 Compute and normalize weights for each particle according to w1(i)g(y1|S1w,(i),λ1(i)) as in (7).

2. Run iteration t(t≥2):

2.1 Sampling:fori=1→Npdo

Sample Stw,(i),λt(i)f(·,·|St1w,(i),λt1(i)) according to (8).end for

2.2 Update the importance weight

wt(i)wt1(i)gyt|Stw,(i),λt(i). (15)

2.3 Normalize the weights. Calculate the posteriori probability of st and obtain the estimate s^t.

2.4 Resampling:ifKeff=k=1Np(wt(i))21Nthresholdthen

Draw Npsamples {S¯tw,(j),λ¯t(j),j=1,,Np} from {Stw,(i),λt(i),i=1,,Np} with probabilities proportional to {wt(i),i=1,,Np}. Assign equal weight to each particle, w¯t(i)=1/Np.end if

Since Stw in the HMM states are discrete with a finite alphabet, and the transitions of Stw and λt are independent according to (8), it is possible to Rao-Blackwellize the ParticleCall algorithm. Rao-Blackwellization is used to marginalize part of the states in the particle filter, hence reducing the number of needed particles Np[16]. We marginalize the discrete states Stw and reduce the hidden process to λt, while relying on the particle filter to calculate p(λ1:t|y1:t).

The original posterior distribution of the states can be expressed as

pλ1:t,S1:tw|y1:t=pS1:tw|y1:t,λ1:tpλ1:t|y1:t. (16)

Since p(λ1:t|y1:t)p(yt|y1:t1,λ1:t)p(λt|λt1(i)), where λt1(i) is a sample from p(λ1:t−1|y1:t1), we can state the Rao-Blackwellized ParticleCall algorithm as below.

Algorithm 2

Rao-Blackwellized ParticleCall algorithm

1. Initialization:

1.1 Initialize particles:fori=1→Npdo

Sample λ1(i) from a Gaussian distribution with mean 1, and the variance calculated using Bustard’s estimates of λin the first 10 test cycles.end for

1.2 Compute and normalize weights for each particle according to w1(i)g(y1|λ1(i))S1wg(y1|S1w,λ1(i)).

1.3 Calculate the discrete distribution p(S1w|y1,λ1(i)) for each i.

2. Run iteration t(t≥2):

2.1 Sampling:fori=1→Npdo

Sample λt(i)f(·|λt1(i)).end for

2.2 Update the importance weight wt(i)wt1(i)g(yt|y1:t1,λ1:t(i)). and normalize the weights.

2.3 Resample if KeffNthreshold

2.4 Update p(Stw|y1:t,λ1:t(i))fori=1→Npdo

Update p(Stw|y1:t,λ1:t(i)) using p(St1w|y1:t1,λ1:t1(i)) and λt(i).end for

In step 2.2 of Algorithm 2, the quantity g(yt|y1:t1,λ1:t(i)) can be obtained by marginalizing over discrete states Stw and St1w,

gyt|y1:t1,λ1:t(i)=Stwpyt|y1:t1,Stw,λ1:t(i)pStw|y1:t1,λ1:t(i)=Stwpyt|Stw,λt(i)St1wpStw|St1w,y1:t1,λ1:t(i)×pSt1w|y1:t1,λ1:t(i), (17)

where p(yt|Stw,λt(i)) is the observation density, p(St1w|y1:t1,λ1:t(i))=p(St1w|y1:t1,λ1:t1(i)) due to the independence of the state transitions, and p(Stw|St1w,y1:t1,λ1:t(i))=p(Stw|St1w) due to the Markov property and the independence of the state transitions.

In step 2.4 of Algorithm 2, the update equation is obtained as

pStw|y1:t,λ1:t(i)pStw,yt,λt(i)|y1:t1,λ1:t1(i)=St1wpyt,Stw,λt(i)|y1:t1,St1w,λ1:t1(i)×pSt1w|y1:t1,λ1:t1(i)=St1wpyt|Stw,λt(i),y1:t1,St1w,λ1:t1(i)×pStw,λt(i)|y1:t1,St1w,λ1:t1(i)×pSt1w|y1:t1,λ1:t1(i)=pyt|Stw,λt(i)pλt(i)|λt1(i)×St1wpStw|St1wpSt1w|y1:t1,λ1:t1(i) (18)

Parameter estimation

To determine the set of parameters Λneeded to run the proposed ParticleCall base calling algorithm, one could rely on the MCMC implementation of the EM algorithm (MCEM) proposed in [8]. In section Results and discussion, we demonstrate the performance of the ParticleCall algorithm that relies on the MCEM parameter estimation scheme. Note, however, that the MCMC sampling strategy employed by MCEM requires a lengthy burn-in period and a very large sample size to perform the expectation step. Therefore, the MCEM parameter estimation scheme is computationally rather intensive and requires significant computational resources if it is to be used for processing large sequencing data sets. As an alternative, we develop an EM parameter estimation scheme which relies on the proposed HMM and uses samples generated by a particle filter to evaluate the expectation of the likelihood function. We refer to this algorithm as the particle filter EM (PFEM). The speed and accuracy of the proposed scheme is practically sound for use in next generation sequencing platforms.

Assumptions on parameters

Recall that the set of parameters needed to run ParticleCall is Λ={pqd1:Lα1:Lσ1:LK1:LΣ1:L}. The phasing and prephasing parameters p and q are assumed to be the same for each sequencing lane and are estimated using the same procedure as Bustard (see, e.g., [8]). The remaining parameters are assumed to be cycle-dependent and need to be estimated for each tile. The cycle-dependency assumption on the parameters can lead to a substantial improvement in the base-calling accuracy [5]. In order to avoid over-fitting, we assume that parameters remain constant within a short window of cycles and then change to a different set of values. To track the changes in the parameters, we first divide the total read length L into several non-overlapping windows and then perform our parameter estimation window-by-window. To further reduce the number of parameters and improve the estimation efficiency, we assume that the parameters d1:L and σ1:L are uniformly distributed over an interval and incorporate them into the hidden states of the HMM model. Therefore, only the mean and variance of these parameters, i.e., dmean, dvar, σmean, and σvarneed to be estimated. Computational results demonstrate that these two assumptions does not affect the accuracy of base-calling.

Particle filter EM algorithm

In the early sequencing cycles, effects of phasing and prephasing are relatively small. Therefore, we may ignore phasing and prephasing to facilitate straight-forward computation of the initial estimates of the remaining parameters. In particular, the signal generated in the early cycles t is approximated as

xt=λtst. (19)

Replacing (2) by (10) leads to a simplified model that allows for straightforward base calling and inference of the parameters by means of linear regression. We use these values to obtain the estimates of dmean, dvar, σmean, and σvar, and to initialize the remaining parameters α, K, Σ, in the particle filter EM parameter estimation procedure.

The parameter estimation is performed window-by-window and is conducted using n reads randomly chosen from a tile (in our experiments, we use n=200). Assume the window length is w, and denote the window index by m. The particle filter EM (PFEM) algorithm finds parameters for one window and then uses these values to initialize the search for parameters in the next window. We illustrate the procedure for the first window here (the same procedure is repeated in the following windows). Let Λ1i={αi,Ki,Σi} denote the set of parameters for window 1 in the ith iteration of the EM scheme. The estimate of Λ1i is given by

Λ1i=argmaxΛ1L1Λ1i1, (20)

where L1(Λ1i1)=j=1nL1,j(Λ1i1) is the sum of the log-likelihood functions over the reads in the training set. The log-likelihood function for each read, L1,j(Λ1i1), is obtained as

L1,j(Λ1i1)=logP(y1:w|Λ1i1)=ElogP(y1:w,s1:w,λ1:w|Λ1i1), (21)

where the expectation is taken with respect to P(s1:w,λ1:w|y1:w,Λ1i1). We rely on an SISR particle filtering scheme to generate equally weighted sample trajectories from P(s1:w,λ1:w|y1:w,Λ1i1). Based on (7) and (8), we calculate logP(y1:w,s1:w,λ1:w|Λ1i1) for these samples and compute their average to approximate the expectation in (12). The maximization (11) is performed by solving equations obtained after taking gradients of L1(Λ1i1) over the parameters and setting them to 0. In our experiment, the PFEM parameter estimation scheme performs 30 EM iterations and uses 600 samples from the particle filter for each window.

Results and discussion

The proposed method is evaluated on a data set obtained by sequencing phiX174 bacteriophage using Illumina Genome Analyzer II with the cycle length 76. This is a short genome with a known sequence which enables reliable performance comparison of different base-calling techniques. We tested ParticleCall and several other algorithms on a tile containing 77337 reads, and present the results here. All the codes are written in C and the tests are run on a desktop with an Intel Core i7 4-core 3GHz processor.

Performance of ParticleCall

The base calling error rates are computed by aligning the reads to the reference genome and evaluating frequency of mismatches. Reads that could not be aligned to the reference with at least 70% matches are discarded. Note that the error rates and speed of the proposed ParticleCall algorithm and the parameter estimation scheme are affected by the parameters l, r, particle number Np, and parameter estimation window length w. We ran ParticleCall with l=r∈{1,2,4}. Increasing l and r beyond l=r=1 did not affect the performance while it significantly slowed down the algorithm. This is due to small values of the phasing and prephasing probabilities, which are estimated to be p=3.54×10−8 and q=0.00335. Therefore, in the remainder of the paper, we set l=r=1. The accuracy of base-calling for different Npis shown in Table 1. As seen there, for the original ParticleCall algorithm, Np=800 leads to high performance with reasonable speed. Rao-Blackwellized ParticleCall can achieve the same accuracy with fewer particles (in particular, Np=300); however, its effective running time is 3 times that of the original ParticleCall with the same performance. This is because the Rao-Blackwellization steps in (9) and (9) require evaluating a sum over all possible Stw (43=64 for our choice l=r=1), resulting in a fairly large number of basic operations needed to calculate exact distribution of the discrete variables. Therefore, for further performance comparisons, we rely on the original ParticleCall algorithm (formalized as Algorithm 1). Table 2 shows the ParticleCall base calling error rate and parameter estimation times for different window lengths w. In the remainder of the paper, we set w=5 as it leads to desirable performance/speed characteristics of the algorithm.

Table 1.

Comparison of ParticleCall with differentNp

Method Np error rate base-calling time (min)  
ParticleCall (via MCEM)
400
0.0126
46
 
 
800
0.0124
88
 
 
1200
0.0124
130
 
ParticleCall (via PFEM)
400
0.0128
46
 
 
800
0.0125
91
 
 
1200
0.0125
133
 
Rao-Blackwellized ParticleCall (via MCEM)
100
0.0128
103
 
 
200
0.0125
190
 
 
300
0.0124
287
 
  400 0.0124 386  

ParticleCall is run using parameters obtained via the MCEM parameter estimation scheme as well as via the PFEM parameter estimation algorithm proposed in this paper. Rao-Blackwellized ParticleCall is run using parameters via the MCEM parameter estimation scheme.

Table 2.

ParticleCall parameter estimation

 
 
parameter estimation
Window length w base-calling error rate time (min)
4
0.0125
50
5
0.0125
39
6
0.0127
29
7 0.0130 25

ParticleCall base-calling error rate and the parameter estimation time of the proposed PFEM parameter estimation algorithm.

Performance comparison of different algorithms

The error rates and speed of the proposed ParticleCall algorithm are compared with those of BayesCall, naiveBayesCall, Rolexa, and Bustard. We run ParticleCall both with parameters provided by the computationally intensive MCEM algorithm as well as with those inferred by the PFEM parameter estimation scheme proposed in this paper. The results are reported in Table 3. Note that Rolexa generally outputs the so-called IUPAC codes, unlike all the other considered algorithms which provide sequences of nucleotides A, C, G, and T. To allow a comparison, we enforce Rolexa to output sequences of nucleotides as well. The comparison of per-cycle error rates is shown in Figure 2. It can be seen from Table 3 and Figure 2 that ParticleCall, BayesCall and naiveBayesCall all have improved base-calling accuracy compared to Bustard. BayesCall is highly accurate but relatively slow – it requires approximately 4 hours to complete base-calling for one tile of the data. naiveBayesCall significantly improves base-calling speed over BayesCall but it does so at the expense of incurring higher error rate. Our ParticleCall base-calling algorithm has the same accuracy as BayesCall, while being roughly 3 times faster. Figure 2 shows that both ParticleCall and BayesCall are more accurate than naiveBayesCall in the early cycles and improve over Bustard in all cycles. Note that Bustard outperforms Rolexa, which is consistent with the results in [5]. Moreover, we see from Table 3 that performing parameter estimation via the MCEM algorithm proposed in [8] requires 19 hours, while the particle filter implementation of the EM estimation scheme proposed in this paper takes only 39 minutes. As evident from Table 3, running ParticleCall with parameters obtained by the PFEM scheme leads to only a minor performance degradation compared to running it with parameters obtained by the MCEM algorithm. Running ParticleCall base calling along with the PFEM parameter estimation scheme takes about 2 hours per tile, which is 9 times faster than the total time required by the less accurate naiveBayesCall.

Table 3.

Comparison of error rates and speed

 
 
base-calling
parameter estimation
 
Method error rate time (min) time (min)  
Bustard
0.0152
2 (total)
 
 
Rolexa
0.0170
35 (total)
 
 
naiveBayesCall
0.0132
21
1139
 
BayesCall
0.0124
231
1139
 
ParticleCall
       
(via MCEM)
0.0124
88
1139
 
ParticleCall
       
(via PFEM) 0.0125 91 39  

The base-calling error rate and the running times of different algorithms. ParticleCall is run using parameters obtained via the MCEM parameter estimation scheme as well as via the PFEM parameter estimation algorithm proposed in this paper. For Bustard and Rolexa, only the total running times are reported.

Figure 2.

Figure 2

Per-cycle error rates of ParticleCall, BayesCall, naiveBayesCall, Rolexa and Bustard. The figure compares the per-cycle error rates of different base-calling algorithms. ParticleCall and BayesCall are the most accurate ones.

Quality scores

Quality scores are used to characterize confidence in the outcome of the base-calling procedures. They are computed as part of the analysis of the acquired raw data and may be used to filter out reads of suspect quality, or to shorten the reads if the quality scores of individual bases fall below certain thresholds. They can also provide confidence information for downstream analysis including sequence assembly and SNP and genotype calling. Frequently used are the so-called phred quality scores, which were originally developed to assess the quality of the conventional Sanger sequencing and automate large-scale sequencing projects. Phred scores are also often provided by the algorithms used for base-calling in next generation sequencing platforms. Formally, the phred score for a called base in the cycle t, s^t, is defined as

Qphred(s^t)=10log10P(s^tst). (22)

Essentially, Qphred(s^t) is the scaled logarithm of the error probability. Higher quality scores imply smaller probability of the base-calling error. For the proposed ParticleCall algorithm, probability of correctly calling a base can be obtained from the posteriori probability as

P(s^tst)=1pst|y1:t. (23)

Quality scores can be used to compare the discrimination ability of different algorithms. The discrimination score D(ε) at error tolerance ε is defined as the ratio of the correctly called bases having P(s^tst)<𝜖 (i.e., the quality score higher than 10log10(𝜖)) to all called bases. Figure 3 compares the discrimination ability of ParticleCall, BayesCall, naiveBayesCall and Bustard. It shows that for a reasonable error tolerance ε, ParticleCall with parameters obtained through MCEM has better discrimination ability than BayesCall, naiveBayesCall and Bustard, while ParticleCall with parameters obtained through PFEM has discrimination ability close to naiveBayesCall and better than other algorithms. In other words, when a small cutoff error tolerance ε is set and all the bases with quality scores below εare considered invalid, ParticleCall provides the most accurate results among the considered base-calling schemes.

Figure 3.

Figure 3

Discrimination abilityD(ε) of quality scores vs error tolerance. The figure shows the percentage of correctly called bases under different error tolerance ε.

Effects of improved base-calling accuracy on de novo sequence assembly

In shotgun sequencing, a long target sequence is oversampled by a library of randomly fragmented copies of the target, and the overlaps between short reads obtained by a high-throughput platform are used to assemble the target. In de novo assembly, the target is reconstructed without consulting any reference [21,22]. Performance of assembly algorithms highly depends on the accuracy of base calling. To demonstrate the effects of base-calling accuracy on assembly, we apply the Velvet assembly algorithm [22] on reads provided by Bustard, Rolexa, naiveBayesCall, BayesCall, and ParticleCall. In particular, we randomly subsample the set of reads provided by each of the base calling algorithms to emulate 5X, 10X, 15X, and 20X coverage. Then we run Velvet on each of the subsets, and evaluate commonly used metrics that quantify the quality of the assembly procedure. Specifically, we evaluate the maximum contig length and the N50 contig length. The described procedure is repeated 200 times to obtain average values of these two quality metrics. The results are shown in Table 4. As can be seen there, ParticleCall provides the largest N50 and maximum contig length among all of the considered base calling schemes, for all of the considered coverages.

Table 4.

de novo assembly results

 
 
 
 
 
 
 
 
 
ParticleCall
ParticleCall
Coverage
Bustard
Rolexa
naiveBayesCall
BayesCall
via MCEM
via PFEM
  N50 Max N50 Max N50 Max N50 Max N50 Max N50 Max
5X
271
607
259
565
278
604
292
629
299
637
289
632
10X
1169
1750
971
1557
1180
1731
1269
1831
1316
1900
1341
1865
15X
3624
3823
2885
3170
3726
3908
3466
3741
3742
3935
3697
3918
20X 4694 4744 4529 4614 4756 4816 4827 4875 5102 5116 4795 5039

The maximum contig length and N50 length of de novo assembly using Velvet. The average values over 200 experiments are shown in the table.

Conclusions

In this paper we presented ParticleCall, a particle filtering algorithm for base calling in the Illumina’s sequencing-by-synthesis platform. The algorithm is developed by relying on an HMM representation of the sequencing process. Experimental results demonstrate that the ParticleCall base calling algorithm is more accurate than Bustard, Rolexa, and naiveBayesCall. It is as accurate as BayesCall while being significantly faster. Quality score analysis of the reads indicates that ParticleCall has better discrimination ability than BayesCall, naiveBayesCall and Bustard. Moreover, a novel particle filter EM (PFEM) parameter estimation scheme, much faster than the existing Monte Carlo implementation of the EM algorithm, was proposed. When relying on the PFEM scheme, ParticleCall has near-optimal performance while needing much shorter total parameter estimation and base calling time.

Author’s contributions

Algorithms and experiments were designed by Xiaohu Shen (XS) and Haris Vikalo (HV). Algorithm code was implemented and tested by XS. The manuscript was written by XS and HV. Both authors read and approved the final manuscript.

Competing interests

The authors declare that they have no competing interests.

Contributor Information

Xiaohu Shen, Email: xhshen@utexas.edu.

Haris Vikalo, Email: hvikalo@ece.utexas.edu.

Acknowledgements

This work was funded by the National Institute of Health under grant 1R21HG006171-01.

References

  1. Shendure J, Ji H. Next-generation DNA sequencing. Nat Biotechnology. 2008;26:1135. doi: 10.1038/nbt1486. [DOI] [PubMed] [Google Scholar]
  2. Metzker M. Emerging technologies in DNA sequencing. Genome Research. 2005;56:1767. doi: 10.1101/gr.3770505. [DOI] [PubMed] [Google Scholar]
  3. Bentley D. Whole-genome re-sequencing. Curr Opin Genet Dev. 2006;16:545. doi: 10.1016/j.gde.2006.10.009. [DOI] [PubMed] [Google Scholar]
  4. Nielsen R, Paul JS, Alvrechtsen A, Song YS. Genotype and SNP calling from next-generation sequencing data. Nature Reviews. 2011;12:443. doi: 10.1038/nrg2986. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Ledergerber C, Dessimoz C. Base-calling for next-generation sequencing platforms. Briefings in Bioinformatics. 2011;12:489. doi: 10.1093/bib/bbq077. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Rougemont J, Amzallag A, Iseli C, Farinelli L, Xenarios I, Naef F. Probabilistic base calling of solexa sequencing data. BMC Bioinformatics. 2008;9:431. doi: 10.1186/1471-2105-9-431. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Erlich Y, Mitra P, Delabastide M, McCombie W, Hannon G. Alta-Cyclic: a self-optimizing base caller for next-generation sequencing. Nat Methods. 2008;5:679. doi: 10.1038/nmeth.1230. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Kao W, Stevens K, Song Y. BayesCall: A model-based base-calling algorithm for high-throughput short-read sequencing. Genome Research. 2009;19:1884. doi: 10.1101/gr.095299.109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Kao W, Stevens K, Song Y. naiveBayesCall: an efficient model-based base-calling algorithm for high-throughput sequencing. Journal of Computational Biology. 2011;18:365. doi: 10.1089/cmb.2010.0247. [DOI] [PubMed] [Google Scholar]
  10. Fedurco M, Romieu A, Williams S. et al. BTA, a novel reagent for DNA attachment on glass and efficient generation of solid-phase amplified DNA colonies. Nucleic Acids Res. 2006;34(3):e22. doi: 10.1093/nar/gnj023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Turcatti G, Romieu A, Fedurce M. et al. A new class of cleavable fluorescent nucleotides: synthesis and optimization as reversible terminators for DNA sequencing by synthesis. Nucleic Acids Res. 2008;36(4):e25. doi: 10.1093/nar/gkn021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Eddy S. Hidden Markov models. Current Opinion in Structural Biology. 1996;6(3):361. doi: 10.1016/s0959-440x(96)80056-x. [DOI] [PubMed] [Google Scholar]
  13. Doucet A, Wang X. Monte Carlo methods for signal processing: A review in the statistical signal processing context. IEEE Signal Processing Magzine. 2005;22:152. [Google Scholar]
  14. Cappé O, Moulines E, Rydén T. Inference in hidden Markov models. Springer Verlag, New York; 2005. [Google Scholar]
  15. Pitt M, Shephard N. Filtering via simulation: Auxiliary particle filters. Journal of the American Statistical Association. 1999;94:590. [Google Scholar]
  16. Doucet A, Godsill S, Andrieu C. On sequential Monte Carlo sampling methods for Bayesian filtering. Statistics and computing. 2000;10(3):197. [Google Scholar]
  17. Kim S, Shephard N, Chib S. Stochastic volatility: likelihood inference and comparison with ARCH models. The Review of Economic Studies. 1998;65(3):361. [Google Scholar]
  18. Liu J, Chen R. Sequential Monte Carlo methods for dynamic systems. Journal of the American statistical association. 1998;93:1032. [Google Scholar]
  19. Kitagawa G. Monte Carlo filter and smoother for non-Gaussian nonlinear state space models. Journal of computational and graphical statistics. 1996;5:1. [Google Scholar]
  20. Carpenter J, Clifford P, Fearnhead P. Improved particle filter for nonlinear problems. Radar, Sonar and Navigation, IEE Proceedings-, Volume 146, IET. 1999. pp. 2–7.
  21. Butler J, MacCallum I, Kleber M, Shlyakhter I, Belmonte M, Lander E, Nusbaum C, Jaffe D. ALLPATHS: de novo assembly of whole-genome shotgun microreads. Genome Research. 2008;18(5):810. doi: 10.1101/gr.7337908. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Zerbino D, Birney E. Velvet: algorithms for de novo short read assembly using de Bruijn graphs. Genome research. 2008;18(5):821. doi: 10.1101/gr.074492.107. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from BMC Bioinformatics are provided here courtesy of BMC

RESOURCES