Skip to main content
Biomolecular Detection and Quantification logoLink to Biomolecular Detection and Quantification
. 2017 Apr 29;12:1–6. doi: 10.1016/j.bdq.2017.04.001

Methods to determine limit of detection and limit of quantification in quantitative real-time PCR (qPCR)

Amin Forootan a,b,, Robert Sjöback c, Jens Björkman c, Björn Sjögreen b, Lucas Linz d, Mikael Kubista c,e
PMCID: PMC5496743  PMID: 28702366

Abstract

Quantitative Real-Time Polymerase Chain Reaction, better known as qPCR, is the most sensitive and specific technique we have for the detection of nucleic acids. Even though it has been around for more than 30 years and is preferred in research applications, it has yet to win broad acceptance in routine practice. This requires a means to unambiguously assess the performance of specific qPCR analyses. Here we present methods to determine the limit of detection (LoD) and the limit of quantification (LoQ) as applicable to qPCR. These are based on standard statistical methods as recommended by regulatory bodies adapted to qPCR and complemented with a novel approach to estimate the precision of LoD.

Keywords: Limit of detection, Limit of quantification, LoD, LoQ, GenEx software, Real-time PCR, qPCR, Data analysis, Replicates, MIQE, Quality control, Standardization

1. Introduction

Arguably among the most critical performance parameters for a diagnostic procedure are those related to the minimum amount of target that can be detected and quantified [11]. The parameters describing those properties are known as the limit of detection “LoD” and the limit of quantification “LoQ”. Their definitions vary slightly among regulatory bodies and standards organizations [1]. The Clinical Laboratory Standards Institute (www.clsi.org), for example, defines LoD as “the lowest amount of analyte (measurand) in a sample that can be detected with (stated) probability, although perhaps not quantified as an exact value” [2]. In many clinical laboratories and diagnostic applications, LoD is used interchangeably to “sensitivity”, “analytical sensitivity” and “detection limit.” This may, however, be confusing as “sensitivity” is also used in other ways. For example, in some applications “sensitivity” refers to the slope of the calibration curve, which is the definition used by the International Union of Pure and Applied Chemistry (IUPAC). CLSI defines LoQ as “the lowest amount of measurand in a sample that can be quantitatively determined with {stated} acceptable precision and stated, acceptable accuracy, under stated experimental conditions” [2]. An alternative LoQ based on clinical sensitivity and specificity has been proposed for diagnostic purposes [17].

Definitions by CLSI

LoD = the lowest amount of analyte (measurand) in a sample that can be detected with (stated) probability, although perhaps not quantified as an exact value.

LoQ = the lowest amount of measurand in a sample that can be quantitatively determined with {stated} acceptable precision and stated, acceptable accuracy, under stated experimental conditions

By far most measuring techniques generate a signal response that is proportional to the amount of measurand present. For example, measured absorption is proportional to the concentration of the measurand as predicted by the Beer-Lambert law. Linear measurements typically generate a background signal that is observed in the absence of measurand and must be subtracted from the measured values. This background signal limits the sensitivity of the measurement and is used to estimate LoD [3]. Working at 95% confidence, the limit of blank “LoB” is:

LoB=meanblank+1.645×σblank (1)

where σ is the standard deviation, and

LoD=LoB+1.645×σlowconcentrationsample (2)

This is also the recommended estimates in the CLSI guideline EP17 [2]. The σ in Eqs. (1) and (2) refers to the true standard deviation, while SD refers to estimated standard deviation from experiments. Replacing σ for SD requires also replacing 1.645 for the corresponding t-value, which depends on the degree of freedom and, hence, the number of replicates performed).

The above equations assume response is linear and data are normal distributed in linear scale. Small deviations from normal distribution when estimating SD have been discussed [12] but in qPCR, not even the response is linear. The measured Cq values are proportional to the log base 2 (log2) of the concentration of the measurand (or the number of target molecules present), which is a logarithmic response. This has dramatic implications on the analysis and interpretation of the data [4]. For example, no Cq value is obtained when a negative sample is measured, as the response never reaches the threshold line, and the standard deviation (SD) cannot be calculated for any set that includes negative samples. Hence, it is not possible to estimate LoD and LoQ by the standard procedures above. A further complication is that estimating confidence intervals assumes normal distribution. While linear data often are normally distributed in linear scale, qPCR data show normal distribution in logarithmic scale, further disqualifying the conventional approaches. To estimate LoD in qPCR one needs to revert to the definition of LoD In this paper we present the procedure to experimentally determine LoD based on sample replicates and also a novel method to estimate the confidence of the LoD. We also present the procedure to estimate LoQ of a qPCR system.

2. Materials and methods

The qPCR method used as an example to assess performance was ValidPrime [5], which is a optimized probe-based assay targeting a highly conserved, non-transcribed locus present in exactly one copy per haploid human genome. The test material was human genomic DNA (CAT# CHG50, TATAA Biocenter) calibrated against the National Institute of standards and technology (NIST) Human DNA Quantitation Standard (SRM 2372). A 2-fold dilution series was prepared covering the range 1 to 2048 molecules per reaction volume. Each standard sample was analyzed in 64 replicates, except for the most diluted sample, which was analyzed in 128 replicates. Grubb's test [13] was used to identify nine outliers that were removed, leaving 759 data points for the analysis.

In the qPCR reaction, TATAA Probe GrandMaster Mix L-Rox was used and the final concentration of the ValidPrime assay in the reaction was 200 nM of a FAM-labeled probe and 400 nM of each primer. The IntelliQube*1 (LGC Douglas Scientific) was used for all sample and master mix dispensing, thermal cycling, and real-time fluorescence detection, utilizing 1.6 μl reaction volumes. The 2-step qPCR protocol included a 1 min enzyme activation step at 95 °C, followed by 50 cycles of 10 s at 95 °C and 30 s at 60 °C. An auto-baselining method was used when plotting the amplification curves. Cq values were calculated with the IntelliQube software by manually setting a threshold line in the region of exponential amplification across all the amplification plots.

Cq data from the IntelliQube were preprocessed and analyzed using GenEx (MultiD analyses AB).

Coefficient of variation was calculated as:

expSDlnconc21 (3)

assuming log-normal distribution of replicate concentrations [6]. This follows from the fact that if the stochastic variable X has lognormal distribution, then by definition ln(X) is normally distributed with, say, mean μ and standard deviation σ. The distribution function of X is then given by

Fx=PX<x=PlnX<lnx=PlnXμσ<lnxμσ=Φlnxμσ (4)

where Φ is the distribution function of the standardized normal distribution. The probability density function of X is readily obtained as Fx from which the mean eμ+σ2/2 and variance eσ21e2μ+σ2 of X are straightforwardly obtained by integration. The coefficient of variation is defined as the ratio between the standard deviation and the mean, i.e., the coefficient of variation of X becomes eσ21.

Cq values were measured at p different concentrations, ci, i = 1, …, p, with n replicas at each concentration. For simplicity of presentation n is kept constant. The data analysis below is straightforward to generalize to the case where there are different of number of replicas for different concentrations. The resulting Cq values are arranged in a data matrix Cqi,j i = 1, …, p j = 1, …, n and an indicator function

Ii,j=1if   Cqi,j<Co0if   Cqi,j>Co (5)

where Co is a user specified cut-off value. Let zi=j=1nIi,j be the number of detected values at concentration ci. The logistic regression model assumes that the observed zi is binomially distributed, Binn,fi, with

fi=11+eβ0β1xi (6)

where xi denotes log2ci. The two unknown parameters β0 and β1 are approximated by maximum likelihood (ML) estimation. The likelihood function is

L=i=1pzinfizi1finzi=i=1pzineβ0y1+β1y2φ (7)

where y1=i=1pzi and y2=i=1pzixi and φ=i=1pnln1+eβ0+β1xi. Setting the derivatives of L with respect to β0 and β1 to zero gives the system of equations for the ML estimate,

y1i=1pn1+eβ0β1xi=0y2i=1pnxi1+eβ0β1xi=0 (9)

This non-linear system of equations is solved by GenEx 6 [11], using a quasi-Newton method. The ML solution will be denoted by βˆ0 and βˆ1. The logistic regression curve is obtained by plotting

tˆ=11+eβˆ0βˆ1x (10)

versus x = log2c. The hat-notation indicates that tˆ is the ML estimate of the exact t=11+eβ0β1x. The observed values y1 and y2 can be considered as samples from a stochastic variable (Y1,Y2) with distribution function ∼eβ0y1+β1y2φ. The moments of (Y1,Y2) are obtained in terms of partial derivatives of φ with respect to β0 and β1, by differentiating the normalization condition

1=Ceβ0y1+β1y2φdy1dy2 (11)

with respect to β0 and β1. C is here a normalization constant. The ML estimate βˆ0 and βˆ1 and hence, the quantity of interest tˆ, can be interpreted as samples from stochastic variables that depend on (Y1, Y2) through the ML equations. The standard error for tˆ Is given by

σˆ=i,jtyi2φβiβjtyj1/2 (12)

In this formula, and in the formulas below we use the convention the limits of the summations are 0 and 1 for all indexes, and that the expression is evaluated at the ML estimated parameters βˆ0 and βˆ1.

The standard 1–2α confidence interval tˆzασˆ, where zαis a percentile of the normal distribution, approximates the exact confidence interval with accuracy On1/2 as the sample size n grows. The parametric bootstrap confidence intervals (Efron [7], Diciccio and Efron [8]) improve on the standard intervals by taking into account higher moments in the normal approximation. There are several variants of bootstrap confidence intervals, but they all approximate the exact confidence interval with accuracy On1. GenEx implements the ABCq confidence intervals described in Diciccio and Efron 1982. ABC stands for Approximate Bootstrap Confidence interval, and the q subscript indicates quadratic form. The formula for the ABCq confidence interval is

tˆqασˆ,tˆ+qα+σˆ (13)

where

qα+=λ+cqλ2andλ=z0+zα1az0+zα2 (14)

The expression for qα is similar but with zα is replaced by −zα. The three parameters a, cq, and z0 are defined as follows. The acceleration,

a=16σˆ3i,j,k3φβiβjβktyityjtyk (15)

where the limits of summation are 0 to 1 for all indexes. The derivatives are evaluated at the ML estimated values βˆ0 and βˆ1. The quadratic coefficient

cq=12σˆ3i,j,k,l2φβiβk2φβjβl2tyiyjtyktyl (16)

And the bias correction

z0=Φ12ΦaΦγˆ (17)

where Φ is the distribution function of the standardized normal distribution. The parameter γˆ is defined by

γˆ=bˆσˆcq, (18)

where the bias estimate is

bˆ=12i,j2φβiβj2tyiyj (19)

Efron [7], [8] makes repeated use of the formula

jvjfxj=ddhfx0+hv0,x1+hv1|h=0 (20)

to rewrite parts of the expressions for a, cq, and bˆ as directional derivatives. Approximating the directional derivative by numerical differentiation is computationally efficient and avoids the possibly lengthy expressions obtained by performing the differentiation exactly.

3. Results

Starting from a calibrated human genome sample (NIST SRM 2372), a 2-fold dilution series of standard samples was prepared covering the range 1–2048 molecules per reaction volume. Each standard sample was analyzed in 64 replicates, except for the most diluted sample, which was analyzed in 128 replicates. Grubb's test identified nine outliers leaving 759 data points for analysis. Fig. 1 plots the measured Cq values versus the expected amounts of molecules per sample (top). The data are fitted to a straight line, which, however, is only to guide the eye, as the low concentration samples are outside the linear range of the qPCR standard curve. Fig. 1 also shows the measured Cq values in a residual plot relative to the fitted straight line (bottom). The plots show how the spread of replicates increases with decreasing amount of target. This is expected due to sampling noise. As the expected number of target copies decreases, the variation across replicates increases. Although other factors also contribute to variation across replicates [14], sampling noise, which can be modeled by the Poisson distribution, is expected to dominate at very low copy numbers [15], [16]. In Fig. 1 it is also seen that the low concentrated samples deviate from the straight line by having somewhat lower Cq values than expected. This bias is due to some of the replicate samples being negative (“non-detect”) and not considered in the plots, resulting in an apparent lower average Cq.

Fig. 1.

Fig. 1

Standards dilution series. Top: 2-fold serial dilution of a calibrated human genomic DNA sample covering the range from 1 to 2048 target molecules on average per sample. Each sample was analyzed in 64 replicates, except for the most diluted sample, which was analyzed in 128 replicates, using the ValidPrime qPCR assay. Bottom: Residual plot of the positive reads.

4. Limit of detection

LoD for qPCR methods can be estimated from analysis of replicate standard curves [10]. From the definition of LoD follows that working at 95% confidence, LoD is the measurand concentration that produces at least 95% positive replicates. Under error free conditions, when only sampling noise would contribute to replicate variation, LoD at 95% confidence is 3 molecules [9]. For most real samples, LoD is also affected by noise contributed to by sampling, extraction, reverse transcription, and qPCR, and may be substantially higher. Fig. 2 plots the fractions of positive replicates versus the number of target molecules per sample for the ValidPrime/gDNA data. From visual inspection, the LoD at 95% confidence is between 2 and 4 target molecules. Fitting the data with the sigmoidal function:

yi=11+eαβlogci, (21)

allows for interpolation, which gives LoD = 2.5 target molecules. This is even slightly below to the theoretical limit caused by sampling noise. The precision of the estimate can be obtained by resampling of the data (Fig. 2):

LoD(with95%confidenceinterval)=2.02.53.7

Fig. 2.

Fig. 2

Limit of detection. Fractions of positive reads obtained with the ValidPrime assay when analyzing samples containing from 1 to 2048 target molecules on average per sample. The measured fractions are fitted to a sigmoidal curve for the estimation of LoD (solid line). At 95% confidence LoD is 2.5 target molecules. Resampling of the data with recurrence (dashed line) allows estimating the 95% confidence interval for the LoD = 2.0 ≤ 2.5 ≤ 3.7 (intersections of the horizontal line at 0.95 and the three sigmoidal curves).

The confidence interval encompasses the theoretical LoD of 3 molecules.

5. Limit of quantification

The LoQ can also be estimated from the replicate standard curves. This is done by calculating the SD for the responses of the replicate samples at the different concentrations. SD of the data can be calculated in either log (Cq values) or linear scale (relative quantities) and does not assume any particular distribution. In each case, SD reflects the average difference of the measured values to the mean in the same scale. In contrary to SD, calculation of the confidence interval assumes normal distribution. For normally distributed data, 68% of the measured values are expected to be within the mean ± 1 SD, and 95% within the mean ± 2 SD. SD is expected to increase with decreasing target concentration due to sampling noise, which alone produces an SD of 0.25 cycles when the average number of target molecules per analyzed aliquot is 35 (Fig. 3). In practice, other factors such as losses due to adsorption to surfaces, decreasing reaction yields at lower concentrations, less efficient reactions, etc., contribute to the error. Calculating SD of the qPCR data in logarithmic scale, i.e., on the Cq values, has the advantage that data usually are normal distributed and confidence intervals are readily calculated. However, for comparison with other measurement techniques the SD should be converted into linear scale and expressed in percentage of the mean, which is known as the relative standard deviation or the coefficient of variation (CV = 100 × SD/mean) which leads to the following expression for the coefficient of variation:

CVln=1+ESDCq2*ln1+E1 (22)

where CVln is the coefficient of variation for log-normal distributed data as expected for concentrations measured in replicates with qPCR, having a qPCR efficiency: E, and standard deviation of replicate Cq-values: SD(Cq).

Fig. 3.

Fig. 3

SD due to sampling ambiguity. Contribution to the standard deviation from sampling ambiguity modeled by the Poisson distribution. Left graph: distribution of target molecules across aliquots from containers with different concentrations modeled by the Poisson distribution. Right graph: SD of Cq values of replicates assuming sampling error only that can be modeled by the Poisson distribution. A container concentration of 35 target molecules on average per aliquot produces SD = 0.25 cycles.

Fig. 4 shows the CV for the human gDNA samples assayed by ValidPrime as a function of the expected number of target molecules in the standard samples. CV increases with decreasing target concentration as expected due to increasing sampling noise and other factors. As target concentration decreases some of the replicate samples starts to be negative (“non-detect”). These concentrations are per definition below the LoQ, and are indicated in red in Fig. 4. There is also a bias toward lower values in the calculated SD, as the non-detect samples cannot be considered.

Fig. 4.

Fig. 4

Limit of Quantification. Coefficient of variation (CV = 100 × SD/mean) of the back-calculated concentrations of the human genome replicate samples analyzed with the ValidPrime assays. Horizontal red dashed line indicates CV = 35% and vertical green dashed line indicates the lowest concentration of samples with a CV below 35%. Red symbols indicate presence of negative (“non-detect”) samples among the replicates. Those are ignored when calculating SD, resulting in a bias toward lower SD values.

There is no general guidance specifying acceptable threshold value of LoQ. What may be reasonable varies from case to case and depends on the complexity of the samples and the required precision in any decisions being supported by the test. At TATAA Biocenter, unless we have other guidance, we specify LoQ as the lowest concentration where replicates show a CV ≤ 35% on back calculated concentrations. Using the ValidPrime method to analyze human gDNA, we find that the lowest amount of gDNA that produces replicates with a CV ≤ 35% is 16 molecules. Hence, LoQ of the ValidPrime method as applied here is 16 molecules.

6. Discussion

The accuracy of LoD and LoQ estimates depends primarily on the concentration increments between the standard samples, while the precision depends primarily on the number of replicates. In fact, LoQ estimates are restricted to those target concentrations/amounts that were contained in the standard sample. In our example the LoQ is 16 molecules, or perhaps less as this satisfies the CV ≤ 35% criteria, but it does not reach 8 molecules, which was the amount of target molecules in the next diluted standard sample. We currently have no reliable means to interpolate the CV versus target amount data to obtain an estimate of LoQ that would be in between 8 and 16 molecules. The number of replicates performed of each standard sample determines the precision in the SD and CV estimates. If these are poor, because of low number of replicates, the LoQ criteria can accidentally be met at a different concentration leading to an erroneous estimate. Occasionally CV rises to over the set threshold (35% in our case) at a certain target amount, then drops to below this threshold at a lower target amount and then rises again. Such fluctuations are due to imprecision in the estimated CV's usually due to too low a number of replicates. In those cases, LoQ should be taken as the lowest target amount higher than any concentration with a CV exceeding the set threshold. In addition to the criteria above, LoQ can never be lower than LoD. Should experiments give such an estimate, LoQ should be reported equal to LoD.

In our ValidPrime/gDNA example, LoD is less than 4 target molecules but does not reach 2, as the standard samples containing in average 4 target molecules have a positive call rate above 95%, while the standard samples containing in average 2 target molecules have a positive call rate below 95%. By interpolating the measured call rates as a function of the average number of target molecules, we obtain a more precise estimate of LoD. For the ValidPrime/gDNA we estimate LoD by interpolation to 2.5 molecules. This is close to the expected LoD at a 95% positive call rate of 3 molecules predicted by the Poisson distribution describing sampling noise, suggesting that other contributions to noise for the ValidPrime/gDNA qPCR analysis are negligible.

If sufficient number of replicates at each standard's concentration is available, the confidence interval of the LoD estimate can be obtained by resampling of the data with recurrence. A fairly high number of replicates, often at least 20, at each concentration are needed to get convergence. The LoD confidence interval is asymmetric, with a smaller range toward lower target concentrations. For the ValidPrime method analyzing human gDNA, the LoD for 95% positive call rate estimated with a 95% confidence range was: LoD = 2.0 ≤ 2.5 ≤ 3.7 molecules.

The concentration range covered by the standard samples has very low impact on the estimated LoD and LoQ. For the LoQ estimate, only the lowest target amount with a CV below the set threshold and the highest target amount with a CV above the set threshold are used. The initial rough LoD estimate is also based on only two samples: the sample with the lowest target amount that produces positive reads at a rate above the set criteria (typically 95%) and the sample with highest target amount that produces positive reads at a rate below the set criteria. A more precise estimate of LoD can be obtained by interpolating the measured positive rates taking into account more standard samples. But even in this case only a small number of standard samples is considered, as only those that produce fractional positive rates (i.e., above 0 and below 100%) contribute appreciably to the fitting. Therefore, from a practical point of view and cost performance, it is better to narrow the target concentration range of the standards to just cover the expected LoD and LoQ and increase the number of replicates to improve precision. In practice, LoD and LoQ are not known ahead of the experiment. A pragmatic approach is then to obtain an initial rough estimate of LoD and LoQ, perhaps as part of the regular standard curve that is measured when establishing any new method to estimate the PCR efficiency, repeatability and dynamic range, and then in a second experiment narrow down the concentration range and increase the number of replicates for more precise estimates of LoD and LoQ.

Acknowledgement

This project was supported by CZ.1.05/1.1.00/02.0109 BIOCEV provided by ERDF and MEYS and by the European IMI project CANCER-ID.

Handled by Jim Huggett

Footnotes

1

For research use only. The products of LGC Douglas Scientific are not for use in diagnostic procedures.

References

  • 1.Shrivastava Alankar, Gupta Vipin B. Methods for the determination of limit of detection and limit of quantitation of the analytical methods. Chronicles Young Scientists. 2011;2(1):21. [Google Scholar]
  • 2.EP17-A, C.L.S.I . CLSI; Wayne PA: 2004. Protocols for Determination of Limits of Detection and Limits of Quantitation; Approved Guideline. [Google Scholar]
  • 3.Armbruster David A., Terry Pry Limit of blank, limit of detection and limit of quantitation. Clin. Biochem. Rev. 2008;29(Suppl 1):S49–S52. [PMC free article] [PubMed] [Google Scholar]
  • 4.Kubista Mikael. The real-time polymerase chain reaction. Mol. Aspects Med. 2006;27(2):95–125. doi: 10.1016/j.mam.2005.12.007. [DOI] [PubMed] [Google Scholar]
  • 5.Laurell Henrik. Correction of RT–qPCR data for genomic DNA-derived signals with ValidPrime. Nucl. Acids Res. 2012:gkr1259. doi: 10.1093/nar/gkr1259. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Limpert Eckhard, Stahel Werner A., Abbt Markus. Log-normal distributions across the sciences: Keys and clues on the charms of statistics, and how mechanical models resembling gambling machines offer a link to a handy way to characterize log-normal distributions, which can provide deeper insight into variability and probability—normal or log-normal: That is the question. BioScience. 2001;51(5):341–352. [Google Scholar]
  • 7.Efron Bradley. Better bootstrap confidence intervals. J. Am. Stat. Assoc. 1987;82(397):171–185. [Google Scholar]
  • 8.Diciccio Thomas, Efron Bradley. More accurate confidence intervals in exponential families. Biometrika. 1992;79(2):231–245. [Google Scholar]
  • 9.Ståhlberg Anders, Kubista Mikael. The workflow of single-cell expression profiling using quantitative real-time PCR. Expert Rev. Mol. Diagnost. 2014;14(3):323–331. doi: 10.1586/14737159.2014.901154. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Burns M., Valdivia H. Modelling the limit of detection in real-time quantitative PCR. Eur. Food Res. Technol. 2008;226(6):1513–1524. [Google Scholar]
  • 11.Bustin S.A., Benes V., Garson J.A., Hellemans J., Huggett J., Kubista M., Mueller R., Nolan T., Pfaffl M.W., Shipley G.L., Vandesompele J. The MIQE guidelines: minimum information for publication of quantitative real-time PCR experiments. Clin. Chem. 2009 Apr 1;55(4):611–622. doi: 10.1373/clinchem.2008.112797. [DOI] [PubMed] [Google Scholar]
  • 12.Huang S., Wang T., Yang M. The evaluation of statistical methods for estimating the lower limit of detection. Assay Drug Develop. Technol. 2013 Feb 1;11(1):35–43. doi: 10.1089/adt.2011.438. [DOI] [PubMed] [Google Scholar]
  • 13.Grubbs F.E. Procedures for detecting outlying observations in samples. Technometrics. 1969 Feb 1;11(1):1–21. [Google Scholar]
  • 14.Ruijter J.M., Villalba A.R., Hellemans J., Untergasser A., van den Hoff M.J. Removal of between-run variation in a multi-plate qPCR experiment. Biomol. Detect. Quantification. 2015 Sep 30;5:10–14. doi: 10.1016/j.bdq.2015.07.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Tellinghuisen J., Spiess A.N. Bias and imprecision in analysis of real-time quantitative polymerase chain reaction data. Analyt. Chem. 2015 Aug 13;87(17):8925–8931. doi: 10.1021/acs.analchem.5b02057. [DOI] [PubMed] [Google Scholar]
  • 16.Tellinghuisen J., Spiess A.N. Absolute copy number from the statistics of the quantification cycle in replicate quantitative polymerase chain reaction experiments. Analyt. Chem. 2015 Jan 23;87(3):1889–1895. doi: 10.1021/acs.analchem.5b00077. [DOI] [PubMed] [Google Scholar]
  • 17.Nutz S., Döll K., Karlovsky P. Determination of the LOQ in real-time PCR by receiver operating characteristic curve analysis: application to qPCR assays for Fusarium verticillioides and F. proliferatum. Analyt. Bioanalyt. Chem. 2011 Aug 1;401(2):717–726. doi: 10.1007/s00216-011-5089-x. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Biomolecular Detection and Quantification are provided here courtesy of Elsevier

RESOURCES