An official website of the United States government
Here's how you know
Official websites use .gov
A
.gov website belongs to an official
government organization in the United States.
Secure .gov websites use HTTPS
A lock (
) or https:// means you've safely
connected to the .gov website. Share sensitive
information only on official, secure websites.
As a library, NLM provides access to scientific literature. Inclusion in an NLM database does not imply endorsement of, or agreement with,
the contents by NLM or the National Institutes of Health.
Learn more:
PMC Disclaimer
|
PMC Copyright Notice
. Author manuscript; available in PMC: 2022 May 19.
Published in final edited form as: ACS Synth Biol. 2022 Mar 10;11(4):1699–1704. doi: 10.1021/acssynbio.2c00082
Correction to GAMES: A dynamic model development workflow for
rigorous characterization of synthetic genetic systems
After publication, we identified two minor mistakes in the code for the GAMES
workflow. Each subtly affected our analysis of the example case study but had no effect
on the GAMES workflow. Here we describe those mistakes and discuss how the resulting
corrections modify the case study analysis.
Correction 1:
There are two locations in the workflow where noise is added to each data
point in the training data: to generate the parameter estimation method (PEM)
evaluation data, and to calculate the threshold used for the parameter profile
likelihood (PPL). In both locations, the value of the noise added to each data point
is selected from a distribution centered at zero with a standard deviation equal to
the standard error associated with the data point (assuming, for our case study,
that the data point represents the mean value for three replicates). We set the
standard deviation, , for each data point to 0.05, so the standard
error, , is calculated with Equation C1, where is the number of replicates:
(Equation C1)
The original published code omitted the square root operation
inEquation C1, such that the
distribution from which each added noise value was drawn was slightly smaller than
it would have been with the correct distribution. We corrected the code in v1.0.2
and repeated the relevant PEM evaluation and PPL simulations. We noted minor changes
in the results that impact the threshold of the PPL in the example study and
affected the identifiability classification of one parameter for one model in the
example study. This change impacted the interpretation of this one
parameter and the quantitative values of the PPL threshold for all parameters.
The other qualitative interpretations remain the same, and no changes were made
to the GAMES workflow.
With the correction, the PEM evaluation results are very similar to the
previous results. The threshold used to define the PEM evaluation criterion remains
the same (R2 = 0.99), and this threshold is satisfied for all models
(Figure 4c model A, Figure S5a model B, Figure S6a model C, Figure S7a model C).
With the correction, the main difference observed for the PPL results is that
the calculated thresholds for each model are higher than in the original analysis.
This change is consistent with our understanding of the PPL threshold, which is
related to the extent of overfitting that is possible given a model, a training data
set, and the associated measurement error. For models A (Figure 7b, Figure 8b,
Figure S4a) and B
(Figure 8b, Figure S5c), the increased threshold is
the only substantial difference between the corrected results and previous results;
all PPL shapes and parameter classifications remain the same.
The corrected PPL shapes and parameter classifications for model C were also
in agreement with the previous results (Figure
8b, Figure S6c),
with the exception of the parameter m*, which now appears
practically unidentifiable—whereas previously this parameter was deemed
identifiable—as the PPL reaches the threshold in the negative direction but
not in the positive direction. However, this corrected result has no impact on
downstream analysis because m* is still classified as identifiable
for the final model (model D). The classification of m* as
practically unidentifiable for model C is reasonable given that the increased PPL
threshold necessitates that higher m* values be traversed when
determining the PPL. As m* is a ratio between the parameters
m and b, once m* reaches a
sufficiently high value such that m >>
b, increasing m* further has no meaningful
effect on the agreement between the training data and simulated data. This
interpretation explains why m* does not reach the threshold in the
positive direction for model C with the correction included here.
The corrected results for model D are very similar to the previous results
(Figure 8b, Figure S7c). All parameter
classifications remain the same, and all parameters are identifiable. The
qualitative shape of the PPL for m* is similar to the shape
observed for m* in model C (with the correction), but in model D,
the PPL crosses the threshold in both the negative and positive directions. This is
reasonable because model D has fewer free parameters (four free parameters) than
does model C (five free parameters), and therefore model D has a lower calculated
PPL threshold, enabling the PPL for m* to cross the threshold in
the positive direction.
Correction 2:
We also noted a minor mistake in the model D case study. For model D,
an incorrect value for kbind(the fixed value of 1 rather than the reference value of 0.05) was used
to define the reference parameter set and calculate the PPL
threshold. This mistake was corrected before generating the
simulation results reported here. Correcting this value led to some parameter sets
having higher values than values (Figure S7c) because
kbind cannot be fit to the reference parameter value
for each noise realization. However, the resulting reduced model with
kbind = 1 still yields very similar agreement
between the training data and simulated data (Figure S7b), which shows that fixing
kbind to 1 (and not to the reference value of 0.05,
which would be unknown in a practical situation when the reference parameters do not
exist) does not significantly affect the results. This phenomenon, in which some
parameter sets have slightly higher values than values, was also observed in the original results
for all models but to a lesser extent. In general, slightly negative values for
-can be attributed to the optimization algorithm
finding local minima (that have only slightly different values than the global minimum) to define
for some noise realizations.
Conclusions:
these corrections affected our interpretation of the example case study but
had no effect on the GAMES workflow itself. The code used to define the case study
example has been updated and annotated on GitHub: https://github.com/leonardlab/GAMES.
Supplementary Material
Figure S4. Additional PPL results for Model A.
(a) Evaluation of the distribution via a simulation study. 1000
individual noise realizations were generated. Parameters were individually
estimated for each of the noise realizations to calculate
. Reference parameters were used to
calculate . The difference between these values,
, represents the amount of overfitting for
each noise realization. (blue dotted line) was determined by
evaluating the 99% confidence interval ( = 0.01) of the distribution
(. (b) Three-dimensional plot of
km, b, and
m along the unidentifiability associated with
m. The surface is smooth, indicating dependencies
between the three parameters. The logarithm
(log10()) of each parameter is plotted.
Figure S5. PEM evaluation, parameter estimation, and determination of confidence
threshold for Model B.
(a) PEM evaluation criterion with 1000 parameter sets
in the global search and 100 initial guesses. Results are shown only for
parameter sets yielding R2
0.90. The PEM evaluation criterion is
satisfied. (b) Best fit to the training data using the
calibrated parameter set. The visual inspection criterion is satisfied.
Parameter values are in Supplementary Table 2. (c)
Determination of the confidence threshold for PPL calculations
( = 7.7).
Figure S6. PEM evaluation, parameter estimation, and determination of confidence
threshold for Model C.
(a) PEM evaluation criterion with 1000 parameter sets
in the global search and 100 initial guesses. The PEM evaluation criterion
is satisfied. (b) Best fit to the training data using the
calibrated parameter set. The visual inspection criterion is satisfied.
Parameter values are in Supplementary Table 2. (c)
Determination of the confidence threshold for PPL calculations
( = 7.9).
Figure S7. PEM evaluation, parameter estimation, and determination of confidence
threshold for Model D.
(a) PEM evaluation criterion with 1000 parameter sets
in the global search and 100 initial guesses. The PEM evaluation criterion
is satisfied. (b) Best fit to the training data using the
calibrated parameter set. The visual inspection criterion is satisfied.
Parameter values are in Supplementary Table 2. (c)
Determination of the confidence threshold for PPL calculations
( = 7.0).
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Figure S4. Additional PPL results for Model A.
(a) Evaluation of the distribution via a simulation study. 1000
individual noise realizations were generated. Parameters were individually
estimated for each of the noise realizations to calculate
. Reference parameters were used to
calculate . The difference between these values,
, represents the amount of overfitting for
each noise realization. (blue dotted line) was determined by
evaluating the 99% confidence interval ( = 0.01) of the distribution
(. (b) Three-dimensional plot of
km, b, and
m along the unidentifiability associated with
m. The surface is smooth, indicating dependencies
between the three parameters. The logarithm
(log10()) of each parameter is plotted.
Figure S5. PEM evaluation, parameter estimation, and determination of confidence
threshold for Model B.
(a) PEM evaluation criterion with 1000 parameter sets
in the global search and 100 initial guesses. Results are shown only for
parameter sets yielding R2
0.90. The PEM evaluation criterion is
satisfied. (b) Best fit to the training data using the
calibrated parameter set. The visual inspection criterion is satisfied.
Parameter values are in Supplementary Table 2. (c)
Determination of the confidence threshold for PPL calculations
( = 7.7).
Figure S6. PEM evaluation, parameter estimation, and determination of confidence
threshold for Model C.
(a) PEM evaluation criterion with 1000 parameter sets
in the global search and 100 initial guesses. The PEM evaluation criterion
is satisfied. (b) Best fit to the training data using the
calibrated parameter set. The visual inspection criterion is satisfied.
Parameter values are in Supplementary Table 2. (c)
Determination of the confidence threshold for PPL calculations
( = 7.9).
Figure S7. PEM evaluation, parameter estimation, and determination of confidence
threshold for Model D.
(a) PEM evaluation criterion with 1000 parameter sets
in the global search and 100 initial guesses. The PEM evaluation criterion
is satisfied. (b) Best fit to the training data using the
calibrated parameter set. The visual inspection criterion is satisfied.
Parameter values are in Supplementary Table 2. (c)
Determination of the confidence threshold for PPL calculations
( = 7.0).