Skip to main content
NPJ Genomic Medicine logoLink to NPJ Genomic Medicine
. 2020 Oct 23;5:47. doi: 10.1038/s41525-020-00154-9

Best practices for the analytical validation of clinical whole-genome sequencing intended for the diagnosis of germline disease

Christian R Marshall 1,, Shimul Chowdhury 2, Ryan J Taft 3, Mathew S Lebo 4,5, Jillian G Buchan 6,14, Steven M Harrison 4,5, Ross Rowsey 7, Eric W Klee 7,8, Pengfei Liu 9, Elizabeth A Worthey 10,15, Vaidehi Jobanputra 11,12, David Dimmock 2, Hutton M Kearney 7, David Bick 10, Shashikant Kulkarni 9,13, Stacie L Taylor 3, John W Belmont 3, Dimitri J Stavropoulos 1, Niall J Lennon 5; Medical Genome Initiative
PMCID: PMC7585436  PMID: 33110627

Abstract

Whole-genome sequencing (WGS) has shown promise in becoming a first-tier diagnostic test for patients with rare genetic disorders; however, standards addressing the definition and deployment practice of a best-in-class test are lacking. To address these gaps, the Medical Genome Initiative, a consortium of leading healthcare and research organizations in the US and Canada, was formed to expand access to high-quality clinical WGS by publishing best practices. Here, we present consensus recommendations on clinical WGS analytical validation for the diagnosis of individuals with suspected germline disease with a focus on test development, upfront considerations for test design, test validation practices, and metrics to monitor test performance. This work also provides insight into the current state of WGS testing at each member institution, including the utilization of reference and other standards across sites. Importantly, members of this initiative strongly believe that clinical WGS is an appropriate first-tier test for patients with rare genetic disorders, and at minimum is ready to replace chromosomal microarray analysis and whole-exome sequencing. The recommendations presented here should reduce the burden on laboratories introducing WGS into clinical practice, and support safe and effective WGS testing for diagnosis of germline disease.

Subject terms: Genetic testing, Next-generation sequencing, Laboratory techniques and procedures, Genetic testing

Introduction

Advances in next-generation sequencing (NGS) over the past decade have transformed genetic testing by increasing diagnostic yield and decreasing the time to reach a diagnosis15. Targeted NGS multigene panels have come into widespread use and whole-exome sequencing (WES) is a powerful aid in the diagnosis of patients with nonspecific phenotypic features610 and critically ill neonates11, where the differential diagnosis often includes multiple rare genetic disorders12. These approaches, however, have both workflow and test content limitations that may constrain their overall efficacy.

Whole-genome sequencing (WGS) can address many of the technical limitations of other enrichment-based NGS approaches, including improved coverage13,14, and sensitivity for the detection of structural and complex variants15. WGS also enables the identification of noncoding variants, such as pathogenic variations disrupting regulatory regions, noncoding RNAs, and mRNA splicing1618. Emerging uses of WGS include HLA genotyping19, pharmacogenetic testing20, and generation of polygenic risk scores21. Several studies have demonstrated the advantages of WGS for the identification of clinically relevant variants in a wide range of cohorts2226, and have shown the diagnostic superiority of WGS compared with conventional testing in pediatric patients2729 and critically ill infants30,31. As a more efficient test, WGS is poised to replace targeted NGS or WES and chromosomal microarray (CMA), as a first-line laboratory approach in the evaluation of children and adults with a suspected genetic disorder28,32,33. WGS also has the benefit of periodic reanalysis across multiple variant types, which will increase diagnostic efficacy through updated annotation and analysis techniques34.

Although the stage is set for widespread adoption of clinical WGS, technical challenges remain, and standards that address both the definition and the deployment practices of a best-in-class clinical WGS test have not been fully defined. Professional bodies have made progress in providing guidance for clinical WGS test validation35,36, and best practices for benchmarking with reference standards and recommended accuracy measures are beginning to emerge3739. It is important to note, however, that these recommendations do not address the specific challenges related to the setup of clinical WGS.

Scope and methods

To address these challenges, a working group comprised of experts from the Medical Genome Initiative40 was created to develop practical recommendations related to the analytical validation of clinical WGS. We decided to focus on the use of a clinical WGS test for the diagnosis of germline disease and that other applications of WGS (such as testing for somatic variants or cell-free circulating DNA) were considered out of scope. As many of the basic principles of laboratory test validation also apply to WGS, this document is not meant to provide a comprehensive description of all the steps of laboratory test validation, but to rather focus on the specific challenges posed by clinical WGS validation.

To identify areas of group consensus and ultimately develop practical recommendations for clinical laboratories, a survey was created that queried working group members on key topics related to analytical validation, including their own current laboratory practices. Biweekly teleconference meetings over a 12-month period were held to share and discuss these current practices, and determine where consensus could be attained. Notably, finding consensus was often difficult due to the variability in validation approaches and the wide range of quality control metrics used among the laboratories. Nonetheless, these recommendations provided herein are meant to aid laboratory personnel who wish to introduce WGS into clinical practice and, more importantly, to support safe and effective WGS testing for diagnosis of germline disease.

Overview of clinical whole-genome sequencing

All clinical diagnostic testing, including WGS, encompasses the entire process from obtaining a patient specimen to the delivery of a clinical report. The technical and analytical elements of clinical WGS can be separated into three stages: sample preparation, including extraction and library preparation followed by sequence generation (primary); read alignment and variant detection (secondary); and annotation, filtering, prioritization, variant classification, and case interpretation followed by variant confirmation, segregation analysis, and finally reporting (tertiary)41 (Fig. 1). These components are common to all high-throughput sequencing tests and informatics pipelines, but differences in components (e.g., informatics algorithms) will result in differences in data quality and accuracy. The focus of this manuscript is the primary and secondary analyses, as these steps relate directly to the evaluation of test performance for the analytical validation of clinical WGS. Elements critical to establishing analytical validity are described below in three sections: (1) test development and optimization, (2) test validation, and (3) ongoing quality management of the test in clinical use. Major steps and activities in the analytical validation are shown in Fig. 2 with key definitions in Box 1. A summary of the key points and recommendations embedded within each of these sections, as well as future considerations can be found in Table 1.

Fig. 1. Clinical whole-genome sequencing workflow.

Fig. 1

The workflow for clinical WGS involves three major analysis steps spanning wet laboratory and informatics processes: primary (blue) analysis refers to the technical production of DNA sequence data from biological samples through the process of converting raw sequencing instrument signals into nucleotides and sequence reads; secondary (green) analysis refers to the identification of DNA variants through read alignment and variant calling; and tertiary (yellow) analysis refers to variant annotation, filtering and prioritization, classification, interpretation, and reporting. Health record information and phenotype can be mined and converted to Human Phenotype Ontology (HPO) terms to aid variant interpretation. Primary analysis involves the sample, and library preparation and sequencing with base calling followed by extensive quality control (QC). During this stage, genotyping with an orthogonal method (SNP-array or targeted assay) is performed for QC purposes. Secondary analysis involves mapping, read alignment, and variant calling. Different classes of variation (SNVs, SV, CNVs, mitochondrial, and repeat expansions) will use different algorithms that can be run in parallel. Aside from QC of alignment and variant calling, the orthogonal genotyping can be used to ensure no sample mix-up has occurred throughout the workflow. Tertiary analysis begins with the annotation of variants followed by filtering, prioritization, and variant classification depending on the phenotype and clinical indication for testing. Classification of variants according to ACMG guidelines may be automated, but the final interpretation involves human intervention and will ultimately be driven by the case phenotype. Variants are reported based on relevance to the primary indication for testing and secondary, or incidental findings not associated with the reason for testing following any necessary confirmation method. Confirmation may be performed with an orthogonal wet laboratory method or in silico examination of the data based on how the test was validated. Clinical correlation (pink) is performed by the ordering physician, which may involve iterative feedback and collaboration with the laboratory (dotted arrows). Throughout the process, collection of aggregate data will be necessary to generate internal allele frequencies and for sharing of interpreted data with repositories.

Fig. 2. Key steps in the analytical validation of a clinical WGS test.

Fig. 2

Key steps in the analytical validation of clinical WGS include test development optimization, test validation, and quality management. Each step involves activities that lead to defined outcomes.

Table 1.

Summary of key questions and recommendations for the analytical validation of whole-genome sequencing.

Current state/question Consensus Recommendation Comments and future outlook
Test development and optimization Test definition considerations
What variant classes should be reported in a clinical WGS test? Yes

• A clinical whole-genome sequencing test should aim to analyze and report on all detectable variant types (Table 1 and Supplementary Fig. 1)

• At a minimum, we recommend reporting on SNVs, indels, and CNVs.

• Test definitions are evolving and laboratories should further aim to offer reporting of mitochondrial variants, repeat expansions, some structural variants, and selected clinically relevant genes whose analytical assessment is made difficult by pseudogenes or highly homologous sequence.
Test performance considerations
What comparisons are necessary for WGS to replace other tests such as CMA or WES? Yes

• Clinical WGS test performance should meet or exceed that of any tests that it is replacing.

• Current evidence suggests WGS is analytically sufficient to replace WES and CMA (Table 1, Supplementary Table 1 and Supplementary Fig. 3.

• If clinical WGS is deployed with any established gaps in performance compared to current gold standard tests, it should be noted on the test report.

• Robust detection by WGS of some variants is either not equivalent (mosaic SNVs) or has yet to be established (e.g., repeat expansions), but should still be included in the test definition as long as limitations in test sensitivity are defined.

Which variant types should be confirmed before reporting on a WGS test? Yes

• The laboratory should have a strategy in place to define which variants need confirmatory testing before reporting.

• Until the accuracy of more complex variants (SVs, REs, etc.) is equivalent to currently accepted assays, confirmation with an orthogonal method is necessary before reporting.

• As algorithms improve for complex variant calling and data are acquired to support WGS accuracy compared to currently accepted assays, it is expected that orthogonal confirmation will not be necessary.
Upfront considerations for test design
How should labs define and evaluate high-quality genome coverage and callability? Some

• Metrics that measure genome completeness should be used to define the performance of clinical WGS, and include overall depth and evenness of coverage.

• These should be monitored with respect to callable regions of the genome and the related calling accuracy for each variant type compared to orthogonally investigated truth sets.

• An assessment of callable regions should use depth of coverage, base quality, and mapping quality.

• Although consensus was achieved on the concept of the evaluation of genome coverage and callability, consistent methodology, and universal cutoffs could not be established. Expected and suggested values, and ranges are shown in Table 3.
What reference standards and positive controls should be used for performing clinical WGS validation? Yes

• Reference standards are useful for the evaluation of calling accuracy across variant type, size, and location.

• Analytical validation of clinical WGS should include publicly available reference standards (e.g., NIST and platinum genomes) in addition to commercially available and laboratory-held positive controls for each variant type (see Supplementary Table 3)

• Reference standards are not sufficient on their own for validation and laboratory-held positive controls derived from the same specimen type should also be used.
How many controls should be used for variant types commonly addressed by the field vs those were standards are still evolving (REs)? Yes

• For commonly addressed variants like small variants, a low minimal number of controls can be utilized if they include well-accepted reference standards.

• For more complex and emerging variant types like RE, a large number of samples are necessary (see Supplementary Table 3).

• As algorithms and reference standard datasets improve, it is expected that fewer samples will be needed for test validation.
Test validation Performance metrics, variant type, and genomic context
What factors should be considered when deciding which metrics to evaluate during test validation? Some • The analytical framework should include metrics that account for genome complexity, with special attention to sequence content and variant type. • Small variants and CNVs have different calling constraints that can be affected differently by low-complexity sequence.
Which performance metrics should be utilized to ensure the accuracy of clinical WGS? Some

• Use of GA4GH and FDA recommendations for sensitivity or PPA and precision or TPPV, and lower bound of the 95% CI when truth sets are available.

• Repeatability, reproducibility, and limits of detection should be tested.

• PPA and NPA against positive controls assessed using a precedent technology is recommended when standard truth sets are not available (see Supplementary Table 1 for example).

• Precision or TPPV is a more useful metric than specificity due to the large number of true negatives expected from clinical WGS.
What limitations affecting variant calling performance should be considered for a WGS test? Some

Variant class:

• Mosaics: limitation for detection of mosaic variants is low due to low overall read depth.

• Structural variants: limitations in the detection of balanced changes (translocation and inversions).

• Repeat expansions: limitation in accurately determining expanded sizes.

Variant size:

• Indels/SV: limitations of calling accuracy for larger indels >20 bp and smaller SVs <1000 bp.

Genomic context:

• Limitations of accuracy for all variant types in regions of high homology, low complexity, and other technically challenging regions.

• Consensus on limitations for calling performance by variant type was achieved; however, consensus on specific definitions related to variant size and genomic context was not.

• Regions that are defined as technically challenging should be documented and made available to those ordering the test.

Sample number and type for validation
How many samples and what types are needed for validation? Some

• Reference standards are sufficient to assess global accuracy of small variants, but should be supplemented with positive controls.

• Beyond small variants, more positive controls are needed; these should reflect the most common pathogenic loci or variants.

• Little consensus on the number of positive controls needed.

• Number of samples needed for emerging uses of clinical WGS will evolve as more laboratories validate clinical WGS.

Quality management Control samples
What should be included in ongoing quality control of a clinical WGS test? Yes

• Identification of a comprehensive set of performance metrics and continual monitoring.

• Use of positive controls on a periodic basis.

Less reliance on positive controls for each run; more on the continual monitoring of defined metrics.
Sequencing quality and performance metrics
What sequencing metrics should be monitored for performance? No

• General agreement in some of the sequencing and analysis metrics that should be used for pass/fail (Table 3).

• Unable to reach consensus on which metrics should be used for sample level QC and monitoring, and the corresponding thresholds that need to be met.

Standards for metrics calculations should allow for consensus on thresholds.

Box 1. Definitions of key terms.

Term Definition
Analytical validity A measure of the accuracy with which a test predicts a genetic change.
Callable region (callability) Regions of the genome where accurate single-nucleotide variant genotype can be reliably derived. Typically expressed as a percentage of non-N reference calls with a passing genotype across a target (whole genome, OMIM genes).
Completeness Proportion of the genome, or a select region of interest (e.g., exons), that have sufficient, high-quality sequencing reads to enable identification of variants.
Negative percent agreement (NPA) Equivalent to specificity. The proportion of correct calls in the absence of a variant, reflecting the frequency of false positives (FP). Calculated as the number of true negatives (TN) detected divided by the total number of positions where a variant is absent (TN plus FP).
No-call or invalid call A position within the testing interval where no variant call can be been made.
Orthogonal confirmation Verification of a specific variant call using a different testing modality.
Positive percent agreement (PPA) Equivalent to recall/sensitivity. Ability of the test to correctly identify variants that are present in a sample, reflecting the frequency of false negatives (FN). Calculated as the number of known variants detected (true positives; TP) divided by the total number of known variants tested (TP plus FN).
Precision Equivalent to TPPV. The fraction of variant calls that match the expected, reflecting the number of FP per test. Calculated as the number of TP divided by the total number of positive calls made (TP plus FP).
Predicted zygosity In diploid organisms, one allele is inherited from the male parent and one from the female parent. Zygosity is a description of whether those two alleles have identical or different DNA sequences.
Read depth A measure of the number of sequence reads that are aligned to a specific base or locus.
Repeatability The percent agreement between the results of successive tests carried out under the same conditions of measurement.
Reproducibility The percent agreement between the results of tests under a variety (e.g., different operators, machines, and time frames).
Sensitivity or recall Equivalent to PPA. Ability of the test to correctly identify variants that are present in a sample, reflecting the frequency of FN. Calculated as the number of known variants detected (TP) divided by the total number of known variants tested (TP plus FN).
Specificity Equivalent to NPA. The proportion of correct calls in the absence of a variant, reflecting the frequency of FP. Calculated as the number of TN detected divided by the total number of positions where a variant is absent (TN plus FP).
Technical positive predictive value (TPPV) Equivalent to precision. The fraction of variant calls that match the expected, reflecting the number of FP per test. Calculated as the number of TP divided by the total number of positive calls made (TP plus FP).

Test development and optimization

There are several components of clinical WGS test design that should be taken into consideration as part of test development and optimization. Here, we focus our discussion on some of the unique aspects of clinical WGS, including the test’s definition, test performance comparisons to current methodologies, and upfront considerations for test design. Other components such as sample and library preparation, sequencing methodology, sequence analysis, and annotation are discussed in more detail in the Supplementary Discussion.

Test definition considerations

Analytical validation requirements will vary based on test definition, which includes both technical considerations and the intended use in a patient population. Although clinical WGS may be used for multiple indications (e.g., inherited disorders, cancer, and healthy individuals), this document focuses on using clinical WGS for individuals with a suspected monogenic germline disorder as the primary use case. The principles of analytical validity described here, however, are applicable to all uses of clinical WGS.

Establishing a test definition for clinical WGS designed to diagnose germline disorders can be challenging for laboratories due to the complexity of the test. Clinical WGS tests are predicated on a specific test definition that delineates both the variant types to be reported and the regions of the genome that will be interrogated (including any limitations), which may vary depending on the variant type. The challenge, due to the comprehensive nature of WGS variant detection, is whether the test definition should be agnostic to phenotype and based on the classes of variants detected or defined for a specific phenotype, since specific loci can be interrogated and reported. The most effective use of genome sequencing at this time is in the evaluation of clinical presentations with a broad range of potential genetic etiologies. However, since it is possible to interrogate specific loci and associated variant types with clinical WGS (e.g., SMN1 deletions for SMA or FMR1 expansions for Fragile X), test definitions will broaden in scope and evolve as analytical performance improves.

Classes of clinically relevant genetic variation detectable by clinical WGS are summarized in Table 2, and include single-nucleotide variants (SNVs), small deletions, duplications, insertions (indels), structural variation (SV), including copy number variation (CNV) and balanced rearrangements, mitochondrial (MT) variants, and repeat expansions (REs)15. The accuracy of detection for some of these variant classes is well established, whereas other classes are technically possible but data demonstrating sufficient detection accuracy are still emerging. A clinical WGS test should aim, wherever possible, to analyze and report on all possible detectable variant types. We recommend SNVs, indels, and CNVs as a viable minimally appropriate set of variants for a WGS test. Laboratories should further aim to offer reporting of MT variants, REs, some structural variants, and selected clinically relevant genes whose analytical assessment is made difficult by pseudogenes or highly homologous sequence (Table 1 and Supplementary Fig. 1). We note that laboratories may not be able to validate all classes of variation prior to initial launch of clinical WGS, and that a phased approach to validation and subsequent test offering may be necessary. Ultimately, the laboratory must provide clear test definitions and identify factors affecting reportable variant types to ordering physicians. For example, if using a specimen source expected to yield limited DNA quantity, PCR for library preparation may be required, and reporting of CNVs42 and REs43 will be adversely affected.

Table 2.

Variant types detectable and reportable from clinical WGS.

Variant type Gene(s) (if applicable) Disorder(s) References (if applicable)

SNVs and small insertions and deletionsa

(1–50 base pairs)

N/A Heritable disease

Zook et al.37

Eberle et al.58

Copy number variationa (deletions and duplications) N/A Heritable disease including known microdeletion/duplication syndromes

Gross et al.33

Stavropoulos et al.27

Lindstrand et al.45

Mitochondrial variationb (SNVs, deletions, duplications, and heteroplasmy of at least 5%) N/A Known mitochondrial disorders Duan et al.56
Structural variantsb N/A Heritable disease including those caused by translocations, inversions, and other genomic rearrangements Lindstrand et al.45
Repeat expansionsc FMR1 Fragile X and related disorders Dolzhenko et al.43
HTT Huntington disease
SCA1 Spinocerebellar ataxia 1
DMPK Myotonic dystrophy 1
C9orf72 Amyotrophic lateral sclerosis
Selected pseudogenesc SMN1 and SMN2 Spinal muscular atrophy Chen et al.57
CYP21A2 21-Hydroxylase deficiency
CYP2D6 Codeine sensitivity
HBA1 and HBA2 Alpha thalassemia
PMS2 Colorectal cancer
PKD1 Polycystic kidney disease 1

aRecommended minimum variant types for clinical validation of WGS. Copy number variation is defined here as unbalanced changes (deletions and duplications) that are at the resolution of chromosomal microarray analysis.

bSome initiative groups have clinically validated. Structural variants are defined here as any genomic alteration >50 base pairs, including balanced and unbalanced changes.

cExamples of targeted loci that could be validated and reported as part of a clinical WGS test.

Test performance considerations

Regardless of the variant types a laboratory may choose to report, a thorough performance comparison between the WGS test and any current testing methodology is warranted to demonstrate that the analytical performance is sufficient for clinical use. Clinical WGS test performance should aim to meet or exceed that of any tests that it is replacing. If clinical WGS is deployed with any established gaps in performance compared to current reference standard tests, it should be noted on the test report (see Table 1). The most immediate and obvious use of clinical WGS is replacement of genome-wide tests, such as WES and CMA. WGS has been shown to be analytically superior to WES for the detection of variants affecting protein function32,44, and there is emerging evidence that the analytical detection of CNVs from WGS is at least equivalent to CMA27,33,45 (Supplementary Table 1).

For the detection of some variant types, it is important to acknowledge that clinical WGS may not be equivalent to current methods and that robust detection has yet to be established. For example, detection of low-level mosaicism represents an important limitation of clinical WGS (at 40× mean depth) compared to WES or targeted panels, where loss of performance may be a significant issue for some indications (e.g., epileptic encephalopathy)46. As previously mentioned, although more complex variant types like those mentioned above (e.g., MT variants with varying levels of heteroplasmy, REs, etc.) can be identified using WGS, we recognize that in some cases the detection accuracy of these variant types may not yet be equivalent to currently accepted assays. There is still inherent value to including these variant classes in the test definition of clinical WGS to ensure as complete a test as possible, as long as limitations in test sensitivity are clearly defined. As with any genetic assay, the test definition should clearly state that a negative report in these instances does not preclude a diagnosis. Laboratories planning to report on complex variant types must include the test limitations in the report, and have a detailed confirmatory test strategy in place. It is consensus of this Initiative that confirmatory testing of these variant types using an orthogonal method is necessary before reporting (Table 1).

Upfront considerations for test design

Upfront considerations for WGS test design, such as sample and library preparation, sequencing methodology, sequence analysis, and annotation generally follow current guidelines35,36,47 and are discussed in the Supplementary Discussion. More complex test design considerations that are specific to clinical WGS, such as evaluation of metrics to determine suitable WGS test coverage, and the number and type of samples necessary for validation are discussed below.

Evaluation of genome coverage, completeness, and callability

Defining and evaluating high-quality genome coverage is one of the most important considerations in clinical WGS test development, since it directly relates to the amount of data required to accurately identify variants of interest. Metrics that measure genome completeness should be used to define the performance of clinical WGS, and include overall depth and evenness of coverage. These measures should be monitored with respect to callable regions of the genome and related calling accuracy for each variant type compared to orthogonally investigated truth sets (Table 1). While universal cutoffs are not yet established, a combination of depth of coverage, base quality, and mapping quality is recommended to assess callability48. The majority of laboratories in this initiative calculate both raw and usable coverage, the latter metric relating to reads used in variant detection and excluding poorly mapped reads, low-quality base pairs, and overlapping paired reads. All sites have evaluated the performance of clinical WGS, using varying mean depth of coverage, and assessed the completeness and accuracy of variant calling in specific target files, such as a reference standard, or comparison to the method clinical WGS is replacing (e.g., WES; Supplementary Figs. 2, 3). Variability in assessment methodology can result in differences in metrics and cutoffs (Table 3); however, when genome completeness was assessed across three of the sites in this initiative using reference standards the values ranged from 97.7–98.1%, suggesting some consistency in sequencing genomes across laboratories (Supplementary Table 2). If the laboratory is providing WGS from different DNA sources, these evaluations should be completed for each specimen type.

Table 3.

Metrics for clinical whole-genome sequencing.

Metric Description Type (threshold) or typical expected value
Examples of pass/fail metrics
 Sample identity Concordance with genotype (orthogonal and/or family structure when available). Pass/fail (match)
 Contaminationa The estimated level of sample cross-individual contamination based on a genotype-free estimation. Pass/fail (≤2%)
 Gb ≥ Q30b Total aligned gigabases (Gb) of data with base quality score >Q30. Pass/fail (≥80 Gb)
 Autosome mean coveragec The mean coverage across human autosomes, after all filters are applied. Pass/fail (≥30)
 % Callabilityd Percent of non-N reference positions in autosomal chromosomes with a passing genotype call. Pass/fail (>95%)
Examples of metrics to monitor
 %Q30 bases total The percentage of bases that meet Q30 scores. ≥85%
 20×%e The fraction of non-N autosome bases that attained at least 20× sequence coverage in post-filtering bases. ≥90%
 PF reads aligned % The percentage of passing filter (PF) reads that align to the reference sequence. >98%
 PF aligned Q20 basesf The number of bases aligned to the reference sequence in PF reads that were mapped at high quality and where the base call quality was Q20 or higher. >1.0E + 11
 Adapter-dimer % The fraction of PF reads that are unaligned and match to a known adapter sequence right from the start of the read. <0.2%
 Chimera % The percentage of reads that map outside of a maximum insert size (usually 100 kb) or that have the two ends mapping to different chromosomes. <1%
 Duplication % The percentage of mapped sequence that is marked as duplicate. <10%
 Median insert sizeg The median insert size of all paired end reads where both ends mapped to the same chromosome. >300 bp
 Excluded total % The percentage of aligned bases excluded due to all filters. <15%

aLaboratories in the Medical Genome Initiative use a threshold of <1% for germline WGS from peripheral blood.

bSome laboratories in the Medical Genome Initiative use a similar metric of ≥85 Gb unaligned Q30 sequence.

cLaboratories in the Medical Genome Initiative use either 30× or 40× mean coverage as a cutoff.

dCallability or the fraction of the genome where accurate calls can be made can be calculated in different ways. The description in the table represents one way to calculate callability, but there are others including using the percentage of base pairs that reach a read depth (RD) of 20, with base quality (BQ) and mapping quality (MQ) of 20.

eMeasure of completeness. Depth of coverage at 15× also used by some laboratories along with a mapping quality cutoff (>10). Targets will also vary; laboratories may measure across genome, exome, OMIM morbid map genes, and positions or exons with known pathogenic regions.

fSome laboratories in the Medical Genome Initiative use Q10 Bases.

gMean insert size may also be used. The mean insert size is the “core” of the distribution. Artifactual outliers in the distribution often cause calculation of nonsensical mean and standard deviation values. To avoid this, the distribution is first trimmed to a “core” distribution of ±N median absolute deviations around the median insert size. By default, N = 10, but this is configurable.

Reference standard materials and positive controls

High-quality reference standard materials and positive controls with associated truth datasets are a necessary resource for laboratories offering clinical WGS. The analytical validation of clinical WGS should include publicly available reference standards in addition to commercially available and laboratory-held positive controls for each variant type. For variant types commonly addressed by the field, including SNVs and indels, a low minimal number of controls can be utilized if these include well-accepted reference standards. For variant types where standards are still evolving (e.g., REs), a larger number of samples should be employed (Table 1). The National Institute of Standards and Technology (NIST) NA12878 genome and Platinum Genomes are routinely utilized by NGS laboratories seeking to establish WGS analytical validity47. These genomes have the benefit of thousands of variants that have been curated and confirmed across many technologies49,50. Within this initiative, all groups have used NA12878 for validation, and most groups also utilize the Ashkenazi Jewish and Chinese ancestry trios from the Personal Genome Project that are available, as reference materials with variant benchmarks37 (Supplementary Table 3).

The ability to subcategorize analytical performance by variant type is another benefit of using well-characterized reference materials. Genome-wide estimates of sensitivity often mask poor performance in certain sequence contexts or across different variant attributes. For example, the sensitivity of large indel detection (>10 bp) in regions of high homology will be poorer compared to detection of smaller indels in less complex regions. Understanding performance in difficult regions of the genome is important for accurately representing the limitations of the assay, and setting benchmarks against which new analytical tools and methods can be developed. The Global Alliance for Genomics and Health (GA4GH) Benchmarking Team recently developed tools (https://github.com/ga4gh/benchmarking-tools) to evaluate performance in this way. Currently, all members of this initiative have incorporated or intend to use the results of such an analysis in their analytical validation study.

Reference standard materials alone are not sufficient to establish validity of a test, however. For example, both the specimen and disease context must also be taken into consideration when sourcing samples for a validation study. For clinical WGS laboratories in this consortium, specimen context has included determination of the acceptable sample types (e.g., blood, saliva, and tissue) with associated representative positive controls. Some pathogenic variants, including short tandem repeats, low copy repeats, SVs with breakpoints within nonunique sequences, paralogs, and pseudogenes, occur in regions of the genome that are difficult to sequence, align, and map. If analysis and reporting of these loci is planned, the laboratory should perform validation assessments on samples with these specific variant types to determine robustness. Since performance expectations may not be well established for these variants, a large number of positive controls should be used (see below and Supplementary Table 3).

Test validation

Clinical WGS requires a multifaceted approach to analytical validation due to the large number of rare genetic disease loci, the number and different classes of variation that can be detected, and the genomic context-driven variability in variant calling accuracy. Traditional summary statistics defining performance metrics across the entire assay are necessary, but not sufficient. The analytical validation framework should include metrics that account for genome complexity, with special attention to sequence content and variant type (Table 1). For example, sequence level and copy number variants have different calling constraints that can be affected differently by low-complexity sequence. Specific test validation recommendations that address these and other clinical WGS-specific validation requirements are discussed in detail below. Other considerations that are not unique to clinical WGS include sequencing bias, repeatability and reproducibility, limits of detection, interference, and regions of homology are discussed in the Supplementary Discussion along with disease-specific variant validation (e.g., SMA testing), software validation, and test modification and updates.

Performance metrics, variant type, and genomic context

Analytical validation is the first step in ensuring diagnostic accuracy and is classically measured in terms of sensitivity (recall) and specificity. However, this initiative agrees with current recommendations from the GA4GH to use precision as a more useful metric than specificity, owing to the large number of true negatives expected by clinical WGS38. The FDA suggests similar, albeit slightly different metrics, for validation of NGS assays, including positive percent agreement (PPA; sensitivity), negative percent agreement (NPA; specificity), and technical positive predictive value (TPPV; equivalent to precision above), as well as reporting the lower bound of the 95% confidence interval (CI)39. Relevant definitions and calculations are provided in Box 1.

This initiative recommends following published guidelines as described above, as the performance metrics are generally applicable to clinical WGS. In addition to the global metrics of accuracy (sensitivity, precision), repeatability (technical replicates performed under identical conditions), reproducibility (comparison of results across instruments), and limits of detection assessment (e.g., mosaic SNVs) should also be measured. For SNV and indels, gold standard reference data are available, as described above and can be used to calculate performance metrics47. Other variant types may not have standard truth sets available, so comparative metrics should be confined to PPA and NPA against laboratory or commercially acquired samples assessed, using a precedent technology. Laboratories may also consider creating virtual datasets and analytically mixed specimens for validation of the variant types that may not have standard truth sets available.

Performance thresholds should be predetermined and matched to clinical requirements for low diagnostic error rates. Flexibility in performance thresholds at the stage of variant calling may be acceptable, as long as these deviations are documented and laboratory procedures include additional confirmatory assessments. These can include additional bioinformatics analyses, manual inspection by analysts, and orthogonal laboratory testing. The amount of data being examined in a clinical WGS test requires that confirmatory methods be restricted to small subsets of the data with potentially high clinical impact. No calls and invalid calls should not be used in calculations of sensitivity, precision, or TPPV in the validation of variant calling. Instead, these should be documented separately as part of the accuracy of the test and, where possible, genomic intervals that routinely have low map quality and coverage should be flagged in the clinical WGS test definition.

Identification of different variant types require unique calling algorithms, resulting in differences in analytical performance. Further stratification by size is warranted for some common variant types to provide greater insight into overall test performance. For example, GA4GH recommends binning insertions, deletions, and duplications into size bins of <50, 50–200, and >200 bp (ref. 51), although it is important to note that most laboratories in this initiative assess additional smaller bins (Supplementary Fig. 4). For CNVs, size bins, and minimum cutoffs are similar to the maximum resolution of current clinical CMA, which can vary from 20 to 100 kb, depending on the platform used. Laboratories in this initiative that currently offer CNVs as part of the test report events at the resolution of CMA using a depth-based CNV caller, whereas smaller CNV events often require split or anomalous read pair information partnered with a depth assessment52.

Variant calling performance can be affected by the sequence context of the region itself, or, in the case of large variants, the surrounding bases. Currently, there are no best practices for the identification of systematically problematic regions or comprehensive population-level truth datasets, but all members of this initiative have developed internal methods to identify such regions. These include regions where clinical WGS may perform poorly, including paralogous genes, which are excluded from the test definition in order to guide appropriate clinical ordering. The initiative also recommends that regions identified as systematically problematic, or that negatively affect variant calling tied to particular variant types, are documented as part of the test validation and made available to ordering clinicians upon request. Some resources already exist for annotation of genes with high homology, and can be used as a starting point (https://www.ncbi.nlm.nih.gov/books/NBK535152/). Limitations affecting variant calling performance observed during validation should be clearly stated on the report and should include reference to variant types, sizes, and genomic context (Table 1).

Sample number and type for validation

The number of samples and specimen types required for clinical WGS validation links back to the test definition and the variant types or known disease loci that the laboratory intends to report. It is not technically or practically feasible to validate all possible pathogenic variants genome wide. Thus, we recommend that the number of samples required for validation be guided by variant type or the targeted locus being interrogated. For small variants (SNVs and indels), members of this initiative agree that the repeatable and accurate assessment of genome reference standards is sufficient to establish global accuracy, but this should be supplemented with patient positive controls containing a range of clinically relevant variants. Interestingly, the number of positive controls used by laboratories in this consortium for small variants varied between 10 and 85 (Supplementary Table 3), reflecting a broad range of practice amongst laboratories.

Validation of variant types beyond small variants require increased numbers of positive controls, and should include the most commonly affected genes, loci, or pathogenic variants if targeting a specific locus. The number of specific variants that should be assessed may vary according to variant type, genomic context, and the availability of appropriate reference samples. Where possible adhering to a statistically rigorous approach similar to that outlined by Jennings et al.53, which incorporates a confidence level of detection and required probability of detection, is recommended. When applying this method and requiring a 95% reliability with 95% CI, at least 59 variants should be used in the performance assessment, as has been previously published36,53. Taking CNV validation as an example, members of this initiative have used between 7 and 42 positive controls (Supplementary Table 3), and included common microdeletion and duplication syndromes (Supplementary Table 1). For other emerging uses of WGS that specifically target loci (e.g., targeted RE or SMN1), many more positive and negative controls are necessary to assess accuracy. As test scope continues to broaden, we expect consensus to emerge on the recommendation of the number of controls required for validation based on the experience of this initiative and others in the community.

Quality management

As with any laboratory test, groups performing clinical WGS should have a robust quality management program in place for quality control and quality assurance, following applicable regulatory guidance from CLIA (www.cdc.gov/clia), CAP (https://www.cap.org/), and ISO (www.iso.org). Much of the guidance from these regulatory bodies is broadly applicable to any laboratory test, including clinical WGS, and is not discussed here. Rather, we touch on a few points to consider for clinical WGS test quality management focusing on control samples, sample identity, library preparation, sequencing quality and performance metrics, and bioinformatics quality assurance. A list of sequencing and performance metrics examples (many of which are discussed in the following sections) can be found in Table 3. This table features a brief description of each metric, as well as suggested cutoffs or ranges for metrics considered pass/fail and those that should be monitored.

Control samples

One of the biggest challenges for laboratories offering clinical WGS is the application of controls to comply with regulatory guidelines. Guidelines recommend the use of positive, negative, and sensitivity controls (e.g., CAP Molecular Pathology Checklist, August 2018—MOL.34229 Controls Qualitative Assays) to ensure that all steps of the assay are successfully executed without contamination. Ongoing quality control of a clinical whole-genome test should include identification of a comprehensive set of performance metrics, continual monitoring of these metrics across samples over time, and the use of positive controls on a periodic basis dependent on overall sample volume (Table 1). Although the inclusion of a control reference standard in every sequencing run is ideal, it is not practical or financially viable for a laboratory performing clinical WGS. Moreover, the use of positive and negative controls may be informative for the overall performance of a sequencing run, but will not be reflective of sample-specific differences and may incorrectly indicate adequate test performance.

There are additional positive and negative control strategies that some laboratories may choose to employ. Some of the groups in the initiative use PhiX, which represents the empirical measure of sequencing error rate. For variant positive controls, one approach is the use of low-level spike-ins of well-characterized positive control samples that include a spectrum of variants in each sequencing run. Similarly, some groups in the initiative are exploring the use of synthetic spike-in constructs, including Sequins54, which can be added to a run at a low level (<1% of reads) and enable a performance assessment that can serve as a process control for at least some variant types. Within this initiative, most groups run a reference standard at periodic intervals, and check for deviations from expected calling accuracy and concordance with previously run samples.

Sample identity

Given the multistep processes needed to generate a final result, a sample identity tracking procedure within the laboratory is recommended during tube and instrument transfers, and to confirm the integrity of the final results. Implementing this tracking procedure will mitigate the risk of sample mix-up through the analytical steps of the assay, but will not necessarily detect other pre-analytical issues, such as labeling or sample collection errors. Although there is no standard method employed by initiative members, examples of sample tracking include comparison of WGS data to SNPs genotyped with a multiplex assay or custom microarray, STR marker analysis, or spike-in methodology. Regardless of sample tracking method, discordance in genotype between WGS and orthogonal testing data results in failure of the test (Table 3). Methods similar to those described above should be used when case-parent trios are sequenced, or when other family members are included in the clinical testing strategy. Formal checks for Mendelian errors to establish parentage and to assess relatedness among other family members should be computed using standard methods.

Library preparation

The yield and quality (e.g., fluorometry and size range) of the DNA should have defined criteria for acceptance that allows a DNA sample to be passed to library preparation and sequencing. For clinical WGS, sample pooling and molecular barcoding is utilized in the majority of laboratories. Some platforms benefit from a dual-barcoding strategy (i.e., a barcode on each end of the library molecule) to reduce the possibility of barcode hopping on the flowcell55. Quality metrics (e.g., library concentration) with acceptance thresholds must be established and the results from each sample must be documented. For sample and library preparation, procedures are needed to detect and interpret systematic drops in quality and/or the percentage of samples meeting minimum quality requirements. A control for library preparation may be used to monitor quality, and troubleshoot preparation versus sample issues and a non-template control can be used to monitor systematic contamination.

Sequencing quality and performance metrics

Test run quality metrics and performance thresholds for clinical WGS should be assessed at the sample level as part of quality control. A quality assurance program should periodically monitor quality metrics over time and identify trends in test performance related to reagent quality, equipment performance, and technical staff.

Clinical WGS sample level quality metrics describe whether the biological specimen and end-to-end test are technically adequate (i.e., whether the test provides the expected analytical sensitivity and technical positive predictive value for all variant types (SNVs, indels, CNVs, and SVs)) within the reportable range of the genome established during test validation.

Quality metrics are calculated for every run of the instrument, and after alignment and variant calling (see Supplementary Discussion for expanded description). Test development optimization and validation processes establish which metrics are reviewed for every sample, but it can be challenging for laboratories to determine appropriate thresholds. Examples of sequencing quality and performance metrics used by members of this consortium to evaluate WGS for pass/fail and monitoring are listed in Table 3. Important metrics for passing samples include the total gigabases (Gb; >Q30) produced per sample, the alignment rate of purity-filtered bases (PF reads aligned %), the predicted usable coverage of the genome (mean autosomal coverage), proportion of reads that are duplicates (% duplication), the % callability (positions with passing genotype call), and any evidence of sample contamination (% contamination). For clinical WGS, it is particularly important to monitor global mapping metrics and assess clinically significant loci for completeness (e.g., OMIM genes and ClinVar pathogenic variants).

Mean coverage and completeness of coverage are commonly used metrics, but as discussed previously, these may be calculated differently across groups (see previous section on coverage evaluation). It is important to note that at the time of publication, the initiative was unable reach a high level of consensus as to which metrics should be used and the corresponding thresholds that need to be met to qualify, as a passing clinical WGS test. There was general agreement on the types of measures that are important (Table 3), but often these were calculated in different ways, which made reaching consensus difficult. This is likely a reflection of the evolving technology and the way in which each group validated testing in the absence of accepted guidelines. More data and laboratory experience are needed before consensus on performance metrics thresholds that define a clinical WGS test can be established.

Bioinformatics quality assurance

Clinical bioinformatics pipelines developed for the analysis of clinical WGS tests are complex, and require a robust quality assurance program for both ongoing monitoring of metrics and pipeline updates36. Due to the continual updating of software versions (e.g., read aligners and variant callers) and data sources for annotation (e.g., OMIM, Clinvar, etc.) the development, validation, and deployment cycles can be challenging for laboratories. Pipeline versions need to be revalidated when updated (see Supplementary Discussion “Software validation”) and a system to track versions with parameters and implementation date must be employed. All code changes need to be documented along with versions of data sources. Pipelines can be tested with reference standards to ensure that they are reproducible and complete without errors.

Summary

Clinical WGS is poised to become a first-tier test for the diagnosis of those individuals with suspected genetic disease. Although some guidelines are beginning to emerge that offer recommendations for the analytical validation of genome testing, specific challenges related to the setup, and deployment of clinical WGS are not addressed. In this document, we aimed to address these gaps through consensus recommendations for the analytical validation of clinical WGS, based on the experiences of members of the Medical Genome Initiative. We focused on providing practical advice for test development optimization, validation practices, and ongoing quality management for the deployment of clinical WGS. Even amongst members within the initiative, it was challenging to come to a consensus on specific recommendations since there are often different, but equally valid approaches to the analytical validation of WGS. Another reason for the lack of consensus is the rapid advancement of the field; the process of WGS is continually being updated and improved meaning laboratories are often at different stages of implementation. However, members of this initiative agreed upon the endorsement of clinical WGS as a viable first-tier test for individuals with rare disorders, and that it should replace CMA and WES.

The recommendations provided here are meant to represent a snapshot of the current state of the field, and we expect best practices to continue to evolve. Although reaching consensus on specific validation-related practice was not always possible, a sentiment shared by all groups was that establishing standards in clinical WGS is difficult but critically important. Collaborative efforts and communication both within, and among research and healthcare institutions are essential to establishing guidelines and standards to increase access to high-quality clinical WGS, while minimizing patient risk. It is clear that much work is needed in the community to establish clear consensus around many of the analytical principles that define a valid clinical genome test.

To this end, our group is committed to providing best practices on clinical WGS topics both upstream and downstream from analytical validity, including genome interpretation, data infrastructure, and clinical utility measures.

Supplementary information

Supplementary Information (544.3KB, pdf)

Acknowledgements

The authors thank Michael Eberle, Mar Gonzàlez-Porta, Cinthya Zepeda Mendoza, Cherisse Marcou, Jaime Lopez, R. Tanner Hagelstrom, and Kirsten Curnow for critical comments and review of the manuscript.

Author contributions

C.R.M. prepared the manuscript. C.R.M., J.W.B., N.J.L., S.C., and R.J.T. conceived the idea and wrote the paper. S.C., C.R.M., D.J.S., N.J.L., M.S.L., and S.L.T. collected and analyzed survey data. S.C., N.J.L., M.S.L., C.R.M., V.J., and P.L. contributed original data and analysis tools. J.G.B., S.M.H., R.R., E.W.K., D.J.S., S.C., N.J.L., M.S.L., and S.L.T. contributed writing of manuscript sections. H.M.K., D.B., D.D., S.K., and E.A.W. provided design advice and critical review of the manuscript.

Data availability

The data analyzed during the current study are available from the corresponding author on reasonable request.

Competing interests

S.L.T., R.J.T., and J.W.B. are current employees and shareholders of Illumina Inc.

Footnotes

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

A full list of members and their affiliations appears in the Supplementary Information.

Supplementary information

Supplementary information is available for this paper at 10.1038/s41525-020-00154-9.

References

  • 1.Boycott K, et al. The clinical application of genome-wide sequencing for monogenic diseases in Canada: position statement of the Canadian College of Medical Geneticists. J. Med. Genet. 2015;52:431–437. doi: 10.1136/jmedgenet-2015-103144. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.ACMG. Points to consider in the clinical application of genomic sequencing. Genet. Med. 2012;14:759–761. doi: 10.1038/gim.2012.74. [DOI] [PubMed] [Google Scholar]
  • 3.Gullapalli RR, et al. Clinical integration of next-generation sequencing technology. Clin. Lab. Med. 2012;32:585–599. doi: 10.1016/j.cll.2012.07.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Matthijs G, et al. Guidelines for diagnostic next-generation sequencing. Eur. J. Hum. Genet. 2016;24:2–5. doi: 10.1038/ejhg.2015.226. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Vrijenhoek T, et al. Next-generation sequencing-based genome diagnostics across clinical genetics centers: implementation choices and their effects. Eur. J. Hum. Genet. 2015;23:1142–1150. doi: 10.1038/ejhg.2014.279. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Farwell KD, et al. Enhanced utility of family-centered diagnostic exome sequencing with inheritance model-based analysis: results from 500 unselected families with undiagnosed genetic conditions. Genet. Med. 2015;17:578–586. doi: 10.1038/gim.2014.154. [DOI] [PubMed] [Google Scholar]
  • 7.Lee H, et al. Clinical exome sequencing for genetic identification of rare Mendelian disorders. JAMA. 2014;312:1880–1887. doi: 10.1001/jama.2014.14604. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Yang Y, et al. Molecular findings among patients referred for clinical whole-exome sequencing. JAMA. 2014;312:1870–1879. doi: 10.1001/jama.2014.14601. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Wright CF, et al. Genetic diagnosis of developmental disorders in the DDD study: a scalable analysis of genome-wide research data. Lancet. 2015;385:1305–1314. doi: 10.1016/S0140-6736(14)61705-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Stark Z, et al. A prospective evaluation of whole-exome sequencing as a first-tier molecular test in infants with suspected monogenic disorders. Genet. Med. 2016;18:1090–1096. doi: 10.1038/gim.2016.1. [DOI] [PubMed] [Google Scholar]
  • 11.Meng L, et al. Use of exome sequencing for infants in intensive care units: ascertainment of severe single-gene disorders and effect on medical management. JAMA Pediatr. 2017;171:e173438. doi: 10.1001/jamapediatrics.2017.3438. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Delaney SK, et al. Toward clinical genomics in everyday medicine: perspectives and recommendations. Expert Rev. Mol. Diagn. 2016;16:521–532. doi: 10.1586/14737159.2016.1146593. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Belkadi A, et al. Whole-genome sequencing is more powerful than whole-exome sequencing for detecting exome variants. Proc. Natl Acad. Sci. USA. 2015;112:5473–5478. doi: 10.1073/pnas.1418631112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Lelieveld SH, Spielmann M, Mundlos S, Veltman JA, Gilissen C. Comparison of exome and genome sequencing technologies for the complete capture of protein-coding regions. Hum. Mutat. 2015;36:815–822. doi: 10.1002/humu.22813. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Bick D, Jones M, Taylor SL, Taft RJ, Belmont J. Case for genome sequencing in infants and children with rare, undiagnosed or genetic diseases. J. Med. Genet. 2019;56:783–791. doi: 10.1136/jmedgenet-2019-106111. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Weedon MN, et al. Recessive mutations in a distal PTF1A enhancer cause isolated pancreatic agenesis. Nat. Genet. 2014;46:61–64. doi: 10.1038/ng.2826. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Merico D, et al. Compound heterozygous mutations in the noncoding RNU4ATAC cause Roifman Syndrome by disrupting minor intron splicing. Nat. Commun. 2015;6:8718. doi: 10.1038/ncomms9718. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Jaganathan K, et al. Predicting splicing from primary sequence with deep learning. Cell. 2019;176:535–548.e524. doi: 10.1016/j.cell.2018.12.015. [DOI] [PubMed] [Google Scholar]
  • 19.Hayashi S, et al. ALPHLARD: a Bayesian method for analyzing HLA genes from whole genome sequence data. BMC Genomics. 2018;19:790. doi: 10.1186/s12864-018-5169-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Cohn I, et al. Genome sequencing as a platform for pharmacogenetic genotyping: a pediatric cohort study. NPJ Genom. Med. 2017;2:19. doi: 10.1038/s41525-017-0021-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Khera AV, et al. Genome-wide polygenic scores for common diseases identify individuals with risk equivalent to monogenic mutations. Nat. Genet. 2018;50:1219–1224. doi: 10.1038/s41588-018-0183-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Carss KJ, et al. Comprehensive rare variant analysis via whole-genome sequencing to determine the molecular pathology of inherited retinal disease. Am. J. Hum. Genet. 2017;100:75–90. doi: 10.1016/j.ajhg.2016.12.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Gilissen C, et al. Genome sequencing identifies major causes of severe intellectual disability. Nature. 2014;511:344–347. doi: 10.1038/nature13394. [DOI] [PubMed] [Google Scholar]
  • 24.Yuen RK, et al. Whole-genome sequencing of quartet families with autism spectrum disorder. Nat. Med. 2015;21:185–191. doi: 10.1038/nm.3792. [DOI] [PubMed] [Google Scholar]
  • 25.Taylor JC, et al. Factors influencing success of clinical genome sequencing across a broad spectrum of disorders. Nat. Genet. 2015;47:717–726. doi: 10.1038/ng.3304. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Scocchia A, et al. Clinical whole genome sequencing as a first-tier test at a resource-limited dysmorphology clinic in Mexico. NPJ Genom. Med. 2019;4:5. doi: 10.1038/s41525-018-0076-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Stavropoulos DJ, et al. Whole genome sequencing expands diagnostic utility and improves clinical management in pediatric medicine. NPJ Genom. Med. 2016;1:15012. doi: 10.1038/npjgenmed.2015.12. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Clark MM, et al. Meta-analysis of the diagnostic and clinical utility of genome and exome sequencing and chromosomal microarray in children with suspected genetic diseases. NPJ Genom. Med. 2018;3:16. doi: 10.1038/s41525-018-0053-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Soden SE, et al. Effectiveness of exome and genome sequencing guided by acuity of illness for diagnosis of neurodevelopmental disorders. Sci. Transl. Med. 2014;6:265ra168. doi: 10.1126/scitranslmed.3010076. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Farnaes L, et al. Rapid whole-genome sequencing decreases infant morbidity and cost of hospitalization. NPJ Genom. Med. 2018;3:10. doi: 10.1038/s41525-018-0049-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Saunders CJ, et al. Rapid whole-genome sequencing for genetic disease diagnosis in neonatal intensive care units. Sci. Transl. Med. 2012;4:154ra135. doi: 10.1126/scitranslmed.3004041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Lionel AC, et al. Improved diagnostic yield compared with targeted gene sequencing panels suggests a role for whole-genome sequencing as a first-tier genetic test. Genet. Med. 2017;20:435–443. doi: 10.1038/gim.2017.119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Gross AM, et al. Copy-number variants in clinical genome sequencing: deployment and interpretation for rare and undiagnosed disease. Genet. Med. 2018;21:1121–1130. doi: 10.1038/s41436-018-0295-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Costain G, et al. Periodic reanalysis of whole-genome sequencing data enhances the diagnostic advantage over standard clinical genetic testing. Eur. J. Hum. Genet. 2018;26:740–744. doi: 10.1038/s41431-018-0114-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Aziz N, et al. College of American Pathologists’ laboratory standards for next-generation sequencing clinical tests. Arch. Pathol. Lab. Med. 2015;139:481–493. doi: 10.5858/arpa.2014-0250-CP. [DOI] [PubMed] [Google Scholar]
  • 36.Roy S, et al. Standards and guidelines for validating next-generation sequencing bioinformatics pipelines: a joint recommendation of the association for molecular pathology and the College of American Pathologists. J. Mol. Diagn. 2018;20:4–27. doi: 10.1016/j.jmoldx.2017.11.003. [DOI] [PubMed] [Google Scholar]
  • 37.Zook JM, et al. An open resource for accurately benchmarking small variant and reference calls. Nat. Biotechnol. 2019;37:561–566. doi: 10.1038/s41587-019-0074-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Krusche P, et al. Best practices for benchmarking germline small-variant calls in human genomes. Nat. Biotechnol. 2019;37:555–560. doi: 10.1038/s41587-019-0054-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.FDA. Considerations for Design, Developmment, and Analytical Validation of Next Generation Sequencing (NGS) - Based In Vitro Diagnostics (IVDs) Intended to Aid in the Diagnosis of Suspected Germline Diseases. (ed US Food and Drug Administration, 2018).
  • 40.Marshall CR, et al. The Medical Genome Initiative: moving whole-genome sequencing for rare disease diagnosis to the clinic. Genome Med. 2020;12:48. doi: 10.1186/s13073-020-00748-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Richards S, et al. Standards and guidelines for the interpretation of sequence variants: a joint consensus recommendation of the American College of Medical Genetics and Genomics and the Association for Molecular Pathology. Genet Med. 2015;17:405–424. doi: 10.1038/gim.2015.30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Trost B, et al. A comprehensive workflow for read depth-based identification of copy-number variation from whole-genome sequence data. Am. J. Hum. Genet. 2018;102:142–155. doi: 10.1016/j.ajhg.2017.12.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Dolzhenko E, et al. Detection of long repeat expansions from PCR-free whole-genome sequence data. Genome Res. 2017;27:1895–1903. doi: 10.1101/gr.225672.117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Kingsmore SF, et al. A randomized, controlled trial of the analytic and diagnostic performance of singleton and trio, rapid genome and exome sequencing in Ill infants. Am. J. Hum. Genet. 2019;105:719–733. doi: 10.1016/j.ajhg.2019.08.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Lindstrand A, et al. From cytogenetics to cytogenomics: whole-genome sequencing as a first-line test comprehensively captures the diverse spectrum of disease-causing genetic variation underlying intellectual disability. Genome Med. 2019;11:68–68. doi: 10.1186/s13073-019-0675-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.D’Gama AM, Walsh CA. Somatic mosaicism and neurodevelopmental disease. Nat. Neurosci. 2018;21:1504–1514. doi: 10.1038/s41593-018-0257-3. [DOI] [PubMed] [Google Scholar]
  • 47.Rehm HL, et al. ACMG clinical laboratory standards for next-generation sequencing. Genet. Med. 2013;15:733–747. doi: 10.1038/gim.2013.92. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Goldfeder, R. L. & Ashley, E. A. A precision metric for clinical genome sequencing. Preprint at 10.1101/051490 (2016).
  • 49.Zook JM, et al. Integrating human sequence data sets provides a resource of benchmark SNP and indel genotype calls. Nat. Biotechnol. 2014;32:246–251. doi: 10.1038/nbt.2835. [DOI] [PubMed] [Google Scholar]
  • 50.Zook J, et al. Reproducible integration of multiple sequencing datasets to form high-confidence SNP, indel, and reference calls for five human genome reference materials. Nat. Biotechnol. 2019;37:561–566. doi: 10.1038/s41587-019-0074-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.GA4GH. Benchmarking Performance Stratification for SNVs and Small Indels.https://github.com/ga4gh/benchmarking-tools/blob/master/doc/standards/GA4GHBenchmarkingPerformanceStratification.md (2017).
  • 52.Hehir-Kwa JY, Pfundt R, Veltman JA. Exome sequencing and whole genome sequencing for the detection of copy number variation. Expert Rev. Mol. Diagn. 2015;15:1023–1032. doi: 10.1586/14737159.2015.1053467. [DOI] [PubMed] [Google Scholar]
  • 53.Jennings LJ, et al. Guidelines for validation of next-generation sequencing-based oncology panels: a joint consensus recommendation of the association for molecular pathology and College of American Pathologists. J. Mol. Diagn. 2017;19:341–365. doi: 10.1016/j.jmoldx.2017.01.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Hardwick SA, Deveson IW, Mercer TR. Reference standards for next-generation sequencing. Nat. Rev. Genet. 2017;18:473–484. doi: 10.1038/nrg.2017.44. [DOI] [PubMed] [Google Scholar]
  • 55.Costello M, et al. Characterization and remediation of sample index swaps by non-redundant dual indexing on massively parallel sequencing platforms. BMC Genomics. 2018;19:332. doi: 10.1186/s12864-018-4703-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Duan M, et al. Evaluating heteroplasmic variations of the mitochondrial genome from whole genome sequencing data. Gene. 2019;699:145–154. doi: 10.1016/j.gene.2019.03.016. [DOI] [PubMed] [Google Scholar]
  • 57.Chen X, et al. Spinal muscular atrophy diagnosis and carrier screening from genome sequencing data. Genet Med. 2020;22:945–953. doi: 10.1038/s41436-020-0754-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Eberle MA, et al. A reference data set of 5.4 million phased human variants validated by genetic inheritance from sequencing a three-generation 17-member pedigree. Genome Res. 2017;27:157–164. doi: 10.1101/gr.210500.116. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Information (544.3KB, pdf)

Data Availability Statement

The data analyzed during the current study are available from the corresponding author on reasonable request.


Articles from NPJ Genomic Medicine are provided here courtesy of Nature Publishing Group

RESOURCES