The higher throughput and lower per-base cost of next-generation sequencing (NGS) as compared to Sanger sequencing has led to its rapid adoption in clinical testing. The number of laboratories offering NGS-based tests has also grown considerably in the past few years, despite the fact that specific Clinical Laboratory Improvement Amendments of 1988/College of American Pathologists (CAP) laboratory standards had not yet been developed to regulate this technology.
To develop a checklist for clinical testing using NGS technology that sets standards for the analytic wet bench process and for bioinformatics or “dry bench” analyses. As NGS-based clinical tests are new to diagnostic testing and are of much greater complexity than traditional Sanger sequencing–based tests, there is an urgent need to develop new regulatory standards for laboratories offering these tests.
To develop the necessary regulatory framework for NGS and to facilitate appropriate adoption of this technology for clinical testing, CAP formed a committee in 2011, the NGS Work Group, to deliberate upon the contents to be included in the checklist.
—A total of 18 laboratory accreditation checklist requirements for the analytic wet bench process and bioinformatics analysis processes have been included within CAP's molecular pathology checklist (MOL).
This report describes the important issues considered by the CAP committee during the development of the new checklist requirements, which address documentation, validation, quality assurance, confirmatory testing, exception logs, monitoring of upgrades, variant interpretation and reporting, incidental findings, data storage, version traceability, and data transfer confidentiality.
DNA sequencing has evolved from Maxam-Gilbert1 and Sanger2,3 methods in the 1970s to a set of technologies that are collectively referred to as next-generation sequencing (NGS).4–12 The primary difference between NGS and first-generation technologies is that sequencing of millions of short fragments of DNA occurs in parallel instead of one DNA fragment at a time. Sequencing of DNA as a clinical test became routinely possible only after the automation of Sanger sequencing methods introduced in the mid-1990s, which used capillary gel electrophoresis with fluorescence-based detection.13,14 The throughput of NGS far surpasses that of automated Sanger sequencing. The higher throughput and lower per-base cost of NGS have contributed to its rapid adoption in clinical testing,15 despite the fact that several aspects of NGS analysis have much higher complexity. Examples include the acquisition and storage of data sets that far exceed those commonly generated in a Clinical Laboratory Improvement Amendments of 1988 (CLIA) laboratory and downstream challenges in computation and interpretation. Areas in which NGS testing is being applied currently include inherited diseases, solid tumors, hematologic malignancies, infectious diseases, human leukocyte antigen analysis, and noninvasive prenatal screening to detect fetal chromosome defects.
The number of laboratories offering NGS testing has grown considerably in the past few years, despite the fact that specific CLIA/College of American Pathologists (CAP) laboratory standards had not yet been developed to regulate this technology. To address this need, the CAP formed an ad hoc committee, the NGS Work Group, to develop the first set of clinical laboratory standards for this nascent technology. Given that NGS-based testing represents an evolving technology with continued improvements in instrumentation, sequencing chemistries, and bioinformatic and computational analyses, the work group aimed to develop standards that provide a necessary regulatory framework for clinical NGS tests (which to date are laboratory-developed tests) without inhibiting further adoption of NGS-based testing technology.
Next-generation sequencing incorporates 2 processes: (1) the analytic wet bench process and (2) bioinformatics analysis of sequence data. The wet bench component generally includes any or all of the following processes: handling of patient samples, extraction of nucleic acids, fragmentation, barcoding (molecular indexing) of patient samples, enrichment of targets for exome or gene panels, adapter ligation, amplification, library preparation, flow cell loading, and generation of sequence reads. Sequence generation is almost entirely automated and the output consists of millions to billions of short sequence reads. The wet bench workflow is followed by intensive computational and bioinformatics analyses that use a variety of algorithms to map and align the short sequence reads to a linear reference human genome sequence. After mapping and alignment, variant calls are made at locations where nucleotides differ from the reference sequence. Separate processes develop content needed to analyze the clinical relevance of variants, either singly or in combination, relative to their contribution to a given clinical phenotype. For individual patient cases, identified variants are evaluated against annotated content to infer the potential for impairments to normal gene function (eg, premature transcript or protein truncation, impact of nonsynonymous amino acid changes to protein function, or alternative splicing). Interpretation requires integrating genomic findings with the patient's clinical phenotype in order to make an informed decision regarding causality and correlation of the deleterious mutation(s) with the patient's disease. The mapping, alignment, variant calling, and variant annotation steps, and, to some degree, clinical interpretation (if decision support tools are used), comprise the overall bioinformatics analysis workflow.
The CAP NGS Work Group approached the analytic wet bench process and the bioinformatics or “dry bench” analyses as 2 discrete processes requiring separate considerations for standards. This division was leveraged to support the fact that some laboratories use external facilities to conduct either portion of NGS-based testing. In a laboratory offering the entire process from wet bench through bioinformatics analysis, clinical validation of their test will incorporate the validation of both parts. A total of 18 laboratory accreditation checklist requirements for the analytic wet bench process and bioinformatics analysis processes have been included within CAP's molecular pathology checklist (MOL). The NGS checklist items include new standards for documentation, validation, quality assurance, confirmatory testing, exception logs, monitoring of upgrades, variant interpretation and reporting, incidental findings, data storage, version traceability, and data transfer confidentiality. As described in this report, the work group's goal was to initially develop foundational accreditation requirements for NGS that could be applied across multiple testing areas including inherited disorders, molecular oncology, and infectious diseases. It was anticipated that once foundational requirements were in place, there would be the need to subsequently develop additional, discipline-specific (eg, molecular oncology) NGS checklist requirements, and this is further addressed in the “Comment” section. This report describes important issues considered by the NGS Work Group during the development of each of the new checklist requirements. In addition, this report serves as a supplement to the CAP NGS checklist requirements and therefore the contents are closely aligned to each requirement for the 2014 checklist.
WET BENCH ANALYTIC PROCESS
NGS Wet Bench Process Documentation
The Laboratory Uses a Standard Operating Procedure to Document the Analytic Wet Bench Process Used to Generate NGS Data
The detailed documentation of the wet bench processes is a critical part of quality assessment in the clinical laboratory. All standard operating protocols of DNA/RNA sample preparation, fragmentation, library preparation, barcoding (molecular indexing), sample pooling, and sequence generation must be documented so that each step and subsequent manipulations can be traced. This includes documentation of all methods and reagents as well as instruments, instrument software, and versions used throughout the wet bench process. In addition, controls used need to be described. A few examples will be highlighted below. Targeted NGS assays (such as multigene panels or exome sequencing) allow selective capture of genomic regions of interest before sequencing, and detailed information regarding the captured region(s) (using genomic coordinates of capture probes and lists of genes) and target-enrichment protocols should be documented. Clinical laboratories that process different types of samples (eg, blood, formalin-fixed paraffin-embedded specimens) should develop standard operating procedures (SOPs) for each validated sample type. The reagents and protocols used for pooled analysis of patient specimens must be detailed and should include the sequence information of the barcodes used for each patient sample. Metrics and quality control parameters used to assess run performance must also be documented. Commonly used metrics include the percentage of reads mapping to the target region, the fraction of bases meeting specified quality and coverage thresholds, and average coverage/base and target region. The laboratory must define and document acceptance and rejection criteria for the wet bench process inclusive of sample preparation and sequencing. It is critical to determine and summarize regions that failed analysis (eg, due to inadequate coverage) if they are not covered by orthogonal technologies (such as Sanger sequencing).
Evidence of compliance for this requirement includes a written SOP that describes the analytic wet bench process and the ability to demonstrate that the laboratory follows its policies and procedures.
NGS Wet Bench Process Validation
The Laboratory Validates the Analytic Wet Bench Process and Revalidates the Entire Process and/or Confirms the Performance of the Components of the Process as Satisfactory When Modifications Are Made. The Extent of Revalidation and/or Confirmation Is Modification Dependent.
Like all laboratory-developed tests in molecular diagnostics and other areas of the clinical laboratory, analytic performance of NGS procedures must be internally validated before clinical implementation. Next-generation sequencing analysis is a complex procedure with many steps within the wet bench workflow. Each step needs to be individually optimized to empirically determine optimal assay conditions and analysis settings. Once those are in place, an analytic validation must be performed for the whole test in a “beginning-to-end” fashion, including the entire wet bench process as well as the bioinformatic analyses. Essential performance characteristics that need to be determined during the validation are the analytic sensitivity and specificity, accuracy (the degree of closeness of measurements to the actual [true] value), precision (reproducibility and reliability), and limit of detection (if applicable). As for any molecular assay, validation should also be conducted independently for each accepted specimen type (blood, saliva, tissue, etc). Next-generation sequencing tests are typically designed to interrogate large and multiple regions of the genome, and its use can range from mutational hotspots for oncology applications to gene panels to exomes or genomes. As a consequence, NGS permits the detection of novel as well as known sequence variants, which necessitates a comprehensive approach to be able to determine test performance with adequate confidence. Because it is not possible to validate all theoretically possible variants that can occur, it is necessary to use a combination of a “methods-based” 16 and “analyte-specific” validation approach for determining a test's analytic performance. Consulting the published literature for studies regarding the accuracy of the relevant NGS platform can be useful to inform the laboratory's own validation work. In most cases, variants will have been identified via Sanger sequencing, considered (at least for now) the gold standard comparative technique. However, variant validation information may also be obtained from oligonucleotide microarray genotyping data in some cases. Several professional organizations have issued guidance regarding validation of molecular tests and, more recently, NGS tests in specific to which the reader is referred.17–21
As the NGS Work Group debated NGS validation requirements, the concept of requiring a minimum number of samples for inclusion in a validation was extensively discussed. It was concluded that adding a minimum sample number requirement was premature given the ongoing evolution of NGS technology and the diversity of applications being implemented in diagnostic laboratories. Further, the concern existed that establishing a minimum sample number requirement may result in laboratories conducting an insufficient validation for a given NGS diagnostic application. The work group noted that NGS validations reported in the literature have varied considerably in sample number size (eg, ∼20–80 plus samples),22–32 reflecting that individual laboratories are on a validation “learning curve.” The total number of samples that needs to be run to appropriately validate an NGS test is driven partly by the size of the test (larger assayed regions will have more variants available for deriving their technical performance), by the number of specific analytes (variants) that need to be assessed, by the possible requirement to determine limit of detection across a range of allele frequencies, and by the number of runs and samples needed to determine precision. At this juncture in time, the NGS Work Group concluded that statistical considerations with regard to the number of samples cannot be universally or comprehensively applied across the numerous assays that are possible when using NGS (eg, amplicon versus targeted capture; small numbers of genes versus exome or genome; inherited disease versus oncology versus infectious disease) as the sequencing methodology. Therefore, we have described different scenarios (eg, samples needed for methods-based approach, samples needed to assess reproducibility and reliability, and clinical samples used to assess diagnostic specificity and sensitivity), each of which will necessitate samples whose numbers will vary with the context of each assay. We emphasized the principles of validation in the requirements and several analytic performance parameters as highlighted below.
Analytic sensitivity can be assessed by using a methods-based approach that aims at maximizing the number of sequence variants that are compared to a gold standard method to increase confidence of analytic performance. These values may then be extrapolated to all bases. For this methods-based approach, pathogenicity of analyzed variants does not matter as this has no bearing on their technical detectability. However, it is important to determine this “baseline” performance by using as many different genomic regions as possible, as sequence context can be an important influence. In addition, laboratories should determine analytic performance separately for all variant types that are relevant for the test (eg, single nucleotide variants, indels, copy number variants, structural variants, homopolymers). Approaches to maximize the number of appropriately identified variants may include cumulative analysis of different in-house–developed tests (eg, different gene panels), provided that they rely on identical protocols. In addition, several publicly available databases provide exome/genome-wide variant calls that can be used in the clinical validation efforts (eg, HapMap or 1000 Genomes). In addition, the Centers for Disease Control and Prevention and National Center for Biotechnology Information have collaborated to establish a Web browser to facilitate access to 2 well-sequenced genomes (NA12878 and NA19240)33 and to provide access to clinical-grade targeted data sets (gene panels) and exome/genome-wide data sets created by various laboratories.34 These databases provide access to large sets of variants that can aid in deriving technical performance specifications. However, an analyte-specific validation may be necessary in addition to the more global methods-based approach when the NGS test includes genes that are known to harbor well-known, disease-causing variants. In such cases, it is important to include traditional positive controls with patient samples, including relevant variants (eg, p.F508del in CFTR) to demonstrate adequate detection by that NGS test. Analytic specificity is often calculated by using “negative” samples (ie, samples that have no pathogenic variant) to determine the fraction that is correctly identified as negative. However, this concept does not work well for NGS-based tests. Once again, a methods-based approach can be leveraged to calculate analytic specificity across the assayed region, for example, by determining the false-positive rate (fraction of variants detected that are incorrect calls). It is also useful to determine the average number of false-positive calls for the regions tested in a clinical sample. Note that the analytic specificity accounts for numerous sources of type I error, including base-calling error, errors due to misalignment, and variant-calling errors. Determining the limit of detection is important for assays that interrogate samples with heterogeneous genotypes (eg, tumor specimens, maternal blood used for noninvasive prenatal testing for fetal aneuploidies, and mosaic specimens). This can be challenging given that Sanger sequencing, which is often used as an alternate gold standard technology during validation, is not as sensitive as NGS. Sample “mixing” experiments (eg, dilution of samples with known allele frequencies) may be used to assess the limit of detection of each variant type. Precision (interrun and intrarun variability) should be determined by using at least 3 samples. For tests that are performed with single-lane sequencers, intrarun variability may be determined by using bar-coded replicates of the same sample.
Homologous sequences such as pseudogenes can interfere with accurate variant calling and therefore pose significant challenges for correctly analyzing affected genes. An upfront bioinformatics homology analysis is useful to determine possible interference by homologous sequences. In addition, read-mapping quality can be used to identify problematic regions. If such genes are included in the NGS test, the laboratory must devise a method to ensure that identified variants are not due to pseudogene sequence and must document the accuracy of the method. When pooled sequencing of bar-coded samples is performed, the laboratory must document that individual sample identity is maintained throughout the wet bench process.
The extent of revalidation and confirmation is dependent on the magnitude of the introduced changes and their potential consequences. For example, minor changes, such as the introduction of a new lot of capture reagent that has already undergone comprehensive validation, can be addressed by confirming adequate performance. In this example, it would be deemed acceptable if the laboratory sequences a previously tested sample and documents that the main run metrics (eg, coverage, read quality) are unchanged and that the same results are obtained. Conversely, a major change, such as the introduction of a new sequencing platform or different target enrichment method, would require a more extensive revalidation.
NGS Wet Bench Process—Quality Management Program
The Laboratory Follows a Documented Quality Management Program for the NGS Analytic Wet Bench Process.
CAP-accredited laboratories must develop and follow a quality management plan. The CAP All Common Checklist (COM) applies to every part of a multispecialty laboratory and includes entire sections on Quality Management and Test Method Performance. However, NGS Wet Bench Process—Quality Management Program was added to the NGS portion of the checklist to highlight the particular needs of laboratories performing NGS. No two quality management programs are alike. Each is shaped by the laboratory's scope, clinical market, and expertise, and the laboratory director is given wide latitude in the design of the quality assurance program. The design of the program must be written, and compliance with that design documented. A good quality assurance program for laboratories performing NGS will include the following attributes35,36 :
The quality assurance program follows the path of workflow. The programs should assess preanalytic steps occurring before NGS, analytic testing, and postanalytic processes used in sequence analysis through reporting.
The NGS quality program should be integrated within the institution's overall quality assurance program. If it is part of a larger institution, such as a hospital or medical center, the NGS quality program should fit well within its overall context.
The program should address common problems that arise in the course of testing. “Problems” include events that can affect the test result or its clinical use as well as nonconformance with the laboratory's own policies and procedures. Documentation includes both review of the effectiveness of corrective actions taken and the revision of policies and procedures intended to prevent recurrence.
The overall goal of the quality program aims to ensure that testing is clinically relevant. This is particularly important for tests such as NGS, for which no comparative analytic result of greater sensitivity may exist. The appropriateness of test orders and analytic decisions must be grounded in medical science and evidence.
The program should also encourage laboratory employees to communicate concerns about the quality of laboratory testing. The investigation of employee complaints and suggestions must be a part of the quality assurance program.
NGS Confirmatory Testing
The Laboratory Has a Policy That Documents Indications for Confirmatory Testing of Reported Variants.
While the accuracy of NGS technologies is continuing to improve, it is widely accepted that most NGS-based sequencing assays will yield false-positive and false-negative results. CAP preferred to give laboratories performing NGS-based assays flexibility in determining when confirmatory testing should be performed, how this testing is performed, and whether to recommend confirmatory studies for follow-up testing for additional family members, which may or may not be NGS based. For example, some laboratories might determine during validation studies that confirmatory testing of identified variants was not necessary owing to the very high coverage achieved by their assay (ie, 1000× coverage of a single-gene NGS-based assay) and/or very high confidence in the identified variants.37,38 However, others may find that they need confirmatory testing by an alternative method to achieve the desired confidence in the variants that are reported. Some laboratories might decide that they will perform confirmatory testing on variants for a predetermined trial period and then reevaluate this decision at a later date. Each laboratory performing NGS must have a policy in place that clearly documents indications for confirmatory testing and/or documents how their assay validation determined that such testing was not required. Laboratories must be able to document compliance with their confirmatory testing policy and show evidence of ongoing monitoring of their NGS assay(s) to ensure that the benchmarks achieved during the validation process are maintained during the routine performance of NGS-based clinical testing and variant reporting. CAP also desired to give laboratories flexibility in deciding the methods used to perform any needed confirmatory testing. Although Sanger sequencing is likely to be the method most commonly chosen for confirmatory testing of NGS-identified variants, CAP did not want to mandate such testing in order to provide clinical laboratories with the flexibility to use other appropriate confirmatory testing methods consistent with the existing expertise of the laboratory and the type and frequency of variants requiring confirmation (ie, allele-specific polymerase chain reaction, melting curve analysis, other NGS-based method).
Methods, Instrument(s), and Reagents Used for Processing and Analyzing a Sample (or Batch of Samples) Can Be Identified and Traced in the Laboratory's Records
Comprehensive records of laboratory assay “runs” are essential to document the conditions and events associated with the complex processes and algorithms involved in the performance and interpretation of clinical NGS–based analyses. Accordingly, such archived information must be maintained within an overarching framework where all reagents, primers, sequencing chemistries, and platforms used for the analysis of each patient sample are traceable. Such records must contain a description of the test performed including the nature of the targeted sequence (eg, genome, exome, specific genes for targeted panels, transcriptome, or methylome) and depth of coverage (eg, range and average). It is also necessary to cite details of the analysis, including any publications or Web sites (with dates accessed) describing the pertinent parameters or other information and/or notations relative to the testing and reporting processes. While all details of the analysis need not be included in the patient report, it is critical that the laboratory maintain a documentation system from which detailed information regarding the analysis of individual patient specimens can be obtained.
The Laboratory Maintains an Exception Log for Patient Samples Where Steps Used in the NGS Analytic Wet Bench Process Deviate From Standard Operating Procedures
The laboratory must document any deviation from the SOP along with an explanation for the deviation, and the resulting outcome. Examples of anticipated deviations may include altered processing upon receipt of a suboptimal specimen, changes to the library preparation, and sequencing of libraries with suboptimal concentrations.
Exceptions may pertain to specimen quality and to the analytic process. At the time of specimen accessioning, an assessment is made as to whether or not a sample is in optimal condition for testing. If there is a concern, this can be documented on the worksheet or on a pending log and communicated to a supervisor or laboratory director. The director may decide to proceed with the testing, but should communicate the issue to the ordering physician and document this communication electronically or on the worksheet. One example of such a scenario is a sample that was not transported under optimal conditions. A decision may be made to process the sample and to proceed with subsequent testing only if the DNA specimen is found to be adequate.
Issues related to specific steps of the wet bench procedure should be reported to the laboratory supervisor or the director of the laboratory. It can then be assessed whether or not the testing was compromised and if the testing can be completed. If, after troubleshooting, the testing is assessed as satisfactory, the results can be interpreted by the laboratory director, provided that the quality controls of the run and the sample results are deemed adequate. All aspects of the testing issue(s) should be thoroughly documented in an “exception log,” including the troubleshooting, the resolution, and the pertinent communications (especially regarding who was involved and who was informed by whom and on what date), and may also be incorporated into the monthly quality assurance report.
On occasion, the laboratory SOP itself may have to be revised to improve phrasing, to make process steps more clear, or to remove small inaccuracies in order to optimize the protocol. In such cases, the proposed correction should ideally be supported by at least 2 additional individuals, including the laboratory supervisor and either the technologist who developed the assay or a reference technologist. Any such corrections must be approved, signed, and dated by the director of the laboratory. This is not an exception log issue per se but rather a correction in the manner the assay is described.
Monitoring of Upgrades
The Laboratory Has a Policy for Monitoring, Implementing, and Documenting Upgrades to Instruments, Sequencing Chemistries, and Reagents or Kits Used to Generate NGS Data
Laboratories must be aware of upgrades to ensure that they are not using obsolete methods. The laboratory must implement a policy to monitor and implement upgrades to instruments, sequencing chemistries, and reagents or kits used to generate NGS data. The policy should address how laboratories performing NGS-based testing can ensure that they are using the most up-to-date sample library preparation as appropriate for that assay, clonal fragment amplification, and sequencing methods in this rapidly evolving environment provided that these newer methods have been validated by the laboratory to improve the quality, reproducibility, and accuracy of the assay. The policy should also address the methods used to monitor upgrades and when a relevant upgrade(s) will be implemented and further validated before productive clinical use. For example, the laboratory's policy may be to monitor and implement upgrades at specified intervals (such as quarterly, biannually, or annually), depending on the relevance of the new upgrade for enhancing assay performance. Additionally, since the implementation of upgrades may require revalidation of the entire wet bench process, or at least the relevant steps, it may be convenient to set time intervals accordingly.
A variety of open-source and commercial bioinformatics algorithms and software is available for analyzing NGS data.39 While these tools continue to improve, they each have strengths and weaknesses with respect to their performance in diagnostic applications. Operationally, the bioinformatics processes applied to NGS data can be conceptualized into 3 major steps. First, is the generation of a sequence read file consisting of a linear nucleotide sequence (eg, ACTGGCA), with each nucleotide assigned a numerical value (termed its base quality score) that correlates to its predicted accuracy. The generation of sequence read files uses instrument-specific software that analyzes several physical parameters, such as signal to noise ratios, during the sequencing run. Sequence read files are usually configured in the FASTQ file format, which contains the compilation of individual sequence reads, each with its own identifier, and an associated base quality score for each nucleotide. FASTQ files have become a dominant form of information exchange in the field of NGS. The next step consists of aligning the sequence reads to a reference sequence, typically a human genome reference sequence, to identify differences between the patient sequence reads and the reference. Identified variants may include single nucleotide variants, insertions and deletions, copy number variants, and other structural variations (translocations, inversions, etc). Identified variants are then annotated to provide information regarding their impact on gene and protein function. Separate processes within the laboratory implement, or otherwise develop, curated content for assessing the clinical relevance of particular variants to a given disease or condition. Lastly, annotated variants are interpreted within the context of the patient's phenotype to render a clinical report. For gene panels and exome or genome sequences, the large list of annotated variants is typically reduced by excluding variants with a higher population frequency and by focusing on rare variants that are of greatest predicted deleterious impact that correlate with patient phenotype.40,41 When analyzing exome or genome sequences within a family unit, variant prioritization typically takes into account variant cosegregation within the family, based on affected versus unaffected family members. Variant prioritization during the tertiary step uses previous knowledge of association of variants and disease within public or private databases of human mutation, such as the Human Gene Mutation Database (HGMD),42 Online Mendelian Inheritance in Man (OMIM),43 and/or other disease/locus-specific databases.44
Developing a cohesive diagnostic pipeline that incorporates bioinformatics steps, and content development for variant annotation, usually requires the integration of multiple algorithms and software applications. As such, laboratories must empirically determine which algorithms and associated bioinformatics tools to apply to each diagnostic application. An iterative pilot process commonly uses known patient samples and training data sets, which may be synthetic or from prior cases, to test algorithms and software parameters. Having established a working set of bioinformatics tools and parameters, the laboratory performs a bioinformatics validation with a larger set of samples to determine analytic sensitivity and specificity for the types of variants assayed (eg, single nucleotide variants, insertions and deletions, homopolymer or repetitive sequences, or copy number variants) and reproducibility (ie, concordance within and across runs, instruments, and technical personnel). The samples used for validation will contain previously confirmed variants, or the identified variants may be confirmed post bioinformatics analysis. The validation may confirm that the bioinformatics tools and parameters are performing satisfactorily (eg, high specificity and sensitivity if the assay is a stand-alone assay for variant detection and reporting versus high sensitivity if it is a screening assay followed by a second assay that is used for confirmation) per laboratory requirements and clinical criteria for reporting, or adjustments or alternative tools may need to be further evaluated.
Once a satisfactory bioinformatics validation has been achieved, translation of the NGS assay into the clinical laboratory requires that laboratories document all aspects of the bioinformatics processes used for clinical diagnostics and implement a quality management program for these steps. Further highlights of the bioinformatics requirements for NGS are discussed below.
NGS Bioinformatics Pipeline Documentation
The Laboratory Uses an SOP to Document the Bioinformatics Pipeline Used to Analyze, Interpret, and Report NGS Results.
Laboratories must document all algorithms, software, and databases (referred to as components) used in the analysis, interpretation, and reporting of NGS results.45 The versions of each of these components in the overall bioinformatics pipeline must be recorded and traceable for each patient result (Version Control). For each component, the laboratory may use a baseline, default installation, or may customize the pipeline by using alternate configuration parameters in deploying individual bioinformatics tools or in running specific algorithms. In either case, laboratories must document any customizations that vary from default configuration or should indicate which parameters, cutoffs, and values are used. Most NGS bioinformatics analyses are conducted by aligning sequence reads to a reference sequence. The reference sequence version number and assembly details need to be identified. When describing the bioinformatics pipeline, laboratories should document the overall workflow of data analysis and include the input and output files for each process step. For each step, laboratories should also develop and document quality control parameters for optimal performance. For example, in the primary step, a laboratory would determine acceptable criteria such as the number of reads passing instrument-specified quality filters. Criteria for variant calling are essential and parameters that are invoked include thresholds for read coverage depth, variant quality scores, and allelic read percentages. Each of these requirements applies to multigene panel applications as well as to exome and genome sequencing. Laboratories should also document the bioinformatics processes that are used for reducing a large variant data set to a list of causal and/or candidate genes and/or variants. For example, in inherited disease assays, laboratories should document approaches used to identify recessive, dominant, and de novo variants. Evidence of compliance for this requirement would be demonstration of appropriate documentation and that the laboratory follows its outlined procedures.
NGS Bioinformatics Pipeline Validation
The Laboratory Validates the Bioinformatics Pipeline and Revalidates the Entire Pipeline and/or Confirms the Performance of the Components of the Pipeline as Satisfactory When Modifications Are Made. The Extent of Revalidation and/or Confirmation Is Modification Dependent.
As with wet bench processes, laboratories use an iterative process during the establishment of a bioinformatics pipeline that involves analyzing sequence read files containing known variants and demonstration that the pipeline can identify the variants.17 For laboratories offering the entire process from wet bench through bioinformatics analysis, the validation of the bioinformatics pipeline should be included in the overall test validation. Once the laboratory has developed and empirically determined optimal performance, and performed adequate testing of its pipeline, the next step is to perform and document a comprehensive validation, again using sequence reads generated from samples with variants that cover the spectrum of the diagnostic testing that the laboratory intends to perform. These steps are essential for both in-house–developed tools and in those cases where a vendor-provided tool or pipeline is used in a manner where it is locked down, for example, the laboratory does not modify or alter any components or parameters of the underlying tools. As with wet bench processes, a sufficient number of samples need to be analyzed to assess the pipeline's analytic and diagnostic sensitivity and specificity as well as the assay's reproducibility. The number of samples assessed should be determined from the assay. Parameters such as the number of genes assessed, which regions of a gene are assessed, and types of variants that need to be detected should ultimately be used to determine the number of control, well-characterized samples (eg, HapMap samples or cell lines with known inherent or engineered variants) and previously analyzed diagnostic samples. The presence of pseudogene sequences and other sequences highly homologous to the target are known to interfere with accurate sequence mapping, alignment, and, by extension, variant calling. The degree of interference, if applicable in a given diagnostic assay, needs to be determined. While it may be possible to address the challenge of coalignment of highly homologous sequences bioinformatically, laboratories may need to set up independent alternative method assays for these problematic regions. The NGS Work Group acknowledged that it was not feasible to comprehensively and exhaustively define the error rates for false positives and false negatives in variant calls. However, the laboratory should assess the error rates for several representative examples by variant type. These rates may be assessed analytically by using well-characterized, control samples. False-positive error rates can be ascertained by sequencing using an alternative method. False-negative error rates are more difficult to ascertain because they may originate from several sources, including insufficient read coverage in a given target region, a lack of variant calling due to parameters such as lower variant quality scores, or a distribution of sequence read directions (ie, forward and reverse reads) that does not meet quality control criteria. In the process of validating variant calls by an alternative method, it may be possible to analyze flanking regions to determine if the variant calling pipeline is identifying all possible variants. For example, when using Sanger sequencing to confirm a variant, primer pairs can be designed to sequence the region of the variant as well as generous portions of flanking regions, which can then be inspected for the presence of variants and correlated with those identified by the bioinformatics pipeline.
A now common practice in NGS is the use of molecular barcodes or indexes during the preparation of libraries. Indexed sequences need to be validated with respect to their uniqueness in a pool and the pipeline must be able to accurately bin (segregate) such indexed sequences. In the analysis of indexed and pooled samples, it is essential to establish criteria for retention or exclusion of sequence reads. For example, some laboratories will only accept sequence reads with indexes in which the index sequence is identical to the index that was used during library preparation. Other indexed reads that do not align in a completely identical fashion are not assigned to the respective sample. The monitoring of the percentage of indexed reads that maintained full identity can be a measure of the presence of contamination from other index sequences. For those assays in which limit of detection is relevant, such as identification of somatic mutations in tumor samples, the bioinformatics pipeline needs to be assessed for that parameter. One approach that can be used to validate limit of detection is to sequence samples with decreasing concentrations of target variants that have been created from a cell line or DNA dilution series.
Validation of the bioinformatics pipeline for identification of variants is application specific and the above discussion is broadly pertinent, with the exception of the limit of detection analyses being specific to samples with heterogeneous genotypes. When using exome and genome sequencing for causal and candidate gene identification, the laboratory must additionally validate its bioinformatics pipeline for this purpose. For example, in the case of inherited diseases, laboratories may approach this by analyzing sequence read sets with known pathogenic variants that are present in several deleterious variant configurations, such as recessive, dominant, and de novo.
Once a bioinformatics pipeline has been validated to meet laboratory requirements and has been implemented, revalidation is required when any changes are made in the pipeline. A practical approach that can be used to revalidate a sequencing pipeline is to use sequence read files from the original validation and simply reanalyze them with the new parameters. This approach may result in identical, smaller, or larger numbers of identified variants and these findings would need to be confirmed. For exome and genome sequencing, changes in bioinformatics pipelines can also result in a new list of presumptive causal or candidate genes. Evidence of compliance for a bioinformatics pipeline validation/revalidation would include the records of validation and any subsequent revalidation and their documented approval for clinical use.
NGS Bioinformatics Pipeline—Quality Management Program
The Laboratory Has a Documented Quality Management Program for the NGS Bioinformatics Pipeline
The routine application of a validated bioinformatics pipeline must be accompanied by monitoring of laboratory-determined quality control metrics.46 Divergence from expected quality metrics during the analysis of clinical samples requires investigation and resolution. Some examples include the following situations: the bioinformatics output of NGS data analysis may demonstrate that an insufficient number of sequence reads passed the expected or required base quality score threshold. Alternatively, the number of variants identified in a data set may deviate substantively from an expected value, based on prior information regarding known frequencies of variation in the human genome. Another example may be an inappropriately high number of indexed sequencing reads that cannot be specifically segregated. Such deviations may indicate a technical aberration or process failure occurring during technical wet bench procedures or during a step in the bioinformatics pipeline. An appropriate quality management program provides the structure and process for investigating these divergences to pinpoint possible causes, and institute appropriate corrective measures. Laboratories must maintain a record of deviations from expected results and document the investigative measures that were used to determine the cause as well as the corrective measures that were implemented. Evidence of compliance would include documentation of monitoring quality control metrics as well as records describing any divergences, including appropriate investigative measures and subsequent corrective actions.
The Laboratory Has a Policy for Monitoring, Documenting, and Implementing Patch-Releases, Upgrades, and Other Updates to the Bioinformatics Pipeline
This checklist item addresses the requirement for laboratories to establish and follow a procedure for identifying and implementing updates to components of the bioinformatics pipeline. Next-generation sequencing bioinformatics pipelines often use multiple packages of open-source software with additional scripts and databases for managing content and aspects of analysis and reporting. Owing to the ongoing evolution of the field, laboratories must have a policy for monitoring updates, patch-releases, and other upgrades to the bioinformatics pipeline. This policy should also address when such updates will be implemented. For example, the laboratory may decide to do this at the time of the update release or at specified intervals (such as quarterly, biannually, or annually), depending on the nature and relevant urgency of the update. Since such updates require revalidation of some or all of the bioinformatics pipeline (see “Bioinformatics Process” section on validation/revalidation), the latter approach of incorporating updates at set intervals may be more efficient, although again this depends on the update. Finally, the laboratory should maintain records that clearly document regular monitoring and implementation of updates. It should be emphasized that this requirement mandates a policy, but it is up to the discretion of the laboratory director if and when a particular update should be incorporated into the laboratory's bioinformatics pipeline.
The Laboratory Has a Policy Regarding the Storage of Input, Intermediate, and Final Data Files Generated by the Bioinformatics Pipeline
Laboratories must establish and follow a procedure for the storage of data files generated by the bioinformatics pipeline. Large data files are generated by NGS and the associated data analysis, including flow cell imaging files, sequence read files containing base calls and associated quality scores, other intermediate files generated after subsequent analysis steps, and variant text files. It is generally not practical to retain all such files for an extended period, so this checklist requirement mandates that the laboratory establish a policy for data storage that specifies data file retention times and which files will be retained after a final report has been generated. The NGS Work Group recommends that, if feasible, laboratories retain sequence files with corresponding quality scores (eg, FASTQ files47) or retain an archival format from which these files can be regenerated (eg, BAM files48). These formats will likely allow reanalysis at a later date, if indicated. For genome or other large-scale sequencing data, retention of FASTQ files or standard archival formats for long periods of time may be cost prohibitive with current storage technologies; however, newer compression formats provide one near-term solution.49,50 How long to store such files is a more complicated decision that depends on numerous issues, including the size of the data set, laboratory storage capacity, medicolegal considerations, as well as other institutional, local, state, or national requirements for data storage. Finally, it should be emphasized that the laboratory's policy for data storage and file retention times must be in accordance with local, state, and national requirements for storage of data.
The Specific Version(s) of the Bioinformatics Pipeline Used to Generate NGS Data Files Are Traceable for Each Patient Report
The specific versions of each component and, where available, associated configurations (eg, command line parameters or other configuration items) of the bioinformatics pipeline used to generate NGS data should be traceable for each patient report. As noted before, the bioinformatics pipeline for analyzing NGS data, especially when based primarily on open-source software, is often composed of a combination of different software packages, scripts, and databases. The performance of a single software package or script and the composition of an internal or external database can significantly impact the overall performance of the bioinformatics pipeline. Consequently, it is important for the laboratory to be able to connect each patient report to the particular bioinformatics pipeline used to generate the report. For in-house–generated scripts and software packages, changes in the script or software should also be documented, but documentation of each component of the pipeline does not need to appear in the patient report. Rather, it is acceptable to refer to the pipeline as a whole, using a laboratory-specific designation (eg, NGS Pipeline v1.0.1). Laboratory-specific designations should be unique to a single combination of pipeline components and configurations. Therefore, any change to a different version of a software package, script, or internal or external database, or change to the configuration of any software, would require a new unique laboratory-specific designation and would require assay revalidation.
The Laboratory Maintains an Exception Log for Patient Cases Where Steps Used in the Bioinformatics Pipeline Deviate From Standard Operating Procedures
Deviations from the laboratory SOP during any step used in the bioinformatics pipeline are documented in an exception log file, including any alterations in software packages, script, version number, database, command line, or parameters.51 Any failures arising during the bioinformatics process should also be recorded in the exception log and include documentation of the issues, the results of any investigations of these issues, any corrective actions taken, and pertinent communications, with sign-off by the laboratory director or designee. The exception log is also required to retain links to the patient reports, and the laboratory director may choose to communicate any clinically relevant SOP deviations to the ordering physician. Exception log documentation may also be incorporated into the monthly quality assurance report.
Deviation such as needing to rerun the analytic pipeline due to network, computer, or storage failure or memory issues, to run a particular step with different parameters or cutoffs than that used to validate the assay, must be documented along with the outcome and explanation. For example, a laboratory may need to alter settings on specific tools or components of its bioinformatics pipeline to adequately analyze particular regions or variants in a given patient case. The reason for the deviation should be described in the exception log, as well as the specific components of the deviation. Each deviation should be linked to the associated patient case and be reviewed by the appropriate laboratory director or designee(s). As warranted, the deviation, or aspects of it, may be included in the final report or in specific communication with the ordering physician.
Deviations related to bugs or failures in the bioinformatics pipelines also need to be recorded in the exception log. The bug, affected cases, and proposed corrective action must be approved, signed, and dated by the laboratory director or designee. Outright failures of the bioinformatics pipelines, which could result from hardware as well as software or operator error, should also be recorded in the exception log to document errors that may have occurred in analyzing individual patient cases.
Evidence of compliance for the exception log requires the ability to demonstrate appropriate documentation of review of the exception log by the laboratory director, demonstration that the laboratory records any issue arising during the bioinformatics procedure, and adequate documentation of subsequent corrective actions taken as a result of these reviews.
NGS Data Transfer Confidentiality Policy
The Laboratory Has a Policy and Procedures Describing Processes to Ensure That Internal and External Storage and Transfer of Sequencing Data Maintains Patient Confidentiality and Security
Next-generation sequencing generates significant amounts of data, particularly of gene sequences that, with other information such as name, date of birth, medical record number, and other components of protected health information, can potentially be used to identify individual patients. Laboratories must establish rigorous processes to ensure the protection and privacy of this information. Laboratories need robust policies regarding the transfer of genomic information to other health care entities and third-party vendors such as those providing cloud-based computing resources or reference laboratory services. Procedures to ensure confidentiality should include data encryption, secure data transfer, user authentication with controlled access to protected health information, and audit trails that track the transmission of data as well as the receiving entities and/or users. Laboratories should also follow standard requirements in the Health Insurance Portability and Accountability Act,52 such as establishing business agreements with external vendors that include sufficient due diligence to verify that appropriate methods are used to ensure confidentiality in the sending and receipt of patient clinical and genomic data.
Interpretation and Reporting of Sequence Variants Follows Professional Organizations' Recommendations and Guidelines
With the adoption of NGS technology, clinical laboratories are expanding their test menus from single gene testing to gene panels, and more recently, to exome and whole genome sequencing. It is evident that laboratories using NGS-based tests will come across a multitude of novel variants that have not been previously reported or classified as being causative of disease.
Currently, most laboratories report gene variants by using the Human Genome Variation Society nomenclature guidelines (www.hgvs.org; accessed January 6, 2014) and follow variant classification guidelines from the American College of Medical Genetics (ACMG)53 for inherited diseases. The nomenclature guidelines were originally published by Antonarakis and den Dunnen in 2001,54 but significant changes and additions have been made since then. Therefore, it is recommended that laboratories use the Web-based versions of these guidelines as they represent the latest revisions. The ACMG guidelines for variant classification are currently under revision and the new version will include interpretation guidelines. It is recommended that the ACMG classification system for inherited diseases be used as reference to increase consistency in variant classification. Disease and gene-specific modification may be necessary but should be documented. For other clinical genomic testing (eg, tumor or pathogen diagnosis), the laboratory should use its best judgment to categorize variants and adopt guidelines as they emerge.
Laboratories must also be aware of the lack of consensus in how transcript versions are used for variant numbering, an area that creates confusion in the literature, and can do the same in clinical reports. An example of this is provided for multiple transcripts produced from the MUTYH gene, which has complicated the nomenclature used to describe mutations identified in the gene. The 2 major transcripts are hMYHα1 (NM_012222.2) and hMYHα3 (NM_001048171.1), encoding polypeptides of 546 and 535 amino acids, respectively. The hMYHα3 transcript is 33 nucleotides shorter than the hMYHα1 transcript and results from alternative splicing of exon 3, which eliminates 11 amino acids from the 5′ end of exon 3 (GMIAECPGAPA). All other codons, and therefore amino acids, are identical between the 2 isoforms. Most literature uses the hMYHα3 variant when naming mutations; however, some reports do use the full-length transcript (hMYHα1). When reporting results for MUTYH testing, or comparing reports from different laboratories, it is imperative to note which transcript has been used to name the alteration(s) found in the gene. The laboratories have called the 2 most common alterations in this gene p.Tyr165Cys and p.Gly382Asp using the reference transcript hMYHα3 (NM_001048171.1), and p.Tyr176Cys and p.Gly393Asp using the reference transcript hMYHα1 (NM_012222.2). Laboratories should have a mechanism to monitor for such changes and use great caution when any variant designation changes are made in clinical reports. It is therefore useful to provide the transcript accession number and version along with the protein syntax in clinical reports to help avoid confusion.
Accurate interpretation of the combination of sequence variants observed in a specimen is a critical component of clinical testing, as it integrates variants that are potentially disruptive for gene function with the patient's clinical phenotype in order to determine whether identified variants may be causative for the disease for which the patient is undergoing testing. During the past few years, various terminologies have been used in clinical reports to denote the consequence of sequence changes, including pathogenic, deleterious, and disease associated, with qualifiers such as possible, probably, and likely, or VUS and VOUS (variant of unknown clinical significance and variant of uncertain clinical significance, respectively). Standardized sequence variant guidelines have been recommended for inherited diseases, while those for tumor or infectious pathogen diagnosis are still under flux. For inherited diseases, the most commonly applied classification is divided into 5 categories: (1) pathogenic, (2) likely pathogenic, (3) uncertain clinical significance, (4) likely benign, and (5) benign.
Laboratories should be cautious when interpreting the potential clinical consequences of sequence changes and carefully consider evidence for disease causation, frequency in the general population (including race/ ethnicity considerations), and functional studies. With the freely available exome and genome sequencing data from many large-scale projects (eg, Exome Variant Server, 1000 Genomes Project), laboratories should make use of relevant databases and computational tools39,55,56 (Table) when interpreting the effects of sequence changes. For large-scale tests such as exome and genome sequencing, it is also critical to assess the evidence implicating a gene in disease, understand the types of variants that have been implicated in disease as well as the mode of inheritance if known. For example, a novel loss-of-function variant in a gene with no established role in disease or a well-characterized variant spectrum should not be assumed to cause disease, despite the severity of the predicted impact to the protein.
Reporting of Incidental Genetic Findings
The Laboratory Has a Policy for Reporting Incidental Genetic Findings Unrelated to the Clinical Purpose for Testing
Clinically significant genetic findings that are unrelated to the phenotype for testing can occur when performing single gene, gene panel, exome, and whole genome sequencing. Limiting sequence analysis to a panel of genes that are relevant to the diagnosis of a particular disease state (either with targeted sequencing or targeted bioinformatics analysis) may limit, but not eliminate, the potential for incidental findings.57–64 This may include identification of variants relevant to autosomal dominant disease, carrier status for recessive diseases, predisposition to adult-onset dominant conditions (including cancer and neurodegenerative conditions), and drug response alleles commonly known as pharmacogenetic markers. Laboratories embarking on use of NGS for clinical testing should be aware of the potential for finding incidental, clinically significant results and should have a policy in place for whether and how these results will be reported for those assays where such incidental findings are expected (eg, exome). The recently published ACMG recommendations for reporting medically actionable incidental findings include a minimum gene list for which, if a known mutation is found, it should be reported.57 Laboratories may choose to follow the ACMG recommendations but are not expected to necessarily only report findings in these genes. Laboratories may also develop their own policies regarding return of incidental results. If the laboratory's policy is not to report incidental findings or to limit reporting to a subset of variants related to a particular disease state, this should be clearly stated in the laboratory report for assays where incidental findings are expected.
Ethical considerations must also be taken into account when deciding whether to reveal certain genetic information to patients. The level of risk associated with disclosing incidental findings depends on the severity of the disease, clinical actionability, and other risk-benefit indicators. For example, common disease risk alleles, such as for type 2 diabetes or cardiovascular disease, which have a small effect size (low relative risks), or pharmacogenetic risk information, may have different severity of consequence, compared to genetic information indicating a predisposition to cancer or a Mendelian disorder that may or may not be medically treatable. All of these facets must be considered before returning results to patients. Laboratories performing large-scale genomic sequencing analysis for clinical testing should be aware of efforts to study the medical and ethical implications of returning incidental results of NGS and consider these when developing their reporting policies.
NGS Test Referral Policy
The Laboratory Has a Policy for Selection of Reference Laboratories and Other Service Providers for NGS Test Referral. Referral May Include the Entire NGS Test Process or Only the Wet Bench or Bioinformatics Processes
The complexity of NGS clinical testing with challenging data analysis has led to some interesting transformation of the traditional laboratory testing model. Although most clinical laboratories are conducting the full NGS testing process in-house, there are a growing number of laboratories that have started outsourcing parts of the testing workflow to external facilities. For example, CLIA laboratories that do not have extensive data analysis capabilities are sending out the bioinformatics data analysis to external facilities. On the other hand, bioinformatic vendor companies, or institutions that are entering the clinical testing arena and have sophisticated data analysis capabilities but do not have a CLIA laboratory facility to process DNA for sequence generation, are outsourcing the wet bench workflow to external laboratories.
With this emerging trend in fragmentation of the clinical workflow, the NGS Work Group decided to include a new requirement for the 2014 version of the NGS Checklist concerning test referral policy. The CAP Laboratory General Checklist requirement 41350 states that the laboratory director or designee is responsible for the selection of an external reference laboratory or other service provider. It is expected that the laboratory director or designee ensure the quality of performance of the external NGS wet bench and/or bioinformatics service provider. Some of the specific aspects that the laboratory director needs to consider in selecting external reference laboratories or service providers are to ensure that (1) the turnaround times are acceptable for the clinical needs for which testing is being done; (2) the external laboratory providing the analytic wet bench information, (ie, sequence generation) is a CLIA-certified laboratory or a laboratory meeting the selection criteria as per CAP requirements; and (3) the quality of the results from the external bioinformatics service provider is verified to be accurate and of high standards. For evidence of compliance, copies of valid CLIA certificates from CLIA-certified external reference laboratories are required for those who outsource the wet bench sequencing workflow. Copies of in-house validation of non–CLIA-certified entities providing bioinformatics analysis are required for those entities that outsource the bioinformatics workflow.
The translation of NGS from basic to clinical research and adoption for clinical diagnostics has occurred over a relatively short period of time. A growing number of clinical laboratories are implementing NGS-based diagnostic assays, mostly in the form of multigene panels, although an increasing number of laboratories are performing exome and genome sequencing. CAP identified that the adoption of NGS by clinical laboratories required the development of accreditation requirements specific to NGS. This report highlights the content of the accreditation requirements that were developed by the CAP NGS Work Group in an effort to respond to the clinical adoption of NGS and to articulate the perspectives of the NGS Work Group in developing the requirements. The NGS field continues to evolve at a rapid pace and this evolution reflects continuing improvements in NGS instrumentation, chemistries, and analytic tools and software. The work group took the perspective that the NGS accreditation requirements should strike a balance that would ensure patient safety and also foster the responsible adoption of NGS in clinical laboratories, using fundamental diagnostic laboratory principles. These principles include documentation of process steps, demonstration that laboratory procedures have been well characterized through proper validation, and institution of quality management programs. Given the reality of this rapidly changing technology, and because NGS-based tests can be varied depending on type of tests and scale of analysis, the NGS Work Group developed general requirements for a variety of NGS clinical testing scenarios to allow laboratories flexibility and latitude in individual approaches to meeting the requirements, while at the same time providing much needed regulatory standards. Some examples used in this article regarding nomenclature, interpretation, and incidental findings highlight and are most applicable to NGS for inherited disease testing. Practically, it was more straightforward to develop laboratory standards for these topics as they have been previously addressed in accreditation requirements for single gene tests in inherited disease. Notably, guidance documents on topics of nomenclature, interpretation of variants, and incidental findings have been previously published. The work group elected not to prematurely introduce requirements on NGS topics that would benefit from further technology evolution and/or deliberation and consensus building at the professional society level for specific clinical disciplines. An example is that guidelines for interpretation of somatic variants need to be developed before introducing a correlate accreditation requirement. Whereas this topic is not unique to NGS-based molecular oncology testing, NGS multigene panel and exome testing have heightened the need for somatic variant interpretation guidelines. Additional, future topics for accreditation requirements for NGS-based molecular oncology could include the role of preanalytic sample assessment and processing of oncology samples and their influence on NGS testing results. Preanalytic metrics, such as formalin-fixed, paraffin-embedded variables influencing NGS results, have not been established in the field and therefore this would be another example where the committee felt it would be premature to include specific requirements. Further accreditation requirements could address the quantitative aspects of both molecular oncology and infectious diseases NGS testing. In essence, there are an emerging number of discipline-specific topics for new accreditation requirements that will be introduced as the field of NGS testing matures and consensus on practice is built through professional experience.
The NGS Work Group acknowledged that accreditation requirements for NGS would need to be revisited and revised as part of an ongoing process as the field of NGS-based diagnostics evolved and matured. With the publication of the accreditation requirements, CAP has subsequently fielded many questions from laboratories as they implement NGS. Specifically, laboratories are periodically seeking clarification with respect to the requirements, and the questions being posed are providing the work group with feedback that is guiding our discussions focused on improving and clarifying these requirements. In addition, the NGS work group has recognized that the initial requirements could not cover all of the NGS applications that clinical laboratories might pursue. Operationally, the NGS work group meets throughout the year via teleconferences and face-to-face meetings. As a part of these activities, discussions focused on revising and expanding accreditation requirements occur. Within the CAP accreditation requirements revision and publication cycle, the NGS work group has the ability to formally submit revised and expanded NGS requirements. The current requirements reflect those that have undergone 1 formal cycle of revisions, and as of this writing, additional revisions and expansions are underway. Again, feedback and experience from the field will be incorporated into the work group's current and future deliberations, resulting in further explanation and expansion of these requirements in future versions of the NGS-related accreditation requirements.
The authors have no relevant financial interest in the products or companies described in this article.