- Introduction
- Chapter 1 From Sequence to Care: The Clinical Genomics Landscape
- Chapter 2 DNA 101 for Clinicians: Genes, Variants, and Mechanisms
- Chapter 3 Clinical-Grade Sequencing Platforms: Panels, Exomes, and Genomes
- Chapter 4 Somatic Sequencing in Oncology: Tumor Profiling and Liquid Biopsy
- Chapter 5 Long-Read, RNA, and Methylation Sequencing in the Clinic
- Chapter 6 Quality, Validation, and Accreditation: Meeting CLIA/CAP Standards
- Chapter 7 Bioinformatics Pipelines and Quality Control for Diagnostic Reporting
- Chapter 8 Databases and Evidence: gnomAD, ClinVar, and Structured Literature Curation
- Chapter 9 Variant Interpretation with ACMG/AMP and Disease-Specific Guidelines
- Chapter 10 Functional Studies and Model Systems to Resolve Uncertainty
- Chapter 11 Reporting Essentials: Clarity, Actionability, and Limitations
- Chapter 12 Incidental and Secondary Findings: Policies, Pitfalls, and Practice
- Chapter 13 Pharmacogenomics: Optimizing Drug Choice and Dosing
- Chapter 14 Rare Disease Diagnostics and Trio Analysis
- Chapter 15 Reproductive and Prenatal Genomics: Carrier Screening, NIPS, and Fetal Sequencing
- Chapter 16 Infectious Disease Genomics: Pathogen Identification and Antimicrobial Resistance
- Chapter 17 Cardiovascular, Neurologic, and Metabolic Disorders: High-Value Use Cases
- Chapter 18 Pediatric vs Adult Genomic Care: Consent, Assent, and Longitudinal Considerations
- Chapter 19 Genetic Counseling in the Genomic Era: Communication and Cascade Testing
- Chapter 20 Integrating Genomics into Clinical Workflows and the EHR
- Chapter 21 Molecular Tumor Boards and Multidisciplinary Decision-Making
- Chapter 22 Health System Implementation: LIMS, Staffing, and Turnaround Time
- Chapter 23 Data Governance, Privacy, and Security in Precision Medicine
- Chapter 24 Equity, Access, and Reimbursement: Real-World Barriers and Solutions
- Chapter 25 Measuring Impact and Looking Ahead: Outcomes, Trials, and Future Horizons
Genomes at Work: Translating Sequencing into Patient Care
Table of Contents
Introduction
Genomic sequencing has moved from the research bench into clinics, wards, and pharmacies, where it increasingly informs diagnosis, risk stratification, and therapy. Yet the path from raw sequence data to a confident, clinically actionable recommendation is complex. It requires technology choices that meet clinical standards, rigorous interpretation grounded in evidence, and workflows that respect time pressures, patient preferences, and ethical boundaries. This book was written to help clinicians, genetic counselors, laboratorians, pharmacists, and health system leaders navigate that path with clarity and confidence.
Genomes at Work focuses on translation—what it takes to use sequencing results to improve individual patient care. We begin with the building blocks: the biology of genes and variants; the advantages and limitations of targeted panels, exomes, genomes, and emerging modalities like long-read, RNA, and methylation sequencing; and the quality systems that distinguish clinical-grade testing from research assays. Alongside technical fundamentals, we emphasize the practicalities of bioinformatics pipelines, validation, and quality control, so that readers can recognize a robust report and know where uncertainty may lurk.
Interpretation is the beating heart of clinical genomics. Throughout the book, we unpack frameworks such as ACMG/AMP criteria, illustrate the role of population databases and disease-specific guidelines, and show when functional evidence can resolve a variant of uncertain significance. We devote dedicated chapters to the art and craft of reporting—clear statements of actionability, explicit limitations, and transparent rationales—because a well-constructed report is a clinical tool, not a data dump.
Real-world care rarely fits a textbook mold, so case studies thread through the chapters. In oncology, we explore how tumor and liquid biopsy results guide targeted therapy and trial enrollment. In rare disease, we consider trio-based analyses, reanalysis over time, and the power of orthogonal evidence. We cover reproductive genetics, infectious disease applications, and high-yield use cases in cardiology, neurology, and metabolism. At each step, therapeutic decision-making is foregrounded: what to do now, what to consider next, and when to refer, re-test, or revisit.
Implementation at scale brings its own challenges. We examine clinic–lab interfaces, molecular tumor boards, and integration with the electronic health record, including decision support and structured data capture. We discuss staffing models, laboratory information management systems, turnaround time, and stewardship of limited resources. Because precision medicine succeeds only when it is equitable and sustainable, we address reimbursement realities, access barriers, and strategies to reduce disparities in testing and downstream care.
Finally, genomics in practice raises profound ethical questions. We provide pragmatic guidance on consent, assent, and return of results across the lifespan; policies for incidental and secondary findings; privacy and data security; and communication strategies that respect uncertainty while supporting shared decision-making. The concluding chapters focus on measurement: how to evaluate clinical utility, track outcomes, design pragmatic trials, and continuously improve programs as evidence and technologies evolve.
Whether you are ordering your first genomic test, counseling families through complex results, building a service line, or refining a mature program, this book aims to be a hands-on companion. Read it cover to cover for a comprehensive view, or dip into specific chapters as needs arise. Above all, we hope to help you put genomes to work—reliably, ethically, and effectively—for the patients who trust us with their care.
CHAPTER ONE: From Sequence to Care: The Clinical Genomics Landscape
The journey from a tube of blood or a slice of tumor to a treatment decision is shorter than it used to be, but it is still more complicated than most clinicians expect. A few decades ago, sequencing a human genome was an international project with an astronomical price tag and a timeline measured in years. Today, a clinical genome can be generated in days for a fraction of the cost, and targeted panels deliver answers even faster. The promise is seductive: identify the cause of a child’s developmental delay, match a patient with metastatic cancer to a life-extending therapy, or spare someone an adverse drug reaction before it happens. Yet the gap between a raw data file and a confident recommendation remains full of decisions, judgments, and trade-offs.
Clinical genomics is not a single test but a spectrum of assays, each with different biological scope and clinical intent. A targeted panel might examine dozens of genes known to cause a specific phenotype, delivering high depth of coverage and rapid turnaround. An exome sequences the protein-coding regions of roughly 20,000 genes, trading breadth for lower cost and a more complex interpretation workload. A genome spans the entire DNA sequence, offering the most comprehensive view but also the greatest challenge in filtering and prioritizing findings. Each of these can be performed on germline DNA (inherited variants), somatic DNA (tumor-specific changes), or both, and they can be augmented with RNA sequencing to assess splicing and gene expression, methylation profiling to map epigenetic states, or long-read technologies that resolve complex regions inaccessible to short-read instruments.
The fact that we can do these tests does not automatically make them useful. Clinical utility is the bedrock concept: does the test result change patient management in a way that improves outcomes? This is not a purely technical question. It depends on the clinical context, the pre-test probability that a genetic cause is present, the availability of effective interventions, and the patient’s goals and values. For many conditions, the diagnostic yield of comprehensive sequencing is substantial but not infinite; a negative result is as meaningful as a positive one, guiding clinicians to look elsewhere. Even when a pathogenic variant is identified, downstream actions—therapy selection, surveillance, family testing—must be available, feasible, and acceptable to the patient.
The clinical laboratory’s role is central. In the United States, laboratory-developed tests or FDA-approved kits are performed under the Clinical Laboratory Improvement Amendments (CLIA) framework, often supplemented by College of American Pathologists (CAP) accreditation. These standards govern personnel qualifications, quality control, proficiency testing, and validation. Similar frameworks exist globally, such as ISO standards and national regulations in the European Union. The key point for ordering clinicians is straightforward: not all sequencing is created equal. A research-grade assay may be informative, but a clinical-grade test provides documented analytical validity, clinical validation, and a structured report that supports medical decision-making.
Analytical validity asks whether the test accurately measures what it claims to measure. For sequencing, this includes coverage (how many times each base is read), uniformity across the genome, and the ability to call different types of variants, from single nucleotide variants to small insertions and deletions, and, increasingly, copy number changes and structural variants. It also includes sensitivity to mosaicism, where only a fraction of cells carry a variant, and the ability to sequence repetitive regions, which are often problematic for short-read technologies. Clinical validation, in turn, demonstrates that the test performs well in the intended clinical population and that results can be interpreted and acted upon appropriately.
The interpretation pipeline is where raw variants are transformed into clinical knowledge. After quality control, reads are aligned to a reference genome and variant calling is performed. Filters remove common polymorphisms by comparing population databases such as gnomAD, while computational predictors estimate the functional impact of amino acid changes or splicing disruptions. Evidence sources—disease databases like ClinVar, literature curation, and functional studies—are then integrated. Clinicians need not be bioinformaticians, but they should understand that the pipeline includes many steps where assumptions and thresholds can affect the final report, and that these should be transparent and auditable.
Not all findings are expected. Incidental or secondary findings are variants discovered outside the test’s primary indication. Policies on whether to look for them, what conditions to include, and how to return results vary by institution and jurisdiction. Professional guidelines propose a minimum list of genes associated with medically actionable conditions, but patients may opt in or out, and the logistics of consent, counseling, and follow-up are nontrivial. A lab may issue a separate report for secondary findings, and clinicians must be prepared to discuss their meaning, the limits of the analysis, and the options for further evaluation or declination.
In oncology, genomic testing often targets somatic alterations within a tumor to guide therapy. The scope ranges from single-gene tests, such as EGFR mutation analysis in lung cancer, to large panels that detect fusions and copy number changes. Liquid biopsy—analysis of circulating tumor DNA in plasma—has expanded the reach of somatic testing by enabling minimally invasive monitoring and detection of resistance mutations. Interpreting tumor variants requires careful distinction between pathogenic driver events and passenger mutations, and results may prompt targeted therapy, immunotherapy consideration, or eligibility for a clinical trial. Turnaround time matters in advanced disease; therefore, many centers prioritize time-to-result over breadth when clinical urgency is high.
Reproductive and prenatal applications illustrate how genomics spans health and life stages. Carrier screening panels identify couples at risk of transmitting recessive disorders, allowing reproductive planning or prenatal diagnosis. Noninvasive prenatal testing analyzes cell-free fetal DNA in maternal blood for aneuploidies and selected microdeletions, offering higher sensitivity and specificity than traditional serum screening. In some cases, fetal exome or genome sequencing is considered for severe ultrasonographic abnormalities. These scenarios raise distinct consent and counseling needs, particularly regarding variant classification in the fetus, potential detection of maternal conditions, and the implications for future pregnancies.
Inherited disease testing also highlights the importance of family context. Trio sequencing—analyzing the patient and both parents—can clarify inheritance patterns, distinguish de novo from inherited variants, and assist in filtering large numbers of candidate variants. For many pediatric disorders, the diagnostic yield of trio exome or genome is significantly higher than proband-only testing, and a diagnosis can redirect care immediately, from syndrome-specific surveillance to surgical planning. In adult-onset conditions, the same data may inform risk-reducing strategies for family members, but must be balanced against insurance discrimination concerns and psychological impact.
Pharmacogenomics is the branch of clinical genomics focused on drug response. Variants in genes such as CYP2D6, CYP2C19, and HLA-B can influence drug metabolism, efficacy, and risk of adverse events. Pre-emptive pharmacogenomic testing is increasingly being integrated into practice, with results placed in the electronic health record to inform prescribing at the point of care. Implementation requires careful translation of star alleles and diplotypes into clinical recommendations, and understanding that ethnicity-specific allele frequencies and haplotype structures can affect interpretation. The goal is simple but impactful: the right drug at the right dose for the right patient at the right time.
Infectious disease genomics has grown rapidly, particularly in microbiology laboratories. Pathogen identification using targeted or whole-genome sequencing can outperform traditional methods, especially in slow-growing or fastidious organisms. More importantly, detecting resistance determinants and transmission clusters guides antimicrobial stewardship and infection control. For example, sequencing Mycobacterium tuberculosis can identify drug resistance patterns that directly inform regimen selection. In hospital outbreaks, rapid sequencing can link cases and inform source tracing. These applications demand robust contamination controls and timely reporting, because clinicians often need answers within hours, not days.
Genomic medicine is as much about workflow as it is about science. A test that sits on a shelf unused because it is cumbersome to order, slow to return results, or difficult to interpret offers no benefit. Health systems must design pathways that fit clinical realities: clear indications for testing, efficient sample collection and transport, an informatics backbone that delivers results to the right clinician at the right time, and mechanisms for follow-up. The electronic health record is central, but integration is often partial. Structured data capture and decision support tools can nudge practice toward guideline-concordant use and make downstream analysis of outcomes feasible.
Another operational pillar is the laboratory information management system, or LIMS. It tracks samples from accessioning through sequencing and reporting, ensures chain of custody, and records quality metrics. It also manages turnaround time monitoring, customer service queries, and the resolution of cancelled or inconclusive tests. Staffing models are equally important. Clinical genomics requires a multidisciplinary team: molecular pathologists, clinical laboratory geneticists, bioinformaticians, genetic counselors, and IT specialists. Many institutions use molecular tumor boards or similar forums to discuss complex cases and align interpretation with therapeutic options and clinical trial opportunities.
Reimbursement influences adoption and equity. Coverage decisions vary by payer, condition, and test type. Some panels are well supported when evidence demonstrates improved outcomes; others face hurdles when clinical utility is uncertain. Costs extend beyond the assay itself to pre-test counseling, interpretation, and post-test management. Health systems that invest in genomics must be able to articulate value: earlier diagnosis, avoidance of ineffective therapies, reduced complications, and improved quality of life. Transparent documentation of indications and outcomes helps build the case for sustainable reimbursement and broader access.
Ethical, legal, and social considerations are woven through every stage. Consent is more than a signature; it is an ongoing conversation about what will be tested, how results will be used, and who will have access to data. Pediatric testing raises issues of assent and future autonomy. Data privacy and security are paramount; de-identification, controlled access, and encryption are baseline protections. The potential for discrimination and stigmatization must be acknowledged, as must the risk of uncertain findings that cause anxiety without clear guidance. Respecting patient preferences and local regulations is not a barrier to genomics but a prerequisite for trust.
Equity and access are persistent challenges. Genomic medicine has historically concentrated in academic centers and affluent communities. Barriers include cost, lack of provider awareness, limited genetic counseling in rural areas, language and cultural differences, and historical injustices that shape trust. Solutions involve broadening indications for testing, funding patient navigation, telehealth models for counseling, and diverse genomic databases that reflect global ancestry. Embedding genomics into primary care, with appropriate supports, can democratize access while maintaining quality. When appropriate, clinicians should consider whether a patient might benefit from research participation, while maintaining clear boundaries between research and clinical care.
Real-world case studies will appear throughout this book, but it is helpful to sketch the kinds of decisions that arise. Consider a child with developmental delay and dysmorphic features: targeted testing may detect a well-known syndrome, but an exome or genome may be needed when the presentation is atypical. A patient with metastatic non–small cell lung cancer: an upfront panel may reveal an actionable fusion; when the tumor sheds poorly, a liquid biopsy may be the only option. An adult referred for preventive care with a family history of sudden cardiac death: a cardiomyopathy panel might identify a pathogenic variant that prompts cascade testing of relatives and lifestyle modifications. In each scenario, the clinician must match the test to the question, anticipate limitations, and plan for how results will inform management.
What should clinicians watch for when adopting or ordering genomic tests? Look for evidence of analytical and clinical validation, clear documentation of indications, and transparency about the limitations of the assay. Ask about turnaround time and what happens if a variant of uncertain significance is identified. Expect reports that state confidence, actionability, and next steps, not just annotations. Confirm that genetic counseling is available before and after testing, especially for germline findings with family implications. Finally, remember that genomics is dynamic: variants may be reclassified as evidence accumulates, and reanalysis of existing data can yield new insights.
The landscape is expanding, but the fundamentals remain constant. Genomics offers powerful tools, not universal solutions. The best outcomes come from disciplined use of the right test at the right time, high-quality interpretation, and thoughtful integration into patient care. With that orientation, we now move from the big picture to the basics: DNA, genes, and the mechanisms by which variants cause disease—the subject of the next chapter.
CHAPTER TWO: DNA 101 for Clinicians: Genes, Variants, and Mechanisms
Genomics is the study of the entire set of DNA in an organism, the genome. In practice, clinical genomics focuses on specific parts of that genome that are most likely to influence health: genes and the sequences that regulate them. For clinicians, the central idea is simple yet powerful. DNA encodes instructions for making proteins and RNA molecules that carry out cellular functions. When those instructions change, function can be altered, leading to disease or changes in drug response. Understanding the language of DNA and the consequences of its variation turns raw sequencing data into clinical insight rather than noise.
The genome is organized into chromosomes, which are long DNA molecules packaged with proteins. Humans have 23 pairs of chromosomes, including one pair of sex chromosomes. The majority of the genome does not code for proteins; regulatory elements, noncoding RNAs, and repetitive sequences occupy much of the sequence. Genes are segments of DNA that are transcribed into RNA and, in most cases, translated into proteins. Each gene has a promoter region that initiates transcription, exons that make up the mature messenger RNA, and introns that are removed during splicing. The boundaries between these regions influence how variants affect gene products.
The genetic code translates nucleotide triplets, called codons, into amino acids. There are 64 possible codons encoding 20 amino acids plus stop signals. The code is redundant; multiple codons can specify the same amino acid, which buffers some variation. A change in a codon may have no effect if it yields the same amino acid, a conservative substitution if the new amino acid is similar, or a disruptive change if the chemistry or size of the protein’s building block is altered. Start and stop codons define the beginning and end of translation; variants that affect these boundaries can have outsized effects on protein length and function.
Variants are alterations from a reference sequence and are described using standardized nomenclature, typically HGVS. A variant might be as small as a single nucleotide change or as large as a deletion or duplication spanning many genes. In clinical reports, variants are categorized as pathogenic, likely pathogenic, uncertain, likely benign, or benign. Pathogenic variants are known to cause disease; uncertain variants, or VUS, lack sufficient evidence for classification. This framework helps clinicians gauge confidence and plan next steps, but it also requires awareness that classification can change as new data accumulate.
There are several common types of variants. Single nucleotide variants change a single base and can alter an amino acid or disrupt splicing. Insertions and deletions, called indels, add or remove one or more bases. If the indel’s length is not a multiple of three, it shifts the reading frame, often leading to a premature stop codon and a truncated protein. Larger structural variants include deletions, duplications, inversions, and translocations, which can affect single genes or entire chromosomal segments. These may be detectable by sequencing but often require different analysis strategies than single nucleotide changes.
Variants can be categorized by their location and functional effect. Missense variants change an amino acid. Nonsense variants introduce a premature stop, typically producing a nonfunctional protein. Splice-site variants occur near exon–intron boundaries and can disrupt normal splicing, sometimes causing exon skipping or intron retention. Synonymous variants do not change the amino acid, but they can still affect splicing or mRNA stability. Regulatory variants in promoters or enhancers can alter how much protein is made, which may be as impactful as changes to the protein sequence itself.
Genetic inheritance patterns describe how variants are transmitted. Autosomal dominant conditions often arise from a single altered copy of a gene, while autosomal recessive conditions require variants in both copies. X-linked inheritance affects genes on the X chromosome and can present differently in males and females due to dosage differences. Mitochondrial DNA is inherited maternally and can cause disorders with variable tissue involvement. Understanding inheritance helps interpret family histories and informs recurrence risks for relatives and offspring.
The clinical consequences of variants are influenced by penetrance and expressivity. Penetrance refers to the likelihood that a person with a pathogenic variant will manifest the associated condition; it can be complete, high, or incomplete. Expressivity describes the range and severity of features among individuals with the same variant. Some variants have age-dependent penetrance, such as those associated with adult-onset neurodegenerative disorders. Environmental factors, modifier genes, and epigenetic influences can also shape phenotype, reminding us that a genotype does not always predict a precise clinical picture.
It is essential to distinguish germline from somatic variants. Germline variants are present in every cell because they are inherited or occur early in embryonic development. They underlie hereditary disorders and influence pharmacogenomics. Somatic variants arise later in life, often in specific tissues such as tumors. They can be caused by environmental exposures, replication errors, or clonal expansion. Two people can have different somatic variants in their tumors, and the same somatic variant may be found in different cancers. Somatic variants are generally not inherited by offspring.
Mosaicism adds another layer of complexity. Mosaicism occurs when a variant is present in only a subset of an individual’s cells. If the variant arises after fertilization, the person has a mixture of cells with and without the variant. The clinical impact depends on which tissues are affected and the proportion of cells carrying the variant. Mosaicism can be challenging to detect by sequencing, particularly if the variant allele fraction is low. In some cases, testing multiple tissues, such as blood and skin, can increase sensitivity.
The distinction between gain-of-function and loss-of-function mechanisms is clinically relevant. Loss-of-function variants typically reduce or eliminate protein activity; this can occur through nonsense changes, frameshifts, or deletions. Gain-of-function variants increase protein activity or confer a new, harmful function; they can lead to overactive signaling pathways or abnormal protein aggregation. For some genes, the same variant can cause disease through a dominant-negative mechanism, where the altered protein interferes with the function of the normal protein from the other allele.
Variants that affect splicing illustrate how subtle changes can have dramatic effects. The canonical splice sites at exon–intron boundaries are highly conserved, and variants here usually disrupt splicing. However, variants deeper within introns or exons can create or destroy splice enhancers or silencers, leading to aberrant splicing even when the canonical sites are intact. RNA sequencing can help confirm splicing consequences, but prediction algorithms are often used in clinical pipelines to prioritize variants that may affect splicing, especially when the phenotype fits but no obvious coding variant is found.
Regulatory variants in noncoding regions highlight the importance of context. Promoters, enhancers, and insulators coordinate gene expression in time and space. A variant that alters a transcription factor binding site can reduce or increase gene expression, sometimes leading to disease without changing the protein sequence. Regulatory elements may act over long distances, affecting genes far away. Understanding these mechanisms can explain phenotypes that arise despite a normal exome, and it underscores why whole-genome sequencing, which captures noncoding regions, is becoming more clinically relevant.
Copy number variation refers to changes in the number of copies of a gene or genomic segment. Deletions and duplications can alter gene dosage, leading to haploinsufficiency or triplosensitivity. Some conditions, such as 22q11.2 deletion syndrome, involve multiple genes and have characteristic features. Detecting copy number variants from sequencing data is feasible but requires specialized analysis; different sequencing strategies may be needed to confidently call large deletions or duplications. Clinical interpretation must consider gene dosage sensitivity and the known phenotypes associated with specific segments.
Structural variants can rearrange the genome, creating fusion genes or disrupting regulatory architecture. Balanced translocations may not alter copy number but can place a gene under a new promoter, causing aberrant expression. In cancer, fusion oncogenes are common drivers; in hereditary disease, translocations can disrupt tumor suppressor genes. Sequencing technologies differ in their ability to detect structural variants; some methods excel at small variants, while long-read or optical mapping approaches better resolve large rearrangements. Recognizing the strengths and limits of the assay helps interpret the absence or presence of certain variant types.
The reference genome is a crucial tool but is not perfect. It is a composite sequence that may not reflect any single individual’s DNA. It is also incomplete in certain regions, such as telomeres, centromeres, and highly repetitive segments. Some variants are difficult to map to the reference due to sequence similarity, which can cause false calls. In clinical practice, laboratories must be aware of reference-specific artifacts and regions of low confidence. As the human pangenome reference expands, more of these challenges will be mitigated, but awareness remains essential.
Allele frequency is a key filter in variant interpretation. If a variant is common in the general population, it is unlikely to cause a severe Mendelian disease, unless the condition has late onset or variable expressivity. Population databases such as gnomAD provide frequencies across ancestries. Variants that are rare or absent in healthy populations rise to the top of the list for pathogenicity. However, some conditions have founder variants that are common in specific populations. Ethnicity-specific context is therefore critical for accurate interpretation and counseling.
Computational predictors estimate the functional impact of variants. Tools like PolyPhen, SIFT, CADD, and REVEL integrate evolutionary conservation, biochemical properties, and genomic features to assign a score. Splicing predictors, such as SpliceAI, assess the likelihood that a variant disrupts splicing. These tools are useful but not definitive; they generate probabilities, not clinical truths. High scores can raise suspicion, but evidence from disease databases, family segregation, and functional studies is required for classification. Overreliance on predictions is a common pitfall.
Some variants are inherited in a way that affects clinical management beyond the individual. Autosomal recessive conditions often involve two variants in the same gene, which may be of different types. For example, one allele might have a missense change and the other a frameshift. Compound heterozygosity means two different pathogenic variants are present on the two alleles, leading to disease. In contrast, homozygosity refers to the same variant on both alleles. Understanding these patterns informs recurrence risk for offspring and helps guide family testing.
Allelic heterogeneity describes how different variants in the same gene can cause the same condition. For example, pathogenic variants in CFTR can cause cystic fibrosis through many distinct mutations. Locus heterogeneity occurs when the same disease can be caused by variants in different genes. For instance, retinitis pigmentosa can arise from variants in numerous genes. This diversity complicates gene selection for testing and interpretation, emphasizing the value of comprehensive assays when the phenotype is broad and multiple genes could be involved.
Variable expressivity and pleiotropy further complicate predictions. Variable expressivity means individuals with the same variant show a range of severity or features. Pleiotropy describes a single gene influencing multiple organ systems. A variant may lead to a syndrome with heart, eye, and skin manifestations. Clinically, this affects differential diagnosis; a patient presenting with only one feature may still carry a pathogenic variant that explains the broader syndrome. Recognizing patterns across systems is crucial in selecting genes and interpreting variants.
Penetrance can be age-dependent or influenced by environmental factors. Some variants confer risk but only manifest later, such as in hereditary cancer syndromes or neurodegenerative diseases. This affects screening strategies and surveillance schedules. In oncology, a somatic variant might be present in a tumor for years before it becomes actionable, and a germline variant might predispose to cancer with varying age of onset. Understanding the natural history of associated conditions guides the timing and intensity of interventions.
Imprinting is an epigenetic phenomenon where gene expression depends on parental origin. Variants in imprinted regions can lead to disease only when inherited from a specific parent. Prader-Willi and Angelman syndromes are classic examples caused by deletions in an imprinted region on chromosome 15, with contrasting phenotypes depending on which parent contributed the deletion. Imprinting errors can also occur de novo. Clinical interpretation must account for parental origin when imprinting is relevant.
Mitochondrial genetics differ from nuclear inheritance. Mitochondrial DNA is maternally inherited, and cells contain many copies of it. Pathogenic variants can be homoplasmic, present in all copies, or heteroplasmic, present in a subset. The proportion of variant mitochondria can change over time and varies by tissue, influencing symptom severity. Testing often requires specialized approaches to quantify heteroplasmy. Counseling must address the maternal transmission risk and variable penetrance.
Genomic context influences variant effect. A variant may behave differently depending on the background of other variants in the genome. Modifier genes can ameliorate or worsen the phenotype of a primary pathogenic variant. In some diseases, certain combinations of variants produce synthetic effects or alter pathways in unexpected ways. While this adds complexity, it also offers therapeutic avenues: modifying downstream pathways may compensate for a primary defect.
Pharmacogenomics connects gene variation to drug metabolism and response. Genes encoding drug-metabolizing enzymes, transporters, and targets can harbor variants that alter dosing or drug choice. For example, CYP2D6 metabolizes many drugs, and copy number variation as well as single nucleotide changes can lead to poor, intermediate, normal, or ultrarapid metabolism. HLA-B*57:01 increases the risk of severe hypersensitivity to abacavir. Recognizing these mechanisms enables proactive prescribing, reducing adverse events and improving efficacy.
Pathogenic mechanisms guide therapeutic strategies. Some diseases are amenable to enzyme replacement therapy if the defect leads to enzyme deficiency. Others involve disrupted protein interactions that can be rescued by small molecules, such as read-through agents that allow translation to bypass premature stop codons. Gene therapy aims to deliver a functional copy of a gene or correct the underlying sequence. Understanding whether a variant is loss-of-function or gain-of-function informs which approach is likely to succeed and which clinical trials are appropriate.
Penetrance modifiers and environmental context can be leveraged clinically. In conditions where penetrance is incomplete, lifestyle changes or prophylactic measures may prevent disease. For example, individuals with certain inherited cancer variants may benefit from enhanced screening or risk-reducing surgeries. In metabolic disorders, dietary modifications can circumvent enzyme deficiencies. The same variant may behave differently in different environments, which is why clinicians should consider exposures, medications, and comorbidities when evaluating a patient’s phenotype and genomic results.
There are common misconceptions that can lead to errors in clinical practice. A variant that is rare is not automatically pathogenic; it could be benign but uncommon. Conversely, a variant reported as benign in databases might be reclassified later if new evidence emerges. Functional assays, while helpful, are not always available or standardized. Computational scores are tools, not authorities. Finally, sequencing depth and coverage limits mean that some variants, particularly in complex regions or at low allele fractions, may be missed or misinterpreted.
Clinicians can apply these concepts directly when reviewing reports. Look for the variant type and location, the predicted effect on protein or splicing, and the evidence cited for classification. Check population frequency, conservation, and whether functional data support the interpretation. Consider the patient’s phenotype and family history to gauge consistency. If a VUS is reported, ask whether additional testing, reanalysis after a period of time, or family segregation studies might clarify its significance. This stepwise approach turns genomic information into actionable clinical knowledge.
Throughout the following chapters, we will explore how these biological principles translate into practical tools and workflows. The type of sequencing assay chosen will determine which variants can be detected and with what confidence. Bioinformatics pipelines will turn raw reads into candidate variants, and evidence sources will help classify them. Clinical scenarios will show how understanding genes, variants, and mechanisms informs decisions at the bedside and in the clinic. With this foundation in how DNA works and how variants behave, the stage is set for the technical chapters that follow.
CHAPTER THREE: Clinical-Grade Sequencing Platforms: Panels, Exomes, and Genomes
Clinical genomics begins with a choice: which parts of the genome to sequence, how deeply to read them, and with what technology. Those decisions shape what can be detected, how confidently it is called, and how quickly results reach clinicians. The field offers three broad approaches—targeted panels, exomes, and genomes—each with advantages and trade-offs. The “best” test is the one that answers the clinical question efficiently, accurately, and ethically, while respecting cost, time, and patient preferences.
A targeted gene panel is the scalpel of sequencing. It interrogates a curated set of genes relevant to a specific phenotype, such as cardiomyopathies, epilepsy, or hereditary cancer. Because the assay focuses on a small region, it can achieve very high coverage—often hundreds of times per base—making it easier to detect variants with confidence, even in samples with low DNA quality or low tumor purity. Turnaround times are typically shorter, and interpretation is more straightforward because the differential diagnosis is constrained to a known set of genes. Panels are also easier to validate, maintain, and update within a laboratory quality system.
However, the precision of a panel comes at the cost of breadth. If the patient’s condition is caused by a gene not included on the panel, or by a gene recently discovered, the test may yield a negative result while the underlying diagnosis remains unknown. Panels can also be limited in the types of variants they detect; some are optimized for single nucleotide variants and small insertions and deletions, but may not robustly capture copy number changes or structural rearrangements. For clinicians, the key question is whether the phenotype fits neatly within the covered genes or whether broader exploration is warranted.
Exome sequencing targets the exome, the set of all protein-coding exons in the genome, which comprises roughly 1–2% of the total DNA. Although that seems small, it includes about 20,000 genes and is where a large proportion of known Mendelian disease-causing variants reside. Exomes offer a balance between breadth and cost. They are particularly useful when the phenotype is broad, when multiple differential diagnoses are plausible, or when a patient has been tested with a panel and remains undiagnosed. With exome data, laboratories can reanalyze over time as new genes and disease associations emerge.
There are limitations. Coverage across the exome is rarely uniform; some regions are difficult to capture or sequence due to guanine–cytosine content or repetitive sequence. Genes with high homology or pseudogenes can be problematic, leading to ambiguous mapping. Not all pathogenic variants lie in coding regions; deep intronic or regulatory variants may be missed by an exome. Moreover, the interpretive burden is heavier. An exome generates thousands of variants that must be filtered against population databases and computational predictors, then correlated with the patient’s phenotype. The clinical context must be clear to avoid “fishing expeditions” that yield uncertain results.
Whole genome sequencing is the most comprehensive approach. It aims to read nearly every base in the genome, capturing coding variants, deep intronic changes, regulatory elements, and structural variants in a single assay. Modern instruments and efficient pipelines have reduced costs and turnaround times, making genomes increasingly viable for clinical use. Genomes provide superior sensitivity for copy number changes, balanced rearrangements, and regions that are problematic for exome capture. For patients with complex or unsolved cases, a genome can be a powerful tool. It is also the foundation for applications beyond single-gene diagnostics, such as pharmacogenomics and risk stratification across multiple systems.
The trade-off is complexity. Genomes produce gigabytes of data and require robust computational infrastructure and sophisticated interpretation frameworks. Filtering to relevant variants without missing rare or novel disease genes is challenging. Noncoding variants are numerous but evidence for pathogenicity is often limited, increasing the chance of uncertain findings. Incidental and secondary findings are more common, and counseling demands greater nuance. A genome may be the right test when the clinical question requires maximum sensitivity and breadth, but it is not a universal first step for every patient.
The choice between germline, somatic, and matched tumor–normal sequencing depends on the clinical scenario. Germline sequencing uses DNA from blood or saliva to identify inherited variants relevant to hereditary disease and pharmacogenomics. Somatic sequencing analyzes tumor tissue to find acquired mutations that drive cancer and inform therapy. A matched tumor–normal pair, where normal tissue is also sequenced, helps distinguish true somatic events from inherited variants. This approach is standard in oncology but can be resource intensive. In some cases, a single tumor-only test is sufficient if the findings are clearly somatic and clinically actionable, but caution is required to avoid misclassifying germline variants as somatic.
Sample quality and type profoundly influence platform selection. Fresh frozen tumor tissue is ideal for somatic analysis because it preserves nucleic acids and often has higher tumor content. Formalin-fixed paraffin-embedded tissue is ubiquitous but can degrade DNA and introduce artifacts, such as C-to-T transitions, that complicate variant calling. Blood or saliva for germline testing is generally robust, but hemolysis, contamination, or low cell yield can reduce DNA quality. For prenatal testing, cell-free fetal DNA from maternal plasma requires specialized handling and assays. Liquid biopsy for circulating tumor DNA demands careful extraction and analytic techniques to detect low-frequency somatic variants against a background of cell-free normal DNA.
Sequencing chemistry and instrument choice also matter. Short-read platforms dominate the clinical landscape because they offer high throughput, accuracy, and well-established workflows. They excel at calling single nucleotide variants and small indels, and they can detect copy number changes and some structural variants with appropriate algorithms. However, short reads struggle with highly repetitive regions, complex structural variants, and phasing—determining whether two variants are on the same chromosome. Long-read sequencing, developed by companies such as PacBio and Oxford Nanopore, generates reads spanning tens to hundreds of kilobases, resolving complex regions and phasing variants across long distances. This can be critical for genes with pseudogenes or structural changes.
Each technology has characteristic error profiles. Short-read sequencing typically has high base-level accuracy, but it can miscall indels in homopolymer runs and has difficulty with regions of extreme GC content. Long-read technologies historically had lower per-base accuracy, but recent improvements have narrowed the gap. Nanopore sequencing detects changes in electrical current as DNA passes through a pore; this method is sensitive to modifications such as methylation, which can be an advantage for epigenomic analyses but requires careful basecalling. PacBio’s HiFi reads offer long reads with high accuracy by circularizing and sequencing the same molecule multiple times. Laboratories must validate each platform and chemistry for the intended clinical use, documenting performance characteristics such as sensitivity, precision, and reproducibility.
Target enrichment strategies are central to panel and exome workflows. Hybridization capture uses synthetic probes to pull down desired regions before sequencing. It offers high specificity and uniform coverage, and it is scalable. Amplicon-based methods amplify regions of interest via PCR; they are fast and require less DNA, but can introduce biases and have higher rates of amplification artifacts. Some labs use a hybrid approach, combining capture for breadth and amplicon for depth in key regions. The choice of enrichment affects coverage uniformity, which in turn impacts variant detection, especially for copy number analysis and low-quality samples.
Somatic testing often requires higher sequencing depth than germline testing because tumor samples may contain only a fraction of DNA with the mutation. For example, a tumor with 30% tumor purity might require depths of 500× or more to confidently detect a variant present in a minority of cells. In contrast, germline testing typically uses depths of 30–100×, which is sufficient for high-confidence calling of heterozygous variants. Depth must be balanced against cost and data management. Excessively high depth can improve sensitivity but may also increase noise if not carefully controlled. Laboratories establish depth thresholds and allelic fraction cutoffs as part of validation.
Accuracy is not only about depth; uniformity is equally important. Even with high average depth, “cold spots” with low coverage can lead to missed variants. Laboratories monitor metrics such as the percentage of targeted bases covered at a minimum threshold, often 20× for germline or 100× for somatic. Uniformity is influenced by capture efficiency, GC content, and repetitive sequences. During validation, labs identify regions with historically poor performance and decide whether to exclude them or apply orthogonal methods. Clinicians should be aware that no assay is perfect, and reports may include disclaimers about regions with lower confidence.
Detecting different variant types requires tailored approaches. Single nucleotide variants are relatively straightforward for most platforms. Insertions and deletions of a few bases are also routinely called, although large indels can be missed by short reads. Copy number variants are inferred by comparing read depth across targeted regions or the entire genome; this works well for large changes but is less sensitive for small duplications or deletions. Structural variants, such as translocations or inversions, can be detected from discordant read pairs and split reads, but long-read sequencing is superior for resolving precise breakpoints and balanced rearrangements. In oncology, fusion genes are important drivers; specialized assays or RNA sequencing may be needed to capture them robustly.
The detection of mosaicism—variants present in only a subset of cells—depends on both technology and analysis. Short-read sequencing can detect mosaicism at allele fractions as low as a few percent if depth is high and noise is low. However, distinguishing true low-level variants from sequencing artifacts is challenging. Long-read technologies may help by providing full-length reads that reduce ambiguous mapping, but they currently have lower throughput. In practice, detecting mosaicism often requires orthogonal methods, such as deep sequencing of specific regions or testing multiple tissues. Clinicians should interpret negative results cautiously in cases where mosaicism is suspected.
RNA sequencing adds a functional dimension to DNA-based testing. While not yet a routine first-line diagnostic for most inherited diseases, RNA-seq is increasingly used to resolve variants of uncertain significance that may affect splicing or expression. It can confirm exon skipping, intron retention, or cryptic splice sites predicted by algorithms. In tumors, RNA sequencing can identify fusion genes and assess expression levels of oncogenes or tumor suppressors. However, RNA is less stable than DNA, requires careful preservation, and is not expressed uniformly across tissues. The choice to add RNA testing depends on the clinical question and the availability of appropriate samples.
Methylation sequencing maps DNA modifications, primarily 5-methylcytosine, which regulate gene expression without altering the underlying sequence. It has emerged as a powerful tool in neurodevelopmental disorders and imprinting conditions, where abnormal methylation patterns can cause disease even without a coding variant. It also aids in tumor classification, as many cancers exhibit characteristic methylation profiles. Bisulfite conversion is traditionally used to detect methylation, but enzymatic approaches and direct nanopore detection are improving accuracy and preserving DNA integrity. Methylation assays require specialized bioinformatics and interpretation frameworks, and they are typically ordered when the phenotype suggests an epigenetic mechanism.
Long-read sequencing is gradually entering clinical practice, particularly for regions and variant types that short reads handle poorly. Genes with high homology, such as CYP2D6, benefit from long reads that can span repetitive regions and differentiate paralogs. Complex structural variants, pathogenic repeat expansions, and phased haplotypes are other areas of strength. The technology remains more expensive and computationally demanding than short-read sequencing, and validation resources are less mature. Nevertheless, some clinical laboratories use long-read assays as a targeted problem-solving tool or as part of research studies that transition into clinical care.
Hybrid approaches combine strengths. A laboratory might use a short-read genome for breadth and supplement it with targeted long-read sequencing of problematic loci. Alternatively, an exome may be paired with a focused panel to achieve high depth in key genes. In oncology, clinicians may order a tumor panel upfront for speed and add liquid biopsy monitoring later. These strategies reflect the reality that no single assay is perfect. The goal is to answer the clinical question efficiently, and that often means layering methods rather than choosing a one-size-fits-all solution.
Clinical-grade sequencing is defined not only by technology but also by rigorous quality systems. In the United States, CLIA mandates standards for personnel, proficiency testing, quality control, and validation. CAP accreditation adds additional requirements and inspections. Internationally, ISO 15189 and national regulations provide similar frameworks. These standards ensure that analytical procedures are documented, reproducible, and monitored. They require laboratories to track metrics like contamination rates, sequencing yield, and batch effects. They also mandate proficiency testing, where blinded samples are run to confirm accuracy compared to peer labs.
Analytical validation is the process of establishing that the test performs as intended. For a sequencing assay, this includes determining sensitivity, specificity, precision, and reproducibility across the full range of sample types and variant classes the lab plans to report. Validation is not a one-time event; it is periodically re-evaluated when platforms change, reagents are updated, or new variant types are added. Documentation is extensive, and laboratories often publish validation data or make it available to customers. Clinicians should ask for validation summaries when adopting a new test or lab, particularly for complex assays like genomes or methylation profiling.
The difference between laboratory-developed tests and FDA-approved kits is an important consideration. Laboratory-developed tests are designed and validated by the lab itself under CLIA, offering flexibility to adapt to new genes and technologies. FDA-approved kits are locked reagent systems with labeling that specifies intended use and performance characteristics. Both can be high quality, but they differ in how they are updated. A lab can quickly add a newly discovered gene to an LDT, while an FDA kit may require a new submission. Clinicians must weigh the need for cutting-edge content against the desire for regulatory pre-market review.
Bioinformatics is the invisible engine behind sequencing. It begins with raw data, processes it through alignment and variant calling, and finishes with annotation and filtering. Each step involves choices that affect the final report. Alignment algorithms determine how reads map to the reference genome, especially in repetitive regions. Variant callers differ in how they handle noise, allele balance, and sample mixtures. Annotation adds gene symbols, predicted effects, population frequencies, and in silico scores. Pipelines must be version controlled and audited, and changes to software can alter results. A robust laboratory documents its pipeline and validates it just like wet-lab procedures.
Quality control is embedded throughout the workflow. Pre-analytic QC assesses sample integrity using metrics like DNA concentration, fragment size, and purity. During sequencing, labs monitor cluster density, read depth, uniformity, and error rates. After alignment, metrics such as mapping rate and coverage distribution are evaluated. Variant calling includes additional filters to remove likely artifacts. Many labs use control samples with known variants as “truth sets” to verify that the pipeline works correctly. When QC fails, the sample may be repeated or reported as insufficient. Clinicians should expect a report to mention whether the sample met quality thresholds and to note any limitations.
Turnaround time is a practical concern that affects clinical utility. Panels can often be completed in one to two weeks, exomes in two to three weeks, and genomes within three to four weeks, depending on lab volume and workflow. In urgent cases, such as neonatal critical illness or rapid tumor profiling, expedited pipelines can shorten this to days. However, faster turnaround often requires trade-offs, such as prioritizing certain genes or using rapid sequencing chemistry that may have different performance characteristics. Clear communication between clinicians and laboratories about time sensitivity helps guide appropriate test selection.
Cost and reimbursement influence test selection and access. Panels are generally the least expensive, followed by exomes and genomes, though prices continue to fall. The total cost includes pre-test counseling, the assay itself, interpretation, reporting, and post-test management. Payer policies vary widely; some cover exomes for pediatric developmental disorders but not for adults with nonspecific symptoms. Oncology panels are more likely to be covered if they guide FDA-approved therapies or trial eligibility. Health systems must articulate value to justify investment, including improved diagnostic yield, reduced diagnostic odysseys, and more appropriate therapy selection.
Ethical considerations are embedded in platform choice. Broader assays increase the likelihood of incidental findings, which require consent, counseling, and infrastructure to return results responsibly. In pediatric patients, parents may consent, but assent should be considered as children mature. For prenatal testing, the scope of analysis raises questions about what conditions to include and how to handle variants of uncertain significance in a fetus. Patients should understand the limits of the assay and the possibility that a result will be uncertain. The ability to decline secondary findings or to opt for restricted analysis should be offered where feasible.
Equity and access are practical realities shaped by platform and logistics. Panels are more feasible in resource-limited settings due to lower cost and computational requirements. Exomes and genomes require greater infrastructure, bioinformatics expertise, and genetic counseling resources. Telehealth and partnerships with centralized laboratories can expand access, but shipping samples across borders raises legal and privacy issues. Global genomic databases are still skewed toward European ancestries, which affects variant interpretation for underrepresented populations. Selecting assays that match local capacity and investing in diverse reference data are necessary steps toward equitable genomics.
Several pitfalls deserve attention. Overreliance on panels may miss emerging genes or noncoding causes. Exomes and genomes can drown clinicians in uncertain findings if the clinical question is not sharply defined. A negative result does not exclude disease; it may reflect assay limits, mosaicism, or novel gene discovery. Variants may be reclassified, and reanalysis of existing data can yield new answers. Incidental findings may reveal stigmatizing information that the patient did not expect. Finally, choosing a test without considering the ability to act on results can lead to frustration; the assay should be matched to available interventions and patient goals.
Selecting a sequencing platform is a clinical decision that balances the question, the patient, and the system. A child with a clear syndrome may benefit from a targeted panel with rapid turnaround and high sensitivity. An individual with a complex, multisystem presentation may need an exome or genome to capture a rare or novel gene. A patient with metastatic cancer needs a fast, somatic assay that focuses on actionable alterations. Pharmacogenomic questions may be addressed with a focused panel or incorporated into a broader genome. Throughout, communication with patients about benefits, limits, and uncertainties is essential.
As platforms evolve, new capabilities will expand what is feasible. Pangenome references will improve mapping in complex regions. Long-read sequencing may become routine for structural variants and repeat expansions. Integrated multi-omic assays—combining DNA, RNA, and methylation—will provide richer portraits of disease mechanisms. The challenge will be to implement these advances with the same rigor that defines clinical-grade sequencing today: validation, quality control, transparent reporting, and ethical care. The next chapters will explore how these data are interpreted and turned into reports that clinicians can trust and act on.
This is a sample preview. The complete book contains 27 sections.