![]()
Artificial intelligence (AI), including machine learning (ML) and generative AI, has the potential to transform laboratory medicine by enhancing diagnostic accuracy, improving efficiency, and enabling more precise, data-driven clinical decision-making. As AI tools are increasingly embedded in laboratory workflows, clinical decision support systems, risk prediction models, and population health applications, their influence on patient care continues to expand.
The safety and effectiveness of these tools, however, depend on how they are developed, implemented, and overseen within clinical practice. In laboratory medicine, AI systems rely on complex diagnostic data and are increasingly positioned to influence clinical interpretation and decision-making, raising important considerations related to data quality, professional oversight, and governance.
This document focuses on the use of AI tools for prediction or classification that rely on laboratory data. It describes policy considerations that affect the performance, reliability, and equity of AI systems in laboratory medicine. Issues such as the use of generative AI applications, individual data privacy, and policies regarding patient disclosure when AI models contribute to clinical decision-making are outside the scope of this discussion.
Key Recommendations:
|
Clinical laboratories operate under rigorous quality management systems that encompass method validation and continuous performance monitoring. These frameworks ensure that the billions of test results reported each year in the United States are accurate, dependable, and clinically meaningful. As AI systems are increasingly applied to analyzing laboratory data, embedded in laboratory workflows, clinical decision support tools, risk prediction models, and population health applications—the need for similarly robust oversight becomes even more critical.
For AI to be safely and effectively implemented in laboratory medicine, several key issues must be addressed. These include establishing appropriate oversight mechanisms, improving data harmonization, mitigating bias, and ensuring robust validation and monitoring of AI systems.
AI and ML technologies vary widely in their potential to affect patient outcomes. Tools that directly influence diagnosis, treatment decisions, or test interpretation carry higher risks than those automating administrative or operational functions. Risk-based oversight approaches that scale to the severity and likelihood of potential harm should therefore form the foundation of regulatory policy for these tools.
In clinical laboratories, the Clinical Laboratory Improvement Amendments (CLIA) already require laboratories to validate and monitor test systems to ensure their accuracy, precision, reportable ranges, and reference intervals before and during clinical use. Although current CLIA regulations do not explicitly address AI systems, modernizing CLIA can provide the needed oversight for laboratory-developed AI services without creating a duplicative regulatory framework.
AI systems that use laboratory data often assume that inputs are comparable across different sites and methods. In practice, this is not always the case. Different assays, platforms, and calibration approaches can produce different numeric results for the same analyte, even when each method is individually accurate. Without harmonization, these differences can:
Global and national guidance on AI in health stress that limited, low-quality, or inconsistent data can produce biased inferences and unsafe results in clinical applications. In laboratory medicine, ongoing harmonization initiatives—for example, the CDC’s standardization programs for cholesterol and hemoglobin A1c—show that aligning methods and reference systems improves comparability and underpins evidence-based interpretation.
Bias is one of the most widely recognized risks of AI/ML applications in healthcare. In laboratory medicine, two categories of bias-related risk are especially important:
A variety of strategies should be considered to mitigate these issues, such as
Implementation of such measures could ensure that laboratory-based AI promotes equity, fairness, and patient trust.
Under CLIA, clinical laboratories must validate new test systems before use and conduct ongoing quality monitoring—including daily quality control (QC), proficiency testing, and trend analysis—to ensure that performance remains within acceptable limits over time. AI systems that rely on or influence laboratory data introduce similar risks. Notably, models that continuously update or “learn” from new inputs can drift in accuracy, calibration, or bias over time. Additionally, even if AI models are kept constant, changes in analytical instrument performance or patient characteristics over time may degrade their ability to produce accurate diagnostic results.
To effectively verify and monitor AI performance, laboratory professionals require sufficient access to the relevant data and model information from developers. Independent evaluation of a system’s performance without transparency is challenging and can undermine ongoing quality assurance efforts. To address these considerations, several policy needs should be considered:
Oversight should correspond to the level of risk and potential impact an AI tool may have on patient outcomes. High-risk diagnostic applications—particularly those that influence clinical decisions or rely heavily on laboratory data should undergo robust validation, maintain transparency, and incorporate continuous monitoring. Lower-risk tools may warrant a more streamlined oversight process consistent with their reduced potential for patient harm.
Congress and federal agencies need to adopt and implement policies that ensure AI/ML clinical systems that are safe, accurate, and efficient.