For Dr. Kristen Eckler, Director of Clinical Content at UpToDate®, having AI to support patient care decisions urgently requires a clear, discerning eye.
“It’s not an accident that the Hippocratic Oath is, ‘Above all else, do no harm.’ I think that is by far the best thing that we could possibly do right now: Don’t screw it up.”
AI tool usage is rapidly spreading among clinicians and care teams, and understanding how AI sources its answers, thinks, and is governed is critical for any individual or organization.
When reviewing any clinical information, there are three key questions clinicians need to evaluate:
- What information is being presented, and what trusted source is it grounded in?
- How was that information interpreted, validated, and judged to be clinically reliable?
- How can it be applied appropriately in the context of an actual patient care decision?
With AI involved: Who is responsible for building, validating, monitoring, and improving the technology generating the guidance that could influence patient care?
For the clinical team at UpToDate—and many across the industry—monitoring the quality of AI responses should always fall on trusted, human expertise. It’s critical for AI to source from expert-generated or reviewed content that has been synthesized with the wider body of evidence and industry guidance.
Research findings need to be viewed within clinical context
Evidence needs to be reviewed and synthesized within the wider clinical context. Individual research studies can have bias from authors or have ties to businesses and publishing authors or become outdated due to newer evidence. While providers undoubtedly have the ability to critically evaluate sources, they may not have time at the point of care to do so or to compare it to the wider body of relevant research and clinical knowledge. The AI serving them responses must be trusted to do so in that moment, so the clinician can make the most effective care decision.
“The challenge isn’t just access to information, it’s discernment,” says Dr. Sheila Bond, Director of Clinical Content Strategy. “This type of work in clinical AI tools requires fluency in evidence-based medicine, critical thinking, and openness to create meaningful knowledge in new ways.”
Determining the most effective form of treatment or the right drug dosage requires precision, context, and accountability. Generative AI responses must do more than summarize information; they must be evaluated through a clinical lens:
- Is the tool delivering clinically appropriate guidance, or simply generating a plausible response based on language prediction?
- Is it drawing from the most current and relevant evidence, or relying on outdated versions of the literature?
- Is it reflecting the full body of evidence, including studies that may challenge or qualify a single finding?
Additionally, AI tools are still sharing incorrect medical information and advice despite referencing authoritative sources.
“This technology is still evolving,” says Dr. Bond, “but we have set a new way in terms of looking for the benchmarks and the features and the standards that you can look to so it's reliable and validatable.”