Skip to main content
  1. Healthcare AI Standard of Care/

Gastroenterology AI Standard of Care: Colonoscopy AI, Polyp Detection, and Liability

Table of Contents

AI at the Scope: The New Frontier of GI Liability
#

Gastroenterology has become the second major clinical frontier for AI in medicine, following radiology. With multiple FDA-cleared computer-aided detection (CADe) systems now in routine use during colonoscopies, endoscopists face novel liability questions: What happens when AI misses a polyp that becomes cancer? What if AI misclassifies a polyp, leading to inadequate follow-up? And critically, does AI assistance create a new standard of care that makes non-AI colonoscopy legally indefensible?

This guide examines the emerging standard of care for AI use in gastroenterology: what FDA clearance means for colonoscopy AI, how liability is allocated between endoscopists and AI systems, what the American Gastroenterological Association recommends, and how GI practitioners can protect themselves while leveraging AI’s genuine benefits.

Key Gastroenterology AI Statistics
  • First FDA clearance: GI Genius (Medtronic), 2021, first AI colonoscopy device
  • Multiple platforms: GI Genius, Fujifilm CAD EYE, EndoScreener, CADDIE (Olympus, 2024)
  • 44 RCTs: Meta-analysis confirms CADe effectiveness for polyp detection
  • No US reimbursement: Unlike Japan (Feb 2024), no add-on payment for CADe in US
  • Deskilling risk: Lancet 2025 study raises concerns about AI dependence

FDA-Cleared Colonoscopy AI: The Current Landscape
#

Timeline of Major Approvals
#

YearDeviceManufacturerSignificance
2021GI GeniusMedtronicFirst FDA-cleared AI colonoscopy device
2022CAD EYEFujifilmMajor competitor enters market
2023EndoScreenerWision AIAdditional options available
2024CADDIEOlympus/OdinFirst cloud-based AI endoscopy system

What These Systems Do
#

Computer-Aided Detection (CADe):

  • Real-time video analysis during colonoscopy
  • Highlights suspected polyps with visual markers
  • Alerts endoscopist to areas requiring attention
  • Does NOT make diagnostic decisions

Computer-Aided Diagnosis (CADx):

  • Predicts polyp histology (hyperplastic vs adenomatous)
  • Still emerging, not yet standard of care
  • Could enable “diagnose and leave” strategies
  • Higher liability exposure than CADe
First FDA-cleared colonoscopy AI (GI Genius)
Randomized controlled trials in 2024 meta-analysis
US reimbursement for CADe (no add-on payment)

What FDA Clearance Does NOT Mean
#

As with radiology AI, FDA 510(k) clearance for colonoscopy AI does NOT guarantee:

No Generalizability Proof:

  • Performance may vary across patient populations
  • Training data may not represent diverse populations
  • Real-world performance often differs from trial conditions

No Comparative Superiority:

  • Cleared as “substantially equivalent” to predicate device
  • Does not mean AI outperforms expert endoscopists
  • Individual endoscopist skill still matters

No Outcome Improvement Guarantee:

  • Detects more polyps, but impact on cancer prevention unclear
  • Increased detection may include clinically insignificant lesions
  • Long-term mortality benefit not yet established
The Detection Paradox

CADe systems consistently detect more polyps than unassisted colonoscopy. But more detection doesn’t automatically mean better outcomes. Increased detection may lead to:

  • Over-treatment of clinically insignificant lesions
  • Increased procedure time and patient burden
  • False confidence in comprehensive screening
  • Potential for more interval cancers if non-AI follow-up is extended based on “AI-assisted” false assurance

Clinical Evidence: What the Research Shows
#

The 2024 Meta-Analysis
#

A comprehensive meta-analysis published in the Annals of Internal Medicine (2024) by Soleymanjahi et al. provides the strongest evidence base:

Scope:

  • 44 randomized controlled trials
  • Comparing standard colonoscopy vs CADe-assisted colonoscopy
  • Platforms: GI Genius, Fujifilm CAD EYE, EndoScreener, YOLO-based systems

Key Findings:

  • CADe significantly increases adenoma detection rate (ADR)
  • Improved detection of sessile serrated lesions
  • Little to no adverse events directly attributable to CADe use
  • Evidence supports conditional recommendation for CADe use

Limitations and Gaps
#

Despite positive findings, significant gaps remain:

GapImplication
Short follow-upImpact on colorectal cancer mortality unknown
Selected populationsMay not generalize to all patient demographics
Expert endoscopistsBenefit may be smaller for high-skill practitioners
Cost-effectivenessNot established, especially without reimbursement

AGA Guidelines: The Emerging Standard of Care
#

2024-2025 Draft Clinical Guidelines
#

The American Gastroenterological Association (AGA) released draft clinical guidelines on AI-assistance in colonoscopy:

Conditional Recommendation:

  • AGA conditionally recommends CADe for polyp detection in adults
  • “Conditional” reflects still-developing evidence base
  • Acknowledges potential benefits while recognizing limitations

Key Guideline Points:

  1. CADe is appropriate for routine screening colonoscopy
  2. Not required to meet standard of care (yet)
  3. Training recommended before clinical use
  4. Documentation of AI use should be standard practice
  5. Human judgment remains primary:AI is advisory
What ‘Conditional Recommendation’ Means for Liability

A conditional recommendation means the AGA believes benefits likely outweigh harms, but the evidence isn’t definitive. For liability purposes:

  • Not using CADe is not automatically negligent (yet)
  • Using CADe doesn’t shield you from missed polyp liability
  • Standard of care is evolving, stay current with guidelines
  • Document your reasoning whether you use AI or not

Implementation Requirements
#

If using CADe, the AGA and general medical AI standards require:

Pre-Deployment:

  • Ensure equipment meets manufacturer specifications
  • Train all operators on system use
  • Establish protocols for AI alerts
  • Inform patients about AI assistance

During Procedure:

  • Use AI as intended by manufacturer
  • Maintain standard withdrawal times
  • Don’t over-rely on AI for comprehensive inspection
  • Document AI alerts and responses

Post-Procedure:

  • Report outcomes for quality monitoring
  • Track adenoma detection rates with/without AI
  • Report adverse events or failures to vendor and FDA

The Liability Framework
#

The Core Liability Scenario
#

A detailed liability scenario from Clinical Gastroenterology and Hepatology (2024) illustrates the emerging risk:

A 50-year-old patient presents for screening colonoscopy. The endoscopist identifies polyps, and a CADx tool provides a “hyperplastic” diagnosis prediction. Based on this prediction and the endoscopist’s clinical impression, one polyp is left in place with follow-up recommended in 10 years. Seven years later, the patient is diagnosed with metastatic colon cancer. A lawsuit contends that the interval cancer was caused by misdiagnosis of the polyp.

Key Questions:

  • Who is liable, the endoscopist, the AI vendor, or both?
  • Did relying on CADx meet the standard of care?
  • Should the polyp have been removed regardless of AI prediction?

Liability Allocation
#

PartyPotential LiabilityDefense
EndoscopistPrimary liability for clinical decisionsFollowed standard of care, documented reasoning
AI VendorProduct liability for defective designFDA clearance, labeled limitations
Hospital/PracticeNegligent credentialing, equipment maintenanceProper training, maintenance protocols
Equipment ManufacturerComponent defects affecting AI performanceMet specifications, proper integration

The “Diagnose and Leave” Problem
#

CADx systems that predict polyp histology create unique risks:

Current Standard: Remove all polyps, send for pathology

AI-Enabled Future: Leave predicted hyperplastic polyps, resect only adenomas

Liability Risk: If a “left” polyp becomes cancer, the endoscopist faces:

  • Claim of negligent reliance on unproven technology
  • Allegation of departing from established standard of care
  • Difficulty defending AI prediction error as beyond control
CADx Is NOT Standard of Care (2025)

As of 2025, no professional society endorses “diagnose and leave” strategies based solely on AI CADx predictions. Until guidelines change:

  • Removing suspicious polyps remains the standard
  • Relying on CADx to leave polyps in place is legally risky
  • Document extensively if deviating from removal standard

The Deskilling Crisis
#

The Lancet 2025 Study
#

A concerning study in The Lancet Gastroenterology & Hepatology (2025) examined how continuous AI exposure affects endoscopist behavior:

Study Design:

  • Retrospective, observational study
  • Four endoscopy centers in Poland
  • Assessed performance when AI was NOT in use after regular AI exposure

Key Concern: The study raises the question: Do endoscopists who regularly use AI become dependent on it, performing worse when AI is unavailable?

Implications:

  • “Automation complacency” may affect GI as it has aviation and radiology
  • Skills may atrophy with AI reliance
  • System failures or unavailability create heightened risk
  • Training must maintain non-AI competency

Protecting Against Deskilling
#

For Individual Endoscopists:

  • Maintain regular non-AI colonoscopy practice
  • Conduct periodic self-assessment without AI assistance
  • Document AI-independent competency

For Practices and Hospitals:

  • Establish rotation schedules with/without AI
  • Track ADR both with and without AI assistance
  • Plan for AI system downtime or failure

Reimbursement: The US Gap
#

Current Status (2025)
#

JurisdictionCADe Reimbursement
United StatesNo specific add-on payment
JapanAdd-on payment approved (Feb 2024)
EuropeVaries by country, limited reimbursement

Why This Matters for Liability
#

The lack of US reimbursement creates a problematic dynamic:

Cost Pressure:

  • Practices bear full cost of AI systems
  • No additional payment for AI-assisted procedures
  • Economic incentive to use AI without full training/integration

Adoption Variability:

  • Well-resourced practices adopt AI; others cannot
  • Creates potential “two-tier” standard of care
  • Rural and safety-net hospitals may lag

Documentation Gap:

  • Without reimbursement coding, AI use may be poorly documented
  • Difficult to track outcomes and quality
  • Liability exposure from incomplete records

Practical Guidance for Endoscopists
#

Using CADe Appropriately
#

DO:

  • Treat CADe as an adjunct, not a replacement for careful inspection
  • Maintain standard withdrawal times (6+ minutes)
  • Respond thoughtfully to AI alerts, investigate, don’t just dismiss
  • Document AI alerts and your response
  • Remove suspicious polyps regardless of AI classification

DON’T:

  • Assume AI has detected all polyps
  • Reduce your own vigilance because “AI is watching”
  • Use CADx predictions to justify leaving polyps in place
  • Skip areas because AI didn’t flag them
  • Fail to document when you override AI

Documentation Best Practices
#

Every AI-assisted colonoscopy report should include:

  1. AI System Used: Name and version
  2. AI Alerts: Number and locations
  3. Your Response: Actions taken for each alert
  4. Overrides: When you disagreed with AI and why
  5. Findings: Standard polyp documentation
  6. Limitations: Any technical issues affecting AI performance

When AI Isn’t Available
#

Maintain competency for non-AI colonoscopy:

  • System downtime happens
  • Not all facilities have AI
  • Your skills shouldn’t atrophy
  • Document that AI was not used if unavailable

Future Directions
#

CADx Evolution
#

Computer-aided diagnosis is advancing rapidly:

Potential Benefits:

  • Resect-and-discard for diminutive polyps (cost savings)
  • Optical diagnosis reducing pathology burden
  • Real-time risk stratification

Barriers to Adoption:

  • Liability exposure remains high
  • No professional society endorsement yet
  • Performance varies across polyp types
  • Requires very high accuracy (>90% sensitivity/specificity)

Capsule Endoscopy AI
#

AI is expanding beyond colonoscopy:

  • Automated reading of capsule endoscopy images
  • Detection of small bowel pathology
  • Reduces reading time from hours to minutes
  • Liability framework still developing

Integration with EHR
#

Future developments may include:

  • Automated procedure documentation
  • Risk-adjusted follow-up recommendations
  • Population health management integration
  • Quality metric tracking

Frequently Asked Questions
#

Is AI-assisted colonoscopy now the standard of care?

Not yet. The AGA’s conditional recommendation supports CADe use but does not mandate it. Performing colonoscopy without AI assistance is not currently negligent per se. However, the standard is evolving, endoscopists should stay current with guidelines and be prepared to justify their approach. Document your reasoning whether you use AI or not.

Who is liable if AI misses a polyp that becomes cancer?

The endoscopist remains primarily liable for clinical decisions, including missed findings. AI is currently advisory, it assists but doesn’t replace human judgment. However, if AI contributed to a miss (e.g., the endoscopist relied on AI’s failure to alert), product liability claims against the AI vendor may also be viable. Multi-party liability is increasingly likely in AI-assisted procedure cases.

Can I use CADx predictions to leave polyps in place?

This is legally risky in 2025. No professional society currently endorses “diagnose and leave” strategies based solely on AI CADx predictions. The standard of care remains removal and pathological examination of suspicious polyps. If you leave a polyp based on AI prediction and it becomes cancer, you face significant liability exposure for departing from established practice.

What should I document about AI use during colonoscopy?

Document: (1) the AI system used (name, version), (2) all AI alerts and their locations, (3) your response to each alert, (4) any instances where you overrode or disagreed with AI, (5) any technical issues affecting AI performance, and (6) if AI was unavailable, note that as well. Comprehensive documentation protects you whether AI helps or fails.

Does using AI protect me from malpractice claims for missed polyps?

No. AI assistance does not immunize you from liability. You remain responsible for thorough inspection and appropriate clinical judgment. AI may actually increase your liability if you demonstrably over-relied on it, reduced your vigilance, or failed to respond appropriately to AI alerts. Use AI as an adjunct that enhances your practice, not a crutch that replaces your skills.

What is the 'deskilling' concern with colonoscopy AI?

Research suggests that endoscopists who regularly use AI may become dependent on it, potentially performing worse when AI is unavailable. This “automation complacency” parallels concerns in radiology and aviation. Protect yourself by maintaining non-AI practice, conducting periodic self-assessment without AI, and ensuring your skills don’t atrophy through AI dependence.

Related Resources#

Healthcare AI
#

AI Liability Framework
#

Professional Guidance
#


Questions About AI in Your GI Practice?

As colonoscopy AI becomes ubiquitous, endoscopists face new liability questions. From CADe implementation to CADx risks, understanding the evolving standard of care is essential for protecting both your patients and your practice. Whether you're evaluating AI adoption, facing a missed polyp claim, or developing AI governance policies, specialized guidance can help navigate this rapidly changing landscape.

Contact Us

Related

AI Medical Device Adverse Events & Liability

Executive Summary # AI medical devices are proliferating faster than regulatory infrastructure can track their failures. With over 1,200 FDA-authorized AI devices and a 14% increase in AI-related malpractice claims since 2022, understanding the liability landscape has never been more critical.

Cardiology AI Standard of Care: ECG Analysis, Risk Prediction, and Liability

AI Transforms Cardiovascular Care # Cardiology has become a major frontier for artificial intelligence in medicine. From AI algorithms that detect arrhythmias on ECGs to predictive models forecasting heart failure readmission, these systems are reshaping how cardiovascular disease is diagnosed, monitored, and managed. But with transformation comes liability questions: When an AI misses atrial fibrillation and the patient suffers a stroke, who is responsible?

Dermatology AI Standard of Care: Skin Cancer Detection, Melanoma Screening, and Liability

AI Enters the Skin Cancer Screening Revolution # Skin cancer is the most common cancer in the United States, yet approximately 25% of cases are misdiagnosed. In January 2024, the FDA authorized DermaSensor, the first AI-enabled dermatologic device cleared for use by non-specialists, opening a new frontier for skin cancer detection in primary care settings.

Emergency Medicine AI Standard of Care: Sepsis Prediction, ED Triage, and Clinical Decision Support Liability

AI in the Emergency Department: Time-Critical Decisions # Emergency medicine is where AI meets life-or-death decisions in real time. From sepsis prediction algorithms to triage decision support, AI promises to help emergency physicians identify critically ill patients faster and allocate resources more effectively. In April 2024, the FDA authorized the first AI diagnostic tool for sepsis, a condition that kills over 350,000 Americans annually.

Endocrinology AI Standard of Care: Diabetes Management, Insulin Dosing, and Metabolic Monitoring

AI Transforms Diabetes and Metabolic Care # Endocrinology, particularly diabetes management, has become one of the most AI-intensive medical specialties. From continuous glucose monitors that predict hypoglycemia 20 minutes in advance to closed-loop “artificial pancreas” systems that automatically adjust insulin delivery, AI is fundamentally reshaping how metabolic diseases are managed.

Hematology AI Standard of Care: Blood Cancer Diagnostics, Transfusion Management, and Coagulation Analysis

AI Transforms Blood Disorder Diagnosis and Treatment # Hematology, the study of blood and blood-forming organs, sits at a critical intersection of AI advancement. From digital microscopy systems that classify leukemia subtypes in seconds to algorithms predicting transfusion needs and optimizing anticoagulation therapy, AI is fundamentally changing how blood disorders are diagnosed and managed.