Why AI Vendor Due Diligence Matters#
Selecting an AI vendor isn’t like choosing traditional software. When you deploy a third-party AI system, you’re not just buying a product, you’re inheriting its biases, its security vulnerabilities, its training data decisions, and potentially its legal liabilities.
The Mobley v. Workday ruling demonstrated that companies can face discrimination claims based on AI decisions made by their vendors’ systems. The AI standard of care is evolving to hold organizations accountable for the AI systems they deploy, regardless of who built them.
The stakes are high:
- Regulatory penalties under emerging AI laws
- Discrimination and civil rights liability
- Data breach and privacy violations
- Reputational damage from AI failures
- Contractual disputes when systems underperform
This guide provides a comprehensive framework for evaluating AI vendors before you sign, and monitoring them after deployment.
Technical Evaluation Criteria#
Model Architecture and Performance#
Before signing any contract, you need to understand what you’re actually getting. Many vendors are vague about their technical capabilities, and vagueness is a red flag.
Core Technical Questions#
- What model architecture powers the system? (proprietary, fine-tuned open-source, API wrapper)
- What training data was used? Can the vendor provide documentation?
- What are the model’s known limitations and failure modes?
- What accuracy metrics does the vendor report, and how were they measured?
- Has the model been independently validated or audited?
- What is the model’s latency and throughput under production loads?
- How does the model handle edge cases and out-of-distribution inputs?
- What version control exists for model updates?
Red Flags in Technical Claims#
“Our AI is 99% accurate”, Accuracy without context is meaningless. Accurate at what task? Measured on what data? What’s the false positive vs. false negative rate? Demand specificity.
“Proprietary technology”, Some legitimate AI involves trade secrets, but complete opacity should concern you. At minimum, vendors should explain the general approach and validation methodology.
“We use cutting-edge AI”, Buzzwords without substance suggest marketing over engineering. Ask for technical documentation, not brochures.
No documentation of failure modes, Every AI system fails. Vendors who can’t articulate how their system fails either don’t understand it or are hiding problems.
Integration and Infrastructure#
Technical Integration Checklist#
- What APIs does the system expose? RESTful? GraphQL? Batch processing?
- What authentication and authorization mechanisms are supported?
- How does the system handle rate limiting and throttling?
- What data formats are required for input/output?
- Is on-premises deployment available, or cloud-only?
- What uptime SLA does the vendor guarantee?
- What disaster recovery and business continuity plans exist?
- How are model updates deployed? Can you control update timing?
- What rollback capabilities exist if an update causes problems?
Explainability and Transparency#
For high-stakes decisions, hiring, lending, healthcare, criminal justice, you may be legally required to explain AI decisions to affected individuals. Even where not legally mandated, explainability is increasingly expected.
Explainability Questions#
- Does the system provide decision explanations?
- Are explanations human-readable or technical-only?
- Can explanations be provided at both individual and aggregate levels?
- Has the explanation methodology been validated?
- Can the system identify which features most influenced specific decisions?
Legal and Contractual Considerations#
Liability Allocation#
The most important contractual question: When this AI system causes harm, who pays?
Indemnification: Does the vendor indemnify you for:
- Intellectual property infringement (training data issues)
- Discrimination claims arising from AI decisions
- Data breaches caused by vendor systems
- Regulatory fines related to AI compliance
- Third-party claims from AI outputs
Limitation of Liability: Watch for caps that are too low relative to potential damages. A $100K liability cap is meaningless if the AI causes a $10M discrimination class action.
Insurance Requirements: What insurance does the vendor carry? Minimum recommended:
- Technology E&O: $5M+
- Cyber liability: $5M+
- General commercial liability: $2M+
Intellectual Property Rights#
Training Data Provenance#
Post-NYT v. OpenAI, training data liability is a legitimate concern. If your vendor trained on unlicensed copyrighted material, you could face secondary liability.
Key Questions:
- Can the vendor document the provenance of training data?
- Does the vendor have licenses for copyrighted training materials?
- What indemnification does the vendor provide for IP claims?
- Does the vendor have a process for responding to takedown requests?
Output Ownership#
- Who owns outputs generated by the AI system?
- Can outputs be used for any purpose, or are there restrictions?
- Does the vendor claim any rights to outputs?
- Can outputs be used to train competing systems?
Regulatory Compliance#
Different industries face different AI regulations. Ensure your vendor understands your compliance obligations.
General AI Regulations:
- EU AI Act compliance (if serving EU customers)
- Colorado AI Act requirements (for high-risk decisions)
- NYC Local Law 144 (automated employment decision tools)
- Illinois BIPA (biometric data)
Industry-Specific:
- HIPAA (healthcare AI)
- ECOA/Fair Housing Act (lending/housing AI)
- FCRA (consumer reporting AI)
- EEOC guidance (employment AI)
- FDA regulations (medical device AI)
- SEC/FINRA rules (financial AI)
Contract Terms to Negotiate#
Audit Rights: You need the ability to audit vendor systems, either directly or through third parties. This should include:
- Access to model performance metrics
- Bias audit data
- Security assessment results
- Training data documentation
Change Notification: The vendor should notify you before:
- Material changes to the model
- Changes to training data
- Updates that could affect accuracy or bias
- Changes to data handling practices
Termination Rights: You should be able to exit if:
- The vendor fails audits
- Regulatory requirements change
- Performance degrades below agreed thresholds
- The vendor is acquired by a competitor
Data Rights on Termination: What happens to your data when the contract ends?
- Return of all data in usable format
- Certification of data deletion
- Transition assistance period
Bias and Fairness Assessment#
Pre-Deployment Bias Evaluation#
AI bias isn’t just an ethical concern, it’s a legal liability. Discriminatory AI systems have led to class actions, regulatory enforcement, and reputational crises.
Documentation Review:
- Has the vendor conducted bias testing? Request documentation.
- What demographic groups were tested?
- What fairness metrics were used?
- What disparities were found?
- What mitigation measures were implemented?
- Has the system been audited by an independent third party?
Protected Classes to Evaluate:
- Race and ethnicity
- Gender and gender identity
- Age (40+)
- Disability status
- Religion
- National origin
- Veteran status
- Pregnancy and parental status
- Sexual orientation (where protected)
Fairness Metrics to Request#
Different fairness metrics capture different types of disparities. No single metric is sufficient.
Demographic Parity: Equal selection rates across groups Equalized Odds: Equal true positive and false positive rates across groups Predictive Parity: Equal precision across groups Individual Fairness: Similar individuals receive similar predictions Calibration: Predicted probabilities are accurate across groups
Ask vendors which metrics they use, why they chose those metrics, and what thresholds they consider acceptable.
Ongoing Bias Monitoring#
Bias can emerge after deployment as:
- User populations change
- Model drift occurs
- Feedback loops amplify small initial biases
- New edge cases appear
Require vendors to provide:
- Ongoing demographic performance reports
- Alerting when metrics exceed thresholds
- Quarterly or annual bias re-audits
- Clear escalation procedures when bias is detected
Security and Privacy Review#
Data Security Assessment#
AI systems often process sensitive data at scale. A security breach could expose millions of records.
Certifications and Audits:
- SOC 2 Type II report (request copy)
- ISO 27001 certification
- Penetration testing results (within last 12 months)
- Vulnerability assessment reports
- HIPAA certification (if handling PHI)
- PCI DSS compliance (if handling payment data)
Data Handling:
- Where is data stored? (geography, cloud provider)
- Is data encrypted at rest? (what standard?)
- Is data encrypted in transit? (TLS version)
- Who has access to customer data?
- What logging and monitoring exists?
- How long is data retained?
- What is the data deletion process?
Incident Response:
- What is the vendor’s breach notification timeframe?
- What incident response procedures exist?
- What was the vendor’s breach history? (check news, disclosures)
- Does the vendor carry cyber insurance?
Privacy Compliance#
Data Processing Questions#
- What data does the system collect beyond what’s submitted?
- Is data used to train or improve the vendor’s models?
- Can you opt out of data use for model training?
- Does the system make automated decisions about individuals?
- What data subject rights does the vendor support (access, deletion, correction)?
- How does the vendor handle cross-border data transfers?
Privacy Impact Assessment#
Before deploying any AI system that processes personal data:
- Document what personal data is processed
- Identify the legal basis for processing
- Assess privacy risks to data subjects
- Implement appropriate safeguards
- Document the assessment and decisions
AI-Specific Security Concerns#
Traditional security assessments may miss AI-specific risks:
Prompt Injection: Can malicious inputs manipulate the AI to bypass controls? Model Extraction: Can competitors reverse-engineer the model through queries? Data Poisoning: Could training data be manipulated to compromise the model? Adversarial Inputs: Is the model robust to inputs designed to cause misclassification?
Ask vendors what testing they’ve done for these AI-specific attack vectors.
Ongoing Monitoring Requirements#
Performance Monitoring Framework#
Due diligence doesn’t end at contract signing. You need ongoing monitoring to ensure the system continues to meet standards.
Performance Metrics (track continuously):
- Accuracy/precision/recall vs. baseline
- Latency and response times
- Error rates and types
- User satisfaction/feedback
- Business outcome metrics
Periodic Reviews (quarterly minimum):
- Bias audit results
- Security posture updates
- Compliance status
- Incident reports
- Model update changelog
- Training data changes
Annual Requirements:
- Full third-party audit (for high-risk systems)
- Contract review and renegotiation
- Regulatory compliance reassessment
- Vendor financial health check
Model Drift Detection#
AI models degrade over time as the world changes. Monitor for:
Data Drift: Input data distributions changing from training data Concept Drift: The relationship between inputs and outcomes changing Performance Decay: Gradual accuracy decline Bias Emergence: New disparities appearing in model outputs
Incident Tracking#
Maintain a log of all AI-related incidents:
- What happened
- When it was detected
- How it was resolved
- Root cause analysis
- Preventive measures implemented
This documentation is essential for demonstrating due diligence if the system is later challenged.
Sample RFP Questions#
Use these questions when evaluating AI vendors. Adapt based on your specific use case and risk level.
Technical Capabilities#
- Describe the model architecture and explain why it’s appropriate for this use case.
- What training data was used? Provide documentation of data sources and licensing.
- What validation testing was performed? Provide methodology and results.
- What are the known limitations and failure modes of this system?
- How does the system handle inputs it wasn’t trained on?
- What explainability features does the system provide?
- Describe your model versioning and update process.
Bias and Fairness#
- Has this system been tested for bias? Provide audit documentation.
- What demographic groups were included in bias testing?
- What fairness metrics do you report, and what are current values?
- What bias mitigation measures have been implemented?
- How do you monitor for bias emergence post-deployment?
- Has the system been audited by an independent third party?
Security and Privacy#
- Provide your most recent SOC 2 Type II report.
- Describe your data encryption approach (at rest and in transit).
- What is your incident response procedure and notification timeframe?
- How is customer data used? Is it used for model training?
- What data subject rights do you support?
- Describe your approach to AI-specific security threats (prompt injection, adversarial inputs).
Legal and Compliance#
- What indemnification do you provide for IP infringement claims?
- What indemnification do you provide for discrimination claims?
- What insurance coverage do you carry?
- What regulatory compliance certifications do you hold?
- Describe your audit rights and what access you provide to customers.
- What is your liability cap, and how was it determined?
Support and Operations#
- What is your guaranteed uptime SLA?
- How will you notify us of model updates?
- What support tiers do you offer?
- What training and documentation do you provide?
- What is your process for handling customer-reported issues?
Vendor Evaluation Scorecard#
Use this scorecard to compare vendors systematically.
| Category | Weight | Score (1-5) | Weighted Score |
|---|---|---|---|
| Technical capability | 25% | ||
| Bias/fairness documentation | 20% | ||
| Security posture | 20% | ||
| Legal terms and indemnification | 15% | ||
| Regulatory compliance | 10% | ||
| Support and operations | 10% | ||
| Total | 100% |
Scoring Guide:
- 5: Exceeds requirements, best-in-class
- 4: Meets all requirements, strong offering
- 3: Meets minimum requirements
- 2: Some gaps, requires negotiation
- 1: Significant gaps, high risk
Minimum threshold: Don’t proceed with vendors scoring below 3.0 overall or below 2.0 in any single category.
Frequently Asked Questions#
What if vendors won’t provide bias audit documentation?#
This is a significant red flag. For high-risk AI systems (employment, lending, healthcare), lack of bias documentation is increasingly unacceptable, and may violate emerging regulations like NYC Local Law 144 and the Colorado AI Act.
Options:
- Require independent third-party audit as condition of contract
- Conduct your own bias testing before deployment
- Choose a different vendor with better transparency
- If proceeding, document your decision and rationale carefully
How much liability cap is enough?#
It depends on your exposure. Consider:
- Maximum potential class size affected by AI decisions
- Per-person damages in discrimination/privacy claims
- Regulatory penalty maximums
- Your organization’s risk tolerance
Rule of thumb: The liability cap should be at least 2x your estimated worst-case exposure. For enterprise deployments affecting thousands of individuals, $10M+ caps are increasingly common.
Can we require vendors to carry AI-specific insurance?#
Yes, and you should. AI-specific policies now exist that cover:
- Algorithmic discrimination claims
- AI errors and omissions
- Training data IP infringement
- Regulatory defense costs
Require proof of coverage and name your organization as additional insured.
What if the vendor uses subprocessors for AI services?#
Many vendors wrap third-party AI APIs (OpenAI, Anthropic, etc.) rather than running their own models. This isn’t necessarily bad, but you need to know:
- Who are the subprocessors?
- What contractual protections exist with them?
- How is your data handled by subprocessors?
- Do subprocessors meet your security requirements?
Flow-down your requirements to subprocessors through contract terms.
How do we handle vendors headquartered outside our jurisdiction?#
Key considerations:
- Choice of law and venue for disputes
- Cross-border data transfer mechanisms
- Local regulatory compliance
- Practical enforceability of contract terms
- Insurance coverage across jurisdictions
For high-risk systems, prefer vendors with legal presence in your jurisdiction.
Conclusion#
AI vendor due diligence is not a checkbox exercise, it’s an ongoing risk management discipline. The AI systems you deploy become extensions of your organization, and their failures become your failures.
This checklist provides a starting framework, but adapt it to your specific industry, use case, and risk tolerance. Document your evaluation process thoroughly. When AI liability questions arise, and they will, your due diligence documentation is your first line of defense.
Key Takeaways:
- Technical evaluation isn’t optional. Understand what you’re buying.
- Liability allocation matters. Negotiate indemnification before you need it.
- Bias assessment is a legal requirement for many high-stakes applications.
- Security due diligence must include AI-specific risks.
- Monitoring continues after deployment. Due diligence is ongoing.
The AI standard of care demands that organizations evaluate AI systems before deployment and monitor them continuously. This checklist helps you meet that standard.