Incident Database Integration
Incident Database Integration
Section titled “Incident Database Integration”Risk estimates should be grounded in real-world data when available. This page explains how to use public AI incident databases to calibrate your delegation risk models.
Why Use Incident Data?
Section titled “Why Use Incident Data?”Expert judgment is useful but can be:
- Biased by recent events or personal experience
- Overconfident about well-understood risks
- Underconfident about novel risks
Incident databases provide:
- Base rates for calibration
- Failure mode taxonomies for comprehensive analysis
- Trend data for forecasting
Key Databases
Section titled “Key Databases”AI Incident Database (AIID)
Section titled “AI Incident Database (AIID)”URL: incidentdatabase.ai
Coverage: 1,000+ documented AI incidents across industries
Useful for:
- Understanding failure mode variety
- Finding reference classes for new systems
- Tracking incident trends over time
Limitations:
- Selection bias (only reported/public incidents)
- Variable documentation quality
- No systematic severity scoring
AIAAIC Repository
Section titled “AIAAIC Repository”Coverage: AI and algorithmic incidents with ethical implications
Useful for:
- Bias and fairness incidents
- Regulatory and compliance cases
- Stakeholder harm patterns
CVE Database (for AI security)
Section titled “CVE Database (for AI security)”URL: cve.org
Coverage: Security vulnerabilities including ML/AI systems
Useful for:
- Adversarial attack vectors
- Model security vulnerabilities
- Prompt injection cases
Extracting Base Rates
Section titled “Extracting Base Rates”Method 1: Direct Frequency Estimation
Section titled “Method 1: Direct Frequency Estimation”Count incidents in a reference class:
// Example: LLM chatbot incidents// AIID shows ~50 chatbot-related incidents over 2 years// Estimated ~10,000 production LLM chatbots deployed
incidentsObserved = 50systemsDeployed = 10000yearsObserved = 2
annualIncidentRate = incidentsObserved / (systemsDeployed * yearsObserved)// = 50 / 20000 = 0.0025 = 0.25% per system-year
// With uncertainty (Poisson-based)incidentRate = beta(50 + 1, 20000 - 50 + 1)// Mean: 0.25%, 90% CI: [0.19%, 0.33%]Caveats:
- Underreporting (multiply by 5-20× for total incidents)
- Survivorship bias (failed systems not tracked)
- Deployment estimates are rough
Method 2: Comparative Analysis
Section titled “Method 2: Comparative Analysis”Compare your system to incidents in similar systems:
// Reference: Image classification in healthcare// AIID shows 12 diagnostic AI incidents over 3 years// Approximately 500 deployed systems
referenceRate = 12 / (500 * 3) // ~0.8% per year
// Adjustment for our systemourSystem_adjustment = mixture( [0.5, 1.0, 2.0], // Could be better, same, or worse [0.3, 0.4, 0.3] // Prior on which applies)
ourIncidentRate = referenceRate * ourSystem_adjustmentMethod 3: Trend Extrapolation
Section titled “Method 3: Trend Extrapolation”Use historical trends to forecast:
// LLM incidents by year (example data)// 2022: 10 incidents// 2023: 35 incidents// 2024: 80 incidents (projected)
growthRate = log(80 / 10) / 2 // ~1.0 (100% per year)
// Forecast for 2025projected2025 = 80 * exp(growthRate) // ~220 incidents
// With uncertaintyincidentForecast = lognormal(log(220), 0.5)Severity Estimation from Incidents
Section titled “Severity Estimation from Incidents”Damage Distribution from Reported Incidents
Section titled “Damage Distribution from Reported Incidents”// Categorize AIID incidents by reported/estimated damage// Minor: <$10K (60% of incidents)// Moderate: $10K-$1M (30% of incidents)// Major: >$1M (10% of incidents)
damageSeverity = mixture( lognormal(log(5000), 0.5), // Minor lognormal(log(100000), 0.8), // Moderate lognormal(log(5000000), 1.0), // Major weights: [0.6, 0.3, 0.1])
// Expected damage per incident// ≈ 0.6 × $5K + 0.3 × $100K + 0.1 × $5M = $533KFailure Mode Frequencies
Section titled “Failure Mode Frequencies”From AIID taxonomy:
| Failure Mode | Frequency | Typical Damage |
|---|---|---|
| Incorrect output | 35% | Low-Medium |
| Bias/discrimination | 20% | Medium-High |
| Privacy breach | 15% | High |
| Security exploit | 10% | High |
| Availability failure | 10% | Medium |
| Physical harm | 5% | Very High |
| Other | 5% | Varies |
// Map to harm mode probabilitiesincorrectOutput_share = 0.35bias_share = 0.20privacy_share = 0.15security_share = 0.10availability_share = 0.10physicalHarm_share = 0.05
// Given an incident occurs, damage distributiondamageGivenIncident = mixture( lognormal(log(5000), 0.8), // Incorrect output lognormal(log(50000), 1.0), // Bias lognormal(log(200000), 1.2), // Privacy lognormal(log(100000), 1.0), // Security lognormal(log(30000), 0.7), // Availability lognormal(log(1000000), 1.5), // Physical harm weights: [0.35, 0.20, 0.15, 0.10, 0.10, 0.05])Updating Your Model with Incident Data
Section titled “Updating Your Model with Incident Data”Prior → Posterior with Incident Counts
Section titled “Prior → Posterior with Incident Counts”// Your prior belief about monthly failure ratepriorFailureRate = beta(5, 95) // ~5%
// Observed in similar systems: 3 failures in 100 system-monthsobservedFailures = 3observedSystemMonths = 100
// UpdateposteriorFailureRate = beta(5 + 3, 95 + (100 - 3))// = beta(8, 192)// Mean: 4.0%, 90% CI: [2%, 7%]Adjusting for Reporting Bias
Section titled “Adjusting for Reporting Bias”Incident databases capture only a fraction of actual incidents:
// Reporting rate estimatereportingRate = beta(10, 90) // ~10% of incidents get reported
// If database shows 50 incidentsobservedIncidents = 50estimatedActualIncidents = observedIncidents / mean(reportingRate)// ≈ 500 actual incidents
// With uncertaintyactualIncidents = observedIncidents / reportingRate// Heavy right tail due to uncertain reporting rateDomain-Specific Adjustments
Section titled “Domain-Specific Adjustments”Different domains have different reporting cultures:
| Domain | Estimated Reporting Rate |
|---|---|
| Healthcare | 20-40% (regulated) |
| Finance | 30-50% (regulated) |
| Social media | 5-15% (less regulated) |
| Internal tools | 1-5% (rarely public) |
Building Your Incident Tracking
Section titled “Building Your Incident Tracking”What to Track
Section titled “What to Track”For each incident:
- Date/Time: When did it occur?
- System: Which component(s) failed?
- Failure Mode: What type of failure?
- Detection: How/when was it detected?
- Damage: Estimated cost/impact
- Root Cause: Why did it happen?
- Resolution: How was it fixed?
Tracking Template
Section titled “Tracking Template”## Incident Report #[ID]
**Date**: YYYY-MM-DD**Severity**: Minor / Moderate / Major / Critical**Status**: Open / Investigating / Resolved / Post-mortem
### Summary[One-line description]
### Affected System(s)- [Component 1]- [Component 2]
### Timeline- HH:MM - Issue detected- HH:MM - Investigation started- HH:MM - Root cause identified- HH:MM - Resolution deployed
### Impact- Users affected: [count]- Estimated damage: $[amount]- Damage type: [infrastructure / data / reputation / regulatory]
### Root Cause[Description]
### Failure Mode Classification- [ ] Incorrect output- [ ] Bias/discrimination- [ ] Privacy breach- [ ] Security exploit- [ ] Availability failure- [ ] Physical harm- [ ] Other: [specify]
### Contributing Factors1. [Factor 1]2. [Factor 2]
### Resolution[What was done to fix it]
### Prevention[What will prevent recurrence]Using Your Data
Section titled “Using Your Data”Quarterly analysis:
- Calculate incident rate trends
- Update harm mode probabilities
- Compare to predictions
- Revise model parameters
Annual analysis:
- Full calibration check
- Brier score for probability predictions
- Damage prediction accuracy
- Model revision if needed
Resources
Section titled “Resources”Incident Databases
Section titled “Incident Databases”Analysis Tools
Section titled “Analysis Tools”- AIID Explorer
- Spreadsheet templates for incident tracking
- SQL queries for incident analysis
Further Reading
Section titled “Further Reading”- “Categorizing Variants of AI Incidents” (McGregor et al.)
- “Learning from AI Failures” (Yampolskiy)
- “Incident Response for ML Systems” (Google)